diff --git "a/FineCaption/trainer_state.json" "b/FineCaption/trainer_state.json" new file mode 100644--- /dev/null +++ "b/FineCaption/trainer_state.json" @@ -0,0 +1,28679 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0, + "eval_steps": 500, + "global_step": 4091, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0002444390124663896, + "grad_norm": 20.966943740844727, + "learning_rate": 1.6260162601626018e-07, + "loss": 3.4835, + "step": 1 + }, + { + "epoch": 0.0004888780249327792, + "grad_norm": 18.250547409057617, + "learning_rate": 3.2520325203252037e-07, + "loss": 3.3028, + "step": 2 + }, + { + "epoch": 0.0007333170373991689, + "grad_norm": 18.719200134277344, + "learning_rate": 4.878048780487805e-07, + "loss": 3.249, + "step": 3 + }, + { + "epoch": 0.0009777560498655585, + "grad_norm": 17.085054397583008, + "learning_rate": 6.504065040650407e-07, + "loss": 3.4029, + "step": 4 + }, + { + "epoch": 0.0012221950623319481, + "grad_norm": 18.758522033691406, + "learning_rate": 8.130081300813009e-07, + "loss": 3.3331, + "step": 5 + }, + { + "epoch": 0.0014666340747983377, + "grad_norm": 15.967318534851074, + "learning_rate": 9.75609756097561e-07, + "loss": 3.176, + "step": 6 + }, + { + "epoch": 0.0017110730872647274, + "grad_norm": 16.959604263305664, + "learning_rate": 1.1382113821138213e-06, + "loss": 3.2587, + "step": 7 + }, + { + "epoch": 0.001955512099731117, + "grad_norm": 16.36026954650879, + "learning_rate": 1.3008130081300815e-06, + "loss": 3.1667, + "step": 8 + }, + { + "epoch": 0.002199951112197507, + "grad_norm": 15.372851371765137, + "learning_rate": 1.4634146341463414e-06, + "loss": 3.1927, + "step": 9 + }, + { + "epoch": 0.0024443901246638962, + "grad_norm": 14.929744720458984, + "learning_rate": 1.6260162601626018e-06, + "loss": 3.0885, + "step": 10 + }, + { + "epoch": 0.002688829137130286, + "grad_norm": 12.868640899658203, + "learning_rate": 1.788617886178862e-06, + "loss": 2.7862, + "step": 11 + }, + { + "epoch": 0.0029332681495966755, + "grad_norm": 11.283382415771484, + "learning_rate": 1.951219512195122e-06, + "loss": 2.5843, + "step": 12 + }, + { + "epoch": 0.0031777071620630653, + "grad_norm": 14.134821891784668, + "learning_rate": 2.1138211382113824e-06, + "loss": 2.6906, + "step": 13 + }, + { + "epoch": 0.0034221461745294547, + "grad_norm": 9.865968704223633, + "learning_rate": 2.2764227642276426e-06, + "loss": 2.6543, + "step": 14 + }, + { + "epoch": 0.0036665851869958446, + "grad_norm": 6.860537052154541, + "learning_rate": 2.4390243902439027e-06, + "loss": 2.4197, + "step": 15 + }, + { + "epoch": 0.003911024199462234, + "grad_norm": 6.645073890686035, + "learning_rate": 2.601626016260163e-06, + "loss": 2.2959, + "step": 16 + }, + { + "epoch": 0.004155463211928624, + "grad_norm": 26.320695877075195, + "learning_rate": 2.764227642276423e-06, + "loss": 2.1923, + "step": 17 + }, + { + "epoch": 0.004399902224395014, + "grad_norm": 8.401859283447266, + "learning_rate": 2.926829268292683e-06, + "loss": 2.4102, + "step": 18 + }, + { + "epoch": 0.0046443412368614035, + "grad_norm": 6.93703556060791, + "learning_rate": 3.0894308943089435e-06, + "loss": 2.2069, + "step": 19 + }, + { + "epoch": 0.0048887802493277925, + "grad_norm": 8.31473445892334, + "learning_rate": 3.2520325203252037e-06, + "loss": 2.0198, + "step": 20 + }, + { + "epoch": 0.005133219261794182, + "grad_norm": 5.992040634155273, + "learning_rate": 3.414634146341464e-06, + "loss": 2.1501, + "step": 21 + }, + { + "epoch": 0.005377658274260572, + "grad_norm": 5.812566757202148, + "learning_rate": 3.577235772357724e-06, + "loss": 2.2029, + "step": 22 + }, + { + "epoch": 0.005622097286726962, + "grad_norm": 6.239164352416992, + "learning_rate": 3.7398373983739838e-06, + "loss": 2.0681, + "step": 23 + }, + { + "epoch": 0.005866536299193351, + "grad_norm": 4.529574871063232, + "learning_rate": 3.902439024390244e-06, + "loss": 2.2424, + "step": 24 + }, + { + "epoch": 0.006110975311659741, + "grad_norm": 5.873737335205078, + "learning_rate": 4.0650406504065046e-06, + "loss": 1.997, + "step": 25 + }, + { + "epoch": 0.006355414324126131, + "grad_norm": 5.107316017150879, + "learning_rate": 4.227642276422765e-06, + "loss": 1.9068, + "step": 26 + }, + { + "epoch": 0.0065998533365925205, + "grad_norm": 5.841495513916016, + "learning_rate": 4.390243902439025e-06, + "loss": 1.9078, + "step": 27 + }, + { + "epoch": 0.0068442923490589095, + "grad_norm": 6.587307929992676, + "learning_rate": 4.552845528455285e-06, + "loss": 1.9157, + "step": 28 + }, + { + "epoch": 0.007088731361525299, + "grad_norm": 7.674979209899902, + "learning_rate": 4.715447154471545e-06, + "loss": 1.9183, + "step": 29 + }, + { + "epoch": 0.007333170373991689, + "grad_norm": 4.5966691970825195, + "learning_rate": 4.8780487804878055e-06, + "loss": 2.2402, + "step": 30 + }, + { + "epoch": 0.007577609386458079, + "grad_norm": 6.864888668060303, + "learning_rate": 5.040650406504065e-06, + "loss": 1.7592, + "step": 31 + }, + { + "epoch": 0.007822048398924468, + "grad_norm": 5.788065433502197, + "learning_rate": 5.203252032520326e-06, + "loss": 1.9031, + "step": 32 + }, + { + "epoch": 0.008066487411390858, + "grad_norm": 5.776228904724121, + "learning_rate": 5.365853658536586e-06, + "loss": 1.8899, + "step": 33 + }, + { + "epoch": 0.008310926423857248, + "grad_norm": 5.260758399963379, + "learning_rate": 5.528455284552846e-06, + "loss": 1.6202, + "step": 34 + }, + { + "epoch": 0.008555365436323637, + "grad_norm": 5.431663990020752, + "learning_rate": 5.691056910569106e-06, + "loss": 1.871, + "step": 35 + }, + { + "epoch": 0.008799804448790027, + "grad_norm": 5.790746688842773, + "learning_rate": 5.853658536585366e-06, + "loss": 1.6639, + "step": 36 + }, + { + "epoch": 0.009044243461256417, + "grad_norm": 5.473887920379639, + "learning_rate": 6.016260162601627e-06, + "loss": 1.8165, + "step": 37 + }, + { + "epoch": 0.009288682473722807, + "grad_norm": 5.433403968811035, + "learning_rate": 6.178861788617887e-06, + "loss": 1.6229, + "step": 38 + }, + { + "epoch": 0.009533121486189195, + "grad_norm": 11.307058334350586, + "learning_rate": 6.341463414634147e-06, + "loss": 1.8198, + "step": 39 + }, + { + "epoch": 0.009777560498655585, + "grad_norm": 5.7803568840026855, + "learning_rate": 6.504065040650407e-06, + "loss": 1.7236, + "step": 40 + }, + { + "epoch": 0.010021999511121975, + "grad_norm": 8.382798194885254, + "learning_rate": 6.666666666666667e-06, + "loss": 1.6812, + "step": 41 + }, + { + "epoch": 0.010266438523588365, + "grad_norm": 6.12075138092041, + "learning_rate": 6.829268292682928e-06, + "loss": 1.6711, + "step": 42 + }, + { + "epoch": 0.010510877536054754, + "grad_norm": 6.378519535064697, + "learning_rate": 6.991869918699188e-06, + "loss": 1.5586, + "step": 43 + }, + { + "epoch": 0.010755316548521144, + "grad_norm": 8.325606346130371, + "learning_rate": 7.154471544715448e-06, + "loss": 1.7467, + "step": 44 + }, + { + "epoch": 0.010999755560987534, + "grad_norm": 5.427236557006836, + "learning_rate": 7.317073170731707e-06, + "loss": 1.561, + "step": 45 + }, + { + "epoch": 0.011244194573453924, + "grad_norm": 5.533506870269775, + "learning_rate": 7.4796747967479676e-06, + "loss": 1.6548, + "step": 46 + }, + { + "epoch": 0.011488633585920312, + "grad_norm": 11.618112564086914, + "learning_rate": 7.64227642276423e-06, + "loss": 1.5789, + "step": 47 + }, + { + "epoch": 0.011733072598386702, + "grad_norm": 6.130812644958496, + "learning_rate": 7.804878048780489e-06, + "loss": 1.518, + "step": 48 + }, + { + "epoch": 0.011977511610853092, + "grad_norm": 6.173574447631836, + "learning_rate": 7.967479674796748e-06, + "loss": 1.7447, + "step": 49 + }, + { + "epoch": 0.012221950623319482, + "grad_norm": 5.371044158935547, + "learning_rate": 8.130081300813009e-06, + "loss": 1.7225, + "step": 50 + }, + { + "epoch": 0.012466389635785871, + "grad_norm": 6.262181282043457, + "learning_rate": 8.292682926829268e-06, + "loss": 1.5532, + "step": 51 + }, + { + "epoch": 0.012710828648252261, + "grad_norm": 6.0705671310424805, + "learning_rate": 8.45528455284553e-06, + "loss": 1.5654, + "step": 52 + }, + { + "epoch": 0.012955267660718651, + "grad_norm": 7.2067694664001465, + "learning_rate": 8.617886178861789e-06, + "loss": 1.4073, + "step": 53 + }, + { + "epoch": 0.013199706673185041, + "grad_norm": 5.209422588348389, + "learning_rate": 8.78048780487805e-06, + "loss": 1.6653, + "step": 54 + }, + { + "epoch": 0.01344414568565143, + "grad_norm": 14.538374900817871, + "learning_rate": 8.94308943089431e-06, + "loss": 1.547, + "step": 55 + }, + { + "epoch": 0.013688584698117819, + "grad_norm": 6.08756685256958, + "learning_rate": 9.10569105691057e-06, + "loss": 1.5763, + "step": 56 + }, + { + "epoch": 0.013933023710584209, + "grad_norm": 5.936624050140381, + "learning_rate": 9.268292682926831e-06, + "loss": 1.6264, + "step": 57 + }, + { + "epoch": 0.014177462723050599, + "grad_norm": 4.346157550811768, + "learning_rate": 9.43089430894309e-06, + "loss": 1.4903, + "step": 58 + }, + { + "epoch": 0.014421901735516988, + "grad_norm": 5.19631814956665, + "learning_rate": 9.59349593495935e-06, + "loss": 1.4802, + "step": 59 + }, + { + "epoch": 0.014666340747983378, + "grad_norm": 6.647718906402588, + "learning_rate": 9.756097560975611e-06, + "loss": 2.2965, + "step": 60 + }, + { + "epoch": 0.014910779760449768, + "grad_norm": 4.844165802001953, + "learning_rate": 9.91869918699187e-06, + "loss": 1.4015, + "step": 61 + }, + { + "epoch": 0.015155218772916158, + "grad_norm": 5.066519737243652, + "learning_rate": 1.008130081300813e-05, + "loss": 1.5471, + "step": 62 + }, + { + "epoch": 0.015399657785382548, + "grad_norm": 6.175212383270264, + "learning_rate": 1.024390243902439e-05, + "loss": 1.5756, + "step": 63 + }, + { + "epoch": 0.015644096797848936, + "grad_norm": 5.052133083343506, + "learning_rate": 1.0406504065040652e-05, + "loss": 1.4716, + "step": 64 + }, + { + "epoch": 0.015888535810315327, + "grad_norm": 6.0629353523254395, + "learning_rate": 1.0569105691056911e-05, + "loss": 1.668, + "step": 65 + }, + { + "epoch": 0.016132974822781716, + "grad_norm": 6.251193046569824, + "learning_rate": 1.0731707317073172e-05, + "loss": 1.4667, + "step": 66 + }, + { + "epoch": 0.016377413835248107, + "grad_norm": 7.878139495849609, + "learning_rate": 1.0894308943089431e-05, + "loss": 1.5591, + "step": 67 + }, + { + "epoch": 0.016621852847714495, + "grad_norm": 5.429769039154053, + "learning_rate": 1.1056910569105692e-05, + "loss": 1.4326, + "step": 68 + }, + { + "epoch": 0.016866291860180883, + "grad_norm": 6.907913684844971, + "learning_rate": 1.1219512195121953e-05, + "loss": 1.5594, + "step": 69 + }, + { + "epoch": 0.017110730872647275, + "grad_norm": 5.1678290367126465, + "learning_rate": 1.1382113821138213e-05, + "loss": 1.5511, + "step": 70 + }, + { + "epoch": 0.017355169885113663, + "grad_norm": 5.462667465209961, + "learning_rate": 1.1544715447154474e-05, + "loss": 1.3895, + "step": 71 + }, + { + "epoch": 0.017599608897580055, + "grad_norm": 6.608217239379883, + "learning_rate": 1.1707317073170731e-05, + "loss": 1.487, + "step": 72 + }, + { + "epoch": 0.017844047910046443, + "grad_norm": 5.475932598114014, + "learning_rate": 1.1869918699186992e-05, + "loss": 1.5315, + "step": 73 + }, + { + "epoch": 0.018088486922512834, + "grad_norm": 7.1279802322387695, + "learning_rate": 1.2032520325203254e-05, + "loss": 1.3711, + "step": 74 + }, + { + "epoch": 0.018332925934979222, + "grad_norm": 6.620932579040527, + "learning_rate": 1.2195121951219513e-05, + "loss": 1.4346, + "step": 75 + }, + { + "epoch": 0.018577364947445614, + "grad_norm": 6.356578826904297, + "learning_rate": 1.2357723577235774e-05, + "loss": 1.4521, + "step": 76 + }, + { + "epoch": 0.018821803959912002, + "grad_norm": 5.530941486358643, + "learning_rate": 1.2520325203252033e-05, + "loss": 1.472, + "step": 77 + }, + { + "epoch": 0.01906624297237839, + "grad_norm": 7.7910261154174805, + "learning_rate": 1.2682926829268294e-05, + "loss": 1.4296, + "step": 78 + }, + { + "epoch": 0.019310681984844782, + "grad_norm": 7.19402551651001, + "learning_rate": 1.2845528455284555e-05, + "loss": 1.4176, + "step": 79 + }, + { + "epoch": 0.01955512099731117, + "grad_norm": 6.1136040687561035, + "learning_rate": 1.3008130081300815e-05, + "loss": 1.4165, + "step": 80 + }, + { + "epoch": 0.01979956000977756, + "grad_norm": 5.990948677062988, + "learning_rate": 1.3170731707317076e-05, + "loss": 1.4218, + "step": 81 + }, + { + "epoch": 0.02004399902224395, + "grad_norm": 8.119118690490723, + "learning_rate": 1.3333333333333333e-05, + "loss": 1.4373, + "step": 82 + }, + { + "epoch": 0.02028843803471034, + "grad_norm": 6.3870954513549805, + "learning_rate": 1.3495934959349594e-05, + "loss": 1.3869, + "step": 83 + }, + { + "epoch": 0.02053287704717673, + "grad_norm": 8.236477851867676, + "learning_rate": 1.3658536585365855e-05, + "loss": 1.4277, + "step": 84 + }, + { + "epoch": 0.020777316059643117, + "grad_norm": 17.77801513671875, + "learning_rate": 1.3821138211382115e-05, + "loss": 1.308, + "step": 85 + }, + { + "epoch": 0.02102175507210951, + "grad_norm": 6.244637489318848, + "learning_rate": 1.3983739837398376e-05, + "loss": 1.3706, + "step": 86 + }, + { + "epoch": 0.021266194084575897, + "grad_norm": 6.249431610107422, + "learning_rate": 1.4146341463414635e-05, + "loss": 1.3358, + "step": 87 + }, + { + "epoch": 0.02151063309704229, + "grad_norm": 4.701333999633789, + "learning_rate": 1.4308943089430896e-05, + "loss": 2.2497, + "step": 88 + }, + { + "epoch": 0.021755072109508677, + "grad_norm": 5.575697898864746, + "learning_rate": 1.4471544715447157e-05, + "loss": 1.4671, + "step": 89 + }, + { + "epoch": 0.02199951112197507, + "grad_norm": 5.8239922523498535, + "learning_rate": 1.4634146341463415e-05, + "loss": 1.4833, + "step": 90 + }, + { + "epoch": 0.022243950134441456, + "grad_norm": 6.4747185707092285, + "learning_rate": 1.4796747967479676e-05, + "loss": 1.5262, + "step": 91 + }, + { + "epoch": 0.022488389146907848, + "grad_norm": 6.724368095397949, + "learning_rate": 1.4959349593495935e-05, + "loss": 1.422, + "step": 92 + }, + { + "epoch": 0.022732828159374236, + "grad_norm": 5.154726982116699, + "learning_rate": 1.5121951219512196e-05, + "loss": 1.3532, + "step": 93 + }, + { + "epoch": 0.022977267171840624, + "grad_norm": 4.542200565338135, + "learning_rate": 1.528455284552846e-05, + "loss": 2.1428, + "step": 94 + }, + { + "epoch": 0.023221706184307016, + "grad_norm": 5.880817890167236, + "learning_rate": 1.5447154471544717e-05, + "loss": 1.5996, + "step": 95 + }, + { + "epoch": 0.023466145196773404, + "grad_norm": 6.723649501800537, + "learning_rate": 1.5609756097560978e-05, + "loss": 1.5054, + "step": 96 + }, + { + "epoch": 0.023710584209239795, + "grad_norm": 5.2248382568359375, + "learning_rate": 1.5772357723577235e-05, + "loss": 1.3787, + "step": 97 + }, + { + "epoch": 0.023955023221706184, + "grad_norm": 5.563326835632324, + "learning_rate": 1.5934959349593496e-05, + "loss": 1.3925, + "step": 98 + }, + { + "epoch": 0.024199462234172575, + "grad_norm": 6.171076774597168, + "learning_rate": 1.6097560975609757e-05, + "loss": 1.502, + "step": 99 + }, + { + "epoch": 0.024443901246638963, + "grad_norm": 11.536863327026367, + "learning_rate": 1.6260162601626018e-05, + "loss": 1.4156, + "step": 100 + }, + { + "epoch": 0.024688340259105355, + "grad_norm": 11.269830703735352, + "learning_rate": 1.642276422764228e-05, + "loss": 1.3481, + "step": 101 + }, + { + "epoch": 0.024932779271571743, + "grad_norm": 7.356262683868408, + "learning_rate": 1.6585365853658537e-05, + "loss": 1.4403, + "step": 102 + }, + { + "epoch": 0.02517721828403813, + "grad_norm": 10.883296012878418, + "learning_rate": 1.6747967479674798e-05, + "loss": 2.2085, + "step": 103 + }, + { + "epoch": 0.025421657296504523, + "grad_norm": 6.563252925872803, + "learning_rate": 1.691056910569106e-05, + "loss": 1.4681, + "step": 104 + }, + { + "epoch": 0.02566609630897091, + "grad_norm": 7.629971027374268, + "learning_rate": 1.7073170731707317e-05, + "loss": 1.4373, + "step": 105 + }, + { + "epoch": 0.025910535321437302, + "grad_norm": 12.355507850646973, + "learning_rate": 1.7235772357723578e-05, + "loss": 1.4394, + "step": 106 + }, + { + "epoch": 0.02615497433390369, + "grad_norm": 6.847367286682129, + "learning_rate": 1.739837398373984e-05, + "loss": 1.4561, + "step": 107 + }, + { + "epoch": 0.026399413346370082, + "grad_norm": 7.732123374938965, + "learning_rate": 1.75609756097561e-05, + "loss": 1.4949, + "step": 108 + }, + { + "epoch": 0.02664385235883647, + "grad_norm": 7.845799922943115, + "learning_rate": 1.772357723577236e-05, + "loss": 1.5389, + "step": 109 + }, + { + "epoch": 0.02688829137130286, + "grad_norm": 5.964730739593506, + "learning_rate": 1.788617886178862e-05, + "loss": 1.4011, + "step": 110 + }, + { + "epoch": 0.02713273038376925, + "grad_norm": 6.563551425933838, + "learning_rate": 1.804878048780488e-05, + "loss": 1.3924, + "step": 111 + }, + { + "epoch": 0.027377169396235638, + "grad_norm": 7.457088947296143, + "learning_rate": 1.821138211382114e-05, + "loss": 1.5348, + "step": 112 + }, + { + "epoch": 0.02762160840870203, + "grad_norm": 5.761719703674316, + "learning_rate": 1.83739837398374e-05, + "loss": 1.3819, + "step": 113 + }, + { + "epoch": 0.027866047421168418, + "grad_norm": 10.146748542785645, + "learning_rate": 1.8536585365853663e-05, + "loss": 1.417, + "step": 114 + }, + { + "epoch": 0.02811048643363481, + "grad_norm": 5.558394432067871, + "learning_rate": 1.869918699186992e-05, + "loss": 1.4842, + "step": 115 + }, + { + "epoch": 0.028354925446101197, + "grad_norm": 8.84795093536377, + "learning_rate": 1.886178861788618e-05, + "loss": 1.3574, + "step": 116 + }, + { + "epoch": 0.02859936445856759, + "grad_norm": 5.490208148956299, + "learning_rate": 1.902439024390244e-05, + "loss": 1.357, + "step": 117 + }, + { + "epoch": 0.028843803471033977, + "grad_norm": 7.372227191925049, + "learning_rate": 1.91869918699187e-05, + "loss": 1.3379, + "step": 118 + }, + { + "epoch": 0.029088242483500365, + "grad_norm": 11.52920913696289, + "learning_rate": 1.934959349593496e-05, + "loss": 1.4243, + "step": 119 + }, + { + "epoch": 0.029332681495966757, + "grad_norm": 6.755467414855957, + "learning_rate": 1.9512195121951222e-05, + "loss": 1.5187, + "step": 120 + }, + { + "epoch": 0.029577120508433145, + "grad_norm": 5.93511438369751, + "learning_rate": 1.9674796747967483e-05, + "loss": 1.3194, + "step": 121 + }, + { + "epoch": 0.029821559520899536, + "grad_norm": 4.53206729888916, + "learning_rate": 1.983739837398374e-05, + "loss": 1.3059, + "step": 122 + }, + { + "epoch": 0.030065998533365924, + "grad_norm": 6.463418483734131, + "learning_rate": 2e-05, + "loss": 1.2321, + "step": 123 + }, + { + "epoch": 0.030310437545832316, + "grad_norm": 5.359195232391357, + "learning_rate": 1.9999996865802212e-05, + "loss": 1.3775, + "step": 124 + }, + { + "epoch": 0.030554876558298704, + "grad_norm": 7.653247356414795, + "learning_rate": 1.9999987463210805e-05, + "loss": 1.4078, + "step": 125 + }, + { + "epoch": 0.030799315570765096, + "grad_norm": 5.883333683013916, + "learning_rate": 1.9999971792231676e-05, + "loss": 1.4596, + "step": 126 + }, + { + "epoch": 0.031043754583231484, + "grad_norm": 6.219335079193115, + "learning_rate": 1.999994985287465e-05, + "loss": 1.3468, + "step": 127 + }, + { + "epoch": 0.03128819359569787, + "grad_norm": 9.816319465637207, + "learning_rate": 1.999992164515348e-05, + "loss": 1.2856, + "step": 128 + }, + { + "epoch": 0.03153263260816426, + "grad_norm": 7.612930774688721, + "learning_rate": 1.999988716908584e-05, + "loss": 1.3644, + "step": 129 + }, + { + "epoch": 0.031777071620630655, + "grad_norm": 7.602503299713135, + "learning_rate": 1.999984642469335e-05, + "loss": 1.4115, + "step": 130 + }, + { + "epoch": 0.03202151063309704, + "grad_norm": 5.727510452270508, + "learning_rate": 1.9999799412001547e-05, + "loss": 1.228, + "step": 131 + }, + { + "epoch": 0.03226594964556343, + "grad_norm": 5.721832752227783, + "learning_rate": 1.99997461310399e-05, + "loss": 1.335, + "step": 132 + }, + { + "epoch": 0.03251038865802982, + "grad_norm": 6.187972068786621, + "learning_rate": 1.9999686581841805e-05, + "loss": 1.339, + "step": 133 + }, + { + "epoch": 0.032754827670496214, + "grad_norm": 5.254082202911377, + "learning_rate": 1.9999620764444594e-05, + "loss": 1.2741, + "step": 134 + }, + { + "epoch": 0.0329992666829626, + "grad_norm": 6.423061370849609, + "learning_rate": 1.9999548678889523e-05, + "loss": 1.3004, + "step": 135 + }, + { + "epoch": 0.03324370569542899, + "grad_norm": 4.9685959815979, + "learning_rate": 1.9999470325221778e-05, + "loss": 1.3251, + "step": 136 + }, + { + "epoch": 0.03348814470789538, + "grad_norm": 5.46399450302124, + "learning_rate": 1.999938570349047e-05, + "loss": 1.3795, + "step": 137 + }, + { + "epoch": 0.03373258372036177, + "grad_norm": 4.732756614685059, + "learning_rate": 1.999929481374865e-05, + "loss": 1.2958, + "step": 138 + }, + { + "epoch": 0.03397702273282816, + "grad_norm": 4.804397106170654, + "learning_rate": 1.9999197656053288e-05, + "loss": 1.2627, + "step": 139 + }, + { + "epoch": 0.03422146174529455, + "grad_norm": 4.775611400604248, + "learning_rate": 1.9999094230465285e-05, + "loss": 1.4391, + "step": 140 + }, + { + "epoch": 0.03446590075776094, + "grad_norm": 4.6918721199035645, + "learning_rate": 1.9998984537049476e-05, + "loss": 1.4239, + "step": 141 + }, + { + "epoch": 0.034710339770227326, + "grad_norm": 4.3264899253845215, + "learning_rate": 1.9998868575874616e-05, + "loss": 1.1658, + "step": 142 + }, + { + "epoch": 0.03495477878269372, + "grad_norm": 4.351634502410889, + "learning_rate": 1.99987463470134e-05, + "loss": 1.3883, + "step": 143 + }, + { + "epoch": 0.03519921779516011, + "grad_norm": 5.136422634124756, + "learning_rate": 1.999861785054244e-05, + "loss": 1.438, + "step": 144 + }, + { + "epoch": 0.0354436568076265, + "grad_norm": 5.345577716827393, + "learning_rate": 1.999848308654229e-05, + "loss": 1.4144, + "step": 145 + }, + { + "epoch": 0.035688095820092886, + "grad_norm": 5.832770824432373, + "learning_rate": 1.9998342055097416e-05, + "loss": 1.299, + "step": 146 + }, + { + "epoch": 0.035932534832559274, + "grad_norm": 7.734096527099609, + "learning_rate": 1.999819475629623e-05, + "loss": 1.4003, + "step": 147 + }, + { + "epoch": 0.03617697384502567, + "grad_norm": 5.463601589202881, + "learning_rate": 1.9998041190231063e-05, + "loss": 1.5057, + "step": 148 + }, + { + "epoch": 0.03642141285749206, + "grad_norm": 8.983975410461426, + "learning_rate": 1.9997881356998172e-05, + "loss": 1.3619, + "step": 149 + }, + { + "epoch": 0.036665851869958445, + "grad_norm": 5.394047737121582, + "learning_rate": 1.999771525669775e-05, + "loss": 1.3032, + "step": 150 + }, + { + "epoch": 0.03691029088242483, + "grad_norm": 4.557994842529297, + "learning_rate": 1.9997542889433917e-05, + "loss": 1.2587, + "step": 151 + }, + { + "epoch": 0.03715472989489123, + "grad_norm": 4.662408351898193, + "learning_rate": 1.9997364255314715e-05, + "loss": 1.3185, + "step": 152 + }, + { + "epoch": 0.037399168907357616, + "grad_norm": 7.458238124847412, + "learning_rate": 1.9997179354452126e-05, + "loss": 1.2708, + "step": 153 + }, + { + "epoch": 0.037643607919824004, + "grad_norm": 4.436248302459717, + "learning_rate": 1.9996988186962044e-05, + "loss": 1.3901, + "step": 154 + }, + { + "epoch": 0.03788804693229039, + "grad_norm": 5.528069972991943, + "learning_rate": 1.9996790752964305e-05, + "loss": 1.4166, + "step": 155 + }, + { + "epoch": 0.03813248594475678, + "grad_norm": 4.3356733322143555, + "learning_rate": 1.9996587052582672e-05, + "loss": 1.152, + "step": 156 + }, + { + "epoch": 0.038376924957223175, + "grad_norm": 5.925254821777344, + "learning_rate": 1.9996377085944826e-05, + "loss": 1.5081, + "step": 157 + }, + { + "epoch": 0.038621363969689564, + "grad_norm": 4.844107151031494, + "learning_rate": 1.999616085318239e-05, + "loss": 1.3824, + "step": 158 + }, + { + "epoch": 0.03886580298215595, + "grad_norm": 6.188292026519775, + "learning_rate": 1.9995938354430898e-05, + "loss": 1.3108, + "step": 159 + }, + { + "epoch": 0.03911024199462234, + "grad_norm": 5.222842216491699, + "learning_rate": 1.9995709589829828e-05, + "loss": 1.433, + "step": 160 + }, + { + "epoch": 0.03935468100708873, + "grad_norm": 6.492764949798584, + "learning_rate": 1.9995474559522576e-05, + "loss": 1.3901, + "step": 161 + }, + { + "epoch": 0.03959912001955512, + "grad_norm": 6.479674339294434, + "learning_rate": 1.999523326365647e-05, + "loss": 1.3002, + "step": 162 + }, + { + "epoch": 0.03984355903202151, + "grad_norm": 4.043074607849121, + "learning_rate": 1.999498570238276e-05, + "loss": 2.238, + "step": 163 + }, + { + "epoch": 0.0400879980444879, + "grad_norm": 8.806992530822754, + "learning_rate": 1.999473187585663e-05, + "loss": 1.3225, + "step": 164 + }, + { + "epoch": 0.04033243705695429, + "grad_norm": 5.6869964599609375, + "learning_rate": 1.9994471784237188e-05, + "loss": 1.3458, + "step": 165 + }, + { + "epoch": 0.04057687606942068, + "grad_norm": 7.028317451477051, + "learning_rate": 1.9994205427687473e-05, + "loss": 1.2942, + "step": 166 + }, + { + "epoch": 0.04082131508188707, + "grad_norm": 5.3759684562683105, + "learning_rate": 1.999393280637444e-05, + "loss": 1.3008, + "step": 167 + }, + { + "epoch": 0.04106575409435346, + "grad_norm": 4.7559943199157715, + "learning_rate": 1.999365392046899e-05, + "loss": 1.3133, + "step": 168 + }, + { + "epoch": 0.04131019310681985, + "grad_norm": 4.474522590637207, + "learning_rate": 1.9993368770145927e-05, + "loss": 1.3373, + "step": 169 + }, + { + "epoch": 0.041554632119286235, + "grad_norm": 6.905185699462891, + "learning_rate": 1.9993077355584e-05, + "loss": 1.4465, + "step": 170 + }, + { + "epoch": 0.04179907113175263, + "grad_norm": 5.875603675842285, + "learning_rate": 1.9992779676965884e-05, + "loss": 1.2736, + "step": 171 + }, + { + "epoch": 0.04204351014421902, + "grad_norm": 3.587334156036377, + "learning_rate": 1.999247573447817e-05, + "loss": 2.1556, + "step": 172 + }, + { + "epoch": 0.042287949156685406, + "grad_norm": 6.450631141662598, + "learning_rate": 1.9992165528311385e-05, + "loss": 1.33, + "step": 173 + }, + { + "epoch": 0.042532388169151794, + "grad_norm": 6.471398830413818, + "learning_rate": 1.9991849058659977e-05, + "loss": 1.2372, + "step": 174 + }, + { + "epoch": 0.04277682718161819, + "grad_norm": 6.2006425857543945, + "learning_rate": 1.9991526325722316e-05, + "loss": 1.2998, + "step": 175 + }, + { + "epoch": 0.04302126619408458, + "grad_norm": 4.926456928253174, + "learning_rate": 1.999119732970071e-05, + "loss": 1.3577, + "step": 176 + }, + { + "epoch": 0.043265705206550965, + "grad_norm": 5.1743292808532715, + "learning_rate": 1.9990862070801388e-05, + "loss": 1.3454, + "step": 177 + }, + { + "epoch": 0.04351014421901735, + "grad_norm": 5.5031890869140625, + "learning_rate": 1.9990520549234502e-05, + "loss": 1.3274, + "step": 178 + }, + { + "epoch": 0.04375458323148374, + "grad_norm": 16.7782039642334, + "learning_rate": 1.999017276521413e-05, + "loss": 1.3566, + "step": 179 + }, + { + "epoch": 0.04399902224395014, + "grad_norm": 5.242146015167236, + "learning_rate": 1.9989818718958275e-05, + "loss": 1.3043, + "step": 180 + }, + { + "epoch": 0.044243461256416525, + "grad_norm": 5.954908847808838, + "learning_rate": 1.9989458410688865e-05, + "loss": 1.2932, + "step": 181 + }, + { + "epoch": 0.04448790026888291, + "grad_norm": 4.562159061431885, + "learning_rate": 1.9989091840631767e-05, + "loss": 1.269, + "step": 182 + }, + { + "epoch": 0.0447323392813493, + "grad_norm": 5.205554485321045, + "learning_rate": 1.9988719009016753e-05, + "loss": 1.2356, + "step": 183 + }, + { + "epoch": 0.044976778293815696, + "grad_norm": 6.116245746612549, + "learning_rate": 1.998833991607753e-05, + "loss": 1.3568, + "step": 184 + }, + { + "epoch": 0.045221217306282084, + "grad_norm": 3.484712600708008, + "learning_rate": 1.9987954562051724e-05, + "loss": 2.1914, + "step": 185 + }, + { + "epoch": 0.04546565631874847, + "grad_norm": 4.402046203613281, + "learning_rate": 1.99875629471809e-05, + "loss": 1.33, + "step": 186 + }, + { + "epoch": 0.04571009533121486, + "grad_norm": 6.1173415184021, + "learning_rate": 1.998716507171053e-05, + "loss": 1.3433, + "step": 187 + }, + { + "epoch": 0.04595453434368125, + "grad_norm": 4.56729793548584, + "learning_rate": 1.998676093589002e-05, + "loss": 1.2448, + "step": 188 + }, + { + "epoch": 0.04619897335614764, + "grad_norm": 2.8044798374176025, + "learning_rate": 1.99863505399727e-05, + "loss": 2.1563, + "step": 189 + }, + { + "epoch": 0.04644341236861403, + "grad_norm": 5.482027530670166, + "learning_rate": 1.998593388421582e-05, + "loss": 1.3307, + "step": 190 + }, + { + "epoch": 0.04668785138108042, + "grad_norm": 6.116431713104248, + "learning_rate": 1.9985510968880555e-05, + "loss": 1.3009, + "step": 191 + }, + { + "epoch": 0.04693229039354681, + "grad_norm": 5.399773120880127, + "learning_rate": 1.9985081794232014e-05, + "loss": 1.3656, + "step": 192 + }, + { + "epoch": 0.0471767294060132, + "grad_norm": 4.578831195831299, + "learning_rate": 1.998464636053921e-05, + "loss": 1.3241, + "step": 193 + }, + { + "epoch": 0.04742116841847959, + "grad_norm": 5.7253618240356445, + "learning_rate": 1.9984204668075094e-05, + "loss": 1.1836, + "step": 194 + }, + { + "epoch": 0.04766560743094598, + "grad_norm": 5.640674114227295, + "learning_rate": 1.9983756717116534e-05, + "loss": 1.3452, + "step": 195 + }, + { + "epoch": 0.04791004644341237, + "grad_norm": 5.7181291580200195, + "learning_rate": 1.998330250794433e-05, + "loss": 1.2925, + "step": 196 + }, + { + "epoch": 0.048154485455878755, + "grad_norm": 5.187880039215088, + "learning_rate": 1.9982842040843193e-05, + "loss": 1.3884, + "step": 197 + }, + { + "epoch": 0.04839892446834515, + "grad_norm": 5.154764175415039, + "learning_rate": 1.9982375316101764e-05, + "loss": 1.3073, + "step": 198 + }, + { + "epoch": 0.04864336348081154, + "grad_norm": 3.99664568901062, + "learning_rate": 1.99819023340126e-05, + "loss": 2.2408, + "step": 199 + }, + { + "epoch": 0.048887802493277926, + "grad_norm": 4.218594074249268, + "learning_rate": 1.998142309487219e-05, + "loss": 1.319, + "step": 200 + }, + { + "epoch": 0.049132241505744315, + "grad_norm": 4.567066669464111, + "learning_rate": 1.9980937598980943e-05, + "loss": 1.2632, + "step": 201 + }, + { + "epoch": 0.04937668051821071, + "grad_norm": 4.567723751068115, + "learning_rate": 1.9980445846643176e-05, + "loss": 1.1766, + "step": 202 + }, + { + "epoch": 0.0496211195306771, + "grad_norm": 4.89946985244751, + "learning_rate": 1.9979947838167152e-05, + "loss": 1.3318, + "step": 203 + }, + { + "epoch": 0.049865558543143486, + "grad_norm": 7.729551792144775, + "learning_rate": 1.9979443573865033e-05, + "loss": 1.3179, + "step": 204 + }, + { + "epoch": 0.050109997555609874, + "grad_norm": 7.349613189697266, + "learning_rate": 1.9978933054052913e-05, + "loss": 1.2955, + "step": 205 + }, + { + "epoch": 0.05035443656807626, + "grad_norm": 4.26453161239624, + "learning_rate": 1.997841627905081e-05, + "loss": 1.3304, + "step": 206 + }, + { + "epoch": 0.05059887558054266, + "grad_norm": 7.018398284912109, + "learning_rate": 1.9977893249182654e-05, + "loss": 1.3471, + "step": 207 + }, + { + "epoch": 0.050843314593009045, + "grad_norm": 6.137728691101074, + "learning_rate": 1.9977363964776305e-05, + "loss": 1.3536, + "step": 208 + }, + { + "epoch": 0.05108775360547543, + "grad_norm": 5.061336994171143, + "learning_rate": 1.9976828426163538e-05, + "loss": 1.3652, + "step": 209 + }, + { + "epoch": 0.05133219261794182, + "grad_norm": 5.005093574523926, + "learning_rate": 1.9976286633680048e-05, + "loss": 1.3201, + "step": 210 + }, + { + "epoch": 0.051576631630408216, + "grad_norm": 3.8842194080352783, + "learning_rate": 1.9975738587665455e-05, + "loss": 1.3596, + "step": 211 + }, + { + "epoch": 0.051821070642874605, + "grad_norm": 5.012754440307617, + "learning_rate": 1.9975184288463295e-05, + "loss": 1.2887, + "step": 212 + }, + { + "epoch": 0.05206550965534099, + "grad_norm": 4.832705497741699, + "learning_rate": 1.9974623736421025e-05, + "loss": 1.3391, + "step": 213 + }, + { + "epoch": 0.05230994866780738, + "grad_norm": 5.078959941864014, + "learning_rate": 1.9974056931890018e-05, + "loss": 1.2961, + "step": 214 + }, + { + "epoch": 0.05255438768027377, + "grad_norm": 4.353578090667725, + "learning_rate": 1.9973483875225573e-05, + "loss": 1.2887, + "step": 215 + }, + { + "epoch": 0.052798826692740164, + "grad_norm": 7.96591329574585, + "learning_rate": 1.9972904566786903e-05, + "loss": 1.3251, + "step": 216 + }, + { + "epoch": 0.05304326570520655, + "grad_norm": 5.235666275024414, + "learning_rate": 1.9972319006937143e-05, + "loss": 1.2725, + "step": 217 + }, + { + "epoch": 0.05328770471767294, + "grad_norm": 4.645522594451904, + "learning_rate": 1.9971727196043344e-05, + "loss": 1.3218, + "step": 218 + }, + { + "epoch": 0.05353214373013933, + "grad_norm": 7.124841690063477, + "learning_rate": 1.9971129134476474e-05, + "loss": 1.3267, + "step": 219 + }, + { + "epoch": 0.05377658274260572, + "grad_norm": 6.27811861038208, + "learning_rate": 1.997052482261143e-05, + "loss": 1.2694, + "step": 220 + }, + { + "epoch": 0.05402102175507211, + "grad_norm": 6.613264560699463, + "learning_rate": 1.996991426082701e-05, + "loss": 1.2389, + "step": 221 + }, + { + "epoch": 0.0542654607675385, + "grad_norm": 5.72642183303833, + "learning_rate": 1.996929744950594e-05, + "loss": 1.2542, + "step": 222 + }, + { + "epoch": 0.05450989978000489, + "grad_norm": 6.274509906768799, + "learning_rate": 1.996867438903486e-05, + "loss": 1.3864, + "step": 223 + }, + { + "epoch": 0.054754338792471276, + "grad_norm": 4.827954292297363, + "learning_rate": 1.9968045079804333e-05, + "loss": 1.311, + "step": 224 + }, + { + "epoch": 0.05499877780493767, + "grad_norm": 4.8781280517578125, + "learning_rate": 1.9967409522208835e-05, + "loss": 1.2287, + "step": 225 + }, + { + "epoch": 0.05524321681740406, + "grad_norm": 4.974234104156494, + "learning_rate": 1.996676771664676e-05, + "loss": 1.3661, + "step": 226 + }, + { + "epoch": 0.05548765582987045, + "grad_norm": 4.70881986618042, + "learning_rate": 1.996611966352041e-05, + "loss": 1.2792, + "step": 227 + }, + { + "epoch": 0.055732094842336835, + "grad_norm": 4.771082401275635, + "learning_rate": 1.996546536323602e-05, + "loss": 1.3459, + "step": 228 + }, + { + "epoch": 0.05597653385480322, + "grad_norm": 4.687952995300293, + "learning_rate": 1.9964804816203723e-05, + "loss": 1.2545, + "step": 229 + }, + { + "epoch": 0.05622097286726962, + "grad_norm": 3.7861790657043457, + "learning_rate": 1.9964138022837575e-05, + "loss": 2.1981, + "step": 230 + }, + { + "epoch": 0.056465411879736006, + "grad_norm": 6.665339946746826, + "learning_rate": 1.9963464983555557e-05, + "loss": 1.4111, + "step": 231 + }, + { + "epoch": 0.056709850892202394, + "grad_norm": 5.215484142303467, + "learning_rate": 1.996278569877955e-05, + "loss": 1.3479, + "step": 232 + }, + { + "epoch": 0.05695428990466878, + "grad_norm": 3.6600303649902344, + "learning_rate": 1.996210016893536e-05, + "loss": 1.3194, + "step": 233 + }, + { + "epoch": 0.05719872891713518, + "grad_norm": 6.554864883422852, + "learning_rate": 1.9961408394452703e-05, + "loss": 1.2998, + "step": 234 + }, + { + "epoch": 0.057443167929601566, + "grad_norm": 6.014383316040039, + "learning_rate": 1.9960710375765212e-05, + "loss": 1.2345, + "step": 235 + }, + { + "epoch": 0.057687606942067954, + "grad_norm": 6.356429576873779, + "learning_rate": 1.9960006113310428e-05, + "loss": 1.2493, + "step": 236 + }, + { + "epoch": 0.05793204595453434, + "grad_norm": 10.219972610473633, + "learning_rate": 1.9959295607529818e-05, + "loss": 1.2908, + "step": 237 + }, + { + "epoch": 0.05817648496700073, + "grad_norm": 4.4042558670043945, + "learning_rate": 1.9958578858868747e-05, + "loss": 1.3068, + "step": 238 + }, + { + "epoch": 0.058420923979467125, + "grad_norm": 4.574854373931885, + "learning_rate": 1.9957855867776507e-05, + "loss": 1.3013, + "step": 239 + }, + { + "epoch": 0.05866536299193351, + "grad_norm": 4.1353960037231445, + "learning_rate": 1.9957126634706294e-05, + "loss": 1.2734, + "step": 240 + }, + { + "epoch": 0.0589098020043999, + "grad_norm": 4.964991092681885, + "learning_rate": 1.9956391160115224e-05, + "loss": 1.1964, + "step": 241 + }, + { + "epoch": 0.05915424101686629, + "grad_norm": 6.201467514038086, + "learning_rate": 1.995564944446432e-05, + "loss": 1.2947, + "step": 242 + }, + { + "epoch": 0.059398680029332684, + "grad_norm": 5.684882164001465, + "learning_rate": 1.9954901488218515e-05, + "loss": 1.3692, + "step": 243 + }, + { + "epoch": 0.05964311904179907, + "grad_norm": 4.244503974914551, + "learning_rate": 1.9954147291846663e-05, + "loss": 2.2702, + "step": 244 + }, + { + "epoch": 0.05988755805426546, + "grad_norm": 6.967065334320068, + "learning_rate": 1.9953386855821523e-05, + "loss": 1.2285, + "step": 245 + }, + { + "epoch": 0.06013199706673185, + "grad_norm": 5.459507942199707, + "learning_rate": 1.9952620180619765e-05, + "loss": 1.266, + "step": 246 + }, + { + "epoch": 0.06037643607919824, + "grad_norm": 7.973972320556641, + "learning_rate": 1.995184726672197e-05, + "loss": 1.3239, + "step": 247 + }, + { + "epoch": 0.06062087509166463, + "grad_norm": 4.386531829833984, + "learning_rate": 1.9951068114612637e-05, + "loss": 1.3246, + "step": 248 + }, + { + "epoch": 0.06086531410413102, + "grad_norm": 4.626567840576172, + "learning_rate": 1.995028272478016e-05, + "loss": 1.3018, + "step": 249 + }, + { + "epoch": 0.06110975311659741, + "grad_norm": 4.793629169464111, + "learning_rate": 1.9949491097716856e-05, + "loss": 1.3344, + "step": 250 + }, + { + "epoch": 0.061354192129063796, + "grad_norm": 6.8952412605285645, + "learning_rate": 1.994869323391895e-05, + "loss": 1.3457, + "step": 251 + }, + { + "epoch": 0.06159863114153019, + "grad_norm": 6.089010238647461, + "learning_rate": 1.994788913388658e-05, + "loss": 1.3226, + "step": 252 + }, + { + "epoch": 0.06184307015399658, + "grad_norm": 6.344317436218262, + "learning_rate": 1.9947078798123778e-05, + "loss": 1.2385, + "step": 253 + }, + { + "epoch": 0.06208750916646297, + "grad_norm": 12.935145378112793, + "learning_rate": 1.99462622271385e-05, + "loss": 1.3256, + "step": 254 + }, + { + "epoch": 0.062331948178929356, + "grad_norm": 4.718111991882324, + "learning_rate": 1.9945439421442602e-05, + "loss": 1.2985, + "step": 255 + }, + { + "epoch": 0.06257638719139574, + "grad_norm": 5.535119533538818, + "learning_rate": 1.9944610381551852e-05, + "loss": 1.2133, + "step": 256 + }, + { + "epoch": 0.06282082620386213, + "grad_norm": 4.996159076690674, + "learning_rate": 1.9943775107985925e-05, + "loss": 1.3943, + "step": 257 + }, + { + "epoch": 0.06306526521632852, + "grad_norm": 5.818681240081787, + "learning_rate": 1.9942933601268403e-05, + "loss": 1.2941, + "step": 258 + }, + { + "epoch": 0.06330970422879492, + "grad_norm": 5.313794136047363, + "learning_rate": 1.994208586192678e-05, + "loss": 1.225, + "step": 259 + }, + { + "epoch": 0.06355414324126131, + "grad_norm": 5.420829772949219, + "learning_rate": 1.994123189049245e-05, + "loss": 1.4065, + "step": 260 + }, + { + "epoch": 0.0637985822537277, + "grad_norm": 7.252939224243164, + "learning_rate": 1.9940371687500713e-05, + "loss": 1.2559, + "step": 261 + }, + { + "epoch": 0.06404302126619409, + "grad_norm": 7.47566556930542, + "learning_rate": 1.993950525349078e-05, + "loss": 1.2258, + "step": 262 + }, + { + "epoch": 0.06428746027866047, + "grad_norm": 6.010402679443359, + "learning_rate": 1.993863258900577e-05, + "loss": 1.2764, + "step": 263 + }, + { + "epoch": 0.06453189929112686, + "grad_norm": 5.5107808113098145, + "learning_rate": 1.9937753694592698e-05, + "loss": 1.3388, + "step": 264 + }, + { + "epoch": 0.06477633830359325, + "grad_norm": 4.4688334465026855, + "learning_rate": 1.993686857080249e-05, + "loss": 1.2992, + "step": 265 + }, + { + "epoch": 0.06502077731605964, + "grad_norm": 4.5571818351745605, + "learning_rate": 1.9935977218189982e-05, + "loss": 1.127, + "step": 266 + }, + { + "epoch": 0.06526521632852603, + "grad_norm": 4.901233673095703, + "learning_rate": 1.9935079637313906e-05, + "loss": 1.213, + "step": 267 + }, + { + "epoch": 0.06550965534099243, + "grad_norm": 6.362276077270508, + "learning_rate": 1.9934175828736898e-05, + "loss": 1.3096, + "step": 268 + }, + { + "epoch": 0.06575409435345882, + "grad_norm": 4.998250484466553, + "learning_rate": 1.993326579302551e-05, + "loss": 1.2048, + "step": 269 + }, + { + "epoch": 0.0659985333659252, + "grad_norm": 3.9181408882141113, + "learning_rate": 1.9932349530750173e-05, + "loss": 1.1552, + "step": 270 + }, + { + "epoch": 0.06624297237839159, + "grad_norm": 10.106731414794922, + "learning_rate": 1.9931427042485252e-05, + "loss": 1.1628, + "step": 271 + }, + { + "epoch": 0.06648741139085798, + "grad_norm": 4.483226299285889, + "learning_rate": 1.993049832880899e-05, + "loss": 1.2237, + "step": 272 + }, + { + "epoch": 0.06673185040332437, + "grad_norm": 7.024012088775635, + "learning_rate": 1.992956339030355e-05, + "loss": 1.2936, + "step": 273 + }, + { + "epoch": 0.06697628941579076, + "grad_norm": 5.0647430419921875, + "learning_rate": 1.992862222755498e-05, + "loss": 1.3156, + "step": 274 + }, + { + "epoch": 0.06722072842825715, + "grad_norm": 6.234722137451172, + "learning_rate": 1.992767484115324e-05, + "loss": 1.2481, + "step": 275 + }, + { + "epoch": 0.06746516744072353, + "grad_norm": 5.06250524520874, + "learning_rate": 1.992672123169219e-05, + "loss": 1.177, + "step": 276 + }, + { + "epoch": 0.06770960645318994, + "grad_norm": 6.165891647338867, + "learning_rate": 1.992576139976959e-05, + "loss": 1.2062, + "step": 277 + }, + { + "epoch": 0.06795404546565632, + "grad_norm": 5.269135475158691, + "learning_rate": 1.9924795345987103e-05, + "loss": 1.2643, + "step": 278 + }, + { + "epoch": 0.06819848447812271, + "grad_norm": 10.178529739379883, + "learning_rate": 1.9923823070950282e-05, + "loss": 1.3677, + "step": 279 + }, + { + "epoch": 0.0684429234905891, + "grad_norm": 6.4800310134887695, + "learning_rate": 1.99228445752686e-05, + "loss": 1.2403, + "step": 280 + }, + { + "epoch": 0.06868736250305549, + "grad_norm": 5.289521217346191, + "learning_rate": 1.992185985955541e-05, + "loss": 1.2473, + "step": 281 + }, + { + "epoch": 0.06893180151552188, + "grad_norm": 4.943880081176758, + "learning_rate": 1.9920868924427964e-05, + "loss": 1.2794, + "step": 282 + }, + { + "epoch": 0.06917624052798826, + "grad_norm": 6.226718425750732, + "learning_rate": 1.991987177050743e-05, + "loss": 1.3859, + "step": 283 + }, + { + "epoch": 0.06942067954045465, + "grad_norm": 4.328200817108154, + "learning_rate": 1.991886839841886e-05, + "loss": 1.2455, + "step": 284 + }, + { + "epoch": 0.06966511855292104, + "grad_norm": 4.170522689819336, + "learning_rate": 1.9917858808791205e-05, + "loss": 1.2064, + "step": 285 + }, + { + "epoch": 0.06990955756538744, + "grad_norm": 4.448747158050537, + "learning_rate": 1.9916843002257322e-05, + "loss": 1.3702, + "step": 286 + }, + { + "epoch": 0.07015399657785383, + "grad_norm": 3.4251625537872314, + "learning_rate": 1.991582097945395e-05, + "loss": 1.2967, + "step": 287 + }, + { + "epoch": 0.07039843559032022, + "grad_norm": 4.255934715270996, + "learning_rate": 1.9914792741021742e-05, + "loss": 1.2506, + "step": 288 + }, + { + "epoch": 0.0706428746027866, + "grad_norm": 3.717904567718506, + "learning_rate": 1.9913758287605227e-05, + "loss": 1.2631, + "step": 289 + }, + { + "epoch": 0.070887313615253, + "grad_norm": 3.819291353225708, + "learning_rate": 1.9912717619852856e-05, + "loss": 1.2166, + "step": 290 + }, + { + "epoch": 0.07113175262771938, + "grad_norm": 3.6381824016571045, + "learning_rate": 1.991167073841695e-05, + "loss": 1.1999, + "step": 291 + }, + { + "epoch": 0.07137619164018577, + "grad_norm": 5.513364791870117, + "learning_rate": 1.9910617643953737e-05, + "loss": 1.2485, + "step": 292 + }, + { + "epoch": 0.07162063065265216, + "grad_norm": 7.550871849060059, + "learning_rate": 1.990955833712334e-05, + "loss": 1.1929, + "step": 293 + }, + { + "epoch": 0.07186506966511855, + "grad_norm": 5.627939224243164, + "learning_rate": 1.9908492818589776e-05, + "loss": 1.3323, + "step": 294 + }, + { + "epoch": 0.07210950867758495, + "grad_norm": 3.952164649963379, + "learning_rate": 1.9907421089020956e-05, + "loss": 1.2111, + "step": 295 + }, + { + "epoch": 0.07235394769005134, + "grad_norm": 5.329583168029785, + "learning_rate": 1.9906343149088675e-05, + "loss": 1.2314, + "step": 296 + }, + { + "epoch": 0.07259838670251773, + "grad_norm": 8.233551025390625, + "learning_rate": 1.9905258999468634e-05, + "loss": 1.2911, + "step": 297 + }, + { + "epoch": 0.07284282571498411, + "grad_norm": 5.11754035949707, + "learning_rate": 1.990416864084042e-05, + "loss": 1.189, + "step": 298 + }, + { + "epoch": 0.0730872647274505, + "grad_norm": 6.454300880432129, + "learning_rate": 1.9903072073887507e-05, + "loss": 1.1986, + "step": 299 + }, + { + "epoch": 0.07333170373991689, + "grad_norm": 4.868954181671143, + "learning_rate": 1.9901969299297278e-05, + "loss": 1.2302, + "step": 300 + }, + { + "epoch": 0.07357614275238328, + "grad_norm": 4.498685359954834, + "learning_rate": 1.990086031776099e-05, + "loss": 1.1852, + "step": 301 + }, + { + "epoch": 0.07382058176484967, + "grad_norm": 7.261220932006836, + "learning_rate": 1.9899745129973792e-05, + "loss": 1.261, + "step": 302 + }, + { + "epoch": 0.07406502077731605, + "grad_norm": 3.6971499919891357, + "learning_rate": 1.9898623736634738e-05, + "loss": 1.1989, + "step": 303 + }, + { + "epoch": 0.07430945978978246, + "grad_norm": 10.59734058380127, + "learning_rate": 1.989749613844675e-05, + "loss": 1.2304, + "step": 304 + }, + { + "epoch": 0.07455389880224884, + "grad_norm": 4.786809921264648, + "learning_rate": 1.989636233611666e-05, + "loss": 1.2734, + "step": 305 + }, + { + "epoch": 0.07479833781471523, + "grad_norm": 3.672269821166992, + "learning_rate": 1.9895222330355174e-05, + "loss": 1.2867, + "step": 306 + }, + { + "epoch": 0.07504277682718162, + "grad_norm": 4.053799629211426, + "learning_rate": 1.98940761218769e-05, + "loss": 2.2605, + "step": 307 + }, + { + "epoch": 0.07528721583964801, + "grad_norm": 5.109869480133057, + "learning_rate": 1.989292371140032e-05, + "loss": 1.2306, + "step": 308 + }, + { + "epoch": 0.0755316548521144, + "grad_norm": 3.5280559062957764, + "learning_rate": 1.989176509964781e-05, + "loss": 1.2821, + "step": 309 + }, + { + "epoch": 0.07577609386458078, + "grad_norm": 5.775945663452148, + "learning_rate": 1.989060028734564e-05, + "loss": 1.2577, + "step": 310 + }, + { + "epoch": 0.07602053287704717, + "grad_norm": 4.2715864181518555, + "learning_rate": 1.9889429275223958e-05, + "loss": 1.2305, + "step": 311 + }, + { + "epoch": 0.07626497188951356, + "grad_norm": 3.606166124343872, + "learning_rate": 1.9888252064016798e-05, + "loss": 1.2248, + "step": 312 + }, + { + "epoch": 0.07650941090197996, + "grad_norm": 5.557638168334961, + "learning_rate": 1.9887068654462088e-05, + "loss": 1.2033, + "step": 313 + }, + { + "epoch": 0.07675384991444635, + "grad_norm": 3.988684892654419, + "learning_rate": 1.9885879047301628e-05, + "loss": 2.2842, + "step": 314 + }, + { + "epoch": 0.07699828892691274, + "grad_norm": 4.706129550933838, + "learning_rate": 1.9884683243281117e-05, + "loss": 1.257, + "step": 315 + }, + { + "epoch": 0.07724272793937913, + "grad_norm": 8.253270149230957, + "learning_rate": 1.988348124315013e-05, + "loss": 1.3223, + "step": 316 + }, + { + "epoch": 0.07748716695184552, + "grad_norm": 4.516219615936279, + "learning_rate": 1.9882273047662128e-05, + "loss": 1.2565, + "step": 317 + }, + { + "epoch": 0.0777316059643119, + "grad_norm": 3.8688483238220215, + "learning_rate": 1.9881058657574456e-05, + "loss": 1.2674, + "step": 318 + }, + { + "epoch": 0.07797604497677829, + "grad_norm": 4.626121520996094, + "learning_rate": 1.9879838073648342e-05, + "loss": 1.2049, + "step": 319 + }, + { + "epoch": 0.07822048398924468, + "grad_norm": 5.420572757720947, + "learning_rate": 1.98786112966489e-05, + "loss": 1.2808, + "step": 320 + }, + { + "epoch": 0.07846492300171107, + "grad_norm": 5.866894721984863, + "learning_rate": 1.9877378327345115e-05, + "loss": 1.3138, + "step": 321 + }, + { + "epoch": 0.07870936201417746, + "grad_norm": 5.478062629699707, + "learning_rate": 1.9876139166509863e-05, + "loss": 1.2282, + "step": 322 + }, + { + "epoch": 0.07895380102664386, + "grad_norm": 4.921603202819824, + "learning_rate": 1.9874893814919908e-05, + "loss": 1.2234, + "step": 323 + }, + { + "epoch": 0.07919824003911025, + "grad_norm": 5.043973445892334, + "learning_rate": 1.9873642273355873e-05, + "loss": 1.1653, + "step": 324 + }, + { + "epoch": 0.07944267905157663, + "grad_norm": 5.5380144119262695, + "learning_rate": 1.9872384542602278e-05, + "loss": 1.2197, + "step": 325 + }, + { + "epoch": 0.07968711806404302, + "grad_norm": 4.780150413513184, + "learning_rate": 1.987112062344752e-05, + "loss": 1.1963, + "step": 326 + }, + { + "epoch": 0.07993155707650941, + "grad_norm": 5.572015285491943, + "learning_rate": 1.9869850516683875e-05, + "loss": 1.15, + "step": 327 + }, + { + "epoch": 0.0801759960889758, + "grad_norm": 6.987370491027832, + "learning_rate": 1.9868574223107493e-05, + "loss": 1.1748, + "step": 328 + }, + { + "epoch": 0.08042043510144219, + "grad_norm": 6.550912380218506, + "learning_rate": 1.9867291743518407e-05, + "loss": 1.2382, + "step": 329 + }, + { + "epoch": 0.08066487411390857, + "grad_norm": 6.358066082000732, + "learning_rate": 1.9866003078720526e-05, + "loss": 1.2318, + "step": 330 + }, + { + "epoch": 0.08090931312637496, + "grad_norm": 8.014790534973145, + "learning_rate": 1.9864708229521637e-05, + "loss": 1.2417, + "step": 331 + }, + { + "epoch": 0.08115375213884136, + "grad_norm": 6.014158248901367, + "learning_rate": 1.9863407196733396e-05, + "loss": 1.0414, + "step": 332 + }, + { + "epoch": 0.08139819115130775, + "grad_norm": 6.411825180053711, + "learning_rate": 1.986209998117135e-05, + "loss": 1.2771, + "step": 333 + }, + { + "epoch": 0.08164263016377414, + "grad_norm": 6.666935920715332, + "learning_rate": 1.9860786583654913e-05, + "loss": 1.3119, + "step": 334 + }, + { + "epoch": 0.08188706917624053, + "grad_norm": 5.211427688598633, + "learning_rate": 1.985946700500737e-05, + "loss": 1.2362, + "step": 335 + }, + { + "epoch": 0.08213150818870692, + "grad_norm": 5.023791313171387, + "learning_rate": 1.9858141246055883e-05, + "loss": 1.2924, + "step": 336 + }, + { + "epoch": 0.0823759472011733, + "grad_norm": 4.761742115020752, + "learning_rate": 1.9856809307631498e-05, + "loss": 2.2858, + "step": 337 + }, + { + "epoch": 0.0826203862136397, + "grad_norm": 5.831111907958984, + "learning_rate": 1.9855471190569125e-05, + "loss": 1.1394, + "step": 338 + }, + { + "epoch": 0.08286482522610608, + "grad_norm": 5.819721698760986, + "learning_rate": 1.985412689570754e-05, + "loss": 1.2416, + "step": 339 + }, + { + "epoch": 0.08310926423857247, + "grad_norm": 5.4629645347595215, + "learning_rate": 1.9852776423889414e-05, + "loss": 1.3436, + "step": 340 + }, + { + "epoch": 0.08335370325103887, + "grad_norm": 5.885761260986328, + "learning_rate": 1.9851419775961265e-05, + "loss": 1.1212, + "step": 341 + }, + { + "epoch": 0.08359814226350526, + "grad_norm": 3.1790401935577393, + "learning_rate": 1.98500569527735e-05, + "loss": 2.1646, + "step": 342 + }, + { + "epoch": 0.08384258127597165, + "grad_norm": 6.93152379989624, + "learning_rate": 1.9848687955180386e-05, + "loss": 1.1123, + "step": 343 + }, + { + "epoch": 0.08408702028843804, + "grad_norm": 6.305123805999756, + "learning_rate": 1.9847312784040068e-05, + "loss": 1.2377, + "step": 344 + }, + { + "epoch": 0.08433145930090442, + "grad_norm": 11.483536720275879, + "learning_rate": 1.9845931440214557e-05, + "loss": 1.1454, + "step": 345 + }, + { + "epoch": 0.08457589831337081, + "grad_norm": 4.77037239074707, + "learning_rate": 1.984454392456973e-05, + "loss": 1.3617, + "step": 346 + }, + { + "epoch": 0.0848203373258372, + "grad_norm": 4.1926984786987305, + "learning_rate": 1.9843150237975343e-05, + "loss": 1.2383, + "step": 347 + }, + { + "epoch": 0.08506477633830359, + "grad_norm": 4.398230075836182, + "learning_rate": 1.984175038130501e-05, + "loss": 1.2658, + "step": 348 + }, + { + "epoch": 0.08530921535076998, + "grad_norm": 6.484598159790039, + "learning_rate": 1.984034435543622e-05, + "loss": 1.1849, + "step": 349 + }, + { + "epoch": 0.08555365436323638, + "grad_norm": 4.679854393005371, + "learning_rate": 1.9838932161250318e-05, + "loss": 1.255, + "step": 350 + }, + { + "epoch": 0.08579809337570277, + "grad_norm": 6.1571807861328125, + "learning_rate": 1.9837513799632536e-05, + "loss": 1.2073, + "step": 351 + }, + { + "epoch": 0.08604253238816915, + "grad_norm": 5.911917686462402, + "learning_rate": 1.9836089271471947e-05, + "loss": 1.2561, + "step": 352 + }, + { + "epoch": 0.08628697140063554, + "grad_norm": 4.896956920623779, + "learning_rate": 1.9834658577661505e-05, + "loss": 1.2099, + "step": 353 + }, + { + "epoch": 0.08653141041310193, + "grad_norm": 4.391876220703125, + "learning_rate": 1.983322171909803e-05, + "loss": 1.304, + "step": 354 + }, + { + "epoch": 0.08677584942556832, + "grad_norm": 4.319279670715332, + "learning_rate": 1.9831778696682195e-05, + "loss": 1.0834, + "step": 355 + }, + { + "epoch": 0.0870202884380347, + "grad_norm": 4.364986896514893, + "learning_rate": 1.983032951131855e-05, + "loss": 1.1473, + "step": 356 + }, + { + "epoch": 0.0872647274505011, + "grad_norm": 4.4287028312683105, + "learning_rate": 1.9828874163915496e-05, + "loss": 1.211, + "step": 357 + }, + { + "epoch": 0.08750916646296748, + "grad_norm": 5.178832530975342, + "learning_rate": 1.9827412655385308e-05, + "loss": 1.2318, + "step": 358 + }, + { + "epoch": 0.08775360547543389, + "grad_norm": 3.4217987060546875, + "learning_rate": 1.9825944986644116e-05, + "loss": 1.0763, + "step": 359 + }, + { + "epoch": 0.08799804448790027, + "grad_norm": 6.738720417022705, + "learning_rate": 1.9824471158611907e-05, + "loss": 1.2174, + "step": 360 + }, + { + "epoch": 0.08824248350036666, + "grad_norm": 4.281075954437256, + "learning_rate": 1.982299117221254e-05, + "loss": 1.1825, + "step": 361 + }, + { + "epoch": 0.08848692251283305, + "grad_norm": 3.974607229232788, + "learning_rate": 1.9821505028373727e-05, + "loss": 1.1445, + "step": 362 + }, + { + "epoch": 0.08873136152529944, + "grad_norm": 4.769040584564209, + "learning_rate": 1.9820012728027044e-05, + "loss": 1.128, + "step": 363 + }, + { + "epoch": 0.08897580053776583, + "grad_norm": 4.660463333129883, + "learning_rate": 1.9818514272107923e-05, + "loss": 1.1685, + "step": 364 + }, + { + "epoch": 0.08922023955023221, + "grad_norm": 4.7089338302612305, + "learning_rate": 1.9817009661555654e-05, + "loss": 1.2943, + "step": 365 + }, + { + "epoch": 0.0894646785626986, + "grad_norm": 3.7988078594207764, + "learning_rate": 1.9815498897313387e-05, + "loss": 1.0695, + "step": 366 + }, + { + "epoch": 0.08970911757516499, + "grad_norm": 4.49351692199707, + "learning_rate": 1.981398198032813e-05, + "loss": 1.2104, + "step": 367 + }, + { + "epoch": 0.08995355658763139, + "grad_norm": 4.02097749710083, + "learning_rate": 1.9812458911550748e-05, + "loss": 1.2177, + "step": 368 + }, + { + "epoch": 0.09019799560009778, + "grad_norm": 4.062883377075195, + "learning_rate": 1.9810929691935956e-05, + "loss": 1.1625, + "step": 369 + }, + { + "epoch": 0.09044243461256417, + "grad_norm": 4.29163932800293, + "learning_rate": 1.9809394322442333e-05, + "loss": 1.1793, + "step": 370 + }, + { + "epoch": 0.09068687362503056, + "grad_norm": 9.156668663024902, + "learning_rate": 1.9807852804032306e-05, + "loss": 1.1871, + "step": 371 + }, + { + "epoch": 0.09093131263749694, + "grad_norm": 5.449754238128662, + "learning_rate": 1.9806305137672165e-05, + "loss": 1.1717, + "step": 372 + }, + { + "epoch": 0.09117575164996333, + "grad_norm": 6.121524810791016, + "learning_rate": 1.9804751324332043e-05, + "loss": 1.0906, + "step": 373 + }, + { + "epoch": 0.09142019066242972, + "grad_norm": 6.13816499710083, + "learning_rate": 1.9803191364985937e-05, + "loss": 1.1783, + "step": 374 + }, + { + "epoch": 0.09166462967489611, + "grad_norm": 4.645517826080322, + "learning_rate": 1.9801625260611684e-05, + "loss": 1.1945, + "step": 375 + }, + { + "epoch": 0.0919090686873625, + "grad_norm": 5.3838396072387695, + "learning_rate": 1.9800053012190988e-05, + "loss": 1.2118, + "step": 376 + }, + { + "epoch": 0.0921535076998289, + "grad_norm": 9.308674812316895, + "learning_rate": 1.9798474620709395e-05, + "loss": 1.1243, + "step": 377 + }, + { + "epoch": 0.09239794671229529, + "grad_norm": 4.490804672241211, + "learning_rate": 1.9796890087156295e-05, + "loss": 1.3031, + "step": 378 + }, + { + "epoch": 0.09264238572476167, + "grad_norm": 4.418246746063232, + "learning_rate": 1.9795299412524948e-05, + "loss": 1.2028, + "step": 379 + }, + { + "epoch": 0.09288682473722806, + "grad_norm": 4.358879089355469, + "learning_rate": 1.9793702597812444e-05, + "loss": 1.0912, + "step": 380 + }, + { + "epoch": 0.09313126374969445, + "grad_norm": 4.636160850524902, + "learning_rate": 1.979209964401973e-05, + "loss": 1.1982, + "step": 381 + }, + { + "epoch": 0.09337570276216084, + "grad_norm": 4.515685558319092, + "learning_rate": 1.9790490552151605e-05, + "loss": 1.1392, + "step": 382 + }, + { + "epoch": 0.09362014177462723, + "grad_norm": 4.175724029541016, + "learning_rate": 1.978887532321671e-05, + "loss": 1.2269, + "step": 383 + }, + { + "epoch": 0.09386458078709362, + "grad_norm": 3.605635404586792, + "learning_rate": 1.9787253958227534e-05, + "loss": 1.2386, + "step": 384 + }, + { + "epoch": 0.09410901979956, + "grad_norm": 4.983905792236328, + "learning_rate": 1.9785626458200413e-05, + "loss": 1.2003, + "step": 385 + }, + { + "epoch": 0.0943534588120264, + "grad_norm": 3.675171375274658, + "learning_rate": 1.9783992824155523e-05, + "loss": 1.2266, + "step": 386 + }, + { + "epoch": 0.0945978978244928, + "grad_norm": 4.82205867767334, + "learning_rate": 1.9782353057116902e-05, + "loss": 1.1161, + "step": 387 + }, + { + "epoch": 0.09484233683695918, + "grad_norm": 4.621769428253174, + "learning_rate": 1.978070715811241e-05, + "loss": 1.1658, + "step": 388 + }, + { + "epoch": 0.09508677584942557, + "grad_norm": 4.2349982261657715, + "learning_rate": 1.9779055128173766e-05, + "loss": 1.269, + "step": 389 + }, + { + "epoch": 0.09533121486189196, + "grad_norm": 3.450382947921753, + "learning_rate": 1.9777396968336526e-05, + "loss": 1.169, + "step": 390 + }, + { + "epoch": 0.09557565387435835, + "grad_norm": 3.7534186840057373, + "learning_rate": 1.9775732679640093e-05, + "loss": 1.1282, + "step": 391 + }, + { + "epoch": 0.09582009288682473, + "grad_norm": 4.103036880493164, + "learning_rate": 1.9774062263127706e-05, + "loss": 1.2434, + "step": 392 + }, + { + "epoch": 0.09606453189929112, + "grad_norm": 4.361753463745117, + "learning_rate": 1.9772385719846454e-05, + "loss": 1.15, + "step": 393 + }, + { + "epoch": 0.09630897091175751, + "grad_norm": 4.32541561126709, + "learning_rate": 1.9770703050847254e-05, + "loss": 1.2142, + "step": 394 + }, + { + "epoch": 0.09655340992422391, + "grad_norm": 3.983166456222534, + "learning_rate": 1.976901425718487e-05, + "loss": 1.1632, + "step": 395 + }, + { + "epoch": 0.0967978489366903, + "grad_norm": 3.596562385559082, + "learning_rate": 1.9767319339917908e-05, + "loss": 1.1292, + "step": 396 + }, + { + "epoch": 0.09704228794915669, + "grad_norm": 4.378901481628418, + "learning_rate": 1.976561830010881e-05, + "loss": 1.2467, + "step": 397 + }, + { + "epoch": 0.09728672696162308, + "grad_norm": 4.227049827575684, + "learning_rate": 1.976391113882385e-05, + "loss": 1.166, + "step": 398 + }, + { + "epoch": 0.09753116597408946, + "grad_norm": 5.524839401245117, + "learning_rate": 1.9762197857133148e-05, + "loss": 1.2571, + "step": 399 + }, + { + "epoch": 0.09777560498655585, + "grad_norm": 6.738865852355957, + "learning_rate": 1.9760478456110655e-05, + "loss": 1.23, + "step": 400 + }, + { + "epoch": 0.09802004399902224, + "grad_norm": 5.322516918182373, + "learning_rate": 1.975875293683416e-05, + "loss": 1.2279, + "step": 401 + }, + { + "epoch": 0.09826448301148863, + "grad_norm": 4.052256107330322, + "learning_rate": 1.9757021300385288e-05, + "loss": 1.2415, + "step": 402 + }, + { + "epoch": 0.09850892202395502, + "grad_norm": 3.557264566421509, + "learning_rate": 1.9755283547849496e-05, + "loss": 1.2575, + "step": 403 + }, + { + "epoch": 0.09875336103642142, + "grad_norm": 5.424764156341553, + "learning_rate": 1.9753539680316077e-05, + "loss": 1.2741, + "step": 404 + }, + { + "epoch": 0.09899780004888781, + "grad_norm": 4.176656246185303, + "learning_rate": 1.9751789698878153e-05, + "loss": 1.2557, + "step": 405 + }, + { + "epoch": 0.0992422390613542, + "grad_norm": 3.171238660812378, + "learning_rate": 1.9750033604632683e-05, + "loss": 1.0859, + "step": 406 + }, + { + "epoch": 0.09948667807382058, + "grad_norm": 3.339172840118408, + "learning_rate": 1.974827139868046e-05, + "loss": 1.2761, + "step": 407 + }, + { + "epoch": 0.09973111708628697, + "grad_norm": 3.803142786026001, + "learning_rate": 1.9746503082126102e-05, + "loss": 1.1748, + "step": 408 + }, + { + "epoch": 0.09997555609875336, + "grad_norm": 2.872415781021118, + "learning_rate": 1.974472865607806e-05, + "loss": 2.1624, + "step": 409 + }, + { + "epoch": 0.10021999511121975, + "grad_norm": 4.578094959259033, + "learning_rate": 1.9742948121648614e-05, + "loss": 1.1665, + "step": 410 + }, + { + "epoch": 0.10046443412368614, + "grad_norm": 4.066554546356201, + "learning_rate": 1.9741161479953872e-05, + "loss": 1.2893, + "step": 411 + }, + { + "epoch": 0.10070887313615252, + "grad_norm": 3.227421760559082, + "learning_rate": 1.9739368732113773e-05, + "loss": 1.2414, + "step": 412 + }, + { + "epoch": 0.10095331214861893, + "grad_norm": 4.6151580810546875, + "learning_rate": 1.9737569879252082e-05, + "loss": 1.1905, + "step": 413 + }, + { + "epoch": 0.10119775116108531, + "grad_norm": 3.6246869564056396, + "learning_rate": 1.9735764922496393e-05, + "loss": 1.2155, + "step": 414 + }, + { + "epoch": 0.1014421901735517, + "grad_norm": 3.8012173175811768, + "learning_rate": 1.9733953862978125e-05, + "loss": 1.1976, + "step": 415 + }, + { + "epoch": 0.10168662918601809, + "grad_norm": 6.171485424041748, + "learning_rate": 1.9732136701832515e-05, + "loss": 1.2181, + "step": 416 + }, + { + "epoch": 0.10193106819848448, + "grad_norm": 3.9184229373931885, + "learning_rate": 1.9730313440198637e-05, + "loss": 1.156, + "step": 417 + }, + { + "epoch": 0.10217550721095087, + "grad_norm": 4.018677711486816, + "learning_rate": 1.9728484079219385e-05, + "loss": 1.0827, + "step": 418 + }, + { + "epoch": 0.10241994622341725, + "grad_norm": 3.981010675430298, + "learning_rate": 1.972664862004147e-05, + "loss": 1.315, + "step": 419 + }, + { + "epoch": 0.10266438523588364, + "grad_norm": 3.377622604370117, + "learning_rate": 1.9724807063815432e-05, + "loss": 1.1715, + "step": 420 + }, + { + "epoch": 0.10290882424835003, + "grad_norm": 2.6135737895965576, + "learning_rate": 1.9722959411695636e-05, + "loss": 2.1113, + "step": 421 + }, + { + "epoch": 0.10315326326081643, + "grad_norm": 4.2256550788879395, + "learning_rate": 1.9721105664840253e-05, + "loss": 1.1828, + "step": 422 + }, + { + "epoch": 0.10339770227328282, + "grad_norm": 4.993865966796875, + "learning_rate": 1.9719245824411298e-05, + "loss": 1.1503, + "step": 423 + }, + { + "epoch": 0.10364214128574921, + "grad_norm": 3.423783540725708, + "learning_rate": 1.9717379891574583e-05, + "loss": 1.0649, + "step": 424 + }, + { + "epoch": 0.1038865802982156, + "grad_norm": 3.6675407886505127, + "learning_rate": 1.9715507867499754e-05, + "loss": 1.0481, + "step": 425 + }, + { + "epoch": 0.10413101931068199, + "grad_norm": 3.9582297801971436, + "learning_rate": 1.9713629753360265e-05, + "loss": 1.1743, + "step": 426 + }, + { + "epoch": 0.10437545832314837, + "grad_norm": 4.515329837799072, + "learning_rate": 1.9711745550333392e-05, + "loss": 1.1082, + "step": 427 + }, + { + "epoch": 0.10461989733561476, + "grad_norm": 4.547237873077393, + "learning_rate": 1.9709855259600236e-05, + "loss": 1.2417, + "step": 428 + }, + { + "epoch": 0.10486433634808115, + "grad_norm": 4.494861125946045, + "learning_rate": 1.9707958882345693e-05, + "loss": 1.1873, + "step": 429 + }, + { + "epoch": 0.10510877536054754, + "grad_norm": 7.8653059005737305, + "learning_rate": 1.97060564197585e-05, + "loss": 1.1605, + "step": 430 + }, + { + "epoch": 0.10535321437301394, + "grad_norm": 4.853890895843506, + "learning_rate": 1.970414787303119e-05, + "loss": 1.1746, + "step": 431 + }, + { + "epoch": 0.10559765338548033, + "grad_norm": 3.662851095199585, + "learning_rate": 1.9702233243360114e-05, + "loss": 1.154, + "step": 432 + }, + { + "epoch": 0.10584209239794672, + "grad_norm": 4.81683349609375, + "learning_rate": 1.9700312531945444e-05, + "loss": 1.3095, + "step": 433 + }, + { + "epoch": 0.1060865314104131, + "grad_norm": 4.331052780151367, + "learning_rate": 1.9698385739991147e-05, + "loss": 1.1571, + "step": 434 + }, + { + "epoch": 0.10633097042287949, + "grad_norm": 4.024257659912109, + "learning_rate": 1.9696452868705026e-05, + "loss": 1.1725, + "step": 435 + }, + { + "epoch": 0.10657540943534588, + "grad_norm": 6.647060871124268, + "learning_rate": 1.969451391929867e-05, + "loss": 1.1375, + "step": 436 + }, + { + "epoch": 0.10681984844781227, + "grad_norm": 4.554993152618408, + "learning_rate": 1.9692568892987494e-05, + "loss": 1.1222, + "step": 437 + }, + { + "epoch": 0.10706428746027866, + "grad_norm": 3.893545627593994, + "learning_rate": 1.9690617790990717e-05, + "loss": 1.2332, + "step": 438 + }, + { + "epoch": 0.10730872647274504, + "grad_norm": 2.6785714626312256, + "learning_rate": 1.968866061453137e-05, + "loss": 2.1153, + "step": 439 + }, + { + "epoch": 0.10755316548521145, + "grad_norm": 5.624014854431152, + "learning_rate": 1.9686697364836278e-05, + "loss": 1.2113, + "step": 440 + }, + { + "epoch": 0.10779760449767783, + "grad_norm": 4.759518146514893, + "learning_rate": 1.9684728043136093e-05, + "loss": 1.1685, + "step": 441 + }, + { + "epoch": 0.10804204351014422, + "grad_norm": 4.094371318817139, + "learning_rate": 1.9682752650665268e-05, + "loss": 1.2696, + "step": 442 + }, + { + "epoch": 0.10828648252261061, + "grad_norm": 4.044309139251709, + "learning_rate": 1.9680771188662044e-05, + "loss": 1.1713, + "step": 443 + }, + { + "epoch": 0.108530921535077, + "grad_norm": 3.826819658279419, + "learning_rate": 1.967878365836849e-05, + "loss": 1.1808, + "step": 444 + }, + { + "epoch": 0.10877536054754339, + "grad_norm": 4.536250591278076, + "learning_rate": 1.9676790061030466e-05, + "loss": 1.3316, + "step": 445 + }, + { + "epoch": 0.10901979956000978, + "grad_norm": 3.299072742462158, + "learning_rate": 1.9674790397897633e-05, + "loss": 1.1167, + "step": 446 + }, + { + "epoch": 0.10926423857247616, + "grad_norm": 3.9642040729522705, + "learning_rate": 1.9672784670223465e-05, + "loss": 1.2412, + "step": 447 + }, + { + "epoch": 0.10950867758494255, + "grad_norm": 3.646501302719116, + "learning_rate": 1.967077287926523e-05, + "loss": 1.1585, + "step": 448 + }, + { + "epoch": 0.10975311659740894, + "grad_norm": 4.142199993133545, + "learning_rate": 1.9668755026283995e-05, + "loss": 1.2492, + "step": 449 + }, + { + "epoch": 0.10999755560987534, + "grad_norm": 4.070352554321289, + "learning_rate": 1.9666731112544638e-05, + "loss": 1.2055, + "step": 450 + }, + { + "epoch": 0.11024199462234173, + "grad_norm": 3.4566304683685303, + "learning_rate": 1.966470113931582e-05, + "loss": 1.1793, + "step": 451 + }, + { + "epoch": 0.11048643363480812, + "grad_norm": 3.860222578048706, + "learning_rate": 1.9662665107870007e-05, + "loss": 1.1789, + "step": 452 + }, + { + "epoch": 0.1107308726472745, + "grad_norm": 3.7197413444519043, + "learning_rate": 1.966062301948347e-05, + "loss": 1.1849, + "step": 453 + }, + { + "epoch": 0.1109753116597409, + "grad_norm": 4.1757965087890625, + "learning_rate": 1.965857487543627e-05, + "loss": 1.2825, + "step": 454 + }, + { + "epoch": 0.11121975067220728, + "grad_norm": 4.345134258270264, + "learning_rate": 1.9656520677012264e-05, + "loss": 1.1732, + "step": 455 + }, + { + "epoch": 0.11146418968467367, + "grad_norm": 4.474052429199219, + "learning_rate": 1.965446042549911e-05, + "loss": 1.1597, + "step": 456 + }, + { + "epoch": 0.11170862869714006, + "grad_norm": 3.8579421043395996, + "learning_rate": 1.9652394122188242e-05, + "loss": 1.2017, + "step": 457 + }, + { + "epoch": 0.11195306770960645, + "grad_norm": 5.020146369934082, + "learning_rate": 1.9650321768374908e-05, + "loss": 1.3404, + "step": 458 + }, + { + "epoch": 0.11219750672207285, + "grad_norm": 4.499261379241943, + "learning_rate": 1.9648243365358145e-05, + "loss": 1.298, + "step": 459 + }, + { + "epoch": 0.11244194573453924, + "grad_norm": 4.317763328552246, + "learning_rate": 1.9646158914440775e-05, + "loss": 1.2115, + "step": 460 + }, + { + "epoch": 0.11268638474700562, + "grad_norm": 3.9400923252105713, + "learning_rate": 1.9644068416929417e-05, + "loss": 1.0254, + "step": 461 + }, + { + "epoch": 0.11293082375947201, + "grad_norm": 4.948456764221191, + "learning_rate": 1.964197187413447e-05, + "loss": 1.3035, + "step": 462 + }, + { + "epoch": 0.1131752627719384, + "grad_norm": 6.067734718322754, + "learning_rate": 1.963986928737014e-05, + "loss": 1.1723, + "step": 463 + }, + { + "epoch": 0.11341970178440479, + "grad_norm": 3.6457936763763428, + "learning_rate": 1.96377606579544e-05, + "loss": 1.2246, + "step": 464 + }, + { + "epoch": 0.11366414079687118, + "grad_norm": 4.135344505310059, + "learning_rate": 1.963564598720903e-05, + "loss": 1.1155, + "step": 465 + }, + { + "epoch": 0.11390857980933757, + "grad_norm": 3.4885904788970947, + "learning_rate": 1.963352527645959e-05, + "loss": 1.198, + "step": 466 + }, + { + "epoch": 0.11415301882180395, + "grad_norm": 5.2463483810424805, + "learning_rate": 1.9631398527035424e-05, + "loss": 1.1427, + "step": 467 + }, + { + "epoch": 0.11439745783427036, + "grad_norm": 4.414389610290527, + "learning_rate": 1.962926574026966e-05, + "loss": 1.1839, + "step": 468 + }, + { + "epoch": 0.11464189684673674, + "grad_norm": 3.498870849609375, + "learning_rate": 1.962712691749922e-05, + "loss": 1.1213, + "step": 469 + }, + { + "epoch": 0.11488633585920313, + "grad_norm": 3.4490199089050293, + "learning_rate": 1.962498206006479e-05, + "loss": 2.1561, + "step": 470 + }, + { + "epoch": 0.11513077487166952, + "grad_norm": 2.884737730026245, + "learning_rate": 1.9622831169310864e-05, + "loss": 1.133, + "step": 471 + }, + { + "epoch": 0.11537521388413591, + "grad_norm": 4.203254699707031, + "learning_rate": 1.96206742465857e-05, + "loss": 1.2361, + "step": 472 + }, + { + "epoch": 0.1156196528966023, + "grad_norm": 3.780165195465088, + "learning_rate": 1.961851129324134e-05, + "loss": 1.1643, + "step": 473 + }, + { + "epoch": 0.11586409190906868, + "grad_norm": 6.365428447723389, + "learning_rate": 1.9616342310633615e-05, + "loss": 1.1947, + "step": 474 + }, + { + "epoch": 0.11610853092153507, + "grad_norm": 4.483095169067383, + "learning_rate": 1.9614167300122126e-05, + "loss": 1.0783, + "step": 475 + }, + { + "epoch": 0.11635296993400146, + "grad_norm": 4.36859130859375, + "learning_rate": 1.9611986263070256e-05, + "loss": 1.0763, + "step": 476 + }, + { + "epoch": 0.11659740894646786, + "grad_norm": 6.242516994476318, + "learning_rate": 1.9609799200845163e-05, + "loss": 1.284, + "step": 477 + }, + { + "epoch": 0.11684184795893425, + "grad_norm": 4.574044227600098, + "learning_rate": 1.9607606114817787e-05, + "loss": 1.0097, + "step": 478 + }, + { + "epoch": 0.11708628697140064, + "grad_norm": 4.711122989654541, + "learning_rate": 1.9605407006362843e-05, + "loss": 1.221, + "step": 479 + }, + { + "epoch": 0.11733072598386703, + "grad_norm": 4.41146993637085, + "learning_rate": 1.9603201876858812e-05, + "loss": 1.1135, + "step": 480 + }, + { + "epoch": 0.11757516499633341, + "grad_norm": 3.8882482051849365, + "learning_rate": 1.9600990727687964e-05, + "loss": 1.1542, + "step": 481 + }, + { + "epoch": 0.1178196040087998, + "grad_norm": 3.4748640060424805, + "learning_rate": 1.959877356023633e-05, + "loss": 1.0965, + "step": 482 + }, + { + "epoch": 0.11806404302126619, + "grad_norm": 4.37790060043335, + "learning_rate": 1.9596550375893718e-05, + "loss": 1.1817, + "step": 483 + }, + { + "epoch": 0.11830848203373258, + "grad_norm": 3.8884594440460205, + "learning_rate": 1.959432117605371e-05, + "loss": 1.2542, + "step": 484 + }, + { + "epoch": 0.11855292104619897, + "grad_norm": 3.711740732192993, + "learning_rate": 1.959208596211366e-05, + "loss": 1.0646, + "step": 485 + }, + { + "epoch": 0.11879736005866537, + "grad_norm": 3.075740337371826, + "learning_rate": 1.958984473547468e-05, + "loss": 2.1749, + "step": 486 + }, + { + "epoch": 0.11904179907113176, + "grad_norm": 4.204500198364258, + "learning_rate": 1.958759749754167e-05, + "loss": 1.181, + "step": 487 + }, + { + "epoch": 0.11928623808359815, + "grad_norm": 4.3053741455078125, + "learning_rate": 1.958534424972328e-05, + "loss": 1.1511, + "step": 488 + }, + { + "epoch": 0.11953067709606453, + "grad_norm": 7.22095251083374, + "learning_rate": 1.9583084993431935e-05, + "loss": 1.2187, + "step": 489 + }, + { + "epoch": 0.11977511610853092, + "grad_norm": 4.425522327423096, + "learning_rate": 1.958081973008383e-05, + "loss": 1.1437, + "step": 490 + }, + { + "epoch": 0.12001955512099731, + "grad_norm": 4.114257335662842, + "learning_rate": 1.9578548461098912e-05, + "loss": 1.1693, + "step": 491 + }, + { + "epoch": 0.1202639941334637, + "grad_norm": 4.519433498382568, + "learning_rate": 1.957627118790092e-05, + "loss": 1.2113, + "step": 492 + }, + { + "epoch": 0.12050843314593009, + "grad_norm": 4.922794342041016, + "learning_rate": 1.9573987911917325e-05, + "loss": 1.0789, + "step": 493 + }, + { + "epoch": 0.12075287215839647, + "grad_norm": 4.798068046569824, + "learning_rate": 1.9571698634579377e-05, + "loss": 1.2138, + "step": 494 + }, + { + "epoch": 0.12099731117086288, + "grad_norm": 4.2914533615112305, + "learning_rate": 1.956940335732209e-05, + "loss": 1.1323, + "step": 495 + }, + { + "epoch": 0.12124175018332926, + "grad_norm": 5.3744401931762695, + "learning_rate": 1.956710208158423e-05, + "loss": 1.2722, + "step": 496 + }, + { + "epoch": 0.12148618919579565, + "grad_norm": 3.724177122116089, + "learning_rate": 1.956479480880833e-05, + "loss": 1.0054, + "step": 497 + }, + { + "epoch": 0.12173062820826204, + "grad_norm": 22.230966567993164, + "learning_rate": 1.9562481540440677e-05, + "loss": 1.0581, + "step": 498 + }, + { + "epoch": 0.12197506722072843, + "grad_norm": 5.926990509033203, + "learning_rate": 1.9560162277931326e-05, + "loss": 1.128, + "step": 499 + }, + { + "epoch": 0.12221950623319482, + "grad_norm": 7.3720784187316895, + "learning_rate": 1.9557837022734075e-05, + "loss": 1.1306, + "step": 500 + }, + { + "epoch": 0.1224639452456612, + "grad_norm": 14.880962371826172, + "learning_rate": 1.9555505776306492e-05, + "loss": 1.2239, + "step": 501 + }, + { + "epoch": 0.12270838425812759, + "grad_norm": 4.3613200187683105, + "learning_rate": 1.9553168540109888e-05, + "loss": 1.1388, + "step": 502 + }, + { + "epoch": 0.12295282327059398, + "grad_norm": 5.127346992492676, + "learning_rate": 1.955082531560934e-05, + "loss": 1.0958, + "step": 503 + }, + { + "epoch": 0.12319726228306038, + "grad_norm": 4.747497081756592, + "learning_rate": 1.954847610427367e-05, + "loss": 1.0396, + "step": 504 + }, + { + "epoch": 0.12344170129552677, + "grad_norm": 4.675597667694092, + "learning_rate": 1.954612090757546e-05, + "loss": 1.1731, + "step": 505 + }, + { + "epoch": 0.12368614030799316, + "grad_norm": 34.22259521484375, + "learning_rate": 1.9543759726991037e-05, + "loss": 1.2323, + "step": 506 + }, + { + "epoch": 0.12393057932045955, + "grad_norm": 6.397866725921631, + "learning_rate": 1.954139256400049e-05, + "loss": 0.9992, + "step": 507 + }, + { + "epoch": 0.12417501833292593, + "grad_norm": 17.597700119018555, + "learning_rate": 1.9539019420087643e-05, + "loss": 1.1705, + "step": 508 + }, + { + "epoch": 0.12441945734539232, + "grad_norm": 8.879899978637695, + "learning_rate": 1.9536640296740078e-05, + "loss": 1.249, + "step": 509 + }, + { + "epoch": 0.12466389635785871, + "grad_norm": 14.198099136352539, + "learning_rate": 1.9534255195449123e-05, + "loss": 1.2276, + "step": 510 + }, + { + "epoch": 0.1249083353703251, + "grad_norm": 6.117961883544922, + "learning_rate": 1.9531864117709855e-05, + "loss": 1.2236, + "step": 511 + }, + { + "epoch": 0.1251527743827915, + "grad_norm": 7.427270412445068, + "learning_rate": 1.9529467065021098e-05, + "loss": 1.1852, + "step": 512 + }, + { + "epoch": 0.1253972133952579, + "grad_norm": 8.58948802947998, + "learning_rate": 1.952706403888542e-05, + "loss": 1.1082, + "step": 513 + }, + { + "epoch": 0.12564165240772426, + "grad_norm": 4.219729423522949, + "learning_rate": 1.9524655040809126e-05, + "loss": 1.1912, + "step": 514 + }, + { + "epoch": 0.12588609142019067, + "grad_norm": 4.362913608551025, + "learning_rate": 1.9522240072302275e-05, + "loss": 1.1358, + "step": 515 + }, + { + "epoch": 0.12613053043265704, + "grad_norm": 8.792346954345703, + "learning_rate": 1.9519819134878668e-05, + "loss": 1.1447, + "step": 516 + }, + { + "epoch": 0.12637496944512344, + "grad_norm": 7.836843490600586, + "learning_rate": 1.951739223005584e-05, + "loss": 1.2419, + "step": 517 + }, + { + "epoch": 0.12661940845758984, + "grad_norm": 4.472485542297363, + "learning_rate": 1.9514959359355076e-05, + "loss": 1.2295, + "step": 518 + }, + { + "epoch": 0.12686384747005622, + "grad_norm": 5.330399513244629, + "learning_rate": 1.9512520524301388e-05, + "loss": 1.2025, + "step": 519 + }, + { + "epoch": 0.12710828648252262, + "grad_norm": 5.496823310852051, + "learning_rate": 1.951007572642354e-05, + "loss": 2.1139, + "step": 520 + }, + { + "epoch": 0.127352725494989, + "grad_norm": 5.915637016296387, + "learning_rate": 1.950762496725403e-05, + "loss": 1.2428, + "step": 521 + }, + { + "epoch": 0.1275971645074554, + "grad_norm": 11.7316255569458, + "learning_rate": 1.9505168248329084e-05, + "loss": 1.1854, + "step": 522 + }, + { + "epoch": 0.12784160351992177, + "grad_norm": 8.459104537963867, + "learning_rate": 1.9502705571188675e-05, + "loss": 1.2272, + "step": 523 + }, + { + "epoch": 0.12808604253238817, + "grad_norm": 4.567411422729492, + "learning_rate": 1.9500236937376504e-05, + "loss": 1.0839, + "step": 524 + }, + { + "epoch": 0.12833048154485455, + "grad_norm": 4.574172019958496, + "learning_rate": 1.949776234844001e-05, + "loss": 1.0274, + "step": 525 + }, + { + "epoch": 0.12857492055732095, + "grad_norm": 7.426229476928711, + "learning_rate": 1.949528180593037e-05, + "loss": 1.1825, + "step": 526 + }, + { + "epoch": 0.12881935956978735, + "grad_norm": 4.16442346572876, + "learning_rate": 1.949279531140247e-05, + "loss": 1.1525, + "step": 527 + }, + { + "epoch": 0.12906379858225372, + "grad_norm": 4.447491645812988, + "learning_rate": 1.949030286641496e-05, + "loss": 1.2678, + "step": 528 + }, + { + "epoch": 0.12930823759472013, + "grad_norm": 6.830371379852295, + "learning_rate": 1.948780447253019e-05, + "loss": 1.2301, + "step": 529 + }, + { + "epoch": 0.1295526766071865, + "grad_norm": 3.7037405967712402, + "learning_rate": 1.9485300131314262e-05, + "loss": 2.1888, + "step": 530 + }, + { + "epoch": 0.1297971156196529, + "grad_norm": 6.7080912590026855, + "learning_rate": 1.948278984433699e-05, + "loss": 1.3056, + "step": 531 + }, + { + "epoch": 0.13004155463211928, + "grad_norm": 3.5571670532226562, + "learning_rate": 1.948027361317192e-05, + "loss": 2.1244, + "step": 532 + }, + { + "epoch": 0.13028599364458568, + "grad_norm": 5.456795692443848, + "learning_rate": 1.947775143939633e-05, + "loss": 1.2952, + "step": 533 + }, + { + "epoch": 0.13053043265705205, + "grad_norm": 6.878313064575195, + "learning_rate": 1.947522332459122e-05, + "loss": 1.18, + "step": 534 + }, + { + "epoch": 0.13077487166951846, + "grad_norm": 4.375424385070801, + "learning_rate": 1.9472689270341303e-05, + "loss": 1.1559, + "step": 535 + }, + { + "epoch": 0.13101931068198486, + "grad_norm": 6.418196201324463, + "learning_rate": 1.9470149278235032e-05, + "loss": 1.1485, + "step": 536 + }, + { + "epoch": 0.13126374969445123, + "grad_norm": 4.910287380218506, + "learning_rate": 1.946760334986458e-05, + "loss": 1.0982, + "step": 537 + }, + { + "epoch": 0.13150818870691763, + "grad_norm": 4.6932549476623535, + "learning_rate": 1.946505148682582e-05, + "loss": 1.1912, + "step": 538 + }, + { + "epoch": 0.131752627719384, + "grad_norm": 5.659276962280273, + "learning_rate": 1.9462493690718373e-05, + "loss": 1.1048, + "step": 539 + }, + { + "epoch": 0.1319970667318504, + "grad_norm": 7.565642833709717, + "learning_rate": 1.9459929963145558e-05, + "loss": 1.1765, + "step": 540 + }, + { + "epoch": 0.13224150574431678, + "grad_norm": 7.787476062774658, + "learning_rate": 1.945736030571443e-05, + "loss": 1.2969, + "step": 541 + }, + { + "epoch": 0.13248594475678319, + "grad_norm": 4.8547773361206055, + "learning_rate": 1.9454784720035747e-05, + "loss": 1.0794, + "step": 542 + }, + { + "epoch": 0.13273038376924956, + "grad_norm": 7.971114635467529, + "learning_rate": 1.945220320772399e-05, + "loss": 1.0347, + "step": 543 + }, + { + "epoch": 0.13297482278171596, + "grad_norm": 4.092767238616943, + "learning_rate": 1.944961577039735e-05, + "loss": 2.2551, + "step": 544 + }, + { + "epoch": 0.13321926179418236, + "grad_norm": 6.591254234313965, + "learning_rate": 1.9447022409677736e-05, + "loss": 1.1544, + "step": 545 + }, + { + "epoch": 0.13346370080664874, + "grad_norm": 5.323657035827637, + "learning_rate": 1.944442312719077e-05, + "loss": 1.1388, + "step": 546 + }, + { + "epoch": 0.13370813981911514, + "grad_norm": 4.295138835906982, + "learning_rate": 1.9441817924565787e-05, + "loss": 1.243, + "step": 547 + }, + { + "epoch": 0.13395257883158151, + "grad_norm": 3.4274163246154785, + "learning_rate": 1.9439206803435828e-05, + "loss": 2.145, + "step": 548 + }, + { + "epoch": 0.13419701784404792, + "grad_norm": 6.668155670166016, + "learning_rate": 1.9436589765437646e-05, + "loss": 1.3672, + "step": 549 + }, + { + "epoch": 0.1344414568565143, + "grad_norm": 4.891482830047607, + "learning_rate": 1.9433966812211712e-05, + "loss": 1.2499, + "step": 550 + }, + { + "epoch": 0.1346858958689807, + "grad_norm": 3.2084758281707764, + "learning_rate": 1.9431337945402186e-05, + "loss": 2.0341, + "step": 551 + }, + { + "epoch": 0.13493033488144707, + "grad_norm": 5.467001438140869, + "learning_rate": 1.9428703166656952e-05, + "loss": 1.059, + "step": 552 + }, + { + "epoch": 0.13517477389391347, + "grad_norm": 9.086297035217285, + "learning_rate": 1.942606247762759e-05, + "loss": 1.0457, + "step": 553 + }, + { + "epoch": 0.13541921290637987, + "grad_norm": 7.168754577636719, + "learning_rate": 1.9423415879969395e-05, + "loss": 1.2066, + "step": 554 + }, + { + "epoch": 0.13566365191884625, + "grad_norm": 5.245144844055176, + "learning_rate": 1.942076337534135e-05, + "loss": 1.1832, + "step": 555 + }, + { + "epoch": 0.13590809093131265, + "grad_norm": 5.616425037384033, + "learning_rate": 1.9418104965406158e-05, + "loss": 1.1588, + "step": 556 + }, + { + "epoch": 0.13615252994377902, + "grad_norm": 4.895685195922852, + "learning_rate": 1.941544065183021e-05, + "loss": 1.1814, + "step": 557 + }, + { + "epoch": 0.13639696895624542, + "grad_norm": 4.542455673217773, + "learning_rate": 1.9412770436283604e-05, + "loss": 1.2014, + "step": 558 + }, + { + "epoch": 0.1366414079687118, + "grad_norm": 6.379021167755127, + "learning_rate": 1.941009432044014e-05, + "loss": 1.2843, + "step": 559 + }, + { + "epoch": 0.1368858469811782, + "grad_norm": 9.439204216003418, + "learning_rate": 1.940741230597731e-05, + "loss": 1.1331, + "step": 560 + }, + { + "epoch": 0.13713028599364457, + "grad_norm": 5.6895856857299805, + "learning_rate": 1.9404724394576305e-05, + "loss": 1.1366, + "step": 561 + }, + { + "epoch": 0.13737472500611098, + "grad_norm": 3.5664186477661133, + "learning_rate": 1.940203058792202e-05, + "loss": 1.0722, + "step": 562 + }, + { + "epoch": 0.13761916401857738, + "grad_norm": 11.332030296325684, + "learning_rate": 1.9399330887703035e-05, + "loss": 1.0417, + "step": 563 + }, + { + "epoch": 0.13786360303104375, + "grad_norm": 3.848562717437744, + "learning_rate": 1.9396625295611634e-05, + "loss": 1.062, + "step": 564 + }, + { + "epoch": 0.13810804204351015, + "grad_norm": 5.241700649261475, + "learning_rate": 1.9393913813343784e-05, + "loss": 1.0942, + "step": 565 + }, + { + "epoch": 0.13835248105597653, + "grad_norm": 3.3595571517944336, + "learning_rate": 1.939119644259915e-05, + "loss": 1.12, + "step": 566 + }, + { + "epoch": 0.13859692006844293, + "grad_norm": 6.70857048034668, + "learning_rate": 1.938847318508109e-05, + "loss": 1.0577, + "step": 567 + }, + { + "epoch": 0.1388413590809093, + "grad_norm": 6.597379684448242, + "learning_rate": 1.9385744042496646e-05, + "loss": 1.1119, + "step": 568 + }, + { + "epoch": 0.1390857980933757, + "grad_norm": 6.321605682373047, + "learning_rate": 1.9383009016556555e-05, + "loss": 1.0859, + "step": 569 + }, + { + "epoch": 0.13933023710584208, + "grad_norm": 5.733395576477051, + "learning_rate": 1.9380268108975238e-05, + "loss": 1.1092, + "step": 570 + }, + { + "epoch": 0.13957467611830848, + "grad_norm": 4.5978922843933105, + "learning_rate": 1.9377521321470806e-05, + "loss": 1.1017, + "step": 571 + }, + { + "epoch": 0.13981911513077488, + "grad_norm": 6.471620082855225, + "learning_rate": 1.9374768655765056e-05, + "loss": 1.143, + "step": 572 + }, + { + "epoch": 0.14006355414324126, + "grad_norm": 3.8329930305480957, + "learning_rate": 1.937201011358346e-05, + "loss": 2.1819, + "step": 573 + }, + { + "epoch": 0.14030799315570766, + "grad_norm": 5.367978096008301, + "learning_rate": 1.9369245696655187e-05, + "loss": 1.0577, + "step": 574 + }, + { + "epoch": 0.14055243216817404, + "grad_norm": 5.2052435874938965, + "learning_rate": 1.9366475406713085e-05, + "loss": 1.1255, + "step": 575 + }, + { + "epoch": 0.14079687118064044, + "grad_norm": 9.861551284790039, + "learning_rate": 1.9363699245493678e-05, + "loss": 1.1107, + "step": 576 + }, + { + "epoch": 0.1410413101931068, + "grad_norm": 6.768249034881592, + "learning_rate": 1.9360917214737177e-05, + "loss": 1.1755, + "step": 577 + }, + { + "epoch": 0.1412857492055732, + "grad_norm": 2.9812920093536377, + "learning_rate": 1.935812931618746e-05, + "loss": 2.1717, + "step": 578 + }, + { + "epoch": 0.1415301882180396, + "grad_norm": 4.798773288726807, + "learning_rate": 1.9355335551592104e-05, + "loss": 0.9572, + "step": 579 + }, + { + "epoch": 0.141774627230506, + "grad_norm": 5.733400821685791, + "learning_rate": 1.9352535922702345e-05, + "loss": 1.0916, + "step": 580 + }, + { + "epoch": 0.1420190662429724, + "grad_norm": 3.7753677368164062, + "learning_rate": 1.93497304312731e-05, + "loss": 2.0032, + "step": 581 + }, + { + "epoch": 0.14226350525543877, + "grad_norm": 3.9927875995635986, + "learning_rate": 1.9346919079062965e-05, + "loss": 1.1038, + "step": 582 + }, + { + "epoch": 0.14250794426790517, + "grad_norm": 3.727595329284668, + "learning_rate": 1.9344101867834202e-05, + "loss": 1.0979, + "step": 583 + }, + { + "epoch": 0.14275238328037154, + "grad_norm": 5.125797271728516, + "learning_rate": 1.9341278799352755e-05, + "loss": 1.0498, + "step": 584 + }, + { + "epoch": 0.14299682229283794, + "grad_norm": 6.8079304695129395, + "learning_rate": 1.9338449875388234e-05, + "loss": 1.2344, + "step": 585 + }, + { + "epoch": 0.14324126130530432, + "grad_norm": 6.831772327423096, + "learning_rate": 1.9335615097713923e-05, + "loss": 1.1498, + "step": 586 + }, + { + "epoch": 0.14348570031777072, + "grad_norm": 3.879077196121216, + "learning_rate": 1.933277446810677e-05, + "loss": 1.0912, + "step": 587 + }, + { + "epoch": 0.1437301393302371, + "grad_norm": 5.787622451782227, + "learning_rate": 1.932992798834739e-05, + "loss": 1.2058, + "step": 588 + }, + { + "epoch": 0.1439745783427035, + "grad_norm": 3.4940860271453857, + "learning_rate": 1.9327075660220076e-05, + "loss": 1.2182, + "step": 589 + }, + { + "epoch": 0.1442190173551699, + "grad_norm": 4.091410160064697, + "learning_rate": 1.9324217485512777e-05, + "loss": 1.0812, + "step": 590 + }, + { + "epoch": 0.14446345636763627, + "grad_norm": 6.517455577850342, + "learning_rate": 1.932135346601711e-05, + "loss": 1.083, + "step": 591 + }, + { + "epoch": 0.14470789538010267, + "grad_norm": 7.640650272369385, + "learning_rate": 1.9318483603528358e-05, + "loss": 1.212, + "step": 592 + }, + { + "epoch": 0.14495233439256905, + "grad_norm": 5.1665472984313965, + "learning_rate": 1.931560789984546e-05, + "loss": 1.1266, + "step": 593 + }, + { + "epoch": 0.14519677340503545, + "grad_norm": 10.023889541625977, + "learning_rate": 1.931272635677102e-05, + "loss": 1.2247, + "step": 594 + }, + { + "epoch": 0.14544121241750183, + "grad_norm": 4.338187217712402, + "learning_rate": 1.9309838976111312e-05, + "loss": 1.0899, + "step": 595 + }, + { + "epoch": 0.14568565142996823, + "grad_norm": 7.103562831878662, + "learning_rate": 1.9306945759676254e-05, + "loss": 1.3005, + "step": 596 + }, + { + "epoch": 0.1459300904424346, + "grad_norm": 6.248996734619141, + "learning_rate": 1.9304046709279426e-05, + "loss": 1.2193, + "step": 597 + }, + { + "epoch": 0.146174529454901, + "grad_norm": 3.4053194522857666, + "learning_rate": 1.9301141826738074e-05, + "loss": 1.1569, + "step": 598 + }, + { + "epoch": 0.1464189684673674, + "grad_norm": 3.0260212421417236, + "learning_rate": 1.9298231113873086e-05, + "loss": 2.0633, + "step": 599 + }, + { + "epoch": 0.14666340747983378, + "grad_norm": 4.802820205688477, + "learning_rate": 1.929531457250902e-05, + "loss": 1.1888, + "step": 600 + }, + { + "epoch": 0.14690784649230018, + "grad_norm": 4.366469383239746, + "learning_rate": 1.9292392204474075e-05, + "loss": 1.1144, + "step": 601 + }, + { + "epoch": 0.14715228550476656, + "grad_norm": 6.060830593109131, + "learning_rate": 1.9289464011600106e-05, + "loss": 1.0817, + "step": 602 + }, + { + "epoch": 0.14739672451723296, + "grad_norm": 10.182448387145996, + "learning_rate": 1.9286529995722624e-05, + "loss": 1.0787, + "step": 603 + }, + { + "epoch": 0.14764116352969933, + "grad_norm": 3.8352510929107666, + "learning_rate": 1.928359015868078e-05, + "loss": 1.1872, + "step": 604 + }, + { + "epoch": 0.14788560254216573, + "grad_norm": 3.9045841693878174, + "learning_rate": 1.928064450231739e-05, + "loss": 1.2423, + "step": 605 + }, + { + "epoch": 0.1481300415546321, + "grad_norm": 4.883039951324463, + "learning_rate": 1.9277693028478895e-05, + "loss": 1.2389, + "step": 606 + }, + { + "epoch": 0.1483744805670985, + "grad_norm": 4.2947773933410645, + "learning_rate": 1.9274735739015405e-05, + "loss": 1.1433, + "step": 607 + }, + { + "epoch": 0.1486189195795649, + "grad_norm": 4.876698970794678, + "learning_rate": 1.9271772635780664e-05, + "loss": 1.1115, + "step": 608 + }, + { + "epoch": 0.1488633585920313, + "grad_norm": 4.491443157196045, + "learning_rate": 1.926880372063206e-05, + "loss": 1.1474, + "step": 609 + }, + { + "epoch": 0.1491077976044977, + "grad_norm": 3.7161436080932617, + "learning_rate": 1.926582899543063e-05, + "loss": 1.1142, + "step": 610 + }, + { + "epoch": 0.14935223661696406, + "grad_norm": 4.214755058288574, + "learning_rate": 1.9262848462041046e-05, + "loss": 1.0922, + "step": 611 + }, + { + "epoch": 0.14959667562943046, + "grad_norm": 9.900253295898438, + "learning_rate": 1.9259862122331627e-05, + "loss": 1.0447, + "step": 612 + }, + { + "epoch": 0.14984111464189684, + "grad_norm": 5.963320732116699, + "learning_rate": 1.9256869978174326e-05, + "loss": 1.0681, + "step": 613 + }, + { + "epoch": 0.15008555365436324, + "grad_norm": 4.601866722106934, + "learning_rate": 1.9253872031444742e-05, + "loss": 0.944, + "step": 614 + }, + { + "epoch": 0.15032999266682961, + "grad_norm": 4.92673921585083, + "learning_rate": 1.9250868284022102e-05, + "loss": 1.0355, + "step": 615 + }, + { + "epoch": 0.15057443167929602, + "grad_norm": 5.88585090637207, + "learning_rate": 1.9247858737789276e-05, + "loss": 2.0564, + "step": 616 + }, + { + "epoch": 0.15081887069176242, + "grad_norm": 9.797651290893555, + "learning_rate": 1.9244843394632765e-05, + "loss": 1.2278, + "step": 617 + }, + { + "epoch": 0.1510633097042288, + "grad_norm": 5.198269844055176, + "learning_rate": 1.9241822256442707e-05, + "loss": 1.172, + "step": 618 + }, + { + "epoch": 0.1513077487166952, + "grad_norm": 3.963874340057373, + "learning_rate": 1.9238795325112867e-05, + "loss": 1.1404, + "step": 619 + }, + { + "epoch": 0.15155218772916157, + "grad_norm": 5.644730091094971, + "learning_rate": 1.9235762602540654e-05, + "loss": 1.0872, + "step": 620 + }, + { + "epoch": 0.15179662674162797, + "grad_norm": 4.740919589996338, + "learning_rate": 1.923272409062709e-05, + "loss": 1.1895, + "step": 621 + }, + { + "epoch": 0.15204106575409435, + "grad_norm": 8.003108024597168, + "learning_rate": 1.922967979127684e-05, + "loss": 1.219, + "step": 622 + }, + { + "epoch": 0.15228550476656075, + "grad_norm": 14.207559585571289, + "learning_rate": 1.9226629706398186e-05, + "loss": 1.1041, + "step": 623 + }, + { + "epoch": 0.15252994377902712, + "grad_norm": 3.9512312412261963, + "learning_rate": 1.9223573837903043e-05, + "loss": 1.246, + "step": 624 + }, + { + "epoch": 0.15277438279149352, + "grad_norm": 5.806921005249023, + "learning_rate": 1.9220512187706956e-05, + "loss": 1.0669, + "step": 625 + }, + { + "epoch": 0.15301882180395993, + "grad_norm": 6.516802787780762, + "learning_rate": 1.921744475772908e-05, + "loss": 1.1033, + "step": 626 + }, + { + "epoch": 0.1532632608164263, + "grad_norm": 6.30975341796875, + "learning_rate": 1.921437154989221e-05, + "loss": 1.1297, + "step": 627 + }, + { + "epoch": 0.1535076998288927, + "grad_norm": 7.9424147605896, + "learning_rate": 1.921129256612275e-05, + "loss": 1.1077, + "step": 628 + }, + { + "epoch": 0.15375213884135908, + "grad_norm": 5.780880928039551, + "learning_rate": 1.9208207808350727e-05, + "loss": 1.1417, + "step": 629 + }, + { + "epoch": 0.15399657785382548, + "grad_norm": 8.841265678405762, + "learning_rate": 1.920511727850979e-05, + "loss": 1.0083, + "step": 630 + }, + { + "epoch": 0.15424101686629185, + "grad_norm": 13.319135665893555, + "learning_rate": 1.920202097853721e-05, + "loss": 1.2188, + "step": 631 + }, + { + "epoch": 0.15448545587875825, + "grad_norm": 5.12705659866333, + "learning_rate": 1.9198918910373864e-05, + "loss": 1.1483, + "step": 632 + }, + { + "epoch": 0.15472989489122463, + "grad_norm": 7.4151835441589355, + "learning_rate": 1.9195811075964253e-05, + "loss": 1.2291, + "step": 633 + }, + { + "epoch": 0.15497433390369103, + "grad_norm": 5.743813514709473, + "learning_rate": 1.919269747725649e-05, + "loss": 1.0871, + "step": 634 + }, + { + "epoch": 0.15521877291615743, + "grad_norm": 6.581920146942139, + "learning_rate": 1.918957811620231e-05, + "loss": 1.1378, + "step": 635 + }, + { + "epoch": 0.1554632119286238, + "grad_norm": 8.724961280822754, + "learning_rate": 1.918645299475704e-05, + "loss": 1.0898, + "step": 636 + }, + { + "epoch": 0.1557076509410902, + "grad_norm": 7.256657600402832, + "learning_rate": 1.9183322114879633e-05, + "loss": 1.015, + "step": 637 + }, + { + "epoch": 0.15595208995355658, + "grad_norm": 5.943508148193359, + "learning_rate": 1.9180185478532655e-05, + "loss": 1.2436, + "step": 638 + }, + { + "epoch": 0.15619652896602298, + "grad_norm": 14.481973648071289, + "learning_rate": 1.9177043087682266e-05, + "loss": 1.0872, + "step": 639 + }, + { + "epoch": 0.15644096797848936, + "grad_norm": 11.159737586975098, + "learning_rate": 1.9173894944298245e-05, + "loss": 1.2359, + "step": 640 + }, + { + "epoch": 0.15668540699095576, + "grad_norm": 7.978509426116943, + "learning_rate": 1.917074105035397e-05, + "loss": 1.1235, + "step": 641 + }, + { + "epoch": 0.15692984600342214, + "grad_norm": 12.450584411621094, + "learning_rate": 1.916758140782643e-05, + "loss": 1.1161, + "step": 642 + }, + { + "epoch": 0.15717428501588854, + "grad_norm": 4.7206525802612305, + "learning_rate": 1.916441601869621e-05, + "loss": 1.1474, + "step": 643 + }, + { + "epoch": 0.1574187240283549, + "grad_norm": 12.776074409484863, + "learning_rate": 1.9161244884947503e-05, + "loss": 1.1997, + "step": 644 + }, + { + "epoch": 0.1576631630408213, + "grad_norm": 14.507882118225098, + "learning_rate": 1.9158068008568102e-05, + "loss": 1.084, + "step": 645 + }, + { + "epoch": 0.15790760205328772, + "grad_norm": 18.697959899902344, + "learning_rate": 1.9154885391549397e-05, + "loss": 1.0914, + "step": 646 + }, + { + "epoch": 0.1581520410657541, + "grad_norm": 5.158618450164795, + "learning_rate": 1.9151697035886384e-05, + "loss": 1.2494, + "step": 647 + }, + { + "epoch": 0.1583964800782205, + "grad_norm": 6.806406497955322, + "learning_rate": 1.914850294357764e-05, + "loss": 1.1049, + "step": 648 + }, + { + "epoch": 0.15864091909068687, + "grad_norm": 13.390410423278809, + "learning_rate": 1.9145303116625357e-05, + "loss": 1.1129, + "step": 649 + }, + { + "epoch": 0.15888535810315327, + "grad_norm": 4.751810550689697, + "learning_rate": 1.914209755703531e-05, + "loss": 1.1234, + "step": 650 + }, + { + "epoch": 0.15912979711561964, + "grad_norm": 6.872396469116211, + "learning_rate": 1.9138886266816868e-05, + "loss": 1.0624, + "step": 651 + }, + { + "epoch": 0.15937423612808604, + "grad_norm": 9.056709289550781, + "learning_rate": 1.9135669247983e-05, + "loss": 1.1375, + "step": 652 + }, + { + "epoch": 0.15961867514055242, + "grad_norm": 8.196880340576172, + "learning_rate": 1.9132446502550257e-05, + "loss": 1.1987, + "step": 653 + }, + { + "epoch": 0.15986311415301882, + "grad_norm": 6.337855815887451, + "learning_rate": 1.9129218032538787e-05, + "loss": 1.1538, + "step": 654 + }, + { + "epoch": 0.16010755316548522, + "grad_norm": 9.169740676879883, + "learning_rate": 1.9125983839972317e-05, + "loss": 1.1671, + "step": 655 + }, + { + "epoch": 0.1603519921779516, + "grad_norm": 4.739553928375244, + "learning_rate": 1.9122743926878173e-05, + "loss": 1.0563, + "step": 656 + }, + { + "epoch": 0.160596431190418, + "grad_norm": 10.157617568969727, + "learning_rate": 1.9119498295287256e-05, + "loss": 1.174, + "step": 657 + }, + { + "epoch": 0.16084087020288437, + "grad_norm": 6.214486598968506, + "learning_rate": 1.911624694723406e-05, + "loss": 1.1447, + "step": 658 + }, + { + "epoch": 0.16108530921535077, + "grad_norm": 4.878256320953369, + "learning_rate": 1.9112989884756655e-05, + "loss": 1.1352, + "step": 659 + }, + { + "epoch": 0.16132974822781715, + "grad_norm": 5.432155609130859, + "learning_rate": 1.9109727109896695e-05, + "loss": 1.2467, + "step": 660 + }, + { + "epoch": 0.16157418724028355, + "grad_norm": 5.435805320739746, + "learning_rate": 1.9106458624699425e-05, + "loss": 1.203, + "step": 661 + }, + { + "epoch": 0.16181862625274993, + "grad_norm": 4.5998759269714355, + "learning_rate": 1.910318443121365e-05, + "loss": 1.1275, + "step": 662 + }, + { + "epoch": 0.16206306526521633, + "grad_norm": 9.564048767089844, + "learning_rate": 1.9099904531491776e-05, + "loss": 1.1447, + "step": 663 + }, + { + "epoch": 0.16230750427768273, + "grad_norm": 6.536534309387207, + "learning_rate": 1.909661892758976e-05, + "loss": 1.1537, + "step": 664 + }, + { + "epoch": 0.1625519432901491, + "grad_norm": 4.478946685791016, + "learning_rate": 1.9093327621567162e-05, + "loss": 1.1499, + "step": 665 + }, + { + "epoch": 0.1627963823026155, + "grad_norm": 8.710973739624023, + "learning_rate": 1.9090030615487093e-05, + "loss": 1.1229, + "step": 666 + }, + { + "epoch": 0.16304082131508188, + "grad_norm": 3.5803980827331543, + "learning_rate": 1.908672791141625e-05, + "loss": 2.077, + "step": 667 + }, + { + "epoch": 0.16328526032754828, + "grad_norm": 7.929402828216553, + "learning_rate": 1.90834195114249e-05, + "loss": 1.1464, + "step": 668 + }, + { + "epoch": 0.16352969934001466, + "grad_norm": 5.926062107086182, + "learning_rate": 1.9080105417586878e-05, + "loss": 1.196, + "step": 669 + }, + { + "epoch": 0.16377413835248106, + "grad_norm": 4.997158050537109, + "learning_rate": 1.9076785631979592e-05, + "loss": 1.192, + "step": 670 + }, + { + "epoch": 0.16401857736494743, + "grad_norm": 4.977759838104248, + "learning_rate": 1.907346015668401e-05, + "loss": 1.0964, + "step": 671 + }, + { + "epoch": 0.16426301637741383, + "grad_norm": 17.051420211791992, + "learning_rate": 1.907012899378467e-05, + "loss": 1.0237, + "step": 672 + }, + { + "epoch": 0.16450745538988024, + "grad_norm": 5.6634297370910645, + "learning_rate": 1.9066792145369685e-05, + "loss": 1.1316, + "step": 673 + }, + { + "epoch": 0.1647518944023466, + "grad_norm": 4.748595714569092, + "learning_rate": 1.9063449613530712e-05, + "loss": 1.0597, + "step": 674 + }, + { + "epoch": 0.164996333414813, + "grad_norm": 7.550394058227539, + "learning_rate": 1.9060101400362998e-05, + "loss": 1.0291, + "step": 675 + }, + { + "epoch": 0.1652407724272794, + "grad_norm": 7.80336856842041, + "learning_rate": 1.905674750796532e-05, + "loss": 1.1255, + "step": 676 + }, + { + "epoch": 0.1654852114397458, + "grad_norm": 4.666580677032471, + "learning_rate": 1.905338793844004e-05, + "loss": 1.0285, + "step": 677 + }, + { + "epoch": 0.16572965045221216, + "grad_norm": 4.3687896728515625, + "learning_rate": 1.9050022693893062e-05, + "loss": 1.1289, + "step": 678 + }, + { + "epoch": 0.16597408946467856, + "grad_norm": 3.923128366470337, + "learning_rate": 1.9046651776433863e-05, + "loss": 1.1788, + "step": 679 + }, + { + "epoch": 0.16621852847714494, + "grad_norm": 5.142563343048096, + "learning_rate": 1.904327518817546e-05, + "loss": 1.0525, + "step": 680 + }, + { + "epoch": 0.16646296748961134, + "grad_norm": 4.094109535217285, + "learning_rate": 1.9039892931234434e-05, + "loss": 1.2607, + "step": 681 + }, + { + "epoch": 0.16670740650207774, + "grad_norm": 4.689309120178223, + "learning_rate": 1.903650500773092e-05, + "loss": 0.9916, + "step": 682 + }, + { + "epoch": 0.16695184551454412, + "grad_norm": 7.4928693771362305, + "learning_rate": 1.90331114197886e-05, + "loss": 1.1558, + "step": 683 + }, + { + "epoch": 0.16719628452701052, + "grad_norm": 7.847235679626465, + "learning_rate": 1.902971216953471e-05, + "loss": 1.105, + "step": 684 + }, + { + "epoch": 0.1674407235394769, + "grad_norm": 4.171519756317139, + "learning_rate": 1.9026307259100037e-05, + "loss": 1.0784, + "step": 685 + }, + { + "epoch": 0.1676851625519433, + "grad_norm": 4.570870876312256, + "learning_rate": 1.9022896690618908e-05, + "loss": 1.118, + "step": 686 + }, + { + "epoch": 0.16792960156440967, + "grad_norm": 4.4747748374938965, + "learning_rate": 1.9019480466229203e-05, + "loss": 1.0661, + "step": 687 + }, + { + "epoch": 0.16817404057687607, + "grad_norm": 4.1525959968566895, + "learning_rate": 1.9016058588072354e-05, + "loss": 1.0394, + "step": 688 + }, + { + "epoch": 0.16841847958934245, + "grad_norm": 4.377383232116699, + "learning_rate": 1.9012631058293322e-05, + "loss": 1.137, + "step": 689 + }, + { + "epoch": 0.16866291860180885, + "grad_norm": 4.25969123840332, + "learning_rate": 1.9009197879040617e-05, + "loss": 1.1238, + "step": 690 + }, + { + "epoch": 0.16890735761427525, + "grad_norm": 4.3925299644470215, + "learning_rate": 1.9005759052466303e-05, + "loss": 1.0514, + "step": 691 + }, + { + "epoch": 0.16915179662674162, + "grad_norm": 6.636261940002441, + "learning_rate": 1.900231458072596e-05, + "loss": 1.0674, + "step": 692 + }, + { + "epoch": 0.16939623563920803, + "grad_norm": 5.294342517852783, + "learning_rate": 1.899886446597872e-05, + "loss": 1.1533, + "step": 693 + }, + { + "epoch": 0.1696406746516744, + "grad_norm": 4.914679527282715, + "learning_rate": 1.899540871038726e-05, + "loss": 1.0943, + "step": 694 + }, + { + "epoch": 0.1698851136641408, + "grad_norm": 4.830330848693848, + "learning_rate": 1.899194731611778e-05, + "loss": 1.0313, + "step": 695 + }, + { + "epoch": 0.17012955267660718, + "grad_norm": 4.74468994140625, + "learning_rate": 1.8988480285340018e-05, + "loss": 1.1535, + "step": 696 + }, + { + "epoch": 0.17037399168907358, + "grad_norm": 6.699695110321045, + "learning_rate": 1.8985007620227246e-05, + "loss": 1.0901, + "step": 697 + }, + { + "epoch": 0.17061843070153995, + "grad_norm": 7.729903697967529, + "learning_rate": 1.898152932295627e-05, + "loss": 1.0657, + "step": 698 + }, + { + "epoch": 0.17086286971400635, + "grad_norm": 4.743852138519287, + "learning_rate": 1.897804539570742e-05, + "loss": 1.0182, + "step": 699 + }, + { + "epoch": 0.17110730872647276, + "grad_norm": 4.740683555603027, + "learning_rate": 1.897455584066456e-05, + "loss": 1.0612, + "step": 700 + }, + { + "epoch": 0.17135174773893913, + "grad_norm": 3.3871266841888428, + "learning_rate": 1.897106066001509e-05, + "loss": 1.0377, + "step": 701 + }, + { + "epoch": 0.17159618675140553, + "grad_norm": 5.6066083908081055, + "learning_rate": 1.8967559855949913e-05, + "loss": 1.2205, + "step": 702 + }, + { + "epoch": 0.1718406257638719, + "grad_norm": 7.7547101974487305, + "learning_rate": 1.8964053430663485e-05, + "loss": 1.1098, + "step": 703 + }, + { + "epoch": 0.1720850647763383, + "grad_norm": 3.957638740539551, + "learning_rate": 1.8960541386353765e-05, + "loss": 1.0606, + "step": 704 + }, + { + "epoch": 0.17232950378880468, + "grad_norm": 5.753738880157471, + "learning_rate": 1.8957023725222238e-05, + "loss": 1.081, + "step": 705 + }, + { + "epoch": 0.17257394280127109, + "grad_norm": 5.259015083312988, + "learning_rate": 1.8953500449473924e-05, + "loss": 1.0121, + "step": 706 + }, + { + "epoch": 0.17281838181373746, + "grad_norm": 4.496668338775635, + "learning_rate": 1.894997156131734e-05, + "loss": 1.0718, + "step": 707 + }, + { + "epoch": 0.17306282082620386, + "grad_norm": 14.881460189819336, + "learning_rate": 1.8946437062964543e-05, + "loss": 1.1767, + "step": 708 + }, + { + "epoch": 0.17330725983867026, + "grad_norm": 5.021279811859131, + "learning_rate": 1.894289695663109e-05, + "loss": 1.1781, + "step": 709 + }, + { + "epoch": 0.17355169885113664, + "grad_norm": 5.705527305603027, + "learning_rate": 1.893935124453606e-05, + "loss": 1.0158, + "step": 710 + }, + { + "epoch": 0.17379613786360304, + "grad_norm": 4.499573230743408, + "learning_rate": 1.8935799928902046e-05, + "loss": 1.168, + "step": 711 + }, + { + "epoch": 0.1740405768760694, + "grad_norm": 4.802264213562012, + "learning_rate": 1.8932243011955154e-05, + "loss": 1.1318, + "step": 712 + }, + { + "epoch": 0.17428501588853582, + "grad_norm": 12.059589385986328, + "learning_rate": 1.8928680495925e-05, + "loss": 0.9681, + "step": 713 + }, + { + "epoch": 0.1745294549010022, + "grad_norm": 5.503899574279785, + "learning_rate": 1.8925112383044713e-05, + "loss": 1.1116, + "step": 714 + }, + { + "epoch": 0.1747738939134686, + "grad_norm": 5.018351078033447, + "learning_rate": 1.8921538675550923e-05, + "loss": 1.098, + "step": 715 + }, + { + "epoch": 0.17501833292593497, + "grad_norm": 6.688838958740234, + "learning_rate": 1.891795937568377e-05, + "loss": 1.1197, + "step": 716 + }, + { + "epoch": 0.17526277193840137, + "grad_norm": 13.338314056396484, + "learning_rate": 1.8914374485686905e-05, + "loss": 1.0497, + "step": 717 + }, + { + "epoch": 0.17550721095086777, + "grad_norm": 11.007012367248535, + "learning_rate": 1.8910784007807477e-05, + "loss": 1.1531, + "step": 718 + }, + { + "epoch": 0.17575164996333414, + "grad_norm": 9.250398635864258, + "learning_rate": 1.890718794429614e-05, + "loss": 1.047, + "step": 719 + }, + { + "epoch": 0.17599608897580055, + "grad_norm": 12.425487518310547, + "learning_rate": 1.8903586297407052e-05, + "loss": 1.0937, + "step": 720 + }, + { + "epoch": 0.17624052798826692, + "grad_norm": 5.152442932128906, + "learning_rate": 1.8899979069397858e-05, + "loss": 1.1518, + "step": 721 + }, + { + "epoch": 0.17648496700073332, + "grad_norm": 6.493432998657227, + "learning_rate": 1.8896366262529718e-05, + "loss": 1.0677, + "step": 722 + }, + { + "epoch": 0.1767294060131997, + "grad_norm": 4.106019020080566, + "learning_rate": 1.8892747879067284e-05, + "loss": 1.1456, + "step": 723 + }, + { + "epoch": 0.1769738450256661, + "grad_norm": 8.680675506591797, + "learning_rate": 1.8889123921278703e-05, + "loss": 1.1829, + "step": 724 + }, + { + "epoch": 0.17721828403813247, + "grad_norm": 3.9373128414154053, + "learning_rate": 1.8885494391435607e-05, + "loss": 1.069, + "step": 725 + }, + { + "epoch": 0.17746272305059888, + "grad_norm": 4.0960822105407715, + "learning_rate": 1.8881859291813137e-05, + "loss": 1.258, + "step": 726 + }, + { + "epoch": 0.17770716206306528, + "grad_norm": 7.045689105987549, + "learning_rate": 1.8878218624689913e-05, + "loss": 1.172, + "step": 727 + }, + { + "epoch": 0.17795160107553165, + "grad_norm": 4.111807823181152, + "learning_rate": 1.887457239234805e-05, + "loss": 1.1729, + "step": 728 + }, + { + "epoch": 0.17819604008799805, + "grad_norm": 4.325777053833008, + "learning_rate": 1.887092059707315e-05, + "loss": 1.1321, + "step": 729 + }, + { + "epoch": 0.17844047910046443, + "grad_norm": 4.5247578620910645, + "learning_rate": 1.8867263241154302e-05, + "loss": 1.0469, + "step": 730 + }, + { + "epoch": 0.17868491811293083, + "grad_norm": 3.589571237564087, + "learning_rate": 1.8863600326884085e-05, + "loss": 1.198, + "step": 731 + }, + { + "epoch": 0.1789293571253972, + "grad_norm": 3.886864185333252, + "learning_rate": 1.8859931856558555e-05, + "loss": 1.0563, + "step": 732 + }, + { + "epoch": 0.1791737961378636, + "grad_norm": 4.039812088012695, + "learning_rate": 1.8856257832477255e-05, + "loss": 1.0757, + "step": 733 + }, + { + "epoch": 0.17941823515032998, + "grad_norm": 4.089430809020996, + "learning_rate": 1.8852578256943212e-05, + "loss": 1.1413, + "step": 734 + }, + { + "epoch": 0.17966267416279638, + "grad_norm": 4.228828430175781, + "learning_rate": 1.8848893132262927e-05, + "loss": 1.2184, + "step": 735 + }, + { + "epoch": 0.17990711317526278, + "grad_norm": 6.709991931915283, + "learning_rate": 1.884520246074638e-05, + "loss": 1.2658, + "step": 736 + }, + { + "epoch": 0.18015155218772916, + "grad_norm": 5.3593430519104, + "learning_rate": 1.884150624470703e-05, + "loss": 1.1778, + "step": 737 + }, + { + "epoch": 0.18039599120019556, + "grad_norm": 4.252145767211914, + "learning_rate": 1.8837804486461818e-05, + "loss": 1.0652, + "step": 738 + }, + { + "epoch": 0.18064043021266193, + "grad_norm": 4.08242130279541, + "learning_rate": 1.8834097188331143e-05, + "loss": 1.1207, + "step": 739 + }, + { + "epoch": 0.18088486922512834, + "grad_norm": 6.409980297088623, + "learning_rate": 1.8830384352638896e-05, + "loss": 1.1498, + "step": 740 + }, + { + "epoch": 0.1811293082375947, + "grad_norm": 5.007962226867676, + "learning_rate": 1.882666598171242e-05, + "loss": 1.1071, + "step": 741 + }, + { + "epoch": 0.1813737472500611, + "grad_norm": 3.8679728507995605, + "learning_rate": 1.8822942077882542e-05, + "loss": 1.0612, + "step": 742 + }, + { + "epoch": 0.1816181862625275, + "grad_norm": 4.374696731567383, + "learning_rate": 1.881921264348355e-05, + "loss": 1.1656, + "step": 743 + }, + { + "epoch": 0.1818626252749939, + "grad_norm": 3.6539933681488037, + "learning_rate": 1.8815477680853208e-05, + "loss": 1.0958, + "step": 744 + }, + { + "epoch": 0.1821070642874603, + "grad_norm": 3.358370542526245, + "learning_rate": 1.8811737192332725e-05, + "loss": 1.0235, + "step": 745 + }, + { + "epoch": 0.18235150329992666, + "grad_norm": 4.081916332244873, + "learning_rate": 1.8807991180266798e-05, + "loss": 1.0765, + "step": 746 + }, + { + "epoch": 0.18259594231239307, + "grad_norm": 4.941745758056641, + "learning_rate": 1.8804239647003574e-05, + "loss": 1.1526, + "step": 747 + }, + { + "epoch": 0.18284038132485944, + "grad_norm": 4.099913597106934, + "learning_rate": 1.880048259489466e-05, + "loss": 1.2486, + "step": 748 + }, + { + "epoch": 0.18308482033732584, + "grad_norm": 3.721834897994995, + "learning_rate": 1.8796720026295122e-05, + "loss": 1.1316, + "step": 749 + }, + { + "epoch": 0.18332925934979222, + "grad_norm": 3.963992118835449, + "learning_rate": 1.8792951943563492e-05, + "loss": 0.9799, + "step": 750 + }, + { + "epoch": 0.18357369836225862, + "grad_norm": 3.802098274230957, + "learning_rate": 1.8789178349061755e-05, + "loss": 1.1526, + "step": 751 + }, + { + "epoch": 0.183818137374725, + "grad_norm": 3.5054547786712646, + "learning_rate": 1.8785399245155345e-05, + "loss": 1.0229, + "step": 752 + }, + { + "epoch": 0.1840625763871914, + "grad_norm": 3.7871315479278564, + "learning_rate": 1.8781614634213153e-05, + "loss": 1.0775, + "step": 753 + }, + { + "epoch": 0.1843070153996578, + "grad_norm": 3.387770891189575, + "learning_rate": 1.8777824518607527e-05, + "loss": 1.0812, + "step": 754 + }, + { + "epoch": 0.18455145441212417, + "grad_norm": 3.360201120376587, + "learning_rate": 1.8774028900714257e-05, + "loss": 1.0478, + "step": 755 + }, + { + "epoch": 0.18479589342459057, + "grad_norm": 5.499178409576416, + "learning_rate": 1.877022778291259e-05, + "loss": 1.1209, + "step": 756 + }, + { + "epoch": 0.18504033243705695, + "grad_norm": 3.671175718307495, + "learning_rate": 1.8766421167585216e-05, + "loss": 1.1505, + "step": 757 + }, + { + "epoch": 0.18528477144952335, + "grad_norm": 4.1667256355285645, + "learning_rate": 1.8762609057118268e-05, + "loss": 1.1308, + "step": 758 + }, + { + "epoch": 0.18552921046198972, + "grad_norm": 4.283543109893799, + "learning_rate": 1.8758791453901337e-05, + "loss": 1.1366, + "step": 759 + }, + { + "epoch": 0.18577364947445613, + "grad_norm": 5.065645217895508, + "learning_rate": 1.8754968360327438e-05, + "loss": 1.2085, + "step": 760 + }, + { + "epoch": 0.1860180884869225, + "grad_norm": 5.341877460479736, + "learning_rate": 1.8751139778793043e-05, + "loss": 1.1089, + "step": 761 + }, + { + "epoch": 0.1862625274993889, + "grad_norm": 4.320544242858887, + "learning_rate": 1.8747305711698058e-05, + "loss": 1.1308, + "step": 762 + }, + { + "epoch": 0.1865069665118553, + "grad_norm": 3.871718168258667, + "learning_rate": 1.8743466161445823e-05, + "loss": 1.0672, + "step": 763 + }, + { + "epoch": 0.18675140552432168, + "grad_norm": 4.854724884033203, + "learning_rate": 1.8739621130443125e-05, + "loss": 1.0795, + "step": 764 + }, + { + "epoch": 0.18699584453678808, + "grad_norm": 4.116591930389404, + "learning_rate": 1.8735770621100182e-05, + "loss": 1.1431, + "step": 765 + }, + { + "epoch": 0.18724028354925445, + "grad_norm": 11.736309051513672, + "learning_rate": 1.8731914635830643e-05, + "loss": 1.0883, + "step": 766 + }, + { + "epoch": 0.18748472256172086, + "grad_norm": 3.832092046737671, + "learning_rate": 1.872805317705159e-05, + "loss": 1.1814, + "step": 767 + }, + { + "epoch": 0.18772916157418723, + "grad_norm": 4.503589153289795, + "learning_rate": 1.872418624718354e-05, + "loss": 1.1775, + "step": 768 + }, + { + "epoch": 0.18797360058665363, + "grad_norm": 3.3327720165252686, + "learning_rate": 1.8720313848650448e-05, + "loss": 1.158, + "step": 769 + }, + { + "epoch": 0.18821803959912, + "grad_norm": 6.3881754875183105, + "learning_rate": 1.871643598387967e-05, + "loss": 1.1307, + "step": 770 + }, + { + "epoch": 0.1884624786115864, + "grad_norm": 4.862054824829102, + "learning_rate": 1.871255265530201e-05, + "loss": 1.1961, + "step": 771 + }, + { + "epoch": 0.1887069176240528, + "grad_norm": 3.6286745071411133, + "learning_rate": 1.8708663865351696e-05, + "loss": 0.9553, + "step": 772 + }, + { + "epoch": 0.18895135663651919, + "grad_norm": 3.490654706954956, + "learning_rate": 1.8704769616466377e-05, + "loss": 1.1614, + "step": 773 + }, + { + "epoch": 0.1891957956489856, + "grad_norm": 3.3287625312805176, + "learning_rate": 1.8700869911087115e-05, + "loss": 2.1641, + "step": 774 + }, + { + "epoch": 0.18944023466145196, + "grad_norm": 4.803013324737549, + "learning_rate": 1.8696964751658405e-05, + "loss": 1.0372, + "step": 775 + }, + { + "epoch": 0.18968467367391836, + "grad_norm": 3.6098735332489014, + "learning_rate": 1.8693054140628154e-05, + "loss": 1.0666, + "step": 776 + }, + { + "epoch": 0.18992911268638474, + "grad_norm": 4.2922749519348145, + "learning_rate": 1.8689138080447687e-05, + "loss": 1.1183, + "step": 777 + }, + { + "epoch": 0.19017355169885114, + "grad_norm": 4.80598258972168, + "learning_rate": 1.8685216573571744e-05, + "loss": 1.1296, + "step": 778 + }, + { + "epoch": 0.19041799071131751, + "grad_norm": 3.6623916625976562, + "learning_rate": 1.8681289622458485e-05, + "loss": 1.0789, + "step": 779 + }, + { + "epoch": 0.19066242972378392, + "grad_norm": 4.169750690460205, + "learning_rate": 1.8677357229569474e-05, + "loss": 1.1334, + "step": 780 + }, + { + "epoch": 0.19090686873625032, + "grad_norm": 5.41149377822876, + "learning_rate": 1.8673419397369693e-05, + "loss": 1.1136, + "step": 781 + }, + { + "epoch": 0.1911513077487167, + "grad_norm": 3.4953672885894775, + "learning_rate": 1.8669476128327534e-05, + "loss": 2.1952, + "step": 782 + }, + { + "epoch": 0.1913957467611831, + "grad_norm": 4.066995620727539, + "learning_rate": 1.8665527424914786e-05, + "loss": 1.1636, + "step": 783 + }, + { + "epoch": 0.19164018577364947, + "grad_norm": 4.137174129486084, + "learning_rate": 1.8661573289606657e-05, + "loss": 1.0459, + "step": 784 + }, + { + "epoch": 0.19188462478611587, + "grad_norm": 4.325610637664795, + "learning_rate": 1.8657613724881755e-05, + "loss": 0.9766, + "step": 785 + }, + { + "epoch": 0.19212906379858224, + "grad_norm": 3.6171977519989014, + "learning_rate": 1.8653648733222096e-05, + "loss": 1.042, + "step": 786 + }, + { + "epoch": 0.19237350281104865, + "grad_norm": 3.619912624359131, + "learning_rate": 1.8649678317113084e-05, + "loss": 1.1192, + "step": 787 + }, + { + "epoch": 0.19261794182351502, + "grad_norm": 4.169992923736572, + "learning_rate": 1.864570247904354e-05, + "loss": 1.1468, + "step": 788 + }, + { + "epoch": 0.19286238083598142, + "grad_norm": 4.603641510009766, + "learning_rate": 1.8641721221505674e-05, + "loss": 1.1097, + "step": 789 + }, + { + "epoch": 0.19310681984844782, + "grad_norm": 4.259536266326904, + "learning_rate": 1.86377345469951e-05, + "loss": 1.1036, + "step": 790 + }, + { + "epoch": 0.1933512588609142, + "grad_norm": 3.3828346729278564, + "learning_rate": 1.863374245801082e-05, + "loss": 1.0612, + "step": 791 + }, + { + "epoch": 0.1935956978733806, + "grad_norm": 4.6406941413879395, + "learning_rate": 1.862974495705523e-05, + "loss": 0.9907, + "step": 792 + }, + { + "epoch": 0.19384013688584698, + "grad_norm": 3.43007755279541, + "learning_rate": 1.862574204663413e-05, + "loss": 1.0577, + "step": 793 + }, + { + "epoch": 0.19408457589831338, + "grad_norm": 8.210151672363281, + "learning_rate": 1.8621733729256694e-05, + "loss": 1.052, + "step": 794 + }, + { + "epoch": 0.19432901491077975, + "grad_norm": 3.7519443035125732, + "learning_rate": 1.8617720007435497e-05, + "loss": 1.1144, + "step": 795 + }, + { + "epoch": 0.19457345392324615, + "grad_norm": 5.4279069900512695, + "learning_rate": 1.8613700883686503e-05, + "loss": 1.1851, + "step": 796 + }, + { + "epoch": 0.19481789293571253, + "grad_norm": 4.4938130378723145, + "learning_rate": 1.860967636052905e-05, + "loss": 1.1729, + "step": 797 + }, + { + "epoch": 0.19506233194817893, + "grad_norm": 3.714663028717041, + "learning_rate": 1.8605646440485877e-05, + "loss": 1.0932, + "step": 798 + }, + { + "epoch": 0.19530677096064533, + "grad_norm": 3.895505666732788, + "learning_rate": 1.860161112608309e-05, + "loss": 1.1341, + "step": 799 + }, + { + "epoch": 0.1955512099731117, + "grad_norm": 5.3650641441345215, + "learning_rate": 1.8597570419850184e-05, + "loss": 1.0481, + "step": 800 + }, + { + "epoch": 0.1957956489855781, + "grad_norm": 4.210098743438721, + "learning_rate": 1.8593524324320035e-05, + "loss": 1.1164, + "step": 801 + }, + { + "epoch": 0.19604008799804448, + "grad_norm": 4.8966851234436035, + "learning_rate": 1.85894728420289e-05, + "loss": 1.1498, + "step": 802 + }, + { + "epoch": 0.19628452701051088, + "grad_norm": 4.228818416595459, + "learning_rate": 1.8585415975516405e-05, + "loss": 1.0885, + "step": 803 + }, + { + "epoch": 0.19652896602297726, + "grad_norm": 3.544618606567383, + "learning_rate": 1.858135372732555e-05, + "loss": 1.1276, + "step": 804 + }, + { + "epoch": 0.19677340503544366, + "grad_norm": 3.545396327972412, + "learning_rate": 1.8577286100002723e-05, + "loss": 1.0031, + "step": 805 + }, + { + "epoch": 0.19701784404791003, + "grad_norm": 3.896831512451172, + "learning_rate": 1.8573213096097666e-05, + "loss": 0.9596, + "step": 806 + }, + { + "epoch": 0.19726228306037644, + "grad_norm": 5.555988788604736, + "learning_rate": 1.8569134718163496e-05, + "loss": 1.1155, + "step": 807 + }, + { + "epoch": 0.19750672207284284, + "grad_norm": 4.680154800415039, + "learning_rate": 1.8565050968756714e-05, + "loss": 1.1159, + "step": 808 + }, + { + "epoch": 0.1977511610853092, + "grad_norm": 3.59993314743042, + "learning_rate": 1.8560961850437163e-05, + "loss": 2.0246, + "step": 809 + }, + { + "epoch": 0.19799560009777561, + "grad_norm": 6.578675746917725, + "learning_rate": 1.8556867365768073e-05, + "loss": 1.057, + "step": 810 + }, + { + "epoch": 0.198240039110242, + "grad_norm": 4.831800937652588, + "learning_rate": 1.855276751731602e-05, + "loss": 1.0949, + "step": 811 + }, + { + "epoch": 0.1984844781227084, + "grad_norm": 4.584733963012695, + "learning_rate": 1.854866230765096e-05, + "loss": 1.1775, + "step": 812 + }, + { + "epoch": 0.19872891713517477, + "grad_norm": 3.855050563812256, + "learning_rate": 1.85445517393462e-05, + "loss": 1.1027, + "step": 813 + }, + { + "epoch": 0.19897335614764117, + "grad_norm": 4.348968029022217, + "learning_rate": 1.85404358149784e-05, + "loss": 1.091, + "step": 814 + }, + { + "epoch": 0.19921779516010754, + "grad_norm": 5.458604335784912, + "learning_rate": 1.8536314537127586e-05, + "loss": 1.0569, + "step": 815 + }, + { + "epoch": 0.19946223417257394, + "grad_norm": 4.697046279907227, + "learning_rate": 1.8532187908377146e-05, + "loss": 1.1723, + "step": 816 + }, + { + "epoch": 0.19970667318504035, + "grad_norm": 3.4961538314819336, + "learning_rate": 1.8528055931313803e-05, + "loss": 2.107, + "step": 817 + }, + { + "epoch": 0.19995111219750672, + "grad_norm": 13.489090919494629, + "learning_rate": 1.8523918608527653e-05, + "loss": 1.0733, + "step": 818 + }, + { + "epoch": 0.20019555120997312, + "grad_norm": 3.1647040843963623, + "learning_rate": 1.8519775942612128e-05, + "loss": 1.061, + "step": 819 + }, + { + "epoch": 0.2004399902224395, + "grad_norm": 5.267470359802246, + "learning_rate": 1.8515627936164015e-05, + "loss": 1.07, + "step": 820 + }, + { + "epoch": 0.2006844292349059, + "grad_norm": 4.1006951332092285, + "learning_rate": 1.8511474591783454e-05, + "loss": 1.1726, + "step": 821 + }, + { + "epoch": 0.20092886824737227, + "grad_norm": 3.6335432529449463, + "learning_rate": 1.850731591207392e-05, + "loss": 1.0347, + "step": 822 + }, + { + "epoch": 0.20117330725983867, + "grad_norm": 4.431300163269043, + "learning_rate": 1.8503151899642237e-05, + "loss": 2.1155, + "step": 823 + }, + { + "epoch": 0.20141774627230505, + "grad_norm": 6.050295829772949, + "learning_rate": 1.8498982557098578e-05, + "loss": 1.1272, + "step": 824 + }, + { + "epoch": 0.20166218528477145, + "grad_norm": 4.173649311065674, + "learning_rate": 1.849480788705645e-05, + "loss": 1.1039, + "step": 825 + }, + { + "epoch": 0.20190662429723785, + "grad_norm": 3.55676531791687, + "learning_rate": 1.8490627892132697e-05, + "loss": 1.1069, + "step": 826 + }, + { + "epoch": 0.20215106330970423, + "grad_norm": 3.58681583404541, + "learning_rate": 1.848644257494751e-05, + "loss": 0.963, + "step": 827 + }, + { + "epoch": 0.20239550232217063, + "grad_norm": 3.541001558303833, + "learning_rate": 1.8482251938124412e-05, + "loss": 1.0795, + "step": 828 + }, + { + "epoch": 0.202639941334637, + "grad_norm": 4.774433135986328, + "learning_rate": 1.8478055984290258e-05, + "loss": 1.0283, + "step": 829 + }, + { + "epoch": 0.2028843803471034, + "grad_norm": 3.2330007553100586, + "learning_rate": 1.8473854716075233e-05, + "loss": 2.0566, + "step": 830 + }, + { + "epoch": 0.20312881935956978, + "grad_norm": 10.420342445373535, + "learning_rate": 1.8469648136112867e-05, + "loss": 1.0285, + "step": 831 + }, + { + "epoch": 0.20337325837203618, + "grad_norm": 9.658334732055664, + "learning_rate": 1.8465436247040005e-05, + "loss": 1.0787, + "step": 832 + }, + { + "epoch": 0.20361769738450256, + "grad_norm": 6.110720634460449, + "learning_rate": 1.846121905149683e-05, + "loss": 1.0487, + "step": 833 + }, + { + "epoch": 0.20386213639696896, + "grad_norm": 3.708970785140991, + "learning_rate": 1.845699655212684e-05, + "loss": 1.1554, + "step": 834 + }, + { + "epoch": 0.20410657540943536, + "grad_norm": 5.930639266967773, + "learning_rate": 1.845276875157687e-05, + "loss": 1.0547, + "step": 835 + }, + { + "epoch": 0.20435101442190173, + "grad_norm": 3.6849205493927, + "learning_rate": 1.8448535652497073e-05, + "loss": 1.0065, + "step": 836 + }, + { + "epoch": 0.20459545343436814, + "grad_norm": 3.9739530086517334, + "learning_rate": 1.844429725754092e-05, + "loss": 1.0815, + "step": 837 + }, + { + "epoch": 0.2048398924468345, + "grad_norm": 5.715768337249756, + "learning_rate": 1.8440053569365204e-05, + "loss": 1.0391, + "step": 838 + }, + { + "epoch": 0.2050843314593009, + "grad_norm": 5.876922130584717, + "learning_rate": 1.8435804590630045e-05, + "loss": 1.0286, + "step": 839 + }, + { + "epoch": 0.20532877047176729, + "grad_norm": 4.035728931427002, + "learning_rate": 1.8431550323998862e-05, + "loss": 1.1854, + "step": 840 + }, + { + "epoch": 0.2055732094842337, + "grad_norm": 3.5158615112304688, + "learning_rate": 1.8427290772138397e-05, + "loss": 1.0352, + "step": 841 + }, + { + "epoch": 0.20581764849670006, + "grad_norm": 5.412172794342041, + "learning_rate": 1.8423025937718712e-05, + "loss": 1.0675, + "step": 842 + }, + { + "epoch": 0.20606208750916646, + "grad_norm": 4.868218421936035, + "learning_rate": 1.841875582341317e-05, + "loss": 1.1832, + "step": 843 + }, + { + "epoch": 0.20630652652163287, + "grad_norm": 3.841078281402588, + "learning_rate": 1.841448043189845e-05, + "loss": 1.1062, + "step": 844 + }, + { + "epoch": 0.20655096553409924, + "grad_norm": 4.0944743156433105, + "learning_rate": 1.8410199765854532e-05, + "loss": 1.0201, + "step": 845 + }, + { + "epoch": 0.20679540454656564, + "grad_norm": 10.717774391174316, + "learning_rate": 1.8405913827964713e-05, + "loss": 1.1188, + "step": 846 + }, + { + "epoch": 0.20703984355903202, + "grad_norm": 4.635489463806152, + "learning_rate": 1.840162262091558e-05, + "loss": 1.0325, + "step": 847 + }, + { + "epoch": 0.20728428257149842, + "grad_norm": 6.215915679931641, + "learning_rate": 1.8397326147397043e-05, + "loss": 1.0543, + "step": 848 + }, + { + "epoch": 0.2075287215839648, + "grad_norm": 4.547093391418457, + "learning_rate": 1.839302441010229e-05, + "loss": 1.1416, + "step": 849 + }, + { + "epoch": 0.2077731605964312, + "grad_norm": 3.299689531326294, + "learning_rate": 1.838871741172783e-05, + "loss": 1.1036, + "step": 850 + }, + { + "epoch": 0.20801759960889757, + "grad_norm": 3.8489363193511963, + "learning_rate": 1.838440515497345e-05, + "loss": 1.1148, + "step": 851 + }, + { + "epoch": 0.20826203862136397, + "grad_norm": 3.808192014694214, + "learning_rate": 1.838008764254225e-05, + "loss": 1.1627, + "step": 852 + }, + { + "epoch": 0.20850647763383037, + "grad_norm": 4.016510009765625, + "learning_rate": 1.8375764877140617e-05, + "loss": 1.0187, + "step": 853 + }, + { + "epoch": 0.20875091664629675, + "grad_norm": 14.063336372375488, + "learning_rate": 1.837143686147823e-05, + "loss": 1.083, + "step": 854 + }, + { + "epoch": 0.20899535565876315, + "grad_norm": 5.032708168029785, + "learning_rate": 1.836710359826806e-05, + "loss": 1.1282, + "step": 855 + }, + { + "epoch": 0.20923979467122952, + "grad_norm": 5.023953914642334, + "learning_rate": 1.836276509022637e-05, + "loss": 1.0746, + "step": 856 + }, + { + "epoch": 0.20948423368369593, + "grad_norm": 4.224484920501709, + "learning_rate": 1.8358421340072704e-05, + "loss": 1.0772, + "step": 857 + }, + { + "epoch": 0.2097286726961623, + "grad_norm": 4.803196430206299, + "learning_rate": 1.83540723505299e-05, + "loss": 2.1353, + "step": 858 + }, + { + "epoch": 0.2099731117086287, + "grad_norm": 4.721566200256348, + "learning_rate": 1.8349718124324075e-05, + "loss": 0.9563, + "step": 859 + }, + { + "epoch": 0.21021755072109508, + "grad_norm": 3.717606782913208, + "learning_rate": 1.834535866418463e-05, + "loss": 1.1166, + "step": 860 + }, + { + "epoch": 0.21046198973356148, + "grad_norm": 4.486362934112549, + "learning_rate": 1.8340993972844252e-05, + "loss": 1.0245, + "step": 861 + }, + { + "epoch": 0.21070642874602788, + "grad_norm": 3.6472747325897217, + "learning_rate": 1.8336624053038898e-05, + "loss": 0.9977, + "step": 862 + }, + { + "epoch": 0.21095086775849425, + "grad_norm": 6.7217230796813965, + "learning_rate": 1.8332248907507808e-05, + "loss": 1.1201, + "step": 863 + }, + { + "epoch": 0.21119530677096066, + "grad_norm": 5.946406841278076, + "learning_rate": 1.8327868538993492e-05, + "loss": 1.1357, + "step": 864 + }, + { + "epoch": 0.21143974578342703, + "grad_norm": 3.6695261001586914, + "learning_rate": 1.8323482950241743e-05, + "loss": 1.0761, + "step": 865 + }, + { + "epoch": 0.21168418479589343, + "grad_norm": 3.1539971828460693, + "learning_rate": 1.831909214400162e-05, + "loss": 1.0073, + "step": 866 + }, + { + "epoch": 0.2119286238083598, + "grad_norm": 4.463747501373291, + "learning_rate": 1.8314696123025456e-05, + "loss": 1.1237, + "step": 867 + }, + { + "epoch": 0.2121730628208262, + "grad_norm": 4.581067085266113, + "learning_rate": 1.8310294890068845e-05, + "loss": 1.0942, + "step": 868 + }, + { + "epoch": 0.21241750183329258, + "grad_norm": 3.73099684715271, + "learning_rate": 1.830588844789066e-05, + "loss": 1.0187, + "step": 869 + }, + { + "epoch": 0.21266194084575898, + "grad_norm": 3.665910243988037, + "learning_rate": 1.830147679925303e-05, + "loss": 1.0703, + "step": 870 + }, + { + "epoch": 0.2129063798582254, + "grad_norm": 4.117025852203369, + "learning_rate": 1.8297059946921357e-05, + "loss": 0.9757, + "step": 871 + }, + { + "epoch": 0.21315081887069176, + "grad_norm": 3.8741068840026855, + "learning_rate": 1.829263789366429e-05, + "loss": 1.1434, + "step": 872 + }, + { + "epoch": 0.21339525788315816, + "grad_norm": 3.788287401199341, + "learning_rate": 1.828821064225375e-05, + "loss": 0.9807, + "step": 873 + }, + { + "epoch": 0.21363969689562454, + "grad_norm": 3.6013906002044678, + "learning_rate": 1.8283778195464913e-05, + "loss": 1.0818, + "step": 874 + }, + { + "epoch": 0.21388413590809094, + "grad_norm": 4.018670558929443, + "learning_rate": 1.8279340556076218e-05, + "loss": 1.1252, + "step": 875 + }, + { + "epoch": 0.2141285749205573, + "grad_norm": 4.4719557762146, + "learning_rate": 1.8274897726869342e-05, + "loss": 1.127, + "step": 876 + }, + { + "epoch": 0.21437301393302371, + "grad_norm": 6.977454662322998, + "learning_rate": 1.8270449710629236e-05, + "loss": 1.092, + "step": 877 + }, + { + "epoch": 0.2146174529454901, + "grad_norm": 4.391088485717773, + "learning_rate": 1.8265996510144085e-05, + "loss": 1.1899, + "step": 878 + }, + { + "epoch": 0.2148618919579565, + "grad_norm": 4.422701358795166, + "learning_rate": 1.8261538128205337e-05, + "loss": 1.0594, + "step": 879 + }, + { + "epoch": 0.2151063309704229, + "grad_norm": 6.5446953773498535, + "learning_rate": 1.825707456760768e-05, + "loss": 1.0841, + "step": 880 + }, + { + "epoch": 0.21535076998288927, + "grad_norm": 4.518584251403809, + "learning_rate": 1.8252605831149052e-05, + "loss": 1.0258, + "step": 881 + }, + { + "epoch": 0.21559520899535567, + "grad_norm": 3.3467774391174316, + "learning_rate": 1.824813192163063e-05, + "loss": 2.1104, + "step": 882 + }, + { + "epoch": 0.21583964800782204, + "grad_norm": 4.550435543060303, + "learning_rate": 1.8243652841856842e-05, + "loss": 1.093, + "step": 883 + }, + { + "epoch": 0.21608408702028845, + "grad_norm": 5.360644817352295, + "learning_rate": 1.823916859463535e-05, + "loss": 1.1999, + "step": 884 + }, + { + "epoch": 0.21632852603275482, + "grad_norm": 4.890727519989014, + "learning_rate": 1.8234679182777055e-05, + "loss": 0.9747, + "step": 885 + }, + { + "epoch": 0.21657296504522122, + "grad_norm": 4.1767659187316895, + "learning_rate": 1.82301846090961e-05, + "loss": 1.0555, + "step": 886 + }, + { + "epoch": 0.2168174040576876, + "grad_norm": 3.9458184242248535, + "learning_rate": 1.8225684876409863e-05, + "loss": 1.1683, + "step": 887 + }, + { + "epoch": 0.217061843070154, + "grad_norm": 3.597496747970581, + "learning_rate": 1.8221179987538952e-05, + "loss": 1.1083, + "step": 888 + }, + { + "epoch": 0.2173062820826204, + "grad_norm": 4.229506015777588, + "learning_rate": 1.8216669945307214e-05, + "loss": 1.1957, + "step": 889 + }, + { + "epoch": 0.21755072109508677, + "grad_norm": 3.556558132171631, + "learning_rate": 1.8212154752541718e-05, + "loss": 1.1348, + "step": 890 + }, + { + "epoch": 0.21779516010755318, + "grad_norm": 3.491957664489746, + "learning_rate": 1.8207634412072765e-05, + "loss": 1.1947, + "step": 891 + }, + { + "epoch": 0.21803959912001955, + "grad_norm": 4.08765983581543, + "learning_rate": 1.8203108926733886e-05, + "loss": 1.0128, + "step": 892 + }, + { + "epoch": 0.21828403813248595, + "grad_norm": 4.435181617736816, + "learning_rate": 1.8198578299361832e-05, + "loss": 1.0455, + "step": 893 + }, + { + "epoch": 0.21852847714495233, + "grad_norm": 4.2444634437561035, + "learning_rate": 1.819404253279658e-05, + "loss": 1.0579, + "step": 894 + }, + { + "epoch": 0.21877291615741873, + "grad_norm": 3.8789174556732178, + "learning_rate": 1.818950162988133e-05, + "loss": 1.0078, + "step": 895 + }, + { + "epoch": 0.2190173551698851, + "grad_norm": 4.655381679534912, + "learning_rate": 1.8184955593462494e-05, + "loss": 1.1297, + "step": 896 + }, + { + "epoch": 0.2192617941823515, + "grad_norm": 16.207721710205078, + "learning_rate": 1.818040442638971e-05, + "loss": 1.1169, + "step": 897 + }, + { + "epoch": 0.21950623319481788, + "grad_norm": 4.304232120513916, + "learning_rate": 1.817584813151584e-05, + "loss": 1.1648, + "step": 898 + }, + { + "epoch": 0.21975067220728428, + "grad_norm": 2.8558313846588135, + "learning_rate": 1.8171286711696935e-05, + "loss": 2.0964, + "step": 899 + }, + { + "epoch": 0.21999511121975068, + "grad_norm": 6.2625322341918945, + "learning_rate": 1.8166720169792276e-05, + "loss": 1.1234, + "step": 900 + }, + { + "epoch": 0.22023955023221706, + "grad_norm": 3.8653852939605713, + "learning_rate": 1.816214850866436e-05, + "loss": 0.9417, + "step": 901 + }, + { + "epoch": 0.22048398924468346, + "grad_norm": 4.717000961303711, + "learning_rate": 1.815757173117888e-05, + "loss": 1.0726, + "step": 902 + }, + { + "epoch": 0.22072842825714983, + "grad_norm": 4.075290203094482, + "learning_rate": 1.8152989840204734e-05, + "loss": 1.1515, + "step": 903 + }, + { + "epoch": 0.22097286726961624, + "grad_norm": 3.7050132751464844, + "learning_rate": 1.814840283861405e-05, + "loss": 1.0404, + "step": 904 + }, + { + "epoch": 0.2212173062820826, + "grad_norm": 4.526838302612305, + "learning_rate": 1.8143810729282125e-05, + "loss": 1.0268, + "step": 905 + }, + { + "epoch": 0.221461745294549, + "grad_norm": 4.461960315704346, + "learning_rate": 1.8139213515087484e-05, + "loss": 1.0661, + "step": 906 + }, + { + "epoch": 0.22170618430701539, + "grad_norm": 2.5100247859954834, + "learning_rate": 1.813461119891184e-05, + "loss": 2.0317, + "step": 907 + }, + { + "epoch": 0.2219506233194818, + "grad_norm": 4.275705337524414, + "learning_rate": 1.8130003783640106e-05, + "loss": 1.1287, + "step": 908 + }, + { + "epoch": 0.2221950623319482, + "grad_norm": 26.058462142944336, + "learning_rate": 1.8125391272160397e-05, + "loss": 1.1154, + "step": 909 + }, + { + "epoch": 0.22243950134441456, + "grad_norm": 5.324725151062012, + "learning_rate": 1.8120773667364012e-05, + "loss": 1.0256, + "step": 910 + }, + { + "epoch": 0.22268394035688097, + "grad_norm": 4.943587303161621, + "learning_rate": 1.811615097214545e-05, + "loss": 1.1156, + "step": 911 + }, + { + "epoch": 0.22292837936934734, + "grad_norm": 3.439422845840454, + "learning_rate": 1.81115231894024e-05, + "loss": 1.0305, + "step": 912 + }, + { + "epoch": 0.22317281838181374, + "grad_norm": 4.746899604797363, + "learning_rate": 1.8106890322035737e-05, + "loss": 1.0388, + "step": 913 + }, + { + "epoch": 0.22341725739428012, + "grad_norm": 3.8092880249023438, + "learning_rate": 1.810225237294953e-05, + "loss": 0.9694, + "step": 914 + }, + { + "epoch": 0.22366169640674652, + "grad_norm": 4.545948028564453, + "learning_rate": 1.8097609345051027e-05, + "loss": 1.0253, + "step": 915 + }, + { + "epoch": 0.2239061354192129, + "grad_norm": 3.7231202125549316, + "learning_rate": 1.8092961241250658e-05, + "loss": 1.0292, + "step": 916 + }, + { + "epoch": 0.2241505744316793, + "grad_norm": 4.942643642425537, + "learning_rate": 1.8088308064462043e-05, + "loss": 1.1375, + "step": 917 + }, + { + "epoch": 0.2243950134441457, + "grad_norm": 5.857900619506836, + "learning_rate": 1.8083649817601976e-05, + "loss": 1.0794, + "step": 918 + }, + { + "epoch": 0.22463945245661207, + "grad_norm": 4.76950740814209, + "learning_rate": 1.807898650359043e-05, + "loss": 1.1841, + "step": 919 + }, + { + "epoch": 0.22488389146907847, + "grad_norm": 3.846597194671631, + "learning_rate": 1.8074318125350553e-05, + "loss": 1.1022, + "step": 920 + }, + { + "epoch": 0.22512833048154485, + "grad_norm": 3.865078926086426, + "learning_rate": 1.8069644685808673e-05, + "loss": 0.996, + "step": 921 + }, + { + "epoch": 0.22537276949401125, + "grad_norm": 5.097954273223877, + "learning_rate": 1.8064966187894284e-05, + "loss": 1.1416, + "step": 922 + }, + { + "epoch": 0.22561720850647762, + "grad_norm": 5.553801536560059, + "learning_rate": 1.8060282634540053e-05, + "loss": 0.9644, + "step": 923 + }, + { + "epoch": 0.22586164751894403, + "grad_norm": 5.003941059112549, + "learning_rate": 1.8055594028681817e-05, + "loss": 1.0668, + "step": 924 + }, + { + "epoch": 0.2261060865314104, + "grad_norm": 4.66826057434082, + "learning_rate": 1.805090037325858e-05, + "loss": 1.1768, + "step": 925 + }, + { + "epoch": 0.2263505255438768, + "grad_norm": 6.5088396072387695, + "learning_rate": 1.8046201671212515e-05, + "loss": 1.033, + "step": 926 + }, + { + "epoch": 0.2265949645563432, + "grad_norm": 4.333165645599365, + "learning_rate": 1.804149792548895e-05, + "loss": 1.0737, + "step": 927 + }, + { + "epoch": 0.22683940356880958, + "grad_norm": 5.561150074005127, + "learning_rate": 1.8036789139036375e-05, + "loss": 1.0841, + "step": 928 + }, + { + "epoch": 0.22708384258127598, + "grad_norm": 4.797734260559082, + "learning_rate": 1.803207531480645e-05, + "loss": 1.1351, + "step": 929 + }, + { + "epoch": 0.22732828159374235, + "grad_norm": 5.050440311431885, + "learning_rate": 1.8027356455753984e-05, + "loss": 1.0342, + "step": 930 + }, + { + "epoch": 0.22757272060620876, + "grad_norm": 5.201828479766846, + "learning_rate": 1.8022632564836948e-05, + "loss": 1.0864, + "step": 931 + }, + { + "epoch": 0.22781715961867513, + "grad_norm": 4.372162342071533, + "learning_rate": 1.801790364501646e-05, + "loss": 0.9967, + "step": 932 + }, + { + "epoch": 0.22806159863114153, + "grad_norm": 3.2812650203704834, + "learning_rate": 1.801316969925679e-05, + "loss": 2.1303, + "step": 933 + }, + { + "epoch": 0.2283060376436079, + "grad_norm": 4.838235378265381, + "learning_rate": 1.8008430730525372e-05, + "loss": 1.084, + "step": 934 + }, + { + "epoch": 0.2285504766560743, + "grad_norm": 6.401941776275635, + "learning_rate": 1.8003686741792773e-05, + "loss": 0.9873, + "step": 935 + }, + { + "epoch": 0.2287949156685407, + "grad_norm": 5.116730690002441, + "learning_rate": 1.7998937736032714e-05, + "loss": 1.028, + "step": 936 + }, + { + "epoch": 0.22903935468100708, + "grad_norm": 3.7749273777008057, + "learning_rate": 1.799418371622206e-05, + "loss": 1.1115, + "step": 937 + }, + { + "epoch": 0.2292837936934735, + "grad_norm": 6.529001235961914, + "learning_rate": 1.798942468534082e-05, + "loss": 1.0567, + "step": 938 + }, + { + "epoch": 0.22952823270593986, + "grad_norm": 5.243895530700684, + "learning_rate": 1.7984660646372138e-05, + "loss": 0.9885, + "step": 939 + }, + { + "epoch": 0.22977267171840626, + "grad_norm": 6.02579927444458, + "learning_rate": 1.7979891602302305e-05, + "loss": 1.0929, + "step": 940 + }, + { + "epoch": 0.23001711073087264, + "grad_norm": 10.071123123168945, + "learning_rate": 1.797511755612075e-05, + "loss": 0.9637, + "step": 941 + }, + { + "epoch": 0.23026154974333904, + "grad_norm": 5.939101696014404, + "learning_rate": 1.797033851082003e-05, + "loss": 1.1494, + "step": 942 + }, + { + "epoch": 0.2305059887558054, + "grad_norm": 4.889463424682617, + "learning_rate": 1.7965554469395837e-05, + "loss": 1.1101, + "step": 943 + }, + { + "epoch": 0.23075042776827182, + "grad_norm": 5.173099517822266, + "learning_rate": 1.7960765434847007e-05, + "loss": 1.0926, + "step": 944 + }, + { + "epoch": 0.23099486678073822, + "grad_norm": 9.772001266479492, + "learning_rate": 1.7955971410175484e-05, + "loss": 1.0194, + "step": 945 + }, + { + "epoch": 0.2312393057932046, + "grad_norm": 4.739356517791748, + "learning_rate": 1.7951172398386363e-05, + "loss": 1.0346, + "step": 946 + }, + { + "epoch": 0.231483744805671, + "grad_norm": 11.065056800842285, + "learning_rate": 1.7946368402487845e-05, + "loss": 1.091, + "step": 947 + }, + { + "epoch": 0.23172818381813737, + "grad_norm": 4.65624475479126, + "learning_rate": 1.7941559425491276e-05, + "loss": 1.1244, + "step": 948 + }, + { + "epoch": 0.23197262283060377, + "grad_norm": 6.162562370300293, + "learning_rate": 1.79367454704111e-05, + "loss": 1.0596, + "step": 949 + }, + { + "epoch": 0.23221706184307014, + "grad_norm": 4.951538562774658, + "learning_rate": 1.7931926540264907e-05, + "loss": 1.0474, + "step": 950 + }, + { + "epoch": 0.23246150085553655, + "grad_norm": 4.932707786560059, + "learning_rate": 1.7927102638073384e-05, + "loss": 1.0991, + "step": 951 + }, + { + "epoch": 0.23270593986800292, + "grad_norm": 5.941732883453369, + "learning_rate": 1.7922273766860346e-05, + "loss": 1.054, + "step": 952 + }, + { + "epoch": 0.23295037888046932, + "grad_norm": 5.31721830368042, + "learning_rate": 1.7917439929652723e-05, + "loss": 1.1043, + "step": 953 + }, + { + "epoch": 0.23319481789293572, + "grad_norm": 6.00404691696167, + "learning_rate": 1.7912601129480554e-05, + "loss": 1.0955, + "step": 954 + }, + { + "epoch": 0.2334392569054021, + "grad_norm": 5.6774187088012695, + "learning_rate": 1.7907757369376984e-05, + "loss": 1.0019, + "step": 955 + }, + { + "epoch": 0.2336836959178685, + "grad_norm": 11.581239700317383, + "learning_rate": 1.7902908652378287e-05, + "loss": 1.0713, + "step": 956 + }, + { + "epoch": 0.23392813493033487, + "grad_norm": 8.59800910949707, + "learning_rate": 1.789805498152382e-05, + "loss": 1.0113, + "step": 957 + }, + { + "epoch": 0.23417257394280128, + "grad_norm": 6.038421154022217, + "learning_rate": 1.7893196359856058e-05, + "loss": 1.0019, + "step": 958 + }, + { + "epoch": 0.23441701295526765, + "grad_norm": 15.316848754882812, + "learning_rate": 1.7888332790420582e-05, + "loss": 1.1193, + "step": 959 + }, + { + "epoch": 0.23466145196773405, + "grad_norm": 5.798333168029785, + "learning_rate": 1.7883464276266064e-05, + "loss": 0.8781, + "step": 960 + }, + { + "epoch": 0.23490589098020043, + "grad_norm": 9.919295310974121, + "learning_rate": 1.7878590820444283e-05, + "loss": 1.08, + "step": 961 + }, + { + "epoch": 0.23515032999266683, + "grad_norm": 4.5314812660217285, + "learning_rate": 1.7873712426010112e-05, + "loss": 0.9922, + "step": 962 + }, + { + "epoch": 0.23539476900513323, + "grad_norm": 5.200216770172119, + "learning_rate": 1.786882909602153e-05, + "loss": 1.059, + "step": 963 + }, + { + "epoch": 0.2356392080175996, + "grad_norm": 12.14441967010498, + "learning_rate": 1.786394083353959e-05, + "loss": 0.9602, + "step": 964 + }, + { + "epoch": 0.235883647030066, + "grad_norm": 6.1346964836120605, + "learning_rate": 1.7859047641628452e-05, + "loss": 1.0439, + "step": 965 + }, + { + "epoch": 0.23612808604253238, + "grad_norm": 11.949640274047852, + "learning_rate": 1.7854149523355366e-05, + "loss": 1.1008, + "step": 966 + }, + { + "epoch": 0.23637252505499878, + "grad_norm": 5.413366794586182, + "learning_rate": 1.7849246481790664e-05, + "loss": 1.0515, + "step": 967 + }, + { + "epoch": 0.23661696406746516, + "grad_norm": 4.006432056427002, + "learning_rate": 1.7844338520007765e-05, + "loss": 1.0281, + "step": 968 + }, + { + "epoch": 0.23686140307993156, + "grad_norm": 4.715639114379883, + "learning_rate": 1.783942564108318e-05, + "loss": 1.0322, + "step": 969 + }, + { + "epoch": 0.23710584209239793, + "grad_norm": 6.650883674621582, + "learning_rate": 1.7834507848096488e-05, + "loss": 1.1547, + "step": 970 + }, + { + "epoch": 0.23735028110486434, + "grad_norm": 3.7897934913635254, + "learning_rate": 1.7829585144130356e-05, + "loss": 1.0635, + "step": 971 + }, + { + "epoch": 0.23759472011733074, + "grad_norm": 7.736684799194336, + "learning_rate": 1.7824657532270535e-05, + "loss": 1.0343, + "step": 972 + }, + { + "epoch": 0.2378391591297971, + "grad_norm": 4.9490275382995605, + "learning_rate": 1.7819725015605843e-05, + "loss": 1.0995, + "step": 973 + }, + { + "epoch": 0.23808359814226351, + "grad_norm": 4.268904209136963, + "learning_rate": 1.781478759722818e-05, + "loss": 1.0989, + "step": 974 + }, + { + "epoch": 0.2383280371547299, + "grad_norm": 4.651836395263672, + "learning_rate": 1.7809845280232512e-05, + "loss": 1.0715, + "step": 975 + }, + { + "epoch": 0.2385724761671963, + "grad_norm": 11.762102127075195, + "learning_rate": 1.7804898067716883e-05, + "loss": 1.0578, + "step": 976 + }, + { + "epoch": 0.23881691517966266, + "grad_norm": 5.454673767089844, + "learning_rate": 1.7799945962782397e-05, + "loss": 1.0425, + "step": 977 + }, + { + "epoch": 0.23906135419212907, + "grad_norm": 5.125723838806152, + "learning_rate": 1.7794988968533232e-05, + "loss": 1.0211, + "step": 978 + }, + { + "epoch": 0.23930579320459544, + "grad_norm": 5.660276412963867, + "learning_rate": 1.779002708807662e-05, + "loss": 1.0719, + "step": 979 + }, + { + "epoch": 0.23955023221706184, + "grad_norm": 4.565948963165283, + "learning_rate": 1.778506032452288e-05, + "loss": 1.0852, + "step": 980 + }, + { + "epoch": 0.23979467122952824, + "grad_norm": 5.478196144104004, + "learning_rate": 1.7780088680985365e-05, + "loss": 1.0543, + "step": 981 + }, + { + "epoch": 0.24003911024199462, + "grad_norm": 4.445716381072998, + "learning_rate": 1.7775112160580497e-05, + "loss": 1.1229, + "step": 982 + }, + { + "epoch": 0.24028354925446102, + "grad_norm": 5.014064311981201, + "learning_rate": 1.777013076642776e-05, + "loss": 1.0659, + "step": 983 + }, + { + "epoch": 0.2405279882669274, + "grad_norm": 6.775639057159424, + "learning_rate": 1.7765144501649687e-05, + "loss": 1.0905, + "step": 984 + }, + { + "epoch": 0.2407724272793938, + "grad_norm": 7.390527725219727, + "learning_rate": 1.776015336937187e-05, + "loss": 1.008, + "step": 985 + }, + { + "epoch": 0.24101686629186017, + "grad_norm": 5.768934726715088, + "learning_rate": 1.7755157372722938e-05, + "loss": 0.9349, + "step": 986 + }, + { + "epoch": 0.24126130530432657, + "grad_norm": 3.9665780067443848, + "learning_rate": 1.775015651483459e-05, + "loss": 0.9906, + "step": 987 + }, + { + "epoch": 0.24150574431679295, + "grad_norm": 5.45059061050415, + "learning_rate": 1.7745150798841554e-05, + "loss": 1.0664, + "step": 988 + }, + { + "epoch": 0.24175018332925935, + "grad_norm": 8.767844200134277, + "learning_rate": 1.774014022788162e-05, + "loss": 1.1122, + "step": 989 + }, + { + "epoch": 0.24199462234172575, + "grad_norm": 5.439845085144043, + "learning_rate": 1.7735124805095602e-05, + "loss": 0.9101, + "step": 990 + }, + { + "epoch": 0.24223906135419213, + "grad_norm": 4.81063175201416, + "learning_rate": 1.773010453362737e-05, + "loss": 1.0126, + "step": 991 + }, + { + "epoch": 0.24248350036665853, + "grad_norm": 6.504293441772461, + "learning_rate": 1.7725079416623828e-05, + "loss": 0.9735, + "step": 992 + }, + { + "epoch": 0.2427279393791249, + "grad_norm": 9.619270324707031, + "learning_rate": 1.772004945723492e-05, + "loss": 1.1093, + "step": 993 + }, + { + "epoch": 0.2429723783915913, + "grad_norm": 4.560388565063477, + "learning_rate": 1.7715014658613626e-05, + "loss": 1.0558, + "step": 994 + }, + { + "epoch": 0.24321681740405768, + "grad_norm": 7.408061981201172, + "learning_rate": 1.7709975023915948e-05, + "loss": 1.1495, + "step": 995 + }, + { + "epoch": 0.24346125641652408, + "grad_norm": 4.253489017486572, + "learning_rate": 1.7704930556300936e-05, + "loss": 1.0827, + "step": 996 + }, + { + "epoch": 0.24370569542899045, + "grad_norm": 4.0216546058654785, + "learning_rate": 1.7699881258930658e-05, + "loss": 1.0846, + "step": 997 + }, + { + "epoch": 0.24395013444145686, + "grad_norm": 5.678346633911133, + "learning_rate": 1.7694827134970217e-05, + "loss": 1.1458, + "step": 998 + }, + { + "epoch": 0.24419457345392326, + "grad_norm": 4.36787748336792, + "learning_rate": 1.7689768187587732e-05, + "loss": 1.0336, + "step": 999 + }, + { + "epoch": 0.24443901246638963, + "grad_norm": 3.6349778175354004, + "learning_rate": 1.7684704419954356e-05, + "loss": 1.07, + "step": 1000 + }, + { + "epoch": 0.24468345147885603, + "grad_norm": 4.574316501617432, + "learning_rate": 1.7679635835244256e-05, + "loss": 1.0281, + "step": 1001 + }, + { + "epoch": 0.2449278904913224, + "grad_norm": 4.44944953918457, + "learning_rate": 1.7674562436634627e-05, + "loss": 1.1225, + "step": 1002 + }, + { + "epoch": 0.2451723295037888, + "grad_norm": 3.7810378074645996, + "learning_rate": 1.766948422730567e-05, + "loss": 2.1173, + "step": 1003 + }, + { + "epoch": 0.24541676851625518, + "grad_norm": 6.347598552703857, + "learning_rate": 1.766440121044061e-05, + "loss": 1.0719, + "step": 1004 + }, + { + "epoch": 0.2456612075287216, + "grad_norm": 4.953799247741699, + "learning_rate": 1.765931338922568e-05, + "loss": 0.9977, + "step": 1005 + }, + { + "epoch": 0.24590564654118796, + "grad_norm": 6.32512092590332, + "learning_rate": 1.7654220766850134e-05, + "loss": 1.044, + "step": 1006 + }, + { + "epoch": 0.24615008555365436, + "grad_norm": 2.5741055011749268, + "learning_rate": 1.7649123346506223e-05, + "loss": 2.0905, + "step": 1007 + }, + { + "epoch": 0.24639452456612077, + "grad_norm": 2.821791410446167, + "learning_rate": 1.7644021131389216e-05, + "loss": 2.1265, + "step": 1008 + }, + { + "epoch": 0.24663896357858714, + "grad_norm": 3.9031708240509033, + "learning_rate": 1.763891412469738e-05, + "loss": 1.0813, + "step": 1009 + }, + { + "epoch": 0.24688340259105354, + "grad_norm": 4.954384803771973, + "learning_rate": 1.763380232963199e-05, + "loss": 1.0133, + "step": 1010 + }, + { + "epoch": 0.24712784160351992, + "grad_norm": 3.2676358222961426, + "learning_rate": 1.762868574939732e-05, + "loss": 1.0316, + "step": 1011 + }, + { + "epoch": 0.24737228061598632, + "grad_norm": 7.159889221191406, + "learning_rate": 1.7623564387200648e-05, + "loss": 1.119, + "step": 1012 + }, + { + "epoch": 0.2476167196284527, + "grad_norm": 3.8670740127563477, + "learning_rate": 1.761843824625225e-05, + "loss": 1.1125, + "step": 1013 + }, + { + "epoch": 0.2478611586409191, + "grad_norm": 4.3193817138671875, + "learning_rate": 1.7613307329765382e-05, + "loss": 1.0935, + "step": 1014 + }, + { + "epoch": 0.24810559765338547, + "grad_norm": 4.301325798034668, + "learning_rate": 1.7608171640956312e-05, + "loss": 0.9862, + "step": 1015 + }, + { + "epoch": 0.24835003666585187, + "grad_norm": 3.985088348388672, + "learning_rate": 1.7603031183044296e-05, + "loss": 1.046, + "step": 1016 + }, + { + "epoch": 0.24859447567831827, + "grad_norm": 6.649075031280518, + "learning_rate": 1.7597885959251574e-05, + "loss": 1.104, + "step": 1017 + }, + { + "epoch": 0.24883891469078465, + "grad_norm": 3.7859950065612793, + "learning_rate": 1.759273597280337e-05, + "loss": 0.8895, + "step": 1018 + }, + { + "epoch": 0.24908335370325105, + "grad_norm": 4.199967384338379, + "learning_rate": 1.758758122692791e-05, + "loss": 0.9918, + "step": 1019 + }, + { + "epoch": 0.24932779271571742, + "grad_norm": 4.479875087738037, + "learning_rate": 1.7582421724856383e-05, + "loss": 1.0098, + "step": 1020 + }, + { + "epoch": 0.24957223172818382, + "grad_norm": 3.7866857051849365, + "learning_rate": 1.7577257469822976e-05, + "loss": 1.0731, + "step": 1021 + }, + { + "epoch": 0.2498166707406502, + "grad_norm": 3.81478214263916, + "learning_rate": 1.7572088465064847e-05, + "loss": 0.9534, + "step": 1022 + }, + { + "epoch": 0.2500611097531166, + "grad_norm": 4.332945346832275, + "learning_rate": 1.756691471382213e-05, + "loss": 1.047, + "step": 1023 + }, + { + "epoch": 0.250305548765583, + "grad_norm": 4.55991268157959, + "learning_rate": 1.7561736219337935e-05, + "loss": 1.0302, + "step": 1024 + }, + { + "epoch": 0.2505499877780494, + "grad_norm": 3.5179877281188965, + "learning_rate": 1.7556552984858353e-05, + "loss": 1.0994, + "step": 1025 + }, + { + "epoch": 0.2507944267905158, + "grad_norm": 5.125848770141602, + "learning_rate": 1.7551365013632437e-05, + "loss": 1.0657, + "step": 1026 + }, + { + "epoch": 0.2510388658029822, + "grad_norm": 4.182981014251709, + "learning_rate": 1.7546172308912213e-05, + "loss": 1.0652, + "step": 1027 + }, + { + "epoch": 0.2512833048154485, + "grad_norm": 5.154917240142822, + "learning_rate": 1.7540974873952673e-05, + "loss": 1.11, + "step": 1028 + }, + { + "epoch": 0.25152774382791493, + "grad_norm": 4.645609378814697, + "learning_rate": 1.7535772712011775e-05, + "loss": 1.0179, + "step": 1029 + }, + { + "epoch": 0.25177218284038133, + "grad_norm": 4.523293495178223, + "learning_rate": 1.7530565826350444e-05, + "loss": 1.0007, + "step": 1030 + }, + { + "epoch": 0.25201662185284773, + "grad_norm": 4.768362998962402, + "learning_rate": 1.7525354220232558e-05, + "loss": 1.0095, + "step": 1031 + }, + { + "epoch": 0.2522610608653141, + "grad_norm": 4.03495454788208, + "learning_rate": 1.752013789692496e-05, + "loss": 1.0007, + "step": 1032 + }, + { + "epoch": 0.2525054998777805, + "grad_norm": 4.036981105804443, + "learning_rate": 1.7514916859697443e-05, + "loss": 1.0643, + "step": 1033 + }, + { + "epoch": 0.2527499388902469, + "grad_norm": 4.7859978675842285, + "learning_rate": 1.7509691111822764e-05, + "loss": 1.0359, + "step": 1034 + }, + { + "epoch": 0.2529943779027133, + "grad_norm": 4.358851909637451, + "learning_rate": 1.750446065657663e-05, + "loss": 0.9784, + "step": 1035 + }, + { + "epoch": 0.2532388169151797, + "grad_norm": 4.325262546539307, + "learning_rate": 1.7499225497237686e-05, + "loss": 1.0047, + "step": 1036 + }, + { + "epoch": 0.25348325592764603, + "grad_norm": 5.185552597045898, + "learning_rate": 1.7493985637087555e-05, + "loss": 1.14, + "step": 1037 + }, + { + "epoch": 0.25372769494011244, + "grad_norm": 4.701370716094971, + "learning_rate": 1.7488741079410775e-05, + "loss": 2.2303, + "step": 1038 + }, + { + "epoch": 0.25397213395257884, + "grad_norm": 6.095943927764893, + "learning_rate": 1.7483491827494846e-05, + "loss": 1.1368, + "step": 1039 + }, + { + "epoch": 0.25421657296504524, + "grad_norm": 4.193117141723633, + "learning_rate": 1.7478237884630208e-05, + "loss": 1.2079, + "step": 1040 + }, + { + "epoch": 0.2544610119775116, + "grad_norm": 4.067698001861572, + "learning_rate": 1.747297925411024e-05, + "loss": 1.045, + "step": 1041 + }, + { + "epoch": 0.254705450989978, + "grad_norm": 4.788492202758789, + "learning_rate": 1.746771593923126e-05, + "loss": 0.9934, + "step": 1042 + }, + { + "epoch": 0.2549498900024444, + "grad_norm": 8.858131408691406, + "learning_rate": 1.7462447943292522e-05, + "loss": 1.0462, + "step": 1043 + }, + { + "epoch": 0.2551943290149108, + "grad_norm": 3.593235492706299, + "learning_rate": 1.745717526959621e-05, + "loss": 1.0151, + "step": 1044 + }, + { + "epoch": 0.2554387680273772, + "grad_norm": 4.977777004241943, + "learning_rate": 1.7451897921447448e-05, + "loss": 1.0593, + "step": 1045 + }, + { + "epoch": 0.25568320703984354, + "grad_norm": 4.215940952301025, + "learning_rate": 1.7446615902154288e-05, + "loss": 1.0583, + "step": 1046 + }, + { + "epoch": 0.25592764605230994, + "grad_norm": 3.8196914196014404, + "learning_rate": 1.7441329215027707e-05, + "loss": 1.0829, + "step": 1047 + }, + { + "epoch": 0.25617208506477634, + "grad_norm": 4.272761821746826, + "learning_rate": 1.743603786338161e-05, + "loss": 1.108, + "step": 1048 + }, + { + "epoch": 0.25641652407724275, + "grad_norm": 3.1501529216766357, + "learning_rate": 1.7430741850532824e-05, + "loss": 0.9738, + "step": 1049 + }, + { + "epoch": 0.2566609630897091, + "grad_norm": 3.941812753677368, + "learning_rate": 1.74254411798011e-05, + "loss": 1.1132, + "step": 1050 + }, + { + "epoch": 0.2569054021021755, + "grad_norm": 4.7528276443481445, + "learning_rate": 1.742013585450911e-05, + "loss": 1.144, + "step": 1051 + }, + { + "epoch": 0.2571498411146419, + "grad_norm": 7.925037384033203, + "learning_rate": 1.741482587798244e-05, + "loss": 2.2551, + "step": 1052 + }, + { + "epoch": 0.2573942801271083, + "grad_norm": 5.1193318367004395, + "learning_rate": 1.7409511253549592e-05, + "loss": 1.0979, + "step": 1053 + }, + { + "epoch": 0.2576387191395747, + "grad_norm": 5.097321033477783, + "learning_rate": 1.7404191984541987e-05, + "loss": 1.0282, + "step": 1054 + }, + { + "epoch": 0.25788315815204105, + "grad_norm": 4.32500696182251, + "learning_rate": 1.7398868074293947e-05, + "loss": 1.1019, + "step": 1055 + }, + { + "epoch": 0.25812759716450745, + "grad_norm": 6.229393005371094, + "learning_rate": 1.7393539526142717e-05, + "loss": 1.034, + "step": 1056 + }, + { + "epoch": 0.25837203617697385, + "grad_norm": 4.147944927215576, + "learning_rate": 1.7388206343428438e-05, + "loss": 1.0432, + "step": 1057 + }, + { + "epoch": 0.25861647518944025, + "grad_norm": 3.581871271133423, + "learning_rate": 1.7382868529494158e-05, + "loss": 1.0871, + "step": 1058 + }, + { + "epoch": 0.2588609142019066, + "grad_norm": 5.094482421875, + "learning_rate": 1.7377526087685832e-05, + "loss": 1.0374, + "step": 1059 + }, + { + "epoch": 0.259105353214373, + "grad_norm": 3.791944980621338, + "learning_rate": 1.7372179021352317e-05, + "loss": 1.1247, + "step": 1060 + }, + { + "epoch": 0.2593497922268394, + "grad_norm": 4.173458099365234, + "learning_rate": 1.736682733384536e-05, + "loss": 0.9955, + "step": 1061 + }, + { + "epoch": 0.2595942312393058, + "grad_norm": 3.895758628845215, + "learning_rate": 1.7361471028519613e-05, + "loss": 1.1296, + "step": 1062 + }, + { + "epoch": 0.2598386702517722, + "grad_norm": 3.5435473918914795, + "learning_rate": 1.735611010873262e-05, + "loss": 1.0162, + "step": 1063 + }, + { + "epoch": 0.26008310926423855, + "grad_norm": 6.77610969543457, + "learning_rate": 1.7350744577844816e-05, + "loss": 0.95, + "step": 1064 + }, + { + "epoch": 0.26032754827670496, + "grad_norm": 7.926026344299316, + "learning_rate": 1.734537443921953e-05, + "loss": 1.0358, + "step": 1065 + }, + { + "epoch": 0.26057198728917136, + "grad_norm": 3.391551971435547, + "learning_rate": 1.733999969622298e-05, + "loss": 0.9246, + "step": 1066 + }, + { + "epoch": 0.26081642630163776, + "grad_norm": 4.00184440612793, + "learning_rate": 1.733462035222426e-05, + "loss": 1.0923, + "step": 1067 + }, + { + "epoch": 0.2610608653141041, + "grad_norm": 6.604822158813477, + "learning_rate": 1.7329236410595364e-05, + "loss": 0.9171, + "step": 1068 + }, + { + "epoch": 0.2613053043265705, + "grad_norm": 4.701949596405029, + "learning_rate": 1.7323847874711152e-05, + "loss": 1.0092, + "step": 1069 + }, + { + "epoch": 0.2615497433390369, + "grad_norm": 4.597169399261475, + "learning_rate": 1.7318454747949375e-05, + "loss": 1.0115, + "step": 1070 + }, + { + "epoch": 0.2617941823515033, + "grad_norm": 5.872190952301025, + "learning_rate": 1.7313057033690662e-05, + "loss": 1.0412, + "step": 1071 + }, + { + "epoch": 0.2620386213639697, + "grad_norm": 5.632815837860107, + "learning_rate": 1.7307654735318505e-05, + "loss": 1.0499, + "step": 1072 + }, + { + "epoch": 0.26228306037643606, + "grad_norm": 5.916876792907715, + "learning_rate": 1.730224785621929e-05, + "loss": 1.0165, + "step": 1073 + }, + { + "epoch": 0.26252749938890246, + "grad_norm": 3.9583590030670166, + "learning_rate": 1.7296836399782253e-05, + "loss": 0.983, + "step": 1074 + }, + { + "epoch": 0.26277193840136887, + "grad_norm": 3.7609782218933105, + "learning_rate": 1.729142036939951e-05, + "loss": 1.0606, + "step": 1075 + }, + { + "epoch": 0.26301637741383527, + "grad_norm": 8.059402465820312, + "learning_rate": 1.7285999768466048e-05, + "loss": 1.0308, + "step": 1076 + }, + { + "epoch": 0.2632608164263016, + "grad_norm": 4.188074111938477, + "learning_rate": 1.728057460037971e-05, + "loss": 1.0676, + "step": 1077 + }, + { + "epoch": 0.263505255438768, + "grad_norm": 3.7996623516082764, + "learning_rate": 1.7275144868541208e-05, + "loss": 1.0604, + "step": 1078 + }, + { + "epoch": 0.2637496944512344, + "grad_norm": 5.376227378845215, + "learning_rate": 1.7269710576354112e-05, + "loss": 1.1378, + "step": 1079 + }, + { + "epoch": 0.2639941334637008, + "grad_norm": 4.061453819274902, + "learning_rate": 1.7264271727224847e-05, + "loss": 1.0861, + "step": 1080 + }, + { + "epoch": 0.2642385724761672, + "grad_norm": 3.80562686920166, + "learning_rate": 1.7258828324562705e-05, + "loss": 1.0534, + "step": 1081 + }, + { + "epoch": 0.26448301148863357, + "grad_norm": 4.415918350219727, + "learning_rate": 1.725338037177983e-05, + "loss": 0.9853, + "step": 1082 + }, + { + "epoch": 0.26472745050109997, + "grad_norm": 4.435208797454834, + "learning_rate": 1.72479278722912e-05, + "loss": 1.1745, + "step": 1083 + }, + { + "epoch": 0.26497188951356637, + "grad_norm": 3.257791042327881, + "learning_rate": 1.7242470829514674e-05, + "loss": 2.1818, + "step": 1084 + }, + { + "epoch": 0.2652163285260328, + "grad_norm": 4.946976184844971, + "learning_rate": 1.723700924687093e-05, + "loss": 1.2528, + "step": 1085 + }, + { + "epoch": 0.2654607675384991, + "grad_norm": 4.8052802085876465, + "learning_rate": 1.7231543127783506e-05, + "loss": 1.034, + "step": 1086 + }, + { + "epoch": 0.2657052065509655, + "grad_norm": 3.2481181621551514, + "learning_rate": 1.7226072475678783e-05, + "loss": 1.1201, + "step": 1087 + }, + { + "epoch": 0.2659496455634319, + "grad_norm": 3.8651785850524902, + "learning_rate": 1.7220597293985987e-05, + "loss": 0.9795, + "step": 1088 + }, + { + "epoch": 0.2661940845758983, + "grad_norm": 4.134873867034912, + "learning_rate": 1.7215117586137174e-05, + "loss": 2.0171, + "step": 1089 + }, + { + "epoch": 0.26643852358836473, + "grad_norm": 3.266214609146118, + "learning_rate": 1.7209633355567242e-05, + "loss": 0.9662, + "step": 1090 + }, + { + "epoch": 0.2666829626008311, + "grad_norm": 4.433269023895264, + "learning_rate": 1.7204144605713922e-05, + "loss": 1.0065, + "step": 1091 + }, + { + "epoch": 0.2669274016132975, + "grad_norm": 3.932711601257324, + "learning_rate": 1.7198651340017783e-05, + "loss": 0.9736, + "step": 1092 + }, + { + "epoch": 0.2671718406257639, + "grad_norm": 4.1653733253479, + "learning_rate": 1.7193153561922214e-05, + "loss": 0.9304, + "step": 1093 + }, + { + "epoch": 0.2674162796382303, + "grad_norm": 6.558185577392578, + "learning_rate": 1.718765127487345e-05, + "loss": 1.0107, + "step": 1094 + }, + { + "epoch": 0.2676607186506966, + "grad_norm": 4.074313640594482, + "learning_rate": 1.718214448232054e-05, + "loss": 1.1244, + "step": 1095 + }, + { + "epoch": 0.26790515766316303, + "grad_norm": 4.0094733238220215, + "learning_rate": 1.7176633187715353e-05, + "loss": 1.0114, + "step": 1096 + }, + { + "epoch": 0.26814959667562943, + "grad_norm": 4.766077041625977, + "learning_rate": 1.7171117394512594e-05, + "loss": 1.0161, + "step": 1097 + }, + { + "epoch": 0.26839403568809583, + "grad_norm": 3.321650743484497, + "learning_rate": 1.7165597106169775e-05, + "loss": 0.9981, + "step": 1098 + }, + { + "epoch": 0.26863847470056224, + "grad_norm": 3.749901056289673, + "learning_rate": 1.716007232614723e-05, + "loss": 1.0639, + "step": 1099 + }, + { + "epoch": 0.2688829137130286, + "grad_norm": 4.82611608505249, + "learning_rate": 1.7154543057908116e-05, + "loss": 1.0381, + "step": 1100 + }, + { + "epoch": 0.269127352725495, + "grad_norm": 7.40592622756958, + "learning_rate": 1.7149009304918392e-05, + "loss": 0.9762, + "step": 1101 + }, + { + "epoch": 0.2693717917379614, + "grad_norm": 3.8617756366729736, + "learning_rate": 1.7143471070646834e-05, + "loss": 1.0359, + "step": 1102 + }, + { + "epoch": 0.2696162307504278, + "grad_norm": 3.7181596755981445, + "learning_rate": 1.713792835856503e-05, + "loss": 1.0627, + "step": 1103 + }, + { + "epoch": 0.26986066976289413, + "grad_norm": 5.8879714012146, + "learning_rate": 1.7132381172147368e-05, + "loss": 0.9501, + "step": 1104 + }, + { + "epoch": 0.27010510877536054, + "grad_norm": 3.9147109985351562, + "learning_rate": 1.7126829514871043e-05, + "loss": 1.0481, + "step": 1105 + }, + { + "epoch": 0.27034954778782694, + "grad_norm": 3.318326234817505, + "learning_rate": 1.7121273390216052e-05, + "loss": 1.0857, + "step": 1106 + }, + { + "epoch": 0.27059398680029334, + "grad_norm": 3.6183900833129883, + "learning_rate": 1.71157128016652e-05, + "loss": 1.0572, + "step": 1107 + }, + { + "epoch": 0.27083842581275974, + "grad_norm": 15.032169342041016, + "learning_rate": 1.711014775270408e-05, + "loss": 1.0478, + "step": 1108 + }, + { + "epoch": 0.2710828648252261, + "grad_norm": 3.319171190261841, + "learning_rate": 1.7104578246821083e-05, + "loss": 0.9196, + "step": 1109 + }, + { + "epoch": 0.2713273038376925, + "grad_norm": 3.304271697998047, + "learning_rate": 1.70990042875074e-05, + "loss": 2.085, + "step": 1110 + }, + { + "epoch": 0.2715717428501589, + "grad_norm": 4.110102653503418, + "learning_rate": 1.7093425878257007e-05, + "loss": 0.9727, + "step": 1111 + }, + { + "epoch": 0.2718161818626253, + "grad_norm": 6.215137958526611, + "learning_rate": 1.708784302256667e-05, + "loss": 1.1239, + "step": 1112 + }, + { + "epoch": 0.27206062087509164, + "grad_norm": 5.920506477355957, + "learning_rate": 1.7082255723935944e-05, + "loss": 1.1007, + "step": 1113 + }, + { + "epoch": 0.27230505988755804, + "grad_norm": 6.019047260284424, + "learning_rate": 1.7076663985867174e-05, + "loss": 1.1096, + "step": 1114 + }, + { + "epoch": 0.27254949890002445, + "grad_norm": 3.635051727294922, + "learning_rate": 1.7071067811865477e-05, + "loss": 1.1436, + "step": 1115 + }, + { + "epoch": 0.27279393791249085, + "grad_norm": 4.673436164855957, + "learning_rate": 1.706546720543876e-05, + "loss": 0.9838, + "step": 1116 + }, + { + "epoch": 0.27303837692495725, + "grad_norm": 2.6015164852142334, + "learning_rate": 1.7059862170097696e-05, + "loss": 2.0027, + "step": 1117 + }, + { + "epoch": 0.2732828159374236, + "grad_norm": 4.968930244445801, + "learning_rate": 1.7054252709355754e-05, + "loss": 0.9975, + "step": 1118 + }, + { + "epoch": 0.27352725494989, + "grad_norm": 2.573199510574341, + "learning_rate": 1.7048638826729163e-05, + "loss": 2.0794, + "step": 1119 + }, + { + "epoch": 0.2737716939623564, + "grad_norm": 4.521797180175781, + "learning_rate": 1.704302052573692e-05, + "loss": 0.9813, + "step": 1120 + }, + { + "epoch": 0.2740161329748228, + "grad_norm": 3.8551506996154785, + "learning_rate": 1.7037397809900807e-05, + "loss": 1.1006, + "step": 1121 + }, + { + "epoch": 0.27426057198728915, + "grad_norm": 3.2936413288116455, + "learning_rate": 1.703177068274536e-05, + "loss": 1.0544, + "step": 1122 + }, + { + "epoch": 0.27450501099975555, + "grad_norm": 3.572744846343994, + "learning_rate": 1.7026139147797887e-05, + "loss": 1.0277, + "step": 1123 + }, + { + "epoch": 0.27474945001222195, + "grad_norm": 3.7930381298065186, + "learning_rate": 1.7020503208588456e-05, + "loss": 1.0485, + "step": 1124 + }, + { + "epoch": 0.27499388902468835, + "grad_norm": 3.3152592182159424, + "learning_rate": 1.7014862868649894e-05, + "loss": 0.9606, + "step": 1125 + }, + { + "epoch": 0.27523832803715476, + "grad_norm": 4.435396671295166, + "learning_rate": 1.7009218131517795e-05, + "loss": 1.0386, + "step": 1126 + }, + { + "epoch": 0.2754827670496211, + "grad_norm": 3.8358654975891113, + "learning_rate": 1.7003569000730498e-05, + "loss": 1.0104, + "step": 1127 + }, + { + "epoch": 0.2757272060620875, + "grad_norm": 4.017136096954346, + "learning_rate": 1.6997915479829104e-05, + "loss": 0.9864, + "step": 1128 + }, + { + "epoch": 0.2759716450745539, + "grad_norm": 4.429943084716797, + "learning_rate": 1.6992257572357465e-05, + "loss": 1.0166, + "step": 1129 + }, + { + "epoch": 0.2762160840870203, + "grad_norm": 4.847922325134277, + "learning_rate": 1.6986595281862177e-05, + "loss": 1.0326, + "step": 1130 + }, + { + "epoch": 0.27646052309948665, + "grad_norm": 2.732870101928711, + "learning_rate": 1.698092861189259e-05, + "loss": 2.1184, + "step": 1131 + }, + { + "epoch": 0.27670496211195306, + "grad_norm": 3.975236654281616, + "learning_rate": 1.69752575660008e-05, + "loss": 1.0131, + "step": 1132 + }, + { + "epoch": 0.27694940112441946, + "grad_norm": 4.296026229858398, + "learning_rate": 1.696958214774164e-05, + "loss": 0.9962, + "step": 1133 + }, + { + "epoch": 0.27719384013688586, + "grad_norm": 3.756542921066284, + "learning_rate": 1.6963902360672684e-05, + "loss": 1.0905, + "step": 1134 + }, + { + "epoch": 0.27743827914935226, + "grad_norm": 3.982203960418701, + "learning_rate": 1.695821820835425e-05, + "loss": 1.1649, + "step": 1135 + }, + { + "epoch": 0.2776827181618186, + "grad_norm": 4.048048973083496, + "learning_rate": 1.6952529694349397e-05, + "loss": 1.0912, + "step": 1136 + }, + { + "epoch": 0.277927157174285, + "grad_norm": 3.5503664016723633, + "learning_rate": 1.6946836822223895e-05, + "loss": 0.9794, + "step": 1137 + }, + { + "epoch": 0.2781715961867514, + "grad_norm": 2.99806547164917, + "learning_rate": 1.6941139595546273e-05, + "loss": 0.9834, + "step": 1138 + }, + { + "epoch": 0.2784160351992178, + "grad_norm": 3.380628824234009, + "learning_rate": 1.693543801788777e-05, + "loss": 1.0707, + "step": 1139 + }, + { + "epoch": 0.27866047421168416, + "grad_norm": 4.29387903213501, + "learning_rate": 1.6929732092822372e-05, + "loss": 1.0751, + "step": 1140 + }, + { + "epoch": 0.27890491322415056, + "grad_norm": 3.4037857055664062, + "learning_rate": 1.6924021823926766e-05, + "loss": 1.031, + "step": 1141 + }, + { + "epoch": 0.27914935223661697, + "grad_norm": 4.047512054443359, + "learning_rate": 1.6918307214780382e-05, + "loss": 1.1181, + "step": 1142 + }, + { + "epoch": 0.27939379124908337, + "grad_norm": 3.380204677581787, + "learning_rate": 1.6912588268965363e-05, + "loss": 1.1261, + "step": 1143 + }, + { + "epoch": 0.27963823026154977, + "grad_norm": 3.7261037826538086, + "learning_rate": 1.6906864990066563e-05, + "loss": 1.0642, + "step": 1144 + }, + { + "epoch": 0.2798826692740161, + "grad_norm": 3.2923035621643066, + "learning_rate": 1.690113738167157e-05, + "loss": 1.1016, + "step": 1145 + }, + { + "epoch": 0.2801271082864825, + "grad_norm": 3.73152756690979, + "learning_rate": 1.689540544737067e-05, + "loss": 0.9956, + "step": 1146 + }, + { + "epoch": 0.2803715472989489, + "grad_norm": 3.534151077270508, + "learning_rate": 1.688966919075687e-05, + "loss": 1.0825, + "step": 1147 + }, + { + "epoch": 0.2806159863114153, + "grad_norm": 3.604247808456421, + "learning_rate": 1.688392861542587e-05, + "loss": 0.9772, + "step": 1148 + }, + { + "epoch": 0.28086042532388167, + "grad_norm": 2.977870464324951, + "learning_rate": 1.6878183724976107e-05, + "loss": 0.9825, + "step": 1149 + }, + { + "epoch": 0.28110486433634807, + "grad_norm": 4.251736164093018, + "learning_rate": 1.6872434523008697e-05, + "loss": 0.9403, + "step": 1150 + }, + { + "epoch": 0.2813493033488145, + "grad_norm": 5.256099700927734, + "learning_rate": 1.6866681013127466e-05, + "loss": 1.0369, + "step": 1151 + }, + { + "epoch": 0.2815937423612809, + "grad_norm": 4.620885848999023, + "learning_rate": 1.6860923198938945e-05, + "loss": 1.1105, + "step": 1152 + }, + { + "epoch": 0.2818381813737473, + "grad_norm": 5.067963600158691, + "learning_rate": 1.6855161084052358e-05, + "loss": 1.02, + "step": 1153 + }, + { + "epoch": 0.2820826203862136, + "grad_norm": 3.8163278102874756, + "learning_rate": 1.6849394672079625e-05, + "loss": 0.9542, + "step": 1154 + }, + { + "epoch": 0.28232705939868, + "grad_norm": 4.469657897949219, + "learning_rate": 1.6843623966635364e-05, + "loss": 0.9513, + "step": 1155 + }, + { + "epoch": 0.2825714984111464, + "grad_norm": 3.72050404548645, + "learning_rate": 1.683784897133688e-05, + "loss": 1.0148, + "step": 1156 + }, + { + "epoch": 0.28281593742361283, + "grad_norm": 5.091887474060059, + "learning_rate": 1.683206968980417e-05, + "loss": 0.9674, + "step": 1157 + }, + { + "epoch": 0.2830603764360792, + "grad_norm": 3.263423442840576, + "learning_rate": 1.6826286125659913e-05, + "loss": 1.0073, + "step": 1158 + }, + { + "epoch": 0.2833048154485456, + "grad_norm": 4.400149345397949, + "learning_rate": 1.682049828252948e-05, + "loss": 1.0492, + "step": 1159 + }, + { + "epoch": 0.283549254461012, + "grad_norm": 3.1265735626220703, + "learning_rate": 1.6814706164040913e-05, + "loss": 1.0499, + "step": 1160 + }, + { + "epoch": 0.2837936934734784, + "grad_norm": 4.412779331207275, + "learning_rate": 1.6808909773824952e-05, + "loss": 0.9605, + "step": 1161 + }, + { + "epoch": 0.2840381324859448, + "grad_norm": 4.249974250793457, + "learning_rate": 1.6803109115515e-05, + "loss": 1.0151, + "step": 1162 + }, + { + "epoch": 0.28428257149841113, + "grad_norm": 3.751030445098877, + "learning_rate": 1.679730419274713e-05, + "loss": 1.029, + "step": 1163 + }, + { + "epoch": 0.28452701051087753, + "grad_norm": 2.881483316421509, + "learning_rate": 1.679149500916011e-05, + "loss": 2.0311, + "step": 1164 + }, + { + "epoch": 0.28477144952334393, + "grad_norm": 6.882498264312744, + "learning_rate": 1.6785681568395353e-05, + "loss": 1.0385, + "step": 1165 + }, + { + "epoch": 0.28501588853581034, + "grad_norm": 3.647284507751465, + "learning_rate": 1.677986387409696e-05, + "loss": 1.1402, + "step": 1166 + }, + { + "epoch": 0.2852603275482767, + "grad_norm": 4.305856227874756, + "learning_rate": 1.67740419299117e-05, + "loss": 1.0261, + "step": 1167 + }, + { + "epoch": 0.2855047665607431, + "grad_norm": 3.294780731201172, + "learning_rate": 1.6768215739488986e-05, + "loss": 0.9288, + "step": 1168 + }, + { + "epoch": 0.2857492055732095, + "grad_norm": 3.736102342605591, + "learning_rate": 1.676238530648091e-05, + "loss": 0.8749, + "step": 1169 + }, + { + "epoch": 0.2859936445856759, + "grad_norm": 4.358032703399658, + "learning_rate": 1.6756550634542218e-05, + "loss": 1.0092, + "step": 1170 + }, + { + "epoch": 0.2862380835981423, + "grad_norm": 3.4615347385406494, + "learning_rate": 1.675071172733031e-05, + "loss": 1.1267, + "step": 1171 + }, + { + "epoch": 0.28648252261060864, + "grad_norm": 3.5754973888397217, + "learning_rate": 1.674486858850525e-05, + "loss": 2.041, + "step": 1172 + }, + { + "epoch": 0.28672696162307504, + "grad_norm": 3.58276629447937, + "learning_rate": 1.6739021221729743e-05, + "loss": 0.9799, + "step": 1173 + }, + { + "epoch": 0.28697140063554144, + "grad_norm": 2.8018088340759277, + "learning_rate": 1.673316963066915e-05, + "loss": 0.9621, + "step": 1174 + }, + { + "epoch": 0.28721583964800784, + "grad_norm": 3.283891201019287, + "learning_rate": 1.6727313818991485e-05, + "loss": 1.0314, + "step": 1175 + }, + { + "epoch": 0.2874602786604742, + "grad_norm": 3.2494401931762695, + "learning_rate": 1.6721453790367396e-05, + "loss": 1.0646, + "step": 1176 + }, + { + "epoch": 0.2877047176729406, + "grad_norm": 3.952605724334717, + "learning_rate": 1.6715589548470187e-05, + "loss": 1.0435, + "step": 1177 + }, + { + "epoch": 0.287949156685407, + "grad_norm": 5.484105110168457, + "learning_rate": 1.670972109697579e-05, + "loss": 0.9748, + "step": 1178 + }, + { + "epoch": 0.2881935956978734, + "grad_norm": 3.5420234203338623, + "learning_rate": 1.6703848439562787e-05, + "loss": 0.9723, + "step": 1179 + }, + { + "epoch": 0.2884380347103398, + "grad_norm": 3.4928793907165527, + "learning_rate": 1.6697971579912392e-05, + "loss": 1.0251, + "step": 1180 + }, + { + "epoch": 0.28868247372280614, + "grad_norm": 3.41666579246521, + "learning_rate": 1.669209052170845e-05, + "loss": 1.0639, + "step": 1181 + }, + { + "epoch": 0.28892691273527255, + "grad_norm": 4.020567417144775, + "learning_rate": 1.6686205268637444e-05, + "loss": 1.0303, + "step": 1182 + }, + { + "epoch": 0.28917135174773895, + "grad_norm": 3.600407361984253, + "learning_rate": 1.668031582438848e-05, + "loss": 1.0419, + "step": 1183 + }, + { + "epoch": 0.28941579076020535, + "grad_norm": 3.676823377609253, + "learning_rate": 1.6674422192653305e-05, + "loss": 1.0094, + "step": 1184 + }, + { + "epoch": 0.2896602297726717, + "grad_norm": 3.4822778701782227, + "learning_rate": 1.666852437712626e-05, + "loss": 1.0277, + "step": 1185 + }, + { + "epoch": 0.2899046687851381, + "grad_norm": 3.104646921157837, + "learning_rate": 1.6662622381504353e-05, + "loss": 0.9741, + "step": 1186 + }, + { + "epoch": 0.2901491077976045, + "grad_norm": 3.453314781188965, + "learning_rate": 1.6656716209487175e-05, + "loss": 0.9603, + "step": 1187 + }, + { + "epoch": 0.2903935468100709, + "grad_norm": 4.3645524978637695, + "learning_rate": 1.6650805864776946e-05, + "loss": 1.0422, + "step": 1188 + }, + { + "epoch": 0.2906379858225373, + "grad_norm": 3.804931402206421, + "learning_rate": 1.664489135107851e-05, + "loss": 0.9541, + "step": 1189 + }, + { + "epoch": 0.29088242483500365, + "grad_norm": 3.7183544635772705, + "learning_rate": 1.6638972672099317e-05, + "loss": 0.9893, + "step": 1190 + }, + { + "epoch": 0.29112686384747005, + "grad_norm": 3.7971720695495605, + "learning_rate": 1.6633049831549424e-05, + "loss": 1.1019, + "step": 1191 + }, + { + "epoch": 0.29137130285993645, + "grad_norm": 3.7743401527404785, + "learning_rate": 1.6627122833141514e-05, + "loss": 0.9558, + "step": 1192 + }, + { + "epoch": 0.29161574187240286, + "grad_norm": 4.806497097015381, + "learning_rate": 1.6621191680590853e-05, + "loss": 0.9999, + "step": 1193 + }, + { + "epoch": 0.2918601808848692, + "grad_norm": 3.7626373767852783, + "learning_rate": 1.6615256377615324e-05, + "loss": 1.0013, + "step": 1194 + }, + { + "epoch": 0.2921046198973356, + "grad_norm": 44.19645690917969, + "learning_rate": 1.660931692793541e-05, + "loss": 1.0507, + "step": 1195 + }, + { + "epoch": 0.292349058909802, + "grad_norm": 3.0135140419006348, + "learning_rate": 1.6603373335274194e-05, + "loss": 2.162, + "step": 1196 + }, + { + "epoch": 0.2925934979222684, + "grad_norm": 3.414067029953003, + "learning_rate": 1.6597425603357357e-05, + "loss": 1.0797, + "step": 1197 + }, + { + "epoch": 0.2928379369347348, + "grad_norm": 4.47196626663208, + "learning_rate": 1.659147373591317e-05, + "loss": 0.865, + "step": 1198 + }, + { + "epoch": 0.29308237594720116, + "grad_norm": 4.882946014404297, + "learning_rate": 1.6585517736672498e-05, + "loss": 0.9721, + "step": 1199 + }, + { + "epoch": 0.29332681495966756, + "grad_norm": 4.430570125579834, + "learning_rate": 1.6579557609368803e-05, + "loss": 1.0138, + "step": 1200 + }, + { + "epoch": 0.29357125397213396, + "grad_norm": 3.383298397064209, + "learning_rate": 1.657359335773812e-05, + "loss": 0.9532, + "step": 1201 + }, + { + "epoch": 0.29381569298460036, + "grad_norm": 4.805011749267578, + "learning_rate": 1.656762498551908e-05, + "loss": 1.0646, + "step": 1202 + }, + { + "epoch": 0.2940601319970667, + "grad_norm": 3.1969213485717773, + "learning_rate": 1.65616524964529e-05, + "loss": 1.0311, + "step": 1203 + }, + { + "epoch": 0.2943045710095331, + "grad_norm": 3.598453998565674, + "learning_rate": 1.655567589428337e-05, + "loss": 0.9863, + "step": 1204 + }, + { + "epoch": 0.2945490100219995, + "grad_norm": 2.7105185985565186, + "learning_rate": 1.6549695182756857e-05, + "loss": 2.0363, + "step": 1205 + }, + { + "epoch": 0.2947934490344659, + "grad_norm": 4.3872599601745605, + "learning_rate": 1.654371036562231e-05, + "loss": 0.9133, + "step": 1206 + }, + { + "epoch": 0.2950378880469323, + "grad_norm": 4.585229873657227, + "learning_rate": 1.653772144663125e-05, + "loss": 1.1049, + "step": 1207 + }, + { + "epoch": 0.29528232705939866, + "grad_norm": 4.361045837402344, + "learning_rate": 1.6531728429537766e-05, + "loss": 1.1456, + "step": 1208 + }, + { + "epoch": 0.29552676607186507, + "grad_norm": 4.613654136657715, + "learning_rate": 1.6525731318098526e-05, + "loss": 1.0033, + "step": 1209 + }, + { + "epoch": 0.29577120508433147, + "grad_norm": 3.5829994678497314, + "learning_rate": 1.651973011607275e-05, + "loss": 0.9729, + "step": 1210 + }, + { + "epoch": 0.29601564409679787, + "grad_norm": 3.3050808906555176, + "learning_rate": 1.6513724827222225e-05, + "loss": 1.0571, + "step": 1211 + }, + { + "epoch": 0.2962600831092642, + "grad_norm": 3.048600673675537, + "learning_rate": 1.6507715455311308e-05, + "loss": 0.9471, + "step": 1212 + }, + { + "epoch": 0.2965045221217306, + "grad_norm": 3.568232297897339, + "learning_rate": 1.6501702004106914e-05, + "loss": 0.9478, + "step": 1213 + }, + { + "epoch": 0.296748961134197, + "grad_norm": 3.4384896755218506, + "learning_rate": 1.6495684477378507e-05, + "loss": 1.0493, + "step": 1214 + }, + { + "epoch": 0.2969934001466634, + "grad_norm": 3.981926918029785, + "learning_rate": 1.6489662878898115e-05, + "loss": 0.9608, + "step": 1215 + }, + { + "epoch": 0.2972378391591298, + "grad_norm": 3.6841025352478027, + "learning_rate": 1.6483637212440312e-05, + "loss": 0.9997, + "step": 1216 + }, + { + "epoch": 0.29748227817159617, + "grad_norm": 3.6886446475982666, + "learning_rate": 1.6477607481782222e-05, + "loss": 0.9751, + "step": 1217 + }, + { + "epoch": 0.2977267171840626, + "grad_norm": 3.5062296390533447, + "learning_rate": 1.6471573690703523e-05, + "loss": 0.9781, + "step": 1218 + }, + { + "epoch": 0.297971156196529, + "grad_norm": 4.244730472564697, + "learning_rate": 1.6465535842986433e-05, + "loss": 1.0561, + "step": 1219 + }, + { + "epoch": 0.2982155952089954, + "grad_norm": 3.478639602661133, + "learning_rate": 1.645949394241571e-05, + "loss": 1.0163, + "step": 1220 + }, + { + "epoch": 0.2984600342214617, + "grad_norm": 3.335463523864746, + "learning_rate": 1.645344799277866e-05, + "loss": 0.9661, + "step": 1221 + }, + { + "epoch": 0.2987044732339281, + "grad_norm": 3.1549289226531982, + "learning_rate": 1.6447397997865124e-05, + "loss": 0.902, + "step": 1222 + }, + { + "epoch": 0.2989489122463945, + "grad_norm": 3.6085996627807617, + "learning_rate": 1.6441343961467473e-05, + "loss": 1.0333, + "step": 1223 + }, + { + "epoch": 0.29919335125886093, + "grad_norm": 4.483367443084717, + "learning_rate": 1.643528588738062e-05, + "loss": 0.9929, + "step": 1224 + }, + { + "epoch": 0.29943779027132733, + "grad_norm": 3.825385808944702, + "learning_rate": 1.642922377940201e-05, + "loss": 1.0265, + "step": 1225 + }, + { + "epoch": 0.2996822292837937, + "grad_norm": 3.291959285736084, + "learning_rate": 1.6423157641331604e-05, + "loss": 0.9699, + "step": 1226 + }, + { + "epoch": 0.2999266682962601, + "grad_norm": 4.078056335449219, + "learning_rate": 1.64170874769719e-05, + "loss": 0.9933, + "step": 1227 + }, + { + "epoch": 0.3001711073087265, + "grad_norm": 3.92166805267334, + "learning_rate": 1.641101329012792e-05, + "loss": 0.9115, + "step": 1228 + }, + { + "epoch": 0.3004155463211929, + "grad_norm": 4.900821208953857, + "learning_rate": 1.6404935084607202e-05, + "loss": 0.9236, + "step": 1229 + }, + { + "epoch": 0.30065998533365923, + "grad_norm": 2.878377914428711, + "learning_rate": 1.6398852864219807e-05, + "loss": 2.0352, + "step": 1230 + }, + { + "epoch": 0.30090442434612563, + "grad_norm": 3.8259389400482178, + "learning_rate": 1.639276663277831e-05, + "loss": 1.0943, + "step": 1231 + }, + { + "epoch": 0.30114886335859203, + "grad_norm": 4.231867790222168, + "learning_rate": 1.63866763940978e-05, + "loss": 0.9894, + "step": 1232 + }, + { + "epoch": 0.30139330237105844, + "grad_norm": 5.750034809112549, + "learning_rate": 1.6380582151995886e-05, + "loss": 1.0019, + "step": 1233 + }, + { + "epoch": 0.30163774138352484, + "grad_norm": 3.387482166290283, + "learning_rate": 1.6374483910292675e-05, + "loss": 0.9352, + "step": 1234 + }, + { + "epoch": 0.3018821803959912, + "grad_norm": 4.67958402633667, + "learning_rate": 1.6368381672810788e-05, + "loss": 1.0685, + "step": 1235 + }, + { + "epoch": 0.3021266194084576, + "grad_norm": 4.168334007263184, + "learning_rate": 1.6362275443375343e-05, + "loss": 0.8781, + "step": 1236 + }, + { + "epoch": 0.302371058420924, + "grad_norm": 3.400838851928711, + "learning_rate": 1.6356165225813975e-05, + "loss": 0.9975, + "step": 1237 + }, + { + "epoch": 0.3026154974333904, + "grad_norm": 3.742551326751709, + "learning_rate": 1.6350051023956806e-05, + "loss": 0.9164, + "step": 1238 + }, + { + "epoch": 0.30285993644585674, + "grad_norm": 4.234443187713623, + "learning_rate": 1.6343932841636455e-05, + "loss": 1.1713, + "step": 1239 + }, + { + "epoch": 0.30310437545832314, + "grad_norm": 3.2706210613250732, + "learning_rate": 1.633781068268805e-05, + "loss": 1.0239, + "step": 1240 + }, + { + "epoch": 0.30334881447078954, + "grad_norm": 3.5965492725372314, + "learning_rate": 1.6331684550949197e-05, + "loss": 0.9647, + "step": 1241 + }, + { + "epoch": 0.30359325348325594, + "grad_norm": 3.227299213409424, + "learning_rate": 1.6325554450259997e-05, + "loss": 1.0027, + "step": 1242 + }, + { + "epoch": 0.30383769249572234, + "grad_norm": 4.168364524841309, + "learning_rate": 1.631942038446304e-05, + "loss": 0.9438, + "step": 1243 + }, + { + "epoch": 0.3040821315081887, + "grad_norm": 3.197082042694092, + "learning_rate": 1.6313282357403407e-05, + "loss": 1.0672, + "step": 1244 + }, + { + "epoch": 0.3043265705206551, + "grad_norm": 3.9803450107574463, + "learning_rate": 1.630714037292865e-05, + "loss": 1.0494, + "step": 1245 + }, + { + "epoch": 0.3045710095331215, + "grad_norm": 3.74851393699646, + "learning_rate": 1.6300994434888805e-05, + "loss": 1.072, + "step": 1246 + }, + { + "epoch": 0.3048154485455879, + "grad_norm": 3.272641658782959, + "learning_rate": 1.6294844547136394e-05, + "loss": 1.0109, + "step": 1247 + }, + { + "epoch": 0.30505988755805424, + "grad_norm": 3.4987287521362305, + "learning_rate": 1.628869071352641e-05, + "loss": 0.9238, + "step": 1248 + }, + { + "epoch": 0.30530432657052065, + "grad_norm": 3.6942923069000244, + "learning_rate": 1.628253293791632e-05, + "loss": 0.9616, + "step": 1249 + }, + { + "epoch": 0.30554876558298705, + "grad_norm": 4.471048355102539, + "learning_rate": 1.627637122416606e-05, + "loss": 0.9523, + "step": 1250 + }, + { + "epoch": 0.30579320459545345, + "grad_norm": 3.160109758377075, + "learning_rate": 1.627020557613803e-05, + "loss": 0.9743, + "step": 1251 + }, + { + "epoch": 0.30603764360791985, + "grad_norm": 3.819319248199463, + "learning_rate": 1.6264035997697118e-05, + "loss": 1.0451, + "step": 1252 + }, + { + "epoch": 0.3062820826203862, + "grad_norm": 4.501501083374023, + "learning_rate": 1.625786249271064e-05, + "loss": 1.0177, + "step": 1253 + }, + { + "epoch": 0.3065265216328526, + "grad_norm": 3.6937527656555176, + "learning_rate": 1.6251685065048408e-05, + "loss": 0.9424, + "step": 1254 + }, + { + "epoch": 0.306770960645319, + "grad_norm": 3.759077548980713, + "learning_rate": 1.624550371858267e-05, + "loss": 1.1124, + "step": 1255 + }, + { + "epoch": 0.3070153996577854, + "grad_norm": 2.905395984649658, + "learning_rate": 1.6239318457188146e-05, + "loss": 2.2317, + "step": 1256 + }, + { + "epoch": 0.30725983867025175, + "grad_norm": 3.678001880645752, + "learning_rate": 1.6233129284741992e-05, + "loss": 1.1202, + "step": 1257 + }, + { + "epoch": 0.30750427768271815, + "grad_norm": 4.4421186447143555, + "learning_rate": 1.6226936205123835e-05, + "loss": 1.0008, + "step": 1258 + }, + { + "epoch": 0.30774871669518455, + "grad_norm": 3.2402305603027344, + "learning_rate": 1.6220739222215738e-05, + "loss": 0.9972, + "step": 1259 + }, + { + "epoch": 0.30799315570765096, + "grad_norm": 3.374966859817505, + "learning_rate": 1.621453833990222e-05, + "loss": 0.9716, + "step": 1260 + }, + { + "epoch": 0.30823759472011736, + "grad_norm": 3.3255765438079834, + "learning_rate": 1.6208333562070232e-05, + "loss": 1.0439, + "step": 1261 + }, + { + "epoch": 0.3084820337325837, + "grad_norm": 3.0071613788604736, + "learning_rate": 1.6202124892609175e-05, + "loss": 0.8857, + "step": 1262 + }, + { + "epoch": 0.3087264727450501, + "grad_norm": 3.1655237674713135, + "learning_rate": 1.6195912335410896e-05, + "loss": 0.9687, + "step": 1263 + }, + { + "epoch": 0.3089709117575165, + "grad_norm": 3.5941660404205322, + "learning_rate": 1.618969589436967e-05, + "loss": 1.0306, + "step": 1264 + }, + { + "epoch": 0.3092153507699829, + "grad_norm": 3.753988742828369, + "learning_rate": 1.6183475573382196e-05, + "loss": 0.9762, + "step": 1265 + }, + { + "epoch": 0.30945978978244926, + "grad_norm": 3.202784538269043, + "learning_rate": 1.6177251376347632e-05, + "loss": 1.0616, + "step": 1266 + }, + { + "epoch": 0.30970422879491566, + "grad_norm": 2.9768574237823486, + "learning_rate": 1.6171023307167545e-05, + "loss": 1.047, + "step": 1267 + }, + { + "epoch": 0.30994866780738206, + "grad_norm": 3.3893463611602783, + "learning_rate": 1.6164791369745935e-05, + "loss": 0.9366, + "step": 1268 + }, + { + "epoch": 0.31019310681984846, + "grad_norm": 3.303011178970337, + "learning_rate": 1.615855556798923e-05, + "loss": 0.9906, + "step": 1269 + }, + { + "epoch": 0.31043754583231487, + "grad_norm": 2.891843557357788, + "learning_rate": 1.615231590580627e-05, + "loss": 0.8647, + "step": 1270 + }, + { + "epoch": 0.3106819848447812, + "grad_norm": 3.257838487625122, + "learning_rate": 1.614607238710833e-05, + "loss": 0.8725, + "step": 1271 + }, + { + "epoch": 0.3109264238572476, + "grad_norm": 4.155395984649658, + "learning_rate": 1.613982501580909e-05, + "loss": 0.9151, + "step": 1272 + }, + { + "epoch": 0.311170862869714, + "grad_norm": 3.5042197704315186, + "learning_rate": 1.613357379582465e-05, + "loss": 0.9406, + "step": 1273 + }, + { + "epoch": 0.3114153018821804, + "grad_norm": 4.06215763092041, + "learning_rate": 1.612731873107352e-05, + "loss": 1.0614, + "step": 1274 + }, + { + "epoch": 0.31165974089464676, + "grad_norm": 3.6693708896636963, + "learning_rate": 1.612105982547663e-05, + "loss": 0.9339, + "step": 1275 + }, + { + "epoch": 0.31190417990711317, + "grad_norm": 3.698263645172119, + "learning_rate": 1.6114797082957304e-05, + "loss": 0.9648, + "step": 1276 + }, + { + "epoch": 0.31214861891957957, + "grad_norm": 3.2180988788604736, + "learning_rate": 1.6108530507441273e-05, + "loss": 2.0458, + "step": 1277 + }, + { + "epoch": 0.31239305793204597, + "grad_norm": 4.171623229980469, + "learning_rate": 1.610226010285668e-05, + "loss": 1.0276, + "step": 1278 + }, + { + "epoch": 0.3126374969445123, + "grad_norm": 3.3476104736328125, + "learning_rate": 1.6095985873134063e-05, + "loss": 1.054, + "step": 1279 + }, + { + "epoch": 0.3128819359569787, + "grad_norm": 3.6903676986694336, + "learning_rate": 1.6089707822206353e-05, + "loss": 0.8384, + "step": 1280 + }, + { + "epoch": 0.3131263749694451, + "grad_norm": 3.4239535331726074, + "learning_rate": 1.6083425954008883e-05, + "loss": 1.0724, + "step": 1281 + }, + { + "epoch": 0.3133708139819115, + "grad_norm": 3.497863292694092, + "learning_rate": 1.6077140272479378e-05, + "loss": 1.0717, + "step": 1282 + }, + { + "epoch": 0.3136152529943779, + "grad_norm": 3.6051435470581055, + "learning_rate": 1.607085078155795e-05, + "loss": 0.9298, + "step": 1283 + }, + { + "epoch": 0.31385969200684427, + "grad_norm": 3.1399309635162354, + "learning_rate": 1.6064557485187104e-05, + "loss": 0.9527, + "step": 1284 + }, + { + "epoch": 0.3141041310193107, + "grad_norm": 3.3739049434661865, + "learning_rate": 1.605826038731172e-05, + "loss": 0.9738, + "step": 1285 + }, + { + "epoch": 0.3143485700317771, + "grad_norm": 3.3782050609588623, + "learning_rate": 1.6051959491879072e-05, + "loss": 0.9883, + "step": 1286 + }, + { + "epoch": 0.3145930090442435, + "grad_norm": 3.374809503555298, + "learning_rate": 1.6045654802838812e-05, + "loss": 0.9174, + "step": 1287 + }, + { + "epoch": 0.3148374480567098, + "grad_norm": 3.3350439071655273, + "learning_rate": 1.603934632414297e-05, + "loss": 0.9236, + "step": 1288 + }, + { + "epoch": 0.3150818870691762, + "grad_norm": 3.8035969734191895, + "learning_rate": 1.6033034059745943e-05, + "loss": 0.8416, + "step": 1289 + }, + { + "epoch": 0.3153263260816426, + "grad_norm": 3.8607382774353027, + "learning_rate": 1.6026718013604514e-05, + "loss": 1.0511, + "step": 1290 + }, + { + "epoch": 0.31557076509410903, + "grad_norm": 3.493785858154297, + "learning_rate": 1.602039818967783e-05, + "loss": 1.0373, + "step": 1291 + }, + { + "epoch": 0.31581520410657543, + "grad_norm": 3.8863096237182617, + "learning_rate": 1.6014074591927405e-05, + "loss": 0.9358, + "step": 1292 + }, + { + "epoch": 0.3160596431190418, + "grad_norm": 3.0080204010009766, + "learning_rate": 1.6007747224317124e-05, + "loss": 0.9329, + "step": 1293 + }, + { + "epoch": 0.3163040821315082, + "grad_norm": 4.005444526672363, + "learning_rate": 1.6001416090813224e-05, + "loss": 1.0229, + "step": 1294 + }, + { + "epoch": 0.3165485211439746, + "grad_norm": 3.2476863861083984, + "learning_rate": 1.5995081195384314e-05, + "loss": 0.9798, + "step": 1295 + }, + { + "epoch": 0.316792960156441, + "grad_norm": 3.350329637527466, + "learning_rate": 1.598874254200136e-05, + "loss": 1.0878, + "step": 1296 + }, + { + "epoch": 0.31703739916890733, + "grad_norm": 3.887662649154663, + "learning_rate": 1.5982400134637674e-05, + "loss": 1.1045, + "step": 1297 + }, + { + "epoch": 0.31728183818137373, + "grad_norm": 3.1978490352630615, + "learning_rate": 1.5976053977268934e-05, + "loss": 0.9608, + "step": 1298 + }, + { + "epoch": 0.31752627719384013, + "grad_norm": 3.234931707382202, + "learning_rate": 1.5969704073873157e-05, + "loss": 1.1088, + "step": 1299 + }, + { + "epoch": 0.31777071620630654, + "grad_norm": 3.477632761001587, + "learning_rate": 1.5963350428430722e-05, + "loss": 0.9836, + "step": 1300 + }, + { + "epoch": 0.31801515521877294, + "grad_norm": 3.838231325149536, + "learning_rate": 1.5956993044924334e-05, + "loss": 1.0023, + "step": 1301 + }, + { + "epoch": 0.3182595942312393, + "grad_norm": 3.980022430419922, + "learning_rate": 1.595063192733906e-05, + "loss": 1.0221, + "step": 1302 + }, + { + "epoch": 0.3185040332437057, + "grad_norm": 3.580064296722412, + "learning_rate": 1.5944267079662302e-05, + "loss": 1.0318, + "step": 1303 + }, + { + "epoch": 0.3187484722561721, + "grad_norm": 3.1863508224487305, + "learning_rate": 1.5937898505883794e-05, + "loss": 1.0015, + "step": 1304 + }, + { + "epoch": 0.3189929112686385, + "grad_norm": 3.451911211013794, + "learning_rate": 1.5931526209995608e-05, + "loss": 0.9728, + "step": 1305 + }, + { + "epoch": 0.31923735028110484, + "grad_norm": 2.908261299133301, + "learning_rate": 1.5925150195992156e-05, + "loss": 2.0991, + "step": 1306 + }, + { + "epoch": 0.31948178929357124, + "grad_norm": 3.784224033355713, + "learning_rate": 1.5918770467870174e-05, + "loss": 1.0714, + "step": 1307 + }, + { + "epoch": 0.31972622830603764, + "grad_norm": 3.7832584381103516, + "learning_rate": 1.5912387029628726e-05, + "loss": 1.0196, + "step": 1308 + }, + { + "epoch": 0.31997066731850404, + "grad_norm": 3.057114601135254, + "learning_rate": 1.5905999885269208e-05, + "loss": 1.0451, + "step": 1309 + }, + { + "epoch": 0.32021510633097044, + "grad_norm": 3.717481851577759, + "learning_rate": 1.5899609038795333e-05, + "loss": 0.8955, + "step": 1310 + }, + { + "epoch": 0.3204595453434368, + "grad_norm": 3.5934057235717773, + "learning_rate": 1.589321449421313e-05, + "loss": 0.9301, + "step": 1311 + }, + { + "epoch": 0.3207039843559032, + "grad_norm": 3.7299716472625732, + "learning_rate": 1.5886816255530964e-05, + "loss": 0.9594, + "step": 1312 + }, + { + "epoch": 0.3209484233683696, + "grad_norm": 3.0106871128082275, + "learning_rate": 1.5880414326759493e-05, + "loss": 1.0206, + "step": 1313 + }, + { + "epoch": 0.321192862380836, + "grad_norm": 3.8183209896087646, + "learning_rate": 1.5874008711911706e-05, + "loss": 0.9222, + "step": 1314 + }, + { + "epoch": 0.32143730139330234, + "grad_norm": 3.292402744293213, + "learning_rate": 1.5867599415002896e-05, + "loss": 0.9671, + "step": 1315 + }, + { + "epoch": 0.32168174040576875, + "grad_norm": 3.2833142280578613, + "learning_rate": 1.5861186440050663e-05, + "loss": 1.0283, + "step": 1316 + }, + { + "epoch": 0.32192617941823515, + "grad_norm": 3.5037293434143066, + "learning_rate": 1.5854769791074906e-05, + "loss": 1.0508, + "step": 1317 + }, + { + "epoch": 0.32217061843070155, + "grad_norm": 3.340787172317505, + "learning_rate": 1.5848349472097848e-05, + "loss": 0.9573, + "step": 1318 + }, + { + "epoch": 0.32241505744316795, + "grad_norm": 8.840167999267578, + "learning_rate": 1.5841925487143986e-05, + "loss": 0.9666, + "step": 1319 + }, + { + "epoch": 0.3226594964556343, + "grad_norm": 3.3367087841033936, + "learning_rate": 1.583549784024014e-05, + "loss": 0.9763, + "step": 1320 + }, + { + "epoch": 0.3229039354681007, + "grad_norm": 3.5148043632507324, + "learning_rate": 1.5829066535415402e-05, + "loss": 0.9383, + "step": 1321 + }, + { + "epoch": 0.3231483744805671, + "grad_norm": 3.5361037254333496, + "learning_rate": 1.5822631576701174e-05, + "loss": 0.977, + "step": 1322 + }, + { + "epoch": 0.3233928134930335, + "grad_norm": 3.2863235473632812, + "learning_rate": 1.5816192968131138e-05, + "loss": 0.9872, + "step": 1323 + }, + { + "epoch": 0.32363725250549985, + "grad_norm": 3.653019428253174, + "learning_rate": 1.5809750713741277e-05, + "loss": 0.8894, + "step": 1324 + }, + { + "epoch": 0.32388169151796625, + "grad_norm": 2.8651626110076904, + "learning_rate": 1.5803304817569847e-05, + "loss": 0.9137, + "step": 1325 + }, + { + "epoch": 0.32412613053043265, + "grad_norm": 2.9790525436401367, + "learning_rate": 1.579685528365739e-05, + "loss": 0.9451, + "step": 1326 + }, + { + "epoch": 0.32437056954289906, + "grad_norm": 3.41646146774292, + "learning_rate": 1.5790402116046725e-05, + "loss": 1.0006, + "step": 1327 + }, + { + "epoch": 0.32461500855536546, + "grad_norm": 3.762275218963623, + "learning_rate": 1.578394531878296e-05, + "loss": 0.9529, + "step": 1328 + }, + { + "epoch": 0.3248594475678318, + "grad_norm": 2.90238618850708, + "learning_rate": 1.5777484895913462e-05, + "loss": 0.9205, + "step": 1329 + }, + { + "epoch": 0.3251038865802982, + "grad_norm": 3.016227960586548, + "learning_rate": 1.5771020851487893e-05, + "loss": 0.9908, + "step": 1330 + }, + { + "epoch": 0.3253483255927646, + "grad_norm": 3.5532524585723877, + "learning_rate": 1.576455318955816e-05, + "loss": 0.9425, + "step": 1331 + }, + { + "epoch": 0.325592764605231, + "grad_norm": 3.1646616458892822, + "learning_rate": 1.5758081914178457e-05, + "loss": 0.8908, + "step": 1332 + }, + { + "epoch": 0.32583720361769736, + "grad_norm": 3.4908621311187744, + "learning_rate": 1.5751607029405226e-05, + "loss": 0.985, + "step": 1333 + }, + { + "epoch": 0.32608164263016376, + "grad_norm": 3.6697299480438232, + "learning_rate": 1.574512853929719e-05, + "loss": 1.0036, + "step": 1334 + }, + { + "epoch": 0.32632608164263016, + "grad_norm": 4.173372268676758, + "learning_rate": 1.573864644791532e-05, + "loss": 0.9331, + "step": 1335 + }, + { + "epoch": 0.32657052065509656, + "grad_norm": 2.790220260620117, + "learning_rate": 1.5732160759322847e-05, + "loss": 1.9576, + "step": 1336 + }, + { + "epoch": 0.32681495966756297, + "grad_norm": 4.081017971038818, + "learning_rate": 1.5725671477585255e-05, + "loss": 1.003, + "step": 1337 + }, + { + "epoch": 0.3270593986800293, + "grad_norm": 4.018110275268555, + "learning_rate": 1.5719178606770283e-05, + "loss": 1.0492, + "step": 1338 + }, + { + "epoch": 0.3273038376924957, + "grad_norm": 3.4949355125427246, + "learning_rate": 1.5712682150947926e-05, + "loss": 0.9217, + "step": 1339 + }, + { + "epoch": 0.3275482767049621, + "grad_norm": 3.6490862369537354, + "learning_rate": 1.5706182114190412e-05, + "loss": 0.8558, + "step": 1340 + }, + { + "epoch": 0.3277927157174285, + "grad_norm": 3.302030563354492, + "learning_rate": 1.569967850057222e-05, + "loss": 0.9701, + "step": 1341 + }, + { + "epoch": 0.32803715472989486, + "grad_norm": 3.829453706741333, + "learning_rate": 1.5693171314170078e-05, + "loss": 0.9252, + "step": 1342 + }, + { + "epoch": 0.32828159374236127, + "grad_norm": 2.87931227684021, + "learning_rate": 1.568666055906295e-05, + "loss": 0.9877, + "step": 1343 + }, + { + "epoch": 0.32852603275482767, + "grad_norm": 2.9058079719543457, + "learning_rate": 1.5680146239332024e-05, + "loss": 2.1484, + "step": 1344 + }, + { + "epoch": 0.32877047176729407, + "grad_norm": 3.9817657470703125, + "learning_rate": 1.567362835906074e-05, + "loss": 0.9291, + "step": 1345 + }, + { + "epoch": 0.32901491077976047, + "grad_norm": 3.298549175262451, + "learning_rate": 1.5667106922334763e-05, + "loss": 0.986, + "step": 1346 + }, + { + "epoch": 0.3292593497922268, + "grad_norm": 3.3551080226898193, + "learning_rate": 1.5660581933241994e-05, + "loss": 0.8962, + "step": 1347 + }, + { + "epoch": 0.3295037888046932, + "grad_norm": 2.3159947395324707, + "learning_rate": 1.565405339587254e-05, + "loss": 1.9325, + "step": 1348 + }, + { + "epoch": 0.3297482278171596, + "grad_norm": 3.5716497898101807, + "learning_rate": 1.564752131431876e-05, + "loss": 0.9268, + "step": 1349 + }, + { + "epoch": 0.329992666829626, + "grad_norm": 3.3860905170440674, + "learning_rate": 1.5640985692675213e-05, + "loss": 0.9784, + "step": 1350 + }, + { + "epoch": 0.33023710584209237, + "grad_norm": 3.5350513458251953, + "learning_rate": 1.5634446535038688e-05, + "loss": 1.0282, + "step": 1351 + }, + { + "epoch": 0.3304815448545588, + "grad_norm": 3.6100711822509766, + "learning_rate": 1.5627903845508184e-05, + "loss": 0.9305, + "step": 1352 + }, + { + "epoch": 0.3307259838670252, + "grad_norm": 3.686244487762451, + "learning_rate": 1.5621357628184924e-05, + "loss": 0.8673, + "step": 1353 + }, + { + "epoch": 0.3309704228794916, + "grad_norm": 2.6438686847686768, + "learning_rate": 1.5614807887172332e-05, + "loss": 2.0463, + "step": 1354 + }, + { + "epoch": 0.331214861891958, + "grad_norm": 3.5230236053466797, + "learning_rate": 1.5608254626576046e-05, + "loss": 0.8949, + "step": 1355 + }, + { + "epoch": 0.3314593009044243, + "grad_norm": 3.89335036277771, + "learning_rate": 1.5601697850503907e-05, + "loss": 0.8568, + "step": 1356 + }, + { + "epoch": 0.3317037399168907, + "grad_norm": 3.775359630584717, + "learning_rate": 1.5595137563065965e-05, + "loss": 1.0217, + "step": 1357 + }, + { + "epoch": 0.33194817892935713, + "grad_norm": 4.173966407775879, + "learning_rate": 1.5588573768374464e-05, + "loss": 0.9904, + "step": 1358 + }, + { + "epoch": 0.33219261794182353, + "grad_norm": 4.213604927062988, + "learning_rate": 1.5582006470543855e-05, + "loss": 1.0392, + "step": 1359 + }, + { + "epoch": 0.3324370569542899, + "grad_norm": 3.8599865436553955, + "learning_rate": 1.5575435673690774e-05, + "loss": 1.0709, + "step": 1360 + }, + { + "epoch": 0.3326814959667563, + "grad_norm": 3.6756699085235596, + "learning_rate": 1.556886138193406e-05, + "loss": 0.8283, + "step": 1361 + }, + { + "epoch": 0.3329259349792227, + "grad_norm": 3.2679405212402344, + "learning_rate": 1.556228359939474e-05, + "loss": 1.0301, + "step": 1362 + }, + { + "epoch": 0.3331703739916891, + "grad_norm": 3.217700719833374, + "learning_rate": 1.5555702330196024e-05, + "loss": 0.9494, + "step": 1363 + }, + { + "epoch": 0.3334148130041555, + "grad_norm": 3.4018821716308594, + "learning_rate": 1.5549117578463316e-05, + "loss": 0.8486, + "step": 1364 + }, + { + "epoch": 0.33365925201662183, + "grad_norm": 3.5787065029144287, + "learning_rate": 1.55425293483242e-05, + "loss": 1.0495, + "step": 1365 + }, + { + "epoch": 0.33390369102908823, + "grad_norm": 3.589322566986084, + "learning_rate": 1.5535937643908435e-05, + "loss": 0.948, + "step": 1366 + }, + { + "epoch": 0.33414813004155464, + "grad_norm": 3.325261116027832, + "learning_rate": 1.5529342469347958e-05, + "loss": 0.9936, + "step": 1367 + }, + { + "epoch": 0.33439256905402104, + "grad_norm": 3.6978461742401123, + "learning_rate": 1.5522743828776898e-05, + "loss": 0.9901, + "step": 1368 + }, + { + "epoch": 0.3346370080664874, + "grad_norm": 3.383035182952881, + "learning_rate": 1.551614172633154e-05, + "loss": 1.0426, + "step": 1369 + }, + { + "epoch": 0.3348814470789538, + "grad_norm": 3.634697198867798, + "learning_rate": 1.5509536166150334e-05, + "loss": 0.9485, + "step": 1370 + }, + { + "epoch": 0.3351258860914202, + "grad_norm": 3.330472230911255, + "learning_rate": 1.5502927152373913e-05, + "loss": 0.9106, + "step": 1371 + }, + { + "epoch": 0.3353703251038866, + "grad_norm": 2.659522771835327, + "learning_rate": 1.5496314689145072e-05, + "loss": 0.8889, + "step": 1372 + }, + { + "epoch": 0.335614764116353, + "grad_norm": 3.5730326175689697, + "learning_rate": 1.548969878060876e-05, + "loss": 0.9087, + "step": 1373 + }, + { + "epoch": 0.33585920312881934, + "grad_norm": 3.6332874298095703, + "learning_rate": 1.5483079430912088e-05, + "loss": 0.9751, + "step": 1374 + }, + { + "epoch": 0.33610364214128574, + "grad_norm": 3.5821280479431152, + "learning_rate": 1.5476456644204328e-05, + "loss": 0.9511, + "step": 1375 + }, + { + "epoch": 0.33634808115375214, + "grad_norm": 3.1708550453186035, + "learning_rate": 1.5469830424636906e-05, + "loss": 0.9146, + "step": 1376 + }, + { + "epoch": 0.33659252016621855, + "grad_norm": 3.0782763957977295, + "learning_rate": 1.54632007763634e-05, + "loss": 2.0277, + "step": 1377 + }, + { + "epoch": 0.3368369591786849, + "grad_norm": 3.970358371734619, + "learning_rate": 1.5456567703539532e-05, + "loss": 1.0644, + "step": 1378 + }, + { + "epoch": 0.3370813981911513, + "grad_norm": 4.540128231048584, + "learning_rate": 1.544993121032318e-05, + "loss": 0.8518, + "step": 1379 + }, + { + "epoch": 0.3373258372036177, + "grad_norm": 3.16619873046875, + "learning_rate": 1.5443291300874353e-05, + "loss": 0.9677, + "step": 1380 + }, + { + "epoch": 0.3375702762160841, + "grad_norm": 4.074929237365723, + "learning_rate": 1.5436647979355214e-05, + "loss": 0.9551, + "step": 1381 + }, + { + "epoch": 0.3378147152285505, + "grad_norm": 3.1865148544311523, + "learning_rate": 1.5430001249930057e-05, + "loss": 0.9697, + "step": 1382 + }, + { + "epoch": 0.33805915424101685, + "grad_norm": 3.514235019683838, + "learning_rate": 1.5423351116765316e-05, + "loss": 1.0022, + "step": 1383 + }, + { + "epoch": 0.33830359325348325, + "grad_norm": 3.3248844146728516, + "learning_rate": 1.541669758402956e-05, + "loss": 1.0536, + "step": 1384 + }, + { + "epoch": 0.33854803226594965, + "grad_norm": 3.2770636081695557, + "learning_rate": 1.5410040655893484e-05, + "loss": 1.0274, + "step": 1385 + }, + { + "epoch": 0.33879247127841605, + "grad_norm": 3.009883403778076, + "learning_rate": 1.5403380336529914e-05, + "loss": 0.8489, + "step": 1386 + }, + { + "epoch": 0.3390369102908824, + "grad_norm": 3.4994874000549316, + "learning_rate": 1.53967166301138e-05, + "loss": 1.0033, + "step": 1387 + }, + { + "epoch": 0.3392813493033488, + "grad_norm": 3.02510666847229, + "learning_rate": 1.5390049540822223e-05, + "loss": 0.9407, + "step": 1388 + }, + { + "epoch": 0.3395257883158152, + "grad_norm": 3.872352361679077, + "learning_rate": 1.5383379072834373e-05, + "loss": 0.9573, + "step": 1389 + }, + { + "epoch": 0.3397702273282816, + "grad_norm": 3.6080379486083984, + "learning_rate": 1.537670523033156e-05, + "loss": 0.9771, + "step": 1390 + }, + { + "epoch": 0.340014666340748, + "grad_norm": 3.544776678085327, + "learning_rate": 1.5370028017497217e-05, + "loss": 1.0244, + "step": 1391 + }, + { + "epoch": 0.34025910535321435, + "grad_norm": 3.672668933868408, + "learning_rate": 1.536334743851689e-05, + "loss": 0.8461, + "step": 1392 + }, + { + "epoch": 0.34050354436568075, + "grad_norm": 3.6929850578308105, + "learning_rate": 1.5356663497578223e-05, + "loss": 0.9417, + "step": 1393 + }, + { + "epoch": 0.34074798337814716, + "grad_norm": 3.087480306625366, + "learning_rate": 1.5349976198870974e-05, + "loss": 0.9396, + "step": 1394 + }, + { + "epoch": 0.34099242239061356, + "grad_norm": 3.1820108890533447, + "learning_rate": 1.534328554658701e-05, + "loss": 0.9691, + "step": 1395 + }, + { + "epoch": 0.3412368614030799, + "grad_norm": 4.0018630027771, + "learning_rate": 1.5336591544920297e-05, + "loss": 0.9179, + "step": 1396 + }, + { + "epoch": 0.3414813004155463, + "grad_norm": 3.438683271408081, + "learning_rate": 1.53298941980669e-05, + "loss": 1.0616, + "step": 1397 + }, + { + "epoch": 0.3417257394280127, + "grad_norm": 3.162782669067383, + "learning_rate": 1.5323193510224973e-05, + "loss": 0.9266, + "step": 1398 + }, + { + "epoch": 0.3419701784404791, + "grad_norm": 3.303999185562134, + "learning_rate": 1.5316489485594785e-05, + "loss": 0.9208, + "step": 1399 + }, + { + "epoch": 0.3422146174529455, + "grad_norm": 3.0878005027770996, + "learning_rate": 1.5309782128378676e-05, + "loss": 0.9152, + "step": 1400 + }, + { + "epoch": 0.34245905646541186, + "grad_norm": 3.1952340602874756, + "learning_rate": 1.5303071442781083e-05, + "loss": 0.9222, + "step": 1401 + }, + { + "epoch": 0.34270349547787826, + "grad_norm": 3.550318479537964, + "learning_rate": 1.5296357433008532e-05, + "loss": 0.8696, + "step": 1402 + }, + { + "epoch": 0.34294793449034466, + "grad_norm": 3.1809489727020264, + "learning_rate": 1.5289640103269626e-05, + "loss": 1.0164, + "step": 1403 + }, + { + "epoch": 0.34319237350281107, + "grad_norm": 3.1642727851867676, + "learning_rate": 1.5282919457775058e-05, + "loss": 0.8859, + "step": 1404 + }, + { + "epoch": 0.3434368125152774, + "grad_norm": 3.2544121742248535, + "learning_rate": 1.5276195500737592e-05, + "loss": 0.8058, + "step": 1405 + }, + { + "epoch": 0.3436812515277438, + "grad_norm": 2.817103862762451, + "learning_rate": 1.5269468236372068e-05, + "loss": 2.0819, + "step": 1406 + }, + { + "epoch": 0.3439256905402102, + "grad_norm": 3.803536891937256, + "learning_rate": 1.5262737668895403e-05, + "loss": 0.9362, + "step": 1407 + }, + { + "epoch": 0.3441701295526766, + "grad_norm": 3.5028889179229736, + "learning_rate": 1.5256003802526586e-05, + "loss": 0.8609, + "step": 1408 + }, + { + "epoch": 0.344414568565143, + "grad_norm": 3.3524909019470215, + "learning_rate": 1.5249266641486666e-05, + "loss": 0.9216, + "step": 1409 + }, + { + "epoch": 0.34465900757760937, + "grad_norm": 3.9344277381896973, + "learning_rate": 1.5242526189998768e-05, + "loss": 0.9408, + "step": 1410 + }, + { + "epoch": 0.34490344659007577, + "grad_norm": 4.254050254821777, + "learning_rate": 1.5235782452288068e-05, + "loss": 1.0183, + "step": 1411 + }, + { + "epoch": 0.34514788560254217, + "grad_norm": 2.2866740226745605, + "learning_rate": 1.5229035432581812e-05, + "loss": 1.9729, + "step": 1412 + }, + { + "epoch": 0.3453923246150086, + "grad_norm": 3.434361696243286, + "learning_rate": 1.5222285135109291e-05, + "loss": 0.9577, + "step": 1413 + }, + { + "epoch": 0.3456367636274749, + "grad_norm": 3.364114284515381, + "learning_rate": 1.5215531564101866e-05, + "loss": 0.9309, + "step": 1414 + }, + { + "epoch": 0.3458812026399413, + "grad_norm": 3.732811212539673, + "learning_rate": 1.5208774723792943e-05, + "loss": 0.8709, + "step": 1415 + }, + { + "epoch": 0.3461256416524077, + "grad_norm": 3.326068878173828, + "learning_rate": 1.5202014618417973e-05, + "loss": 0.9526, + "step": 1416 + }, + { + "epoch": 0.3463700806648741, + "grad_norm": 3.0876729488372803, + "learning_rate": 1.5195251252214461e-05, + "loss": 0.9559, + "step": 1417 + }, + { + "epoch": 0.3466145196773405, + "grad_norm": 3.168135404586792, + "learning_rate": 1.5188484629421946e-05, + "loss": 0.9279, + "step": 1418 + }, + { + "epoch": 0.3468589586898069, + "grad_norm": 3.341728925704956, + "learning_rate": 1.518171475428202e-05, + "loss": 1.0006, + "step": 1419 + }, + { + "epoch": 0.3471033977022733, + "grad_norm": 3.315312385559082, + "learning_rate": 1.5174941631038307e-05, + "loss": 1.0684, + "step": 1420 + }, + { + "epoch": 0.3473478367147397, + "grad_norm": 4.451924800872803, + "learning_rate": 1.5168165263936472e-05, + "loss": 0.873, + "step": 1421 + }, + { + "epoch": 0.3475922757272061, + "grad_norm": 3.256350517272949, + "learning_rate": 1.5161385657224203e-05, + "loss": 1.0304, + "step": 1422 + }, + { + "epoch": 0.3478367147396724, + "grad_norm": 3.359121084213257, + "learning_rate": 1.5154602815151233e-05, + "loss": 1.0106, + "step": 1423 + }, + { + "epoch": 0.3480811537521388, + "grad_norm": 3.1137757301330566, + "learning_rate": 1.514781674196931e-05, + "loss": 1.0017, + "step": 1424 + }, + { + "epoch": 0.34832559276460523, + "grad_norm": 3.178013324737549, + "learning_rate": 1.5141027441932217e-05, + "loss": 0.8749, + "step": 1425 + }, + { + "epoch": 0.34857003177707163, + "grad_norm": 4.293848037719727, + "learning_rate": 1.5134234919295757e-05, + "loss": 0.9248, + "step": 1426 + }, + { + "epoch": 0.34881447078953803, + "grad_norm": 3.428175449371338, + "learning_rate": 1.5127439178317747e-05, + "loss": 1.0673, + "step": 1427 + }, + { + "epoch": 0.3490589098020044, + "grad_norm": 3.1072657108306885, + "learning_rate": 1.5120640223258027e-05, + "loss": 0.9458, + "step": 1428 + }, + { + "epoch": 0.3493033488144708, + "grad_norm": 3.426907777786255, + "learning_rate": 1.5113838058378454e-05, + "loss": 0.9146, + "step": 1429 + }, + { + "epoch": 0.3495477878269372, + "grad_norm": 3.247302293777466, + "learning_rate": 1.5107032687942891e-05, + "loss": 0.8704, + "step": 1430 + }, + { + "epoch": 0.3497922268394036, + "grad_norm": 3.252124786376953, + "learning_rate": 1.5100224116217217e-05, + "loss": 0.9551, + "step": 1431 + }, + { + "epoch": 0.35003666585186993, + "grad_norm": 3.9852957725524902, + "learning_rate": 1.5093412347469312e-05, + "loss": 0.9488, + "step": 1432 + }, + { + "epoch": 0.35028110486433633, + "grad_norm": 3.5851423740386963, + "learning_rate": 1.5086597385969062e-05, + "loss": 1.0187, + "step": 1433 + }, + { + "epoch": 0.35052554387680274, + "grad_norm": 3.946223735809326, + "learning_rate": 1.5079779235988355e-05, + "loss": 0.973, + "step": 1434 + }, + { + "epoch": 0.35076998288926914, + "grad_norm": 3.861586093902588, + "learning_rate": 1.5072957901801075e-05, + "loss": 0.911, + "step": 1435 + }, + { + "epoch": 0.35101442190173554, + "grad_norm": 3.4354403018951416, + "learning_rate": 1.5066133387683106e-05, + "loss": 0.8669, + "step": 1436 + }, + { + "epoch": 0.3512588609142019, + "grad_norm": 3.3271026611328125, + "learning_rate": 1.5059305697912323e-05, + "loss": 0.8781, + "step": 1437 + }, + { + "epoch": 0.3515032999266683, + "grad_norm": 3.2607266902923584, + "learning_rate": 1.5052474836768594e-05, + "loss": 0.9524, + "step": 1438 + }, + { + "epoch": 0.3517477389391347, + "grad_norm": 3.6708407402038574, + "learning_rate": 1.5045640808533772e-05, + "loss": 1.0256, + "step": 1439 + }, + { + "epoch": 0.3519921779516011, + "grad_norm": 3.570585250854492, + "learning_rate": 1.5038803617491695e-05, + "loss": 0.9217, + "step": 1440 + }, + { + "epoch": 0.35223661696406744, + "grad_norm": 3.3812408447265625, + "learning_rate": 1.5031963267928185e-05, + "loss": 0.8215, + "step": 1441 + }, + { + "epoch": 0.35248105597653384, + "grad_norm": 3.565835952758789, + "learning_rate": 1.5025119764131044e-05, + "loss": 0.9133, + "step": 1442 + }, + { + "epoch": 0.35272549498900024, + "grad_norm": 4.318790912628174, + "learning_rate": 1.501827311039005e-05, + "loss": 0.9073, + "step": 1443 + }, + { + "epoch": 0.35296993400146665, + "grad_norm": 3.476274251937866, + "learning_rate": 1.501142331099696e-05, + "loss": 0.9068, + "step": 1444 + }, + { + "epoch": 0.35321437301393305, + "grad_norm": 3.3230934143066406, + "learning_rate": 1.5004570370245492e-05, + "loss": 0.988, + "step": 1445 + }, + { + "epoch": 0.3534588120263994, + "grad_norm": 3.812248706817627, + "learning_rate": 1.4997714292431347e-05, + "loss": 1.0304, + "step": 1446 + }, + { + "epoch": 0.3537032510388658, + "grad_norm": 3.7100956439971924, + "learning_rate": 1.4990855081852185e-05, + "loss": 0.9584, + "step": 1447 + }, + { + "epoch": 0.3539476900513322, + "grad_norm": 2.9337382316589355, + "learning_rate": 1.4983992742807627e-05, + "loss": 0.8695, + "step": 1448 + }, + { + "epoch": 0.3541921290637986, + "grad_norm": 3.947782039642334, + "learning_rate": 1.4977127279599264e-05, + "loss": 1.0827, + "step": 1449 + }, + { + "epoch": 0.35443656807626495, + "grad_norm": 3.16684889793396, + "learning_rate": 1.4970258696530635e-05, + "loss": 0.8774, + "step": 1450 + }, + { + "epoch": 0.35468100708873135, + "grad_norm": 4.034370422363281, + "learning_rate": 1.4963386997907242e-05, + "loss": 0.949, + "step": 1451 + }, + { + "epoch": 0.35492544610119775, + "grad_norm": 3.9608681201934814, + "learning_rate": 1.4956512188036536e-05, + "loss": 0.9403, + "step": 1452 + }, + { + "epoch": 0.35516988511366415, + "grad_norm": 3.7968034744262695, + "learning_rate": 1.4949634271227918e-05, + "loss": 0.9495, + "step": 1453 + }, + { + "epoch": 0.35541432412613055, + "grad_norm": 3.5002903938293457, + "learning_rate": 1.4942753251792746e-05, + "loss": 0.9865, + "step": 1454 + }, + { + "epoch": 0.3556587631385969, + "grad_norm": 3.36322021484375, + "learning_rate": 1.4935869134044308e-05, + "loss": 0.8611, + "step": 1455 + }, + { + "epoch": 0.3559032021510633, + "grad_norm": 4.71673583984375, + "learning_rate": 1.4928981922297842e-05, + "loss": 0.8905, + "step": 1456 + }, + { + "epoch": 0.3561476411635297, + "grad_norm": 3.0979466438293457, + "learning_rate": 1.4922091620870527e-05, + "loss": 0.9622, + "step": 1457 + }, + { + "epoch": 0.3563920801759961, + "grad_norm": 3.6254913806915283, + "learning_rate": 1.4915198234081479e-05, + "loss": 1.0715, + "step": 1458 + }, + { + "epoch": 0.35663651918846245, + "grad_norm": 3.1062965393066406, + "learning_rate": 1.4908301766251739e-05, + "loss": 0.8824, + "step": 1459 + }, + { + "epoch": 0.35688095820092886, + "grad_norm": 3.2139036655426025, + "learning_rate": 1.4901402221704287e-05, + "loss": 0.9883, + "step": 1460 + }, + { + "epoch": 0.35712539721339526, + "grad_norm": 3.6782898902893066, + "learning_rate": 1.4894499604764035e-05, + "loss": 0.9123, + "step": 1461 + }, + { + "epoch": 0.35736983622586166, + "grad_norm": 2.6064562797546387, + "learning_rate": 1.4887593919757814e-05, + "loss": 2.121, + "step": 1462 + }, + { + "epoch": 0.35761427523832806, + "grad_norm": 4.038392066955566, + "learning_rate": 1.4880685171014382e-05, + "loss": 0.9432, + "step": 1463 + }, + { + "epoch": 0.3578587142507944, + "grad_norm": 4.281519412994385, + "learning_rate": 1.4873773362864413e-05, + "loss": 0.9428, + "step": 1464 + }, + { + "epoch": 0.3581031532632608, + "grad_norm": 4.620762825012207, + "learning_rate": 1.4866858499640504e-05, + "loss": 0.9179, + "step": 1465 + }, + { + "epoch": 0.3583475922757272, + "grad_norm": 3.9565887451171875, + "learning_rate": 1.4859940585677165e-05, + "loss": 0.9347, + "step": 1466 + }, + { + "epoch": 0.3585920312881936, + "grad_norm": 3.9693334102630615, + "learning_rate": 1.4853019625310813e-05, + "loss": 0.9412, + "step": 1467 + }, + { + "epoch": 0.35883647030065996, + "grad_norm": 3.7491869926452637, + "learning_rate": 1.4846095622879784e-05, + "loss": 0.921, + "step": 1468 + }, + { + "epoch": 0.35908090931312636, + "grad_norm": 2.97538161277771, + "learning_rate": 1.483916858272432e-05, + "loss": 0.8116, + "step": 1469 + }, + { + "epoch": 0.35932534832559276, + "grad_norm": 3.518476724624634, + "learning_rate": 1.4832238509186559e-05, + "loss": 0.957, + "step": 1470 + }, + { + "epoch": 0.35956978733805917, + "grad_norm": 3.6826512813568115, + "learning_rate": 1.4825305406610547e-05, + "loss": 0.8955, + "step": 1471 + }, + { + "epoch": 0.35981422635052557, + "grad_norm": 3.420494794845581, + "learning_rate": 1.4818369279342227e-05, + "loss": 0.957, + "step": 1472 + }, + { + "epoch": 0.3600586653629919, + "grad_norm": 3.42960262298584, + "learning_rate": 1.4811430131729436e-05, + "loss": 0.918, + "step": 1473 + }, + { + "epoch": 0.3603031043754583, + "grad_norm": 2.91977596282959, + "learning_rate": 1.480448796812191e-05, + "loss": 0.9817, + "step": 1474 + }, + { + "epoch": 0.3605475433879247, + "grad_norm": 5.7857489585876465, + "learning_rate": 1.4797542792871267e-05, + "loss": 0.8361, + "step": 1475 + }, + { + "epoch": 0.3607919824003911, + "grad_norm": 4.280693531036377, + "learning_rate": 1.4790594610331019e-05, + "loss": 0.8621, + "step": 1476 + }, + { + "epoch": 0.36103642141285747, + "grad_norm": 3.5429012775421143, + "learning_rate": 1.4783643424856564e-05, + "loss": 0.9482, + "step": 1477 + }, + { + "epoch": 0.36128086042532387, + "grad_norm": 2.4663522243499756, + "learning_rate": 1.477668924080518e-05, + "loss": 2.0677, + "step": 1478 + }, + { + "epoch": 0.36152529943779027, + "grad_norm": 2.56671404838562, + "learning_rate": 1.4769732062536022e-05, + "loss": 2.0516, + "step": 1479 + }, + { + "epoch": 0.3617697384502567, + "grad_norm": 5.004128932952881, + "learning_rate": 1.4762771894410129e-05, + "loss": 0.9523, + "step": 1480 + }, + { + "epoch": 0.3620141774627231, + "grad_norm": 3.212191104888916, + "learning_rate": 1.4755808740790403e-05, + "loss": 0.8761, + "step": 1481 + }, + { + "epoch": 0.3622586164751894, + "grad_norm": 4.195330619812012, + "learning_rate": 1.4748842606041626e-05, + "loss": 0.9086, + "step": 1482 + }, + { + "epoch": 0.3625030554876558, + "grad_norm": 3.0373518466949463, + "learning_rate": 1.4741873494530452e-05, + "loss": 0.8122, + "step": 1483 + }, + { + "epoch": 0.3627474945001222, + "grad_norm": 2.4052069187164307, + "learning_rate": 1.4734901410625392e-05, + "loss": 2.074, + "step": 1484 + }, + { + "epoch": 0.3629919335125886, + "grad_norm": 3.4554340839385986, + "learning_rate": 1.4727926358696825e-05, + "loss": 1.0138, + "step": 1485 + }, + { + "epoch": 0.363236372525055, + "grad_norm": 2.9953768253326416, + "learning_rate": 1.4720948343116987e-05, + "loss": 0.9359, + "step": 1486 + }, + { + "epoch": 0.3634808115375214, + "grad_norm": 3.070591926574707, + "learning_rate": 1.4713967368259981e-05, + "loss": 1.0384, + "step": 1487 + }, + { + "epoch": 0.3637252505499878, + "grad_norm": 3.7442331314086914, + "learning_rate": 1.4706983438501745e-05, + "loss": 0.9505, + "step": 1488 + }, + { + "epoch": 0.3639696895624542, + "grad_norm": 3.5531771183013916, + "learning_rate": 1.4699996558220095e-05, + "loss": 1.0305, + "step": 1489 + }, + { + "epoch": 0.3642141285749206, + "grad_norm": 3.0107712745666504, + "learning_rate": 1.4693006731794682e-05, + "loss": 1.0072, + "step": 1490 + }, + { + "epoch": 0.36445856758738693, + "grad_norm": 3.122729539871216, + "learning_rate": 1.4686013963607e-05, + "loss": 0.9986, + "step": 1491 + }, + { + "epoch": 0.36470300659985333, + "grad_norm": 2.9052510261535645, + "learning_rate": 1.4679018258040399e-05, + "loss": 0.8143, + "step": 1492 + }, + { + "epoch": 0.36494744561231973, + "grad_norm": 4.365767478942871, + "learning_rate": 1.4672019619480056e-05, + "loss": 1.0574, + "step": 1493 + }, + { + "epoch": 0.36519188462478613, + "grad_norm": 3.220187187194824, + "learning_rate": 1.4665018052313004e-05, + "loss": 0.8821, + "step": 1494 + }, + { + "epoch": 0.3654363236372525, + "grad_norm": 3.0632996559143066, + "learning_rate": 1.4658013560928098e-05, + "loss": 0.9838, + "step": 1495 + }, + { + "epoch": 0.3656807626497189, + "grad_norm": 3.8072338104248047, + "learning_rate": 1.4651006149716027e-05, + "loss": 1.0137, + "step": 1496 + }, + { + "epoch": 0.3659252016621853, + "grad_norm": 3.2164018154144287, + "learning_rate": 1.4643995823069318e-05, + "loss": 1.0438, + "step": 1497 + }, + { + "epoch": 0.3661696406746517, + "grad_norm": 3.714524269104004, + "learning_rate": 1.4636982585382321e-05, + "loss": 1.0036, + "step": 1498 + }, + { + "epoch": 0.3664140796871181, + "grad_norm": 3.291309118270874, + "learning_rate": 1.4629966441051208e-05, + "loss": 0.977, + "step": 1499 + }, + { + "epoch": 0.36665851869958443, + "grad_norm": 4.559157848358154, + "learning_rate": 1.462294739447398e-05, + "loss": 0.9867, + "step": 1500 + }, + { + "epoch": 0.36690295771205084, + "grad_norm": 3.373988389968872, + "learning_rate": 1.4615925450050448e-05, + "loss": 0.9148, + "step": 1501 + }, + { + "epoch": 0.36714739672451724, + "grad_norm": 3.6921863555908203, + "learning_rate": 1.4608900612182246e-05, + "loss": 0.8028, + "step": 1502 + }, + { + "epoch": 0.36739183573698364, + "grad_norm": 3.494563579559326, + "learning_rate": 1.4601872885272824e-05, + "loss": 0.8157, + "step": 1503 + }, + { + "epoch": 0.36763627474945, + "grad_norm": 3.2356550693511963, + "learning_rate": 1.4594842273727434e-05, + "loss": 0.9164, + "step": 1504 + }, + { + "epoch": 0.3678807137619164, + "grad_norm": 3.729564666748047, + "learning_rate": 1.4587808781953146e-05, + "loss": 0.8828, + "step": 1505 + }, + { + "epoch": 0.3681251527743828, + "grad_norm": 2.970707416534424, + "learning_rate": 1.4580772414358829e-05, + "loss": 0.8681, + "step": 1506 + }, + { + "epoch": 0.3683695917868492, + "grad_norm": 3.6039421558380127, + "learning_rate": 1.4573733175355152e-05, + "loss": 0.9996, + "step": 1507 + }, + { + "epoch": 0.3686140307993156, + "grad_norm": 4.286959171295166, + "learning_rate": 1.4566691069354595e-05, + "loss": 0.9506, + "step": 1508 + }, + { + "epoch": 0.36885846981178194, + "grad_norm": 3.340758800506592, + "learning_rate": 1.4559646100771428e-05, + "loss": 0.8819, + "step": 1509 + }, + { + "epoch": 0.36910290882424834, + "grad_norm": 4.261744022369385, + "learning_rate": 1.4552598274021713e-05, + "loss": 1.0301, + "step": 1510 + }, + { + "epoch": 0.36934734783671475, + "grad_norm": 3.2351458072662354, + "learning_rate": 1.4545547593523308e-05, + "loss": 0.9546, + "step": 1511 + }, + { + "epoch": 0.36959178684918115, + "grad_norm": 3.675766706466675, + "learning_rate": 1.4538494063695861e-05, + "loss": 0.9193, + "step": 1512 + }, + { + "epoch": 0.3698362258616475, + "grad_norm": 4.477265357971191, + "learning_rate": 1.4531437688960796e-05, + "loss": 0.989, + "step": 1513 + }, + { + "epoch": 0.3700806648741139, + "grad_norm": 4.235387325286865, + "learning_rate": 1.4524378473741335e-05, + "loss": 0.9246, + "step": 1514 + }, + { + "epoch": 0.3703251038865803, + "grad_norm": 4.069041728973389, + "learning_rate": 1.4517316422462471e-05, + "loss": 0.8349, + "step": 1515 + }, + { + "epoch": 0.3705695428990467, + "grad_norm": 4.676631450653076, + "learning_rate": 1.4510251539550978e-05, + "loss": 0.9418, + "step": 1516 + }, + { + "epoch": 0.3708139819115131, + "grad_norm": 3.579474687576294, + "learning_rate": 1.4503183829435403e-05, + "loss": 0.8662, + "step": 1517 + }, + { + "epoch": 0.37105842092397945, + "grad_norm": 3.939366340637207, + "learning_rate": 1.4496113296546068e-05, + "loss": 0.9567, + "step": 1518 + }, + { + "epoch": 0.37130285993644585, + "grad_norm": 3.652257204055786, + "learning_rate": 1.4489039945315059e-05, + "loss": 0.826, + "step": 1519 + }, + { + "epoch": 0.37154729894891225, + "grad_norm": 3.49208664894104, + "learning_rate": 1.4481963780176235e-05, + "loss": 0.815, + "step": 1520 + }, + { + "epoch": 0.37179173796137865, + "grad_norm": 4.103240013122559, + "learning_rate": 1.4474884805565217e-05, + "loss": 0.8812, + "step": 1521 + }, + { + "epoch": 0.372036176973845, + "grad_norm": 3.390845537185669, + "learning_rate": 1.4467803025919383e-05, + "loss": 0.9073, + "step": 1522 + }, + { + "epoch": 0.3722806159863114, + "grad_norm": 4.154443740844727, + "learning_rate": 1.4460718445677877e-05, + "loss": 0.9916, + "step": 1523 + }, + { + "epoch": 0.3725250549987778, + "grad_norm": 3.4157791137695312, + "learning_rate": 1.4453631069281593e-05, + "loss": 0.9102, + "step": 1524 + }, + { + "epoch": 0.3727694940112442, + "grad_norm": 3.5285110473632812, + "learning_rate": 1.4446540901173176e-05, + "loss": 0.8537, + "step": 1525 + }, + { + "epoch": 0.3730139330237106, + "grad_norm": 4.130982875823975, + "learning_rate": 1.4439447945797026e-05, + "loss": 0.8885, + "step": 1526 + }, + { + "epoch": 0.37325837203617696, + "grad_norm": 3.806924819946289, + "learning_rate": 1.4432352207599288e-05, + "loss": 0.9369, + "step": 1527 + }, + { + "epoch": 0.37350281104864336, + "grad_norm": 3.0304319858551025, + "learning_rate": 1.442525369102785e-05, + "loss": 0.884, + "step": 1528 + }, + { + "epoch": 0.37374725006110976, + "grad_norm": 3.4081881046295166, + "learning_rate": 1.4418152400532344e-05, + "loss": 0.9057, + "step": 1529 + }, + { + "epoch": 0.37399168907357616, + "grad_norm": 3.4484832286834717, + "learning_rate": 1.441104834056414e-05, + "loss": 0.9393, + "step": 1530 + }, + { + "epoch": 0.3742361280860425, + "grad_norm": 3.4003255367279053, + "learning_rate": 1.4403941515576344e-05, + "loss": 0.8726, + "step": 1531 + }, + { + "epoch": 0.3744805670985089, + "grad_norm": 4.160417079925537, + "learning_rate": 1.4396831930023796e-05, + "loss": 0.9071, + "step": 1532 + }, + { + "epoch": 0.3747250061109753, + "grad_norm": 3.3831288814544678, + "learning_rate": 1.4389719588363062e-05, + "loss": 0.8404, + "step": 1533 + }, + { + "epoch": 0.3749694451234417, + "grad_norm": 3.3936948776245117, + "learning_rate": 1.4382604495052444e-05, + "loss": 0.9059, + "step": 1534 + }, + { + "epoch": 0.3752138841359081, + "grad_norm": 4.137432098388672, + "learning_rate": 1.437548665455196e-05, + "loss": 0.7942, + "step": 1535 + }, + { + "epoch": 0.37545832314837446, + "grad_norm": 4.356401443481445, + "learning_rate": 1.4368366071323354e-05, + "loss": 0.9009, + "step": 1536 + }, + { + "epoch": 0.37570276216084086, + "grad_norm": 3.207725763320923, + "learning_rate": 1.4361242749830093e-05, + "loss": 0.8208, + "step": 1537 + }, + { + "epoch": 0.37594720117330727, + "grad_norm": 3.3889639377593994, + "learning_rate": 1.435411669453735e-05, + "loss": 0.9573, + "step": 1538 + }, + { + "epoch": 0.37619164018577367, + "grad_norm": 4.115054130554199, + "learning_rate": 1.4346987909912024e-05, + "loss": 0.92, + "step": 1539 + }, + { + "epoch": 0.37643607919824, + "grad_norm": 3.934098720550537, + "learning_rate": 1.4339856400422721e-05, + "loss": 0.8831, + "step": 1540 + }, + { + "epoch": 0.3766805182107064, + "grad_norm": 4.6525559425354, + "learning_rate": 1.4332722170539748e-05, + "loss": 0.843, + "step": 1541 + }, + { + "epoch": 0.3769249572231728, + "grad_norm": 3.4838201999664307, + "learning_rate": 1.4325585224735125e-05, + "loss": 0.9346, + "step": 1542 + }, + { + "epoch": 0.3771693962356392, + "grad_norm": 3.2306668758392334, + "learning_rate": 1.431844556748257e-05, + "loss": 0.8231, + "step": 1543 + }, + { + "epoch": 0.3774138352481056, + "grad_norm": 3.608369827270508, + "learning_rate": 1.4311303203257505e-05, + "loss": 0.9087, + "step": 1544 + }, + { + "epoch": 0.37765827426057197, + "grad_norm": 3.370871067047119, + "learning_rate": 1.4304158136537047e-05, + "loss": 0.9316, + "step": 1545 + }, + { + "epoch": 0.37790271327303837, + "grad_norm": 3.6224355697631836, + "learning_rate": 1.4297010371800002e-05, + "loss": 0.8198, + "step": 1546 + }, + { + "epoch": 0.3781471522855048, + "grad_norm": 2.598627805709839, + "learning_rate": 1.4289859913526876e-05, + "loss": 2.0515, + "step": 1547 + }, + { + "epoch": 0.3783915912979712, + "grad_norm": 3.645094394683838, + "learning_rate": 1.4282706766199859e-05, + "loss": 0.9776, + "step": 1548 + }, + { + "epoch": 0.3786360303104375, + "grad_norm": 3.49668288230896, + "learning_rate": 1.4275550934302822e-05, + "loss": 0.9527, + "step": 1549 + }, + { + "epoch": 0.3788804693229039, + "grad_norm": 3.822556972503662, + "learning_rate": 1.426839242232133e-05, + "loss": 0.9844, + "step": 1550 + }, + { + "epoch": 0.3791249083353703, + "grad_norm": 3.3121814727783203, + "learning_rate": 1.4261231234742618e-05, + "loss": 0.8436, + "step": 1551 + }, + { + "epoch": 0.3793693473478367, + "grad_norm": 3.4897799491882324, + "learning_rate": 1.4254067376055602e-05, + "loss": 0.9729, + "step": 1552 + }, + { + "epoch": 0.37961378636030313, + "grad_norm": 3.6050734519958496, + "learning_rate": 1.424690085075087e-05, + "loss": 0.6889, + "step": 1553 + }, + { + "epoch": 0.3798582253727695, + "grad_norm": 3.686805248260498, + "learning_rate": 1.4239731663320688e-05, + "loss": 0.8947, + "step": 1554 + }, + { + "epoch": 0.3801026643852359, + "grad_norm": 3.1977193355560303, + "learning_rate": 1.4232559818258983e-05, + "loss": 0.9087, + "step": 1555 + }, + { + "epoch": 0.3803471033977023, + "grad_norm": 5.1845526695251465, + "learning_rate": 1.4225385320061353e-05, + "loss": 0.8625, + "step": 1556 + }, + { + "epoch": 0.3805915424101687, + "grad_norm": 3.603829860687256, + "learning_rate": 1.4218208173225057e-05, + "loss": 0.9319, + "step": 1557 + }, + { + "epoch": 0.38083598142263503, + "grad_norm": 3.646444797515869, + "learning_rate": 1.4211028382249013e-05, + "loss": 0.8586, + "step": 1558 + }, + { + "epoch": 0.38108042043510143, + "grad_norm": 4.110007286071777, + "learning_rate": 1.4203845951633799e-05, + "loss": 0.9234, + "step": 1559 + }, + { + "epoch": 0.38132485944756783, + "grad_norm": 4.331524848937988, + "learning_rate": 1.4196660885881646e-05, + "loss": 0.9518, + "step": 1560 + }, + { + "epoch": 0.38156929846003423, + "grad_norm": 3.5555384159088135, + "learning_rate": 1.4189473189496437e-05, + "loss": 0.8741, + "step": 1561 + }, + { + "epoch": 0.38181373747250064, + "grad_norm": 3.843526840209961, + "learning_rate": 1.4182282866983708e-05, + "loss": 0.8401, + "step": 1562 + }, + { + "epoch": 0.382058176484967, + "grad_norm": 3.4764273166656494, + "learning_rate": 1.4175089922850633e-05, + "loss": 0.9026, + "step": 1563 + }, + { + "epoch": 0.3823026154974334, + "grad_norm": 3.949537754058838, + "learning_rate": 1.4167894361606038e-05, + "loss": 0.8916, + "step": 1564 + }, + { + "epoch": 0.3825470545098998, + "grad_norm": 3.3268473148345947, + "learning_rate": 1.4160696187760383e-05, + "loss": 0.8919, + "step": 1565 + }, + { + "epoch": 0.3827914935223662, + "grad_norm": 4.385059356689453, + "learning_rate": 1.4153495405825768e-05, + "loss": 1.0173, + "step": 1566 + }, + { + "epoch": 0.38303593253483254, + "grad_norm": 4.0300445556640625, + "learning_rate": 1.4146292020315926e-05, + "loss": 0.9705, + "step": 1567 + }, + { + "epoch": 0.38328037154729894, + "grad_norm": 3.9647634029388428, + "learning_rate": 1.413908603574623e-05, + "loss": 0.9537, + "step": 1568 + }, + { + "epoch": 0.38352481055976534, + "grad_norm": 3.9260358810424805, + "learning_rate": 1.413187745663367e-05, + "loss": 0.8883, + "step": 1569 + }, + { + "epoch": 0.38376924957223174, + "grad_norm": 3.1854300498962402, + "learning_rate": 1.4124666287496873e-05, + "loss": 0.9824, + "step": 1570 + }, + { + "epoch": 0.38401368858469814, + "grad_norm": 3.3261890411376953, + "learning_rate": 1.4117452532856084e-05, + "loss": 0.9699, + "step": 1571 + }, + { + "epoch": 0.3842581275971645, + "grad_norm": 3.499786615371704, + "learning_rate": 1.4110236197233164e-05, + "loss": 0.8683, + "step": 1572 + }, + { + "epoch": 0.3845025666096309, + "grad_norm": 3.24682879447937, + "learning_rate": 1.4103017285151607e-05, + "loss": 0.9277, + "step": 1573 + }, + { + "epoch": 0.3847470056220973, + "grad_norm": 6.736312389373779, + "learning_rate": 1.4095795801136509e-05, + "loss": 0.9788, + "step": 1574 + }, + { + "epoch": 0.3849914446345637, + "grad_norm": 3.485842227935791, + "learning_rate": 1.408857174971458e-05, + "loss": 0.818, + "step": 1575 + }, + { + "epoch": 0.38523588364703004, + "grad_norm": 3.612020969390869, + "learning_rate": 1.4081345135414144e-05, + "loss": 1.0562, + "step": 1576 + }, + { + "epoch": 0.38548032265949644, + "grad_norm": 3.5550265312194824, + "learning_rate": 1.4074115962765125e-05, + "loss": 0.9049, + "step": 1577 + }, + { + "epoch": 0.38572476167196285, + "grad_norm": 3.0933775901794434, + "learning_rate": 1.4066884236299057e-05, + "loss": 0.8222, + "step": 1578 + }, + { + "epoch": 0.38596920068442925, + "grad_norm": 3.570155143737793, + "learning_rate": 1.4059649960549071e-05, + "loss": 0.8068, + "step": 1579 + }, + { + "epoch": 0.38621363969689565, + "grad_norm": 3.3882038593292236, + "learning_rate": 1.4052413140049898e-05, + "loss": 0.874, + "step": 1580 + }, + { + "epoch": 0.386458078709362, + "grad_norm": 3.5525708198547363, + "learning_rate": 1.4045173779337866e-05, + "loss": 0.9939, + "step": 1581 + }, + { + "epoch": 0.3867025177218284, + "grad_norm": 4.007676124572754, + "learning_rate": 1.4037931882950888e-05, + "loss": 0.8346, + "step": 1582 + }, + { + "epoch": 0.3869469567342948, + "grad_norm": 3.0331296920776367, + "learning_rate": 1.4030687455428473e-05, + "loss": 0.9635, + "step": 1583 + }, + { + "epoch": 0.3871913957467612, + "grad_norm": 15.911253929138184, + "learning_rate": 1.4023440501311713e-05, + "loss": 0.8254, + "step": 1584 + }, + { + "epoch": 0.38743583475922755, + "grad_norm": 3.9656739234924316, + "learning_rate": 1.4016191025143289e-05, + "loss": 0.9775, + "step": 1585 + }, + { + "epoch": 0.38768027377169395, + "grad_norm": 3.3272018432617188, + "learning_rate": 1.400893903146746e-05, + "loss": 0.9645, + "step": 1586 + }, + { + "epoch": 0.38792471278416035, + "grad_norm": 4.041081428527832, + "learning_rate": 1.4001684524830057e-05, + "loss": 0.8346, + "step": 1587 + }, + { + "epoch": 0.38816915179662675, + "grad_norm": 3.2119650840759277, + "learning_rate": 1.3994427509778499e-05, + "loss": 0.8845, + "step": 1588 + }, + { + "epoch": 0.38841359080909316, + "grad_norm": 2.7611374855041504, + "learning_rate": 1.3987167990861763e-05, + "loss": 0.8931, + "step": 1589 + }, + { + "epoch": 0.3886580298215595, + "grad_norm": 3.178654670715332, + "learning_rate": 1.3979905972630405e-05, + "loss": 0.9309, + "step": 1590 + }, + { + "epoch": 0.3889024688340259, + "grad_norm": 4.22986364364624, + "learning_rate": 1.3972641459636548e-05, + "loss": 0.7984, + "step": 1591 + }, + { + "epoch": 0.3891469078464923, + "grad_norm": 3.732235908508301, + "learning_rate": 1.396537445643387e-05, + "loss": 0.8091, + "step": 1592 + }, + { + "epoch": 0.3893913468589587, + "grad_norm": 3.6782584190368652, + "learning_rate": 1.3958104967577627e-05, + "loss": 1.053, + "step": 1593 + }, + { + "epoch": 0.38963578587142506, + "grad_norm": 3.923552989959717, + "learning_rate": 1.3950832997624612e-05, + "loss": 0.8426, + "step": 1594 + }, + { + "epoch": 0.38988022488389146, + "grad_norm": 3.2947030067443848, + "learning_rate": 1.3943558551133186e-05, + "loss": 0.9391, + "step": 1595 + }, + { + "epoch": 0.39012466389635786, + "grad_norm": 3.272634506225586, + "learning_rate": 1.3936281632663263e-05, + "loss": 0.886, + "step": 1596 + }, + { + "epoch": 0.39036910290882426, + "grad_norm": 4.148128986358643, + "learning_rate": 1.39290022467763e-05, + "loss": 1.045, + "step": 1597 + }, + { + "epoch": 0.39061354192129066, + "grad_norm": 3.9834187030792236, + "learning_rate": 1.3921720398035306e-05, + "loss": 0.8731, + "step": 1598 + }, + { + "epoch": 0.390857980933757, + "grad_norm": 3.3036372661590576, + "learning_rate": 1.3914436091004829e-05, + "loss": 0.8053, + "step": 1599 + }, + { + "epoch": 0.3911024199462234, + "grad_norm": 3.2304863929748535, + "learning_rate": 1.3907149330250962e-05, + "loss": 0.8049, + "step": 1600 + }, + { + "epoch": 0.3913468589586898, + "grad_norm": 3.1071078777313232, + "learning_rate": 1.3899860120341338e-05, + "loss": 0.7691, + "step": 1601 + }, + { + "epoch": 0.3915912979711562, + "grad_norm": 3.505664110183716, + "learning_rate": 1.3892568465845118e-05, + "loss": 0.8412, + "step": 1602 + }, + { + "epoch": 0.39183573698362256, + "grad_norm": 3.2916007041931152, + "learning_rate": 1.3885274371333001e-05, + "loss": 0.8779, + "step": 1603 + }, + { + "epoch": 0.39208017599608896, + "grad_norm": 3.703036069869995, + "learning_rate": 1.3877977841377217e-05, + "loss": 0.8388, + "step": 1604 + }, + { + "epoch": 0.39232461500855537, + "grad_norm": 3.327202320098877, + "learning_rate": 1.3870678880551516e-05, + "loss": 0.9072, + "step": 1605 + }, + { + "epoch": 0.39256905402102177, + "grad_norm": 5.386071681976318, + "learning_rate": 1.3863377493431177e-05, + "loss": 0.975, + "step": 1606 + }, + { + "epoch": 0.39281349303348817, + "grad_norm": 3.7295799255371094, + "learning_rate": 1.3856073684592994e-05, + "loss": 0.7813, + "step": 1607 + }, + { + "epoch": 0.3930579320459545, + "grad_norm": 3.692769765853882, + "learning_rate": 1.3848767458615286e-05, + "loss": 0.8892, + "step": 1608 + }, + { + "epoch": 0.3933023710584209, + "grad_norm": 3.072274923324585, + "learning_rate": 1.3841458820077889e-05, + "loss": 0.8592, + "step": 1609 + }, + { + "epoch": 0.3935468100708873, + "grad_norm": 3.234300136566162, + "learning_rate": 1.3834147773562143e-05, + "loss": 1.0064, + "step": 1610 + }, + { + "epoch": 0.3937912490833537, + "grad_norm": 4.414983749389648, + "learning_rate": 1.3826834323650899e-05, + "loss": 0.8768, + "step": 1611 + }, + { + "epoch": 0.39403568809582007, + "grad_norm": 3.7046587467193604, + "learning_rate": 1.3819518474928519e-05, + "loss": 0.8014, + "step": 1612 + }, + { + "epoch": 0.39428012710828647, + "grad_norm": 3.0449559688568115, + "learning_rate": 1.3812200231980869e-05, + "loss": 0.9361, + "step": 1613 + }, + { + "epoch": 0.3945245661207529, + "grad_norm": 3.53568696975708, + "learning_rate": 1.380487959939531e-05, + "loss": 0.9733, + "step": 1614 + }, + { + "epoch": 0.3947690051332193, + "grad_norm": 3.7599360942840576, + "learning_rate": 1.37975565817607e-05, + "loss": 0.9341, + "step": 1615 + }, + { + "epoch": 0.3950134441456857, + "grad_norm": 3.2119531631469727, + "learning_rate": 1.3790231183667403e-05, + "loss": 0.926, + "step": 1616 + }, + { + "epoch": 0.395257883158152, + "grad_norm": 3.0126593112945557, + "learning_rate": 1.3782903409707266e-05, + "loss": 0.8564, + "step": 1617 + }, + { + "epoch": 0.3955023221706184, + "grad_norm": 3.4946446418762207, + "learning_rate": 1.3775573264473629e-05, + "loss": 0.8601, + "step": 1618 + }, + { + "epoch": 0.3957467611830848, + "grad_norm": 3.624980926513672, + "learning_rate": 1.3768240752561315e-05, + "loss": 0.9043, + "step": 1619 + }, + { + "epoch": 0.39599120019555123, + "grad_norm": 3.868273973464966, + "learning_rate": 1.3760905878566633e-05, + "loss": 0.9251, + "step": 1620 + }, + { + "epoch": 0.3962356392080176, + "grad_norm": 3.1690731048583984, + "learning_rate": 1.3753568647087372e-05, + "loss": 0.8974, + "step": 1621 + }, + { + "epoch": 0.396480078220484, + "grad_norm": 4.424831867218018, + "learning_rate": 1.37462290627228e-05, + "loss": 0.8481, + "step": 1622 + }, + { + "epoch": 0.3967245172329504, + "grad_norm": 3.5461859703063965, + "learning_rate": 1.3738887130073655e-05, + "loss": 0.8403, + "step": 1623 + }, + { + "epoch": 0.3969689562454168, + "grad_norm": 3.7361207008361816, + "learning_rate": 1.3731542853742155e-05, + "loss": 0.8354, + "step": 1624 + }, + { + "epoch": 0.3972133952578832, + "grad_norm": 2.770488977432251, + "learning_rate": 1.3724196238331983e-05, + "loss": 2.013, + "step": 1625 + }, + { + "epoch": 0.39745783427034953, + "grad_norm": 2.9239935874938965, + "learning_rate": 1.3716847288448287e-05, + "loss": 0.7827, + "step": 1626 + }, + { + "epoch": 0.39770227328281593, + "grad_norm": 3.373499870300293, + "learning_rate": 1.370949600869768e-05, + "loss": 0.9397, + "step": 1627 + }, + { + "epoch": 0.39794671229528233, + "grad_norm": 3.6640830039978027, + "learning_rate": 1.3702142403688234e-05, + "loss": 0.7888, + "step": 1628 + }, + { + "epoch": 0.39819115130774874, + "grad_norm": 3.442188024520874, + "learning_rate": 1.3694786478029478e-05, + "loss": 0.8488, + "step": 1629 + }, + { + "epoch": 0.3984355903202151, + "grad_norm": 3.294236183166504, + "learning_rate": 1.3687428236332401e-05, + "loss": 0.8168, + "step": 1630 + }, + { + "epoch": 0.3986800293326815, + "grad_norm": 3.681321620941162, + "learning_rate": 1.3680067683209438e-05, + "loss": 0.8385, + "step": 1631 + }, + { + "epoch": 0.3989244683451479, + "grad_norm": 5.7639384269714355, + "learning_rate": 1.3672704823274472e-05, + "loss": 0.8752, + "step": 1632 + }, + { + "epoch": 0.3991689073576143, + "grad_norm": 4.043473720550537, + "learning_rate": 1.366533966114284e-05, + "loss": 0.8396, + "step": 1633 + }, + { + "epoch": 0.3994133463700807, + "grad_norm": 3.78446626663208, + "learning_rate": 1.3657972201431315e-05, + "loss": 0.9433, + "step": 1634 + }, + { + "epoch": 0.39965778538254704, + "grad_norm": 3.113582134246826, + "learning_rate": 1.3650602448758113e-05, + "loss": 0.8771, + "step": 1635 + }, + { + "epoch": 0.39990222439501344, + "grad_norm": 3.1159162521362305, + "learning_rate": 1.3643230407742883e-05, + "loss": 0.7404, + "step": 1636 + }, + { + "epoch": 0.40014666340747984, + "grad_norm": 2.6976401805877686, + "learning_rate": 1.3635856083006715e-05, + "loss": 2.1143, + "step": 1637 + }, + { + "epoch": 0.40039110241994624, + "grad_norm": 3.4253532886505127, + "learning_rate": 1.3628479479172128e-05, + "loss": 0.6934, + "step": 1638 + }, + { + "epoch": 0.4006355414324126, + "grad_norm": 3.560701847076416, + "learning_rate": 1.3621100600863066e-05, + "loss": 0.9006, + "step": 1639 + }, + { + "epoch": 0.400879980444879, + "grad_norm": 3.8821518421173096, + "learning_rate": 1.3613719452704906e-05, + "loss": 0.7986, + "step": 1640 + }, + { + "epoch": 0.4011244194573454, + "grad_norm": 3.5182454586029053, + "learning_rate": 1.3606336039324439e-05, + "loss": 0.8107, + "step": 1641 + }, + { + "epoch": 0.4013688584698118, + "grad_norm": 3.8955798149108887, + "learning_rate": 1.3598950365349884e-05, + "loss": 0.885, + "step": 1642 + }, + { + "epoch": 0.4016132974822782, + "grad_norm": 3.7581288814544678, + "learning_rate": 1.3591562435410873e-05, + "loss": 0.7875, + "step": 1643 + }, + { + "epoch": 0.40185773649474454, + "grad_norm": 4.369556427001953, + "learning_rate": 1.3584172254138452e-05, + "loss": 0.9033, + "step": 1644 + }, + { + "epoch": 0.40210217550721095, + "grad_norm": 5.720822811126709, + "learning_rate": 1.357677982616508e-05, + "loss": 0.9166, + "step": 1645 + }, + { + "epoch": 0.40234661451967735, + "grad_norm": 3.371997117996216, + "learning_rate": 1.356938515612462e-05, + "loss": 0.9141, + "step": 1646 + }, + { + "epoch": 0.40259105353214375, + "grad_norm": 4.52249002456665, + "learning_rate": 1.3561988248652346e-05, + "loss": 0.9192, + "step": 1647 + }, + { + "epoch": 0.4028354925446101, + "grad_norm": 3.647069215774536, + "learning_rate": 1.3554589108384937e-05, + "loss": 0.9729, + "step": 1648 + }, + { + "epoch": 0.4030799315570765, + "grad_norm": 3.2838327884674072, + "learning_rate": 1.3547187739960458e-05, + "loss": 0.8376, + "step": 1649 + }, + { + "epoch": 0.4033243705695429, + "grad_norm": 4.263354778289795, + "learning_rate": 1.3539784148018387e-05, + "loss": 0.9311, + "step": 1650 + }, + { + "epoch": 0.4035688095820093, + "grad_norm": 3.7330758571624756, + "learning_rate": 1.353237833719958e-05, + "loss": 0.8295, + "step": 1651 + }, + { + "epoch": 0.4038132485944757, + "grad_norm": 3.026618003845215, + "learning_rate": 1.3524970312146305e-05, + "loss": 0.9112, + "step": 1652 + }, + { + "epoch": 0.40405768760694205, + "grad_norm": 3.332324266433716, + "learning_rate": 1.3517560077502191e-05, + "loss": 0.8801, + "step": 1653 + }, + { + "epoch": 0.40430212661940845, + "grad_norm": 3.2940218448638916, + "learning_rate": 1.3510147637912276e-05, + "loss": 0.9387, + "step": 1654 + }, + { + "epoch": 0.40454656563187485, + "grad_norm": 3.100508213043213, + "learning_rate": 1.3502732998022965e-05, + "loss": 0.8721, + "step": 1655 + }, + { + "epoch": 0.40479100464434126, + "grad_norm": 3.2187600135803223, + "learning_rate": 1.3495316162482051e-05, + "loss": 0.773, + "step": 1656 + }, + { + "epoch": 0.4050354436568076, + "grad_norm": 3.792860984802246, + "learning_rate": 1.3487897135938697e-05, + "loss": 0.7794, + "step": 1657 + }, + { + "epoch": 0.405279882669274, + "grad_norm": 4.043926239013672, + "learning_rate": 1.3480475923043445e-05, + "loss": 0.845, + "step": 1658 + }, + { + "epoch": 0.4055243216817404, + "grad_norm": 3.7164793014526367, + "learning_rate": 1.3473052528448203e-05, + "loss": 0.9005, + "step": 1659 + }, + { + "epoch": 0.4057687606942068, + "grad_norm": 4.479531288146973, + "learning_rate": 1.346562695680625e-05, + "loss": 0.8135, + "step": 1660 + }, + { + "epoch": 0.4060131997066732, + "grad_norm": 3.599146604537964, + "learning_rate": 1.3458199212772227e-05, + "loss": 0.7886, + "step": 1661 + }, + { + "epoch": 0.40625763871913956, + "grad_norm": 3.1297788619995117, + "learning_rate": 1.3450769301002135e-05, + "loss": 0.8802, + "step": 1662 + }, + { + "epoch": 0.40650207773160596, + "grad_norm": 3.232038974761963, + "learning_rate": 1.3443337226153343e-05, + "loss": 0.888, + "step": 1663 + }, + { + "epoch": 0.40674651674407236, + "grad_norm": 3.5411183834075928, + "learning_rate": 1.3435902992884567e-05, + "loss": 0.88, + "step": 1664 + }, + { + "epoch": 0.40699095575653876, + "grad_norm": 4.1356682777404785, + "learning_rate": 1.3428466605855874e-05, + "loss": 0.8242, + "step": 1665 + }, + { + "epoch": 0.4072353947690051, + "grad_norm": 4.684967517852783, + "learning_rate": 1.342102806972869e-05, + "loss": 0.837, + "step": 1666 + }, + { + "epoch": 0.4074798337814715, + "grad_norm": 4.071098327636719, + "learning_rate": 1.3413587389165783e-05, + "loss": 0.9001, + "step": 1667 + }, + { + "epoch": 0.4077242727939379, + "grad_norm": 3.038487434387207, + "learning_rate": 1.3406144568831265e-05, + "loss": 0.8962, + "step": 1668 + }, + { + "epoch": 0.4079687118064043, + "grad_norm": 3.531371593475342, + "learning_rate": 1.3398699613390593e-05, + "loss": 0.89, + "step": 1669 + }, + { + "epoch": 0.4082131508188707, + "grad_norm": 3.3554468154907227, + "learning_rate": 1.3391252527510554e-05, + "loss": 0.9256, + "step": 1670 + }, + { + "epoch": 0.40845758983133706, + "grad_norm": 4.093055725097656, + "learning_rate": 1.3383803315859281e-05, + "loss": 0.8241, + "step": 1671 + }, + { + "epoch": 0.40870202884380347, + "grad_norm": 3.222688913345337, + "learning_rate": 1.3376351983106233e-05, + "loss": 0.9334, + "step": 1672 + }, + { + "epoch": 0.40894646785626987, + "grad_norm": 3.072899341583252, + "learning_rate": 1.3368898533922202e-05, + "loss": 0.7888, + "step": 1673 + }, + { + "epoch": 0.40919090686873627, + "grad_norm": 3.434206008911133, + "learning_rate": 1.3361442972979301e-05, + "loss": 0.9011, + "step": 1674 + }, + { + "epoch": 0.4094353458812026, + "grad_norm": 3.513918161392212, + "learning_rate": 1.3353985304950974e-05, + "loss": 0.8411, + "step": 1675 + }, + { + "epoch": 0.409679784893669, + "grad_norm": 3.3258190155029297, + "learning_rate": 1.3346525534511978e-05, + "loss": 0.846, + "step": 1676 + }, + { + "epoch": 0.4099242239061354, + "grad_norm": 4.432790756225586, + "learning_rate": 1.3339063666338396e-05, + "loss": 0.8694, + "step": 1677 + }, + { + "epoch": 0.4101686629186018, + "grad_norm": 3.975558042526245, + "learning_rate": 1.333159970510762e-05, + "loss": 0.9027, + "step": 1678 + }, + { + "epoch": 0.4104131019310682, + "grad_norm": 3.4570248126983643, + "learning_rate": 1.3324133655498361e-05, + "loss": 0.8495, + "step": 1679 + }, + { + "epoch": 0.41065754094353457, + "grad_norm": 3.605822801589966, + "learning_rate": 1.3316665522190628e-05, + "loss": 0.8228, + "step": 1680 + }, + { + "epoch": 0.410901979956001, + "grad_norm": 3.1482040882110596, + "learning_rate": 1.3309195309865746e-05, + "loss": 0.7748, + "step": 1681 + }, + { + "epoch": 0.4111464189684674, + "grad_norm": 3.5796356201171875, + "learning_rate": 1.3301723023206336e-05, + "loss": 0.8505, + "step": 1682 + }, + { + "epoch": 0.4113908579809338, + "grad_norm": 3.5911896228790283, + "learning_rate": 1.329424866689633e-05, + "loss": 0.8742, + "step": 1683 + }, + { + "epoch": 0.4116352969934001, + "grad_norm": 2.871198892593384, + "learning_rate": 1.3286772245620942e-05, + "loss": 0.8436, + "step": 1684 + }, + { + "epoch": 0.4118797360058665, + "grad_norm": 3.2517623901367188, + "learning_rate": 1.3279293764066693e-05, + "loss": 0.8314, + "step": 1685 + }, + { + "epoch": 0.41212417501833293, + "grad_norm": 3.427194833755493, + "learning_rate": 1.3271813226921388e-05, + "loss": 0.8339, + "step": 1686 + }, + { + "epoch": 0.41236861403079933, + "grad_norm": 4.1440582275390625, + "learning_rate": 1.3264330638874128e-05, + "loss": 0.9002, + "step": 1687 + }, + { + "epoch": 0.41261305304326573, + "grad_norm": 3.1286559104919434, + "learning_rate": 1.3256846004615293e-05, + "loss": 0.7998, + "step": 1688 + }, + { + "epoch": 0.4128574920557321, + "grad_norm": 3.3335187435150146, + "learning_rate": 1.3249359328836549e-05, + "loss": 0.8403, + "step": 1689 + }, + { + "epoch": 0.4131019310681985, + "grad_norm": 3.547224998474121, + "learning_rate": 1.3241870616230839e-05, + "loss": 0.9001, + "step": 1690 + }, + { + "epoch": 0.4133463700806649, + "grad_norm": 3.3161604404449463, + "learning_rate": 1.3234379871492381e-05, + "loss": 0.7957, + "step": 1691 + }, + { + "epoch": 0.4135908090931313, + "grad_norm": 4.421625137329102, + "learning_rate": 1.3226887099316678e-05, + "loss": 0.8809, + "step": 1692 + }, + { + "epoch": 0.41383524810559763, + "grad_norm": 3.8033032417297363, + "learning_rate": 1.3219392304400489e-05, + "loss": 0.8683, + "step": 1693 + }, + { + "epoch": 0.41407968711806403, + "grad_norm": 3.9016315937042236, + "learning_rate": 1.3211895491441853e-05, + "loss": 0.8793, + "step": 1694 + }, + { + "epoch": 0.41432412613053043, + "grad_norm": 3.785693407058716, + "learning_rate": 1.3204396665140066e-05, + "loss": 0.895, + "step": 1695 + }, + { + "epoch": 0.41456856514299684, + "grad_norm": 3.5152273178100586, + "learning_rate": 1.3196895830195691e-05, + "loss": 0.7666, + "step": 1696 + }, + { + "epoch": 0.41481300415546324, + "grad_norm": 3.0873780250549316, + "learning_rate": 1.3189392991310545e-05, + "loss": 0.8339, + "step": 1697 + }, + { + "epoch": 0.4150574431679296, + "grad_norm": 4.051246643066406, + "learning_rate": 1.3181888153187705e-05, + "loss": 0.9735, + "step": 1698 + }, + { + "epoch": 0.415301882180396, + "grad_norm": 3.4806783199310303, + "learning_rate": 1.3174381320531504e-05, + "loss": 0.7755, + "step": 1699 + }, + { + "epoch": 0.4155463211928624, + "grad_norm": 3.6442301273345947, + "learning_rate": 1.3166872498047515e-05, + "loss": 0.8943, + "step": 1700 + }, + { + "epoch": 0.4157907602053288, + "grad_norm": 4.0153374671936035, + "learning_rate": 1.315936169044257e-05, + "loss": 0.884, + "step": 1701 + }, + { + "epoch": 0.41603519921779514, + "grad_norm": 3.542607545852661, + "learning_rate": 1.3151848902424743e-05, + "loss": 0.9262, + "step": 1702 + }, + { + "epoch": 0.41627963823026154, + "grad_norm": 3.4306719303131104, + "learning_rate": 1.3144334138703341e-05, + "loss": 0.9948, + "step": 1703 + }, + { + "epoch": 0.41652407724272794, + "grad_norm": 3.551039934158325, + "learning_rate": 1.3136817403988918e-05, + "loss": 0.7998, + "step": 1704 + }, + { + "epoch": 0.41676851625519434, + "grad_norm": 3.1957240104675293, + "learning_rate": 1.3129298702993256e-05, + "loss": 0.929, + "step": 1705 + }, + { + "epoch": 0.41701295526766075, + "grad_norm": 3.0402817726135254, + "learning_rate": 1.3121778040429382e-05, + "loss": 0.8153, + "step": 1706 + }, + { + "epoch": 0.4172573942801271, + "grad_norm": 2.92474102973938, + "learning_rate": 1.311425542101154e-05, + "loss": 0.8377, + "step": 1707 + }, + { + "epoch": 0.4175018332925935, + "grad_norm": 3.0902931690216064, + "learning_rate": 1.3106730849455205e-05, + "loss": 0.8092, + "step": 1708 + }, + { + "epoch": 0.4177462723050599, + "grad_norm": 2.998044967651367, + "learning_rate": 1.3099204330477078e-05, + "loss": 0.8138, + "step": 1709 + }, + { + "epoch": 0.4179907113175263, + "grad_norm": 3.3012495040893555, + "learning_rate": 1.3091675868795077e-05, + "loss": 0.697, + "step": 1710 + }, + { + "epoch": 0.41823515032999264, + "grad_norm": 3.38061261177063, + "learning_rate": 1.3084145469128343e-05, + "loss": 0.8677, + "step": 1711 + }, + { + "epoch": 0.41847958934245905, + "grad_norm": 3.9767093658447266, + "learning_rate": 1.3076613136197225e-05, + "loss": 0.8228, + "step": 1712 + }, + { + "epoch": 0.41872402835492545, + "grad_norm": 3.446218252182007, + "learning_rate": 1.306907887472329e-05, + "loss": 0.7189, + "step": 1713 + }, + { + "epoch": 0.41896846736739185, + "grad_norm": 4.28263521194458, + "learning_rate": 1.3061542689429308e-05, + "loss": 0.7953, + "step": 1714 + }, + { + "epoch": 0.41921290637985825, + "grad_norm": 3.218669891357422, + "learning_rate": 1.305400458503926e-05, + "loss": 0.7847, + "step": 1715 + }, + { + "epoch": 0.4194573453923246, + "grad_norm": 4.574066638946533, + "learning_rate": 1.3046464566278325e-05, + "loss": 0.7932, + "step": 1716 + }, + { + "epoch": 0.419701784404791, + "grad_norm": 3.627200126647949, + "learning_rate": 1.303892263787289e-05, + "loss": 0.9059, + "step": 1717 + }, + { + "epoch": 0.4199462234172574, + "grad_norm": 3.5440592765808105, + "learning_rate": 1.3031378804550533e-05, + "loss": 0.7798, + "step": 1718 + }, + { + "epoch": 0.4201906624297238, + "grad_norm": 3.743830442428589, + "learning_rate": 1.3023833071040026e-05, + "loss": 0.8675, + "step": 1719 + }, + { + "epoch": 0.42043510144219015, + "grad_norm": 4.3647356033325195, + "learning_rate": 1.3016285442071332e-05, + "loss": 0.8428, + "step": 1720 + }, + { + "epoch": 0.42067954045465655, + "grad_norm": 3.2686777114868164, + "learning_rate": 1.3008735922375607e-05, + "loss": 0.7906, + "step": 1721 + }, + { + "epoch": 0.42092397946712296, + "grad_norm": 4.280836582183838, + "learning_rate": 1.3001184516685181e-05, + "loss": 0.8492, + "step": 1722 + }, + { + "epoch": 0.42116841847958936, + "grad_norm": 3.6485018730163574, + "learning_rate": 1.2993631229733584e-05, + "loss": 0.845, + "step": 1723 + }, + { + "epoch": 0.42141285749205576, + "grad_norm": 3.2848522663116455, + "learning_rate": 1.2986076066255504e-05, + "loss": 0.7116, + "step": 1724 + }, + { + "epoch": 0.4216572965045221, + "grad_norm": 3.6487104892730713, + "learning_rate": 1.2978519030986827e-05, + "loss": 0.8676, + "step": 1725 + }, + { + "epoch": 0.4219017355169885, + "grad_norm": 3.091217279434204, + "learning_rate": 1.2970960128664597e-05, + "loss": 0.8372, + "step": 1726 + }, + { + "epoch": 0.4221461745294549, + "grad_norm": 2.7542967796325684, + "learning_rate": 1.2963399364027031e-05, + "loss": 1.9989, + "step": 1727 + }, + { + "epoch": 0.4223906135419213, + "grad_norm": 3.3845884799957275, + "learning_rate": 1.2955836741813519e-05, + "loss": 0.7785, + "step": 1728 + }, + { + "epoch": 0.42263505255438766, + "grad_norm": 3.99977970123291, + "learning_rate": 1.294827226676461e-05, + "loss": 0.8772, + "step": 1729 + }, + { + "epoch": 0.42287949156685406, + "grad_norm": 3.4157869815826416, + "learning_rate": 1.2940705943622013e-05, + "loss": 0.8963, + "step": 1730 + }, + { + "epoch": 0.42312393057932046, + "grad_norm": 4.183679580688477, + "learning_rate": 1.2933137777128607e-05, + "loss": 0.8847, + "step": 1731 + }, + { + "epoch": 0.42336836959178686, + "grad_norm": 3.5591185092926025, + "learning_rate": 1.2925567772028412e-05, + "loss": 0.7772, + "step": 1732 + }, + { + "epoch": 0.42361280860425327, + "grad_norm": 4.990171909332275, + "learning_rate": 1.2917995933066604e-05, + "loss": 0.901, + "step": 1733 + }, + { + "epoch": 0.4238572476167196, + "grad_norm": 2.8711509704589844, + "learning_rate": 1.2910422264989519e-05, + "loss": 0.6467, + "step": 1734 + }, + { + "epoch": 0.424101686629186, + "grad_norm": 3.38092041015625, + "learning_rate": 1.2902846772544625e-05, + "loss": 0.9127, + "step": 1735 + }, + { + "epoch": 0.4243461256416524, + "grad_norm": 3.018601894378662, + "learning_rate": 1.2895269460480544e-05, + "loss": 0.8054, + "step": 1736 + }, + { + "epoch": 0.4245905646541188, + "grad_norm": 3.373168706893921, + "learning_rate": 1.2887690333547034e-05, + "loss": 0.8284, + "step": 1737 + }, + { + "epoch": 0.42483500366658516, + "grad_norm": 3.377891778945923, + "learning_rate": 1.2880109396494993e-05, + "loss": 0.8189, + "step": 1738 + }, + { + "epoch": 0.42507944267905157, + "grad_norm": 4.429141044616699, + "learning_rate": 1.287252665407645e-05, + "loss": 0.7497, + "step": 1739 + }, + { + "epoch": 0.42532388169151797, + "grad_norm": 3.5740861892700195, + "learning_rate": 1.2864942111044567e-05, + "loss": 0.7699, + "step": 1740 + }, + { + "epoch": 0.42556832070398437, + "grad_norm": 3.383955478668213, + "learning_rate": 1.2857355772153637e-05, + "loss": 0.747, + "step": 1741 + }, + { + "epoch": 0.4258127597164508, + "grad_norm": 3.5648624897003174, + "learning_rate": 1.2849767642159079e-05, + "loss": 0.9059, + "step": 1742 + }, + { + "epoch": 0.4260571987289171, + "grad_norm": 3.2955615520477295, + "learning_rate": 1.2842177725817433e-05, + "loss": 0.8256, + "step": 1743 + }, + { + "epoch": 0.4263016377413835, + "grad_norm": 3.646454334259033, + "learning_rate": 1.2834586027886355e-05, + "loss": 0.8469, + "step": 1744 + }, + { + "epoch": 0.4265460767538499, + "grad_norm": 2.9300601482391357, + "learning_rate": 1.2826992553124628e-05, + "loss": 2.0021, + "step": 1745 + }, + { + "epoch": 0.4267905157663163, + "grad_norm": 2.608389139175415, + "learning_rate": 1.2819397306292136e-05, + "loss": 2.051, + "step": 1746 + }, + { + "epoch": 0.42703495477878267, + "grad_norm": 3.829514265060425, + "learning_rate": 1.2811800292149881e-05, + "loss": 0.786, + "step": 1747 + }, + { + "epoch": 0.4272793937912491, + "grad_norm": 3.398343563079834, + "learning_rate": 1.2804201515459975e-05, + "loss": 0.9001, + "step": 1748 + }, + { + "epoch": 0.4275238328037155, + "grad_norm": 3.9880435466766357, + "learning_rate": 1.279660098098563e-05, + "loss": 0.8368, + "step": 1749 + }, + { + "epoch": 0.4277682718161819, + "grad_norm": 3.6510822772979736, + "learning_rate": 1.2788998693491163e-05, + "loss": 0.7814, + "step": 1750 + }, + { + "epoch": 0.4280127108286483, + "grad_norm": 5.199392795562744, + "learning_rate": 1.2781394657741988e-05, + "loss": 0.7836, + "step": 1751 + }, + { + "epoch": 0.4282571498411146, + "grad_norm": 3.309724807739258, + "learning_rate": 1.2773788878504614e-05, + "loss": 0.8544, + "step": 1752 + }, + { + "epoch": 0.42850158885358103, + "grad_norm": 4.230286121368408, + "learning_rate": 1.2766181360546646e-05, + "loss": 0.8382, + "step": 1753 + }, + { + "epoch": 0.42874602786604743, + "grad_norm": 3.4013397693634033, + "learning_rate": 1.2758572108636775e-05, + "loss": 0.7617, + "step": 1754 + }, + { + "epoch": 0.42899046687851383, + "grad_norm": 3.3231332302093506, + "learning_rate": 1.2750961127544782e-05, + "loss": 0.8446, + "step": 1755 + }, + { + "epoch": 0.4292349058909802, + "grad_norm": 3.5105698108673096, + "learning_rate": 1.2743348422041532e-05, + "loss": 0.7687, + "step": 1756 + }, + { + "epoch": 0.4294793449034466, + "grad_norm": 3.3357815742492676, + "learning_rate": 1.2735733996898972e-05, + "loss": 0.8778, + "step": 1757 + }, + { + "epoch": 0.429723783915913, + "grad_norm": 3.030113935470581, + "learning_rate": 1.2728117856890121e-05, + "loss": 2.0554, + "step": 1758 + }, + { + "epoch": 0.4299682229283794, + "grad_norm": 4.042571067810059, + "learning_rate": 1.2720500006789079e-05, + "loss": 0.7558, + "step": 1759 + }, + { + "epoch": 0.4302126619408458, + "grad_norm": 4.2148237228393555, + "learning_rate": 1.2712880451371015e-05, + "loss": 0.8378, + "step": 1760 + }, + { + "epoch": 0.43045710095331213, + "grad_norm": 3.0944976806640625, + "learning_rate": 1.2705259195412168e-05, + "loss": 0.9255, + "step": 1761 + }, + { + "epoch": 0.43070153996577853, + "grad_norm": 3.7010090351104736, + "learning_rate": 1.269763624368984e-05, + "loss": 0.8578, + "step": 1762 + }, + { + "epoch": 0.43094597897824494, + "grad_norm": 3.246903419494629, + "learning_rate": 1.2690011600982401e-05, + "loss": 0.8243, + "step": 1763 + }, + { + "epoch": 0.43119041799071134, + "grad_norm": 3.145646333694458, + "learning_rate": 1.2682385272069281e-05, + "loss": 0.8352, + "step": 1764 + }, + { + "epoch": 0.4314348570031777, + "grad_norm": 4.118160247802734, + "learning_rate": 1.2674757261730963e-05, + "loss": 0.7928, + "step": 1765 + }, + { + "epoch": 0.4316792960156441, + "grad_norm": 4.461591720581055, + "learning_rate": 1.2667127574748985e-05, + "loss": 0.8024, + "step": 1766 + }, + { + "epoch": 0.4319237350281105, + "grad_norm": 3.4589874744415283, + "learning_rate": 1.2659496215905937e-05, + "loss": 0.7367, + "step": 1767 + }, + { + "epoch": 0.4321681740405769, + "grad_norm": 3.478851556777954, + "learning_rate": 1.2651863189985455e-05, + "loss": 0.8295, + "step": 1768 + }, + { + "epoch": 0.4324126130530433, + "grad_norm": 2.5998263359069824, + "learning_rate": 1.2644228501772225e-05, + "loss": 1.8908, + "step": 1769 + }, + { + "epoch": 0.43265705206550964, + "grad_norm": 3.7075815200805664, + "learning_rate": 1.2636592156051965e-05, + "loss": 0.8881, + "step": 1770 + }, + { + "epoch": 0.43290149107797604, + "grad_norm": 3.9388482570648193, + "learning_rate": 1.2628954157611449e-05, + "loss": 0.8284, + "step": 1771 + }, + { + "epoch": 0.43314593009044244, + "grad_norm": 6.109549045562744, + "learning_rate": 1.2621314511238469e-05, + "loss": 0.9496, + "step": 1772 + }, + { + "epoch": 0.43339036910290885, + "grad_norm": 2.9666237831115723, + "learning_rate": 1.2613673221721859e-05, + "loss": 0.7682, + "step": 1773 + }, + { + "epoch": 0.4336348081153752, + "grad_norm": 3.7829387187957764, + "learning_rate": 1.260603029385148e-05, + "loss": 0.9357, + "step": 1774 + }, + { + "epoch": 0.4338792471278416, + "grad_norm": 3.2909271717071533, + "learning_rate": 1.2598385732418226e-05, + "loss": 0.8327, + "step": 1775 + }, + { + "epoch": 0.434123686140308, + "grad_norm": 3.5076942443847656, + "learning_rate": 1.2590739542214007e-05, + "loss": 0.82, + "step": 1776 + }, + { + "epoch": 0.4343681251527744, + "grad_norm": 3.365180015563965, + "learning_rate": 1.2583091728031757e-05, + "loss": 0.7442, + "step": 1777 + }, + { + "epoch": 0.4346125641652408, + "grad_norm": 14.855374336242676, + "learning_rate": 1.257544229466543e-05, + "loss": 0.7882, + "step": 1778 + }, + { + "epoch": 0.43485700317770715, + "grad_norm": 3.5813956260681152, + "learning_rate": 1.2567791246909995e-05, + "loss": 1.0035, + "step": 1779 + }, + { + "epoch": 0.43510144219017355, + "grad_norm": 3.7195217609405518, + "learning_rate": 1.2560138589561427e-05, + "loss": 0.7808, + "step": 1780 + }, + { + "epoch": 0.43534588120263995, + "grad_norm": 3.236727714538574, + "learning_rate": 1.255248432741672e-05, + "loss": 0.8204, + "step": 1781 + }, + { + "epoch": 0.43559032021510635, + "grad_norm": 3.5511107444763184, + "learning_rate": 1.2544828465273864e-05, + "loss": 0.8246, + "step": 1782 + }, + { + "epoch": 0.4358347592275727, + "grad_norm": 3.321305990219116, + "learning_rate": 1.2537171007931859e-05, + "loss": 0.7877, + "step": 1783 + }, + { + "epoch": 0.4360791982400391, + "grad_norm": 3.3074100017547607, + "learning_rate": 1.2529511960190699e-05, + "loss": 0.8372, + "step": 1784 + }, + { + "epoch": 0.4363236372525055, + "grad_norm": 3.4279422760009766, + "learning_rate": 1.252185132685138e-05, + "loss": 0.8411, + "step": 1785 + }, + { + "epoch": 0.4365680762649719, + "grad_norm": 3.2827301025390625, + "learning_rate": 1.2514189112715888e-05, + "loss": 0.7474, + "step": 1786 + }, + { + "epoch": 0.4368125152774383, + "grad_norm": 3.407957077026367, + "learning_rate": 1.2506525322587207e-05, + "loss": 0.7737, + "step": 1787 + }, + { + "epoch": 0.43705695428990465, + "grad_norm": 4.449306011199951, + "learning_rate": 1.2498859961269299e-05, + "loss": 0.802, + "step": 1788 + }, + { + "epoch": 0.43730139330237106, + "grad_norm": 3.125731945037842, + "learning_rate": 1.2491193033567117e-05, + "loss": 0.8348, + "step": 1789 + }, + { + "epoch": 0.43754583231483746, + "grad_norm": 4.144693374633789, + "learning_rate": 1.2483524544286596e-05, + "loss": 0.7007, + "step": 1790 + }, + { + "epoch": 0.43779027132730386, + "grad_norm": 4.416677951812744, + "learning_rate": 1.2475854498234647e-05, + "loss": 0.8256, + "step": 1791 + }, + { + "epoch": 0.4380347103397702, + "grad_norm": 4.807706356048584, + "learning_rate": 1.2468182900219158e-05, + "loss": 0.7557, + "step": 1792 + }, + { + "epoch": 0.4382791493522366, + "grad_norm": 3.2665696144104004, + "learning_rate": 1.2460509755048991e-05, + "loss": 0.6903, + "step": 1793 + }, + { + "epoch": 0.438523588364703, + "grad_norm": 4.200536251068115, + "learning_rate": 1.2452835067533975e-05, + "loss": 0.785, + "step": 1794 + }, + { + "epoch": 0.4387680273771694, + "grad_norm": 3.9921391010284424, + "learning_rate": 1.2445158842484913e-05, + "loss": 0.9047, + "step": 1795 + }, + { + "epoch": 0.43901246638963576, + "grad_norm": 5.3824849128723145, + "learning_rate": 1.2437481084713558e-05, + "loss": 0.8092, + "step": 1796 + }, + { + "epoch": 0.43925690540210216, + "grad_norm": 4.081234455108643, + "learning_rate": 1.242980179903264e-05, + "loss": 0.858, + "step": 1797 + }, + { + "epoch": 0.43950134441456856, + "grad_norm": 4.144099712371826, + "learning_rate": 1.2422120990255836e-05, + "loss": 0.8211, + "step": 1798 + }, + { + "epoch": 0.43974578342703496, + "grad_norm": 2.637526273727417, + "learning_rate": 1.2414438663197782e-05, + "loss": 1.9785, + "step": 1799 + }, + { + "epoch": 0.43999022243950137, + "grad_norm": 3.5604233741760254, + "learning_rate": 1.2406754822674062e-05, + "loss": 0.7446, + "step": 1800 + }, + { + "epoch": 0.4402346614519677, + "grad_norm": 3.7411844730377197, + "learning_rate": 1.239906947350121e-05, + "loss": 1.0626, + "step": 1801 + }, + { + "epoch": 0.4404791004644341, + "grad_norm": 3.6301281452178955, + "learning_rate": 1.2391382620496715e-05, + "loss": 0.8541, + "step": 1802 + }, + { + "epoch": 0.4407235394769005, + "grad_norm": 3.982541799545288, + "learning_rate": 1.2383694268478992e-05, + "loss": 0.8327, + "step": 1803 + }, + { + "epoch": 0.4409679784893669, + "grad_norm": 4.221193790435791, + "learning_rate": 1.2376004422267407e-05, + "loss": 0.9032, + "step": 1804 + }, + { + "epoch": 0.44121241750183327, + "grad_norm": 3.3037431240081787, + "learning_rate": 1.2368313086682262e-05, + "loss": 0.8737, + "step": 1805 + }, + { + "epoch": 0.44145685651429967, + "grad_norm": 3.748661756515503, + "learning_rate": 1.2360620266544787e-05, + "loss": 0.9294, + "step": 1806 + }, + { + "epoch": 0.44170129552676607, + "grad_norm": 3.5204269886016846, + "learning_rate": 1.2352925966677147e-05, + "loss": 0.772, + "step": 1807 + }, + { + "epoch": 0.44194573453923247, + "grad_norm": 3.786822557449341, + "learning_rate": 1.2345230191902435e-05, + "loss": 0.7596, + "step": 1808 + }, + { + "epoch": 0.4421901735516989, + "grad_norm": 4.062258720397949, + "learning_rate": 1.2337532947044664e-05, + "loss": 0.853, + "step": 1809 + }, + { + "epoch": 0.4424346125641652, + "grad_norm": 3.43574857711792, + "learning_rate": 1.2329834236928774e-05, + "loss": 0.6149, + "step": 1810 + }, + { + "epoch": 0.4426790515766316, + "grad_norm": 5.103590965270996, + "learning_rate": 1.2322134066380622e-05, + "loss": 0.746, + "step": 1811 + }, + { + "epoch": 0.442923490589098, + "grad_norm": 4.332845211029053, + "learning_rate": 1.231443244022698e-05, + "loss": 0.75, + "step": 1812 + }, + { + "epoch": 0.4431679296015644, + "grad_norm": 3.292994737625122, + "learning_rate": 1.2306729363295529e-05, + "loss": 0.9762, + "step": 1813 + }, + { + "epoch": 0.44341236861403077, + "grad_norm": 3.4701523780822754, + "learning_rate": 1.2299024840414861e-05, + "loss": 0.7983, + "step": 1814 + }, + { + "epoch": 0.4436568076264972, + "grad_norm": 4.202922821044922, + "learning_rate": 1.2291318876414477e-05, + "loss": 0.8687, + "step": 1815 + }, + { + "epoch": 0.4439012466389636, + "grad_norm": 3.164367437362671, + "learning_rate": 1.2283611476124784e-05, + "loss": 0.8103, + "step": 1816 + }, + { + "epoch": 0.44414568565143, + "grad_norm": 3.6044301986694336, + "learning_rate": 1.2275902644377082e-05, + "loss": 0.8641, + "step": 1817 + }, + { + "epoch": 0.4443901246638964, + "grad_norm": 3.4861881732940674, + "learning_rate": 1.2268192386003572e-05, + "loss": 0.9032, + "step": 1818 + }, + { + "epoch": 0.4446345636763627, + "grad_norm": 3.0395853519439697, + "learning_rate": 1.226048070583735e-05, + "loss": 0.73, + "step": 1819 + }, + { + "epoch": 0.44487900268882913, + "grad_norm": 3.807711124420166, + "learning_rate": 1.22527676087124e-05, + "loss": 0.8286, + "step": 1820 + }, + { + "epoch": 0.44512344170129553, + "grad_norm": 3.1775054931640625, + "learning_rate": 1.22450530994636e-05, + "loss": 0.773, + "step": 1821 + }, + { + "epoch": 0.44536788071376193, + "grad_norm": 11.516867637634277, + "learning_rate": 1.2237337182926706e-05, + "loss": 0.8356, + "step": 1822 + }, + { + "epoch": 0.4456123197262283, + "grad_norm": 3.5438990592956543, + "learning_rate": 1.222961986393836e-05, + "loss": 0.9018, + "step": 1823 + }, + { + "epoch": 0.4458567587386947, + "grad_norm": 3.439138412475586, + "learning_rate": 1.2221901147336086e-05, + "loss": 0.759, + "step": 1824 + }, + { + "epoch": 0.4461011977511611, + "grad_norm": 3.017598867416382, + "learning_rate": 1.2214181037958274e-05, + "loss": 0.8366, + "step": 1825 + }, + { + "epoch": 0.4463456367636275, + "grad_norm": 3.530103921890259, + "learning_rate": 1.2206459540644205e-05, + "loss": 0.7565, + "step": 1826 + }, + { + "epoch": 0.4465900757760939, + "grad_norm": 3.458371639251709, + "learning_rate": 1.2198736660234009e-05, + "loss": 0.766, + "step": 1827 + }, + { + "epoch": 0.44683451478856023, + "grad_norm": 3.8791873455047607, + "learning_rate": 1.2191012401568698e-05, + "loss": 0.8749, + "step": 1828 + }, + { + "epoch": 0.44707895380102664, + "grad_norm": 3.9143688678741455, + "learning_rate": 1.2183286769490143e-05, + "loss": 0.833, + "step": 1829 + }, + { + "epoch": 0.44732339281349304, + "grad_norm": 3.308911085128784, + "learning_rate": 1.2175559768841071e-05, + "loss": 0.7712, + "step": 1830 + }, + { + "epoch": 0.44756783182595944, + "grad_norm": 4.858282089233398, + "learning_rate": 1.2167831404465078e-05, + "loss": 0.9879, + "step": 1831 + }, + { + "epoch": 0.4478122708384258, + "grad_norm": 3.2932534217834473, + "learning_rate": 1.2160101681206602e-05, + "loss": 0.9242, + "step": 1832 + }, + { + "epoch": 0.4480567098508922, + "grad_norm": 3.069490671157837, + "learning_rate": 1.2152370603910946e-05, + "loss": 0.817, + "step": 1833 + }, + { + "epoch": 0.4483011488633586, + "grad_norm": 3.297757863998413, + "learning_rate": 1.214463817742425e-05, + "loss": 0.7234, + "step": 1834 + }, + { + "epoch": 0.448545587875825, + "grad_norm": 3.113157272338867, + "learning_rate": 1.2136904406593507e-05, + "loss": 0.8785, + "step": 1835 + }, + { + "epoch": 0.4487900268882914, + "grad_norm": 2.865623950958252, + "learning_rate": 1.2129169296266552e-05, + "loss": 2.0621, + "step": 1836 + }, + { + "epoch": 0.44903446590075774, + "grad_norm": 3.2112600803375244, + "learning_rate": 1.2121432851292057e-05, + "loss": 0.7704, + "step": 1837 + }, + { + "epoch": 0.44927890491322414, + "grad_norm": 2.9245505332946777, + "learning_rate": 1.2113695076519529e-05, + "loss": 0.848, + "step": 1838 + }, + { + "epoch": 0.44952334392569054, + "grad_norm": 3.3004794120788574, + "learning_rate": 1.2105955976799316e-05, + "loss": 0.7734, + "step": 1839 + }, + { + "epoch": 0.44976778293815695, + "grad_norm": 3.367562770843506, + "learning_rate": 1.2098215556982583e-05, + "loss": 0.8719, + "step": 1840 + }, + { + "epoch": 0.4500122219506233, + "grad_norm": 2.2853074073791504, + "learning_rate": 1.2090473821921343e-05, + "loss": 2.0348, + "step": 1841 + }, + { + "epoch": 0.4502566609630897, + "grad_norm": 4.120794773101807, + "learning_rate": 1.2082730776468414e-05, + "loss": 0.8574, + "step": 1842 + }, + { + "epoch": 0.4505010999755561, + "grad_norm": 2.4619553089141846, + "learning_rate": 1.2074986425477447e-05, + "loss": 2.0985, + "step": 1843 + }, + { + "epoch": 0.4507455389880225, + "grad_norm": 3.2559115886688232, + "learning_rate": 1.2067240773802907e-05, + "loss": 0.7199, + "step": 1844 + }, + { + "epoch": 0.4509899780004889, + "grad_norm": 3.8532941341400146, + "learning_rate": 1.205949382630007e-05, + "loss": 0.8172, + "step": 1845 + }, + { + "epoch": 0.45123441701295525, + "grad_norm": 3.294890880584717, + "learning_rate": 1.2051745587825036e-05, + "loss": 0.8261, + "step": 1846 + }, + { + "epoch": 0.45147885602542165, + "grad_norm": 2.884343147277832, + "learning_rate": 1.2043996063234707e-05, + "loss": 0.8833, + "step": 1847 + }, + { + "epoch": 0.45172329503788805, + "grad_norm": 3.017421245574951, + "learning_rate": 1.2036245257386783e-05, + "loss": 0.7939, + "step": 1848 + }, + { + "epoch": 0.45196773405035445, + "grad_norm": 3.6880812644958496, + "learning_rate": 1.2028493175139784e-05, + "loss": 0.7836, + "step": 1849 + }, + { + "epoch": 0.4522121730628208, + "grad_norm": 3.3114538192749023, + "learning_rate": 1.2020739821353022e-05, + "loss": 0.815, + "step": 1850 + }, + { + "epoch": 0.4524566120752872, + "grad_norm": 3.5703442096710205, + "learning_rate": 1.2012985200886602e-05, + "loss": 0.8905, + "step": 1851 + }, + { + "epoch": 0.4527010510877536, + "grad_norm": 2.905712604522705, + "learning_rate": 1.2005229318601429e-05, + "loss": 0.8272, + "step": 1852 + }, + { + "epoch": 0.45294549010022, + "grad_norm": 2.9451382160186768, + "learning_rate": 1.1997472179359196e-05, + "loss": 0.8074, + "step": 1853 + }, + { + "epoch": 0.4531899291126864, + "grad_norm": 3.0270562171936035, + "learning_rate": 1.1989713788022383e-05, + "loss": 0.7081, + "step": 1854 + }, + { + "epoch": 0.45343436812515275, + "grad_norm": 3.377389907836914, + "learning_rate": 1.1981954149454263e-05, + "loss": 0.7165, + "step": 1855 + }, + { + "epoch": 0.45367880713761916, + "grad_norm": 3.476201295852661, + "learning_rate": 1.1974193268518874e-05, + "loss": 0.8254, + "step": 1856 + }, + { + "epoch": 0.45392324615008556, + "grad_norm": 3.7356038093566895, + "learning_rate": 1.1966431150081053e-05, + "loss": 0.7156, + "step": 1857 + }, + { + "epoch": 0.45416768516255196, + "grad_norm": 2.9921445846557617, + "learning_rate": 1.1958667799006395e-05, + "loss": 0.8301, + "step": 1858 + }, + { + "epoch": 0.4544121241750183, + "grad_norm": 3.6033852100372314, + "learning_rate": 1.1950903220161286e-05, + "loss": 0.7017, + "step": 1859 + }, + { + "epoch": 0.4546565631874847, + "grad_norm": 3.0926873683929443, + "learning_rate": 1.194313741841286e-05, + "loss": 0.8492, + "step": 1860 + }, + { + "epoch": 0.4549010021999511, + "grad_norm": 4.124162197113037, + "learning_rate": 1.1935370398629033e-05, + "loss": 0.8546, + "step": 1861 + }, + { + "epoch": 0.4551454412124175, + "grad_norm": 3.167569398880005, + "learning_rate": 1.1927602165678485e-05, + "loss": 0.7296, + "step": 1862 + }, + { + "epoch": 0.4553898802248839, + "grad_norm": 3.3567657470703125, + "learning_rate": 1.1919832724430642e-05, + "loss": 0.8011, + "step": 1863 + }, + { + "epoch": 0.45563431923735026, + "grad_norm": 4.348156452178955, + "learning_rate": 1.1912062079755704e-05, + "loss": 0.797, + "step": 1864 + }, + { + "epoch": 0.45587875824981666, + "grad_norm": 3.06976056098938, + "learning_rate": 1.1904290236524618e-05, + "loss": 0.8518, + "step": 1865 + }, + { + "epoch": 0.45612319726228306, + "grad_norm": 3.9246575832366943, + "learning_rate": 1.1896517199609083e-05, + "loss": 0.7095, + "step": 1866 + }, + { + "epoch": 0.45636763627474947, + "grad_norm": 3.9467039108276367, + "learning_rate": 1.1888742973881544e-05, + "loss": 0.8455, + "step": 1867 + }, + { + "epoch": 0.4566120752872158, + "grad_norm": 3.6636672019958496, + "learning_rate": 1.1880967564215194e-05, + "loss": 0.865, + "step": 1868 + }, + { + "epoch": 0.4568565142996822, + "grad_norm": 3.345817804336548, + "learning_rate": 1.1873190975483965e-05, + "loss": 0.7664, + "step": 1869 + }, + { + "epoch": 0.4571009533121486, + "grad_norm": 3.5963964462280273, + "learning_rate": 1.1865413212562537e-05, + "loss": 0.7425, + "step": 1870 + }, + { + "epoch": 0.457345392324615, + "grad_norm": 3.2574336528778076, + "learning_rate": 1.185763428032631e-05, + "loss": 0.78, + "step": 1871 + }, + { + "epoch": 0.4575898313370814, + "grad_norm": 3.143759250640869, + "learning_rate": 1.1849854183651435e-05, + "loss": 2.088, + "step": 1872 + }, + { + "epoch": 0.45783427034954777, + "grad_norm": 3.23978590965271, + "learning_rate": 1.1842072927414781e-05, + "loss": 0.763, + "step": 1873 + }, + { + "epoch": 0.45807870936201417, + "grad_norm": 3.4633617401123047, + "learning_rate": 1.183429051649395e-05, + "loss": 0.8203, + "step": 1874 + }, + { + "epoch": 0.45832314837448057, + "grad_norm": 3.5833747386932373, + "learning_rate": 1.1826506955767259e-05, + "loss": 0.5899, + "step": 1875 + }, + { + "epoch": 0.458567587386947, + "grad_norm": 2.921069860458374, + "learning_rate": 1.1818722250113755e-05, + "loss": 0.6378, + "step": 1876 + }, + { + "epoch": 0.4588120263994133, + "grad_norm": 3.160431385040283, + "learning_rate": 1.18109364044132e-05, + "loss": 0.7744, + "step": 1877 + }, + { + "epoch": 0.4590564654118797, + "grad_norm": 2.5235769748687744, + "learning_rate": 1.180314942354607e-05, + "loss": 2.0303, + "step": 1878 + }, + { + "epoch": 0.4593009044243461, + "grad_norm": 3.5756592750549316, + "learning_rate": 1.179536131239355e-05, + "loss": 0.8258, + "step": 1879 + }, + { + "epoch": 0.4595453434368125, + "grad_norm": 4.374162673950195, + "learning_rate": 1.1787572075837539e-05, + "loss": 0.8376, + "step": 1880 + }, + { + "epoch": 0.4597897824492789, + "grad_norm": 4.300260066986084, + "learning_rate": 1.1779781718760641e-05, + "loss": 0.867, + "step": 1881 + }, + { + "epoch": 0.4600342214617453, + "grad_norm": 3.3397669792175293, + "learning_rate": 1.1771990246046154e-05, + "loss": 0.7503, + "step": 1882 + }, + { + "epoch": 0.4602786604742117, + "grad_norm": 3.846641778945923, + "learning_rate": 1.1764197662578087e-05, + "loss": 0.7351, + "step": 1883 + }, + { + "epoch": 0.4605230994866781, + "grad_norm": 4.7137980461120605, + "learning_rate": 1.1756403973241136e-05, + "loss": 0.7328, + "step": 1884 + }, + { + "epoch": 0.4607675384991445, + "grad_norm": 3.4643044471740723, + "learning_rate": 1.1748609182920694e-05, + "loss": 0.7494, + "step": 1885 + }, + { + "epoch": 0.4610119775116108, + "grad_norm": 4.772416591644287, + "learning_rate": 1.1740813296502845e-05, + "loss": 0.8827, + "step": 1886 + }, + { + "epoch": 0.46125641652407723, + "grad_norm": 3.249156951904297, + "learning_rate": 1.1733016318874357e-05, + "loss": 0.9389, + "step": 1887 + }, + { + "epoch": 0.46150085553654363, + "grad_norm": 3.579883575439453, + "learning_rate": 1.1725218254922685e-05, + "loss": 0.7083, + "step": 1888 + }, + { + "epoch": 0.46174529454901003, + "grad_norm": 3.744875192642212, + "learning_rate": 1.1717419109535967e-05, + "loss": 0.7988, + "step": 1889 + }, + { + "epoch": 0.46198973356147643, + "grad_norm": 3.765515089035034, + "learning_rate": 1.1709618887603013e-05, + "loss": 0.8993, + "step": 1890 + }, + { + "epoch": 0.4622341725739428, + "grad_norm": 4.36591100692749, + "learning_rate": 1.1701817594013312e-05, + "loss": 0.6889, + "step": 1891 + }, + { + "epoch": 0.4624786115864092, + "grad_norm": 2.7193212509155273, + "learning_rate": 1.1694015233657023e-05, + "loss": 0.8109, + "step": 1892 + }, + { + "epoch": 0.4627230505988756, + "grad_norm": 4.089737415313721, + "learning_rate": 1.1686211811424971e-05, + "loss": 0.8103, + "step": 1893 + }, + { + "epoch": 0.462967489611342, + "grad_norm": 4.676167964935303, + "learning_rate": 1.1678407332208652e-05, + "loss": 0.7724, + "step": 1894 + }, + { + "epoch": 0.46321192862380833, + "grad_norm": 2.849379062652588, + "learning_rate": 1.1670601800900225e-05, + "loss": 0.8262, + "step": 1895 + }, + { + "epoch": 0.46345636763627474, + "grad_norm": 3.5339317321777344, + "learning_rate": 1.1662795222392503e-05, + "loss": 0.7615, + "step": 1896 + }, + { + "epoch": 0.46370080664874114, + "grad_norm": 3.3167636394500732, + "learning_rate": 1.1654987601578958e-05, + "loss": 0.8335, + "step": 1897 + }, + { + "epoch": 0.46394524566120754, + "grad_norm": 4.1246771812438965, + "learning_rate": 1.1647178943353716e-05, + "loss": 0.7546, + "step": 1898 + }, + { + "epoch": 0.46418968467367394, + "grad_norm": 3.220127582550049, + "learning_rate": 1.1639369252611552e-05, + "loss": 0.7727, + "step": 1899 + }, + { + "epoch": 0.4644341236861403, + "grad_norm": 3.6973886489868164, + "learning_rate": 1.1631558534247894e-05, + "loss": 0.8367, + "step": 1900 + }, + { + "epoch": 0.4646785626986067, + "grad_norm": 3.421121835708618, + "learning_rate": 1.1623746793158803e-05, + "loss": 0.7907, + "step": 1901 + }, + { + "epoch": 0.4649230017110731, + "grad_norm": 5.688394546508789, + "learning_rate": 1.161593403424099e-05, + "loss": 0.7756, + "step": 1902 + }, + { + "epoch": 0.4651674407235395, + "grad_norm": 3.4840052127838135, + "learning_rate": 1.1608120262391803e-05, + "loss": 0.7949, + "step": 1903 + }, + { + "epoch": 0.46541187973600584, + "grad_norm": 3.7205164432525635, + "learning_rate": 1.1600305482509215e-05, + "loss": 0.8698, + "step": 1904 + }, + { + "epoch": 0.46565631874847224, + "grad_norm": 3.2122421264648438, + "learning_rate": 1.159248969949185e-05, + "loss": 0.9021, + "step": 1905 + }, + { + "epoch": 0.46590075776093864, + "grad_norm": 3.5042636394500732, + "learning_rate": 1.1584672918238948e-05, + "loss": 0.9091, + "step": 1906 + }, + { + "epoch": 0.46614519677340505, + "grad_norm": 3.6216835975646973, + "learning_rate": 1.1576855143650372e-05, + "loss": 0.7081, + "step": 1907 + }, + { + "epoch": 0.46638963578587145, + "grad_norm": 3.4164111614227295, + "learning_rate": 1.1569036380626615e-05, + "loss": 0.869, + "step": 1908 + }, + { + "epoch": 0.4666340747983378, + "grad_norm": 3.370806932449341, + "learning_rate": 1.1561216634068785e-05, + "loss": 0.7781, + "step": 1909 + }, + { + "epoch": 0.4668785138108042, + "grad_norm": 3.592364549636841, + "learning_rate": 1.155339590887861e-05, + "loss": 0.7527, + "step": 1910 + }, + { + "epoch": 0.4671229528232706, + "grad_norm": 3.2882182598114014, + "learning_rate": 1.1545574209958433e-05, + "loss": 0.8143, + "step": 1911 + }, + { + "epoch": 0.467367391835737, + "grad_norm": 2.8944430351257324, + "learning_rate": 1.15377515422112e-05, + "loss": 0.7112, + "step": 1912 + }, + { + "epoch": 0.46761183084820335, + "grad_norm": 3.555100202560425, + "learning_rate": 1.152992791054047e-05, + "loss": 0.6917, + "step": 1913 + }, + { + "epoch": 0.46785626986066975, + "grad_norm": 2.828277349472046, + "learning_rate": 1.1522103319850406e-05, + "loss": 0.7013, + "step": 1914 + }, + { + "epoch": 0.46810070887313615, + "grad_norm": 3.50907039642334, + "learning_rate": 1.1514277775045768e-05, + "loss": 0.7952, + "step": 1915 + }, + { + "epoch": 0.46834514788560255, + "grad_norm": 6.130204677581787, + "learning_rate": 1.1506451281031918e-05, + "loss": 0.804, + "step": 1916 + }, + { + "epoch": 0.46858958689806895, + "grad_norm": 3.4739468097686768, + "learning_rate": 1.1498623842714816e-05, + "loss": 0.8219, + "step": 1917 + }, + { + "epoch": 0.4688340259105353, + "grad_norm": 4.35533332824707, + "learning_rate": 1.1490795465001001e-05, + "loss": 0.73, + "step": 1918 + }, + { + "epoch": 0.4690784649230017, + "grad_norm": 3.3043723106384277, + "learning_rate": 1.148296615279762e-05, + "loss": 0.8029, + "step": 1919 + }, + { + "epoch": 0.4693229039354681, + "grad_norm": 3.4134469032287598, + "learning_rate": 1.1475135911012392e-05, + "loss": 0.787, + "step": 1920 + }, + { + "epoch": 0.4695673429479345, + "grad_norm": 3.1891961097717285, + "learning_rate": 1.1467304744553618e-05, + "loss": 0.6445, + "step": 1921 + }, + { + "epoch": 0.46981178196040085, + "grad_norm": 3.495227098464966, + "learning_rate": 1.1459472658330188e-05, + "loss": 0.7557, + "step": 1922 + }, + { + "epoch": 0.47005622097286726, + "grad_norm": 2.7130374908447266, + "learning_rate": 1.1451639657251564e-05, + "loss": 2.031, + "step": 1923 + }, + { + "epoch": 0.47030065998533366, + "grad_norm": 3.4243030548095703, + "learning_rate": 1.1443805746227776e-05, + "loss": 0.8233, + "step": 1924 + }, + { + "epoch": 0.47054509899780006, + "grad_norm": 3.749325752258301, + "learning_rate": 1.143597093016943e-05, + "loss": 0.7684, + "step": 1925 + }, + { + "epoch": 0.47078953801026646, + "grad_norm": 3.5366694927215576, + "learning_rate": 1.1428135213987706e-05, + "loss": 0.8376, + "step": 1926 + }, + { + "epoch": 0.4710339770227328, + "grad_norm": 3.4410207271575928, + "learning_rate": 1.1420298602594334e-05, + "loss": 0.8641, + "step": 1927 + }, + { + "epoch": 0.4712784160351992, + "grad_norm": 3.6780996322631836, + "learning_rate": 1.1412461100901614e-05, + "loss": 0.7465, + "step": 1928 + }, + { + "epoch": 0.4715228550476656, + "grad_norm": 3.1814169883728027, + "learning_rate": 1.14046227138224e-05, + "loss": 0.7965, + "step": 1929 + }, + { + "epoch": 0.471767294060132, + "grad_norm": 3.3467376232147217, + "learning_rate": 1.1396783446270105e-05, + "loss": 0.7431, + "step": 1930 + }, + { + "epoch": 0.47201173307259836, + "grad_norm": 2.4791579246520996, + "learning_rate": 1.1388943303158692e-05, + "loss": 2.0101, + "step": 1931 + }, + { + "epoch": 0.47225617208506476, + "grad_norm": 3.1137938499450684, + "learning_rate": 1.1381102289402675e-05, + "loss": 0.9033, + "step": 1932 + }, + { + "epoch": 0.47250061109753116, + "grad_norm": 3.520462989807129, + "learning_rate": 1.1373260409917104e-05, + "loss": 0.8201, + "step": 1933 + }, + { + "epoch": 0.47274505010999757, + "grad_norm": 2.4430835247039795, + "learning_rate": 1.136541766961759e-05, + "loss": 2.0582, + "step": 1934 + }, + { + "epoch": 0.47298948912246397, + "grad_norm": 3.2520508766174316, + "learning_rate": 1.1357574073420265e-05, + "loss": 0.7676, + "step": 1935 + }, + { + "epoch": 0.4732339281349303, + "grad_norm": 3.3971803188323975, + "learning_rate": 1.1349729626241807e-05, + "loss": 0.8358, + "step": 1936 + }, + { + "epoch": 0.4734783671473967, + "grad_norm": 3.3537838459014893, + "learning_rate": 1.1341884332999431e-05, + "loss": 0.923, + "step": 1937 + }, + { + "epoch": 0.4737228061598631, + "grad_norm": 3.855825424194336, + "learning_rate": 1.1334038198610868e-05, + "loss": 0.8022, + "step": 1938 + }, + { + "epoch": 0.4739672451723295, + "grad_norm": 3.2089316844940186, + "learning_rate": 1.1326191227994392e-05, + "loss": 0.7227, + "step": 1939 + }, + { + "epoch": 0.47421168418479587, + "grad_norm": 3.6445322036743164, + "learning_rate": 1.1318343426068795e-05, + "loss": 0.8884, + "step": 1940 + }, + { + "epoch": 0.47445612319726227, + "grad_norm": 3.0271670818328857, + "learning_rate": 1.1310494797753382e-05, + "loss": 0.618, + "step": 1941 + }, + { + "epoch": 0.47470056220972867, + "grad_norm": 4.566411018371582, + "learning_rate": 1.1302645347967993e-05, + "loss": 0.776, + "step": 1942 + }, + { + "epoch": 0.4749450012221951, + "grad_norm": 3.470848321914673, + "learning_rate": 1.1294795081632968e-05, + "loss": 0.8234, + "step": 1943 + }, + { + "epoch": 0.4751894402346615, + "grad_norm": 3.8001720905303955, + "learning_rate": 1.128694400366917e-05, + "loss": 0.7269, + "step": 1944 + }, + { + "epoch": 0.4754338792471278, + "grad_norm": 3.591465950012207, + "learning_rate": 1.1279092118997958e-05, + "loss": 0.7565, + "step": 1945 + }, + { + "epoch": 0.4756783182595942, + "grad_norm": 3.489041328430176, + "learning_rate": 1.1271239432541208e-05, + "loss": 0.7692, + "step": 1946 + }, + { + "epoch": 0.4759227572720606, + "grad_norm": 2.8316056728363037, + "learning_rate": 1.1263385949221294e-05, + "loss": 0.8811, + "step": 1947 + }, + { + "epoch": 0.47616719628452703, + "grad_norm": 3.651979684829712, + "learning_rate": 1.1255531673961091e-05, + "loss": 0.7406, + "step": 1948 + }, + { + "epoch": 0.4764116352969934, + "grad_norm": 3.323721170425415, + "learning_rate": 1.1247676611683964e-05, + "loss": 0.7923, + "step": 1949 + }, + { + "epoch": 0.4766560743094598, + "grad_norm": 4.642440319061279, + "learning_rate": 1.1239820767313788e-05, + "loss": 0.7638, + "step": 1950 + }, + { + "epoch": 0.4769005133219262, + "grad_norm": 3.692936897277832, + "learning_rate": 1.1231964145774906e-05, + "loss": 0.7265, + "step": 1951 + }, + { + "epoch": 0.4771449523343926, + "grad_norm": 4.496062755584717, + "learning_rate": 1.1224106751992164e-05, + "loss": 0.7691, + "step": 1952 + }, + { + "epoch": 0.477389391346859, + "grad_norm": 3.4846792221069336, + "learning_rate": 1.1216248590890886e-05, + "loss": 0.7858, + "step": 1953 + }, + { + "epoch": 0.47763383035932533, + "grad_norm": 4.358302116394043, + "learning_rate": 1.1208389667396879e-05, + "loss": 0.8424, + "step": 1954 + }, + { + "epoch": 0.47787826937179173, + "grad_norm": 3.3236751556396484, + "learning_rate": 1.120052998643643e-05, + "loss": 0.8013, + "step": 1955 + }, + { + "epoch": 0.47812270838425813, + "grad_norm": 4.438330173492432, + "learning_rate": 1.119266955293629e-05, + "loss": 0.7891, + "step": 1956 + }, + { + "epoch": 0.47836714739672453, + "grad_norm": 4.703066825866699, + "learning_rate": 1.1184808371823696e-05, + "loss": 0.7928, + "step": 1957 + }, + { + "epoch": 0.4786115864091909, + "grad_norm": 4.193386554718018, + "learning_rate": 1.1176946448026348e-05, + "loss": 0.8049, + "step": 1958 + }, + { + "epoch": 0.4788560254216573, + "grad_norm": 3.0352625846862793, + "learning_rate": 1.1169083786472407e-05, + "loss": 0.7952, + "step": 1959 + }, + { + "epoch": 0.4791004644341237, + "grad_norm": 3.2521796226501465, + "learning_rate": 1.1161220392090503e-05, + "loss": 0.6706, + "step": 1960 + }, + { + "epoch": 0.4793449034465901, + "grad_norm": 3.6830103397369385, + "learning_rate": 1.1153356269809721e-05, + "loss": 0.7813, + "step": 1961 + }, + { + "epoch": 0.4795893424590565, + "grad_norm": 4.002415180206299, + "learning_rate": 1.1145491424559607e-05, + "loss": 0.7248, + "step": 1962 + }, + { + "epoch": 0.47983378147152284, + "grad_norm": 3.116703987121582, + "learning_rate": 1.1137625861270151e-05, + "loss": 0.7807, + "step": 1963 + }, + { + "epoch": 0.48007822048398924, + "grad_norm": 2.9933393001556396, + "learning_rate": 1.1129759584871802e-05, + "loss": 2.0964, + "step": 1964 + }, + { + "epoch": 0.48032265949645564, + "grad_norm": 3.455176830291748, + "learning_rate": 1.1121892600295456e-05, + "loss": 0.7856, + "step": 1965 + }, + { + "epoch": 0.48056709850892204, + "grad_norm": 2.866671085357666, + "learning_rate": 1.1114024912472449e-05, + "loss": 2.049, + "step": 1966 + }, + { + "epoch": 0.4808115375213884, + "grad_norm": 3.603726625442505, + "learning_rate": 1.1106156526334559e-05, + "loss": 0.7538, + "step": 1967 + }, + { + "epoch": 0.4810559765338548, + "grad_norm": 2.406770706176758, + "learning_rate": 1.1098287446814001e-05, + "loss": 1.944, + "step": 1968 + }, + { + "epoch": 0.4813004155463212, + "grad_norm": 4.4387712478637695, + "learning_rate": 1.1090417678843423e-05, + "loss": 0.9268, + "step": 1969 + }, + { + "epoch": 0.4815448545587876, + "grad_norm": 3.8196306228637695, + "learning_rate": 1.1082547227355911e-05, + "loss": 0.8084, + "step": 1970 + }, + { + "epoch": 0.481789293571254, + "grad_norm": 3.659147024154663, + "learning_rate": 1.1074676097284973e-05, + "loss": 0.7973, + "step": 1971 + }, + { + "epoch": 0.48203373258372034, + "grad_norm": 4.428442001342773, + "learning_rate": 1.1066804293564541e-05, + "loss": 0.8604, + "step": 1972 + }, + { + "epoch": 0.48227817159618674, + "grad_norm": 4.519818305969238, + "learning_rate": 1.1058931821128982e-05, + "loss": 0.6759, + "step": 1973 + }, + { + "epoch": 0.48252261060865315, + "grad_norm": 3.2999141216278076, + "learning_rate": 1.1051058684913067e-05, + "loss": 0.7477, + "step": 1974 + }, + { + "epoch": 0.48276704962111955, + "grad_norm": 3.5467281341552734, + "learning_rate": 1.1043184889851992e-05, + "loss": 0.7753, + "step": 1975 + }, + { + "epoch": 0.4830114886335859, + "grad_norm": 3.5430102348327637, + "learning_rate": 1.1035310440881359e-05, + "loss": 0.8758, + "step": 1976 + }, + { + "epoch": 0.4832559276460523, + "grad_norm": 4.608526229858398, + "learning_rate": 1.1027435342937189e-05, + "loss": 0.7839, + "step": 1977 + }, + { + "epoch": 0.4835003666585187, + "grad_norm": 3.1568756103515625, + "learning_rate": 1.10195596009559e-05, + "loss": 0.6923, + "step": 1978 + }, + { + "epoch": 0.4837448056709851, + "grad_norm": 3.5793933868408203, + "learning_rate": 1.1011683219874324e-05, + "loss": 0.8686, + "step": 1979 + }, + { + "epoch": 0.4839892446834515, + "grad_norm": 4.421467304229736, + "learning_rate": 1.1003806204629683e-05, + "loss": 0.7134, + "step": 1980 + }, + { + "epoch": 0.48423368369591785, + "grad_norm": 4.978440284729004, + "learning_rate": 1.0995928560159608e-05, + "loss": 0.6484, + "step": 1981 + }, + { + "epoch": 0.48447812270838425, + "grad_norm": 3.6223433017730713, + "learning_rate": 1.0988050291402113e-05, + "loss": 0.7283, + "step": 1982 + }, + { + "epoch": 0.48472256172085065, + "grad_norm": 3.2677972316741943, + "learning_rate": 1.098017140329561e-05, + "loss": 0.7855, + "step": 1983 + }, + { + "epoch": 0.48496700073331706, + "grad_norm": 3.3966641426086426, + "learning_rate": 1.0972291900778892e-05, + "loss": 0.8409, + "step": 1984 + }, + { + "epoch": 0.4852114397457834, + "grad_norm": 2.9886679649353027, + "learning_rate": 1.0964411788791156e-05, + "loss": 0.7301, + "step": 1985 + }, + { + "epoch": 0.4854558787582498, + "grad_norm": 3.46282958984375, + "learning_rate": 1.0956531072271959e-05, + "loss": 0.7322, + "step": 1986 + }, + { + "epoch": 0.4857003177707162, + "grad_norm": 3.8131372928619385, + "learning_rate": 1.0948649756161246e-05, + "loss": 0.7173, + "step": 1987 + }, + { + "epoch": 0.4859447567831826, + "grad_norm": 3.8756089210510254, + "learning_rate": 1.0940767845399341e-05, + "loss": 0.786, + "step": 1988 + }, + { + "epoch": 0.486189195795649, + "grad_norm": 5.3649725914001465, + "learning_rate": 1.0932885344926931e-05, + "loss": 0.8385, + "step": 1989 + }, + { + "epoch": 0.48643363480811536, + "grad_norm": 3.4769184589385986, + "learning_rate": 1.0925002259685089e-05, + "loss": 0.7209, + "step": 1990 + }, + { + "epoch": 0.48667807382058176, + "grad_norm": 4.807114601135254, + "learning_rate": 1.0917118594615237e-05, + "loss": 0.7383, + "step": 1991 + }, + { + "epoch": 0.48692251283304816, + "grad_norm": 2.774432897567749, + "learning_rate": 1.0909234354659173e-05, + "loss": 0.7314, + "step": 1992 + }, + { + "epoch": 0.48716695184551456, + "grad_norm": 3.699162006378174, + "learning_rate": 1.0901349544759046e-05, + "loss": 0.8145, + "step": 1993 + }, + { + "epoch": 0.4874113908579809, + "grad_norm": 3.1336519718170166, + "learning_rate": 1.0893464169857368e-05, + "loss": 0.7637, + "step": 1994 + }, + { + "epoch": 0.4876558298704473, + "grad_norm": 3.9432170391082764, + "learning_rate": 1.0885578234897003e-05, + "loss": 0.883, + "step": 1995 + }, + { + "epoch": 0.4879002688829137, + "grad_norm": 2.927171230316162, + "learning_rate": 1.0877691744821171e-05, + "loss": 0.719, + "step": 1996 + }, + { + "epoch": 0.4881447078953801, + "grad_norm": 3.5491185188293457, + "learning_rate": 1.0869804704573435e-05, + "loss": 0.7115, + "step": 1997 + }, + { + "epoch": 0.4883891469078465, + "grad_norm": 3.6908175945281982, + "learning_rate": 1.08619171190977e-05, + "loss": 0.8634, + "step": 1998 + }, + { + "epoch": 0.48863358592031286, + "grad_norm": 3.3738112449645996, + "learning_rate": 1.0854028993338222e-05, + "loss": 0.664, + "step": 1999 + }, + { + "epoch": 0.48887802493277926, + "grad_norm": 3.688183307647705, + "learning_rate": 1.0846140332239584e-05, + "loss": 0.8448, + "step": 2000 + }, + { + "epoch": 0.48912246394524567, + "grad_norm": 3.9575772285461426, + "learning_rate": 1.0838251140746717e-05, + "loss": 0.8176, + "step": 2001 + }, + { + "epoch": 0.48936690295771207, + "grad_norm": 3.4544336795806885, + "learning_rate": 1.0830361423804872e-05, + "loss": 0.7331, + "step": 2002 + }, + { + "epoch": 0.4896113419701784, + "grad_norm": 2.768550157546997, + "learning_rate": 1.082247118635964e-05, + "loss": 0.8275, + "step": 2003 + }, + { + "epoch": 0.4898557809826448, + "grad_norm": 3.2085041999816895, + "learning_rate": 1.0814580433356933e-05, + "loss": 0.717, + "step": 2004 + }, + { + "epoch": 0.4901002199951112, + "grad_norm": 3.5018882751464844, + "learning_rate": 1.0806689169742988e-05, + "loss": 0.6366, + "step": 2005 + }, + { + "epoch": 0.4903446590075776, + "grad_norm": 3.169766426086426, + "learning_rate": 1.0798797400464361e-05, + "loss": 0.7642, + "step": 2006 + }, + { + "epoch": 0.490589098020044, + "grad_norm": 3.3302865028381348, + "learning_rate": 1.0790905130467923e-05, + "loss": 0.801, + "step": 2007 + }, + { + "epoch": 0.49083353703251037, + "grad_norm": 4.264266490936279, + "learning_rate": 1.0783012364700865e-05, + "loss": 0.7877, + "step": 2008 + }, + { + "epoch": 0.49107797604497677, + "grad_norm": 2.8109889030456543, + "learning_rate": 1.077511910811068e-05, + "loss": 0.7575, + "step": 2009 + }, + { + "epoch": 0.4913224150574432, + "grad_norm": 3.112183094024658, + "learning_rate": 1.0767225365645178e-05, + "loss": 0.7362, + "step": 2010 + }, + { + "epoch": 0.4915668540699096, + "grad_norm": 3.1281116008758545, + "learning_rate": 1.0759331142252463e-05, + "loss": 0.6488, + "step": 2011 + }, + { + "epoch": 0.4918112930823759, + "grad_norm": 3.025792121887207, + "learning_rate": 1.0751436442880953e-05, + "loss": 0.7768, + "step": 2012 + }, + { + "epoch": 0.4920557320948423, + "grad_norm": 3.433082342147827, + "learning_rate": 1.0743541272479356e-05, + "loss": 0.7959, + "step": 2013 + }, + { + "epoch": 0.4923001711073087, + "grad_norm": 3.600435495376587, + "learning_rate": 1.0735645635996676e-05, + "loss": 0.7833, + "step": 2014 + }, + { + "epoch": 0.49254461011977513, + "grad_norm": 3.4164845943450928, + "learning_rate": 1.072774953838221e-05, + "loss": 0.6478, + "step": 2015 + }, + { + "epoch": 0.49278904913224153, + "grad_norm": 3.6529622077941895, + "learning_rate": 1.0719852984585546e-05, + "loss": 0.766, + "step": 2016 + }, + { + "epoch": 0.4930334881447079, + "grad_norm": 4.316786289215088, + "learning_rate": 1.0711955979556555e-05, + "loss": 0.8255, + "step": 2017 + }, + { + "epoch": 0.4932779271571743, + "grad_norm": 2.940098762512207, + "learning_rate": 1.0704058528245389e-05, + "loss": 0.7522, + "step": 2018 + }, + { + "epoch": 0.4935223661696407, + "grad_norm": 4.512490272521973, + "learning_rate": 1.0696160635602488e-05, + "loss": 0.8231, + "step": 2019 + }, + { + "epoch": 0.4937668051821071, + "grad_norm": 3.141284227371216, + "learning_rate": 1.0688262306578566e-05, + "loss": 0.9031, + "step": 2020 + }, + { + "epoch": 0.49401124419457343, + "grad_norm": 4.085526466369629, + "learning_rate": 1.0680363546124599e-05, + "loss": 0.7519, + "step": 2021 + }, + { + "epoch": 0.49425568320703983, + "grad_norm": 2.9810550212860107, + "learning_rate": 1.067246435919185e-05, + "loss": 0.6218, + "step": 2022 + }, + { + "epoch": 0.49450012221950623, + "grad_norm": 2.8658111095428467, + "learning_rate": 1.0664564750731838e-05, + "loss": 0.6835, + "step": 2023 + }, + { + "epoch": 0.49474456123197263, + "grad_norm": 3.3811020851135254, + "learning_rate": 1.0656664725696352e-05, + "loss": 0.822, + "step": 2024 + }, + { + "epoch": 0.49498900024443904, + "grad_norm": 3.2734344005584717, + "learning_rate": 1.064876428903744e-05, + "loss": 0.835, + "step": 2025 + }, + { + "epoch": 0.4952334392569054, + "grad_norm": 3.8637382984161377, + "learning_rate": 1.0640863445707408e-05, + "loss": 0.7283, + "step": 2026 + }, + { + "epoch": 0.4954778782693718, + "grad_norm": 3.770451307296753, + "learning_rate": 1.0632962200658816e-05, + "loss": 0.7497, + "step": 2027 + }, + { + "epoch": 0.4957223172818382, + "grad_norm": 3.022958755493164, + "learning_rate": 1.0625060558844478e-05, + "loss": 2.0316, + "step": 2028 + }, + { + "epoch": 0.4959667562943046, + "grad_norm": 3.7786545753479004, + "learning_rate": 1.0617158525217457e-05, + "loss": 0.762, + "step": 2029 + }, + { + "epoch": 0.49621119530677094, + "grad_norm": 3.3836238384246826, + "learning_rate": 1.0609256104731057e-05, + "loss": 0.7447, + "step": 2030 + }, + { + "epoch": 0.49645563431923734, + "grad_norm": 3.8169052600860596, + "learning_rate": 1.060135330233883e-05, + "loss": 0.6994, + "step": 2031 + }, + { + "epoch": 0.49670007333170374, + "grad_norm": 3.290356397628784, + "learning_rate": 1.0593450122994567e-05, + "loss": 0.6643, + "step": 2032 + }, + { + "epoch": 0.49694451234417014, + "grad_norm": 2.974830150604248, + "learning_rate": 1.0585546571652288e-05, + "loss": 0.7194, + "step": 2033 + }, + { + "epoch": 0.49718895135663654, + "grad_norm": 2.969827651977539, + "learning_rate": 1.0577642653266253e-05, + "loss": 0.8106, + "step": 2034 + }, + { + "epoch": 0.4974333903691029, + "grad_norm": 3.1830921173095703, + "learning_rate": 1.0569738372790956e-05, + "loss": 0.7061, + "step": 2035 + }, + { + "epoch": 0.4976778293815693, + "grad_norm": 2.936081886291504, + "learning_rate": 1.0561833735181108e-05, + "loss": 2.0002, + "step": 2036 + }, + { + "epoch": 0.4979222683940357, + "grad_norm": 2.954249143600464, + "learning_rate": 1.055392874539165e-05, + "loss": 0.7313, + "step": 2037 + }, + { + "epoch": 0.4981667074065021, + "grad_norm": 3.6443800926208496, + "learning_rate": 1.0546023408377743e-05, + "loss": 0.8397, + "step": 2038 + }, + { + "epoch": 0.49841114641896844, + "grad_norm": 3.4366455078125, + "learning_rate": 1.0538117729094766e-05, + "loss": 0.7084, + "step": 2039 + }, + { + "epoch": 0.49865558543143484, + "grad_norm": 2.2596499919891357, + "learning_rate": 1.0530211712498306e-05, + "loss": 1.9601, + "step": 2040 + }, + { + "epoch": 0.49890002444390125, + "grad_norm": 3.0489089488983154, + "learning_rate": 1.0522305363544172e-05, + "loss": 0.7073, + "step": 2041 + }, + { + "epoch": 0.49914446345636765, + "grad_norm": 4.950798511505127, + "learning_rate": 1.0514398687188372e-05, + "loss": 0.8156, + "step": 2042 + }, + { + "epoch": 0.49938890246883405, + "grad_norm": 3.044152021408081, + "learning_rate": 1.0506491688387128e-05, + "loss": 0.6842, + "step": 2043 + }, + { + "epoch": 0.4996333414813004, + "grad_norm": 3.6759657859802246, + "learning_rate": 1.0498584372096858e-05, + "loss": 0.7377, + "step": 2044 + }, + { + "epoch": 0.4998777804937668, + "grad_norm": 2.9382901191711426, + "learning_rate": 1.0490676743274181e-05, + "loss": 0.749, + "step": 2045 + }, + { + "epoch": 0.5001222195062331, + "grad_norm": 3.570497989654541, + "learning_rate": 1.0482768806875911e-05, + "loss": 0.8168, + "step": 2046 + }, + { + "epoch": 0.5003666585186995, + "grad_norm": 3.9341888427734375, + "learning_rate": 1.0474860567859058e-05, + "loss": 0.7277, + "step": 2047 + }, + { + "epoch": 0.500611097531166, + "grad_norm": 7.264518737792969, + "learning_rate": 1.0466952031180813e-05, + "loss": 0.6585, + "step": 2048 + }, + { + "epoch": 0.5008555365436324, + "grad_norm": 3.4198923110961914, + "learning_rate": 1.0459043201798563e-05, + "loss": 0.8649, + "step": 2049 + }, + { + "epoch": 0.5010999755560988, + "grad_norm": 3.1877851486206055, + "learning_rate": 1.0451134084669876e-05, + "loss": 0.6722, + "step": 2050 + }, + { + "epoch": 0.5013444145685652, + "grad_norm": 3.139477014541626, + "learning_rate": 1.04432246847525e-05, + "loss": 0.761, + "step": 2051 + }, + { + "epoch": 0.5015888535810316, + "grad_norm": 3.5756216049194336, + "learning_rate": 1.0435315007004357e-05, + "loss": 0.7999, + "step": 2052 + }, + { + "epoch": 0.501833292593498, + "grad_norm": 2.963740587234497, + "learning_rate": 1.0427405056383548e-05, + "loss": 0.7494, + "step": 2053 + }, + { + "epoch": 0.5020777316059644, + "grad_norm": 2.5787642002105713, + "learning_rate": 1.0419494837848343e-05, + "loss": 1.9864, + "step": 2054 + }, + { + "epoch": 0.5023221706184307, + "grad_norm": 3.3031718730926514, + "learning_rate": 1.0411584356357178e-05, + "loss": 0.7257, + "step": 2055 + }, + { + "epoch": 0.502566609630897, + "grad_norm": 3.0101404190063477, + "learning_rate": 1.0403673616868659e-05, + "loss": 0.8019, + "step": 2056 + }, + { + "epoch": 0.5028110486433635, + "grad_norm": 2.8920490741729736, + "learning_rate": 1.0395762624341544e-05, + "loss": 0.7606, + "step": 2057 + }, + { + "epoch": 0.5030554876558299, + "grad_norm": 3.88023042678833, + "learning_rate": 1.0387851383734763e-05, + "loss": 0.7647, + "step": 2058 + }, + { + "epoch": 0.5032999266682963, + "grad_norm": 3.9406485557556152, + "learning_rate": 1.0379939900007394e-05, + "loss": 0.7223, + "step": 2059 + }, + { + "epoch": 0.5035443656807627, + "grad_norm": 3.2017197608947754, + "learning_rate": 1.0372028178118665e-05, + "loss": 0.7909, + "step": 2060 + }, + { + "epoch": 0.5037888046932291, + "grad_norm": 3.2319176197052, + "learning_rate": 1.0364116223027956e-05, + "loss": 0.6608, + "step": 2061 + }, + { + "epoch": 0.5040332437056955, + "grad_norm": 3.2559704780578613, + "learning_rate": 1.0356204039694795e-05, + "loss": 0.7812, + "step": 2062 + }, + { + "epoch": 0.5042776827181619, + "grad_norm": 3.1422622203826904, + "learning_rate": 1.0348291633078851e-05, + "loss": 0.7641, + "step": 2063 + }, + { + "epoch": 0.5045221217306282, + "grad_norm": 3.148240804672241, + "learning_rate": 1.0340379008139933e-05, + "loss": 0.7373, + "step": 2064 + }, + { + "epoch": 0.5047665607430946, + "grad_norm": 3.469862699508667, + "learning_rate": 1.0332466169837992e-05, + "loss": 0.811, + "step": 2065 + }, + { + "epoch": 0.505010999755561, + "grad_norm": 3.4699959754943848, + "learning_rate": 1.03245531231331e-05, + "loss": 0.6805, + "step": 2066 + }, + { + "epoch": 0.5052554387680274, + "grad_norm": 4.243104934692383, + "learning_rate": 1.0316639872985471e-05, + "loss": 0.7452, + "step": 2067 + }, + { + "epoch": 0.5054998777804938, + "grad_norm": 3.567861557006836, + "learning_rate": 1.0308726424355446e-05, + "loss": 0.7222, + "step": 2068 + }, + { + "epoch": 0.5057443167929602, + "grad_norm": 3.9209423065185547, + "learning_rate": 1.0300812782203486e-05, + "loss": 0.7033, + "step": 2069 + }, + { + "epoch": 0.5059887558054266, + "grad_norm": 4.002355098724365, + "learning_rate": 1.0292898951490174e-05, + "loss": 0.6573, + "step": 2070 + }, + { + "epoch": 0.506233194817893, + "grad_norm": 3.7021231651306152, + "learning_rate": 1.0284984937176213e-05, + "loss": 0.7672, + "step": 2071 + }, + { + "epoch": 0.5064776338303594, + "grad_norm": 2.4699697494506836, + "learning_rate": 1.0277070744222419e-05, + "loss": 2.0226, + "step": 2072 + }, + { + "epoch": 0.5067220728428257, + "grad_norm": 4.644677639007568, + "learning_rate": 1.0269156377589722e-05, + "loss": 0.7015, + "step": 2073 + }, + { + "epoch": 0.5069665118552921, + "grad_norm": 3.0466513633728027, + "learning_rate": 1.026124184223916e-05, + "loss": 0.8128, + "step": 2074 + }, + { + "epoch": 0.5072109508677585, + "grad_norm": 4.4984354972839355, + "learning_rate": 1.025332714313188e-05, + "loss": 0.7525, + "step": 2075 + }, + { + "epoch": 0.5074553898802249, + "grad_norm": 4.234147071838379, + "learning_rate": 1.0245412285229124e-05, + "loss": 0.7366, + "step": 2076 + }, + { + "epoch": 0.5076998288926913, + "grad_norm": 3.7217657566070557, + "learning_rate": 1.023749727349224e-05, + "loss": 0.7647, + "step": 2077 + }, + { + "epoch": 0.5079442679051577, + "grad_norm": 3.2942984104156494, + "learning_rate": 1.0229582112882668e-05, + "loss": 0.8248, + "step": 2078 + }, + { + "epoch": 0.5081887069176241, + "grad_norm": 3.738570213317871, + "learning_rate": 1.0221666808361947e-05, + "loss": 0.5877, + "step": 2079 + }, + { + "epoch": 0.5084331459300905, + "grad_norm": 3.184098720550537, + "learning_rate": 1.0213751364891702e-05, + "loss": 0.6752, + "step": 2080 + }, + { + "epoch": 0.5086775849425569, + "grad_norm": 3.1130599975585938, + "learning_rate": 1.0205835787433645e-05, + "loss": 0.6944, + "step": 2081 + }, + { + "epoch": 0.5089220239550232, + "grad_norm": 2.9968113899230957, + "learning_rate": 1.0197920080949574e-05, + "loss": 0.8001, + "step": 2082 + }, + { + "epoch": 0.5091664629674896, + "grad_norm": 3.2675774097442627, + "learning_rate": 1.0190004250401369e-05, + "loss": 0.7529, + "step": 2083 + }, + { + "epoch": 0.509410901979956, + "grad_norm": 2.6524808406829834, + "learning_rate": 1.0182088300750983e-05, + "loss": 1.9075, + "step": 2084 + }, + { + "epoch": 0.5096553409924224, + "grad_norm": 3.6236817836761475, + "learning_rate": 1.0174172236960447e-05, + "loss": 0.8717, + "step": 2085 + }, + { + "epoch": 0.5098997800048888, + "grad_norm": 3.115124225616455, + "learning_rate": 1.0166256063991861e-05, + "loss": 0.6924, + "step": 2086 + }, + { + "epoch": 0.5101442190173552, + "grad_norm": 3.547389030456543, + "learning_rate": 1.0158339786807401e-05, + "loss": 0.6893, + "step": 2087 + }, + { + "epoch": 0.5103886580298216, + "grad_norm": 3.550128221511841, + "learning_rate": 1.0150423410369294e-05, + "loss": 0.6903, + "step": 2088 + }, + { + "epoch": 0.510633097042288, + "grad_norm": 3.768012523651123, + "learning_rate": 1.0142506939639847e-05, + "loss": 0.8392, + "step": 2089 + }, + { + "epoch": 0.5108775360547544, + "grad_norm": 2.932687759399414, + "learning_rate": 1.013459037958141e-05, + "loss": 0.7111, + "step": 2090 + }, + { + "epoch": 0.5111219750672207, + "grad_norm": 3.274658441543579, + "learning_rate": 1.0126673735156402e-05, + "loss": 0.8295, + "step": 2091 + }, + { + "epoch": 0.5113664140796871, + "grad_norm": 4.2612624168396, + "learning_rate": 1.0118757011327285e-05, + "loss": 0.8573, + "step": 2092 + }, + { + "epoch": 0.5116108530921535, + "grad_norm": 2.732764959335327, + "learning_rate": 1.0110840213056575e-05, + "loss": 0.6972, + "step": 2093 + }, + { + "epoch": 0.5118552921046199, + "grad_norm": 5.253904342651367, + "learning_rate": 1.0102923345306834e-05, + "loss": 0.7623, + "step": 2094 + }, + { + "epoch": 0.5120997311170863, + "grad_norm": 2.9598002433776855, + "learning_rate": 1.009500641304067e-05, + "loss": 0.5901, + "step": 2095 + }, + { + "epoch": 0.5123441701295527, + "grad_norm": 7.8216776847839355, + "learning_rate": 1.0087089421220727e-05, + "loss": 0.8268, + "step": 2096 + }, + { + "epoch": 0.5125886091420191, + "grad_norm": 3.346170663833618, + "learning_rate": 1.0079172374809689e-05, + "loss": 0.6633, + "step": 2097 + }, + { + "epoch": 0.5128330481544855, + "grad_norm": 3.266075372695923, + "learning_rate": 1.0071255278770276e-05, + "loss": 0.8366, + "step": 2098 + }, + { + "epoch": 0.5130774871669519, + "grad_norm": 3.055297613143921, + "learning_rate": 1.0063338138065235e-05, + "loss": 0.6402, + "step": 2099 + }, + { + "epoch": 0.5133219261794182, + "grad_norm": 5.6008830070495605, + "learning_rate": 1.0055420957657344e-05, + "loss": 0.7894, + "step": 2100 + }, + { + "epoch": 0.5135663651918846, + "grad_norm": 2.906796455383301, + "learning_rate": 1.0047503742509405e-05, + "loss": 0.7169, + "step": 2101 + }, + { + "epoch": 0.513810804204351, + "grad_norm": 3.6960554122924805, + "learning_rate": 1.0039586497584241e-05, + "loss": 0.7333, + "step": 2102 + }, + { + "epoch": 0.5140552432168174, + "grad_norm": 3.835498094558716, + "learning_rate": 1.0031669227844693e-05, + "loss": 0.7082, + "step": 2103 + }, + { + "epoch": 0.5142996822292838, + "grad_norm": 3.1887574195861816, + "learning_rate": 1.0023751938253618e-05, + "loss": 0.8432, + "step": 2104 + }, + { + "epoch": 0.5145441212417502, + "grad_norm": 3.2100727558135986, + "learning_rate": 1.0015834633773893e-05, + "loss": 0.6839, + "step": 2105 + }, + { + "epoch": 0.5147885602542166, + "grad_norm": 4.217765808105469, + "learning_rate": 1.0007917319368392e-05, + "loss": 0.7569, + "step": 2106 + }, + { + "epoch": 0.515032999266683, + "grad_norm": 4.070542812347412, + "learning_rate": 1e-05, + "loss": 0.753, + "step": 2107 + }, + { + "epoch": 0.5152774382791494, + "grad_norm": 3.809004783630371, + "learning_rate": 9.99208268063161e-06, + "loss": 0.7754, + "step": 2108 + }, + { + "epoch": 0.5155218772916157, + "grad_norm": 2.945784568786621, + "learning_rate": 9.98416536622611e-06, + "loss": 0.7516, + "step": 2109 + }, + { + "epoch": 0.5157663163040821, + "grad_norm": 3.0451085567474365, + "learning_rate": 9.976248061746383e-06, + "loss": 0.8051, + "step": 2110 + }, + { + "epoch": 0.5160107553165485, + "grad_norm": 4.2851786613464355, + "learning_rate": 9.968330772155312e-06, + "loss": 0.8682, + "step": 2111 + }, + { + "epoch": 0.5162551943290149, + "grad_norm": 3.8674421310424805, + "learning_rate": 9.960413502415764e-06, + "loss": 0.7483, + "step": 2112 + }, + { + "epoch": 0.5164996333414813, + "grad_norm": 3.2832260131835938, + "learning_rate": 9.9524962574906e-06, + "loss": 0.7464, + "step": 2113 + }, + { + "epoch": 0.5167440723539477, + "grad_norm": 3.5348575115203857, + "learning_rate": 9.944579042342661e-06, + "loss": 0.886, + "step": 2114 + }, + { + "epoch": 0.5169885113664141, + "grad_norm": 3.5706822872161865, + "learning_rate": 9.936661861934765e-06, + "loss": 0.6823, + "step": 2115 + }, + { + "epoch": 0.5172329503788805, + "grad_norm": 3.5740959644317627, + "learning_rate": 9.928744721229726e-06, + "loss": 0.7477, + "step": 2116 + }, + { + "epoch": 0.5174773893913469, + "grad_norm": 3.688131093978882, + "learning_rate": 9.920827625190311e-06, + "loss": 0.8044, + "step": 2117 + }, + { + "epoch": 0.5177218284038132, + "grad_norm": 2.408571481704712, + "learning_rate": 9.912910578779275e-06, + "loss": 1.9138, + "step": 2118 + }, + { + "epoch": 0.5179662674162796, + "grad_norm": 3.5460824966430664, + "learning_rate": 9.904993586959333e-06, + "loss": 0.858, + "step": 2119 + }, + { + "epoch": 0.518210706428746, + "grad_norm": 2.603878974914551, + "learning_rate": 9.89707665469317e-06, + "loss": 1.9577, + "step": 2120 + }, + { + "epoch": 0.5184551454412124, + "grad_norm": 4.377145290374756, + "learning_rate": 9.889159786943428e-06, + "loss": 0.7671, + "step": 2121 + }, + { + "epoch": 0.5186995844536788, + "grad_norm": 3.42803692817688, + "learning_rate": 9.881242988672719e-06, + "loss": 0.7304, + "step": 2122 + }, + { + "epoch": 0.5189440234661452, + "grad_norm": 2.933485269546509, + "learning_rate": 9.873326264843601e-06, + "loss": 0.7568, + "step": 2123 + }, + { + "epoch": 0.5191884624786116, + "grad_norm": 3.03100323677063, + "learning_rate": 9.865409620418593e-06, + "loss": 0.6751, + "step": 2124 + }, + { + "epoch": 0.519432901491078, + "grad_norm": 3.503622055053711, + "learning_rate": 9.857493060360157e-06, + "loss": 0.8314, + "step": 2125 + }, + { + "epoch": 0.5196773405035444, + "grad_norm": 3.3902294635772705, + "learning_rate": 9.84957658963071e-06, + "loss": 0.7963, + "step": 2126 + }, + { + "epoch": 0.5199217795160107, + "grad_norm": 3.721219062805176, + "learning_rate": 9.841660213192606e-06, + "loss": 0.7359, + "step": 2127 + }, + { + "epoch": 0.5201662185284771, + "grad_norm": 3.8676271438598633, + "learning_rate": 9.833743936008144e-06, + "loss": 0.7679, + "step": 2128 + }, + { + "epoch": 0.5204106575409435, + "grad_norm": 2.977606773376465, + "learning_rate": 9.825827763039558e-06, + "loss": 0.7072, + "step": 2129 + }, + { + "epoch": 0.5206550965534099, + "grad_norm": 4.037961006164551, + "learning_rate": 9.817911699249022e-06, + "loss": 0.7505, + "step": 2130 + }, + { + "epoch": 0.5208995355658763, + "grad_norm": 3.039395332336426, + "learning_rate": 9.809995749598633e-06, + "loss": 0.6771, + "step": 2131 + }, + { + "epoch": 0.5211439745783427, + "grad_norm": 4.734142303466797, + "learning_rate": 9.802079919050426e-06, + "loss": 0.6907, + "step": 2132 + }, + { + "epoch": 0.5213884135908091, + "grad_norm": 3.565530300140381, + "learning_rate": 9.794164212566355e-06, + "loss": 0.5967, + "step": 2133 + }, + { + "epoch": 0.5216328526032755, + "grad_norm": 4.300550937652588, + "learning_rate": 9.7862486351083e-06, + "loss": 0.7463, + "step": 2134 + }, + { + "epoch": 0.5218772916157419, + "grad_norm": 4.79160737991333, + "learning_rate": 9.778333191638055e-06, + "loss": 0.7749, + "step": 2135 + }, + { + "epoch": 0.5221217306282082, + "grad_norm": 5.3668951988220215, + "learning_rate": 9.770417887117333e-06, + "loss": 0.7261, + "step": 2136 + }, + { + "epoch": 0.5223661696406746, + "grad_norm": 2.6563315391540527, + "learning_rate": 9.762502726507764e-06, + "loss": 2.0795, + "step": 2137 + }, + { + "epoch": 0.522610608653141, + "grad_norm": 3.5295982360839844, + "learning_rate": 9.75458771477088e-06, + "loss": 0.6785, + "step": 2138 + }, + { + "epoch": 0.5228550476656074, + "grad_norm": 3.7802164554595947, + "learning_rate": 9.746672856868124e-06, + "loss": 0.9156, + "step": 2139 + }, + { + "epoch": 0.5230994866780738, + "grad_norm": 3.2856857776641846, + "learning_rate": 9.738758157760841e-06, + "loss": 0.791, + "step": 2140 + }, + { + "epoch": 0.5233439256905402, + "grad_norm": 3.8294010162353516, + "learning_rate": 9.730843622410282e-06, + "loss": 0.7145, + "step": 2141 + }, + { + "epoch": 0.5235883647030066, + "grad_norm": 3.379720687866211, + "learning_rate": 9.722929255777583e-06, + "loss": 0.7279, + "step": 2142 + }, + { + "epoch": 0.523832803715473, + "grad_norm": 4.032857894897461, + "learning_rate": 9.715015062823789e-06, + "loss": 0.7665, + "step": 2143 + }, + { + "epoch": 0.5240772427279394, + "grad_norm": 3.1897528171539307, + "learning_rate": 9.707101048509828e-06, + "loss": 0.8118, + "step": 2144 + }, + { + "epoch": 0.5243216817404057, + "grad_norm": 2.891777992248535, + "learning_rate": 9.699187217796518e-06, + "loss": 0.7128, + "step": 2145 + }, + { + "epoch": 0.5245661207528721, + "grad_norm": 2.616403818130493, + "learning_rate": 9.691273575644554e-06, + "loss": 0.702, + "step": 2146 + }, + { + "epoch": 0.5248105597653385, + "grad_norm": 3.5809245109558105, + "learning_rate": 9.68336012701453e-06, + "loss": 0.833, + "step": 2147 + }, + { + "epoch": 0.5250549987778049, + "grad_norm": 4.2926025390625, + "learning_rate": 9.675446876866903e-06, + "loss": 0.6003, + "step": 2148 + }, + { + "epoch": 0.5252994377902713, + "grad_norm": 3.5323801040649414, + "learning_rate": 9.667533830162013e-06, + "loss": 0.6544, + "step": 2149 + }, + { + "epoch": 0.5255438768027377, + "grad_norm": 2.839102029800415, + "learning_rate": 9.65962099186007e-06, + "loss": 0.753, + "step": 2150 + }, + { + "epoch": 0.5257883158152041, + "grad_norm": 4.1956706047058105, + "learning_rate": 9.651708366921152e-06, + "loss": 0.909, + "step": 2151 + }, + { + "epoch": 0.5260327548276705, + "grad_norm": 3.9112610816955566, + "learning_rate": 9.643795960305207e-06, + "loss": 0.7354, + "step": 2152 + }, + { + "epoch": 0.5262771938401369, + "grad_norm": 3.4937121868133545, + "learning_rate": 9.635883776972046e-06, + "loss": 0.7002, + "step": 2153 + }, + { + "epoch": 0.5265216328526032, + "grad_norm": 3.4174258708953857, + "learning_rate": 9.62797182188134e-06, + "loss": 0.679, + "step": 2154 + }, + { + "epoch": 0.5267660718650696, + "grad_norm": 3.6268627643585205, + "learning_rate": 9.620060099992609e-06, + "loss": 0.5634, + "step": 2155 + }, + { + "epoch": 0.527010510877536, + "grad_norm": 3.2625319957733154, + "learning_rate": 9.612148616265238e-06, + "loss": 0.7009, + "step": 2156 + }, + { + "epoch": 0.5272549498900024, + "grad_norm": 3.6257452964782715, + "learning_rate": 9.604237375658458e-06, + "loss": 0.8636, + "step": 2157 + }, + { + "epoch": 0.5274993889024688, + "grad_norm": 3.004514694213867, + "learning_rate": 9.596326383131346e-06, + "loss": 0.6685, + "step": 2158 + }, + { + "epoch": 0.5277438279149352, + "grad_norm": 3.018902063369751, + "learning_rate": 9.588415643642827e-06, + "loss": 0.7041, + "step": 2159 + }, + { + "epoch": 0.5279882669274016, + "grad_norm": 4.501762390136719, + "learning_rate": 9.580505162151662e-06, + "loss": 0.8137, + "step": 2160 + }, + { + "epoch": 0.528232705939868, + "grad_norm": 3.087576150894165, + "learning_rate": 9.572594943616457e-06, + "loss": 0.7622, + "step": 2161 + }, + { + "epoch": 0.5284771449523344, + "grad_norm": 4.186478614807129, + "learning_rate": 9.564684992995645e-06, + "loss": 0.7919, + "step": 2162 + }, + { + "epoch": 0.5287215839648007, + "grad_norm": 4.6504693031311035, + "learning_rate": 9.556775315247502e-06, + "loss": 0.7596, + "step": 2163 + }, + { + "epoch": 0.5289660229772671, + "grad_norm": 3.2655417919158936, + "learning_rate": 9.548865915330124e-06, + "loss": 0.618, + "step": 2164 + }, + { + "epoch": 0.5292104619897335, + "grad_norm": 3.9398417472839355, + "learning_rate": 9.540956798201439e-06, + "loss": 0.6942, + "step": 2165 + }, + { + "epoch": 0.5294549010021999, + "grad_norm": 3.884324550628662, + "learning_rate": 9.53304796881919e-06, + "loss": 0.7633, + "step": 2166 + }, + { + "epoch": 0.5296993400146663, + "grad_norm": 3.5051238536834717, + "learning_rate": 9.525139432140946e-06, + "loss": 0.6823, + "step": 2167 + }, + { + "epoch": 0.5299437790271327, + "grad_norm": 2.708531379699707, + "learning_rate": 9.51723119312409e-06, + "loss": 2.0782, + "step": 2168 + }, + { + "epoch": 0.5301882180395991, + "grad_norm": 3.4217796325683594, + "learning_rate": 9.50932325672582e-06, + "loss": 0.7089, + "step": 2169 + }, + { + "epoch": 0.5304326570520655, + "grad_norm": 3.7905852794647217, + "learning_rate": 9.501415627903143e-06, + "loss": 0.8027, + "step": 2170 + }, + { + "epoch": 0.530677096064532, + "grad_norm": 3.344539165496826, + "learning_rate": 9.493508311612874e-06, + "loss": 0.6425, + "step": 2171 + }, + { + "epoch": 0.5309215350769982, + "grad_norm": 3.495182752609253, + "learning_rate": 9.485601312811631e-06, + "loss": 0.7988, + "step": 2172 + }, + { + "epoch": 0.5311659740894646, + "grad_norm": 3.6051738262176514, + "learning_rate": 9.477694636455833e-06, + "loss": 0.7847, + "step": 2173 + }, + { + "epoch": 0.531410413101931, + "grad_norm": 3.1293435096740723, + "learning_rate": 9.4697882875017e-06, + "loss": 0.7358, + "step": 2174 + }, + { + "epoch": 0.5316548521143974, + "grad_norm": 3.203247308731079, + "learning_rate": 9.46188227090524e-06, + "loss": 0.75, + "step": 2175 + }, + { + "epoch": 0.5318992911268638, + "grad_norm": 3.34112548828125, + "learning_rate": 9.45397659162226e-06, + "loss": 0.6814, + "step": 2176 + }, + { + "epoch": 0.5321437301393303, + "grad_norm": 3.279837131500244, + "learning_rate": 9.44607125460835e-06, + "loss": 0.6884, + "step": 2177 + }, + { + "epoch": 0.5323881691517967, + "grad_norm": 4.138041019439697, + "learning_rate": 9.438166264818892e-06, + "loss": 0.797, + "step": 2178 + }, + { + "epoch": 0.532632608164263, + "grad_norm": 2.98604416847229, + "learning_rate": 9.430261627209044e-06, + "loss": 0.6787, + "step": 2179 + }, + { + "epoch": 0.5328770471767295, + "grad_norm": 3.883176565170288, + "learning_rate": 9.422357346733747e-06, + "loss": 0.8277, + "step": 2180 + }, + { + "epoch": 0.5331214861891957, + "grad_norm": 4.455817699432373, + "learning_rate": 9.414453428347715e-06, + "loss": 0.8231, + "step": 2181 + }, + { + "epoch": 0.5333659252016621, + "grad_norm": 3.1506121158599854, + "learning_rate": 9.406549877005438e-06, + "loss": 0.6412, + "step": 2182 + }, + { + "epoch": 0.5336103642141286, + "grad_norm": 3.284975528717041, + "learning_rate": 9.398646697661173e-06, + "loss": 0.7799, + "step": 2183 + }, + { + "epoch": 0.533854803226595, + "grad_norm": 3.78600811958313, + "learning_rate": 9.390743895268945e-06, + "loss": 0.6762, + "step": 2184 + }, + { + "epoch": 0.5340992422390614, + "grad_norm": 3.901350975036621, + "learning_rate": 9.382841474782546e-06, + "loss": 0.7135, + "step": 2185 + }, + { + "epoch": 0.5343436812515278, + "grad_norm": 4.45625114440918, + "learning_rate": 9.374939441155525e-06, + "loss": 0.8253, + "step": 2186 + }, + { + "epoch": 0.5345881202639942, + "grad_norm": 3.3413562774658203, + "learning_rate": 9.367037799341187e-06, + "loss": 0.7585, + "step": 2187 + }, + { + "epoch": 0.5348325592764606, + "grad_norm": 4.1224517822265625, + "learning_rate": 9.359136554292596e-06, + "loss": 0.8461, + "step": 2188 + }, + { + "epoch": 0.535076998288927, + "grad_norm": 3.8622636795043945, + "learning_rate": 9.351235710962565e-06, + "loss": 0.7622, + "step": 2189 + }, + { + "epoch": 0.5353214373013933, + "grad_norm": 3.3691866397857666, + "learning_rate": 9.343335274303651e-06, + "loss": 0.7139, + "step": 2190 + }, + { + "epoch": 0.5355658763138597, + "grad_norm": 3.879357099533081, + "learning_rate": 9.335435249268165e-06, + "loss": 0.6645, + "step": 2191 + }, + { + "epoch": 0.5358103153263261, + "grad_norm": 3.245859146118164, + "learning_rate": 9.327535640808154e-06, + "loss": 0.8594, + "step": 2192 + }, + { + "epoch": 0.5360547543387925, + "grad_norm": 3.7075939178466797, + "learning_rate": 9.3196364538754e-06, + "loss": 0.8228, + "step": 2193 + }, + { + "epoch": 0.5362991933512589, + "grad_norm": 3.3815085887908936, + "learning_rate": 9.311737693421436e-06, + "loss": 0.7622, + "step": 2194 + }, + { + "epoch": 0.5365436323637253, + "grad_norm": 3.013307571411133, + "learning_rate": 9.303839364397512e-06, + "loss": 0.6551, + "step": 2195 + }, + { + "epoch": 0.5367880713761917, + "grad_norm": 3.158292770385742, + "learning_rate": 9.295941471754611e-06, + "loss": 0.6397, + "step": 2196 + }, + { + "epoch": 0.5370325103886581, + "grad_norm": 3.4648733139038086, + "learning_rate": 9.288044020443449e-06, + "loss": 0.6763, + "step": 2197 + }, + { + "epoch": 0.5372769494011245, + "grad_norm": 4.360973834991455, + "learning_rate": 9.280147015414458e-06, + "loss": 0.7493, + "step": 2198 + }, + { + "epoch": 0.5375213884135908, + "grad_norm": 3.1984996795654297, + "learning_rate": 9.272250461617794e-06, + "loss": 0.7879, + "step": 2199 + }, + { + "epoch": 0.5377658274260572, + "grad_norm": 4.430309295654297, + "learning_rate": 9.264354364003327e-06, + "loss": 0.6783, + "step": 2200 + }, + { + "epoch": 0.5380102664385236, + "grad_norm": 3.2827811241149902, + "learning_rate": 9.256458727520648e-06, + "loss": 0.6597, + "step": 2201 + }, + { + "epoch": 0.53825470545099, + "grad_norm": 3.689610719680786, + "learning_rate": 9.24856355711905e-06, + "loss": 0.7053, + "step": 2202 + }, + { + "epoch": 0.5384991444634564, + "grad_norm": 3.9610626697540283, + "learning_rate": 9.24066885774754e-06, + "loss": 0.5845, + "step": 2203 + }, + { + "epoch": 0.5387435834759228, + "grad_norm": 3.752939462661743, + "learning_rate": 9.232774634354827e-06, + "loss": 0.8217, + "step": 2204 + }, + { + "epoch": 0.5389880224883892, + "grad_norm": 3.2424635887145996, + "learning_rate": 9.224880891889324e-06, + "loss": 0.7209, + "step": 2205 + }, + { + "epoch": 0.5392324615008556, + "grad_norm": 3.151031017303467, + "learning_rate": 9.21698763529914e-06, + "loss": 0.6904, + "step": 2206 + }, + { + "epoch": 0.539476900513322, + "grad_norm": 3.540400981903076, + "learning_rate": 9.209094869532082e-06, + "loss": 0.5895, + "step": 2207 + }, + { + "epoch": 0.5397213395257883, + "grad_norm": 2.9011690616607666, + "learning_rate": 9.201202599535639e-06, + "loss": 0.7095, + "step": 2208 + }, + { + "epoch": 0.5399657785382547, + "grad_norm": 3.394333600997925, + "learning_rate": 9.193310830257012e-06, + "loss": 0.7982, + "step": 2209 + }, + { + "epoch": 0.5402102175507211, + "grad_norm": 2.996051788330078, + "learning_rate": 9.185419566643068e-06, + "loss": 0.6493, + "step": 2210 + }, + { + "epoch": 0.5404546565631875, + "grad_norm": 3.2841668128967285, + "learning_rate": 9.177528813640362e-06, + "loss": 0.6623, + "step": 2211 + }, + { + "epoch": 0.5406990955756539, + "grad_norm": 3.1686737537384033, + "learning_rate": 9.16963857619513e-06, + "loss": 0.6757, + "step": 2212 + }, + { + "epoch": 0.5409435345881203, + "grad_norm": 3.0455615520477295, + "learning_rate": 9.161748859253288e-06, + "loss": 0.7552, + "step": 2213 + }, + { + "epoch": 0.5411879736005867, + "grad_norm": 3.6713058948516846, + "learning_rate": 9.153859667760418e-06, + "loss": 0.6814, + "step": 2214 + }, + { + "epoch": 0.5414324126130531, + "grad_norm": 3.267914295196533, + "learning_rate": 9.145971006661783e-06, + "loss": 0.6431, + "step": 2215 + }, + { + "epoch": 0.5416768516255195, + "grad_norm": 3.4913010597229004, + "learning_rate": 9.138082880902302e-06, + "loss": 0.7237, + "step": 2216 + }, + { + "epoch": 0.5419212906379858, + "grad_norm": 3.234109878540039, + "learning_rate": 9.13019529542657e-06, + "loss": 0.6613, + "step": 2217 + }, + { + "epoch": 0.5421657296504522, + "grad_norm": 3.133761405944824, + "learning_rate": 9.12230825517883e-06, + "loss": 0.7178, + "step": 2218 + }, + { + "epoch": 0.5424101686629186, + "grad_norm": 4.280647277832031, + "learning_rate": 9.114421765103e-06, + "loss": 0.7539, + "step": 2219 + }, + { + "epoch": 0.542654607675385, + "grad_norm": 3.2164087295532227, + "learning_rate": 9.106535830142638e-06, + "loss": 0.6519, + "step": 2220 + }, + { + "epoch": 0.5428990466878514, + "grad_norm": 3.8115792274475098, + "learning_rate": 9.098650455240959e-06, + "loss": 0.9071, + "step": 2221 + }, + { + "epoch": 0.5431434857003178, + "grad_norm": 3.463438034057617, + "learning_rate": 9.090765645340832e-06, + "loss": 0.768, + "step": 2222 + }, + { + "epoch": 0.5433879247127842, + "grad_norm": 3.009864330291748, + "learning_rate": 9.082881405384767e-06, + "loss": 0.7107, + "step": 2223 + }, + { + "epoch": 0.5436323637252506, + "grad_norm": 3.3293442726135254, + "learning_rate": 9.074997740314913e-06, + "loss": 0.7244, + "step": 2224 + }, + { + "epoch": 0.543876802737717, + "grad_norm": 3.3943569660186768, + "learning_rate": 9.067114655073069e-06, + "loss": 0.6464, + "step": 2225 + }, + { + "epoch": 0.5441212417501833, + "grad_norm": 3.474684000015259, + "learning_rate": 9.059232154600662e-06, + "loss": 0.7722, + "step": 2226 + }, + { + "epoch": 0.5443656807626497, + "grad_norm": 2.908451795578003, + "learning_rate": 9.051350243838757e-06, + "loss": 1.9873, + "step": 2227 + }, + { + "epoch": 0.5446101197751161, + "grad_norm": 3.428748846054077, + "learning_rate": 9.043468927728045e-06, + "loss": 0.6635, + "step": 2228 + }, + { + "epoch": 0.5448545587875825, + "grad_norm": 3.169649124145508, + "learning_rate": 9.035588211208847e-06, + "loss": 0.6984, + "step": 2229 + }, + { + "epoch": 0.5450989978000489, + "grad_norm": 2.6431148052215576, + "learning_rate": 9.02770809922111e-06, + "loss": 0.6158, + "step": 2230 + }, + { + "epoch": 0.5453434368125153, + "grad_norm": 3.05291748046875, + "learning_rate": 9.019828596704394e-06, + "loss": 0.6906, + "step": 2231 + }, + { + "epoch": 0.5455878758249817, + "grad_norm": 3.4386649131774902, + "learning_rate": 9.011949708597892e-06, + "loss": 0.607, + "step": 2232 + }, + { + "epoch": 0.5458323148374481, + "grad_norm": 3.2652945518493652, + "learning_rate": 9.004071439840396e-06, + "loss": 0.7593, + "step": 2233 + }, + { + "epoch": 0.5460767538499145, + "grad_norm": 3.2680203914642334, + "learning_rate": 8.996193795370319e-06, + "loss": 0.7699, + "step": 2234 + }, + { + "epoch": 0.5463211928623808, + "grad_norm": 3.967542886734009, + "learning_rate": 8.98831678012568e-06, + "loss": 0.7487, + "step": 2235 + }, + { + "epoch": 0.5465656318748472, + "grad_norm": 2.661180257797241, + "learning_rate": 8.980440399044103e-06, + "loss": 1.9339, + "step": 2236 + }, + { + "epoch": 0.5468100708873136, + "grad_norm": 2.546757698059082, + "learning_rate": 8.972564657062816e-06, + "loss": 2.0937, + "step": 2237 + }, + { + "epoch": 0.54705450989978, + "grad_norm": 3.3631649017333984, + "learning_rate": 8.964689559118646e-06, + "loss": 0.7624, + "step": 2238 + }, + { + "epoch": 0.5472989489122464, + "grad_norm": 2.922140121459961, + "learning_rate": 8.956815110148012e-06, + "loss": 0.6668, + "step": 2239 + }, + { + "epoch": 0.5475433879247128, + "grad_norm": 2.4277820587158203, + "learning_rate": 8.948941315086936e-06, + "loss": 1.9969, + "step": 2240 + }, + { + "epoch": 0.5477878269371792, + "grad_norm": 3.292299747467041, + "learning_rate": 8.941068178871021e-06, + "loss": 0.6945, + "step": 2241 + }, + { + "epoch": 0.5480322659496456, + "grad_norm": 3.2512614727020264, + "learning_rate": 8.933195706435459e-06, + "loss": 0.6505, + "step": 2242 + }, + { + "epoch": 0.548276704962112, + "grad_norm": 3.8018834590911865, + "learning_rate": 8.92532390271503e-06, + "loss": 0.7485, + "step": 2243 + }, + { + "epoch": 0.5485211439745783, + "grad_norm": 3.6477644443511963, + "learning_rate": 8.917452772644092e-06, + "loss": 0.6455, + "step": 2244 + }, + { + "epoch": 0.5487655829870447, + "grad_norm": 3.821967124938965, + "learning_rate": 8.90958232115658e-06, + "loss": 0.6256, + "step": 2245 + }, + { + "epoch": 0.5490100219995111, + "grad_norm": 3.82509708404541, + "learning_rate": 8.901712553186002e-06, + "loss": 0.8019, + "step": 2246 + }, + { + "epoch": 0.5492544610119775, + "grad_norm": 3.373044729232788, + "learning_rate": 8.893843473665443e-06, + "loss": 0.6713, + "step": 2247 + }, + { + "epoch": 0.5494989000244439, + "grad_norm": 3.1013381481170654, + "learning_rate": 8.885975087527553e-06, + "loss": 0.7236, + "step": 2248 + }, + { + "epoch": 0.5497433390369103, + "grad_norm": 3.8983516693115234, + "learning_rate": 8.878107399704546e-06, + "loss": 0.7146, + "step": 2249 + }, + { + "epoch": 0.5499877780493767, + "grad_norm": 3.556209087371826, + "learning_rate": 8.870240415128201e-06, + "loss": 0.6534, + "step": 2250 + }, + { + "epoch": 0.5502322170618431, + "grad_norm": 3.2743749618530273, + "learning_rate": 8.862374138729854e-06, + "loss": 0.7089, + "step": 2251 + }, + { + "epoch": 0.5504766560743095, + "grad_norm": 3.252896308898926, + "learning_rate": 8.8545085754404e-06, + "loss": 0.7129, + "step": 2252 + }, + { + "epoch": 0.5507210950867758, + "grad_norm": 3.328610897064209, + "learning_rate": 8.846643730190284e-06, + "loss": 0.6465, + "step": 2253 + }, + { + "epoch": 0.5509655340992422, + "grad_norm": 3.4283292293548584, + "learning_rate": 8.838779607909502e-06, + "loss": 0.6214, + "step": 2254 + }, + { + "epoch": 0.5512099731117086, + "grad_norm": 2.920253038406372, + "learning_rate": 8.830916213527593e-06, + "loss": 0.6444, + "step": 2255 + }, + { + "epoch": 0.551454412124175, + "grad_norm": 3.2998158931732178, + "learning_rate": 8.823053551973653e-06, + "loss": 0.6866, + "step": 2256 + }, + { + "epoch": 0.5516988511366414, + "grad_norm": 3.3970022201538086, + "learning_rate": 8.815191628176304e-06, + "loss": 0.6624, + "step": 2257 + }, + { + "epoch": 0.5519432901491078, + "grad_norm": 3.831857919692993, + "learning_rate": 8.807330447063712e-06, + "loss": 0.7256, + "step": 2258 + }, + { + "epoch": 0.5521877291615742, + "grad_norm": 4.044900417327881, + "learning_rate": 8.799470013563573e-06, + "loss": 0.639, + "step": 2259 + }, + { + "epoch": 0.5524321681740406, + "grad_norm": 3.4633517265319824, + "learning_rate": 8.791610332603123e-06, + "loss": 0.5559, + "step": 2260 + }, + { + "epoch": 0.552676607186507, + "grad_norm": 4.289843559265137, + "learning_rate": 8.783751409109116e-06, + "loss": 0.7395, + "step": 2261 + }, + { + "epoch": 0.5529210461989733, + "grad_norm": 3.948843002319336, + "learning_rate": 8.77589324800784e-06, + "loss": 0.8218, + "step": 2262 + }, + { + "epoch": 0.5531654852114397, + "grad_norm": 4.675017356872559, + "learning_rate": 8.768035854225098e-06, + "loss": 0.8025, + "step": 2263 + }, + { + "epoch": 0.5534099242239061, + "grad_norm": 3.471562385559082, + "learning_rate": 8.760179232686217e-06, + "loss": 0.7199, + "step": 2264 + }, + { + "epoch": 0.5536543632363725, + "grad_norm": 3.258531093597412, + "learning_rate": 8.752323388316038e-06, + "loss": 0.725, + "step": 2265 + }, + { + "epoch": 0.5538988022488389, + "grad_norm": 3.5345466136932373, + "learning_rate": 8.744468326038914e-06, + "loss": 0.6563, + "step": 2266 + }, + { + "epoch": 0.5541432412613053, + "grad_norm": 3.502413272857666, + "learning_rate": 8.73661405077871e-06, + "loss": 0.5881, + "step": 2267 + }, + { + "epoch": 0.5543876802737717, + "grad_norm": 5.4487810134887695, + "learning_rate": 8.728760567458797e-06, + "loss": 0.7555, + "step": 2268 + }, + { + "epoch": 0.5546321192862381, + "grad_norm": 3.4690463542938232, + "learning_rate": 8.720907881002047e-06, + "loss": 0.8153, + "step": 2269 + }, + { + "epoch": 0.5548765582987045, + "grad_norm": 3.3505942821502686, + "learning_rate": 8.713055996330832e-06, + "loss": 0.6743, + "step": 2270 + }, + { + "epoch": 0.5551209973111708, + "grad_norm": 3.491990089416504, + "learning_rate": 8.705204918367032e-06, + "loss": 0.6761, + "step": 2271 + }, + { + "epoch": 0.5553654363236372, + "grad_norm": 4.659776210784912, + "learning_rate": 8.697354652032009e-06, + "loss": 0.6884, + "step": 2272 + }, + { + "epoch": 0.5556098753361036, + "grad_norm": 3.3480889797210693, + "learning_rate": 8.689505202246618e-06, + "loss": 0.6282, + "step": 2273 + }, + { + "epoch": 0.55585431434857, + "grad_norm": 3.3003580570220947, + "learning_rate": 8.681656573931209e-06, + "loss": 0.6511, + "step": 2274 + }, + { + "epoch": 0.5560987533610364, + "grad_norm": 2.958965539932251, + "learning_rate": 8.67380877200561e-06, + "loss": 0.5956, + "step": 2275 + }, + { + "epoch": 0.5563431923735028, + "grad_norm": 3.247917890548706, + "learning_rate": 8.665961801389133e-06, + "loss": 0.6243, + "step": 2276 + }, + { + "epoch": 0.5565876313859692, + "grad_norm": 3.247648000717163, + "learning_rate": 8.658115667000572e-06, + "loss": 0.7174, + "step": 2277 + }, + { + "epoch": 0.5568320703984356, + "grad_norm": 3.5899839401245117, + "learning_rate": 8.650270373758194e-06, + "loss": 0.7944, + "step": 2278 + }, + { + "epoch": 0.557076509410902, + "grad_norm": 3.4315006732940674, + "learning_rate": 8.642425926579738e-06, + "loss": 0.6849, + "step": 2279 + }, + { + "epoch": 0.5573209484233683, + "grad_norm": 4.766323089599609, + "learning_rate": 8.634582330382414e-06, + "loss": 0.8128, + "step": 2280 + }, + { + "epoch": 0.5575653874358347, + "grad_norm": 3.2827932834625244, + "learning_rate": 8.626739590082897e-06, + "loss": 0.6284, + "step": 2281 + }, + { + "epoch": 0.5578098264483011, + "grad_norm": 3.4646546840667725, + "learning_rate": 8.61889771059733e-06, + "loss": 0.578, + "step": 2282 + }, + { + "epoch": 0.5580542654607675, + "grad_norm": 3.7662670612335205, + "learning_rate": 8.611056696841313e-06, + "loss": 0.7601, + "step": 2283 + }, + { + "epoch": 0.5582987044732339, + "grad_norm": 2.9055752754211426, + "learning_rate": 8.6032165537299e-06, + "loss": 0.7847, + "step": 2284 + }, + { + "epoch": 0.5585431434857003, + "grad_norm": 4.044349670410156, + "learning_rate": 8.595377286177606e-06, + "loss": 0.7565, + "step": 2285 + }, + { + "epoch": 0.5587875824981667, + "grad_norm": 2.8701465129852295, + "learning_rate": 8.587538899098388e-06, + "loss": 1.9964, + "step": 2286 + }, + { + "epoch": 0.5590320215106331, + "grad_norm": 3.0263373851776123, + "learning_rate": 8.579701397405668e-06, + "loss": 0.635, + "step": 2287 + }, + { + "epoch": 0.5592764605230995, + "grad_norm": 2.837200403213501, + "learning_rate": 8.571864786012296e-06, + "loss": 2.005, + "step": 2288 + }, + { + "epoch": 0.5595208995355658, + "grad_norm": 5.424970626831055, + "learning_rate": 8.564029069830569e-06, + "loss": 0.6735, + "step": 2289 + }, + { + "epoch": 0.5597653385480322, + "grad_norm": 2.9481616020202637, + "learning_rate": 8.556194253772226e-06, + "loss": 0.5457, + "step": 2290 + }, + { + "epoch": 0.5600097775604986, + "grad_norm": 4.5754241943359375, + "learning_rate": 8.54836034274844e-06, + "loss": 0.7518, + "step": 2291 + }, + { + "epoch": 0.560254216572965, + "grad_norm": 3.5060579776763916, + "learning_rate": 8.540527341669814e-06, + "loss": 0.6221, + "step": 2292 + }, + { + "epoch": 0.5604986555854314, + "grad_norm": 3.9225127696990967, + "learning_rate": 8.532695255446384e-06, + "loss": 0.8294, + "step": 2293 + }, + { + "epoch": 0.5607430945978978, + "grad_norm": 4.013699054718018, + "learning_rate": 8.524864088987613e-06, + "loss": 0.6485, + "step": 2294 + }, + { + "epoch": 0.5609875336103642, + "grad_norm": 3.0791900157928467, + "learning_rate": 8.517033847202383e-06, + "loss": 0.6977, + "step": 2295 + }, + { + "epoch": 0.5612319726228306, + "grad_norm": 3.0367400646209717, + "learning_rate": 8.509204534999002e-06, + "loss": 0.6768, + "step": 2296 + }, + { + "epoch": 0.561476411635297, + "grad_norm": 3.0239598751068115, + "learning_rate": 8.501376157285191e-06, + "loss": 0.6619, + "step": 2297 + }, + { + "epoch": 0.5617208506477633, + "grad_norm": 4.120974063873291, + "learning_rate": 8.493548718968085e-06, + "loss": 0.6598, + "step": 2298 + }, + { + "epoch": 0.5619652896602297, + "grad_norm": 4.155196666717529, + "learning_rate": 8.485722224954237e-06, + "loss": 0.6761, + "step": 2299 + }, + { + "epoch": 0.5622097286726961, + "grad_norm": 3.427410364151001, + "learning_rate": 8.477896680149597e-06, + "loss": 0.5893, + "step": 2300 + }, + { + "epoch": 0.5624541676851625, + "grad_norm": 2.950836658477783, + "learning_rate": 8.47007208945953e-06, + "loss": 0.7799, + "step": 2301 + }, + { + "epoch": 0.562698606697629, + "grad_norm": 3.4842567443847656, + "learning_rate": 8.4622484577888e-06, + "loss": 0.6295, + "step": 2302 + }, + { + "epoch": 0.5629430457100953, + "grad_norm": 3.7331862449645996, + "learning_rate": 8.454425790041567e-06, + "loss": 0.7276, + "step": 2303 + }, + { + "epoch": 0.5631874847225617, + "grad_norm": 3.904407501220703, + "learning_rate": 8.44660409112139e-06, + "loss": 0.7983, + "step": 2304 + }, + { + "epoch": 0.5634319237350282, + "grad_norm": 3.3740532398223877, + "learning_rate": 8.438783365931216e-06, + "loss": 0.7281, + "step": 2305 + }, + { + "epoch": 0.5636763627474946, + "grad_norm": 3.833692789077759, + "learning_rate": 8.430963619373388e-06, + "loss": 0.8742, + "step": 2306 + }, + { + "epoch": 0.5639208017599608, + "grad_norm": 3.350278377532959, + "learning_rate": 8.423144856349631e-06, + "loss": 0.6639, + "step": 2307 + }, + { + "epoch": 0.5641652407724272, + "grad_norm": 2.7863752841949463, + "learning_rate": 8.415327081761056e-06, + "loss": 0.4901, + "step": 2308 + }, + { + "epoch": 0.5644096797848936, + "grad_norm": 3.5297372341156006, + "learning_rate": 8.407510300508151e-06, + "loss": 0.6328, + "step": 2309 + }, + { + "epoch": 0.56465411879736, + "grad_norm": 3.285212278366089, + "learning_rate": 8.399694517490787e-06, + "loss": 0.7632, + "step": 2310 + }, + { + "epoch": 0.5648985578098265, + "grad_norm": 3.537651777267456, + "learning_rate": 8.391879737608202e-06, + "loss": 0.6741, + "step": 2311 + }, + { + "epoch": 0.5651429968222929, + "grad_norm": 3.2823824882507324, + "learning_rate": 8.384065965759014e-06, + "loss": 0.7633, + "step": 2312 + }, + { + "epoch": 0.5653874358347593, + "grad_norm": 3.2051095962524414, + "learning_rate": 8.3762532068412e-06, + "loss": 0.6253, + "step": 2313 + }, + { + "epoch": 0.5656318748472257, + "grad_norm": 3.437474489212036, + "learning_rate": 8.368441465752111e-06, + "loss": 0.7761, + "step": 2314 + }, + { + "epoch": 0.5658763138596921, + "grad_norm": 3.0659470558166504, + "learning_rate": 8.36063074738845e-06, + "loss": 0.6317, + "step": 2315 + }, + { + "epoch": 0.5661207528721584, + "grad_norm": 6.433424472808838, + "learning_rate": 8.352821056646289e-06, + "loss": 0.7217, + "step": 2316 + }, + { + "epoch": 0.5663651918846248, + "grad_norm": 3.3285651206970215, + "learning_rate": 8.345012398421046e-06, + "loss": 0.7531, + "step": 2317 + }, + { + "epoch": 0.5666096308970912, + "grad_norm": 3.584637403488159, + "learning_rate": 8.3372047776075e-06, + "loss": 0.7354, + "step": 2318 + }, + { + "epoch": 0.5668540699095576, + "grad_norm": 3.155935764312744, + "learning_rate": 8.329398199099778e-06, + "loss": 0.7217, + "step": 2319 + }, + { + "epoch": 0.567098508922024, + "grad_norm": 3.0519256591796875, + "learning_rate": 8.321592667791351e-06, + "loss": 0.6979, + "step": 2320 + }, + { + "epoch": 0.5673429479344904, + "grad_norm": 2.8886446952819824, + "learning_rate": 8.313788188575032e-06, + "loss": 0.7433, + "step": 2321 + }, + { + "epoch": 0.5675873869469568, + "grad_norm": 3.074049472808838, + "learning_rate": 8.30598476634298e-06, + "loss": 0.743, + "step": 2322 + }, + { + "epoch": 0.5678318259594232, + "grad_norm": 3.1361405849456787, + "learning_rate": 8.29818240598669e-06, + "loss": 0.7218, + "step": 2323 + }, + { + "epoch": 0.5680762649718896, + "grad_norm": 3.6520795822143555, + "learning_rate": 8.290381112396989e-06, + "loss": 0.8034, + "step": 2324 + }, + { + "epoch": 0.5683207039843559, + "grad_norm": 3.4125802516937256, + "learning_rate": 8.282580890464034e-06, + "loss": 0.8232, + "step": 2325 + }, + { + "epoch": 0.5685651429968223, + "grad_norm": 4.007071018218994, + "learning_rate": 8.274781745077316e-06, + "loss": 0.5776, + "step": 2326 + }, + { + "epoch": 0.5688095820092887, + "grad_norm": 3.6633801460266113, + "learning_rate": 8.266983681125647e-06, + "loss": 0.6866, + "step": 2327 + }, + { + "epoch": 0.5690540210217551, + "grad_norm": 2.8978915214538574, + "learning_rate": 8.25918670349716e-06, + "loss": 0.5959, + "step": 2328 + }, + { + "epoch": 0.5692984600342215, + "grad_norm": 3.001453399658203, + "learning_rate": 8.25139081707931e-06, + "loss": 0.694, + "step": 2329 + }, + { + "epoch": 0.5695428990466879, + "grad_norm": 2.951714038848877, + "learning_rate": 8.24359602675887e-06, + "loss": 0.7599, + "step": 2330 + }, + { + "epoch": 0.5697873380591543, + "grad_norm": 6.370380878448486, + "learning_rate": 8.23580233742192e-06, + "loss": 0.6479, + "step": 2331 + }, + { + "epoch": 0.5700317770716207, + "grad_norm": 3.31969952583313, + "learning_rate": 8.228009753953845e-06, + "loss": 0.7751, + "step": 2332 + }, + { + "epoch": 0.5702762160840871, + "grad_norm": 3.5400986671447754, + "learning_rate": 8.22021828123936e-06, + "loss": 0.6758, + "step": 2333 + }, + { + "epoch": 0.5705206550965534, + "grad_norm": 3.052528142929077, + "learning_rate": 8.21242792416246e-06, + "loss": 0.5639, + "step": 2334 + }, + { + "epoch": 0.5707650941090198, + "grad_norm": 3.4429566860198975, + "learning_rate": 8.20463868760645e-06, + "loss": 0.7557, + "step": 2335 + }, + { + "epoch": 0.5710095331214862, + "grad_norm": 3.9706220626831055, + "learning_rate": 8.196850576453933e-06, + "loss": 0.7045, + "step": 2336 + }, + { + "epoch": 0.5712539721339526, + "grad_norm": 3.4687039852142334, + "learning_rate": 8.189063595586804e-06, + "loss": 0.6398, + "step": 2337 + }, + { + "epoch": 0.571498411146419, + "grad_norm": 3.567106246948242, + "learning_rate": 8.181277749886248e-06, + "loss": 0.639, + "step": 2338 + }, + { + "epoch": 0.5717428501588854, + "grad_norm": 4.738640785217285, + "learning_rate": 8.173493044232745e-06, + "loss": 0.7687, + "step": 2339 + }, + { + "epoch": 0.5719872891713518, + "grad_norm": 3.808171510696411, + "learning_rate": 8.165709483506054e-06, + "loss": 0.7441, + "step": 2340 + }, + { + "epoch": 0.5722317281838182, + "grad_norm": 3.3038058280944824, + "learning_rate": 8.15792707258522e-06, + "loss": 0.6759, + "step": 2341 + }, + { + "epoch": 0.5724761671962846, + "grad_norm": 3.4772489070892334, + "learning_rate": 8.150145816348567e-06, + "loss": 0.6799, + "step": 2342 + }, + { + "epoch": 0.5727206062087509, + "grad_norm": 3.959152936935425, + "learning_rate": 8.142365719673693e-06, + "loss": 0.7665, + "step": 2343 + }, + { + "epoch": 0.5729650452212173, + "grad_norm": 4.0073137283325195, + "learning_rate": 8.13458678743747e-06, + "loss": 0.7992, + "step": 2344 + }, + { + "epoch": 0.5732094842336837, + "grad_norm": 4.013628005981445, + "learning_rate": 8.126809024516038e-06, + "loss": 0.6077, + "step": 2345 + }, + { + "epoch": 0.5734539232461501, + "grad_norm": 3.3716201782226562, + "learning_rate": 8.119032435784811e-06, + "loss": 0.7101, + "step": 2346 + }, + { + "epoch": 0.5736983622586165, + "grad_norm": 4.247351169586182, + "learning_rate": 8.111257026118461e-06, + "loss": 0.6254, + "step": 2347 + }, + { + "epoch": 0.5739428012710829, + "grad_norm": 3.6460940837860107, + "learning_rate": 8.103482800390919e-06, + "loss": 0.695, + "step": 2348 + }, + { + "epoch": 0.5741872402835493, + "grad_norm": 2.9894654750823975, + "learning_rate": 8.09570976347538e-06, + "loss": 0.6567, + "step": 2349 + }, + { + "epoch": 0.5744316792960157, + "grad_norm": 3.431727409362793, + "learning_rate": 8.087937920244296e-06, + "loss": 0.5698, + "step": 2350 + }, + { + "epoch": 0.5746761183084821, + "grad_norm": 3.010704755783081, + "learning_rate": 8.08016727556936e-06, + "loss": 0.7224, + "step": 2351 + }, + { + "epoch": 0.5749205573209484, + "grad_norm": 2.5566749572753906, + "learning_rate": 8.07239783432152e-06, + "loss": 0.5189, + "step": 2352 + }, + { + "epoch": 0.5751649963334148, + "grad_norm": 4.0545220375061035, + "learning_rate": 8.064629601370968e-06, + "loss": 0.6559, + "step": 2353 + }, + { + "epoch": 0.5754094353458812, + "grad_norm": 3.776262044906616, + "learning_rate": 8.056862581587143e-06, + "loss": 0.63, + "step": 2354 + }, + { + "epoch": 0.5756538743583476, + "grad_norm": 2.8388962745666504, + "learning_rate": 8.04909677983872e-06, + "loss": 0.7652, + "step": 2355 + }, + { + "epoch": 0.575898313370814, + "grad_norm": 3.862236976623535, + "learning_rate": 8.041332200993606e-06, + "loss": 0.6503, + "step": 2356 + }, + { + "epoch": 0.5761427523832804, + "grad_norm": 3.48097825050354, + "learning_rate": 8.03356884991895e-06, + "loss": 0.7618, + "step": 2357 + }, + { + "epoch": 0.5763871913957468, + "grad_norm": 2.9636313915252686, + "learning_rate": 8.02580673148113e-06, + "loss": 0.6997, + "step": 2358 + }, + { + "epoch": 0.5766316304082132, + "grad_norm": 7.425884246826172, + "learning_rate": 8.018045850545744e-06, + "loss": 0.6347, + "step": 2359 + }, + { + "epoch": 0.5768760694206796, + "grad_norm": 2.9083340167999268, + "learning_rate": 8.01028621197762e-06, + "loss": 0.533, + "step": 2360 + }, + { + "epoch": 0.5771205084331459, + "grad_norm": 3.161489725112915, + "learning_rate": 8.002527820640809e-06, + "loss": 0.7049, + "step": 2361 + }, + { + "epoch": 0.5773649474456123, + "grad_norm": 4.079695701599121, + "learning_rate": 7.994770681398576e-06, + "loss": 0.766, + "step": 2362 + }, + { + "epoch": 0.5776093864580787, + "grad_norm": 3.2374866008758545, + "learning_rate": 7.987014799113398e-06, + "loss": 0.7038, + "step": 2363 + }, + { + "epoch": 0.5778538254705451, + "grad_norm": 3.604962110519409, + "learning_rate": 7.97926017864698e-06, + "loss": 0.7046, + "step": 2364 + }, + { + "epoch": 0.5780982644830115, + "grad_norm": 4.064101219177246, + "learning_rate": 7.971506824860216e-06, + "loss": 0.5358, + "step": 2365 + }, + { + "epoch": 0.5783427034954779, + "grad_norm": 3.2200396060943604, + "learning_rate": 7.963754742613218e-06, + "loss": 0.7982, + "step": 2366 + }, + { + "epoch": 0.5785871425079443, + "grad_norm": 3.961225986480713, + "learning_rate": 7.956003936765298e-06, + "loss": 0.7388, + "step": 2367 + }, + { + "epoch": 0.5788315815204107, + "grad_norm": 3.4508423805236816, + "learning_rate": 7.948254412174966e-06, + "loss": 2.0356, + "step": 2368 + }, + { + "epoch": 0.5790760205328771, + "grad_norm": 4.107931613922119, + "learning_rate": 7.940506173699933e-06, + "loss": 0.6808, + "step": 2369 + }, + { + "epoch": 0.5793204595453434, + "grad_norm": 3.6741394996643066, + "learning_rate": 7.932759226197098e-06, + "loss": 0.6325, + "step": 2370 + }, + { + "epoch": 0.5795648985578098, + "grad_norm": 3.243452548980713, + "learning_rate": 7.925013574522556e-06, + "loss": 0.5824, + "step": 2371 + }, + { + "epoch": 0.5798093375702762, + "grad_norm": 3.2412142753601074, + "learning_rate": 7.91726922353159e-06, + "loss": 0.6886, + "step": 2372 + }, + { + "epoch": 0.5800537765827426, + "grad_norm": 3.364396333694458, + "learning_rate": 7.90952617807866e-06, + "loss": 0.6811, + "step": 2373 + }, + { + "epoch": 0.580298215595209, + "grad_norm": 3.435480833053589, + "learning_rate": 7.901784443017419e-06, + "loss": 0.6123, + "step": 2374 + }, + { + "epoch": 0.5805426546076754, + "grad_norm": 3.247061252593994, + "learning_rate": 7.894044023200691e-06, + "loss": 0.652, + "step": 2375 + }, + { + "epoch": 0.5807870936201418, + "grad_norm": 4.547329425811768, + "learning_rate": 7.886304923480476e-06, + "loss": 0.7178, + "step": 2376 + }, + { + "epoch": 0.5810315326326082, + "grad_norm": 2.973055601119995, + "learning_rate": 7.878567148707948e-06, + "loss": 0.761, + "step": 2377 + }, + { + "epoch": 0.5812759716450746, + "grad_norm": 4.504457473754883, + "learning_rate": 7.870830703733451e-06, + "loss": 0.6513, + "step": 2378 + }, + { + "epoch": 0.5815204106575409, + "grad_norm": 3.269958734512329, + "learning_rate": 7.86309559340649e-06, + "loss": 0.6604, + "step": 2379 + }, + { + "epoch": 0.5817648496700073, + "grad_norm": 2.929678201675415, + "learning_rate": 7.85536182257575e-06, + "loss": 0.6358, + "step": 2380 + }, + { + "epoch": 0.5820092886824737, + "grad_norm": 4.424942970275879, + "learning_rate": 7.847629396089054e-06, + "loss": 0.6125, + "step": 2381 + }, + { + "epoch": 0.5822537276949401, + "grad_norm": 3.2533557415008545, + "learning_rate": 7.839898318793398e-06, + "loss": 0.6891, + "step": 2382 + }, + { + "epoch": 0.5824981667074065, + "grad_norm": 5.905025959014893, + "learning_rate": 7.832168595534925e-06, + "loss": 0.6394, + "step": 2383 + }, + { + "epoch": 0.5827426057198729, + "grad_norm": 3.163419723510742, + "learning_rate": 7.82444023115893e-06, + "loss": 0.5843, + "step": 2384 + }, + { + "epoch": 0.5829870447323393, + "grad_norm": 4.014256954193115, + "learning_rate": 7.81671323050986e-06, + "loss": 0.5822, + "step": 2385 + }, + { + "epoch": 0.5832314837448057, + "grad_norm": 3.2915475368499756, + "learning_rate": 7.808987598431303e-06, + "loss": 0.6776, + "step": 2386 + }, + { + "epoch": 0.5834759227572721, + "grad_norm": 3.713676929473877, + "learning_rate": 7.801263339765992e-06, + "loss": 0.8284, + "step": 2387 + }, + { + "epoch": 0.5837203617697384, + "grad_norm": 3.4072582721710205, + "learning_rate": 7.793540459355798e-06, + "loss": 0.7618, + "step": 2388 + }, + { + "epoch": 0.5839648007822048, + "grad_norm": 2.850114583969116, + "learning_rate": 7.785818962041727e-06, + "loss": 0.7429, + "step": 2389 + }, + { + "epoch": 0.5842092397946712, + "grad_norm": 3.4230334758758545, + "learning_rate": 7.778098852663918e-06, + "loss": 0.6766, + "step": 2390 + }, + { + "epoch": 0.5844536788071376, + "grad_norm": 2.526958465576172, + "learning_rate": 7.770380136061643e-06, + "loss": 2.0334, + "step": 2391 + }, + { + "epoch": 0.584698117819604, + "grad_norm": 5.009162902832031, + "learning_rate": 7.762662817073297e-06, + "loss": 0.6014, + "step": 2392 + }, + { + "epoch": 0.5849425568320704, + "grad_norm": 3.332958221435547, + "learning_rate": 7.754946900536403e-06, + "loss": 0.6578, + "step": 2393 + }, + { + "epoch": 0.5851869958445368, + "grad_norm": 3.8514368534088135, + "learning_rate": 7.747232391287602e-06, + "loss": 0.7524, + "step": 2394 + }, + { + "epoch": 0.5854314348570032, + "grad_norm": 5.737468719482422, + "learning_rate": 7.739519294162652e-06, + "loss": 0.6417, + "step": 2395 + }, + { + "epoch": 0.5856758738694696, + "grad_norm": 3.6034488677978516, + "learning_rate": 7.73180761399643e-06, + "loss": 0.7302, + "step": 2396 + }, + { + "epoch": 0.5859203128819359, + "grad_norm": 3.2893946170806885, + "learning_rate": 7.72409735562292e-06, + "loss": 0.6442, + "step": 2397 + }, + { + "epoch": 0.5861647518944023, + "grad_norm": 5.023566246032715, + "learning_rate": 7.716388523875219e-06, + "loss": 0.7578, + "step": 2398 + }, + { + "epoch": 0.5864091909068687, + "grad_norm": 3.2663068771362305, + "learning_rate": 7.708681123585525e-06, + "loss": 0.6888, + "step": 2399 + }, + { + "epoch": 0.5866536299193351, + "grad_norm": 3.55702805519104, + "learning_rate": 7.700975159585142e-06, + "loss": 0.7193, + "step": 2400 + }, + { + "epoch": 0.5868980689318015, + "grad_norm": 3.774721622467041, + "learning_rate": 7.693270636704476e-06, + "loss": 0.7504, + "step": 2401 + }, + { + "epoch": 0.5871425079442679, + "grad_norm": 5.629326820373535, + "learning_rate": 7.685567559773024e-06, + "loss": 0.7078, + "step": 2402 + }, + { + "epoch": 0.5873869469567343, + "grad_norm": 3.1027450561523438, + "learning_rate": 7.67786593361938e-06, + "loss": 0.7157, + "step": 2403 + }, + { + "epoch": 0.5876313859692007, + "grad_norm": 3.616337537765503, + "learning_rate": 7.670165763071227e-06, + "loss": 0.6647, + "step": 2404 + }, + { + "epoch": 0.5878758249816671, + "grad_norm": 3.252295732498169, + "learning_rate": 7.662467052955337e-06, + "loss": 0.7219, + "step": 2405 + }, + { + "epoch": 0.5881202639941334, + "grad_norm": 3.382500171661377, + "learning_rate": 7.65476980809757e-06, + "loss": 0.5509, + "step": 2406 + }, + { + "epoch": 0.5883647030065998, + "grad_norm": 3.338959217071533, + "learning_rate": 7.647074033322856e-06, + "loss": 0.6593, + "step": 2407 + }, + { + "epoch": 0.5886091420190662, + "grad_norm": 2.3664376735687256, + "learning_rate": 7.639379733455219e-06, + "loss": 1.9463, + "step": 2408 + }, + { + "epoch": 0.5888535810315326, + "grad_norm": 2.9018778800964355, + "learning_rate": 7.631686913317743e-06, + "loss": 0.6245, + "step": 2409 + }, + { + "epoch": 0.589098020043999, + "grad_norm": 3.5737507343292236, + "learning_rate": 7.623995577732594e-06, + "loss": 0.6687, + "step": 2410 + }, + { + "epoch": 0.5893424590564654, + "grad_norm": 2.874079704284668, + "learning_rate": 7.616305731521009e-06, + "loss": 0.7341, + "step": 2411 + }, + { + "epoch": 0.5895868980689318, + "grad_norm": 4.047252655029297, + "learning_rate": 7.6086173795032866e-06, + "loss": 0.6604, + "step": 2412 + }, + { + "epoch": 0.5898313370813982, + "grad_norm": 3.585425615310669, + "learning_rate": 7.600930526498788e-06, + "loss": 0.7895, + "step": 2413 + }, + { + "epoch": 0.5900757760938646, + "grad_norm": 2.9749667644500732, + "learning_rate": 7.593245177325941e-06, + "loss": 0.5783, + "step": 2414 + }, + { + "epoch": 0.5903202151063309, + "grad_norm": 2.7515406608581543, + "learning_rate": 7.585561336802221e-06, + "loss": 0.6703, + "step": 2415 + }, + { + "epoch": 0.5905646541187973, + "grad_norm": 3.3146510124206543, + "learning_rate": 7.577879009744165e-06, + "loss": 0.6058, + "step": 2416 + }, + { + "epoch": 0.5908090931312637, + "grad_norm": 3.073467969894409, + "learning_rate": 7.570198200967363e-06, + "loss": 0.5668, + "step": 2417 + }, + { + "epoch": 0.5910535321437301, + "grad_norm": 3.7749462127685547, + "learning_rate": 7.562518915286443e-06, + "loss": 0.7409, + "step": 2418 + }, + { + "epoch": 0.5912979711561965, + "grad_norm": 3.177457571029663, + "learning_rate": 7.554841157515091e-06, + "loss": 0.7827, + "step": 2419 + }, + { + "epoch": 0.5915424101686629, + "grad_norm": 2.8870596885681152, + "learning_rate": 7.547164932466028e-06, + "loss": 0.7256, + "step": 2420 + }, + { + "epoch": 0.5917868491811293, + "grad_norm": 3.333930253982544, + "learning_rate": 7.539490244951013e-06, + "loss": 0.6556, + "step": 2421 + }, + { + "epoch": 0.5920312881935957, + "grad_norm": 6.634274482727051, + "learning_rate": 7.531817099780846e-06, + "loss": 0.69, + "step": 2422 + }, + { + "epoch": 0.5922757272060621, + "grad_norm": 3.813068151473999, + "learning_rate": 7.524145501765357e-06, + "loss": 0.6719, + "step": 2423 + }, + { + "epoch": 0.5925201662185284, + "grad_norm": 3.0771398544311523, + "learning_rate": 7.516475455713409e-06, + "loss": 0.6068, + "step": 2424 + }, + { + "epoch": 0.5927646052309948, + "grad_norm": 3.213043689727783, + "learning_rate": 7.508806966432882e-06, + "loss": 0.6483, + "step": 2425 + }, + { + "epoch": 0.5930090442434612, + "grad_norm": 4.564669132232666, + "learning_rate": 7.501140038730701e-06, + "loss": 0.5449, + "step": 2426 + }, + { + "epoch": 0.5932534832559276, + "grad_norm": 3.1842901706695557, + "learning_rate": 7.493474677412795e-06, + "loss": 0.6448, + "step": 2427 + }, + { + "epoch": 0.593497922268394, + "grad_norm": 2.83215594291687, + "learning_rate": 7.4858108872841115e-06, + "loss": 0.5718, + "step": 2428 + }, + { + "epoch": 0.5937423612808604, + "grad_norm": 3.178879499435425, + "learning_rate": 7.4781486731486215e-06, + "loss": 0.6779, + "step": 2429 + }, + { + "epoch": 0.5939868002933268, + "grad_norm": 3.813783884048462, + "learning_rate": 7.470488039809304e-06, + "loss": 0.7745, + "step": 2430 + }, + { + "epoch": 0.5942312393057932, + "grad_norm": 4.4895195960998535, + "learning_rate": 7.462828992068144e-06, + "loss": 0.6709, + "step": 2431 + }, + { + "epoch": 0.5944756783182596, + "grad_norm": 3.431575059890747, + "learning_rate": 7.455171534726138e-06, + "loss": 0.8067, + "step": 2432 + }, + { + "epoch": 0.5947201173307259, + "grad_norm": 3.2519891262054443, + "learning_rate": 7.447515672583282e-06, + "loss": 0.7389, + "step": 2433 + }, + { + "epoch": 0.5949645563431923, + "grad_norm": 3.4478063583374023, + "learning_rate": 7.4398614104385734e-06, + "loss": 0.7747, + "step": 2434 + }, + { + "epoch": 0.5952089953556587, + "grad_norm": 3.40944504737854, + "learning_rate": 7.432208753090009e-06, + "loss": 0.7563, + "step": 2435 + }, + { + "epoch": 0.5954534343681251, + "grad_norm": 3.2093870639801025, + "learning_rate": 7.4245577053345715e-06, + "loss": 0.7171, + "step": 2436 + }, + { + "epoch": 0.5956978733805915, + "grad_norm": 3.222412347793579, + "learning_rate": 7.416908271968246e-06, + "loss": 0.6656, + "step": 2437 + }, + { + "epoch": 0.595942312393058, + "grad_norm": 2.9758498668670654, + "learning_rate": 7.409260457785999e-06, + "loss": 0.7187, + "step": 2438 + }, + { + "epoch": 0.5961867514055244, + "grad_norm": 2.8184404373168945, + "learning_rate": 7.401614267581779e-06, + "loss": 0.6505, + "step": 2439 + }, + { + "epoch": 0.5964311904179908, + "grad_norm": 3.585016965866089, + "learning_rate": 7.393969706148524e-06, + "loss": 0.7019, + "step": 2440 + }, + { + "epoch": 0.5966756294304572, + "grad_norm": 4.0438995361328125, + "learning_rate": 7.386326778278142e-06, + "loss": 0.6532, + "step": 2441 + }, + { + "epoch": 0.5969200684429234, + "grad_norm": 3.1334328651428223, + "learning_rate": 7.378685488761533e-06, + "loss": 0.7686, + "step": 2442 + }, + { + "epoch": 0.5971645074553898, + "grad_norm": 3.9417247772216797, + "learning_rate": 7.371045842388552e-06, + "loss": 0.8383, + "step": 2443 + }, + { + "epoch": 0.5974089464678562, + "grad_norm": 4.557409763336182, + "learning_rate": 7.363407843948034e-06, + "loss": 0.6166, + "step": 2444 + }, + { + "epoch": 0.5976533854803227, + "grad_norm": 3.150182008743286, + "learning_rate": 7.355771498227777e-06, + "loss": 0.6707, + "step": 2445 + }, + { + "epoch": 0.597897824492789, + "grad_norm": 5.719577312469482, + "learning_rate": 7.3481368100145464e-06, + "loss": 0.7095, + "step": 2446 + }, + { + "epoch": 0.5981422635052555, + "grad_norm": 3.6504716873168945, + "learning_rate": 7.340503784094066e-06, + "loss": 0.6926, + "step": 2447 + }, + { + "epoch": 0.5983867025177219, + "grad_norm": 4.071256160736084, + "learning_rate": 7.332872425251017e-06, + "loss": 0.6978, + "step": 2448 + }, + { + "epoch": 0.5986311415301883, + "grad_norm": 3.5719997882843018, + "learning_rate": 7.325242738269039e-06, + "loss": 0.5985, + "step": 2449 + }, + { + "epoch": 0.5988755805426547, + "grad_norm": 2.8013484477996826, + "learning_rate": 7.317614727930722e-06, + "loss": 0.5605, + "step": 2450 + }, + { + "epoch": 0.599120019555121, + "grad_norm": 2.548478126525879, + "learning_rate": 7.3099883990176025e-06, + "loss": 1.9373, + "step": 2451 + }, + { + "epoch": 0.5993644585675874, + "grad_norm": 2.671673059463501, + "learning_rate": 7.302363756310165e-06, + "loss": 1.9074, + "step": 2452 + }, + { + "epoch": 0.5996088975800538, + "grad_norm": 2.897136926651001, + "learning_rate": 7.294740804587838e-06, + "loss": 0.6139, + "step": 2453 + }, + { + "epoch": 0.5998533365925202, + "grad_norm": 3.2317192554473877, + "learning_rate": 7.287119548628992e-06, + "loss": 0.7463, + "step": 2454 + }, + { + "epoch": 0.6000977756049866, + "grad_norm": 2.8740546703338623, + "learning_rate": 7.279499993210928e-06, + "loss": 0.6704, + "step": 2455 + }, + { + "epoch": 0.600342214617453, + "grad_norm": 3.8139729499816895, + "learning_rate": 7.27188214310988e-06, + "loss": 0.6726, + "step": 2456 + }, + { + "epoch": 0.6005866536299194, + "grad_norm": 2.5996851921081543, + "learning_rate": 7.26426600310103e-06, + "loss": 0.561, + "step": 2457 + }, + { + "epoch": 0.6008310926423858, + "grad_norm": 3.3971922397613525, + "learning_rate": 7.256651577958469e-06, + "loss": 0.7901, + "step": 2458 + }, + { + "epoch": 0.6010755316548522, + "grad_norm": 3.3375139236450195, + "learning_rate": 7.24903887245522e-06, + "loss": 0.596, + "step": 2459 + }, + { + "epoch": 0.6013199706673185, + "grad_norm": 3.5682437419891357, + "learning_rate": 7.241427891363228e-06, + "loss": 0.7292, + "step": 2460 + }, + { + "epoch": 0.6015644096797849, + "grad_norm": 2.866892099380493, + "learning_rate": 7.233818639453358e-06, + "loss": 0.6654, + "step": 2461 + }, + { + "epoch": 0.6018088486922513, + "grad_norm": 2.9490864276885986, + "learning_rate": 7.22621112149539e-06, + "loss": 0.5401, + "step": 2462 + }, + { + "epoch": 0.6020532877047177, + "grad_norm": 2.8294754028320312, + "learning_rate": 7.218605342258016e-06, + "loss": 0.6467, + "step": 2463 + }, + { + "epoch": 0.6022977267171841, + "grad_norm": 3.224865436553955, + "learning_rate": 7.21100130650884e-06, + "loss": 0.7148, + "step": 2464 + }, + { + "epoch": 0.6025421657296505, + "grad_norm": 4.785635471343994, + "learning_rate": 7.203399019014374e-06, + "loss": 0.6656, + "step": 2465 + }, + { + "epoch": 0.6027866047421169, + "grad_norm": 3.1482491493225098, + "learning_rate": 7.195798484540028e-06, + "loss": 0.6414, + "step": 2466 + }, + { + "epoch": 0.6030310437545833, + "grad_norm": 4.730430603027344, + "learning_rate": 7.188199707850123e-06, + "loss": 0.7205, + "step": 2467 + }, + { + "epoch": 0.6032754827670497, + "grad_norm": 3.990285873413086, + "learning_rate": 7.1806026937078695e-06, + "loss": 0.6635, + "step": 2468 + }, + { + "epoch": 0.603519921779516, + "grad_norm": 3.3099493980407715, + "learning_rate": 7.173007446875375e-06, + "loss": 0.7286, + "step": 2469 + }, + { + "epoch": 0.6037643607919824, + "grad_norm": 3.4665327072143555, + "learning_rate": 7.165413972113646e-06, + "loss": 0.7668, + "step": 2470 + }, + { + "epoch": 0.6040087998044488, + "grad_norm": 3.216830015182495, + "learning_rate": 7.15782227418257e-06, + "loss": 0.6273, + "step": 2471 + }, + { + "epoch": 0.6042532388169152, + "grad_norm": 2.6301674842834473, + "learning_rate": 7.150232357840921e-06, + "loss": 0.5984, + "step": 2472 + }, + { + "epoch": 0.6044976778293816, + "grad_norm": 3.5538887977600098, + "learning_rate": 7.1426442278463626e-06, + "loss": 0.7371, + "step": 2473 + }, + { + "epoch": 0.604742116841848, + "grad_norm": 4.289004802703857, + "learning_rate": 7.135057888955435e-06, + "loss": 0.6264, + "step": 2474 + }, + { + "epoch": 0.6049865558543144, + "grad_norm": 3.42405366897583, + "learning_rate": 7.127473345923553e-06, + "loss": 0.6621, + "step": 2475 + }, + { + "epoch": 0.6052309948667808, + "grad_norm": 3.0243327617645264, + "learning_rate": 7.1198906035050105e-06, + "loss": 0.7344, + "step": 2476 + }, + { + "epoch": 0.6054754338792472, + "grad_norm": 2.5014827251434326, + "learning_rate": 7.112309666452968e-06, + "loss": 2.0079, + "step": 2477 + }, + { + "epoch": 0.6057198728917135, + "grad_norm": 2.786853313446045, + "learning_rate": 7.10473053951946e-06, + "loss": 0.7084, + "step": 2478 + }, + { + "epoch": 0.6059643119041799, + "grad_norm": 3.303546667098999, + "learning_rate": 7.097153227455379e-06, + "loss": 0.7626, + "step": 2479 + }, + { + "epoch": 0.6062087509166463, + "grad_norm": 3.2983829975128174, + "learning_rate": 7.089577735010485e-06, + "loss": 0.7953, + "step": 2480 + }, + { + "epoch": 0.6064531899291127, + "grad_norm": 3.149487257003784, + "learning_rate": 7.0820040669333975e-06, + "loss": 0.7116, + "step": 2481 + }, + { + "epoch": 0.6066976289415791, + "grad_norm": 3.133228063583374, + "learning_rate": 7.074432227971593e-06, + "loss": 0.6555, + "step": 2482 + }, + { + "epoch": 0.6069420679540455, + "grad_norm": 3.6981680393218994, + "learning_rate": 7.066862222871397e-06, + "loss": 0.5982, + "step": 2483 + }, + { + "epoch": 0.6071865069665119, + "grad_norm": 4.039612770080566, + "learning_rate": 7.059294056377988e-06, + "loss": 0.5946, + "step": 2484 + }, + { + "epoch": 0.6074309459789783, + "grad_norm": 3.019282579421997, + "learning_rate": 7.051727733235395e-06, + "loss": 0.6346, + "step": 2485 + }, + { + "epoch": 0.6076753849914447, + "grad_norm": 2.6989247798919678, + "learning_rate": 7.044163258186485e-06, + "loss": 0.6808, + "step": 2486 + }, + { + "epoch": 0.607919824003911, + "grad_norm": 3.9518232345581055, + "learning_rate": 7.036600635972971e-06, + "loss": 0.7237, + "step": 2487 + }, + { + "epoch": 0.6081642630163774, + "grad_norm": 4.522764682769775, + "learning_rate": 7.029039871335406e-06, + "loss": 0.6655, + "step": 2488 + }, + { + "epoch": 0.6084087020288438, + "grad_norm": 3.1290531158447266, + "learning_rate": 7.021480969013176e-06, + "loss": 0.745, + "step": 2489 + }, + { + "epoch": 0.6086531410413102, + "grad_norm": 2.9655075073242188, + "learning_rate": 7.0139239337444956e-06, + "loss": 0.7403, + "step": 2490 + }, + { + "epoch": 0.6088975800537766, + "grad_norm": 3.183637857437134, + "learning_rate": 7.006368770266421e-06, + "loss": 0.6196, + "step": 2491 + }, + { + "epoch": 0.609142019066243, + "grad_norm": 3.102365016937256, + "learning_rate": 6.99881548331482e-06, + "loss": 0.6792, + "step": 2492 + }, + { + "epoch": 0.6093864580787094, + "grad_norm": 3.733527660369873, + "learning_rate": 6.9912640776243974e-06, + "loss": 0.6448, + "step": 2493 + }, + { + "epoch": 0.6096308970911758, + "grad_norm": 2.7660675048828125, + "learning_rate": 6.983714557928672e-06, + "loss": 0.6567, + "step": 2494 + }, + { + "epoch": 0.6098753361036422, + "grad_norm": 4.2180938720703125, + "learning_rate": 6.976166928959978e-06, + "loss": 0.7289, + "step": 2495 + }, + { + "epoch": 0.6101197751161085, + "grad_norm": 2.9590981006622314, + "learning_rate": 6.96862119544947e-06, + "loss": 0.6344, + "step": 2496 + }, + { + "epoch": 0.6103642141285749, + "grad_norm": 5.102367401123047, + "learning_rate": 6.961077362127111e-06, + "loss": 0.6535, + "step": 2497 + }, + { + "epoch": 0.6106086531410413, + "grad_norm": 2.679868221282959, + "learning_rate": 6.953535433721677e-06, + "loss": 0.6962, + "step": 2498 + }, + { + "epoch": 0.6108530921535077, + "grad_norm": 4.007815837860107, + "learning_rate": 6.945995414960744e-06, + "loss": 0.8283, + "step": 2499 + }, + { + "epoch": 0.6110975311659741, + "grad_norm": 4.890908241271973, + "learning_rate": 6.938457310570696e-06, + "loss": 0.6869, + "step": 2500 + }, + { + "epoch": 0.6113419701784405, + "grad_norm": 3.021857261657715, + "learning_rate": 6.930921125276715e-06, + "loss": 0.6152, + "step": 2501 + }, + { + "epoch": 0.6115864091909069, + "grad_norm": 2.963789701461792, + "learning_rate": 6.923386863802779e-06, + "loss": 0.5795, + "step": 2502 + }, + { + "epoch": 0.6118308482033733, + "grad_norm": 3.034787178039551, + "learning_rate": 6.915854530871657e-06, + "loss": 0.5507, + "step": 2503 + }, + { + "epoch": 0.6120752872158397, + "grad_norm": 2.8483729362487793, + "learning_rate": 6.9083241312049235e-06, + "loss": 0.7486, + "step": 2504 + }, + { + "epoch": 0.612319726228306, + "grad_norm": 3.7511532306671143, + "learning_rate": 6.900795669522924e-06, + "loss": 0.8166, + "step": 2505 + }, + { + "epoch": 0.6125641652407724, + "grad_norm": 3.2344181537628174, + "learning_rate": 6.893269150544797e-06, + "loss": 0.6897, + "step": 2506 + }, + { + "epoch": 0.6128086042532388, + "grad_norm": 3.738460063934326, + "learning_rate": 6.885744578988463e-06, + "loss": 0.7133, + "step": 2507 + }, + { + "epoch": 0.6130530432657052, + "grad_norm": 5.127227783203125, + "learning_rate": 6.878221959570621e-06, + "loss": 0.595, + "step": 2508 + }, + { + "epoch": 0.6132974822781716, + "grad_norm": 4.383499622344971, + "learning_rate": 6.870701297006746e-06, + "loss": 0.6411, + "step": 2509 + }, + { + "epoch": 0.613541921290638, + "grad_norm": 3.2642929553985596, + "learning_rate": 6.8631825960110866e-06, + "loss": 0.7823, + "step": 2510 + }, + { + "epoch": 0.6137863603031044, + "grad_norm": 3.017092704772949, + "learning_rate": 6.855665861296662e-06, + "loss": 0.6362, + "step": 2511 + }, + { + "epoch": 0.6140307993155708, + "grad_norm": 3.371415138244629, + "learning_rate": 6.84815109757526e-06, + "loss": 0.6656, + "step": 2512 + }, + { + "epoch": 0.6142752383280372, + "grad_norm": 3.385425329208374, + "learning_rate": 6.84063830955743e-06, + "loss": 0.6934, + "step": 2513 + }, + { + "epoch": 0.6145196773405035, + "grad_norm": 3.7038681507110596, + "learning_rate": 6.833127501952488e-06, + "loss": 0.5947, + "step": 2514 + }, + { + "epoch": 0.6147641163529699, + "grad_norm": 2.9765119552612305, + "learning_rate": 6.8256186794685016e-06, + "loss": 0.5329, + "step": 2515 + }, + { + "epoch": 0.6150085553654363, + "grad_norm": 2.7302298545837402, + "learning_rate": 6.818111846812299e-06, + "loss": 0.6509, + "step": 2516 + }, + { + "epoch": 0.6152529943779027, + "grad_norm": 3.5046465396881104, + "learning_rate": 6.810607008689461e-06, + "loss": 0.6428, + "step": 2517 + }, + { + "epoch": 0.6154974333903691, + "grad_norm": 3.45648455619812, + "learning_rate": 6.803104169804311e-06, + "loss": 0.707, + "step": 2518 + }, + { + "epoch": 0.6157418724028355, + "grad_norm": 3.033381938934326, + "learning_rate": 6.795603334859935e-06, + "loss": 0.6832, + "step": 2519 + }, + { + "epoch": 0.6159863114153019, + "grad_norm": 3.4825875759124756, + "learning_rate": 6.788104508558147e-06, + "loss": 0.6363, + "step": 2520 + }, + { + "epoch": 0.6162307504277683, + "grad_norm": 3.1465296745300293, + "learning_rate": 6.78060769559951e-06, + "loss": 0.6199, + "step": 2521 + }, + { + "epoch": 0.6164751894402347, + "grad_norm": 3.8383195400238037, + "learning_rate": 6.773112900683323e-06, + "loss": 0.6372, + "step": 2522 + }, + { + "epoch": 0.616719628452701, + "grad_norm": 3.987447738647461, + "learning_rate": 6.7656201285076195e-06, + "loss": 0.7747, + "step": 2523 + }, + { + "epoch": 0.6169640674651674, + "grad_norm": 2.686070680618286, + "learning_rate": 6.758129383769165e-06, + "loss": 0.6127, + "step": 2524 + }, + { + "epoch": 0.6172085064776338, + "grad_norm": 2.738743543624878, + "learning_rate": 6.7506406711634545e-06, + "loss": 0.6823, + "step": 2525 + }, + { + "epoch": 0.6174529454901002, + "grad_norm": 3.0114638805389404, + "learning_rate": 6.743153995384709e-06, + "loss": 0.7053, + "step": 2526 + }, + { + "epoch": 0.6176973845025666, + "grad_norm": 3.4111175537109375, + "learning_rate": 6.735669361125874e-06, + "loss": 0.64, + "step": 2527 + }, + { + "epoch": 0.617941823515033, + "grad_norm": 2.8889999389648438, + "learning_rate": 6.728186773078614e-06, + "loss": 0.7055, + "step": 2528 + }, + { + "epoch": 0.6181862625274994, + "grad_norm": 3.439328670501709, + "learning_rate": 6.720706235933312e-06, + "loss": 0.5602, + "step": 2529 + }, + { + "epoch": 0.6184307015399658, + "grad_norm": 2.9301159381866455, + "learning_rate": 6.713227754379063e-06, + "loss": 0.7021, + "step": 2530 + }, + { + "epoch": 0.6186751405524322, + "grad_norm": 3.2925527095794678, + "learning_rate": 6.705751333103676e-06, + "loss": 0.6667, + "step": 2531 + }, + { + "epoch": 0.6189195795648985, + "grad_norm": 3.848553419113159, + "learning_rate": 6.6982769767936676e-06, + "loss": 0.7464, + "step": 2532 + }, + { + "epoch": 0.6191640185773649, + "grad_norm": 3.1005401611328125, + "learning_rate": 6.690804690134259e-06, + "loss": 0.6674, + "step": 2533 + }, + { + "epoch": 0.6194084575898313, + "grad_norm": 3.32075834274292, + "learning_rate": 6.6833344778093735e-06, + "loss": 0.6011, + "step": 2534 + }, + { + "epoch": 0.6196528966022977, + "grad_norm": 3.6397931575775146, + "learning_rate": 6.675866344501641e-06, + "loss": 0.727, + "step": 2535 + }, + { + "epoch": 0.6198973356147641, + "grad_norm": 3.8641738891601562, + "learning_rate": 6.668400294892379e-06, + "loss": 0.7457, + "step": 2536 + }, + { + "epoch": 0.6201417746272305, + "grad_norm": 2.848632335662842, + "learning_rate": 6.660936333661605e-06, + "loss": 0.667, + "step": 2537 + }, + { + "epoch": 0.6203862136396969, + "grad_norm": 2.7623350620269775, + "learning_rate": 6.653474465488023e-06, + "loss": 0.6282, + "step": 2538 + }, + { + "epoch": 0.6206306526521633, + "grad_norm": 3.0257644653320312, + "learning_rate": 6.64601469504903e-06, + "loss": 0.5238, + "step": 2539 + }, + { + "epoch": 0.6208750916646297, + "grad_norm": 2.996584892272949, + "learning_rate": 6.638557027020703e-06, + "loss": 0.6428, + "step": 2540 + }, + { + "epoch": 0.621119530677096, + "grad_norm": 3.1068155765533447, + "learning_rate": 6.631101466077801e-06, + "loss": 0.6324, + "step": 2541 + }, + { + "epoch": 0.6213639696895624, + "grad_norm": 4.016943454742432, + "learning_rate": 6.623648016893769e-06, + "loss": 0.6182, + "step": 2542 + }, + { + "epoch": 0.6216084087020288, + "grad_norm": 3.138049602508545, + "learning_rate": 6.616196684140721e-06, + "loss": 0.5747, + "step": 2543 + }, + { + "epoch": 0.6218528477144952, + "grad_norm": 2.8710036277770996, + "learning_rate": 6.60874747248945e-06, + "loss": 0.6195, + "step": 2544 + }, + { + "epoch": 0.6220972867269616, + "grad_norm": 4.811293601989746, + "learning_rate": 6.601300386609412e-06, + "loss": 0.6909, + "step": 2545 + }, + { + "epoch": 0.622341725739428, + "grad_norm": 4.38634729385376, + "learning_rate": 6.593855431168739e-06, + "loss": 0.593, + "step": 2546 + }, + { + "epoch": 0.6225861647518944, + "grad_norm": 3.7089486122131348, + "learning_rate": 6.5864126108342215e-06, + "loss": 0.6747, + "step": 2547 + }, + { + "epoch": 0.6228306037643608, + "grad_norm": 2.638113260269165, + "learning_rate": 6.578971930271313e-06, + "loss": 0.5938, + "step": 2548 + }, + { + "epoch": 0.6230750427768272, + "grad_norm": 3.1624960899353027, + "learning_rate": 6.571533394144127e-06, + "loss": 0.6043, + "step": 2549 + }, + { + "epoch": 0.6233194817892935, + "grad_norm": 3.2632787227630615, + "learning_rate": 6.5640970071154356e-06, + "loss": 0.6845, + "step": 2550 + }, + { + "epoch": 0.6235639208017599, + "grad_norm": 3.290696382522583, + "learning_rate": 6.556662773846658e-06, + "loss": 0.7496, + "step": 2551 + }, + { + "epoch": 0.6238083598142263, + "grad_norm": 3.2277283668518066, + "learning_rate": 6.549230698997865e-06, + "loss": 0.7165, + "step": 2552 + }, + { + "epoch": 0.6240527988266927, + "grad_norm": 3.4584851264953613, + "learning_rate": 6.541800787227776e-06, + "loss": 0.7162, + "step": 2553 + }, + { + "epoch": 0.6242972378391591, + "grad_norm": 2.970109462738037, + "learning_rate": 6.534373043193751e-06, + "loss": 0.661, + "step": 2554 + }, + { + "epoch": 0.6245416768516255, + "grad_norm": 4.127514362335205, + "learning_rate": 6.526947471551799e-06, + "loss": 0.6906, + "step": 2555 + }, + { + "epoch": 0.6247861158640919, + "grad_norm": 3.361604690551758, + "learning_rate": 6.519524076956558e-06, + "loss": 0.5777, + "step": 2556 + }, + { + "epoch": 0.6250305548765583, + "grad_norm": 2.528740167617798, + "learning_rate": 6.512102864061305e-06, + "loss": 1.9661, + "step": 2557 + }, + { + "epoch": 0.6252749938890246, + "grad_norm": 2.5687520503997803, + "learning_rate": 6.504683837517953e-06, + "loss": 0.5021, + "step": 2558 + }, + { + "epoch": 0.625519432901491, + "grad_norm": 3.5585079193115234, + "learning_rate": 6.497267001977037e-06, + "loss": 0.6798, + "step": 2559 + }, + { + "epoch": 0.6257638719139574, + "grad_norm": 3.9767425060272217, + "learning_rate": 6.489852362087727e-06, + "loss": 0.6911, + "step": 2560 + }, + { + "epoch": 0.6260083109264238, + "grad_norm": 3.0334603786468506, + "learning_rate": 6.48243992249781e-06, + "loss": 0.5789, + "step": 2561 + }, + { + "epoch": 0.6262527499388902, + "grad_norm": 2.5939929485321045, + "learning_rate": 6.4750296878537004e-06, + "loss": 0.5324, + "step": 2562 + }, + { + "epoch": 0.6264971889513566, + "grad_norm": 2.6828372478485107, + "learning_rate": 6.4676216628004204e-06, + "loss": 0.5582, + "step": 2563 + }, + { + "epoch": 0.626741627963823, + "grad_norm": 2.956088066101074, + "learning_rate": 6.460215851981617e-06, + "loss": 0.6962, + "step": 2564 + }, + { + "epoch": 0.6269860669762894, + "grad_norm": 3.4469640254974365, + "learning_rate": 6.452812260039542e-06, + "loss": 0.8145, + "step": 2565 + }, + { + "epoch": 0.6272305059887558, + "grad_norm": 2.3670003414154053, + "learning_rate": 6.445410891615067e-06, + "loss": 1.9115, + "step": 2566 + }, + { + "epoch": 0.6274749450012221, + "grad_norm": 2.997681140899658, + "learning_rate": 6.4380117513476545e-06, + "loss": 0.5838, + "step": 2567 + }, + { + "epoch": 0.6277193840136885, + "grad_norm": 2.9319982528686523, + "learning_rate": 6.430614843875384e-06, + "loss": 0.6775, + "step": 2568 + }, + { + "epoch": 0.6279638230261549, + "grad_norm": 2.865135431289673, + "learning_rate": 6.423220173834923e-06, + "loss": 0.6572, + "step": 2569 + }, + { + "epoch": 0.6282082620386213, + "grad_norm": 2.836655616760254, + "learning_rate": 6.415827745861552e-06, + "loss": 0.6656, + "step": 2570 + }, + { + "epoch": 0.6284527010510877, + "grad_norm": 3.49206805229187, + "learning_rate": 6.40843756458913e-06, + "loss": 0.6767, + "step": 2571 + }, + { + "epoch": 0.6286971400635541, + "grad_norm": 2.854118585586548, + "learning_rate": 6.401049634650119e-06, + "loss": 0.6471, + "step": 2572 + }, + { + "epoch": 0.6289415790760206, + "grad_norm": 3.155752420425415, + "learning_rate": 6.393663960675565e-06, + "loss": 0.7015, + "step": 2573 + }, + { + "epoch": 0.629186018088487, + "grad_norm": 2.570833206176758, + "learning_rate": 6.386280547295099e-06, + "loss": 0.5743, + "step": 2574 + }, + { + "epoch": 0.6294304571009534, + "grad_norm": 2.7275097370147705, + "learning_rate": 6.378899399136937e-06, + "loss": 0.5756, + "step": 2575 + }, + { + "epoch": 0.6296748961134196, + "grad_norm": 3.417999267578125, + "learning_rate": 6.371520520827876e-06, + "loss": 0.6533, + "step": 2576 + }, + { + "epoch": 0.629919335125886, + "grad_norm": 2.962523937225342, + "learning_rate": 6.3641439169932895e-06, + "loss": 0.7356, + "step": 2577 + }, + { + "epoch": 0.6301637741383525, + "grad_norm": 7.3755269050598145, + "learning_rate": 6.356769592257122e-06, + "loss": 0.7784, + "step": 2578 + }, + { + "epoch": 0.6304082131508189, + "grad_norm": 2.9884088039398193, + "learning_rate": 6.349397551241894e-06, + "loss": 0.6326, + "step": 2579 + }, + { + "epoch": 0.6306526521632853, + "grad_norm": 3.609339952468872, + "learning_rate": 6.342027798568686e-06, + "loss": 0.6522, + "step": 2580 + }, + { + "epoch": 0.6308970911757517, + "grad_norm": 3.33388090133667, + "learning_rate": 6.3346603388571605e-06, + "loss": 0.7649, + "step": 2581 + }, + { + "epoch": 0.6311415301882181, + "grad_norm": 2.7813782691955566, + "learning_rate": 6.3272951767255284e-06, + "loss": 0.731, + "step": 2582 + }, + { + "epoch": 0.6313859692006845, + "grad_norm": 3.0105741024017334, + "learning_rate": 6.319932316790565e-06, + "loss": 0.8428, + "step": 2583 + }, + { + "epoch": 0.6316304082131509, + "grad_norm": 3.0146050453186035, + "learning_rate": 6.3125717636676005e-06, + "loss": 0.6955, + "step": 2584 + }, + { + "epoch": 0.6318748472256172, + "grad_norm": 2.816695213317871, + "learning_rate": 6.305213521970523e-06, + "loss": 0.7129, + "step": 2585 + }, + { + "epoch": 0.6321192862380836, + "grad_norm": 2.3536274433135986, + "learning_rate": 6.297857596311769e-06, + "loss": 1.9621, + "step": 2586 + }, + { + "epoch": 0.63236372525055, + "grad_norm": 3.4241769313812256, + "learning_rate": 6.290503991302324e-06, + "loss": 0.6107, + "step": 2587 + }, + { + "epoch": 0.6326081642630164, + "grad_norm": 2.7386038303375244, + "learning_rate": 6.283152711551714e-06, + "loss": 0.6496, + "step": 2588 + }, + { + "epoch": 0.6328526032754828, + "grad_norm": 5.59810209274292, + "learning_rate": 6.275803761668019e-06, + "loss": 0.6817, + "step": 2589 + }, + { + "epoch": 0.6330970422879492, + "grad_norm": 2.6066410541534424, + "learning_rate": 6.2684571462578465e-06, + "loss": 0.703, + "step": 2590 + }, + { + "epoch": 0.6333414813004156, + "grad_norm": 3.150466203689575, + "learning_rate": 6.261112869926348e-06, + "loss": 0.7259, + "step": 2591 + }, + { + "epoch": 0.633585920312882, + "grad_norm": 2.94368052482605, + "learning_rate": 6.253770937277206e-06, + "loss": 0.6505, + "step": 2592 + }, + { + "epoch": 0.6338303593253484, + "grad_norm": 3.3030643463134766, + "learning_rate": 6.246431352912632e-06, + "loss": 0.6825, + "step": 2593 + }, + { + "epoch": 0.6340747983378147, + "grad_norm": 2.6119279861450195, + "learning_rate": 6.239094121433371e-06, + "loss": 0.6397, + "step": 2594 + }, + { + "epoch": 0.6343192373502811, + "grad_norm": 3.283381700515747, + "learning_rate": 6.231759247438689e-06, + "loss": 0.7639, + "step": 2595 + }, + { + "epoch": 0.6345636763627475, + "grad_norm": 2.4869697093963623, + "learning_rate": 6.224426735526372e-06, + "loss": 0.5081, + "step": 2596 + }, + { + "epoch": 0.6348081153752139, + "grad_norm": 3.226597785949707, + "learning_rate": 6.217096590292734e-06, + "loss": 0.59, + "step": 2597 + }, + { + "epoch": 0.6350525543876803, + "grad_norm": 2.7441442012786865, + "learning_rate": 6.2097688163325974e-06, + "loss": 0.5933, + "step": 2598 + }, + { + "epoch": 0.6352969934001467, + "grad_norm": 2.381624937057495, + "learning_rate": 6.202443418239301e-06, + "loss": 1.9185, + "step": 2599 + }, + { + "epoch": 0.6355414324126131, + "grad_norm": 2.9569995403289795, + "learning_rate": 6.195120400604695e-06, + "loss": 0.7283, + "step": 2600 + }, + { + "epoch": 0.6357858714250795, + "grad_norm": 3.2580981254577637, + "learning_rate": 6.187799768019134e-06, + "loss": 0.5857, + "step": 2601 + }, + { + "epoch": 0.6360303104375459, + "grad_norm": 2.321307897567749, + "learning_rate": 6.180481525071482e-06, + "loss": 1.9926, + "step": 2602 + }, + { + "epoch": 0.6362747494500122, + "grad_norm": 3.270001173019409, + "learning_rate": 6.173165676349103e-06, + "loss": 0.6934, + "step": 2603 + }, + { + "epoch": 0.6365191884624786, + "grad_norm": 3.450073003768921, + "learning_rate": 6.165852226437861e-06, + "loss": 0.6653, + "step": 2604 + }, + { + "epoch": 0.636763627474945, + "grad_norm": 3.1501264572143555, + "learning_rate": 6.158541179922113e-06, + "loss": 0.8334, + "step": 2605 + }, + { + "epoch": 0.6370080664874114, + "grad_norm": 4.02630615234375, + "learning_rate": 6.1512325413847154e-06, + "loss": 0.787, + "step": 2606 + }, + { + "epoch": 0.6372525054998778, + "grad_norm": 2.958813190460205, + "learning_rate": 6.143926315407011e-06, + "loss": 0.646, + "step": 2607 + }, + { + "epoch": 0.6374969445123442, + "grad_norm": 3.098252773284912, + "learning_rate": 6.13662250656883e-06, + "loss": 0.6546, + "step": 2608 + }, + { + "epoch": 0.6377413835248106, + "grad_norm": 3.045335292816162, + "learning_rate": 6.1293211194484895e-06, + "loss": 0.6385, + "step": 2609 + }, + { + "epoch": 0.637985822537277, + "grad_norm": 6.5176568031311035, + "learning_rate": 6.122022158622787e-06, + "loss": 0.6316, + "step": 2610 + }, + { + "epoch": 0.6382302615497434, + "grad_norm": 2.843134880065918, + "learning_rate": 6.114725628666997e-06, + "loss": 0.7969, + "step": 2611 + }, + { + "epoch": 0.6384747005622097, + "grad_norm": 2.9319586753845215, + "learning_rate": 6.107431534154882e-06, + "loss": 0.6795, + "step": 2612 + }, + { + "epoch": 0.6387191395746761, + "grad_norm": 3.498922824859619, + "learning_rate": 6.1001398796586644e-06, + "loss": 0.668, + "step": 2613 + }, + { + "epoch": 0.6389635785871425, + "grad_norm": 3.7721149921417236, + "learning_rate": 6.092850669749038e-06, + "loss": 0.7027, + "step": 2614 + }, + { + "epoch": 0.6392080175996089, + "grad_norm": 3.661775827407837, + "learning_rate": 6.085563908995174e-06, + "loss": 0.7541, + "step": 2615 + }, + { + "epoch": 0.6394524566120753, + "grad_norm": 2.810243606567383, + "learning_rate": 6.0782796019646984e-06, + "loss": 0.6738, + "step": 2616 + }, + { + "epoch": 0.6396968956245417, + "grad_norm": 2.9646449089050293, + "learning_rate": 6.070997753223703e-06, + "loss": 0.5183, + "step": 2617 + }, + { + "epoch": 0.6399413346370081, + "grad_norm": 3.056952714920044, + "learning_rate": 6.063718367336739e-06, + "loss": 0.5632, + "step": 2618 + }, + { + "epoch": 0.6401857736494745, + "grad_norm": 2.522547483444214, + "learning_rate": 6.056441448866817e-06, + "loss": 0.7386, + "step": 2619 + }, + { + "epoch": 0.6404302126619409, + "grad_norm": 2.832892894744873, + "learning_rate": 6.0491670023753914e-06, + "loss": 0.6423, + "step": 2620 + }, + { + "epoch": 0.6406746516744072, + "grad_norm": 2.9263675212860107, + "learning_rate": 6.041895032422377e-06, + "loss": 0.638, + "step": 2621 + }, + { + "epoch": 0.6409190906868736, + "grad_norm": 3.260471820831299, + "learning_rate": 6.03462554356613e-06, + "loss": 0.6096, + "step": 2622 + }, + { + "epoch": 0.64116352969934, + "grad_norm": 3.9631731510162354, + "learning_rate": 6.027358540363457e-06, + "loss": 0.6807, + "step": 2623 + }, + { + "epoch": 0.6414079687118064, + "grad_norm": 3.1918716430664062, + "learning_rate": 6.020094027369598e-06, + "loss": 0.7503, + "step": 2624 + }, + { + "epoch": 0.6416524077242728, + "grad_norm": 2.8288021087646484, + "learning_rate": 6.012832009138243e-06, + "loss": 0.6821, + "step": 2625 + }, + { + "epoch": 0.6418968467367392, + "grad_norm": 2.9716744422912598, + "learning_rate": 6.005572490221507e-06, + "loss": 0.6767, + "step": 2626 + }, + { + "epoch": 0.6421412857492056, + "grad_norm": 4.20635986328125, + "learning_rate": 5.998315475169942e-06, + "loss": 0.6081, + "step": 2627 + }, + { + "epoch": 0.642385724761672, + "grad_norm": 2.957223415374756, + "learning_rate": 5.9910609685325415e-06, + "loss": 0.7475, + "step": 2628 + }, + { + "epoch": 0.6426301637741384, + "grad_norm": 2.925553798675537, + "learning_rate": 5.983808974856711e-06, + "loss": 0.6628, + "step": 2629 + }, + { + "epoch": 0.6428746027866047, + "grad_norm": 3.361935615539551, + "learning_rate": 5.976559498688288e-06, + "loss": 0.6694, + "step": 2630 + }, + { + "epoch": 0.6431190417990711, + "grad_norm": 2.64375376701355, + "learning_rate": 5.969312544571529e-06, + "loss": 0.5882, + "step": 2631 + }, + { + "epoch": 0.6433634808115375, + "grad_norm": 3.661879777908325, + "learning_rate": 5.962068117049115e-06, + "loss": 0.5921, + "step": 2632 + }, + { + "epoch": 0.6436079198240039, + "grad_norm": 3.00673770904541, + "learning_rate": 5.9548262206621375e-06, + "loss": 0.7094, + "step": 2633 + }, + { + "epoch": 0.6438523588364703, + "grad_norm": 3.405564546585083, + "learning_rate": 5.947586859950103e-06, + "loss": 0.7097, + "step": 2634 + }, + { + "epoch": 0.6440967978489367, + "grad_norm": 3.2848403453826904, + "learning_rate": 5.94035003945093e-06, + "loss": 0.7331, + "step": 2635 + }, + { + "epoch": 0.6443412368614031, + "grad_norm": 3.589048385620117, + "learning_rate": 5.9331157637009466e-06, + "loss": 0.6206, + "step": 2636 + }, + { + "epoch": 0.6445856758738695, + "grad_norm": 3.4497900009155273, + "learning_rate": 5.925884037234879e-06, + "loss": 0.6693, + "step": 2637 + }, + { + "epoch": 0.6448301148863359, + "grad_norm": 3.1013028621673584, + "learning_rate": 5.918654864585859e-06, + "loss": 0.7106, + "step": 2638 + }, + { + "epoch": 0.6450745538988022, + "grad_norm": 2.397865056991577, + "learning_rate": 5.911428250285422e-06, + "loss": 1.8904, + "step": 2639 + }, + { + "epoch": 0.6453189929112686, + "grad_norm": 4.107816219329834, + "learning_rate": 5.904204198863492e-06, + "loss": 0.645, + "step": 2640 + }, + { + "epoch": 0.645563431923735, + "grad_norm": 3.6357409954071045, + "learning_rate": 5.8969827148483935e-06, + "loss": 0.6926, + "step": 2641 + }, + { + "epoch": 0.6458078709362014, + "grad_norm": 3.5307352542877197, + "learning_rate": 5.889763802766835e-06, + "loss": 0.6974, + "step": 2642 + }, + { + "epoch": 0.6460523099486678, + "grad_norm": 2.94061279296875, + "learning_rate": 5.88254746714392e-06, + "loss": 0.6028, + "step": 2643 + }, + { + "epoch": 0.6462967489611342, + "grad_norm": 2.9789211750030518, + "learning_rate": 5.87533371250313e-06, + "loss": 0.6147, + "step": 2644 + }, + { + "epoch": 0.6465411879736006, + "grad_norm": 3.805901050567627, + "learning_rate": 5.868122543366332e-06, + "loss": 0.7588, + "step": 2645 + }, + { + "epoch": 0.646785626986067, + "grad_norm": 2.642486095428467, + "learning_rate": 5.860913964253774e-06, + "loss": 0.6221, + "step": 2646 + }, + { + "epoch": 0.6470300659985334, + "grad_norm": 2.385053873062134, + "learning_rate": 5.853707979684077e-06, + "loss": 1.9012, + "step": 2647 + }, + { + "epoch": 0.6472745050109997, + "grad_norm": 3.222817897796631, + "learning_rate": 5.846504594174236e-06, + "loss": 0.6914, + "step": 2648 + }, + { + "epoch": 0.6475189440234661, + "grad_norm": 2.6899101734161377, + "learning_rate": 5.8393038122396205e-06, + "loss": 0.6335, + "step": 2649 + }, + { + "epoch": 0.6477633830359325, + "grad_norm": 2.827629327774048, + "learning_rate": 5.832105638393968e-06, + "loss": 0.4106, + "step": 2650 + }, + { + "epoch": 0.6480078220483989, + "grad_norm": 3.3581056594848633, + "learning_rate": 5.824910077149372e-06, + "loss": 0.6776, + "step": 2651 + }, + { + "epoch": 0.6482522610608653, + "grad_norm": 3.319446325302124, + "learning_rate": 5.817717133016298e-06, + "loss": 0.6217, + "step": 2652 + }, + { + "epoch": 0.6484967000733317, + "grad_norm": 3.3250958919525146, + "learning_rate": 5.810526810503567e-06, + "loss": 0.638, + "step": 2653 + }, + { + "epoch": 0.6487411390857981, + "grad_norm": 3.6977198123931885, + "learning_rate": 5.80333911411836e-06, + "loss": 0.6203, + "step": 2654 + }, + { + "epoch": 0.6489855780982645, + "grad_norm": 2.956408739089966, + "learning_rate": 5.796154048366209e-06, + "loss": 0.5279, + "step": 2655 + }, + { + "epoch": 0.6492300171107309, + "grad_norm": 3.3112738132476807, + "learning_rate": 5.7889716177509945e-06, + "loss": 0.6822, + "step": 2656 + }, + { + "epoch": 0.6494744561231972, + "grad_norm": 3.304767370223999, + "learning_rate": 5.7817918267749505e-06, + "loss": 0.552, + "step": 2657 + }, + { + "epoch": 0.6497188951356636, + "grad_norm": 3.368194103240967, + "learning_rate": 5.774614679938651e-06, + "loss": 0.7749, + "step": 2658 + }, + { + "epoch": 0.64996333414813, + "grad_norm": 3.636293411254883, + "learning_rate": 5.76744018174102e-06, + "loss": 0.7058, + "step": 2659 + }, + { + "epoch": 0.6502077731605964, + "grad_norm": 2.703385829925537, + "learning_rate": 5.760268336679315e-06, + "loss": 0.6515, + "step": 2660 + }, + { + "epoch": 0.6504522121730628, + "grad_norm": 9.689297676086426, + "learning_rate": 5.753099149249133e-06, + "loss": 0.6111, + "step": 2661 + }, + { + "epoch": 0.6506966511855292, + "grad_norm": 3.1031436920166016, + "learning_rate": 5.7459326239444025e-06, + "loss": 0.6971, + "step": 2662 + }, + { + "epoch": 0.6509410901979956, + "grad_norm": 3.6608095169067383, + "learning_rate": 5.738768765257386e-06, + "loss": 0.6315, + "step": 2663 + }, + { + "epoch": 0.651185529210462, + "grad_norm": 4.4993743896484375, + "learning_rate": 5.731607577678675e-06, + "loss": 0.6462, + "step": 2664 + }, + { + "epoch": 0.6514299682229284, + "grad_norm": 3.131669282913208, + "learning_rate": 5.724449065697182e-06, + "loss": 0.6861, + "step": 2665 + }, + { + "epoch": 0.6516744072353947, + "grad_norm": 4.2920756340026855, + "learning_rate": 5.717293233800148e-06, + "loss": 0.6796, + "step": 2666 + }, + { + "epoch": 0.6519188462478611, + "grad_norm": 3.0195488929748535, + "learning_rate": 5.7101400864731294e-06, + "loss": 0.6626, + "step": 2667 + }, + { + "epoch": 0.6521632852603275, + "grad_norm": 3.632920742034912, + "learning_rate": 5.702989628200004e-06, + "loss": 0.6894, + "step": 2668 + }, + { + "epoch": 0.6524077242727939, + "grad_norm": 4.062261581420898, + "learning_rate": 5.69584186346296e-06, + "loss": 0.7573, + "step": 2669 + }, + { + "epoch": 0.6526521632852603, + "grad_norm": 3.829195499420166, + "learning_rate": 5.688696796742501e-06, + "loss": 0.7282, + "step": 2670 + }, + { + "epoch": 0.6528966022977267, + "grad_norm": 3.4981908798217773, + "learning_rate": 5.681554432517435e-06, + "loss": 0.5847, + "step": 2671 + }, + { + "epoch": 0.6531410413101931, + "grad_norm": 2.9106080532073975, + "learning_rate": 5.674414775264878e-06, + "loss": 0.4843, + "step": 2672 + }, + { + "epoch": 0.6533854803226595, + "grad_norm": 2.600187301635742, + "learning_rate": 5.667277829460254e-06, + "loss": 1.965, + "step": 2673 + }, + { + "epoch": 0.6536299193351259, + "grad_norm": 2.7137207984924316, + "learning_rate": 5.6601435995772814e-06, + "loss": 0.6175, + "step": 2674 + }, + { + "epoch": 0.6538743583475922, + "grad_norm": 3.5779716968536377, + "learning_rate": 5.6530120900879774e-06, + "loss": 0.6038, + "step": 2675 + }, + { + "epoch": 0.6541187973600586, + "grad_norm": 3.0512759685516357, + "learning_rate": 5.6458833054626514e-06, + "loss": 0.6629, + "step": 2676 + }, + { + "epoch": 0.654363236372525, + "grad_norm": 2.890873908996582, + "learning_rate": 5.638757250169913e-06, + "loss": 0.5274, + "step": 2677 + }, + { + "epoch": 0.6546076753849914, + "grad_norm": 2.889688491821289, + "learning_rate": 5.6316339286766496e-06, + "loss": 0.609, + "step": 2678 + }, + { + "epoch": 0.6548521143974578, + "grad_norm": 4.647675037384033, + "learning_rate": 5.6245133454480446e-06, + "loss": 0.6843, + "step": 2679 + }, + { + "epoch": 0.6550965534099242, + "grad_norm": 3.1891238689422607, + "learning_rate": 5.61739550494756e-06, + "loss": 0.6044, + "step": 2680 + }, + { + "epoch": 0.6553409924223906, + "grad_norm": 2.9441967010498047, + "learning_rate": 5.610280411636941e-06, + "loss": 0.6735, + "step": 2681 + }, + { + "epoch": 0.655585431434857, + "grad_norm": 3.5355992317199707, + "learning_rate": 5.603168069976209e-06, + "loss": 0.6165, + "step": 2682 + }, + { + "epoch": 0.6558298704473234, + "grad_norm": 3.233180522918701, + "learning_rate": 5.5960584844236565e-06, + "loss": 0.6039, + "step": 2683 + }, + { + "epoch": 0.6560743094597897, + "grad_norm": 2.8662421703338623, + "learning_rate": 5.588951659435861e-06, + "loss": 0.6255, + "step": 2684 + }, + { + "epoch": 0.6563187484722561, + "grad_norm": 3.419118642807007, + "learning_rate": 5.581847599467658e-06, + "loss": 0.7289, + "step": 2685 + }, + { + "epoch": 0.6565631874847225, + "grad_norm": 3.948251485824585, + "learning_rate": 5.574746308972153e-06, + "loss": 0.5889, + "step": 2686 + }, + { + "epoch": 0.6568076264971889, + "grad_norm": 2.250866174697876, + "learning_rate": 5.5676477924007165e-06, + "loss": 1.9108, + "step": 2687 + }, + { + "epoch": 0.6570520655096553, + "grad_norm": 3.220052719116211, + "learning_rate": 5.5605520542029776e-06, + "loss": 0.6972, + "step": 2688 + }, + { + "epoch": 0.6572965045221217, + "grad_norm": 3.2889387607574463, + "learning_rate": 5.553459098826828e-06, + "loss": 0.5567, + "step": 2689 + }, + { + "epoch": 0.6575409435345881, + "grad_norm": 3.8833532333374023, + "learning_rate": 5.546368930718411e-06, + "loss": 0.6254, + "step": 2690 + }, + { + "epoch": 0.6577853825470545, + "grad_norm": 4.197533130645752, + "learning_rate": 5.539281554322126e-06, + "loss": 0.7118, + "step": 2691 + }, + { + "epoch": 0.6580298215595209, + "grad_norm": 3.916963815689087, + "learning_rate": 5.532196974080619e-06, + "loss": 0.6149, + "step": 2692 + }, + { + "epoch": 0.6582742605719872, + "grad_norm": 2.631883144378662, + "learning_rate": 5.5251151944347825e-06, + "loss": 1.8987, + "step": 2693 + }, + { + "epoch": 0.6585186995844536, + "grad_norm": 3.2310924530029297, + "learning_rate": 5.518036219823765e-06, + "loss": 0.6472, + "step": 2694 + }, + { + "epoch": 0.65876313859692, + "grad_norm": 3.2843449115753174, + "learning_rate": 5.510960054684942e-06, + "loss": 0.7493, + "step": 2695 + }, + { + "epoch": 0.6590075776093864, + "grad_norm": 3.5885109901428223, + "learning_rate": 5.503886703453933e-06, + "loss": 0.5283, + "step": 2696 + }, + { + "epoch": 0.6592520166218528, + "grad_norm": 4.44291877746582, + "learning_rate": 5.496816170564597e-06, + "loss": 0.6773, + "step": 2697 + }, + { + "epoch": 0.6594964556343192, + "grad_norm": 2.9985857009887695, + "learning_rate": 5.4897484604490225e-06, + "loss": 0.7804, + "step": 2698 + }, + { + "epoch": 0.6597408946467856, + "grad_norm": 2.802274227142334, + "learning_rate": 5.482683577537529e-06, + "loss": 0.6248, + "step": 2699 + }, + { + "epoch": 0.659985333659252, + "grad_norm": 3.206056594848633, + "learning_rate": 5.475621526258665e-06, + "loss": 0.6502, + "step": 2700 + }, + { + "epoch": 0.6602297726717185, + "grad_norm": 3.777484655380249, + "learning_rate": 5.468562311039205e-06, + "loss": 0.6926, + "step": 2701 + }, + { + "epoch": 0.6604742116841847, + "grad_norm": 3.5440683364868164, + "learning_rate": 5.461505936304144e-06, + "loss": 0.5646, + "step": 2702 + }, + { + "epoch": 0.6607186506966511, + "grad_norm": 3.01408314704895, + "learning_rate": 5.4544524064766944e-06, + "loss": 0.5474, + "step": 2703 + }, + { + "epoch": 0.6609630897091175, + "grad_norm": 4.224313735961914, + "learning_rate": 5.4474017259782866e-06, + "loss": 0.6911, + "step": 2704 + }, + { + "epoch": 0.661207528721584, + "grad_norm": 3.4431660175323486, + "learning_rate": 5.440353899228572e-06, + "loss": 0.6834, + "step": 2705 + }, + { + "epoch": 0.6614519677340504, + "grad_norm": 3.106137990951538, + "learning_rate": 5.4333089306454045e-06, + "loss": 0.5567, + "step": 2706 + }, + { + "epoch": 0.6616964067465168, + "grad_norm": 3.4715335369110107, + "learning_rate": 5.426266824644848e-06, + "loss": 0.7525, + "step": 2707 + }, + { + "epoch": 0.6619408457589832, + "grad_norm": 2.917834758758545, + "learning_rate": 5.419227585641174e-06, + "loss": 0.462, + "step": 2708 + }, + { + "epoch": 0.6621852847714496, + "grad_norm": 4.073702812194824, + "learning_rate": 5.412191218046856e-06, + "loss": 0.5448, + "step": 2709 + }, + { + "epoch": 0.662429723783916, + "grad_norm": 2.8708159923553467, + "learning_rate": 5.405157726272567e-06, + "loss": 0.6345, + "step": 2710 + }, + { + "epoch": 0.6626741627963822, + "grad_norm": 2.726679563522339, + "learning_rate": 5.3981271147271786e-06, + "loss": 0.7013, + "step": 2711 + }, + { + "epoch": 0.6629186018088487, + "grad_norm": 2.372623920440674, + "learning_rate": 5.391099387817756e-06, + "loss": 0.6221, + "step": 2712 + }, + { + "epoch": 0.663163040821315, + "grad_norm": 2.783414363861084, + "learning_rate": 5.384074549949554e-06, + "loss": 0.6144, + "step": 2713 + }, + { + "epoch": 0.6634074798337815, + "grad_norm": 3.0188844203948975, + "learning_rate": 5.377052605526023e-06, + "loss": 0.73, + "step": 2714 + }, + { + "epoch": 0.6636519188462479, + "grad_norm": 2.9153764247894287, + "learning_rate": 5.370033558948793e-06, + "loss": 0.5721, + "step": 2715 + }, + { + "epoch": 0.6638963578587143, + "grad_norm": 2.798769235610962, + "learning_rate": 5.363017414617682e-06, + "loss": 0.6387, + "step": 2716 + }, + { + "epoch": 0.6641407968711807, + "grad_norm": 2.9460997581481934, + "learning_rate": 5.356004176930683e-06, + "loss": 0.6611, + "step": 2717 + }, + { + "epoch": 0.6643852358836471, + "grad_norm": 2.4899098873138428, + "learning_rate": 5.348993850283974e-06, + "loss": 1.9264, + "step": 2718 + }, + { + "epoch": 0.6646296748961135, + "grad_norm": 3.263526439666748, + "learning_rate": 5.341986439071907e-06, + "loss": 0.6537, + "step": 2719 + }, + { + "epoch": 0.6648741139085798, + "grad_norm": 3.4318578243255615, + "learning_rate": 5.334981947686995e-06, + "loss": 0.6641, + "step": 2720 + }, + { + "epoch": 0.6651185529210462, + "grad_norm": 3.2904953956604004, + "learning_rate": 5.327980380519942e-06, + "loss": 0.7114, + "step": 2721 + }, + { + "epoch": 0.6653629919335126, + "grad_norm": 2.738497734069824, + "learning_rate": 5.320981741959603e-06, + "loss": 0.53, + "step": 2722 + }, + { + "epoch": 0.665607430945979, + "grad_norm": 6.3192267417907715, + "learning_rate": 5.313986036393e-06, + "loss": 0.6404, + "step": 2723 + }, + { + "epoch": 0.6658518699584454, + "grad_norm": 2.9573323726654053, + "learning_rate": 5.306993268205319e-06, + "loss": 0.5863, + "step": 2724 + }, + { + "epoch": 0.6660963089709118, + "grad_norm": 3.060486078262329, + "learning_rate": 5.300003441779904e-06, + "loss": 0.6007, + "step": 2725 + }, + { + "epoch": 0.6663407479833782, + "grad_norm": 3.0728816986083984, + "learning_rate": 5.293016561498255e-06, + "loss": 0.7026, + "step": 2726 + }, + { + "epoch": 0.6665851869958446, + "grad_norm": 2.4286727905273438, + "learning_rate": 5.286032631740023e-06, + "loss": 1.9812, + "step": 2727 + }, + { + "epoch": 0.666829626008311, + "grad_norm": 2.3266477584838867, + "learning_rate": 5.279051656883014e-06, + "loss": 1.9323, + "step": 2728 + }, + { + "epoch": 0.6670740650207773, + "grad_norm": 4.493717670440674, + "learning_rate": 5.272073641303177e-06, + "loss": 0.7852, + "step": 2729 + }, + { + "epoch": 0.6673185040332437, + "grad_norm": 5.459487438201904, + "learning_rate": 5.26509858937461e-06, + "loss": 0.7235, + "step": 2730 + }, + { + "epoch": 0.6675629430457101, + "grad_norm": 3.1351940631866455, + "learning_rate": 5.25812650546955e-06, + "loss": 0.5807, + "step": 2731 + }, + { + "epoch": 0.6678073820581765, + "grad_norm": 3.1465470790863037, + "learning_rate": 5.251157393958376e-06, + "loss": 0.6398, + "step": 2732 + }, + { + "epoch": 0.6680518210706429, + "grad_norm": 2.907792329788208, + "learning_rate": 5.244191259209602e-06, + "loss": 0.5835, + "step": 2733 + }, + { + "epoch": 0.6682962600831093, + "grad_norm": 2.9785854816436768, + "learning_rate": 5.237228105589877e-06, + "loss": 0.6426, + "step": 2734 + }, + { + "epoch": 0.6685406990955757, + "grad_norm": 3.418694257736206, + "learning_rate": 5.230267937463977e-06, + "loss": 0.6216, + "step": 2735 + }, + { + "epoch": 0.6687851381080421, + "grad_norm": 2.9922711849212646, + "learning_rate": 5.223310759194822e-06, + "loss": 0.7244, + "step": 2736 + }, + { + "epoch": 0.6690295771205085, + "grad_norm": 2.6697285175323486, + "learning_rate": 5.216356575143436e-06, + "loss": 0.5847, + "step": 2737 + }, + { + "epoch": 0.6692740161329748, + "grad_norm": 2.942906141281128, + "learning_rate": 5.209405389668982e-06, + "loss": 0.6909, + "step": 2738 + }, + { + "epoch": 0.6695184551454412, + "grad_norm": 3.1385350227355957, + "learning_rate": 5.202457207128736e-06, + "loss": 0.5378, + "step": 2739 + }, + { + "epoch": 0.6697628941579076, + "grad_norm": 3.10185170173645, + "learning_rate": 5.195512031878093e-06, + "loss": 0.5174, + "step": 2740 + }, + { + "epoch": 0.670007333170374, + "grad_norm": 3.0209953784942627, + "learning_rate": 5.188569868270566e-06, + "loss": 0.6357, + "step": 2741 + }, + { + "epoch": 0.6702517721828404, + "grad_norm": 3.6313841342926025, + "learning_rate": 5.181630720657776e-06, + "loss": 0.6426, + "step": 2742 + }, + { + "epoch": 0.6704962111953068, + "grad_norm": 3.3282594680786133, + "learning_rate": 5.1746945933894555e-06, + "loss": 0.6618, + "step": 2743 + }, + { + "epoch": 0.6707406502077732, + "grad_norm": 2.451758623123169, + "learning_rate": 5.167761490813444e-06, + "loss": 0.5011, + "step": 2744 + }, + { + "epoch": 0.6709850892202396, + "grad_norm": 3.1366803646087646, + "learning_rate": 5.160831417275683e-06, + "loss": 0.636, + "step": 2745 + }, + { + "epoch": 0.671229528232706, + "grad_norm": 2.826927661895752, + "learning_rate": 5.153904377120218e-06, + "loss": 0.5909, + "step": 2746 + }, + { + "epoch": 0.6714739672451723, + "grad_norm": 2.7017712593078613, + "learning_rate": 5.146980374689192e-06, + "loss": 0.5149, + "step": 2747 + }, + { + "epoch": 0.6717184062576387, + "grad_norm": 2.9147655963897705, + "learning_rate": 5.140059414322842e-06, + "loss": 0.6561, + "step": 2748 + }, + { + "epoch": 0.6719628452701051, + "grad_norm": 3.0260162353515625, + "learning_rate": 5.133141500359502e-06, + "loss": 0.672, + "step": 2749 + }, + { + "epoch": 0.6722072842825715, + "grad_norm": 3.405888557434082, + "learning_rate": 5.1262266371355914e-06, + "loss": 0.6075, + "step": 2750 + }, + { + "epoch": 0.6724517232950379, + "grad_norm": 3.4034671783447266, + "learning_rate": 5.11931482898562e-06, + "loss": 0.7385, + "step": 2751 + }, + { + "epoch": 0.6726961623075043, + "grad_norm": 3.760180950164795, + "learning_rate": 5.112406080242186e-06, + "loss": 0.6619, + "step": 2752 + }, + { + "epoch": 0.6729406013199707, + "grad_norm": 3.1953327655792236, + "learning_rate": 5.105500395235965e-06, + "loss": 0.6422, + "step": 2753 + }, + { + "epoch": 0.6731850403324371, + "grad_norm": 3.0213558673858643, + "learning_rate": 5.098597778295713e-06, + "loss": 0.51, + "step": 2754 + }, + { + "epoch": 0.6734294793449035, + "grad_norm": 3.2140281200408936, + "learning_rate": 5.091698233748264e-06, + "loss": 0.6254, + "step": 2755 + }, + { + "epoch": 0.6736739183573698, + "grad_norm": 2.8674018383026123, + "learning_rate": 5.0848017659185254e-06, + "loss": 0.6284, + "step": 2756 + }, + { + "epoch": 0.6739183573698362, + "grad_norm": 3.3768415451049805, + "learning_rate": 5.077908379129475e-06, + "loss": 0.6839, + "step": 2757 + }, + { + "epoch": 0.6741627963823026, + "grad_norm": 3.359588861465454, + "learning_rate": 5.071018077702161e-06, + "loss": 0.658, + "step": 2758 + }, + { + "epoch": 0.674407235394769, + "grad_norm": 2.669429063796997, + "learning_rate": 5.064130865955696e-06, + "loss": 0.5982, + "step": 2759 + }, + { + "epoch": 0.6746516744072354, + "grad_norm": 2.9091298580169678, + "learning_rate": 5.057246748207258e-06, + "loss": 0.5159, + "step": 2760 + }, + { + "epoch": 0.6748961134197018, + "grad_norm": 3.143890857696533, + "learning_rate": 5.050365728772084e-06, + "loss": 0.6629, + "step": 2761 + }, + { + "epoch": 0.6751405524321682, + "grad_norm": 3.617204189300537, + "learning_rate": 5.043487811963469e-06, + "loss": 0.7287, + "step": 2762 + }, + { + "epoch": 0.6753849914446346, + "grad_norm": 3.3309438228607178, + "learning_rate": 5.0366130020927625e-06, + "loss": 0.6314, + "step": 2763 + }, + { + "epoch": 0.675629430457101, + "grad_norm": 4.037367343902588, + "learning_rate": 5.02974130346937e-06, + "loss": 0.6331, + "step": 2764 + }, + { + "epoch": 0.6758738694695673, + "grad_norm": 2.586611747741699, + "learning_rate": 5.022872720400741e-06, + "loss": 0.5245, + "step": 2765 + }, + { + "epoch": 0.6761183084820337, + "grad_norm": 3.3242573738098145, + "learning_rate": 5.0160072571923725e-06, + "loss": 0.6345, + "step": 2766 + }, + { + "epoch": 0.6763627474945001, + "grad_norm": 2.9784135818481445, + "learning_rate": 5.009144918147817e-06, + "loss": 0.6177, + "step": 2767 + }, + { + "epoch": 0.6766071865069665, + "grad_norm": 3.950590133666992, + "learning_rate": 5.002285707568653e-06, + "loss": 0.5661, + "step": 2768 + }, + { + "epoch": 0.6768516255194329, + "grad_norm": 3.056154251098633, + "learning_rate": 4.995429629754509e-06, + "loss": 0.5672, + "step": 2769 + }, + { + "epoch": 0.6770960645318993, + "grad_norm": 2.899587392807007, + "learning_rate": 4.988576689003044e-06, + "loss": 0.5638, + "step": 2770 + }, + { + "epoch": 0.6773405035443657, + "grad_norm": 3.3658998012542725, + "learning_rate": 4.981726889609952e-06, + "loss": 0.6199, + "step": 2771 + }, + { + "epoch": 0.6775849425568321, + "grad_norm": 3.157362699508667, + "learning_rate": 4.974880235868959e-06, + "loss": 0.6285, + "step": 2772 + }, + { + "epoch": 0.6778293815692985, + "grad_norm": 3.2186131477355957, + "learning_rate": 4.968036732071818e-06, + "loss": 0.6394, + "step": 2773 + }, + { + "epoch": 0.6780738205817648, + "grad_norm": 3.415682077407837, + "learning_rate": 4.9611963825083085e-06, + "loss": 0.5819, + "step": 2774 + }, + { + "epoch": 0.6783182595942312, + "grad_norm": 3.1678550243377686, + "learning_rate": 4.954359191466232e-06, + "loss": 0.6095, + "step": 2775 + }, + { + "epoch": 0.6785626986066976, + "grad_norm": 3.0142929553985596, + "learning_rate": 4.947525163231408e-06, + "loss": 0.615, + "step": 2776 + }, + { + "epoch": 0.678807137619164, + "grad_norm": 2.398094415664673, + "learning_rate": 4.940694302087679e-06, + "loss": 1.9171, + "step": 2777 + }, + { + "epoch": 0.6790515766316304, + "grad_norm": 2.8858699798583984, + "learning_rate": 4.933866612316898e-06, + "loss": 0.6783, + "step": 2778 + }, + { + "epoch": 0.6792960156440968, + "grad_norm": 3.315502166748047, + "learning_rate": 4.9270420981989295e-06, + "loss": 0.6088, + "step": 2779 + }, + { + "epoch": 0.6795404546565632, + "grad_norm": 2.7693212032318115, + "learning_rate": 4.92022076401165e-06, + "loss": 0.6648, + "step": 2780 + }, + { + "epoch": 0.6797848936690296, + "grad_norm": 3.115082263946533, + "learning_rate": 4.913402614030944e-06, + "loss": 0.5976, + "step": 2781 + }, + { + "epoch": 0.680029332681496, + "grad_norm": 3.340075969696045, + "learning_rate": 4.9065876525306884e-06, + "loss": 0.7053, + "step": 2782 + }, + { + "epoch": 0.6802737716939623, + "grad_norm": 2.9971578121185303, + "learning_rate": 4.899775883782784e-06, + "loss": 0.6259, + "step": 2783 + }, + { + "epoch": 0.6805182107064287, + "grad_norm": 2.8590915203094482, + "learning_rate": 4.89296731205711e-06, + "loss": 0.6496, + "step": 2784 + }, + { + "epoch": 0.6807626497188951, + "grad_norm": 3.2005624771118164, + "learning_rate": 4.8861619416215475e-06, + "loss": 0.5618, + "step": 2785 + }, + { + "epoch": 0.6810070887313615, + "grad_norm": 5.067270278930664, + "learning_rate": 4.8793597767419755e-06, + "loss": 0.5195, + "step": 2786 + }, + { + "epoch": 0.6812515277438279, + "grad_norm": 2.8245551586151123, + "learning_rate": 4.872560821682256e-06, + "loss": 0.5953, + "step": 2787 + }, + { + "epoch": 0.6814959667562943, + "grad_norm": 2.979853630065918, + "learning_rate": 4.8657650807042465e-06, + "loss": 0.6692, + "step": 2788 + }, + { + "epoch": 0.6817404057687607, + "grad_norm": 3.309493064880371, + "learning_rate": 4.858972558067784e-06, + "loss": 0.5598, + "step": 2789 + }, + { + "epoch": 0.6819848447812271, + "grad_norm": 3.0199952125549316, + "learning_rate": 4.852183258030691e-06, + "loss": 0.6, + "step": 2790 + }, + { + "epoch": 0.6822292837936935, + "grad_norm": 3.136849880218506, + "learning_rate": 4.84539718484877e-06, + "loss": 0.5988, + "step": 2791 + }, + { + "epoch": 0.6824737228061598, + "grad_norm": 3.2955336570739746, + "learning_rate": 4.8386143427757995e-06, + "loss": 0.6135, + "step": 2792 + }, + { + "epoch": 0.6827181618186262, + "grad_norm": 4.007350921630859, + "learning_rate": 4.831834736063533e-06, + "loss": 0.4945, + "step": 2793 + }, + { + "epoch": 0.6829626008310926, + "grad_norm": 3.997565507888794, + "learning_rate": 4.825058368961696e-06, + "loss": 0.7764, + "step": 2794 + }, + { + "epoch": 0.683207039843559, + "grad_norm": 3.0956525802612305, + "learning_rate": 4.818285245717984e-06, + "loss": 0.6211, + "step": 2795 + }, + { + "epoch": 0.6834514788560254, + "grad_norm": 3.000917911529541, + "learning_rate": 4.811515370578059e-06, + "loss": 0.5148, + "step": 2796 + }, + { + "epoch": 0.6836959178684918, + "grad_norm": 3.562018394470215, + "learning_rate": 4.804748747785542e-06, + "loss": 0.5863, + "step": 2797 + }, + { + "epoch": 0.6839403568809582, + "grad_norm": 3.1512999534606934, + "learning_rate": 4.7979853815820275e-06, + "loss": 0.6411, + "step": 2798 + }, + { + "epoch": 0.6841847958934246, + "grad_norm": 3.354006290435791, + "learning_rate": 4.791225276207058e-06, + "loss": 0.5797, + "step": 2799 + }, + { + "epoch": 0.684429234905891, + "grad_norm": 2.866476535797119, + "learning_rate": 4.784468435898134e-06, + "loss": 0.6366, + "step": 2800 + }, + { + "epoch": 0.6846736739183573, + "grad_norm": 2.913511037826538, + "learning_rate": 4.77771486489071e-06, + "loss": 0.6318, + "step": 2801 + }, + { + "epoch": 0.6849181129308237, + "grad_norm": 2.909332036972046, + "learning_rate": 4.7709645674181925e-06, + "loss": 0.6117, + "step": 2802 + }, + { + "epoch": 0.6851625519432901, + "grad_norm": 2.971118450164795, + "learning_rate": 4.764217547711935e-06, + "loss": 0.6834, + "step": 2803 + }, + { + "epoch": 0.6854069909557565, + "grad_norm": 2.961076021194458, + "learning_rate": 4.757473810001235e-06, + "loss": 0.6081, + "step": 2804 + }, + { + "epoch": 0.6856514299682229, + "grad_norm": 3.154080867767334, + "learning_rate": 4.750733358513335e-06, + "loss": 0.6245, + "step": 2805 + }, + { + "epoch": 0.6858958689806893, + "grad_norm": 3.5119946002960205, + "learning_rate": 4.743996197473417e-06, + "loss": 0.6834, + "step": 2806 + }, + { + "epoch": 0.6861403079931557, + "grad_norm": 3.775698661804199, + "learning_rate": 4.7372623311045995e-06, + "loss": 0.6074, + "step": 2807 + }, + { + "epoch": 0.6863847470056221, + "grad_norm": 2.907153606414795, + "learning_rate": 4.730531763627937e-06, + "loss": 0.6108, + "step": 2808 + }, + { + "epoch": 0.6866291860180885, + "grad_norm": 2.9769084453582764, + "learning_rate": 4.723804499262414e-06, + "loss": 0.5833, + "step": 2809 + }, + { + "epoch": 0.6868736250305548, + "grad_norm": 4.168307781219482, + "learning_rate": 4.7170805422249474e-06, + "loss": 0.6263, + "step": 2810 + }, + { + "epoch": 0.6871180640430212, + "grad_norm": 3.363607883453369, + "learning_rate": 4.710359896730379e-06, + "loss": 0.4936, + "step": 2811 + }, + { + "epoch": 0.6873625030554876, + "grad_norm": 3.9064033031463623, + "learning_rate": 4.7036425669914756e-06, + "loss": 0.6862, + "step": 2812 + }, + { + "epoch": 0.687606942067954, + "grad_norm": 2.661390542984009, + "learning_rate": 4.696928557218919e-06, + "loss": 0.5281, + "step": 2813 + }, + { + "epoch": 0.6878513810804204, + "grad_norm": 3.045010566711426, + "learning_rate": 4.690217871621327e-06, + "loss": 0.5213, + "step": 2814 + }, + { + "epoch": 0.6880958200928868, + "grad_norm": 3.699244976043701, + "learning_rate": 4.683510514405219e-06, + "loss": 0.6995, + "step": 2815 + }, + { + "epoch": 0.6883402591053532, + "grad_norm": 3.2344930171966553, + "learning_rate": 4.676806489775028e-06, + "loss": 0.7577, + "step": 2816 + }, + { + "epoch": 0.6885846981178196, + "grad_norm": 2.8136589527130127, + "learning_rate": 4.670105801933106e-06, + "loss": 0.6066, + "step": 2817 + }, + { + "epoch": 0.688829137130286, + "grad_norm": 3.5094377994537354, + "learning_rate": 4.663408455079707e-06, + "loss": 0.6551, + "step": 2818 + }, + { + "epoch": 0.6890735761427523, + "grad_norm": 3.238950490951538, + "learning_rate": 4.6567144534129926e-06, + "loss": 0.655, + "step": 2819 + }, + { + "epoch": 0.6893180151552187, + "grad_norm": 2.555016040802002, + "learning_rate": 4.65002380112903e-06, + "loss": 1.9487, + "step": 2820 + }, + { + "epoch": 0.6895624541676851, + "grad_norm": 3.152003049850464, + "learning_rate": 4.643336502421783e-06, + "loss": 0.5787, + "step": 2821 + }, + { + "epoch": 0.6898068931801515, + "grad_norm": 2.7996885776519775, + "learning_rate": 4.636652561483116e-06, + "loss": 0.6433, + "step": 2822 + }, + { + "epoch": 0.6900513321926179, + "grad_norm": 4.192671775817871, + "learning_rate": 4.629971982502786e-06, + "loss": 0.5745, + "step": 2823 + }, + { + "epoch": 0.6902957712050843, + "grad_norm": 3.3025522232055664, + "learning_rate": 4.623294769668445e-06, + "loss": 0.6802, + "step": 2824 + }, + { + "epoch": 0.6905402102175507, + "grad_norm": 3.519629955291748, + "learning_rate": 4.616620927165635e-06, + "loss": 0.647, + "step": 2825 + }, + { + "epoch": 0.6907846492300171, + "grad_norm": 3.209341049194336, + "learning_rate": 4.6099504591777845e-06, + "loss": 0.6831, + "step": 2826 + }, + { + "epoch": 0.6910290882424835, + "grad_norm": 2.624837636947632, + "learning_rate": 4.603283369886205e-06, + "loss": 0.5011, + "step": 2827 + }, + { + "epoch": 0.6912735272549498, + "grad_norm": 3.348609685897827, + "learning_rate": 4.596619663470089e-06, + "loss": 0.6547, + "step": 2828 + }, + { + "epoch": 0.6915179662674162, + "grad_norm": 4.262418746948242, + "learning_rate": 4.5899593441065195e-06, + "loss": 0.6114, + "step": 2829 + }, + { + "epoch": 0.6917624052798826, + "grad_norm": 3.3939056396484375, + "learning_rate": 4.583302415970443e-06, + "loss": 0.6676, + "step": 2830 + }, + { + "epoch": 0.692006844292349, + "grad_norm": 2.8519091606140137, + "learning_rate": 4.576648883234686e-06, + "loss": 0.5483, + "step": 2831 + }, + { + "epoch": 0.6922512833048154, + "grad_norm": 3.068394184112549, + "learning_rate": 4.5699987500699465e-06, + "loss": 0.7529, + "step": 2832 + }, + { + "epoch": 0.6924957223172818, + "grad_norm": 3.2849552631378174, + "learning_rate": 4.563352020644791e-06, + "loss": 0.7754, + "step": 2833 + }, + { + "epoch": 0.6927401613297482, + "grad_norm": 3.2050352096557617, + "learning_rate": 4.556708699125652e-06, + "loss": 0.5716, + "step": 2834 + }, + { + "epoch": 0.6929846003422147, + "grad_norm": 2.885378360748291, + "learning_rate": 4.550068789676826e-06, + "loss": 0.6766, + "step": 2835 + }, + { + "epoch": 0.693229039354681, + "grad_norm": 3.097829818725586, + "learning_rate": 4.5434322964604715e-06, + "loss": 0.6388, + "step": 2836 + }, + { + "epoch": 0.6934734783671473, + "grad_norm": 3.3307607173919678, + "learning_rate": 4.536799223636604e-06, + "loss": 0.6683, + "step": 2837 + }, + { + "epoch": 0.6937179173796137, + "grad_norm": 3.108682155609131, + "learning_rate": 4.530169575363098e-06, + "loss": 0.5499, + "step": 2838 + }, + { + "epoch": 0.6939623563920801, + "grad_norm": 2.598700761795044, + "learning_rate": 4.523543355795676e-06, + "loss": 0.5555, + "step": 2839 + }, + { + "epoch": 0.6942067954045466, + "grad_norm": 2.4821794033050537, + "learning_rate": 4.516920569087918e-06, + "loss": 0.535, + "step": 2840 + }, + { + "epoch": 0.694451234417013, + "grad_norm": 2.8600430488586426, + "learning_rate": 4.510301219391245e-06, + "loss": 0.6832, + "step": 2841 + }, + { + "epoch": 0.6946956734294794, + "grad_norm": 3.905129909515381, + "learning_rate": 4.503685310854931e-06, + "loss": 0.7206, + "step": 2842 + }, + { + "epoch": 0.6949401124419458, + "grad_norm": 2.983491897583008, + "learning_rate": 4.497072847626087e-06, + "loss": 0.6094, + "step": 2843 + }, + { + "epoch": 0.6951845514544122, + "grad_norm": 2.955796003341675, + "learning_rate": 4.490463833849669e-06, + "loss": 0.5527, + "step": 2844 + }, + { + "epoch": 0.6954289904668786, + "grad_norm": 3.2205801010131836, + "learning_rate": 4.483858273668464e-06, + "loss": 0.6138, + "step": 2845 + }, + { + "epoch": 0.6956734294793449, + "grad_norm": 2.5980865955352783, + "learning_rate": 4.477256171223103e-06, + "loss": 1.8764, + "step": 2846 + }, + { + "epoch": 0.6959178684918113, + "grad_norm": 3.737715482711792, + "learning_rate": 4.470657530652042e-06, + "loss": 0.6366, + "step": 2847 + }, + { + "epoch": 0.6961623075042777, + "grad_norm": 2.8147330284118652, + "learning_rate": 4.464062356091571e-06, + "loss": 0.4609, + "step": 2848 + }, + { + "epoch": 0.6964067465167441, + "grad_norm": 2.979423999786377, + "learning_rate": 4.457470651675806e-06, + "loss": 0.5153, + "step": 2849 + }, + { + "epoch": 0.6966511855292105, + "grad_norm": 3.7111847400665283, + "learning_rate": 4.450882421536688e-06, + "loss": 0.7031, + "step": 2850 + }, + { + "epoch": 0.6968956245416769, + "grad_norm": 3.0900397300720215, + "learning_rate": 4.444297669803981e-06, + "loss": 0.7322, + "step": 2851 + }, + { + "epoch": 0.6971400635541433, + "grad_norm": 3.054860830307007, + "learning_rate": 4.437716400605263e-06, + "loss": 0.5775, + "step": 2852 + }, + { + "epoch": 0.6973845025666097, + "grad_norm": 3.381854772567749, + "learning_rate": 4.431138618065942e-06, + "loss": 0.7313, + "step": 2853 + }, + { + "epoch": 0.6976289415790761, + "grad_norm": 7.616814613342285, + "learning_rate": 4.424564326309228e-06, + "loss": 0.5338, + "step": 2854 + }, + { + "epoch": 0.6978733805915424, + "grad_norm": 2.912473440170288, + "learning_rate": 4.417993529456148e-06, + "loss": 0.536, + "step": 2855 + }, + { + "epoch": 0.6981178196040088, + "grad_norm": 2.8949167728424072, + "learning_rate": 4.411426231625537e-06, + "loss": 0.6394, + "step": 2856 + }, + { + "epoch": 0.6983622586164752, + "grad_norm": 3.7534406185150146, + "learning_rate": 4.404862436934037e-06, + "loss": 0.6242, + "step": 2857 + }, + { + "epoch": 0.6986066976289416, + "grad_norm": 2.8737173080444336, + "learning_rate": 4.3983021494960944e-06, + "loss": 0.6676, + "step": 2858 + }, + { + "epoch": 0.698851136641408, + "grad_norm": 2.802374839782715, + "learning_rate": 4.391745373423957e-06, + "loss": 0.6221, + "step": 2859 + }, + { + "epoch": 0.6990955756538744, + "grad_norm": 2.5841798782348633, + "learning_rate": 4.38519211282767e-06, + "loss": 0.5938, + "step": 2860 + }, + { + "epoch": 0.6993400146663408, + "grad_norm": 3.130535364151001, + "learning_rate": 4.378642371815078e-06, + "loss": 0.5083, + "step": 2861 + }, + { + "epoch": 0.6995844536788072, + "grad_norm": 2.743894338607788, + "learning_rate": 4.372096154491814e-06, + "loss": 0.6127, + "step": 2862 + }, + { + "epoch": 0.6998288926912736, + "grad_norm": 4.366552352905273, + "learning_rate": 4.365553464961313e-06, + "loss": 0.6796, + "step": 2863 + }, + { + "epoch": 0.7000733317037399, + "grad_norm": 3.4443304538726807, + "learning_rate": 4.359014307324787e-06, + "loss": 0.5831, + "step": 2864 + }, + { + "epoch": 0.7003177707162063, + "grad_norm": 3.2634642124176025, + "learning_rate": 4.3524786856812406e-06, + "loss": 0.6987, + "step": 2865 + }, + { + "epoch": 0.7005622097286727, + "grad_norm": 5.027645587921143, + "learning_rate": 4.3459466041274575e-06, + "loss": 0.6459, + "step": 2866 + }, + { + "epoch": 0.7008066487411391, + "grad_norm": 3.121417284011841, + "learning_rate": 4.339418066758008e-06, + "loss": 0.6671, + "step": 2867 + }, + { + "epoch": 0.7010510877536055, + "grad_norm": 3.2320797443389893, + "learning_rate": 4.332893077665234e-06, + "loss": 0.6006, + "step": 2868 + }, + { + "epoch": 0.7012955267660719, + "grad_norm": 3.52091646194458, + "learning_rate": 4.32637164093926e-06, + "loss": 0.7388, + "step": 2869 + }, + { + "epoch": 0.7015399657785383, + "grad_norm": 2.5063316822052, + "learning_rate": 4.319853760667978e-06, + "loss": 1.8042, + "step": 2870 + }, + { + "epoch": 0.7017844047910047, + "grad_norm": 2.862105131149292, + "learning_rate": 4.313339440937055e-06, + "loss": 0.51, + "step": 2871 + }, + { + "epoch": 0.7020288438034711, + "grad_norm": 2.7071707248687744, + "learning_rate": 4.3068286858299235e-06, + "loss": 0.609, + "step": 2872 + }, + { + "epoch": 0.7022732828159374, + "grad_norm": 3.4979820251464844, + "learning_rate": 4.300321499427782e-06, + "loss": 0.5811, + "step": 2873 + }, + { + "epoch": 0.7025177218284038, + "grad_norm": 2.958627223968506, + "learning_rate": 4.293817885809594e-06, + "loss": 0.5751, + "step": 2874 + }, + { + "epoch": 0.7027621608408702, + "grad_norm": 3.80709171295166, + "learning_rate": 4.287317849052075e-06, + "loss": 0.6346, + "step": 2875 + }, + { + "epoch": 0.7030065998533366, + "grad_norm": 3.485647678375244, + "learning_rate": 4.280821393229716e-06, + "loss": 0.5961, + "step": 2876 + }, + { + "epoch": 0.703251038865803, + "grad_norm": 3.2261605262756348, + "learning_rate": 4.274328522414747e-06, + "loss": 0.5644, + "step": 2877 + }, + { + "epoch": 0.7034954778782694, + "grad_norm": 2.664689779281616, + "learning_rate": 4.2678392406771555e-06, + "loss": 0.4838, + "step": 2878 + }, + { + "epoch": 0.7037399168907358, + "grad_norm": 3.4998104572296143, + "learning_rate": 4.261353552084682e-06, + "loss": 0.5976, + "step": 2879 + }, + { + "epoch": 0.7039843559032022, + "grad_norm": 3.702908515930176, + "learning_rate": 4.254871460702811e-06, + "loss": 0.6867, + "step": 2880 + }, + { + "epoch": 0.7042287949156686, + "grad_norm": 3.084543466567993, + "learning_rate": 4.248392970594774e-06, + "loss": 0.6221, + "step": 2881 + }, + { + "epoch": 0.7044732339281349, + "grad_norm": 3.0470752716064453, + "learning_rate": 4.241918085821547e-06, + "loss": 0.6252, + "step": 2882 + }, + { + "epoch": 0.7047176729406013, + "grad_norm": 2.600609540939331, + "learning_rate": 4.2354468104418415e-06, + "loss": 1.8867, + "step": 2883 + }, + { + "epoch": 0.7049621119530677, + "grad_norm": 4.214003562927246, + "learning_rate": 4.228979148512109e-06, + "loss": 0.7683, + "step": 2884 + }, + { + "epoch": 0.7052065509655341, + "grad_norm": 6.833065032958984, + "learning_rate": 4.2225151040865375e-06, + "loss": 0.659, + "step": 2885 + }, + { + "epoch": 0.7054509899780005, + "grad_norm": 3.0870535373687744, + "learning_rate": 4.216054681217044e-06, + "loss": 0.5151, + "step": 2886 + }, + { + "epoch": 0.7056954289904669, + "grad_norm": 2.8286168575286865, + "learning_rate": 4.209597883953278e-06, + "loss": 0.4955, + "step": 2887 + }, + { + "epoch": 0.7059398680029333, + "grad_norm": 2.5923023223876953, + "learning_rate": 4.203144716342617e-06, + "loss": 1.9329, + "step": 2888 + }, + { + "epoch": 0.7061843070153997, + "grad_norm": 3.287935733795166, + "learning_rate": 4.196695182430157e-06, + "loss": 0.5925, + "step": 2889 + }, + { + "epoch": 0.7064287460278661, + "grad_norm": 3.587629556655884, + "learning_rate": 4.190249286258723e-06, + "loss": 0.7135, + "step": 2890 + }, + { + "epoch": 0.7066731850403324, + "grad_norm": 2.9305241107940674, + "learning_rate": 4.18380703186886e-06, + "loss": 0.5792, + "step": 2891 + }, + { + "epoch": 0.7069176240527988, + "grad_norm": 3.006047248840332, + "learning_rate": 4.1773684232988285e-06, + "loss": 0.5638, + "step": 2892 + }, + { + "epoch": 0.7071620630652652, + "grad_norm": 2.8415372371673584, + "learning_rate": 4.170933464584601e-06, + "loss": 0.5779, + "step": 2893 + }, + { + "epoch": 0.7074065020777316, + "grad_norm": 3.2072343826293945, + "learning_rate": 4.1645021597598635e-06, + "loss": 0.6026, + "step": 2894 + }, + { + "epoch": 0.707650941090198, + "grad_norm": 2.8006460666656494, + "learning_rate": 4.158074512856014e-06, + "loss": 0.5016, + "step": 2895 + }, + { + "epoch": 0.7078953801026644, + "grad_norm": 3.4516077041625977, + "learning_rate": 4.151650527902155e-06, + "loss": 0.6544, + "step": 2896 + }, + { + "epoch": 0.7081398191151308, + "grad_norm": 2.9812870025634766, + "learning_rate": 4.145230208925094e-06, + "loss": 0.6984, + "step": 2897 + }, + { + "epoch": 0.7083842581275972, + "grad_norm": 11.79186725616455, + "learning_rate": 4.138813559949341e-06, + "loss": 0.7448, + "step": 2898 + }, + { + "epoch": 0.7086286971400636, + "grad_norm": 3.4640963077545166, + "learning_rate": 4.132400584997106e-06, + "loss": 0.6699, + "step": 2899 + }, + { + "epoch": 0.7088731361525299, + "grad_norm": 2.5082526206970215, + "learning_rate": 4.125991288088294e-06, + "loss": 0.5293, + "step": 2900 + }, + { + "epoch": 0.7091175751649963, + "grad_norm": 2.9462575912475586, + "learning_rate": 4.1195856732405094e-06, + "loss": 0.6258, + "step": 2901 + }, + { + "epoch": 0.7093620141774627, + "grad_norm": 3.4112930297851562, + "learning_rate": 4.1131837444690404e-06, + "loss": 0.6114, + "step": 2902 + }, + { + "epoch": 0.7096064531899291, + "grad_norm": 2.7927472591400146, + "learning_rate": 4.106785505786872e-06, + "loss": 0.4668, + "step": 2903 + }, + { + "epoch": 0.7098508922023955, + "grad_norm": 3.160207748413086, + "learning_rate": 4.100390961204672e-06, + "loss": 0.5719, + "step": 2904 + }, + { + "epoch": 0.7100953312148619, + "grad_norm": 2.6186273097991943, + "learning_rate": 4.094000114730795e-06, + "loss": 0.5201, + "step": 2905 + }, + { + "epoch": 0.7103397702273283, + "grad_norm": 3.6046929359436035, + "learning_rate": 4.087612970371273e-06, + "loss": 0.6772, + "step": 2906 + }, + { + "epoch": 0.7105842092397947, + "grad_norm": 3.1790590286254883, + "learning_rate": 4.081229532129826e-06, + "loss": 0.7094, + "step": 2907 + }, + { + "epoch": 0.7108286482522611, + "grad_norm": 3.196411371231079, + "learning_rate": 4.0748498040078446e-06, + "loss": 0.6551, + "step": 2908 + }, + { + "epoch": 0.7110730872647274, + "grad_norm": 3.279223918914795, + "learning_rate": 4.068473790004393e-06, + "loss": 0.6332, + "step": 2909 + }, + { + "epoch": 0.7113175262771938, + "grad_norm": 3.0042917728424072, + "learning_rate": 4.062101494116209e-06, + "loss": 0.5151, + "step": 2910 + }, + { + "epoch": 0.7115619652896602, + "grad_norm": 2.7973415851593018, + "learning_rate": 4.055732920337699e-06, + "loss": 0.5038, + "step": 2911 + }, + { + "epoch": 0.7118064043021266, + "grad_norm": 3.4121341705322266, + "learning_rate": 4.04936807266094e-06, + "loss": 0.5919, + "step": 2912 + }, + { + "epoch": 0.712050843314593, + "grad_norm": 3.2722136974334717, + "learning_rate": 4.043006955075667e-06, + "loss": 0.69, + "step": 2913 + }, + { + "epoch": 0.7122952823270594, + "grad_norm": 3.4652504920959473, + "learning_rate": 4.036649571569282e-06, + "loss": 0.6469, + "step": 2914 + }, + { + "epoch": 0.7125397213395258, + "grad_norm": 3.436591863632202, + "learning_rate": 4.030295926126845e-06, + "loss": 0.6366, + "step": 2915 + }, + { + "epoch": 0.7127841603519922, + "grad_norm": 3.649648666381836, + "learning_rate": 4.02394602273107e-06, + "loss": 0.5092, + "step": 2916 + }, + { + "epoch": 0.7130285993644586, + "grad_norm": 3.596709966659546, + "learning_rate": 4.0175998653623295e-06, + "loss": 0.5792, + "step": 2917 + }, + { + "epoch": 0.7132730383769249, + "grad_norm": 3.119182586669922, + "learning_rate": 4.011257457998645e-06, + "loss": 0.5226, + "step": 2918 + }, + { + "epoch": 0.7135174773893913, + "grad_norm": 3.1880149841308594, + "learning_rate": 4.004918804615689e-06, + "loss": 0.5484, + "step": 2919 + }, + { + "epoch": 0.7137619164018577, + "grad_norm": 2.373765230178833, + "learning_rate": 3.998583909186781e-06, + "loss": 1.9726, + "step": 2920 + }, + { + "epoch": 0.7140063554143241, + "grad_norm": 2.5934154987335205, + "learning_rate": 3.992252775682877e-06, + "loss": 0.5649, + "step": 2921 + }, + { + "epoch": 0.7142507944267905, + "grad_norm": 3.5194554328918457, + "learning_rate": 3.985925408072594e-06, + "loss": 0.6557, + "step": 2922 + }, + { + "epoch": 0.7144952334392569, + "grad_norm": 3.8668274879455566, + "learning_rate": 3.979601810322169e-06, + "loss": 0.5971, + "step": 2923 + }, + { + "epoch": 0.7147396724517233, + "grad_norm": 3.060781240463257, + "learning_rate": 3.973281986395485e-06, + "loss": 0.7926, + "step": 2924 + }, + { + "epoch": 0.7149841114641897, + "grad_norm": 3.3815793991088867, + "learning_rate": 3.966965940254057e-06, + "loss": 0.6712, + "step": 2925 + }, + { + "epoch": 0.7152285504766561, + "grad_norm": 3.3370070457458496, + "learning_rate": 3.960653675857033e-06, + "loss": 0.6374, + "step": 2926 + }, + { + "epoch": 0.7154729894891224, + "grad_norm": 3.052048444747925, + "learning_rate": 3.954345197161188e-06, + "loss": 0.6032, + "step": 2927 + }, + { + "epoch": 0.7157174285015888, + "grad_norm": 2.9592669010162354, + "learning_rate": 3.948040508120929e-06, + "loss": 0.5775, + "step": 2928 + }, + { + "epoch": 0.7159618675140552, + "grad_norm": 2.4977145195007324, + "learning_rate": 3.941739612688284e-06, + "loss": 0.5427, + "step": 2929 + }, + { + "epoch": 0.7162063065265216, + "grad_norm": 2.9116251468658447, + "learning_rate": 3.935442514812901e-06, + "loss": 0.6394, + "step": 2930 + }, + { + "epoch": 0.716450745538988, + "grad_norm": 2.811952590942383, + "learning_rate": 3.929149218442052e-06, + "loss": 0.5343, + "step": 2931 + }, + { + "epoch": 0.7166951845514544, + "grad_norm": 3.4379796981811523, + "learning_rate": 3.922859727520625e-06, + "loss": 0.6813, + "step": 2932 + }, + { + "epoch": 0.7169396235639208, + "grad_norm": 3.4804670810699463, + "learning_rate": 3.91657404599112e-06, + "loss": 0.624, + "step": 2933 + }, + { + "epoch": 0.7171840625763872, + "grad_norm": 2.6202759742736816, + "learning_rate": 3.910292177793651e-06, + "loss": 0.5678, + "step": 2934 + }, + { + "epoch": 0.7174285015888536, + "grad_norm": 3.9853694438934326, + "learning_rate": 3.904014126865943e-06, + "loss": 0.6223, + "step": 2935 + }, + { + "epoch": 0.7176729406013199, + "grad_norm": 3.015868902206421, + "learning_rate": 3.897739897143324e-06, + "loss": 0.5786, + "step": 2936 + }, + { + "epoch": 0.7179173796137863, + "grad_norm": 2.8636083602905273, + "learning_rate": 3.891469492558728e-06, + "loss": 0.6114, + "step": 2937 + }, + { + "epoch": 0.7181618186262527, + "grad_norm": 3.039456844329834, + "learning_rate": 3.8852029170427e-06, + "loss": 0.6155, + "step": 2938 + }, + { + "epoch": 0.7184062576387191, + "grad_norm": 2.8267924785614014, + "learning_rate": 3.878940174523371e-06, + "loss": 0.6414, + "step": 2939 + }, + { + "epoch": 0.7186506966511855, + "grad_norm": 2.8125078678131104, + "learning_rate": 3.872681268926479e-06, + "loss": 0.6549, + "step": 2940 + }, + { + "epoch": 0.7188951356636519, + "grad_norm": 3.337916135787964, + "learning_rate": 3.866426204175353e-06, + "loss": 0.5881, + "step": 2941 + }, + { + "epoch": 0.7191395746761183, + "grad_norm": 2.848170757293701, + "learning_rate": 3.860174984190914e-06, + "loss": 0.6312, + "step": 2942 + }, + { + "epoch": 0.7193840136885847, + "grad_norm": 2.944838047027588, + "learning_rate": 3.8539276128916736e-06, + "loss": 0.6264, + "step": 2943 + }, + { + "epoch": 0.7196284527010511, + "grad_norm": 3.2724449634552, + "learning_rate": 3.847684094193733e-06, + "loss": 0.6204, + "step": 2944 + }, + { + "epoch": 0.7198728917135174, + "grad_norm": 3.8080027103424072, + "learning_rate": 3.841444432010775e-06, + "loss": 0.625, + "step": 2945 + }, + { + "epoch": 0.7201173307259838, + "grad_norm": 3.054201602935791, + "learning_rate": 3.835208630254068e-06, + "loss": 0.7294, + "step": 2946 + }, + { + "epoch": 0.7203617697384502, + "grad_norm": 3.6399037837982178, + "learning_rate": 3.828976692832458e-06, + "loss": 0.5939, + "step": 2947 + }, + { + "epoch": 0.7206062087509166, + "grad_norm": 3.149604320526123, + "learning_rate": 3.822748623652372e-06, + "loss": 0.6319, + "step": 2948 + }, + { + "epoch": 0.720850647763383, + "grad_norm": 3.125087261199951, + "learning_rate": 3.816524426617807e-06, + "loss": 0.5891, + "step": 2949 + }, + { + "epoch": 0.7210950867758494, + "grad_norm": 3.2292184829711914, + "learning_rate": 3.810304105630338e-06, + "loss": 0.561, + "step": 2950 + }, + { + "epoch": 0.7213395257883158, + "grad_norm": 2.8233754634857178, + "learning_rate": 3.804087664589108e-06, + "loss": 0.4755, + "step": 2951 + }, + { + "epoch": 0.7215839648007822, + "grad_norm": 3.1551756858825684, + "learning_rate": 3.7978751073908237e-06, + "loss": 0.6226, + "step": 2952 + }, + { + "epoch": 0.7218284038132486, + "grad_norm": 4.270998001098633, + "learning_rate": 3.791666437929771e-06, + "loss": 0.5689, + "step": 2953 + }, + { + "epoch": 0.7220728428257149, + "grad_norm": 3.086409568786621, + "learning_rate": 3.7854616600977834e-06, + "loss": 0.669, + "step": 2954 + }, + { + "epoch": 0.7223172818381813, + "grad_norm": 3.408801555633545, + "learning_rate": 3.7792607777842626e-06, + "loss": 0.6059, + "step": 2955 + }, + { + "epoch": 0.7225617208506477, + "grad_norm": 3.2890207767486572, + "learning_rate": 3.773063794876167e-06, + "loss": 0.7136, + "step": 2956 + }, + { + "epoch": 0.7228061598631141, + "grad_norm": 3.042243242263794, + "learning_rate": 3.7668707152580097e-06, + "loss": 0.5738, + "step": 2957 + }, + { + "epoch": 0.7230505988755805, + "grad_norm": 2.7547099590301514, + "learning_rate": 3.760681542811859e-06, + "loss": 0.5578, + "step": 2958 + }, + { + "epoch": 0.7232950378880469, + "grad_norm": 3.3390941619873047, + "learning_rate": 3.754496281417331e-06, + "loss": 0.6282, + "step": 2959 + }, + { + "epoch": 0.7235394769005133, + "grad_norm": 3.3198680877685547, + "learning_rate": 3.7483149349515957e-06, + "loss": 0.5808, + "step": 2960 + }, + { + "epoch": 0.7237839159129797, + "grad_norm": 2.6300973892211914, + "learning_rate": 3.742137507289363e-06, + "loss": 0.538, + "step": 2961 + }, + { + "epoch": 0.7240283549254461, + "grad_norm": 3.129347085952759, + "learning_rate": 3.7359640023028888e-06, + "loss": 0.6722, + "step": 2962 + }, + { + "epoch": 0.7242727939379124, + "grad_norm": 2.856234550476074, + "learning_rate": 3.729794423861971e-06, + "loss": 0.6394, + "step": 2963 + }, + { + "epoch": 0.7245172329503788, + "grad_norm": 3.1484363079071045, + "learning_rate": 3.7236287758339464e-06, + "loss": 0.6045, + "step": 2964 + }, + { + "epoch": 0.7247616719628452, + "grad_norm": 3.6585779190063477, + "learning_rate": 3.717467062083685e-06, + "loss": 0.5096, + "step": 2965 + }, + { + "epoch": 0.7250061109753116, + "grad_norm": 3.5617711544036865, + "learning_rate": 3.7113092864735943e-06, + "loss": 0.6079, + "step": 2966 + }, + { + "epoch": 0.725250549987778, + "grad_norm": 3.1096692085266113, + "learning_rate": 3.7051554528636114e-06, + "loss": 0.6274, + "step": 2967 + }, + { + "epoch": 0.7254949890002445, + "grad_norm": 3.7432024478912354, + "learning_rate": 3.699005565111198e-06, + "loss": 0.6443, + "step": 2968 + }, + { + "epoch": 0.7257394280127109, + "grad_norm": 3.277007579803467, + "learning_rate": 3.6928596270713558e-06, + "loss": 0.6341, + "step": 2969 + }, + { + "epoch": 0.7259838670251773, + "grad_norm": 2.8448712825775146, + "learning_rate": 3.6867176425965967e-06, + "loss": 0.6601, + "step": 2970 + }, + { + "epoch": 0.7262283060376437, + "grad_norm": 2.983934164047241, + "learning_rate": 3.680579615536961e-06, + "loss": 0.6822, + "step": 2971 + }, + { + "epoch": 0.72647274505011, + "grad_norm": 2.7468302249908447, + "learning_rate": 3.6744455497400055e-06, + "loss": 0.5229, + "step": 2972 + }, + { + "epoch": 0.7267171840625763, + "grad_norm": 3.0838937759399414, + "learning_rate": 3.6683154490508064e-06, + "loss": 0.7091, + "step": 2973 + }, + { + "epoch": 0.7269616230750428, + "grad_norm": 2.580299139022827, + "learning_rate": 3.662189317311953e-06, + "loss": 0.6071, + "step": 2974 + }, + { + "epoch": 0.7272060620875092, + "grad_norm": 3.2007927894592285, + "learning_rate": 3.6560671583635467e-06, + "loss": 0.6529, + "step": 2975 + }, + { + "epoch": 0.7274505010999756, + "grad_norm": 2.868992805480957, + "learning_rate": 3.6499489760431993e-06, + "loss": 1.8465, + "step": 2976 + }, + { + "epoch": 0.727694940112442, + "grad_norm": 3.0807933807373047, + "learning_rate": 3.643834774186029e-06, + "loss": 0.6774, + "step": 2977 + }, + { + "epoch": 0.7279393791249084, + "grad_norm": 3.0739243030548096, + "learning_rate": 3.6377245566246597e-06, + "loss": 0.5611, + "step": 2978 + }, + { + "epoch": 0.7281838181373748, + "grad_norm": 3.160921812057495, + "learning_rate": 3.631618327189218e-06, + "loss": 0.5293, + "step": 2979 + }, + { + "epoch": 0.7284282571498412, + "grad_norm": 7.900955677032471, + "learning_rate": 3.6255160897073293e-06, + "loss": 0.6552, + "step": 2980 + }, + { + "epoch": 0.7286726961623075, + "grad_norm": 2.859971046447754, + "learning_rate": 3.6194178480041174e-06, + "loss": 0.5905, + "step": 2981 + }, + { + "epoch": 0.7289171351747739, + "grad_norm": 3.280907154083252, + "learning_rate": 3.6133236059022016e-06, + "loss": 0.6404, + "step": 2982 + }, + { + "epoch": 0.7291615741872403, + "grad_norm": 4.005970478057861, + "learning_rate": 3.607233367221691e-06, + "loss": 0.6535, + "step": 2983 + }, + { + "epoch": 0.7294060131997067, + "grad_norm": 2.936605453491211, + "learning_rate": 3.601147135780194e-06, + "loss": 0.5508, + "step": 2984 + }, + { + "epoch": 0.7296504522121731, + "grad_norm": 2.9298524856567383, + "learning_rate": 3.5950649153927997e-06, + "loss": 0.5346, + "step": 2985 + }, + { + "epoch": 0.7298948912246395, + "grad_norm": 3.077277421951294, + "learning_rate": 3.588986709872082e-06, + "loss": 0.5944, + "step": 2986 + }, + { + "epoch": 0.7301393302371059, + "grad_norm": 3.2504830360412598, + "learning_rate": 3.5829125230281014e-06, + "loss": 0.6482, + "step": 2987 + }, + { + "epoch": 0.7303837692495723, + "grad_norm": 2.85105037689209, + "learning_rate": 3.5768423586684e-06, + "loss": 0.6188, + "step": 2988 + }, + { + "epoch": 0.7306282082620387, + "grad_norm": 2.931598663330078, + "learning_rate": 3.5707762205979946e-06, + "loss": 0.6613, + "step": 2989 + }, + { + "epoch": 0.730872647274505, + "grad_norm": 3.099318265914917, + "learning_rate": 3.564714112619383e-06, + "loss": 0.5855, + "step": 2990 + }, + { + "epoch": 0.7311170862869714, + "grad_norm": 2.8876891136169434, + "learning_rate": 3.558656038532532e-06, + "loss": 0.533, + "step": 2991 + }, + { + "epoch": 0.7313615252994378, + "grad_norm": 2.955615997314453, + "learning_rate": 3.5526020021348827e-06, + "loss": 0.5758, + "step": 2992 + }, + { + "epoch": 0.7316059643119042, + "grad_norm": 3.9695706367492676, + "learning_rate": 3.5465520072213455e-06, + "loss": 0.6492, + "step": 2993 + }, + { + "epoch": 0.7318504033243706, + "grad_norm": 3.853733777999878, + "learning_rate": 3.5405060575842953e-06, + "loss": 0.5827, + "step": 2994 + }, + { + "epoch": 0.732094842336837, + "grad_norm": 3.007627487182617, + "learning_rate": 3.5344641570135742e-06, + "loss": 0.672, + "step": 2995 + }, + { + "epoch": 0.7323392813493034, + "grad_norm": 3.5044028759002686, + "learning_rate": 3.5284263092964823e-06, + "loss": 0.668, + "step": 2996 + }, + { + "epoch": 0.7325837203617698, + "grad_norm": 3.6388285160064697, + "learning_rate": 3.5223925182177833e-06, + "loss": 0.6515, + "step": 2997 + }, + { + "epoch": 0.7328281593742362, + "grad_norm": 3.2376761436462402, + "learning_rate": 3.516362787559695e-06, + "loss": 0.5645, + "step": 2998 + }, + { + "epoch": 0.7330725983867025, + "grad_norm": 4.141454219818115, + "learning_rate": 3.510337121101888e-06, + "loss": 0.4661, + "step": 2999 + }, + { + "epoch": 0.7333170373991689, + "grad_norm": 3.248368740081787, + "learning_rate": 3.5043155226214954e-06, + "loss": 0.6578, + "step": 3000 + }, + { + "epoch": 0.7335614764116353, + "grad_norm": 2.848895788192749, + "learning_rate": 3.4982979958930896e-06, + "loss": 0.5574, + "step": 3001 + }, + { + "epoch": 0.7338059154241017, + "grad_norm": 3.017939805984497, + "learning_rate": 3.4922845446886944e-06, + "loss": 0.5379, + "step": 3002 + }, + { + "epoch": 0.7340503544365681, + "grad_norm": 3.1488208770751953, + "learning_rate": 3.48627517277778e-06, + "loss": 0.5726, + "step": 3003 + }, + { + "epoch": 0.7342947934490345, + "grad_norm": 3.541123151779175, + "learning_rate": 3.4802698839272564e-06, + "loss": 0.5333, + "step": 3004 + }, + { + "epoch": 0.7345392324615009, + "grad_norm": 3.4302146434783936, + "learning_rate": 3.4742686819014782e-06, + "loss": 0.5916, + "step": 3005 + }, + { + "epoch": 0.7347836714739673, + "grad_norm": 3.442814350128174, + "learning_rate": 3.468271570462235e-06, + "loss": 0.7797, + "step": 3006 + }, + { + "epoch": 0.7350281104864337, + "grad_norm": 3.1694416999816895, + "learning_rate": 3.462278553368753e-06, + "loss": 0.5926, + "step": 3007 + }, + { + "epoch": 0.7352725494989, + "grad_norm": 5.13250207901001, + "learning_rate": 3.4562896343776943e-06, + "loss": 0.7076, + "step": 3008 + }, + { + "epoch": 0.7355169885113664, + "grad_norm": 3.201740264892578, + "learning_rate": 3.4503048172431495e-06, + "loss": 0.5783, + "step": 3009 + }, + { + "epoch": 0.7357614275238328, + "grad_norm": 3.3980233669281006, + "learning_rate": 3.4443241057166342e-06, + "loss": 0.6619, + "step": 3010 + }, + { + "epoch": 0.7360058665362992, + "grad_norm": 3.5091662406921387, + "learning_rate": 3.4383475035471026e-06, + "loss": 0.5663, + "step": 3011 + }, + { + "epoch": 0.7362503055487656, + "grad_norm": 3.1590633392333984, + "learning_rate": 3.432375014480922e-06, + "loss": 0.5789, + "step": 3012 + }, + { + "epoch": 0.736494744561232, + "grad_norm": 2.7551000118255615, + "learning_rate": 3.426406642261885e-06, + "loss": 1.8702, + "step": 3013 + }, + { + "epoch": 0.7367391835736984, + "grad_norm": 2.7853903770446777, + "learning_rate": 3.4204423906312023e-06, + "loss": 0.5757, + "step": 3014 + }, + { + "epoch": 0.7369836225861648, + "grad_norm": 2.52852201461792, + "learning_rate": 3.4144822633275042e-06, + "loss": 1.845, + "step": 3015 + }, + { + "epoch": 0.7372280615986312, + "grad_norm": 3.035095691680908, + "learning_rate": 3.408526264086833e-06, + "loss": 0.4472, + "step": 3016 + }, + { + "epoch": 0.7374725006110975, + "grad_norm": 3.2289512157440186, + "learning_rate": 3.402574396642646e-06, + "loss": 0.5751, + "step": 3017 + }, + { + "epoch": 0.7377169396235639, + "grad_norm": 5.053186893463135, + "learning_rate": 3.3966266647258085e-06, + "loss": 0.5418, + "step": 3018 + }, + { + "epoch": 0.7379613786360303, + "grad_norm": 3.862535238265991, + "learning_rate": 3.3906830720645943e-06, + "loss": 0.452, + "step": 3019 + }, + { + "epoch": 0.7382058176484967, + "grad_norm": 6.835580348968506, + "learning_rate": 3.3847436223846788e-06, + "loss": 0.605, + "step": 3020 + }, + { + "epoch": 0.7384502566609631, + "grad_norm": 2.5224106311798096, + "learning_rate": 3.378808319409149e-06, + "loss": 0.5702, + "step": 3021 + }, + { + "epoch": 0.7386946956734295, + "grad_norm": 2.9877500534057617, + "learning_rate": 3.372877166858488e-06, + "loss": 0.6121, + "step": 3022 + }, + { + "epoch": 0.7389391346858959, + "grad_norm": 2.6402807235717773, + "learning_rate": 3.3669501684505734e-06, + "loss": 2.0057, + "step": 3023 + }, + { + "epoch": 0.7391835736983623, + "grad_norm": 3.599905490875244, + "learning_rate": 3.3610273279006842e-06, + "loss": 0.5901, + "step": 3024 + }, + { + "epoch": 0.7394280127108287, + "grad_norm": 3.2533516883850098, + "learning_rate": 3.3551086489214913e-06, + "loss": 0.5949, + "step": 3025 + }, + { + "epoch": 0.739672451723295, + "grad_norm": 3.3682477474212646, + "learning_rate": 3.349194135223056e-06, + "loss": 0.5901, + "step": 3026 + }, + { + "epoch": 0.7399168907357614, + "grad_norm": 2.728986978530884, + "learning_rate": 3.343283790512829e-06, + "loss": 0.699, + "step": 3027 + }, + { + "epoch": 0.7401613297482278, + "grad_norm": 2.769089460372925, + "learning_rate": 3.3373776184956485e-06, + "loss": 0.7134, + "step": 3028 + }, + { + "epoch": 0.7404057687606942, + "grad_norm": 2.7086477279663086, + "learning_rate": 3.331475622873738e-06, + "loss": 0.5526, + "step": 3029 + }, + { + "epoch": 0.7406502077731606, + "grad_norm": 3.8077518939971924, + "learning_rate": 3.3255778073467004e-06, + "loss": 0.5683, + "step": 3030 + }, + { + "epoch": 0.740894646785627, + "grad_norm": 3.039224147796631, + "learning_rate": 3.319684175611517e-06, + "loss": 0.5612, + "step": 3031 + }, + { + "epoch": 0.7411390857980934, + "grad_norm": 3.579586982727051, + "learning_rate": 3.3137947313625563e-06, + "loss": 0.6918, + "step": 3032 + }, + { + "epoch": 0.7413835248105598, + "grad_norm": 3.2064623832702637, + "learning_rate": 3.30790947829155e-06, + "loss": 0.5335, + "step": 3033 + }, + { + "epoch": 0.7416279638230262, + "grad_norm": 6.851931571960449, + "learning_rate": 3.3020284200876095e-06, + "loss": 0.5767, + "step": 3034 + }, + { + "epoch": 0.7418724028354925, + "grad_norm": 3.6696505546569824, + "learning_rate": 3.296151560437214e-06, + "loss": 0.6169, + "step": 3035 + }, + { + "epoch": 0.7421168418479589, + "grad_norm": 3.404021978378296, + "learning_rate": 3.290278903024211e-06, + "loss": 0.6475, + "step": 3036 + }, + { + "epoch": 0.7423612808604253, + "grad_norm": 3.687025308609009, + "learning_rate": 3.284410451529816e-06, + "loss": 0.6712, + "step": 3037 + }, + { + "epoch": 0.7426057198728917, + "grad_norm": 3.5287423133850098, + "learning_rate": 3.2785462096326047e-06, + "loss": 1.9579, + "step": 3038 + }, + { + "epoch": 0.7428501588853581, + "grad_norm": 2.978653907775879, + "learning_rate": 3.2726861810085163e-06, + "loss": 0.514, + "step": 3039 + }, + { + "epoch": 0.7430945978978245, + "grad_norm": 3.3357956409454346, + "learning_rate": 3.26683036933085e-06, + "loss": 0.5666, + "step": 3040 + }, + { + "epoch": 0.7433390369102909, + "grad_norm": 3.517115354537964, + "learning_rate": 3.2609787782702595e-06, + "loss": 0.6104, + "step": 3041 + }, + { + "epoch": 0.7435834759227573, + "grad_norm": 2.289231538772583, + "learning_rate": 3.255131411494753e-06, + "loss": 2.043, + "step": 3042 + }, + { + "epoch": 0.7438279149352237, + "grad_norm": 2.9872188568115234, + "learning_rate": 3.2492882726696907e-06, + "loss": 0.6355, + "step": 3043 + }, + { + "epoch": 0.74407235394769, + "grad_norm": 8.818699836730957, + "learning_rate": 3.243449365457785e-06, + "loss": 0.5859, + "step": 3044 + }, + { + "epoch": 0.7443167929601564, + "grad_norm": 2.9520797729492188, + "learning_rate": 3.237614693519089e-06, + "loss": 0.6683, + "step": 3045 + }, + { + "epoch": 0.7445612319726228, + "grad_norm": 3.3904762268066406, + "learning_rate": 3.2317842605110126e-06, + "loss": 0.7299, + "step": 3046 + }, + { + "epoch": 0.7448056709850892, + "grad_norm": 2.8226022720336914, + "learning_rate": 3.225958070088301e-06, + "loss": 0.5884, + "step": 3047 + }, + { + "epoch": 0.7450501099975556, + "grad_norm": 2.8809454441070557, + "learning_rate": 3.2201361259030374e-06, + "loss": 0.5376, + "step": 3048 + }, + { + "epoch": 0.745294549010022, + "grad_norm": 2.6132972240448, + "learning_rate": 3.2143184316046485e-06, + "loss": 0.5241, + "step": 3049 + }, + { + "epoch": 0.7455389880224884, + "grad_norm": 3.0091755390167236, + "learning_rate": 3.2085049908398945e-06, + "loss": 0.5732, + "step": 3050 + }, + { + "epoch": 0.7457834270349548, + "grad_norm": 3.3569703102111816, + "learning_rate": 3.2026958072528715e-06, + "loss": 0.5624, + "step": 3051 + }, + { + "epoch": 0.7460278660474212, + "grad_norm": 3.4056758880615234, + "learning_rate": 3.1968908844850045e-06, + "loss": 0.5958, + "step": 3052 + }, + { + "epoch": 0.7462723050598875, + "grad_norm": 3.654890298843384, + "learning_rate": 3.191090226175049e-06, + "loss": 0.663, + "step": 3053 + }, + { + "epoch": 0.7465167440723539, + "grad_norm": 4.088568210601807, + "learning_rate": 3.1852938359590857e-06, + "loss": 0.665, + "step": 3054 + }, + { + "epoch": 0.7467611830848203, + "grad_norm": 2.4330692291259766, + "learning_rate": 3.179501717470522e-06, + "loss": 2.0035, + "step": 3055 + }, + { + "epoch": 0.7470056220972867, + "grad_norm": 3.3720357418060303, + "learning_rate": 3.1737138743400896e-06, + "loss": 0.6575, + "step": 3056 + }, + { + "epoch": 0.7472500611097531, + "grad_norm": 2.899686336517334, + "learning_rate": 3.167930310195834e-06, + "loss": 0.6562, + "step": 3057 + }, + { + "epoch": 0.7474945001222195, + "grad_norm": 3.022397518157959, + "learning_rate": 3.162151028663123e-06, + "loss": 0.6167, + "step": 3058 + }, + { + "epoch": 0.7477389391346859, + "grad_norm": 5.976170063018799, + "learning_rate": 3.15637603336464e-06, + "loss": 0.5957, + "step": 3059 + }, + { + "epoch": 0.7479833781471523, + "grad_norm": 3.8271026611328125, + "learning_rate": 3.1506053279203787e-06, + "loss": 0.5749, + "step": 3060 + }, + { + "epoch": 0.7482278171596187, + "grad_norm": 3.036123275756836, + "learning_rate": 3.1448389159476433e-06, + "loss": 0.5961, + "step": 3061 + }, + { + "epoch": 0.748472256172085, + "grad_norm": 3.2765238285064697, + "learning_rate": 3.139076801061056e-06, + "loss": 0.6276, + "step": 3062 + }, + { + "epoch": 0.7487166951845514, + "grad_norm": 8.13082218170166, + "learning_rate": 3.1333189868725343e-06, + "loss": 0.5269, + "step": 3063 + }, + { + "epoch": 0.7489611341970178, + "grad_norm": 2.5719962120056152, + "learning_rate": 3.1275654769913043e-06, + "loss": 1.9888, + "step": 3064 + }, + { + "epoch": 0.7492055732094842, + "grad_norm": 2.6839728355407715, + "learning_rate": 3.1218162750238933e-06, + "loss": 0.482, + "step": 3065 + }, + { + "epoch": 0.7494500122219506, + "grad_norm": 3.614717960357666, + "learning_rate": 3.1160713845741287e-06, + "loss": 0.5976, + "step": 3066 + }, + { + "epoch": 0.749694451234417, + "grad_norm": 3.5247299671173096, + "learning_rate": 3.110330809243134e-06, + "loss": 0.6967, + "step": 3067 + }, + { + "epoch": 0.7499388902468834, + "grad_norm": 3.888062000274658, + "learning_rate": 3.1045945526293307e-06, + "loss": 0.5281, + "step": 3068 + }, + { + "epoch": 0.7501833292593498, + "grad_norm": 2.962456464767456, + "learning_rate": 3.098862618328431e-06, + "loss": 1.8584, + "step": 3069 + }, + { + "epoch": 0.7504277682718162, + "grad_norm": 2.798931121826172, + "learning_rate": 3.0931350099334366e-06, + "loss": 0.5392, + "step": 3070 + }, + { + "epoch": 0.7506722072842825, + "grad_norm": 3.3497190475463867, + "learning_rate": 3.087411731034641e-06, + "loss": 0.6611, + "step": 3071 + }, + { + "epoch": 0.7509166462967489, + "grad_norm": 3.342142343521118, + "learning_rate": 3.08169278521962e-06, + "loss": 0.6632, + "step": 3072 + }, + { + "epoch": 0.7511610853092153, + "grad_norm": 2.9381072521209717, + "learning_rate": 3.0759781760732365e-06, + "loss": 0.6726, + "step": 3073 + }, + { + "epoch": 0.7514055243216817, + "grad_norm": 3.3888957500457764, + "learning_rate": 3.070267907177633e-06, + "loss": 0.6559, + "step": 3074 + }, + { + "epoch": 0.7516499633341481, + "grad_norm": 2.966024875640869, + "learning_rate": 3.064561982112232e-06, + "loss": 0.6128, + "step": 3075 + }, + { + "epoch": 0.7518944023466145, + "grad_norm": 3.589439630508423, + "learning_rate": 3.0588604044537297e-06, + "loss": 0.6941, + "step": 3076 + }, + { + "epoch": 0.7521388413590809, + "grad_norm": 2.973039388656616, + "learning_rate": 3.053163177776106e-06, + "loss": 0.6058, + "step": 3077 + }, + { + "epoch": 0.7523832803715473, + "grad_norm": 3.746828079223633, + "learning_rate": 3.047470305650607e-06, + "loss": 0.5882, + "step": 3078 + }, + { + "epoch": 0.7526277193840137, + "grad_norm": 3.47501277923584, + "learning_rate": 3.0417817916457482e-06, + "loss": 0.5496, + "step": 3079 + }, + { + "epoch": 0.75287215839648, + "grad_norm": 3.0161173343658447, + "learning_rate": 3.036097639327317e-06, + "loss": 0.6409, + "step": 3080 + }, + { + "epoch": 0.7531165974089464, + "grad_norm": 2.6342554092407227, + "learning_rate": 3.0304178522583626e-06, + "loss": 0.5368, + "step": 3081 + }, + { + "epoch": 0.7533610364214128, + "grad_norm": 3.435028553009033, + "learning_rate": 3.0247424339992017e-06, + "loss": 0.6462, + "step": 3082 + }, + { + "epoch": 0.7536054754338792, + "grad_norm": 3.231520175933838, + "learning_rate": 3.0190713881074106e-06, + "loss": 0.5733, + "step": 3083 + }, + { + "epoch": 0.7538499144463456, + "grad_norm": 3.6398332118988037, + "learning_rate": 3.0134047181378256e-06, + "loss": 0.6683, + "step": 3084 + }, + { + "epoch": 0.754094353458812, + "grad_norm": 3.8633766174316406, + "learning_rate": 3.007742427642539e-06, + "loss": 0.5601, + "step": 3085 + }, + { + "epoch": 0.7543387924712784, + "grad_norm": 3.340566635131836, + "learning_rate": 3.0020845201708993e-06, + "loss": 0.4947, + "step": 3086 + }, + { + "epoch": 0.7545832314837448, + "grad_norm": 3.265387773513794, + "learning_rate": 2.9964309992695053e-06, + "loss": 0.5801, + "step": 3087 + }, + { + "epoch": 0.7548276704962112, + "grad_norm": 2.477468490600586, + "learning_rate": 2.9907818684822088e-06, + "loss": 1.9359, + "step": 3088 + }, + { + "epoch": 0.7550721095086775, + "grad_norm": 2.758582592010498, + "learning_rate": 2.985137131350109e-06, + "loss": 0.432, + "step": 3089 + }, + { + "epoch": 0.7553165485211439, + "grad_norm": 3.1395211219787598, + "learning_rate": 2.979496791411548e-06, + "loss": 0.5712, + "step": 3090 + }, + { + "epoch": 0.7555609875336103, + "grad_norm": 3.1856343746185303, + "learning_rate": 2.973860852202117e-06, + "loss": 0.7134, + "step": 3091 + }, + { + "epoch": 0.7558054265460767, + "grad_norm": 3.088649034500122, + "learning_rate": 2.9682293172546407e-06, + "loss": 0.6315, + "step": 3092 + }, + { + "epoch": 0.7560498655585431, + "grad_norm": 3.3430135250091553, + "learning_rate": 2.962602190099193e-06, + "loss": 0.629, + "step": 3093 + }, + { + "epoch": 0.7562943045710095, + "grad_norm": 4.353748321533203, + "learning_rate": 2.9569794742630796e-06, + "loss": 0.6365, + "step": 3094 + }, + { + "epoch": 0.756538743583476, + "grad_norm": 3.6114869117736816, + "learning_rate": 2.9513611732708404e-06, + "loss": 0.7775, + "step": 3095 + }, + { + "epoch": 0.7567831825959423, + "grad_norm": 2.5576424598693848, + "learning_rate": 2.9457472906442476e-06, + "loss": 1.9289, + "step": 3096 + }, + { + "epoch": 0.7570276216084088, + "grad_norm": 3.31587553024292, + "learning_rate": 2.9401378299023043e-06, + "loss": 0.5658, + "step": 3097 + }, + { + "epoch": 0.757272060620875, + "grad_norm": 3.4183433055877686, + "learning_rate": 2.9345327945612446e-06, + "loss": 0.512, + "step": 3098 + }, + { + "epoch": 0.7575164996333414, + "grad_norm": 3.2171711921691895, + "learning_rate": 2.9289321881345257e-06, + "loss": 0.7227, + "step": 3099 + }, + { + "epoch": 0.7577609386458078, + "grad_norm": 5.540745258331299, + "learning_rate": 2.9233360141328283e-06, + "loss": 0.6218, + "step": 3100 + }, + { + "epoch": 0.7580053776582742, + "grad_norm": 3.9040966033935547, + "learning_rate": 2.917744276064056e-06, + "loss": 0.4521, + "step": 3101 + }, + { + "epoch": 0.7582498166707407, + "grad_norm": 3.151887893676758, + "learning_rate": 2.9121569774333327e-06, + "loss": 0.6599, + "step": 3102 + }, + { + "epoch": 0.758494255683207, + "grad_norm": 3.7008299827575684, + "learning_rate": 2.9065741217429964e-06, + "loss": 0.593, + "step": 3103 + }, + { + "epoch": 0.7587386946956735, + "grad_norm": 4.508377552032471, + "learning_rate": 2.9009957124926026e-06, + "loss": 0.5348, + "step": 3104 + }, + { + "epoch": 0.7589831337081399, + "grad_norm": 2.980142116546631, + "learning_rate": 2.8954217531789186e-06, + "loss": 0.5423, + "step": 3105 + }, + { + "epoch": 0.7592275727206063, + "grad_norm": 2.6276297569274902, + "learning_rate": 2.8898522472959234e-06, + "loss": 0.581, + "step": 3106 + }, + { + "epoch": 0.7594720117330725, + "grad_norm": 3.8187081813812256, + "learning_rate": 2.8842871983348e-06, + "loss": 0.6095, + "step": 3107 + }, + { + "epoch": 0.759716450745539, + "grad_norm": 3.7661819458007812, + "learning_rate": 2.878726609783947e-06, + "loss": 0.5473, + "step": 3108 + }, + { + "epoch": 0.7599608897580054, + "grad_norm": 5.643442153930664, + "learning_rate": 2.8731704851289598e-06, + "loss": 0.5268, + "step": 3109 + }, + { + "epoch": 0.7602053287704718, + "grad_norm": 3.5928304195404053, + "learning_rate": 2.8676188278526353e-06, + "loss": 0.5524, + "step": 3110 + }, + { + "epoch": 0.7604497677829382, + "grad_norm": 3.473266124725342, + "learning_rate": 2.8620716414349714e-06, + "loss": 0.4903, + "step": 3111 + }, + { + "epoch": 0.7606942067954046, + "grad_norm": 3.6228229999542236, + "learning_rate": 2.856528929353166e-06, + "loss": 0.6172, + "step": 3112 + }, + { + "epoch": 0.760938645807871, + "grad_norm": 3.694929361343384, + "learning_rate": 2.85099069508161e-06, + "loss": 0.584, + "step": 3113 + }, + { + "epoch": 0.7611830848203374, + "grad_norm": 4.231311798095703, + "learning_rate": 2.8454569420918865e-06, + "loss": 0.7133, + "step": 3114 + }, + { + "epoch": 0.7614275238328038, + "grad_norm": 4.651334285736084, + "learning_rate": 2.8399276738527716e-06, + "loss": 0.5553, + "step": 3115 + }, + { + "epoch": 0.7616719628452701, + "grad_norm": 3.3225133419036865, + "learning_rate": 2.8344028938302292e-06, + "loss": 0.5635, + "step": 3116 + }, + { + "epoch": 0.7619164018577365, + "grad_norm": 2.9594101905822754, + "learning_rate": 2.8288826054874095e-06, + "loss": 0.6375, + "step": 3117 + }, + { + "epoch": 0.7621608408702029, + "grad_norm": 3.052440643310547, + "learning_rate": 2.823366812284649e-06, + "loss": 0.5714, + "step": 3118 + }, + { + "epoch": 0.7624052798826693, + "grad_norm": 4.923966884613037, + "learning_rate": 2.817855517679464e-06, + "loss": 0.6374, + "step": 3119 + }, + { + "epoch": 0.7626497188951357, + "grad_norm": 3.6699352264404297, + "learning_rate": 2.812348725126551e-06, + "loss": 0.5905, + "step": 3120 + }, + { + "epoch": 0.7628941579076021, + "grad_norm": 3.297264575958252, + "learning_rate": 2.806846438077787e-06, + "loss": 0.6301, + "step": 3121 + }, + { + "epoch": 0.7631385969200685, + "grad_norm": 2.899012327194214, + "learning_rate": 2.801348659982224e-06, + "loss": 0.4651, + "step": 3122 + }, + { + "epoch": 0.7633830359325349, + "grad_norm": 3.3183135986328125, + "learning_rate": 2.795855394286081e-06, + "loss": 0.5737, + "step": 3123 + }, + { + "epoch": 0.7636274749450013, + "grad_norm": 3.0558860301971436, + "learning_rate": 2.7903666444327615e-06, + "loss": 0.5875, + "step": 3124 + }, + { + "epoch": 0.7638719139574676, + "grad_norm": 5.014186859130859, + "learning_rate": 2.7848824138628295e-06, + "loss": 0.595, + "step": 3125 + }, + { + "epoch": 0.764116352969934, + "grad_norm": 3.420767307281494, + "learning_rate": 2.779402706014016e-06, + "loss": 0.5561, + "step": 3126 + }, + { + "epoch": 0.7643607919824004, + "grad_norm": 3.7397305965423584, + "learning_rate": 2.7739275243212183e-06, + "loss": 0.5283, + "step": 3127 + }, + { + "epoch": 0.7646052309948668, + "grad_norm": 2.853084087371826, + "learning_rate": 2.7684568722164983e-06, + "loss": 0.6438, + "step": 3128 + }, + { + "epoch": 0.7648496700073332, + "grad_norm": 3.4784562587738037, + "learning_rate": 2.7629907531290767e-06, + "loss": 0.6024, + "step": 3129 + }, + { + "epoch": 0.7650941090197996, + "grad_norm": 3.2651922702789307, + "learning_rate": 2.7575291704853325e-06, + "loss": 0.6737, + "step": 3130 + }, + { + "epoch": 0.765338548032266, + "grad_norm": 3.167628526687622, + "learning_rate": 2.7520721277088023e-06, + "loss": 0.6024, + "step": 3131 + }, + { + "epoch": 0.7655829870447324, + "grad_norm": 3.006815195083618, + "learning_rate": 2.746619628220176e-06, + "loss": 0.5909, + "step": 3132 + }, + { + "epoch": 0.7658274260571988, + "grad_norm": 3.065413236618042, + "learning_rate": 2.7411716754372964e-06, + "loss": 0.6586, + "step": 3133 + }, + { + "epoch": 0.7660718650696651, + "grad_norm": 2.7193782329559326, + "learning_rate": 2.7357282727751567e-06, + "loss": 2.0259, + "step": 3134 + }, + { + "epoch": 0.7663163040821315, + "grad_norm": 2.904202938079834, + "learning_rate": 2.730289423645894e-06, + "loss": 0.5203, + "step": 3135 + }, + { + "epoch": 0.7665607430945979, + "grad_norm": 3.3061468601226807, + "learning_rate": 2.724855131458798e-06, + "loss": 0.6414, + "step": 3136 + }, + { + "epoch": 0.7668051821070643, + "grad_norm": 2.839189291000366, + "learning_rate": 2.719425399620296e-06, + "loss": 0.6055, + "step": 3137 + }, + { + "epoch": 0.7670496211195307, + "grad_norm": 3.020749807357788, + "learning_rate": 2.714000231533954e-06, + "loss": 0.4995, + "step": 3138 + }, + { + "epoch": 0.7672940601319971, + "grad_norm": 3.7392027378082275, + "learning_rate": 2.7085796306004907e-06, + "loss": 0.5281, + "step": 3139 + }, + { + "epoch": 0.7675384991444635, + "grad_norm": 2.9909677505493164, + "learning_rate": 2.7031636002177496e-06, + "loss": 0.5481, + "step": 3140 + }, + { + "epoch": 0.7677829381569299, + "grad_norm": 2.932992935180664, + "learning_rate": 2.697752143780713e-06, + "loss": 0.6252, + "step": 3141 + }, + { + "epoch": 0.7680273771693963, + "grad_norm": 3.954922676086426, + "learning_rate": 2.6923452646814942e-06, + "loss": 0.6438, + "step": 3142 + }, + { + "epoch": 0.7682718161818626, + "grad_norm": 3.9304919242858887, + "learning_rate": 2.6869429663093405e-06, + "loss": 1.8718, + "step": 3143 + }, + { + "epoch": 0.768516255194329, + "grad_norm": 7.261805057525635, + "learning_rate": 2.681545252050626e-06, + "loss": 0.56, + "step": 3144 + }, + { + "epoch": 0.7687606942067954, + "grad_norm": 3.9659855365753174, + "learning_rate": 2.676152125288851e-06, + "loss": 0.5806, + "step": 3145 + }, + { + "epoch": 0.7690051332192618, + "grad_norm": 3.0885088443756104, + "learning_rate": 2.67076358940464e-06, + "loss": 0.6062, + "step": 3146 + }, + { + "epoch": 0.7692495722317282, + "grad_norm": 3.743597984313965, + "learning_rate": 2.665379647775743e-06, + "loss": 0.5626, + "step": 3147 + }, + { + "epoch": 0.7694940112441946, + "grad_norm": 3.111335039138794, + "learning_rate": 2.660000303777025e-06, + "loss": 0.5573, + "step": 3148 + }, + { + "epoch": 0.769738450256661, + "grad_norm": 3.303645372390747, + "learning_rate": 2.6546255607804726e-06, + "loss": 0.6009, + "step": 3149 + }, + { + "epoch": 0.7699828892691274, + "grad_norm": 3.0187478065490723, + "learning_rate": 2.649255422155188e-06, + "loss": 0.526, + "step": 3150 + }, + { + "epoch": 0.7702273282815938, + "grad_norm": 2.943671464920044, + "learning_rate": 2.643889891267386e-06, + "loss": 0.6326, + "step": 3151 + }, + { + "epoch": 0.7704717672940601, + "grad_norm": 3.4020495414733887, + "learning_rate": 2.638528971480393e-06, + "loss": 0.6332, + "step": 3152 + }, + { + "epoch": 0.7707162063065265, + "grad_norm": 4.747247219085693, + "learning_rate": 2.633172666154645e-06, + "loss": 0.6612, + "step": 3153 + }, + { + "epoch": 0.7709606453189929, + "grad_norm": 3.2169346809387207, + "learning_rate": 2.627820978647686e-06, + "loss": 0.642, + "step": 3154 + }, + { + "epoch": 0.7712050843314593, + "grad_norm": 2.937433958053589, + "learning_rate": 2.6224739123141684e-06, + "loss": 0.4554, + "step": 3155 + }, + { + "epoch": 0.7714495233439257, + "grad_norm": 3.4266116619110107, + "learning_rate": 2.6171314705058438e-06, + "loss": 0.5838, + "step": 3156 + }, + { + "epoch": 0.7716939623563921, + "grad_norm": 3.0606844425201416, + "learning_rate": 2.611793656571565e-06, + "loss": 0.5011, + "step": 3157 + }, + { + "epoch": 0.7719384013688585, + "grad_norm": 3.9658639430999756, + "learning_rate": 2.606460473857285e-06, + "loss": 0.6529, + "step": 3158 + }, + { + "epoch": 0.7721828403813249, + "grad_norm": 3.4948606491088867, + "learning_rate": 2.601131925706054e-06, + "loss": 0.6081, + "step": 3159 + }, + { + "epoch": 0.7724272793937913, + "grad_norm": 3.4018239974975586, + "learning_rate": 2.5958080154580167e-06, + "loss": 0.7337, + "step": 3160 + }, + { + "epoch": 0.7726717184062576, + "grad_norm": 4.360661029815674, + "learning_rate": 2.5904887464504115e-06, + "loss": 0.5395, + "step": 3161 + }, + { + "epoch": 0.772916157418724, + "grad_norm": 2.925231695175171, + "learning_rate": 2.5851741220175643e-06, + "loss": 0.566, + "step": 3162 + }, + { + "epoch": 0.7731605964311904, + "grad_norm": 5.550756931304932, + "learning_rate": 2.5798641454908945e-06, + "loss": 0.6263, + "step": 3163 + }, + { + "epoch": 0.7734050354436568, + "grad_norm": 3.855212926864624, + "learning_rate": 2.5745588201989037e-06, + "loss": 0.7732, + "step": 3164 + }, + { + "epoch": 0.7736494744561232, + "grad_norm": 3.303528070449829, + "learning_rate": 2.569258149467181e-06, + "loss": 0.4347, + "step": 3165 + }, + { + "epoch": 0.7738939134685896, + "grad_norm": 2.3593063354492188, + "learning_rate": 2.5639621366183953e-06, + "loss": 1.8563, + "step": 3166 + }, + { + "epoch": 0.774138352481056, + "grad_norm": 3.5858919620513916, + "learning_rate": 2.558670784972298e-06, + "loss": 0.6312, + "step": 3167 + }, + { + "epoch": 0.7743827914935224, + "grad_norm": 5.1975016593933105, + "learning_rate": 2.5533840978457135e-06, + "loss": 0.6655, + "step": 3168 + }, + { + "epoch": 0.7746272305059888, + "grad_norm": 3.6032609939575195, + "learning_rate": 2.5481020785525534e-06, + "loss": 0.53, + "step": 3169 + }, + { + "epoch": 0.7748716695184551, + "grad_norm": 3.0232839584350586, + "learning_rate": 2.542824730403792e-06, + "loss": 0.5679, + "step": 3170 + }, + { + "epoch": 0.7751161085309215, + "grad_norm": 3.177489995956421, + "learning_rate": 2.537552056707483e-06, + "loss": 0.6288, + "step": 3171 + }, + { + "epoch": 0.7753605475433879, + "grad_norm": 2.9517323970794678, + "learning_rate": 2.5322840607687426e-06, + "loss": 0.6599, + "step": 3172 + }, + { + "epoch": 0.7756049865558543, + "grad_norm": 3.787276268005371, + "learning_rate": 2.5270207458897622e-06, + "loss": 0.51, + "step": 3173 + }, + { + "epoch": 0.7758494255683207, + "grad_norm": 3.143080234527588, + "learning_rate": 2.5217621153697947e-06, + "loss": 0.585, + "step": 3174 + }, + { + "epoch": 0.7760938645807871, + "grad_norm": 2.5998079776763916, + "learning_rate": 2.5165081725051564e-06, + "loss": 1.8594, + "step": 3175 + }, + { + "epoch": 0.7763383035932535, + "grad_norm": 3.082580327987671, + "learning_rate": 2.511258920589229e-06, + "loss": 0.5744, + "step": 3176 + }, + { + "epoch": 0.7765827426057199, + "grad_norm": 4.161632061004639, + "learning_rate": 2.50601436291245e-06, + "loss": 0.5662, + "step": 3177 + }, + { + "epoch": 0.7768271816181863, + "grad_norm": 3.5269548892974854, + "learning_rate": 2.5007745027623164e-06, + "loss": 0.7999, + "step": 3178 + }, + { + "epoch": 0.7770716206306526, + "grad_norm": 2.985814332962036, + "learning_rate": 2.4955393434233756e-06, + "loss": 0.4679, + "step": 3179 + }, + { + "epoch": 0.777316059643119, + "grad_norm": 2.7241947650909424, + "learning_rate": 2.4903088881772384e-06, + "loss": 0.5001, + "step": 3180 + }, + { + "epoch": 0.7775604986555854, + "grad_norm": 3.6743762493133545, + "learning_rate": 2.4850831403025597e-06, + "loss": 0.5835, + "step": 3181 + }, + { + "epoch": 0.7778049376680518, + "grad_norm": 3.5109803676605225, + "learning_rate": 2.4798621030750438e-06, + "loss": 0.5953, + "step": 3182 + }, + { + "epoch": 0.7780493766805182, + "grad_norm": 2.7164785861968994, + "learning_rate": 2.4746457797674438e-06, + "loss": 1.8747, + "step": 3183 + }, + { + "epoch": 0.7782938156929846, + "grad_norm": 4.3310723304748535, + "learning_rate": 2.4694341736495576e-06, + "loss": 0.6783, + "step": 3184 + }, + { + "epoch": 0.778538254705451, + "grad_norm": 3.274315118789673, + "learning_rate": 2.4642272879882245e-06, + "loss": 0.6919, + "step": 3185 + }, + { + "epoch": 0.7787826937179174, + "grad_norm": 2.5512311458587646, + "learning_rate": 2.4590251260473288e-06, + "loss": 0.5595, + "step": 3186 + }, + { + "epoch": 0.7790271327303838, + "grad_norm": 3.0792834758758545, + "learning_rate": 2.45382769108779e-06, + "loss": 0.6065, + "step": 3187 + }, + { + "epoch": 0.7792715717428501, + "grad_norm": 2.90177321434021, + "learning_rate": 2.4486349863675664e-06, + "loss": 0.5647, + "step": 3188 + }, + { + "epoch": 0.7795160107553165, + "grad_norm": 2.9708359241485596, + "learning_rate": 2.443447015141648e-06, + "loss": 0.4932, + "step": 3189 + }, + { + "epoch": 0.7797604497677829, + "grad_norm": 3.1886961460113525, + "learning_rate": 2.4382637806620646e-06, + "loss": 0.5723, + "step": 3190 + }, + { + "epoch": 0.7800048887802493, + "grad_norm": 3.5020089149475098, + "learning_rate": 2.433085286177872e-06, + "loss": 0.5639, + "step": 3191 + }, + { + "epoch": 0.7802493277927157, + "grad_norm": 4.378829002380371, + "learning_rate": 2.4279115349351546e-06, + "loss": 0.6527, + "step": 3192 + }, + { + "epoch": 0.7804937668051821, + "grad_norm": 4.593955993652344, + "learning_rate": 2.4227425301770236e-06, + "loss": 0.7164, + "step": 3193 + }, + { + "epoch": 0.7807382058176485, + "grad_norm": 2.77014422416687, + "learning_rate": 2.417578275143616e-06, + "loss": 0.4451, + "step": 3194 + }, + { + "epoch": 0.7809826448301149, + "grad_norm": 3.2085678577423096, + "learning_rate": 2.4124187730720916e-06, + "loss": 0.6105, + "step": 3195 + }, + { + "epoch": 0.7812270838425813, + "grad_norm": 2.853844404220581, + "learning_rate": 2.4072640271966297e-06, + "loss": 0.5077, + "step": 3196 + }, + { + "epoch": 0.7814715228550476, + "grad_norm": 2.981623888015747, + "learning_rate": 2.402114040748429e-06, + "loss": 0.5778, + "step": 3197 + }, + { + "epoch": 0.781715961867514, + "grad_norm": 3.194746971130371, + "learning_rate": 2.396968816955706e-06, + "loss": 0.7014, + "step": 3198 + }, + { + "epoch": 0.7819604008799804, + "grad_norm": 2.9603145122528076, + "learning_rate": 2.3918283590436897e-06, + "loss": 0.4981, + "step": 3199 + }, + { + "epoch": 0.7822048398924468, + "grad_norm": 3.691871166229248, + "learning_rate": 2.3866926702346182e-06, + "loss": 0.495, + "step": 3200 + }, + { + "epoch": 0.7824492789049132, + "grad_norm": 3.7968385219573975, + "learning_rate": 2.381561753747753e-06, + "loss": 0.5113, + "step": 3201 + }, + { + "epoch": 0.7826937179173796, + "grad_norm": 2.7656779289245605, + "learning_rate": 2.3764356127993495e-06, + "loss": 2.0152, + "step": 3202 + }, + { + "epoch": 0.782938156929846, + "grad_norm": 3.0234978199005127, + "learning_rate": 2.3713142506026786e-06, + "loss": 0.6202, + "step": 3203 + }, + { + "epoch": 0.7831825959423124, + "grad_norm": 3.1162590980529785, + "learning_rate": 2.36619767036801e-06, + "loss": 0.6806, + "step": 3204 + }, + { + "epoch": 0.7834270349547788, + "grad_norm": 3.5407638549804688, + "learning_rate": 2.361085875302621e-06, + "loss": 0.621, + "step": 3205 + }, + { + "epoch": 0.7836714739672451, + "grad_norm": 4.021968841552734, + "learning_rate": 2.3559788686107853e-06, + "loss": 0.5917, + "step": 3206 + }, + { + "epoch": 0.7839159129797115, + "grad_norm": 3.0466601848602295, + "learning_rate": 2.3508766534937776e-06, + "loss": 0.5675, + "step": 3207 + }, + { + "epoch": 0.7841603519921779, + "grad_norm": 2.976064920425415, + "learning_rate": 2.3457792331498672e-06, + "loss": 0.5782, + "step": 3208 + }, + { + "epoch": 0.7844047910046443, + "grad_norm": 2.7482686042785645, + "learning_rate": 2.3406866107743196e-06, + "loss": 0.5479, + "step": 3209 + }, + { + "epoch": 0.7846492300171107, + "grad_norm": 3.396845817565918, + "learning_rate": 2.3355987895593923e-06, + "loss": 0.556, + "step": 3210 + }, + { + "epoch": 0.7848936690295771, + "grad_norm": 2.758816957473755, + "learning_rate": 2.330515772694333e-06, + "loss": 0.5656, + "step": 3211 + }, + { + "epoch": 0.7851381080420435, + "grad_norm": 2.695256233215332, + "learning_rate": 2.325437563365376e-06, + "loss": 0.4653, + "step": 3212 + }, + { + "epoch": 0.7853825470545099, + "grad_norm": 5.168088436126709, + "learning_rate": 2.320364164755745e-06, + "loss": 0.6637, + "step": 3213 + }, + { + "epoch": 0.7856269860669763, + "grad_norm": 3.181368350982666, + "learning_rate": 2.3152955800456468e-06, + "loss": 0.613, + "step": 3214 + }, + { + "epoch": 0.7858714250794426, + "grad_norm": 6.01630973815918, + "learning_rate": 2.3102318124122714e-06, + "loss": 0.6279, + "step": 3215 + }, + { + "epoch": 0.786115864091909, + "grad_norm": 3.028567314147949, + "learning_rate": 2.3051728650297856e-06, + "loss": 0.5401, + "step": 3216 + }, + { + "epoch": 0.7863603031043754, + "grad_norm": 3.762362480163574, + "learning_rate": 2.3001187410693427e-06, + "loss": 0.5725, + "step": 3217 + }, + { + "epoch": 0.7866047421168418, + "grad_norm": 3.6007094383239746, + "learning_rate": 2.295069443699065e-06, + "loss": 0.6648, + "step": 3218 + }, + { + "epoch": 0.7868491811293082, + "grad_norm": 2.9934427738189697, + "learning_rate": 2.290024976084052e-06, + "loss": 0.5656, + "step": 3219 + }, + { + "epoch": 0.7870936201417746, + "grad_norm": 3.62497615814209, + "learning_rate": 2.2849853413863756e-06, + "loss": 0.6877, + "step": 3220 + }, + { + "epoch": 0.787338059154241, + "grad_norm": 3.790687084197998, + "learning_rate": 2.279950542765078e-06, + "loss": 0.6179, + "step": 3221 + }, + { + "epoch": 0.7875824981667074, + "grad_norm": 2.8185369968414307, + "learning_rate": 2.274920583376171e-06, + "loss": 0.5653, + "step": 3222 + }, + { + "epoch": 0.7878269371791738, + "grad_norm": 3.2843177318573, + "learning_rate": 2.26989546637263e-06, + "loss": 0.5257, + "step": 3223 + }, + { + "epoch": 0.7880713761916401, + "grad_norm": 2.947882890701294, + "learning_rate": 2.2648751949044e-06, + "loss": 0.6428, + "step": 3224 + }, + { + "epoch": 0.7883158152041065, + "grad_norm": 3.8019323348999023, + "learning_rate": 2.2598597721183834e-06, + "loss": 0.5644, + "step": 3225 + }, + { + "epoch": 0.7885602542165729, + "grad_norm": 4.020700454711914, + "learning_rate": 2.2548492011584465e-06, + "loss": 0.5394, + "step": 3226 + }, + { + "epoch": 0.7888046932290393, + "grad_norm": 3.4065825939178467, + "learning_rate": 2.2498434851654125e-06, + "loss": 0.7011, + "step": 3227 + }, + { + "epoch": 0.7890491322415057, + "grad_norm": 3.1583659648895264, + "learning_rate": 2.244842627277064e-06, + "loss": 0.7297, + "step": 3228 + }, + { + "epoch": 0.7892935712539721, + "grad_norm": 2.93677020072937, + "learning_rate": 2.2398466306281364e-06, + "loss": 0.6397, + "step": 3229 + }, + { + "epoch": 0.7895380102664386, + "grad_norm": 3.476424217224121, + "learning_rate": 2.2348554983503156e-06, + "loss": 0.6933, + "step": 3230 + }, + { + "epoch": 0.789782449278905, + "grad_norm": 2.351652145385742, + "learning_rate": 2.2298692335722403e-06, + "loss": 1.98, + "step": 3231 + }, + { + "epoch": 0.7900268882913714, + "grad_norm": 4.299272537231445, + "learning_rate": 2.224887839419503e-06, + "loss": 0.6083, + "step": 3232 + }, + { + "epoch": 0.7902713273038376, + "grad_norm": 3.0552446842193604, + "learning_rate": 2.219911319014637e-06, + "loss": 0.5848, + "step": 3233 + }, + { + "epoch": 0.790515766316304, + "grad_norm": 2.289609909057617, + "learning_rate": 2.2149396754771203e-06, + "loss": 1.9677, + "step": 3234 + }, + { + "epoch": 0.7907602053287704, + "grad_norm": 3.8874871730804443, + "learning_rate": 2.209972911923377e-06, + "loss": 0.6953, + "step": 3235 + }, + { + "epoch": 0.7910046443412369, + "grad_norm": 2.7373220920562744, + "learning_rate": 2.205011031466772e-06, + "loss": 0.5882, + "step": 3236 + }, + { + "epoch": 0.7912490833537033, + "grad_norm": 3.414679527282715, + "learning_rate": 2.200054037217605e-06, + "loss": 0.6221, + "step": 3237 + }, + { + "epoch": 0.7914935223661697, + "grad_norm": 3.6899285316467285, + "learning_rate": 2.1951019322831193e-06, + "loss": 0.6806, + "step": 3238 + }, + { + "epoch": 0.7917379613786361, + "grad_norm": 3.0427722930908203, + "learning_rate": 2.190154719767489e-06, + "loss": 0.6509, + "step": 3239 + }, + { + "epoch": 0.7919824003911025, + "grad_norm": 2.772542953491211, + "learning_rate": 2.1852124027718212e-06, + "loss": 0.4565, + "step": 3240 + }, + { + "epoch": 0.7922268394035689, + "grad_norm": 3.280536651611328, + "learning_rate": 2.1802749843941583e-06, + "loss": 0.5821, + "step": 3241 + }, + { + "epoch": 0.7924712784160352, + "grad_norm": 3.199457883834839, + "learning_rate": 2.1753424677294675e-06, + "loss": 0.6618, + "step": 3242 + }, + { + "epoch": 0.7927157174285016, + "grad_norm": 3.418578624725342, + "learning_rate": 2.170414855869647e-06, + "loss": 0.5916, + "step": 3243 + }, + { + "epoch": 0.792960156440968, + "grad_norm": 3.6555192470550537, + "learning_rate": 2.1654921519035176e-06, + "loss": 0.5722, + "step": 3244 + }, + { + "epoch": 0.7932045954534344, + "grad_norm": 2.9484596252441406, + "learning_rate": 2.1605743589168247e-06, + "loss": 0.6156, + "step": 3245 + }, + { + "epoch": 0.7934490344659008, + "grad_norm": 3.5383124351501465, + "learning_rate": 2.155661479992236e-06, + "loss": 0.5778, + "step": 3246 + }, + { + "epoch": 0.7936934734783672, + "grad_norm": 5.903968811035156, + "learning_rate": 2.1507535182093353e-06, + "loss": 0.5943, + "step": 3247 + }, + { + "epoch": 0.7939379124908336, + "grad_norm": 2.7591583728790283, + "learning_rate": 2.145850476644634e-06, + "loss": 0.444, + "step": 3248 + }, + { + "epoch": 0.7941823515033, + "grad_norm": 2.4827725887298584, + "learning_rate": 2.1409523583715486e-06, + "loss": 0.4999, + "step": 3249 + }, + { + "epoch": 0.7944267905157664, + "grad_norm": 3.5222556591033936, + "learning_rate": 2.1360591664604125e-06, + "loss": 0.5139, + "step": 3250 + }, + { + "epoch": 0.7946712295282327, + "grad_norm": 2.6247849464416504, + "learning_rate": 2.1311709039784734e-06, + "loss": 0.5401, + "step": 3251 + }, + { + "epoch": 0.7949156685406991, + "grad_norm": 3.2825536727905273, + "learning_rate": 2.126287573989888e-06, + "loss": 0.6492, + "step": 3252 + }, + { + "epoch": 0.7951601075531655, + "grad_norm": 2.7209300994873047, + "learning_rate": 2.121409179555719e-06, + "loss": 0.5069, + "step": 3253 + }, + { + "epoch": 0.7954045465656319, + "grad_norm": 3.4337806701660156, + "learning_rate": 2.116535723733938e-06, + "loss": 0.7743, + "step": 3254 + }, + { + "epoch": 0.7956489855780983, + "grad_norm": 3.2049851417541504, + "learning_rate": 2.11166720957942e-06, + "loss": 0.5545, + "step": 3255 + }, + { + "epoch": 0.7958934245905647, + "grad_norm": 2.9724299907684326, + "learning_rate": 2.1068036401439418e-06, + "loss": 0.6282, + "step": 3256 + }, + { + "epoch": 0.7961378636030311, + "grad_norm": 2.6269307136535645, + "learning_rate": 2.1019450184761814e-06, + "loss": 0.5668, + "step": 3257 + }, + { + "epoch": 0.7963823026154975, + "grad_norm": 3.181718349456787, + "learning_rate": 2.097091347621715e-06, + "loss": 0.5737, + "step": 3258 + }, + { + "epoch": 0.7966267416279639, + "grad_norm": 3.2268974781036377, + "learning_rate": 2.092242630623016e-06, + "loss": 0.6092, + "step": 3259 + }, + { + "epoch": 0.7968711806404302, + "grad_norm": 3.338627576828003, + "learning_rate": 2.08739887051945e-06, + "loss": 0.5365, + "step": 3260 + }, + { + "epoch": 0.7971156196528966, + "grad_norm": 2.914188861846924, + "learning_rate": 2.0825600703472814e-06, + "loss": 0.5473, + "step": 3261 + }, + { + "epoch": 0.797360058665363, + "grad_norm": 3.5348644256591797, + "learning_rate": 2.0777262331396543e-06, + "loss": 0.6275, + "step": 3262 + }, + { + "epoch": 0.7976044976778294, + "grad_norm": 3.0596249103546143, + "learning_rate": 2.0728973619266178e-06, + "loss": 0.5902, + "step": 3263 + }, + { + "epoch": 0.7978489366902958, + "grad_norm": 2.831759452819824, + "learning_rate": 2.0680734597350947e-06, + "loss": 0.5543, + "step": 3264 + }, + { + "epoch": 0.7980933757027622, + "grad_norm": 3.386136054992676, + "learning_rate": 2.0632545295888996e-06, + "loss": 0.5854, + "step": 3265 + }, + { + "epoch": 0.7983378147152286, + "grad_norm": 2.5762999057769775, + "learning_rate": 2.058440574508727e-06, + "loss": 0.4227, + "step": 3266 + }, + { + "epoch": 0.798582253727695, + "grad_norm": 2.781954288482666, + "learning_rate": 2.0536315975121545e-06, + "loss": 0.4883, + "step": 3267 + }, + { + "epoch": 0.7988266927401614, + "grad_norm": 3.280689239501953, + "learning_rate": 2.04882760161364e-06, + "loss": 0.518, + "step": 3268 + }, + { + "epoch": 0.7990711317526277, + "grad_norm": 2.670980453491211, + "learning_rate": 2.044028589824517e-06, + "loss": 0.4825, + "step": 3269 + }, + { + "epoch": 0.7993155707650941, + "grad_norm": 2.9630038738250732, + "learning_rate": 2.0392345651529964e-06, + "loss": 0.5242, + "step": 3270 + }, + { + "epoch": 0.7995600097775605, + "grad_norm": 2.9382364749908447, + "learning_rate": 2.0344455306041633e-06, + "loss": 0.5933, + "step": 3271 + }, + { + "epoch": 0.7998044487900269, + "grad_norm": 3.124269723892212, + "learning_rate": 2.029661489179974e-06, + "loss": 0.5831, + "step": 3272 + }, + { + "epoch": 0.8000488878024933, + "grad_norm": 3.2077784538269043, + "learning_rate": 2.0248824438792526e-06, + "loss": 0.6423, + "step": 3273 + }, + { + "epoch": 0.8002933268149597, + "grad_norm": 2.9261975288391113, + "learning_rate": 2.0201083976976966e-06, + "loss": 0.5352, + "step": 3274 + }, + { + "epoch": 0.8005377658274261, + "grad_norm": 2.8065948486328125, + "learning_rate": 2.015339353627865e-06, + "loss": 0.7017, + "step": 3275 + }, + { + "epoch": 0.8007822048398925, + "grad_norm": 2.950518846511841, + "learning_rate": 2.0105753146591846e-06, + "loss": 0.5949, + "step": 3276 + }, + { + "epoch": 0.8010266438523589, + "grad_norm": 2.621652126312256, + "learning_rate": 2.0058162837779437e-06, + "loss": 1.9721, + "step": 3277 + }, + { + "epoch": 0.8012710828648252, + "grad_norm": 2.55247163772583, + "learning_rate": 2.001062263967286e-06, + "loss": 1.8001, + "step": 3278 + }, + { + "epoch": 0.8015155218772916, + "grad_norm": 3.304582118988037, + "learning_rate": 1.9963132582072277e-06, + "loss": 0.4885, + "step": 3279 + }, + { + "epoch": 0.801759960889758, + "grad_norm": 3.22530198097229, + "learning_rate": 1.9915692694746293e-06, + "loss": 0.5259, + "step": 3280 + }, + { + "epoch": 0.8020043999022244, + "grad_norm": 6.259829521179199, + "learning_rate": 1.98683030074321e-06, + "loss": 0.5966, + "step": 3281 + }, + { + "epoch": 0.8022488389146908, + "grad_norm": 3.880714178085327, + "learning_rate": 1.9820963549835435e-06, + "loss": 0.588, + "step": 3282 + }, + { + "epoch": 0.8024932779271572, + "grad_norm": 3.169184923171997, + "learning_rate": 1.9773674351630543e-06, + "loss": 0.5565, + "step": 3283 + }, + { + "epoch": 0.8027377169396236, + "grad_norm": 3.4683644771575928, + "learning_rate": 1.972643544246017e-06, + "loss": 0.5455, + "step": 3284 + }, + { + "epoch": 0.80298215595209, + "grad_norm": 2.546109437942505, + "learning_rate": 1.967924685193552e-06, + "loss": 1.8964, + "step": 3285 + }, + { + "epoch": 0.8032265949645564, + "grad_norm": 7.747804164886475, + "learning_rate": 1.963210860963628e-06, + "loss": 0.6601, + "step": 3286 + }, + { + "epoch": 0.8034710339770227, + "grad_norm": 3.7255680561065674, + "learning_rate": 1.958502074511055e-06, + "loss": 0.6702, + "step": 3287 + }, + { + "epoch": 0.8037154729894891, + "grad_norm": 2.965513229370117, + "learning_rate": 1.9537983287874883e-06, + "loss": 0.6289, + "step": 3288 + }, + { + "epoch": 0.8039599120019555, + "grad_norm": 2.916699171066284, + "learning_rate": 1.949099626741422e-06, + "loss": 0.5857, + "step": 3289 + }, + { + "epoch": 0.8042043510144219, + "grad_norm": 2.9211907386779785, + "learning_rate": 1.9444059713181864e-06, + "loss": 0.6084, + "step": 3290 + }, + { + "epoch": 0.8044487900268883, + "grad_norm": 3.1928536891937256, + "learning_rate": 1.939717365459952e-06, + "loss": 0.5481, + "step": 3291 + }, + { + "epoch": 0.8046932290393547, + "grad_norm": 3.2966573238372803, + "learning_rate": 1.935033812105721e-06, + "loss": 0.4839, + "step": 3292 + }, + { + "epoch": 0.8049376680518211, + "grad_norm": 2.993467330932617, + "learning_rate": 1.9303553141913287e-06, + "loss": 0.6134, + "step": 3293 + }, + { + "epoch": 0.8051821070642875, + "grad_norm": 3.431793212890625, + "learning_rate": 1.9256818746494487e-06, + "loss": 0.5922, + "step": 3294 + }, + { + "epoch": 0.8054265460767539, + "grad_norm": 2.649465560913086, + "learning_rate": 1.9210134964095728e-06, + "loss": 0.4745, + "step": 3295 + }, + { + "epoch": 0.8056709850892202, + "grad_norm": 8.001269340515137, + "learning_rate": 1.9163501823980258e-06, + "loss": 0.5893, + "step": 3296 + }, + { + "epoch": 0.8059154241016866, + "grad_norm": 3.9312362670898438, + "learning_rate": 1.911691935537958e-06, + "loss": 0.5305, + "step": 3297 + }, + { + "epoch": 0.806159863114153, + "grad_norm": 3.161888360977173, + "learning_rate": 1.9070387587493433e-06, + "loss": 0.5117, + "step": 3298 + }, + { + "epoch": 0.8064043021266194, + "grad_norm": 2.8150248527526855, + "learning_rate": 1.9023906549489767e-06, + "loss": 0.6931, + "step": 3299 + }, + { + "epoch": 0.8066487411390858, + "grad_norm": 2.817641019821167, + "learning_rate": 1.8977476270504725e-06, + "loss": 0.6476, + "step": 3300 + }, + { + "epoch": 0.8068931801515522, + "grad_norm": 3.453606605529785, + "learning_rate": 1.8931096779642644e-06, + "loss": 0.595, + "step": 3301 + }, + { + "epoch": 0.8071376191640186, + "grad_norm": 3.230611801147461, + "learning_rate": 1.8884768105976037e-06, + "loss": 0.5715, + "step": 3302 + }, + { + "epoch": 0.807382058176485, + "grad_norm": 4.638473033905029, + "learning_rate": 1.8838490278545528e-06, + "loss": 0.5706, + "step": 3303 + }, + { + "epoch": 0.8076264971889514, + "grad_norm": 3.2550625801086426, + "learning_rate": 1.8792263326359917e-06, + "loss": 0.6181, + "step": 3304 + }, + { + "epoch": 0.8078709362014177, + "grad_norm": 3.597907066345215, + "learning_rate": 1.8746087278396074e-06, + "loss": 0.6596, + "step": 3305 + }, + { + "epoch": 0.8081153752138841, + "grad_norm": 3.790290594100952, + "learning_rate": 1.8699962163598962e-06, + "loss": 0.6337, + "step": 3306 + }, + { + "epoch": 0.8083598142263505, + "grad_norm": 4.040053844451904, + "learning_rate": 1.8653888010881637e-06, + "loss": 0.5535, + "step": 3307 + }, + { + "epoch": 0.8086042532388169, + "grad_norm": 3.4842941761016846, + "learning_rate": 1.8607864849125202e-06, + "loss": 0.6042, + "step": 3308 + }, + { + "epoch": 0.8088486922512833, + "grad_norm": 2.6666111946105957, + "learning_rate": 1.8561892707178765e-06, + "loss": 0.419, + "step": 3309 + }, + { + "epoch": 0.8090931312637497, + "grad_norm": 2.8186697959899902, + "learning_rate": 1.8515971613859541e-06, + "loss": 0.5402, + "step": 3310 + }, + { + "epoch": 0.8093375702762161, + "grad_norm": 3.2007358074188232, + "learning_rate": 1.847010159795265e-06, + "loss": 2.0194, + "step": 3311 + }, + { + "epoch": 0.8095820092886825, + "grad_norm": 3.150097608566284, + "learning_rate": 1.8424282688211247e-06, + "loss": 0.6157, + "step": 3312 + }, + { + "epoch": 0.8098264483011489, + "grad_norm": 3.6688637733459473, + "learning_rate": 1.8378514913356437e-06, + "loss": 0.6479, + "step": 3313 + }, + { + "epoch": 0.8100708873136152, + "grad_norm": 2.9342734813690186, + "learning_rate": 1.833279830207726e-06, + "loss": 0.5581, + "step": 3314 + }, + { + "epoch": 0.8103153263260816, + "grad_norm": 2.9184458255767822, + "learning_rate": 1.82871328830307e-06, + "loss": 0.5614, + "step": 3315 + }, + { + "epoch": 0.810559765338548, + "grad_norm": 3.491520881652832, + "learning_rate": 1.8241518684841642e-06, + "loss": 0.5316, + "step": 3316 + }, + { + "epoch": 0.8108042043510144, + "grad_norm": 3.5354225635528564, + "learning_rate": 1.8195955736102888e-06, + "loss": 0.5644, + "step": 3317 + }, + { + "epoch": 0.8110486433634808, + "grad_norm": 3.7676517963409424, + "learning_rate": 1.8150444065375084e-06, + "loss": 0.5796, + "step": 3318 + }, + { + "epoch": 0.8112930823759472, + "grad_norm": 3.5893468856811523, + "learning_rate": 1.8104983701186751e-06, + "loss": 0.5001, + "step": 3319 + }, + { + "epoch": 0.8115375213884136, + "grad_norm": 2.6374058723449707, + "learning_rate": 1.8059574672034242e-06, + "loss": 0.5068, + "step": 3320 + }, + { + "epoch": 0.81178196040088, + "grad_norm": 3.215869188308716, + "learning_rate": 1.8014217006381728e-06, + "loss": 0.4949, + "step": 3321 + }, + { + "epoch": 0.8120263994133464, + "grad_norm": 4.275416851043701, + "learning_rate": 1.7968910732661193e-06, + "loss": 0.5926, + "step": 3322 + }, + { + "epoch": 0.8122708384258127, + "grad_norm": 3.5583596229553223, + "learning_rate": 1.7923655879272395e-06, + "loss": 0.5681, + "step": 3323 + }, + { + "epoch": 0.8125152774382791, + "grad_norm": 2.748497724533081, + "learning_rate": 1.7878452474582852e-06, + "loss": 0.626, + "step": 3324 + }, + { + "epoch": 0.8127597164507455, + "grad_norm": 3.1607143878936768, + "learning_rate": 1.783330054692789e-06, + "loss": 0.4609, + "step": 3325 + }, + { + "epoch": 0.8130041554632119, + "grad_norm": 3.6746819019317627, + "learning_rate": 1.77882001246105e-06, + "loss": 0.596, + "step": 3326 + }, + { + "epoch": 0.8132485944756783, + "grad_norm": 3.008025884628296, + "learning_rate": 1.7743151235901401e-06, + "loss": 0.5775, + "step": 3327 + }, + { + "epoch": 0.8134930334881447, + "grad_norm": 3.740696907043457, + "learning_rate": 1.769815390903904e-06, + "loss": 0.5793, + "step": 3328 + }, + { + "epoch": 0.8137374725006111, + "grad_norm": 4.715339660644531, + "learning_rate": 1.7653208172229497e-06, + "loss": 0.7491, + "step": 3329 + }, + { + "epoch": 0.8139819115130775, + "grad_norm": 3.7908732891082764, + "learning_rate": 1.7608314053646558e-06, + "loss": 0.5596, + "step": 3330 + }, + { + "epoch": 0.8142263505255439, + "grad_norm": 3.267061471939087, + "learning_rate": 1.7563471581431623e-06, + "loss": 0.5367, + "step": 3331 + }, + { + "epoch": 0.8144707895380102, + "grad_norm": 2.9402544498443604, + "learning_rate": 1.751868078369372e-06, + "loss": 0.5918, + "step": 3332 + }, + { + "epoch": 0.8147152285504766, + "grad_norm": 2.7814300060272217, + "learning_rate": 1.7473941688509511e-06, + "loss": 0.4855, + "step": 3333 + }, + { + "epoch": 0.814959667562943, + "grad_norm": 3.9253361225128174, + "learning_rate": 1.7429254323923217e-06, + "loss": 0.5666, + "step": 3334 + }, + { + "epoch": 0.8152041065754094, + "grad_norm": 3.1928324699401855, + "learning_rate": 1.7384618717946656e-06, + "loss": 0.6218, + "step": 3335 + }, + { + "epoch": 0.8154485455878758, + "grad_norm": 3.4124715328216553, + "learning_rate": 1.7340034898559178e-06, + "loss": 0.5449, + "step": 3336 + }, + { + "epoch": 0.8156929846003422, + "grad_norm": 3.0815014839172363, + "learning_rate": 1.7295502893707672e-06, + "loss": 0.5887, + "step": 3337 + }, + { + "epoch": 0.8159374236128086, + "grad_norm": 3.380671262741089, + "learning_rate": 1.72510227313066e-06, + "loss": 0.6312, + "step": 3338 + }, + { + "epoch": 0.816181862625275, + "grad_norm": 4.207968711853027, + "learning_rate": 1.7206594439237867e-06, + "loss": 0.5467, + "step": 3339 + }, + { + "epoch": 0.8164263016377414, + "grad_norm": 3.0521678924560547, + "learning_rate": 1.7162218045350886e-06, + "loss": 0.5587, + "step": 3340 + }, + { + "epoch": 0.8166707406502077, + "grad_norm": 3.4466350078582764, + "learning_rate": 1.7117893577462541e-06, + "loss": 0.5124, + "step": 3341 + }, + { + "epoch": 0.8169151796626741, + "grad_norm": 3.064985513687134, + "learning_rate": 1.7073621063357149e-06, + "loss": 0.5861, + "step": 3342 + }, + { + "epoch": 0.8171596186751405, + "grad_norm": 3.235666275024414, + "learning_rate": 1.7029400530786478e-06, + "loss": 0.642, + "step": 3343 + }, + { + "epoch": 0.8174040576876069, + "grad_norm": 3.1638059616088867, + "learning_rate": 1.698523200746972e-06, + "loss": 0.6843, + "step": 3344 + }, + { + "epoch": 0.8176484967000733, + "grad_norm": 3.068049669265747, + "learning_rate": 1.6941115521093433e-06, + "loss": 0.5732, + "step": 3345 + }, + { + "epoch": 0.8178929357125397, + "grad_norm": 4.2525506019592285, + "learning_rate": 1.6897051099311578e-06, + "loss": 0.5456, + "step": 3346 + }, + { + "epoch": 0.8181373747250061, + "grad_norm": 2.8963608741760254, + "learning_rate": 1.6853038769745466e-06, + "loss": 0.5593, + "step": 3347 + }, + { + "epoch": 0.8183818137374725, + "grad_norm": 2.282418727874756, + "learning_rate": 1.6809078559983815e-06, + "loss": 1.9292, + "step": 3348 + }, + { + "epoch": 0.8186262527499389, + "grad_norm": 3.5449211597442627, + "learning_rate": 1.6765170497582583e-06, + "loss": 0.6164, + "step": 3349 + }, + { + "epoch": 0.8188706917624052, + "grad_norm": 2.9903504848480225, + "learning_rate": 1.6721314610065098e-06, + "loss": 0.4726, + "step": 3350 + }, + { + "epoch": 0.8191151307748716, + "grad_norm": 3.019068717956543, + "learning_rate": 1.6677510924921958e-06, + "loss": 0.6304, + "step": 3351 + }, + { + "epoch": 0.819359569787338, + "grad_norm": 3.0083367824554443, + "learning_rate": 1.6633759469611033e-06, + "loss": 0.6567, + "step": 3352 + }, + { + "epoch": 0.8196040087998044, + "grad_norm": 3.0527002811431885, + "learning_rate": 1.6590060271557485e-06, + "loss": 0.5369, + "step": 3353 + }, + { + "epoch": 0.8198484478122708, + "grad_norm": 3.012636661529541, + "learning_rate": 1.654641335815369e-06, + "loss": 0.6116, + "step": 3354 + }, + { + "epoch": 0.8200928868247372, + "grad_norm": 3.206631898880005, + "learning_rate": 1.6502818756759275e-06, + "loss": 0.5986, + "step": 3355 + }, + { + "epoch": 0.8203373258372036, + "grad_norm": 3.120879888534546, + "learning_rate": 1.6459276494701037e-06, + "loss": 0.6617, + "step": 3356 + }, + { + "epoch": 0.82058176484967, + "grad_norm": 4.448589324951172, + "learning_rate": 1.6415786599272998e-06, + "loss": 0.5226, + "step": 3357 + }, + { + "epoch": 0.8208262038621364, + "grad_norm": 3.174437999725342, + "learning_rate": 1.637234909773633e-06, + "loss": 0.6756, + "step": 3358 + }, + { + "epoch": 0.8210706428746027, + "grad_norm": 2.672804117202759, + "learning_rate": 1.6328964017319415e-06, + "loss": 1.9117, + "step": 3359 + }, + { + "epoch": 0.8213150818870691, + "grad_norm": 3.058718204498291, + "learning_rate": 1.6285631385217715e-06, + "loss": 0.6112, + "step": 3360 + }, + { + "epoch": 0.8215595208995355, + "grad_norm": 3.089479923248291, + "learning_rate": 1.6242351228593833e-06, + "loss": 0.567, + "step": 3361 + }, + { + "epoch": 0.821803959912002, + "grad_norm": 3.1912524700164795, + "learning_rate": 1.6199123574577501e-06, + "loss": 0.5626, + "step": 3362 + }, + { + "epoch": 0.8220483989244683, + "grad_norm": 2.8705239295959473, + "learning_rate": 1.6155948450265501e-06, + "loss": 0.6154, + "step": 3363 + }, + { + "epoch": 0.8222928379369348, + "grad_norm": 2.5632574558258057, + "learning_rate": 1.6112825882721727e-06, + "loss": 0.4688, + "step": 3364 + }, + { + "epoch": 0.8225372769494012, + "grad_norm": 3.148527145385742, + "learning_rate": 1.6069755898977102e-06, + "loss": 0.5945, + "step": 3365 + }, + { + "epoch": 0.8227817159618676, + "grad_norm": 3.0986287593841553, + "learning_rate": 1.6026738526029584e-06, + "loss": 0.6132, + "step": 3366 + }, + { + "epoch": 0.823026154974334, + "grad_norm": 3.508009672164917, + "learning_rate": 1.5983773790844193e-06, + "loss": 0.6409, + "step": 3367 + }, + { + "epoch": 0.8232705939868002, + "grad_norm": 2.3688530921936035, + "learning_rate": 1.5940861720352897e-06, + "loss": 0.4476, + "step": 3368 + }, + { + "epoch": 0.8235150329992666, + "grad_norm": 4.0446271896362305, + "learning_rate": 1.5898002341454688e-06, + "loss": 0.5516, + "step": 3369 + }, + { + "epoch": 0.823759472011733, + "grad_norm": 3.177182197570801, + "learning_rate": 1.5855195681015522e-06, + "loss": 0.5728, + "step": 3370 + }, + { + "epoch": 0.8240039110241995, + "grad_norm": 3.1529734134674072, + "learning_rate": 1.5812441765868292e-06, + "loss": 0.5391, + "step": 3371 + }, + { + "epoch": 0.8242483500366659, + "grad_norm": 3.328282594680786, + "learning_rate": 1.5769740622812878e-06, + "loss": 0.5315, + "step": 3372 + }, + { + "epoch": 0.8244927890491323, + "grad_norm": 3.349891185760498, + "learning_rate": 1.5727092278616018e-06, + "loss": 0.4858, + "step": 3373 + }, + { + "epoch": 0.8247372280615987, + "grad_norm": 3.391483783721924, + "learning_rate": 1.5684496760011402e-06, + "loss": 0.6607, + "step": 3374 + }, + { + "epoch": 0.8249816670740651, + "grad_norm": 3.472534656524658, + "learning_rate": 1.5641954093699562e-06, + "loss": 0.6045, + "step": 3375 + }, + { + "epoch": 0.8252261060865315, + "grad_norm": 3.5358104705810547, + "learning_rate": 1.5599464306347944e-06, + "loss": 0.4378, + "step": 3376 + }, + { + "epoch": 0.8254705450989978, + "grad_norm": 2.7267298698425293, + "learning_rate": 1.5557027424590808e-06, + "loss": 0.4642, + "step": 3377 + }, + { + "epoch": 0.8257149841114642, + "grad_norm": 3.7782864570617676, + "learning_rate": 1.551464347502929e-06, + "loss": 0.5535, + "step": 3378 + }, + { + "epoch": 0.8259594231239306, + "grad_norm": 2.8620481491088867, + "learning_rate": 1.5472312484231322e-06, + "loss": 0.6043, + "step": 3379 + }, + { + "epoch": 0.826203862136397, + "grad_norm": 2.610053777694702, + "learning_rate": 1.5430034478731626e-06, + "loss": 0.5234, + "step": 3380 + }, + { + "epoch": 0.8264483011488634, + "grad_norm": 3.294963836669922, + "learning_rate": 1.5387809485031745e-06, + "loss": 0.7222, + "step": 3381 + }, + { + "epoch": 0.8266927401613298, + "grad_norm": 3.2242040634155273, + "learning_rate": 1.5345637529599976e-06, + "loss": 0.6174, + "step": 3382 + }, + { + "epoch": 0.8269371791737962, + "grad_norm": 3.8701202869415283, + "learning_rate": 1.5303518638871361e-06, + "loss": 0.614, + "step": 3383 + }, + { + "epoch": 0.8271816181862626, + "grad_norm": 2.8776087760925293, + "learning_rate": 1.5261452839247693e-06, + "loss": 0.5321, + "step": 3384 + }, + { + "epoch": 0.827426057198729, + "grad_norm": 3.0642764568328857, + "learning_rate": 1.5219440157097476e-06, + "loss": 0.4576, + "step": 3385 + }, + { + "epoch": 0.8276704962111953, + "grad_norm": 3.0332086086273193, + "learning_rate": 1.517748061875589e-06, + "loss": 0.5447, + "step": 3386 + }, + { + "epoch": 0.8279149352236617, + "grad_norm": 3.202911138534546, + "learning_rate": 1.5135574250524898e-06, + "loss": 0.5537, + "step": 3387 + }, + { + "epoch": 0.8281593742361281, + "grad_norm": 2.650573253631592, + "learning_rate": 1.509372107867303e-06, + "loss": 0.5489, + "step": 3388 + }, + { + "epoch": 0.8284038132485945, + "grad_norm": 3.583961248397827, + "learning_rate": 1.5051921129435521e-06, + "loss": 0.581, + "step": 3389 + }, + { + "epoch": 0.8286482522610609, + "grad_norm": 3.015089750289917, + "learning_rate": 1.5010174429014223e-06, + "loss": 0.4641, + "step": 3390 + }, + { + "epoch": 0.8288926912735273, + "grad_norm": 3.078062057495117, + "learning_rate": 1.4968481003577628e-06, + "loss": 0.5389, + "step": 3391 + }, + { + "epoch": 0.8291371302859937, + "grad_norm": 2.9358878135681152, + "learning_rate": 1.4926840879260818e-06, + "loss": 0.4643, + "step": 3392 + }, + { + "epoch": 0.8293815692984601, + "grad_norm": 2.809329032897949, + "learning_rate": 1.4885254082165479e-06, + "loss": 0.5133, + "step": 3393 + }, + { + "epoch": 0.8296260083109265, + "grad_norm": 3.1469004154205322, + "learning_rate": 1.4843720638359848e-06, + "loss": 0.6365, + "step": 3394 + }, + { + "epoch": 0.8298704473233928, + "grad_norm": 3.38547420501709, + "learning_rate": 1.4802240573878735e-06, + "loss": 0.4907, + "step": 3395 + }, + { + "epoch": 0.8301148863358592, + "grad_norm": 2.985200881958008, + "learning_rate": 1.4760813914723493e-06, + "loss": 0.5556, + "step": 3396 + }, + { + "epoch": 0.8303593253483256, + "grad_norm": 3.3409643173217773, + "learning_rate": 1.4719440686861975e-06, + "loss": 0.7045, + "step": 3397 + }, + { + "epoch": 0.830603764360792, + "grad_norm": 3.254333257675171, + "learning_rate": 1.4678120916228577e-06, + "loss": 0.6291, + "step": 3398 + }, + { + "epoch": 0.8308482033732584, + "grad_norm": 3.3871660232543945, + "learning_rate": 1.4636854628724152e-06, + "loss": 0.5017, + "step": 3399 + }, + { + "epoch": 0.8310926423857248, + "grad_norm": 2.778360366821289, + "learning_rate": 1.4595641850216035e-06, + "loss": 0.4855, + "step": 3400 + }, + { + "epoch": 0.8313370813981912, + "grad_norm": 3.3966450691223145, + "learning_rate": 1.4554482606538044e-06, + "loss": 0.7041, + "step": 3401 + }, + { + "epoch": 0.8315815204106576, + "grad_norm": 3.475071430206299, + "learning_rate": 1.4513376923490397e-06, + "loss": 0.7118, + "step": 3402 + }, + { + "epoch": 0.831825959423124, + "grad_norm": 2.6346867084503174, + "learning_rate": 1.447232482683979e-06, + "loss": 2.0269, + "step": 3403 + }, + { + "epoch": 0.8320703984355903, + "grad_norm": 3.1301729679107666, + "learning_rate": 1.4431326342319297e-06, + "loss": 0.5734, + "step": 3404 + }, + { + "epoch": 0.8323148374480567, + "grad_norm": 3.976536512374878, + "learning_rate": 1.439038149562838e-06, + "loss": 0.5534, + "step": 3405 + }, + { + "epoch": 0.8325592764605231, + "grad_norm": 3.441725730895996, + "learning_rate": 1.4349490312432879e-06, + "loss": 0.623, + "step": 3406 + }, + { + "epoch": 0.8328037154729895, + "grad_norm": 2.3651645183563232, + "learning_rate": 1.4308652818365031e-06, + "loss": 1.8833, + "step": 3407 + }, + { + "epoch": 0.8330481544854559, + "grad_norm": 2.689514636993408, + "learning_rate": 1.4267869039023375e-06, + "loss": 1.9719, + "step": 3408 + }, + { + "epoch": 0.8332925934979223, + "grad_norm": 2.6300761699676514, + "learning_rate": 1.4227138999972801e-06, + "loss": 0.4334, + "step": 3409 + }, + { + "epoch": 0.8335370325103887, + "grad_norm": 3.1888070106506348, + "learning_rate": 1.41864627267445e-06, + "loss": 0.5239, + "step": 3410 + }, + { + "epoch": 0.8337814715228551, + "grad_norm": 3.563365936279297, + "learning_rate": 1.4145840244835985e-06, + "loss": 0.5463, + "step": 3411 + }, + { + "epoch": 0.8340259105353215, + "grad_norm": 3.6456291675567627, + "learning_rate": 1.4105271579711021e-06, + "loss": 0.5268, + "step": 3412 + }, + { + "epoch": 0.8342703495477878, + "grad_norm": 3.0830070972442627, + "learning_rate": 1.4064756756799669e-06, + "loss": 0.7429, + "step": 3413 + }, + { + "epoch": 0.8345147885602542, + "grad_norm": 2.9476635456085205, + "learning_rate": 1.40242958014982e-06, + "loss": 0.4965, + "step": 3414 + }, + { + "epoch": 0.8347592275727206, + "grad_norm": 3.2751858234405518, + "learning_rate": 1.398388873916916e-06, + "loss": 0.7193, + "step": 3415 + }, + { + "epoch": 0.835003666585187, + "grad_norm": 3.2210662364959717, + "learning_rate": 1.3943535595141288e-06, + "loss": 0.6244, + "step": 3416 + }, + { + "epoch": 0.8352481055976534, + "grad_norm": 3.0521740913391113, + "learning_rate": 1.3903236394709507e-06, + "loss": 0.5066, + "step": 3417 + }, + { + "epoch": 0.8354925446101198, + "grad_norm": 2.9845948219299316, + "learning_rate": 1.3862991163134997e-06, + "loss": 0.4988, + "step": 3418 + }, + { + "epoch": 0.8357369836225862, + "grad_norm": 2.6232802867889404, + "learning_rate": 1.3822799925645036e-06, + "loss": 0.435, + "step": 3419 + }, + { + "epoch": 0.8359814226350526, + "grad_norm": 3.771132230758667, + "learning_rate": 1.378266270743308e-06, + "loss": 0.6209, + "step": 3420 + }, + { + "epoch": 0.836225861647519, + "grad_norm": 2.725372314453125, + "learning_rate": 1.3742579533658729e-06, + "loss": 0.6027, + "step": 3421 + }, + { + "epoch": 0.8364703006599853, + "grad_norm": 3.415480613708496, + "learning_rate": 1.3702550429447704e-06, + "loss": 0.6796, + "step": 3422 + }, + { + "epoch": 0.8367147396724517, + "grad_norm": 3.163527488708496, + "learning_rate": 1.3662575419891822e-06, + "loss": 0.6231, + "step": 3423 + }, + { + "epoch": 0.8369591786849181, + "grad_norm": 2.838290214538574, + "learning_rate": 1.3622654530049018e-06, + "loss": 0.4933, + "step": 3424 + }, + { + "epoch": 0.8372036176973845, + "grad_norm": 3.074998378753662, + "learning_rate": 1.3582787784943264e-06, + "loss": 0.574, + "step": 3425 + }, + { + "epoch": 0.8374480567098509, + "grad_norm": 2.949556589126587, + "learning_rate": 1.3542975209564624e-06, + "loss": 0.4989, + "step": 3426 + }, + { + "epoch": 0.8376924957223173, + "grad_norm": 3.073032855987549, + "learning_rate": 1.3503216828869192e-06, + "loss": 0.5717, + "step": 3427 + }, + { + "epoch": 0.8379369347347837, + "grad_norm": 3.236557722091675, + "learning_rate": 1.34635126677791e-06, + "loss": 0.6438, + "step": 3428 + }, + { + "epoch": 0.8381813737472501, + "grad_norm": 3.076246500015259, + "learning_rate": 1.342386275118247e-06, + "loss": 0.5371, + "step": 3429 + }, + { + "epoch": 0.8384258127597165, + "grad_norm": 3.4420979022979736, + "learning_rate": 1.3384267103933467e-06, + "loss": 0.5614, + "step": 3430 + }, + { + "epoch": 0.8386702517721828, + "grad_norm": 3.550089120864868, + "learning_rate": 1.3344725750852183e-06, + "loss": 0.5129, + "step": 3431 + }, + { + "epoch": 0.8389146907846492, + "grad_norm": 2.726255178451538, + "learning_rate": 1.3305238716724712e-06, + "loss": 0.5836, + "step": 3432 + }, + { + "epoch": 0.8391591297971156, + "grad_norm": 3.8522744178771973, + "learning_rate": 1.3265806026303063e-06, + "loss": 0.5566, + "step": 3433 + }, + { + "epoch": 0.839403568809582, + "grad_norm": 3.801321506500244, + "learning_rate": 1.3226427704305257e-06, + "loss": 0.653, + "step": 3434 + }, + { + "epoch": 0.8396480078220484, + "grad_norm": 3.028841733932495, + "learning_rate": 1.3187103775415156e-06, + "loss": 0.5406, + "step": 3435 + }, + { + "epoch": 0.8398924468345148, + "grad_norm": 2.624055862426758, + "learning_rate": 1.3147834264282566e-06, + "loss": 0.4619, + "step": 3436 + }, + { + "epoch": 0.8401368858469812, + "grad_norm": 3.205291271209717, + "learning_rate": 1.3108619195523153e-06, + "loss": 0.6027, + "step": 3437 + }, + { + "epoch": 0.8403813248594476, + "grad_norm": 2.5827434062957764, + "learning_rate": 1.3069458593718487e-06, + "loss": 0.5382, + "step": 3438 + }, + { + "epoch": 0.840625763871914, + "grad_norm": 2.9252283573150635, + "learning_rate": 1.3030352483415965e-06, + "loss": 0.601, + "step": 3439 + }, + { + "epoch": 0.8408702028843803, + "grad_norm": 3.566788911819458, + "learning_rate": 1.2991300889128867e-06, + "loss": 0.5668, + "step": 3440 + }, + { + "epoch": 0.8411146418968467, + "grad_norm": 3.534867286682129, + "learning_rate": 1.2952303835336256e-06, + "loss": 0.6072, + "step": 3441 + }, + { + "epoch": 0.8413590809093131, + "grad_norm": 2.8029935359954834, + "learning_rate": 1.2913361346483044e-06, + "loss": 0.5564, + "step": 3442 + }, + { + "epoch": 0.8416035199217795, + "grad_norm": 3.5112040042877197, + "learning_rate": 1.2874473446979917e-06, + "loss": 0.5928, + "step": 3443 + }, + { + "epoch": 0.8418479589342459, + "grad_norm": 3.430260181427002, + "learning_rate": 1.2835640161203344e-06, + "loss": 0.595, + "step": 3444 + }, + { + "epoch": 0.8420923979467123, + "grad_norm": 3.0886754989624023, + "learning_rate": 1.2796861513495572e-06, + "loss": 0.7083, + "step": 3445 + }, + { + "epoch": 0.8423368369591787, + "grad_norm": 2.705249547958374, + "learning_rate": 1.2758137528164593e-06, + "loss": 0.5929, + "step": 3446 + }, + { + "epoch": 0.8425812759716451, + "grad_norm": 2.9894611835479736, + "learning_rate": 1.2719468229484133e-06, + "loss": 0.4958, + "step": 3447 + }, + { + "epoch": 0.8428257149841115, + "grad_norm": 3.4993841648101807, + "learning_rate": 1.2680853641693603e-06, + "loss": 0.5782, + "step": 3448 + }, + { + "epoch": 0.8430701539965778, + "grad_norm": 4.66158390045166, + "learning_rate": 1.2642293788998206e-06, + "loss": 0.5414, + "step": 3449 + }, + { + "epoch": 0.8433145930090442, + "grad_norm": 2.817009925842285, + "learning_rate": 1.2603788695568765e-06, + "loss": 0.5355, + "step": 3450 + }, + { + "epoch": 0.8435590320215106, + "grad_norm": 2.995457649230957, + "learning_rate": 1.2565338385541792e-06, + "loss": 0.5915, + "step": 3451 + }, + { + "epoch": 0.843803471033977, + "grad_norm": 3.3391265869140625, + "learning_rate": 1.2526942883019466e-06, + "loss": 0.671, + "step": 3452 + }, + { + "epoch": 0.8440479100464434, + "grad_norm": 2.5185141563415527, + "learning_rate": 1.24886022120696e-06, + "loss": 1.9917, + "step": 3453 + }, + { + "epoch": 0.8442923490589098, + "grad_norm": 3.22603440284729, + "learning_rate": 1.2450316396725647e-06, + "loss": 0.6184, + "step": 3454 + }, + { + "epoch": 0.8445367880713762, + "grad_norm": 3.209512948989868, + "learning_rate": 1.2412085460986668e-06, + "loss": 0.6196, + "step": 3455 + }, + { + "epoch": 0.8447812270838426, + "grad_norm": 3.446316957473755, + "learning_rate": 1.2373909428817333e-06, + "loss": 0.4349, + "step": 3456 + }, + { + "epoch": 0.845025666096309, + "grad_norm": 3.9681384563446045, + "learning_rate": 1.233578832414788e-06, + "loss": 0.5544, + "step": 3457 + }, + { + "epoch": 0.8452701051087753, + "grad_norm": 2.933584213256836, + "learning_rate": 1.2297722170874138e-06, + "loss": 0.5319, + "step": 3458 + }, + { + "epoch": 0.8455145441212417, + "grad_norm": 2.817800760269165, + "learning_rate": 1.2259710992857464e-06, + "loss": 0.5167, + "step": 3459 + }, + { + "epoch": 0.8457589831337081, + "grad_norm": 3.023679256439209, + "learning_rate": 1.222175481392478e-06, + "loss": 0.5529, + "step": 3460 + }, + { + "epoch": 0.8460034221461745, + "grad_norm": 3.3433263301849365, + "learning_rate": 1.2183853657868504e-06, + "loss": 0.5096, + "step": 3461 + }, + { + "epoch": 0.8462478611586409, + "grad_norm": 3.537362813949585, + "learning_rate": 1.2146007548446593e-06, + "loss": 0.5955, + "step": 3462 + }, + { + "epoch": 0.8464923001711073, + "grad_norm": 3.5287539958953857, + "learning_rate": 1.2108216509382487e-06, + "loss": 0.6435, + "step": 3463 + }, + { + "epoch": 0.8467367391835737, + "grad_norm": 3.43084716796875, + "learning_rate": 1.2070480564365084e-06, + "loss": 0.6399, + "step": 3464 + }, + { + "epoch": 0.8469811781960401, + "grad_norm": 3.1973459720611572, + "learning_rate": 1.2032799737048795e-06, + "loss": 0.5751, + "step": 3465 + }, + { + "epoch": 0.8472256172085065, + "grad_norm": 3.478426218032837, + "learning_rate": 1.1995174051053438e-06, + "loss": 0.5172, + "step": 3466 + }, + { + "epoch": 0.8474700562209728, + "grad_norm": 3.005103588104248, + "learning_rate": 1.195760352996429e-06, + "loss": 0.5434, + "step": 3467 + }, + { + "epoch": 0.8477144952334392, + "grad_norm": 3.503016948699951, + "learning_rate": 1.1920088197332037e-06, + "loss": 0.6582, + "step": 3468 + }, + { + "epoch": 0.8479589342459056, + "grad_norm": 3.18108868598938, + "learning_rate": 1.1882628076672765e-06, + "loss": 0.5774, + "step": 3469 + }, + { + "epoch": 0.848203373258372, + "grad_norm": 3.7590348720550537, + "learning_rate": 1.184522319146797e-06, + "loss": 0.5038, + "step": 3470 + }, + { + "epoch": 0.8484478122708384, + "grad_norm": 3.0808441638946533, + "learning_rate": 1.1807873565164507e-06, + "loss": 0.5621, + "step": 3471 + }, + { + "epoch": 0.8486922512833048, + "grad_norm": 3.629685640335083, + "learning_rate": 1.1770579221174606e-06, + "loss": 0.5175, + "step": 3472 + }, + { + "epoch": 0.8489366902957712, + "grad_norm": 2.5867950916290283, + "learning_rate": 1.173334018287583e-06, + "loss": 1.8647, + "step": 3473 + }, + { + "epoch": 0.8491811293082376, + "grad_norm": 3.691901445388794, + "learning_rate": 1.1696156473611076e-06, + "loss": 0.477, + "step": 3474 + }, + { + "epoch": 0.849425568320704, + "grad_norm": 3.2322871685028076, + "learning_rate": 1.1659028116688575e-06, + "loss": 0.6267, + "step": 3475 + }, + { + "epoch": 0.8496700073331703, + "grad_norm": 7.0858259201049805, + "learning_rate": 1.1621955135381846e-06, + "loss": 0.676, + "step": 3476 + }, + { + "epoch": 0.8499144463456367, + "grad_norm": 2.888760805130005, + "learning_rate": 1.158493755292971e-06, + "loss": 0.4867, + "step": 3477 + }, + { + "epoch": 0.8501588853581031, + "grad_norm": 2.3710877895355225, + "learning_rate": 1.1547975392536237e-06, + "loss": 1.944, + "step": 3478 + }, + { + "epoch": 0.8504033243705695, + "grad_norm": 2.343113422393799, + "learning_rate": 1.151106867737075e-06, + "loss": 1.8545, + "step": 3479 + }, + { + "epoch": 0.8506477633830359, + "grad_norm": 3.9684882164001465, + "learning_rate": 1.1474217430567891e-06, + "loss": 0.6167, + "step": 3480 + }, + { + "epoch": 0.8508922023955023, + "grad_norm": 2.7422499656677246, + "learning_rate": 1.1437421675227457e-06, + "loss": 0.51, + "step": 3481 + }, + { + "epoch": 0.8511366414079687, + "grad_norm": 3.3167271614074707, + "learning_rate": 1.1400681434414474e-06, + "loss": 0.5601, + "step": 3482 + }, + { + "epoch": 0.8513810804204351, + "grad_norm": 3.22688627243042, + "learning_rate": 1.1363996731159188e-06, + "loss": 0.5387, + "step": 3483 + }, + { + "epoch": 0.8516255194329015, + "grad_norm": 2.901665449142456, + "learning_rate": 1.1327367588457005e-06, + "loss": 0.5228, + "step": 3484 + }, + { + "epoch": 0.8518699584453678, + "grad_norm": 3.9015395641326904, + "learning_rate": 1.1290794029268547e-06, + "loss": 0.6894, + "step": 3485 + }, + { + "epoch": 0.8521143974578342, + "grad_norm": 3.6681385040283203, + "learning_rate": 1.1254276076519555e-06, + "loss": 0.652, + "step": 3486 + }, + { + "epoch": 0.8523588364703006, + "grad_norm": 2.9650065898895264, + "learning_rate": 1.1217813753100914e-06, + "loss": 0.5841, + "step": 3487 + }, + { + "epoch": 0.852603275482767, + "grad_norm": 3.3798346519470215, + "learning_rate": 1.1181407081868667e-06, + "loss": 0.6314, + "step": 3488 + }, + { + "epoch": 0.8528477144952334, + "grad_norm": 3.1485350131988525, + "learning_rate": 1.1145056085643957e-06, + "loss": 0.6248, + "step": 3489 + }, + { + "epoch": 0.8530921535076998, + "grad_norm": 2.994342803955078, + "learning_rate": 1.1108760787213014e-06, + "loss": 1.967, + "step": 3490 + }, + { + "epoch": 0.8533365925201662, + "grad_norm": 2.749772071838379, + "learning_rate": 1.107252120932717e-06, + "loss": 0.5881, + "step": 3491 + }, + { + "epoch": 0.8535810315326327, + "grad_norm": 2.9013586044311523, + "learning_rate": 1.1036337374702833e-06, + "loss": 0.5159, + "step": 3492 + }, + { + "epoch": 0.853825470545099, + "grad_norm": 3.1439692974090576, + "learning_rate": 1.100020930602146e-06, + "loss": 0.5081, + "step": 3493 + }, + { + "epoch": 0.8540699095575653, + "grad_norm": 3.1041696071624756, + "learning_rate": 1.0964137025929556e-06, + "loss": 0.4956, + "step": 3494 + }, + { + "epoch": 0.8543143485700317, + "grad_norm": 3.3415110111236572, + "learning_rate": 1.0928120557038612e-06, + "loss": 0.595, + "step": 3495 + }, + { + "epoch": 0.8545587875824981, + "grad_norm": 5.222829341888428, + "learning_rate": 1.089215992192525e-06, + "loss": 0.7418, + "step": 3496 + }, + { + "epoch": 0.8548032265949645, + "grad_norm": 2.5324299335479736, + "learning_rate": 1.0856255143130967e-06, + "loss": 2.0127, + "step": 3497 + }, + { + "epoch": 0.855047665607431, + "grad_norm": 2.8876936435699463, + "learning_rate": 1.0820406243162319e-06, + "loss": 0.5061, + "step": 3498 + }, + { + "epoch": 0.8552921046198974, + "grad_norm": 2.9262752532958984, + "learning_rate": 1.0784613244490816e-06, + "loss": 0.5307, + "step": 3499 + }, + { + "epoch": 0.8555365436323638, + "grad_norm": 3.3010175228118896, + "learning_rate": 1.0748876169552902e-06, + "loss": 0.5321, + "step": 3500 + }, + { + "epoch": 0.8557809826448302, + "grad_norm": 2.5630874633789062, + "learning_rate": 1.0713195040750012e-06, + "loss": 1.9169, + "step": 3501 + }, + { + "epoch": 0.8560254216572966, + "grad_norm": 3.1026952266693115, + "learning_rate": 1.0677569880448479e-06, + "loss": 0.5615, + "step": 3502 + }, + { + "epoch": 0.8562698606697629, + "grad_norm": 3.3464791774749756, + "learning_rate": 1.0642000710979573e-06, + "loss": 0.5555, + "step": 3503 + }, + { + "epoch": 0.8565142996822293, + "grad_norm": 3.168520450592041, + "learning_rate": 1.0606487554639445e-06, + "loss": 0.568, + "step": 3504 + }, + { + "epoch": 0.8567587386946957, + "grad_norm": 5.3720855712890625, + "learning_rate": 1.0571030433689145e-06, + "loss": 0.6874, + "step": 3505 + }, + { + "epoch": 0.8570031777071621, + "grad_norm": 2.557122230529785, + "learning_rate": 1.0535629370354594e-06, + "loss": 2.0013, + "step": 3506 + }, + { + "epoch": 0.8572476167196285, + "grad_norm": 4.0707173347473145, + "learning_rate": 1.0500284386826598e-06, + "loss": 0.6974, + "step": 3507 + }, + { + "epoch": 0.8574920557320949, + "grad_norm": 2.953801393508911, + "learning_rate": 1.046499550526079e-06, + "loss": 0.5197, + "step": 3508 + }, + { + "epoch": 0.8577364947445613, + "grad_norm": 2.8796932697296143, + "learning_rate": 1.0429762747777627e-06, + "loss": 0.4481, + "step": 3509 + }, + { + "epoch": 0.8579809337570277, + "grad_norm": 2.8739702701568604, + "learning_rate": 1.0394586136462392e-06, + "loss": 0.4557, + "step": 3510 + }, + { + "epoch": 0.8582253727694941, + "grad_norm": 2.9083657264709473, + "learning_rate": 1.035946569336519e-06, + "loss": 0.5681, + "step": 3511 + }, + { + "epoch": 0.8584698117819604, + "grad_norm": 3.1416075229644775, + "learning_rate": 1.0324401440500887e-06, + "loss": 0.5766, + "step": 3512 + }, + { + "epoch": 0.8587142507944268, + "grad_norm": 3.763436794281006, + "learning_rate": 1.0289393399849156e-06, + "loss": 0.4242, + "step": 3513 + }, + { + "epoch": 0.8589586898068932, + "grad_norm": 3.609952211380005, + "learning_rate": 1.0254441593354413e-06, + "loss": 0.6058, + "step": 3514 + }, + { + "epoch": 0.8592031288193596, + "grad_norm": 2.9349989891052246, + "learning_rate": 1.0219546042925842e-06, + "loss": 0.6789, + "step": 3515 + }, + { + "epoch": 0.859447567831826, + "grad_norm": 2.800530195236206, + "learning_rate": 1.0184706770437337e-06, + "loss": 0.5381, + "step": 3516 + }, + { + "epoch": 0.8596920068442924, + "grad_norm": 3.2204642295837402, + "learning_rate": 1.0149923797727556e-06, + "loss": 0.5504, + "step": 3517 + }, + { + "epoch": 0.8599364458567588, + "grad_norm": 2.8866260051727295, + "learning_rate": 1.0115197146599843e-06, + "loss": 0.4532, + "step": 3518 + }, + { + "epoch": 0.8601808848692252, + "grad_norm": 2.6637918949127197, + "learning_rate": 1.0080526838822214e-06, + "loss": 1.9445, + "step": 3519 + }, + { + "epoch": 0.8604253238816916, + "grad_norm": 3.248420000076294, + "learning_rate": 1.0045912896127408e-06, + "loss": 0.5784, + "step": 3520 + }, + { + "epoch": 0.8606697628941579, + "grad_norm": 3.388120174407959, + "learning_rate": 1.0011355340212802e-06, + "loss": 0.549, + "step": 3521 + }, + { + "epoch": 0.8609142019066243, + "grad_norm": 2.701521635055542, + "learning_rate": 9.976854192740437e-07, + "loss": 0.4971, + "step": 3522 + }, + { + "epoch": 0.8611586409190907, + "grad_norm": 2.9408528804779053, + "learning_rate": 9.942409475337012e-07, + "loss": 0.5046, + "step": 3523 + }, + { + "epoch": 0.8614030799315571, + "grad_norm": 3.63046932220459, + "learning_rate": 9.908021209593832e-07, + "loss": 0.5875, + "step": 3524 + }, + { + "epoch": 0.8616475189440235, + "grad_norm": 2.625814437866211, + "learning_rate": 9.873689417066823e-07, + "loss": 0.4405, + "step": 3525 + }, + { + "epoch": 0.8618919579564899, + "grad_norm": 4.167357444763184, + "learning_rate": 9.839414119276501e-07, + "loss": 0.6265, + "step": 3526 + }, + { + "epoch": 0.8621363969689563, + "grad_norm": 2.6651597023010254, + "learning_rate": 9.805195337707962e-07, + "loss": 1.9388, + "step": 3527 + }, + { + "epoch": 0.8623808359814227, + "grad_norm": 4.415526866912842, + "learning_rate": 9.771033093810945e-07, + "loss": 0.6068, + "step": 3528 + }, + { + "epoch": 0.8626252749938891, + "grad_norm": 3.076693058013916, + "learning_rate": 9.736927408999653e-07, + "loss": 0.4294, + "step": 3529 + }, + { + "epoch": 0.8628697140063554, + "grad_norm": 2.883120059967041, + "learning_rate": 9.702878304652908e-07, + "loss": 0.6395, + "step": 3530 + }, + { + "epoch": 0.8631141530188218, + "grad_norm": 3.9754087924957275, + "learning_rate": 9.668885802114002e-07, + "loss": 0.6143, + "step": 3531 + }, + { + "epoch": 0.8633585920312882, + "grad_norm": 3.4229605197906494, + "learning_rate": 9.634949922690806e-07, + "loss": 0.5813, + "step": 3532 + }, + { + "epoch": 0.8636030310437546, + "grad_norm": 3.067798137664795, + "learning_rate": 9.601070687655667e-07, + "loss": 0.5549, + "step": 3533 + }, + { + "epoch": 0.863847470056221, + "grad_norm": 3.216827392578125, + "learning_rate": 9.56724811824542e-07, + "loss": 0.5268, + "step": 3534 + }, + { + "epoch": 0.8640919090686874, + "grad_norm": 2.9125301837921143, + "learning_rate": 9.533482235661395e-07, + "loss": 1.9737, + "step": 3535 + }, + { + "epoch": 0.8643363480811538, + "grad_norm": 2.989281177520752, + "learning_rate": 9.499773061069384e-07, + "loss": 0.5109, + "step": 3536 + }, + { + "epoch": 0.8645807870936202, + "grad_norm": 2.971723794937134, + "learning_rate": 9.466120615599628e-07, + "loss": 0.5246, + "step": 3537 + }, + { + "epoch": 0.8648252261060866, + "grad_norm": 2.9526774883270264, + "learning_rate": 9.432524920346819e-07, + "loss": 0.5929, + "step": 3538 + }, + { + "epoch": 0.8650696651185529, + "grad_norm": 3.23332142829895, + "learning_rate": 9.398985996370058e-07, + "loss": 0.5422, + "step": 3539 + }, + { + "epoch": 0.8653141041310193, + "grad_norm": 3.561800479888916, + "learning_rate": 9.36550386469287e-07, + "loss": 0.6373, + "step": 3540 + }, + { + "epoch": 0.8655585431434857, + "grad_norm": 3.2027857303619385, + "learning_rate": 9.33207854630317e-07, + "loss": 0.5397, + "step": 3541 + }, + { + "epoch": 0.8658029821559521, + "grad_norm": 3.2876229286193848, + "learning_rate": 9.298710062153304e-07, + "loss": 0.518, + "step": 3542 + }, + { + "epoch": 0.8660474211684185, + "grad_norm": 4.465147972106934, + "learning_rate": 9.265398433159933e-07, + "loss": 0.5502, + "step": 3543 + }, + { + "epoch": 0.8662918601808849, + "grad_norm": 4.417576313018799, + "learning_rate": 9.232143680204109e-07, + "loss": 0.5574, + "step": 3544 + }, + { + "epoch": 0.8665362991933513, + "grad_norm": 2.8578336238861084, + "learning_rate": 9.198945824131222e-07, + "loss": 0.4954, + "step": 3545 + }, + { + "epoch": 0.8667807382058177, + "grad_norm": 2.9006593227386475, + "learning_rate": 9.165804885751006e-07, + "loss": 0.5072, + "step": 3546 + }, + { + "epoch": 0.8670251772182841, + "grad_norm": 2.952204465866089, + "learning_rate": 9.132720885837509e-07, + "loss": 0.5391, + "step": 3547 + }, + { + "epoch": 0.8672696162307504, + "grad_norm": 2.539088726043701, + "learning_rate": 9.099693845129087e-07, + "loss": 1.8635, + "step": 3548 + }, + { + "epoch": 0.8675140552432168, + "grad_norm": 2.843809127807617, + "learning_rate": 9.066723784328412e-07, + "loss": 0.6246, + "step": 3549 + }, + { + "epoch": 0.8677584942556832, + "grad_norm": 3.1896400451660156, + "learning_rate": 9.03381072410241e-07, + "loss": 0.4663, + "step": 3550 + }, + { + "epoch": 0.8680029332681496, + "grad_norm": 3.474761724472046, + "learning_rate": 9.000954685082286e-07, + "loss": 0.4892, + "step": 3551 + }, + { + "epoch": 0.868247372280616, + "grad_norm": 3.1440491676330566, + "learning_rate": 8.968155687863512e-07, + "loss": 0.5951, + "step": 3552 + }, + { + "epoch": 0.8684918112930824, + "grad_norm": 3.480292797088623, + "learning_rate": 8.935413753005784e-07, + "loss": 0.6139, + "step": 3553 + }, + { + "epoch": 0.8687362503055488, + "grad_norm": 2.9116463661193848, + "learning_rate": 8.902728901033065e-07, + "loss": 0.6055, + "step": 3554 + }, + { + "epoch": 0.8689806893180152, + "grad_norm": 3.0884416103363037, + "learning_rate": 8.870101152433497e-07, + "loss": 0.5688, + "step": 3555 + }, + { + "epoch": 0.8692251283304816, + "grad_norm": 3.200479030609131, + "learning_rate": 8.837530527659443e-07, + "loss": 0.6305, + "step": 3556 + }, + { + "epoch": 0.8694695673429479, + "grad_norm": 2.9874043464660645, + "learning_rate": 8.805017047127451e-07, + "loss": 0.5621, + "step": 3557 + }, + { + "epoch": 0.8697140063554143, + "grad_norm": 3.9589836597442627, + "learning_rate": 8.772560731218293e-07, + "loss": 0.6652, + "step": 3558 + }, + { + "epoch": 0.8699584453678807, + "grad_norm": 3.272484540939331, + "learning_rate": 8.74016160027683e-07, + "loss": 0.6541, + "step": 3559 + }, + { + "epoch": 0.8702028843803471, + "grad_norm": 3.113635778427124, + "learning_rate": 8.707819674612151e-07, + "loss": 0.5202, + "step": 3560 + }, + { + "epoch": 0.8704473233928135, + "grad_norm": 3.079364776611328, + "learning_rate": 8.675534974497435e-07, + "loss": 0.5399, + "step": 3561 + }, + { + "epoch": 0.8706917624052799, + "grad_norm": 3.283508777618408, + "learning_rate": 8.643307520170018e-07, + "loss": 0.5759, + "step": 3562 + }, + { + "epoch": 0.8709362014177463, + "grad_norm": 3.144463539123535, + "learning_rate": 8.611137331831331e-07, + "loss": 0.4659, + "step": 3563 + }, + { + "epoch": 0.8711806404302127, + "grad_norm": 3.242224931716919, + "learning_rate": 8.579024429646932e-07, + "loss": 0.5753, + "step": 3564 + }, + { + "epoch": 0.8714250794426791, + "grad_norm": 3.6955604553222656, + "learning_rate": 8.54696883374646e-07, + "loss": 0.5807, + "step": 3565 + }, + { + "epoch": 0.8716695184551454, + "grad_norm": 3.042665719985962, + "learning_rate": 8.514970564223623e-07, + "loss": 0.5616, + "step": 3566 + }, + { + "epoch": 0.8719139574676118, + "grad_norm": 4.710071563720703, + "learning_rate": 8.483029641136209e-07, + "loss": 0.6149, + "step": 3567 + }, + { + "epoch": 0.8721583964800782, + "grad_norm": 2.4829370975494385, + "learning_rate": 8.451146084506046e-07, + "loss": 1.8073, + "step": 3568 + }, + { + "epoch": 0.8724028354925446, + "grad_norm": 3.0676817893981934, + "learning_rate": 8.419319914319002e-07, + "loss": 0.54, + "step": 3569 + }, + { + "epoch": 0.872647274505011, + "grad_norm": 3.504573345184326, + "learning_rate": 8.387551150524998e-07, + "loss": 0.4817, + "step": 3570 + }, + { + "epoch": 0.8728917135174774, + "grad_norm": 3.500375270843506, + "learning_rate": 8.355839813037936e-07, + "loss": 0.5654, + "step": 3571 + }, + { + "epoch": 0.8731361525299438, + "grad_norm": 2.4557383060455322, + "learning_rate": 8.324185921735728e-07, + "loss": 1.9117, + "step": 3572 + }, + { + "epoch": 0.8733805915424102, + "grad_norm": 2.697639226913452, + "learning_rate": 8.292589496460323e-07, + "loss": 0.4705, + "step": 3573 + }, + { + "epoch": 0.8736250305548766, + "grad_norm": 3.3228890895843506, + "learning_rate": 8.26105055701758e-07, + "loss": 0.5901, + "step": 3574 + }, + { + "epoch": 0.8738694695673429, + "grad_norm": 3.011554479598999, + "learning_rate": 8.229569123177361e-07, + "loss": 0.6151, + "step": 3575 + }, + { + "epoch": 0.8741139085798093, + "grad_norm": 2.707702875137329, + "learning_rate": 8.198145214673481e-07, + "loss": 0.5116, + "step": 3576 + }, + { + "epoch": 0.8743583475922757, + "grad_norm": 3.353950023651123, + "learning_rate": 8.166778851203672e-07, + "loss": 0.646, + "step": 3577 + }, + { + "epoch": 0.8746027866047421, + "grad_norm": 2.3703274726867676, + "learning_rate": 8.135470052429628e-07, + "loss": 1.8762, + "step": 3578 + }, + { + "epoch": 0.8748472256172085, + "grad_norm": 3.6735572814941406, + "learning_rate": 8.10421883797694e-07, + "loss": 0.6678, + "step": 3579 + }, + { + "epoch": 0.8750916646296749, + "grad_norm": 3.0874695777893066, + "learning_rate": 8.073025227435093e-07, + "loss": 0.6327, + "step": 3580 + }, + { + "epoch": 0.8753361036421413, + "grad_norm": 2.686279296875, + "learning_rate": 8.041889240357493e-07, + "loss": 0.5964, + "step": 3581 + }, + { + "epoch": 0.8755805426546077, + "grad_norm": 3.0673916339874268, + "learning_rate": 8.0108108962614e-07, + "loss": 0.6688, + "step": 3582 + }, + { + "epoch": 0.875824981667074, + "grad_norm": 3.7942118644714355, + "learning_rate": 7.979790214627937e-07, + "loss": 0.5957, + "step": 3583 + }, + { + "epoch": 0.8760694206795404, + "grad_norm": 4.573256015777588, + "learning_rate": 7.948827214902122e-07, + "loss": 0.5473, + "step": 3584 + }, + { + "epoch": 0.8763138596920068, + "grad_norm": 3.371189832687378, + "learning_rate": 7.917921916492765e-07, + "loss": 0.5922, + "step": 3585 + }, + { + "epoch": 0.8765582987044732, + "grad_norm": 3.7442123889923096, + "learning_rate": 7.887074338772538e-07, + "loss": 0.5704, + "step": 3586 + }, + { + "epoch": 0.8768027377169396, + "grad_norm": 3.203345775604248, + "learning_rate": 7.856284501077927e-07, + "loss": 0.5671, + "step": 3587 + }, + { + "epoch": 0.877047176729406, + "grad_norm": 2.4070796966552734, + "learning_rate": 7.825552422709192e-07, + "loss": 0.4202, + "step": 3588 + }, + { + "epoch": 0.8772916157418724, + "grad_norm": 2.3015387058258057, + "learning_rate": 7.79487812293046e-07, + "loss": 1.8715, + "step": 3589 + }, + { + "epoch": 0.8775360547543388, + "grad_norm": 3.0606741905212402, + "learning_rate": 7.764261620969571e-07, + "loss": 0.6387, + "step": 3590 + }, + { + "epoch": 0.8777804937668052, + "grad_norm": 2.820016622543335, + "learning_rate": 7.733702936018162e-07, + "loss": 0.4976, + "step": 3591 + }, + { + "epoch": 0.8780249327792715, + "grad_norm": 3.0214173793792725, + "learning_rate": 7.703202087231632e-07, + "loss": 0.5504, + "step": 3592 + }, + { + "epoch": 0.8782693717917379, + "grad_norm": 3.4106690883636475, + "learning_rate": 7.672759093729121e-07, + "loss": 0.5901, + "step": 3593 + }, + { + "epoch": 0.8785138108042043, + "grad_norm": 2.8477838039398193, + "learning_rate": 7.642373974593487e-07, + "loss": 0.4521, + "step": 3594 + }, + { + "epoch": 0.8787582498166707, + "grad_norm": 3.21529221534729, + "learning_rate": 7.612046748871327e-07, + "loss": 0.6287, + "step": 3595 + }, + { + "epoch": 0.8790026888291371, + "grad_norm": 2.869361639022827, + "learning_rate": 7.581777435572956e-07, + "loss": 0.5404, + "step": 3596 + }, + { + "epoch": 0.8792471278416035, + "grad_norm": 2.8519232273101807, + "learning_rate": 7.551566053672377e-07, + "loss": 0.4786, + "step": 3597 + }, + { + "epoch": 0.8794915668540699, + "grad_norm": 2.9038546085357666, + "learning_rate": 7.521412622107271e-07, + "loss": 0.5005, + "step": 3598 + }, + { + "epoch": 0.8797360058665363, + "grad_norm": 11.24041748046875, + "learning_rate": 7.49131715977901e-07, + "loss": 0.5871, + "step": 3599 + }, + { + "epoch": 0.8799804448790027, + "grad_norm": 2.5400288105010986, + "learning_rate": 7.461279685552614e-07, + "loss": 1.9806, + "step": 3600 + }, + { + "epoch": 0.880224883891469, + "grad_norm": 2.685638666152954, + "learning_rate": 7.431300218256754e-07, + "loss": 0.5443, + "step": 3601 + }, + { + "epoch": 0.8804693229039354, + "grad_norm": 3.251020669937134, + "learning_rate": 7.401378776683754e-07, + "loss": 0.614, + "step": 3602 + }, + { + "epoch": 0.8807137619164018, + "grad_norm": 2.709386110305786, + "learning_rate": 7.371515379589555e-07, + "loss": 0.4861, + "step": 3603 + }, + { + "epoch": 0.8809582009288682, + "grad_norm": 4.262940406799316, + "learning_rate": 7.341710045693717e-07, + "loss": 0.6105, + "step": 3604 + }, + { + "epoch": 0.8812026399413346, + "grad_norm": 3.71225643157959, + "learning_rate": 7.311962793679417e-07, + "loss": 0.7028, + "step": 3605 + }, + { + "epoch": 0.881447078953801, + "grad_norm": 2.9761314392089844, + "learning_rate": 7.282273642193383e-07, + "loss": 0.6167, + "step": 3606 + }, + { + "epoch": 0.8816915179662674, + "grad_norm": 2.59669828414917, + "learning_rate": 7.252642609845973e-07, + "loss": 0.4977, + "step": 3607 + }, + { + "epoch": 0.8819359569787338, + "grad_norm": 3.502819776535034, + "learning_rate": 7.223069715211073e-07, + "loss": 0.6519, + "step": 3608 + }, + { + "epoch": 0.8821803959912002, + "grad_norm": 2.915161609649658, + "learning_rate": 7.193554976826145e-07, + "loss": 0.5265, + "step": 3609 + }, + { + "epoch": 0.8824248350036665, + "grad_norm": 3.6396989822387695, + "learning_rate": 7.16409841319221e-07, + "loss": 0.6412, + "step": 3610 + }, + { + "epoch": 0.8826692740161329, + "grad_norm": 3.3161749839782715, + "learning_rate": 7.13470004277379e-07, + "loss": 0.6229, + "step": 3611 + }, + { + "epoch": 0.8829137130285993, + "grad_norm": 4.260815620422363, + "learning_rate": 7.105359883998953e-07, + "loss": 0.4812, + "step": 3612 + }, + { + "epoch": 0.8831581520410657, + "grad_norm": 4.785272598266602, + "learning_rate": 7.076077955259275e-07, + "loss": 0.5726, + "step": 3613 + }, + { + "epoch": 0.8834025910535321, + "grad_norm": 2.9758827686309814, + "learning_rate": 7.046854274909809e-07, + "loss": 0.4628, + "step": 3614 + }, + { + "epoch": 0.8836470300659985, + "grad_norm": 2.9017090797424316, + "learning_rate": 7.017688861269145e-07, + "loss": 0.5333, + "step": 3615 + }, + { + "epoch": 0.8838914690784649, + "grad_norm": 3.9723899364471436, + "learning_rate": 6.988581732619293e-07, + "loss": 0.5544, + "step": 3616 + }, + { + "epoch": 0.8841359080909313, + "grad_norm": 3.3946757316589355, + "learning_rate": 6.959532907205758e-07, + "loss": 0.6218, + "step": 3617 + }, + { + "epoch": 0.8843803471033977, + "grad_norm": 2.7967257499694824, + "learning_rate": 6.930542403237506e-07, + "loss": 0.5608, + "step": 3618 + }, + { + "epoch": 0.884624786115864, + "grad_norm": 2.9319894313812256, + "learning_rate": 6.901610238886891e-07, + "loss": 0.5659, + "step": 3619 + }, + { + "epoch": 0.8848692251283304, + "grad_norm": 3.273327350616455, + "learning_rate": 6.872736432289794e-07, + "loss": 0.6086, + "step": 3620 + }, + { + "epoch": 0.8851136641407968, + "grad_norm": 3.4320311546325684, + "learning_rate": 6.843921001545429e-07, + "loss": 0.5059, + "step": 3621 + }, + { + "epoch": 0.8853581031532632, + "grad_norm": 3.1227705478668213, + "learning_rate": 6.815163964716453e-07, + "loss": 0.4984, + "step": 3622 + }, + { + "epoch": 0.8856025421657296, + "grad_norm": 3.0057742595672607, + "learning_rate": 6.786465339828918e-07, + "loss": 0.6151, + "step": 3623 + }, + { + "epoch": 0.885846981178196, + "grad_norm": 3.298367500305176, + "learning_rate": 6.757825144872255e-07, + "loss": 0.6156, + "step": 3624 + }, + { + "epoch": 0.8860914201906624, + "grad_norm": 3.862901210784912, + "learning_rate": 6.729243397799268e-07, + "loss": 0.6285, + "step": 3625 + }, + { + "epoch": 0.8863358592031289, + "grad_norm": 3.6718273162841797, + "learning_rate": 6.700720116526116e-07, + "loss": 0.4351, + "step": 3626 + }, + { + "epoch": 0.8865802982155953, + "grad_norm": 3.213244676589966, + "learning_rate": 6.672255318932342e-07, + "loss": 0.5081, + "step": 3627 + }, + { + "epoch": 0.8868247372280615, + "grad_norm": 3.0694515705108643, + "learning_rate": 6.643849022860794e-07, + "loss": 0.6884, + "step": 3628 + }, + { + "epoch": 0.887069176240528, + "grad_norm": 3.5863587856292725, + "learning_rate": 6.615501246117673e-07, + "loss": 0.6471, + "step": 3629 + }, + { + "epoch": 0.8873136152529943, + "grad_norm": 3.412667989730835, + "learning_rate": 6.587212006472465e-07, + "loss": 0.5116, + "step": 3630 + }, + { + "epoch": 0.8875580542654607, + "grad_norm": 3.131528615951538, + "learning_rate": 6.558981321658009e-07, + "loss": 0.529, + "step": 3631 + }, + { + "epoch": 0.8878024932779272, + "grad_norm": 3.2789711952209473, + "learning_rate": 6.530809209370403e-07, + "loss": 0.5665, + "step": 3632 + }, + { + "epoch": 0.8880469322903936, + "grad_norm": 3.286752700805664, + "learning_rate": 6.502695687269034e-07, + "loss": 0.661, + "step": 3633 + }, + { + "epoch": 0.88829137130286, + "grad_norm": 3.495955467224121, + "learning_rate": 6.474640772976571e-07, + "loss": 0.6199, + "step": 3634 + }, + { + "epoch": 0.8885358103153264, + "grad_norm": 2.993870973587036, + "learning_rate": 6.446644484078968e-07, + "loss": 0.5167, + "step": 3635 + }, + { + "epoch": 0.8887802493277928, + "grad_norm": 2.757718324661255, + "learning_rate": 6.418706838125388e-07, + "loss": 0.5215, + "step": 3636 + }, + { + "epoch": 0.889024688340259, + "grad_norm": 3.5278823375701904, + "learning_rate": 6.390827852628256e-07, + "loss": 0.5353, + "step": 3637 + }, + { + "epoch": 0.8892691273527255, + "grad_norm": 3.1759135723114014, + "learning_rate": 6.363007545063227e-07, + "loss": 0.5631, + "step": 3638 + }, + { + "epoch": 0.8895135663651919, + "grad_norm": 2.910161256790161, + "learning_rate": 6.33524593286915e-07, + "loss": 0.6147, + "step": 3639 + }, + { + "epoch": 0.8897580053776583, + "grad_norm": 3.1023874282836914, + "learning_rate": 6.307543033448127e-07, + "loss": 0.5945, + "step": 3640 + }, + { + "epoch": 0.8900024443901247, + "grad_norm": 40.76762390136719, + "learning_rate": 6.279898864165423e-07, + "loss": 0.5745, + "step": 3641 + }, + { + "epoch": 0.8902468834025911, + "grad_norm": 4.0469231605529785, + "learning_rate": 6.252313442349489e-07, + "loss": 0.5816, + "step": 3642 + }, + { + "epoch": 0.8904913224150575, + "grad_norm": 3.096689462661743, + "learning_rate": 6.22478678529197e-07, + "loss": 0.5289, + "step": 3643 + }, + { + "epoch": 0.8907357614275239, + "grad_norm": 3.436739921569824, + "learning_rate": 6.197318910247652e-07, + "loss": 0.6384, + "step": 3644 + }, + { + "epoch": 0.8909802004399903, + "grad_norm": 3.6540892124176025, + "learning_rate": 6.169909834434484e-07, + "loss": 0.6279, + "step": 3645 + }, + { + "epoch": 0.8912246394524566, + "grad_norm": 3.2602572441101074, + "learning_rate": 6.14255957503358e-07, + "loss": 0.5458, + "step": 3646 + }, + { + "epoch": 0.891469078464923, + "grad_norm": 3.537839651107788, + "learning_rate": 6.115268149189147e-07, + "loss": 0.6928, + "step": 3647 + }, + { + "epoch": 0.8917135174773894, + "grad_norm": 6.434822082519531, + "learning_rate": 6.088035574008532e-07, + "loss": 0.57, + "step": 3648 + }, + { + "epoch": 0.8919579564898558, + "grad_norm": 2.3490285873413086, + "learning_rate": 6.060861866562196e-07, + "loss": 1.9858, + "step": 3649 + }, + { + "epoch": 0.8922023955023222, + "grad_norm": 3.2424468994140625, + "learning_rate": 6.033747043883675e-07, + "loss": 0.6247, + "step": 3650 + }, + { + "epoch": 0.8924468345147886, + "grad_norm": 3.5415916442871094, + "learning_rate": 6.006691122969644e-07, + "loss": 0.5527, + "step": 3651 + }, + { + "epoch": 0.892691273527255, + "grad_norm": 4.1693572998046875, + "learning_rate": 5.979694120779811e-07, + "loss": 0.6999, + "step": 3652 + }, + { + "epoch": 0.8929357125397214, + "grad_norm": 3.3961987495422363, + "learning_rate": 5.952756054236953e-07, + "loss": 0.4981, + "step": 3653 + }, + { + "epoch": 0.8931801515521878, + "grad_norm": 3.0379087924957275, + "learning_rate": 5.925876940226926e-07, + "loss": 0.7162, + "step": 3654 + }, + { + "epoch": 0.8934245905646541, + "grad_norm": 3.3472681045532227, + "learning_rate": 5.899056795598635e-07, + "loss": 0.5355, + "step": 3655 + }, + { + "epoch": 0.8936690295771205, + "grad_norm": 3.012376308441162, + "learning_rate": 5.872295637163983e-07, + "loss": 0.5354, + "step": 3656 + }, + { + "epoch": 0.8939134685895869, + "grad_norm": 2.4523651599884033, + "learning_rate": 5.845593481697931e-07, + "loss": 1.8777, + "step": 3657 + }, + { + "epoch": 0.8941579076020533, + "grad_norm": 3.6199769973754883, + "learning_rate": 5.818950345938457e-07, + "loss": 0.6562, + "step": 3658 + }, + { + "epoch": 0.8944023466145197, + "grad_norm": 3.089425563812256, + "learning_rate": 5.792366246586512e-07, + "loss": 0.5348, + "step": 3659 + }, + { + "epoch": 0.8946467856269861, + "grad_norm": 3.0774762630462646, + "learning_rate": 5.765841200306088e-07, + "loss": 0.7085, + "step": 3660 + }, + { + "epoch": 0.8948912246394525, + "grad_norm": 3.439373731613159, + "learning_rate": 5.739375223724108e-07, + "loss": 0.6117, + "step": 3661 + }, + { + "epoch": 0.8951356636519189, + "grad_norm": 2.7386152744293213, + "learning_rate": 5.712968333430502e-07, + "loss": 0.6133, + "step": 3662 + }, + { + "epoch": 0.8953801026643853, + "grad_norm": 3.5213887691497803, + "learning_rate": 5.686620545978161e-07, + "loss": 0.5565, + "step": 3663 + }, + { + "epoch": 0.8956245416768516, + "grad_norm": 3.4529879093170166, + "learning_rate": 5.660331877882897e-07, + "loss": 0.6187, + "step": 3664 + }, + { + "epoch": 0.895868980689318, + "grad_norm": 2.5605435371398926, + "learning_rate": 5.634102345623526e-07, + "loss": 0.5098, + "step": 3665 + }, + { + "epoch": 0.8961134197017844, + "grad_norm": 3.1192750930786133, + "learning_rate": 5.607931965641733e-07, + "loss": 0.4693, + "step": 3666 + }, + { + "epoch": 0.8963578587142508, + "grad_norm": 2.7218759059906006, + "learning_rate": 5.581820754342138e-07, + "loss": 0.4582, + "step": 3667 + }, + { + "epoch": 0.8966022977267172, + "grad_norm": 2.4536046981811523, + "learning_rate": 5.555768728092304e-07, + "loss": 1.8286, + "step": 3668 + }, + { + "epoch": 0.8968467367391836, + "grad_norm": 3.203338384628296, + "learning_rate": 5.529775903222656e-07, + "loss": 0.5028, + "step": 3669 + }, + { + "epoch": 0.89709117575165, + "grad_norm": 4.010106563568115, + "learning_rate": 5.50384229602653e-07, + "loss": 0.6932, + "step": 3670 + }, + { + "epoch": 0.8973356147641164, + "grad_norm": 3.7111124992370605, + "learning_rate": 5.477967922760141e-07, + "loss": 0.6522, + "step": 3671 + }, + { + "epoch": 0.8975800537765828, + "grad_norm": 3.3121304512023926, + "learning_rate": 5.452152799642551e-07, + "loss": 0.5106, + "step": 3672 + }, + { + "epoch": 0.8978244927890491, + "grad_norm": 2.946337938308716, + "learning_rate": 5.426396942855727e-07, + "loss": 0.4833, + "step": 3673 + }, + { + "epoch": 0.8980689318015155, + "grad_norm": 3.6235878467559814, + "learning_rate": 5.400700368544442e-07, + "loss": 0.6588, + "step": 3674 + }, + { + "epoch": 0.8983133708139819, + "grad_norm": 3.8211817741394043, + "learning_rate": 5.375063092816313e-07, + "loss": 0.5454, + "step": 3675 + }, + { + "epoch": 0.8985578098264483, + "grad_norm": 3.652818202972412, + "learning_rate": 5.349485131741828e-07, + "loss": 0.6498, + "step": 3676 + }, + { + "epoch": 0.8988022488389147, + "grad_norm": 3.070875406265259, + "learning_rate": 5.323966501354261e-07, + "loss": 0.4917, + "step": 3677 + }, + { + "epoch": 0.8990466878513811, + "grad_norm": 2.9537672996520996, + "learning_rate": 5.298507217649684e-07, + "loss": 0.525, + "step": 3678 + }, + { + "epoch": 0.8992911268638475, + "grad_norm": 3.348336935043335, + "learning_rate": 5.273107296586987e-07, + "loss": 0.5026, + "step": 3679 + }, + { + "epoch": 0.8995355658763139, + "grad_norm": 3.548139810562134, + "learning_rate": 5.247766754087846e-07, + "loss": 0.5901, + "step": 3680 + }, + { + "epoch": 0.8997800048887803, + "grad_norm": 3.775935411453247, + "learning_rate": 5.222485606036709e-07, + "loss": 0.6468, + "step": 3681 + }, + { + "epoch": 0.9000244439012466, + "grad_norm": 3.2515830993652344, + "learning_rate": 5.197263868280821e-07, + "loss": 0.66, + "step": 3682 + }, + { + "epoch": 0.900268882913713, + "grad_norm": 2.97196888923645, + "learning_rate": 5.172101556630149e-07, + "loss": 0.5048, + "step": 3683 + }, + { + "epoch": 0.9005133219261794, + "grad_norm": 3.2715892791748047, + "learning_rate": 5.146998686857419e-07, + "loss": 0.5409, + "step": 3684 + }, + { + "epoch": 0.9007577609386458, + "grad_norm": 3.282442092895508, + "learning_rate": 5.121955274698098e-07, + "loss": 0.5975, + "step": 3685 + }, + { + "epoch": 0.9010021999511122, + "grad_norm": 3.06097674369812, + "learning_rate": 5.09697133585042e-07, + "loss": 0.5095, + "step": 3686 + }, + { + "epoch": 0.9012466389635786, + "grad_norm": 3.0278491973876953, + "learning_rate": 5.072046885975291e-07, + "loss": 0.5366, + "step": 3687 + }, + { + "epoch": 0.901491077976045, + "grad_norm": 4.123024940490723, + "learning_rate": 5.047181940696333e-07, + "loss": 0.5028, + "step": 3688 + }, + { + "epoch": 0.9017355169885114, + "grad_norm": 3.111257553100586, + "learning_rate": 5.02237651559988e-07, + "loss": 0.5759, + "step": 3689 + }, + { + "epoch": 0.9019799560009778, + "grad_norm": 3.1422526836395264, + "learning_rate": 4.997630626234962e-07, + "loss": 0.4266, + "step": 3690 + }, + { + "epoch": 0.9022243950134441, + "grad_norm": 3.3071560859680176, + "learning_rate": 4.972944288113268e-07, + "loss": 0.6064, + "step": 3691 + }, + { + "epoch": 0.9024688340259105, + "grad_norm": 2.9000163078308105, + "learning_rate": 4.948317516709177e-07, + "loss": 1.9377, + "step": 3692 + }, + { + "epoch": 0.9027132730383769, + "grad_norm": 2.9188854694366455, + "learning_rate": 4.923750327459731e-07, + "loss": 0.5908, + "step": 3693 + }, + { + "epoch": 0.9029577120508433, + "grad_norm": 3.1524181365966797, + "learning_rate": 4.899242735764609e-07, + "loss": 0.563, + "step": 3694 + }, + { + "epoch": 0.9032021510633097, + "grad_norm": 4.9470367431640625, + "learning_rate": 4.874794756986145e-07, + "loss": 0.5097, + "step": 3695 + }, + { + "epoch": 0.9034465900757761, + "grad_norm": 3.1351184844970703, + "learning_rate": 4.85040640644927e-07, + "loss": 0.5333, + "step": 3696 + }, + { + "epoch": 0.9036910290882425, + "grad_norm": 3.649569034576416, + "learning_rate": 4.826077699441611e-07, + "loss": 0.6437, + "step": 3697 + }, + { + "epoch": 0.9039354681007089, + "grad_norm": 2.842440366744995, + "learning_rate": 4.801808651213346e-07, + "loss": 0.5591, + "step": 3698 + }, + { + "epoch": 0.9041799071131753, + "grad_norm": 2.9804258346557617, + "learning_rate": 4.777599276977262e-07, + "loss": 0.6262, + "step": 3699 + }, + { + "epoch": 0.9044243461256416, + "grad_norm": 4.149869918823242, + "learning_rate": 4.7534495919087767e-07, + "loss": 0.6289, + "step": 3700 + }, + { + "epoch": 0.904668785138108, + "grad_norm": 2.6993746757507324, + "learning_rate": 4.729359611145845e-07, + "loss": 1.899, + "step": 3701 + }, + { + "epoch": 0.9049132241505744, + "grad_norm": 2.8985257148742676, + "learning_rate": 4.705329349789034e-07, + "loss": 0.4559, + "step": 3702 + }, + { + "epoch": 0.9051576631630408, + "grad_norm": 2.9051637649536133, + "learning_rate": 4.6813588229014605e-07, + "loss": 0.5752, + "step": 3703 + }, + { + "epoch": 0.9054021021755072, + "grad_norm": 3.1530659198760986, + "learning_rate": 4.6574480455087836e-07, + "loss": 0.5387, + "step": 3704 + }, + { + "epoch": 0.9056465411879736, + "grad_norm": 3.1871185302734375, + "learning_rate": 4.6335970325992463e-07, + "loss": 0.5843, + "step": 3705 + }, + { + "epoch": 0.90589098020044, + "grad_norm": 2.9980342388153076, + "learning_rate": 4.60980579912359e-07, + "loss": 0.5831, + "step": 3706 + }, + { + "epoch": 0.9061354192129064, + "grad_norm": 3.6054625511169434, + "learning_rate": 4.5860743599951186e-07, + "loss": 0.5821, + "step": 3707 + }, + { + "epoch": 0.9063798582253728, + "grad_norm": 3.908937931060791, + "learning_rate": 4.5624027300896214e-07, + "loss": 0.5487, + "step": 3708 + }, + { + "epoch": 0.9066242972378391, + "grad_norm": 3.200122833251953, + "learning_rate": 4.538790924245418e-07, + "loss": 0.4827, + "step": 3709 + }, + { + "epoch": 0.9068687362503055, + "grad_norm": 3.2075066566467285, + "learning_rate": 4.515238957263313e-07, + "loss": 0.5491, + "step": 3710 + }, + { + "epoch": 0.9071131752627719, + "grad_norm": 2.8623270988464355, + "learning_rate": 4.49174684390663e-07, + "loss": 0.6853, + "step": 3711 + }, + { + "epoch": 0.9073576142752383, + "grad_norm": 3.1399548053741455, + "learning_rate": 4.4683145989011356e-07, + "loss": 0.6352, + "step": 3712 + }, + { + "epoch": 0.9076020532877047, + "grad_norm": 3.4346516132354736, + "learning_rate": 4.4449422369351013e-07, + "loss": 0.7099, + "step": 3713 + }, + { + "epoch": 0.9078464923001711, + "grad_norm": 2.989849805831909, + "learning_rate": 4.421629772659242e-07, + "loss": 0.4438, + "step": 3714 + }, + { + "epoch": 0.9080909313126375, + "grad_norm": 3.322847604751587, + "learning_rate": 4.3983772206867446e-07, + "loss": 0.6186, + "step": 3715 + }, + { + "epoch": 0.9083353703251039, + "grad_norm": 3.3168976306915283, + "learning_rate": 4.375184595593207e-07, + "loss": 0.4834, + "step": 3716 + }, + { + "epoch": 0.9085798093375703, + "grad_norm": 3.5158891677856445, + "learning_rate": 4.35205191191671e-07, + "loss": 0.6321, + "step": 3717 + }, + { + "epoch": 0.9088242483500366, + "grad_norm": 2.9080677032470703, + "learning_rate": 4.3289791841577113e-07, + "loss": 0.5509, + "step": 3718 + }, + { + "epoch": 0.909068687362503, + "grad_norm": 3.4545187950134277, + "learning_rate": 4.305966426779118e-07, + "loss": 0.5999, + "step": 3719 + }, + { + "epoch": 0.9093131263749694, + "grad_norm": 3.0747053623199463, + "learning_rate": 4.283013654206236e-07, + "loss": 0.5349, + "step": 3720 + }, + { + "epoch": 0.9095575653874358, + "grad_norm": 2.9126248359680176, + "learning_rate": 4.260120880826768e-07, + "loss": 0.5253, + "step": 3721 + }, + { + "epoch": 0.9098020043999022, + "grad_norm": 3.5154640674591064, + "learning_rate": 4.2372881209908234e-07, + "loss": 0.6083, + "step": 3722 + }, + { + "epoch": 0.9100464434123686, + "grad_norm": 3.2760703563690186, + "learning_rate": 4.214515389010865e-07, + "loss": 0.4783, + "step": 3723 + }, + { + "epoch": 0.910290882424835, + "grad_norm": 3.1350185871124268, + "learning_rate": 4.1918026991617513e-07, + "loss": 0.4299, + "step": 3724 + }, + { + "epoch": 0.9105353214373014, + "grad_norm": 3.5196869373321533, + "learning_rate": 4.1691500656806825e-07, + "loss": 0.557, + "step": 3725 + }, + { + "epoch": 0.9107797604497678, + "grad_norm": 3.4376039505004883, + "learning_rate": 4.146557502767246e-07, + "loss": 0.5684, + "step": 3726 + }, + { + "epoch": 0.9110241994622341, + "grad_norm": 2.774477243423462, + "learning_rate": 4.1240250245833226e-07, + "loss": 0.4675, + "step": 3727 + }, + { + "epoch": 0.9112686384747005, + "grad_norm": 2.8976266384124756, + "learning_rate": 4.101552645253182e-07, + "loss": 0.5628, + "step": 3728 + }, + { + "epoch": 0.9115130774871669, + "grad_norm": 3.259958505630493, + "learning_rate": 4.079140378863411e-07, + "loss": 0.5347, + "step": 3729 + }, + { + "epoch": 0.9117575164996333, + "grad_norm": 3.4463083744049072, + "learning_rate": 4.056788239462894e-07, + "loss": 0.6018, + "step": 3730 + }, + { + "epoch": 0.9120019555120997, + "grad_norm": 3.383847951889038, + "learning_rate": 4.034496241062824e-07, + "loss": 0.473, + "step": 3731 + }, + { + "epoch": 0.9122463945245661, + "grad_norm": 3.5528504848480225, + "learning_rate": 4.0122643976367225e-07, + "loss": 0.5487, + "step": 3732 + }, + { + "epoch": 0.9124908335370325, + "grad_norm": 3.103224039077759, + "learning_rate": 3.9900927231203866e-07, + "loss": 0.7557, + "step": 3733 + }, + { + "epoch": 0.9127352725494989, + "grad_norm": 3.223209857940674, + "learning_rate": 3.967981231411888e-07, + "loss": 0.5294, + "step": 3734 + }, + { + "epoch": 0.9129797115619653, + "grad_norm": 2.878547191619873, + "learning_rate": 3.945929936371595e-07, + "loss": 0.5665, + "step": 3735 + }, + { + "epoch": 0.9132241505744316, + "grad_norm": 3.334857225418091, + "learning_rate": 3.9239388518221287e-07, + "loss": 0.4665, + "step": 3736 + }, + { + "epoch": 0.913468589586898, + "grad_norm": 3.6171045303344727, + "learning_rate": 3.9020079915483733e-07, + "loss": 0.5931, + "step": 3737 + }, + { + "epoch": 0.9137130285993644, + "grad_norm": 3.136127471923828, + "learning_rate": 3.8801373692974655e-07, + "loss": 0.549, + "step": 3738 + }, + { + "epoch": 0.9139574676118308, + "grad_norm": 4.839405536651611, + "learning_rate": 3.8583269987787607e-07, + "loss": 0.5256, + "step": 3739 + }, + { + "epoch": 0.9142019066242972, + "grad_norm": 3.2706949710845947, + "learning_rate": 3.836576893663868e-07, + "loss": 0.5917, + "step": 3740 + }, + { + "epoch": 0.9144463456367636, + "grad_norm": 2.8269474506378174, + "learning_rate": 3.8148870675866145e-07, + "loss": 0.5479, + "step": 3741 + }, + { + "epoch": 0.91469078464923, + "grad_norm": 3.574561357498169, + "learning_rate": 3.7932575341430467e-07, + "loss": 0.6755, + "step": 3742 + }, + { + "epoch": 0.9149352236616964, + "grad_norm": 4.047536849975586, + "learning_rate": 3.7716883068913857e-07, + "loss": 0.5917, + "step": 3743 + }, + { + "epoch": 0.9151796626741628, + "grad_norm": 3.0007340908050537, + "learning_rate": 3.7501793993521164e-07, + "loss": 0.5201, + "step": 3744 + }, + { + "epoch": 0.9154241016866291, + "grad_norm": 3.1705756187438965, + "learning_rate": 3.7287308250078535e-07, + "loss": 0.6814, + "step": 3745 + }, + { + "epoch": 0.9156685406990955, + "grad_norm": 6.387424468994141, + "learning_rate": 3.7073425973034095e-07, + "loss": 0.6554, + "step": 3746 + }, + { + "epoch": 0.9159129797115619, + "grad_norm": 3.1905219554901123, + "learning_rate": 3.6860147296457814e-07, + "loss": 0.5053, + "step": 3747 + }, + { + "epoch": 0.9161574187240283, + "grad_norm": 5.621198654174805, + "learning_rate": 3.664747235404109e-07, + "loss": 0.5023, + "step": 3748 + }, + { + "epoch": 0.9164018577364947, + "grad_norm": 3.318756341934204, + "learning_rate": 3.643540127909695e-07, + "loss": 0.5744, + "step": 3749 + }, + { + "epoch": 0.9166462967489611, + "grad_norm": 2.960386276245117, + "learning_rate": 3.6223934204560165e-07, + "loss": 0.452, + "step": 3750 + }, + { + "epoch": 0.9168907357614275, + "grad_norm": 3.3074445724487305, + "learning_rate": 3.601307126298648e-07, + "loss": 0.6977, + "step": 3751 + }, + { + "epoch": 0.917135174773894, + "grad_norm": 3.0012786388397217, + "learning_rate": 3.580281258655316e-07, + "loss": 0.6826, + "step": 3752 + }, + { + "epoch": 0.9173796137863603, + "grad_norm": 3.6927969455718994, + "learning_rate": 3.559315830705856e-07, + "loss": 0.5505, + "step": 3753 + }, + { + "epoch": 0.9176240527988266, + "grad_norm": 3.082348585128784, + "learning_rate": 3.5384108555922446e-07, + "loss": 1.8339, + "step": 3754 + }, + { + "epoch": 0.917868491811293, + "grad_norm": 2.996356725692749, + "learning_rate": 3.517566346418544e-07, + "loss": 0.5314, + "step": 3755 + }, + { + "epoch": 0.9181129308237594, + "grad_norm": 2.9821434020996094, + "learning_rate": 3.496782316250913e-07, + "loss": 0.5003, + "step": 3756 + }, + { + "epoch": 0.9183573698362258, + "grad_norm": 3.2172985076904297, + "learning_rate": 3.476058778117608e-07, + "loss": 0.5269, + "step": 3757 + }, + { + "epoch": 0.9186018088486922, + "grad_norm": 4.517338752746582, + "learning_rate": 3.4553957450089383e-07, + "loss": 0.5819, + "step": 3758 + }, + { + "epoch": 0.9188462478611586, + "grad_norm": 3.4000778198242188, + "learning_rate": 3.434793229877353e-07, + "loss": 0.5547, + "step": 3759 + }, + { + "epoch": 0.919090686873625, + "grad_norm": 3.2452781200408936, + "learning_rate": 3.41425124563729e-07, + "loss": 0.6142, + "step": 3760 + }, + { + "epoch": 0.9193351258860915, + "grad_norm": 3.8453309535980225, + "learning_rate": 3.3937698051653034e-07, + "loss": 0.5169, + "step": 3761 + }, + { + "epoch": 0.9195795648985579, + "grad_norm": 3.3375754356384277, + "learning_rate": 3.373348921299946e-07, + "loss": 0.4744, + "step": 3762 + }, + { + "epoch": 0.9198240039110241, + "grad_norm": 3.6361775398254395, + "learning_rate": 3.352988606841845e-07, + "loss": 0.59, + "step": 3763 + }, + { + "epoch": 0.9200684429234905, + "grad_norm": 2.8407204151153564, + "learning_rate": 3.3326888745536576e-07, + "loss": 0.5369, + "step": 3764 + }, + { + "epoch": 0.920312881935957, + "grad_norm": 4.514292240142822, + "learning_rate": 3.312449737160051e-07, + "loss": 0.4623, + "step": 3765 + }, + { + "epoch": 0.9205573209484234, + "grad_norm": 2.8216071128845215, + "learning_rate": 3.2922712073477194e-07, + "loss": 0.4855, + "step": 3766 + }, + { + "epoch": 0.9208017599608898, + "grad_norm": 3.497938394546509, + "learning_rate": 3.272153297765368e-07, + "loss": 0.582, + "step": 3767 + }, + { + "epoch": 0.9210461989733562, + "grad_norm": 3.3942174911499023, + "learning_rate": 3.252096021023687e-07, + "loss": 0.5727, + "step": 3768 + }, + { + "epoch": 0.9212906379858226, + "grad_norm": 2.20690655708313, + "learning_rate": 3.2320993896953756e-07, + "loss": 1.8639, + "step": 3769 + }, + { + "epoch": 0.921535076998289, + "grad_norm": 2.8550937175750732, + "learning_rate": 3.2121634163151285e-07, + "loss": 0.5327, + "step": 3770 + }, + { + "epoch": 0.9217795160107554, + "grad_norm": 3.311756134033203, + "learning_rate": 3.1922881133795827e-07, + "loss": 0.6002, + "step": 3771 + }, + { + "epoch": 0.9220239550232217, + "grad_norm": 3.338653564453125, + "learning_rate": 3.172473493347361e-07, + "loss": 0.4517, + "step": 3772 + }, + { + "epoch": 0.922268394035688, + "grad_norm": 3.409890651702881, + "learning_rate": 3.152719568639073e-07, + "loss": 0.5154, + "step": 3773 + }, + { + "epoch": 0.9225128330481545, + "grad_norm": 3.480623960494995, + "learning_rate": 3.133026351637236e-07, + "loss": 0.5594, + "step": 3774 + }, + { + "epoch": 0.9227572720606209, + "grad_norm": 5.149979591369629, + "learning_rate": 3.113393854686353e-07, + "loss": 0.6494, + "step": 3775 + }, + { + "epoch": 0.9230017110730873, + "grad_norm": 3.3873095512390137, + "learning_rate": 3.093822090092857e-07, + "loss": 0.5442, + "step": 3776 + }, + { + "epoch": 0.9232461500855537, + "grad_norm": 3.5246260166168213, + "learning_rate": 3.0743110701250913e-07, + "loss": 0.5315, + "step": 3777 + }, + { + "epoch": 0.9234905890980201, + "grad_norm": 2.8095595836639404, + "learning_rate": 3.054860807013327e-07, + "loss": 1.9355, + "step": 3778 + }, + { + "epoch": 0.9237350281104865, + "grad_norm": 3.0907297134399414, + "learning_rate": 3.035471312949778e-07, + "loss": 0.6488, + "step": 3779 + }, + { + "epoch": 0.9239794671229529, + "grad_norm": 4.124264717102051, + "learning_rate": 3.016142600088534e-07, + "loss": 0.5738, + "step": 3780 + }, + { + "epoch": 0.9242239061354192, + "grad_norm": 3.186234712600708, + "learning_rate": 2.996874680545603e-07, + "loss": 0.6444, + "step": 3781 + }, + { + "epoch": 0.9244683451478856, + "grad_norm": 2.7483363151550293, + "learning_rate": 2.9776675663988785e-07, + "loss": 0.5399, + "step": 3782 + }, + { + "epoch": 0.924712784160352, + "grad_norm": 3.520253896713257, + "learning_rate": 2.9585212696881307e-07, + "loss": 0.4852, + "step": 3783 + }, + { + "epoch": 0.9249572231728184, + "grad_norm": 3.1906232833862305, + "learning_rate": 2.939435802415025e-07, + "loss": 0.5869, + "step": 3784 + }, + { + "epoch": 0.9252016621852848, + "grad_norm": 3.0981388092041016, + "learning_rate": 2.920411176543081e-07, + "loss": 0.6187, + "step": 3785 + }, + { + "epoch": 0.9254461011977512, + "grad_norm": 2.660426616668701, + "learning_rate": 2.9014474039976923e-07, + "loss": 0.4867, + "step": 3786 + }, + { + "epoch": 0.9256905402102176, + "grad_norm": 2.9685277938842773, + "learning_rate": 2.8825444966661064e-07, + "loss": 0.4, + "step": 3787 + }, + { + "epoch": 0.925934979222684, + "grad_norm": 9.820734024047852, + "learning_rate": 2.8637024663974e-07, + "loss": 0.5051, + "step": 3788 + }, + { + "epoch": 0.9261794182351504, + "grad_norm": 2.707178831100464, + "learning_rate": 2.844921325002492e-07, + "loss": 2.026, + "step": 3789 + }, + { + "epoch": 0.9264238572476167, + "grad_norm": 4.109099864959717, + "learning_rate": 2.8262010842541874e-07, + "loss": 0.6936, + "step": 3790 + }, + { + "epoch": 0.9266682962600831, + "grad_norm": 3.1422524452209473, + "learning_rate": 2.8075417558870333e-07, + "loss": 0.5485, + "step": 3791 + }, + { + "epoch": 0.9269127352725495, + "grad_norm": 4.1273698806762695, + "learning_rate": 2.7889433515974616e-07, + "loss": 0.5213, + "step": 3792 + }, + { + "epoch": 0.9271571742850159, + "grad_norm": 2.814871072769165, + "learning_rate": 2.7704058830436696e-07, + "loss": 0.5673, + "step": 3793 + }, + { + "epoch": 0.9274016132974823, + "grad_norm": 2.826104164123535, + "learning_rate": 2.7519293618456845e-07, + "loss": 0.4887, + "step": 3794 + }, + { + "epoch": 0.9276460523099487, + "grad_norm": 4.0700764656066895, + "learning_rate": 2.733513799585319e-07, + "loss": 0.6081, + "step": 3795 + }, + { + "epoch": 0.9278904913224151, + "grad_norm": 3.63450026512146, + "learning_rate": 2.7151592078061726e-07, + "loss": 0.4312, + "step": 3796 + }, + { + "epoch": 0.9281349303348815, + "grad_norm": 3.3018198013305664, + "learning_rate": 2.6968655980136425e-07, + "loss": 0.5232, + "step": 3797 + }, + { + "epoch": 0.9283793693473479, + "grad_norm": 3.0066380500793457, + "learning_rate": 2.678632981674867e-07, + "loss": 0.5133, + "step": 3798 + }, + { + "epoch": 0.9286238083598142, + "grad_norm": 2.7969233989715576, + "learning_rate": 2.660461370218792e-07, + "loss": 0.6212, + "step": 3799 + }, + { + "epoch": 0.9288682473722806, + "grad_norm": 2.9382822513580322, + "learning_rate": 2.6423507750360844e-07, + "loss": 0.5063, + "step": 3800 + }, + { + "epoch": 0.929112686384747, + "grad_norm": 3.175105571746826, + "learning_rate": 2.624301207479185e-07, + "loss": 0.5232, + "step": 3801 + }, + { + "epoch": 0.9293571253972134, + "grad_norm": 3.261991024017334, + "learning_rate": 2.6063126788622883e-07, + "loss": 0.7234, + "step": 3802 + }, + { + "epoch": 0.9296015644096798, + "grad_norm": 3.413081407546997, + "learning_rate": 2.588385200461307e-07, + "loss": 0.6532, + "step": 3803 + }, + { + "epoch": 0.9298460034221462, + "grad_norm": 3.2571053504943848, + "learning_rate": 2.570518783513887e-07, + "loss": 0.5096, + "step": 3804 + }, + { + "epoch": 0.9300904424346126, + "grad_norm": 2.5494961738586426, + "learning_rate": 2.552713439219412e-07, + "loss": 1.8917, + "step": 3805 + }, + { + "epoch": 0.930334881447079, + "grad_norm": 3.4841763973236084, + "learning_rate": 2.5349691787389885e-07, + "loss": 0.5292, + "step": 3806 + }, + { + "epoch": 0.9305793204595454, + "grad_norm": 3.9227099418640137, + "learning_rate": 2.517286013195397e-07, + "loss": 0.6536, + "step": 3807 + }, + { + "epoch": 0.9308237594720117, + "grad_norm": 3.9732210636138916, + "learning_rate": 2.4996639536731706e-07, + "loss": 0.7448, + "step": 3808 + }, + { + "epoch": 0.9310681984844781, + "grad_norm": 3.2098023891448975, + "learning_rate": 2.482103011218495e-07, + "loss": 0.6618, + "step": 3809 + }, + { + "epoch": 0.9313126374969445, + "grad_norm": 3.286810874938965, + "learning_rate": 2.464603196839277e-07, + "loss": 0.6486, + "step": 3810 + }, + { + "epoch": 0.9315570765094109, + "grad_norm": 3.226602792739868, + "learning_rate": 2.447164521505074e-07, + "loss": 0.6143, + "step": 3811 + }, + { + "epoch": 0.9318015155218773, + "grad_norm": 4.43597936630249, + "learning_rate": 2.4297869961471544e-07, + "loss": 0.5931, + "step": 3812 + }, + { + "epoch": 0.9320459545343437, + "grad_norm": 4.943711757659912, + "learning_rate": 2.4124706316584255e-07, + "loss": 0.6558, + "step": 3813 + }, + { + "epoch": 0.9322903935468101, + "grad_norm": 2.977363109588623, + "learning_rate": 2.3952154388934833e-07, + "loss": 0.5582, + "step": 3814 + }, + { + "epoch": 0.9325348325592765, + "grad_norm": 3.788539409637451, + "learning_rate": 2.3780214286685533e-07, + "loss": 0.674, + "step": 3815 + }, + { + "epoch": 0.9327792715717429, + "grad_norm": 2.4482150077819824, + "learning_rate": 2.360888611761536e-07, + "loss": 1.9792, + "step": 3816 + }, + { + "epoch": 0.9330237105842092, + "grad_norm": 2.869302988052368, + "learning_rate": 2.3438169989119408e-07, + "loss": 0.4668, + "step": 3817 + }, + { + "epoch": 0.9332681495966756, + "grad_norm": 3.960190773010254, + "learning_rate": 2.32680660082093e-07, + "loss": 0.5912, + "step": 3818 + }, + { + "epoch": 0.933512588609142, + "grad_norm": 3.486650228500366, + "learning_rate": 2.3098574281513185e-07, + "loss": 0.3871, + "step": 3819 + }, + { + "epoch": 0.9337570276216084, + "grad_norm": 3.2709710597991943, + "learning_rate": 2.292969491527486e-07, + "loss": 0.4614, + "step": 3820 + }, + { + "epoch": 0.9340014666340748, + "grad_norm": 3.7016708850860596, + "learning_rate": 2.276142801535486e-07, + "loss": 0.488, + "step": 3821 + }, + { + "epoch": 0.9342459056465412, + "grad_norm": 3.7128171920776367, + "learning_rate": 2.259377368722937e-07, + "loss": 0.5591, + "step": 3822 + }, + { + "epoch": 0.9344903446590076, + "grad_norm": 2.55253267288208, + "learning_rate": 2.2426732035990772e-07, + "loss": 0.4151, + "step": 3823 + }, + { + "epoch": 0.934734783671474, + "grad_norm": 4.539761543273926, + "learning_rate": 2.226030316634753e-07, + "loss": 0.5469, + "step": 3824 + }, + { + "epoch": 0.9349792226839404, + "grad_norm": 4.752309322357178, + "learning_rate": 2.2094487182623747e-07, + "loss": 0.7591, + "step": 3825 + }, + { + "epoch": 0.9352236616964067, + "grad_norm": 3.1529886722564697, + "learning_rate": 2.192928418875928e-07, + "loss": 0.5054, + "step": 3826 + }, + { + "epoch": 0.9354681007088731, + "grad_norm": 3.589515209197998, + "learning_rate": 2.1764694288310185e-07, + "loss": 0.6118, + "step": 3827 + }, + { + "epoch": 0.9357125397213395, + "grad_norm": 3.063481330871582, + "learning_rate": 2.1600717584447706e-07, + "loss": 0.5649, + "step": 3828 + }, + { + "epoch": 0.9359569787338059, + "grad_norm": 3.080091953277588, + "learning_rate": 2.143735417995907e-07, + "loss": 0.5806, + "step": 3829 + }, + { + "epoch": 0.9362014177462723, + "grad_norm": 3.8041274547576904, + "learning_rate": 2.1274604177246804e-07, + "loss": 0.5447, + "step": 3830 + }, + { + "epoch": 0.9364458567587387, + "grad_norm": 3.509629726409912, + "learning_rate": 2.1112467678329197e-07, + "loss": 0.5552, + "step": 3831 + }, + { + "epoch": 0.9366902957712051, + "grad_norm": 3.3342905044555664, + "learning_rate": 2.095094478483961e-07, + "loss": 0.5797, + "step": 3832 + }, + { + "epoch": 0.9369347347836715, + "grad_norm": 2.862881898880005, + "learning_rate": 2.079003559802706e-07, + "loss": 0.4585, + "step": 3833 + }, + { + "epoch": 0.9371791737961379, + "grad_norm": 2.8512535095214844, + "learning_rate": 2.0629740218755857e-07, + "loss": 0.4295, + "step": 3834 + }, + { + "epoch": 0.9374236128086042, + "grad_norm": 2.828467607498169, + "learning_rate": 2.0470058747505516e-07, + "loss": 0.4893, + "step": 3835 + }, + { + "epoch": 0.9376680518210706, + "grad_norm": 3.393681526184082, + "learning_rate": 2.031099128437064e-07, + "loss": 0.6233, + "step": 3836 + }, + { + "epoch": 0.937912490833537, + "grad_norm": 2.908700942993164, + "learning_rate": 2.0152537929061023e-07, + "loss": 0.6536, + "step": 3837 + }, + { + "epoch": 0.9381569298460034, + "grad_norm": 2.6761109828948975, + "learning_rate": 1.9994698780901434e-07, + "loss": 0.4504, + "step": 3838 + }, + { + "epoch": 0.9384013688584698, + "grad_norm": 3.427279472351074, + "learning_rate": 1.983747393883173e-07, + "loss": 0.524, + "step": 3839 + }, + { + "epoch": 0.9386458078709362, + "grad_norm": 3.2629621028900146, + "learning_rate": 1.9680863501406744e-07, + "loss": 0.5726, + "step": 3840 + }, + { + "epoch": 0.9388902468834026, + "grad_norm": 2.9898974895477295, + "learning_rate": 1.9524867566795945e-07, + "loss": 0.4985, + "step": 3841 + }, + { + "epoch": 0.939134685895869, + "grad_norm": 3.20027756690979, + "learning_rate": 1.9369486232783896e-07, + "loss": 0.5411, + "step": 3842 + }, + { + "epoch": 0.9393791249083354, + "grad_norm": 5.192112922668457, + "learning_rate": 1.921471959676957e-07, + "loss": 0.5168, + "step": 3843 + }, + { + "epoch": 0.9396235639208017, + "grad_norm": 3.0395219326019287, + "learning_rate": 1.9060567755767035e-07, + "loss": 0.546, + "step": 3844 + }, + { + "epoch": 0.9398680029332681, + "grad_norm": 3.2350528240203857, + "learning_rate": 1.890703080640466e-07, + "loss": 0.5259, + "step": 3845 + }, + { + "epoch": 0.9401124419457345, + "grad_norm": 4.550525188446045, + "learning_rate": 1.8754108844925455e-07, + "loss": 0.615, + "step": 3846 + }, + { + "epoch": 0.9403568809582009, + "grad_norm": 2.9899308681488037, + "learning_rate": 1.8601801967187083e-07, + "loss": 0.4783, + "step": 3847 + }, + { + "epoch": 0.9406013199706673, + "grad_norm": 3.2106664180755615, + "learning_rate": 1.8450110268661392e-07, + "loss": 0.5835, + "step": 3848 + }, + { + "epoch": 0.9408457589831337, + "grad_norm": 3.7672276496887207, + "learning_rate": 1.8299033844434878e-07, + "loss": 0.575, + "step": 3849 + }, + { + "epoch": 0.9410901979956001, + "grad_norm": 3.14738130569458, + "learning_rate": 1.8148572789208007e-07, + "loss": 0.6041, + "step": 3850 + }, + { + "epoch": 0.9413346370080665, + "grad_norm": 3.2783117294311523, + "learning_rate": 1.7998727197295785e-07, + "loss": 0.475, + "step": 3851 + }, + { + "epoch": 0.9415790760205329, + "grad_norm": 3.5579917430877686, + "learning_rate": 1.7849497162627516e-07, + "loss": 0.6771, + "step": 3852 + }, + { + "epoch": 0.9418235150329992, + "grad_norm": 2.999138355255127, + "learning_rate": 1.7700882778746374e-07, + "loss": 0.556, + "step": 3853 + }, + { + "epoch": 0.9420679540454656, + "grad_norm": 4.949743270874023, + "learning_rate": 1.755288413880951e-07, + "loss": 0.3769, + "step": 3854 + }, + { + "epoch": 0.942312393057932, + "grad_norm": 2.9086415767669678, + "learning_rate": 1.7405501335588825e-07, + "loss": 0.4424, + "step": 3855 + }, + { + "epoch": 0.9425568320703984, + "grad_norm": 3.3956940174102783, + "learning_rate": 1.7258734461469307e-07, + "loss": 0.6627, + "step": 3856 + }, + { + "epoch": 0.9428012710828648, + "grad_norm": 3.2270140647888184, + "learning_rate": 1.7112583608450361e-07, + "loss": 0.6194, + "step": 3857 + }, + { + "epoch": 0.9430457100953312, + "grad_norm": 3.090681552886963, + "learning_rate": 1.6967048868145154e-07, + "loss": 0.5783, + "step": 3858 + }, + { + "epoch": 0.9432901491077976, + "grad_norm": 2.989459991455078, + "learning_rate": 1.6822130331780485e-07, + "loss": 0.549, + "step": 3859 + }, + { + "epoch": 0.943534588120264, + "grad_norm": 3.7176594734191895, + "learning_rate": 1.6677828090197245e-07, + "loss": 0.672, + "step": 3860 + }, + { + "epoch": 0.9437790271327304, + "grad_norm": 2.7825515270233154, + "learning_rate": 1.6534142233849527e-07, + "loss": 0.4756, + "step": 3861 + }, + { + "epoch": 0.9440234661451967, + "grad_norm": 5.053201675415039, + "learning_rate": 1.6391072852805613e-07, + "loss": 0.6182, + "step": 3862 + }, + { + "epoch": 0.9442679051576631, + "grad_norm": 3.03202486038208, + "learning_rate": 1.6248620036746764e-07, + "loss": 0.5124, + "step": 3863 + }, + { + "epoch": 0.9445123441701295, + "grad_norm": 3.8698859214782715, + "learning_rate": 1.6106783874968114e-07, + "loss": 0.5856, + "step": 3864 + }, + { + "epoch": 0.9447567831825959, + "grad_norm": 3.8033907413482666, + "learning_rate": 1.5965564456378313e-07, + "loss": 0.4567, + "step": 3865 + }, + { + "epoch": 0.9450012221950623, + "grad_norm": 3.2855918407440186, + "learning_rate": 1.582496186949911e-07, + "loss": 0.5178, + "step": 3866 + }, + { + "epoch": 0.9452456612075287, + "grad_norm": 3.8047099113464355, + "learning_rate": 1.5684976202465786e-07, + "loss": 0.5535, + "step": 3867 + }, + { + "epoch": 0.9454901002199951, + "grad_norm": 4.872805118560791, + "learning_rate": 1.554560754302703e-07, + "loss": 0.6506, + "step": 3868 + }, + { + "epoch": 0.9457345392324615, + "grad_norm": 3.565657138824463, + "learning_rate": 1.5406855978544634e-07, + "loss": 0.5024, + "step": 3869 + }, + { + "epoch": 0.9459789782449279, + "grad_norm": 3.1939427852630615, + "learning_rate": 1.5268721595993464e-07, + "loss": 0.5648, + "step": 3870 + }, + { + "epoch": 0.9462234172573942, + "grad_norm": 4.939906120300293, + "learning_rate": 1.5131204481961592e-07, + "loss": 0.4766, + "step": 3871 + }, + { + "epoch": 0.9464678562698606, + "grad_norm": 3.5011534690856934, + "learning_rate": 1.4994304722650176e-07, + "loss": 0.556, + "step": 3872 + }, + { + "epoch": 0.946712295282327, + "grad_norm": 3.5804851055145264, + "learning_rate": 1.4858022403873573e-07, + "loss": 0.5861, + "step": 3873 + }, + { + "epoch": 0.9469567342947934, + "grad_norm": 3.398738384246826, + "learning_rate": 1.472235761105878e-07, + "loss": 0.5581, + "step": 3874 + }, + { + "epoch": 0.9472011733072598, + "grad_norm": 3.1239571571350098, + "learning_rate": 1.4587310429245883e-07, + "loss": 0.6228, + "step": 3875 + }, + { + "epoch": 0.9474456123197262, + "grad_norm": 3.6954193115234375, + "learning_rate": 1.4452880943087722e-07, + "loss": 0.468, + "step": 3876 + }, + { + "epoch": 0.9476900513321926, + "grad_norm": 2.738016128540039, + "learning_rate": 1.4319069236850224e-07, + "loss": 0.542, + "step": 3877 + }, + { + "epoch": 0.947934490344659, + "grad_norm": 3.486159086227417, + "learning_rate": 1.418587539441174e-07, + "loss": 0.6804, + "step": 3878 + }, + { + "epoch": 0.9481789293571254, + "grad_norm": 3.5900747776031494, + "learning_rate": 1.4053299499263374e-07, + "loss": 0.7054, + "step": 3879 + }, + { + "epoch": 0.9484233683695917, + "grad_norm": 3.0891590118408203, + "learning_rate": 1.392134163450898e-07, + "loss": 0.5256, + "step": 3880 + }, + { + "epoch": 0.9486678073820581, + "grad_norm": 5.227392673492432, + "learning_rate": 1.3790001882865056e-07, + "loss": 0.5639, + "step": 3881 + }, + { + "epoch": 0.9489122463945245, + "grad_norm": 4.298923969268799, + "learning_rate": 1.365928032666042e-07, + "loss": 0.5737, + "step": 3882 + }, + { + "epoch": 0.9491566854069909, + "grad_norm": 3.931917667388916, + "learning_rate": 1.3529177047836628e-07, + "loss": 0.5685, + "step": 3883 + }, + { + "epoch": 0.9494011244194573, + "grad_norm": 2.9893269538879395, + "learning_rate": 1.3399692127947562e-07, + "loss": 0.6057, + "step": 3884 + }, + { + "epoch": 0.9496455634319237, + "grad_norm": 3.0450754165649414, + "learning_rate": 1.3270825648159403e-07, + "loss": 0.605, + "step": 3885 + }, + { + "epoch": 0.9498900024443901, + "grad_norm": 3.5307817459106445, + "learning_rate": 1.3142577689250756e-07, + "loss": 0.5407, + "step": 3886 + }, + { + "epoch": 0.9501344414568565, + "grad_norm": 3.128784656524658, + "learning_rate": 1.3014948331612653e-07, + "loss": 0.4743, + "step": 3887 + }, + { + "epoch": 0.950378880469323, + "grad_norm": 4.463705062866211, + "learning_rate": 1.2887937655248096e-07, + "loss": 0.5312, + "step": 3888 + }, + { + "epoch": 0.9506233194817892, + "grad_norm": 3.3671658039093018, + "learning_rate": 1.2761545739772396e-07, + "loss": 0.5913, + "step": 3889 + }, + { + "epoch": 0.9508677584942556, + "grad_norm": 3.701791286468506, + "learning_rate": 1.2635772664413072e-07, + "loss": 0.6569, + "step": 3890 + }, + { + "epoch": 0.951112197506722, + "grad_norm": 3.6596016883850098, + "learning_rate": 1.251061850800961e-07, + "loss": 0.5157, + "step": 3891 + }, + { + "epoch": 0.9513566365191884, + "grad_norm": 3.160101890563965, + "learning_rate": 1.23860833490137e-07, + "loss": 0.6493, + "step": 3892 + }, + { + "epoch": 0.9516010755316548, + "grad_norm": 2.915860652923584, + "learning_rate": 1.226216726548879e-07, + "loss": 0.5418, + "step": 3893 + }, + { + "epoch": 0.9518455145441213, + "grad_norm": 3.316563844680786, + "learning_rate": 1.213887033511041e-07, + "loss": 0.5572, + "step": 3894 + }, + { + "epoch": 0.9520899535565877, + "grad_norm": 3.3448314666748047, + "learning_rate": 1.2016192635165958e-07, + "loss": 0.5751, + "step": 3895 + }, + { + "epoch": 0.9523343925690541, + "grad_norm": 3.304323196411133, + "learning_rate": 1.1894134242554589e-07, + "loss": 0.5206, + "step": 3896 + }, + { + "epoch": 0.9525788315815205, + "grad_norm": 3.19252347946167, + "learning_rate": 1.1772695233787545e-07, + "loss": 0.6348, + "step": 3897 + }, + { + "epoch": 0.9528232705939867, + "grad_norm": 3.2109947204589844, + "learning_rate": 1.1651875684987268e-07, + "loss": 0.6405, + "step": 3898 + }, + { + "epoch": 0.9530677096064532, + "grad_norm": 3.1668834686279297, + "learning_rate": 1.1531675671888621e-07, + "loss": 0.5971, + "step": 3899 + }, + { + "epoch": 0.9533121486189196, + "grad_norm": 2.928603172302246, + "learning_rate": 1.1412095269837442e-07, + "loss": 0.5108, + "step": 3900 + }, + { + "epoch": 0.953556587631386, + "grad_norm": 3.901209831237793, + "learning_rate": 1.1293134553791551e-07, + "loss": 0.5936, + "step": 3901 + }, + { + "epoch": 0.9538010266438524, + "grad_norm": 3.3114547729492188, + "learning_rate": 1.1174793598320299e-07, + "loss": 0.5659, + "step": 3902 + }, + { + "epoch": 0.9540454656563188, + "grad_norm": 2.952723503112793, + "learning_rate": 1.1057072477604347e-07, + "loss": 0.5262, + "step": 3903 + }, + { + "epoch": 0.9542899046687852, + "grad_norm": 3.1195309162139893, + "learning_rate": 1.0939971265436e-07, + "loss": 0.5468, + "step": 3904 + }, + { + "epoch": 0.9545343436812516, + "grad_norm": 2.4515929222106934, + "learning_rate": 1.0823490035218986e-07, + "loss": 1.9989, + "step": 3905 + }, + { + "epoch": 0.954778782693718, + "grad_norm": 2.760552406311035, + "learning_rate": 1.0707628859968344e-07, + "loss": 0.5545, + "step": 3906 + }, + { + "epoch": 0.9550232217061843, + "grad_norm": 3.261154890060425, + "learning_rate": 1.0592387812310312e-07, + "loss": 0.424, + "step": 3907 + }, + { + "epoch": 0.9552676607186507, + "grad_norm": 3.056074380874634, + "learning_rate": 1.0477766964482772e-07, + "loss": 0.5899, + "step": 3908 + }, + { + "epoch": 0.9555120997311171, + "grad_norm": 4.080880165100098, + "learning_rate": 1.0363766388334362e-07, + "loss": 0.5603, + "step": 3909 + }, + { + "epoch": 0.9557565387435835, + "grad_norm": 3.1623666286468506, + "learning_rate": 1.0250386155325254e-07, + "loss": 0.6329, + "step": 3910 + }, + { + "epoch": 0.9560009777560499, + "grad_norm": 3.1618571281433105, + "learning_rate": 1.0137626336526596e-07, + "loss": 0.7661, + "step": 3911 + }, + { + "epoch": 0.9562454167685163, + "grad_norm": 3.1398682594299316, + "learning_rate": 1.0025487002620848e-07, + "loss": 0.5826, + "step": 3912 + }, + { + "epoch": 0.9564898557809827, + "grad_norm": 4.327209949493408, + "learning_rate": 9.913968223901227e-08, + "loss": 0.6099, + "step": 3913 + }, + { + "epoch": 0.9567342947934491, + "grad_norm": 2.97302508354187, + "learning_rate": 9.803070070272258e-08, + "loss": 0.5439, + "step": 3914 + }, + { + "epoch": 0.9569787338059155, + "grad_norm": 4.734954357147217, + "learning_rate": 9.692792611249224e-08, + "loss": 0.5333, + "step": 3915 + }, + { + "epoch": 0.9572231728183818, + "grad_norm": 3.317031145095825, + "learning_rate": 9.583135915958386e-08, + "loss": 0.5139, + "step": 3916 + }, + { + "epoch": 0.9574676118308482, + "grad_norm": 3.1889259815216064, + "learning_rate": 9.47410005313687e-08, + "loss": 0.651, + "step": 3917 + }, + { + "epoch": 0.9577120508433146, + "grad_norm": 3.311048984527588, + "learning_rate": 9.365685091132782e-08, + "loss": 0.567, + "step": 3918 + }, + { + "epoch": 0.957956489855781, + "grad_norm": 2.774733543395996, + "learning_rate": 9.257891097904759e-08, + "loss": 0.4583, + "step": 3919 + }, + { + "epoch": 0.9582009288682474, + "grad_norm": 2.797528028488159, + "learning_rate": 9.15071814102253e-08, + "loss": 0.5282, + "step": 3920 + }, + { + "epoch": 0.9584453678807138, + "grad_norm": 3.390241861343384, + "learning_rate": 9.044166287666134e-08, + "loss": 0.4714, + "step": 3921 + }, + { + "epoch": 0.9586898068931802, + "grad_norm": 3.394399404525757, + "learning_rate": 8.938235604626589e-08, + "loss": 0.5817, + "step": 3922 + }, + { + "epoch": 0.9589342459056466, + "grad_norm": 4.014432907104492, + "learning_rate": 8.832926158305443e-08, + "loss": 0.6375, + "step": 3923 + }, + { + "epoch": 0.959178684918113, + "grad_norm": 2.896803379058838, + "learning_rate": 8.728238014714785e-08, + "loss": 0.5929, + "step": 3924 + }, + { + "epoch": 0.9594231239305793, + "grad_norm": 3.5837626457214355, + "learning_rate": 8.624171239477342e-08, + "loss": 0.4898, + "step": 3925 + }, + { + "epoch": 0.9596675629430457, + "grad_norm": 2.9800174236297607, + "learning_rate": 8.520725897826265e-08, + "loss": 0.5695, + "step": 3926 + }, + { + "epoch": 0.9599120019555121, + "grad_norm": 3.6687331199645996, + "learning_rate": 8.417902054605131e-08, + "loss": 0.5069, + "step": 3927 + }, + { + "epoch": 0.9601564409679785, + "grad_norm": 3.7042768001556396, + "learning_rate": 8.315699774268048e-08, + "loss": 0.5197, + "step": 3928 + }, + { + "epoch": 0.9604008799804449, + "grad_norm": 3.6040287017822266, + "learning_rate": 8.214119120879437e-08, + "loss": 0.4721, + "step": 3929 + }, + { + "epoch": 0.9606453189929113, + "grad_norm": 3.8857340812683105, + "learning_rate": 8.11316015811403e-08, + "loss": 0.5595, + "step": 3930 + }, + { + "epoch": 0.9608897580053777, + "grad_norm": 2.8456661701202393, + "learning_rate": 8.012822949256981e-08, + "loss": 0.659, + "step": 3931 + }, + { + "epoch": 0.9611341970178441, + "grad_norm": 3.1799733638763428, + "learning_rate": 7.913107557203648e-08, + "loss": 0.6267, + "step": 3932 + }, + { + "epoch": 0.9613786360303105, + "grad_norm": 3.4712483882904053, + "learning_rate": 7.814014044459362e-08, + "loss": 0.622, + "step": 3933 + }, + { + "epoch": 0.9616230750427768, + "grad_norm": 3.23189115524292, + "learning_rate": 7.715542473140103e-08, + "loss": 0.4777, + "step": 3934 + }, + { + "epoch": 0.9618675140552432, + "grad_norm": 2.996572971343994, + "learning_rate": 7.617692904971718e-08, + "loss": 0.4598, + "step": 3935 + }, + { + "epoch": 0.9621119530677096, + "grad_norm": 3.7120778560638428, + "learning_rate": 7.520465401290033e-08, + "loss": 0.5708, + "step": 3936 + }, + { + "epoch": 0.962356392080176, + "grad_norm": 2.813230514526367, + "learning_rate": 7.423860023041185e-08, + "loss": 0.4853, + "step": 3937 + }, + { + "epoch": 0.9626008310926424, + "grad_norm": 2.9314701557159424, + "learning_rate": 7.327876830781288e-08, + "loss": 0.5752, + "step": 3938 + }, + { + "epoch": 0.9628452701051088, + "grad_norm": 3.0445821285247803, + "learning_rate": 7.232515884676327e-08, + "loss": 0.6062, + "step": 3939 + }, + { + "epoch": 0.9630897091175752, + "grad_norm": 6.091731548309326, + "learning_rate": 7.137777244502375e-08, + "loss": 0.5332, + "step": 3940 + }, + { + "epoch": 0.9633341481300416, + "grad_norm": 3.6530470848083496, + "learning_rate": 7.043660969645261e-08, + "loss": 0.5968, + "step": 3941 + }, + { + "epoch": 0.963578587142508, + "grad_norm": 3.23164439201355, + "learning_rate": 6.950167119100903e-08, + "loss": 0.5435, + "step": 3942 + }, + { + "epoch": 0.9638230261549743, + "grad_norm": 3.90278959274292, + "learning_rate": 6.857295751474979e-08, + "loss": 0.5857, + "step": 3943 + }, + { + "epoch": 0.9640674651674407, + "grad_norm": 3.4893884658813477, + "learning_rate": 6.765046924982699e-08, + "loss": 0.5718, + "step": 3944 + }, + { + "epoch": 0.9643119041799071, + "grad_norm": 2.482055425643921, + "learning_rate": 6.673420697449473e-08, + "loss": 1.9335, + "step": 3945 + }, + { + "epoch": 0.9645563431923735, + "grad_norm": 3.010075092315674, + "learning_rate": 6.582417126310359e-08, + "loss": 0.5077, + "step": 3946 + }, + { + "epoch": 0.9648007822048399, + "grad_norm": 2.269382953643799, + "learning_rate": 6.492036268609725e-08, + "loss": 1.8825, + "step": 3947 + }, + { + "epoch": 0.9650452212173063, + "grad_norm": 4.631083011627197, + "learning_rate": 6.402278181002031e-08, + "loss": 0.5548, + "step": 3948 + }, + { + "epoch": 0.9652896602297727, + "grad_norm": 3.2476084232330322, + "learning_rate": 6.313142919751159e-08, + "loss": 0.5657, + "step": 3949 + }, + { + "epoch": 0.9655340992422391, + "grad_norm": 3.203571081161499, + "learning_rate": 6.224630540730525e-08, + "loss": 0.4498, + "step": 3950 + }, + { + "epoch": 0.9657785382547055, + "grad_norm": 3.3352339267730713, + "learning_rate": 6.136741099423416e-08, + "loss": 0.4863, + "step": 3951 + }, + { + "epoch": 0.9660229772671718, + "grad_norm": 3.1167092323303223, + "learning_rate": 6.049474650922094e-08, + "loss": 0.6072, + "step": 3952 + }, + { + "epoch": 0.9662674162796382, + "grad_norm": 3.6150765419006348, + "learning_rate": 5.962831249928913e-08, + "loss": 0.5307, + "step": 3953 + }, + { + "epoch": 0.9665118552921046, + "grad_norm": 3.6788432598114014, + "learning_rate": 5.8768109507553185e-08, + "loss": 0.4531, + "step": 3954 + }, + { + "epoch": 0.966756294304571, + "grad_norm": 3.4593963623046875, + "learning_rate": 5.791413807322066e-08, + "loss": 0.4987, + "step": 3955 + }, + { + "epoch": 0.9670007333170374, + "grad_norm": 3.258887767791748, + "learning_rate": 5.7066398731596694e-08, + "loss": 0.646, + "step": 3956 + }, + { + "epoch": 0.9672451723295038, + "grad_norm": 3.145399570465088, + "learning_rate": 5.6224892014077324e-08, + "loss": 0.6184, + "step": 3957 + }, + { + "epoch": 0.9674896113419702, + "grad_norm": 3.165806531906128, + "learning_rate": 5.5389618448150606e-08, + "loss": 0.5939, + "step": 3958 + }, + { + "epoch": 0.9677340503544366, + "grad_norm": 3.396052360534668, + "learning_rate": 5.456057855740215e-08, + "loss": 0.4596, + "step": 3959 + }, + { + "epoch": 0.967978489366903, + "grad_norm": 5.122045993804932, + "learning_rate": 5.3737772861502945e-08, + "loss": 0.6347, + "step": 3960 + }, + { + "epoch": 0.9682229283793693, + "grad_norm": 3.333050489425659, + "learning_rate": 5.2921201876223737e-08, + "loss": 0.6124, + "step": 3961 + }, + { + "epoch": 0.9684673673918357, + "grad_norm": 2.9732086658477783, + "learning_rate": 5.2110866113421754e-08, + "loss": 0.5867, + "step": 3962 + }, + { + "epoch": 0.9687118064043021, + "grad_norm": 2.6987709999084473, + "learning_rate": 5.1306766081048456e-08, + "loss": 0.4852, + "step": 3963 + }, + { + "epoch": 0.9689562454167685, + "grad_norm": 3.627744436264038, + "learning_rate": 5.050890228314509e-08, + "loss": 0.7174, + "step": 3964 + }, + { + "epoch": 0.9692006844292349, + "grad_norm": 3.2260825634002686, + "learning_rate": 4.971727521984382e-08, + "loss": 0.4968, + "step": 3965 + }, + { + "epoch": 0.9694451234417013, + "grad_norm": 3.5571961402893066, + "learning_rate": 4.8931885387367706e-08, + "loss": 0.5191, + "step": 3966 + }, + { + "epoch": 0.9696895624541677, + "grad_norm": 3.4485998153686523, + "learning_rate": 4.815273327803183e-08, + "loss": 0.567, + "step": 3967 + }, + { + "epoch": 0.9699340014666341, + "grad_norm": 4.239404201507568, + "learning_rate": 4.7379819380236614e-08, + "loss": 0.6098, + "step": 3968 + }, + { + "epoch": 0.9701784404791005, + "grad_norm": 3.378331422805786, + "learning_rate": 4.6613144178477844e-08, + "loss": 0.4949, + "step": 3969 + }, + { + "epoch": 0.9704228794915668, + "grad_norm": 3.5867037773132324, + "learning_rate": 4.585270815333665e-08, + "loss": 0.5589, + "step": 3970 + }, + { + "epoch": 0.9706673185040332, + "grad_norm": 3.320458173751831, + "learning_rate": 4.5098511781485056e-08, + "loss": 0.5532, + "step": 3971 + }, + { + "epoch": 0.9709117575164996, + "grad_norm": 3.188258171081543, + "learning_rate": 4.435055553568268e-08, + "loss": 0.6084, + "step": 3972 + }, + { + "epoch": 0.971156196528966, + "grad_norm": 3.1021206378936768, + "learning_rate": 4.360883988477782e-08, + "loss": 0.6218, + "step": 3973 + }, + { + "epoch": 0.9714006355414324, + "grad_norm": 3.52012300491333, + "learning_rate": 4.2873365293706335e-08, + "loss": 0.5511, + "step": 3974 + }, + { + "epoch": 0.9716450745538988, + "grad_norm": 3.305572032928467, + "learning_rate": 4.2144132223495e-08, + "loss": 0.5442, + "step": 3975 + }, + { + "epoch": 0.9718895135663652, + "grad_norm": 2.587251663208008, + "learning_rate": 4.1421141131254835e-08, + "loss": 1.852, + "step": 3976 + }, + { + "epoch": 0.9721339525788316, + "grad_norm": 2.945148229598999, + "learning_rate": 4.070439247018554e-08, + "loss": 0.5045, + "step": 3977 + }, + { + "epoch": 0.972378391591298, + "grad_norm": 2.8457350730895996, + "learning_rate": 3.999388668957216e-08, + "loss": 0.5637, + "step": 3978 + }, + { + "epoch": 0.9726228306037643, + "grad_norm": 3.0436813831329346, + "learning_rate": 3.928962423479066e-08, + "loss": 0.4438, + "step": 3979 + }, + { + "epoch": 0.9728672696162307, + "grad_norm": 3.1909372806549072, + "learning_rate": 3.859160554729791e-08, + "loss": 0.6261, + "step": 3980 + }, + { + "epoch": 0.9731117086286971, + "grad_norm": 2.574711322784424, + "learning_rate": 3.789983106464057e-08, + "loss": 1.8583, + "step": 3981 + }, + { + "epoch": 0.9733561476411635, + "grad_norm": 3.3103384971618652, + "learning_rate": 3.7214301220450666e-08, + "loss": 0.5687, + "step": 3982 + }, + { + "epoch": 0.9736005866536299, + "grad_norm": 3.413031816482544, + "learning_rate": 3.653501644444446e-08, + "loss": 0.6336, + "step": 3983 + }, + { + "epoch": 0.9738450256660963, + "grad_norm": 3.0209238529205322, + "learning_rate": 3.586197716242579e-08, + "loss": 0.5029, + "step": 3984 + }, + { + "epoch": 0.9740894646785627, + "grad_norm": 4.465044021606445, + "learning_rate": 3.519518379628162e-08, + "loss": 0.7279, + "step": 3985 + }, + { + "epoch": 0.9743339036910291, + "grad_norm": 3.354797124862671, + "learning_rate": 3.453463676398427e-08, + "loss": 0.5322, + "step": 3986 + }, + { + "epoch": 0.9745783427034955, + "grad_norm": 9.969941139221191, + "learning_rate": 3.3880336479590324e-08, + "loss": 0.5312, + "step": 3987 + }, + { + "epoch": 0.9748227817159618, + "grad_norm": 3.3734261989593506, + "learning_rate": 3.32322833532428e-08, + "loss": 0.5122, + "step": 3988 + }, + { + "epoch": 0.9750672207284282, + "grad_norm": 3.315612554550171, + "learning_rate": 3.259047779116453e-08, + "loss": 0.511, + "step": 3989 + }, + { + "epoch": 0.9753116597408946, + "grad_norm": 3.0684654712677, + "learning_rate": 3.195492019566704e-08, + "loss": 0.602, + "step": 3990 + }, + { + "epoch": 0.975556098753361, + "grad_norm": 4.556325435638428, + "learning_rate": 3.132561096514164e-08, + "loss": 0.6097, + "step": 3991 + }, + { + "epoch": 0.9758005377658274, + "grad_norm": 3.0438623428344727, + "learning_rate": 3.07025504940639e-08, + "loss": 0.4931, + "step": 3992 + }, + { + "epoch": 0.9760449767782938, + "grad_norm": 3.167351245880127, + "learning_rate": 3.0085739172993623e-08, + "loss": 0.4976, + "step": 3993 + }, + { + "epoch": 0.9762894157907602, + "grad_norm": 2.8869566917419434, + "learning_rate": 2.9475177388573752e-08, + "loss": 1.9206, + "step": 3994 + }, + { + "epoch": 0.9765338548032266, + "grad_norm": 2.6476495265960693, + "learning_rate": 2.8870865523525916e-08, + "loss": 0.478, + "step": 3995 + }, + { + "epoch": 0.976778293815693, + "grad_norm": 2.7549493312835693, + "learning_rate": 2.8272803956658212e-08, + "loss": 0.6107, + "step": 3996 + }, + { + "epoch": 0.9770227328281593, + "grad_norm": 3.1494834423065186, + "learning_rate": 2.7680993062858543e-08, + "loss": 0.4316, + "step": 3997 + }, + { + "epoch": 0.9772671718406257, + "grad_norm": 3.2198588848114014, + "learning_rate": 2.7095433213097933e-08, + "loss": 0.5957, + "step": 3998 + }, + { + "epoch": 0.9775116108530921, + "grad_norm": 3.3557252883911133, + "learning_rate": 2.651612477442833e-08, + "loss": 0.6265, + "step": 3999 + }, + { + "epoch": 0.9777560498655585, + "grad_norm": 4.323390007019043, + "learning_rate": 2.5943068109983706e-08, + "loss": 0.5282, + "step": 4000 + }, + { + "epoch": 0.9780004888780249, + "grad_norm": 2.7845654487609863, + "learning_rate": 2.5376263578977823e-08, + "loss": 0.5089, + "step": 4001 + }, + { + "epoch": 0.9782449278904913, + "grad_norm": 3.303093910217285, + "learning_rate": 2.4815711536706478e-08, + "loss": 0.5188, + "step": 4002 + }, + { + "epoch": 0.9784893669029577, + "grad_norm": 3.255608558654785, + "learning_rate": 2.4261412334546373e-08, + "loss": 0.644, + "step": 4003 + }, + { + "epoch": 0.9787338059154241, + "grad_norm": 3.067213773727417, + "learning_rate": 2.3713366319954023e-08, + "loss": 0.4166, + "step": 4004 + }, + { + "epoch": 0.9789782449278905, + "grad_norm": 3.1581287384033203, + "learning_rate": 2.3171573836464623e-08, + "loss": 0.6115, + "step": 4005 + }, + { + "epoch": 0.9792226839403568, + "grad_norm": 4.095109939575195, + "learning_rate": 2.2636035223697616e-08, + "loss": 0.5978, + "step": 4006 + }, + { + "epoch": 0.9794671229528232, + "grad_norm": 2.942018747329712, + "learning_rate": 2.210675081734892e-08, + "loss": 0.5768, + "step": 4007 + }, + { + "epoch": 0.9797115619652896, + "grad_norm": 3.5556507110595703, + "learning_rate": 2.1583720949194253e-08, + "loss": 0.5152, + "step": 4008 + }, + { + "epoch": 0.979956000977756, + "grad_norm": 2.95241379737854, + "learning_rate": 2.1066945947090244e-08, + "loss": 0.5812, + "step": 4009 + }, + { + "epoch": 0.9802004399902224, + "grad_norm": 3.049710988998413, + "learning_rate": 2.055642613497111e-08, + "loss": 0.5166, + "step": 4010 + }, + { + "epoch": 0.9804448790026888, + "grad_norm": 3.40008807182312, + "learning_rate": 2.0052161832850858e-08, + "loss": 0.4201, + "step": 4011 + }, + { + "epoch": 0.9806893180151552, + "grad_norm": 2.8084328174591064, + "learning_rate": 1.9554153356823313e-08, + "loss": 0.5247, + "step": 4012 + }, + { + "epoch": 0.9809337570276216, + "grad_norm": 3.105757236480713, + "learning_rate": 1.9062401019059874e-08, + "loss": 0.467, + "step": 4013 + }, + { + "epoch": 0.981178196040088, + "grad_norm": 2.792382001876831, + "learning_rate": 1.857690512780952e-08, + "loss": 0.6131, + "step": 4014 + }, + { + "epoch": 0.9814226350525543, + "grad_norm": 3.0735909938812256, + "learning_rate": 1.8097665987399925e-08, + "loss": 0.5773, + "step": 4015 + }, + { + "epoch": 0.9816670740650207, + "grad_norm": 3.092674493789673, + "learning_rate": 1.762468389823857e-08, + "loss": 0.6127, + "step": 4016 + }, + { + "epoch": 0.9819115130774871, + "grad_norm": 2.8229806423187256, + "learning_rate": 1.7157959156808292e-08, + "loss": 0.6363, + "step": 4017 + }, + { + "epoch": 0.9821559520899535, + "grad_norm": 3.3986260890960693, + "learning_rate": 1.6697492055671725e-08, + "loss": 0.5013, + "step": 4018 + }, + { + "epoch": 0.98240039110242, + "grad_norm": 3.2098350524902344, + "learning_rate": 1.6243282883465772e-08, + "loss": 0.5575, + "step": 4019 + }, + { + "epoch": 0.9826448301148863, + "grad_norm": 3.488957643508911, + "learning_rate": 1.579533192490823e-08, + "loss": 0.5848, + "step": 4020 + }, + { + "epoch": 0.9828892691273527, + "grad_norm": 4.478259086608887, + "learning_rate": 1.5353639460793378e-08, + "loss": 0.6147, + "step": 4021 + }, + { + "epoch": 0.9831337081398192, + "grad_norm": 4.556621074676514, + "learning_rate": 1.4918205767989746e-08, + "loss": 0.5966, + "step": 4022 + }, + { + "epoch": 0.9833781471522856, + "grad_norm": 3.862208604812622, + "learning_rate": 1.4489031119444553e-08, + "loss": 0.4832, + "step": 4023 + }, + { + "epoch": 0.9836225861647518, + "grad_norm": 3.3213610649108887, + "learning_rate": 1.4066115784181488e-08, + "loss": 0.5607, + "step": 4024 + }, + { + "epoch": 0.9838670251772182, + "grad_norm": 2.868886947631836, + "learning_rate": 1.364946002730183e-08, + "loss": 0.5182, + "step": 4025 + }, + { + "epoch": 0.9841114641896846, + "grad_norm": 3.2725284099578857, + "learning_rate": 1.3239064109981103e-08, + "loss": 0.6052, + "step": 4026 + }, + { + "epoch": 0.984355903202151, + "grad_norm": 3.655763626098633, + "learning_rate": 1.2834928289472415e-08, + "loss": 0.5215, + "step": 4027 + }, + { + "epoch": 0.9846003422146175, + "grad_norm": 3.8058059215545654, + "learning_rate": 1.2437052819102013e-08, + "loss": 0.5269, + "step": 4028 + }, + { + "epoch": 0.9848447812270839, + "grad_norm": 3.2667770385742188, + "learning_rate": 1.2045437948275952e-08, + "loss": 0.5415, + "step": 4029 + }, + { + "epoch": 0.9850892202395503, + "grad_norm": 2.810737133026123, + "learning_rate": 1.166008392247342e-08, + "loss": 1.9476, + "step": 4030 + }, + { + "epoch": 0.9853336592520167, + "grad_norm": 3.1225063800811768, + "learning_rate": 1.1280990983248975e-08, + "loss": 0.5263, + "step": 4031 + }, + { + "epoch": 0.9855780982644831, + "grad_norm": 2.9405412673950195, + "learning_rate": 1.090815936823475e-08, + "loss": 1.9106, + "step": 4032 + }, + { + "epoch": 0.9858225372769494, + "grad_norm": 2.3317549228668213, + "learning_rate": 1.0541589311133804e-08, + "loss": 1.8014, + "step": 4033 + }, + { + "epoch": 0.9860669762894158, + "grad_norm": 3.156327486038208, + "learning_rate": 1.0181281041728996e-08, + "loss": 0.5655, + "step": 4034 + }, + { + "epoch": 0.9863114153018822, + "grad_norm": 2.7583189010620117, + "learning_rate": 9.827234785874107e-09, + "loss": 0.6005, + "step": 4035 + }, + { + "epoch": 0.9865558543143486, + "grad_norm": 3.0450844764709473, + "learning_rate": 9.479450765500497e-09, + "loss": 0.4669, + "step": 4036 + }, + { + "epoch": 0.986800293326815, + "grad_norm": 3.0492610931396484, + "learning_rate": 9.137929198612672e-09, + "loss": 0.5395, + "step": 4037 + }, + { + "epoch": 0.9870447323392814, + "grad_norm": 3.032101631164551, + "learning_rate": 8.802670299290494e-09, + "loss": 0.5397, + "step": 4038 + }, + { + "epoch": 0.9872891713517478, + "grad_norm": 3.1941475868225098, + "learning_rate": 8.473674277685862e-09, + "loss": 0.5214, + "step": 4039 + }, + { + "epoch": 0.9875336103642142, + "grad_norm": 3.856670379638672, + "learning_rate": 8.150941340027142e-09, + "loss": 0.551, + "step": 4040 + }, + { + "epoch": 0.9877780493766806, + "grad_norm": 3.1291728019714355, + "learning_rate": 7.834471688616952e-09, + "loss": 0.5206, + "step": 4041 + }, + { + "epoch": 0.9880224883891469, + "grad_norm": 3.016200304031372, + "learning_rate": 7.524265521831054e-09, + "loss": 0.5845, + "step": 4042 + }, + { + "epoch": 0.9882669274016133, + "grad_norm": 3.971785306930542, + "learning_rate": 7.220323034117238e-09, + "loss": 0.5915, + "step": 4043 + }, + { + "epoch": 0.9885113664140797, + "grad_norm": 3.125352382659912, + "learning_rate": 6.9226444159997684e-09, + "loss": 0.4734, + "step": 4044 + }, + { + "epoch": 0.9887558054265461, + "grad_norm": 2.851304531097412, + "learning_rate": 6.631229854076049e-09, + "loss": 0.5954, + "step": 4045 + }, + { + "epoch": 0.9890002444390125, + "grad_norm": 2.7313387393951416, + "learning_rate": 6.346079531015514e-09, + "loss": 1.8465, + "step": 4046 + }, + { + "epoch": 0.9892446834514789, + "grad_norm": 2.8360304832458496, + "learning_rate": 6.067193625560741e-09, + "loss": 0.488, + "step": 4047 + }, + { + "epoch": 0.9894891224639453, + "grad_norm": 2.8465449810028076, + "learning_rate": 5.794572312530777e-09, + "loss": 0.4853, + "step": 4048 + }, + { + "epoch": 0.9897335614764117, + "grad_norm": 3.0731120109558105, + "learning_rate": 5.528215762813372e-09, + "loss": 0.5777, + "step": 4049 + }, + { + "epoch": 0.9899780004888781, + "grad_norm": 3.0907604694366455, + "learning_rate": 5.268124143371633e-09, + "loss": 0.6588, + "step": 4050 + }, + { + "epoch": 0.9902224395013444, + "grad_norm": 3.1686196327209473, + "learning_rate": 5.014297617242925e-09, + "loss": 0.57, + "step": 4051 + }, + { + "epoch": 0.9904668785138108, + "grad_norm": 3.1961419582366943, + "learning_rate": 4.766736343533307e-09, + "loss": 0.5639, + "step": 4052 + }, + { + "epoch": 0.9907113175262772, + "grad_norm": 3.2424230575561523, + "learning_rate": 4.525440477425314e-09, + "loss": 0.605, + "step": 4053 + }, + { + "epoch": 0.9909557565387436, + "grad_norm": 3.204475164413452, + "learning_rate": 4.290410170173509e-09, + "loss": 0.4764, + "step": 4054 + }, + { + "epoch": 0.99120019555121, + "grad_norm": 10.434322357177734, + "learning_rate": 4.061645569103379e-09, + "loss": 0.6154, + "step": 4055 + }, + { + "epoch": 0.9914446345636764, + "grad_norm": 3.9556498527526855, + "learning_rate": 3.839146817612438e-09, + "loss": 0.5348, + "step": 4056 + }, + { + "epoch": 0.9916890735761428, + "grad_norm": 3.443718671798706, + "learning_rate": 3.6229140551735655e-09, + "loss": 0.5136, + "step": 4057 + }, + { + "epoch": 0.9919335125886092, + "grad_norm": 2.643991708755493, + "learning_rate": 3.4129474173294486e-09, + "loss": 0.4989, + "step": 4058 + }, + { + "epoch": 0.9921779516010756, + "grad_norm": 3.180394411087036, + "learning_rate": 3.209247035694807e-09, + "loss": 0.4559, + "step": 4059 + }, + { + "epoch": 0.9924223906135419, + "grad_norm": 3.485826015472412, + "learning_rate": 3.0118130379575005e-09, + "loss": 0.5038, + "step": 4060 + }, + { + "epoch": 0.9926668296260083, + "grad_norm": 3.4516539573669434, + "learning_rate": 2.8206455478774206e-09, + "loss": 0.5237, + "step": 4061 + }, + { + "epoch": 0.9929112686384747, + "grad_norm": 3.217087507247925, + "learning_rate": 2.6357446852853797e-09, + "loss": 0.5465, + "step": 4062 + }, + { + "epoch": 0.9931557076509411, + "grad_norm": 3.096763849258423, + "learning_rate": 2.45711056608533e-09, + "loss": 0.5782, + "step": 4063 + }, + { + "epoch": 0.9934001466634075, + "grad_norm": 3.3539602756500244, + "learning_rate": 2.284743302251036e-09, + "loss": 0.5717, + "step": 4064 + }, + { + "epoch": 0.9936445856758739, + "grad_norm": 3.5747156143188477, + "learning_rate": 2.1186430018294015e-09, + "loss": 0.4844, + "step": 4065 + }, + { + "epoch": 0.9938890246883403, + "grad_norm": 3.2676870822906494, + "learning_rate": 1.9588097689393627e-09, + "loss": 0.6203, + "step": 4066 + }, + { + "epoch": 0.9941334637008067, + "grad_norm": 3.1657159328460693, + "learning_rate": 1.805243703770776e-09, + "loss": 0.5333, + "step": 4067 + }, + { + "epoch": 0.9943779027132731, + "grad_norm": 2.9021496772766113, + "learning_rate": 1.6579449025844186e-09, + "loss": 0.5604, + "step": 4068 + }, + { + "epoch": 0.9946223417257394, + "grad_norm": 2.9915456771850586, + "learning_rate": 1.5169134577119882e-09, + "loss": 0.5709, + "step": 4069 + }, + { + "epoch": 0.9948667807382058, + "grad_norm": 3.2231616973876953, + "learning_rate": 1.3821494575594341e-09, + "loss": 0.519, + "step": 4070 + }, + { + "epoch": 0.9951112197506722, + "grad_norm": 4.008378505706787, + "learning_rate": 1.2536529866014058e-09, + "loss": 0.5854, + "step": 4071 + }, + { + "epoch": 0.9953556587631386, + "grad_norm": 3.2395238876342773, + "learning_rate": 1.1314241253834734e-09, + "loss": 0.5486, + "step": 4072 + }, + { + "epoch": 0.995600097775605, + "grad_norm": 3.368342161178589, + "learning_rate": 1.0154629505254587e-09, + "loss": 0.513, + "step": 4073 + }, + { + "epoch": 0.9958445367880714, + "grad_norm": 3.0852789878845215, + "learning_rate": 9.057695347158835e-10, + "loss": 0.5585, + "step": 4074 + }, + { + "epoch": 0.9960889758005378, + "grad_norm": 2.6465818881988525, + "learning_rate": 8.023439467141902e-10, + "loss": 1.8768, + "step": 4075 + }, + { + "epoch": 0.9963334148130042, + "grad_norm": 5.087937355041504, + "learning_rate": 7.051862513518526e-10, + "loss": 0.6375, + "step": 4076 + }, + { + "epoch": 0.9965778538254706, + "grad_norm": 3.305978536605835, + "learning_rate": 6.142965095312648e-10, + "loss": 0.6294, + "step": 4077 + }, + { + "epoch": 0.9968222928379369, + "grad_norm": 3.551265001296997, + "learning_rate": 5.296747782257416e-10, + "loss": 0.5115, + "step": 4078 + }, + { + "epoch": 0.9970667318504033, + "grad_norm": 3.1084768772125244, + "learning_rate": 4.51321110479519e-10, + "loss": 0.5921, + "step": 4079 + }, + { + "epoch": 0.9973111708628697, + "grad_norm": 3.146742820739746, + "learning_rate": 3.7923555540886336e-10, + "loss": 0.4884, + "step": 4080 + }, + { + "epoch": 0.9975556098753361, + "grad_norm": 3.0563406944274902, + "learning_rate": 3.1341815819763146e-10, + "loss": 0.5111, + "step": 4081 + }, + { + "epoch": 0.9978000488878025, + "grad_norm": 3.5022075176239014, + "learning_rate": 2.5386896010393126e-10, + "loss": 0.6503, + "step": 4082 + }, + { + "epoch": 0.9980444879002689, + "grad_norm": 3.6591250896453857, + "learning_rate": 2.0058799845568134e-10, + "loss": 0.6007, + "step": 4083 + }, + { + "epoch": 0.9982889269127353, + "grad_norm": 3.736536979675293, + "learning_rate": 1.5357530665172094e-10, + "loss": 0.5575, + "step": 4084 + }, + { + "epoch": 0.9985333659252017, + "grad_norm": 3.080501079559326, + "learning_rate": 1.128309141606998e-10, + "loss": 0.5473, + "step": 4085 + }, + { + "epoch": 0.9987778049376681, + "grad_norm": 3.0575947761535645, + "learning_rate": 7.835484652329861e-11, + "loss": 0.4791, + "step": 4086 + }, + { + "epoch": 0.9990222439501344, + "grad_norm": 3.32004976272583, + "learning_rate": 5.014712535111876e-11, + "loss": 0.5526, + "step": 4087 + }, + { + "epoch": 0.9992666829626008, + "grad_norm": 3.2589986324310303, + "learning_rate": 2.8207768324461928e-11, + "loss": 0.5164, + "step": 4088 + }, + { + "epoch": 0.9995111219750672, + "grad_norm": 3.316433906555176, + "learning_rate": 1.2536789196770927e-11, + "loss": 0.5081, + "step": 4089 + }, + { + "epoch": 0.9997555609875336, + "grad_norm": 3.078674077987671, + "learning_rate": 3.134197790188864e-12, + "loss": 0.64, + "step": 4090 + }, + { + "epoch": 1.0, + "grad_norm": 3.979902982711792, + "learning_rate": 0.0, + "loss": 0.7255, + "step": 4091 + }, + { + "epoch": 1.0, + "step": 4091, + "total_flos": 1.4976433291127685e+18, + "train_loss": 0.902596783321838, + "train_runtime": 27008.3323, + "train_samples_per_second": 7.269, + "train_steps_per_second": 0.151 + } + ], + "logging_steps": 1.0, + "max_steps": 4091, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 3500000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.4976433291127685e+18, + "train_batch_size": 6, + "trial_name": null, + "trial_params": null +}