diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,41790 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.9999371266897202, + "eval_steps": 500, + "global_step": 5964, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00016766216074609662, + "grad_norm": 6.196194648742676, + "learning_rate": 1.11731843575419e-07, + "loss": 3.11, + "step": 1 + }, + { + "epoch": 0.00033532432149219324, + "grad_norm": 6.196194648742676, + "learning_rate": 1.11731843575419e-07, + "loss": 3.0615, + "step": 2 + }, + { + "epoch": 0.0005029864822382899, + "grad_norm": 6.196194648742676, + "learning_rate": 1.11731843575419e-07, + "loss": 2.9922, + "step": 3 + }, + { + "epoch": 0.0006706486429843865, + "grad_norm": 6.196194648742676, + "learning_rate": 1.11731843575419e-07, + "loss": 2.9874, + "step": 4 + }, + { + "epoch": 0.0008383108037304831, + "grad_norm": 5.285811901092529, + "learning_rate": 2.23463687150838e-07, + "loss": 2.7919, + "step": 5 + }, + { + "epoch": 0.0010059729644765797, + "grad_norm": 5.377310752868652, + "learning_rate": 3.3519553072625703e-07, + "loss": 2.9085, + "step": 6 + }, + { + "epoch": 0.0011736351252226763, + "grad_norm": 4.72886848449707, + "learning_rate": 4.46927374301676e-07, + "loss": 2.6002, + "step": 7 + }, + { + "epoch": 0.001341297285968773, + "grad_norm": 4.93556022644043, + "learning_rate": 5.58659217877095e-07, + "loss": 2.4875, + "step": 8 + }, + { + "epoch": 0.0015089594467148696, + "grad_norm": 5.479576587677002, + "learning_rate": 6.703910614525141e-07, + "loss": 2.8652, + "step": 9 + }, + { + "epoch": 0.0016766216074609662, + "grad_norm": 5.086491584777832, + "learning_rate": 7.82122905027933e-07, + "loss": 2.6703, + "step": 10 + }, + { + "epoch": 0.0018442837682070628, + "grad_norm": 5.002711296081543, + "learning_rate": 8.93854748603352e-07, + "loss": 2.6001, + "step": 11 + }, + { + "epoch": 0.0020119459289531594, + "grad_norm": 6.0489115715026855, + "learning_rate": 1.005586592178771e-06, + "loss": 2.7088, + "step": 12 + }, + { + "epoch": 0.002179608089699256, + "grad_norm": 6.025905132293701, + "learning_rate": 1.11731843575419e-06, + "loss": 2.4601, + "step": 13 + }, + { + "epoch": 0.0023472702504453526, + "grad_norm": 5.1472296714782715, + "learning_rate": 1.229050279329609e-06, + "loss": 2.1754, + "step": 14 + }, + { + "epoch": 0.002514932411191449, + "grad_norm": 4.403075695037842, + "learning_rate": 1.3407821229050281e-06, + "loss": 1.8755, + "step": 15 + }, + { + "epoch": 0.002682594571937546, + "grad_norm": 3.35719633102417, + "learning_rate": 1.4525139664804472e-06, + "loss": 1.7016, + "step": 16 + }, + { + "epoch": 0.0028502567326836423, + "grad_norm": 4.584776878356934, + "learning_rate": 1.564245810055866e-06, + "loss": 1.7616, + "step": 17 + }, + { + "epoch": 0.003017918893429739, + "grad_norm": 2.5632901191711426, + "learning_rate": 1.675977653631285e-06, + "loss": 1.6525, + "step": 18 + }, + { + "epoch": 0.0031855810541758355, + "grad_norm": 2.6150174140930176, + "learning_rate": 1.787709497206704e-06, + "loss": 1.5917, + "step": 19 + }, + { + "epoch": 0.0033532432149219324, + "grad_norm": 2.1641123294830322, + "learning_rate": 1.899441340782123e-06, + "loss": 1.4867, + "step": 20 + }, + { + "epoch": 0.0035209053756680288, + "grad_norm": 2.4881510734558105, + "learning_rate": 2.011173184357542e-06, + "loss": 1.627, + "step": 21 + }, + { + "epoch": 0.0036885675364141256, + "grad_norm": 1.2478208541870117, + "learning_rate": 2.1229050279329612e-06, + "loss": 1.3723, + "step": 22 + }, + { + "epoch": 0.003856229697160222, + "grad_norm": 1.4412287473678589, + "learning_rate": 2.23463687150838e-06, + "loss": 1.4862, + "step": 23 + }, + { + "epoch": 0.004023891857906319, + "grad_norm": 1.1230992078781128, + "learning_rate": 2.3463687150837993e-06, + "loss": 1.3752, + "step": 24 + }, + { + "epoch": 0.004191554018652416, + "grad_norm": 1.0392589569091797, + "learning_rate": 2.458100558659218e-06, + "loss": 1.3647, + "step": 25 + }, + { + "epoch": 0.004359216179398512, + "grad_norm": 0.9552459716796875, + "learning_rate": 2.569832402234637e-06, + "loss": 1.3772, + "step": 26 + }, + { + "epoch": 0.0045268783401446085, + "grad_norm": 1.024999737739563, + "learning_rate": 2.6815642458100562e-06, + "loss": 1.2639, + "step": 27 + }, + { + "epoch": 0.004694540500890705, + "grad_norm": 0.9447226524353027, + "learning_rate": 2.793296089385475e-06, + "loss": 1.4179, + "step": 28 + }, + { + "epoch": 0.004862202661636802, + "grad_norm": 1.181274175643921, + "learning_rate": 2.9050279329608943e-06, + "loss": 1.5209, + "step": 29 + }, + { + "epoch": 0.005029864822382898, + "grad_norm": 0.858210563659668, + "learning_rate": 3.016759776536313e-06, + "loss": 1.3232, + "step": 30 + }, + { + "epoch": 0.005197526983128995, + "grad_norm": 1.189065933227539, + "learning_rate": 3.128491620111732e-06, + "loss": 1.5529, + "step": 31 + }, + { + "epoch": 0.005365189143875092, + "grad_norm": 0.9217438697814941, + "learning_rate": 3.240223463687151e-06, + "loss": 1.3698, + "step": 32 + }, + { + "epoch": 0.005532851304621189, + "grad_norm": 1.0276325941085815, + "learning_rate": 3.35195530726257e-06, + "loss": 1.4141, + "step": 33 + }, + { + "epoch": 0.005700513465367285, + "grad_norm": 0.8150102496147156, + "learning_rate": 3.4636871508379893e-06, + "loss": 1.2984, + "step": 34 + }, + { + "epoch": 0.005868175626113381, + "grad_norm": 0.7057741284370422, + "learning_rate": 3.575418994413408e-06, + "loss": 1.3079, + "step": 35 + }, + { + "epoch": 0.006035837786859478, + "grad_norm": 0.8169935345649719, + "learning_rate": 3.687150837988827e-06, + "loss": 1.4512, + "step": 36 + }, + { + "epoch": 0.006203499947605575, + "grad_norm": 0.61067134141922, + "learning_rate": 3.798882681564246e-06, + "loss": 1.1787, + "step": 37 + }, + { + "epoch": 0.006371162108351671, + "grad_norm": 0.7273681163787842, + "learning_rate": 3.910614525139665e-06, + "loss": 1.3199, + "step": 38 + }, + { + "epoch": 0.006538824269097768, + "grad_norm": 0.6396476030349731, + "learning_rate": 4.022346368715084e-06, + "loss": 1.1523, + "step": 39 + }, + { + "epoch": 0.006706486429843865, + "grad_norm": 0.5737459659576416, + "learning_rate": 4.134078212290504e-06, + "loss": 1.1673, + "step": 40 + }, + { + "epoch": 0.0068741485905899615, + "grad_norm": 0.719789445400238, + "learning_rate": 4.2458100558659224e-06, + "loss": 1.3278, + "step": 41 + }, + { + "epoch": 0.0070418107513360575, + "grad_norm": 0.5431291460990906, + "learning_rate": 4.357541899441341e-06, + "loss": 1.2429, + "step": 42 + }, + { + "epoch": 0.007209472912082154, + "grad_norm": 0.6092455387115479, + "learning_rate": 4.46927374301676e-06, + "loss": 1.1985, + "step": 43 + }, + { + "epoch": 0.007377135072828251, + "grad_norm": 0.5443903803825378, + "learning_rate": 4.581005586592179e-06, + "loss": 1.1565, + "step": 44 + }, + { + "epoch": 0.007544797233574348, + "grad_norm": 0.5537682175636292, + "learning_rate": 4.692737430167599e-06, + "loss": 1.2249, + "step": 45 + }, + { + "epoch": 0.007712459394320444, + "grad_norm": 0.6771538853645325, + "learning_rate": 4.8044692737430175e-06, + "loss": 1.1947, + "step": 46 + }, + { + "epoch": 0.007880121555066542, + "grad_norm": 0.5913676023483276, + "learning_rate": 4.916201117318436e-06, + "loss": 1.1952, + "step": 47 + }, + { + "epoch": 0.008047783715812638, + "grad_norm": 0.5791601538658142, + "learning_rate": 5.027932960893855e-06, + "loss": 1.2931, + "step": 48 + }, + { + "epoch": 0.008215445876558734, + "grad_norm": 0.6004194617271423, + "learning_rate": 5.139664804469274e-06, + "loss": 1.207, + "step": 49 + }, + { + "epoch": 0.008383108037304831, + "grad_norm": 0.6252710223197937, + "learning_rate": 5.251396648044693e-06, + "loss": 1.2427, + "step": 50 + }, + { + "epoch": 0.008550770198050927, + "grad_norm": 0.5790568590164185, + "learning_rate": 5.3631284916201125e-06, + "loss": 1.2262, + "step": 51 + }, + { + "epoch": 0.008718432358797023, + "grad_norm": 0.5642638802528381, + "learning_rate": 5.474860335195531e-06, + "loss": 1.1652, + "step": 52 + }, + { + "epoch": 0.008886094519543121, + "grad_norm": 0.5329627394676208, + "learning_rate": 5.58659217877095e-06, + "loss": 1.1425, + "step": 53 + }, + { + "epoch": 0.009053756680289217, + "grad_norm": 0.6454995274543762, + "learning_rate": 5.698324022346369e-06, + "loss": 1.2716, + "step": 54 + }, + { + "epoch": 0.009221418841035315, + "grad_norm": 0.6814166307449341, + "learning_rate": 5.810055865921789e-06, + "loss": 1.1937, + "step": 55 + }, + { + "epoch": 0.00938908100178141, + "grad_norm": 0.6075778007507324, + "learning_rate": 5.9217877094972075e-06, + "loss": 1.2141, + "step": 56 + }, + { + "epoch": 0.009556743162527507, + "grad_norm": 0.53324955701828, + "learning_rate": 6.033519553072626e-06, + "loss": 1.1361, + "step": 57 + }, + { + "epoch": 0.009724405323273604, + "grad_norm": 0.5701301693916321, + "learning_rate": 6.145251396648045e-06, + "loss": 1.1547, + "step": 58 + }, + { + "epoch": 0.0098920674840197, + "grad_norm": 0.4995803236961365, + "learning_rate": 6.256983240223464e-06, + "loss": 1.1433, + "step": 59 + }, + { + "epoch": 0.010059729644765796, + "grad_norm": 0.5348702669143677, + "learning_rate": 6.368715083798883e-06, + "loss": 1.1044, + "step": 60 + }, + { + "epoch": 0.010227391805511894, + "grad_norm": 0.5916696190834045, + "learning_rate": 6.480446927374302e-06, + "loss": 1.1822, + "step": 61 + }, + { + "epoch": 0.01039505396625799, + "grad_norm": 0.5305972695350647, + "learning_rate": 6.592178770949721e-06, + "loss": 1.2197, + "step": 62 + }, + { + "epoch": 0.010562716127004088, + "grad_norm": 0.7340630888938904, + "learning_rate": 6.70391061452514e-06, + "loss": 1.2318, + "step": 63 + }, + { + "epoch": 0.010730378287750184, + "grad_norm": 0.5850639939308167, + "learning_rate": 6.815642458100559e-06, + "loss": 1.2176, + "step": 64 + }, + { + "epoch": 0.01089804044849628, + "grad_norm": 0.6341410279273987, + "learning_rate": 6.927374301675979e-06, + "loss": 1.1692, + "step": 65 + }, + { + "epoch": 0.011065702609242377, + "grad_norm": 0.766492486000061, + "learning_rate": 7.0391061452513975e-06, + "loss": 1.147, + "step": 66 + }, + { + "epoch": 0.011233364769988473, + "grad_norm": 0.541501522064209, + "learning_rate": 7.150837988826816e-06, + "loss": 1.2227, + "step": 67 + }, + { + "epoch": 0.01140102693073457, + "grad_norm": 0.47469985485076904, + "learning_rate": 7.262569832402235e-06, + "loss": 1.1507, + "step": 68 + }, + { + "epoch": 0.011568689091480667, + "grad_norm": 0.6132412552833557, + "learning_rate": 7.374301675977654e-06, + "loss": 1.282, + "step": 69 + }, + { + "epoch": 0.011736351252226763, + "grad_norm": 0.4676038920879364, + "learning_rate": 7.486033519553073e-06, + "loss": 1.1221, + "step": 70 + }, + { + "epoch": 0.01190401341297286, + "grad_norm": 0.5179970264434814, + "learning_rate": 7.597765363128492e-06, + "loss": 1.0446, + "step": 71 + }, + { + "epoch": 0.012071675573718956, + "grad_norm": 0.4430614113807678, + "learning_rate": 7.709497206703911e-06, + "loss": 1.1906, + "step": 72 + }, + { + "epoch": 0.012239337734465052, + "grad_norm": 0.4925226867198944, + "learning_rate": 7.82122905027933e-06, + "loss": 1.1505, + "step": 73 + }, + { + "epoch": 0.01240699989521115, + "grad_norm": 0.5351929664611816, + "learning_rate": 7.932960893854749e-06, + "loss": 1.207, + "step": 74 + }, + { + "epoch": 0.012574662055957246, + "grad_norm": 0.5194462537765503, + "learning_rate": 8.044692737430168e-06, + "loss": 1.1883, + "step": 75 + }, + { + "epoch": 0.012742324216703342, + "grad_norm": 0.5400429964065552, + "learning_rate": 8.156424581005588e-06, + "loss": 1.0012, + "step": 76 + }, + { + "epoch": 0.01290998637744944, + "grad_norm": 0.47353821992874146, + "learning_rate": 8.268156424581007e-06, + "loss": 1.171, + "step": 77 + }, + { + "epoch": 0.013077648538195536, + "grad_norm": 0.6433570384979248, + "learning_rate": 8.379888268156426e-06, + "loss": 1.2229, + "step": 78 + }, + { + "epoch": 0.013245310698941633, + "grad_norm": 0.5411203503608704, + "learning_rate": 8.491620111731845e-06, + "loss": 1.0776, + "step": 79 + }, + { + "epoch": 0.01341297285968773, + "grad_norm": 0.6546565294265747, + "learning_rate": 8.603351955307264e-06, + "loss": 1.1872, + "step": 80 + }, + { + "epoch": 0.013580635020433825, + "grad_norm": 0.6882161498069763, + "learning_rate": 8.715083798882683e-06, + "loss": 1.1454, + "step": 81 + }, + { + "epoch": 0.013748297181179923, + "grad_norm": 0.5556113123893738, + "learning_rate": 8.826815642458101e-06, + "loss": 1.1364, + "step": 82 + }, + { + "epoch": 0.013915959341926019, + "grad_norm": 0.6302836537361145, + "learning_rate": 8.93854748603352e-06, + "loss": 1.2659, + "step": 83 + }, + { + "epoch": 0.014083621502672115, + "grad_norm": 0.8605783581733704, + "learning_rate": 9.050279329608939e-06, + "loss": 1.1607, + "step": 84 + }, + { + "epoch": 0.014251283663418213, + "grad_norm": 0.5417123436927795, + "learning_rate": 9.162011173184358e-06, + "loss": 1.1393, + "step": 85 + }, + { + "epoch": 0.014418945824164309, + "grad_norm": 0.6310127973556519, + "learning_rate": 9.273743016759777e-06, + "loss": 1.3032, + "step": 86 + }, + { + "epoch": 0.014586607984910405, + "grad_norm": 0.5595749020576477, + "learning_rate": 9.385474860335197e-06, + "loss": 1.1699, + "step": 87 + }, + { + "epoch": 0.014754270145656502, + "grad_norm": 0.5508741736412048, + "learning_rate": 9.497206703910616e-06, + "loss": 1.0422, + "step": 88 + }, + { + "epoch": 0.014921932306402598, + "grad_norm": 0.44198355078697205, + "learning_rate": 9.608938547486035e-06, + "loss": 1.006, + "step": 89 + }, + { + "epoch": 0.015089594467148696, + "grad_norm": 0.612446129322052, + "learning_rate": 9.720670391061454e-06, + "loss": 1.2126, + "step": 90 + }, + { + "epoch": 0.015257256627894792, + "grad_norm": 0.534264087677002, + "learning_rate": 9.832402234636873e-06, + "loss": 1.0954, + "step": 91 + }, + { + "epoch": 0.015424918788640888, + "grad_norm": 0.4617749750614166, + "learning_rate": 9.944134078212291e-06, + "loss": 1.0987, + "step": 92 + }, + { + "epoch": 0.015592580949386986, + "grad_norm": 0.5459811687469482, + "learning_rate": 1.005586592178771e-05, + "loss": 1.1247, + "step": 93 + }, + { + "epoch": 0.015760243110133083, + "grad_norm": 0.4880772829055786, + "learning_rate": 1.0167597765363129e-05, + "loss": 1.0422, + "step": 94 + }, + { + "epoch": 0.01592790527087918, + "grad_norm": 0.5880354046821594, + "learning_rate": 1.0279329608938548e-05, + "loss": 1.166, + "step": 95 + }, + { + "epoch": 0.016095567431625275, + "grad_norm": 0.5863614678382874, + "learning_rate": 1.0391061452513967e-05, + "loss": 1.1016, + "step": 96 + }, + { + "epoch": 0.01626322959237137, + "grad_norm": 0.5670218467712402, + "learning_rate": 1.0502793296089386e-05, + "loss": 0.9635, + "step": 97 + }, + { + "epoch": 0.016430891753117467, + "grad_norm": 0.8734570741653442, + "learning_rate": 1.0614525139664806e-05, + "loss": 1.0499, + "step": 98 + }, + { + "epoch": 0.016598553913863563, + "grad_norm": 0.6028108596801758, + "learning_rate": 1.0726256983240225e-05, + "loss": 1.2074, + "step": 99 + }, + { + "epoch": 0.016766216074609663, + "grad_norm": 0.5182443857192993, + "learning_rate": 1.0837988826815644e-05, + "loss": 1.0133, + "step": 100 + }, + { + "epoch": 0.01693387823535576, + "grad_norm": 0.7909883260726929, + "learning_rate": 1.0949720670391063e-05, + "loss": 1.1455, + "step": 101 + }, + { + "epoch": 0.017101540396101855, + "grad_norm": 0.6897411942481995, + "learning_rate": 1.1061452513966481e-05, + "loss": 1.0995, + "step": 102 + }, + { + "epoch": 0.01726920255684795, + "grad_norm": 0.6529523134231567, + "learning_rate": 1.11731843575419e-05, + "loss": 1.0004, + "step": 103 + }, + { + "epoch": 0.017436864717594047, + "grad_norm": 0.68169766664505, + "learning_rate": 1.1284916201117319e-05, + "loss": 1.1065, + "step": 104 + }, + { + "epoch": 0.017604526878340146, + "grad_norm": 0.7525808811187744, + "learning_rate": 1.1396648044692738e-05, + "loss": 1.1017, + "step": 105 + }, + { + "epoch": 0.017772189039086242, + "grad_norm": 0.6268847584724426, + "learning_rate": 1.1508379888268157e-05, + "loss": 1.0249, + "step": 106 + }, + { + "epoch": 0.017939851199832338, + "grad_norm": 0.6089746952056885, + "learning_rate": 1.1620111731843577e-05, + "loss": 1.1194, + "step": 107 + }, + { + "epoch": 0.018107513360578434, + "grad_norm": 0.8341772556304932, + "learning_rate": 1.1731843575418994e-05, + "loss": 1.1703, + "step": 108 + }, + { + "epoch": 0.01827517552132453, + "grad_norm": 0.5683283805847168, + "learning_rate": 1.1843575418994415e-05, + "loss": 1.0827, + "step": 109 + }, + { + "epoch": 0.01844283768207063, + "grad_norm": 0.6581299304962158, + "learning_rate": 1.1955307262569834e-05, + "loss": 1.091, + "step": 110 + }, + { + "epoch": 0.018610499842816725, + "grad_norm": 0.489067018032074, + "learning_rate": 1.2067039106145253e-05, + "loss": 1.0926, + "step": 111 + }, + { + "epoch": 0.01877816200356282, + "grad_norm": 0.76224684715271, + "learning_rate": 1.2178770949720671e-05, + "loss": 1.0664, + "step": 112 + }, + { + "epoch": 0.018945824164308917, + "grad_norm": 0.5604745149612427, + "learning_rate": 1.229050279329609e-05, + "loss": 1.1076, + "step": 113 + }, + { + "epoch": 0.019113486325055013, + "grad_norm": 0.6854090094566345, + "learning_rate": 1.2402234636871509e-05, + "loss": 1.128, + "step": 114 + }, + { + "epoch": 0.01928114848580111, + "grad_norm": 0.7920408844947815, + "learning_rate": 1.2513966480446928e-05, + "loss": 1.2003, + "step": 115 + }, + { + "epoch": 0.01944881064654721, + "grad_norm": 0.6446617841720581, + "learning_rate": 1.2625698324022347e-05, + "loss": 1.1067, + "step": 116 + }, + { + "epoch": 0.019616472807293305, + "grad_norm": 0.5878286361694336, + "learning_rate": 1.2737430167597766e-05, + "loss": 1.116, + "step": 117 + }, + { + "epoch": 0.0197841349680394, + "grad_norm": 0.46466970443725586, + "learning_rate": 1.2849162011173186e-05, + "loss": 1.0498, + "step": 118 + }, + { + "epoch": 0.019951797128785496, + "grad_norm": 0.6509464383125305, + "learning_rate": 1.2960893854748603e-05, + "loss": 1.1395, + "step": 119 + }, + { + "epoch": 0.020119459289531592, + "grad_norm": 0.6141960024833679, + "learning_rate": 1.3072625698324024e-05, + "loss": 0.9784, + "step": 120 + }, + { + "epoch": 0.020287121450277692, + "grad_norm": 0.6657926440238953, + "learning_rate": 1.3184357541899443e-05, + "loss": 1.0352, + "step": 121 + }, + { + "epoch": 0.020454783611023788, + "grad_norm": 0.5377758741378784, + "learning_rate": 1.3296089385474861e-05, + "loss": 1.0195, + "step": 122 + }, + { + "epoch": 0.020622445771769884, + "grad_norm": 0.5579245686531067, + "learning_rate": 1.340782122905028e-05, + "loss": 1.0121, + "step": 123 + }, + { + "epoch": 0.02079010793251598, + "grad_norm": 0.5483793020248413, + "learning_rate": 1.3519553072625699e-05, + "loss": 1.0246, + "step": 124 + }, + { + "epoch": 0.020957770093262076, + "grad_norm": 0.5603605508804321, + "learning_rate": 1.3631284916201118e-05, + "loss": 1.0742, + "step": 125 + }, + { + "epoch": 0.021125432254008175, + "grad_norm": 0.761964738368988, + "learning_rate": 1.3743016759776537e-05, + "loss": 1.0959, + "step": 126 + }, + { + "epoch": 0.02129309441475427, + "grad_norm": 0.49273210763931274, + "learning_rate": 1.3854748603351957e-05, + "loss": 1.2079, + "step": 127 + }, + { + "epoch": 0.021460756575500367, + "grad_norm": 0.4556741714477539, + "learning_rate": 1.3966480446927374e-05, + "loss": 1.0563, + "step": 128 + }, + { + "epoch": 0.021628418736246463, + "grad_norm": 0.6324474811553955, + "learning_rate": 1.4078212290502795e-05, + "loss": 1.002, + "step": 129 + }, + { + "epoch": 0.02179608089699256, + "grad_norm": 0.5139722228050232, + "learning_rate": 1.4189944134078212e-05, + "loss": 1.0051, + "step": 130 + }, + { + "epoch": 0.021963743057738655, + "grad_norm": 0.5473005175590515, + "learning_rate": 1.4301675977653633e-05, + "loss": 1.051, + "step": 131 + }, + { + "epoch": 0.022131405218484754, + "grad_norm": 0.6989631056785583, + "learning_rate": 1.4413407821229052e-05, + "loss": 1.0833, + "step": 132 + }, + { + "epoch": 0.02229906737923085, + "grad_norm": 0.6231628656387329, + "learning_rate": 1.452513966480447e-05, + "loss": 1.1625, + "step": 133 + }, + { + "epoch": 0.022466729539976946, + "grad_norm": 0.7199480533599854, + "learning_rate": 1.463687150837989e-05, + "loss": 1.0238, + "step": 134 + }, + { + "epoch": 0.022634391700723042, + "grad_norm": 0.5824499130249023, + "learning_rate": 1.4748603351955308e-05, + "loss": 1.0499, + "step": 135 + }, + { + "epoch": 0.02280205386146914, + "grad_norm": 0.6677570343017578, + "learning_rate": 1.4860335195530729e-05, + "loss": 1.1293, + "step": 136 + }, + { + "epoch": 0.022969716022215238, + "grad_norm": 0.5944182276725769, + "learning_rate": 1.4972067039106146e-05, + "loss": 1.0386, + "step": 137 + }, + { + "epoch": 0.023137378182961334, + "grad_norm": 1.3920981884002686, + "learning_rate": 1.5083798882681566e-05, + "loss": 1.1251, + "step": 138 + }, + { + "epoch": 0.02330504034370743, + "grad_norm": 0.7003018260002136, + "learning_rate": 1.5195530726256983e-05, + "loss": 1.0233, + "step": 139 + }, + { + "epoch": 0.023472702504453526, + "grad_norm": 0.6967798471450806, + "learning_rate": 1.5307262569832404e-05, + "loss": 1.0562, + "step": 140 + }, + { + "epoch": 0.02364036466519962, + "grad_norm": 0.5572317838668823, + "learning_rate": 1.5418994413407823e-05, + "loss": 0.9461, + "step": 141 + }, + { + "epoch": 0.02380802682594572, + "grad_norm": 0.7279749512672424, + "learning_rate": 1.553072625698324e-05, + "loss": 1.0506, + "step": 142 + }, + { + "epoch": 0.023975688986691817, + "grad_norm": 0.5048472285270691, + "learning_rate": 1.564245810055866e-05, + "loss": 1.048, + "step": 143 + }, + { + "epoch": 0.024143351147437913, + "grad_norm": 0.6082873344421387, + "learning_rate": 1.575418994413408e-05, + "loss": 1.0414, + "step": 144 + }, + { + "epoch": 0.02431101330818401, + "grad_norm": 0.5643455386161804, + "learning_rate": 1.5865921787709498e-05, + "loss": 1.0404, + "step": 145 + }, + { + "epoch": 0.024478675468930105, + "grad_norm": 0.46841081976890564, + "learning_rate": 1.5977653631284917e-05, + "loss": 1.0697, + "step": 146 + }, + { + "epoch": 0.0246463376296762, + "grad_norm": 0.5768006443977356, + "learning_rate": 1.6089385474860336e-05, + "loss": 1.1046, + "step": 147 + }, + { + "epoch": 0.0248139997904223, + "grad_norm": 0.6741750836372375, + "learning_rate": 1.6201117318435755e-05, + "loss": 1.0681, + "step": 148 + }, + { + "epoch": 0.024981661951168396, + "grad_norm": 0.6723300814628601, + "learning_rate": 1.6312849162011177e-05, + "loss": 1.1534, + "step": 149 + }, + { + "epoch": 0.025149324111914492, + "grad_norm": 0.5499408841133118, + "learning_rate": 1.6424581005586592e-05, + "loss": 1.0647, + "step": 150 + }, + { + "epoch": 0.025316986272660588, + "grad_norm": 0.5709015727043152, + "learning_rate": 1.6536312849162014e-05, + "loss": 0.9977, + "step": 151 + }, + { + "epoch": 0.025484648433406684, + "grad_norm": 0.6413765549659729, + "learning_rate": 1.664804469273743e-05, + "loss": 1.1216, + "step": 152 + }, + { + "epoch": 0.025652310594152784, + "grad_norm": 0.566363513469696, + "learning_rate": 1.6759776536312852e-05, + "loss": 1.1001, + "step": 153 + }, + { + "epoch": 0.02581997275489888, + "grad_norm": 0.5891416072845459, + "learning_rate": 1.687150837988827e-05, + "loss": 0.9475, + "step": 154 + }, + { + "epoch": 0.025987634915644976, + "grad_norm": 0.6749016046524048, + "learning_rate": 1.698324022346369e-05, + "loss": 1.1254, + "step": 155 + }, + { + "epoch": 0.02615529707639107, + "grad_norm": 0.5689506530761719, + "learning_rate": 1.709497206703911e-05, + "loss": 1.0695, + "step": 156 + }, + { + "epoch": 0.026322959237137167, + "grad_norm": 0.8832252621650696, + "learning_rate": 1.7206703910614527e-05, + "loss": 1.0325, + "step": 157 + }, + { + "epoch": 0.026490621397883267, + "grad_norm": 0.7964456081390381, + "learning_rate": 1.7318435754189946e-05, + "loss": 1.0879, + "step": 158 + }, + { + "epoch": 0.026658283558629363, + "grad_norm": 0.5532066822052002, + "learning_rate": 1.7430167597765365e-05, + "loss": 1.0784, + "step": 159 + }, + { + "epoch": 0.02682594571937546, + "grad_norm": 0.6335783004760742, + "learning_rate": 1.7541899441340784e-05, + "loss": 1.0252, + "step": 160 + }, + { + "epoch": 0.026993607880121555, + "grad_norm": 0.5445144176483154, + "learning_rate": 1.7653631284916203e-05, + "loss": 1.1504, + "step": 161 + }, + { + "epoch": 0.02716127004086765, + "grad_norm": 0.6088312268257141, + "learning_rate": 1.776536312849162e-05, + "loss": 1.0816, + "step": 162 + }, + { + "epoch": 0.027328932201613747, + "grad_norm": 0.6831774711608887, + "learning_rate": 1.787709497206704e-05, + "loss": 1.035, + "step": 163 + }, + { + "epoch": 0.027496594362359846, + "grad_norm": 0.4983123242855072, + "learning_rate": 1.798882681564246e-05, + "loss": 1.0276, + "step": 164 + }, + { + "epoch": 0.027664256523105942, + "grad_norm": 0.5578044056892395, + "learning_rate": 1.8100558659217878e-05, + "loss": 1.0178, + "step": 165 + }, + { + "epoch": 0.027831918683852038, + "grad_norm": 0.7284736633300781, + "learning_rate": 1.8212290502793297e-05, + "loss": 0.9363, + "step": 166 + }, + { + "epoch": 0.027999580844598134, + "grad_norm": 0.43957066535949707, + "learning_rate": 1.8324022346368716e-05, + "loss": 1.0298, + "step": 167 + }, + { + "epoch": 0.02816724300534423, + "grad_norm": 0.46956202387809753, + "learning_rate": 1.8435754189944135e-05, + "loss": 1.0963, + "step": 168 + }, + { + "epoch": 0.02833490516609033, + "grad_norm": 0.6402130126953125, + "learning_rate": 1.8547486033519553e-05, + "loss": 1.0679, + "step": 169 + }, + { + "epoch": 0.028502567326836425, + "grad_norm": 0.6925402283668518, + "learning_rate": 1.8659217877094972e-05, + "loss": 1.0334, + "step": 170 + }, + { + "epoch": 0.02867022948758252, + "grad_norm": 1.1133993864059448, + "learning_rate": 1.8770949720670394e-05, + "loss": 0.9692, + "step": 171 + }, + { + "epoch": 0.028837891648328617, + "grad_norm": 0.5076237916946411, + "learning_rate": 1.888268156424581e-05, + "loss": 1.1312, + "step": 172 + }, + { + "epoch": 0.029005553809074713, + "grad_norm": 0.6855918169021606, + "learning_rate": 1.8994413407821232e-05, + "loss": 1.0913, + "step": 173 + }, + { + "epoch": 0.02917321596982081, + "grad_norm": 0.5001389384269714, + "learning_rate": 1.910614525139665e-05, + "loss": 1.1183, + "step": 174 + }, + { + "epoch": 0.02934087813056691, + "grad_norm": 0.5716369152069092, + "learning_rate": 1.921787709497207e-05, + "loss": 0.9813, + "step": 175 + }, + { + "epoch": 0.029508540291313005, + "grad_norm": 0.7187888622283936, + "learning_rate": 1.932960893854749e-05, + "loss": 0.9594, + "step": 176 + }, + { + "epoch": 0.0296762024520591, + "grad_norm": 0.5117635726928711, + "learning_rate": 1.9441340782122907e-05, + "loss": 1.0199, + "step": 177 + }, + { + "epoch": 0.029843864612805197, + "grad_norm": 0.5199819803237915, + "learning_rate": 1.9553072625698326e-05, + "loss": 1.0014, + "step": 178 + }, + { + "epoch": 0.030011526773551293, + "grad_norm": 0.5007014870643616, + "learning_rate": 1.9664804469273745e-05, + "loss": 0.9336, + "step": 179 + }, + { + "epoch": 0.030179188934297392, + "grad_norm": 0.5830110907554626, + "learning_rate": 1.9776536312849164e-05, + "loss": 1.0689, + "step": 180 + }, + { + "epoch": 0.030346851095043488, + "grad_norm": 0.527374804019928, + "learning_rate": 1.9888268156424583e-05, + "loss": 1.0273, + "step": 181 + }, + { + "epoch": 0.030514513255789584, + "grad_norm": 0.5847291350364685, + "learning_rate": 2e-05, + "loss": 1.0122, + "step": 182 + }, + { + "epoch": 0.03068217541653568, + "grad_norm": 0.5212209820747375, + "learning_rate": 1.9999998525438088e-05, + "loss": 0.8963, + "step": 183 + }, + { + "epoch": 0.030849837577281776, + "grad_norm": 0.49605804681777954, + "learning_rate": 1.999999410175278e-05, + "loss": 0.9395, + "step": 184 + }, + { + "epoch": 0.031017499738027875, + "grad_norm": 0.47218137979507446, + "learning_rate": 1.9999986728945384e-05, + "loss": 1.0157, + "step": 185 + }, + { + "epoch": 0.03118516189877397, + "grad_norm": 0.7420831322669983, + "learning_rate": 1.9999976407018074e-05, + "loss": 1.0173, + "step": 186 + }, + { + "epoch": 0.031352824059520064, + "grad_norm": 0.48486003279685974, + "learning_rate": 1.9999963135973894e-05, + "loss": 0.9915, + "step": 187 + }, + { + "epoch": 0.03152048622026617, + "grad_norm": 0.6362403035163879, + "learning_rate": 1.9999946915816756e-05, + "loss": 1.0975, + "step": 188 + }, + { + "epoch": 0.03168814838101226, + "grad_norm": 0.5725786685943604, + "learning_rate": 1.9999927746551447e-05, + "loss": 1.0043, + "step": 189 + }, + { + "epoch": 0.03185581054175836, + "grad_norm": 0.5909467339515686, + "learning_rate": 1.9999905628183617e-05, + "loss": 1.0289, + "step": 190 + }, + { + "epoch": 0.032023472702504455, + "grad_norm": 0.5642744898796082, + "learning_rate": 1.999988056071979e-05, + "loss": 1.1165, + "step": 191 + }, + { + "epoch": 0.03219113486325055, + "grad_norm": 0.3727734386920929, + "learning_rate": 1.999985254416736e-05, + "loss": 0.9655, + "step": 192 + }, + { + "epoch": 0.03235879702399665, + "grad_norm": 0.48998305201530457, + "learning_rate": 1.9999821578534588e-05, + "loss": 1.1287, + "step": 193 + }, + { + "epoch": 0.03252645918474274, + "grad_norm": 0.44185349345207214, + "learning_rate": 1.9999787663830606e-05, + "loss": 0.9565, + "step": 194 + }, + { + "epoch": 0.03269412134548884, + "grad_norm": 0.40309658646583557, + "learning_rate": 1.9999750800065415e-05, + "loss": 1.0181, + "step": 195 + }, + { + "epoch": 0.032861783506234934, + "grad_norm": 0.5723433494567871, + "learning_rate": 1.9999710987249892e-05, + "loss": 1.0667, + "step": 196 + }, + { + "epoch": 0.03302944566698103, + "grad_norm": 0.5408050417900085, + "learning_rate": 1.9999668225395772e-05, + "loss": 0.9858, + "step": 197 + }, + { + "epoch": 0.033197107827727126, + "grad_norm": 0.83476322889328, + "learning_rate": 1.9999622514515667e-05, + "loss": 1.0893, + "step": 198 + }, + { + "epoch": 0.03336476998847323, + "grad_norm": 0.8037353754043579, + "learning_rate": 1.999957385462306e-05, + "loss": 0.9974, + "step": 199 + }, + { + "epoch": 0.033532432149219325, + "grad_norm": 0.7400439381599426, + "learning_rate": 1.9999522245732302e-05, + "loss": 1.1179, + "step": 200 + }, + { + "epoch": 0.03370009430996542, + "grad_norm": 0.5144438743591309, + "learning_rate": 1.999946768785861e-05, + "loss": 1.0092, + "step": 201 + }, + { + "epoch": 0.03386775647071152, + "grad_norm": 0.5504910349845886, + "learning_rate": 1.9999410181018074e-05, + "loss": 0.9403, + "step": 202 + }, + { + "epoch": 0.03403541863145761, + "grad_norm": 0.5855805277824402, + "learning_rate": 1.999934972522766e-05, + "loss": 1.0515, + "step": 203 + }, + { + "epoch": 0.03420308079220371, + "grad_norm": 0.7673451900482178, + "learning_rate": 1.999928632050519e-05, + "loss": 1.1424, + "step": 204 + }, + { + "epoch": 0.034370742952949805, + "grad_norm": 0.7211079597473145, + "learning_rate": 1.999921996686936e-05, + "loss": 0.9997, + "step": 205 + }, + { + "epoch": 0.0345384051136959, + "grad_norm": 0.4259258806705475, + "learning_rate": 1.9999150664339745e-05, + "loss": 0.9601, + "step": 206 + }, + { + "epoch": 0.034706067274442, + "grad_norm": 0.4799496531486511, + "learning_rate": 1.9999078412936784e-05, + "loss": 1.0027, + "step": 207 + }, + { + "epoch": 0.03487372943518809, + "grad_norm": 0.5094309449195862, + "learning_rate": 1.999900321268178e-05, + "loss": 0.9668, + "step": 208 + }, + { + "epoch": 0.035041391595934196, + "grad_norm": 0.49225088953971863, + "learning_rate": 1.9998925063596915e-05, + "loss": 1.0153, + "step": 209 + }, + { + "epoch": 0.03520905375668029, + "grad_norm": 0.4400315582752228, + "learning_rate": 1.9998843965705235e-05, + "loss": 1.0868, + "step": 210 + }, + { + "epoch": 0.03537671591742639, + "grad_norm": 0.6707110404968262, + "learning_rate": 1.9998759919030653e-05, + "loss": 1.0686, + "step": 211 + }, + { + "epoch": 0.035544378078172484, + "grad_norm": 0.4518424868583679, + "learning_rate": 1.999867292359796e-05, + "loss": 1.0012, + "step": 212 + }, + { + "epoch": 0.03571204023891858, + "grad_norm": 0.39341041445732117, + "learning_rate": 1.999858297943281e-05, + "loss": 0.9752, + "step": 213 + }, + { + "epoch": 0.035879702399664676, + "grad_norm": 0.42496252059936523, + "learning_rate": 1.9998490086561727e-05, + "loss": 1.008, + "step": 214 + }, + { + "epoch": 0.03604736456041077, + "grad_norm": 0.5031048059463501, + "learning_rate": 1.999839424501211e-05, + "loss": 1.024, + "step": 215 + }, + { + "epoch": 0.03621502672115687, + "grad_norm": 0.4959554970264435, + "learning_rate": 1.9998295454812218e-05, + "loss": 0.9674, + "step": 216 + }, + { + "epoch": 0.036382688881902964, + "grad_norm": 0.5379063487052917, + "learning_rate": 1.9998193715991192e-05, + "loss": 1.0666, + "step": 217 + }, + { + "epoch": 0.03655035104264906, + "grad_norm": 0.46192410588264465, + "learning_rate": 1.9998089028579035e-05, + "loss": 0.9727, + "step": 218 + }, + { + "epoch": 0.036718013203395156, + "grad_norm": 0.5664047002792358, + "learning_rate": 1.9997981392606617e-05, + "loss": 1.0543, + "step": 219 + }, + { + "epoch": 0.03688567536414126, + "grad_norm": 0.7212116122245789, + "learning_rate": 1.9997870808105683e-05, + "loss": 1.07, + "step": 220 + }, + { + "epoch": 0.037053337524887354, + "grad_norm": 0.712087869644165, + "learning_rate": 1.9997757275108847e-05, + "loss": 1.0069, + "step": 221 + }, + { + "epoch": 0.03722099968563345, + "grad_norm": 0.6852487921714783, + "learning_rate": 1.999764079364959e-05, + "loss": 1.015, + "step": 222 + }, + { + "epoch": 0.037388661846379546, + "grad_norm": 0.7390941977500916, + "learning_rate": 1.9997521363762264e-05, + "loss": 1.0333, + "step": 223 + }, + { + "epoch": 0.03755632400712564, + "grad_norm": 0.5429960489273071, + "learning_rate": 1.999739898548209e-05, + "loss": 1.0331, + "step": 224 + }, + { + "epoch": 0.03772398616787174, + "grad_norm": 0.5059285759925842, + "learning_rate": 1.999727365884516e-05, + "loss": 1.0684, + "step": 225 + }, + { + "epoch": 0.037891648328617834, + "grad_norm": 0.546054482460022, + "learning_rate": 1.999714538388843e-05, + "loss": 1.0447, + "step": 226 + }, + { + "epoch": 0.03805931048936393, + "grad_norm": 0.420764297246933, + "learning_rate": 1.9997014160649736e-05, + "loss": 1.033, + "step": 227 + }, + { + "epoch": 0.038226972650110026, + "grad_norm": 0.4457850754261017, + "learning_rate": 1.9996879989167777e-05, + "loss": 1.0114, + "step": 228 + }, + { + "epoch": 0.03839463481085612, + "grad_norm": 0.5681815147399902, + "learning_rate": 1.999674286948212e-05, + "loss": 1.0674, + "step": 229 + }, + { + "epoch": 0.03856229697160222, + "grad_norm": 0.4577525854110718, + "learning_rate": 1.9996602801633197e-05, + "loss": 1.0182, + "step": 230 + }, + { + "epoch": 0.03872995913234832, + "grad_norm": 0.47984087467193604, + "learning_rate": 1.999645978566233e-05, + "loss": 1.0901, + "step": 231 + }, + { + "epoch": 0.03889762129309442, + "grad_norm": 0.5177239179611206, + "learning_rate": 1.9996313821611682e-05, + "loss": 1.0182, + "step": 232 + }, + { + "epoch": 0.03906528345384051, + "grad_norm": 0.5124174952507019, + "learning_rate": 1.9996164909524306e-05, + "loss": 0.9901, + "step": 233 + }, + { + "epoch": 0.03923294561458661, + "grad_norm": 0.5485550761222839, + "learning_rate": 1.9996013049444117e-05, + "loss": 0.9861, + "step": 234 + }, + { + "epoch": 0.039400607775332705, + "grad_norm": 1.155792474746704, + "learning_rate": 1.9995858241415904e-05, + "loss": 1.0383, + "step": 235 + }, + { + "epoch": 0.0395682699360788, + "grad_norm": 0.46704602241516113, + "learning_rate": 1.999570048548532e-05, + "loss": 1.002, + "step": 236 + }, + { + "epoch": 0.0397359320968249, + "grad_norm": 0.5665890574455261, + "learning_rate": 1.9995539781698882e-05, + "loss": 1.0785, + "step": 237 + }, + { + "epoch": 0.03990359425757099, + "grad_norm": 0.41134896874427795, + "learning_rate": 1.9995376130103992e-05, + "loss": 0.8948, + "step": 238 + }, + { + "epoch": 0.04007125641831709, + "grad_norm": 0.47926095128059387, + "learning_rate": 1.9995209530748913e-05, + "loss": 0.9718, + "step": 239 + }, + { + "epoch": 0.040238918579063185, + "grad_norm": 0.4371228814125061, + "learning_rate": 1.9995039983682774e-05, + "loss": 1.0359, + "step": 240 + }, + { + "epoch": 0.04040658073980929, + "grad_norm": 0.5196617245674133, + "learning_rate": 1.9994867488955573e-05, + "loss": 1.0706, + "step": 241 + }, + { + "epoch": 0.040574242900555384, + "grad_norm": 0.4669308364391327, + "learning_rate": 1.9994692046618188e-05, + "loss": 0.9539, + "step": 242 + }, + { + "epoch": 0.04074190506130148, + "grad_norm": 0.5073533654212952, + "learning_rate": 1.9994513656722356e-05, + "loss": 1.0063, + "step": 243 + }, + { + "epoch": 0.040909567222047576, + "grad_norm": 0.47835734486579895, + "learning_rate": 1.9994332319320688e-05, + "loss": 0.9411, + "step": 244 + }, + { + "epoch": 0.04107722938279367, + "grad_norm": 0.5108445286750793, + "learning_rate": 1.9994148034466655e-05, + "loss": 0.9498, + "step": 245 + }, + { + "epoch": 0.04124489154353977, + "grad_norm": 0.45715704560279846, + "learning_rate": 1.9993960802214617e-05, + "loss": 1.0476, + "step": 246 + }, + { + "epoch": 0.041412553704285864, + "grad_norm": 0.5002694129943848, + "learning_rate": 1.9993770622619784e-05, + "loss": 1.0674, + "step": 247 + }, + { + "epoch": 0.04158021586503196, + "grad_norm": 0.659797728061676, + "learning_rate": 1.9993577495738245e-05, + "loss": 1.0106, + "step": 248 + }, + { + "epoch": 0.041747878025778055, + "grad_norm": 0.69830322265625, + "learning_rate": 1.999338142162695e-05, + "loss": 1.0719, + "step": 249 + }, + { + "epoch": 0.04191554018652415, + "grad_norm": 0.6348698735237122, + "learning_rate": 1.9993182400343727e-05, + "loss": 1.0302, + "step": 250 + }, + { + "epoch": 0.04208320234727025, + "grad_norm": 0.5063202977180481, + "learning_rate": 1.9992980431947276e-05, + "loss": 1.115, + "step": 251 + }, + { + "epoch": 0.04225086450801635, + "grad_norm": 0.43055394291877747, + "learning_rate": 1.9992775516497148e-05, + "loss": 0.9884, + "step": 252 + }, + { + "epoch": 0.042418526668762446, + "grad_norm": 0.48402297496795654, + "learning_rate": 1.9992567654053785e-05, + "loss": 1.0252, + "step": 253 + }, + { + "epoch": 0.04258618882950854, + "grad_norm": 0.48005810379981995, + "learning_rate": 1.9992356844678485e-05, + "loss": 0.9396, + "step": 254 + }, + { + "epoch": 0.04275385099025464, + "grad_norm": 0.5253524780273438, + "learning_rate": 1.9992143088433417e-05, + "loss": 1.0266, + "step": 255 + }, + { + "epoch": 0.042921513151000734, + "grad_norm": 0.4629131853580475, + "learning_rate": 1.9991926385381623e-05, + "loss": 1.0202, + "step": 256 + }, + { + "epoch": 0.04308917531174683, + "grad_norm": 0.675323486328125, + "learning_rate": 1.9991706735587006e-05, + "loss": 1.007, + "step": 257 + }, + { + "epoch": 0.043256837472492926, + "grad_norm": 0.49460679292678833, + "learning_rate": 1.999148413911435e-05, + "loss": 0.983, + "step": 258 + }, + { + "epoch": 0.04342449963323902, + "grad_norm": 0.5188019871711731, + "learning_rate": 1.9991258596029297e-05, + "loss": 1.1205, + "step": 259 + }, + { + "epoch": 0.04359216179398512, + "grad_norm": 0.5086391568183899, + "learning_rate": 1.9991030106398367e-05, + "loss": 1.0457, + "step": 260 + }, + { + "epoch": 0.043759823954731214, + "grad_norm": 0.5028776526451111, + "learning_rate": 1.999079867028894e-05, + "loss": 0.9295, + "step": 261 + }, + { + "epoch": 0.04392748611547731, + "grad_norm": 0.6068524718284607, + "learning_rate": 1.999056428776927e-05, + "loss": 1.1401, + "step": 262 + }, + { + "epoch": 0.04409514827622341, + "grad_norm": 0.49584200978279114, + "learning_rate": 1.999032695890848e-05, + "loss": 0.974, + "step": 263 + }, + { + "epoch": 0.04426281043696951, + "grad_norm": 0.47098278999328613, + "learning_rate": 1.9990086683776564e-05, + "loss": 0.9915, + "step": 264 + }, + { + "epoch": 0.044430472597715605, + "grad_norm": 0.6156361699104309, + "learning_rate": 1.9989843462444377e-05, + "loss": 1.0524, + "step": 265 + }, + { + "epoch": 0.0445981347584617, + "grad_norm": 0.7538841962814331, + "learning_rate": 1.998959729498365e-05, + "loss": 1.0742, + "step": 266 + }, + { + "epoch": 0.0447657969192078, + "grad_norm": 0.6614930629730225, + "learning_rate": 1.9989348181466987e-05, + "loss": 0.9725, + "step": 267 + }, + { + "epoch": 0.04493345907995389, + "grad_norm": 0.7805516719818115, + "learning_rate": 1.9989096121967842e-05, + "loss": 1.0216, + "step": 268 + }, + { + "epoch": 0.04510112124069999, + "grad_norm": 0.6585594415664673, + "learning_rate": 1.9988841116560566e-05, + "loss": 0.9367, + "step": 269 + }, + { + "epoch": 0.045268783401446085, + "grad_norm": 0.5541887879371643, + "learning_rate": 1.9988583165320347e-05, + "loss": 1.0547, + "step": 270 + }, + { + "epoch": 0.04543644556219218, + "grad_norm": 0.7834230065345764, + "learning_rate": 1.998832226832327e-05, + "loss": 0.9246, + "step": 271 + }, + { + "epoch": 0.04560410772293828, + "grad_norm": 0.5578247308731079, + "learning_rate": 1.9988058425646268e-05, + "loss": 0.9464, + "step": 272 + }, + { + "epoch": 0.04577176988368437, + "grad_norm": 0.5382641553878784, + "learning_rate": 1.9987791637367157e-05, + "loss": 1.0635, + "step": 273 + }, + { + "epoch": 0.045939432044430475, + "grad_norm": 0.7065907716751099, + "learning_rate": 1.998752190356462e-05, + "loss": 0.9919, + "step": 274 + }, + { + "epoch": 0.04610709420517657, + "grad_norm": 0.5382146239280701, + "learning_rate": 1.99872492243182e-05, + "loss": 1.0498, + "step": 275 + }, + { + "epoch": 0.04627475636592267, + "grad_norm": 0.5285945534706116, + "learning_rate": 1.9986973599708305e-05, + "loss": 0.9463, + "step": 276 + }, + { + "epoch": 0.04644241852666876, + "grad_norm": 0.5281641483306885, + "learning_rate": 1.9986695029816237e-05, + "loss": 0.984, + "step": 277 + }, + { + "epoch": 0.04661008068741486, + "grad_norm": 0.5835294723510742, + "learning_rate": 1.9986413514724136e-05, + "loss": 1.0335, + "step": 278 + }, + { + "epoch": 0.046777742848160955, + "grad_norm": 0.4561820924282074, + "learning_rate": 1.9986129054515035e-05, + "loss": 1.0397, + "step": 279 + }, + { + "epoch": 0.04694540500890705, + "grad_norm": 0.5885711312294006, + "learning_rate": 1.9985841649272814e-05, + "loss": 1.0512, + "step": 280 + }, + { + "epoch": 0.04711306716965315, + "grad_norm": 0.441677451133728, + "learning_rate": 1.998555129908224e-05, + "loss": 1.0904, + "step": 281 + }, + { + "epoch": 0.04728072933039924, + "grad_norm": 0.4709782302379608, + "learning_rate": 1.9985258004028938e-05, + "loss": 1.099, + "step": 282 + }, + { + "epoch": 0.04744839149114534, + "grad_norm": 0.5236688256263733, + "learning_rate": 1.998496176419941e-05, + "loss": 0.976, + "step": 283 + }, + { + "epoch": 0.04761605365189144, + "grad_norm": 0.4500700831413269, + "learning_rate": 1.998466257968101e-05, + "loss": 0.9947, + "step": 284 + }, + { + "epoch": 0.04778371581263754, + "grad_norm": 0.44306880235671997, + "learning_rate": 1.998436045056198e-05, + "loss": 1.0322, + "step": 285 + }, + { + "epoch": 0.047951377973383634, + "grad_norm": 0.44918814301490784, + "learning_rate": 1.9984055376931414e-05, + "loss": 0.9553, + "step": 286 + }, + { + "epoch": 0.04811904013412973, + "grad_norm": 0.5018520951271057, + "learning_rate": 1.998374735887929e-05, + "loss": 1.0741, + "step": 287 + }, + { + "epoch": 0.048286702294875826, + "grad_norm": 0.4231759011745453, + "learning_rate": 1.9983436396496442e-05, + "loss": 1.0914, + "step": 288 + }, + { + "epoch": 0.04845436445562192, + "grad_norm": 0.5387656688690186, + "learning_rate": 1.9983122489874582e-05, + "loss": 0.9482, + "step": 289 + }, + { + "epoch": 0.04862202661636802, + "grad_norm": 0.460317999124527, + "learning_rate": 1.9982805639106276e-05, + "loss": 1.0508, + "step": 290 + }, + { + "epoch": 0.048789688777114114, + "grad_norm": 0.5162121653556824, + "learning_rate": 1.9982485844284972e-05, + "loss": 1.0266, + "step": 291 + }, + { + "epoch": 0.04895735093786021, + "grad_norm": 0.7197742462158203, + "learning_rate": 1.9982163105504983e-05, + "loss": 0.9918, + "step": 292 + }, + { + "epoch": 0.049125013098606306, + "grad_norm": 0.6081408262252808, + "learning_rate": 1.9981837422861487e-05, + "loss": 0.9567, + "step": 293 + }, + { + "epoch": 0.0492926752593524, + "grad_norm": 0.5756159424781799, + "learning_rate": 1.9981508796450532e-05, + "loss": 0.9618, + "step": 294 + }, + { + "epoch": 0.049460337420098505, + "grad_norm": 0.43800628185272217, + "learning_rate": 1.998117722636903e-05, + "loss": 1.0086, + "step": 295 + }, + { + "epoch": 0.0496279995808446, + "grad_norm": 0.6097040772438049, + "learning_rate": 1.9980842712714774e-05, + "loss": 0.9752, + "step": 296 + }, + { + "epoch": 0.0497956617415907, + "grad_norm": 0.5512593984603882, + "learning_rate": 1.9980505255586408e-05, + "loss": 1.0035, + "step": 297 + }, + { + "epoch": 0.04996332390233679, + "grad_norm": 0.6284067034721375, + "learning_rate": 1.9980164855083454e-05, + "loss": 0.9961, + "step": 298 + }, + { + "epoch": 0.05013098606308289, + "grad_norm": 0.30609074234962463, + "learning_rate": 1.9979821511306308e-05, + "loss": 0.9692, + "step": 299 + }, + { + "epoch": 0.050298648223828984, + "grad_norm": 0.58941650390625, + "learning_rate": 1.9979475224356214e-05, + "loss": 1.0258, + "step": 300 + }, + { + "epoch": 0.05046631038457508, + "grad_norm": 0.4312822222709656, + "learning_rate": 1.9979125994335306e-05, + "loss": 1.1427, + "step": 301 + }, + { + "epoch": 0.050633972545321176, + "grad_norm": 0.4537133276462555, + "learning_rate": 1.997877382134657e-05, + "loss": 0.9259, + "step": 302 + }, + { + "epoch": 0.05080163470606727, + "grad_norm": 0.4248863458633423, + "learning_rate": 1.9978418705493868e-05, + "loss": 1.036, + "step": 303 + }, + { + "epoch": 0.05096929686681337, + "grad_norm": 0.5460931658744812, + "learning_rate": 1.9978060646881933e-05, + "loss": 1.002, + "step": 304 + }, + { + "epoch": 0.051136959027559464, + "grad_norm": 0.5437471866607666, + "learning_rate": 1.9977699645616353e-05, + "loss": 1.0424, + "step": 305 + }, + { + "epoch": 0.05130462118830557, + "grad_norm": 0.6186936497688293, + "learning_rate": 1.9977335701803597e-05, + "loss": 0.9514, + "step": 306 + }, + { + "epoch": 0.05147228334905166, + "grad_norm": 0.6375221014022827, + "learning_rate": 1.9976968815550994e-05, + "loss": 1.0237, + "step": 307 + }, + { + "epoch": 0.05163994550979776, + "grad_norm": 0.5780304670333862, + "learning_rate": 1.9976598986966746e-05, + "loss": 1.0126, + "step": 308 + }, + { + "epoch": 0.051807607670543855, + "grad_norm": 0.47898033261299133, + "learning_rate": 1.9976226216159916e-05, + "loss": 1.0855, + "step": 309 + }, + { + "epoch": 0.05197526983128995, + "grad_norm": 0.6292722225189209, + "learning_rate": 1.9975850503240444e-05, + "loss": 0.9622, + "step": 310 + }, + { + "epoch": 0.05214293199203605, + "grad_norm": 0.6259632110595703, + "learning_rate": 1.997547184831913e-05, + "loss": 0.952, + "step": 311 + }, + { + "epoch": 0.05231059415278214, + "grad_norm": 0.4796934425830841, + "learning_rate": 1.9975090251507637e-05, + "loss": 0.9743, + "step": 312 + }, + { + "epoch": 0.05247825631352824, + "grad_norm": 0.4715990424156189, + "learning_rate": 1.9974705712918515e-05, + "loss": 0.9663, + "step": 313 + }, + { + "epoch": 0.052645918474274335, + "grad_norm": 0.45795300602912903, + "learning_rate": 1.9974318232665164e-05, + "loss": 1.0288, + "step": 314 + }, + { + "epoch": 0.05281358063502043, + "grad_norm": 0.4572800099849701, + "learning_rate": 1.9973927810861855e-05, + "loss": 0.9768, + "step": 315 + }, + { + "epoch": 0.052981242795766534, + "grad_norm": 0.6992179751396179, + "learning_rate": 1.9973534447623727e-05, + "loss": 1.0511, + "step": 316 + }, + { + "epoch": 0.05314890495651263, + "grad_norm": 0.6290198564529419, + "learning_rate": 1.9973138143066793e-05, + "loss": 1.0728, + "step": 317 + }, + { + "epoch": 0.053316567117258726, + "grad_norm": 0.4092872142791748, + "learning_rate": 1.997273889730792e-05, + "loss": 0.9257, + "step": 318 + }, + { + "epoch": 0.05348422927800482, + "grad_norm": 0.5989061594009399, + "learning_rate": 1.997233671046486e-05, + "loss": 0.9965, + "step": 319 + }, + { + "epoch": 0.05365189143875092, + "grad_norm": 0.70281583070755, + "learning_rate": 1.997193158265622e-05, + "loss": 1.0463, + "step": 320 + }, + { + "epoch": 0.053819553599497014, + "grad_norm": 0.6267127990722656, + "learning_rate": 1.9971523514001473e-05, + "loss": 1.0258, + "step": 321 + }, + { + "epoch": 0.05398721576024311, + "grad_norm": 0.6525517106056213, + "learning_rate": 1.9971112504620968e-05, + "loss": 1.0238, + "step": 322 + }, + { + "epoch": 0.054154877920989206, + "grad_norm": 0.5730092525482178, + "learning_rate": 1.9970698554635913e-05, + "loss": 0.9462, + "step": 323 + }, + { + "epoch": 0.0543225400817353, + "grad_norm": 0.5112807750701904, + "learning_rate": 1.997028166416839e-05, + "loss": 1.0223, + "step": 324 + }, + { + "epoch": 0.0544902022424814, + "grad_norm": 0.6277716755867004, + "learning_rate": 1.996986183334134e-05, + "loss": 0.9783, + "step": 325 + }, + { + "epoch": 0.054657864403227494, + "grad_norm": 0.4657522439956665, + "learning_rate": 1.9969439062278588e-05, + "loss": 1.0838, + "step": 326 + }, + { + "epoch": 0.054825526563973596, + "grad_norm": 0.5641884803771973, + "learning_rate": 1.9969013351104803e-05, + "loss": 0.9719, + "step": 327 + }, + { + "epoch": 0.05499318872471969, + "grad_norm": 0.6853857636451721, + "learning_rate": 1.9968584699945534e-05, + "loss": 1.0499, + "step": 328 + }, + { + "epoch": 0.05516085088546579, + "grad_norm": 0.594561755657196, + "learning_rate": 1.9968153108927198e-05, + "loss": 0.9929, + "step": 329 + }, + { + "epoch": 0.055328513046211884, + "grad_norm": 0.43680334091186523, + "learning_rate": 1.996771857817708e-05, + "loss": 0.9872, + "step": 330 + }, + { + "epoch": 0.05549617520695798, + "grad_norm": 0.37571197748184204, + "learning_rate": 1.9967281107823326e-05, + "loss": 0.951, + "step": 331 + }, + { + "epoch": 0.055663837367704076, + "grad_norm": 0.4594593942165375, + "learning_rate": 1.9966840697994945e-05, + "loss": 0.9354, + "step": 332 + }, + { + "epoch": 0.05583149952845017, + "grad_norm": 0.48202845454216003, + "learning_rate": 1.996639734882183e-05, + "loss": 0.9478, + "step": 333 + }, + { + "epoch": 0.05599916168919627, + "grad_norm": 0.4156875014305115, + "learning_rate": 1.996595106043472e-05, + "loss": 0.962, + "step": 334 + }, + { + "epoch": 0.056166823849942364, + "grad_norm": 0.5083615779876709, + "learning_rate": 1.9965501832965238e-05, + "loss": 1.0278, + "step": 335 + }, + { + "epoch": 0.05633448601068846, + "grad_norm": 0.5085647702217102, + "learning_rate": 1.9965049666545863e-05, + "loss": 0.9764, + "step": 336 + }, + { + "epoch": 0.056502148171434556, + "grad_norm": 0.5676213502883911, + "learning_rate": 1.9964594561309946e-05, + "loss": 1.0438, + "step": 337 + }, + { + "epoch": 0.05666981033218066, + "grad_norm": 0.5562122464179993, + "learning_rate": 1.9964136517391708e-05, + "loss": 0.9988, + "step": 338 + }, + { + "epoch": 0.056837472492926755, + "grad_norm": 0.5095664858818054, + "learning_rate": 1.9963675534926224e-05, + "loss": 1.0729, + "step": 339 + }, + { + "epoch": 0.05700513465367285, + "grad_norm": 0.5423064231872559, + "learning_rate": 1.9963211614049443e-05, + "loss": 1.0693, + "step": 340 + }, + { + "epoch": 0.05717279681441895, + "grad_norm": 0.666354775428772, + "learning_rate": 1.996274475489819e-05, + "loss": 0.9466, + "step": 341 + }, + { + "epoch": 0.05734045897516504, + "grad_norm": 0.5196070075035095, + "learning_rate": 1.9962274957610136e-05, + "loss": 0.9676, + "step": 342 + }, + { + "epoch": 0.05750812113591114, + "grad_norm": 0.6126438975334167, + "learning_rate": 1.996180222232384e-05, + "loss": 0.9033, + "step": 343 + }, + { + "epoch": 0.057675783296657235, + "grad_norm": 0.5262427926063538, + "learning_rate": 1.9961326549178713e-05, + "loss": 0.9979, + "step": 344 + }, + { + "epoch": 0.05784344545740333, + "grad_norm": 0.5726452469825745, + "learning_rate": 1.9960847938315037e-05, + "loss": 0.9771, + "step": 345 + }, + { + "epoch": 0.05801110761814943, + "grad_norm": 0.5337097644805908, + "learning_rate": 1.996036638987396e-05, + "loss": 0.9814, + "step": 346 + }, + { + "epoch": 0.05817876977889552, + "grad_norm": 0.6877371072769165, + "learning_rate": 1.99598819039975e-05, + "loss": 1.1398, + "step": 347 + }, + { + "epoch": 0.05834643193964162, + "grad_norm": 0.6705992221832275, + "learning_rate": 1.995939448082853e-05, + "loss": 1.0732, + "step": 348 + }, + { + "epoch": 0.05851409410038772, + "grad_norm": 0.7965999841690063, + "learning_rate": 1.9958904120510807e-05, + "loss": 1.0088, + "step": 349 + }, + { + "epoch": 0.05868175626113382, + "grad_norm": 0.6121879816055298, + "learning_rate": 1.995841082318894e-05, + "loss": 1.0457, + "step": 350 + }, + { + "epoch": 0.058849418421879914, + "grad_norm": 0.892564594745636, + "learning_rate": 1.9957914589008405e-05, + "loss": 0.9857, + "step": 351 + }, + { + "epoch": 0.05901708058262601, + "grad_norm": 0.4690881669521332, + "learning_rate": 1.9957415418115552e-05, + "loss": 0.9926, + "step": 352 + }, + { + "epoch": 0.059184742743372105, + "grad_norm": 0.8930359482765198, + "learning_rate": 1.995691331065759e-05, + "loss": 1.0861, + "step": 353 + }, + { + "epoch": 0.0593524049041182, + "grad_norm": 0.6299958825111389, + "learning_rate": 1.99564082667826e-05, + "loss": 1.0111, + "step": 354 + }, + { + "epoch": 0.0595200670648643, + "grad_norm": 0.6326963901519775, + "learning_rate": 1.9955900286639524e-05, + "loss": 1.0335, + "step": 355 + }, + { + "epoch": 0.05968772922561039, + "grad_norm": 0.5321698188781738, + "learning_rate": 1.9955389370378172e-05, + "loss": 0.9373, + "step": 356 + }, + { + "epoch": 0.05985539138635649, + "grad_norm": 0.42701229453086853, + "learning_rate": 1.9954875518149216e-05, + "loss": 1.0024, + "step": 357 + }, + { + "epoch": 0.060023053547102585, + "grad_norm": 0.7688174247741699, + "learning_rate": 1.99543587301042e-05, + "loss": 1.0012, + "step": 358 + }, + { + "epoch": 0.06019071570784869, + "grad_norm": 0.47423213720321655, + "learning_rate": 1.9953839006395534e-05, + "loss": 0.9813, + "step": 359 + }, + { + "epoch": 0.060358377868594784, + "grad_norm": 0.48588648438453674, + "learning_rate": 1.995331634717649e-05, + "loss": 1.0102, + "step": 360 + }, + { + "epoch": 0.06052604002934088, + "grad_norm": 0.586348295211792, + "learning_rate": 1.99527907526012e-05, + "loss": 1.0298, + "step": 361 + }, + { + "epoch": 0.060693702190086976, + "grad_norm": 0.6227160692214966, + "learning_rate": 1.995226222282468e-05, + "loss": 1.0071, + "step": 362 + }, + { + "epoch": 0.06086136435083307, + "grad_norm": 0.6064775586128235, + "learning_rate": 1.9951730758002787e-05, + "loss": 1.014, + "step": 363 + }, + { + "epoch": 0.06102902651157917, + "grad_norm": 0.6879925727844238, + "learning_rate": 1.9951196358292266e-05, + "loss": 1.0248, + "step": 364 + }, + { + "epoch": 0.061196688672325264, + "grad_norm": 0.4487518072128296, + "learning_rate": 1.9950659023850716e-05, + "loss": 1.0143, + "step": 365 + }, + { + "epoch": 0.06136435083307136, + "grad_norm": 0.4098510444164276, + "learning_rate": 1.99501187548366e-05, + "loss": 0.9518, + "step": 366 + }, + { + "epoch": 0.061532012993817456, + "grad_norm": 0.49687305092811584, + "learning_rate": 1.9949575551409258e-05, + "loss": 0.9737, + "step": 367 + }, + { + "epoch": 0.06169967515456355, + "grad_norm": 0.43834513425827026, + "learning_rate": 1.9949029413728877e-05, + "loss": 0.9448, + "step": 368 + }, + { + "epoch": 0.06186733731530965, + "grad_norm": 0.4889039993286133, + "learning_rate": 1.9948480341956524e-05, + "loss": 1.0808, + "step": 369 + }, + { + "epoch": 0.06203499947605575, + "grad_norm": 0.4841388165950775, + "learning_rate": 1.9947928336254132e-05, + "loss": 1.0362, + "step": 370 + }, + { + "epoch": 0.06220266163680185, + "grad_norm": 0.6277660131454468, + "learning_rate": 1.9947373396784487e-05, + "loss": 1.0221, + "step": 371 + }, + { + "epoch": 0.06237032379754794, + "grad_norm": 0.5336534380912781, + "learning_rate": 1.9946815523711253e-05, + "loss": 1.0528, + "step": 372 + }, + { + "epoch": 0.06253798595829403, + "grad_norm": 0.999744713306427, + "learning_rate": 1.9946254717198954e-05, + "loss": 1.1037, + "step": 373 + }, + { + "epoch": 0.06270564811904013, + "grad_norm": 0.6568665504455566, + "learning_rate": 1.9945690977412972e-05, + "loss": 1.0035, + "step": 374 + }, + { + "epoch": 0.06287331027978624, + "grad_norm": 0.4865848422050476, + "learning_rate": 1.9945124304519568e-05, + "loss": 1.0624, + "step": 375 + }, + { + "epoch": 0.06304097244053233, + "grad_norm": 0.4973253607749939, + "learning_rate": 1.9944554698685853e-05, + "loss": 0.924, + "step": 376 + }, + { + "epoch": 0.06320863460127843, + "grad_norm": 0.7405445575714111, + "learning_rate": 1.9943982160079823e-05, + "loss": 1.013, + "step": 377 + }, + { + "epoch": 0.06337629676202453, + "grad_norm": 0.5829219222068787, + "learning_rate": 1.9943406688870316e-05, + "loss": 0.9768, + "step": 378 + }, + { + "epoch": 0.06354395892277062, + "grad_norm": 0.4399314820766449, + "learning_rate": 1.9942828285227047e-05, + "loss": 1.0749, + "step": 379 + }, + { + "epoch": 0.06371162108351672, + "grad_norm": 0.47130849957466125, + "learning_rate": 1.9942246949320598e-05, + "loss": 0.917, + "step": 380 + }, + { + "epoch": 0.06387928324426281, + "grad_norm": 0.46881669759750366, + "learning_rate": 1.9941662681322412e-05, + "loss": 0.9698, + "step": 381 + }, + { + "epoch": 0.06404694540500891, + "grad_norm": 0.4304395318031311, + "learning_rate": 1.9941075481404795e-05, + "loss": 1.0417, + "step": 382 + }, + { + "epoch": 0.064214607565755, + "grad_norm": 0.6960583925247192, + "learning_rate": 1.994048534974092e-05, + "loss": 1.0315, + "step": 383 + }, + { + "epoch": 0.0643822697265011, + "grad_norm": 0.6453326940536499, + "learning_rate": 1.9939892286504822e-05, + "loss": 0.9886, + "step": 384 + }, + { + "epoch": 0.0645499318872472, + "grad_norm": 0.4545916020870209, + "learning_rate": 1.993929629187141e-05, + "loss": 0.9806, + "step": 385 + }, + { + "epoch": 0.0647175940479933, + "grad_norm": 0.38227584958076477, + "learning_rate": 1.993869736601644e-05, + "loss": 0.9928, + "step": 386 + }, + { + "epoch": 0.06488525620873939, + "grad_norm": 0.48465222120285034, + "learning_rate": 1.9938095509116547e-05, + "loss": 1.1353, + "step": 387 + }, + { + "epoch": 0.06505291836948549, + "grad_norm": 0.4780482351779938, + "learning_rate": 1.9937490721349232e-05, + "loss": 0.9548, + "step": 388 + }, + { + "epoch": 0.06522058053023158, + "grad_norm": 0.5209558010101318, + "learning_rate": 1.9936883002892843e-05, + "loss": 1.0472, + "step": 389 + }, + { + "epoch": 0.06538824269097768, + "grad_norm": 0.48167288303375244, + "learning_rate": 1.9936272353926616e-05, + "loss": 1.0512, + "step": 390 + }, + { + "epoch": 0.06555590485172377, + "grad_norm": 0.4132109582424164, + "learning_rate": 1.9935658774630632e-05, + "loss": 0.9558, + "step": 391 + }, + { + "epoch": 0.06572356701246987, + "grad_norm": 0.562962532043457, + "learning_rate": 1.993504226518584e-05, + "loss": 0.9254, + "step": 392 + }, + { + "epoch": 0.06589122917321596, + "grad_norm": 0.4468546211719513, + "learning_rate": 1.9934422825774064e-05, + "loss": 0.9164, + "step": 393 + }, + { + "epoch": 0.06605889133396206, + "grad_norm": 0.4178194999694824, + "learning_rate": 1.9933800456577978e-05, + "loss": 0.9262, + "step": 394 + }, + { + "epoch": 0.06622655349470816, + "grad_norm": 0.7586342096328735, + "learning_rate": 1.993317515778113e-05, + "loss": 0.9454, + "step": 395 + }, + { + "epoch": 0.06639421565545425, + "grad_norm": 0.5593826770782471, + "learning_rate": 1.9932546929567924e-05, + "loss": 1.0024, + "step": 396 + }, + { + "epoch": 0.06656187781620036, + "grad_norm": 0.4449445903301239, + "learning_rate": 1.993191577212364e-05, + "loss": 0.9994, + "step": 397 + }, + { + "epoch": 0.06672953997694646, + "grad_norm": 0.5252183675765991, + "learning_rate": 1.993128168563441e-05, + "loss": 1.0497, + "step": 398 + }, + { + "epoch": 0.06689720213769255, + "grad_norm": 0.45754343271255493, + "learning_rate": 1.9930644670287233e-05, + "loss": 1.0029, + "step": 399 + }, + { + "epoch": 0.06706486429843865, + "grad_norm": 0.4065295159816742, + "learning_rate": 1.993000472626997e-05, + "loss": 0.8735, + "step": 400 + }, + { + "epoch": 0.06723252645918475, + "grad_norm": 0.4539814591407776, + "learning_rate": 1.9929361853771355e-05, + "loss": 1.017, + "step": 401 + }, + { + "epoch": 0.06740018861993084, + "grad_norm": 0.46643343567848206, + "learning_rate": 1.9928716052980977e-05, + "loss": 1.0082, + "step": 402 + }, + { + "epoch": 0.06756785078067694, + "grad_norm": 0.48987117409706116, + "learning_rate": 1.9928067324089286e-05, + "loss": 0.9141, + "step": 403 + }, + { + "epoch": 0.06773551294142303, + "grad_norm": 0.3898268938064575, + "learning_rate": 1.9927415667287605e-05, + "loss": 1.0585, + "step": 404 + }, + { + "epoch": 0.06790317510216913, + "grad_norm": 0.5273199081420898, + "learning_rate": 1.9926761082768116e-05, + "loss": 1.0478, + "step": 405 + }, + { + "epoch": 0.06807083726291523, + "grad_norm": 0.5008173584938049, + "learning_rate": 1.992610357072386e-05, + "loss": 1.0472, + "step": 406 + }, + { + "epoch": 0.06823849942366132, + "grad_norm": 0.42762380838394165, + "learning_rate": 1.992544313134875e-05, + "loss": 0.9769, + "step": 407 + }, + { + "epoch": 0.06840616158440742, + "grad_norm": 0.43059566617012024, + "learning_rate": 1.9924779764837553e-05, + "loss": 1.0213, + "step": 408 + }, + { + "epoch": 0.06857382374515351, + "grad_norm": 0.485988050699234, + "learning_rate": 1.992411347138591e-05, + "loss": 1.0403, + "step": 409 + }, + { + "epoch": 0.06874148590589961, + "grad_norm": 0.5233361721038818, + "learning_rate": 1.9923444251190314e-05, + "loss": 0.9312, + "step": 410 + }, + { + "epoch": 0.0689091480666457, + "grad_norm": 0.7353261709213257, + "learning_rate": 1.992277210444813e-05, + "loss": 1.0113, + "step": 411 + }, + { + "epoch": 0.0690768102273918, + "grad_norm": 0.4786561131477356, + "learning_rate": 1.9922097031357575e-05, + "loss": 0.8914, + "step": 412 + }, + { + "epoch": 0.0692444723881379, + "grad_norm": 0.6206673383712769, + "learning_rate": 1.9921419032117752e-05, + "loss": 1.0199, + "step": 413 + }, + { + "epoch": 0.069412134548884, + "grad_norm": 0.5633043646812439, + "learning_rate": 1.9920738106928593e-05, + "loss": 1.0299, + "step": 414 + }, + { + "epoch": 0.06957979670963009, + "grad_norm": 0.6748209595680237, + "learning_rate": 1.992005425599092e-05, + "loss": 0.9454, + "step": 415 + }, + { + "epoch": 0.06974745887037619, + "grad_norm": 0.6763722896575928, + "learning_rate": 1.9919367479506413e-05, + "loss": 1.1172, + "step": 416 + }, + { + "epoch": 0.06991512103112228, + "grad_norm": 0.6160325407981873, + "learning_rate": 1.9918677777677605e-05, + "loss": 1.0262, + "step": 417 + }, + { + "epoch": 0.07008278319186839, + "grad_norm": 0.6006767749786377, + "learning_rate": 1.99179851507079e-05, + "loss": 1.0134, + "step": 418 + }, + { + "epoch": 0.07025044535261449, + "grad_norm": 0.49446508288383484, + "learning_rate": 1.9917289598801556e-05, + "loss": 1.0113, + "step": 419 + }, + { + "epoch": 0.07041810751336058, + "grad_norm": 0.44796285033226013, + "learning_rate": 1.991659112216371e-05, + "loss": 0.9279, + "step": 420 + }, + { + "epoch": 0.07058576967410668, + "grad_norm": 0.4072938561439514, + "learning_rate": 1.9915889721000347e-05, + "loss": 0.9871, + "step": 421 + }, + { + "epoch": 0.07075343183485278, + "grad_norm": 0.5549907088279724, + "learning_rate": 1.991518539551832e-05, + "loss": 0.9974, + "step": 422 + }, + { + "epoch": 0.07092109399559887, + "grad_norm": 0.4595836102962494, + "learning_rate": 1.991447814592534e-05, + "loss": 1.0345, + "step": 423 + }, + { + "epoch": 0.07108875615634497, + "grad_norm": 0.422370582818985, + "learning_rate": 1.9913767972429984e-05, + "loss": 0.9588, + "step": 424 + }, + { + "epoch": 0.07125641831709106, + "grad_norm": 0.4242878258228302, + "learning_rate": 1.9913054875241694e-05, + "loss": 0.9386, + "step": 425 + }, + { + "epoch": 0.07142408047783716, + "grad_norm": 0.5106265544891357, + "learning_rate": 1.991233885457077e-05, + "loss": 0.9666, + "step": 426 + }, + { + "epoch": 0.07159174263858326, + "grad_norm": 0.4680672585964203, + "learning_rate": 1.9911619910628374e-05, + "loss": 0.9116, + "step": 427 + }, + { + "epoch": 0.07175940479932935, + "grad_norm": 0.5930494666099548, + "learning_rate": 1.9910898043626534e-05, + "loss": 1.035, + "step": 428 + }, + { + "epoch": 0.07192706696007545, + "grad_norm": 0.7918307781219482, + "learning_rate": 1.9910173253778136e-05, + "loss": 1.0634, + "step": 429 + }, + { + "epoch": 0.07209472912082154, + "grad_norm": 0.4641052186489105, + "learning_rate": 1.990944554129693e-05, + "loss": 0.834, + "step": 430 + }, + { + "epoch": 0.07226239128156764, + "grad_norm": 0.46474114060401917, + "learning_rate": 1.9908714906397525e-05, + "loss": 0.8806, + "step": 431 + }, + { + "epoch": 0.07243005344231374, + "grad_norm": 0.4513774812221527, + "learning_rate": 1.99079813492954e-05, + "loss": 0.9827, + "step": 432 + }, + { + "epoch": 0.07259771560305983, + "grad_norm": 0.45238038897514343, + "learning_rate": 1.990724487020688e-05, + "loss": 0.9823, + "step": 433 + }, + { + "epoch": 0.07276537776380593, + "grad_norm": 0.44659870862960815, + "learning_rate": 1.9906505469349176e-05, + "loss": 0.8166, + "step": 434 + }, + { + "epoch": 0.07293303992455202, + "grad_norm": 0.4721660912036896, + "learning_rate": 1.9905763146940332e-05, + "loss": 0.947, + "step": 435 + }, + { + "epoch": 0.07310070208529812, + "grad_norm": 0.38887742161750793, + "learning_rate": 1.990501790319928e-05, + "loss": 0.935, + "step": 436 + }, + { + "epoch": 0.07326836424604422, + "grad_norm": 0.472125381231308, + "learning_rate": 1.9904269738345793e-05, + "loss": 1.0113, + "step": 437 + }, + { + "epoch": 0.07343602640679031, + "grad_norm": 0.4801134467124939, + "learning_rate": 1.9903518652600518e-05, + "loss": 0.9506, + "step": 438 + }, + { + "epoch": 0.07360368856753641, + "grad_norm": 0.47314056754112244, + "learning_rate": 1.990276464618496e-05, + "loss": 0.916, + "step": 439 + }, + { + "epoch": 0.07377135072828252, + "grad_norm": 0.5533985495567322, + "learning_rate": 1.9902007719321483e-05, + "loss": 1.0233, + "step": 440 + }, + { + "epoch": 0.07393901288902861, + "grad_norm": 0.47966158390045166, + "learning_rate": 1.9901247872233316e-05, + "loss": 0.9955, + "step": 441 + }, + { + "epoch": 0.07410667504977471, + "grad_norm": 0.5108001232147217, + "learning_rate": 1.9900485105144544e-05, + "loss": 0.9569, + "step": 442 + }, + { + "epoch": 0.0742743372105208, + "grad_norm": 0.5768658518791199, + "learning_rate": 1.989971941828012e-05, + "loss": 0.9923, + "step": 443 + }, + { + "epoch": 0.0744419993712669, + "grad_norm": 0.5834493041038513, + "learning_rate": 1.9898950811865854e-05, + "loss": 0.9778, + "step": 444 + }, + { + "epoch": 0.074609661532013, + "grad_norm": 0.6800981760025024, + "learning_rate": 1.9898179286128415e-05, + "loss": 1.0887, + "step": 445 + }, + { + "epoch": 0.07477732369275909, + "grad_norm": 0.4957737624645233, + "learning_rate": 1.9897404841295337e-05, + "loss": 0.959, + "step": 446 + }, + { + "epoch": 0.07494498585350519, + "grad_norm": 0.4654386341571808, + "learning_rate": 1.9896627477595017e-05, + "loss": 1.0396, + "step": 447 + }, + { + "epoch": 0.07511264801425128, + "grad_norm": 0.4682818353176117, + "learning_rate": 1.9895847195256702e-05, + "loss": 0.9909, + "step": 448 + }, + { + "epoch": 0.07528031017499738, + "grad_norm": 0.4938904643058777, + "learning_rate": 1.9895063994510512e-05, + "loss": 1.0329, + "step": 449 + }, + { + "epoch": 0.07544797233574348, + "grad_norm": 0.5164246559143066, + "learning_rate": 1.989427787558742e-05, + "loss": 1.005, + "step": 450 + }, + { + "epoch": 0.07561563449648957, + "grad_norm": 0.7106466889381409, + "learning_rate": 1.9893488838719267e-05, + "loss": 1.0256, + "step": 451 + }, + { + "epoch": 0.07578329665723567, + "grad_norm": 0.6531050205230713, + "learning_rate": 1.9892696884138743e-05, + "loss": 1.0237, + "step": 452 + }, + { + "epoch": 0.07595095881798176, + "grad_norm": 0.7170326709747314, + "learning_rate": 1.989190201207941e-05, + "loss": 0.9953, + "step": 453 + }, + { + "epoch": 0.07611862097872786, + "grad_norm": 0.5520920753479004, + "learning_rate": 1.9891104222775683e-05, + "loss": 1.0664, + "step": 454 + }, + { + "epoch": 0.07628628313947396, + "grad_norm": 0.5298479795455933, + "learning_rate": 1.9890303516462842e-05, + "loss": 0.867, + "step": 455 + }, + { + "epoch": 0.07645394530022005, + "grad_norm": 0.7908779978752136, + "learning_rate": 1.9889499893377024e-05, + "loss": 0.9758, + "step": 456 + }, + { + "epoch": 0.07662160746096615, + "grad_norm": 0.649056613445282, + "learning_rate": 1.988869335375523e-05, + "loss": 1.0339, + "step": 457 + }, + { + "epoch": 0.07678926962171224, + "grad_norm": 0.6440086960792542, + "learning_rate": 1.988788389783531e-05, + "loss": 0.8257, + "step": 458 + }, + { + "epoch": 0.07695693178245834, + "grad_norm": 0.498965322971344, + "learning_rate": 1.9887071525855993e-05, + "loss": 1.0239, + "step": 459 + }, + { + "epoch": 0.07712459394320444, + "grad_norm": 0.6430321931838989, + "learning_rate": 1.9886256238056854e-05, + "loss": 0.9629, + "step": 460 + }, + { + "epoch": 0.07729225610395055, + "grad_norm": 0.5625649094581604, + "learning_rate": 1.9885438034678327e-05, + "loss": 0.915, + "step": 461 + }, + { + "epoch": 0.07745991826469664, + "grad_norm": 0.58819580078125, + "learning_rate": 1.9884616915961717e-05, + "loss": 1.0797, + "step": 462 + }, + { + "epoch": 0.07762758042544274, + "grad_norm": 0.6026577949523926, + "learning_rate": 1.9883792882149177e-05, + "loss": 0.8432, + "step": 463 + }, + { + "epoch": 0.07779524258618883, + "grad_norm": 0.5222564339637756, + "learning_rate": 1.988296593348373e-05, + "loss": 0.9467, + "step": 464 + }, + { + "epoch": 0.07796290474693493, + "grad_norm": 0.616072952747345, + "learning_rate": 1.988213607020925e-05, + "loss": 1.0728, + "step": 465 + }, + { + "epoch": 0.07813056690768103, + "grad_norm": 0.4086941182613373, + "learning_rate": 1.988130329257047e-05, + "loss": 1.0023, + "step": 466 + }, + { + "epoch": 0.07829822906842712, + "grad_norm": 0.60639888048172, + "learning_rate": 1.9880467600812992e-05, + "loss": 0.9698, + "step": 467 + }, + { + "epoch": 0.07846589122917322, + "grad_norm": 0.6214800477027893, + "learning_rate": 1.9879628995183274e-05, + "loss": 1.0132, + "step": 468 + }, + { + "epoch": 0.07863355338991931, + "grad_norm": 0.4865536689758301, + "learning_rate": 1.9878787475928624e-05, + "loss": 0.8728, + "step": 469 + }, + { + "epoch": 0.07880121555066541, + "grad_norm": 0.6627867221832275, + "learning_rate": 1.9877943043297223e-05, + "loss": 1.0299, + "step": 470 + }, + { + "epoch": 0.0789688777114115, + "grad_norm": 0.5963670611381531, + "learning_rate": 1.98770956975381e-05, + "loss": 1.0561, + "step": 471 + }, + { + "epoch": 0.0791365398721576, + "grad_norm": 0.48536255955696106, + "learning_rate": 1.987624543890115e-05, + "loss": 1.0183, + "step": 472 + }, + { + "epoch": 0.0793042020329037, + "grad_norm": 0.4555172622203827, + "learning_rate": 1.9875392267637123e-05, + "loss": 0.9699, + "step": 473 + }, + { + "epoch": 0.0794718641936498, + "grad_norm": 0.5513790845870972, + "learning_rate": 1.9874536183997636e-05, + "loss": 1.035, + "step": 474 + }, + { + "epoch": 0.07963952635439589, + "grad_norm": 0.5619701743125916, + "learning_rate": 1.987367718823515e-05, + "loss": 1.0642, + "step": 475 + }, + { + "epoch": 0.07980718851514199, + "grad_norm": 0.43956467509269714, + "learning_rate": 1.9872815280602996e-05, + "loss": 1.0078, + "step": 476 + }, + { + "epoch": 0.07997485067588808, + "grad_norm": 0.40225303173065186, + "learning_rate": 1.9871950461355365e-05, + "loss": 0.9142, + "step": 477 + }, + { + "epoch": 0.08014251283663418, + "grad_norm": 0.49438419938087463, + "learning_rate": 1.9871082730747298e-05, + "loss": 0.9941, + "step": 478 + }, + { + "epoch": 0.08031017499738027, + "grad_norm": 0.6986750364303589, + "learning_rate": 1.9870212089034707e-05, + "loss": 0.9752, + "step": 479 + }, + { + "epoch": 0.08047783715812637, + "grad_norm": 0.5081155300140381, + "learning_rate": 1.9869338536474345e-05, + "loss": 0.9865, + "step": 480 + }, + { + "epoch": 0.08064549931887247, + "grad_norm": 0.5153716802597046, + "learning_rate": 1.986846207332384e-05, + "loss": 1.1058, + "step": 481 + }, + { + "epoch": 0.08081316147961858, + "grad_norm": 0.5596706867218018, + "learning_rate": 1.986758269984167e-05, + "loss": 0.9267, + "step": 482 + }, + { + "epoch": 0.08098082364036467, + "grad_norm": 0.486183762550354, + "learning_rate": 1.9866700416287174e-05, + "loss": 1.0091, + "step": 483 + }, + { + "epoch": 0.08114848580111077, + "grad_norm": 0.5393663644790649, + "learning_rate": 1.986581522292055e-05, + "loss": 1.0414, + "step": 484 + }, + { + "epoch": 0.08131614796185686, + "grad_norm": 0.46591389179229736, + "learning_rate": 1.9864927120002845e-05, + "loss": 0.882, + "step": 485 + }, + { + "epoch": 0.08148381012260296, + "grad_norm": 0.5534307360649109, + "learning_rate": 1.9864036107795978e-05, + "loss": 0.8958, + "step": 486 + }, + { + "epoch": 0.08165147228334906, + "grad_norm": 0.4817509353160858, + "learning_rate": 1.986314218656272e-05, + "loss": 0.9899, + "step": 487 + }, + { + "epoch": 0.08181913444409515, + "grad_norm": 0.4864092767238617, + "learning_rate": 1.98622453565667e-05, + "loss": 0.9151, + "step": 488 + }, + { + "epoch": 0.08198679660484125, + "grad_norm": 0.581265389919281, + "learning_rate": 1.9861345618072395e-05, + "loss": 0.9738, + "step": 489 + }, + { + "epoch": 0.08215445876558734, + "grad_norm": 0.37302395701408386, + "learning_rate": 1.986044297134516e-05, + "loss": 0.9457, + "step": 490 + }, + { + "epoch": 0.08232212092633344, + "grad_norm": 0.5029395222663879, + "learning_rate": 1.9859537416651197e-05, + "loss": 0.9713, + "step": 491 + }, + { + "epoch": 0.08248978308707954, + "grad_norm": 0.7264549136161804, + "learning_rate": 1.9858628954257554e-05, + "loss": 1.0336, + "step": 492 + }, + { + "epoch": 0.08265744524782563, + "grad_norm": 0.5837662816047668, + "learning_rate": 1.985771758443216e-05, + "loss": 0.9516, + "step": 493 + }, + { + "epoch": 0.08282510740857173, + "grad_norm": 0.4827551245689392, + "learning_rate": 1.9856803307443782e-05, + "loss": 0.9188, + "step": 494 + }, + { + "epoch": 0.08299276956931782, + "grad_norm": 0.5622090101242065, + "learning_rate": 1.9855886123562055e-05, + "loss": 0.9516, + "step": 495 + }, + { + "epoch": 0.08316043173006392, + "grad_norm": 0.5088374614715576, + "learning_rate": 1.9854966033057462e-05, + "loss": 0.9087, + "step": 496 + }, + { + "epoch": 0.08332809389081001, + "grad_norm": 0.5393046140670776, + "learning_rate": 1.985404303620136e-05, + "loss": 1.0514, + "step": 497 + }, + { + "epoch": 0.08349575605155611, + "grad_norm": 0.6230491995811462, + "learning_rate": 1.985311713326594e-05, + "loss": 1.0803, + "step": 498 + }, + { + "epoch": 0.0836634182123022, + "grad_norm": 0.47609272599220276, + "learning_rate": 1.9852188324524272e-05, + "loss": 0.8345, + "step": 499 + }, + { + "epoch": 0.0838310803730483, + "grad_norm": 0.6452209949493408, + "learning_rate": 1.9851256610250268e-05, + "loss": 0.9531, + "step": 500 + }, + { + "epoch": 0.0839987425337944, + "grad_norm": 0.4989963173866272, + "learning_rate": 1.9850321990718703e-05, + "loss": 1.0362, + "step": 501 + }, + { + "epoch": 0.0841664046945405, + "grad_norm": 0.5229847431182861, + "learning_rate": 1.984938446620521e-05, + "loss": 1.0242, + "step": 502 + }, + { + "epoch": 0.08433406685528659, + "grad_norm": 0.4766591489315033, + "learning_rate": 1.984844403698627e-05, + "loss": 1.058, + "step": 503 + }, + { + "epoch": 0.0845017290160327, + "grad_norm": 0.4473041594028473, + "learning_rate": 1.9847500703339238e-05, + "loss": 1.0586, + "step": 504 + }, + { + "epoch": 0.0846693911767788, + "grad_norm": 0.45140060782432556, + "learning_rate": 1.9846554465542306e-05, + "loss": 0.9494, + "step": 505 + }, + { + "epoch": 0.08483705333752489, + "grad_norm": 0.44374144077301025, + "learning_rate": 1.9845605323874532e-05, + "loss": 0.947, + "step": 506 + }, + { + "epoch": 0.08500471549827099, + "grad_norm": 0.41401800513267517, + "learning_rate": 1.9844653278615836e-05, + "loss": 0.9302, + "step": 507 + }, + { + "epoch": 0.08517237765901708, + "grad_norm": 0.5847159624099731, + "learning_rate": 1.984369833004698e-05, + "loss": 0.9366, + "step": 508 + }, + { + "epoch": 0.08534003981976318, + "grad_norm": 0.547697126865387, + "learning_rate": 1.9842740478449594e-05, + "loss": 1.0562, + "step": 509 + }, + { + "epoch": 0.08550770198050928, + "grad_norm": 0.5329285860061646, + "learning_rate": 1.9841779724106165e-05, + "loss": 0.9999, + "step": 510 + }, + { + "epoch": 0.08567536414125537, + "grad_norm": 0.5062839984893799, + "learning_rate": 1.9840816067300022e-05, + "loss": 1.045, + "step": 511 + }, + { + "epoch": 0.08584302630200147, + "grad_norm": 0.457762748003006, + "learning_rate": 1.9839849508315365e-05, + "loss": 1.0371, + "step": 512 + }, + { + "epoch": 0.08601068846274756, + "grad_norm": 0.5547719597816467, + "learning_rate": 1.9838880047437243e-05, + "loss": 1.0181, + "step": 513 + }, + { + "epoch": 0.08617835062349366, + "grad_norm": 0.5168294310569763, + "learning_rate": 1.983790768495156e-05, + "loss": 0.961, + "step": 514 + }, + { + "epoch": 0.08634601278423976, + "grad_norm": 0.6367170214653015, + "learning_rate": 1.9836932421145084e-05, + "loss": 0.999, + "step": 515 + }, + { + "epoch": 0.08651367494498585, + "grad_norm": 0.6939343214035034, + "learning_rate": 1.9835954256305423e-05, + "loss": 0.9431, + "step": 516 + }, + { + "epoch": 0.08668133710573195, + "grad_norm": 0.503060519695282, + "learning_rate": 1.9834973190721057e-05, + "loss": 0.877, + "step": 517 + }, + { + "epoch": 0.08684899926647804, + "grad_norm": 0.6143543124198914, + "learning_rate": 1.9833989224681314e-05, + "loss": 1.016, + "step": 518 + }, + { + "epoch": 0.08701666142722414, + "grad_norm": 0.4443577826023102, + "learning_rate": 1.9833002358476375e-05, + "loss": 0.9724, + "step": 519 + }, + { + "epoch": 0.08718432358797024, + "grad_norm": 0.5362754464149475, + "learning_rate": 1.983201259239728e-05, + "loss": 1.0051, + "step": 520 + }, + { + "epoch": 0.08735198574871633, + "grad_norm": 0.4462275505065918, + "learning_rate": 1.9831019926735924e-05, + "loss": 0.9377, + "step": 521 + }, + { + "epoch": 0.08751964790946243, + "grad_norm": 0.8423423767089844, + "learning_rate": 1.9830024361785055e-05, + "loss": 0.9989, + "step": 522 + }, + { + "epoch": 0.08768731007020852, + "grad_norm": 0.7546736001968384, + "learning_rate": 1.982902589783828e-05, + "loss": 1.0704, + "step": 523 + }, + { + "epoch": 0.08785497223095462, + "grad_norm": 0.536176323890686, + "learning_rate": 1.9828024535190056e-05, + "loss": 0.9193, + "step": 524 + }, + { + "epoch": 0.08802263439170073, + "grad_norm": 0.8669360876083374, + "learning_rate": 1.98270202741357e-05, + "loss": 0.9705, + "step": 525 + }, + { + "epoch": 0.08819029655244683, + "grad_norm": 0.5635377764701843, + "learning_rate": 1.9826013114971375e-05, + "loss": 0.9177, + "step": 526 + }, + { + "epoch": 0.08835795871319292, + "grad_norm": 0.5467398166656494, + "learning_rate": 1.9825003057994112e-05, + "loss": 0.9935, + "step": 527 + }, + { + "epoch": 0.08852562087393902, + "grad_norm": 0.45431026816368103, + "learning_rate": 1.9823990103501783e-05, + "loss": 1.0419, + "step": 528 + }, + { + "epoch": 0.08869328303468511, + "grad_norm": 0.5135222673416138, + "learning_rate": 1.9822974251793126e-05, + "loss": 0.9543, + "step": 529 + }, + { + "epoch": 0.08886094519543121, + "grad_norm": 0.4904458522796631, + "learning_rate": 1.9821955503167726e-05, + "loss": 1.0024, + "step": 530 + }, + { + "epoch": 0.0890286073561773, + "grad_norm": 0.3739793002605438, + "learning_rate": 1.9820933857926024e-05, + "loss": 0.9023, + "step": 531 + }, + { + "epoch": 0.0891962695169234, + "grad_norm": 0.4209132492542267, + "learning_rate": 1.9819909316369314e-05, + "loss": 1.0788, + "step": 532 + }, + { + "epoch": 0.0893639316776695, + "grad_norm": 0.5051009654998779, + "learning_rate": 1.9818881878799755e-05, + "loss": 1.0887, + "step": 533 + }, + { + "epoch": 0.0895315938384156, + "grad_norm": 0.43670955300331116, + "learning_rate": 1.981785154552034e-05, + "loss": 0.9689, + "step": 534 + }, + { + "epoch": 0.08969925599916169, + "grad_norm": 0.36185288429260254, + "learning_rate": 1.9816818316834933e-05, + "loss": 0.9555, + "step": 535 + }, + { + "epoch": 0.08986691815990779, + "grad_norm": 0.565312922000885, + "learning_rate": 1.981578219304824e-05, + "loss": 1.1048, + "step": 536 + }, + { + "epoch": 0.09003458032065388, + "grad_norm": 0.37618112564086914, + "learning_rate": 1.981474317446584e-05, + "loss": 0.8761, + "step": 537 + }, + { + "epoch": 0.09020224248139998, + "grad_norm": 0.47744205594062805, + "learning_rate": 1.9813701261394136e-05, + "loss": 0.9031, + "step": 538 + }, + { + "epoch": 0.09036990464214607, + "grad_norm": 0.3989582359790802, + "learning_rate": 1.981265645414041e-05, + "loss": 0.9217, + "step": 539 + }, + { + "epoch": 0.09053756680289217, + "grad_norm": 0.5119574069976807, + "learning_rate": 1.9811608753012792e-05, + "loss": 0.9871, + "step": 540 + }, + { + "epoch": 0.09070522896363827, + "grad_norm": 0.6921777725219727, + "learning_rate": 1.981055815832025e-05, + "loss": 1.056, + "step": 541 + }, + { + "epoch": 0.09087289112438436, + "grad_norm": 0.4229326546192169, + "learning_rate": 1.9809504670372634e-05, + "loss": 0.9696, + "step": 542 + }, + { + "epoch": 0.09104055328513046, + "grad_norm": 0.47192320227622986, + "learning_rate": 1.9808448289480614e-05, + "loss": 1.0248, + "step": 543 + }, + { + "epoch": 0.09120821544587655, + "grad_norm": 0.5146031379699707, + "learning_rate": 1.9807389015955742e-05, + "loss": 0.9254, + "step": 544 + }, + { + "epoch": 0.09137587760662265, + "grad_norm": 0.48524725437164307, + "learning_rate": 1.9806326850110406e-05, + "loss": 0.9548, + "step": 545 + }, + { + "epoch": 0.09154353976736875, + "grad_norm": 0.48073190450668335, + "learning_rate": 1.980526179225785e-05, + "loss": 0.9626, + "step": 546 + }, + { + "epoch": 0.09171120192811485, + "grad_norm": 0.6683670878410339, + "learning_rate": 1.980419384271217e-05, + "loss": 1.0316, + "step": 547 + }, + { + "epoch": 0.09187886408886095, + "grad_norm": 0.511667788028717, + "learning_rate": 1.9803123001788327e-05, + "loss": 0.9347, + "step": 548 + }, + { + "epoch": 0.09204652624960705, + "grad_norm": 0.4302925765514374, + "learning_rate": 1.980204926980212e-05, + "loss": 0.9375, + "step": 549 + }, + { + "epoch": 0.09221418841035314, + "grad_norm": 0.6657047271728516, + "learning_rate": 1.980097264707021e-05, + "loss": 1.0072, + "step": 550 + }, + { + "epoch": 0.09238185057109924, + "grad_norm": 0.5272813439369202, + "learning_rate": 1.97998931339101e-05, + "loss": 0.9524, + "step": 551 + }, + { + "epoch": 0.09254951273184533, + "grad_norm": 0.5764918327331543, + "learning_rate": 1.9798810730640153e-05, + "loss": 0.973, + "step": 552 + }, + { + "epoch": 0.09271717489259143, + "grad_norm": 0.41316327452659607, + "learning_rate": 1.9797725437579586e-05, + "loss": 0.9984, + "step": 553 + }, + { + "epoch": 0.09288483705333753, + "grad_norm": 0.40722402930259705, + "learning_rate": 1.9796637255048462e-05, + "loss": 0.82, + "step": 554 + }, + { + "epoch": 0.09305249921408362, + "grad_norm": 0.3617995083332062, + "learning_rate": 1.9795546183367708e-05, + "loss": 0.9146, + "step": 555 + }, + { + "epoch": 0.09322016137482972, + "grad_norm": 0.40822163224220276, + "learning_rate": 1.979445222285908e-05, + "loss": 0.963, + "step": 556 + }, + { + "epoch": 0.09338782353557581, + "grad_norm": 0.5854674577713013, + "learning_rate": 1.9793355373845214e-05, + "loss": 1.0111, + "step": 557 + }, + { + "epoch": 0.09355548569632191, + "grad_norm": 0.44850680232048035, + "learning_rate": 1.9792255636649574e-05, + "loss": 0.9146, + "step": 558 + }, + { + "epoch": 0.093723147857068, + "grad_norm": 0.4834839999675751, + "learning_rate": 1.9791153011596497e-05, + "loss": 1.0255, + "step": 559 + }, + { + "epoch": 0.0938908100178141, + "grad_norm": 0.5060991048812866, + "learning_rate": 1.9790047499011152e-05, + "loss": 1.0309, + "step": 560 + }, + { + "epoch": 0.0940584721785602, + "grad_norm": 0.5379565954208374, + "learning_rate": 1.9788939099219572e-05, + "loss": 0.9369, + "step": 561 + }, + { + "epoch": 0.0942261343393063, + "grad_norm": 0.44498369097709656, + "learning_rate": 1.9787827812548634e-05, + "loss": 0.9539, + "step": 562 + }, + { + "epoch": 0.09439379650005239, + "grad_norm": 0.45467501878738403, + "learning_rate": 1.9786713639326076e-05, + "loss": 0.9934, + "step": 563 + }, + { + "epoch": 0.09456145866079849, + "grad_norm": 0.454647034406662, + "learning_rate": 1.978559657988048e-05, + "loss": 0.8791, + "step": 564 + }, + { + "epoch": 0.09472912082154458, + "grad_norm": 0.4199789762496948, + "learning_rate": 1.9784476634541276e-05, + "loss": 0.9348, + "step": 565 + }, + { + "epoch": 0.09489678298229068, + "grad_norm": 0.34197133779525757, + "learning_rate": 1.9783353803638757e-05, + "loss": 0.9424, + "step": 566 + }, + { + "epoch": 0.09506444514303677, + "grad_norm": 0.49100300669670105, + "learning_rate": 1.978222808750405e-05, + "loss": 0.9444, + "step": 567 + }, + { + "epoch": 0.09523210730378288, + "grad_norm": 0.5817572474479675, + "learning_rate": 1.978109948646915e-05, + "loss": 0.9855, + "step": 568 + }, + { + "epoch": 0.09539976946452898, + "grad_norm": 0.6111281514167786, + "learning_rate": 1.97799680008669e-05, + "loss": 0.9756, + "step": 569 + }, + { + "epoch": 0.09556743162527508, + "grad_norm": 0.49359261989593506, + "learning_rate": 1.977883363103098e-05, + "loss": 0.9946, + "step": 570 + }, + { + "epoch": 0.09573509378602117, + "grad_norm": 0.5452761650085449, + "learning_rate": 1.9777696377295933e-05, + "loss": 1.0238, + "step": 571 + }, + { + "epoch": 0.09590275594676727, + "grad_norm": 0.4995993375778198, + "learning_rate": 1.9776556239997146e-05, + "loss": 0.9368, + "step": 572 + }, + { + "epoch": 0.09607041810751336, + "grad_norm": 0.5189254283905029, + "learning_rate": 1.9775413219470864e-05, + "loss": 0.9531, + "step": 573 + }, + { + "epoch": 0.09623808026825946, + "grad_norm": 0.6281453371047974, + "learning_rate": 1.9774267316054177e-05, + "loss": 0.9533, + "step": 574 + }, + { + "epoch": 0.09640574242900556, + "grad_norm": 0.8713141083717346, + "learning_rate": 1.9773118530085024e-05, + "loss": 0.8697, + "step": 575 + }, + { + "epoch": 0.09657340458975165, + "grad_norm": 0.6054361462593079, + "learning_rate": 1.97719668619022e-05, + "loss": 0.9665, + "step": 576 + }, + { + "epoch": 0.09674106675049775, + "grad_norm": 0.5086394548416138, + "learning_rate": 1.9770812311845344e-05, + "loss": 0.9154, + "step": 577 + }, + { + "epoch": 0.09690872891124384, + "grad_norm": 0.6285595297813416, + "learning_rate": 1.9769654880254944e-05, + "loss": 0.876, + "step": 578 + }, + { + "epoch": 0.09707639107198994, + "grad_norm": 0.4077214300632477, + "learning_rate": 1.9768494567472348e-05, + "loss": 0.8753, + "step": 579 + }, + { + "epoch": 0.09724405323273604, + "grad_norm": 0.47626930475234985, + "learning_rate": 1.976733137383974e-05, + "loss": 0.9437, + "step": 580 + }, + { + "epoch": 0.09741171539348213, + "grad_norm": 0.48559632897377014, + "learning_rate": 1.9766165299700164e-05, + "loss": 1.0168, + "step": 581 + }, + { + "epoch": 0.09757937755422823, + "grad_norm": 0.582771897315979, + "learning_rate": 1.9764996345397505e-05, + "loss": 0.8995, + "step": 582 + }, + { + "epoch": 0.09774703971497432, + "grad_norm": 0.5317685008049011, + "learning_rate": 1.976382451127651e-05, + "loss": 0.8447, + "step": 583 + }, + { + "epoch": 0.09791470187572042, + "grad_norm": 0.5318000912666321, + "learning_rate": 1.976264979768276e-05, + "loss": 1.0047, + "step": 584 + }, + { + "epoch": 0.09808236403646652, + "grad_norm": 0.6099218130111694, + "learning_rate": 1.97614722049627e-05, + "loss": 0.9661, + "step": 585 + }, + { + "epoch": 0.09825002619721261, + "grad_norm": 0.6967294812202454, + "learning_rate": 1.976029173346361e-05, + "loss": 1.0169, + "step": 586 + }, + { + "epoch": 0.09841768835795871, + "grad_norm": 0.741862952709198, + "learning_rate": 1.975910838353363e-05, + "loss": 1.0529, + "step": 587 + }, + { + "epoch": 0.0985853505187048, + "grad_norm": 0.5964228510856628, + "learning_rate": 1.9757922155521737e-05, + "loss": 1.0295, + "step": 588 + }, + { + "epoch": 0.09875301267945091, + "grad_norm": 0.5316452383995056, + "learning_rate": 1.9756733049777772e-05, + "loss": 1.0512, + "step": 589 + }, + { + "epoch": 0.09892067484019701, + "grad_norm": 0.533889889717102, + "learning_rate": 1.9755541066652416e-05, + "loss": 0.9739, + "step": 590 + }, + { + "epoch": 0.0990883370009431, + "grad_norm": 0.5988102555274963, + "learning_rate": 1.9754346206497197e-05, + "loss": 0.8895, + "step": 591 + }, + { + "epoch": 0.0992559991616892, + "grad_norm": 0.38310664892196655, + "learning_rate": 1.9753148469664496e-05, + "loss": 1.0405, + "step": 592 + }, + { + "epoch": 0.0994236613224353, + "grad_norm": 0.46506670117378235, + "learning_rate": 1.975194785650754e-05, + "loss": 0.975, + "step": 593 + }, + { + "epoch": 0.0995913234831814, + "grad_norm": 0.47853362560272217, + "learning_rate": 1.975074436738041e-05, + "loss": 0.9476, + "step": 594 + }, + { + "epoch": 0.09975898564392749, + "grad_norm": 0.5202714204788208, + "learning_rate": 1.9749538002638015e-05, + "loss": 0.98, + "step": 595 + }, + { + "epoch": 0.09992664780467359, + "grad_norm": 0.49216657876968384, + "learning_rate": 1.9748328762636138e-05, + "loss": 0.919, + "step": 596 + }, + { + "epoch": 0.10009430996541968, + "grad_norm": 0.37900805473327637, + "learning_rate": 1.9747116647731404e-05, + "loss": 0.9919, + "step": 597 + }, + { + "epoch": 0.10026197212616578, + "grad_norm": 0.5223008394241333, + "learning_rate": 1.9745901658281267e-05, + "loss": 0.9399, + "step": 598 + }, + { + "epoch": 0.10042963428691187, + "grad_norm": 0.6181545257568359, + "learning_rate": 1.9744683794644052e-05, + "loss": 1.0172, + "step": 599 + }, + { + "epoch": 0.10059729644765797, + "grad_norm": 0.5866363048553467, + "learning_rate": 1.974346305717892e-05, + "loss": 0.9147, + "step": 600 + }, + { + "epoch": 0.10076495860840406, + "grad_norm": 0.5286931991577148, + "learning_rate": 1.974223944624588e-05, + "loss": 1.0886, + "step": 601 + }, + { + "epoch": 0.10093262076915016, + "grad_norm": 0.7786131501197815, + "learning_rate": 1.974101296220579e-05, + "loss": 1.0236, + "step": 602 + }, + { + "epoch": 0.10110028292989626, + "grad_norm": 0.6727932095527649, + "learning_rate": 1.9739783605420357e-05, + "loss": 0.9602, + "step": 603 + }, + { + "epoch": 0.10126794509064235, + "grad_norm": 0.5062850713729858, + "learning_rate": 1.9738551376252134e-05, + "loss": 0.9491, + "step": 604 + }, + { + "epoch": 0.10143560725138845, + "grad_norm": 0.6286392211914062, + "learning_rate": 1.973731627506452e-05, + "loss": 1.0341, + "step": 605 + }, + { + "epoch": 0.10160326941213454, + "grad_norm": 0.7011929750442505, + "learning_rate": 1.973607830222176e-05, + "loss": 0.9357, + "step": 606 + }, + { + "epoch": 0.10177093157288064, + "grad_norm": 0.5614083409309387, + "learning_rate": 1.9734837458088947e-05, + "loss": 0.9866, + "step": 607 + }, + { + "epoch": 0.10193859373362674, + "grad_norm": 0.5135806202888489, + "learning_rate": 1.9733593743032024e-05, + "loss": 0.9159, + "step": 608 + }, + { + "epoch": 0.10210625589437283, + "grad_norm": 0.47957277297973633, + "learning_rate": 1.9732347157417776e-05, + "loss": 0.9339, + "step": 609 + }, + { + "epoch": 0.10227391805511893, + "grad_norm": 0.4980842173099518, + "learning_rate": 1.9731097701613843e-05, + "loss": 0.9543, + "step": 610 + }, + { + "epoch": 0.10244158021586504, + "grad_norm": 0.6782700419425964, + "learning_rate": 1.9729845375988694e-05, + "loss": 1.0212, + "step": 611 + }, + { + "epoch": 0.10260924237661113, + "grad_norm": 0.47850656509399414, + "learning_rate": 1.972859018091166e-05, + "loss": 0.9725, + "step": 612 + }, + { + "epoch": 0.10277690453735723, + "grad_norm": 0.5600951313972473, + "learning_rate": 1.9727332116752918e-05, + "loss": 0.9931, + "step": 613 + }, + { + "epoch": 0.10294456669810333, + "grad_norm": 0.445838987827301, + "learning_rate": 1.9726071183883482e-05, + "loss": 0.996, + "step": 614 + }, + { + "epoch": 0.10311222885884942, + "grad_norm": 0.6184260845184326, + "learning_rate": 1.9724807382675215e-05, + "loss": 0.9265, + "step": 615 + }, + { + "epoch": 0.10327989101959552, + "grad_norm": 0.8456107378005981, + "learning_rate": 1.9723540713500834e-05, + "loss": 0.9978, + "step": 616 + }, + { + "epoch": 0.10344755318034161, + "grad_norm": 0.5166464447975159, + "learning_rate": 1.972227117673389e-05, + "loss": 0.9357, + "step": 617 + }, + { + "epoch": 0.10361521534108771, + "grad_norm": 0.5970107913017273, + "learning_rate": 1.972099877274879e-05, + "loss": 0.9209, + "step": 618 + }, + { + "epoch": 0.1037828775018338, + "grad_norm": 0.4248644709587097, + "learning_rate": 1.971972350192077e-05, + "loss": 0.8596, + "step": 619 + }, + { + "epoch": 0.1039505396625799, + "grad_norm": 0.6106349229812622, + "learning_rate": 1.971844536462594e-05, + "loss": 0.9357, + "step": 620 + }, + { + "epoch": 0.104118201823326, + "grad_norm": 0.4165198802947998, + "learning_rate": 1.9717164361241225e-05, + "loss": 0.9382, + "step": 621 + }, + { + "epoch": 0.1042858639840721, + "grad_norm": 0.5961109399795532, + "learning_rate": 1.9715880492144415e-05, + "loss": 1.0102, + "step": 622 + }, + { + "epoch": 0.10445352614481819, + "grad_norm": 0.39271023869514465, + "learning_rate": 1.9714593757714143e-05, + "loss": 1.008, + "step": 623 + }, + { + "epoch": 0.10462118830556429, + "grad_norm": 0.4923511743545532, + "learning_rate": 1.9713304158329873e-05, + "loss": 1.0765, + "step": 624 + }, + { + "epoch": 0.10478885046631038, + "grad_norm": 0.49212756752967834, + "learning_rate": 1.9712011694371932e-05, + "loss": 0.9187, + "step": 625 + }, + { + "epoch": 0.10495651262705648, + "grad_norm": 0.4398774802684784, + "learning_rate": 1.971071636622148e-05, + "loss": 0.9804, + "step": 626 + }, + { + "epoch": 0.10512417478780257, + "grad_norm": 0.4320284426212311, + "learning_rate": 1.9709418174260523e-05, + "loss": 0.8478, + "step": 627 + }, + { + "epoch": 0.10529183694854867, + "grad_norm": 0.5619915127754211, + "learning_rate": 1.9708117118871918e-05, + "loss": 0.96, + "step": 628 + }, + { + "epoch": 0.10545949910929477, + "grad_norm": 0.4957972466945648, + "learning_rate": 1.9706813200439357e-05, + "loss": 0.9265, + "step": 629 + }, + { + "epoch": 0.10562716127004086, + "grad_norm": 0.41308099031448364, + "learning_rate": 1.970550641934739e-05, + "loss": 0.9377, + "step": 630 + }, + { + "epoch": 0.10579482343078696, + "grad_norm": 0.43232348561286926, + "learning_rate": 1.9704196775981397e-05, + "loss": 0.958, + "step": 631 + }, + { + "epoch": 0.10596248559153307, + "grad_norm": 0.5194548964500427, + "learning_rate": 1.970288427072761e-05, + "loss": 0.9631, + "step": 632 + }, + { + "epoch": 0.10613014775227916, + "grad_norm": 0.5242116451263428, + "learning_rate": 1.9701568903973104e-05, + "loss": 0.9468, + "step": 633 + }, + { + "epoch": 0.10629780991302526, + "grad_norm": 0.6986731886863708, + "learning_rate": 1.9700250676105793e-05, + "loss": 0.9996, + "step": 634 + }, + { + "epoch": 0.10646547207377136, + "grad_norm": 0.6663616895675659, + "learning_rate": 1.9698929587514444e-05, + "loss": 0.9063, + "step": 635 + }, + { + "epoch": 0.10663313423451745, + "grad_norm": 0.49162620306015015, + "learning_rate": 1.9697605638588655e-05, + "loss": 1.01, + "step": 636 + }, + { + "epoch": 0.10680079639526355, + "grad_norm": 0.5097907185554504, + "learning_rate": 1.9696278829718882e-05, + "loss": 0.9027, + "step": 637 + }, + { + "epoch": 0.10696845855600964, + "grad_norm": 0.46596378087997437, + "learning_rate": 1.9694949161296416e-05, + "loss": 0.9775, + "step": 638 + }, + { + "epoch": 0.10713612071675574, + "grad_norm": 0.4176481068134308, + "learning_rate": 1.969361663371339e-05, + "loss": 0.8596, + "step": 639 + }, + { + "epoch": 0.10730378287750184, + "grad_norm": 0.4506107568740845, + "learning_rate": 1.9692281247362785e-05, + "loss": 0.9667, + "step": 640 + }, + { + "epoch": 0.10747144503824793, + "grad_norm": 0.38935619592666626, + "learning_rate": 1.9690943002638425e-05, + "loss": 1.032, + "step": 641 + }, + { + "epoch": 0.10763910719899403, + "grad_norm": 0.45161691308021545, + "learning_rate": 1.968960189993497e-05, + "loss": 0.988, + "step": 642 + }, + { + "epoch": 0.10780676935974012, + "grad_norm": 0.6230657696723938, + "learning_rate": 1.968825793964793e-05, + "loss": 0.9048, + "step": 643 + }, + { + "epoch": 0.10797443152048622, + "grad_norm": 0.5937588810920715, + "learning_rate": 1.9686911122173657e-05, + "loss": 0.9231, + "step": 644 + }, + { + "epoch": 0.10814209368123232, + "grad_norm": 0.6096629500389099, + "learning_rate": 1.9685561447909342e-05, + "loss": 0.92, + "step": 645 + }, + { + "epoch": 0.10830975584197841, + "grad_norm": 0.6745415329933167, + "learning_rate": 1.9684208917253018e-05, + "loss": 0.9689, + "step": 646 + }, + { + "epoch": 0.10847741800272451, + "grad_norm": 0.5356251001358032, + "learning_rate": 1.968285353060357e-05, + "loss": 0.9586, + "step": 647 + }, + { + "epoch": 0.1086450801634706, + "grad_norm": 0.5918893218040466, + "learning_rate": 1.9681495288360716e-05, + "loss": 0.941, + "step": 648 + }, + { + "epoch": 0.1088127423242167, + "grad_norm": 0.46468257904052734, + "learning_rate": 1.9680134190925015e-05, + "loss": 0.9289, + "step": 649 + }, + { + "epoch": 0.1089804044849628, + "grad_norm": 0.46366575360298157, + "learning_rate": 1.9678770238697876e-05, + "loss": 0.9416, + "step": 650 + }, + { + "epoch": 0.10914806664570889, + "grad_norm": 0.5816429257392883, + "learning_rate": 1.9677403432081543e-05, + "loss": 0.9545, + "step": 651 + }, + { + "epoch": 0.10931572880645499, + "grad_norm": 0.9998892545700073, + "learning_rate": 1.9676033771479102e-05, + "loss": 0.951, + "step": 652 + }, + { + "epoch": 0.10948339096720108, + "grad_norm": 0.381295382976532, + "learning_rate": 1.9674661257294487e-05, + "loss": 1.0326, + "step": 653 + }, + { + "epoch": 0.10965105312794719, + "grad_norm": 0.46346166729927063, + "learning_rate": 1.9673285889932468e-05, + "loss": 1.0274, + "step": 654 + }, + { + "epoch": 0.10981871528869329, + "grad_norm": 0.44738730788230896, + "learning_rate": 1.9671907669798655e-05, + "loss": 0.8813, + "step": 655 + }, + { + "epoch": 0.10998637744943938, + "grad_norm": 0.5271283388137817, + "learning_rate": 1.9670526597299505e-05, + "loss": 0.9463, + "step": 656 + }, + { + "epoch": 0.11015403961018548, + "grad_norm": 0.54680335521698, + "learning_rate": 1.966914267284232e-05, + "loss": 0.9622, + "step": 657 + }, + { + "epoch": 0.11032170177093158, + "grad_norm": 0.5334712862968445, + "learning_rate": 1.966775589683522e-05, + "loss": 0.9911, + "step": 658 + }, + { + "epoch": 0.11048936393167767, + "grad_norm": 0.5559065341949463, + "learning_rate": 1.96663662696872e-05, + "loss": 0.9833, + "step": 659 + }, + { + "epoch": 0.11065702609242377, + "grad_norm": 0.5740633606910706, + "learning_rate": 1.9664973791808064e-05, + "loss": 0.8797, + "step": 660 + }, + { + "epoch": 0.11082468825316986, + "grad_norm": 0.4243875741958618, + "learning_rate": 1.966357846360848e-05, + "loss": 0.9998, + "step": 661 + }, + { + "epoch": 0.11099235041391596, + "grad_norm": 0.5085628032684326, + "learning_rate": 1.9662180285499944e-05, + "loss": 0.9774, + "step": 662 + }, + { + "epoch": 0.11116001257466206, + "grad_norm": 0.54347163438797, + "learning_rate": 1.96607792578948e-05, + "loss": 1.0116, + "step": 663 + }, + { + "epoch": 0.11132767473540815, + "grad_norm": 0.5442708134651184, + "learning_rate": 1.965937538120622e-05, + "loss": 0.9071, + "step": 664 + }, + { + "epoch": 0.11149533689615425, + "grad_norm": 0.5275413990020752, + "learning_rate": 1.9657968655848232e-05, + "loss": 0.9896, + "step": 665 + }, + { + "epoch": 0.11166299905690034, + "grad_norm": 0.3896923065185547, + "learning_rate": 1.9656559082235697e-05, + "loss": 0.9375, + "step": 666 + }, + { + "epoch": 0.11183066121764644, + "grad_norm": 0.4918743073940277, + "learning_rate": 1.965514666078431e-05, + "loss": 0.9335, + "step": 667 + }, + { + "epoch": 0.11199832337839254, + "grad_norm": 0.5577867031097412, + "learning_rate": 1.9653731391910614e-05, + "loss": 0.9044, + "step": 668 + }, + { + "epoch": 0.11216598553913863, + "grad_norm": 0.4111005961894989, + "learning_rate": 1.9652313276031995e-05, + "loss": 0.9077, + "step": 669 + }, + { + "epoch": 0.11233364769988473, + "grad_norm": 0.576390266418457, + "learning_rate": 1.9650892313566667e-05, + "loss": 1.0825, + "step": 670 + }, + { + "epoch": 0.11250130986063082, + "grad_norm": 0.5298464298248291, + "learning_rate": 1.9649468504933685e-05, + "loss": 0.9854, + "step": 671 + }, + { + "epoch": 0.11266897202137692, + "grad_norm": 0.5587629675865173, + "learning_rate": 1.964804185055296e-05, + "loss": 0.909, + "step": 672 + }, + { + "epoch": 0.11283663418212302, + "grad_norm": 0.6600515246391296, + "learning_rate": 1.9646612350845223e-05, + "loss": 0.9134, + "step": 673 + }, + { + "epoch": 0.11300429634286911, + "grad_norm": 0.5140649676322937, + "learning_rate": 1.9645180006232046e-05, + "loss": 1.0089, + "step": 674 + }, + { + "epoch": 0.11317195850361522, + "grad_norm": 0.7896112203598022, + "learning_rate": 1.9643744817135856e-05, + "loss": 0.8986, + "step": 675 + }, + { + "epoch": 0.11333962066436132, + "grad_norm": 0.5896207094192505, + "learning_rate": 1.9642306783979902e-05, + "loss": 0.9698, + "step": 676 + }, + { + "epoch": 0.11350728282510741, + "grad_norm": 0.5102990865707397, + "learning_rate": 1.9640865907188273e-05, + "loss": 0.993, + "step": 677 + }, + { + "epoch": 0.11367494498585351, + "grad_norm": 0.6648029685020447, + "learning_rate": 1.9639422187185913e-05, + "loss": 1.0052, + "step": 678 + }, + { + "epoch": 0.1138426071465996, + "grad_norm": 0.5018223524093628, + "learning_rate": 1.9637975624398583e-05, + "loss": 0.9741, + "step": 679 + }, + { + "epoch": 0.1140102693073457, + "grad_norm": 0.32673925161361694, + "learning_rate": 1.9636526219252898e-05, + "loss": 0.9169, + "step": 680 + }, + { + "epoch": 0.1141779314680918, + "grad_norm": 0.39858177304267883, + "learning_rate": 1.9635073972176303e-05, + "loss": 0.9473, + "step": 681 + }, + { + "epoch": 0.1143455936288379, + "grad_norm": 0.4796540439128876, + "learning_rate": 1.9633618883597087e-05, + "loss": 1.0437, + "step": 682 + }, + { + "epoch": 0.11451325578958399, + "grad_norm": 0.517875611782074, + "learning_rate": 1.963216095394437e-05, + "loss": 0.9306, + "step": 683 + }, + { + "epoch": 0.11468091795033009, + "grad_norm": 0.756864607334137, + "learning_rate": 1.9630700183648112e-05, + "loss": 0.8807, + "step": 684 + }, + { + "epoch": 0.11484858011107618, + "grad_norm": 0.4738442897796631, + "learning_rate": 1.9629236573139118e-05, + "loss": 0.9638, + "step": 685 + }, + { + "epoch": 0.11501624227182228, + "grad_norm": 0.45979225635528564, + "learning_rate": 1.9627770122849017e-05, + "loss": 0.8824, + "step": 686 + }, + { + "epoch": 0.11518390443256837, + "grad_norm": 0.47601425647735596, + "learning_rate": 1.9626300833210293e-05, + "loss": 0.9283, + "step": 687 + }, + { + "epoch": 0.11535156659331447, + "grad_norm": 0.49816030263900757, + "learning_rate": 1.962482870465625e-05, + "loss": 0.8787, + "step": 688 + }, + { + "epoch": 0.11551922875406057, + "grad_norm": 0.5461615920066833, + "learning_rate": 1.9623353737621035e-05, + "loss": 0.9864, + "step": 689 + }, + { + "epoch": 0.11568689091480666, + "grad_norm": 0.5288472771644592, + "learning_rate": 1.9621875932539643e-05, + "loss": 0.9758, + "step": 690 + }, + { + "epoch": 0.11585455307555276, + "grad_norm": 0.6684693098068237, + "learning_rate": 1.9620395289847894e-05, + "loss": 0.905, + "step": 691 + }, + { + "epoch": 0.11602221523629885, + "grad_norm": 0.46418723464012146, + "learning_rate": 1.9618911809982445e-05, + "loss": 0.9914, + "step": 692 + }, + { + "epoch": 0.11618987739704495, + "grad_norm": 0.677675187587738, + "learning_rate": 1.9617425493380796e-05, + "loss": 1.0004, + "step": 693 + }, + { + "epoch": 0.11635753955779105, + "grad_norm": 0.646614134311676, + "learning_rate": 1.9615936340481276e-05, + "loss": 1.0012, + "step": 694 + }, + { + "epoch": 0.11652520171853714, + "grad_norm": 0.8820503354072571, + "learning_rate": 1.9614444351723056e-05, + "loss": 0.8818, + "step": 695 + }, + { + "epoch": 0.11669286387928324, + "grad_norm": 0.6305606961250305, + "learning_rate": 1.9612949527546147e-05, + "loss": 0.9922, + "step": 696 + }, + { + "epoch": 0.11686052604002935, + "grad_norm": 0.664298415184021, + "learning_rate": 1.9611451868391384e-05, + "loss": 0.86, + "step": 697 + }, + { + "epoch": 0.11702818820077544, + "grad_norm": 0.845890462398529, + "learning_rate": 1.9609951374700443e-05, + "loss": 0.9133, + "step": 698 + }, + { + "epoch": 0.11719585036152154, + "grad_norm": 0.5247465372085571, + "learning_rate": 1.9608448046915852e-05, + "loss": 0.9349, + "step": 699 + }, + { + "epoch": 0.11736351252226764, + "grad_norm": 0.48808741569519043, + "learning_rate": 1.9606941885480947e-05, + "loss": 0.9186, + "step": 700 + }, + { + "epoch": 0.11753117468301373, + "grad_norm": 0.4177757501602173, + "learning_rate": 1.960543289083992e-05, + "loss": 0.9658, + "step": 701 + }, + { + "epoch": 0.11769883684375983, + "grad_norm": 0.4671350419521332, + "learning_rate": 1.9603921063437795e-05, + "loss": 0.9344, + "step": 702 + }, + { + "epoch": 0.11786649900450592, + "grad_norm": 0.4618290960788727, + "learning_rate": 1.960240640372042e-05, + "loss": 0.965, + "step": 703 + }, + { + "epoch": 0.11803416116525202, + "grad_norm": 0.5182496905326843, + "learning_rate": 1.9600888912134495e-05, + "loss": 0.9024, + "step": 704 + }, + { + "epoch": 0.11820182332599811, + "grad_norm": 0.5245487689971924, + "learning_rate": 1.9599368589127542e-05, + "loss": 0.933, + "step": 705 + }, + { + "epoch": 0.11836948548674421, + "grad_norm": 0.4822204113006592, + "learning_rate": 1.9597845435147924e-05, + "loss": 0.9824, + "step": 706 + }, + { + "epoch": 0.1185371476474903, + "grad_norm": 0.7162145972251892, + "learning_rate": 1.959631945064484e-05, + "loss": 1.0036, + "step": 707 + }, + { + "epoch": 0.1187048098082364, + "grad_norm": 0.4953795075416565, + "learning_rate": 1.9594790636068323e-05, + "loss": 0.897, + "step": 708 + }, + { + "epoch": 0.1188724719689825, + "grad_norm": 0.6701470017433167, + "learning_rate": 1.9593258991869235e-05, + "loss": 0.9, + "step": 709 + }, + { + "epoch": 0.1190401341297286, + "grad_norm": 0.72853022813797, + "learning_rate": 1.959172451849928e-05, + "loss": 0.9866, + "step": 710 + }, + { + "epoch": 0.11920779629047469, + "grad_norm": 0.5792762637138367, + "learning_rate": 1.9590187216410987e-05, + "loss": 0.9393, + "step": 711 + }, + { + "epoch": 0.11937545845122079, + "grad_norm": 0.5983287692070007, + "learning_rate": 1.9588647086057734e-05, + "loss": 1.029, + "step": 712 + }, + { + "epoch": 0.11954312061196688, + "grad_norm": 0.7085567712783813, + "learning_rate": 1.958710412789372e-05, + "loss": 0.9128, + "step": 713 + }, + { + "epoch": 0.11971078277271298, + "grad_norm": 0.44527891278266907, + "learning_rate": 1.9585558342373986e-05, + "loss": 0.9395, + "step": 714 + }, + { + "epoch": 0.11987844493345907, + "grad_norm": 0.5031890273094177, + "learning_rate": 1.9584009729954395e-05, + "loss": 1.034, + "step": 715 + }, + { + "epoch": 0.12004610709420517, + "grad_norm": 0.4327860176563263, + "learning_rate": 1.9582458291091664e-05, + "loss": 0.9596, + "step": 716 + }, + { + "epoch": 0.12021376925495127, + "grad_norm": 0.642092764377594, + "learning_rate": 1.9580904026243322e-05, + "loss": 0.9595, + "step": 717 + }, + { + "epoch": 0.12038143141569738, + "grad_norm": 0.6110759973526001, + "learning_rate": 1.9579346935867743e-05, + "loss": 0.9861, + "step": 718 + }, + { + "epoch": 0.12054909357644347, + "grad_norm": 0.46162545680999756, + "learning_rate": 1.9577787020424135e-05, + "loss": 0.8916, + "step": 719 + }, + { + "epoch": 0.12071675573718957, + "grad_norm": 0.5048706531524658, + "learning_rate": 1.9576224280372537e-05, + "loss": 0.9519, + "step": 720 + }, + { + "epoch": 0.12088441789793566, + "grad_norm": 0.44191446900367737, + "learning_rate": 1.9574658716173817e-05, + "loss": 1.0365, + "step": 721 + }, + { + "epoch": 0.12105208005868176, + "grad_norm": 0.7060146331787109, + "learning_rate": 1.9573090328289675e-05, + "loss": 0.9745, + "step": 722 + }, + { + "epoch": 0.12121974221942786, + "grad_norm": 0.48769238591194153, + "learning_rate": 1.957151911718266e-05, + "loss": 1.0441, + "step": 723 + }, + { + "epoch": 0.12138740438017395, + "grad_norm": 0.4783252775669098, + "learning_rate": 1.9569945083316133e-05, + "loss": 0.9536, + "step": 724 + }, + { + "epoch": 0.12155506654092005, + "grad_norm": 0.5853663682937622, + "learning_rate": 1.9568368227154297e-05, + "loss": 0.8748, + "step": 725 + }, + { + "epoch": 0.12172272870166614, + "grad_norm": 0.5817577242851257, + "learning_rate": 1.9566788549162188e-05, + "loss": 1.0118, + "step": 726 + }, + { + "epoch": 0.12189039086241224, + "grad_norm": 0.44570523500442505, + "learning_rate": 1.9565206049805673e-05, + "loss": 0.934, + "step": 727 + }, + { + "epoch": 0.12205805302315834, + "grad_norm": 0.7238737940788269, + "learning_rate": 1.9563620729551448e-05, + "loss": 0.9746, + "step": 728 + }, + { + "epoch": 0.12222571518390443, + "grad_norm": 0.43183931708335876, + "learning_rate": 1.9562032588867047e-05, + "loss": 0.9409, + "step": 729 + }, + { + "epoch": 0.12239337734465053, + "grad_norm": 0.4638485014438629, + "learning_rate": 1.956044162822083e-05, + "loss": 0.8582, + "step": 730 + }, + { + "epoch": 0.12256103950539662, + "grad_norm": 0.5277942419052124, + "learning_rate": 1.955884784808199e-05, + "loss": 0.899, + "step": 731 + }, + { + "epoch": 0.12272870166614272, + "grad_norm": 0.5024052858352661, + "learning_rate": 1.9557251248920556e-05, + "loss": 1.0152, + "step": 732 + }, + { + "epoch": 0.12289636382688882, + "grad_norm": 0.46881332993507385, + "learning_rate": 1.9555651831207383e-05, + "loss": 0.9594, + "step": 733 + }, + { + "epoch": 0.12306402598763491, + "grad_norm": 0.6736809015274048, + "learning_rate": 1.955404959541416e-05, + "loss": 1.0201, + "step": 734 + }, + { + "epoch": 0.12323168814838101, + "grad_norm": 1.2368195056915283, + "learning_rate": 1.95524445420134e-05, + "loss": 0.8936, + "step": 735 + }, + { + "epoch": 0.1233993503091271, + "grad_norm": 0.5060524344444275, + "learning_rate": 1.9550836671478465e-05, + "loss": 0.8747, + "step": 736 + }, + { + "epoch": 0.1235670124698732, + "grad_norm": 0.7231465578079224, + "learning_rate": 1.9549225984283528e-05, + "loss": 0.8913, + "step": 737 + }, + { + "epoch": 0.1237346746306193, + "grad_norm": 0.5625017881393433, + "learning_rate": 1.9547612480903598e-05, + "loss": 1.0382, + "step": 738 + }, + { + "epoch": 0.1239023367913654, + "grad_norm": 0.6068775057792664, + "learning_rate": 1.9545996161814526e-05, + "loss": 0.9309, + "step": 739 + }, + { + "epoch": 0.1240699989521115, + "grad_norm": 0.5323880314826965, + "learning_rate": 1.9544377027492974e-05, + "loss": 0.861, + "step": 740 + }, + { + "epoch": 0.1242376611128576, + "grad_norm": 0.5942120552062988, + "learning_rate": 1.954275507841646e-05, + "loss": 0.9093, + "step": 741 + }, + { + "epoch": 0.1244053232736037, + "grad_norm": 0.5026261210441589, + "learning_rate": 1.9541130315063297e-05, + "loss": 0.9982, + "step": 742 + }, + { + "epoch": 0.12457298543434979, + "grad_norm": 0.5589470863342285, + "learning_rate": 1.9539502737912664e-05, + "loss": 0.9119, + "step": 743 + }, + { + "epoch": 0.12474064759509589, + "grad_norm": 0.7846218347549438, + "learning_rate": 1.9537872347444544e-05, + "loss": 0.9607, + "step": 744 + }, + { + "epoch": 0.12490830975584198, + "grad_norm": 0.5878773927688599, + "learning_rate": 1.9536239144139767e-05, + "loss": 0.9941, + "step": 745 + }, + { + "epoch": 0.12507597191658806, + "grad_norm": 0.5514938235282898, + "learning_rate": 1.9534603128479976e-05, + "loss": 0.9315, + "step": 746 + }, + { + "epoch": 0.12524363407733416, + "grad_norm": 0.7384706735610962, + "learning_rate": 1.953296430094766e-05, + "loss": 0.9754, + "step": 747 + }, + { + "epoch": 0.12541129623808026, + "grad_norm": 0.45522984862327576, + "learning_rate": 1.953132266202613e-05, + "loss": 0.9331, + "step": 748 + }, + { + "epoch": 0.12557895839882635, + "grad_norm": 0.6133016347885132, + "learning_rate": 1.9529678212199517e-05, + "loss": 0.9913, + "step": 749 + }, + { + "epoch": 0.12574662055957248, + "grad_norm": 0.6371902227401733, + "learning_rate": 1.9528030951952796e-05, + "loss": 1.0075, + "step": 750 + }, + { + "epoch": 0.12591428272031857, + "grad_norm": 0.5542172789573669, + "learning_rate": 1.9526380881771766e-05, + "loss": 0.9307, + "step": 751 + }, + { + "epoch": 0.12608194488106467, + "grad_norm": 0.5045576095581055, + "learning_rate": 1.952472800214305e-05, + "loss": 0.9441, + "step": 752 + }, + { + "epoch": 0.12624960704181076, + "grad_norm": 0.4427657723426819, + "learning_rate": 1.95230723135541e-05, + "loss": 0.8946, + "step": 753 + }, + { + "epoch": 0.12641726920255686, + "grad_norm": 0.4645458161830902, + "learning_rate": 1.9521413816493206e-05, + "loss": 0.9801, + "step": 754 + }, + { + "epoch": 0.12658493136330295, + "grad_norm": 0.4842035174369812, + "learning_rate": 1.9519752511449476e-05, + "loss": 0.9389, + "step": 755 + }, + { + "epoch": 0.12675259352404905, + "grad_norm": 0.4109826683998108, + "learning_rate": 1.9518088398912846e-05, + "loss": 0.9059, + "step": 756 + }, + { + "epoch": 0.12692025568479515, + "grad_norm": 0.4934842586517334, + "learning_rate": 1.9516421479374088e-05, + "loss": 0.9971, + "step": 757 + }, + { + "epoch": 0.12708791784554124, + "grad_norm": 0.3934910297393799, + "learning_rate": 1.9514751753324795e-05, + "loss": 0.9402, + "step": 758 + }, + { + "epoch": 0.12725558000628734, + "grad_norm": 0.45803117752075195, + "learning_rate": 1.951307922125739e-05, + "loss": 0.9056, + "step": 759 + }, + { + "epoch": 0.12742324216703343, + "grad_norm": 0.4900297522544861, + "learning_rate": 1.9511403883665126e-05, + "loss": 0.892, + "step": 760 + }, + { + "epoch": 0.12759090432777953, + "grad_norm": 0.5113170146942139, + "learning_rate": 1.9509725741042077e-05, + "loss": 0.9232, + "step": 761 + }, + { + "epoch": 0.12775856648852563, + "grad_norm": 0.3888622522354126, + "learning_rate": 1.9508044793883152e-05, + "loss": 0.8747, + "step": 762 + }, + { + "epoch": 0.12792622864927172, + "grad_norm": 0.5268617272377014, + "learning_rate": 1.9506361042684077e-05, + "loss": 0.8486, + "step": 763 + }, + { + "epoch": 0.12809389081001782, + "grad_norm": 0.49528518319129944, + "learning_rate": 1.950467448794142e-05, + "loss": 0.8963, + "step": 764 + }, + { + "epoch": 0.12826155297076391, + "grad_norm": 0.38182708621025085, + "learning_rate": 1.9502985130152557e-05, + "loss": 0.9232, + "step": 765 + }, + { + "epoch": 0.12842921513151, + "grad_norm": 0.4788253903388977, + "learning_rate": 1.950129296981571e-05, + "loss": 0.9111, + "step": 766 + }, + { + "epoch": 0.1285968772922561, + "grad_norm": 0.49691474437713623, + "learning_rate": 1.949959800742991e-05, + "loss": 0.99, + "step": 767 + }, + { + "epoch": 0.1287645394530022, + "grad_norm": 0.7026716470718384, + "learning_rate": 1.949790024349503e-05, + "loss": 0.9412, + "step": 768 + }, + { + "epoch": 0.1289322016137483, + "grad_norm": 0.3846409320831299, + "learning_rate": 1.949619967851175e-05, + "loss": 0.9823, + "step": 769 + }, + { + "epoch": 0.1290998637744944, + "grad_norm": 0.5087264776229858, + "learning_rate": 1.94944963129816e-05, + "loss": 0.9425, + "step": 770 + }, + { + "epoch": 0.1292675259352405, + "grad_norm": 0.42938852310180664, + "learning_rate": 1.9492790147406918e-05, + "loss": 0.9304, + "step": 771 + }, + { + "epoch": 0.1294351880959866, + "grad_norm": 0.4374798834323883, + "learning_rate": 1.9491081182290877e-05, + "loss": 0.8703, + "step": 772 + }, + { + "epoch": 0.12960285025673268, + "grad_norm": 0.3993895947933197, + "learning_rate": 1.9489369418137464e-05, + "loss": 0.8903, + "step": 773 + }, + { + "epoch": 0.12977051241747878, + "grad_norm": 0.4784952700138092, + "learning_rate": 1.9487654855451507e-05, + "loss": 1.0, + "step": 774 + }, + { + "epoch": 0.12993817457822487, + "grad_norm": 0.48840826749801636, + "learning_rate": 1.948593749473865e-05, + "loss": 0.8849, + "step": 775 + }, + { + "epoch": 0.13010583673897097, + "grad_norm": 0.42742201685905457, + "learning_rate": 1.9484217336505362e-05, + "loss": 0.992, + "step": 776 + }, + { + "epoch": 0.13027349889971707, + "grad_norm": 0.4182666838169098, + "learning_rate": 1.948249438125894e-05, + "loss": 1.0784, + "step": 777 + }, + { + "epoch": 0.13044116106046316, + "grad_norm": 0.3841330111026764, + "learning_rate": 1.9480768629507508e-05, + "loss": 0.9422, + "step": 778 + }, + { + "epoch": 0.13060882322120926, + "grad_norm": 0.3590227961540222, + "learning_rate": 1.9479040081760005e-05, + "loss": 0.9938, + "step": 779 + }, + { + "epoch": 0.13077648538195535, + "grad_norm": 0.5238223075866699, + "learning_rate": 1.9477308738526207e-05, + "loss": 0.8615, + "step": 780 + }, + { + "epoch": 0.13094414754270145, + "grad_norm": 0.423281192779541, + "learning_rate": 1.9475574600316704e-05, + "loss": 0.9103, + "step": 781 + }, + { + "epoch": 0.13111180970344755, + "grad_norm": 0.38820210099220276, + "learning_rate": 1.947383766764292e-05, + "loss": 0.8673, + "step": 782 + }, + { + "epoch": 0.13127947186419364, + "grad_norm": 0.6232205033302307, + "learning_rate": 1.9472097941017092e-05, + "loss": 0.8361, + "step": 783 + }, + { + "epoch": 0.13144713402493974, + "grad_norm": 0.5535816550254822, + "learning_rate": 1.947035542095229e-05, + "loss": 0.9535, + "step": 784 + }, + { + "epoch": 0.13161479618568583, + "grad_norm": 0.4846842885017395, + "learning_rate": 1.9468610107962405e-05, + "loss": 0.9282, + "step": 785 + }, + { + "epoch": 0.13178245834643193, + "grad_norm": 0.8760911822319031, + "learning_rate": 1.9466862002562154e-05, + "loss": 0.9522, + "step": 786 + }, + { + "epoch": 0.13195012050717803, + "grad_norm": 0.5710975527763367, + "learning_rate": 1.946511110526707e-05, + "loss": 1.0379, + "step": 787 + }, + { + "epoch": 0.13211778266792412, + "grad_norm": 0.7084416747093201, + "learning_rate": 1.9463357416593517e-05, + "loss": 1.0698, + "step": 788 + }, + { + "epoch": 0.13228544482867022, + "grad_norm": 0.6290233135223389, + "learning_rate": 1.946160093705868e-05, + "loss": 0.8534, + "step": 789 + }, + { + "epoch": 0.1324531069894163, + "grad_norm": 0.5018370747566223, + "learning_rate": 1.945984166718056e-05, + "loss": 1.0086, + "step": 790 + }, + { + "epoch": 0.1326207691501624, + "grad_norm": 0.46618857979774475, + "learning_rate": 1.9458079607478e-05, + "loss": 0.903, + "step": 791 + }, + { + "epoch": 0.1327884313109085, + "grad_norm": 0.40250569581985474, + "learning_rate": 1.9456314758470642e-05, + "loss": 0.9104, + "step": 792 + }, + { + "epoch": 0.13295609347165463, + "grad_norm": 0.44152161478996277, + "learning_rate": 1.9454547120678966e-05, + "loss": 0.9364, + "step": 793 + }, + { + "epoch": 0.13312375563240073, + "grad_norm": 0.4875357151031494, + "learning_rate": 1.945277669462427e-05, + "loss": 0.9933, + "step": 794 + }, + { + "epoch": 0.13329141779314682, + "grad_norm": 0.43597495555877686, + "learning_rate": 1.9451003480828677e-05, + "loss": 1.0492, + "step": 795 + }, + { + "epoch": 0.13345907995389292, + "grad_norm": 0.5917737483978271, + "learning_rate": 1.9449227479815126e-05, + "loss": 0.9141, + "step": 796 + }, + { + "epoch": 0.133626742114639, + "grad_norm": 0.6467670202255249, + "learning_rate": 1.9447448692107383e-05, + "loss": 0.9828, + "step": 797 + }, + { + "epoch": 0.1337944042753851, + "grad_norm": 0.6712422966957092, + "learning_rate": 1.9445667118230037e-05, + "loss": 0.9484, + "step": 798 + }, + { + "epoch": 0.1339620664361312, + "grad_norm": 0.4396202564239502, + "learning_rate": 1.944388275870849e-05, + "loss": 1.0194, + "step": 799 + }, + { + "epoch": 0.1341297285968773, + "grad_norm": 0.5531654357910156, + "learning_rate": 1.9442095614068977e-05, + "loss": 0.9144, + "step": 800 + }, + { + "epoch": 0.1342973907576234, + "grad_norm": 0.5850380063056946, + "learning_rate": 1.944030568483855e-05, + "loss": 0.8749, + "step": 801 + }, + { + "epoch": 0.1344650529183695, + "grad_norm": 0.5013009309768677, + "learning_rate": 1.943851297154508e-05, + "loss": 0.92, + "step": 802 + }, + { + "epoch": 0.1346327150791156, + "grad_norm": 0.501086413860321, + "learning_rate": 1.9436717474717253e-05, + "loss": 0.9618, + "step": 803 + }, + { + "epoch": 0.13480037723986169, + "grad_norm": 0.4982043206691742, + "learning_rate": 1.9434919194884593e-05, + "loss": 0.9695, + "step": 804 + }, + { + "epoch": 0.13496803940060778, + "grad_norm": 0.3764241635799408, + "learning_rate": 1.9433118132577432e-05, + "loss": 0.8995, + "step": 805 + }, + { + "epoch": 0.13513570156135388, + "grad_norm": 0.533713161945343, + "learning_rate": 1.9431314288326925e-05, + "loss": 1.018, + "step": 806 + }, + { + "epoch": 0.13530336372209997, + "grad_norm": 0.5478500127792358, + "learning_rate": 1.942950766266505e-05, + "loss": 0.9551, + "step": 807 + }, + { + "epoch": 0.13547102588284607, + "grad_norm": 0.6805524826049805, + "learning_rate": 1.9427698256124598e-05, + "loss": 1.0809, + "step": 808 + }, + { + "epoch": 0.13563868804359216, + "grad_norm": 0.48846960067749023, + "learning_rate": 1.9425886069239192e-05, + "loss": 0.9434, + "step": 809 + }, + { + "epoch": 0.13580635020433826, + "grad_norm": 0.5114566683769226, + "learning_rate": 1.9424071102543262e-05, + "loss": 0.982, + "step": 810 + }, + { + "epoch": 0.13597401236508436, + "grad_norm": 0.5536603927612305, + "learning_rate": 1.942225335657207e-05, + "loss": 1.0784, + "step": 811 + }, + { + "epoch": 0.13614167452583045, + "grad_norm": 0.5240349173545837, + "learning_rate": 1.9420432831861688e-05, + "loss": 0.8194, + "step": 812 + }, + { + "epoch": 0.13630933668657655, + "grad_norm": 0.42566990852355957, + "learning_rate": 1.941860952894901e-05, + "loss": 0.9778, + "step": 813 + }, + { + "epoch": 0.13647699884732264, + "grad_norm": 0.49338826537132263, + "learning_rate": 1.9416783448371758e-05, + "loss": 0.8738, + "step": 814 + }, + { + "epoch": 0.13664466100806874, + "grad_norm": 0.5864494442939758, + "learning_rate": 1.941495459066846e-05, + "loss": 0.9206, + "step": 815 + }, + { + "epoch": 0.13681232316881484, + "grad_norm": 0.3878153860569, + "learning_rate": 1.9413122956378466e-05, + "loss": 0.9598, + "step": 816 + }, + { + "epoch": 0.13697998532956093, + "grad_norm": 0.9809596538543701, + "learning_rate": 1.941128854604195e-05, + "loss": 0.9102, + "step": 817 + }, + { + "epoch": 0.13714764749030703, + "grad_norm": 0.5030829906463623, + "learning_rate": 1.9409451360199907e-05, + "loss": 0.9994, + "step": 818 + }, + { + "epoch": 0.13731530965105312, + "grad_norm": 0.510537326335907, + "learning_rate": 1.9407611399394145e-05, + "loss": 0.8467, + "step": 819 + }, + { + "epoch": 0.13748297181179922, + "grad_norm": 0.6032730937004089, + "learning_rate": 1.9405768664167284e-05, + "loss": 0.978, + "step": 820 + }, + { + "epoch": 0.13765063397254532, + "grad_norm": 0.5233103632926941, + "learning_rate": 1.9403923155062776e-05, + "loss": 0.9526, + "step": 821 + }, + { + "epoch": 0.1378182961332914, + "grad_norm": 0.6112736463546753, + "learning_rate": 1.9402074872624883e-05, + "loss": 0.9314, + "step": 822 + }, + { + "epoch": 0.1379859582940375, + "grad_norm": 0.6209530234336853, + "learning_rate": 1.9400223817398684e-05, + "loss": 1.0253, + "step": 823 + }, + { + "epoch": 0.1381536204547836, + "grad_norm": 0.5491402745246887, + "learning_rate": 1.939836998993008e-05, + "loss": 0.919, + "step": 824 + }, + { + "epoch": 0.1383212826155297, + "grad_norm": 0.8800770044326782, + "learning_rate": 1.939651339076579e-05, + "loss": 0.8918, + "step": 825 + }, + { + "epoch": 0.1384889447762758, + "grad_norm": 0.48260605335235596, + "learning_rate": 1.939465402045335e-05, + "loss": 1.0064, + "step": 826 + }, + { + "epoch": 0.1386566069370219, + "grad_norm": 0.7165107131004333, + "learning_rate": 1.9392791879541096e-05, + "loss": 0.8179, + "step": 827 + }, + { + "epoch": 0.138824269097768, + "grad_norm": 0.4312284290790558, + "learning_rate": 1.9390926968578213e-05, + "loss": 0.9617, + "step": 828 + }, + { + "epoch": 0.13899193125851408, + "grad_norm": 0.4455448091030121, + "learning_rate": 1.9389059288114684e-05, + "loss": 0.8876, + "step": 829 + }, + { + "epoch": 0.13915959341926018, + "grad_norm": 0.55988609790802, + "learning_rate": 1.9387188838701304e-05, + "loss": 0.9553, + "step": 830 + }, + { + "epoch": 0.13932725558000628, + "grad_norm": 0.5089917778968811, + "learning_rate": 1.93853156208897e-05, + "loss": 0.8604, + "step": 831 + }, + { + "epoch": 0.13949491774075237, + "grad_norm": 0.487514466047287, + "learning_rate": 1.9383439635232296e-05, + "loss": 0.9055, + "step": 832 + }, + { + "epoch": 0.13966257990149847, + "grad_norm": 0.4499186873435974, + "learning_rate": 1.9381560882282354e-05, + "loss": 0.991, + "step": 833 + }, + { + "epoch": 0.13983024206224456, + "grad_norm": 0.43431034684181213, + "learning_rate": 1.9379679362593937e-05, + "loss": 0.9932, + "step": 834 + }, + { + "epoch": 0.13999790422299066, + "grad_norm": 0.7048651576042175, + "learning_rate": 1.937779507672193e-05, + "loss": 0.9272, + "step": 835 + }, + { + "epoch": 0.14016556638373678, + "grad_norm": 0.9253151416778564, + "learning_rate": 1.9375908025222033e-05, + "loss": 0.9153, + "step": 836 + }, + { + "epoch": 0.14033322854448288, + "grad_norm": 0.9569368362426758, + "learning_rate": 1.9374018208650755e-05, + "loss": 0.9495, + "step": 837 + }, + { + "epoch": 0.14050089070522898, + "grad_norm": 1.2080894708633423, + "learning_rate": 1.9372125627565435e-05, + "loss": 0.9293, + "step": 838 + }, + { + "epoch": 0.14066855286597507, + "grad_norm": 0.41559672355651855, + "learning_rate": 1.937023028252421e-05, + "loss": 0.8253, + "step": 839 + }, + { + "epoch": 0.14083621502672117, + "grad_norm": 1.0294303894042969, + "learning_rate": 1.9368332174086046e-05, + "loss": 1.0341, + "step": 840 + }, + { + "epoch": 0.14100387718746726, + "grad_norm": 0.47193440794944763, + "learning_rate": 1.936643130281072e-05, + "loss": 0.9289, + "step": 841 + }, + { + "epoch": 0.14117153934821336, + "grad_norm": 0.499761164188385, + "learning_rate": 1.9364527669258815e-05, + "loss": 0.9112, + "step": 842 + }, + { + "epoch": 0.14133920150895946, + "grad_norm": 0.40940871834754944, + "learning_rate": 1.9362621273991747e-05, + "loss": 0.8562, + "step": 843 + }, + { + "epoch": 0.14150686366970555, + "grad_norm": 0.897140622138977, + "learning_rate": 1.936071211757173e-05, + "loss": 0.9378, + "step": 844 + }, + { + "epoch": 0.14167452583045165, + "grad_norm": 0.6589300632476807, + "learning_rate": 1.935880020056179e-05, + "loss": 0.8975, + "step": 845 + }, + { + "epoch": 0.14184218799119774, + "grad_norm": 0.5597459077835083, + "learning_rate": 1.935688552352579e-05, + "loss": 0.9226, + "step": 846 + }, + { + "epoch": 0.14200985015194384, + "grad_norm": 0.7280226945877075, + "learning_rate": 1.9354968087028383e-05, + "loss": 0.8995, + "step": 847 + }, + { + "epoch": 0.14217751231268994, + "grad_norm": 0.8181997537612915, + "learning_rate": 1.9353047891635044e-05, + "loss": 1.011, + "step": 848 + }, + { + "epoch": 0.14234517447343603, + "grad_norm": 0.5827728509902954, + "learning_rate": 1.9351124937912065e-05, + "loss": 1.0122, + "step": 849 + }, + { + "epoch": 0.14251283663418213, + "grad_norm": 0.47748494148254395, + "learning_rate": 1.9349199226426552e-05, + "loss": 1.0305, + "step": 850 + }, + { + "epoch": 0.14268049879492822, + "grad_norm": 0.40899279713630676, + "learning_rate": 1.9347270757746416e-05, + "loss": 0.9256, + "step": 851 + }, + { + "epoch": 0.14284816095567432, + "grad_norm": 0.4111090302467346, + "learning_rate": 1.9345339532440387e-05, + "loss": 0.9297, + "step": 852 + }, + { + "epoch": 0.14301582311642042, + "grad_norm": 0.48824769258499146, + "learning_rate": 1.9343405551078008e-05, + "loss": 0.9565, + "step": 853 + }, + { + "epoch": 0.1431834852771665, + "grad_norm": 0.5168846249580383, + "learning_rate": 1.9341468814229635e-05, + "loss": 0.9212, + "step": 854 + }, + { + "epoch": 0.1433511474379126, + "grad_norm": 0.5405130386352539, + "learning_rate": 1.9339529322466432e-05, + "loss": 1.0312, + "step": 855 + }, + { + "epoch": 0.1435188095986587, + "grad_norm": 0.4492095112800598, + "learning_rate": 1.9337587076360387e-05, + "loss": 0.865, + "step": 856 + }, + { + "epoch": 0.1436864717594048, + "grad_norm": 0.562528133392334, + "learning_rate": 1.9335642076484283e-05, + "loss": 0.9668, + "step": 857 + }, + { + "epoch": 0.1438541339201509, + "grad_norm": 0.6639419198036194, + "learning_rate": 1.9333694323411732e-05, + "loss": 0.9625, + "step": 858 + }, + { + "epoch": 0.144021796080897, + "grad_norm": 0.42684265971183777, + "learning_rate": 1.933174381771715e-05, + "loss": 0.855, + "step": 859 + }, + { + "epoch": 0.1441894582416431, + "grad_norm": 0.6701211333274841, + "learning_rate": 1.932979055997576e-05, + "loss": 1.0239, + "step": 860 + }, + { + "epoch": 0.14435712040238918, + "grad_norm": 0.619564950466156, + "learning_rate": 1.9327834550763605e-05, + "loss": 0.9425, + "step": 861 + }, + { + "epoch": 0.14452478256313528, + "grad_norm": 0.6924566626548767, + "learning_rate": 1.9325875790657537e-05, + "loss": 1.0193, + "step": 862 + }, + { + "epoch": 0.14469244472388137, + "grad_norm": 0.4808378219604492, + "learning_rate": 1.9323914280235218e-05, + "loss": 0.9406, + "step": 863 + }, + { + "epoch": 0.14486010688462747, + "grad_norm": 0.797642171382904, + "learning_rate": 1.9321950020075117e-05, + "loss": 0.9621, + "step": 864 + }, + { + "epoch": 0.14502776904537357, + "grad_norm": 0.5004295110702515, + "learning_rate": 1.9319983010756527e-05, + "loss": 0.848, + "step": 865 + }, + { + "epoch": 0.14519543120611966, + "grad_norm": 0.5292854905128479, + "learning_rate": 1.931801325285954e-05, + "loss": 0.9103, + "step": 866 + }, + { + "epoch": 0.14536309336686576, + "grad_norm": 0.5634934306144714, + "learning_rate": 1.931604074696506e-05, + "loss": 0.9064, + "step": 867 + }, + { + "epoch": 0.14553075552761185, + "grad_norm": 0.5075427889823914, + "learning_rate": 1.9314065493654806e-05, + "loss": 1.054, + "step": 868 + }, + { + "epoch": 0.14569841768835795, + "grad_norm": 0.46562308073043823, + "learning_rate": 1.9312087493511304e-05, + "loss": 0.8924, + "step": 869 + }, + { + "epoch": 0.14586607984910405, + "grad_norm": 0.6002721786499023, + "learning_rate": 1.931010674711789e-05, + "loss": 0.9316, + "step": 870 + }, + { + "epoch": 0.14603374200985014, + "grad_norm": 0.5705669522285461, + "learning_rate": 1.930812325505871e-05, + "loss": 0.903, + "step": 871 + }, + { + "epoch": 0.14620140417059624, + "grad_norm": 0.43029841780662537, + "learning_rate": 1.930613701791872e-05, + "loss": 0.9611, + "step": 872 + }, + { + "epoch": 0.14636906633134233, + "grad_norm": 0.4147214889526367, + "learning_rate": 1.930414803628369e-05, + "loss": 0.8383, + "step": 873 + }, + { + "epoch": 0.14653672849208843, + "grad_norm": 0.6177957653999329, + "learning_rate": 1.930215631074019e-05, + "loss": 0.9271, + "step": 874 + }, + { + "epoch": 0.14670439065283453, + "grad_norm": 0.7204621434211731, + "learning_rate": 1.930016184187561e-05, + "loss": 0.9782, + "step": 875 + }, + { + "epoch": 0.14687205281358062, + "grad_norm": 0.5664174556732178, + "learning_rate": 1.9298164630278137e-05, + "loss": 0.9013, + "step": 876 + }, + { + "epoch": 0.14703971497432672, + "grad_norm": 0.5182133316993713, + "learning_rate": 1.929616467653678e-05, + "loss": 0.9522, + "step": 877 + }, + { + "epoch": 0.14720737713507281, + "grad_norm": 0.5646714568138123, + "learning_rate": 1.929416198124135e-05, + "loss": 0.8977, + "step": 878 + }, + { + "epoch": 0.14737503929581894, + "grad_norm": 0.4593978524208069, + "learning_rate": 1.9292156544982456e-05, + "loss": 0.8403, + "step": 879 + }, + { + "epoch": 0.14754270145656503, + "grad_norm": 0.49368441104888916, + "learning_rate": 1.929014836835154e-05, + "loss": 0.957, + "step": 880 + }, + { + "epoch": 0.14771036361731113, + "grad_norm": 0.7045102715492249, + "learning_rate": 1.928813745194083e-05, + "loss": 0.9834, + "step": 881 + }, + { + "epoch": 0.14787802577805723, + "grad_norm": 0.5111392140388489, + "learning_rate": 1.9286123796343373e-05, + "loss": 0.8888, + "step": 882 + }, + { + "epoch": 0.14804568793880332, + "grad_norm": 0.3912787437438965, + "learning_rate": 1.9284107402153022e-05, + "loss": 0.8511, + "step": 883 + }, + { + "epoch": 0.14821335009954942, + "grad_norm": 0.665562093257904, + "learning_rate": 1.928208826996443e-05, + "loss": 1.0157, + "step": 884 + }, + { + "epoch": 0.14838101226029551, + "grad_norm": 0.653226375579834, + "learning_rate": 1.9280066400373073e-05, + "loss": 0.9063, + "step": 885 + }, + { + "epoch": 0.1485486744210416, + "grad_norm": 0.42588213086128235, + "learning_rate": 1.9278041793975216e-05, + "loss": 1.0176, + "step": 886 + }, + { + "epoch": 0.1487163365817877, + "grad_norm": 0.465458482503891, + "learning_rate": 1.927601445136795e-05, + "loss": 0.9025, + "step": 887 + }, + { + "epoch": 0.1488839987425338, + "grad_norm": 0.505429208278656, + "learning_rate": 1.927398437314916e-05, + "loss": 0.9376, + "step": 888 + }, + { + "epoch": 0.1490516609032799, + "grad_norm": 0.43974682688713074, + "learning_rate": 1.9271951559917537e-05, + "loss": 1.0323, + "step": 889 + }, + { + "epoch": 0.149219323064026, + "grad_norm": 0.5050548911094666, + "learning_rate": 1.9269916012272586e-05, + "loss": 0.9432, + "step": 890 + }, + { + "epoch": 0.1493869852247721, + "grad_norm": 0.5680622458457947, + "learning_rate": 1.9267877730814615e-05, + "loss": 0.9627, + "step": 891 + }, + { + "epoch": 0.14955464738551819, + "grad_norm": 0.43889379501342773, + "learning_rate": 1.9265836716144744e-05, + "loss": 0.9195, + "step": 892 + }, + { + "epoch": 0.14972230954626428, + "grad_norm": 0.4966212511062622, + "learning_rate": 1.9263792968864882e-05, + "loss": 0.9697, + "step": 893 + }, + { + "epoch": 0.14988997170701038, + "grad_norm": 0.44300034642219543, + "learning_rate": 1.9261746489577767e-05, + "loss": 1.0088, + "step": 894 + }, + { + "epoch": 0.15005763386775647, + "grad_norm": 0.6189342141151428, + "learning_rate": 1.9259697278886926e-05, + "loss": 0.9452, + "step": 895 + }, + { + "epoch": 0.15022529602850257, + "grad_norm": 0.5580673813819885, + "learning_rate": 1.9257645337396692e-05, + "loss": 0.8635, + "step": 896 + }, + { + "epoch": 0.15039295818924867, + "grad_norm": 0.47836774587631226, + "learning_rate": 1.9255590665712214e-05, + "loss": 0.9098, + "step": 897 + }, + { + "epoch": 0.15056062034999476, + "grad_norm": 0.5996960997581482, + "learning_rate": 1.9253533264439436e-05, + "loss": 0.9834, + "step": 898 + }, + { + "epoch": 0.15072828251074086, + "grad_norm": 0.4515504240989685, + "learning_rate": 1.9251473134185116e-05, + "loss": 0.9444, + "step": 899 + }, + { + "epoch": 0.15089594467148695, + "grad_norm": 0.41490310430526733, + "learning_rate": 1.924941027555681e-05, + "loss": 1.0415, + "step": 900 + }, + { + "epoch": 0.15106360683223305, + "grad_norm": 0.40346625447273254, + "learning_rate": 1.924734468916288e-05, + "loss": 0.903, + "step": 901 + }, + { + "epoch": 0.15123126899297915, + "grad_norm": 0.5119394063949585, + "learning_rate": 1.9245276375612493e-05, + "loss": 0.8893, + "step": 902 + }, + { + "epoch": 0.15139893115372524, + "grad_norm": 0.4831201136112213, + "learning_rate": 1.924320533551562e-05, + "loss": 1.0228, + "step": 903 + }, + { + "epoch": 0.15156659331447134, + "grad_norm": 0.4201967418193817, + "learning_rate": 1.9241131569483035e-05, + "loss": 0.893, + "step": 904 + }, + { + "epoch": 0.15173425547521743, + "grad_norm": 0.5925395488739014, + "learning_rate": 1.923905507812632e-05, + "loss": 0.9479, + "step": 905 + }, + { + "epoch": 0.15190191763596353, + "grad_norm": 0.45901793241500854, + "learning_rate": 1.923697586205786e-05, + "loss": 0.861, + "step": 906 + }, + { + "epoch": 0.15206957979670963, + "grad_norm": 0.510019063949585, + "learning_rate": 1.9234893921890834e-05, + "loss": 0.8774, + "step": 907 + }, + { + "epoch": 0.15223724195745572, + "grad_norm": 0.6654001474380493, + "learning_rate": 1.923280925823924e-05, + "loss": 0.9092, + "step": 908 + }, + { + "epoch": 0.15240490411820182, + "grad_norm": 0.49889883399009705, + "learning_rate": 1.9230721871717862e-05, + "loss": 0.8707, + "step": 909 + }, + { + "epoch": 0.1525725662789479, + "grad_norm": 0.5877993106842041, + "learning_rate": 1.9228631762942307e-05, + "loss": 1.0557, + "step": 910 + }, + { + "epoch": 0.152740228439694, + "grad_norm": 0.3399173319339752, + "learning_rate": 1.9226538932528967e-05, + "loss": 0.9035, + "step": 911 + }, + { + "epoch": 0.1529078906004401, + "grad_norm": 0.3397526741027832, + "learning_rate": 1.9224443381095047e-05, + "loss": 1.0377, + "step": 912 + }, + { + "epoch": 0.1530755527611862, + "grad_norm": 0.5500420928001404, + "learning_rate": 1.9222345109258546e-05, + "loss": 0.9871, + "step": 913 + }, + { + "epoch": 0.1532432149219323, + "grad_norm": 0.6797765493392944, + "learning_rate": 1.9220244117638277e-05, + "loss": 0.8941, + "step": 914 + }, + { + "epoch": 0.1534108770826784, + "grad_norm": 0.3991592526435852, + "learning_rate": 1.9218140406853846e-05, + "loss": 0.9538, + "step": 915 + }, + { + "epoch": 0.1535785392434245, + "grad_norm": 0.43551596999168396, + "learning_rate": 1.9216033977525658e-05, + "loss": 0.9516, + "step": 916 + }, + { + "epoch": 0.15374620140417058, + "grad_norm": 0.4320926070213318, + "learning_rate": 1.9213924830274934e-05, + "loss": 0.8031, + "step": 917 + }, + { + "epoch": 0.15391386356491668, + "grad_norm": 0.3962501287460327, + "learning_rate": 1.921181296572368e-05, + "loss": 1.0334, + "step": 918 + }, + { + "epoch": 0.15408152572566278, + "grad_norm": 0.484610378742218, + "learning_rate": 1.9209698384494714e-05, + "loss": 0.9991, + "step": 919 + }, + { + "epoch": 0.15424918788640887, + "grad_norm": 0.5024212598800659, + "learning_rate": 1.9207581087211654e-05, + "loss": 0.8489, + "step": 920 + }, + { + "epoch": 0.154416850047155, + "grad_norm": 0.4835463762283325, + "learning_rate": 1.9205461074498917e-05, + "loss": 0.905, + "step": 921 + }, + { + "epoch": 0.1545845122079011, + "grad_norm": 0.3671019673347473, + "learning_rate": 1.9203338346981718e-05, + "loss": 0.9507, + "step": 922 + }, + { + "epoch": 0.1547521743686472, + "grad_norm": 0.7113335132598877, + "learning_rate": 1.9201212905286074e-05, + "loss": 0.8627, + "step": 923 + }, + { + "epoch": 0.15491983652939328, + "grad_norm": 0.4007314145565033, + "learning_rate": 1.919908475003881e-05, + "loss": 0.982, + "step": 924 + }, + { + "epoch": 0.15508749869013938, + "grad_norm": 0.48794025182724, + "learning_rate": 1.9196953881867544e-05, + "loss": 1.0267, + "step": 925 + }, + { + "epoch": 0.15525516085088548, + "grad_norm": 0.7895486950874329, + "learning_rate": 1.9194820301400695e-05, + "loss": 0.9659, + "step": 926 + }, + { + "epoch": 0.15542282301163157, + "grad_norm": 0.6803207397460938, + "learning_rate": 1.919268400926748e-05, + "loss": 0.9211, + "step": 927 + }, + { + "epoch": 0.15559048517237767, + "grad_norm": 0.41591545939445496, + "learning_rate": 1.9190545006097917e-05, + "loss": 0.9046, + "step": 928 + }, + { + "epoch": 0.15575814733312376, + "grad_norm": 0.5918329954147339, + "learning_rate": 1.918840329252283e-05, + "loss": 0.9703, + "step": 929 + }, + { + "epoch": 0.15592580949386986, + "grad_norm": 0.6053893566131592, + "learning_rate": 1.9186258869173834e-05, + "loss": 0.9772, + "step": 930 + }, + { + "epoch": 0.15609347165461596, + "grad_norm": 0.47089946269989014, + "learning_rate": 1.9184111736683343e-05, + "loss": 0.9331, + "step": 931 + }, + { + "epoch": 0.15626113381536205, + "grad_norm": 0.4194715619087219, + "learning_rate": 1.918196189568458e-05, + "loss": 0.9214, + "step": 932 + }, + { + "epoch": 0.15642879597610815, + "grad_norm": 0.3285825252532959, + "learning_rate": 1.917980934681155e-05, + "loss": 0.9413, + "step": 933 + }, + { + "epoch": 0.15659645813685424, + "grad_norm": 0.43966519832611084, + "learning_rate": 1.9177654090699074e-05, + "loss": 0.8261, + "step": 934 + }, + { + "epoch": 0.15676412029760034, + "grad_norm": 0.42831242084503174, + "learning_rate": 1.9175496127982764e-05, + "loss": 0.9752, + "step": 935 + }, + { + "epoch": 0.15693178245834644, + "grad_norm": 0.34641575813293457, + "learning_rate": 1.9173335459299025e-05, + "loss": 0.9415, + "step": 936 + }, + { + "epoch": 0.15709944461909253, + "grad_norm": 0.30951234698295593, + "learning_rate": 1.9171172085285064e-05, + "loss": 0.8722, + "step": 937 + }, + { + "epoch": 0.15726710677983863, + "grad_norm": 0.4109880328178406, + "learning_rate": 1.9169006006578897e-05, + "loss": 0.942, + "step": 938 + }, + { + "epoch": 0.15743476894058472, + "grad_norm": 0.34938591718673706, + "learning_rate": 1.9166837223819317e-05, + "loss": 0.8811, + "step": 939 + }, + { + "epoch": 0.15760243110133082, + "grad_norm": 0.42489537596702576, + "learning_rate": 1.9164665737645927e-05, + "loss": 0.9891, + "step": 940 + }, + { + "epoch": 0.15777009326207692, + "grad_norm": 0.43413758277893066, + "learning_rate": 1.9162491548699128e-05, + "loss": 0.8942, + "step": 941 + }, + { + "epoch": 0.157937755422823, + "grad_norm": 0.5286901593208313, + "learning_rate": 1.9160314657620116e-05, + "loss": 1.0414, + "step": 942 + }, + { + "epoch": 0.1581054175835691, + "grad_norm": 0.4752057194709778, + "learning_rate": 1.9158135065050878e-05, + "loss": 0.8457, + "step": 943 + }, + { + "epoch": 0.1582730797443152, + "grad_norm": 0.4286821186542511, + "learning_rate": 1.9155952771634207e-05, + "loss": 0.9238, + "step": 944 + }, + { + "epoch": 0.1584407419050613, + "grad_norm": 0.5006504654884338, + "learning_rate": 1.9153767778013686e-05, + "loss": 0.9863, + "step": 945 + }, + { + "epoch": 0.1586084040658074, + "grad_norm": 0.9059145450592041, + "learning_rate": 1.91515800848337e-05, + "loss": 0.9411, + "step": 946 + }, + { + "epoch": 0.1587760662265535, + "grad_norm": 0.663236141204834, + "learning_rate": 1.9149389692739424e-05, + "loss": 0.8345, + "step": 947 + }, + { + "epoch": 0.1589437283872996, + "grad_norm": 0.4279334843158722, + "learning_rate": 1.9147196602376833e-05, + "loss": 0.9397, + "step": 948 + }, + { + "epoch": 0.15911139054804568, + "grad_norm": 0.5905395746231079, + "learning_rate": 1.9145000814392696e-05, + "loss": 0.9753, + "step": 949 + }, + { + "epoch": 0.15927905270879178, + "grad_norm": 0.5215145945549011, + "learning_rate": 1.9142802329434574e-05, + "loss": 0.8985, + "step": 950 + }, + { + "epoch": 0.15944671486953788, + "grad_norm": 0.5803624987602234, + "learning_rate": 1.9140601148150833e-05, + "loss": 0.9656, + "step": 951 + }, + { + "epoch": 0.15961437703028397, + "grad_norm": 0.4757272005081177, + "learning_rate": 1.913839727119063e-05, + "loss": 0.8953, + "step": 952 + }, + { + "epoch": 0.15978203919103007, + "grad_norm": 0.4840550720691681, + "learning_rate": 1.9136190699203913e-05, + "loss": 0.9365, + "step": 953 + }, + { + "epoch": 0.15994970135177616, + "grad_norm": 0.4999079704284668, + "learning_rate": 1.9133981432841424e-05, + "loss": 0.9202, + "step": 954 + }, + { + "epoch": 0.16011736351252226, + "grad_norm": 0.5819324851036072, + "learning_rate": 1.913176947275471e-05, + "loss": 0.9156, + "step": 955 + }, + { + "epoch": 0.16028502567326836, + "grad_norm": 0.6553800106048584, + "learning_rate": 1.9129554819596098e-05, + "loss": 0.9624, + "step": 956 + }, + { + "epoch": 0.16045268783401445, + "grad_norm": 0.4883859157562256, + "learning_rate": 1.912733747401872e-05, + "loss": 0.9026, + "step": 957 + }, + { + "epoch": 0.16062034999476055, + "grad_norm": 0.4642502963542938, + "learning_rate": 1.91251174366765e-05, + "loss": 0.9512, + "step": 958 + }, + { + "epoch": 0.16078801215550664, + "grad_norm": 0.38176947832107544, + "learning_rate": 1.9122894708224152e-05, + "loss": 0.8772, + "step": 959 + }, + { + "epoch": 0.16095567431625274, + "grad_norm": 0.4411686062812805, + "learning_rate": 1.912066928931719e-05, + "loss": 0.8728, + "step": 960 + }, + { + "epoch": 0.16112333647699884, + "grad_norm": 0.6528915762901306, + "learning_rate": 1.9118441180611912e-05, + "loss": 0.9028, + "step": 961 + }, + { + "epoch": 0.16129099863774493, + "grad_norm": 0.4659099578857422, + "learning_rate": 1.911621038276542e-05, + "loss": 0.9866, + "step": 962 + }, + { + "epoch": 0.16145866079849103, + "grad_norm": 0.6679427623748779, + "learning_rate": 1.9113976896435603e-05, + "loss": 0.8638, + "step": 963 + }, + { + "epoch": 0.16162632295923715, + "grad_norm": 0.48949748277664185, + "learning_rate": 1.9111740722281137e-05, + "loss": 1.0119, + "step": 964 + }, + { + "epoch": 0.16179398511998325, + "grad_norm": 0.5931878685951233, + "learning_rate": 1.910950186096151e-05, + "loss": 1.075, + "step": 965 + }, + { + "epoch": 0.16196164728072934, + "grad_norm": 0.6350274085998535, + "learning_rate": 1.910726031313698e-05, + "loss": 0.9303, + "step": 966 + }, + { + "epoch": 0.16212930944147544, + "grad_norm": 0.41805300116539, + "learning_rate": 1.910501607946861e-05, + "loss": 0.9238, + "step": 967 + }, + { + "epoch": 0.16229697160222153, + "grad_norm": 0.5107613801956177, + "learning_rate": 1.910276916061825e-05, + "loss": 0.9229, + "step": 968 + }, + { + "epoch": 0.16246463376296763, + "grad_norm": 0.44543734192848206, + "learning_rate": 1.910051955724855e-05, + "loss": 0.9944, + "step": 969 + }, + { + "epoch": 0.16263229592371373, + "grad_norm": 0.5454578399658203, + "learning_rate": 1.9098267270022943e-05, + "loss": 0.9248, + "step": 970 + }, + { + "epoch": 0.16279995808445982, + "grad_norm": 0.5694398283958435, + "learning_rate": 1.9096012299605654e-05, + "loss": 0.9687, + "step": 971 + }, + { + "epoch": 0.16296762024520592, + "grad_norm": 0.5694398283958435, + "learning_rate": 1.9096012299605654e-05, + "loss": 0.9867, + "step": 972 + }, + { + "epoch": 0.16313528240595201, + "grad_norm": 0.49293121695518494, + "learning_rate": 1.90937546466617e-05, + "loss": 0.8689, + "step": 973 + }, + { + "epoch": 0.1633029445666981, + "grad_norm": 0.545941948890686, + "learning_rate": 1.9091494311856903e-05, + "loss": 0.9003, + "step": 974 + }, + { + "epoch": 0.1634706067274442, + "grad_norm": 0.47112664580345154, + "learning_rate": 1.9089231295857846e-05, + "loss": 1.0334, + "step": 975 + }, + { + "epoch": 0.1636382688881903, + "grad_norm": 0.49594390392303467, + "learning_rate": 1.9086965599331938e-05, + "loss": 0.8357, + "step": 976 + }, + { + "epoch": 0.1638059310489364, + "grad_norm": 0.655060887336731, + "learning_rate": 1.9084697222947344e-05, + "loss": 0.9273, + "step": 977 + }, + { + "epoch": 0.1639735932096825, + "grad_norm": 0.682763934135437, + "learning_rate": 1.9082426167373053e-05, + "loss": 0.9753, + "step": 978 + }, + { + "epoch": 0.1641412553704286, + "grad_norm": 0.5177135467529297, + "learning_rate": 1.9080152433278816e-05, + "loss": 0.9426, + "step": 979 + }, + { + "epoch": 0.1643089175311747, + "grad_norm": 0.6700261831283569, + "learning_rate": 1.9077876021335188e-05, + "loss": 1.0105, + "step": 980 + }, + { + "epoch": 0.16447657969192078, + "grad_norm": 0.4972025752067566, + "learning_rate": 1.907559693221351e-05, + "loss": 0.9862, + "step": 981 + }, + { + "epoch": 0.16464424185266688, + "grad_norm": 0.6664692759513855, + "learning_rate": 1.9073315166585916e-05, + "loss": 0.9289, + "step": 982 + }, + { + "epoch": 0.16481190401341297, + "grad_norm": 0.6527049541473389, + "learning_rate": 1.907103072512533e-05, + "loss": 0.923, + "step": 983 + }, + { + "epoch": 0.16497956617415907, + "grad_norm": 0.4614576995372772, + "learning_rate": 1.9068743608505454e-05, + "loss": 0.9671, + "step": 984 + }, + { + "epoch": 0.16514722833490517, + "grad_norm": 0.43237075209617615, + "learning_rate": 1.9066453817400793e-05, + "loss": 0.9164, + "step": 985 + }, + { + "epoch": 0.16531489049565126, + "grad_norm": 0.4315905272960663, + "learning_rate": 1.9064161352486638e-05, + "loss": 0.9128, + "step": 986 + }, + { + "epoch": 0.16548255265639736, + "grad_norm": 0.5184720158576965, + "learning_rate": 1.9061866214439054e-05, + "loss": 0.8276, + "step": 987 + }, + { + "epoch": 0.16565021481714345, + "grad_norm": 0.687303364276886, + "learning_rate": 1.905956840393491e-05, + "loss": 0.9414, + "step": 988 + }, + { + "epoch": 0.16581787697788955, + "grad_norm": 0.472570538520813, + "learning_rate": 1.9057267921651865e-05, + "loss": 1.0119, + "step": 989 + }, + { + "epoch": 0.16598553913863565, + "grad_norm": 0.5595750212669373, + "learning_rate": 1.9054964768268354e-05, + "loss": 0.9516, + "step": 990 + }, + { + "epoch": 0.16615320129938174, + "grad_norm": 0.4138607084751129, + "learning_rate": 1.905265894446361e-05, + "loss": 0.9961, + "step": 991 + }, + { + "epoch": 0.16632086346012784, + "grad_norm": 0.5740602016448975, + "learning_rate": 1.905035045091764e-05, + "loss": 0.9427, + "step": 992 + }, + { + "epoch": 0.16648852562087393, + "grad_norm": 0.7034726142883301, + "learning_rate": 1.9048039288311256e-05, + "loss": 1.0249, + "step": 993 + }, + { + "epoch": 0.16665618778162003, + "grad_norm": 0.567058265209198, + "learning_rate": 1.904572545732605e-05, + "loss": 0.9528, + "step": 994 + }, + { + "epoch": 0.16682384994236613, + "grad_norm": 0.43043065071105957, + "learning_rate": 1.904340895864439e-05, + "loss": 0.9099, + "step": 995 + }, + { + "epoch": 0.16699151210311222, + "grad_norm": 0.49964940547943115, + "learning_rate": 1.9041089792949446e-05, + "loss": 0.9711, + "step": 996 + }, + { + "epoch": 0.16715917426385832, + "grad_norm": 0.7605466246604919, + "learning_rate": 1.9038767960925168e-05, + "loss": 0.9072, + "step": 997 + }, + { + "epoch": 0.1673268364246044, + "grad_norm": 0.5706338882446289, + "learning_rate": 1.9036443463256295e-05, + "loss": 0.9962, + "step": 998 + }, + { + "epoch": 0.1674944985853505, + "grad_norm": 0.7213522791862488, + "learning_rate": 1.9034116300628346e-05, + "loss": 0.9282, + "step": 999 + }, + { + "epoch": 0.1676621607460966, + "grad_norm": 0.5890035629272461, + "learning_rate": 1.9031786473727633e-05, + "loss": 0.9342, + "step": 1000 + }, + { + "epoch": 0.1678298229068427, + "grad_norm": 0.5635486841201782, + "learning_rate": 1.902945398324125e-05, + "loss": 0.9039, + "step": 1001 + }, + { + "epoch": 0.1679974850675888, + "grad_norm": 0.5275377035140991, + "learning_rate": 1.902711882985708e-05, + "loss": 0.9226, + "step": 1002 + }, + { + "epoch": 0.1681651472283349, + "grad_norm": 0.8132424354553223, + "learning_rate": 1.902478101426378e-05, + "loss": 0.8846, + "step": 1003 + }, + { + "epoch": 0.168332809389081, + "grad_norm": 0.4180213510990143, + "learning_rate": 1.902244053715081e-05, + "loss": 0.9278, + "step": 1004 + }, + { + "epoch": 0.16850047154982709, + "grad_norm": 0.520513117313385, + "learning_rate": 1.9020097399208403e-05, + "loss": 1.0251, + "step": 1005 + }, + { + "epoch": 0.16866813371057318, + "grad_norm": 0.6166831851005554, + "learning_rate": 1.9017751601127575e-05, + "loss": 0.9955, + "step": 1006 + }, + { + "epoch": 0.1688357958713193, + "grad_norm": 0.6414588689804077, + "learning_rate": 1.901540314360014e-05, + "loss": 0.8512, + "step": 1007 + }, + { + "epoch": 0.1690034580320654, + "grad_norm": 0.5270276069641113, + "learning_rate": 1.9013052027318677e-05, + "loss": 0.951, + "step": 1008 + }, + { + "epoch": 0.1691711201928115, + "grad_norm": 0.4635254740715027, + "learning_rate": 1.9010698252976566e-05, + "loss": 0.8338, + "step": 1009 + }, + { + "epoch": 0.1693387823535576, + "grad_norm": 0.4911115765571594, + "learning_rate": 1.9008341821267965e-05, + "loss": 0.8872, + "step": 1010 + }, + { + "epoch": 0.1695064445143037, + "grad_norm": 0.45328113436698914, + "learning_rate": 1.9005982732887812e-05, + "loss": 0.8935, + "step": 1011 + }, + { + "epoch": 0.16967410667504979, + "grad_norm": 0.6181873083114624, + "learning_rate": 1.9003620988531827e-05, + "loss": 0.9842, + "step": 1012 + }, + { + "epoch": 0.16984176883579588, + "grad_norm": 0.7268185615539551, + "learning_rate": 1.9001256588896526e-05, + "loss": 0.8971, + "step": 1013 + }, + { + "epoch": 0.17000943099654198, + "grad_norm": 0.6407479047775269, + "learning_rate": 1.8998889534679196e-05, + "loss": 0.8073, + "step": 1014 + }, + { + "epoch": 0.17017709315728807, + "grad_norm": 0.6562261581420898, + "learning_rate": 1.8996519826577907e-05, + "loss": 0.9595, + "step": 1015 + }, + { + "epoch": 0.17034475531803417, + "grad_norm": 0.44220224022865295, + "learning_rate": 1.8994147465291526e-05, + "loss": 0.9315, + "step": 1016 + }, + { + "epoch": 0.17051241747878026, + "grad_norm": 0.4324677288532257, + "learning_rate": 1.8991772451519678e-05, + "loss": 0.8583, + "step": 1017 + }, + { + "epoch": 0.17068007963952636, + "grad_norm": 0.5509170889854431, + "learning_rate": 1.8989394785962795e-05, + "loss": 0.9205, + "step": 1018 + }, + { + "epoch": 0.17084774180027246, + "grad_norm": 0.5897940397262573, + "learning_rate": 1.8987014469322075e-05, + "loss": 0.8807, + "step": 1019 + }, + { + "epoch": 0.17101540396101855, + "grad_norm": 0.6917343735694885, + "learning_rate": 1.89846315022995e-05, + "loss": 1.0349, + "step": 1020 + }, + { + "epoch": 0.17118306612176465, + "grad_norm": 0.44609513878822327, + "learning_rate": 1.898224588559784e-05, + "loss": 0.8867, + "step": 1021 + }, + { + "epoch": 0.17135072828251074, + "grad_norm": 0.6210997700691223, + "learning_rate": 1.8979857619920646e-05, + "loss": 0.9736, + "step": 1022 + }, + { + "epoch": 0.17151839044325684, + "grad_norm": 0.5329290628433228, + "learning_rate": 1.8977466705972244e-05, + "loss": 0.9225, + "step": 1023 + }, + { + "epoch": 0.17168605260400294, + "grad_norm": 0.5414072871208191, + "learning_rate": 1.8975073144457742e-05, + "loss": 0.9704, + "step": 1024 + }, + { + "epoch": 0.17185371476474903, + "grad_norm": 0.42674532532691956, + "learning_rate": 1.8972676936083035e-05, + "loss": 0.8851, + "step": 1025 + }, + { + "epoch": 0.17202137692549513, + "grad_norm": 0.7599685788154602, + "learning_rate": 1.8970278081554792e-05, + "loss": 0.9601, + "step": 1026 + }, + { + "epoch": 0.17218903908624122, + "grad_norm": 0.41676852107048035, + "learning_rate": 1.8967876581580468e-05, + "loss": 0.9191, + "step": 1027 + }, + { + "epoch": 0.17235670124698732, + "grad_norm": 0.8608080148696899, + "learning_rate": 1.8965472436868288e-05, + "loss": 0.8661, + "step": 1028 + }, + { + "epoch": 0.17252436340773342, + "grad_norm": 0.6654183864593506, + "learning_rate": 1.896306564812727e-05, + "loss": 0.9127, + "step": 1029 + }, + { + "epoch": 0.1726920255684795, + "grad_norm": 0.6694945693016052, + "learning_rate": 1.8960656216067204e-05, + "loss": 1.0001, + "step": 1030 + }, + { + "epoch": 0.1728596877292256, + "grad_norm": 0.7841224074363708, + "learning_rate": 1.8958244141398664e-05, + "loss": 0.9673, + "step": 1031 + }, + { + "epoch": 0.1730273498899717, + "grad_norm": 0.47580161690711975, + "learning_rate": 1.8955829424832995e-05, + "loss": 0.928, + "step": 1032 + }, + { + "epoch": 0.1731950120507178, + "grad_norm": 0.7303934097290039, + "learning_rate": 1.895341206708233e-05, + "loss": 0.8929, + "step": 1033 + }, + { + "epoch": 0.1733626742114639, + "grad_norm": 0.8481614589691162, + "learning_rate": 1.8950992068859584e-05, + "loss": 0.9073, + "step": 1034 + }, + { + "epoch": 0.17353033637221, + "grad_norm": 0.632799506187439, + "learning_rate": 1.8948569430878435e-05, + "loss": 1.0129, + "step": 1035 + }, + { + "epoch": 0.1736979985329561, + "grad_norm": 0.7045972347259521, + "learning_rate": 1.894614415385335e-05, + "loss": 0.9188, + "step": 1036 + }, + { + "epoch": 0.17386566069370218, + "grad_norm": 0.486174613237381, + "learning_rate": 1.8943716238499575e-05, + "loss": 1.0025, + "step": 1037 + }, + { + "epoch": 0.17403332285444828, + "grad_norm": 0.3715253174304962, + "learning_rate": 1.8941285685533136e-05, + "loss": 0.9962, + "step": 1038 + }, + { + "epoch": 0.17420098501519438, + "grad_norm": 0.38294509053230286, + "learning_rate": 1.893885249567083e-05, + "loss": 0.8867, + "step": 1039 + }, + { + "epoch": 0.17436864717594047, + "grad_norm": 0.5372864603996277, + "learning_rate": 1.8936416669630236e-05, + "loss": 0.9098, + "step": 1040 + }, + { + "epoch": 0.17453630933668657, + "grad_norm": 0.42620623111724854, + "learning_rate": 1.8933978208129705e-05, + "loss": 0.9706, + "step": 1041 + }, + { + "epoch": 0.17470397149743266, + "grad_norm": 0.4851018488407135, + "learning_rate": 1.8931537111888378e-05, + "loss": 1.0786, + "step": 1042 + }, + { + "epoch": 0.17487163365817876, + "grad_norm": 0.4234909117221832, + "learning_rate": 1.8929093381626153e-05, + "loss": 1.0391, + "step": 1043 + }, + { + "epoch": 0.17503929581892486, + "grad_norm": 0.5287904739379883, + "learning_rate": 1.8926647018063724e-05, + "loss": 0.9263, + "step": 1044 + }, + { + "epoch": 0.17520695797967095, + "grad_norm": 0.5778496861457825, + "learning_rate": 1.8924198021922553e-05, + "loss": 0.9885, + "step": 1045 + }, + { + "epoch": 0.17537462014041705, + "grad_norm": 0.4014924466609955, + "learning_rate": 1.8921746393924877e-05, + "loss": 0.8806, + "step": 1046 + }, + { + "epoch": 0.17554228230116314, + "grad_norm": 0.5376038551330566, + "learning_rate": 1.8919292134793718e-05, + "loss": 0.9444, + "step": 1047 + }, + { + "epoch": 0.17570994446190924, + "grad_norm": 0.43011462688446045, + "learning_rate": 1.8916835245252854e-05, + "loss": 0.9794, + "step": 1048 + }, + { + "epoch": 0.17587760662265534, + "grad_norm": 0.38411054015159607, + "learning_rate": 1.891437572602687e-05, + "loss": 0.9942, + "step": 1049 + }, + { + "epoch": 0.17604526878340146, + "grad_norm": 0.46081578731536865, + "learning_rate": 1.8911913577841096e-05, + "loss": 0.9074, + "step": 1050 + }, + { + "epoch": 0.17621293094414756, + "grad_norm": 0.3966897428035736, + "learning_rate": 1.8909448801421653e-05, + "loss": 0.9153, + "step": 1051 + }, + { + "epoch": 0.17638059310489365, + "grad_norm": 0.40592241287231445, + "learning_rate": 1.8906981397495434e-05, + "loss": 0.9857, + "step": 1052 + }, + { + "epoch": 0.17654825526563975, + "grad_norm": 0.47412779927253723, + "learning_rate": 1.8904511366790106e-05, + "loss": 0.959, + "step": 1053 + }, + { + "epoch": 0.17671591742638584, + "grad_norm": 0.40784403681755066, + "learning_rate": 1.8902038710034113e-05, + "loss": 0.9044, + "step": 1054 + }, + { + "epoch": 0.17688357958713194, + "grad_norm": 0.416376531124115, + "learning_rate": 1.8899563427956674e-05, + "loss": 0.9814, + "step": 1055 + }, + { + "epoch": 0.17705124174787804, + "grad_norm": 0.5241764187812805, + "learning_rate": 1.889708552128778e-05, + "loss": 0.9262, + "step": 1056 + }, + { + "epoch": 0.17721890390862413, + "grad_norm": 0.467724084854126, + "learning_rate": 1.889460499075819e-05, + "loss": 0.9341, + "step": 1057 + }, + { + "epoch": 0.17738656606937023, + "grad_norm": 0.4077341854572296, + "learning_rate": 1.8892121837099452e-05, + "loss": 0.9989, + "step": 1058 + }, + { + "epoch": 0.17755422823011632, + "grad_norm": 0.7256616950035095, + "learning_rate": 1.8889636061043874e-05, + "loss": 0.9983, + "step": 1059 + }, + { + "epoch": 0.17772189039086242, + "grad_norm": 0.49016696214675903, + "learning_rate": 1.8887147663324545e-05, + "loss": 0.8761, + "step": 1060 + }, + { + "epoch": 0.17788955255160852, + "grad_norm": 0.6254734396934509, + "learning_rate": 1.8884656644675318e-05, + "loss": 1.001, + "step": 1061 + }, + { + "epoch": 0.1780572147123546, + "grad_norm": 0.5470951199531555, + "learning_rate": 1.888216300583083e-05, + "loss": 0.8073, + "step": 1062 + }, + { + "epoch": 0.1782248768731007, + "grad_norm": 0.49495676159858704, + "learning_rate": 1.887966674752649e-05, + "loss": 0.9055, + "step": 1063 + }, + { + "epoch": 0.1783925390338468, + "grad_norm": 0.4782699644565582, + "learning_rate": 1.8877167870498466e-05, + "loss": 1.0146, + "step": 1064 + }, + { + "epoch": 0.1785602011945929, + "grad_norm": 0.8720697164535522, + "learning_rate": 1.8874666375483713e-05, + "loss": 0.8812, + "step": 1065 + }, + { + "epoch": 0.178727863355339, + "grad_norm": 0.5451815724372864, + "learning_rate": 1.8872162263219955e-05, + "loss": 0.9154, + "step": 1066 + }, + { + "epoch": 0.1788955255160851, + "grad_norm": 0.48717665672302246, + "learning_rate": 1.886965553444568e-05, + "loss": 0.9923, + "step": 1067 + }, + { + "epoch": 0.1790631876768312, + "grad_norm": 0.4174070358276367, + "learning_rate": 1.886714618990016e-05, + "loss": 0.9578, + "step": 1068 + }, + { + "epoch": 0.17923084983757728, + "grad_norm": 0.38399094343185425, + "learning_rate": 1.8864634230323426e-05, + "loss": 0.9129, + "step": 1069 + }, + { + "epoch": 0.17939851199832338, + "grad_norm": 0.6723623871803284, + "learning_rate": 1.886211965645629e-05, + "loss": 0.9536, + "step": 1070 + }, + { + "epoch": 0.17956617415906947, + "grad_norm": 0.5474948287010193, + "learning_rate": 1.8859602469040327e-05, + "loss": 0.9327, + "step": 1071 + }, + { + "epoch": 0.17973383631981557, + "grad_norm": 0.6205623149871826, + "learning_rate": 1.885708266881789e-05, + "loss": 0.9177, + "step": 1072 + }, + { + "epoch": 0.17990149848056167, + "grad_norm": 0.3894779086112976, + "learning_rate": 1.8854560256532098e-05, + "loss": 0.9197, + "step": 1073 + }, + { + "epoch": 0.18006916064130776, + "grad_norm": 0.7599223256111145, + "learning_rate": 1.8852035232926848e-05, + "loss": 0.8938, + "step": 1074 + }, + { + "epoch": 0.18023682280205386, + "grad_norm": 0.5655592679977417, + "learning_rate": 1.884950759874679e-05, + "loss": 0.8273, + "step": 1075 + }, + { + "epoch": 0.18040448496279995, + "grad_norm": 0.5135934948921204, + "learning_rate": 1.8846977354737356e-05, + "loss": 0.9494, + "step": 1076 + }, + { + "epoch": 0.18057214712354605, + "grad_norm": 0.7166675925254822, + "learning_rate": 1.8844444501644755e-05, + "loss": 0.8252, + "step": 1077 + }, + { + "epoch": 0.18073980928429215, + "grad_norm": 0.7380692958831787, + "learning_rate": 1.8841909040215947e-05, + "loss": 0.8823, + "step": 1078 + }, + { + "epoch": 0.18090747144503824, + "grad_norm": 0.6297410130500793, + "learning_rate": 1.883937097119868e-05, + "loss": 1.0003, + "step": 1079 + }, + { + "epoch": 0.18107513360578434, + "grad_norm": 0.48514869809150696, + "learning_rate": 1.883683029534145e-05, + "loss": 0.9187, + "step": 1080 + }, + { + "epoch": 0.18124279576653043, + "grad_norm": 0.586363673210144, + "learning_rate": 1.8834287013393547e-05, + "loss": 0.9495, + "step": 1081 + }, + { + "epoch": 0.18141045792727653, + "grad_norm": 0.5589828491210938, + "learning_rate": 1.883174112610501e-05, + "loss": 0.9236, + "step": 1082 + }, + { + "epoch": 0.18157812008802263, + "grad_norm": 0.5361699461936951, + "learning_rate": 1.882919263422665e-05, + "loss": 0.9043, + "step": 1083 + }, + { + "epoch": 0.18174578224876872, + "grad_norm": 0.48993849754333496, + "learning_rate": 1.8826641538510055e-05, + "loss": 0.9285, + "step": 1084 + }, + { + "epoch": 0.18191344440951482, + "grad_norm": 0.5430667996406555, + "learning_rate": 1.8824087839707572e-05, + "loss": 1.0018, + "step": 1085 + }, + { + "epoch": 0.18208110657026091, + "grad_norm": 0.5209212899208069, + "learning_rate": 1.8821531538572317e-05, + "loss": 0.9375, + "step": 1086 + }, + { + "epoch": 0.182248768731007, + "grad_norm": 0.4663066267967224, + "learning_rate": 1.8818972635858174e-05, + "loss": 0.9695, + "step": 1087 + }, + { + "epoch": 0.1824164308917531, + "grad_norm": 0.4286949634552002, + "learning_rate": 1.88164111323198e-05, + "loss": 0.9385, + "step": 1088 + }, + { + "epoch": 0.1825840930524992, + "grad_norm": 0.47834473848342896, + "learning_rate": 1.8813847028712612e-05, + "loss": 0.9313, + "step": 1089 + }, + { + "epoch": 0.1827517552132453, + "grad_norm": 0.49863681197166443, + "learning_rate": 1.8811280325792788e-05, + "loss": 1.0002, + "step": 1090 + }, + { + "epoch": 0.1829194173739914, + "grad_norm": 0.4201532304286957, + "learning_rate": 1.8808711024317295e-05, + "loss": 0.8969, + "step": 1091 + }, + { + "epoch": 0.1830870795347375, + "grad_norm": 0.4612746238708496, + "learning_rate": 1.880613912504384e-05, + "loss": 0.9271, + "step": 1092 + }, + { + "epoch": 0.18325474169548361, + "grad_norm": 0.4401395916938782, + "learning_rate": 1.8803564628730916e-05, + "loss": 0.8983, + "step": 1093 + }, + { + "epoch": 0.1834224038562297, + "grad_norm": 0.4388650059700012, + "learning_rate": 1.8800987536137767e-05, + "loss": 0.8806, + "step": 1094 + }, + { + "epoch": 0.1835900660169758, + "grad_norm": 0.6108338832855225, + "learning_rate": 1.8798407848024415e-05, + "loss": 0.9898, + "step": 1095 + }, + { + "epoch": 0.1837577281777219, + "grad_norm": 0.5061869621276855, + "learning_rate": 1.8795825565151637e-05, + "loss": 0.8857, + "step": 1096 + }, + { + "epoch": 0.183925390338468, + "grad_norm": 0.44790008664131165, + "learning_rate": 1.8793240688280986e-05, + "loss": 0.9221, + "step": 1097 + }, + { + "epoch": 0.1840930524992141, + "grad_norm": 0.3312665820121765, + "learning_rate": 1.8790653218174766e-05, + "loss": 0.9456, + "step": 1098 + }, + { + "epoch": 0.1842607146599602, + "grad_norm": 0.44701552391052246, + "learning_rate": 1.8788063155596062e-05, + "loss": 0.8792, + "step": 1099 + }, + { + "epoch": 0.18442837682070629, + "grad_norm": 0.4169693887233734, + "learning_rate": 1.8785470501308714e-05, + "loss": 0.893, + "step": 1100 + }, + { + "epoch": 0.18459603898145238, + "grad_norm": 0.4645668566226959, + "learning_rate": 1.8782875256077322e-05, + "loss": 0.9022, + "step": 1101 + }, + { + "epoch": 0.18476370114219848, + "grad_norm": 0.39160627126693726, + "learning_rate": 1.8780277420667262e-05, + "loss": 0.9049, + "step": 1102 + }, + { + "epoch": 0.18493136330294457, + "grad_norm": 0.5148125886917114, + "learning_rate": 1.877767699584467e-05, + "loss": 0.9455, + "step": 1103 + }, + { + "epoch": 0.18509902546369067, + "grad_norm": 0.3958742022514343, + "learning_rate": 1.8775073982376434e-05, + "loss": 0.8458, + "step": 1104 + }, + { + "epoch": 0.18526668762443677, + "grad_norm": 0.5639967918395996, + "learning_rate": 1.8772468381030222e-05, + "loss": 0.8476, + "step": 1105 + }, + { + "epoch": 0.18543434978518286, + "grad_norm": 0.5112168788909912, + "learning_rate": 1.876986019257446e-05, + "loss": 0.846, + "step": 1106 + }, + { + "epoch": 0.18560201194592896, + "grad_norm": 0.44852420687675476, + "learning_rate": 1.876724941777833e-05, + "loss": 0.974, + "step": 1107 + }, + { + "epoch": 0.18576967410667505, + "grad_norm": 0.46299490332603455, + "learning_rate": 1.876463605741178e-05, + "loss": 0.9148, + "step": 1108 + }, + { + "epoch": 0.18593733626742115, + "grad_norm": 0.7084307074546814, + "learning_rate": 1.876202011224553e-05, + "loss": 0.9549, + "step": 1109 + }, + { + "epoch": 0.18610499842816725, + "grad_norm": 0.4639042019844055, + "learning_rate": 1.8759401583051048e-05, + "loss": 0.9632, + "step": 1110 + }, + { + "epoch": 0.18627266058891334, + "grad_norm": 0.5113834738731384, + "learning_rate": 1.8756780470600574e-05, + "loss": 0.9011, + "step": 1111 + }, + { + "epoch": 0.18644032274965944, + "grad_norm": 0.6037055850028992, + "learning_rate": 1.8754156775667108e-05, + "loss": 0.9357, + "step": 1112 + }, + { + "epoch": 0.18660798491040553, + "grad_norm": 0.5472172498703003, + "learning_rate": 1.8751530499024403e-05, + "loss": 0.9359, + "step": 1113 + }, + { + "epoch": 0.18677564707115163, + "grad_norm": 0.42744895815849304, + "learning_rate": 1.8748901641446987e-05, + "loss": 0.9693, + "step": 1114 + }, + { + "epoch": 0.18694330923189773, + "grad_norm": 0.5853201150894165, + "learning_rate": 1.874627020371014e-05, + "loss": 0.8754, + "step": 1115 + }, + { + "epoch": 0.18711097139264382, + "grad_norm": 0.5467081069946289, + "learning_rate": 1.874363618658991e-05, + "loss": 0.9653, + "step": 1116 + }, + { + "epoch": 0.18727863355338992, + "grad_norm": 0.47121307253837585, + "learning_rate": 1.87409995908631e-05, + "loss": 0.9778, + "step": 1117 + }, + { + "epoch": 0.187446295714136, + "grad_norm": 0.5530045628547668, + "learning_rate": 1.8738360417307265e-05, + "loss": 0.995, + "step": 1118 + }, + { + "epoch": 0.1876139578748821, + "grad_norm": 0.5050699710845947, + "learning_rate": 1.873571866670074e-05, + "loss": 0.9575, + "step": 1119 + }, + { + "epoch": 0.1877816200356282, + "grad_norm": 0.48247194290161133, + "learning_rate": 1.873307433982261e-05, + "loss": 0.992, + "step": 1120 + }, + { + "epoch": 0.1879492821963743, + "grad_norm": 0.5888962149620056, + "learning_rate": 1.8730427437452717e-05, + "loss": 0.9089, + "step": 1121 + }, + { + "epoch": 0.1881169443571204, + "grad_norm": 0.43505725264549255, + "learning_rate": 1.8727777960371662e-05, + "loss": 0.9586, + "step": 1122 + }, + { + "epoch": 0.1882846065178665, + "grad_norm": 0.3793167471885681, + "learning_rate": 1.8725125909360816e-05, + "loss": 0.8689, + "step": 1123 + }, + { + "epoch": 0.1884522686786126, + "grad_norm": 0.42412588000297546, + "learning_rate": 1.8722471285202295e-05, + "loss": 0.9496, + "step": 1124 + }, + { + "epoch": 0.18861993083935868, + "grad_norm": 0.7520803213119507, + "learning_rate": 1.871981408867898e-05, + "loss": 0.9198, + "step": 1125 + }, + { + "epoch": 0.18878759300010478, + "grad_norm": 0.4949362277984619, + "learning_rate": 1.8717154320574522e-05, + "loss": 0.8231, + "step": 1126 + }, + { + "epoch": 0.18895525516085088, + "grad_norm": 0.5005122423171997, + "learning_rate": 1.8714491981673307e-05, + "loss": 0.9228, + "step": 1127 + }, + { + "epoch": 0.18912291732159697, + "grad_norm": 0.5874606966972351, + "learning_rate": 1.87118270727605e-05, + "loss": 0.8301, + "step": 1128 + }, + { + "epoch": 0.18929057948234307, + "grad_norm": 0.5143774747848511, + "learning_rate": 1.870915959462201e-05, + "loss": 0.9073, + "step": 1129 + }, + { + "epoch": 0.18945824164308916, + "grad_norm": 0.922320544719696, + "learning_rate": 1.8706489548044508e-05, + "loss": 0.9736, + "step": 1130 + }, + { + "epoch": 0.18962590380383526, + "grad_norm": 0.8636284470558167, + "learning_rate": 1.8703816933815433e-05, + "loss": 0.9078, + "step": 1131 + }, + { + "epoch": 0.18979356596458136, + "grad_norm": 1.1362345218658447, + "learning_rate": 1.8701141752722966e-05, + "loss": 0.9348, + "step": 1132 + }, + { + "epoch": 0.18996122812532745, + "grad_norm": 0.692942202091217, + "learning_rate": 1.8698464005556047e-05, + "loss": 0.8462, + "step": 1133 + }, + { + "epoch": 0.19012889028607355, + "grad_norm": 0.6303272843360901, + "learning_rate": 1.8695783693104386e-05, + "loss": 0.9073, + "step": 1134 + }, + { + "epoch": 0.19029655244681964, + "grad_norm": 0.502835214138031, + "learning_rate": 1.8693100816158436e-05, + "loss": 0.9079, + "step": 1135 + }, + { + "epoch": 0.19046421460756577, + "grad_norm": 0.4576420485973358, + "learning_rate": 1.869041537550941e-05, + "loss": 0.9106, + "step": 1136 + }, + { + "epoch": 0.19063187676831186, + "grad_norm": 0.8908449411392212, + "learning_rate": 1.8687727371949275e-05, + "loss": 0.9147, + "step": 1137 + }, + { + "epoch": 0.19079953892905796, + "grad_norm": 1.049439787864685, + "learning_rate": 1.868503680627076e-05, + "loss": 1.0857, + "step": 1138 + }, + { + "epoch": 0.19096720108980406, + "grad_norm": 0.6737858653068542, + "learning_rate": 1.868234367926735e-05, + "loss": 0.918, + "step": 1139 + }, + { + "epoch": 0.19113486325055015, + "grad_norm": 0.5657393932342529, + "learning_rate": 1.8679647991733268e-05, + "loss": 0.9149, + "step": 1140 + }, + { + "epoch": 0.19130252541129625, + "grad_norm": 0.5046471953392029, + "learning_rate": 1.8676949744463523e-05, + "loss": 0.9214, + "step": 1141 + }, + { + "epoch": 0.19147018757204234, + "grad_norm": 0.5004892349243164, + "learning_rate": 1.867424893825385e-05, + "loss": 0.9502, + "step": 1142 + }, + { + "epoch": 0.19163784973278844, + "grad_norm": 0.4683850109577179, + "learning_rate": 1.8671545573900755e-05, + "loss": 0.7789, + "step": 1143 + }, + { + "epoch": 0.19180551189353454, + "grad_norm": 0.48335880041122437, + "learning_rate": 1.866883965220149e-05, + "loss": 0.8971, + "step": 1144 + }, + { + "epoch": 0.19197317405428063, + "grad_norm": 0.3865186274051666, + "learning_rate": 1.866613117395407e-05, + "loss": 0.975, + "step": 1145 + }, + { + "epoch": 0.19214083621502673, + "grad_norm": 0.44036880135536194, + "learning_rate": 1.8663420139957252e-05, + "loss": 0.9938, + "step": 1146 + }, + { + "epoch": 0.19230849837577282, + "grad_norm": 0.5232735276222229, + "learning_rate": 1.866070655101056e-05, + "loss": 0.9507, + "step": 1147 + }, + { + "epoch": 0.19247616053651892, + "grad_norm": 0.4845692217350006, + "learning_rate": 1.8657990407914265e-05, + "loss": 0.9477, + "step": 1148 + }, + { + "epoch": 0.19264382269726502, + "grad_norm": 0.3743135929107666, + "learning_rate": 1.865527171146938e-05, + "loss": 0.9608, + "step": 1149 + }, + { + "epoch": 0.1928114848580111, + "grad_norm": 0.7610765695571899, + "learning_rate": 1.8652550462477697e-05, + "loss": 0.9364, + "step": 1150 + }, + { + "epoch": 0.1929791470187572, + "grad_norm": 0.35837090015411377, + "learning_rate": 1.864982666174174e-05, + "loss": 0.8771, + "step": 1151 + }, + { + "epoch": 0.1931468091795033, + "grad_norm": 0.6423548460006714, + "learning_rate": 1.864710031006479e-05, + "loss": 0.9066, + "step": 1152 + }, + { + "epoch": 0.1933144713402494, + "grad_norm": 0.3354455232620239, + "learning_rate": 1.864437140825088e-05, + "loss": 0.873, + "step": 1153 + }, + { + "epoch": 0.1934821335009955, + "grad_norm": 0.48024141788482666, + "learning_rate": 1.8641639957104805e-05, + "loss": 0.9756, + "step": 1154 + }, + { + "epoch": 0.1936497956617416, + "grad_norm": 0.7103967666625977, + "learning_rate": 1.8638905957432097e-05, + "loss": 0.9366, + "step": 1155 + }, + { + "epoch": 0.1938174578224877, + "grad_norm": 0.5881415009498596, + "learning_rate": 1.8636169410039046e-05, + "loss": 0.9805, + "step": 1156 + }, + { + "epoch": 0.19398511998323378, + "grad_norm": 0.5831557512283325, + "learning_rate": 1.86334303157327e-05, + "loss": 0.8257, + "step": 1157 + }, + { + "epoch": 0.19415278214397988, + "grad_norm": 0.6262730956077576, + "learning_rate": 1.8630688675320844e-05, + "loss": 0.8544, + "step": 1158 + }, + { + "epoch": 0.19432044430472598, + "grad_norm": 0.3318983018398285, + "learning_rate": 1.8627944489612025e-05, + "loss": 0.8899, + "step": 1159 + }, + { + "epoch": 0.19448810646547207, + "grad_norm": 0.4775957465171814, + "learning_rate": 1.862519775941554e-05, + "loss": 0.9246, + "step": 1160 + }, + { + "epoch": 0.19465576862621817, + "grad_norm": 0.5421665906906128, + "learning_rate": 1.862244848554143e-05, + "loss": 0.9443, + "step": 1161 + }, + { + "epoch": 0.19482343078696426, + "grad_norm": 0.48920223116874695, + "learning_rate": 1.8619696668800494e-05, + "loss": 0.8853, + "step": 1162 + }, + { + "epoch": 0.19499109294771036, + "grad_norm": 0.4863426089286804, + "learning_rate": 1.8616942310004273e-05, + "loss": 0.9421, + "step": 1163 + }, + { + "epoch": 0.19515875510845646, + "grad_norm": 0.47465115785598755, + "learning_rate": 1.8614185409965062e-05, + "loss": 0.951, + "step": 1164 + }, + { + "epoch": 0.19532641726920255, + "grad_norm": 0.6428936123847961, + "learning_rate": 1.8611425969495908e-05, + "loss": 0.8917, + "step": 1165 + }, + { + "epoch": 0.19549407942994865, + "grad_norm": 0.48652878403663635, + "learning_rate": 1.86086639894106e-05, + "loss": 1.004, + "step": 1166 + }, + { + "epoch": 0.19566174159069474, + "grad_norm": 0.6992968916893005, + "learning_rate": 1.8605899470523683e-05, + "loss": 0.8919, + "step": 1167 + }, + { + "epoch": 0.19582940375144084, + "grad_norm": 0.8496128916740417, + "learning_rate": 1.8603132413650443e-05, + "loss": 0.8623, + "step": 1168 + }, + { + "epoch": 0.19599706591218694, + "grad_norm": 0.5947402119636536, + "learning_rate": 1.8600362819606927e-05, + "loss": 0.7653, + "step": 1169 + }, + { + "epoch": 0.19616472807293303, + "grad_norm": 0.6545706391334534, + "learning_rate": 1.859759068920992e-05, + "loss": 0.8367, + "step": 1170 + }, + { + "epoch": 0.19633239023367913, + "grad_norm": 0.7118186950683594, + "learning_rate": 1.8594816023276954e-05, + "loss": 0.9847, + "step": 1171 + }, + { + "epoch": 0.19650005239442522, + "grad_norm": 0.4567353427410126, + "learning_rate": 1.8592038822626314e-05, + "loss": 0.9423, + "step": 1172 + }, + { + "epoch": 0.19666771455517132, + "grad_norm": 0.5413332581520081, + "learning_rate": 1.8589259088077033e-05, + "loss": 0.9664, + "step": 1173 + }, + { + "epoch": 0.19683537671591742, + "grad_norm": 0.4965800940990448, + "learning_rate": 1.858647682044889e-05, + "loss": 0.8317, + "step": 1174 + }, + { + "epoch": 0.1970030388766635, + "grad_norm": 0.4861052930355072, + "learning_rate": 1.8583692020562404e-05, + "loss": 0.9442, + "step": 1175 + }, + { + "epoch": 0.1971707010374096, + "grad_norm": 0.8544037342071533, + "learning_rate": 1.8580904689238855e-05, + "loss": 0.9905, + "step": 1176 + }, + { + "epoch": 0.1973383631981557, + "grad_norm": 0.49847543239593506, + "learning_rate": 1.8578114827300253e-05, + "loss": 0.8249, + "step": 1177 + }, + { + "epoch": 0.19750602535890183, + "grad_norm": 0.3586089611053467, + "learning_rate": 1.8575322435569368e-05, + "loss": 0.8953, + "step": 1178 + }, + { + "epoch": 0.19767368751964792, + "grad_norm": 0.4339447021484375, + "learning_rate": 1.857252751486971e-05, + "loss": 0.969, + "step": 1179 + }, + { + "epoch": 0.19784134968039402, + "grad_norm": 0.5465442538261414, + "learning_rate": 1.856973006602554e-05, + "loss": 1.0242, + "step": 1180 + }, + { + "epoch": 0.19800901184114011, + "grad_norm": 0.45056402683258057, + "learning_rate": 1.8566930089861853e-05, + "loss": 0.9243, + "step": 1181 + }, + { + "epoch": 0.1981766740018862, + "grad_norm": 0.5189386010169983, + "learning_rate": 1.85641275872044e-05, + "loss": 0.9065, + "step": 1182 + }, + { + "epoch": 0.1983443361626323, + "grad_norm": 0.3600766062736511, + "learning_rate": 1.8561322558879672e-05, + "loss": 0.8328, + "step": 1183 + }, + { + "epoch": 0.1985119983233784, + "grad_norm": 0.35876795649528503, + "learning_rate": 1.855851500571491e-05, + "loss": 0.9278, + "step": 1184 + }, + { + "epoch": 0.1986796604841245, + "grad_norm": 0.46552401781082153, + "learning_rate": 1.8555704928538093e-05, + "loss": 1.0202, + "step": 1185 + }, + { + "epoch": 0.1988473226448706, + "grad_norm": 0.5195321440696716, + "learning_rate": 1.8552892328177953e-05, + "loss": 0.962, + "step": 1186 + }, + { + "epoch": 0.1990149848056167, + "grad_norm": 0.4392522871494293, + "learning_rate": 1.8550077205463953e-05, + "loss": 0.9341, + "step": 1187 + }, + { + "epoch": 0.1991826469663628, + "grad_norm": 0.44318604469299316, + "learning_rate": 1.8547259561226313e-05, + "loss": 1.0245, + "step": 1188 + }, + { + "epoch": 0.19935030912710888, + "grad_norm": 0.4302159249782562, + "learning_rate": 1.8544439396295987e-05, + "loss": 0.9287, + "step": 1189 + }, + { + "epoch": 0.19951797128785498, + "grad_norm": 0.4559105634689331, + "learning_rate": 1.8541616711504676e-05, + "loss": 1.0109, + "step": 1190 + }, + { + "epoch": 0.19968563344860107, + "grad_norm": 0.49366477131843567, + "learning_rate": 1.853879150768483e-05, + "loss": 0.9105, + "step": 1191 + }, + { + "epoch": 0.19985329560934717, + "grad_norm": 0.41833993792533875, + "learning_rate": 1.8535963785669637e-05, + "loss": 0.9037, + "step": 1192 + }, + { + "epoch": 0.20002095777009327, + "grad_norm": 0.4915224015712738, + "learning_rate": 1.8533133546293026e-05, + "loss": 0.9391, + "step": 1193 + }, + { + "epoch": 0.20018861993083936, + "grad_norm": 0.3822900652885437, + "learning_rate": 1.8530300790389663e-05, + "loss": 0.9825, + "step": 1194 + }, + { + "epoch": 0.20035628209158546, + "grad_norm": 0.5990911722183228, + "learning_rate": 1.8527465518794967e-05, + "loss": 0.9609, + "step": 1195 + }, + { + "epoch": 0.20052394425233155, + "grad_norm": 0.6043885350227356, + "learning_rate": 1.85246277323451e-05, + "loss": 0.9627, + "step": 1196 + }, + { + "epoch": 0.20069160641307765, + "grad_norm": 0.5229475498199463, + "learning_rate": 1.8521787431876954e-05, + "loss": 0.9089, + "step": 1197 + }, + { + "epoch": 0.20085926857382375, + "grad_norm": 0.6026935577392578, + "learning_rate": 1.851894461822817e-05, + "loss": 0.8736, + "step": 1198 + }, + { + "epoch": 0.20102693073456984, + "grad_norm": 0.5354313850402832, + "learning_rate": 1.8516099292237132e-05, + "loss": 0.9127, + "step": 1199 + }, + { + "epoch": 0.20119459289531594, + "grad_norm": 0.5791041254997253, + "learning_rate": 1.8513251454742957e-05, + "loss": 0.9913, + "step": 1200 + }, + { + "epoch": 0.20136225505606203, + "grad_norm": 0.5280284285545349, + "learning_rate": 1.8510401106585508e-05, + "loss": 0.8859, + "step": 1201 + }, + { + "epoch": 0.20152991721680813, + "grad_norm": 0.43773534893989563, + "learning_rate": 1.8507548248605393e-05, + "loss": 0.9555, + "step": 1202 + }, + { + "epoch": 0.20169757937755423, + "grad_norm": 0.49584686756134033, + "learning_rate": 1.850469288164395e-05, + "loss": 0.9496, + "step": 1203 + }, + { + "epoch": 0.20186524153830032, + "grad_norm": 0.4935537278652191, + "learning_rate": 1.8501835006543264e-05, + "loss": 0.8697, + "step": 1204 + }, + { + "epoch": 0.20203290369904642, + "grad_norm": 0.4067463278770447, + "learning_rate": 1.8498974624146164e-05, + "loss": 0.9278, + "step": 1205 + }, + { + "epoch": 0.2022005658597925, + "grad_norm": 0.48257601261138916, + "learning_rate": 1.84961117352962e-05, + "loss": 0.8886, + "step": 1206 + }, + { + "epoch": 0.2023682280205386, + "grad_norm": 0.38043829798698425, + "learning_rate": 1.849324634083768e-05, + "loss": 0.8953, + "step": 1207 + }, + { + "epoch": 0.2025358901812847, + "grad_norm": 0.5581970810890198, + "learning_rate": 1.8490378441615645e-05, + "loss": 0.91, + "step": 1208 + }, + { + "epoch": 0.2027035523420308, + "grad_norm": 0.47461193799972534, + "learning_rate": 1.8487508038475873e-05, + "loss": 0.9231, + "step": 1209 + }, + { + "epoch": 0.2028712145027769, + "grad_norm": 0.5092021226882935, + "learning_rate": 1.848463513226488e-05, + "loss": 0.9152, + "step": 1210 + }, + { + "epoch": 0.203038876663523, + "grad_norm": 0.48253050446510315, + "learning_rate": 1.8481759723829922e-05, + "loss": 0.9013, + "step": 1211 + }, + { + "epoch": 0.2032065388242691, + "grad_norm": 0.49508121609687805, + "learning_rate": 1.8478881814019e-05, + "loss": 0.8866, + "step": 1212 + }, + { + "epoch": 0.20337420098501519, + "grad_norm": 0.4377021789550781, + "learning_rate": 1.8476001403680832e-05, + "loss": 0.961, + "step": 1213 + }, + { + "epoch": 0.20354186314576128, + "grad_norm": 0.8740084171295166, + "learning_rate": 1.8473118493664897e-05, + "loss": 0.9397, + "step": 1214 + }, + { + "epoch": 0.20370952530650738, + "grad_norm": 0.49391335248947144, + "learning_rate": 1.8470233084821393e-05, + "loss": 0.8889, + "step": 1215 + }, + { + "epoch": 0.20387718746725347, + "grad_norm": 0.5895149111747742, + "learning_rate": 1.8467345178001272e-05, + "loss": 0.8346, + "step": 1216 + }, + { + "epoch": 0.20404484962799957, + "grad_norm": 0.48423007130622864, + "learning_rate": 1.84644547740562e-05, + "loss": 0.9015, + "step": 1217 + }, + { + "epoch": 0.20421251178874567, + "grad_norm": 0.34172114729881287, + "learning_rate": 1.8461561873838612e-05, + "loss": 0.8922, + "step": 1218 + }, + { + "epoch": 0.20438017394949176, + "grad_norm": 0.40491271018981934, + "learning_rate": 1.8458666478201643e-05, + "loss": 0.927, + "step": 1219 + }, + { + "epoch": 0.20454783611023786, + "grad_norm": 0.5788916349411011, + "learning_rate": 1.845576858799919e-05, + "loss": 0.9363, + "step": 1220 + }, + { + "epoch": 0.20471549827098398, + "grad_norm": 0.5173137784004211, + "learning_rate": 1.8452868204085873e-05, + "loss": 0.8754, + "step": 1221 + }, + { + "epoch": 0.20488316043173008, + "grad_norm": 0.38726314902305603, + "learning_rate": 1.8449965327317053e-05, + "loss": 0.8359, + "step": 1222 + }, + { + "epoch": 0.20505082259247617, + "grad_norm": 0.43636006116867065, + "learning_rate": 1.8447059958548822e-05, + "loss": 0.9459, + "step": 1223 + }, + { + "epoch": 0.20521848475322227, + "grad_norm": 0.4201851785182953, + "learning_rate": 1.8444152098638014e-05, + "loss": 0.92, + "step": 1224 + }, + { + "epoch": 0.20538614691396836, + "grad_norm": 0.541737973690033, + "learning_rate": 1.8441241748442186e-05, + "loss": 0.8944, + "step": 1225 + }, + { + "epoch": 0.20555380907471446, + "grad_norm": 0.4758997857570648, + "learning_rate": 1.8438328908819643e-05, + "loss": 0.8966, + "step": 1226 + }, + { + "epoch": 0.20572147123546056, + "grad_norm": 0.5539448857307434, + "learning_rate": 1.8435413580629412e-05, + "loss": 0.9706, + "step": 1227 + }, + { + "epoch": 0.20588913339620665, + "grad_norm": 0.56991046667099, + "learning_rate": 1.8432495764731265e-05, + "loss": 0.976, + "step": 1228 + }, + { + "epoch": 0.20605679555695275, + "grad_norm": 0.4244685173034668, + "learning_rate": 1.8429575461985693e-05, + "loss": 0.8839, + "step": 1229 + }, + { + "epoch": 0.20622445771769884, + "grad_norm": 0.6971105337142944, + "learning_rate": 1.8426652673253943e-05, + "loss": 0.9441, + "step": 1230 + }, + { + "epoch": 0.20639211987844494, + "grad_norm": 0.5528003573417664, + "learning_rate": 1.8423727399397968e-05, + "loss": 0.9105, + "step": 1231 + }, + { + "epoch": 0.20655978203919104, + "grad_norm": 0.3812488913536072, + "learning_rate": 1.8420799641280478e-05, + "loss": 0.9143, + "step": 1232 + }, + { + "epoch": 0.20672744419993713, + "grad_norm": 0.423740953207016, + "learning_rate": 1.8417869399764898e-05, + "loss": 0.9908, + "step": 1233 + }, + { + "epoch": 0.20689510636068323, + "grad_norm": 0.47226041555404663, + "learning_rate": 1.8414936675715395e-05, + "loss": 0.8505, + "step": 1234 + }, + { + "epoch": 0.20706276852142932, + "grad_norm": 0.8200578093528748, + "learning_rate": 1.8412001469996867e-05, + "loss": 0.9519, + "step": 1235 + }, + { + "epoch": 0.20723043068217542, + "grad_norm": 0.3981587290763855, + "learning_rate": 1.840906378347494e-05, + "loss": 0.8226, + "step": 1236 + }, + { + "epoch": 0.20739809284292152, + "grad_norm": 0.5681799054145813, + "learning_rate": 1.840612361701598e-05, + "loss": 0.9721, + "step": 1237 + }, + { + "epoch": 0.2075657550036676, + "grad_norm": 0.617257297039032, + "learning_rate": 1.8403180971487067e-05, + "loss": 0.8837, + "step": 1238 + }, + { + "epoch": 0.2077334171644137, + "grad_norm": 0.7258968949317932, + "learning_rate": 1.8400235847756035e-05, + "loss": 0.8497, + "step": 1239 + }, + { + "epoch": 0.2079010793251598, + "grad_norm": 0.6802220344543457, + "learning_rate": 1.839728824669143e-05, + "loss": 1.0061, + "step": 1240 + }, + { + "epoch": 0.2080687414859059, + "grad_norm": 0.4653562903404236, + "learning_rate": 1.839433816916254e-05, + "loss": 0.9083, + "step": 1241 + }, + { + "epoch": 0.208236403646652, + "grad_norm": 0.5622283220291138, + "learning_rate": 1.839138561603938e-05, + "loss": 0.891, + "step": 1242 + }, + { + "epoch": 0.2084040658073981, + "grad_norm": 0.4902152121067047, + "learning_rate": 1.8388430588192693e-05, + "loss": 0.876, + "step": 1243 + }, + { + "epoch": 0.2085717279681442, + "grad_norm": 0.7380679845809937, + "learning_rate": 1.838547308649395e-05, + "loss": 0.9175, + "step": 1244 + }, + { + "epoch": 0.20873939012889028, + "grad_norm": 0.7325884699821472, + "learning_rate": 1.838251311181536e-05, + "loss": 0.849, + "step": 1245 + }, + { + "epoch": 0.20890705228963638, + "grad_norm": 0.5089899897575378, + "learning_rate": 1.8379550665029856e-05, + "loss": 0.8693, + "step": 1246 + }, + { + "epoch": 0.20907471445038248, + "grad_norm": 0.48446375131607056, + "learning_rate": 1.8376585747011094e-05, + "loss": 0.9291, + "step": 1247 + }, + { + "epoch": 0.20924237661112857, + "grad_norm": 0.4304530620574951, + "learning_rate": 1.837361835863347e-05, + "loss": 0.9757, + "step": 1248 + }, + { + "epoch": 0.20941003877187467, + "grad_norm": 0.4314531087875366, + "learning_rate": 1.8370648500772107e-05, + "loss": 0.9058, + "step": 1249 + }, + { + "epoch": 0.20957770093262076, + "grad_norm": 0.425742506980896, + "learning_rate": 1.8367676174302848e-05, + "loss": 0.8532, + "step": 1250 + }, + { + "epoch": 0.20974536309336686, + "grad_norm": 0.5282460451126099, + "learning_rate": 1.8364701380102267e-05, + "loss": 0.9287, + "step": 1251 + }, + { + "epoch": 0.20991302525411296, + "grad_norm": 0.43143585324287415, + "learning_rate": 1.8361724119047675e-05, + "loss": 0.918, + "step": 1252 + }, + { + "epoch": 0.21008068741485905, + "grad_norm": 0.4454243779182434, + "learning_rate": 1.8358744392017095e-05, + "loss": 0.9913, + "step": 1253 + }, + { + "epoch": 0.21024834957560515, + "grad_norm": 0.41314229369163513, + "learning_rate": 1.835576219988929e-05, + "loss": 0.9382, + "step": 1254 + }, + { + "epoch": 0.21041601173635124, + "grad_norm": 0.49109718203544617, + "learning_rate": 1.8352777543543745e-05, + "loss": 0.937, + "step": 1255 + }, + { + "epoch": 0.21058367389709734, + "grad_norm": 0.4162750244140625, + "learning_rate": 1.834979042386067e-05, + "loss": 0.8294, + "step": 1256 + }, + { + "epoch": 0.21075133605784344, + "grad_norm": 0.5138880014419556, + "learning_rate": 1.8346800841721006e-05, + "loss": 0.9124, + "step": 1257 + }, + { + "epoch": 0.21091899821858953, + "grad_norm": 0.44537249207496643, + "learning_rate": 1.8343808798006417e-05, + "loss": 0.8992, + "step": 1258 + }, + { + "epoch": 0.21108666037933563, + "grad_norm": 0.5039196014404297, + "learning_rate": 1.8340814293599293e-05, + "loss": 0.9855, + "step": 1259 + }, + { + "epoch": 0.21125432254008172, + "grad_norm": 0.46184661984443665, + "learning_rate": 1.8337817329382754e-05, + "loss": 0.9403, + "step": 1260 + }, + { + "epoch": 0.21142198470082782, + "grad_norm": 0.556057333946228, + "learning_rate": 1.8334817906240638e-05, + "loss": 0.8827, + "step": 1261 + }, + { + "epoch": 0.21158964686157392, + "grad_norm": 0.5662396550178528, + "learning_rate": 1.8331816025057508e-05, + "loss": 0.9287, + "step": 1262 + }, + { + "epoch": 0.21175730902232, + "grad_norm": 0.4389037489891052, + "learning_rate": 1.8328811686718667e-05, + "loss": 0.8198, + "step": 1263 + }, + { + "epoch": 0.21192497118306614, + "grad_norm": 0.4806714951992035, + "learning_rate": 1.832580489211012e-05, + "loss": 0.8745, + "step": 1264 + }, + { + "epoch": 0.21209263334381223, + "grad_norm": 0.6693044900894165, + "learning_rate": 1.8322795642118618e-05, + "loss": 0.8701, + "step": 1265 + }, + { + "epoch": 0.21226029550455833, + "grad_norm": 0.5517094135284424, + "learning_rate": 1.8319783937631622e-05, + "loss": 0.8977, + "step": 1266 + }, + { + "epoch": 0.21242795766530442, + "grad_norm": 0.5779098272323608, + "learning_rate": 1.8316769779537316e-05, + "loss": 0.9396, + "step": 1267 + }, + { + "epoch": 0.21259561982605052, + "grad_norm": 0.6421261429786682, + "learning_rate": 1.831375316872462e-05, + "loss": 0.9319, + "step": 1268 + }, + { + "epoch": 0.21276328198679662, + "grad_norm": 0.4507690966129303, + "learning_rate": 1.8310734106083166e-05, + "loss": 0.9084, + "step": 1269 + }, + { + "epoch": 0.2129309441475427, + "grad_norm": 0.4391637146472931, + "learning_rate": 1.830771259250331e-05, + "loss": 0.8856, + "step": 1270 + }, + { + "epoch": 0.2130986063082888, + "grad_norm": 0.7781029939651489, + "learning_rate": 1.8304688628876147e-05, + "loss": 0.9225, + "step": 1271 + }, + { + "epoch": 0.2132662684690349, + "grad_norm": 0.5514547824859619, + "learning_rate": 1.8301662216093464e-05, + "loss": 0.9248, + "step": 1272 + }, + { + "epoch": 0.213433930629781, + "grad_norm": 0.5290319919586182, + "learning_rate": 1.82986333550478e-05, + "loss": 0.8452, + "step": 1273 + }, + { + "epoch": 0.2136015927905271, + "grad_norm": 0.48711273074150085, + "learning_rate": 1.8295602046632396e-05, + "loss": 0.9225, + "step": 1274 + }, + { + "epoch": 0.2137692549512732, + "grad_norm": 0.8355472087860107, + "learning_rate": 1.8292568291741228e-05, + "loss": 0.9159, + "step": 1275 + }, + { + "epoch": 0.2139369171120193, + "grad_norm": 0.6240776181221008, + "learning_rate": 1.8289532091268985e-05, + "loss": 0.8713, + "step": 1276 + }, + { + "epoch": 0.21410457927276538, + "grad_norm": 0.42136117815971375, + "learning_rate": 1.8286493446111077e-05, + "loss": 0.8471, + "step": 1277 + }, + { + "epoch": 0.21427224143351148, + "grad_norm": 0.3959744870662689, + "learning_rate": 1.8283452357163644e-05, + "loss": 0.8487, + "step": 1278 + }, + { + "epoch": 0.21443990359425757, + "grad_norm": 0.43854543566703796, + "learning_rate": 1.828040882532354e-05, + "loss": 0.9884, + "step": 1279 + }, + { + "epoch": 0.21460756575500367, + "grad_norm": 0.5910789966583252, + "learning_rate": 1.827736285148834e-05, + "loss": 0.9265, + "step": 1280 + }, + { + "epoch": 0.21477522791574977, + "grad_norm": 0.6404376029968262, + "learning_rate": 1.8274314436556336e-05, + "loss": 0.9823, + "step": 1281 + }, + { + "epoch": 0.21494289007649586, + "grad_norm": 0.598853349685669, + "learning_rate": 1.827126358142654e-05, + "loss": 0.9459, + "step": 1282 + }, + { + "epoch": 0.21511055223724196, + "grad_norm": 0.4955078363418579, + "learning_rate": 1.82682102869987e-05, + "loss": 0.8622, + "step": 1283 + }, + { + "epoch": 0.21527821439798805, + "grad_norm": 0.5622814297676086, + "learning_rate": 1.8265154554173257e-05, + "loss": 0.9037, + "step": 1284 + }, + { + "epoch": 0.21544587655873415, + "grad_norm": 0.4194563925266266, + "learning_rate": 1.8262096383851396e-05, + "loss": 0.9338, + "step": 1285 + }, + { + "epoch": 0.21561353871948025, + "grad_norm": 0.47616180777549744, + "learning_rate": 1.8259035776935003e-05, + "loss": 0.8984, + "step": 1286 + }, + { + "epoch": 0.21578120088022634, + "grad_norm": 0.42187511920928955, + "learning_rate": 1.8255972734326685e-05, + "loss": 0.8744, + "step": 1287 + }, + { + "epoch": 0.21594886304097244, + "grad_norm": 0.47959572076797485, + "learning_rate": 1.8252907256929777e-05, + "loss": 1.0063, + "step": 1288 + }, + { + "epoch": 0.21611652520171853, + "grad_norm": 0.5419632792472839, + "learning_rate": 1.8249839345648324e-05, + "loss": 0.9908, + "step": 1289 + }, + { + "epoch": 0.21628418736246463, + "grad_norm": 0.47485166788101196, + "learning_rate": 1.8246769001387092e-05, + "loss": 0.9448, + "step": 1290 + }, + { + "epoch": 0.21645184952321073, + "grad_norm": 0.4475467801094055, + "learning_rate": 1.8243696225051564e-05, + "loss": 0.9025, + "step": 1291 + }, + { + "epoch": 0.21661951168395682, + "grad_norm": 0.5631753206253052, + "learning_rate": 1.8240621017547938e-05, + "loss": 0.8794, + "step": 1292 + }, + { + "epoch": 0.21678717384470292, + "grad_norm": 0.5549768209457397, + "learning_rate": 1.8237543379783133e-05, + "loss": 0.9901, + "step": 1293 + }, + { + "epoch": 0.21695483600544901, + "grad_norm": 0.36228808760643005, + "learning_rate": 1.8234463312664782e-05, + "loss": 0.9285, + "step": 1294 + }, + { + "epoch": 0.2171224981661951, + "grad_norm": 0.48294028639793396, + "learning_rate": 1.8231380817101234e-05, + "loss": 0.8196, + "step": 1295 + }, + { + "epoch": 0.2172901603269412, + "grad_norm": 0.6337847113609314, + "learning_rate": 1.8228295894001554e-05, + "loss": 0.8088, + "step": 1296 + }, + { + "epoch": 0.2174578224876873, + "grad_norm": 0.5867326259613037, + "learning_rate": 1.8225208544275526e-05, + "loss": 0.9791, + "step": 1297 + }, + { + "epoch": 0.2176254846484334, + "grad_norm": 0.4840776026248932, + "learning_rate": 1.822211876883365e-05, + "loss": 1.0371, + "step": 1298 + }, + { + "epoch": 0.2177931468091795, + "grad_norm": 0.49656978249549866, + "learning_rate": 1.821902656858713e-05, + "loss": 0.8567, + "step": 1299 + }, + { + "epoch": 0.2179608089699256, + "grad_norm": 0.5669172406196594, + "learning_rate": 1.8215931944447904e-05, + "loss": 0.9545, + "step": 1300 + }, + { + "epoch": 0.2181284711306717, + "grad_norm": 0.4737217128276825, + "learning_rate": 1.8212834897328614e-05, + "loss": 0.8801, + "step": 1301 + }, + { + "epoch": 0.21829613329141778, + "grad_norm": 0.49199214577674866, + "learning_rate": 1.820973542814261e-05, + "loss": 0.9038, + "step": 1302 + }, + { + "epoch": 0.21846379545216388, + "grad_norm": 0.7255045771598816, + "learning_rate": 1.820663353780397e-05, + "loss": 0.8827, + "step": 1303 + }, + { + "epoch": 0.21863145761290997, + "grad_norm": 0.8148097991943359, + "learning_rate": 1.820352922722748e-05, + "loss": 0.9634, + "step": 1304 + }, + { + "epoch": 0.21879911977365607, + "grad_norm": 0.6887121200561523, + "learning_rate": 1.8200422497328635e-05, + "loss": 0.9764, + "step": 1305 + }, + { + "epoch": 0.21896678193440217, + "grad_norm": 0.6535807251930237, + "learning_rate": 1.8197313349023655e-05, + "loss": 0.8518, + "step": 1306 + }, + { + "epoch": 0.2191344440951483, + "grad_norm": 0.4110938012599945, + "learning_rate": 1.819420178322946e-05, + "loss": 0.94, + "step": 1307 + }, + { + "epoch": 0.21930210625589439, + "grad_norm": 0.5408376455307007, + "learning_rate": 1.8191087800863692e-05, + "loss": 0.9591, + "step": 1308 + }, + { + "epoch": 0.21946976841664048, + "grad_norm": 0.5167989730834961, + "learning_rate": 1.8187971402844706e-05, + "loss": 0.8543, + "step": 1309 + }, + { + "epoch": 0.21963743057738658, + "grad_norm": 0.5233778357505798, + "learning_rate": 1.818485259009156e-05, + "loss": 0.9628, + "step": 1310 + }, + { + "epoch": 0.21980509273813267, + "grad_norm": 0.5276911854743958, + "learning_rate": 1.818173136352403e-05, + "loss": 0.9192, + "step": 1311 + }, + { + "epoch": 0.21997275489887877, + "grad_norm": 0.43043678998947144, + "learning_rate": 1.8178607724062613e-05, + "loss": 1.0198, + "step": 1312 + }, + { + "epoch": 0.22014041705962487, + "grad_norm": 0.47249215841293335, + "learning_rate": 1.8175481672628505e-05, + "loss": 0.9738, + "step": 1313 + }, + { + "epoch": 0.22030807922037096, + "grad_norm": 0.4545799791812897, + "learning_rate": 1.8172353210143613e-05, + "loss": 0.9664, + "step": 1314 + }, + { + "epoch": 0.22047574138111706, + "grad_norm": 0.6046268939971924, + "learning_rate": 1.8169222337530567e-05, + "loss": 0.8992, + "step": 1315 + }, + { + "epoch": 0.22064340354186315, + "grad_norm": 0.4744549095630646, + "learning_rate": 1.816608905571269e-05, + "loss": 1.0001, + "step": 1316 + }, + { + "epoch": 0.22081106570260925, + "grad_norm": 0.4846677780151367, + "learning_rate": 1.8162953365614037e-05, + "loss": 0.9084, + "step": 1317 + }, + { + "epoch": 0.22097872786335535, + "grad_norm": 0.5149449110031128, + "learning_rate": 1.8159815268159348e-05, + "loss": 0.8979, + "step": 1318 + }, + { + "epoch": 0.22114639002410144, + "grad_norm": 0.5075791478157043, + "learning_rate": 1.8156674764274102e-05, + "loss": 0.8869, + "step": 1319 + }, + { + "epoch": 0.22131405218484754, + "grad_norm": 0.3952072262763977, + "learning_rate": 1.8153531854884465e-05, + "loss": 0.8001, + "step": 1320 + }, + { + "epoch": 0.22148171434559363, + "grad_norm": 0.39093104004859924, + "learning_rate": 1.815038654091732e-05, + "loss": 0.9095, + "step": 1321 + }, + { + "epoch": 0.22164937650633973, + "grad_norm": 0.40068456530570984, + "learning_rate": 1.814723882330026e-05, + "loss": 1.0543, + "step": 1322 + }, + { + "epoch": 0.22181703866708583, + "grad_norm": 0.46475377678871155, + "learning_rate": 1.814408870296158e-05, + "loss": 1.0002, + "step": 1323 + }, + { + "epoch": 0.22198470082783192, + "grad_norm": 0.5081574320793152, + "learning_rate": 1.81409361808303e-05, + "loss": 0.8861, + "step": 1324 + }, + { + "epoch": 0.22215236298857802, + "grad_norm": 0.4871772229671478, + "learning_rate": 1.813778125783613e-05, + "loss": 0.9433, + "step": 1325 + }, + { + "epoch": 0.2223200251493241, + "grad_norm": 0.6324952840805054, + "learning_rate": 1.81346239349095e-05, + "loss": 0.9318, + "step": 1326 + }, + { + "epoch": 0.2224876873100702, + "grad_norm": 0.4739683270454407, + "learning_rate": 1.813146421298154e-05, + "loss": 0.9027, + "step": 1327 + }, + { + "epoch": 0.2226553494708163, + "grad_norm": 0.44335031509399414, + "learning_rate": 1.8128302092984095e-05, + "loss": 0.952, + "step": 1328 + }, + { + "epoch": 0.2228230116315624, + "grad_norm": 0.6963682174682617, + "learning_rate": 1.812513757584971e-05, + "loss": 0.9993, + "step": 1329 + }, + { + "epoch": 0.2229906737923085, + "grad_norm": 0.46847036480903625, + "learning_rate": 1.8121970662511646e-05, + "loss": 0.9906, + "step": 1330 + }, + { + "epoch": 0.2231583359530546, + "grad_norm": 0.4766705632209778, + "learning_rate": 1.8118801353903856e-05, + "loss": 0.9885, + "step": 1331 + }, + { + "epoch": 0.2233259981138007, + "grad_norm": 0.4485360085964203, + "learning_rate": 1.8115629650961014e-05, + "loss": 0.7972, + "step": 1332 + }, + { + "epoch": 0.22349366027454678, + "grad_norm": 0.7110408544540405, + "learning_rate": 1.8112455554618495e-05, + "loss": 0.9015, + "step": 1333 + }, + { + "epoch": 0.22366132243529288, + "grad_norm": 0.4494708776473999, + "learning_rate": 1.810927906581238e-05, + "loss": 0.9505, + "step": 1334 + }, + { + "epoch": 0.22382898459603898, + "grad_norm": 1.1168867349624634, + "learning_rate": 1.8106100185479446e-05, + "loss": 1.0235, + "step": 1335 + }, + { + "epoch": 0.22399664675678507, + "grad_norm": 0.4510136842727661, + "learning_rate": 1.8102918914557196e-05, + "loss": 0.9523, + "step": 1336 + }, + { + "epoch": 0.22416430891753117, + "grad_norm": 0.5218761563301086, + "learning_rate": 1.809973525398382e-05, + "loss": 0.8929, + "step": 1337 + }, + { + "epoch": 0.22433197107827726, + "grad_norm": 0.5287281274795532, + "learning_rate": 1.8096549204698223e-05, + "loss": 0.941, + "step": 1338 + }, + { + "epoch": 0.22449963323902336, + "grad_norm": 0.6634778380393982, + "learning_rate": 1.8093360767640003e-05, + "loss": 0.9209, + "step": 1339 + }, + { + "epoch": 0.22466729539976946, + "grad_norm": 0.5254518985748291, + "learning_rate": 1.8090169943749477e-05, + "loss": 1.0379, + "step": 1340 + }, + { + "epoch": 0.22483495756051555, + "grad_norm": 1.0332894325256348, + "learning_rate": 1.808697673396765e-05, + "loss": 0.7821, + "step": 1341 + }, + { + "epoch": 0.22500261972126165, + "grad_norm": 0.6242459416389465, + "learning_rate": 1.808378113923625e-05, + "loss": 0.9408, + "step": 1342 + }, + { + "epoch": 0.22517028188200774, + "grad_norm": 0.4452417194843292, + "learning_rate": 1.8080583160497695e-05, + "loss": 0.9338, + "step": 1343 + }, + { + "epoch": 0.22533794404275384, + "grad_norm": 0.4866446554660797, + "learning_rate": 1.80773827986951e-05, + "loss": 0.9156, + "step": 1344 + }, + { + "epoch": 0.22550560620349994, + "grad_norm": 0.6305269002914429, + "learning_rate": 1.80741800547723e-05, + "loss": 0.9429, + "step": 1345 + }, + { + "epoch": 0.22567326836424603, + "grad_norm": 0.4827277362346649, + "learning_rate": 1.8070974929673818e-05, + "loss": 0.8953, + "step": 1346 + }, + { + "epoch": 0.22584093052499213, + "grad_norm": 0.7621358036994934, + "learning_rate": 1.806776742434489e-05, + "loss": 0.994, + "step": 1347 + }, + { + "epoch": 0.22600859268573822, + "grad_norm": 0.5351371765136719, + "learning_rate": 1.8064557539731446e-05, + "loss": 0.8527, + "step": 1348 + }, + { + "epoch": 0.22617625484648432, + "grad_norm": 0.5965827107429504, + "learning_rate": 1.8061345276780124e-05, + "loss": 0.942, + "step": 1349 + }, + { + "epoch": 0.22634391700723044, + "grad_norm": 0.5161508321762085, + "learning_rate": 1.805813063643825e-05, + "loss": 0.9561, + "step": 1350 + }, + { + "epoch": 0.22651157916797654, + "grad_norm": 0.40499433875083923, + "learning_rate": 1.805491361965388e-05, + "loss": 0.8948, + "step": 1351 + }, + { + "epoch": 0.22667924132872264, + "grad_norm": 0.6148918271064758, + "learning_rate": 1.8051694227375734e-05, + "loss": 0.8553, + "step": 1352 + }, + { + "epoch": 0.22684690348946873, + "grad_norm": 0.7514647841453552, + "learning_rate": 1.804847246055326e-05, + "loss": 0.9548, + "step": 1353 + }, + { + "epoch": 0.22701456565021483, + "grad_norm": 0.6081562638282776, + "learning_rate": 1.8045248320136593e-05, + "loss": 0.9345, + "step": 1354 + }, + { + "epoch": 0.22718222781096092, + "grad_norm": 0.6093937158584595, + "learning_rate": 1.8042021807076575e-05, + "loss": 0.8476, + "step": 1355 + }, + { + "epoch": 0.22734988997170702, + "grad_norm": 0.7178387641906738, + "learning_rate": 1.8038792922324743e-05, + "loss": 0.8914, + "step": 1356 + }, + { + "epoch": 0.22751755213245312, + "grad_norm": 0.6035656332969666, + "learning_rate": 1.8035561666833336e-05, + "loss": 0.9777, + "step": 1357 + }, + { + "epoch": 0.2276852142931992, + "grad_norm": 0.7128610014915466, + "learning_rate": 1.803232804155529e-05, + "loss": 0.9712, + "step": 1358 + }, + { + "epoch": 0.2278528764539453, + "grad_norm": 0.5069277882575989, + "learning_rate": 1.8029092047444238e-05, + "loss": 0.9525, + "step": 1359 + }, + { + "epoch": 0.2280205386146914, + "grad_norm": 0.5084493160247803, + "learning_rate": 1.8025853685454522e-05, + "loss": 0.9595, + "step": 1360 + }, + { + "epoch": 0.2281882007754375, + "grad_norm": 0.5395858287811279, + "learning_rate": 1.8022612956541174e-05, + "loss": 0.9214, + "step": 1361 + }, + { + "epoch": 0.2283558629361836, + "grad_norm": 0.6916206479072571, + "learning_rate": 1.801936986165992e-05, + "loss": 0.9464, + "step": 1362 + }, + { + "epoch": 0.2285235250969297, + "grad_norm": 0.532524585723877, + "learning_rate": 1.801612440176719e-05, + "loss": 0.8434, + "step": 1363 + }, + { + "epoch": 0.2286911872576758, + "grad_norm": 0.6037734150886536, + "learning_rate": 1.8012876577820113e-05, + "loss": 0.9723, + "step": 1364 + }, + { + "epoch": 0.22885884941842188, + "grad_norm": 0.678534209728241, + "learning_rate": 1.800962639077651e-05, + "loss": 0.8821, + "step": 1365 + }, + { + "epoch": 0.22902651157916798, + "grad_norm": 0.48763179779052734, + "learning_rate": 1.8006373841594905e-05, + "loss": 0.9745, + "step": 1366 + }, + { + "epoch": 0.22919417373991408, + "grad_norm": 0.4831984341144562, + "learning_rate": 1.800311893123451e-05, + "loss": 0.8353, + "step": 1367 + }, + { + "epoch": 0.22936183590066017, + "grad_norm": 0.5547946095466614, + "learning_rate": 1.7999861660655247e-05, + "loss": 0.9186, + "step": 1368 + }, + { + "epoch": 0.22952949806140627, + "grad_norm": 0.4802793860435486, + "learning_rate": 1.7996602030817713e-05, + "loss": 0.9354, + "step": 1369 + }, + { + "epoch": 0.22969716022215236, + "grad_norm": 0.49225011467933655, + "learning_rate": 1.7993340042683222e-05, + "loss": 0.9488, + "step": 1370 + }, + { + "epoch": 0.22986482238289846, + "grad_norm": 0.5383260250091553, + "learning_rate": 1.7990075697213775e-05, + "loss": 0.8672, + "step": 1371 + }, + { + "epoch": 0.23003248454364456, + "grad_norm": 0.5191199779510498, + "learning_rate": 1.798680899537206e-05, + "loss": 0.974, + "step": 1372 + }, + { + "epoch": 0.23020014670439065, + "grad_norm": 0.45761701464653015, + "learning_rate": 1.7983539938121476e-05, + "loss": 0.8985, + "step": 1373 + }, + { + "epoch": 0.23036780886513675, + "grad_norm": 0.4727153778076172, + "learning_rate": 1.7980268526426105e-05, + "loss": 0.9525, + "step": 1374 + }, + { + "epoch": 0.23053547102588284, + "grad_norm": 0.4912465512752533, + "learning_rate": 1.7976994761250727e-05, + "loss": 1.0002, + "step": 1375 + }, + { + "epoch": 0.23070313318662894, + "grad_norm": 0.41561660170555115, + "learning_rate": 1.7973718643560817e-05, + "loss": 0.9725, + "step": 1376 + }, + { + "epoch": 0.23087079534737504, + "grad_norm": 0.4833761155605316, + "learning_rate": 1.797044017432254e-05, + "loss": 0.9952, + "step": 1377 + }, + { + "epoch": 0.23103845750812113, + "grad_norm": 0.6341785788536072, + "learning_rate": 1.7967159354502764e-05, + "loss": 0.8741, + "step": 1378 + }, + { + "epoch": 0.23120611966886723, + "grad_norm": 0.41993364691734314, + "learning_rate": 1.7963876185069032e-05, + "loss": 0.9431, + "step": 1379 + }, + { + "epoch": 0.23137378182961332, + "grad_norm": 0.43000006675720215, + "learning_rate": 1.7960590666989604e-05, + "loss": 0.818, + "step": 1380 + }, + { + "epoch": 0.23154144399035942, + "grad_norm": 0.4882097840309143, + "learning_rate": 1.795730280123341e-05, + "loss": 0.8718, + "step": 1381 + }, + { + "epoch": 0.23170910615110552, + "grad_norm": 0.7146399617195129, + "learning_rate": 1.795401258877009e-05, + "loss": 0.8447, + "step": 1382 + }, + { + "epoch": 0.2318767683118516, + "grad_norm": 0.5772314667701721, + "learning_rate": 1.795072003056996e-05, + "loss": 0.7995, + "step": 1383 + }, + { + "epoch": 0.2320444304725977, + "grad_norm": 0.655830442905426, + "learning_rate": 1.7947425127604045e-05, + "loss": 0.9063, + "step": 1384 + }, + { + "epoch": 0.2322120926333438, + "grad_norm": 1.0778709650039673, + "learning_rate": 1.7944127880844043e-05, + "loss": 0.8961, + "step": 1385 + }, + { + "epoch": 0.2323797547940899, + "grad_norm": 0.36804503202438354, + "learning_rate": 1.794082829126236e-05, + "loss": 0.8971, + "step": 1386 + }, + { + "epoch": 0.232547416954836, + "grad_norm": 0.4572069048881531, + "learning_rate": 1.7937526359832088e-05, + "loss": 0.8715, + "step": 1387 + }, + { + "epoch": 0.2327150791155821, + "grad_norm": 1.2500030994415283, + "learning_rate": 1.7934222087526996e-05, + "loss": 0.8778, + "step": 1388 + }, + { + "epoch": 0.2328827412763282, + "grad_norm": 0.4822962284088135, + "learning_rate": 1.793091547532157e-05, + "loss": 0.8366, + "step": 1389 + }, + { + "epoch": 0.23305040343707428, + "grad_norm": 0.36731037497520447, + "learning_rate": 1.792760652419096e-05, + "loss": 0.8884, + "step": 1390 + }, + { + "epoch": 0.23321806559782038, + "grad_norm": 0.576995313167572, + "learning_rate": 1.792429523511102e-05, + "loss": 0.9172, + "step": 1391 + }, + { + "epoch": 0.23338572775856647, + "grad_norm": 0.406104177236557, + "learning_rate": 1.792098160905829e-05, + "loss": 0.9602, + "step": 1392 + }, + { + "epoch": 0.2335533899193126, + "grad_norm": 0.6425048112869263, + "learning_rate": 1.7917665647010003e-05, + "loss": 0.9503, + "step": 1393 + }, + { + "epoch": 0.2337210520800587, + "grad_norm": 0.3961374759674072, + "learning_rate": 1.791434734994407e-05, + "loss": 0.9535, + "step": 1394 + }, + { + "epoch": 0.2338887142408048, + "grad_norm": 0.47615012526512146, + "learning_rate": 1.79110267188391e-05, + "loss": 0.8725, + "step": 1395 + }, + { + "epoch": 0.2340563764015509, + "grad_norm": 0.5663509964942932, + "learning_rate": 1.7907703754674392e-05, + "loss": 0.8807, + "step": 1396 + }, + { + "epoch": 0.23422403856229698, + "grad_norm": 0.4356711804866791, + "learning_rate": 1.790437845842993e-05, + "loss": 0.9362, + "step": 1397 + }, + { + "epoch": 0.23439170072304308, + "grad_norm": 0.5933541655540466, + "learning_rate": 1.7901050831086377e-05, + "loss": 0.9572, + "step": 1398 + }, + { + "epoch": 0.23455936288378917, + "grad_norm": 1.4825639724731445, + "learning_rate": 1.78977208736251e-05, + "loss": 0.862, + "step": 1399 + }, + { + "epoch": 0.23472702504453527, + "grad_norm": 0.38333049416542053, + "learning_rate": 1.789438858702814e-05, + "loss": 0.9109, + "step": 1400 + }, + { + "epoch": 0.23489468720528137, + "grad_norm": 0.40966543555259705, + "learning_rate": 1.789105397227823e-05, + "loss": 0.8819, + "step": 1401 + }, + { + "epoch": 0.23506234936602746, + "grad_norm": 0.5122490525245667, + "learning_rate": 1.7887717030358794e-05, + "loss": 0.9017, + "step": 1402 + }, + { + "epoch": 0.23523001152677356, + "grad_norm": 0.6329760551452637, + "learning_rate": 1.788437776225393e-05, + "loss": 0.8879, + "step": 1403 + }, + { + "epoch": 0.23539767368751965, + "grad_norm": 0.3476681113243103, + "learning_rate": 1.7881036168948435e-05, + "loss": 0.84, + "step": 1404 + }, + { + "epoch": 0.23556533584826575, + "grad_norm": 0.44182735681533813, + "learning_rate": 1.7877692251427783e-05, + "loss": 0.7926, + "step": 1405 + }, + { + "epoch": 0.23573299800901185, + "grad_norm": 0.4372468590736389, + "learning_rate": 1.787434601067814e-05, + "loss": 0.8847, + "step": 1406 + }, + { + "epoch": 0.23590066016975794, + "grad_norm": 0.4955695867538452, + "learning_rate": 1.787099744768635e-05, + "loss": 0.9369, + "step": 1407 + }, + { + "epoch": 0.23606832233050404, + "grad_norm": 0.7181636095046997, + "learning_rate": 1.7867646563439947e-05, + "loss": 0.9023, + "step": 1408 + }, + { + "epoch": 0.23623598449125013, + "grad_norm": 0.5251885652542114, + "learning_rate": 1.786429335892715e-05, + "loss": 0.8992, + "step": 1409 + }, + { + "epoch": 0.23640364665199623, + "grad_norm": 0.5547168254852295, + "learning_rate": 1.786093783513686e-05, + "loss": 0.9733, + "step": 1410 + }, + { + "epoch": 0.23657130881274233, + "grad_norm": 0.6549569964408875, + "learning_rate": 1.7857579993058662e-05, + "loss": 0.8548, + "step": 1411 + }, + { + "epoch": 0.23673897097348842, + "grad_norm": 0.6456899046897888, + "learning_rate": 1.7854219833682825e-05, + "loss": 0.9847, + "step": 1412 + }, + { + "epoch": 0.23690663313423452, + "grad_norm": 0.3912792503833771, + "learning_rate": 1.78508573580003e-05, + "loss": 0.8897, + "step": 1413 + }, + { + "epoch": 0.2370742952949806, + "grad_norm": 0.43298226594924927, + "learning_rate": 1.7847492567002723e-05, + "loss": 0.9704, + "step": 1414 + }, + { + "epoch": 0.2372419574557267, + "grad_norm": 0.5298269987106323, + "learning_rate": 1.784412546168242e-05, + "loss": 0.9295, + "step": 1415 + }, + { + "epoch": 0.2374096196164728, + "grad_norm": 0.4604285657405853, + "learning_rate": 1.784075604303238e-05, + "loss": 0.9533, + "step": 1416 + }, + { + "epoch": 0.2375772817772189, + "grad_norm": 0.6257635354995728, + "learning_rate": 1.7837384312046297e-05, + "loss": 0.9461, + "step": 1417 + }, + { + "epoch": 0.237744943937965, + "grad_norm": 0.4810788631439209, + "learning_rate": 1.7834010269718526e-05, + "loss": 0.9202, + "step": 1418 + }, + { + "epoch": 0.2379126060987111, + "grad_norm": 0.6605314016342163, + "learning_rate": 1.7830633917044125e-05, + "loss": 0.9729, + "step": 1419 + }, + { + "epoch": 0.2380802682594572, + "grad_norm": 0.3528065085411072, + "learning_rate": 1.7827255255018813e-05, + "loss": 0.8712, + "step": 1420 + }, + { + "epoch": 0.23824793042020329, + "grad_norm": 0.4199623167514801, + "learning_rate": 1.7823874284639004e-05, + "loss": 0.9097, + "step": 1421 + }, + { + "epoch": 0.23841559258094938, + "grad_norm": 0.5196505188941956, + "learning_rate": 1.782049100690179e-05, + "loss": 0.9383, + "step": 1422 + }, + { + "epoch": 0.23858325474169548, + "grad_norm": 0.5818756818771362, + "learning_rate": 1.781710542280493e-05, + "loss": 0.8194, + "step": 1423 + }, + { + "epoch": 0.23875091690244157, + "grad_norm": 0.43274083733558655, + "learning_rate": 1.781371753334689e-05, + "loss": 0.9057, + "step": 1424 + }, + { + "epoch": 0.23891857906318767, + "grad_norm": 0.35609766840934753, + "learning_rate": 1.7810327339526792e-05, + "loss": 0.8873, + "step": 1425 + }, + { + "epoch": 0.23908624122393377, + "grad_norm": 0.48200127482414246, + "learning_rate": 1.7806934842344444e-05, + "loss": 0.896, + "step": 1426 + }, + { + "epoch": 0.23925390338467986, + "grad_norm": 0.6076840758323669, + "learning_rate": 1.7803540042800347e-05, + "loss": 0.838, + "step": 1427 + }, + { + "epoch": 0.23942156554542596, + "grad_norm": 0.6591511964797974, + "learning_rate": 1.7800142941895658e-05, + "loss": 0.945, + "step": 1428 + }, + { + "epoch": 0.23958922770617205, + "grad_norm": 0.4139047861099243, + "learning_rate": 1.7796743540632226e-05, + "loss": 0.8899, + "step": 1429 + }, + { + "epoch": 0.23975688986691815, + "grad_norm": 0.5683903098106384, + "learning_rate": 1.779334184001258e-05, + "loss": 0.892, + "step": 1430 + }, + { + "epoch": 0.23992455202766425, + "grad_norm": 0.4455464482307434, + "learning_rate": 1.778993784103992e-05, + "loss": 0.9469, + "step": 1431 + }, + { + "epoch": 0.24009221418841034, + "grad_norm": 0.42767488956451416, + "learning_rate": 1.778653154471813e-05, + "loss": 0.887, + "step": 1432 + }, + { + "epoch": 0.24025987634915644, + "grad_norm": 0.6215432286262512, + "learning_rate": 1.778312295205177e-05, + "loss": 0.8905, + "step": 1433 + }, + { + "epoch": 0.24042753850990253, + "grad_norm": 0.4767569899559021, + "learning_rate": 1.7779712064046073e-05, + "loss": 1.0294, + "step": 1434 + }, + { + "epoch": 0.24059520067064866, + "grad_norm": 0.5109009146690369, + "learning_rate": 1.7776298881706955e-05, + "loss": 0.9222, + "step": 1435 + }, + { + "epoch": 0.24076286283139475, + "grad_norm": 0.8203328251838684, + "learning_rate": 1.7772883406041002e-05, + "loss": 0.8387, + "step": 1436 + }, + { + "epoch": 0.24093052499214085, + "grad_norm": 0.5343973636627197, + "learning_rate": 1.7769465638055483e-05, + "loss": 0.926, + "step": 1437 + }, + { + "epoch": 0.24109818715288694, + "grad_norm": 1.1681573390960693, + "learning_rate": 1.776604557875834e-05, + "loss": 0.9941, + "step": 1438 + }, + { + "epoch": 0.24126584931363304, + "grad_norm": 0.5948541164398193, + "learning_rate": 1.7762623229158192e-05, + "loss": 0.9241, + "step": 1439 + }, + { + "epoch": 0.24143351147437914, + "grad_norm": 0.5291740894317627, + "learning_rate": 1.7759198590264327e-05, + "loss": 0.8958, + "step": 1440 + }, + { + "epoch": 0.24160117363512523, + "grad_norm": 0.5829247832298279, + "learning_rate": 1.775577166308672e-05, + "loss": 0.8648, + "step": 1441 + }, + { + "epoch": 0.24176883579587133, + "grad_norm": 0.43644002079963684, + "learning_rate": 1.775234244863601e-05, + "loss": 1.0055, + "step": 1442 + }, + { + "epoch": 0.24193649795661742, + "grad_norm": 0.7420687675476074, + "learning_rate": 1.7748910947923515e-05, + "loss": 0.8401, + "step": 1443 + }, + { + "epoch": 0.24210416011736352, + "grad_norm": 0.5568789839744568, + "learning_rate": 1.774547716196123e-05, + "loss": 0.9209, + "step": 1444 + }, + { + "epoch": 0.24227182227810962, + "grad_norm": 0.47366562485694885, + "learning_rate": 1.7742041091761823e-05, + "loss": 1.0516, + "step": 1445 + }, + { + "epoch": 0.2424394844388557, + "grad_norm": 0.37101438641548157, + "learning_rate": 1.7738602738338625e-05, + "loss": 0.9446, + "step": 1446 + }, + { + "epoch": 0.2426071465996018, + "grad_norm": 0.34895703196525574, + "learning_rate": 1.773516210270565e-05, + "loss": 0.855, + "step": 1447 + }, + { + "epoch": 0.2427748087603479, + "grad_norm": 0.4138221740722656, + "learning_rate": 1.7731719185877593e-05, + "loss": 0.8637, + "step": 1448 + }, + { + "epoch": 0.242942470921094, + "grad_norm": 0.44439229369163513, + "learning_rate": 1.7728273988869807e-05, + "loss": 0.9166, + "step": 1449 + }, + { + "epoch": 0.2431101330818401, + "grad_norm": 0.6057480573654175, + "learning_rate": 1.7724826512698324e-05, + "loss": 0.9553, + "step": 1450 + }, + { + "epoch": 0.2432777952425862, + "grad_norm": 0.5055689811706543, + "learning_rate": 1.7721376758379847e-05, + "loss": 0.8872, + "step": 1451 + }, + { + "epoch": 0.2434454574033323, + "grad_norm": 0.4600638151168823, + "learning_rate": 1.7717924726931754e-05, + "loss": 0.8982, + "step": 1452 + }, + { + "epoch": 0.24361311956407838, + "grad_norm": 0.4982481002807617, + "learning_rate": 1.7714470419372086e-05, + "loss": 0.9145, + "step": 1453 + }, + { + "epoch": 0.24378078172482448, + "grad_norm": 0.47119757533073425, + "learning_rate": 1.7711013836719565e-05, + "loss": 0.9157, + "step": 1454 + }, + { + "epoch": 0.24394844388557058, + "grad_norm": 0.5009276270866394, + "learning_rate": 1.770755497999358e-05, + "loss": 0.9679, + "step": 1455 + }, + { + "epoch": 0.24411610604631667, + "grad_norm": 0.7232698202133179, + "learning_rate": 1.770409385021419e-05, + "loss": 0.8615, + "step": 1456 + }, + { + "epoch": 0.24428376820706277, + "grad_norm": 0.4853869676589966, + "learning_rate": 1.7700630448402125e-05, + "loss": 1.0043, + "step": 1457 + }, + { + "epoch": 0.24445143036780886, + "grad_norm": 0.5485565066337585, + "learning_rate": 1.7697164775578785e-05, + "loss": 0.8516, + "step": 1458 + }, + { + "epoch": 0.24461909252855496, + "grad_norm": 0.6456785798072815, + "learning_rate": 1.769369683276624e-05, + "loss": 0.9454, + "step": 1459 + }, + { + "epoch": 0.24478675468930106, + "grad_norm": 0.5416443347930908, + "learning_rate": 1.7690226620987233e-05, + "loss": 0.9289, + "step": 1460 + }, + { + "epoch": 0.24495441685004715, + "grad_norm": 0.5395210981369019, + "learning_rate": 1.7686754141265162e-05, + "loss": 0.9846, + "step": 1461 + }, + { + "epoch": 0.24512207901079325, + "grad_norm": 0.4691596031188965, + "learning_rate": 1.768327939462412e-05, + "loss": 0.8755, + "step": 1462 + }, + { + "epoch": 0.24528974117153934, + "grad_norm": 0.4843076169490814, + "learning_rate": 1.7679802382088836e-05, + "loss": 0.9427, + "step": 1463 + }, + { + "epoch": 0.24545740333228544, + "grad_norm": 0.4803543984889984, + "learning_rate": 1.7676323104684736e-05, + "loss": 1.0253, + "step": 1464 + }, + { + "epoch": 0.24562506549303154, + "grad_norm": 0.5563539862632751, + "learning_rate": 1.7672841563437897e-05, + "loss": 0.8087, + "step": 1465 + }, + { + "epoch": 0.24579272765377763, + "grad_norm": 0.46070781350135803, + "learning_rate": 1.766935775937507e-05, + "loss": 0.8593, + "step": 1466 + }, + { + "epoch": 0.24596038981452373, + "grad_norm": 0.8026439547538757, + "learning_rate": 1.766587169352367e-05, + "loss": 0.831, + "step": 1467 + }, + { + "epoch": 0.24612805197526982, + "grad_norm": 0.6293075680732727, + "learning_rate": 1.7662383366911785e-05, + "loss": 0.8819, + "step": 1468 + }, + { + "epoch": 0.24629571413601592, + "grad_norm": 0.630046546459198, + "learning_rate": 1.765889278056816e-05, + "loss": 0.8985, + "step": 1469 + }, + { + "epoch": 0.24646337629676202, + "grad_norm": 0.5543246865272522, + "learning_rate": 1.7655399935522216e-05, + "loss": 0.8879, + "step": 1470 + }, + { + "epoch": 0.2466310384575081, + "grad_norm": 0.440186083316803, + "learning_rate": 1.7651904832804038e-05, + "loss": 0.88, + "step": 1471 + }, + { + "epoch": 0.2467987006182542, + "grad_norm": 0.6598295569419861, + "learning_rate": 1.764840747344437e-05, + "loss": 0.9657, + "step": 1472 + }, + { + "epoch": 0.2469663627790003, + "grad_norm": 0.8729094862937927, + "learning_rate": 1.764490785847463e-05, + "loss": 0.8873, + "step": 1473 + }, + { + "epoch": 0.2471340249397464, + "grad_norm": 0.523439884185791, + "learning_rate": 1.76414059889269e-05, + "loss": 0.9577, + "step": 1474 + }, + { + "epoch": 0.2473016871004925, + "grad_norm": 0.41881799697875977, + "learning_rate": 1.7637901865833923e-05, + "loss": 0.9237, + "step": 1475 + }, + { + "epoch": 0.2474693492612386, + "grad_norm": 0.5430641770362854, + "learning_rate": 1.76343954902291e-05, + "loss": 0.8916, + "step": 1476 + }, + { + "epoch": 0.2476370114219847, + "grad_norm": 0.4627753794193268, + "learning_rate": 1.763088686314652e-05, + "loss": 1.0475, + "step": 1477 + }, + { + "epoch": 0.2478046735827308, + "grad_norm": 0.40650758147239685, + "learning_rate": 1.7627375985620902e-05, + "loss": 0.8889, + "step": 1478 + }, + { + "epoch": 0.2479723357434769, + "grad_norm": 0.31283724308013916, + "learning_rate": 1.7623862858687665e-05, + "loss": 0.8052, + "step": 1479 + }, + { + "epoch": 0.248139997904223, + "grad_norm": 0.4834359288215637, + "learning_rate": 1.762034748338286e-05, + "loss": 0.9563, + "step": 1480 + }, + { + "epoch": 0.2483076600649691, + "grad_norm": 0.4991844892501831, + "learning_rate": 1.761682986074323e-05, + "loss": 0.9546, + "step": 1481 + }, + { + "epoch": 0.2484753222257152, + "grad_norm": 0.5470111966133118, + "learning_rate": 1.761330999180615e-05, + "loss": 0.9472, + "step": 1482 + }, + { + "epoch": 0.2486429843864613, + "grad_norm": 0.4790292978286743, + "learning_rate": 1.7609787877609678e-05, + "loss": 0.919, + "step": 1483 + }, + { + "epoch": 0.2488106465472074, + "grad_norm": 0.49103057384490967, + "learning_rate": 1.7606263519192535e-05, + "loss": 0.8835, + "step": 1484 + }, + { + "epoch": 0.24897830870795348, + "grad_norm": 0.5295326709747314, + "learning_rate": 1.760273691759409e-05, + "loss": 0.9874, + "step": 1485 + }, + { + "epoch": 0.24914597086869958, + "grad_norm": 0.3911377489566803, + "learning_rate": 1.7599208073854383e-05, + "loss": 0.8222, + "step": 1486 + }, + { + "epoch": 0.24931363302944567, + "grad_norm": 0.5055460333824158, + "learning_rate": 1.759567698901412e-05, + "loss": 0.9445, + "step": 1487 + }, + { + "epoch": 0.24948129519019177, + "grad_norm": 0.6002017855644226, + "learning_rate": 1.7592143664114656e-05, + "loss": 0.9085, + "step": 1488 + }, + { + "epoch": 0.24964895735093787, + "grad_norm": 0.4354606568813324, + "learning_rate": 1.758860810019801e-05, + "loss": 0.8811, + "step": 1489 + }, + { + "epoch": 0.24981661951168396, + "grad_norm": 0.40272077918052673, + "learning_rate": 1.758507029830687e-05, + "loss": 0.8248, + "step": 1490 + }, + { + "epoch": 0.24998428167243006, + "grad_norm": 0.4683881103992462, + "learning_rate": 1.7581530259484572e-05, + "loss": 0.9636, + "step": 1491 + }, + { + "epoch": 0.2501519438331761, + "grad_norm": 0.501693606376648, + "learning_rate": 1.7577987984775125e-05, + "loss": 0.9065, + "step": 1492 + }, + { + "epoch": 0.2503196059939222, + "grad_norm": 0.5398340821266174, + "learning_rate": 1.7574443475223183e-05, + "loss": 0.904, + "step": 1493 + }, + { + "epoch": 0.2504872681546683, + "grad_norm": 0.404601126909256, + "learning_rate": 1.757089673187406e-05, + "loss": 0.9034, + "step": 1494 + }, + { + "epoch": 0.2506549303154144, + "grad_norm": 0.5416507124900818, + "learning_rate": 1.7567347755773752e-05, + "loss": 0.9015, + "step": 1495 + }, + { + "epoch": 0.2508225924761605, + "grad_norm": 0.611964225769043, + "learning_rate": 1.756379654796888e-05, + "loss": 0.8833, + "step": 1496 + }, + { + "epoch": 0.2509902546369066, + "grad_norm": 0.4463139474391937, + "learning_rate": 1.7560243109506746e-05, + "loss": 0.9851, + "step": 1497 + }, + { + "epoch": 0.2511579167976527, + "grad_norm": 0.547607958316803, + "learning_rate": 1.7556687441435302e-05, + "loss": 0.9473, + "step": 1498 + }, + { + "epoch": 0.2513255789583988, + "grad_norm": 0.5264208912849426, + "learning_rate": 1.7553129544803158e-05, + "loss": 0.8994, + "step": 1499 + }, + { + "epoch": 0.25149324111914495, + "grad_norm": 0.4187546372413635, + "learning_rate": 1.754956942065958e-05, + "loss": 0.9347, + "step": 1500 + }, + { + "epoch": 0.25166090327989105, + "grad_norm": 0.4960138201713562, + "learning_rate": 1.7546007070054497e-05, + "loss": 0.8381, + "step": 1501 + }, + { + "epoch": 0.25182856544063714, + "grad_norm": 0.7610869407653809, + "learning_rate": 1.754244249403849e-05, + "loss": 0.9455, + "step": 1502 + }, + { + "epoch": 0.25199622760138324, + "grad_norm": 0.4258002042770386, + "learning_rate": 1.7538875693662793e-05, + "loss": 0.9081, + "step": 1503 + }, + { + "epoch": 0.25216388976212933, + "grad_norm": 0.6918498873710632, + "learning_rate": 1.7535306669979305e-05, + "loss": 0.8242, + "step": 1504 + }, + { + "epoch": 0.25233155192287543, + "grad_norm": 0.6293070912361145, + "learning_rate": 1.753173542404057e-05, + "loss": 0.9185, + "step": 1505 + }, + { + "epoch": 0.2524992140836215, + "grad_norm": 0.5741620063781738, + "learning_rate": 1.752816195689979e-05, + "loss": 0.9219, + "step": 1506 + }, + { + "epoch": 0.2526668762443676, + "grad_norm": 0.49497202038764954, + "learning_rate": 1.752458626961083e-05, + "loss": 0.9218, + "step": 1507 + }, + { + "epoch": 0.2528345384051137, + "grad_norm": 0.7709310054779053, + "learning_rate": 1.7521008363228208e-05, + "loss": 0.9209, + "step": 1508 + }, + { + "epoch": 0.2530022005658598, + "grad_norm": 0.7194563746452332, + "learning_rate": 1.7517428238807085e-05, + "loss": 0.8209, + "step": 1509 + }, + { + "epoch": 0.2531698627266059, + "grad_norm": 0.6282023191452026, + "learning_rate": 1.751384589740329e-05, + "loss": 0.9013, + "step": 1510 + }, + { + "epoch": 0.253337524887352, + "grad_norm": 0.4894861876964569, + "learning_rate": 1.7510261340073297e-05, + "loss": 0.8998, + "step": 1511 + }, + { + "epoch": 0.2535051870480981, + "grad_norm": 0.9297739863395691, + "learning_rate": 1.7506674567874234e-05, + "loss": 0.9283, + "step": 1512 + }, + { + "epoch": 0.2536728492088442, + "grad_norm": 0.586509644985199, + "learning_rate": 1.7503085581863887e-05, + "loss": 0.9808, + "step": 1513 + }, + { + "epoch": 0.2538405113695903, + "grad_norm": 0.3984715938568115, + "learning_rate": 1.7499494383100697e-05, + "loss": 0.8636, + "step": 1514 + }, + { + "epoch": 0.2540081735303364, + "grad_norm": 0.522193431854248, + "learning_rate": 1.7495900972643745e-05, + "loss": 0.9593, + "step": 1515 + }, + { + "epoch": 0.2541758356910825, + "grad_norm": 0.5000961422920227, + "learning_rate": 1.7492305351552777e-05, + "loss": 1.0069, + "step": 1516 + }, + { + "epoch": 0.2543434978518286, + "grad_norm": 0.42993372678756714, + "learning_rate": 1.7488707520888183e-05, + "loss": 0.9476, + "step": 1517 + }, + { + "epoch": 0.2545111600125747, + "grad_norm": 0.5267705917358398, + "learning_rate": 1.7485107481711014e-05, + "loss": 0.8771, + "step": 1518 + }, + { + "epoch": 0.2546788221733208, + "grad_norm": 0.4390380084514618, + "learning_rate": 1.7481505235082958e-05, + "loss": 0.8721, + "step": 1519 + }, + { + "epoch": 0.25484648433406687, + "grad_norm": 0.5014662742614746, + "learning_rate": 1.7477900782066365e-05, + "loss": 0.9384, + "step": 1520 + }, + { + "epoch": 0.25501414649481297, + "grad_norm": 0.42803215980529785, + "learning_rate": 1.747429412372424e-05, + "loss": 0.8852, + "step": 1521 + }, + { + "epoch": 0.25518180865555906, + "grad_norm": 0.7259057760238647, + "learning_rate": 1.747068526112022e-05, + "loss": 0.9101, + "step": 1522 + }, + { + "epoch": 0.25534947081630516, + "grad_norm": 0.8499658107757568, + "learning_rate": 1.746707419531861e-05, + "loss": 0.87, + "step": 1523 + }, + { + "epoch": 0.25551713297705125, + "grad_norm": 0.6882765889167786, + "learning_rate": 1.7463460927384356e-05, + "loss": 0.9239, + "step": 1524 + }, + { + "epoch": 0.25568479513779735, + "grad_norm": 0.6137887239456177, + "learning_rate": 1.745984545838306e-05, + "loss": 1.059, + "step": 1525 + }, + { + "epoch": 0.25585245729854345, + "grad_norm": 0.5296462774276733, + "learning_rate": 1.745622778938096e-05, + "loss": 0.9238, + "step": 1526 + }, + { + "epoch": 0.25602011945928954, + "grad_norm": 0.42972806096076965, + "learning_rate": 1.7452607921444956e-05, + "loss": 0.8551, + "step": 1527 + }, + { + "epoch": 0.25618778162003564, + "grad_norm": 0.4224143922328949, + "learning_rate": 1.744898585564259e-05, + "loss": 0.8837, + "step": 1528 + }, + { + "epoch": 0.25635544378078173, + "grad_norm": 0.4010677635669708, + "learning_rate": 1.744536159304206e-05, + "loss": 0.8855, + "step": 1529 + }, + { + "epoch": 0.25652310594152783, + "grad_norm": 0.6350672841072083, + "learning_rate": 1.74417351347122e-05, + "loss": 0.8995, + "step": 1530 + }, + { + "epoch": 0.2566907681022739, + "grad_norm": 0.5510476231575012, + "learning_rate": 1.7438106481722495e-05, + "loss": 0.9429, + "step": 1531 + }, + { + "epoch": 0.25685843026302, + "grad_norm": 0.5789257884025574, + "learning_rate": 1.743447563514309e-05, + "loss": 0.8655, + "step": 1532 + }, + { + "epoch": 0.2570260924237661, + "grad_norm": 0.5662457346916199, + "learning_rate": 1.743084259604476e-05, + "loss": 0.8362, + "step": 1533 + }, + { + "epoch": 0.2571937545845122, + "grad_norm": 0.44558578729629517, + "learning_rate": 1.742720736549893e-05, + "loss": 0.8342, + "step": 1534 + }, + { + "epoch": 0.2573614167452583, + "grad_norm": 0.4779844880104065, + "learning_rate": 1.7423569944577677e-05, + "loss": 0.8702, + "step": 1535 + }, + { + "epoch": 0.2575290789060044, + "grad_norm": 0.41661155223846436, + "learning_rate": 1.7419930334353728e-05, + "loss": 0.947, + "step": 1536 + }, + { + "epoch": 0.2576967410667505, + "grad_norm": 0.5755490064620972, + "learning_rate": 1.741628853590044e-05, + "loss": 0.9165, + "step": 1537 + }, + { + "epoch": 0.2578644032274966, + "grad_norm": 0.4679485261440277, + "learning_rate": 1.741264455029183e-05, + "loss": 0.847, + "step": 1538 + }, + { + "epoch": 0.2580320653882427, + "grad_norm": 0.5541486144065857, + "learning_rate": 1.7408998378602555e-05, + "loss": 0.9664, + "step": 1539 + }, + { + "epoch": 0.2581997275489888, + "grad_norm": 0.45861151814460754, + "learning_rate": 1.7405350021907912e-05, + "loss": 0.8682, + "step": 1540 + }, + { + "epoch": 0.2583673897097349, + "grad_norm": 0.3169395923614502, + "learning_rate": 1.7401699481283847e-05, + "loss": 0.8557, + "step": 1541 + }, + { + "epoch": 0.258535051870481, + "grad_norm": 0.5471020340919495, + "learning_rate": 1.739804675780695e-05, + "loss": 0.9523, + "step": 1542 + }, + { + "epoch": 0.2587027140312271, + "grad_norm": 0.6324790120124817, + "learning_rate": 1.7394391852554456e-05, + "loss": 0.9002, + "step": 1543 + }, + { + "epoch": 0.2588703761919732, + "grad_norm": 0.4142170250415802, + "learning_rate": 1.7390734766604244e-05, + "loss": 0.938, + "step": 1544 + }, + { + "epoch": 0.25903803835271927, + "grad_norm": 0.633558988571167, + "learning_rate": 1.738707550103483e-05, + "loss": 0.8781, + "step": 1545 + }, + { + "epoch": 0.25920570051346536, + "grad_norm": 0.7527374029159546, + "learning_rate": 1.738341405692538e-05, + "loss": 0.8925, + "step": 1546 + }, + { + "epoch": 0.25937336267421146, + "grad_norm": 0.6943590044975281, + "learning_rate": 1.7379750435355696e-05, + "loss": 0.8406, + "step": 1547 + }, + { + "epoch": 0.25954102483495756, + "grad_norm": 0.5189474821090698, + "learning_rate": 1.7376084637406222e-05, + "loss": 0.9907, + "step": 1548 + }, + { + "epoch": 0.25970868699570365, + "grad_norm": 0.4832892417907715, + "learning_rate": 1.737241666415806e-05, + "loss": 0.9535, + "step": 1549 + }, + { + "epoch": 0.25987634915644975, + "grad_norm": 0.5611308217048645, + "learning_rate": 1.7368746516692928e-05, + "loss": 1.023, + "step": 1550 + }, + { + "epoch": 0.26004401131719584, + "grad_norm": 0.5159692764282227, + "learning_rate": 1.73650741960932e-05, + "loss": 0.9357, + "step": 1551 + }, + { + "epoch": 0.26021167347794194, + "grad_norm": 0.5313618183135986, + "learning_rate": 1.7361399703441893e-05, + "loss": 1.044, + "step": 1552 + }, + { + "epoch": 0.26037933563868804, + "grad_norm": 0.5279186367988586, + "learning_rate": 1.735772303982266e-05, + "loss": 0.9254, + "step": 1553 + }, + { + "epoch": 0.26054699779943413, + "grad_norm": 0.6467424035072327, + "learning_rate": 1.735404420631979e-05, + "loss": 0.9476, + "step": 1554 + }, + { + "epoch": 0.26071465996018023, + "grad_norm": 0.48677462339401245, + "learning_rate": 1.7350363204018227e-05, + "loss": 0.9348, + "step": 1555 + }, + { + "epoch": 0.2608823221209263, + "grad_norm": 0.7248123288154602, + "learning_rate": 1.7346680034003528e-05, + "loss": 0.9353, + "step": 1556 + }, + { + "epoch": 0.2610499842816724, + "grad_norm": 0.41243961453437805, + "learning_rate": 1.734299469736192e-05, + "loss": 0.9441, + "step": 1557 + }, + { + "epoch": 0.2612176464424185, + "grad_norm": 0.548994779586792, + "learning_rate": 1.7339307195180248e-05, + "loss": 0.9101, + "step": 1558 + }, + { + "epoch": 0.2613853086031646, + "grad_norm": 0.5552588701248169, + "learning_rate": 1.7335617528546e-05, + "loss": 0.9204, + "step": 1559 + }, + { + "epoch": 0.2615529707639107, + "grad_norm": 0.6628120541572571, + "learning_rate": 1.733192569854731e-05, + "loss": 0.9021, + "step": 1560 + }, + { + "epoch": 0.2617206329246568, + "grad_norm": 0.5505399107933044, + "learning_rate": 1.732823170627294e-05, + "loss": 0.8775, + "step": 1561 + }, + { + "epoch": 0.2618882950854029, + "grad_norm": 0.44663092494010925, + "learning_rate": 1.7324535552812298e-05, + "loss": 0.8786, + "step": 1562 + }, + { + "epoch": 0.262055957246149, + "grad_norm": 0.47220563888549805, + "learning_rate": 1.732083723925542e-05, + "loss": 0.9145, + "step": 1563 + }, + { + "epoch": 0.2622236194068951, + "grad_norm": 0.5293234586715698, + "learning_rate": 1.731713676669299e-05, + "loss": 0.9809, + "step": 1564 + }, + { + "epoch": 0.2623912815676412, + "grad_norm": 0.35137319564819336, + "learning_rate": 1.731343413621632e-05, + "loss": 0.924, + "step": 1565 + }, + { + "epoch": 0.2625589437283873, + "grad_norm": 0.38129615783691406, + "learning_rate": 1.7309729348917364e-05, + "loss": 0.9275, + "step": 1566 + }, + { + "epoch": 0.2627266058891334, + "grad_norm": 0.5752395391464233, + "learning_rate": 1.7306022405888705e-05, + "loss": 0.9061, + "step": 1567 + }, + { + "epoch": 0.2628942680498795, + "grad_norm": 0.5390276908874512, + "learning_rate": 1.7302313308223574e-05, + "loss": 0.905, + "step": 1568 + }, + { + "epoch": 0.26306193021062557, + "grad_norm": 0.5569885969161987, + "learning_rate": 1.729860205701582e-05, + "loss": 0.9214, + "step": 1569 + }, + { + "epoch": 0.26322959237137167, + "grad_norm": 0.5757485032081604, + "learning_rate": 1.7294888653359945e-05, + "loss": 0.9978, + "step": 1570 + }, + { + "epoch": 0.26339725453211776, + "grad_norm": 0.5486928224563599, + "learning_rate": 1.7291173098351077e-05, + "loss": 0.9747, + "step": 1571 + }, + { + "epoch": 0.26356491669286386, + "grad_norm": 0.43678364157676697, + "learning_rate": 1.7287455393084972e-05, + "loss": 0.8731, + "step": 1572 + }, + { + "epoch": 0.26373257885360996, + "grad_norm": 0.47250810265541077, + "learning_rate": 1.7283735538658034e-05, + "loss": 0.8917, + "step": 1573 + }, + { + "epoch": 0.26390024101435605, + "grad_norm": 0.4039928615093231, + "learning_rate": 1.728001353616729e-05, + "loss": 0.91, + "step": 1574 + }, + { + "epoch": 0.26406790317510215, + "grad_norm": 0.47751352190971375, + "learning_rate": 1.727628938671041e-05, + "loss": 0.96, + "step": 1575 + }, + { + "epoch": 0.26423556533584824, + "grad_norm": 0.49584338068962097, + "learning_rate": 1.7272563091385694e-05, + "loss": 0.94, + "step": 1576 + }, + { + "epoch": 0.26440322749659434, + "grad_norm": 0.43514296412467957, + "learning_rate": 1.7268834651292062e-05, + "loss": 0.9175, + "step": 1577 + }, + { + "epoch": 0.26457088965734044, + "grad_norm": 0.49915921688079834, + "learning_rate": 1.726510406752908e-05, + "loss": 0.967, + "step": 1578 + }, + { + "epoch": 0.26473855181808653, + "grad_norm": 0.5309367775917053, + "learning_rate": 1.726137134119695e-05, + "loss": 0.8521, + "step": 1579 + }, + { + "epoch": 0.2649062139788326, + "grad_norm": 0.6773521900177002, + "learning_rate": 1.7257636473396496e-05, + "loss": 0.8804, + "step": 1580 + }, + { + "epoch": 0.2650738761395787, + "grad_norm": 0.5327140092849731, + "learning_rate": 1.7253899465229175e-05, + "loss": 0.8907, + "step": 1581 + }, + { + "epoch": 0.2652415383003248, + "grad_norm": 0.3631095588207245, + "learning_rate": 1.725016031779708e-05, + "loss": 0.9123, + "step": 1582 + }, + { + "epoch": 0.2654092004610709, + "grad_norm": 0.47867029905319214, + "learning_rate": 1.7246419032202926e-05, + "loss": 0.8792, + "step": 1583 + }, + { + "epoch": 0.265576862621817, + "grad_norm": 0.47533097863197327, + "learning_rate": 1.7242675609550067e-05, + "loss": 0.8526, + "step": 1584 + }, + { + "epoch": 0.26574452478256316, + "grad_norm": 0.6036980748176575, + "learning_rate": 1.723893005094249e-05, + "loss": 0.8762, + "step": 1585 + }, + { + "epoch": 0.26591218694330926, + "grad_norm": 0.6582639217376709, + "learning_rate": 1.7235182357484802e-05, + "loss": 0.779, + "step": 1586 + }, + { + "epoch": 0.26607984910405535, + "grad_norm": 0.8061196208000183, + "learning_rate": 1.7231432530282246e-05, + "loss": 0.9484, + "step": 1587 + }, + { + "epoch": 0.26624751126480145, + "grad_norm": 0.4974749982357025, + "learning_rate": 1.722768057044069e-05, + "loss": 0.9257, + "step": 1588 + }, + { + "epoch": 0.26641517342554755, + "grad_norm": 0.570415198802948, + "learning_rate": 1.7223926479066636e-05, + "loss": 0.9261, + "step": 1589 + }, + { + "epoch": 0.26658283558629364, + "grad_norm": 1.189063549041748, + "learning_rate": 1.722017025726721e-05, + "loss": 0.9155, + "step": 1590 + }, + { + "epoch": 0.26675049774703974, + "grad_norm": 0.6250978708267212, + "learning_rate": 1.721641190615017e-05, + "loss": 0.8973, + "step": 1591 + }, + { + "epoch": 0.26691815990778583, + "grad_norm": 0.4300852417945862, + "learning_rate": 1.7212651426823895e-05, + "loss": 0.8577, + "step": 1592 + }, + { + "epoch": 0.26708582206853193, + "grad_norm": 0.3604830503463745, + "learning_rate": 1.7208888820397404e-05, + "loss": 0.8551, + "step": 1593 + }, + { + "epoch": 0.267253484229278, + "grad_norm": 0.5577496886253357, + "learning_rate": 1.7205124087980334e-05, + "loss": 1.0528, + "step": 1594 + }, + { + "epoch": 0.2674211463900241, + "grad_norm": 0.5893198251724243, + "learning_rate": 1.7201357230682952e-05, + "loss": 0.8834, + "step": 1595 + }, + { + "epoch": 0.2675888085507702, + "grad_norm": 0.48370787501335144, + "learning_rate": 1.7197588249616147e-05, + "loss": 0.9746, + "step": 1596 + }, + { + "epoch": 0.2677564707115163, + "grad_norm": 0.8386217355728149, + "learning_rate": 1.719381714589144e-05, + "loss": 0.8626, + "step": 1597 + }, + { + "epoch": 0.2679241328722624, + "grad_norm": 0.5362548828125, + "learning_rate": 1.7190043920620982e-05, + "loss": 0.9499, + "step": 1598 + }, + { + "epoch": 0.2680917950330085, + "grad_norm": 0.5042610764503479, + "learning_rate": 1.7186268574917536e-05, + "loss": 0.9054, + "step": 1599 + }, + { + "epoch": 0.2682594571937546, + "grad_norm": 0.40954241156578064, + "learning_rate": 1.71824911098945e-05, + "loss": 0.8414, + "step": 1600 + }, + { + "epoch": 0.2684271193545007, + "grad_norm": 0.8121725916862488, + "learning_rate": 1.7178711526665895e-05, + "loss": 0.8461, + "step": 1601 + }, + { + "epoch": 0.2685947815152468, + "grad_norm": 0.4340389370918274, + "learning_rate": 1.7174929826346373e-05, + "loss": 0.8011, + "step": 1602 + }, + { + "epoch": 0.2687624436759929, + "grad_norm": 0.7912620902061462, + "learning_rate": 1.7171146010051196e-05, + "loss": 0.8803, + "step": 1603 + }, + { + "epoch": 0.268930105836739, + "grad_norm": 0.4555317461490631, + "learning_rate": 1.716736007889626e-05, + "loss": 0.895, + "step": 1604 + }, + { + "epoch": 0.2690977679974851, + "grad_norm": 0.5665902495384216, + "learning_rate": 1.716357203399809e-05, + "loss": 1.0344, + "step": 1605 + }, + { + "epoch": 0.2692654301582312, + "grad_norm": 0.6170781850814819, + "learning_rate": 1.7159781876473816e-05, + "loss": 0.9448, + "step": 1606 + }, + { + "epoch": 0.2694330923189773, + "grad_norm": 0.657977819442749, + "learning_rate": 1.715598960744121e-05, + "loss": 0.9308, + "step": 1607 + }, + { + "epoch": 0.26960075447972337, + "grad_norm": 0.6097431778907776, + "learning_rate": 1.7152195228018662e-05, + "loss": 0.915, + "step": 1608 + }, + { + "epoch": 0.26976841664046947, + "grad_norm": 0.4871729016304016, + "learning_rate": 1.714839873932517e-05, + "loss": 0.8911, + "step": 1609 + }, + { + "epoch": 0.26993607880121556, + "grad_norm": 0.41206493973731995, + "learning_rate": 1.714460014248038e-05, + "loss": 0.8257, + "step": 1610 + }, + { + "epoch": 0.27010374096196166, + "grad_norm": 0.451313316822052, + "learning_rate": 1.7140799438604533e-05, + "loss": 0.9392, + "step": 1611 + }, + { + "epoch": 0.27027140312270775, + "grad_norm": 0.47345075011253357, + "learning_rate": 1.713699662881851e-05, + "loss": 0.8944, + "step": 1612 + }, + { + "epoch": 0.27043906528345385, + "grad_norm": 0.39391496777534485, + "learning_rate": 1.7133191714243805e-05, + "loss": 0.9162, + "step": 1613 + }, + { + "epoch": 0.27060672744419995, + "grad_norm": 0.5918540358543396, + "learning_rate": 1.7129384696002533e-05, + "loss": 0.9649, + "step": 1614 + }, + { + "epoch": 0.27077438960494604, + "grad_norm": 0.8174153566360474, + "learning_rate": 1.712557557521743e-05, + "loss": 0.9903, + "step": 1615 + }, + { + "epoch": 0.27094205176569214, + "grad_norm": 0.5288435220718384, + "learning_rate": 1.712176435301186e-05, + "loss": 0.9516, + "step": 1616 + }, + { + "epoch": 0.27110971392643823, + "grad_norm": 0.46600520610809326, + "learning_rate": 1.7117951030509795e-05, + "loss": 0.9965, + "step": 1617 + }, + { + "epoch": 0.27127737608718433, + "grad_norm": 0.4489824175834656, + "learning_rate": 1.7114135608835826e-05, + "loss": 0.9398, + "step": 1618 + }, + { + "epoch": 0.2714450382479304, + "grad_norm": 0.5334166884422302, + "learning_rate": 1.7110318089115178e-05, + "loss": 0.9217, + "step": 1619 + }, + { + "epoch": 0.2716127004086765, + "grad_norm": 0.5517166256904602, + "learning_rate": 1.7106498472473674e-05, + "loss": 0.8955, + "step": 1620 + }, + { + "epoch": 0.2717803625694226, + "grad_norm": 0.37127768993377686, + "learning_rate": 1.710267676003778e-05, + "loss": 0.8803, + "step": 1621 + }, + { + "epoch": 0.2719480247301687, + "grad_norm": 0.44107526540756226, + "learning_rate": 1.7098852952934555e-05, + "loss": 0.9826, + "step": 1622 + }, + { + "epoch": 0.2721156868909148, + "grad_norm": 0.5912244915962219, + "learning_rate": 1.709502705229169e-05, + "loss": 0.9638, + "step": 1623 + }, + { + "epoch": 0.2722833490516609, + "grad_norm": 0.4062788486480713, + "learning_rate": 1.709119905923749e-05, + "loss": 0.8838, + "step": 1624 + }, + { + "epoch": 0.272451011212407, + "grad_norm": 0.4899587035179138, + "learning_rate": 1.7087368974900876e-05, + "loss": 0.9641, + "step": 1625 + }, + { + "epoch": 0.2726186733731531, + "grad_norm": 0.3973456919193268, + "learning_rate": 1.7083536800411392e-05, + "loss": 0.919, + "step": 1626 + }, + { + "epoch": 0.2727863355338992, + "grad_norm": 0.5032211542129517, + "learning_rate": 1.7079702536899193e-05, + "loss": 0.8664, + "step": 1627 + }, + { + "epoch": 0.2729539976946453, + "grad_norm": 0.8235628604888916, + "learning_rate": 1.7075866185495046e-05, + "loss": 0.9643, + "step": 1628 + }, + { + "epoch": 0.2731216598553914, + "grad_norm": 0.4531996548175812, + "learning_rate": 1.7072027747330348e-05, + "loss": 0.8904, + "step": 1629 + }, + { + "epoch": 0.2732893220161375, + "grad_norm": 0.44512197375297546, + "learning_rate": 1.706818722353709e-05, + "loss": 0.9208, + "step": 1630 + }, + { + "epoch": 0.2734569841768836, + "grad_norm": 0.5449144244194031, + "learning_rate": 1.70643446152479e-05, + "loss": 0.92, + "step": 1631 + }, + { + "epoch": 0.2736246463376297, + "grad_norm": 0.3762809932231903, + "learning_rate": 1.7060499923596002e-05, + "loss": 0.9175, + "step": 1632 + }, + { + "epoch": 0.27379230849837577, + "grad_norm": 0.5805980563163757, + "learning_rate": 1.705665314971525e-05, + "loss": 0.8536, + "step": 1633 + }, + { + "epoch": 0.27395997065912187, + "grad_norm": 0.5688173174858093, + "learning_rate": 1.70528042947401e-05, + "loss": 0.914, + "step": 1634 + }, + { + "epoch": 0.27412763281986796, + "grad_norm": 0.6951828598976135, + "learning_rate": 1.7048953359805637e-05, + "loss": 0.9056, + "step": 1635 + }, + { + "epoch": 0.27429529498061406, + "grad_norm": 0.5278764367103577, + "learning_rate": 1.7045100346047535e-05, + "loss": 0.9174, + "step": 1636 + }, + { + "epoch": 0.27446295714136015, + "grad_norm": 0.42395704984664917, + "learning_rate": 1.7041245254602104e-05, + "loss": 1.0499, + "step": 1637 + }, + { + "epoch": 0.27463061930210625, + "grad_norm": 0.7272180318832397, + "learning_rate": 1.703738808660626e-05, + "loss": 0.8348, + "step": 1638 + }, + { + "epoch": 0.27479828146285235, + "grad_norm": 0.7003971338272095, + "learning_rate": 1.7033528843197523e-05, + "loss": 0.9273, + "step": 1639 + }, + { + "epoch": 0.27496594362359844, + "grad_norm": 0.47901052236557007, + "learning_rate": 1.702966752551404e-05, + "loss": 0.839, + "step": 1640 + }, + { + "epoch": 0.27513360578434454, + "grad_norm": 0.3797691762447357, + "learning_rate": 1.7025804134694552e-05, + "loss": 0.8958, + "step": 1641 + }, + { + "epoch": 0.27530126794509063, + "grad_norm": 0.4849264323711395, + "learning_rate": 1.7021938671878427e-05, + "loss": 0.8831, + "step": 1642 + }, + { + "epoch": 0.27546893010583673, + "grad_norm": 0.3957948088645935, + "learning_rate": 1.701807113820564e-05, + "loss": 0.9219, + "step": 1643 + }, + { + "epoch": 0.2756365922665828, + "grad_norm": 0.46925845742225647, + "learning_rate": 1.7014201534816765e-05, + "loss": 0.9282, + "step": 1644 + }, + { + "epoch": 0.2758042544273289, + "grad_norm": 0.5414051413536072, + "learning_rate": 1.7010329862853007e-05, + "loss": 0.9243, + "step": 1645 + }, + { + "epoch": 0.275971916588075, + "grad_norm": 1.0770642757415771, + "learning_rate": 1.700645612345616e-05, + "loss": 0.9393, + "step": 1646 + }, + { + "epoch": 0.2761395787488211, + "grad_norm": 0.4121866226196289, + "learning_rate": 1.7002580317768646e-05, + "loss": 0.9172, + "step": 1647 + }, + { + "epoch": 0.2763072409095672, + "grad_norm": 0.6341374516487122, + "learning_rate": 1.699870244693348e-05, + "loss": 0.8848, + "step": 1648 + }, + { + "epoch": 0.2764749030703133, + "grad_norm": 0.8026604056358337, + "learning_rate": 1.6994822512094306e-05, + "loss": 0.8162, + "step": 1649 + }, + { + "epoch": 0.2766425652310594, + "grad_norm": 0.78305983543396, + "learning_rate": 1.6990940514395353e-05, + "loss": 0.8461, + "step": 1650 + }, + { + "epoch": 0.2768102273918055, + "grad_norm": 0.7583958506584167, + "learning_rate": 1.6987056454981476e-05, + "loss": 0.9291, + "step": 1651 + }, + { + "epoch": 0.2769778895525516, + "grad_norm": 0.5095685720443726, + "learning_rate": 1.698317033499813e-05, + "loss": 1.0257, + "step": 1652 + }, + { + "epoch": 0.2771455517132977, + "grad_norm": 0.8323339223861694, + "learning_rate": 1.697928215559138e-05, + "loss": 0.9488, + "step": 1653 + }, + { + "epoch": 0.2773132138740438, + "grad_norm": 0.6618877649307251, + "learning_rate": 1.6975391917907903e-05, + "loss": 0.9038, + "step": 1654 + }, + { + "epoch": 0.2774808760347899, + "grad_norm": 0.43667611479759216, + "learning_rate": 1.6971499623094972e-05, + "loss": 0.9315, + "step": 1655 + }, + { + "epoch": 0.277648538195536, + "grad_norm": 0.3366505801677704, + "learning_rate": 1.6967605272300476e-05, + "loss": 0.8708, + "step": 1656 + }, + { + "epoch": 0.2778162003562821, + "grad_norm": 0.6763900518417358, + "learning_rate": 1.6963708866672907e-05, + "loss": 0.863, + "step": 1657 + }, + { + "epoch": 0.27798386251702817, + "grad_norm": 0.41367295384407043, + "learning_rate": 1.695981040736136e-05, + "loss": 0.8219, + "step": 1658 + }, + { + "epoch": 0.27815152467777426, + "grad_norm": 0.831977367401123, + "learning_rate": 1.6955909895515544e-05, + "loss": 0.9022, + "step": 1659 + }, + { + "epoch": 0.27831918683852036, + "grad_norm": 0.670785129070282, + "learning_rate": 1.6952007332285766e-05, + "loss": 0.8842, + "step": 1660 + }, + { + "epoch": 0.27848684899926646, + "grad_norm": 0.759387731552124, + "learning_rate": 1.694810271882294e-05, + "loss": 0.8635, + "step": 1661 + }, + { + "epoch": 0.27865451116001255, + "grad_norm": 0.47069522738456726, + "learning_rate": 1.6944196056278585e-05, + "loss": 0.8868, + "step": 1662 + }, + { + "epoch": 0.27882217332075865, + "grad_norm": 0.4956364631652832, + "learning_rate": 1.6940287345804828e-05, + "loss": 0.9507, + "step": 1663 + }, + { + "epoch": 0.27898983548150474, + "grad_norm": 0.6357293725013733, + "learning_rate": 1.6936376588554386e-05, + "loss": 0.842, + "step": 1664 + }, + { + "epoch": 0.27915749764225084, + "grad_norm": 0.48392024636268616, + "learning_rate": 1.69324637856806e-05, + "loss": 0.8362, + "step": 1665 + }, + { + "epoch": 0.27932515980299694, + "grad_norm": 0.5613908171653748, + "learning_rate": 1.69285489383374e-05, + "loss": 0.9111, + "step": 1666 + }, + { + "epoch": 0.27949282196374303, + "grad_norm": 0.6918414235115051, + "learning_rate": 1.692463204767932e-05, + "loss": 0.8713, + "step": 1667 + }, + { + "epoch": 0.27966048412448913, + "grad_norm": 0.5419053435325623, + "learning_rate": 1.6920713114861507e-05, + "loss": 0.9079, + "step": 1668 + }, + { + "epoch": 0.2798281462852352, + "grad_norm": 0.4282153248786926, + "learning_rate": 1.69167921410397e-05, + "loss": 0.8899, + "step": 1669 + }, + { + "epoch": 0.2799958084459813, + "grad_norm": 0.3876660168170929, + "learning_rate": 1.691286912737024e-05, + "loss": 0.8342, + "step": 1670 + }, + { + "epoch": 0.28016347060672747, + "grad_norm": 0.6596119999885559, + "learning_rate": 1.6908944075010067e-05, + "loss": 0.8435, + "step": 1671 + }, + { + "epoch": 0.28033113276747357, + "grad_norm": 0.9446216225624084, + "learning_rate": 1.690501698511674e-05, + "loss": 0.9391, + "step": 1672 + }, + { + "epoch": 0.28049879492821966, + "grad_norm": 0.44242310523986816, + "learning_rate": 1.69010878588484e-05, + "loss": 0.8816, + "step": 1673 + }, + { + "epoch": 0.28066645708896576, + "grad_norm": 0.45864903926849365, + "learning_rate": 1.6897156697363795e-05, + "loss": 0.9007, + "step": 1674 + }, + { + "epoch": 0.28083411924971186, + "grad_norm": 0.43198034167289734, + "learning_rate": 1.689322350182227e-05, + "loss": 1.0315, + "step": 1675 + }, + { + "epoch": 0.28100178141045795, + "grad_norm": 0.4529312252998352, + "learning_rate": 1.688928827338378e-05, + "loss": 0.9204, + "step": 1676 + }, + { + "epoch": 0.28116944357120405, + "grad_norm": 0.3595108985900879, + "learning_rate": 1.6885351013208874e-05, + "loss": 0.9069, + "step": 1677 + }, + { + "epoch": 0.28133710573195014, + "grad_norm": 0.40380436182022095, + "learning_rate": 1.6881411722458688e-05, + "loss": 0.9084, + "step": 1678 + }, + { + "epoch": 0.28150476789269624, + "grad_norm": 0.3343381881713867, + "learning_rate": 1.6877470402294978e-05, + "loss": 0.8491, + "step": 1679 + }, + { + "epoch": 0.28167243005344234, + "grad_norm": 0.47201022505760193, + "learning_rate": 1.6873527053880082e-05, + "loss": 0.9348, + "step": 1680 + }, + { + "epoch": 0.28184009221418843, + "grad_norm": 0.4359557628631592, + "learning_rate": 1.6869581678376944e-05, + "loss": 0.945, + "step": 1681 + }, + { + "epoch": 0.2820077543749345, + "grad_norm": 0.49326834082603455, + "learning_rate": 1.6865634276949104e-05, + "loss": 0.9073, + "step": 1682 + }, + { + "epoch": 0.2821754165356806, + "grad_norm": 0.4376702904701233, + "learning_rate": 1.68616848507607e-05, + "loss": 0.9262, + "step": 1683 + }, + { + "epoch": 0.2823430786964267, + "grad_norm": 0.4711952209472656, + "learning_rate": 1.685773340097647e-05, + "loss": 0.937, + "step": 1684 + }, + { + "epoch": 0.2825107408571728, + "grad_norm": 0.38133877515792847, + "learning_rate": 1.6853779928761734e-05, + "loss": 0.859, + "step": 1685 + }, + { + "epoch": 0.2826784030179189, + "grad_norm": 0.44120439887046814, + "learning_rate": 1.6849824435282434e-05, + "loss": 0.9153, + "step": 1686 + }, + { + "epoch": 0.282846065178665, + "grad_norm": 0.5790536403656006, + "learning_rate": 1.6845866921705086e-05, + "loss": 0.9605, + "step": 1687 + }, + { + "epoch": 0.2830137273394111, + "grad_norm": 0.40194109082221985, + "learning_rate": 1.684190738919681e-05, + "loss": 0.8764, + "step": 1688 + }, + { + "epoch": 0.2831813895001572, + "grad_norm": 0.466818630695343, + "learning_rate": 1.683794583892533e-05, + "loss": 0.8306, + "step": 1689 + }, + { + "epoch": 0.2833490516609033, + "grad_norm": 0.677361011505127, + "learning_rate": 1.6833982272058942e-05, + "loss": 0.8976, + "step": 1690 + }, + { + "epoch": 0.2835167138216494, + "grad_norm": 0.4113081991672516, + "learning_rate": 1.683001668976656e-05, + "loss": 0.9737, + "step": 1691 + }, + { + "epoch": 0.2836843759823955, + "grad_norm": 0.5291762948036194, + "learning_rate": 1.6826049093217682e-05, + "loss": 0.855, + "step": 1692 + }, + { + "epoch": 0.2838520381431416, + "grad_norm": 0.5291401147842407, + "learning_rate": 1.68220794835824e-05, + "loss": 0.9007, + "step": 1693 + }, + { + "epoch": 0.2840197003038877, + "grad_norm": 0.625732958316803, + "learning_rate": 1.6818107862031404e-05, + "loss": 0.8679, + "step": 1694 + }, + { + "epoch": 0.2841873624646338, + "grad_norm": 0.5087775588035583, + "learning_rate": 1.681413422973597e-05, + "loss": 0.8911, + "step": 1695 + }, + { + "epoch": 0.28435502462537987, + "grad_norm": 0.5665459632873535, + "learning_rate": 1.6810158587867973e-05, + "loss": 0.8136, + "step": 1696 + }, + { + "epoch": 0.28452268678612597, + "grad_norm": 0.5535372495651245, + "learning_rate": 1.6806180937599882e-05, + "loss": 0.9728, + "step": 1697 + }, + { + "epoch": 0.28469034894687206, + "grad_norm": 0.5744376182556152, + "learning_rate": 1.680220128010475e-05, + "loss": 0.9279, + "step": 1698 + }, + { + "epoch": 0.28485801110761816, + "grad_norm": 0.3833550214767456, + "learning_rate": 1.6798219616556234e-05, + "loss": 0.9503, + "step": 1699 + }, + { + "epoch": 0.28502567326836425, + "grad_norm": 0.47775256633758545, + "learning_rate": 1.6794235948128564e-05, + "loss": 0.8974, + "step": 1700 + }, + { + "epoch": 0.28519333542911035, + "grad_norm": 0.3792549669742584, + "learning_rate": 1.679025027599659e-05, + "loss": 0.8856, + "step": 1701 + }, + { + "epoch": 0.28536099758985645, + "grad_norm": 0.7515558004379272, + "learning_rate": 1.678626260133572e-05, + "loss": 0.9097, + "step": 1702 + }, + { + "epoch": 0.28552865975060254, + "grad_norm": 0.4453137218952179, + "learning_rate": 1.678227292532198e-05, + "loss": 0.8997, + "step": 1703 + }, + { + "epoch": 0.28569632191134864, + "grad_norm": 0.35799017548561096, + "learning_rate": 1.6778281249131973e-05, + "loss": 0.8296, + "step": 1704 + }, + { + "epoch": 0.28586398407209473, + "grad_norm": 0.6710291504859924, + "learning_rate": 1.6774287573942885e-05, + "loss": 0.8943, + "step": 1705 + }, + { + "epoch": 0.28603164623284083, + "grad_norm": 0.6353430151939392, + "learning_rate": 1.6770291900932513e-05, + "loss": 1.0383, + "step": 1706 + }, + { + "epoch": 0.2861993083935869, + "grad_norm": 0.40030571818351746, + "learning_rate": 1.6766294231279218e-05, + "loss": 0.9065, + "step": 1707 + }, + { + "epoch": 0.286366970554333, + "grad_norm": 0.644238293170929, + "learning_rate": 1.6762294566161976e-05, + "loss": 0.9221, + "step": 1708 + }, + { + "epoch": 0.2865346327150791, + "grad_norm": 0.528060257434845, + "learning_rate": 1.6758292906760323e-05, + "loss": 0.9305, + "step": 1709 + }, + { + "epoch": 0.2867022948758252, + "grad_norm": 0.5396594405174255, + "learning_rate": 1.675428925425441e-05, + "loss": 0.8952, + "step": 1710 + }, + { + "epoch": 0.2868699570365713, + "grad_norm": 0.5679185390472412, + "learning_rate": 1.6750283609824956e-05, + "loss": 0.9096, + "step": 1711 + }, + { + "epoch": 0.2870376191973174, + "grad_norm": 0.3885863423347473, + "learning_rate": 1.6746275974653278e-05, + "loss": 0.8593, + "step": 1712 + }, + { + "epoch": 0.2872052813580635, + "grad_norm": 0.6562624573707581, + "learning_rate": 1.674226634992128e-05, + "loss": 0.9612, + "step": 1713 + }, + { + "epoch": 0.2873729435188096, + "grad_norm": 0.46650078892707825, + "learning_rate": 1.6738254736811446e-05, + "loss": 0.8997, + "step": 1714 + }, + { + "epoch": 0.2875406056795557, + "grad_norm": 0.4466274380683899, + "learning_rate": 1.673424113650685e-05, + "loss": 0.8345, + "step": 1715 + }, + { + "epoch": 0.2877082678403018, + "grad_norm": 0.5457987189292908, + "learning_rate": 1.673022555019115e-05, + "loss": 0.9359, + "step": 1716 + }, + { + "epoch": 0.2878759300010479, + "grad_norm": 0.4065357744693756, + "learning_rate": 1.6726207979048604e-05, + "loss": 0.9402, + "step": 1717 + }, + { + "epoch": 0.288043592161794, + "grad_norm": 0.4762505292892456, + "learning_rate": 1.6722188424264036e-05, + "loss": 0.9184, + "step": 1718 + }, + { + "epoch": 0.2882112543225401, + "grad_norm": 0.40490788221359253, + "learning_rate": 1.6718166887022856e-05, + "loss": 0.872, + "step": 1719 + }, + { + "epoch": 0.2883789164832862, + "grad_norm": 0.632838785648346, + "learning_rate": 1.6714143368511076e-05, + "loss": 0.9559, + "step": 1720 + }, + { + "epoch": 0.28854657864403227, + "grad_norm": 0.6651273369789124, + "learning_rate": 1.6710117869915272e-05, + "loss": 0.8717, + "step": 1721 + }, + { + "epoch": 0.28871424080477837, + "grad_norm": 0.6503715515136719, + "learning_rate": 1.670609039242262e-05, + "loss": 0.8381, + "step": 1722 + }, + { + "epoch": 0.28888190296552446, + "grad_norm": 1.0909502506256104, + "learning_rate": 1.6702060937220873e-05, + "loss": 0.8412, + "step": 1723 + }, + { + "epoch": 0.28904956512627056, + "grad_norm": 0.9266510605812073, + "learning_rate": 1.6698029505498364e-05, + "loss": 0.8889, + "step": 1724 + }, + { + "epoch": 0.28921722728701665, + "grad_norm": 0.591495156288147, + "learning_rate": 1.669399609844401e-05, + "loss": 0.9035, + "step": 1725 + }, + { + "epoch": 0.28938488944776275, + "grad_norm": 1.0084071159362793, + "learning_rate": 1.668996071724732e-05, + "loss": 0.9163, + "step": 1726 + }, + { + "epoch": 0.28955255160850885, + "grad_norm": 0.6242332458496094, + "learning_rate": 1.6685923363098374e-05, + "loss": 0.9056, + "step": 1727 + }, + { + "epoch": 0.28972021376925494, + "grad_norm": 0.39805349707603455, + "learning_rate": 1.6681884037187836e-05, + "loss": 0.836, + "step": 1728 + }, + { + "epoch": 0.28988787593000104, + "grad_norm": 0.45632416009902954, + "learning_rate": 1.6677842740706958e-05, + "loss": 0.9636, + "step": 1729 + }, + { + "epoch": 0.29005553809074713, + "grad_norm": 0.43758928775787354, + "learning_rate": 1.667379947484756e-05, + "loss": 0.8712, + "step": 1730 + }, + { + "epoch": 0.29022320025149323, + "grad_norm": 0.5538960695266724, + "learning_rate": 1.666975424080206e-05, + "loss": 0.8779, + "step": 1731 + }, + { + "epoch": 0.2903908624122393, + "grad_norm": 0.48360779881477356, + "learning_rate": 1.6665707039763445e-05, + "loss": 0.8818, + "step": 1732 + }, + { + "epoch": 0.2905585245729854, + "grad_norm": 0.5961825251579285, + "learning_rate": 1.666165787292528e-05, + "loss": 0.9156, + "step": 1733 + }, + { + "epoch": 0.2907261867337315, + "grad_norm": 0.6016607284545898, + "learning_rate": 1.6657606741481723e-05, + "loss": 0.9387, + "step": 1734 + }, + { + "epoch": 0.2908938488944776, + "grad_norm": 0.5556625127792358, + "learning_rate": 1.6653553646627494e-05, + "loss": 0.9521, + "step": 1735 + }, + { + "epoch": 0.2910615110552237, + "grad_norm": 0.7546951174736023, + "learning_rate": 1.6649498589557908e-05, + "loss": 0.8334, + "step": 1736 + }, + { + "epoch": 0.2912291732159698, + "grad_norm": 0.6190481781959534, + "learning_rate": 1.6645441571468847e-05, + "loss": 0.9916, + "step": 1737 + }, + { + "epoch": 0.2913968353767159, + "grad_norm": 0.6533049941062927, + "learning_rate": 1.6641382593556776e-05, + "loss": 0.9274, + "step": 1738 + }, + { + "epoch": 0.291564497537462, + "grad_norm": 0.5003221035003662, + "learning_rate": 1.6637321657018742e-05, + "loss": 0.9809, + "step": 1739 + }, + { + "epoch": 0.2917321596982081, + "grad_norm": 0.5371973514556885, + "learning_rate": 1.663325876305236e-05, + "loss": 0.8339, + "step": 1740 + }, + { + "epoch": 0.2918998218589542, + "grad_norm": 0.4703129231929779, + "learning_rate": 1.662919391285583e-05, + "loss": 1.0353, + "step": 1741 + }, + { + "epoch": 0.2920674840197003, + "grad_norm": 0.4482932984828949, + "learning_rate": 1.6625127107627928e-05, + "loss": 1.014, + "step": 1742 + }, + { + "epoch": 0.2922351461804464, + "grad_norm": 0.620073139667511, + "learning_rate": 1.6621058348568008e-05, + "loss": 0.8419, + "step": 1743 + }, + { + "epoch": 0.2924028083411925, + "grad_norm": 0.4446816146373749, + "learning_rate": 1.6616987636875987e-05, + "loss": 0.822, + "step": 1744 + }, + { + "epoch": 0.2925704705019386, + "grad_norm": 0.4312778115272522, + "learning_rate": 1.661291497375238e-05, + "loss": 0.909, + "step": 1745 + }, + { + "epoch": 0.29273813266268467, + "grad_norm": 0.4368993639945984, + "learning_rate": 1.6608840360398257e-05, + "loss": 0.9185, + "step": 1746 + }, + { + "epoch": 0.29290579482343077, + "grad_norm": 0.43704938888549805, + "learning_rate": 1.660476379801528e-05, + "loss": 0.9995, + "step": 1747 + }, + { + "epoch": 0.29307345698417686, + "grad_norm": 0.585857093334198, + "learning_rate": 1.660068528780567e-05, + "loss": 0.9022, + "step": 1748 + }, + { + "epoch": 0.29324111914492296, + "grad_norm": 0.4600761830806732, + "learning_rate": 1.6596604830972234e-05, + "loss": 0.91, + "step": 1749 + }, + { + "epoch": 0.29340878130566905, + "grad_norm": 0.43956708908081055, + "learning_rate": 1.6592522428718348e-05, + "loss": 0.9309, + "step": 1750 + }, + { + "epoch": 0.29357644346641515, + "grad_norm": 0.6964955925941467, + "learning_rate": 1.6588438082247966e-05, + "loss": 0.895, + "step": 1751 + }, + { + "epoch": 0.29374410562716124, + "grad_norm": 0.45925039052963257, + "learning_rate": 1.6584351792765612e-05, + "loss": 0.8934, + "step": 1752 + }, + { + "epoch": 0.29391176778790734, + "grad_norm": 0.5126753449440002, + "learning_rate": 1.658026356147638e-05, + "loss": 0.9728, + "step": 1753 + }, + { + "epoch": 0.29407942994865344, + "grad_norm": 0.4704686999320984, + "learning_rate": 1.6576173389585938e-05, + "loss": 0.9259, + "step": 1754 + }, + { + "epoch": 0.29424709210939953, + "grad_norm": 0.46139395236968994, + "learning_rate": 1.657208127830053e-05, + "loss": 0.9183, + "step": 1755 + }, + { + "epoch": 0.29441475427014563, + "grad_norm": 0.4636790454387665, + "learning_rate": 1.656798722882698e-05, + "loss": 0.8836, + "step": 1756 + }, + { + "epoch": 0.2945824164308918, + "grad_norm": 0.4975752532482147, + "learning_rate": 1.656389124237266e-05, + "loss": 0.8966, + "step": 1757 + }, + { + "epoch": 0.2947500785916379, + "grad_norm": 0.5077047348022461, + "learning_rate": 1.6559793320145536e-05, + "loss": 0.9251, + "step": 1758 + }, + { + "epoch": 0.29491774075238397, + "grad_norm": 0.544331431388855, + "learning_rate": 1.6555693463354133e-05, + "loss": 0.9085, + "step": 1759 + }, + { + "epoch": 0.29508540291313007, + "grad_norm": 0.4727647602558136, + "learning_rate": 1.655159167320755e-05, + "loss": 0.8702, + "step": 1760 + }, + { + "epoch": 0.29525306507387616, + "grad_norm": 0.703439474105835, + "learning_rate": 1.6547487950915458e-05, + "loss": 0.8626, + "step": 1761 + }, + { + "epoch": 0.29542072723462226, + "grad_norm": 0.49445345997810364, + "learning_rate": 1.654338229768809e-05, + "loss": 0.9302, + "step": 1762 + }, + { + "epoch": 0.29558838939536836, + "grad_norm": 0.4604395627975464, + "learning_rate": 1.6539274714736255e-05, + "loss": 0.8253, + "step": 1763 + }, + { + "epoch": 0.29575605155611445, + "grad_norm": 0.8136701583862305, + "learning_rate": 1.6535165203271334e-05, + "loss": 0.8502, + "step": 1764 + }, + { + "epoch": 0.29592371371686055, + "grad_norm": 0.4312807619571686, + "learning_rate": 1.653105376450527e-05, + "loss": 0.9097, + "step": 1765 + }, + { + "epoch": 0.29609137587760664, + "grad_norm": 0.4612438380718231, + "learning_rate": 1.652694039965058e-05, + "loss": 0.9087, + "step": 1766 + }, + { + "epoch": 0.29625903803835274, + "grad_norm": 0.4308323860168457, + "learning_rate": 1.6522825109920342e-05, + "loss": 0.8786, + "step": 1767 + }, + { + "epoch": 0.29642670019909884, + "grad_norm": 0.7085330486297607, + "learning_rate": 1.651870789652821e-05, + "loss": 0.868, + "step": 1768 + }, + { + "epoch": 0.29659436235984493, + "grad_norm": 0.466951459646225, + "learning_rate": 1.6514588760688397e-05, + "loss": 0.8911, + "step": 1769 + }, + { + "epoch": 0.29676202452059103, + "grad_norm": 0.864098310470581, + "learning_rate": 1.6510467703615694e-05, + "loss": 0.8381, + "step": 1770 + }, + { + "epoch": 0.2969296866813371, + "grad_norm": 0.515808641910553, + "learning_rate": 1.6506344726525443e-05, + "loss": 0.8305, + "step": 1771 + }, + { + "epoch": 0.2970973488420832, + "grad_norm": 0.5473104119300842, + "learning_rate": 1.6502219830633567e-05, + "loss": 0.8603, + "step": 1772 + }, + { + "epoch": 0.2972650110028293, + "grad_norm": 0.3563365340232849, + "learning_rate": 1.6498093017156545e-05, + "loss": 0.8859, + "step": 1773 + }, + { + "epoch": 0.2974326731635754, + "grad_norm": 0.5445033311843872, + "learning_rate": 1.649396428731143e-05, + "loss": 0.9604, + "step": 1774 + }, + { + "epoch": 0.2976003353243215, + "grad_norm": 0.40800729393959045, + "learning_rate": 1.6489833642315834e-05, + "loss": 0.8372, + "step": 1775 + }, + { + "epoch": 0.2977679974850676, + "grad_norm": 0.4679819345474243, + "learning_rate": 1.648570108338793e-05, + "loss": 0.9257, + "step": 1776 + }, + { + "epoch": 0.2979356596458137, + "grad_norm": 0.3677198588848114, + "learning_rate": 1.648156661174647e-05, + "loss": 0.8854, + "step": 1777 + }, + { + "epoch": 0.2981033218065598, + "grad_norm": 0.5773995518684387, + "learning_rate": 1.6477430228610755e-05, + "loss": 0.9674, + "step": 1778 + }, + { + "epoch": 0.2982709839673059, + "grad_norm": 0.44452473521232605, + "learning_rate": 1.6473291935200657e-05, + "loss": 0.9222, + "step": 1779 + }, + { + "epoch": 0.298438646128052, + "grad_norm": 0.4076407849788666, + "learning_rate": 1.646915173273661e-05, + "loss": 0.8526, + "step": 1780 + }, + { + "epoch": 0.2986063082887981, + "grad_norm": 0.5405694246292114, + "learning_rate": 1.6465009622439607e-05, + "loss": 0.8831, + "step": 1781 + }, + { + "epoch": 0.2987739704495442, + "grad_norm": 0.44485101103782654, + "learning_rate": 1.6460865605531214e-05, + "loss": 0.949, + "step": 1782 + }, + { + "epoch": 0.2989416326102903, + "grad_norm": 0.529278039932251, + "learning_rate": 1.645671968323355e-05, + "loss": 0.8862, + "step": 1783 + }, + { + "epoch": 0.29910929477103637, + "grad_norm": 0.4891248643398285, + "learning_rate": 1.64525718567693e-05, + "loss": 0.8989, + "step": 1784 + }, + { + "epoch": 0.29927695693178247, + "grad_norm": 0.4654734432697296, + "learning_rate": 1.6448422127361707e-05, + "loss": 0.8702, + "step": 1785 + }, + { + "epoch": 0.29944461909252856, + "grad_norm": 0.5863080024719238, + "learning_rate": 1.6444270496234583e-05, + "loss": 0.9906, + "step": 1786 + }, + { + "epoch": 0.29961228125327466, + "grad_norm": 0.4149811267852783, + "learning_rate": 1.6440116964612287e-05, + "loss": 0.9792, + "step": 1787 + }, + { + "epoch": 0.29977994341402076, + "grad_norm": 0.6241120100021362, + "learning_rate": 1.6435961533719753e-05, + "loss": 1.0157, + "step": 1788 + }, + { + "epoch": 0.29994760557476685, + "grad_norm": 0.677161455154419, + "learning_rate": 1.6431804204782468e-05, + "loss": 0.9572, + "step": 1789 + }, + { + "epoch": 0.30011526773551295, + "grad_norm": 0.5379821062088013, + "learning_rate": 1.6427644979026473e-05, + "loss": 0.9378, + "step": 1790 + }, + { + "epoch": 0.30028292989625904, + "grad_norm": 0.40610626339912415, + "learning_rate": 1.642348385767839e-05, + "loss": 0.9116, + "step": 1791 + }, + { + "epoch": 0.30045059205700514, + "grad_norm": 0.47444257140159607, + "learning_rate": 1.641932084196537e-05, + "loss": 0.7692, + "step": 1792 + }, + { + "epoch": 0.30061825421775124, + "grad_norm": 0.44950932264328003, + "learning_rate": 1.641515593311515e-05, + "loss": 0.8458, + "step": 1793 + }, + { + "epoch": 0.30078591637849733, + "grad_norm": 0.5710813999176025, + "learning_rate": 1.6410989132356005e-05, + "loss": 0.9122, + "step": 1794 + }, + { + "epoch": 0.3009535785392434, + "grad_norm": 0.45583242177963257, + "learning_rate": 1.6406820440916778e-05, + "loss": 0.8478, + "step": 1795 + }, + { + "epoch": 0.3011212406999895, + "grad_norm": 0.5046128630638123, + "learning_rate": 1.6402649860026868e-05, + "loss": 0.9029, + "step": 1796 + }, + { + "epoch": 0.3012889028607356, + "grad_norm": 0.5096794962882996, + "learning_rate": 1.6398477390916235e-05, + "loss": 0.8417, + "step": 1797 + }, + { + "epoch": 0.3014565650214817, + "grad_norm": 0.510442852973938, + "learning_rate": 1.6394303034815387e-05, + "loss": 0.8932, + "step": 1798 + }, + { + "epoch": 0.3016242271822278, + "grad_norm": 0.6885331869125366, + "learning_rate": 1.639012679295539e-05, + "loss": 0.9257, + "step": 1799 + }, + { + "epoch": 0.3017918893429739, + "grad_norm": 0.8317268490791321, + "learning_rate": 1.638594866656788e-05, + "loss": 0.8742, + "step": 1800 + }, + { + "epoch": 0.30195955150372, + "grad_norm": 0.5611824989318848, + "learning_rate": 1.638176865688503e-05, + "loss": 0.8217, + "step": 1801 + }, + { + "epoch": 0.3021272136644661, + "grad_norm": 0.4200962781906128, + "learning_rate": 1.637758676513958e-05, + "loss": 0.8437, + "step": 1802 + }, + { + "epoch": 0.3022948758252122, + "grad_norm": 0.6332334876060486, + "learning_rate": 1.637340299256482e-05, + "loss": 0.9094, + "step": 1803 + }, + { + "epoch": 0.3024625379859583, + "grad_norm": 0.6723257899284363, + "learning_rate": 1.6369217340394595e-05, + "loss": 0.9367, + "step": 1804 + }, + { + "epoch": 0.3026302001467044, + "grad_norm": 0.6654242873191833, + "learning_rate": 1.6365029809863308e-05, + "loss": 0.7906, + "step": 1805 + }, + { + "epoch": 0.3027978623074505, + "grad_norm": 0.5226824283599854, + "learning_rate": 1.6360840402205914e-05, + "loss": 0.9066, + "step": 1806 + }, + { + "epoch": 0.3029655244681966, + "grad_norm": 0.6385287642478943, + "learning_rate": 1.635664911865792e-05, + "loss": 0.9012, + "step": 1807 + }, + { + "epoch": 0.3031331866289427, + "grad_norm": 1.1316041946411133, + "learning_rate": 1.6352455960455385e-05, + "loss": 0.9579, + "step": 1808 + }, + { + "epoch": 0.30330084878968877, + "grad_norm": 0.4610840678215027, + "learning_rate": 1.6348260928834932e-05, + "loss": 0.9515, + "step": 1809 + }, + { + "epoch": 0.30346851095043487, + "grad_norm": 0.4964912533760071, + "learning_rate": 1.634406402503372e-05, + "loss": 0.8182, + "step": 1810 + }, + { + "epoch": 0.30363617311118096, + "grad_norm": 0.6284961700439453, + "learning_rate": 1.6339865250289468e-05, + "loss": 0.8412, + "step": 1811 + }, + { + "epoch": 0.30380383527192706, + "grad_norm": 0.5082619786262512, + "learning_rate": 1.6335664605840446e-05, + "loss": 0.9138, + "step": 1812 + }, + { + "epoch": 0.30397149743267315, + "grad_norm": 0.5302368402481079, + "learning_rate": 1.6331462092925485e-05, + "loss": 0.8233, + "step": 1813 + }, + { + "epoch": 0.30413915959341925, + "grad_norm": 0.4402926564216614, + "learning_rate": 1.6327257712783945e-05, + "loss": 0.8033, + "step": 1814 + }, + { + "epoch": 0.30430682175416535, + "grad_norm": 0.5954879522323608, + "learning_rate": 1.6323051466655756e-05, + "loss": 0.9123, + "step": 1815 + }, + { + "epoch": 0.30447448391491144, + "grad_norm": 0.5569353699684143, + "learning_rate": 1.6318843355781395e-05, + "loss": 0.8772, + "step": 1816 + }, + { + "epoch": 0.30464214607565754, + "grad_norm": 0.5235996246337891, + "learning_rate": 1.6314633381401882e-05, + "loss": 0.8687, + "step": 1817 + }, + { + "epoch": 0.30480980823640363, + "grad_norm": 0.4863227307796478, + "learning_rate": 1.6310421544758794e-05, + "loss": 0.929, + "step": 1818 + }, + { + "epoch": 0.30497747039714973, + "grad_norm": 0.3899315297603607, + "learning_rate": 1.630620784709425e-05, + "loss": 0.8168, + "step": 1819 + }, + { + "epoch": 0.3051451325578958, + "grad_norm": 0.4086456596851349, + "learning_rate": 1.6301992289650925e-05, + "loss": 0.9013, + "step": 1820 + }, + { + "epoch": 0.3053127947186419, + "grad_norm": 0.5903594493865967, + "learning_rate": 1.6297774873672036e-05, + "loss": 0.9235, + "step": 1821 + }, + { + "epoch": 0.305480456879388, + "grad_norm": 0.5880112648010254, + "learning_rate": 1.629355560040135e-05, + "loss": 0.8814, + "step": 1822 + }, + { + "epoch": 0.3056481190401341, + "grad_norm": 0.48934492468833923, + "learning_rate": 1.6289334471083194e-05, + "loss": 0.8804, + "step": 1823 + }, + { + "epoch": 0.3058157812008802, + "grad_norm": 0.48269081115722656, + "learning_rate": 1.6285111486962414e-05, + "loss": 0.8828, + "step": 1824 + }, + { + "epoch": 0.3059834433616263, + "grad_norm": 0.6265968680381775, + "learning_rate": 1.6280886649284432e-05, + "loss": 0.8181, + "step": 1825 + }, + { + "epoch": 0.3061511055223724, + "grad_norm": 0.5036160945892334, + "learning_rate": 1.62766599592952e-05, + "loss": 0.7815, + "step": 1826 + }, + { + "epoch": 0.3063187676831185, + "grad_norm": 0.5734903812408447, + "learning_rate": 1.627243141824123e-05, + "loss": 0.823, + "step": 1827 + }, + { + "epoch": 0.3064864298438646, + "grad_norm": 0.5602980256080627, + "learning_rate": 1.626820102736956e-05, + "loss": 1.0331, + "step": 1828 + }, + { + "epoch": 0.3066540920046107, + "grad_norm": 0.6858996748924255, + "learning_rate": 1.626396878792779e-05, + "loss": 0.8685, + "step": 1829 + }, + { + "epoch": 0.3068217541653568, + "grad_norm": 0.6805505752563477, + "learning_rate": 1.6259734701164058e-05, + "loss": 0.9075, + "step": 1830 + }, + { + "epoch": 0.3069894163261029, + "grad_norm": 0.5220349431037903, + "learning_rate": 1.625549876832705e-05, + "loss": 0.948, + "step": 1831 + }, + { + "epoch": 0.307157078486849, + "grad_norm": 0.4397321939468384, + "learning_rate": 1.6251260990665993e-05, + "loss": 0.8982, + "step": 1832 + }, + { + "epoch": 0.3073247406475951, + "grad_norm": 0.542477011680603, + "learning_rate": 1.624702136943066e-05, + "loss": 0.9823, + "step": 1833 + }, + { + "epoch": 0.30749240280834117, + "grad_norm": 0.45024630427360535, + "learning_rate": 1.6242779905871375e-05, + "loss": 0.901, + "step": 1834 + }, + { + "epoch": 0.30766006496908727, + "grad_norm": 0.6258662939071655, + "learning_rate": 1.623853660123899e-05, + "loss": 0.9651, + "step": 1835 + }, + { + "epoch": 0.30782772712983336, + "grad_norm": 0.47294142842292786, + "learning_rate": 1.6234291456784912e-05, + "loss": 0.9054, + "step": 1836 + }, + { + "epoch": 0.30799538929057946, + "grad_norm": 0.5752155780792236, + "learning_rate": 1.623004447376108e-05, + "loss": 0.9933, + "step": 1837 + }, + { + "epoch": 0.30816305145132555, + "grad_norm": 0.48368701338768005, + "learning_rate": 1.6225795653419995e-05, + "loss": 0.9122, + "step": 1838 + }, + { + "epoch": 0.30833071361207165, + "grad_norm": 0.5220962762832642, + "learning_rate": 1.6221544997014673e-05, + "loss": 0.9381, + "step": 1839 + }, + { + "epoch": 0.30849837577281775, + "grad_norm": 0.5780762434005737, + "learning_rate": 1.621729250579869e-05, + "loss": 1.042, + "step": 1840 + }, + { + "epoch": 0.30866603793356384, + "grad_norm": 0.44534143805503845, + "learning_rate": 1.6213038181026166e-05, + "loss": 0.8988, + "step": 1841 + }, + { + "epoch": 0.30883370009431, + "grad_norm": 0.5983442068099976, + "learning_rate": 1.6208782023951744e-05, + "loss": 0.8925, + "step": 1842 + }, + { + "epoch": 0.3090013622550561, + "grad_norm": 0.3870874345302582, + "learning_rate": 1.620452403583062e-05, + "loss": 0.8957, + "step": 1843 + }, + { + "epoch": 0.3091690244158022, + "grad_norm": 0.46022143959999084, + "learning_rate": 1.620026421791853e-05, + "loss": 0.925, + "step": 1844 + }, + { + "epoch": 0.3093366865765483, + "grad_norm": 0.36380746960639954, + "learning_rate": 1.6196002571471745e-05, + "loss": 0.8479, + "step": 1845 + }, + { + "epoch": 0.3095043487372944, + "grad_norm": 0.6444778442382812, + "learning_rate": 1.619173909774708e-05, + "loss": 0.9413, + "step": 1846 + }, + { + "epoch": 0.3096720108980405, + "grad_norm": 0.36853939294815063, + "learning_rate": 1.618747379800188e-05, + "loss": 0.839, + "step": 1847 + }, + { + "epoch": 0.30983967305878657, + "grad_norm": 0.5956276655197144, + "learning_rate": 1.618320667349404e-05, + "loss": 0.8552, + "step": 1848 + }, + { + "epoch": 0.31000733521953266, + "grad_norm": 0.7519698143005371, + "learning_rate": 1.6178937725481984e-05, + "loss": 0.9885, + "step": 1849 + }, + { + "epoch": 0.31017499738027876, + "grad_norm": 0.7840356826782227, + "learning_rate": 1.6174666955224686e-05, + "loss": 0.9566, + "step": 1850 + }, + { + "epoch": 0.31034265954102486, + "grad_norm": 0.4490809738636017, + "learning_rate": 1.617039436398164e-05, + "loss": 1.0007, + "step": 1851 + }, + { + "epoch": 0.31051032170177095, + "grad_norm": 0.4416768550872803, + "learning_rate": 1.616611995301289e-05, + "loss": 0.8151, + "step": 1852 + }, + { + "epoch": 0.31067798386251705, + "grad_norm": 0.5982283353805542, + "learning_rate": 1.616184372357901e-05, + "loss": 0.9484, + "step": 1853 + }, + { + "epoch": 0.31084564602326314, + "grad_norm": 0.5351338386535645, + "learning_rate": 1.6157565676941115e-05, + "loss": 0.9318, + "step": 1854 + }, + { + "epoch": 0.31101330818400924, + "grad_norm": 0.3872014582157135, + "learning_rate": 1.6153285814360856e-05, + "loss": 1.0033, + "step": 1855 + }, + { + "epoch": 0.31118097034475534, + "grad_norm": 0.48003336787223816, + "learning_rate": 1.6149004137100412e-05, + "loss": 0.8638, + "step": 1856 + }, + { + "epoch": 0.31134863250550143, + "grad_norm": 0.5285705924034119, + "learning_rate": 1.614472064642251e-05, + "loss": 0.9282, + "step": 1857 + }, + { + "epoch": 0.31151629466624753, + "grad_norm": 0.3734569549560547, + "learning_rate": 1.61404353435904e-05, + "loss": 0.8872, + "step": 1858 + }, + { + "epoch": 0.3116839568269936, + "grad_norm": 0.516380786895752, + "learning_rate": 1.613614822986787e-05, + "loss": 0.9805, + "step": 1859 + }, + { + "epoch": 0.3118516189877397, + "grad_norm": 0.47568896412849426, + "learning_rate": 1.6131859306519243e-05, + "loss": 0.8505, + "step": 1860 + }, + { + "epoch": 0.3120192811484858, + "grad_norm": 0.41954758763313293, + "learning_rate": 1.612756857480938e-05, + "loss": 0.7861, + "step": 1861 + }, + { + "epoch": 0.3121869433092319, + "grad_norm": 0.44616973400115967, + "learning_rate": 1.6123276036003664e-05, + "loss": 0.8088, + "step": 1862 + }, + { + "epoch": 0.312354605469978, + "grad_norm": 0.5130486488342285, + "learning_rate": 1.611898169136802e-05, + "loss": 0.9097, + "step": 1863 + }, + { + "epoch": 0.3125222676307241, + "grad_norm": 0.5425180196762085, + "learning_rate": 1.6114685542168912e-05, + "loss": 0.9192, + "step": 1864 + }, + { + "epoch": 0.3126899297914702, + "grad_norm": 0.37737610936164856, + "learning_rate": 1.6110387589673315e-05, + "loss": 0.8592, + "step": 1865 + }, + { + "epoch": 0.3128575919522163, + "grad_norm": 0.3993113040924072, + "learning_rate": 1.610608783514876e-05, + "loss": 0.9123, + "step": 1866 + }, + { + "epoch": 0.3130252541129624, + "grad_norm": 0.42021462321281433, + "learning_rate": 1.6101786279863285e-05, + "loss": 0.9289, + "step": 1867 + }, + { + "epoch": 0.3131929162737085, + "grad_norm": 0.6417856812477112, + "learning_rate": 1.6097482925085482e-05, + "loss": 0.9057, + "step": 1868 + }, + { + "epoch": 0.3133605784344546, + "grad_norm": 0.5079712867736816, + "learning_rate": 1.609317777208446e-05, + "loss": 1.0042, + "step": 1869 + }, + { + "epoch": 0.3135282405952007, + "grad_norm": 0.4228523075580597, + "learning_rate": 1.6088870822129862e-05, + "loss": 0.8974, + "step": 1870 + }, + { + "epoch": 0.3136959027559468, + "grad_norm": 0.487529993057251, + "learning_rate": 1.608456207649186e-05, + "loss": 0.8846, + "step": 1871 + }, + { + "epoch": 0.31386356491669287, + "grad_norm": 0.49924778938293457, + "learning_rate": 1.608025153644116e-05, + "loss": 0.9101, + "step": 1872 + }, + { + "epoch": 0.31403122707743897, + "grad_norm": 0.43718042969703674, + "learning_rate": 1.607593920324899e-05, + "loss": 0.8513, + "step": 1873 + }, + { + "epoch": 0.31419888923818506, + "grad_norm": 0.5099447965621948, + "learning_rate": 1.6071625078187113e-05, + "loss": 0.9156, + "step": 1874 + }, + { + "epoch": 0.31436655139893116, + "grad_norm": 0.47883695363998413, + "learning_rate": 1.6067309162527817e-05, + "loss": 0.7846, + "step": 1875 + }, + { + "epoch": 0.31453421355967726, + "grad_norm": 0.5323141813278198, + "learning_rate": 1.606299145754392e-05, + "loss": 0.9689, + "step": 1876 + }, + { + "epoch": 0.31470187572042335, + "grad_norm": 0.5095372200012207, + "learning_rate": 1.605867196450876e-05, + "loss": 0.7851, + "step": 1877 + }, + { + "epoch": 0.31486953788116945, + "grad_norm": 0.4553854763507843, + "learning_rate": 1.6054350684696216e-05, + "loss": 0.9793, + "step": 1878 + }, + { + "epoch": 0.31503720004191554, + "grad_norm": 0.5465895533561707, + "learning_rate": 1.6050027619380687e-05, + "loss": 0.8846, + "step": 1879 + }, + { + "epoch": 0.31520486220266164, + "grad_norm": 0.42045995593070984, + "learning_rate": 1.60457027698371e-05, + "loss": 0.9413, + "step": 1880 + }, + { + "epoch": 0.31537252436340774, + "grad_norm": 0.6263823509216309, + "learning_rate": 1.6041376137340893e-05, + "loss": 0.8666, + "step": 1881 + }, + { + "epoch": 0.31554018652415383, + "grad_norm": 0.43639567494392395, + "learning_rate": 1.603704772316806e-05, + "loss": 0.8834, + "step": 1882 + }, + { + "epoch": 0.3157078486848999, + "grad_norm": 0.56414794921875, + "learning_rate": 1.6032717528595098e-05, + "loss": 0.9547, + "step": 1883 + }, + { + "epoch": 0.315875510845646, + "grad_norm": 0.38132351636886597, + "learning_rate": 1.602838555489904e-05, + "loss": 0.9057, + "step": 1884 + }, + { + "epoch": 0.3160431730063921, + "grad_norm": 0.3706132769584656, + "learning_rate": 1.6024051803357425e-05, + "loss": 0.8835, + "step": 1885 + }, + { + "epoch": 0.3162108351671382, + "grad_norm": 0.4287024140357971, + "learning_rate": 1.6019716275248342e-05, + "loss": 0.9048, + "step": 1886 + }, + { + "epoch": 0.3163784973278843, + "grad_norm": 0.6251978278160095, + "learning_rate": 1.601537897185039e-05, + "loss": 0.9559, + "step": 1887 + }, + { + "epoch": 0.3165461594886304, + "grad_norm": 0.590323805809021, + "learning_rate": 1.6011039894442685e-05, + "loss": 0.9553, + "step": 1888 + }, + { + "epoch": 0.3167138216493765, + "grad_norm": 0.5125306248664856, + "learning_rate": 1.600669904430489e-05, + "loss": 0.8657, + "step": 1889 + }, + { + "epoch": 0.3168814838101226, + "grad_norm": 0.6732698678970337, + "learning_rate": 1.6002356422717163e-05, + "loss": 0.8711, + "step": 1890 + }, + { + "epoch": 0.3170491459708687, + "grad_norm": 0.5461569428443909, + "learning_rate": 1.59980120309602e-05, + "loss": 0.8593, + "step": 1891 + }, + { + "epoch": 0.3172168081316148, + "grad_norm": 0.5875232219696045, + "learning_rate": 1.599366587031522e-05, + "loss": 0.9037, + "step": 1892 + }, + { + "epoch": 0.3173844702923609, + "grad_norm": 0.5603500008583069, + "learning_rate": 1.5989317942063955e-05, + "loss": 0.8561, + "step": 1893 + }, + { + "epoch": 0.317552132453107, + "grad_norm": 0.4748394787311554, + "learning_rate": 1.598496824748866e-05, + "loss": 0.8718, + "step": 1894 + }, + { + "epoch": 0.3177197946138531, + "grad_norm": 0.43374186754226685, + "learning_rate": 1.5980616787872122e-05, + "loss": 0.8226, + "step": 1895 + }, + { + "epoch": 0.3178874567745992, + "grad_norm": 0.5378707647323608, + "learning_rate": 1.597626356449764e-05, + "loss": 0.7987, + "step": 1896 + }, + { + "epoch": 0.31805511893534527, + "grad_norm": 0.6311928033828735, + "learning_rate": 1.5971908578649024e-05, + "loss": 0.9061, + "step": 1897 + }, + { + "epoch": 0.31822278109609137, + "grad_norm": 0.5331396460533142, + "learning_rate": 1.5967551831610624e-05, + "loss": 0.9141, + "step": 1898 + }, + { + "epoch": 0.31839044325683746, + "grad_norm": 0.3972858786582947, + "learning_rate": 1.596319332466729e-05, + "loss": 0.8557, + "step": 1899 + }, + { + "epoch": 0.31855810541758356, + "grad_norm": 0.5714462995529175, + "learning_rate": 1.5958833059104405e-05, + "loss": 0.9814, + "step": 1900 + }, + { + "epoch": 0.31872576757832966, + "grad_norm": 0.5758419036865234, + "learning_rate": 1.5954471036207865e-05, + "loss": 0.8925, + "step": 1901 + }, + { + "epoch": 0.31889342973907575, + "grad_norm": 0.8500440716743469, + "learning_rate": 1.5950107257264076e-05, + "loss": 0.9505, + "step": 1902 + }, + { + "epoch": 0.31906109189982185, + "grad_norm": 0.4728148877620697, + "learning_rate": 1.5945741723559987e-05, + "loss": 0.8705, + "step": 1903 + }, + { + "epoch": 0.31922875406056794, + "grad_norm": 0.3619634211063385, + "learning_rate": 1.5941374436383035e-05, + "loss": 0.9196, + "step": 1904 + }, + { + "epoch": 0.31939641622131404, + "grad_norm": 0.4979732930660248, + "learning_rate": 1.593700539702119e-05, + "loss": 0.8234, + "step": 1905 + }, + { + "epoch": 0.31956407838206013, + "grad_norm": 0.5817590951919556, + "learning_rate": 1.5932634606762935e-05, + "loss": 0.9044, + "step": 1906 + }, + { + "epoch": 0.31973174054280623, + "grad_norm": 0.516790509223938, + "learning_rate": 1.5928262066897277e-05, + "loss": 0.8855, + "step": 1907 + }, + { + "epoch": 0.3198994027035523, + "grad_norm": 0.5449026226997375, + "learning_rate": 1.5923887778713726e-05, + "loss": 0.9708, + "step": 1908 + }, + { + "epoch": 0.3200670648642984, + "grad_norm": 0.4417110085487366, + "learning_rate": 1.591951174350231e-05, + "loss": 0.957, + "step": 1909 + }, + { + "epoch": 0.3202347270250445, + "grad_norm": 0.5064684748649597, + "learning_rate": 1.5915133962553584e-05, + "loss": 0.9166, + "step": 1910 + }, + { + "epoch": 0.3204023891857906, + "grad_norm": 0.4732873737812042, + "learning_rate": 1.5910754437158608e-05, + "loss": 0.9244, + "step": 1911 + }, + { + "epoch": 0.3205700513465367, + "grad_norm": 0.7517212629318237, + "learning_rate": 1.5906373168608952e-05, + "loss": 0.8813, + "step": 1912 + }, + { + "epoch": 0.3207377135072828, + "grad_norm": 0.44778767228126526, + "learning_rate": 1.5901990158196712e-05, + "loss": 0.862, + "step": 1913 + }, + { + "epoch": 0.3209053756680289, + "grad_norm": 0.470090389251709, + "learning_rate": 1.5897605407214497e-05, + "loss": 0.919, + "step": 1914 + }, + { + "epoch": 0.321073037828775, + "grad_norm": 0.3986078202724457, + "learning_rate": 1.589321891695541e-05, + "loss": 0.9469, + "step": 1915 + }, + { + "epoch": 0.3212406999895211, + "grad_norm": 0.5769183039665222, + "learning_rate": 1.5888830688713096e-05, + "loss": 0.915, + "step": 1916 + }, + { + "epoch": 0.3214083621502672, + "grad_norm": 0.42942318320274353, + "learning_rate": 1.5884440723781687e-05, + "loss": 0.8832, + "step": 1917 + }, + { + "epoch": 0.3215760243110133, + "grad_norm": 0.6780664920806885, + "learning_rate": 1.5880049023455848e-05, + "loss": 0.8358, + "step": 1918 + }, + { + "epoch": 0.3217436864717594, + "grad_norm": 0.3817391097545624, + "learning_rate": 1.5875655589030738e-05, + "loss": 0.9046, + "step": 1919 + }, + { + "epoch": 0.3219113486325055, + "grad_norm": 0.700987696647644, + "learning_rate": 1.5871260421802034e-05, + "loss": 0.8926, + "step": 1920 + }, + { + "epoch": 0.3220790107932516, + "grad_norm": 0.5530411005020142, + "learning_rate": 1.5866863523065933e-05, + "loss": 0.7418, + "step": 1921 + }, + { + "epoch": 0.32224667295399767, + "grad_norm": 0.4336833953857422, + "learning_rate": 1.586246489411913e-05, + "loss": 0.9115, + "step": 1922 + }, + { + "epoch": 0.32241433511474377, + "grad_norm": 0.5864166021347046, + "learning_rate": 1.5858064536258837e-05, + "loss": 1.0381, + "step": 1923 + }, + { + "epoch": 0.32258199727548986, + "grad_norm": 0.5918449759483337, + "learning_rate": 1.5853662450782772e-05, + "loss": 0.889, + "step": 1924 + }, + { + "epoch": 0.32274965943623596, + "grad_norm": 0.43187668919563293, + "learning_rate": 1.5849258638989166e-05, + "loss": 0.8609, + "step": 1925 + }, + { + "epoch": 0.32291732159698205, + "grad_norm": 0.5723666548728943, + "learning_rate": 1.5844853102176757e-05, + "loss": 0.934, + "step": 1926 + }, + { + "epoch": 0.32308498375772815, + "grad_norm": 0.5808163285255432, + "learning_rate": 1.5840445841644793e-05, + "loss": 0.8731, + "step": 1927 + }, + { + "epoch": 0.3232526459184743, + "grad_norm": 0.3876830041408539, + "learning_rate": 1.5836036858693024e-05, + "loss": 0.8578, + "step": 1928 + }, + { + "epoch": 0.3234203080792204, + "grad_norm": 0.8593195080757141, + "learning_rate": 1.583162615462172e-05, + "loss": 0.9692, + "step": 1929 + }, + { + "epoch": 0.3235879702399665, + "grad_norm": 0.5928516387939453, + "learning_rate": 1.5827213730731656e-05, + "loss": 0.9549, + "step": 1930 + }, + { + "epoch": 0.3237556324007126, + "grad_norm": 0.8312473297119141, + "learning_rate": 1.5822799588324102e-05, + "loss": 0.8135, + "step": 1931 + }, + { + "epoch": 0.3239232945614587, + "grad_norm": 0.5945383906364441, + "learning_rate": 1.581838372870085e-05, + "loss": 0.9046, + "step": 1932 + }, + { + "epoch": 0.3240909567222048, + "grad_norm": 0.4078502953052521, + "learning_rate": 1.5813966153164186e-05, + "loss": 0.8857, + "step": 1933 + }, + { + "epoch": 0.3242586188829509, + "grad_norm": 0.38547268509864807, + "learning_rate": 1.580954686301691e-05, + "loss": 0.8854, + "step": 1934 + }, + { + "epoch": 0.324426281043697, + "grad_norm": 0.3948238492012024, + "learning_rate": 1.5805125859562327e-05, + "loss": 0.8447, + "step": 1935 + }, + { + "epoch": 0.32459394320444307, + "grad_norm": 0.4084761142730713, + "learning_rate": 1.5800703144104242e-05, + "loss": 0.8287, + "step": 1936 + }, + { + "epoch": 0.32476160536518917, + "grad_norm": 0.4811871349811554, + "learning_rate": 1.5796278717946972e-05, + "loss": 0.944, + "step": 1937 + }, + { + "epoch": 0.32492926752593526, + "grad_norm": 0.3900909721851349, + "learning_rate": 1.5791852582395334e-05, + "loss": 0.8761, + "step": 1938 + }, + { + "epoch": 0.32509692968668136, + "grad_norm": 0.432222455739975, + "learning_rate": 1.5787424738754652e-05, + "loss": 0.9091, + "step": 1939 + }, + { + "epoch": 0.32526459184742745, + "grad_norm": 0.5341072082519531, + "learning_rate": 1.5782995188330746e-05, + "loss": 0.8926, + "step": 1940 + }, + { + "epoch": 0.32543225400817355, + "grad_norm": 0.4855509400367737, + "learning_rate": 1.5778563932429952e-05, + "loss": 0.8781, + "step": 1941 + }, + { + "epoch": 0.32559991616891965, + "grad_norm": 0.5989813804626465, + "learning_rate": 1.57741309723591e-05, + "loss": 0.8938, + "step": 1942 + }, + { + "epoch": 0.32576757832966574, + "grad_norm": 0.39877378940582275, + "learning_rate": 1.576969630942552e-05, + "loss": 0.9231, + "step": 1943 + }, + { + "epoch": 0.32593524049041184, + "grad_norm": 0.5871772766113281, + "learning_rate": 1.5765259944937058e-05, + "loss": 0.9229, + "step": 1944 + }, + { + "epoch": 0.32610290265115793, + "grad_norm": 0.7102689743041992, + "learning_rate": 1.5760821880202047e-05, + "loss": 0.9258, + "step": 1945 + }, + { + "epoch": 0.32627056481190403, + "grad_norm": 0.5745960474014282, + "learning_rate": 1.575638211652933e-05, + "loss": 0.999, + "step": 1946 + }, + { + "epoch": 0.3264382269726501, + "grad_norm": 0.5958930850028992, + "learning_rate": 1.5751940655228242e-05, + "loss": 0.9169, + "step": 1947 + }, + { + "epoch": 0.3266058891333962, + "grad_norm": 0.3930474817752838, + "learning_rate": 1.574749749760863e-05, + "loss": 0.9572, + "step": 1948 + }, + { + "epoch": 0.3267735512941423, + "grad_norm": 0.9402393698692322, + "learning_rate": 1.5743052644980838e-05, + "loss": 1.0284, + "step": 1949 + }, + { + "epoch": 0.3269412134548884, + "grad_norm": 0.7290217280387878, + "learning_rate": 1.5738606098655708e-05, + "loss": 0.9351, + "step": 1950 + }, + { + "epoch": 0.3271088756156345, + "grad_norm": 0.7580875158309937, + "learning_rate": 1.5734157859944574e-05, + "loss": 0.9613, + "step": 1951 + }, + { + "epoch": 0.3272765377763806, + "grad_norm": 0.5927433371543884, + "learning_rate": 1.5729707930159282e-05, + "loss": 0.8718, + "step": 1952 + }, + { + "epoch": 0.3274441999371267, + "grad_norm": 0.5822391510009766, + "learning_rate": 1.572525631061218e-05, + "loss": 0.8529, + "step": 1953 + }, + { + "epoch": 0.3276118620978728, + "grad_norm": 0.4330624043941498, + "learning_rate": 1.572080300261609e-05, + "loss": 0.8967, + "step": 1954 + }, + { + "epoch": 0.3277795242586189, + "grad_norm": 0.522164523601532, + "learning_rate": 1.5716348007484358e-05, + "loss": 0.913, + "step": 1955 + }, + { + "epoch": 0.327947186419365, + "grad_norm": 0.403029203414917, + "learning_rate": 1.571189132653081e-05, + "loss": 0.83, + "step": 1956 + }, + { + "epoch": 0.3281148485801111, + "grad_norm": 0.9028471112251282, + "learning_rate": 1.5707432961069785e-05, + "loss": 0.9277, + "step": 1957 + }, + { + "epoch": 0.3282825107408572, + "grad_norm": 0.42061564326286316, + "learning_rate": 1.5702972912416103e-05, + "loss": 0.829, + "step": 1958 + }, + { + "epoch": 0.3284501729016033, + "grad_norm": 0.5323939323425293, + "learning_rate": 1.569851118188509e-05, + "loss": 0.8657, + "step": 1959 + }, + { + "epoch": 0.3286178350623494, + "grad_norm": 0.4818108379840851, + "learning_rate": 1.569404777079257e-05, + "loss": 0.9152, + "step": 1960 + }, + { + "epoch": 0.32878549722309547, + "grad_norm": 0.49384453892707825, + "learning_rate": 1.568958268045485e-05, + "loss": 0.9482, + "step": 1961 + }, + { + "epoch": 0.32895315938384156, + "grad_norm": 0.6106820702552795, + "learning_rate": 1.5685115912188745e-05, + "loss": 0.8776, + "step": 1962 + }, + { + "epoch": 0.32912082154458766, + "grad_norm": 0.6082288026809692, + "learning_rate": 1.568064746731156e-05, + "loss": 0.9401, + "step": 1963 + }, + { + "epoch": 0.32928848370533376, + "grad_norm": 0.5787052512168884, + "learning_rate": 1.5676177347141096e-05, + "loss": 0.8468, + "step": 1964 + }, + { + "epoch": 0.32945614586607985, + "grad_norm": 0.4497988820075989, + "learning_rate": 1.5671705552995642e-05, + "loss": 0.9482, + "step": 1965 + }, + { + "epoch": 0.32962380802682595, + "grad_norm": 0.4412933588027954, + "learning_rate": 1.5667232086193994e-05, + "loss": 0.8962, + "step": 1966 + }, + { + "epoch": 0.32979147018757204, + "grad_norm": 0.4735589921474457, + "learning_rate": 1.566275694805542e-05, + "loss": 0.8334, + "step": 1967 + }, + { + "epoch": 0.32995913234831814, + "grad_norm": 0.44961315393447876, + "learning_rate": 1.5658280139899706e-05, + "loss": 0.9887, + "step": 1968 + }, + { + "epoch": 0.33012679450906424, + "grad_norm": 0.4612094759941101, + "learning_rate": 1.5653801663047113e-05, + "loss": 0.9264, + "step": 1969 + }, + { + "epoch": 0.33029445666981033, + "grad_norm": 0.4764600098133087, + "learning_rate": 1.56493215188184e-05, + "loss": 0.889, + "step": 1970 + }, + { + "epoch": 0.33046211883055643, + "grad_norm": 0.4745563268661499, + "learning_rate": 1.5644839708534817e-05, + "loss": 0.924, + "step": 1971 + }, + { + "epoch": 0.3306297809913025, + "grad_norm": 0.3943979740142822, + "learning_rate": 1.5640356233518103e-05, + "loss": 0.9431, + "step": 1972 + }, + { + "epoch": 0.3307974431520486, + "grad_norm": 0.6608482599258423, + "learning_rate": 1.5635871095090495e-05, + "loss": 0.9135, + "step": 1973 + }, + { + "epoch": 0.3309651053127947, + "grad_norm": 0.497452974319458, + "learning_rate": 1.563138429457471e-05, + "loss": 0.8359, + "step": 1974 + }, + { + "epoch": 0.3311327674735408, + "grad_norm": 0.5150594115257263, + "learning_rate": 1.5626895833293965e-05, + "loss": 0.9055, + "step": 1975 + }, + { + "epoch": 0.3313004296342869, + "grad_norm": 0.43873703479766846, + "learning_rate": 1.562240571257196e-05, + "loss": 0.8912, + "step": 1976 + }, + { + "epoch": 0.331468091795033, + "grad_norm": 0.5945590138435364, + "learning_rate": 1.5617913933732892e-05, + "loss": 0.9009, + "step": 1977 + }, + { + "epoch": 0.3316357539557791, + "grad_norm": 0.5829073190689087, + "learning_rate": 1.5613420498101438e-05, + "loss": 0.9412, + "step": 1978 + }, + { + "epoch": 0.3318034161165252, + "grad_norm": 0.6509609818458557, + "learning_rate": 1.5608925407002774e-05, + "loss": 0.9201, + "step": 1979 + }, + { + "epoch": 0.3319710782772713, + "grad_norm": 0.48291072249412537, + "learning_rate": 1.5604428661762548e-05, + "loss": 0.8516, + "step": 1980 + }, + { + "epoch": 0.3321387404380174, + "grad_norm": 0.5338192582130432, + "learning_rate": 1.559993026370691e-05, + "loss": 0.9614, + "step": 1981 + }, + { + "epoch": 0.3323064025987635, + "grad_norm": 0.4796494245529175, + "learning_rate": 1.5595430214162498e-05, + "loss": 0.845, + "step": 1982 + }, + { + "epoch": 0.3324740647595096, + "grad_norm": 0.575904905796051, + "learning_rate": 1.5590928514456428e-05, + "loss": 0.8648, + "step": 1983 + }, + { + "epoch": 0.3326417269202557, + "grad_norm": 0.42729178071022034, + "learning_rate": 1.5586425165916308e-05, + "loss": 0.8632, + "step": 1984 + }, + { + "epoch": 0.33280938908100177, + "grad_norm": 0.3935372233390808, + "learning_rate": 1.5581920169870227e-05, + "loss": 0.9263, + "step": 1985 + }, + { + "epoch": 0.33297705124174787, + "grad_norm": 0.49478182196617126, + "learning_rate": 1.557741352764677e-05, + "loss": 1.0177, + "step": 1986 + }, + { + "epoch": 0.33314471340249396, + "grad_norm": 0.6155440807342529, + "learning_rate": 1.5572905240575004e-05, + "loss": 0.9465, + "step": 1987 + }, + { + "epoch": 0.33331237556324006, + "grad_norm": 0.4374069571495056, + "learning_rate": 1.556839530998447e-05, + "loss": 0.9038, + "step": 1988 + }, + { + "epoch": 0.33348003772398616, + "grad_norm": 0.5493287444114685, + "learning_rate": 1.5563883737205207e-05, + "loss": 0.8903, + "step": 1989 + }, + { + "epoch": 0.33364769988473225, + "grad_norm": 0.5655654072761536, + "learning_rate": 1.5559370523567734e-05, + "loss": 0.8419, + "step": 1990 + }, + { + "epoch": 0.33381536204547835, + "grad_norm": 0.7805976271629333, + "learning_rate": 1.5554855670403057e-05, + "loss": 0.8981, + "step": 1991 + }, + { + "epoch": 0.33398302420622444, + "grad_norm": 0.6328807473182678, + "learning_rate": 1.555033917904265e-05, + "loss": 0.864, + "step": 1992 + }, + { + "epoch": 0.33415068636697054, + "grad_norm": 0.4354403018951416, + "learning_rate": 1.5545821050818494e-05, + "loss": 0.8553, + "step": 1993 + }, + { + "epoch": 0.33431834852771664, + "grad_norm": 0.5558031797409058, + "learning_rate": 1.5541301287063036e-05, + "loss": 0.9043, + "step": 1994 + }, + { + "epoch": 0.33448601068846273, + "grad_norm": 0.4632706940174103, + "learning_rate": 1.5536779889109213e-05, + "loss": 0.8402, + "step": 1995 + }, + { + "epoch": 0.3346536728492088, + "grad_norm": 0.4619845151901245, + "learning_rate": 1.5532256858290442e-05, + "loss": 0.8918, + "step": 1996 + }, + { + "epoch": 0.3348213350099549, + "grad_norm": 0.6596323251724243, + "learning_rate": 1.5527732195940616e-05, + "loss": 0.8812, + "step": 1997 + }, + { + "epoch": 0.334988997170701, + "grad_norm": 0.5146951675415039, + "learning_rate": 1.5523205903394117e-05, + "loss": 0.93, + "step": 1998 + }, + { + "epoch": 0.3351566593314471, + "grad_norm": 0.4785274565219879, + "learning_rate": 1.5518677981985802e-05, + "loss": 0.8647, + "step": 1999 + }, + { + "epoch": 0.3353243214921932, + "grad_norm": 0.4051639139652252, + "learning_rate": 1.5514148433051018e-05, + "loss": 0.8242, + "step": 2000 + }, + { + "epoch": 0.3354919836529393, + "grad_norm": 0.5596480965614319, + "learning_rate": 1.550961725792558e-05, + "loss": 0.8847, + "step": 2001 + }, + { + "epoch": 0.3356596458136854, + "grad_norm": 0.3232559561729431, + "learning_rate": 1.5505084457945786e-05, + "loss": 0.8139, + "step": 2002 + }, + { + "epoch": 0.3358273079744315, + "grad_norm": 0.4186466336250305, + "learning_rate": 1.5500550034448415e-05, + "loss": 0.7907, + "step": 2003 + }, + { + "epoch": 0.3359949701351776, + "grad_norm": 0.44500428438186646, + "learning_rate": 1.549601398877073e-05, + "loss": 0.9744, + "step": 2004 + }, + { + "epoch": 0.3361626322959237, + "grad_norm": 0.5455708503723145, + "learning_rate": 1.549147632225046e-05, + "loss": 0.8561, + "step": 2005 + }, + { + "epoch": 0.3363302944566698, + "grad_norm": 0.43856748938560486, + "learning_rate": 1.5486937036225824e-05, + "loss": 0.8707, + "step": 2006 + }, + { + "epoch": 0.3364979566174159, + "grad_norm": 0.5387969017028809, + "learning_rate": 1.5482396132035515e-05, + "loss": 0.8936, + "step": 2007 + }, + { + "epoch": 0.336665618778162, + "grad_norm": 0.4660685062408447, + "learning_rate": 1.5477853611018697e-05, + "loss": 0.8289, + "step": 2008 + }, + { + "epoch": 0.3368332809389081, + "grad_norm": 0.6854202151298523, + "learning_rate": 1.5473309474515014e-05, + "loss": 0.9128, + "step": 2009 + }, + { + "epoch": 0.33700094309965417, + "grad_norm": 0.6004125475883484, + "learning_rate": 1.5468763723864598e-05, + "loss": 0.9228, + "step": 2010 + }, + { + "epoch": 0.33716860526040027, + "grad_norm": 0.5102965831756592, + "learning_rate": 1.546421636040804e-05, + "loss": 0.855, + "step": 2011 + }, + { + "epoch": 0.33733626742114636, + "grad_norm": 0.38877010345458984, + "learning_rate": 1.5459667385486416e-05, + "loss": 0.9162, + "step": 2012 + }, + { + "epoch": 0.33750392958189246, + "grad_norm": 0.5080621242523193, + "learning_rate": 1.545511680044127e-05, + "loss": 0.9592, + "step": 2013 + }, + { + "epoch": 0.3376715917426386, + "grad_norm": 0.4342464208602905, + "learning_rate": 1.5450564606614635e-05, + "loss": 0.8897, + "step": 2014 + }, + { + "epoch": 0.3378392539033847, + "grad_norm": 0.5745513439178467, + "learning_rate": 1.5446010805348997e-05, + "loss": 1.0136, + "step": 2015 + }, + { + "epoch": 0.3380069160641308, + "grad_norm": 0.4189508259296417, + "learning_rate": 1.5441455397987342e-05, + "loss": 0.9253, + "step": 2016 + }, + { + "epoch": 0.3381745782248769, + "grad_norm": 0.5819382071495056, + "learning_rate": 1.5436898385873105e-05, + "loss": 0.8154, + "step": 2017 + }, + { + "epoch": 0.338342240385623, + "grad_norm": 0.4832051694393158, + "learning_rate": 1.5432339770350212e-05, + "loss": 0.8503, + "step": 2018 + }, + { + "epoch": 0.3385099025463691, + "grad_norm": 0.47551408410072327, + "learning_rate": 1.5427779552763053e-05, + "loss": 0.8631, + "step": 2019 + }, + { + "epoch": 0.3386775647071152, + "grad_norm": 0.5339730978012085, + "learning_rate": 1.5423217734456493e-05, + "loss": 0.9093, + "step": 2020 + }, + { + "epoch": 0.3388452268678613, + "grad_norm": 0.496891051530838, + "learning_rate": 1.5418654316775868e-05, + "loss": 0.9656, + "step": 2021 + }, + { + "epoch": 0.3390128890286074, + "grad_norm": 0.5256526470184326, + "learning_rate": 1.541408930106699e-05, + "loss": 0.8912, + "step": 2022 + }, + { + "epoch": 0.3391805511893535, + "grad_norm": 0.3820645213127136, + "learning_rate": 1.5409522688676132e-05, + "loss": 0.8646, + "step": 2023 + }, + { + "epoch": 0.33934821335009957, + "grad_norm": 0.6550490260124207, + "learning_rate": 1.5404954480950047e-05, + "loss": 0.8541, + "step": 2024 + }, + { + "epoch": 0.33951587551084567, + "grad_norm": 0.41333290934562683, + "learning_rate": 1.540038467923596e-05, + "loss": 0.8607, + "step": 2025 + }, + { + "epoch": 0.33968353767159176, + "grad_norm": 0.5534454584121704, + "learning_rate": 1.539581328488156e-05, + "loss": 0.82, + "step": 2026 + }, + { + "epoch": 0.33985119983233786, + "grad_norm": 0.4226256012916565, + "learning_rate": 1.5391240299235e-05, + "loss": 0.8238, + "step": 2027 + }, + { + "epoch": 0.34001886199308395, + "grad_norm": 0.47660452127456665, + "learning_rate": 1.5386665723644923e-05, + "loss": 0.8789, + "step": 2028 + }, + { + "epoch": 0.34018652415383005, + "grad_norm": 0.6586236953735352, + "learning_rate": 1.5382089559460423e-05, + "loss": 0.9371, + "step": 2029 + }, + { + "epoch": 0.34035418631457615, + "grad_norm": 0.386690229177475, + "learning_rate": 1.5377511808031064e-05, + "loss": 0.8813, + "step": 2030 + }, + { + "epoch": 0.34052184847532224, + "grad_norm": 0.4806725084781647, + "learning_rate": 1.5372932470706885e-05, + "loss": 0.8278, + "step": 2031 + }, + { + "epoch": 0.34068951063606834, + "grad_norm": 0.6102979183197021, + "learning_rate": 1.5368351548838386e-05, + "loss": 0.8384, + "step": 2032 + }, + { + "epoch": 0.34085717279681443, + "grad_norm": 0.5258501172065735, + "learning_rate": 1.5363769043776545e-05, + "loss": 0.8905, + "step": 2033 + }, + { + "epoch": 0.34102483495756053, + "grad_norm": 0.4061277508735657, + "learning_rate": 1.5359184956872785e-05, + "loss": 0.9178, + "step": 2034 + }, + { + "epoch": 0.3411924971183066, + "grad_norm": 0.43954938650131226, + "learning_rate": 1.535459928947903e-05, + "loss": 0.8698, + "step": 2035 + }, + { + "epoch": 0.3413601592790527, + "grad_norm": 0.4769081771373749, + "learning_rate": 1.5350012042947633e-05, + "loss": 0.9357, + "step": 2036 + }, + { + "epoch": 0.3415278214397988, + "grad_norm": 0.7200191617012024, + "learning_rate": 1.5345423218631438e-05, + "loss": 0.8827, + "step": 2037 + }, + { + "epoch": 0.3416954836005449, + "grad_norm": 0.4411308169364929, + "learning_rate": 1.5340832817883743e-05, + "loss": 0.8839, + "step": 2038 + }, + { + "epoch": 0.341863145761291, + "grad_norm": 0.4716808497905731, + "learning_rate": 1.533624084205832e-05, + "loss": 0.8936, + "step": 2039 + }, + { + "epoch": 0.3420308079220371, + "grad_norm": 0.38220396637916565, + "learning_rate": 1.5331647292509394e-05, + "loss": 0.9168, + "step": 2040 + }, + { + "epoch": 0.3421984700827832, + "grad_norm": 0.3923315107822418, + "learning_rate": 1.532705217059166e-05, + "loss": 0.9659, + "step": 2041 + }, + { + "epoch": 0.3423661322435293, + "grad_norm": 0.7197695970535278, + "learning_rate": 1.5322455477660274e-05, + "loss": 0.8764, + "step": 2042 + }, + { + "epoch": 0.3425337944042754, + "grad_norm": 0.577632486820221, + "learning_rate": 1.531785721507086e-05, + "loss": 0.8478, + "step": 2043 + }, + { + "epoch": 0.3427014565650215, + "grad_norm": 0.4102010428905487, + "learning_rate": 1.5313257384179508e-05, + "loss": 0.8483, + "step": 2044 + }, + { + "epoch": 0.3428691187257676, + "grad_norm": 0.37063077092170715, + "learning_rate": 1.5308655986342757e-05, + "loss": 0.9228, + "step": 2045 + }, + { + "epoch": 0.3430367808865137, + "grad_norm": 0.6842082142829895, + "learning_rate": 1.530405302291762e-05, + "loss": 0.8926, + "step": 2046 + }, + { + "epoch": 0.3432044430472598, + "grad_norm": 0.47118207812309265, + "learning_rate": 1.529944849526157e-05, + "loss": 0.9021, + "step": 2047 + }, + { + "epoch": 0.3433721052080059, + "grad_norm": 0.6092045307159424, + "learning_rate": 1.5294842404732535e-05, + "loss": 0.9341, + "step": 2048 + }, + { + "epoch": 0.34353976736875197, + "grad_norm": 0.6156114339828491, + "learning_rate": 1.529023475268891e-05, + "loss": 0.8798, + "step": 2049 + }, + { + "epoch": 0.34370742952949807, + "grad_norm": 0.5023700594902039, + "learning_rate": 1.528562554048955e-05, + "loss": 0.9569, + "step": 2050 + }, + { + "epoch": 0.34387509169024416, + "grad_norm": 0.4121425449848175, + "learning_rate": 1.5281014769493766e-05, + "loss": 0.8448, + "step": 2051 + }, + { + "epoch": 0.34404275385099026, + "grad_norm": 0.5774893760681152, + "learning_rate": 1.527640244106133e-05, + "loss": 0.8717, + "step": 2052 + }, + { + "epoch": 0.34421041601173635, + "grad_norm": 0.5007169842720032, + "learning_rate": 1.5271788556552477e-05, + "loss": 0.9379, + "step": 2053 + }, + { + "epoch": 0.34437807817248245, + "grad_norm": 0.49954840540885925, + "learning_rate": 1.52671731173279e-05, + "loss": 0.8518, + "step": 2054 + }, + { + "epoch": 0.34454574033322855, + "grad_norm": 0.7455732822418213, + "learning_rate": 1.5262556124748754e-05, + "loss": 0.9908, + "step": 2055 + }, + { + "epoch": 0.34471340249397464, + "grad_norm": 0.43593671917915344, + "learning_rate": 1.5257937580176636e-05, + "loss": 0.9346, + "step": 2056 + }, + { + "epoch": 0.34488106465472074, + "grad_norm": 0.4693664014339447, + "learning_rate": 1.525331748497362e-05, + "loss": 0.9401, + "step": 2057 + }, + { + "epoch": 0.34504872681546683, + "grad_norm": 0.45605477690696716, + "learning_rate": 1.5248695840502223e-05, + "loss": 0.8253, + "step": 2058 + }, + { + "epoch": 0.34521638897621293, + "grad_norm": 1.0499744415283203, + "learning_rate": 1.5244072648125432e-05, + "loss": 0.9594, + "step": 2059 + }, + { + "epoch": 0.345384051136959, + "grad_norm": 0.5815116167068481, + "learning_rate": 1.5239447909206679e-05, + "loss": 0.8302, + "step": 2060 + }, + { + "epoch": 0.3455517132977051, + "grad_norm": 0.34735602140426636, + "learning_rate": 1.5234821625109862e-05, + "loss": 0.922, + "step": 2061 + }, + { + "epoch": 0.3457193754584512, + "grad_norm": 1.402817964553833, + "learning_rate": 1.5230193797199324e-05, + "loss": 0.8915, + "step": 2062 + }, + { + "epoch": 0.3458870376191973, + "grad_norm": 0.44540536403656006, + "learning_rate": 1.5225564426839872e-05, + "loss": 0.8578, + "step": 2063 + }, + { + "epoch": 0.3460546997799434, + "grad_norm": 0.46437814831733704, + "learning_rate": 1.522093351539676e-05, + "loss": 0.8542, + "step": 2064 + }, + { + "epoch": 0.3462223619406895, + "grad_norm": 0.39311614632606506, + "learning_rate": 1.5216301064235706e-05, + "loss": 0.7968, + "step": 2065 + }, + { + "epoch": 0.3463900241014356, + "grad_norm": 0.8382595181465149, + "learning_rate": 1.5211667074722875e-05, + "loss": 1.0011, + "step": 2066 + }, + { + "epoch": 0.3465576862621817, + "grad_norm": 0.5507627129554749, + "learning_rate": 1.520703154822489e-05, + "loss": 0.8895, + "step": 2067 + }, + { + "epoch": 0.3467253484229278, + "grad_norm": 0.5316657423973083, + "learning_rate": 1.5202394486108823e-05, + "loss": 0.9189, + "step": 2068 + }, + { + "epoch": 0.3468930105836739, + "grad_norm": 0.6288915276527405, + "learning_rate": 1.5197755889742202e-05, + "loss": 0.915, + "step": 2069 + }, + { + "epoch": 0.34706067274442, + "grad_norm": 0.6252545714378357, + "learning_rate": 1.5193115760493007e-05, + "loss": 0.8947, + "step": 2070 + }, + { + "epoch": 0.3472283349051661, + "grad_norm": 0.5686506628990173, + "learning_rate": 1.5188474099729669e-05, + "loss": 0.8418, + "step": 2071 + }, + { + "epoch": 0.3473959970659122, + "grad_norm": 0.42737486958503723, + "learning_rate": 1.5183830908821069e-05, + "loss": 0.8932, + "step": 2072 + }, + { + "epoch": 0.3475636592266583, + "grad_norm": 0.5029768943786621, + "learning_rate": 1.5179186189136545e-05, + "loss": 0.9382, + "step": 2073 + }, + { + "epoch": 0.34773132138740437, + "grad_norm": 0.6150388717651367, + "learning_rate": 1.517453994204588e-05, + "loss": 0.8753, + "step": 2074 + }, + { + "epoch": 0.34789898354815046, + "grad_norm": 0.40642431378364563, + "learning_rate": 1.5169892168919312e-05, + "loss": 0.9051, + "step": 2075 + }, + { + "epoch": 0.34806664570889656, + "grad_norm": 0.5509570240974426, + "learning_rate": 1.5165242871127523e-05, + "loss": 0.8797, + "step": 2076 + }, + { + "epoch": 0.34823430786964266, + "grad_norm": 0.5052769184112549, + "learning_rate": 1.5160592050041654e-05, + "loss": 0.91, + "step": 2077 + }, + { + "epoch": 0.34840197003038875, + "grad_norm": 0.601817786693573, + "learning_rate": 1.5155939707033287e-05, + "loss": 0.9322, + "step": 2078 + }, + { + "epoch": 0.34856963219113485, + "grad_norm": 0.5623035430908203, + "learning_rate": 1.5151285843474454e-05, + "loss": 0.8976, + "step": 2079 + }, + { + "epoch": 0.34873729435188094, + "grad_norm": 0.5203773975372314, + "learning_rate": 1.5146630460737638e-05, + "loss": 0.8305, + "step": 2080 + }, + { + "epoch": 0.34890495651262704, + "grad_norm": 0.5576394200325012, + "learning_rate": 1.5141973560195768e-05, + "loss": 0.9809, + "step": 2081 + }, + { + "epoch": 0.34907261867337314, + "grad_norm": 0.4212713837623596, + "learning_rate": 1.5137315143222227e-05, + "loss": 0.8525, + "step": 2082 + }, + { + "epoch": 0.34924028083411923, + "grad_norm": 0.7241677641868591, + "learning_rate": 1.5132655211190832e-05, + "loss": 0.8777, + "step": 2083 + }, + { + "epoch": 0.34940794299486533, + "grad_norm": 0.4672906994819641, + "learning_rate": 1.5127993765475859e-05, + "loss": 0.894, + "step": 2084 + }, + { + "epoch": 0.3495756051556114, + "grad_norm": 0.38957101106643677, + "learning_rate": 1.512333080745203e-05, + "loss": 0.9166, + "step": 2085 + }, + { + "epoch": 0.3497432673163575, + "grad_norm": 0.4429856836795807, + "learning_rate": 1.5118666338494501e-05, + "loss": 0.9079, + "step": 2086 + }, + { + "epoch": 0.3499109294771036, + "grad_norm": 0.7306303977966309, + "learning_rate": 1.5114000359978884e-05, + "loss": 0.9717, + "step": 2087 + }, + { + "epoch": 0.3500785916378497, + "grad_norm": 0.45758989453315735, + "learning_rate": 1.5109332873281237e-05, + "loss": 0.8567, + "step": 2088 + }, + { + "epoch": 0.3502462537985958, + "grad_norm": 0.6235557794570923, + "learning_rate": 1.510466387977806e-05, + "loss": 0.8931, + "step": 2089 + }, + { + "epoch": 0.3504139159593419, + "grad_norm": 0.4747142195701599, + "learning_rate": 1.5099993380846291e-05, + "loss": 0.8268, + "step": 2090 + }, + { + "epoch": 0.350581578120088, + "grad_norm": 0.505138635635376, + "learning_rate": 1.509532137786332e-05, + "loss": 0.8828, + "step": 2091 + }, + { + "epoch": 0.3507492402808341, + "grad_norm": 0.3458777964115143, + "learning_rate": 1.5090647872206984e-05, + "loss": 0.8832, + "step": 2092 + }, + { + "epoch": 0.3509169024415802, + "grad_norm": 0.46667036414146423, + "learning_rate": 1.5085972865255556e-05, + "loss": 0.9134, + "step": 2093 + }, + { + "epoch": 0.3510845646023263, + "grad_norm": 0.4600054919719696, + "learning_rate": 1.508129635838775e-05, + "loss": 0.8623, + "step": 2094 + }, + { + "epoch": 0.3512522267630724, + "grad_norm": 0.4440256953239441, + "learning_rate": 1.5076618352982724e-05, + "loss": 0.9679, + "step": 2095 + }, + { + "epoch": 0.3514198889238185, + "grad_norm": 0.48019370436668396, + "learning_rate": 1.5071938850420087e-05, + "loss": 0.9159, + "step": 2096 + }, + { + "epoch": 0.3515875510845646, + "grad_norm": 0.5663225650787354, + "learning_rate": 1.5067257852079876e-05, + "loss": 0.8541, + "step": 2097 + }, + { + "epoch": 0.35175521324531067, + "grad_norm": 0.4951961040496826, + "learning_rate": 1.5062575359342575e-05, + "loss": 0.9002, + "step": 2098 + }, + { + "epoch": 0.3519228754060568, + "grad_norm": 0.4354960024356842, + "learning_rate": 1.5057891373589113e-05, + "loss": 0.9126, + "step": 2099 + }, + { + "epoch": 0.3520905375668029, + "grad_norm": 0.5936130285263062, + "learning_rate": 1.5053205896200856e-05, + "loss": 0.9094, + "step": 2100 + }, + { + "epoch": 0.352258199727549, + "grad_norm": 0.6447715163230896, + "learning_rate": 1.5048518928559608e-05, + "loss": 0.9056, + "step": 2101 + }, + { + "epoch": 0.3524258618882951, + "grad_norm": 0.5037466287612915, + "learning_rate": 1.504383047204761e-05, + "loss": 0.8379, + "step": 2102 + }, + { + "epoch": 0.3525935240490412, + "grad_norm": 0.587268590927124, + "learning_rate": 1.5039140528047548e-05, + "loss": 0.8673, + "step": 2103 + }, + { + "epoch": 0.3527611862097873, + "grad_norm": 1.0174998044967651, + "learning_rate": 1.5034449097942549e-05, + "loss": 0.9398, + "step": 2104 + }, + { + "epoch": 0.3529288483705334, + "grad_norm": 0.7159402966499329, + "learning_rate": 1.5029756183116171e-05, + "loss": 0.8737, + "step": 2105 + }, + { + "epoch": 0.3530965105312795, + "grad_norm": 0.46339261531829834, + "learning_rate": 1.5025061784952408e-05, + "loss": 0.8339, + "step": 2106 + }, + { + "epoch": 0.3532641726920256, + "grad_norm": 0.810058057308197, + "learning_rate": 1.50203659048357e-05, + "loss": 0.8888, + "step": 2107 + }, + { + "epoch": 0.3534318348527717, + "grad_norm": 0.5692547559738159, + "learning_rate": 1.5015668544150925e-05, + "loss": 0.9204, + "step": 2108 + }, + { + "epoch": 0.3535994970135178, + "grad_norm": 0.4281153976917267, + "learning_rate": 1.5010969704283387e-05, + "loss": 0.7818, + "step": 2109 + }, + { + "epoch": 0.3537671591742639, + "grad_norm": 0.4954259693622589, + "learning_rate": 1.5006269386618833e-05, + "loss": 0.9107, + "step": 2110 + }, + { + "epoch": 0.35393482133501, + "grad_norm": 0.5465715527534485, + "learning_rate": 1.5001567592543445e-05, + "loss": 0.9045, + "step": 2111 + }, + { + "epoch": 0.35410248349575607, + "grad_norm": 0.5770376920700073, + "learning_rate": 1.499686432344384e-05, + "loss": 0.8672, + "step": 2112 + }, + { + "epoch": 0.35427014565650217, + "grad_norm": 0.583249568939209, + "learning_rate": 1.499215958070707e-05, + "loss": 0.9409, + "step": 2113 + }, + { + "epoch": 0.35443780781724826, + "grad_norm": 0.5041799545288086, + "learning_rate": 1.4987453365720624e-05, + "loss": 0.8529, + "step": 2114 + }, + { + "epoch": 0.35460546997799436, + "grad_norm": 0.4735868573188782, + "learning_rate": 1.4982745679872421e-05, + "loss": 0.9399, + "step": 2115 + }, + { + "epoch": 0.35477313213874045, + "grad_norm": 0.4119265079498291, + "learning_rate": 1.4978036524550816e-05, + "loss": 0.8372, + "step": 2116 + }, + { + "epoch": 0.35494079429948655, + "grad_norm": 0.4211297929286957, + "learning_rate": 1.49733259011446e-05, + "loss": 0.8324, + "step": 2117 + }, + { + "epoch": 0.35510845646023265, + "grad_norm": 0.5091115236282349, + "learning_rate": 1.496861381104299e-05, + "loss": 0.8963, + "step": 2118 + }, + { + "epoch": 0.35527611862097874, + "grad_norm": 0.5502423048019409, + "learning_rate": 1.4963900255635644e-05, + "loss": 0.8684, + "step": 2119 + }, + { + "epoch": 0.35544378078172484, + "grad_norm": 0.5665629506111145, + "learning_rate": 1.4959185236312642e-05, + "loss": 0.9255, + "step": 2120 + }, + { + "epoch": 0.35561144294247093, + "grad_norm": 0.5827038884162903, + "learning_rate": 1.4954468754464509e-05, + "loss": 0.8963, + "step": 2121 + }, + { + "epoch": 0.35577910510321703, + "grad_norm": 0.47760000824928284, + "learning_rate": 1.4949750811482187e-05, + "loss": 0.9071, + "step": 2122 + }, + { + "epoch": 0.3559467672639631, + "grad_norm": 0.5717841982841492, + "learning_rate": 1.4945031408757058e-05, + "loss": 0.864, + "step": 2123 + }, + { + "epoch": 0.3561144294247092, + "grad_norm": 0.5802021622657776, + "learning_rate": 1.4940310547680938e-05, + "loss": 0.9182, + "step": 2124 + }, + { + "epoch": 0.3562820915854553, + "grad_norm": 0.632815957069397, + "learning_rate": 1.4935588229646057e-05, + "loss": 0.8659, + "step": 2125 + }, + { + "epoch": 0.3564497537462014, + "grad_norm": 1.1413359642028809, + "learning_rate": 1.4930864456045096e-05, + "loss": 0.9987, + "step": 2126 + }, + { + "epoch": 0.3566174159069475, + "grad_norm": 0.5720139145851135, + "learning_rate": 1.4926139228271143e-05, + "loss": 0.8572, + "step": 2127 + }, + { + "epoch": 0.3567850780676936, + "grad_norm": 0.473939448595047, + "learning_rate": 1.492141254771774e-05, + "loss": 0.9373, + "step": 2128 + }, + { + "epoch": 0.3569527402284397, + "grad_norm": 0.6179938316345215, + "learning_rate": 1.4916684415778828e-05, + "loss": 0.9507, + "step": 2129 + }, + { + "epoch": 0.3571204023891858, + "grad_norm": 0.48950061202049255, + "learning_rate": 1.4911954833848805e-05, + "loss": 0.9136, + "step": 2130 + }, + { + "epoch": 0.3572880645499319, + "grad_norm": 0.46943745017051697, + "learning_rate": 1.4907223803322474e-05, + "loss": 0.781, + "step": 2131 + }, + { + "epoch": 0.357455726710678, + "grad_norm": 0.3945772647857666, + "learning_rate": 1.490249132559508e-05, + "loss": 0.9011, + "step": 2132 + }, + { + "epoch": 0.3576233888714241, + "grad_norm": 0.4948115944862366, + "learning_rate": 1.4897757402062285e-05, + "loss": 0.8314, + "step": 2133 + }, + { + "epoch": 0.3577910510321702, + "grad_norm": 0.4736458957195282, + "learning_rate": 1.4893022034120186e-05, + "loss": 0.7921, + "step": 2134 + }, + { + "epoch": 0.3579587131929163, + "grad_norm": 0.618366003036499, + "learning_rate": 1.4888285223165299e-05, + "loss": 0.8579, + "step": 2135 + }, + { + "epoch": 0.3581263753536624, + "grad_norm": 0.7641869783401489, + "learning_rate": 1.4883546970594571e-05, + "loss": 0.9287, + "step": 2136 + }, + { + "epoch": 0.35829403751440847, + "grad_norm": 0.4503091275691986, + "learning_rate": 1.4878807277805366e-05, + "loss": 0.923, + "step": 2137 + }, + { + "epoch": 0.35846169967515457, + "grad_norm": 0.42575526237487793, + "learning_rate": 1.4874066146195485e-05, + "loss": 0.9257, + "step": 2138 + }, + { + "epoch": 0.35862936183590066, + "grad_norm": 0.5589778423309326, + "learning_rate": 1.4869323577163139e-05, + "loss": 0.918, + "step": 2139 + }, + { + "epoch": 0.35879702399664676, + "grad_norm": 0.5460204482078552, + "learning_rate": 1.4864579572106977e-05, + "loss": 0.9419, + "step": 2140 + }, + { + "epoch": 0.35896468615739285, + "grad_norm": 0.4740411937236786, + "learning_rate": 1.485983413242606e-05, + "loss": 0.8739, + "step": 2141 + }, + { + "epoch": 0.35913234831813895, + "grad_norm": 0.7383924722671509, + "learning_rate": 1.4855087259519882e-05, + "loss": 0.9669, + "step": 2142 + }, + { + "epoch": 0.35930001047888505, + "grad_norm": 0.4387388825416565, + "learning_rate": 1.485033895478835e-05, + "loss": 0.91, + "step": 2143 + }, + { + "epoch": 0.35946767263963114, + "grad_norm": 0.41963890194892883, + "learning_rate": 1.48455892196318e-05, + "loss": 0.8902, + "step": 2144 + }, + { + "epoch": 0.35963533480037724, + "grad_norm": 0.4966847598552704, + "learning_rate": 1.4840838055450986e-05, + "loss": 0.9053, + "step": 2145 + }, + { + "epoch": 0.35980299696112333, + "grad_norm": 0.9311870336532593, + "learning_rate": 1.4836085463647088e-05, + "loss": 0.8795, + "step": 2146 + }, + { + "epoch": 0.35997065912186943, + "grad_norm": 0.6284542679786682, + "learning_rate": 1.4831331445621699e-05, + "loss": 0.9396, + "step": 2147 + }, + { + "epoch": 0.3601383212826155, + "grad_norm": 0.39609295129776, + "learning_rate": 1.4826576002776842e-05, + "loss": 0.8391, + "step": 2148 + }, + { + "epoch": 0.3603059834433616, + "grad_norm": 0.7315961718559265, + "learning_rate": 1.4821819136514956e-05, + "loss": 0.9071, + "step": 2149 + }, + { + "epoch": 0.3604736456041077, + "grad_norm": 0.3803544342517853, + "learning_rate": 1.4817060848238902e-05, + "loss": 0.9259, + "step": 2150 + }, + { + "epoch": 0.3606413077648538, + "grad_norm": 0.3933790922164917, + "learning_rate": 1.4812301139351953e-05, + "loss": 0.9043, + "step": 2151 + }, + { + "epoch": 0.3608089699255999, + "grad_norm": 0.5101569294929504, + "learning_rate": 1.4807540011257806e-05, + "loss": 0.9055, + "step": 2152 + }, + { + "epoch": 0.360976632086346, + "grad_norm": 0.4260362982749939, + "learning_rate": 1.4802777465360581e-05, + "loss": 0.8096, + "step": 2153 + }, + { + "epoch": 0.3611442942470921, + "grad_norm": 0.4232743978500366, + "learning_rate": 1.4798013503064809e-05, + "loss": 1.0025, + "step": 2154 + }, + { + "epoch": 0.3613119564078382, + "grad_norm": 0.6433045864105225, + "learning_rate": 1.4793248125775442e-05, + "loss": 0.8771, + "step": 2155 + }, + { + "epoch": 0.3614796185685843, + "grad_norm": 0.42645955085754395, + "learning_rate": 1.478848133489785e-05, + "loss": 0.8817, + "step": 2156 + }, + { + "epoch": 0.3616472807293304, + "grad_norm": 0.5128361582756042, + "learning_rate": 1.4783713131837813e-05, + "loss": 0.9481, + "step": 2157 + }, + { + "epoch": 0.3618149428900765, + "grad_norm": 0.46490219235420227, + "learning_rate": 1.4778943518001543e-05, + "loss": 0.8478, + "step": 2158 + }, + { + "epoch": 0.3619826050508226, + "grad_norm": 0.5935733318328857, + "learning_rate": 1.4774172494795651e-05, + "loss": 0.8164, + "step": 2159 + }, + { + "epoch": 0.3621502672115687, + "grad_norm": 0.40863484144210815, + "learning_rate": 1.476940006362717e-05, + "loss": 0.8704, + "step": 2160 + }, + { + "epoch": 0.3623179293723148, + "grad_norm": 1.119663119316101, + "learning_rate": 1.4764626225903555e-05, + "loss": 0.9581, + "step": 2161 + }, + { + "epoch": 0.36248559153306087, + "grad_norm": 0.5529540181159973, + "learning_rate": 1.4759850983032666e-05, + "loss": 0.8493, + "step": 2162 + }, + { + "epoch": 0.36265325369380697, + "grad_norm": 0.6024941205978394, + "learning_rate": 1.4755074336422778e-05, + "loss": 0.9546, + "step": 2163 + }, + { + "epoch": 0.36282091585455306, + "grad_norm": 0.34549546241760254, + "learning_rate": 1.475029628748259e-05, + "loss": 0.8187, + "step": 2164 + }, + { + "epoch": 0.36298857801529916, + "grad_norm": 0.39582517743110657, + "learning_rate": 1.4745516837621207e-05, + "loss": 0.9273, + "step": 2165 + }, + { + "epoch": 0.36315624017604525, + "grad_norm": 0.47179707884788513, + "learning_rate": 1.4740735988248144e-05, + "loss": 0.9474, + "step": 2166 + }, + { + "epoch": 0.36332390233679135, + "grad_norm": 0.4781091809272766, + "learning_rate": 1.4735953740773334e-05, + "loss": 0.9955, + "step": 2167 + }, + { + "epoch": 0.36349156449753744, + "grad_norm": 0.3955364525318146, + "learning_rate": 1.4731170096607121e-05, + "loss": 0.8759, + "step": 2168 + }, + { + "epoch": 0.36365922665828354, + "grad_norm": 0.5289798974990845, + "learning_rate": 1.4726385057160264e-05, + "loss": 0.9638, + "step": 2169 + }, + { + "epoch": 0.36382688881902964, + "grad_norm": 0.417681485414505, + "learning_rate": 1.4721598623843924e-05, + "loss": 0.8737, + "step": 2170 + }, + { + "epoch": 0.36399455097977573, + "grad_norm": 0.5590136051177979, + "learning_rate": 1.4716810798069683e-05, + "loss": 0.9051, + "step": 2171 + }, + { + "epoch": 0.36416221314052183, + "grad_norm": 0.5966497659683228, + "learning_rate": 1.4712021581249534e-05, + "loss": 0.8726, + "step": 2172 + }, + { + "epoch": 0.3643298753012679, + "grad_norm": 0.5650334358215332, + "learning_rate": 1.4707230974795869e-05, + "loss": 0.9522, + "step": 2173 + }, + { + "epoch": 0.364497537462014, + "grad_norm": 0.46143072843551636, + "learning_rate": 1.47024389801215e-05, + "loss": 0.9797, + "step": 2174 + }, + { + "epoch": 0.3646651996227601, + "grad_norm": 0.4127398133277893, + "learning_rate": 1.4697645598639652e-05, + "loss": 0.9056, + "step": 2175 + }, + { + "epoch": 0.3648328617835062, + "grad_norm": 0.6077963709831238, + "learning_rate": 1.4692850831763943e-05, + "loss": 0.9559, + "step": 2176 + }, + { + "epoch": 0.3650005239442523, + "grad_norm": 0.4098505973815918, + "learning_rate": 1.4688054680908414e-05, + "loss": 0.8225, + "step": 2177 + }, + { + "epoch": 0.3651681861049984, + "grad_norm": 0.5953378677368164, + "learning_rate": 1.4683257147487506e-05, + "loss": 0.8915, + "step": 2178 + }, + { + "epoch": 0.3653358482657445, + "grad_norm": 0.5840544104576111, + "learning_rate": 1.4678458232916075e-05, + "loss": 0.8355, + "step": 2179 + }, + { + "epoch": 0.3655035104264906, + "grad_norm": 0.4457241892814636, + "learning_rate": 1.467365793860938e-05, + "loss": 0.8149, + "step": 2180 + }, + { + "epoch": 0.3656711725872367, + "grad_norm": 0.5155458450317383, + "learning_rate": 1.4668856265983085e-05, + "loss": 0.8703, + "step": 2181 + }, + { + "epoch": 0.3658388347479828, + "grad_norm": 0.5299484133720398, + "learning_rate": 1.466405321645326e-05, + "loss": 0.9161, + "step": 2182 + }, + { + "epoch": 0.3660064969087289, + "grad_norm": 0.44516849517822266, + "learning_rate": 1.4659248791436393e-05, + "loss": 0.9276, + "step": 2183 + }, + { + "epoch": 0.366174159069475, + "grad_norm": 0.45529067516326904, + "learning_rate": 1.465444299234936e-05, + "loss": 0.8972, + "step": 2184 + }, + { + "epoch": 0.36634182123022113, + "grad_norm": 0.4852927625179291, + "learning_rate": 1.4649635820609457e-05, + "loss": 0.8844, + "step": 2185 + }, + { + "epoch": 0.36650948339096723, + "grad_norm": 0.5148529410362244, + "learning_rate": 1.4644827277634367e-05, + "loss": 0.9069, + "step": 2186 + }, + { + "epoch": 0.3666771455517133, + "grad_norm": 0.40305307507514954, + "learning_rate": 1.4640017364842205e-05, + "loss": 0.8885, + "step": 2187 + }, + { + "epoch": 0.3668448077124594, + "grad_norm": 0.4781140685081482, + "learning_rate": 1.4635206083651461e-05, + "loss": 0.8817, + "step": 2188 + }, + { + "epoch": 0.3670124698732055, + "grad_norm": 0.5818206071853638, + "learning_rate": 1.4630393435481047e-05, + "loss": 0.8572, + "step": 2189 + }, + { + "epoch": 0.3671801320339516, + "grad_norm": 0.4082125425338745, + "learning_rate": 1.4625579421750271e-05, + "loss": 0.8723, + "step": 2190 + }, + { + "epoch": 0.3673477941946977, + "grad_norm": 0.4111286997795105, + "learning_rate": 1.4620764043878845e-05, + "loss": 0.8568, + "step": 2191 + }, + { + "epoch": 0.3675154563554438, + "grad_norm": 0.5303593873977661, + "learning_rate": 1.4615947303286888e-05, + "loss": 0.8298, + "step": 2192 + }, + { + "epoch": 0.3676831185161899, + "grad_norm": 0.4056621491909027, + "learning_rate": 1.4611129201394911e-05, + "loss": 0.9032, + "step": 2193 + }, + { + "epoch": 0.367850780676936, + "grad_norm": 0.4418686628341675, + "learning_rate": 1.4606309739623834e-05, + "loss": 0.8864, + "step": 2194 + }, + { + "epoch": 0.3680184428376821, + "grad_norm": 0.4276810586452484, + "learning_rate": 1.4601488919394975e-05, + "loss": 0.9574, + "step": 2195 + }, + { + "epoch": 0.3681861049984282, + "grad_norm": 0.8117255568504333, + "learning_rate": 1.4596666742130055e-05, + "loss": 0.749, + "step": 2196 + }, + { + "epoch": 0.3683537671591743, + "grad_norm": 1.4907840490341187, + "learning_rate": 1.4591843209251191e-05, + "loss": 0.8621, + "step": 2197 + }, + { + "epoch": 0.3685214293199204, + "grad_norm": 0.5102953910827637, + "learning_rate": 1.4587018322180906e-05, + "loss": 0.9216, + "step": 2198 + }, + { + "epoch": 0.3686890914806665, + "grad_norm": 1.0061148405075073, + "learning_rate": 1.458219208234212e-05, + "loss": 0.9426, + "step": 2199 + }, + { + "epoch": 0.36885675364141257, + "grad_norm": 0.6368162631988525, + "learning_rate": 1.4577364491158147e-05, + "loss": 0.9052, + "step": 2200 + }, + { + "epoch": 0.36902441580215867, + "grad_norm": 0.5865437984466553, + "learning_rate": 1.4572535550052702e-05, + "loss": 0.9249, + "step": 2201 + }, + { + "epoch": 0.36919207796290476, + "grad_norm": 0.3822292387485504, + "learning_rate": 1.4567705260449905e-05, + "loss": 0.8256, + "step": 2202 + }, + { + "epoch": 0.36935974012365086, + "grad_norm": 0.6831461787223816, + "learning_rate": 1.4562873623774267e-05, + "loss": 0.851, + "step": 2203 + }, + { + "epoch": 0.36952740228439696, + "grad_norm": 0.6423549652099609, + "learning_rate": 1.4558040641450695e-05, + "loss": 0.8969, + "step": 2204 + }, + { + "epoch": 0.36969506444514305, + "grad_norm": 0.583774745464325, + "learning_rate": 1.4553206314904493e-05, + "loss": 0.9309, + "step": 2205 + }, + { + "epoch": 0.36986272660588915, + "grad_norm": 0.6059913635253906, + "learning_rate": 1.4548370645561372e-05, + "loss": 0.9703, + "step": 2206 + }, + { + "epoch": 0.37003038876663524, + "grad_norm": 0.5601528882980347, + "learning_rate": 1.4543533634847422e-05, + "loss": 0.9313, + "step": 2207 + }, + { + "epoch": 0.37019805092738134, + "grad_norm": 0.39669668674468994, + "learning_rate": 1.4538695284189145e-05, + "loss": 0.9154, + "step": 2208 + }, + { + "epoch": 0.37036571308812744, + "grad_norm": 0.5013231635093689, + "learning_rate": 1.4533855595013424e-05, + "loss": 0.8856, + "step": 2209 + }, + { + "epoch": 0.37053337524887353, + "grad_norm": 0.6057686805725098, + "learning_rate": 1.4529014568747551e-05, + "loss": 0.9103, + "step": 2210 + }, + { + "epoch": 0.3707010374096196, + "grad_norm": 0.5593185424804688, + "learning_rate": 1.4524172206819195e-05, + "loss": 0.901, + "step": 2211 + }, + { + "epoch": 0.3708686995703657, + "grad_norm": 0.5884758234024048, + "learning_rate": 1.4519328510656435e-05, + "loss": 0.9134, + "step": 2212 + }, + { + "epoch": 0.3710363617311118, + "grad_norm": 0.5830038189888, + "learning_rate": 1.4514483481687733e-05, + "loss": 0.9205, + "step": 2213 + }, + { + "epoch": 0.3712040238918579, + "grad_norm": 0.5845561027526855, + "learning_rate": 1.450963712134195e-05, + "loss": 0.9248, + "step": 2214 + }, + { + "epoch": 0.371371686052604, + "grad_norm": 0.4929487705230713, + "learning_rate": 1.4504789431048343e-05, + "loss": 0.8335, + "step": 2215 + }, + { + "epoch": 0.3715393482133501, + "grad_norm": 0.4876137375831604, + "learning_rate": 1.4499940412236549e-05, + "loss": 0.8844, + "step": 2216 + }, + { + "epoch": 0.3717070103740962, + "grad_norm": 0.4443703293800354, + "learning_rate": 1.4495090066336604e-05, + "loss": 0.9326, + "step": 2217 + }, + { + "epoch": 0.3718746725348423, + "grad_norm": 0.5985068082809448, + "learning_rate": 1.4490238394778936e-05, + "loss": 0.9112, + "step": 2218 + }, + { + "epoch": 0.3720423346955884, + "grad_norm": 0.5642960667610168, + "learning_rate": 1.4485385398994364e-05, + "loss": 0.9219, + "step": 2219 + }, + { + "epoch": 0.3722099968563345, + "grad_norm": 0.506014883518219, + "learning_rate": 1.4480531080414096e-05, + "loss": 0.9068, + "step": 2220 + }, + { + "epoch": 0.3723776590170806, + "grad_norm": 0.5465354919433594, + "learning_rate": 1.4475675440469729e-05, + "loss": 0.8949, + "step": 2221 + }, + { + "epoch": 0.3725453211778267, + "grad_norm": 0.4898136854171753, + "learning_rate": 1.4470818480593258e-05, + "loss": 0.9157, + "step": 2222 + }, + { + "epoch": 0.3727129833385728, + "grad_norm": 0.4084881544113159, + "learning_rate": 1.4465960202217053e-05, + "loss": 0.8657, + "step": 2223 + }, + { + "epoch": 0.3728806454993189, + "grad_norm": 0.5694803595542908, + "learning_rate": 1.4461100606773884e-05, + "loss": 0.8568, + "step": 2224 + }, + { + "epoch": 0.37304830766006497, + "grad_norm": 0.41786134243011475, + "learning_rate": 1.4456239695696903e-05, + "loss": 0.96, + "step": 2225 + }, + { + "epoch": 0.37321596982081107, + "grad_norm": 0.49354809522628784, + "learning_rate": 1.4451377470419658e-05, + "loss": 0.9287, + "step": 2226 + }, + { + "epoch": 0.37338363198155716, + "grad_norm": 0.41685160994529724, + "learning_rate": 1.4446513932376076e-05, + "loss": 0.9182, + "step": 2227 + }, + { + "epoch": 0.37355129414230326, + "grad_norm": 0.3885839283466339, + "learning_rate": 1.444164908300047e-05, + "loss": 0.8644, + "step": 2228 + }, + { + "epoch": 0.37371895630304935, + "grad_norm": 0.49852511286735535, + "learning_rate": 1.4436782923727557e-05, + "loss": 0.9, + "step": 2229 + }, + { + "epoch": 0.37388661846379545, + "grad_norm": 0.5106147527694702, + "learning_rate": 1.4431915455992416e-05, + "loss": 0.9022, + "step": 2230 + }, + { + "epoch": 0.37405428062454155, + "grad_norm": 0.5619063973426819, + "learning_rate": 1.4427046681230525e-05, + "loss": 0.9078, + "step": 2231 + }, + { + "epoch": 0.37422194278528764, + "grad_norm": 0.4425818622112274, + "learning_rate": 1.4422176600877751e-05, + "loss": 0.7656, + "step": 2232 + }, + { + "epoch": 0.37438960494603374, + "grad_norm": 0.5203561186790466, + "learning_rate": 1.441730521637034e-05, + "loss": 0.984, + "step": 2233 + }, + { + "epoch": 0.37455726710677983, + "grad_norm": 0.41096168756484985, + "learning_rate": 1.441243252914492e-05, + "loss": 0.9238, + "step": 2234 + }, + { + "epoch": 0.37472492926752593, + "grad_norm": 0.47181493043899536, + "learning_rate": 1.4407558540638509e-05, + "loss": 0.8651, + "step": 2235 + }, + { + "epoch": 0.374892591428272, + "grad_norm": 0.6022822856903076, + "learning_rate": 1.4402683252288508e-05, + "loss": 0.9019, + "step": 2236 + }, + { + "epoch": 0.3750602535890181, + "grad_norm": 0.5729867815971375, + "learning_rate": 1.4397806665532693e-05, + "loss": 0.8741, + "step": 2237 + }, + { + "epoch": 0.3752279157497642, + "grad_norm": 0.5488620400428772, + "learning_rate": 1.4392928781809243e-05, + "loss": 0.9981, + "step": 2238 + }, + { + "epoch": 0.3753955779105103, + "grad_norm": 0.5091243982315063, + "learning_rate": 1.4388049602556692e-05, + "loss": 0.8706, + "step": 2239 + }, + { + "epoch": 0.3755632400712564, + "grad_norm": 0.5202438235282898, + "learning_rate": 1.4383169129213982e-05, + "loss": 0.8039, + "step": 2240 + }, + { + "epoch": 0.3757309022320025, + "grad_norm": 0.474795401096344, + "learning_rate": 1.4378287363220415e-05, + "loss": 0.902, + "step": 2241 + }, + { + "epoch": 0.3758985643927486, + "grad_norm": 0.43491849303245544, + "learning_rate": 1.4373404306015695e-05, + "loss": 0.9261, + "step": 2242 + }, + { + "epoch": 0.3760662265534947, + "grad_norm": 0.6873254776000977, + "learning_rate": 1.4368519959039882e-05, + "loss": 0.8585, + "step": 2243 + }, + { + "epoch": 0.3762338887142408, + "grad_norm": 0.4213145971298218, + "learning_rate": 1.4363634323733442e-05, + "loss": 0.8277, + "step": 2244 + }, + { + "epoch": 0.3764015508749869, + "grad_norm": 0.7943501472473145, + "learning_rate": 1.435874740153721e-05, + "loss": 0.9383, + "step": 2245 + }, + { + "epoch": 0.376569213035733, + "grad_norm": 1.0533539056777954, + "learning_rate": 1.4353859193892392e-05, + "loss": 0.8199, + "step": 2246 + }, + { + "epoch": 0.3767368751964791, + "grad_norm": 0.4386700391769409, + "learning_rate": 1.4348969702240583e-05, + "loss": 0.8748, + "step": 2247 + }, + { + "epoch": 0.3769045373572252, + "grad_norm": 0.5937883257865906, + "learning_rate": 1.434407892802376e-05, + "loss": 0.9957, + "step": 2248 + }, + { + "epoch": 0.3770721995179713, + "grad_norm": 0.6216612458229065, + "learning_rate": 1.4339186872684268e-05, + "loss": 0.8387, + "step": 2249 + }, + { + "epoch": 0.37723986167871737, + "grad_norm": 0.6769209504127502, + "learning_rate": 1.4334293537664836e-05, + "loss": 0.821, + "step": 2250 + }, + { + "epoch": 0.37740752383946347, + "grad_norm": 0.645228922367096, + "learning_rate": 1.4329398924408567e-05, + "loss": 1.0086, + "step": 2251 + }, + { + "epoch": 0.37757518600020956, + "grad_norm": 0.40301021933555603, + "learning_rate": 1.4324503034358946e-05, + "loss": 0.8927, + "step": 2252 + }, + { + "epoch": 0.37774284816095566, + "grad_norm": 0.5435569882392883, + "learning_rate": 1.4319605868959831e-05, + "loss": 0.8832, + "step": 2253 + }, + { + "epoch": 0.37791051032170175, + "grad_norm": 0.7265287637710571, + "learning_rate": 1.4314707429655456e-05, + "loss": 0.891, + "step": 2254 + }, + { + "epoch": 0.37807817248244785, + "grad_norm": 0.4399094879627228, + "learning_rate": 1.4309807717890432e-05, + "loss": 0.9177, + "step": 2255 + }, + { + "epoch": 0.37824583464319395, + "grad_norm": 0.3403279185295105, + "learning_rate": 1.4304906735109746e-05, + "loss": 0.7757, + "step": 2256 + }, + { + "epoch": 0.37841349680394004, + "grad_norm": 0.45156174898147583, + "learning_rate": 1.4300004482758755e-05, + "loss": 0.8753, + "step": 2257 + }, + { + "epoch": 0.37858115896468614, + "grad_norm": 0.780342698097229, + "learning_rate": 1.4295100962283197e-05, + "loss": 0.9102, + "step": 2258 + }, + { + "epoch": 0.37874882112543223, + "grad_norm": 0.684647262096405, + "learning_rate": 1.4290196175129181e-05, + "loss": 0.8894, + "step": 2259 + }, + { + "epoch": 0.37891648328617833, + "grad_norm": 0.5325799584388733, + "learning_rate": 1.4285290122743186e-05, + "loss": 0.9213, + "step": 2260 + }, + { + "epoch": 0.3790841454469244, + "grad_norm": 0.5337144136428833, + "learning_rate": 1.4280382806572071e-05, + "loss": 0.909, + "step": 2261 + }, + { + "epoch": 0.3792518076076705, + "grad_norm": 0.4459291398525238, + "learning_rate": 1.427547422806306e-05, + "loss": 0.9284, + "step": 2262 + }, + { + "epoch": 0.3794194697684166, + "grad_norm": 0.6480666995048523, + "learning_rate": 1.4270564388663761e-05, + "loss": 0.8756, + "step": 2263 + }, + { + "epoch": 0.3795871319291627, + "grad_norm": 0.5705782771110535, + "learning_rate": 1.4265653289822141e-05, + "loss": 0.8864, + "step": 2264 + }, + { + "epoch": 0.3797547940899088, + "grad_norm": 0.45472240447998047, + "learning_rate": 1.4260740932986546e-05, + "loss": 0.8901, + "step": 2265 + }, + { + "epoch": 0.3799224562506549, + "grad_norm": 0.5093509554862976, + "learning_rate": 1.4255827319605686e-05, + "loss": 0.892, + "step": 2266 + }, + { + "epoch": 0.380090118411401, + "grad_norm": 0.522854208946228, + "learning_rate": 1.4250912451128654e-05, + "loss": 1.004, + "step": 2267 + }, + { + "epoch": 0.3802577805721471, + "grad_norm": 0.3982647955417633, + "learning_rate": 1.4245996329004901e-05, + "loss": 0.862, + "step": 2268 + }, + { + "epoch": 0.3804254427328932, + "grad_norm": 0.48918038606643677, + "learning_rate": 1.4241078954684253e-05, + "loss": 0.8579, + "step": 2269 + }, + { + "epoch": 0.3805931048936393, + "grad_norm": 0.3785038888454437, + "learning_rate": 1.4236160329616903e-05, + "loss": 0.8844, + "step": 2270 + }, + { + "epoch": 0.38076076705438544, + "grad_norm": 0.5706347227096558, + "learning_rate": 1.4231240455253419e-05, + "loss": 0.9279, + "step": 2271 + }, + { + "epoch": 0.38092842921513154, + "grad_norm": 0.6968598365783691, + "learning_rate": 1.4226319333044729e-05, + "loss": 0.8614, + "step": 2272 + }, + { + "epoch": 0.38109609137587763, + "grad_norm": 0.5402582287788391, + "learning_rate": 1.4221396964442135e-05, + "loss": 0.9304, + "step": 2273 + }, + { + "epoch": 0.38126375353662373, + "grad_norm": 0.4895389974117279, + "learning_rate": 1.4216473350897301e-05, + "loss": 0.9808, + "step": 2274 + }, + { + "epoch": 0.3814314156973698, + "grad_norm": 0.35909193754196167, + "learning_rate": 1.4211548493862266e-05, + "loss": 0.8544, + "step": 2275 + }, + { + "epoch": 0.3815990778581159, + "grad_norm": 0.5294655561447144, + "learning_rate": 1.4206622394789432e-05, + "loss": 0.8938, + "step": 2276 + }, + { + "epoch": 0.381766740018862, + "grad_norm": 0.4982142746448517, + "learning_rate": 1.4201695055131556e-05, + "loss": 0.8515, + "step": 2277 + }, + { + "epoch": 0.3819344021796081, + "grad_norm": 0.47081637382507324, + "learning_rate": 1.4196766476341783e-05, + "loss": 0.8818, + "step": 2278 + }, + { + "epoch": 0.3821020643403542, + "grad_norm": 0.4863807260990143, + "learning_rate": 1.4191836659873613e-05, + "loss": 0.8882, + "step": 2279 + }, + { + "epoch": 0.3822697265011003, + "grad_norm": 0.5890251994132996, + "learning_rate": 1.4186905607180901e-05, + "loss": 0.885, + "step": 2280 + }, + { + "epoch": 0.3824373886618464, + "grad_norm": 0.6916296482086182, + "learning_rate": 1.4181973319717882e-05, + "loss": 0.8967, + "step": 2281 + }, + { + "epoch": 0.3826050508225925, + "grad_norm": 0.6745121479034424, + "learning_rate": 1.4177039798939143e-05, + "loss": 0.9705, + "step": 2282 + }, + { + "epoch": 0.3827727129833386, + "grad_norm": 0.6650843620300293, + "learning_rate": 1.4172105046299647e-05, + "loss": 0.8422, + "step": 2283 + }, + { + "epoch": 0.3829403751440847, + "grad_norm": 0.5521268248558044, + "learning_rate": 1.4167169063254708e-05, + "loss": 0.8828, + "step": 2284 + }, + { + "epoch": 0.3831080373048308, + "grad_norm": 0.6419956088066101, + "learning_rate": 1.416223185126001e-05, + "loss": 0.8358, + "step": 2285 + }, + { + "epoch": 0.3832756994655769, + "grad_norm": 0.535331130027771, + "learning_rate": 1.4157293411771603e-05, + "loss": 0.9164, + "step": 2286 + }, + { + "epoch": 0.383443361626323, + "grad_norm": 0.5716285109519958, + "learning_rate": 1.4152353746245888e-05, + "loss": 0.9543, + "step": 2287 + }, + { + "epoch": 0.38361102378706907, + "grad_norm": 0.3962705731391907, + "learning_rate": 1.4147412856139638e-05, + "loss": 0.9082, + "step": 2288 + }, + { + "epoch": 0.38377868594781517, + "grad_norm": 0.4236718416213989, + "learning_rate": 1.4142470742909976e-05, + "loss": 0.8863, + "step": 2289 + }, + { + "epoch": 0.38394634810856126, + "grad_norm": 0.5821020603179932, + "learning_rate": 1.4137527408014397e-05, + "loss": 0.9455, + "step": 2290 + }, + { + "epoch": 0.38411401026930736, + "grad_norm": 0.4136497378349304, + "learning_rate": 1.4132582852910753e-05, + "loss": 0.8888, + "step": 2291 + }, + { + "epoch": 0.38428167243005346, + "grad_norm": 0.47119471430778503, + "learning_rate": 1.412763707905725e-05, + "loss": 0.8335, + "step": 2292 + }, + { + "epoch": 0.38444933459079955, + "grad_norm": 0.421060174703598, + "learning_rate": 1.4122690087912462e-05, + "loss": 0.9604, + "step": 2293 + }, + { + "epoch": 0.38461699675154565, + "grad_norm": 0.4163464903831482, + "learning_rate": 1.4117741880935318e-05, + "loss": 0.908, + "step": 2294 + }, + { + "epoch": 0.38478465891229174, + "grad_norm": 0.4435379207134247, + "learning_rate": 1.4112792459585105e-05, + "loss": 0.9262, + "step": 2295 + }, + { + "epoch": 0.38495232107303784, + "grad_norm": 0.8743917346000671, + "learning_rate": 1.4107841825321465e-05, + "loss": 0.8946, + "step": 2296 + }, + { + "epoch": 0.38511998323378394, + "grad_norm": 0.9882781505584717, + "learning_rate": 1.4102889979604403e-05, + "loss": 0.9052, + "step": 2297 + }, + { + "epoch": 0.38528764539453003, + "grad_norm": 0.4133789539337158, + "learning_rate": 1.4097936923894282e-05, + "loss": 0.8854, + "step": 2298 + }, + { + "epoch": 0.3854553075552761, + "grad_norm": 0.4353831112384796, + "learning_rate": 1.4092982659651822e-05, + "loss": 0.8621, + "step": 2299 + }, + { + "epoch": 0.3856229697160222, + "grad_norm": 0.5475447773933411, + "learning_rate": 1.4088027188338088e-05, + "loss": 0.8884, + "step": 2300 + }, + { + "epoch": 0.3857906318767683, + "grad_norm": 0.49370449781417847, + "learning_rate": 1.4083070511414514e-05, + "loss": 0.9254, + "step": 2301 + }, + { + "epoch": 0.3859582940375144, + "grad_norm": 0.5450944900512695, + "learning_rate": 1.4078112630342891e-05, + "loss": 0.8184, + "step": 2302 + }, + { + "epoch": 0.3861259561982605, + "grad_norm": 0.5551537275314331, + "learning_rate": 1.4073153546585352e-05, + "loss": 0.8938, + "step": 2303 + }, + { + "epoch": 0.3862936183590066, + "grad_norm": 0.5214157700538635, + "learning_rate": 1.4068193261604397e-05, + "loss": 0.8185, + "step": 2304 + }, + { + "epoch": 0.3864612805197527, + "grad_norm": 0.5265510678291321, + "learning_rate": 1.4063231776862872e-05, + "loss": 0.9045, + "step": 2305 + }, + { + "epoch": 0.3866289426804988, + "grad_norm": 0.4628002345561981, + "learning_rate": 1.4058269093823982e-05, + "loss": 0.8955, + "step": 2306 + }, + { + "epoch": 0.3867966048412449, + "grad_norm": 0.46717819571495056, + "learning_rate": 1.4053305213951283e-05, + "loss": 0.807, + "step": 2307 + }, + { + "epoch": 0.386964267001991, + "grad_norm": 0.5428999066352844, + "learning_rate": 1.4048340138708684e-05, + "loss": 0.8461, + "step": 2308 + }, + { + "epoch": 0.3871319291627371, + "grad_norm": 0.7314903140068054, + "learning_rate": 1.404337386956045e-05, + "loss": 0.7835, + "step": 2309 + }, + { + "epoch": 0.3872995913234832, + "grad_norm": 0.8051000833511353, + "learning_rate": 1.4038406407971192e-05, + "loss": 0.8377, + "step": 2310 + }, + { + "epoch": 0.3874672534842293, + "grad_norm": 0.7588483691215515, + "learning_rate": 1.4033437755405878e-05, + "loss": 0.9077, + "step": 2311 + }, + { + "epoch": 0.3876349156449754, + "grad_norm": 0.6154654622077942, + "learning_rate": 1.4028467913329826e-05, + "loss": 0.8729, + "step": 2312 + }, + { + "epoch": 0.38780257780572147, + "grad_norm": 0.44144371151924133, + "learning_rate": 1.4023496883208702e-05, + "loss": 0.9389, + "step": 2313 + }, + { + "epoch": 0.38797023996646757, + "grad_norm": 0.4922724664211273, + "learning_rate": 1.4018524666508526e-05, + "loss": 0.8981, + "step": 2314 + }, + { + "epoch": 0.38813790212721366, + "grad_norm": 0.5737454891204834, + "learning_rate": 1.4013551264695663e-05, + "loss": 0.9817, + "step": 2315 + }, + { + "epoch": 0.38830556428795976, + "grad_norm": 0.6628448367118835, + "learning_rate": 1.4008576679236834e-05, + "loss": 0.9185, + "step": 2316 + }, + { + "epoch": 0.38847322644870586, + "grad_norm": 1.202582597732544, + "learning_rate": 1.4003600911599105e-05, + "loss": 0.9302, + "step": 2317 + }, + { + "epoch": 0.38864088860945195, + "grad_norm": 0.4161703586578369, + "learning_rate": 1.3998623963249894e-05, + "loss": 0.8186, + "step": 2318 + }, + { + "epoch": 0.38880855077019805, + "grad_norm": 0.658116340637207, + "learning_rate": 1.3993645835656955e-05, + "loss": 0.7665, + "step": 2319 + }, + { + "epoch": 0.38897621293094414, + "grad_norm": 0.4697672426700592, + "learning_rate": 1.3988666530288413e-05, + "loss": 0.8408, + "step": 2320 + }, + { + "epoch": 0.38914387509169024, + "grad_norm": 0.552117645740509, + "learning_rate": 1.3983686048612716e-05, + "loss": 0.7742, + "step": 2321 + }, + { + "epoch": 0.38931153725243633, + "grad_norm": 0.8014466166496277, + "learning_rate": 1.3978704392098678e-05, + "loss": 0.8477, + "step": 2322 + }, + { + "epoch": 0.38947919941318243, + "grad_norm": 0.6279250383377075, + "learning_rate": 1.3973721562215443e-05, + "loss": 0.8233, + "step": 2323 + }, + { + "epoch": 0.3896468615739285, + "grad_norm": 0.5173192024230957, + "learning_rate": 1.396873756043252e-05, + "loss": 0.912, + "step": 2324 + }, + { + "epoch": 0.3898145237346746, + "grad_norm": 0.5528737306594849, + "learning_rate": 1.396375238821974e-05, + "loss": 0.926, + "step": 2325 + }, + { + "epoch": 0.3899821858954207, + "grad_norm": 0.6598599553108215, + "learning_rate": 1.3958766047047298e-05, + "loss": 0.8262, + "step": 2326 + }, + { + "epoch": 0.3901498480561668, + "grad_norm": 0.5384594798088074, + "learning_rate": 1.3953778538385732e-05, + "loss": 0.8689, + "step": 2327 + }, + { + "epoch": 0.3903175102169129, + "grad_norm": 0.47622594237327576, + "learning_rate": 1.3948789863705914e-05, + "loss": 0.8962, + "step": 2328 + }, + { + "epoch": 0.390485172377659, + "grad_norm": 0.4942217171192169, + "learning_rate": 1.3943800024479068e-05, + "loss": 0.903, + "step": 2329 + }, + { + "epoch": 0.3906528345384051, + "grad_norm": 0.5775742530822754, + "learning_rate": 1.393880902217676e-05, + "loss": 0.835, + "step": 2330 + }, + { + "epoch": 0.3908204966991512, + "grad_norm": 0.42464926838874817, + "learning_rate": 1.3933816858270898e-05, + "loss": 0.945, + "step": 2331 + }, + { + "epoch": 0.3909881588598973, + "grad_norm": 0.3777066767215729, + "learning_rate": 1.3928823534233733e-05, + "loss": 0.8608, + "step": 2332 + }, + { + "epoch": 0.3911558210206434, + "grad_norm": 0.374035507440567, + "learning_rate": 1.3923829051537858e-05, + "loss": 0.9155, + "step": 2333 + }, + { + "epoch": 0.3913234831813895, + "grad_norm": 0.5102998614311218, + "learning_rate": 1.3918833411656205e-05, + "loss": 0.817, + "step": 2334 + }, + { + "epoch": 0.3914911453421356, + "grad_norm": 0.3295595645904541, + "learning_rate": 1.3913836616062056e-05, + "loss": 0.8171, + "step": 2335 + }, + { + "epoch": 0.3916588075028817, + "grad_norm": 0.6316651701927185, + "learning_rate": 1.3908838666229023e-05, + "loss": 0.9449, + "step": 2336 + }, + { + "epoch": 0.3918264696636278, + "grad_norm": 0.8228375911712646, + "learning_rate": 1.3903839563631065e-05, + "loss": 0.7795, + "step": 2337 + }, + { + "epoch": 0.39199413182437387, + "grad_norm": 0.8097789287567139, + "learning_rate": 1.389883930974248e-05, + "loss": 0.89, + "step": 2338 + }, + { + "epoch": 0.39216179398511997, + "grad_norm": 0.4250030517578125, + "learning_rate": 1.3893837906037903e-05, + "loss": 0.918, + "step": 2339 + }, + { + "epoch": 0.39232945614586606, + "grad_norm": 0.6134943962097168, + "learning_rate": 1.3888835353992312e-05, + "loss": 0.8479, + "step": 2340 + }, + { + "epoch": 0.39249711830661216, + "grad_norm": 0.886298418045044, + "learning_rate": 1.388383165508102e-05, + "loss": 0.9473, + "step": 2341 + }, + { + "epoch": 0.39266478046735825, + "grad_norm": 0.378825306892395, + "learning_rate": 1.3878826810779677e-05, + "loss": 0.9453, + "step": 2342 + }, + { + "epoch": 0.39283244262810435, + "grad_norm": 0.5482847094535828, + "learning_rate": 1.3873820822564282e-05, + "loss": 0.9344, + "step": 2343 + }, + { + "epoch": 0.39300010478885045, + "grad_norm": 0.5132451057434082, + "learning_rate": 1.3868813691911154e-05, + "loss": 0.8788, + "step": 2344 + }, + { + "epoch": 0.39316776694959654, + "grad_norm": 0.4483496844768524, + "learning_rate": 1.3863805420296965e-05, + "loss": 0.9336, + "step": 2345 + }, + { + "epoch": 0.39333542911034264, + "grad_norm": 0.4734271168708801, + "learning_rate": 1.3858796009198708e-05, + "loss": 0.9291, + "step": 2346 + }, + { + "epoch": 0.39350309127108873, + "grad_norm": 0.5556327700614929, + "learning_rate": 1.385378546009373e-05, + "loss": 0.8528, + "step": 2347 + }, + { + "epoch": 0.39367075343183483, + "grad_norm": 0.39046409726142883, + "learning_rate": 1.3848773774459696e-05, + "loss": 0.952, + "step": 2348 + }, + { + "epoch": 0.3938384155925809, + "grad_norm": 0.6670508980751038, + "learning_rate": 1.3843760953774616e-05, + "loss": 0.9058, + "step": 2349 + }, + { + "epoch": 0.394006077753327, + "grad_norm": 0.4425426423549652, + "learning_rate": 1.3838746999516836e-05, + "loss": 0.9347, + "step": 2350 + }, + { + "epoch": 0.3941737399140731, + "grad_norm": 0.5918128490447998, + "learning_rate": 1.3833731913165032e-05, + "loss": 0.9125, + "step": 2351 + }, + { + "epoch": 0.3943414020748192, + "grad_norm": 0.3698294460773468, + "learning_rate": 1.3828715696198214e-05, + "loss": 0.8371, + "step": 2352 + }, + { + "epoch": 0.3945090642355653, + "grad_norm": 0.4565315246582031, + "learning_rate": 1.3823698350095726e-05, + "loss": 0.9464, + "step": 2353 + }, + { + "epoch": 0.3946767263963114, + "grad_norm": 0.4425026476383209, + "learning_rate": 1.381867987633725e-05, + "loss": 0.9439, + "step": 2354 + }, + { + "epoch": 0.3948443885570575, + "grad_norm": 0.5444761514663696, + "learning_rate": 1.3813660276402787e-05, + "loss": 0.8747, + "step": 2355 + }, + { + "epoch": 0.39501205071780365, + "grad_norm": 0.542966902256012, + "learning_rate": 1.3808639551772688e-05, + "loss": 0.9186, + "step": 2356 + }, + { + "epoch": 0.39517971287854975, + "grad_norm": 0.42123109102249146, + "learning_rate": 1.3803617703927621e-05, + "loss": 0.8864, + "step": 2357 + }, + { + "epoch": 0.39534737503929585, + "grad_norm": 0.4050232768058777, + "learning_rate": 1.3798594734348596e-05, + "loss": 0.8296, + "step": 2358 + }, + { + "epoch": 0.39551503720004194, + "grad_norm": 0.536188542842865, + "learning_rate": 1.3793570644516947e-05, + "loss": 0.8622, + "step": 2359 + }, + { + "epoch": 0.39568269936078804, + "grad_norm": 0.384074866771698, + "learning_rate": 1.3788545435914337e-05, + "loss": 0.8495, + "step": 2360 + }, + { + "epoch": 0.39585036152153413, + "grad_norm": 0.4254181683063507, + "learning_rate": 1.3783519110022768e-05, + "loss": 0.8757, + "step": 2361 + }, + { + "epoch": 0.39601802368228023, + "grad_norm": 0.45154184103012085, + "learning_rate": 1.3778491668324561e-05, + "loss": 0.8858, + "step": 2362 + }, + { + "epoch": 0.3961856858430263, + "grad_norm": 0.45874765515327454, + "learning_rate": 1.3773463112302374e-05, + "loss": 0.8704, + "step": 2363 + }, + { + "epoch": 0.3963533480037724, + "grad_norm": 0.4734143614768982, + "learning_rate": 1.3768433443439189e-05, + "loss": 0.8793, + "step": 2364 + }, + { + "epoch": 0.3965210101645185, + "grad_norm": 0.5925267338752747, + "learning_rate": 1.3763402663218317e-05, + "loss": 0.8406, + "step": 2365 + }, + { + "epoch": 0.3966886723252646, + "grad_norm": 0.626492440700531, + "learning_rate": 1.3758370773123397e-05, + "loss": 0.9716, + "step": 2366 + }, + { + "epoch": 0.3968563344860107, + "grad_norm": 0.6277036666870117, + "learning_rate": 1.3753337774638397e-05, + "loss": 0.8995, + "step": 2367 + }, + { + "epoch": 0.3970239966467568, + "grad_norm": 0.5515948534011841, + "learning_rate": 1.3748303669247611e-05, + "loss": 0.9342, + "step": 2368 + }, + { + "epoch": 0.3971916588075029, + "grad_norm": 0.8049086928367615, + "learning_rate": 1.374326845843566e-05, + "loss": 0.94, + "step": 2369 + }, + { + "epoch": 0.397359320968249, + "grad_norm": 0.45917946100234985, + "learning_rate": 1.3738232143687489e-05, + "loss": 0.9634, + "step": 2370 + }, + { + "epoch": 0.3975269831289951, + "grad_norm": 0.5680646896362305, + "learning_rate": 1.3733194726488366e-05, + "loss": 0.8937, + "step": 2371 + }, + { + "epoch": 0.3976946452897412, + "grad_norm": 0.5159409046173096, + "learning_rate": 1.3728156208323892e-05, + "loss": 0.8145, + "step": 2372 + }, + { + "epoch": 0.3978623074504873, + "grad_norm": 0.45107534527778625, + "learning_rate": 1.3723116590679986e-05, + "loss": 0.9567, + "step": 2373 + }, + { + "epoch": 0.3980299696112334, + "grad_norm": 0.374700129032135, + "learning_rate": 1.3718075875042895e-05, + "loss": 0.9302, + "step": 2374 + }, + { + "epoch": 0.3981976317719795, + "grad_norm": 0.6136937737464905, + "learning_rate": 1.3713034062899191e-05, + "loss": 0.9222, + "step": 2375 + }, + { + "epoch": 0.3983652939327256, + "grad_norm": 0.4689367115497589, + "learning_rate": 1.3707991155735761e-05, + "loss": 0.8588, + "step": 2376 + }, + { + "epoch": 0.39853295609347167, + "grad_norm": 0.45876532793045044, + "learning_rate": 1.3702947155039826e-05, + "loss": 0.9334, + "step": 2377 + }, + { + "epoch": 0.39870061825421776, + "grad_norm": 0.3927895128726959, + "learning_rate": 1.369790206229892e-05, + "loss": 0.8993, + "step": 2378 + }, + { + "epoch": 0.39886828041496386, + "grad_norm": 0.49190419912338257, + "learning_rate": 1.3692855879000908e-05, + "loss": 0.8762, + "step": 2379 + }, + { + "epoch": 0.39903594257570996, + "grad_norm": 0.5364404320716858, + "learning_rate": 1.3687808606633965e-05, + "loss": 0.9148, + "step": 2380 + }, + { + "epoch": 0.39920360473645605, + "grad_norm": 0.41728052496910095, + "learning_rate": 1.3682760246686604e-05, + "loss": 0.8652, + "step": 2381 + }, + { + "epoch": 0.39937126689720215, + "grad_norm": 0.46292200684547424, + "learning_rate": 1.367771080064764e-05, + "loss": 0.8716, + "step": 2382 + }, + { + "epoch": 0.39953892905794824, + "grad_norm": 0.5097537040710449, + "learning_rate": 1.367266027000622e-05, + "loss": 0.8465, + "step": 2383 + }, + { + "epoch": 0.39970659121869434, + "grad_norm": 0.51020747423172, + "learning_rate": 1.3667608656251808e-05, + "loss": 0.8676, + "step": 2384 + }, + { + "epoch": 0.39987425337944044, + "grad_norm": 0.5306868553161621, + "learning_rate": 1.366255596087419e-05, + "loss": 0.81, + "step": 2385 + }, + { + "epoch": 0.40004191554018653, + "grad_norm": 0.4405503273010254, + "learning_rate": 1.3657502185363467e-05, + "loss": 0.8347, + "step": 2386 + }, + { + "epoch": 0.40020957770093263, + "grad_norm": 0.46693986654281616, + "learning_rate": 1.3652447331210057e-05, + "loss": 0.9545, + "step": 2387 + }, + { + "epoch": 0.4003772398616787, + "grad_norm": 0.7226107716560364, + "learning_rate": 1.3647391399904701e-05, + "loss": 0.9323, + "step": 2388 + }, + { + "epoch": 0.4005449020224248, + "grad_norm": 0.6572045683860779, + "learning_rate": 1.3642334392938458e-05, + "loss": 0.9501, + "step": 2389 + }, + { + "epoch": 0.4007125641831709, + "grad_norm": 0.34955379366874695, + "learning_rate": 1.3637276311802701e-05, + "loss": 0.8603, + "step": 2390 + }, + { + "epoch": 0.400880226343917, + "grad_norm": 0.48697224259376526, + "learning_rate": 1.3632217157989118e-05, + "loss": 0.8229, + "step": 2391 + }, + { + "epoch": 0.4010478885046631, + "grad_norm": 0.4833279252052307, + "learning_rate": 1.3627156932989719e-05, + "loss": 0.9941, + "step": 2392 + }, + { + "epoch": 0.4012155506654092, + "grad_norm": 0.6540115475654602, + "learning_rate": 1.3622095638296827e-05, + "loss": 0.8278, + "step": 2393 + }, + { + "epoch": 0.4013832128261553, + "grad_norm": 0.5749005079269409, + "learning_rate": 1.3617033275403078e-05, + "loss": 0.9305, + "step": 2394 + }, + { + "epoch": 0.4015508749869014, + "grad_norm": 0.6052101850509644, + "learning_rate": 1.3611969845801427e-05, + "loss": 0.8888, + "step": 2395 + }, + { + "epoch": 0.4017185371476475, + "grad_norm": 0.45506125688552856, + "learning_rate": 1.3606905350985143e-05, + "loss": 0.8748, + "step": 2396 + }, + { + "epoch": 0.4018861993083936, + "grad_norm": 0.45200562477111816, + "learning_rate": 1.3601839792447809e-05, + "loss": 0.8185, + "step": 2397 + }, + { + "epoch": 0.4020538614691397, + "grad_norm": 0.38190123438835144, + "learning_rate": 1.3596773171683317e-05, + "loss": 0.827, + "step": 2398 + }, + { + "epoch": 0.4022215236298858, + "grad_norm": 0.4883127510547638, + "learning_rate": 1.3591705490185875e-05, + "loss": 0.895, + "step": 2399 + }, + { + "epoch": 0.4023891857906319, + "grad_norm": 0.4430232644081116, + "learning_rate": 1.3586636749450015e-05, + "loss": 0.8115, + "step": 2400 + }, + { + "epoch": 0.40255684795137797, + "grad_norm": 0.6474736928939819, + "learning_rate": 1.3581566950970562e-05, + "loss": 0.8738, + "step": 2401 + }, + { + "epoch": 0.40272451011212407, + "grad_norm": 0.9033061861991882, + "learning_rate": 1.3576496096242665e-05, + "loss": 0.814, + "step": 2402 + }, + { + "epoch": 0.40289217227287016, + "grad_norm": 0.4570147395133972, + "learning_rate": 1.3571424186761784e-05, + "loss": 0.7655, + "step": 2403 + }, + { + "epoch": 0.40305983443361626, + "grad_norm": 0.7191325426101685, + "learning_rate": 1.3566351224023684e-05, + "loss": 0.9122, + "step": 2404 + }, + { + "epoch": 0.40322749659436236, + "grad_norm": 0.3910098373889923, + "learning_rate": 1.3561277209524449e-05, + "loss": 0.8671, + "step": 2405 + }, + { + "epoch": 0.40339515875510845, + "grad_norm": 0.5069496035575867, + "learning_rate": 1.3556202144760461e-05, + "loss": 0.8933, + "step": 2406 + }, + { + "epoch": 0.40356282091585455, + "grad_norm": 0.400499552488327, + "learning_rate": 1.355112603122843e-05, + "loss": 0.9089, + "step": 2407 + }, + { + "epoch": 0.40373048307660064, + "grad_norm": 0.3381521701812744, + "learning_rate": 1.3546048870425356e-05, + "loss": 0.965, + "step": 2408 + }, + { + "epoch": 0.40389814523734674, + "grad_norm": 0.46304380893707275, + "learning_rate": 1.3540970663848565e-05, + "loss": 0.8994, + "step": 2409 + }, + { + "epoch": 0.40406580739809284, + "grad_norm": 0.6324905157089233, + "learning_rate": 1.3535891412995675e-05, + "loss": 0.8701, + "step": 2410 + }, + { + "epoch": 0.40423346955883893, + "grad_norm": 0.45117953419685364, + "learning_rate": 1.3530811119364623e-05, + "loss": 0.8599, + "step": 2411 + }, + { + "epoch": 0.404401131719585, + "grad_norm": 0.5117924809455872, + "learning_rate": 1.3525729784453648e-05, + "loss": 0.8782, + "step": 2412 + }, + { + "epoch": 0.4045687938803311, + "grad_norm": 0.452868789434433, + "learning_rate": 1.3520647409761305e-05, + "loss": 0.8646, + "step": 2413 + }, + { + "epoch": 0.4047364560410772, + "grad_norm": 0.3938448131084442, + "learning_rate": 1.3515563996786443e-05, + "loss": 0.8669, + "step": 2414 + }, + { + "epoch": 0.4049041182018233, + "grad_norm": 0.5183711647987366, + "learning_rate": 1.3510479547028226e-05, + "loss": 0.8956, + "step": 2415 + }, + { + "epoch": 0.4050717803625694, + "grad_norm": 0.45423707365989685, + "learning_rate": 1.3505394061986122e-05, + "loss": 0.8156, + "step": 2416 + }, + { + "epoch": 0.4052394425233155, + "grad_norm": 0.4229229986667633, + "learning_rate": 1.3500307543159902e-05, + "loss": 0.8449, + "step": 2417 + }, + { + "epoch": 0.4054071046840616, + "grad_norm": 0.4813057780265808, + "learning_rate": 1.3495219992049644e-05, + "loss": 0.9262, + "step": 2418 + }, + { + "epoch": 0.4055747668448077, + "grad_norm": 0.737172544002533, + "learning_rate": 1.349013141015573e-05, + "loss": 0.9281, + "step": 2419 + }, + { + "epoch": 0.4057424290055538, + "grad_norm": 0.5750868320465088, + "learning_rate": 1.3485041798978846e-05, + "loss": 0.9335, + "step": 2420 + }, + { + "epoch": 0.4059100911662999, + "grad_norm": 0.42083829641342163, + "learning_rate": 1.347995116001998e-05, + "loss": 0.7781, + "step": 2421 + }, + { + "epoch": 0.406077753327046, + "grad_norm": 0.5151885151863098, + "learning_rate": 1.3474859494780423e-05, + "loss": 0.909, + "step": 2422 + }, + { + "epoch": 0.4062454154877921, + "grad_norm": 0.4647657871246338, + "learning_rate": 1.3469766804761774e-05, + "loss": 0.8934, + "step": 2423 + }, + { + "epoch": 0.4064130776485382, + "grad_norm": 0.5370908975601196, + "learning_rate": 1.3464673091465929e-05, + "loss": 0.9387, + "step": 2424 + }, + { + "epoch": 0.4065807398092843, + "grad_norm": 0.4686649441719055, + "learning_rate": 1.3459578356395088e-05, + "loss": 0.8882, + "step": 2425 + }, + { + "epoch": 0.40674840197003037, + "grad_norm": 0.6504122018814087, + "learning_rate": 1.345448260105175e-05, + "loss": 0.8825, + "step": 2426 + }, + { + "epoch": 0.40691606413077647, + "grad_norm": 0.5108534097671509, + "learning_rate": 1.344938582693872e-05, + "loss": 0.8653, + "step": 2427 + }, + { + "epoch": 0.40708372629152256, + "grad_norm": 0.4992177188396454, + "learning_rate": 1.3444288035559092e-05, + "loss": 0.8074, + "step": 2428 + }, + { + "epoch": 0.40725138845226866, + "grad_norm": 0.5172570943832397, + "learning_rate": 1.3439189228416271e-05, + "loss": 0.8284, + "step": 2429 + }, + { + "epoch": 0.40741905061301475, + "grad_norm": 0.4258668124675751, + "learning_rate": 1.3434089407013963e-05, + "loss": 0.8868, + "step": 2430 + }, + { + "epoch": 0.40758671277376085, + "grad_norm": 0.4065842628479004, + "learning_rate": 1.3428988572856163e-05, + "loss": 0.908, + "step": 2431 + }, + { + "epoch": 0.40775437493450695, + "grad_norm": 0.6188164353370667, + "learning_rate": 1.3423886727447176e-05, + "loss": 0.9106, + "step": 2432 + }, + { + "epoch": 0.40792203709525304, + "grad_norm": 0.549799382686615, + "learning_rate": 1.3418783872291595e-05, + "loss": 0.9599, + "step": 2433 + }, + { + "epoch": 0.40808969925599914, + "grad_norm": 0.7364528775215149, + "learning_rate": 1.3413680008894312e-05, + "loss": 0.9087, + "step": 2434 + }, + { + "epoch": 0.40825736141674523, + "grad_norm": 0.6526517271995544, + "learning_rate": 1.3408575138760526e-05, + "loss": 0.975, + "step": 2435 + }, + { + "epoch": 0.40842502357749133, + "grad_norm": 0.44261178374290466, + "learning_rate": 1.3403469263395725e-05, + "loss": 0.9273, + "step": 2436 + }, + { + "epoch": 0.4085926857382374, + "grad_norm": 0.45836639404296875, + "learning_rate": 1.339836238430569e-05, + "loss": 0.7632, + "step": 2437 + }, + { + "epoch": 0.4087603478989835, + "grad_norm": 0.39426425099372864, + "learning_rate": 1.3393254502996511e-05, + "loss": 1.0172, + "step": 2438 + }, + { + "epoch": 0.4089280100597296, + "grad_norm": 0.47519373893737793, + "learning_rate": 1.3388145620974558e-05, + "loss": 0.9009, + "step": 2439 + }, + { + "epoch": 0.4090956722204757, + "grad_norm": 0.4381753206253052, + "learning_rate": 1.3383035739746507e-05, + "loss": 0.8718, + "step": 2440 + }, + { + "epoch": 0.4092633343812218, + "grad_norm": 0.5138468742370605, + "learning_rate": 1.3377924860819325e-05, + "loss": 0.8698, + "step": 2441 + }, + { + "epoch": 0.40943099654196796, + "grad_norm": 0.40322327613830566, + "learning_rate": 1.3372812985700274e-05, + "loss": 0.9485, + "step": 2442 + }, + { + "epoch": 0.40959865870271406, + "grad_norm": 0.5753828883171082, + "learning_rate": 1.3367700115896908e-05, + "loss": 0.8841, + "step": 2443 + }, + { + "epoch": 0.40976632086346015, + "grad_norm": 0.3679884970188141, + "learning_rate": 1.3362586252917079e-05, + "loss": 0.8343, + "step": 2444 + }, + { + "epoch": 0.40993398302420625, + "grad_norm": 0.4531792104244232, + "learning_rate": 1.335747139826892e-05, + "loss": 0.9368, + "step": 2445 + }, + { + "epoch": 0.41010164518495235, + "grad_norm": 0.43219149112701416, + "learning_rate": 1.3352355553460873e-05, + "loss": 0.8713, + "step": 2446 + }, + { + "epoch": 0.41026930734569844, + "grad_norm": 0.5404754281044006, + "learning_rate": 1.3347238720001663e-05, + "loss": 0.8641, + "step": 2447 + }, + { + "epoch": 0.41043696950644454, + "grad_norm": 0.47969868779182434, + "learning_rate": 1.3342120899400305e-05, + "loss": 0.8207, + "step": 2448 + }, + { + "epoch": 0.41060463166719063, + "grad_norm": 0.5258057713508606, + "learning_rate": 1.333700209316611e-05, + "loss": 0.9198, + "step": 2449 + }, + { + "epoch": 0.41077229382793673, + "grad_norm": 0.41233959794044495, + "learning_rate": 1.3331882302808679e-05, + "loss": 0.9456, + "step": 2450 + }, + { + "epoch": 0.4109399559886828, + "grad_norm": 0.5792062282562256, + "learning_rate": 1.3326761529837895e-05, + "loss": 0.818, + "step": 2451 + }, + { + "epoch": 0.4111076181494289, + "grad_norm": 0.5087970495223999, + "learning_rate": 1.3321639775763942e-05, + "loss": 0.9785, + "step": 2452 + }, + { + "epoch": 0.411275280310175, + "grad_norm": 0.40101125836372375, + "learning_rate": 1.3316517042097285e-05, + "loss": 0.8717, + "step": 2453 + }, + { + "epoch": 0.4114429424709211, + "grad_norm": 0.4285100996494293, + "learning_rate": 1.3311393330348689e-05, + "loss": 0.9145, + "step": 2454 + }, + { + "epoch": 0.4116106046316672, + "grad_norm": 0.599571943283081, + "learning_rate": 1.3306268642029194e-05, + "loss": 0.9772, + "step": 2455 + }, + { + "epoch": 0.4117782667924133, + "grad_norm": 0.660012423992157, + "learning_rate": 1.3301142978650133e-05, + "loss": 0.8471, + "step": 2456 + }, + { + "epoch": 0.4119459289531594, + "grad_norm": 1.0488002300262451, + "learning_rate": 1.3296016341723135e-05, + "loss": 0.8742, + "step": 2457 + }, + { + "epoch": 0.4121135911139055, + "grad_norm": 0.5106312036514282, + "learning_rate": 1.32908887327601e-05, + "loss": 0.9073, + "step": 2458 + }, + { + "epoch": 0.4122812532746516, + "grad_norm": 0.4380190670490265, + "learning_rate": 1.328576015327323e-05, + "loss": 0.8505, + "step": 2459 + }, + { + "epoch": 0.4124489154353977, + "grad_norm": 0.4387100338935852, + "learning_rate": 1.3280630604775002e-05, + "loss": 0.9409, + "step": 2460 + }, + { + "epoch": 0.4126165775961438, + "grad_norm": 0.6855369210243225, + "learning_rate": 1.3275500088778186e-05, + "loss": 0.8504, + "step": 2461 + }, + { + "epoch": 0.4127842397568899, + "grad_norm": 0.38202857971191406, + "learning_rate": 1.3270368606795833e-05, + "loss": 0.9885, + "step": 2462 + }, + { + "epoch": 0.412951901917636, + "grad_norm": 0.6671926379203796, + "learning_rate": 1.3265236160341282e-05, + "loss": 0.7529, + "step": 2463 + }, + { + "epoch": 0.4131195640783821, + "grad_norm": 0.6485735774040222, + "learning_rate": 1.3260102750928154e-05, + "loss": 0.8242, + "step": 2464 + }, + { + "epoch": 0.41328722623912817, + "grad_norm": 0.3961820602416992, + "learning_rate": 1.3254968380070356e-05, + "loss": 0.8644, + "step": 2465 + }, + { + "epoch": 0.41345488839987427, + "grad_norm": 0.7097799777984619, + "learning_rate": 1.324983304928208e-05, + "loss": 0.8563, + "step": 2466 + }, + { + "epoch": 0.41362255056062036, + "grad_norm": 0.46577221155166626, + "learning_rate": 1.3244696760077793e-05, + "loss": 0.9005, + "step": 2467 + }, + { + "epoch": 0.41379021272136646, + "grad_norm": 0.36700090765953064, + "learning_rate": 1.3239559513972251e-05, + "loss": 0.9534, + "step": 2468 + }, + { + "epoch": 0.41395787488211255, + "grad_norm": 0.37399622797966003, + "learning_rate": 1.3234421312480494e-05, + "loss": 0.9114, + "step": 2469 + }, + { + "epoch": 0.41412553704285865, + "grad_norm": 0.656875491142273, + "learning_rate": 1.3229282157117847e-05, + "loss": 0.9358, + "step": 2470 + }, + { + "epoch": 0.41429319920360475, + "grad_norm": 0.3412625789642334, + "learning_rate": 1.3224142049399896e-05, + "loss": 0.7384, + "step": 2471 + }, + { + "epoch": 0.41446086136435084, + "grad_norm": 0.6255205869674683, + "learning_rate": 1.3219000990842531e-05, + "loss": 0.859, + "step": 2472 + }, + { + "epoch": 0.41462852352509694, + "grad_norm": 0.809914231300354, + "learning_rate": 1.321385898296192e-05, + "loss": 0.7862, + "step": 2473 + }, + { + "epoch": 0.41479618568584303, + "grad_norm": 0.5074305534362793, + "learning_rate": 1.3208716027274493e-05, + "loss": 0.894, + "step": 2474 + }, + { + "epoch": 0.41496384784658913, + "grad_norm": 0.5917060971260071, + "learning_rate": 1.320357212529698e-05, + "loss": 0.8439, + "step": 2475 + }, + { + "epoch": 0.4151315100073352, + "grad_norm": 0.5297524333000183, + "learning_rate": 1.3198427278546376e-05, + "loss": 0.9077, + "step": 2476 + }, + { + "epoch": 0.4152991721680813, + "grad_norm": 0.4922879934310913, + "learning_rate": 1.3193281488539963e-05, + "loss": 0.9503, + "step": 2477 + }, + { + "epoch": 0.4154668343288274, + "grad_norm": 0.4003523886203766, + "learning_rate": 1.3188134756795296e-05, + "loss": 0.8866, + "step": 2478 + }, + { + "epoch": 0.4156344964895735, + "grad_norm": 0.4242881238460541, + "learning_rate": 1.3182987084830213e-05, + "loss": 0.9004, + "step": 2479 + }, + { + "epoch": 0.4158021586503196, + "grad_norm": 0.46138808131217957, + "learning_rate": 1.317783847416282e-05, + "loss": 0.8495, + "step": 2480 + }, + { + "epoch": 0.4159698208110657, + "grad_norm": 0.3736347258090973, + "learning_rate": 1.3172688926311515e-05, + "loss": 0.798, + "step": 2481 + }, + { + "epoch": 0.4161374829718118, + "grad_norm": 0.587462306022644, + "learning_rate": 1.316753844279496e-05, + "loss": 0.8119, + "step": 2482 + }, + { + "epoch": 0.4163051451325579, + "grad_norm": 0.46949222683906555, + "learning_rate": 1.3162387025132088e-05, + "loss": 0.8524, + "step": 2483 + }, + { + "epoch": 0.416472807293304, + "grad_norm": 0.419544517993927, + "learning_rate": 1.315723467484213e-05, + "loss": 0.8296, + "step": 2484 + }, + { + "epoch": 0.4166404694540501, + "grad_norm": 0.5041922926902771, + "learning_rate": 1.3152081393444568e-05, + "loss": 0.9743, + "step": 2485 + }, + { + "epoch": 0.4168081316147962, + "grad_norm": 0.4275999069213867, + "learning_rate": 1.3146927182459172e-05, + "loss": 0.8648, + "step": 2486 + }, + { + "epoch": 0.4169757937755423, + "grad_norm": 0.3998974561691284, + "learning_rate": 1.3141772043405983e-05, + "loss": 0.8869, + "step": 2487 + }, + { + "epoch": 0.4171434559362884, + "grad_norm": 0.5498875379562378, + "learning_rate": 1.3136615977805312e-05, + "loss": 0.8177, + "step": 2488 + }, + { + "epoch": 0.4173111180970345, + "grad_norm": 0.5881977081298828, + "learning_rate": 1.3131458987177752e-05, + "loss": 0.9554, + "step": 2489 + }, + { + "epoch": 0.41747878025778057, + "grad_norm": 0.5910879373550415, + "learning_rate": 1.3126301073044158e-05, + "loss": 0.8834, + "step": 2490 + }, + { + "epoch": 0.41764644241852666, + "grad_norm": 0.456267386674881, + "learning_rate": 1.3121142236925666e-05, + "loss": 0.8794, + "step": 2491 + }, + { + "epoch": 0.41781410457927276, + "grad_norm": 0.3418677747249603, + "learning_rate": 1.3115982480343679e-05, + "loss": 0.7878, + "step": 2492 + }, + { + "epoch": 0.41798176674001886, + "grad_norm": 0.44005241990089417, + "learning_rate": 1.3110821804819877e-05, + "loss": 0.8275, + "step": 2493 + }, + { + "epoch": 0.41814942890076495, + "grad_norm": 0.4577225148677826, + "learning_rate": 1.31056602118762e-05, + "loss": 0.8243, + "step": 2494 + }, + { + "epoch": 0.41831709106151105, + "grad_norm": 0.5882584452629089, + "learning_rate": 1.3100497703034874e-05, + "loss": 0.9296, + "step": 2495 + }, + { + "epoch": 0.41848475322225714, + "grad_norm": 0.4153955578804016, + "learning_rate": 1.3095334279818377e-05, + "loss": 0.852, + "step": 2496 + }, + { + "epoch": 0.41865241538300324, + "grad_norm": 0.6201179027557373, + "learning_rate": 1.3090169943749475e-05, + "loss": 0.8484, + "step": 2497 + }, + { + "epoch": 0.41882007754374934, + "grad_norm": 0.44844475388526917, + "learning_rate": 1.3085004696351192e-05, + "loss": 0.9298, + "step": 2498 + }, + { + "epoch": 0.41898773970449543, + "grad_norm": 0.45476582646369934, + "learning_rate": 1.3079838539146824e-05, + "loss": 0.9391, + "step": 2499 + }, + { + "epoch": 0.41915540186524153, + "grad_norm": 0.5748103857040405, + "learning_rate": 1.3074671473659932e-05, + "loss": 0.8742, + "step": 2500 + }, + { + "epoch": 0.4193230640259876, + "grad_norm": 0.5483225584030151, + "learning_rate": 1.3069503501414349e-05, + "loss": 0.8643, + "step": 2501 + }, + { + "epoch": 0.4194907261867337, + "grad_norm": 0.5988959670066833, + "learning_rate": 1.3064334623934174e-05, + "loss": 0.9111, + "step": 2502 + }, + { + "epoch": 0.4196583883474798, + "grad_norm": 0.5877673625946045, + "learning_rate": 1.3059164842743774e-05, + "loss": 0.9511, + "step": 2503 + }, + { + "epoch": 0.4198260505082259, + "grad_norm": 0.671694278717041, + "learning_rate": 1.305399415936778e-05, + "loss": 0.8705, + "step": 2504 + }, + { + "epoch": 0.419993712668972, + "grad_norm": 0.35245388746261597, + "learning_rate": 1.3048822575331093e-05, + "loss": 0.8582, + "step": 2505 + }, + { + "epoch": 0.4201613748297181, + "grad_norm": 0.4740123152732849, + "learning_rate": 1.3043650092158875e-05, + "loss": 0.8263, + "step": 2506 + }, + { + "epoch": 0.4203290369904642, + "grad_norm": 0.3961458206176758, + "learning_rate": 1.3038476711376558e-05, + "loss": 0.8907, + "step": 2507 + }, + { + "epoch": 0.4204966991512103, + "grad_norm": 0.8107627630233765, + "learning_rate": 1.3033302434509831e-05, + "loss": 0.8103, + "step": 2508 + }, + { + "epoch": 0.4206643613119564, + "grad_norm": 0.48412755131721497, + "learning_rate": 1.3028127263084656e-05, + "loss": 0.8953, + "step": 2509 + }, + { + "epoch": 0.4208320234727025, + "grad_norm": 0.5948150157928467, + "learning_rate": 1.3022951198627254e-05, + "loss": 0.871, + "step": 2510 + }, + { + "epoch": 0.4209996856334486, + "grad_norm": 0.4686228334903717, + "learning_rate": 1.3017774242664112e-05, + "loss": 0.873, + "step": 2511 + }, + { + "epoch": 0.4211673477941947, + "grad_norm": 0.5049168467521667, + "learning_rate": 1.3012596396721978e-05, + "loss": 0.8775, + "step": 2512 + }, + { + "epoch": 0.4213350099549408, + "grad_norm": 0.7438037395477295, + "learning_rate": 1.3007417662327857e-05, + "loss": 0.9154, + "step": 2513 + }, + { + "epoch": 0.42150267211568687, + "grad_norm": 0.4495711922645569, + "learning_rate": 1.3002238041009032e-05, + "loss": 0.9141, + "step": 2514 + }, + { + "epoch": 0.42167033427643297, + "grad_norm": 0.48235267400741577, + "learning_rate": 1.2997057534293032e-05, + "loss": 0.8947, + "step": 2515 + }, + { + "epoch": 0.42183799643717906, + "grad_norm": 0.42142271995544434, + "learning_rate": 1.2991876143707653e-05, + "loss": 0.7854, + "step": 2516 + }, + { + "epoch": 0.42200565859792516, + "grad_norm": 0.6156967878341675, + "learning_rate": 1.2986693870780949e-05, + "loss": 0.8772, + "step": 2517 + }, + { + "epoch": 0.42217332075867126, + "grad_norm": 0.37790626287460327, + "learning_rate": 1.2981510717041244e-05, + "loss": 0.8518, + "step": 2518 + }, + { + "epoch": 0.42234098291941735, + "grad_norm": 0.6372936964035034, + "learning_rate": 1.2976326684017104e-05, + "loss": 0.8625, + "step": 2519 + }, + { + "epoch": 0.42250864508016345, + "grad_norm": 0.6741154193878174, + "learning_rate": 1.297114177323737e-05, + "loss": 0.8152, + "step": 2520 + }, + { + "epoch": 0.42267630724090954, + "grad_norm": 0.41690656542778015, + "learning_rate": 1.2965955986231138e-05, + "loss": 0.8574, + "step": 2521 + }, + { + "epoch": 0.42284396940165564, + "grad_norm": 0.5158107876777649, + "learning_rate": 1.2960769324527758e-05, + "loss": 0.8277, + "step": 2522 + }, + { + "epoch": 0.42301163156240174, + "grad_norm": 0.4730595648288727, + "learning_rate": 1.2955581789656844e-05, + "loss": 0.9815, + "step": 2523 + }, + { + "epoch": 0.42317929372314783, + "grad_norm": 0.5132426619529724, + "learning_rate": 1.2950393383148257e-05, + "loss": 0.8643, + "step": 2524 + }, + { + "epoch": 0.4233469558838939, + "grad_norm": 0.3574404716491699, + "learning_rate": 1.294520410653213e-05, + "loss": 1.0134, + "step": 2525 + }, + { + "epoch": 0.42351461804464, + "grad_norm": 0.45536112785339355, + "learning_rate": 1.2940013961338843e-05, + "loss": 0.8603, + "step": 2526 + }, + { + "epoch": 0.4236822802053861, + "grad_norm": 0.6392506957054138, + "learning_rate": 1.2934822949099032e-05, + "loss": 0.8885, + "step": 2527 + }, + { + "epoch": 0.42384994236613227, + "grad_norm": 0.44098612666130066, + "learning_rate": 1.2929631071343588e-05, + "loss": 0.8565, + "step": 2528 + }, + { + "epoch": 0.42401760452687837, + "grad_norm": 0.7202128171920776, + "learning_rate": 1.2924438329603668e-05, + "loss": 0.8701, + "step": 2529 + }, + { + "epoch": 0.42418526668762446, + "grad_norm": 0.41382378339767456, + "learning_rate": 1.2919244725410673e-05, + "loss": 0.897, + "step": 2530 + }, + { + "epoch": 0.42435292884837056, + "grad_norm": 0.7349418997764587, + "learning_rate": 1.2914050260296258e-05, + "loss": 0.8141, + "step": 2531 + }, + { + "epoch": 0.42452059100911665, + "grad_norm": 0.42445793747901917, + "learning_rate": 1.2908854935792339e-05, + "loss": 0.94, + "step": 2532 + }, + { + "epoch": 0.42468825316986275, + "grad_norm": 0.7116059064865112, + "learning_rate": 1.2903658753431077e-05, + "loss": 0.9679, + "step": 2533 + }, + { + "epoch": 0.42485591533060885, + "grad_norm": 0.4538300633430481, + "learning_rate": 1.2898461714744897e-05, + "loss": 0.8899, + "step": 2534 + }, + { + "epoch": 0.42502357749135494, + "grad_norm": 0.3443276882171631, + "learning_rate": 1.2893263821266462e-05, + "loss": 0.8228, + "step": 2535 + }, + { + "epoch": 0.42519123965210104, + "grad_norm": 0.44806644320487976, + "learning_rate": 1.28880650745287e-05, + "loss": 0.8981, + "step": 2536 + }, + { + "epoch": 0.42535890181284713, + "grad_norm": 0.5360666513442993, + "learning_rate": 1.2882865476064787e-05, + "loss": 0.8946, + "step": 2537 + }, + { + "epoch": 0.42552656397359323, + "grad_norm": 0.4778892397880554, + "learning_rate": 1.2877665027408146e-05, + "loss": 0.834, + "step": 2538 + }, + { + "epoch": 0.4256942261343393, + "grad_norm": 0.3988834023475647, + "learning_rate": 1.2872463730092457e-05, + "loss": 0.9192, + "step": 2539 + }, + { + "epoch": 0.4258618882950854, + "grad_norm": 0.33127453923225403, + "learning_rate": 1.286726158565164e-05, + "loss": 0.8632, + "step": 2540 + }, + { + "epoch": 0.4260295504558315, + "grad_norm": 0.5348523259162903, + "learning_rate": 1.2862058595619882e-05, + "loss": 0.8979, + "step": 2541 + }, + { + "epoch": 0.4261972126165776, + "grad_norm": 0.4789660573005676, + "learning_rate": 1.2856854761531601e-05, + "loss": 0.9459, + "step": 2542 + }, + { + "epoch": 0.4263648747773237, + "grad_norm": 0.45396173000335693, + "learning_rate": 1.2851650084921474e-05, + "loss": 0.9301, + "step": 2543 + }, + { + "epoch": 0.4265325369380698, + "grad_norm": 0.9283210635185242, + "learning_rate": 1.2846444567324428e-05, + "loss": 0.8131, + "step": 2544 + }, + { + "epoch": 0.4267001990988159, + "grad_norm": 0.4070189595222473, + "learning_rate": 1.284123821027563e-05, + "loss": 0.9318, + "step": 2545 + }, + { + "epoch": 0.426867861259562, + "grad_norm": 0.7896355390548706, + "learning_rate": 1.2836031015310504e-05, + "loss": 0.9275, + "step": 2546 + }, + { + "epoch": 0.4270355234203081, + "grad_norm": 0.5288454294204712, + "learning_rate": 1.283082298396471e-05, + "loss": 0.8689, + "step": 2547 + }, + { + "epoch": 0.4272031855810542, + "grad_norm": 0.6114990711212158, + "learning_rate": 1.2825614117774166e-05, + "loss": 0.8482, + "step": 2548 + }, + { + "epoch": 0.4273708477418003, + "grad_norm": 0.4103098213672638, + "learning_rate": 1.282040441827503e-05, + "loss": 0.8395, + "step": 2549 + }, + { + "epoch": 0.4275385099025464, + "grad_norm": 0.47522878646850586, + "learning_rate": 1.2815193887003707e-05, + "loss": 0.8157, + "step": 2550 + }, + { + "epoch": 0.4277061720632925, + "grad_norm": 0.4092075228691101, + "learning_rate": 1.2809982525496845e-05, + "loss": 0.8557, + "step": 2551 + }, + { + "epoch": 0.4278738342240386, + "grad_norm": 0.607429027557373, + "learning_rate": 1.2804770335291344e-05, + "loss": 0.8362, + "step": 2552 + }, + { + "epoch": 0.42804149638478467, + "grad_norm": 0.5187923908233643, + "learning_rate": 1.2799557317924336e-05, + "loss": 1.0054, + "step": 2553 + }, + { + "epoch": 0.42820915854553077, + "grad_norm": 0.6085100769996643, + "learning_rate": 1.2794343474933211e-05, + "loss": 0.8382, + "step": 2554 + }, + { + "epoch": 0.42837682070627686, + "grad_norm": 0.42959046363830566, + "learning_rate": 1.2789128807855592e-05, + "loss": 0.8725, + "step": 2555 + }, + { + "epoch": 0.42854448286702296, + "grad_norm": 0.3355506658554077, + "learning_rate": 1.278391331822935e-05, + "loss": 0.7844, + "step": 2556 + }, + { + "epoch": 0.42871214502776905, + "grad_norm": 0.5724706649780273, + "learning_rate": 1.2778697007592602e-05, + "loss": 0.9136, + "step": 2557 + }, + { + "epoch": 0.42887980718851515, + "grad_norm": 0.4948655366897583, + "learning_rate": 1.2773479877483695e-05, + "loss": 0.7616, + "step": 2558 + }, + { + "epoch": 0.42904746934926125, + "grad_norm": 0.44995033740997314, + "learning_rate": 1.2768261929441228e-05, + "loss": 0.8336, + "step": 2559 + }, + { + "epoch": 0.42921513151000734, + "grad_norm": 0.32069793343544006, + "learning_rate": 1.2763043165004038e-05, + "loss": 0.9137, + "step": 2560 + }, + { + "epoch": 0.42938279367075344, + "grad_norm": 0.4152913987636566, + "learning_rate": 1.2757823585711209e-05, + "loss": 0.8715, + "step": 2561 + }, + { + "epoch": 0.42955045583149953, + "grad_norm": 0.8363330364227295, + "learning_rate": 1.275260319310205e-05, + "loss": 0.8229, + "step": 2562 + }, + { + "epoch": 0.42971811799224563, + "grad_norm": 0.38372570276260376, + "learning_rate": 1.2747381988716127e-05, + "loss": 0.9127, + "step": 2563 + }, + { + "epoch": 0.4298857801529917, + "grad_norm": 0.5683997273445129, + "learning_rate": 1.2742159974093235e-05, + "loss": 0.9956, + "step": 2564 + }, + { + "epoch": 0.4300534423137378, + "grad_norm": 0.6207015514373779, + "learning_rate": 1.2736937150773413e-05, + "loss": 0.8893, + "step": 2565 + }, + { + "epoch": 0.4302211044744839, + "grad_norm": 0.48093733191490173, + "learning_rate": 1.2731713520296931e-05, + "loss": 0.964, + "step": 2566 + }, + { + "epoch": 0.43038876663523, + "grad_norm": 0.5342889428138733, + "learning_rate": 1.2726489084204308e-05, + "loss": 0.9238, + "step": 2567 + }, + { + "epoch": 0.4305564287959761, + "grad_norm": 0.36307162046432495, + "learning_rate": 1.272126384403629e-05, + "loss": 0.8753, + "step": 2568 + }, + { + "epoch": 0.4307240909567222, + "grad_norm": 0.5717389583587646, + "learning_rate": 1.2716037801333871e-05, + "loss": 0.9702, + "step": 2569 + }, + { + "epoch": 0.4308917531174683, + "grad_norm": 0.550253689289093, + "learning_rate": 1.2710810957638267e-05, + "loss": 0.873, + "step": 2570 + }, + { + "epoch": 0.4310594152782144, + "grad_norm": 0.5449841618537903, + "learning_rate": 1.2705583314490947e-05, + "loss": 0.9191, + "step": 2571 + }, + { + "epoch": 0.4312270774389605, + "grad_norm": 0.34667208790779114, + "learning_rate": 1.2700354873433608e-05, + "loss": 0.9353, + "step": 2572 + }, + { + "epoch": 0.4313947395997066, + "grad_norm": 0.4964389204978943, + "learning_rate": 1.2695125636008176e-05, + "loss": 0.8727, + "step": 2573 + }, + { + "epoch": 0.4315624017604527, + "grad_norm": 0.44565391540527344, + "learning_rate": 1.268989560375682e-05, + "loss": 0.8622, + "step": 2574 + }, + { + "epoch": 0.4317300639211988, + "grad_norm": 0.4834344983100891, + "learning_rate": 1.2684664778221943e-05, + "loss": 0.9428, + "step": 2575 + }, + { + "epoch": 0.4318977260819449, + "grad_norm": 0.6531493663787842, + "learning_rate": 1.2679433160946177e-05, + "loss": 0.959, + "step": 2576 + }, + { + "epoch": 0.432065388242691, + "grad_norm": 0.445249080657959, + "learning_rate": 1.2674200753472395e-05, + "loss": 0.8755, + "step": 2577 + }, + { + "epoch": 0.43223305040343707, + "grad_norm": 0.7106694579124451, + "learning_rate": 1.2668967557343695e-05, + "loss": 0.89, + "step": 2578 + }, + { + "epoch": 0.43240071256418317, + "grad_norm": 0.5058774948120117, + "learning_rate": 1.2663733574103415e-05, + "loss": 0.9328, + "step": 2579 + }, + { + "epoch": 0.43256837472492926, + "grad_norm": 0.5595027208328247, + "learning_rate": 1.2658498805295118e-05, + "loss": 0.8226, + "step": 2580 + }, + { + "epoch": 0.43273603688567536, + "grad_norm": 0.6688955426216125, + "learning_rate": 1.2653263252462607e-05, + "loss": 0.9153, + "step": 2581 + }, + { + "epoch": 0.43290369904642145, + "grad_norm": 0.7537813782691956, + "learning_rate": 1.2648026917149902e-05, + "loss": 0.8436, + "step": 2582 + }, + { + "epoch": 0.43307136120716755, + "grad_norm": 0.4780575633049011, + "learning_rate": 1.2642789800901273e-05, + "loss": 0.8802, + "step": 2583 + }, + { + "epoch": 0.43323902336791364, + "grad_norm": 0.6803731918334961, + "learning_rate": 1.2637551905261209e-05, + "loss": 0.8808, + "step": 2584 + }, + { + "epoch": 0.43340668552865974, + "grad_norm": 0.49162110686302185, + "learning_rate": 1.263231323177442e-05, + "loss": 0.8593, + "step": 2585 + }, + { + "epoch": 0.43357434768940584, + "grad_norm": 0.45431092381477356, + "learning_rate": 1.262707378198587e-05, + "loss": 0.817, + "step": 2586 + }, + { + "epoch": 0.43374200985015193, + "grad_norm": 0.464048832654953, + "learning_rate": 1.262183355744073e-05, + "loss": 0.8544, + "step": 2587 + }, + { + "epoch": 0.43390967201089803, + "grad_norm": 0.3684556782245636, + "learning_rate": 1.2616592559684408e-05, + "loss": 0.8597, + "step": 2588 + }, + { + "epoch": 0.4340773341716441, + "grad_norm": 0.514134407043457, + "learning_rate": 1.261135079026254e-05, + "loss": 0.9222, + "step": 2589 + }, + { + "epoch": 0.4342449963323902, + "grad_norm": 0.37602466344833374, + "learning_rate": 1.2606108250720985e-05, + "loss": 0.8417, + "step": 2590 + }, + { + "epoch": 0.4344126584931363, + "grad_norm": 0.5111823081970215, + "learning_rate": 1.260086494260584e-05, + "loss": 0.8551, + "step": 2591 + }, + { + "epoch": 0.4345803206538824, + "grad_norm": 0.4155372381210327, + "learning_rate": 1.2595620867463416e-05, + "loss": 0.9644, + "step": 2592 + }, + { + "epoch": 0.4347479828146285, + "grad_norm": 0.533291757106781, + "learning_rate": 1.2590376026840251e-05, + "loss": 0.8856, + "step": 2593 + }, + { + "epoch": 0.4349156449753746, + "grad_norm": 0.7136791944503784, + "learning_rate": 1.2585130422283129e-05, + "loss": 0.8378, + "step": 2594 + }, + { + "epoch": 0.4350833071361207, + "grad_norm": 0.6195120811462402, + "learning_rate": 1.2579884055339032e-05, + "loss": 0.8113, + "step": 2595 + }, + { + "epoch": 0.4352509692968668, + "grad_norm": 0.8092840313911438, + "learning_rate": 1.2574636927555179e-05, + "loss": 1.0516, + "step": 2596 + }, + { + "epoch": 0.4354186314576129, + "grad_norm": 0.561496376991272, + "learning_rate": 1.2569389040479012e-05, + "loss": 0.8378, + "step": 2597 + }, + { + "epoch": 0.435586293618359, + "grad_norm": 0.45630037784576416, + "learning_rate": 1.2564140395658207e-05, + "loss": 0.8966, + "step": 2598 + }, + { + "epoch": 0.4357539557791051, + "grad_norm": 0.4840919077396393, + "learning_rate": 1.2558890994640644e-05, + "loss": 0.8163, + "step": 2599 + }, + { + "epoch": 0.4359216179398512, + "grad_norm": 0.5131926536560059, + "learning_rate": 1.2553640838974439e-05, + "loss": 0.8156, + "step": 2600 + }, + { + "epoch": 0.4360892801005973, + "grad_norm": 0.5673779845237732, + "learning_rate": 1.2548389930207932e-05, + "loss": 0.8422, + "step": 2601 + }, + { + "epoch": 0.4362569422613434, + "grad_norm": 0.43523696064949036, + "learning_rate": 1.2543138269889679e-05, + "loss": 0.9012, + "step": 2602 + }, + { + "epoch": 0.43642460442208947, + "grad_norm": 0.6371061205863953, + "learning_rate": 1.253788585956846e-05, + "loss": 0.9298, + "step": 2603 + }, + { + "epoch": 0.43659226658283556, + "grad_norm": 0.4277431070804596, + "learning_rate": 1.253263270079327e-05, + "loss": 0.8382, + "step": 2604 + }, + { + "epoch": 0.43675992874358166, + "grad_norm": 0.49343574047088623, + "learning_rate": 1.2527378795113339e-05, + "loss": 0.8996, + "step": 2605 + }, + { + "epoch": 0.43692759090432776, + "grad_norm": 0.3353070616722107, + "learning_rate": 1.2522124144078103e-05, + "loss": 0.9396, + "step": 2606 + }, + { + "epoch": 0.43709525306507385, + "grad_norm": 0.37291547656059265, + "learning_rate": 1.251686874923723e-05, + "loss": 0.8347, + "step": 2607 + }, + { + "epoch": 0.43726291522581995, + "grad_norm": 0.6479389667510986, + "learning_rate": 1.251161261214059e-05, + "loss": 0.7938, + "step": 2608 + }, + { + "epoch": 0.43743057738656604, + "grad_norm": 0.5434373617172241, + "learning_rate": 1.2506355734338293e-05, + "loss": 0.8304, + "step": 2609 + }, + { + "epoch": 0.43759823954731214, + "grad_norm": 0.3810223639011383, + "learning_rate": 1.250109811738066e-05, + "loss": 0.8859, + "step": 2610 + }, + { + "epoch": 0.43776590170805824, + "grad_norm": 0.4282234311103821, + "learning_rate": 1.2495839762818216e-05, + "loss": 0.8614, + "step": 2611 + }, + { + "epoch": 0.43793356386880433, + "grad_norm": 0.4012936055660248, + "learning_rate": 1.249058067220172e-05, + "loss": 0.8971, + "step": 2612 + }, + { + "epoch": 0.4381012260295505, + "grad_norm": 0.3515467047691345, + "learning_rate": 1.2485320847082143e-05, + "loss": 0.9126, + "step": 2613 + }, + { + "epoch": 0.4382688881902966, + "grad_norm": 0.4747259020805359, + "learning_rate": 1.2480060289010677e-05, + "loss": 0.9065, + "step": 2614 + }, + { + "epoch": 0.4384365503510427, + "grad_norm": 0.5856126546859741, + "learning_rate": 1.247479899953872e-05, + "loss": 0.8844, + "step": 2615 + }, + { + "epoch": 0.43860421251178877, + "grad_norm": 0.672063946723938, + "learning_rate": 1.246953698021789e-05, + "loss": 0.858, + "step": 2616 + }, + { + "epoch": 0.43877187467253487, + "grad_norm": 0.5069361329078674, + "learning_rate": 1.2464274232600029e-05, + "loss": 0.8002, + "step": 2617 + }, + { + "epoch": 0.43893953683328096, + "grad_norm": 0.5100276470184326, + "learning_rate": 1.2459010758237178e-05, + "loss": 0.8739, + "step": 2618 + }, + { + "epoch": 0.43910719899402706, + "grad_norm": 0.6383470892906189, + "learning_rate": 1.2453746558681607e-05, + "loss": 0.9281, + "step": 2619 + }, + { + "epoch": 0.43927486115477316, + "grad_norm": 0.38440290093421936, + "learning_rate": 1.2448481635485792e-05, + "loss": 0.8567, + "step": 2620 + }, + { + "epoch": 0.43944252331551925, + "grad_norm": 0.3238489627838135, + "learning_rate": 1.2443215990202424e-05, + "loss": 0.8649, + "step": 2621 + }, + { + "epoch": 0.43961018547626535, + "grad_norm": 0.49014559388160706, + "learning_rate": 1.2437949624384405e-05, + "loss": 0.8069, + "step": 2622 + }, + { + "epoch": 0.43977784763701144, + "grad_norm": 0.8587152361869812, + "learning_rate": 1.2432682539584856e-05, + "loss": 0.8208, + "step": 2623 + }, + { + "epoch": 0.43994550979775754, + "grad_norm": 0.8027268648147583, + "learning_rate": 1.2427414737357097e-05, + "loss": 0.9219, + "step": 2624 + }, + { + "epoch": 0.44011317195850364, + "grad_norm": 0.44230809807777405, + "learning_rate": 1.242214621925468e-05, + "loss": 0.8819, + "step": 2625 + }, + { + "epoch": 0.44028083411924973, + "grad_norm": 0.3373022973537445, + "learning_rate": 1.2416876986831348e-05, + "loss": 0.9194, + "step": 2626 + }, + { + "epoch": 0.4404484962799958, + "grad_norm": 0.613947331905365, + "learning_rate": 1.2411607041641062e-05, + "loss": 0.9163, + "step": 2627 + }, + { + "epoch": 0.4406161584407419, + "grad_norm": 0.42350468039512634, + "learning_rate": 1.2406336385238e-05, + "loss": 0.8645, + "step": 2628 + }, + { + "epoch": 0.440783820601488, + "grad_norm": 0.4935706853866577, + "learning_rate": 1.240106501917654e-05, + "loss": 0.8883, + "step": 2629 + }, + { + "epoch": 0.4409514827622341, + "grad_norm": 0.4629109799861908, + "learning_rate": 1.2395792945011273e-05, + "loss": 0.9361, + "step": 2630 + }, + { + "epoch": 0.4411191449229802, + "grad_norm": 0.4974744915962219, + "learning_rate": 1.2390520164297e-05, + "loss": 0.9175, + "step": 2631 + }, + { + "epoch": 0.4412868070837263, + "grad_norm": 0.6437855958938599, + "learning_rate": 1.2385246678588732e-05, + "loss": 0.899, + "step": 2632 + }, + { + "epoch": 0.4414544692444724, + "grad_norm": 0.4009495675563812, + "learning_rate": 1.2379972489441678e-05, + "loss": 0.8831, + "step": 2633 + }, + { + "epoch": 0.4416221314052185, + "grad_norm": 0.41146427392959595, + "learning_rate": 1.237469759841127e-05, + "loss": 0.8963, + "step": 2634 + }, + { + "epoch": 0.4417897935659646, + "grad_norm": 0.4736334979534149, + "learning_rate": 1.236942200705313e-05, + "loss": 0.915, + "step": 2635 + }, + { + "epoch": 0.4419574557267107, + "grad_norm": 0.3997778296470642, + "learning_rate": 1.2364145716923103e-05, + "loss": 0.8388, + "step": 2636 + }, + { + "epoch": 0.4421251178874568, + "grad_norm": 0.38202333450317383, + "learning_rate": 1.2358868729577232e-05, + "loss": 0.8934, + "step": 2637 + }, + { + "epoch": 0.4422927800482029, + "grad_norm": 0.5045069456100464, + "learning_rate": 1.235359104657176e-05, + "loss": 0.8547, + "step": 2638 + }, + { + "epoch": 0.442460442208949, + "grad_norm": 0.5607749819755554, + "learning_rate": 1.2348312669463145e-05, + "loss": 0.8935, + "step": 2639 + }, + { + "epoch": 0.4426281043696951, + "grad_norm": 0.631809413433075, + "learning_rate": 1.2343033599808044e-05, + "loss": 0.9419, + "step": 2640 + }, + { + "epoch": 0.44279576653044117, + "grad_norm": 0.827242910861969, + "learning_rate": 1.2337753839163324e-05, + "loss": 0.8015, + "step": 2641 + }, + { + "epoch": 0.44296342869118727, + "grad_norm": 0.39556291699409485, + "learning_rate": 1.2332473389086043e-05, + "loss": 0.9154, + "step": 2642 + }, + { + "epoch": 0.44313109085193336, + "grad_norm": 1.0048776865005493, + "learning_rate": 1.2327192251133482e-05, + "loss": 0.9431, + "step": 2643 + }, + { + "epoch": 0.44329875301267946, + "grad_norm": 0.6327106356620789, + "learning_rate": 1.2321910426863108e-05, + "loss": 0.8219, + "step": 2644 + }, + { + "epoch": 0.44346641517342555, + "grad_norm": 0.6693361401557922, + "learning_rate": 1.2316627917832598e-05, + "loss": 0.9393, + "step": 2645 + }, + { + "epoch": 0.44363407733417165, + "grad_norm": 0.5261967778205872, + "learning_rate": 1.2311344725599828e-05, + "loss": 0.8466, + "step": 2646 + }, + { + "epoch": 0.44380173949491775, + "grad_norm": 0.5540294647216797, + "learning_rate": 1.2306060851722877e-05, + "loss": 0.8857, + "step": 2647 + }, + { + "epoch": 0.44396940165566384, + "grad_norm": 0.44563156366348267, + "learning_rate": 1.2300776297760026e-05, + "loss": 0.7964, + "step": 2648 + }, + { + "epoch": 0.44413706381640994, + "grad_norm": 1.2173652648925781, + "learning_rate": 1.2295491065269753e-05, + "loss": 0.9064, + "step": 2649 + }, + { + "epoch": 0.44430472597715603, + "grad_norm": 0.48927780985832214, + "learning_rate": 1.229020515581074e-05, + "loss": 0.9365, + "step": 2650 + }, + { + "epoch": 0.44447238813790213, + "grad_norm": 0.8384155035018921, + "learning_rate": 1.2284918570941871e-05, + "loss": 0.925, + "step": 2651 + }, + { + "epoch": 0.4446400502986482, + "grad_norm": 0.5220843553543091, + "learning_rate": 1.2279631312222218e-05, + "loss": 0.8733, + "step": 2652 + }, + { + "epoch": 0.4448077124593943, + "grad_norm": 1.3837053775787354, + "learning_rate": 1.2274343381211067e-05, + "loss": 0.9314, + "step": 2653 + }, + { + "epoch": 0.4449753746201404, + "grad_norm": 0.49828994274139404, + "learning_rate": 1.2269054779467884e-05, + "loss": 0.8824, + "step": 2654 + }, + { + "epoch": 0.4451430367808865, + "grad_norm": 0.4432735741138458, + "learning_rate": 1.2263765508552354e-05, + "loss": 1.0045, + "step": 2655 + }, + { + "epoch": 0.4453106989416326, + "grad_norm": 0.4565088748931885, + "learning_rate": 1.2258475570024343e-05, + "loss": 0.8646, + "step": 2656 + }, + { + "epoch": 0.4454783611023787, + "grad_norm": 0.587070643901825, + "learning_rate": 1.2253184965443918e-05, + "loss": 0.8346, + "step": 2657 + }, + { + "epoch": 0.4456460232631248, + "grad_norm": 0.4128307104110718, + "learning_rate": 1.2247893696371347e-05, + "loss": 0.8834, + "step": 2658 + }, + { + "epoch": 0.4458136854238709, + "grad_norm": 0.5621363520622253, + "learning_rate": 1.224260176436709e-05, + "loss": 0.8545, + "step": 2659 + }, + { + "epoch": 0.445981347584617, + "grad_norm": 0.46466559171676636, + "learning_rate": 1.2237309170991805e-05, + "loss": 0.8809, + "step": 2660 + }, + { + "epoch": 0.4461490097453631, + "grad_norm": 0.5122204422950745, + "learning_rate": 1.223201591780634e-05, + "loss": 0.8707, + "step": 2661 + }, + { + "epoch": 0.4463166719061092, + "grad_norm": 0.5194559097290039, + "learning_rate": 1.222672200637174e-05, + "loss": 0.9194, + "step": 2662 + }, + { + "epoch": 0.4464843340668553, + "grad_norm": 0.6348057985305786, + "learning_rate": 1.222142743824925e-05, + "loss": 0.7764, + "step": 2663 + }, + { + "epoch": 0.4466519962276014, + "grad_norm": 0.4446607530117035, + "learning_rate": 1.22161322150003e-05, + "loss": 0.9271, + "step": 2664 + }, + { + "epoch": 0.4468196583883475, + "grad_norm": 0.3675079345703125, + "learning_rate": 1.2210836338186515e-05, + "loss": 0.9728, + "step": 2665 + }, + { + "epoch": 0.44698732054909357, + "grad_norm": 0.580569326877594, + "learning_rate": 1.2205539809369719e-05, + "loss": 0.8281, + "step": 2666 + }, + { + "epoch": 0.44715498270983967, + "grad_norm": 0.6098851561546326, + "learning_rate": 1.2200242630111926e-05, + "loss": 0.9218, + "step": 2667 + }, + { + "epoch": 0.44732264487058576, + "grad_norm": 0.8198440074920654, + "learning_rate": 1.2194944801975334e-05, + "loss": 0.8709, + "step": 2668 + }, + { + "epoch": 0.44749030703133186, + "grad_norm": 0.6457405090332031, + "learning_rate": 1.218964632652234e-05, + "loss": 0.8942, + "step": 2669 + }, + { + "epoch": 0.44765796919207795, + "grad_norm": 0.6705514788627625, + "learning_rate": 1.218434720531553e-05, + "loss": 0.878, + "step": 2670 + }, + { + "epoch": 0.44782563135282405, + "grad_norm": 0.3943946063518524, + "learning_rate": 1.2179047439917681e-05, + "loss": 0.8365, + "step": 2671 + }, + { + "epoch": 0.44799329351357015, + "grad_norm": 0.40450936555862427, + "learning_rate": 1.217374703189176e-05, + "loss": 0.9172, + "step": 2672 + }, + { + "epoch": 0.44816095567431624, + "grad_norm": 0.5442905426025391, + "learning_rate": 1.2168445982800923e-05, + "loss": 0.8982, + "step": 2673 + }, + { + "epoch": 0.44832861783506234, + "grad_norm": 0.3890281319618225, + "learning_rate": 1.2163144294208511e-05, + "loss": 0.7854, + "step": 2674 + }, + { + "epoch": 0.44849627999580843, + "grad_norm": 0.41442999243736267, + "learning_rate": 1.2157841967678064e-05, + "loss": 0.8477, + "step": 2675 + }, + { + "epoch": 0.44866394215655453, + "grad_norm": 0.6467515230178833, + "learning_rate": 1.2152539004773298e-05, + "loss": 0.9126, + "step": 2676 + }, + { + "epoch": 0.4488316043173006, + "grad_norm": 0.6503413915634155, + "learning_rate": 1.2147235407058124e-05, + "loss": 0.8516, + "step": 2677 + }, + { + "epoch": 0.4489992664780467, + "grad_norm": 0.4241797924041748, + "learning_rate": 1.2141931176096643e-05, + "loss": 0.8361, + "step": 2678 + }, + { + "epoch": 0.4491669286387928, + "grad_norm": 0.4132331609725952, + "learning_rate": 1.2136626313453136e-05, + "loss": 0.8998, + "step": 2679 + }, + { + "epoch": 0.4493345907995389, + "grad_norm": 0.5378803610801697, + "learning_rate": 1.2131320820692067e-05, + "loss": 0.8882, + "step": 2680 + }, + { + "epoch": 0.449502252960285, + "grad_norm": 0.6029634475708008, + "learning_rate": 1.21260146993781e-05, + "loss": 0.8952, + "step": 2681 + }, + { + "epoch": 0.4496699151210311, + "grad_norm": 0.5718348026275635, + "learning_rate": 1.2120707951076069e-05, + "loss": 0.9178, + "step": 2682 + }, + { + "epoch": 0.4498375772817772, + "grad_norm": 0.839780330657959, + "learning_rate": 1.2115400577351009e-05, + "loss": 0.8232, + "step": 2683 + }, + { + "epoch": 0.4500052394425233, + "grad_norm": 0.4166347086429596, + "learning_rate": 1.2110092579768117e-05, + "loss": 0.8531, + "step": 2684 + }, + { + "epoch": 0.4501729016032694, + "grad_norm": 0.6671404242515564, + "learning_rate": 1.2104783959892798e-05, + "loss": 0.8739, + "step": 2685 + }, + { + "epoch": 0.4503405637640155, + "grad_norm": 0.4798339307308197, + "learning_rate": 1.2099474719290625e-05, + "loss": 0.8764, + "step": 2686 + }, + { + "epoch": 0.4505082259247616, + "grad_norm": 0.40089160203933716, + "learning_rate": 1.2094164859527363e-05, + "loss": 0.8791, + "step": 2687 + }, + { + "epoch": 0.4506758880855077, + "grad_norm": 0.36743858456611633, + "learning_rate": 1.2088854382168946e-05, + "loss": 0.8389, + "step": 2688 + }, + { + "epoch": 0.4508435502462538, + "grad_norm": 0.882380485534668, + "learning_rate": 1.2083543288781512e-05, + "loss": 0.8486, + "step": 2689 + }, + { + "epoch": 0.4510112124069999, + "grad_norm": 0.5218948125839233, + "learning_rate": 1.2078231580931359e-05, + "loss": 0.8741, + "step": 2690 + }, + { + "epoch": 0.45117887456774597, + "grad_norm": 0.5964364409446716, + "learning_rate": 1.2072919260184977e-05, + "loss": 0.9058, + "step": 2691 + }, + { + "epoch": 0.45134653672849206, + "grad_norm": 0.43583714962005615, + "learning_rate": 1.2067606328109038e-05, + "loss": 0.8084, + "step": 2692 + }, + { + "epoch": 0.45151419888923816, + "grad_norm": 0.5721093416213989, + "learning_rate": 1.2062292786270391e-05, + "loss": 0.7833, + "step": 2693 + }, + { + "epoch": 0.45168186104998426, + "grad_norm": 0.43429702520370483, + "learning_rate": 1.2056978636236065e-05, + "loss": 0.8776, + "step": 2694 + }, + { + "epoch": 0.45184952321073035, + "grad_norm": 0.5420209765434265, + "learning_rate": 1.2051663879573265e-05, + "loss": 0.9177, + "step": 2695 + }, + { + "epoch": 0.45201718537147645, + "grad_norm": 0.5308378338813782, + "learning_rate": 1.2046348517849384e-05, + "loss": 0.812, + "step": 2696 + }, + { + "epoch": 0.45218484753222254, + "grad_norm": 0.450838565826416, + "learning_rate": 1.2041032552631984e-05, + "loss": 0.8564, + "step": 2697 + }, + { + "epoch": 0.45235250969296864, + "grad_norm": 0.5462943315505981, + "learning_rate": 1.2035715985488814e-05, + "loss": 0.8829, + "step": 2698 + }, + { + "epoch": 0.4525201718537148, + "grad_norm": 0.41511619091033936, + "learning_rate": 1.2030398817987785e-05, + "loss": 0.8796, + "step": 2699 + }, + { + "epoch": 0.4526878340144609, + "grad_norm": 0.4854419529438019, + "learning_rate": 1.2025081051697007e-05, + "loss": 0.8153, + "step": 2700 + }, + { + "epoch": 0.452855496175207, + "grad_norm": 0.5956982374191284, + "learning_rate": 1.2019762688184753e-05, + "loss": 0.9383, + "step": 2701 + }, + { + "epoch": 0.4530231583359531, + "grad_norm": 0.43768510222435, + "learning_rate": 1.201444372901947e-05, + "loss": 0.7995, + "step": 2702 + }, + { + "epoch": 0.4531908204966992, + "grad_norm": 0.3518020212650299, + "learning_rate": 1.2009124175769784e-05, + "loss": 0.8851, + "step": 2703 + }, + { + "epoch": 0.45335848265744527, + "grad_norm": 0.632449209690094, + "learning_rate": 1.2003804030004502e-05, + "loss": 0.861, + "step": 2704 + }, + { + "epoch": 0.45352614481819137, + "grad_norm": 0.3550005853176117, + "learning_rate": 1.1998483293292602e-05, + "loss": 0.8596, + "step": 2705 + }, + { + "epoch": 0.45369380697893746, + "grad_norm": 0.5096614360809326, + "learning_rate": 1.1993161967203228e-05, + "loss": 0.8511, + "step": 2706 + }, + { + "epoch": 0.45386146913968356, + "grad_norm": 0.43639296293258667, + "learning_rate": 1.1987840053305706e-05, + "loss": 0.8645, + "step": 2707 + }, + { + "epoch": 0.45402913130042966, + "grad_norm": 0.8187398314476013, + "learning_rate": 1.1982517553169544e-05, + "loss": 0.9201, + "step": 2708 + }, + { + "epoch": 0.45419679346117575, + "grad_norm": 0.7677807807922363, + "learning_rate": 1.1977194468364403e-05, + "loss": 0.8638, + "step": 2709 + }, + { + "epoch": 0.45436445562192185, + "grad_norm": 0.42025095224380493, + "learning_rate": 1.1971870800460132e-05, + "loss": 0.802, + "step": 2710 + }, + { + "epoch": 0.45453211778266794, + "grad_norm": 0.43635404109954834, + "learning_rate": 1.196654655102674e-05, + "loss": 0.9023, + "step": 2711 + }, + { + "epoch": 0.45469977994341404, + "grad_norm": 0.46622127294540405, + "learning_rate": 1.1961221721634425e-05, + "loss": 0.9021, + "step": 2712 + }, + { + "epoch": 0.45486744210416014, + "grad_norm": 0.37527352571487427, + "learning_rate": 1.1955896313853535e-05, + "loss": 0.8711, + "step": 2713 + }, + { + "epoch": 0.45503510426490623, + "grad_norm": 0.5640096068382263, + "learning_rate": 1.1950570329254604e-05, + "loss": 0.8739, + "step": 2714 + }, + { + "epoch": 0.4552027664256523, + "grad_norm": 0.6676239967346191, + "learning_rate": 1.1945243769408329e-05, + "loss": 0.8787, + "step": 2715 + }, + { + "epoch": 0.4553704285863984, + "grad_norm": 0.5405741333961487, + "learning_rate": 1.1939916635885575e-05, + "loss": 0.8961, + "step": 2716 + }, + { + "epoch": 0.4555380907471445, + "grad_norm": 0.4485068917274475, + "learning_rate": 1.193458893025739e-05, + "loss": 0.8998, + "step": 2717 + }, + { + "epoch": 0.4557057529078906, + "grad_norm": 0.5062191486358643, + "learning_rate": 1.192926065409497e-05, + "loss": 0.8388, + "step": 2718 + }, + { + "epoch": 0.4558734150686367, + "grad_norm": 0.6288925409317017, + "learning_rate": 1.1923931808969693e-05, + "loss": 0.9418, + "step": 2719 + }, + { + "epoch": 0.4560410772293828, + "grad_norm": 0.47147685289382935, + "learning_rate": 1.1918602396453102e-05, + "loss": 0.7866, + "step": 2720 + }, + { + "epoch": 0.4562087393901289, + "grad_norm": 0.3272930383682251, + "learning_rate": 1.191327241811691e-05, + "loss": 0.8037, + "step": 2721 + }, + { + "epoch": 0.456376401550875, + "grad_norm": 0.4322454333305359, + "learning_rate": 1.1907941875532987e-05, + "loss": 0.8208, + "step": 2722 + }, + { + "epoch": 0.4565440637116211, + "grad_norm": 0.9034909605979919, + "learning_rate": 1.1902610770273379e-05, + "loss": 0.8868, + "step": 2723 + }, + { + "epoch": 0.4567117258723672, + "grad_norm": 0.6835198402404785, + "learning_rate": 1.1897279103910297e-05, + "loss": 0.8235, + "step": 2724 + }, + { + "epoch": 0.4568793880331133, + "grad_norm": 0.5388729572296143, + "learning_rate": 1.1891946878016114e-05, + "loss": 0.8918, + "step": 2725 + }, + { + "epoch": 0.4570470501938594, + "grad_norm": 0.4670707881450653, + "learning_rate": 1.1886614094163368e-05, + "loss": 0.8109, + "step": 2726 + }, + { + "epoch": 0.4572147123546055, + "grad_norm": 0.48207026720046997, + "learning_rate": 1.1881280753924763e-05, + "loss": 0.9336, + "step": 2727 + }, + { + "epoch": 0.4573823745153516, + "grad_norm": 0.589199423789978, + "learning_rate": 1.187594685887317e-05, + "loss": 0.8118, + "step": 2728 + }, + { + "epoch": 0.45755003667609767, + "grad_norm": 0.42270368337631226, + "learning_rate": 1.1870612410581617e-05, + "loss": 0.8729, + "step": 2729 + }, + { + "epoch": 0.45771769883684377, + "grad_norm": 0.4935012459754944, + "learning_rate": 1.18652774106233e-05, + "loss": 0.8112, + "step": 2730 + }, + { + "epoch": 0.45788536099758986, + "grad_norm": 0.4753749966621399, + "learning_rate": 1.185994186057158e-05, + "loss": 0.8847, + "step": 2731 + }, + { + "epoch": 0.45805302315833596, + "grad_norm": 0.48952072858810425, + "learning_rate": 1.185460576199997e-05, + "loss": 0.9115, + "step": 2732 + }, + { + "epoch": 0.45822068531908206, + "grad_norm": 0.4920048117637634, + "learning_rate": 1.1849269116482162e-05, + "loss": 0.9322, + "step": 2733 + }, + { + "epoch": 0.45838834747982815, + "grad_norm": 0.6082614064216614, + "learning_rate": 1.1843931925591987e-05, + "loss": 0.8564, + "step": 2734 + }, + { + "epoch": 0.45855600964057425, + "grad_norm": 0.48946818709373474, + "learning_rate": 1.1838594190903456e-05, + "loss": 0.8717, + "step": 2735 + }, + { + "epoch": 0.45872367180132034, + "grad_norm": 0.6212282180786133, + "learning_rate": 1.1833255913990732e-05, + "loss": 0.9478, + "step": 2736 + }, + { + "epoch": 0.45889133396206644, + "grad_norm": 0.4374319314956665, + "learning_rate": 1.1827917096428137e-05, + "loss": 0.855, + "step": 2737 + }, + { + "epoch": 0.45905899612281253, + "grad_norm": 0.6202477812767029, + "learning_rate": 1.1822577739790153e-05, + "loss": 0.8663, + "step": 2738 + }, + { + "epoch": 0.45922665828355863, + "grad_norm": 0.5401442646980286, + "learning_rate": 1.1817237845651429e-05, + "loss": 0.8311, + "step": 2739 + }, + { + "epoch": 0.4593943204443047, + "grad_norm": 0.6633997559547424, + "learning_rate": 1.1811897415586761e-05, + "loss": 0.8092, + "step": 2740 + }, + { + "epoch": 0.4595619826050508, + "grad_norm": 0.3960711658000946, + "learning_rate": 1.1806556451171108e-05, + "loss": 0.8179, + "step": 2741 + }, + { + "epoch": 0.4597296447657969, + "grad_norm": 0.6744838356971741, + "learning_rate": 1.1801214953979584e-05, + "loss": 0.7789, + "step": 2742 + }, + { + "epoch": 0.459897306926543, + "grad_norm": 0.45512571930885315, + "learning_rate": 1.179587292558747e-05, + "loss": 0.9245, + "step": 2743 + }, + { + "epoch": 0.4600649690872891, + "grad_norm": 0.3947019875049591, + "learning_rate": 1.1790530367570194e-05, + "loss": 0.8824, + "step": 2744 + }, + { + "epoch": 0.4602326312480352, + "grad_norm": 0.4425636827945709, + "learning_rate": 1.1785187281503336e-05, + "loss": 0.844, + "step": 2745 + }, + { + "epoch": 0.4604002934087813, + "grad_norm": 0.5157086849212646, + "learning_rate": 1.1779843668962647e-05, + "loss": 0.9509, + "step": 2746 + }, + { + "epoch": 0.4605679555695274, + "grad_norm": 0.4011687636375427, + "learning_rate": 1.1774499531524018e-05, + "loss": 0.9558, + "step": 2747 + }, + { + "epoch": 0.4607356177302735, + "grad_norm": 0.5148869156837463, + "learning_rate": 1.1769154870763504e-05, + "loss": 0.8244, + "step": 2748 + }, + { + "epoch": 0.4609032798910196, + "grad_norm": 0.6069275736808777, + "learning_rate": 1.1763809688257313e-05, + "loss": 0.9558, + "step": 2749 + }, + { + "epoch": 0.4610709420517657, + "grad_norm": 0.5783160328865051, + "learning_rate": 1.1758463985581803e-05, + "loss": 0.8771, + "step": 2750 + }, + { + "epoch": 0.4612386042125118, + "grad_norm": 0.36110037565231323, + "learning_rate": 1.1753117764313491e-05, + "loss": 0.8755, + "step": 2751 + }, + { + "epoch": 0.4614062663732579, + "grad_norm": 0.3790871202945709, + "learning_rate": 1.1747771026029038e-05, + "loss": 0.7351, + "step": 2752 + }, + { + "epoch": 0.461573928534004, + "grad_norm": 0.6918930411338806, + "learning_rate": 1.1742423772305269e-05, + "loss": 0.8708, + "step": 2753 + }, + { + "epoch": 0.46174159069475007, + "grad_norm": 0.48895999789237976, + "learning_rate": 1.1737076004719155e-05, + "loss": 0.8933, + "step": 2754 + }, + { + "epoch": 0.46190925285549617, + "grad_norm": 0.5753569006919861, + "learning_rate": 1.1731727724847816e-05, + "loss": 0.8738, + "step": 2755 + }, + { + "epoch": 0.46207691501624226, + "grad_norm": 0.418861985206604, + "learning_rate": 1.1726378934268525e-05, + "loss": 0.8168, + "step": 2756 + }, + { + "epoch": 0.46224457717698836, + "grad_norm": 0.4654471278190613, + "learning_rate": 1.172102963455871e-05, + "loss": 0.8289, + "step": 2757 + }, + { + "epoch": 0.46241223933773445, + "grad_norm": 1.0062460899353027, + "learning_rate": 1.1715679827295947e-05, + "loss": 0.9493, + "step": 2758 + }, + { + "epoch": 0.46257990149848055, + "grad_norm": 0.5639843940734863, + "learning_rate": 1.1710329514057956e-05, + "loss": 0.902, + "step": 2759 + }, + { + "epoch": 0.46274756365922665, + "grad_norm": 0.5350931882858276, + "learning_rate": 1.1704978696422614e-05, + "loss": 0.9216, + "step": 2760 + }, + { + "epoch": 0.46291522581997274, + "grad_norm": 0.489963561296463, + "learning_rate": 1.169962737596794e-05, + "loss": 0.7876, + "step": 2761 + }, + { + "epoch": 0.46308288798071884, + "grad_norm": 0.4331090748310089, + "learning_rate": 1.169427555427211e-05, + "loss": 0.8466, + "step": 2762 + }, + { + "epoch": 0.46325055014146493, + "grad_norm": 0.48761695623397827, + "learning_rate": 1.1688923232913436e-05, + "loss": 0.9871, + "step": 2763 + }, + { + "epoch": 0.46341821230221103, + "grad_norm": 0.3835407793521881, + "learning_rate": 1.1683570413470384e-05, + "loss": 0.9213, + "step": 2764 + }, + { + "epoch": 0.4635858744629571, + "grad_norm": 0.4346884787082672, + "learning_rate": 1.1678217097521576e-05, + "loss": 0.8505, + "step": 2765 + }, + { + "epoch": 0.4637535366237032, + "grad_norm": 0.5936617255210876, + "learning_rate": 1.1672863286645763e-05, + "loss": 0.9202, + "step": 2766 + }, + { + "epoch": 0.4639211987844493, + "grad_norm": 0.5256935954093933, + "learning_rate": 1.1667508982421853e-05, + "loss": 0.8315, + "step": 2767 + }, + { + "epoch": 0.4640888609451954, + "grad_norm": 0.5747020244598389, + "learning_rate": 1.1662154186428891e-05, + "loss": 0.8775, + "step": 2768 + }, + { + "epoch": 0.4642565231059415, + "grad_norm": 0.5799922347068787, + "learning_rate": 1.1656798900246082e-05, + "loss": 0.8667, + "step": 2769 + }, + { + "epoch": 0.4644241852666876, + "grad_norm": 0.4185238778591156, + "learning_rate": 1.165144312545276e-05, + "loss": 0.8788, + "step": 2770 + }, + { + "epoch": 0.4645918474274337, + "grad_norm": 0.4439437985420227, + "learning_rate": 1.164608686362841e-05, + "loss": 0.8039, + "step": 2771 + }, + { + "epoch": 0.4647595095881798, + "grad_norm": 0.5659282803535461, + "learning_rate": 1.1640730116352661e-05, + "loss": 0.8726, + "step": 2772 + }, + { + "epoch": 0.4649271717489259, + "grad_norm": 0.33008360862731934, + "learning_rate": 1.1635372885205285e-05, + "loss": 0.8077, + "step": 2773 + }, + { + "epoch": 0.465094833909672, + "grad_norm": 0.4482969343662262, + "learning_rate": 1.1630015171766196e-05, + "loss": 0.8443, + "step": 2774 + }, + { + "epoch": 0.4652624960704181, + "grad_norm": 0.42362096905708313, + "learning_rate": 1.1624656977615445e-05, + "loss": 0.7937, + "step": 2775 + }, + { + "epoch": 0.4654301582311642, + "grad_norm": 0.8205183148384094, + "learning_rate": 1.1619298304333236e-05, + "loss": 0.8775, + "step": 2776 + }, + { + "epoch": 0.4655978203919103, + "grad_norm": 0.46288955211639404, + "learning_rate": 1.1613939153499903e-05, + "loss": 0.8718, + "step": 2777 + }, + { + "epoch": 0.4657654825526564, + "grad_norm": 0.6447383761405945, + "learning_rate": 1.160857952669593e-05, + "loss": 0.8223, + "step": 2778 + }, + { + "epoch": 0.46593314471340247, + "grad_norm": 0.37369903922080994, + "learning_rate": 1.1603219425501935e-05, + "loss": 0.8591, + "step": 2779 + }, + { + "epoch": 0.46610080687414857, + "grad_norm": 0.4112333655357361, + "learning_rate": 1.159785885149868e-05, + "loss": 0.9124, + "step": 2780 + }, + { + "epoch": 0.46626846903489466, + "grad_norm": 0.33953040838241577, + "learning_rate": 1.1592497806267064e-05, + "loss": 0.8197, + "step": 2781 + }, + { + "epoch": 0.46643613119564076, + "grad_norm": 0.44476771354675293, + "learning_rate": 1.1587136291388126e-05, + "loss": 0.9272, + "step": 2782 + }, + { + "epoch": 0.46660379335638685, + "grad_norm": 0.6123186349868774, + "learning_rate": 1.1581774308443042e-05, + "loss": 0.8828, + "step": 2783 + }, + { + "epoch": 0.46677145551713295, + "grad_norm": 0.7039437294006348, + "learning_rate": 1.1576411859013125e-05, + "loss": 0.8523, + "step": 2784 + }, + { + "epoch": 0.4669391176778791, + "grad_norm": 0.5443616509437561, + "learning_rate": 1.1571048944679835e-05, + "loss": 0.8987, + "step": 2785 + }, + { + "epoch": 0.4671067798386252, + "grad_norm": 0.5298369526863098, + "learning_rate": 1.1565685567024754e-05, + "loss": 0.8838, + "step": 2786 + }, + { + "epoch": 0.4672744419993713, + "grad_norm": 0.4926570653915405, + "learning_rate": 1.1560321727629614e-05, + "loss": 0.8281, + "step": 2787 + }, + { + "epoch": 0.4674421041601174, + "grad_norm": 0.34552261233329773, + "learning_rate": 1.1554957428076272e-05, + "loss": 0.8708, + "step": 2788 + }, + { + "epoch": 0.4676097663208635, + "grad_norm": 0.546116054058075, + "learning_rate": 1.1549592669946732e-05, + "loss": 0.8396, + "step": 2789 + }, + { + "epoch": 0.4677774284816096, + "grad_norm": 0.5592318773269653, + "learning_rate": 1.1544227454823124e-05, + "loss": 0.8501, + "step": 2790 + }, + { + "epoch": 0.4679450906423557, + "grad_norm": 0.3868325352668762, + "learning_rate": 1.1538861784287718e-05, + "loss": 0.8726, + "step": 2791 + }, + { + "epoch": 0.4681127528031018, + "grad_norm": 0.5497609972953796, + "learning_rate": 1.1533495659922917e-05, + "loss": 0.9093, + "step": 2792 + }, + { + "epoch": 0.46828041496384787, + "grad_norm": 0.592467188835144, + "learning_rate": 1.1528129083311257e-05, + "loss": 0.9059, + "step": 2793 + }, + { + "epoch": 0.46844807712459396, + "grad_norm": 0.45810258388519287, + "learning_rate": 1.1522762056035407e-05, + "loss": 0.9085, + "step": 2794 + }, + { + "epoch": 0.46861573928534006, + "grad_norm": 0.44684118032455444, + "learning_rate": 1.1517394579678171e-05, + "loss": 0.8506, + "step": 2795 + }, + { + "epoch": 0.46878340144608616, + "grad_norm": 0.5014858841896057, + "learning_rate": 1.1512026655822483e-05, + "loss": 0.8976, + "step": 2796 + }, + { + "epoch": 0.46895106360683225, + "grad_norm": 0.37813401222229004, + "learning_rate": 1.1506658286051415e-05, + "loss": 0.8886, + "step": 2797 + }, + { + "epoch": 0.46911872576757835, + "grad_norm": 0.48495879769325256, + "learning_rate": 1.1501289471948158e-05, + "loss": 0.8978, + "step": 2798 + }, + { + "epoch": 0.46928638792832444, + "grad_norm": 0.528350830078125, + "learning_rate": 1.1495920215096046e-05, + "loss": 0.878, + "step": 2799 + }, + { + "epoch": 0.46945405008907054, + "grad_norm": 0.7721688151359558, + "learning_rate": 1.1490550517078538e-05, + "loss": 0.8757, + "step": 2800 + }, + { + "epoch": 0.46962171224981664, + "grad_norm": 0.5684508085250854, + "learning_rate": 1.1485180379479229e-05, + "loss": 0.8388, + "step": 2801 + }, + { + "epoch": 0.46978937441056273, + "grad_norm": 0.5709981322288513, + "learning_rate": 1.1479809803881829e-05, + "loss": 1.0181, + "step": 2802 + }, + { + "epoch": 0.46995703657130883, + "grad_norm": 0.4858846068382263, + "learning_rate": 1.1474438791870201e-05, + "loss": 0.8368, + "step": 2803 + }, + { + "epoch": 0.4701246987320549, + "grad_norm": 0.4464353919029236, + "learning_rate": 1.146906734502831e-05, + "loss": 0.8901, + "step": 2804 + }, + { + "epoch": 0.470292360892801, + "grad_norm": 0.47529369592666626, + "learning_rate": 1.146369546494027e-05, + "loss": 0.9403, + "step": 2805 + }, + { + "epoch": 0.4704600230535471, + "grad_norm": 0.4896434545516968, + "learning_rate": 1.1458323153190314e-05, + "loss": 0.8692, + "step": 2806 + }, + { + "epoch": 0.4706276852142932, + "grad_norm": 0.44182074069976807, + "learning_rate": 1.14529504113628e-05, + "loss": 0.8377, + "step": 2807 + }, + { + "epoch": 0.4707953473750393, + "grad_norm": 0.495044469833374, + "learning_rate": 1.144757724104222e-05, + "loss": 0.8722, + "step": 2808 + }, + { + "epoch": 0.4709630095357854, + "grad_norm": 0.6132462620735168, + "learning_rate": 1.1442203643813184e-05, + "loss": 0.8313, + "step": 2809 + }, + { + "epoch": 0.4711306716965315, + "grad_norm": 0.726916491985321, + "learning_rate": 1.1436829621260436e-05, + "loss": 0.8299, + "step": 2810 + }, + { + "epoch": 0.4712983338572776, + "grad_norm": 0.6654655933380127, + "learning_rate": 1.1431455174968839e-05, + "loss": 0.968, + "step": 2811 + }, + { + "epoch": 0.4714659960180237, + "grad_norm": 0.5855336785316467, + "learning_rate": 1.1426080306523388e-05, + "loss": 0.8825, + "step": 2812 + }, + { + "epoch": 0.4716336581787698, + "grad_norm": 0.5632380843162537, + "learning_rate": 1.1420705017509197e-05, + "loss": 0.8535, + "step": 2813 + }, + { + "epoch": 0.4718013203395159, + "grad_norm": 0.4087710380554199, + "learning_rate": 1.14153293095115e-05, + "loss": 0.8571, + "step": 2814 + }, + { + "epoch": 0.471968982500262, + "grad_norm": 0.4733235836029053, + "learning_rate": 1.1409953184115668e-05, + "loss": 0.9356, + "step": 2815 + }, + { + "epoch": 0.4721366446610081, + "grad_norm": 0.7453212738037109, + "learning_rate": 1.1404576642907179e-05, + "loss": 0.8702, + "step": 2816 + }, + { + "epoch": 0.47230430682175417, + "grad_norm": 0.3844403326511383, + "learning_rate": 1.1399199687471647e-05, + "loss": 0.8673, + "step": 2817 + }, + { + "epoch": 0.47247196898250027, + "grad_norm": 0.615859866142273, + "learning_rate": 1.13938223193948e-05, + "loss": 0.8172, + "step": 2818 + }, + { + "epoch": 0.47263963114324636, + "grad_norm": 0.43845224380493164, + "learning_rate": 1.1388444540262496e-05, + "loss": 0.8666, + "step": 2819 + }, + { + "epoch": 0.47280729330399246, + "grad_norm": 0.38096320629119873, + "learning_rate": 1.1383066351660698e-05, + "loss": 0.8549, + "step": 2820 + }, + { + "epoch": 0.47297495546473856, + "grad_norm": 0.974331259727478, + "learning_rate": 1.1377687755175511e-05, + "loss": 0.9511, + "step": 2821 + }, + { + "epoch": 0.47314261762548465, + "grad_norm": 0.7325894236564636, + "learning_rate": 1.1372308752393144e-05, + "loss": 0.7975, + "step": 2822 + }, + { + "epoch": 0.47331027978623075, + "grad_norm": 0.39797723293304443, + "learning_rate": 1.1366929344899931e-05, + "loss": 0.8754, + "step": 2823 + }, + { + "epoch": 0.47347794194697684, + "grad_norm": 0.3541925251483917, + "learning_rate": 1.1361549534282332e-05, + "loss": 0.9109, + "step": 2824 + }, + { + "epoch": 0.47364560410772294, + "grad_norm": 0.5293611288070679, + "learning_rate": 1.1356169322126911e-05, + "loss": 0.8164, + "step": 2825 + }, + { + "epoch": 0.47381326626846904, + "grad_norm": 0.5768407583236694, + "learning_rate": 1.135078871002037e-05, + "loss": 0.9299, + "step": 2826 + }, + { + "epoch": 0.47398092842921513, + "grad_norm": 0.46879512071609497, + "learning_rate": 1.1345407699549504e-05, + "loss": 0.8096, + "step": 2827 + }, + { + "epoch": 0.4741485905899612, + "grad_norm": 0.3861348330974579, + "learning_rate": 1.1340026292301252e-05, + "loss": 0.8833, + "step": 2828 + }, + { + "epoch": 0.4743162527507073, + "grad_norm": 0.42656075954437256, + "learning_rate": 1.1334644489862652e-05, + "loss": 0.8739, + "step": 2829 + }, + { + "epoch": 0.4744839149114534, + "grad_norm": 0.3519597053527832, + "learning_rate": 1.1329262293820863e-05, + "loss": 0.839, + "step": 2830 + }, + { + "epoch": 0.4746515770721995, + "grad_norm": 0.48209017515182495, + "learning_rate": 1.132387970576317e-05, + "loss": 0.9045, + "step": 2831 + }, + { + "epoch": 0.4748192392329456, + "grad_norm": 0.7903998494148254, + "learning_rate": 1.1318496727276953e-05, + "loss": 0.7866, + "step": 2832 + }, + { + "epoch": 0.4749869013936917, + "grad_norm": 0.45061150193214417, + "learning_rate": 1.1313113359949726e-05, + "loss": 0.8044, + "step": 2833 + }, + { + "epoch": 0.4751545635544378, + "grad_norm": 0.4440174102783203, + "learning_rate": 1.1307729605369109e-05, + "loss": 0.942, + "step": 2834 + }, + { + "epoch": 0.4753222257151839, + "grad_norm": 0.37487006187438965, + "learning_rate": 1.1302345465122839e-05, + "loss": 0.9064, + "step": 2835 + }, + { + "epoch": 0.47548988787593, + "grad_norm": 0.6825331449508667, + "learning_rate": 1.129696094079876e-05, + "loss": 0.7952, + "step": 2836 + }, + { + "epoch": 0.4756575500366761, + "grad_norm": 0.5929781198501587, + "learning_rate": 1.1291576033984844e-05, + "loss": 0.8484, + "step": 2837 + }, + { + "epoch": 0.4758252121974222, + "grad_norm": 0.568448543548584, + "learning_rate": 1.1286190746269164e-05, + "loss": 0.8992, + "step": 2838 + }, + { + "epoch": 0.4759928743581683, + "grad_norm": 0.6622653007507324, + "learning_rate": 1.1280805079239903e-05, + "loss": 0.9454, + "step": 2839 + }, + { + "epoch": 0.4761605365189144, + "grad_norm": 0.35762348771095276, + "learning_rate": 1.1275419034485365e-05, + "loss": 0.7527, + "step": 2840 + }, + { + "epoch": 0.4763281986796605, + "grad_norm": 0.5092772245407104, + "learning_rate": 1.127003261359396e-05, + "loss": 0.8905, + "step": 2841 + }, + { + "epoch": 0.47649586084040657, + "grad_norm": 0.5554852485656738, + "learning_rate": 1.1264645818154211e-05, + "loss": 0.9333, + "step": 2842 + }, + { + "epoch": 0.47666352300115267, + "grad_norm": 0.40020039677619934, + "learning_rate": 1.125925864975475e-05, + "loss": 0.8556, + "step": 2843 + }, + { + "epoch": 0.47683118516189876, + "grad_norm": 0.5516888499259949, + "learning_rate": 1.125387110998432e-05, + "loss": 0.8682, + "step": 2844 + }, + { + "epoch": 0.47699884732264486, + "grad_norm": 0.6319754719734192, + "learning_rate": 1.1248483200431774e-05, + "loss": 0.8332, + "step": 2845 + }, + { + "epoch": 0.47716650948339095, + "grad_norm": 0.4318239688873291, + "learning_rate": 1.124309492268607e-05, + "loss": 0.7839, + "step": 2846 + }, + { + "epoch": 0.47733417164413705, + "grad_norm": 0.3661859929561615, + "learning_rate": 1.1237706278336284e-05, + "loss": 0.8549, + "step": 2847 + }, + { + "epoch": 0.47750183380488315, + "grad_norm": 0.3942444622516632, + "learning_rate": 1.1232317268971586e-05, + "loss": 0.8146, + "step": 2848 + }, + { + "epoch": 0.47766949596562924, + "grad_norm": 0.570182740688324, + "learning_rate": 1.1226927896181271e-05, + "loss": 0.8495, + "step": 2849 + }, + { + "epoch": 0.47783715812637534, + "grad_norm": 0.46788638830184937, + "learning_rate": 1.1221538161554724e-05, + "loss": 0.8217, + "step": 2850 + }, + { + "epoch": 0.47800482028712143, + "grad_norm": 0.403129518032074, + "learning_rate": 1.1216148066681448e-05, + "loss": 0.8135, + "step": 2851 + }, + { + "epoch": 0.47817248244786753, + "grad_norm": 0.4809132516384125, + "learning_rate": 1.1210757613151044e-05, + "loss": 0.7873, + "step": 2852 + }, + { + "epoch": 0.4783401446086136, + "grad_norm": 0.7033389806747437, + "learning_rate": 1.1205366802553231e-05, + "loss": 0.8182, + "step": 2853 + }, + { + "epoch": 0.4785078067693597, + "grad_norm": 0.4435984790325165, + "learning_rate": 1.1199975636477822e-05, + "loss": 0.9731, + "step": 2854 + }, + { + "epoch": 0.4786754689301058, + "grad_norm": 0.34627559781074524, + "learning_rate": 1.119458411651474e-05, + "loss": 0.9428, + "step": 2855 + }, + { + "epoch": 0.4788431310908519, + "grad_norm": 0.4384484887123108, + "learning_rate": 1.1189192244254006e-05, + "loss": 0.8891, + "step": 2856 + }, + { + "epoch": 0.479010793251598, + "grad_norm": 0.49134066700935364, + "learning_rate": 1.1183800021285756e-05, + "loss": 0.8339, + "step": 2857 + }, + { + "epoch": 0.4791784554123441, + "grad_norm": 0.6400789618492126, + "learning_rate": 1.1178407449200222e-05, + "loss": 0.8377, + "step": 2858 + }, + { + "epoch": 0.4793461175730902, + "grad_norm": 0.5748932957649231, + "learning_rate": 1.1173014529587734e-05, + "loss": 0.8434, + "step": 2859 + }, + { + "epoch": 0.4795137797338363, + "grad_norm": 0.6503258347511292, + "learning_rate": 1.1167621264038742e-05, + "loss": 0.9055, + "step": 2860 + }, + { + "epoch": 0.4796814418945824, + "grad_norm": 0.43417903780937195, + "learning_rate": 1.1162227654143777e-05, + "loss": 0.8923, + "step": 2861 + }, + { + "epoch": 0.4798491040553285, + "grad_norm": 0.4653598368167877, + "learning_rate": 1.1156833701493486e-05, + "loss": 0.8268, + "step": 2862 + }, + { + "epoch": 0.4800167662160746, + "grad_norm": 0.665627658367157, + "learning_rate": 1.1151439407678612e-05, + "loss": 0.8754, + "step": 2863 + }, + { + "epoch": 0.4801844283768207, + "grad_norm": 0.46827465295791626, + "learning_rate": 1.1146044774289997e-05, + "loss": 0.8615, + "step": 2864 + }, + { + "epoch": 0.4803520905375668, + "grad_norm": 0.4687343239784241, + "learning_rate": 1.1140649802918591e-05, + "loss": 0.9117, + "step": 2865 + }, + { + "epoch": 0.4805197526983129, + "grad_norm": 0.4067821502685547, + "learning_rate": 1.113525449515543e-05, + "loss": 0.8897, + "step": 2866 + }, + { + "epoch": 0.48068741485905897, + "grad_norm": 0.5301191210746765, + "learning_rate": 1.1129858852591662e-05, + "loss": 0.8381, + "step": 2867 + }, + { + "epoch": 0.48085507701980507, + "grad_norm": 0.5383304953575134, + "learning_rate": 1.1124462876818526e-05, + "loss": 0.8658, + "step": 2868 + }, + { + "epoch": 0.48102273918055116, + "grad_norm": 0.34599506855010986, + "learning_rate": 1.1119066569427366e-05, + "loss": 0.8396, + "step": 2869 + }, + { + "epoch": 0.4811904013412973, + "grad_norm": 0.40095260739326477, + "learning_rate": 1.1113669932009619e-05, + "loss": 0.824, + "step": 2870 + }, + { + "epoch": 0.4813580635020434, + "grad_norm": 0.5407511591911316, + "learning_rate": 1.1108272966156812e-05, + "loss": 0.7932, + "step": 2871 + }, + { + "epoch": 0.4815257256627895, + "grad_norm": 0.3525823652744293, + "learning_rate": 1.1102875673460592e-05, + "loss": 0.8249, + "step": 2872 + }, + { + "epoch": 0.4816933878235356, + "grad_norm": 0.4650687873363495, + "learning_rate": 1.1097478055512674e-05, + "loss": 0.8797, + "step": 2873 + }, + { + "epoch": 0.4818610499842817, + "grad_norm": 0.5316118001937866, + "learning_rate": 1.1092080113904886e-05, + "loss": 0.8579, + "step": 2874 + }, + { + "epoch": 0.4820287121450278, + "grad_norm": 0.6918269395828247, + "learning_rate": 1.1086681850229153e-05, + "loss": 0.8817, + "step": 2875 + }, + { + "epoch": 0.4821963743057739, + "grad_norm": 0.40541380643844604, + "learning_rate": 1.1081283266077488e-05, + "loss": 0.8493, + "step": 2876 + }, + { + "epoch": 0.48236403646652, + "grad_norm": 0.5759273767471313, + "learning_rate": 1.1075884363041994e-05, + "loss": 0.9329, + "step": 2877 + }, + { + "epoch": 0.4825316986272661, + "grad_norm": 0.739473283290863, + "learning_rate": 1.107048514271488e-05, + "loss": 0.991, + "step": 2878 + }, + { + "epoch": 0.4826993607880122, + "grad_norm": 0.5789638757705688, + "learning_rate": 1.1065085606688442e-05, + "loss": 0.8588, + "step": 2879 + }, + { + "epoch": 0.4828670229487583, + "grad_norm": 0.5917166471481323, + "learning_rate": 1.1059685756555069e-05, + "loss": 0.901, + "step": 2880 + }, + { + "epoch": 0.48303468510950437, + "grad_norm": 0.42208248376846313, + "learning_rate": 1.1054285593907248e-05, + "loss": 0.8927, + "step": 2881 + }, + { + "epoch": 0.48320234727025047, + "grad_norm": 0.6999191641807556, + "learning_rate": 1.1048885120337546e-05, + "loss": 0.938, + "step": 2882 + }, + { + "epoch": 0.48337000943099656, + "grad_norm": 0.6355289816856384, + "learning_rate": 1.1043484337438637e-05, + "loss": 0.9418, + "step": 2883 + }, + { + "epoch": 0.48353767159174266, + "grad_norm": 0.566191554069519, + "learning_rate": 1.1038083246803275e-05, + "loss": 0.9113, + "step": 2884 + }, + { + "epoch": 0.48370533375248875, + "grad_norm": 0.7057574391365051, + "learning_rate": 1.1032681850024307e-05, + "loss": 0.7919, + "step": 2885 + }, + { + "epoch": 0.48387299591323485, + "grad_norm": 0.4619996249675751, + "learning_rate": 1.1027280148694676e-05, + "loss": 0.8256, + "step": 2886 + }, + { + "epoch": 0.48404065807398095, + "grad_norm": 0.941382646560669, + "learning_rate": 1.1021878144407408e-05, + "loss": 0.9087, + "step": 2887 + }, + { + "epoch": 0.48420832023472704, + "grad_norm": 0.39337483048439026, + "learning_rate": 1.1016475838755623e-05, + "loss": 0.8315, + "step": 2888 + }, + { + "epoch": 0.48437598239547314, + "grad_norm": 0.40230581164360046, + "learning_rate": 1.1011073233332525e-05, + "loss": 0.8307, + "step": 2889 + }, + { + "epoch": 0.48454364455621923, + "grad_norm": 0.6783583760261536, + "learning_rate": 1.100567032973141e-05, + "loss": 0.9285, + "step": 2890 + }, + { + "epoch": 0.48471130671696533, + "grad_norm": 0.5566238164901733, + "learning_rate": 1.1000267129545661e-05, + "loss": 0.7715, + "step": 2891 + }, + { + "epoch": 0.4848789688777114, + "grad_norm": 0.6129046082496643, + "learning_rate": 1.0994863634368754e-05, + "loss": 0.8487, + "step": 2892 + }, + { + "epoch": 0.4850466310384575, + "grad_norm": 0.5205916166305542, + "learning_rate": 1.0989459845794234e-05, + "loss": 0.9206, + "step": 2893 + }, + { + "epoch": 0.4852142931992036, + "grad_norm": 0.6074056625366211, + "learning_rate": 1.0984055765415757e-05, + "loss": 0.7813, + "step": 2894 + }, + { + "epoch": 0.4853819553599497, + "grad_norm": 1.0004022121429443, + "learning_rate": 1.0978651394827054e-05, + "loss": 0.8608, + "step": 2895 + }, + { + "epoch": 0.4855496175206958, + "grad_norm": 0.9308025240898132, + "learning_rate": 1.0973246735621932e-05, + "loss": 0.8962, + "step": 2896 + }, + { + "epoch": 0.4857172796814419, + "grad_norm": 0.5531139969825745, + "learning_rate": 1.0967841789394296e-05, + "loss": 0.8099, + "step": 2897 + }, + { + "epoch": 0.485884941842188, + "grad_norm": 0.6492610573768616, + "learning_rate": 1.0962436557738131e-05, + "loss": 0.79, + "step": 2898 + }, + { + "epoch": 0.4860526040029341, + "grad_norm": 0.557978630065918, + "learning_rate": 1.095703104224751e-05, + "loss": 0.7553, + "step": 2899 + }, + { + "epoch": 0.4862202661636802, + "grad_norm": 0.5231502652168274, + "learning_rate": 1.0951625244516584e-05, + "loss": 0.898, + "step": 2900 + }, + { + "epoch": 0.4863879283244263, + "grad_norm": 0.8099695444107056, + "learning_rate": 1.0946219166139587e-05, + "loss": 0.8512, + "step": 2901 + }, + { + "epoch": 0.4865555904851724, + "grad_norm": 0.5336698293685913, + "learning_rate": 1.094081280871084e-05, + "loss": 0.8892, + "step": 2902 + }, + { + "epoch": 0.4867232526459185, + "grad_norm": 0.3855321705341339, + "learning_rate": 1.0935406173824749e-05, + "loss": 0.8501, + "step": 2903 + }, + { + "epoch": 0.4868909148066646, + "grad_norm": 0.3807976245880127, + "learning_rate": 1.0929999263075795e-05, + "loss": 0.8586, + "step": 2904 + }, + { + "epoch": 0.4870585769674107, + "grad_norm": 1.0085023641586304, + "learning_rate": 1.0924592078058538e-05, + "loss": 0.9726, + "step": 2905 + }, + { + "epoch": 0.48722623912815677, + "grad_norm": 0.4902290105819702, + "learning_rate": 1.0919184620367634e-05, + "loss": 0.8192, + "step": 2906 + }, + { + "epoch": 0.48739390128890286, + "grad_norm": 0.5184853672981262, + "learning_rate": 1.0913776891597798e-05, + "loss": 0.9842, + "step": 2907 + }, + { + "epoch": 0.48756156344964896, + "grad_norm": 0.6011456847190857, + "learning_rate": 1.0908368893343843e-05, + "loss": 0.8141, + "step": 2908 + }, + { + "epoch": 0.48772922561039506, + "grad_norm": 0.4382106065750122, + "learning_rate": 1.0902960627200655e-05, + "loss": 0.851, + "step": 2909 + }, + { + "epoch": 0.48789688777114115, + "grad_norm": 0.47010084986686707, + "learning_rate": 1.0897552094763196e-05, + "loss": 0.9802, + "step": 2910 + }, + { + "epoch": 0.48806454993188725, + "grad_norm": 0.46682846546173096, + "learning_rate": 1.0892143297626511e-05, + "loss": 0.8456, + "step": 2911 + }, + { + "epoch": 0.48823221209263334, + "grad_norm": 0.5086984634399414, + "learning_rate": 1.0886734237385718e-05, + "loss": 0.8319, + "step": 2912 + }, + { + "epoch": 0.48839987425337944, + "grad_norm": 0.4189518392086029, + "learning_rate": 1.088132491563602e-05, + "loss": 0.8569, + "step": 2913 + }, + { + "epoch": 0.48856753641412554, + "grad_norm": 0.4165961742401123, + "learning_rate": 1.0875915333972689e-05, + "loss": 0.8158, + "step": 2914 + }, + { + "epoch": 0.48873519857487163, + "grad_norm": 0.6751852035522461, + "learning_rate": 1.0870505493991084e-05, + "loss": 0.8983, + "step": 2915 + }, + { + "epoch": 0.48890286073561773, + "grad_norm": 0.4144243001937866, + "learning_rate": 1.0865095397286623e-05, + "loss": 0.8411, + "step": 2916 + }, + { + "epoch": 0.4890705228963638, + "grad_norm": 0.6219884753227234, + "learning_rate": 1.0859685045454821e-05, + "loss": 0.8788, + "step": 2917 + }, + { + "epoch": 0.4892381850571099, + "grad_norm": 0.4056248366832733, + "learning_rate": 1.0854274440091254e-05, + "loss": 0.8773, + "step": 2918 + }, + { + "epoch": 0.489405847217856, + "grad_norm": 0.47027096152305603, + "learning_rate": 1.0848863582791576e-05, + "loss": 0.8727, + "step": 2919 + }, + { + "epoch": 0.4895735093786021, + "grad_norm": 0.9668684601783752, + "learning_rate": 1.0843452475151515e-05, + "loss": 0.9114, + "step": 2920 + }, + { + "epoch": 0.4897411715393482, + "grad_norm": 0.39250072836875916, + "learning_rate": 1.0838041118766876e-05, + "loss": 0.8781, + "step": 2921 + }, + { + "epoch": 0.4899088337000943, + "grad_norm": 0.8956718444824219, + "learning_rate": 1.0832629515233534e-05, + "loss": 0.9103, + "step": 2922 + }, + { + "epoch": 0.4900764958608404, + "grad_norm": 0.862220823764801, + "learning_rate": 1.0827217666147435e-05, + "loss": 0.8376, + "step": 2923 + }, + { + "epoch": 0.4902441580215865, + "grad_norm": 0.46078476309776306, + "learning_rate": 1.0821805573104607e-05, + "loss": 0.9565, + "step": 2924 + }, + { + "epoch": 0.4904118201823326, + "grad_norm": 0.38189762830734253, + "learning_rate": 1.0816393237701134e-05, + "loss": 0.8986, + "step": 2925 + }, + { + "epoch": 0.4905794823430787, + "grad_norm": 0.46521735191345215, + "learning_rate": 1.081098066153319e-05, + "loss": 0.9397, + "step": 2926 + }, + { + "epoch": 0.4907471445038248, + "grad_norm": 0.41276007890701294, + "learning_rate": 1.0805567846197007e-05, + "loss": 0.8809, + "step": 2927 + }, + { + "epoch": 0.4909148066645709, + "grad_norm": 0.36183953285217285, + "learning_rate": 1.0800154793288888e-05, + "loss": 0.8985, + "step": 2928 + }, + { + "epoch": 0.491082468825317, + "grad_norm": 0.592648983001709, + "learning_rate": 1.0794741504405215e-05, + "loss": 0.921, + "step": 2929 + }, + { + "epoch": 0.49125013098606307, + "grad_norm": 0.7352839708328247, + "learning_rate": 1.078932798114243e-05, + "loss": 0.8357, + "step": 2930 + }, + { + "epoch": 0.49141779314680917, + "grad_norm": 0.39497673511505127, + "learning_rate": 1.078391422509705e-05, + "loss": 0.7776, + "step": 2931 + }, + { + "epoch": 0.49158545530755526, + "grad_norm": 0.4355677366256714, + "learning_rate": 1.0778500237865656e-05, + "loss": 0.87, + "step": 2932 + }, + { + "epoch": 0.49175311746830136, + "grad_norm": 0.403594434261322, + "learning_rate": 1.0773086021044905e-05, + "loss": 0.8779, + "step": 2933 + }, + { + "epoch": 0.49192077962904746, + "grad_norm": 0.4891184866428375, + "learning_rate": 1.0767671576231512e-05, + "loss": 0.8959, + "step": 2934 + }, + { + "epoch": 0.49208844178979355, + "grad_norm": 0.39034304022789, + "learning_rate": 1.0762256905022266e-05, + "loss": 0.8926, + "step": 2935 + }, + { + "epoch": 0.49225610395053965, + "grad_norm": 0.5464780330657959, + "learning_rate": 1.0756842009014016e-05, + "loss": 0.8915, + "step": 2936 + }, + { + "epoch": 0.49242376611128574, + "grad_norm": 0.7192713618278503, + "learning_rate": 1.0751426889803689e-05, + "loss": 0.8104, + "step": 2937 + }, + { + "epoch": 0.49259142827203184, + "grad_norm": 0.3802322447299957, + "learning_rate": 1.0746011548988269e-05, + "loss": 0.853, + "step": 2938 + }, + { + "epoch": 0.49275909043277794, + "grad_norm": 0.6255716681480408, + "learning_rate": 1.07405959881648e-05, + "loss": 0.9386, + "step": 2939 + }, + { + "epoch": 0.49292675259352403, + "grad_norm": 0.5813717842102051, + "learning_rate": 1.073518020893041e-05, + "loss": 0.9199, + "step": 2940 + }, + { + "epoch": 0.4930944147542701, + "grad_norm": 0.6427274942398071, + "learning_rate": 1.0729764212882268e-05, + "loss": 0.696, + "step": 2941 + }, + { + "epoch": 0.4932620769150162, + "grad_norm": 0.34310704469680786, + "learning_rate": 1.0724348001617626e-05, + "loss": 0.918, + "step": 2942 + }, + { + "epoch": 0.4934297390757623, + "grad_norm": 0.5016130805015564, + "learning_rate": 1.0718931576733787e-05, + "loss": 0.8238, + "step": 2943 + }, + { + "epoch": 0.4935974012365084, + "grad_norm": 0.3720301687717438, + "learning_rate": 1.0713514939828122e-05, + "loss": 0.7974, + "step": 2944 + }, + { + "epoch": 0.4937650633972545, + "grad_norm": 0.4041135311126709, + "learning_rate": 1.0708098092498071e-05, + "loss": 0.8674, + "step": 2945 + }, + { + "epoch": 0.4939327255580006, + "grad_norm": 0.47650477290153503, + "learning_rate": 1.070268103634112e-05, + "loss": 0.8526, + "step": 2946 + }, + { + "epoch": 0.4941003877187467, + "grad_norm": 0.5747629404067993, + "learning_rate": 1.0697263772954831e-05, + "loss": 0.8781, + "step": 2947 + }, + { + "epoch": 0.4942680498794928, + "grad_norm": 0.49312344193458557, + "learning_rate": 1.0691846303936822e-05, + "loss": 0.8176, + "step": 2948 + }, + { + "epoch": 0.4944357120402389, + "grad_norm": 0.5730398893356323, + "learning_rate": 1.0686428630884774e-05, + "loss": 0.8855, + "step": 2949 + }, + { + "epoch": 0.494603374200985, + "grad_norm": 0.4680347144603729, + "learning_rate": 1.0681010755396417e-05, + "loss": 0.9677, + "step": 2950 + }, + { + "epoch": 0.4947710363617311, + "grad_norm": 0.4128454327583313, + "learning_rate": 1.0675592679069558e-05, + "loss": 0.8525, + "step": 2951 + }, + { + "epoch": 0.4949386985224772, + "grad_norm": 0.4013599157333374, + "learning_rate": 1.0670174403502051e-05, + "loss": 0.8549, + "step": 2952 + }, + { + "epoch": 0.4951063606832233, + "grad_norm": 0.4626038670539856, + "learning_rate": 1.0664755930291817e-05, + "loss": 0.8399, + "step": 2953 + }, + { + "epoch": 0.4952740228439694, + "grad_norm": 0.3847028911113739, + "learning_rate": 1.0659337261036824e-05, + "loss": 0.9009, + "step": 2954 + }, + { + "epoch": 0.49544168500471547, + "grad_norm": 0.6146700382232666, + "learning_rate": 1.065391839733511e-05, + "loss": 0.9276, + "step": 2955 + }, + { + "epoch": 0.4956093471654616, + "grad_norm": 0.7219106554985046, + "learning_rate": 1.0648499340784765e-05, + "loss": 0.8842, + "step": 2956 + }, + { + "epoch": 0.4957770093262077, + "grad_norm": 0.5040003657341003, + "learning_rate": 1.0643080092983933e-05, + "loss": 0.9594, + "step": 2957 + }, + { + "epoch": 0.4959446714869538, + "grad_norm": 0.4865266978740692, + "learning_rate": 1.0637660655530814e-05, + "loss": 0.8077, + "step": 2958 + }, + { + "epoch": 0.4961123336476999, + "grad_norm": 0.46663138270378113, + "learning_rate": 1.0632241030023678e-05, + "loss": 0.9022, + "step": 2959 + }, + { + "epoch": 0.496279995808446, + "grad_norm": 0.4126185178756714, + "learning_rate": 1.062682121806083e-05, + "loss": 0.8231, + "step": 2960 + }, + { + "epoch": 0.4964476579691921, + "grad_norm": 0.6498703956604004, + "learning_rate": 1.0621401221240646e-05, + "loss": 0.9935, + "step": 2961 + }, + { + "epoch": 0.4966153201299382, + "grad_norm": 0.44782713055610657, + "learning_rate": 1.061598104116154e-05, + "loss": 0.9894, + "step": 2962 + }, + { + "epoch": 0.4967829822906843, + "grad_norm": 0.5257605910301208, + "learning_rate": 1.0610560679422007e-05, + "loss": 0.8628, + "step": 2963 + }, + { + "epoch": 0.4969506444514304, + "grad_norm": 0.40351545810699463, + "learning_rate": 1.0605140137620562e-05, + "loss": 0.9404, + "step": 2964 + }, + { + "epoch": 0.4971183066121765, + "grad_norm": 0.3297235369682312, + "learning_rate": 1.0599719417355801e-05, + "loss": 0.8571, + "step": 2965 + }, + { + "epoch": 0.4972859687729226, + "grad_norm": 0.41913363337516785, + "learning_rate": 1.0594298520226354e-05, + "loss": 0.8634, + "step": 2966 + }, + { + "epoch": 0.4974536309336687, + "grad_norm": 0.7649381756782532, + "learning_rate": 1.0588877447830917e-05, + "loss": 0.8632, + "step": 2967 + }, + { + "epoch": 0.4976212930944148, + "grad_norm": 0.4903692901134491, + "learning_rate": 1.058345620176823e-05, + "loss": 1.0128, + "step": 2968 + }, + { + "epoch": 0.49778895525516087, + "grad_norm": 0.4322664439678192, + "learning_rate": 1.057803478363708e-05, + "loss": 0.9595, + "step": 2969 + }, + { + "epoch": 0.49795661741590697, + "grad_norm": 0.4703931510448456, + "learning_rate": 1.0572613195036319e-05, + "loss": 0.924, + "step": 2970 + }, + { + "epoch": 0.49812427957665306, + "grad_norm": 0.4785899817943573, + "learning_rate": 1.0567191437564832e-05, + "loss": 0.9254, + "step": 2971 + }, + { + "epoch": 0.49829194173739916, + "grad_norm": 0.7141011953353882, + "learning_rate": 1.0561769512821571e-05, + "loss": 0.8905, + "step": 2972 + }, + { + "epoch": 0.49845960389814525, + "grad_norm": 0.4536585807800293, + "learning_rate": 1.055634742240552e-05, + "loss": 0.779, + "step": 2973 + }, + { + "epoch": 0.49862726605889135, + "grad_norm": 0.478220134973526, + "learning_rate": 1.055092516791573e-05, + "loss": 0.8576, + "step": 2974 + }, + { + "epoch": 0.49879492821963745, + "grad_norm": 0.4529150128364563, + "learning_rate": 1.0545502750951282e-05, + "loss": 0.8794, + "step": 2975 + }, + { + "epoch": 0.49896259038038354, + "grad_norm": 0.5789543986320496, + "learning_rate": 1.0540080173111318e-05, + "loss": 0.9309, + "step": 2976 + }, + { + "epoch": 0.49913025254112964, + "grad_norm": 0.47327640652656555, + "learning_rate": 1.0534657435995024e-05, + "loss": 0.9094, + "step": 2977 + }, + { + "epoch": 0.49929791470187573, + "grad_norm": 0.3968491554260254, + "learning_rate": 1.0529234541201631e-05, + "loss": 0.8466, + "step": 2978 + }, + { + "epoch": 0.49946557686262183, + "grad_norm": 0.3362426459789276, + "learning_rate": 1.0523811490330422e-05, + "loss": 0.8066, + "step": 2979 + }, + { + "epoch": 0.4996332390233679, + "grad_norm": 0.4447023570537567, + "learning_rate": 1.0518388284980714e-05, + "loss": 0.8388, + "step": 2980 + }, + { + "epoch": 0.499800901184114, + "grad_norm": 0.594958484172821, + "learning_rate": 1.0512964926751882e-05, + "loss": 0.8109, + "step": 2981 + }, + { + "epoch": 0.4999685633448601, + "grad_norm": 0.49405747652053833, + "learning_rate": 1.0507541417243344e-05, + "loss": 0.9152, + "step": 2982 + }, + { + "epoch": 0.5001362255056062, + "grad_norm": 0.6013819575309753, + "learning_rate": 1.0502117758054556e-05, + "loss": 0.8891, + "step": 2983 + }, + { + "epoch": 0.5003038876663523, + "grad_norm": 0.4540571868419647, + "learning_rate": 1.0496693950785023e-05, + "loss": 0.7649, + "step": 2984 + }, + { + "epoch": 0.5004715498270984, + "grad_norm": 0.5885023474693298, + "learning_rate": 1.0491269997034292e-05, + "loss": 0.8125, + "step": 2985 + }, + { + "epoch": 0.5006392119878444, + "grad_norm": 0.5920272469520569, + "learning_rate": 1.0485845898401959e-05, + "loss": 0.8707, + "step": 2986 + }, + { + "epoch": 0.5008068741485906, + "grad_norm": 0.7279382348060608, + "learning_rate": 1.048042165648765e-05, + "loss": 1.001, + "step": 2987 + }, + { + "epoch": 0.5009745363093366, + "grad_norm": 0.4096040427684784, + "learning_rate": 1.047499727289105e-05, + "loss": 0.8671, + "step": 2988 + }, + { + "epoch": 0.5011421984700828, + "grad_norm": 0.494907408952713, + "learning_rate": 1.0469572749211869e-05, + "loss": 0.8205, + "step": 2989 + }, + { + "epoch": 0.5013098606308288, + "grad_norm": 0.39264318346977234, + "learning_rate": 1.0464148087049872e-05, + "loss": 0.9057, + "step": 2990 + }, + { + "epoch": 0.501477522791575, + "grad_norm": 0.5199148654937744, + "learning_rate": 1.0458723288004858e-05, + "loss": 0.8148, + "step": 2991 + }, + { + "epoch": 0.501645184952321, + "grad_norm": 0.49908947944641113, + "learning_rate": 1.0453298353676662e-05, + "loss": 0.9296, + "step": 2992 + }, + { + "epoch": 0.5018128471130672, + "grad_norm": 0.40984004735946655, + "learning_rate": 1.0447873285665168e-05, + "loss": 0.9216, + "step": 2993 + }, + { + "epoch": 0.5019805092738132, + "grad_norm": 0.8072329759597778, + "learning_rate": 1.04424480855703e-05, + "loss": 0.9464, + "step": 2994 + }, + { + "epoch": 0.5021481714345594, + "grad_norm": 0.47458645701408386, + "learning_rate": 1.0437022754992011e-05, + "loss": 0.9203, + "step": 2995 + }, + { + "epoch": 0.5023158335953054, + "grad_norm": 0.5706420540809631, + "learning_rate": 1.0431597295530299e-05, + "loss": 0.9139, + "step": 2996 + }, + { + "epoch": 0.5024834957560516, + "grad_norm": 0.4910517930984497, + "learning_rate": 1.0426171708785203e-05, + "loss": 0.9184, + "step": 2997 + }, + { + "epoch": 0.5026511579167976, + "grad_norm": 0.4286358058452606, + "learning_rate": 1.0420745996356792e-05, + "loss": 0.8901, + "step": 2998 + }, + { + "epoch": 0.5028188200775437, + "grad_norm": 0.39360329508781433, + "learning_rate": 1.0415320159845175e-05, + "loss": 0.8745, + "step": 2999 + }, + { + "epoch": 0.5029864822382899, + "grad_norm": 0.6680741310119629, + "learning_rate": 1.04098942008505e-05, + "loss": 0.905, + "step": 3000 + }, + { + "epoch": 0.5031541443990359, + "grad_norm": 0.631578803062439, + "learning_rate": 1.040446812097295e-05, + "loss": 0.8432, + "step": 3001 + }, + { + "epoch": 0.5033218065597821, + "grad_norm": 0.6036131381988525, + "learning_rate": 1.0399041921812745e-05, + "loss": 0.926, + "step": 3002 + }, + { + "epoch": 0.5034894687205281, + "grad_norm": 0.4073098301887512, + "learning_rate": 1.0393615604970132e-05, + "loss": 0.8286, + "step": 3003 + }, + { + "epoch": 0.5036571308812743, + "grad_norm": 0.4562547206878662, + "learning_rate": 1.0388189172045407e-05, + "loss": 0.7902, + "step": 3004 + }, + { + "epoch": 0.5038247930420203, + "grad_norm": 0.47056758403778076, + "learning_rate": 1.0382762624638885e-05, + "loss": 0.9353, + "step": 3005 + }, + { + "epoch": 0.5039924552027665, + "grad_norm": 0.43075841665267944, + "learning_rate": 1.0377335964350928e-05, + "loss": 0.7843, + "step": 3006 + }, + { + "epoch": 0.5041601173635125, + "grad_norm": 0.3730616867542267, + "learning_rate": 1.0371909192781922e-05, + "loss": 0.8009, + "step": 3007 + }, + { + "epoch": 0.5043277795242587, + "grad_norm": 0.8039329648017883, + "learning_rate": 1.0366482311532286e-05, + "loss": 0.8716, + "step": 3008 + }, + { + "epoch": 0.5044954416850047, + "grad_norm": 0.5027324557304382, + "learning_rate": 1.0361055322202482e-05, + "loss": 0.8978, + "step": 3009 + }, + { + "epoch": 0.5046631038457509, + "grad_norm": 0.5500113368034363, + "learning_rate": 1.0355628226392994e-05, + "loss": 0.8553, + "step": 3010 + }, + { + "epoch": 0.5048307660064969, + "grad_norm": 0.8844757676124573, + "learning_rate": 1.0350201025704334e-05, + "loss": 0.8621, + "step": 3011 + }, + { + "epoch": 0.504998428167243, + "grad_norm": 0.5123084187507629, + "learning_rate": 1.0344773721737058e-05, + "loss": 0.8796, + "step": 3012 + }, + { + "epoch": 0.5051660903279891, + "grad_norm": 0.5489059090614319, + "learning_rate": 1.0339346316091742e-05, + "loss": 0.911, + "step": 3013 + }, + { + "epoch": 0.5053337524887352, + "grad_norm": 0.7856314778327942, + "learning_rate": 1.0333918810368994e-05, + "loss": 0.8677, + "step": 3014 + }, + { + "epoch": 0.5055014146494813, + "grad_norm": 0.44746580719947815, + "learning_rate": 1.0328491206169456e-05, + "loss": 0.8347, + "step": 3015 + }, + { + "epoch": 0.5056690768102274, + "grad_norm": 0.7760793566703796, + "learning_rate": 1.0323063505093791e-05, + "loss": 0.8911, + "step": 3016 + }, + { + "epoch": 0.5058367389709735, + "grad_norm": 0.7316550612449646, + "learning_rate": 1.03176357087427e-05, + "loss": 0.9219, + "step": 3017 + }, + { + "epoch": 0.5060044011317196, + "grad_norm": 0.4846658706665039, + "learning_rate": 1.0312207818716905e-05, + "loss": 0.8055, + "step": 3018 + }, + { + "epoch": 0.5061720632924657, + "grad_norm": 0.5303776860237122, + "learning_rate": 1.0306779836617156e-05, + "loss": 0.8776, + "step": 3019 + }, + { + "epoch": 0.5063397254532118, + "grad_norm": 0.5125965476036072, + "learning_rate": 1.0301351764044239e-05, + "loss": 0.9409, + "step": 3020 + }, + { + "epoch": 0.5065073876139579, + "grad_norm": 0.39778846502304077, + "learning_rate": 1.0295923602598951e-05, + "loss": 0.7961, + "step": 3021 + }, + { + "epoch": 0.506675049774704, + "grad_norm": 0.42018333077430725, + "learning_rate": 1.0290495353882131e-05, + "loss": 0.8947, + "step": 3022 + }, + { + "epoch": 0.50684271193545, + "grad_norm": 0.5329704284667969, + "learning_rate": 1.0285067019494632e-05, + "loss": 0.8357, + "step": 3023 + }, + { + "epoch": 0.5070103740961962, + "grad_norm": 0.6231377720832825, + "learning_rate": 1.0279638601037338e-05, + "loss": 0.8307, + "step": 3024 + }, + { + "epoch": 0.5071780362569422, + "grad_norm": 0.5728231072425842, + "learning_rate": 1.027421010011116e-05, + "loss": 0.877, + "step": 3025 + }, + { + "epoch": 0.5073456984176884, + "grad_norm": 0.6350905895233154, + "learning_rate": 1.026878151831703e-05, + "loss": 0.9096, + "step": 3026 + }, + { + "epoch": 0.5075133605784344, + "grad_norm": 0.6212195754051208, + "learning_rate": 1.0263352857255899e-05, + "loss": 0.8157, + "step": 3027 + }, + { + "epoch": 0.5076810227391806, + "grad_norm": 0.40822339057922363, + "learning_rate": 1.0257924118528749e-05, + "loss": 0.9032, + "step": 3028 + }, + { + "epoch": 0.5078486848999266, + "grad_norm": 0.364933043718338, + "learning_rate": 1.0252495303736584e-05, + "loss": 0.8313, + "step": 3029 + }, + { + "epoch": 0.5080163470606728, + "grad_norm": 0.46916210651397705, + "learning_rate": 1.0247066414480424e-05, + "loss": 0.9472, + "step": 3030 + }, + { + "epoch": 0.5081840092214188, + "grad_norm": 0.4093315601348877, + "learning_rate": 1.0241637452361323e-05, + "loss": 0.8384, + "step": 3031 + }, + { + "epoch": 0.508351671382165, + "grad_norm": 0.7968176007270813, + "learning_rate": 1.0236208418980345e-05, + "loss": 0.857, + "step": 3032 + }, + { + "epoch": 0.508519333542911, + "grad_norm": 0.599753201007843, + "learning_rate": 1.023077931593858e-05, + "loss": 0.9746, + "step": 3033 + }, + { + "epoch": 0.5086869957036572, + "grad_norm": 0.6237933039665222, + "learning_rate": 1.0225350144837134e-05, + "loss": 0.9419, + "step": 3034 + }, + { + "epoch": 0.5088546578644032, + "grad_norm": 0.41964903473854065, + "learning_rate": 1.021992090727714e-05, + "loss": 0.8519, + "step": 3035 + }, + { + "epoch": 0.5090223200251494, + "grad_norm": 0.5063328742980957, + "learning_rate": 1.021449160485975e-05, + "loss": 0.9173, + "step": 3036 + }, + { + "epoch": 0.5091899821858954, + "grad_norm": 0.5008317828178406, + "learning_rate": 1.0209062239186127e-05, + "loss": 0.8588, + "step": 3037 + }, + { + "epoch": 0.5093576443466415, + "grad_norm": 0.3873981833457947, + "learning_rate": 1.0203632811857462e-05, + "loss": 0.85, + "step": 3038 + }, + { + "epoch": 0.5095253065073876, + "grad_norm": 0.6122221946716309, + "learning_rate": 1.019820332447496e-05, + "loss": 0.8791, + "step": 3039 + }, + { + "epoch": 0.5096929686681337, + "grad_norm": 0.5497280955314636, + "learning_rate": 1.0192773778639842e-05, + "loss": 0.8646, + "step": 3040 + }, + { + "epoch": 0.5098606308288798, + "grad_norm": 0.41970667243003845, + "learning_rate": 1.0187344175953351e-05, + "loss": 0.9264, + "step": 3041 + }, + { + "epoch": 0.5100282929896259, + "grad_norm": 0.45841696858406067, + "learning_rate": 1.018191451801674e-05, + "loss": 0.8689, + "step": 3042 + }, + { + "epoch": 0.510195955150372, + "grad_norm": 0.5520064234733582, + "learning_rate": 1.0176484806431288e-05, + "loss": 0.812, + "step": 3043 + }, + { + "epoch": 0.5103636173111181, + "grad_norm": 0.6513684988021851, + "learning_rate": 1.017105504279828e-05, + "loss": 0.8724, + "step": 3044 + }, + { + "epoch": 0.5105312794718642, + "grad_norm": 0.6604621410369873, + "learning_rate": 1.0165625228719021e-05, + "loss": 0.8985, + "step": 3045 + }, + { + "epoch": 0.5106989416326103, + "grad_norm": 0.692649781703949, + "learning_rate": 1.0160195365794832e-05, + "loss": 0.9538, + "step": 3046 + }, + { + "epoch": 0.5108666037933564, + "grad_norm": 0.5462502837181091, + "learning_rate": 1.0154765455627048e-05, + "loss": 0.901, + "step": 3047 + }, + { + "epoch": 0.5110342659541025, + "grad_norm": 0.625444769859314, + "learning_rate": 1.014933549981701e-05, + "loss": 0.9138, + "step": 3048 + }, + { + "epoch": 0.5112019281148485, + "grad_norm": 0.4703475534915924, + "learning_rate": 1.0143905499966086e-05, + "loss": 0.8508, + "step": 3049 + }, + { + "epoch": 0.5113695902755947, + "grad_norm": 0.522480309009552, + "learning_rate": 1.013847545767565e-05, + "loss": 0.8468, + "step": 3050 + }, + { + "epoch": 0.5115372524363407, + "grad_norm": 0.7132927775382996, + "learning_rate": 1.0133045374547084e-05, + "loss": 0.8493, + "step": 3051 + }, + { + "epoch": 0.5117049145970869, + "grad_norm": 0.5355094075202942, + "learning_rate": 1.0127615252181792e-05, + "loss": 0.8351, + "step": 3052 + }, + { + "epoch": 0.5118725767578329, + "grad_norm": 0.5499727129936218, + "learning_rate": 1.0122185092181178e-05, + "loss": 0.9591, + "step": 3053 + }, + { + "epoch": 0.5120402389185791, + "grad_norm": 0.5338245630264282, + "learning_rate": 1.0116754896146672e-05, + "loss": 0.8053, + "step": 3054 + }, + { + "epoch": 0.5122079010793251, + "grad_norm": 0.4411284029483795, + "learning_rate": 1.01113246656797e-05, + "loss": 0.941, + "step": 3055 + }, + { + "epoch": 0.5123755632400713, + "grad_norm": 0.4244464039802551, + "learning_rate": 1.0105894402381703e-05, + "loss": 0.822, + "step": 3056 + }, + { + "epoch": 0.5125432254008173, + "grad_norm": 0.4807767868041992, + "learning_rate": 1.0100464107854137e-05, + "loss": 0.8257, + "step": 3057 + }, + { + "epoch": 0.5127108875615635, + "grad_norm": 0.4651215672492981, + "learning_rate": 1.0095033783698462e-05, + "loss": 0.8494, + "step": 3058 + }, + { + "epoch": 0.5128785497223095, + "grad_norm": 0.5512310862541199, + "learning_rate": 1.008960343151615e-05, + "loss": 0.8645, + "step": 3059 + }, + { + "epoch": 0.5130462118830557, + "grad_norm": 0.4677385091781616, + "learning_rate": 1.0084173052908675e-05, + "loss": 0.9542, + "step": 3060 + }, + { + "epoch": 0.5132138740438017, + "grad_norm": 0.4788549542427063, + "learning_rate": 1.0078742649477522e-05, + "loss": 0.8827, + "step": 3061 + }, + { + "epoch": 0.5133815362045479, + "grad_norm": 0.5147599577903748, + "learning_rate": 1.0073312222824187e-05, + "loss": 0.8223, + "step": 3062 + }, + { + "epoch": 0.5135491983652939, + "grad_norm": 0.47588101029396057, + "learning_rate": 1.0067881774550175e-05, + "loss": 0.9135, + "step": 3063 + }, + { + "epoch": 0.51371686052604, + "grad_norm": 0.5515173673629761, + "learning_rate": 1.0062451306256981e-05, + "loss": 0.8483, + "step": 3064 + }, + { + "epoch": 0.5138845226867861, + "grad_norm": 0.6544636487960815, + "learning_rate": 1.0057020819546122e-05, + "loss": 0.8364, + "step": 3065 + }, + { + "epoch": 0.5140521848475322, + "grad_norm": 0.8613523244857788, + "learning_rate": 1.0051590316019122e-05, + "loss": 0.8428, + "step": 3066 + }, + { + "epoch": 0.5142198470082783, + "grad_norm": 0.39275920391082764, + "learning_rate": 1.0046159797277497e-05, + "loss": 0.8907, + "step": 3067 + }, + { + "epoch": 0.5143875091690244, + "grad_norm": 0.5242758393287659, + "learning_rate": 1.0040729264922778e-05, + "loss": 0.8082, + "step": 3068 + }, + { + "epoch": 0.5145551713297705, + "grad_norm": 0.6845324635505676, + "learning_rate": 1.0035298720556493e-05, + "loss": 0.911, + "step": 3069 + }, + { + "epoch": 0.5147228334905166, + "grad_norm": 0.5309914946556091, + "learning_rate": 1.0029868165780178e-05, + "loss": 0.9313, + "step": 3070 + }, + { + "epoch": 0.5148904956512627, + "grad_norm": 0.4313521087169647, + "learning_rate": 1.0024437602195371e-05, + "loss": 0.8413, + "step": 3071 + }, + { + "epoch": 0.5150581578120088, + "grad_norm": 0.4510025084018707, + "learning_rate": 1.0019007031403614e-05, + "loss": 0.8322, + "step": 3072 + }, + { + "epoch": 0.5152258199727549, + "grad_norm": 0.6852996945381165, + "learning_rate": 1.0013576455006446e-05, + "loss": 0.9159, + "step": 3073 + }, + { + "epoch": 0.515393482133501, + "grad_norm": 0.5756083726882935, + "learning_rate": 1.0008145874605414e-05, + "loss": 0.9771, + "step": 3074 + }, + { + "epoch": 0.515561144294247, + "grad_norm": 0.8508740067481995, + "learning_rate": 1.0002715291802065e-05, + "loss": 0.8343, + "step": 3075 + }, + { + "epoch": 0.5157288064549932, + "grad_norm": 0.42136839032173157, + "learning_rate": 9.997284708197938e-06, + "loss": 0.8462, + "step": 3076 + }, + { + "epoch": 0.5158964686157392, + "grad_norm": 0.6306905746459961, + "learning_rate": 9.991854125394587e-06, + "loss": 0.9103, + "step": 3077 + }, + { + "epoch": 0.5160641307764854, + "grad_norm": 0.5494988560676575, + "learning_rate": 9.986423544993559e-06, + "loss": 0.9001, + "step": 3078 + }, + { + "epoch": 0.5162317929372314, + "grad_norm": 0.47950515151023865, + "learning_rate": 9.98099296859639e-06, + "loss": 0.8205, + "step": 3079 + }, + { + "epoch": 0.5163994550979776, + "grad_norm": 0.5724183917045593, + "learning_rate": 9.975562397804632e-06, + "loss": 0.8277, + "step": 3080 + }, + { + "epoch": 0.5165671172587236, + "grad_norm": 0.29128915071487427, + "learning_rate": 9.970131834219823e-06, + "loss": 0.8813, + "step": 3081 + }, + { + "epoch": 0.5167347794194698, + "grad_norm": 0.5448590517044067, + "learning_rate": 9.964701279443509e-06, + "loss": 0.8322, + "step": 3082 + }, + { + "epoch": 0.5169024415802158, + "grad_norm": 0.40740934014320374, + "learning_rate": 9.959270735077226e-06, + "loss": 0.8544, + "step": 3083 + }, + { + "epoch": 0.517070103740962, + "grad_norm": 0.5573588013648987, + "learning_rate": 9.953840202722505e-06, + "loss": 0.9403, + "step": 3084 + }, + { + "epoch": 0.5172377659017081, + "grad_norm": 0.48960620164871216, + "learning_rate": 9.94840968398088e-06, + "loss": 0.904, + "step": 3085 + }, + { + "epoch": 0.5174054280624542, + "grad_norm": 0.4078063368797302, + "learning_rate": 9.942979180453878e-06, + "loss": 0.9525, + "step": 3086 + }, + { + "epoch": 0.5175730902232003, + "grad_norm": 0.46629518270492554, + "learning_rate": 9.937548693743024e-06, + "loss": 0.8169, + "step": 3087 + }, + { + "epoch": 0.5177407523839463, + "grad_norm": 0.5462809801101685, + "learning_rate": 9.932118225449832e-06, + "loss": 0.8677, + "step": 3088 + }, + { + "epoch": 0.5179084145446925, + "grad_norm": 0.3972527086734772, + "learning_rate": 9.926687777175815e-06, + "loss": 0.8871, + "step": 3089 + }, + { + "epoch": 0.5180760767054385, + "grad_norm": 0.38695618510246277, + "learning_rate": 9.921257350522481e-06, + "loss": 0.8438, + "step": 3090 + }, + { + "epoch": 0.5182437388661847, + "grad_norm": 0.41372543573379517, + "learning_rate": 9.915826947091328e-06, + "loss": 0.787, + "step": 3091 + }, + { + "epoch": 0.5184114010269307, + "grad_norm": 0.39852261543273926, + "learning_rate": 9.910396568483855e-06, + "loss": 0.8882, + "step": 3092 + }, + { + "epoch": 0.5185790631876769, + "grad_norm": 0.6337873935699463, + "learning_rate": 9.90496621630154e-06, + "loss": 0.9105, + "step": 3093 + }, + { + "epoch": 0.5187467253484229, + "grad_norm": 0.4452924430370331, + "learning_rate": 9.899535892145864e-06, + "loss": 0.8428, + "step": 3094 + }, + { + "epoch": 0.5189143875091691, + "grad_norm": 0.46548423171043396, + "learning_rate": 9.894105597618297e-06, + "loss": 0.8208, + "step": 3095 + }, + { + "epoch": 0.5190820496699151, + "grad_norm": 0.6204050779342651, + "learning_rate": 9.888675334320302e-06, + "loss": 0.8098, + "step": 3096 + }, + { + "epoch": 0.5192497118306613, + "grad_norm": 0.6824650168418884, + "learning_rate": 9.883245103853333e-06, + "loss": 0.9182, + "step": 3097 + }, + { + "epoch": 0.5194173739914073, + "grad_norm": 0.7250338792800903, + "learning_rate": 9.877814907818824e-06, + "loss": 0.9327, + "step": 3098 + }, + { + "epoch": 0.5195850361521535, + "grad_norm": 0.6577723622322083, + "learning_rate": 9.87238474781821e-06, + "loss": 0.8936, + "step": 3099 + }, + { + "epoch": 0.5197526983128995, + "grad_norm": 0.40129998326301575, + "learning_rate": 9.866954625452918e-06, + "loss": 0.8641, + "step": 3100 + }, + { + "epoch": 0.5199203604736456, + "grad_norm": 0.36280500888824463, + "learning_rate": 9.861524542324355e-06, + "loss": 0.8502, + "step": 3101 + }, + { + "epoch": 0.5200880226343917, + "grad_norm": 0.44010069966316223, + "learning_rate": 9.856094500033916e-06, + "loss": 0.8933, + "step": 3102 + }, + { + "epoch": 0.5202556847951378, + "grad_norm": 0.37167131900787354, + "learning_rate": 9.850664500182991e-06, + "loss": 0.8611, + "step": 3103 + }, + { + "epoch": 0.5204233469558839, + "grad_norm": 0.6913161277770996, + "learning_rate": 9.845234544372956e-06, + "loss": 0.9099, + "step": 3104 + }, + { + "epoch": 0.52059100911663, + "grad_norm": 0.506624162197113, + "learning_rate": 9.839804634205168e-06, + "loss": 0.973, + "step": 3105 + }, + { + "epoch": 0.5207586712773761, + "grad_norm": 0.3480541408061981, + "learning_rate": 9.834374771280982e-06, + "loss": 0.8869, + "step": 3106 + }, + { + "epoch": 0.5209263334381222, + "grad_norm": 0.40910738706588745, + "learning_rate": 9.828944957201724e-06, + "loss": 0.8806, + "step": 3107 + }, + { + "epoch": 0.5210939955988683, + "grad_norm": 0.47690507769584656, + "learning_rate": 9.823515193568715e-06, + "loss": 0.818, + "step": 3108 + }, + { + "epoch": 0.5212616577596144, + "grad_norm": 0.6599212288856506, + "learning_rate": 9.818085481983262e-06, + "loss": 0.9207, + "step": 3109 + }, + { + "epoch": 0.5214293199203605, + "grad_norm": 0.47169163823127747, + "learning_rate": 9.812655824046654e-06, + "loss": 0.8209, + "step": 3110 + }, + { + "epoch": 0.5215969820811066, + "grad_norm": 0.5017246603965759, + "learning_rate": 9.807226221360163e-06, + "loss": 0.8716, + "step": 3111 + }, + { + "epoch": 0.5217646442418526, + "grad_norm": 0.38969993591308594, + "learning_rate": 9.801796675525044e-06, + "loss": 0.8849, + "step": 3112 + }, + { + "epoch": 0.5219323064025988, + "grad_norm": 0.6319521069526672, + "learning_rate": 9.796367188142541e-06, + "loss": 0.8002, + "step": 3113 + }, + { + "epoch": 0.5220999685633448, + "grad_norm": 0.5612240433692932, + "learning_rate": 9.790937760813874e-06, + "loss": 0.8045, + "step": 3114 + }, + { + "epoch": 0.522267630724091, + "grad_norm": 0.5317338705062866, + "learning_rate": 9.785508395140255e-06, + "loss": 0.8176, + "step": 3115 + }, + { + "epoch": 0.522435292884837, + "grad_norm": 0.4771307706832886, + "learning_rate": 9.780079092722863e-06, + "loss": 0.8691, + "step": 3116 + }, + { + "epoch": 0.5226029550455832, + "grad_norm": 0.8237624764442444, + "learning_rate": 9.77464985516287e-06, + "loss": 0.8767, + "step": 3117 + }, + { + "epoch": 0.5227706172063292, + "grad_norm": 0.611888587474823, + "learning_rate": 9.769220684061422e-06, + "loss": 0.8261, + "step": 3118 + }, + { + "epoch": 0.5229382793670754, + "grad_norm": 0.7214562296867371, + "learning_rate": 9.76379158101966e-06, + "loss": 0.8596, + "step": 3119 + }, + { + "epoch": 0.5231059415278214, + "grad_norm": 0.6046148538589478, + "learning_rate": 9.75836254763868e-06, + "loss": 0.9288, + "step": 3120 + }, + { + "epoch": 0.5232736036885676, + "grad_norm": 0.7193319797515869, + "learning_rate": 9.752933585519578e-06, + "loss": 0.9434, + "step": 3121 + }, + { + "epoch": 0.5234412658493136, + "grad_norm": 0.539698600769043, + "learning_rate": 9.747504696263417e-06, + "loss": 0.8804, + "step": 3122 + }, + { + "epoch": 0.5236089280100598, + "grad_norm": 0.5887028574943542, + "learning_rate": 9.742075881471251e-06, + "loss": 0.9242, + "step": 3123 + }, + { + "epoch": 0.5237765901708058, + "grad_norm": 0.6475830674171448, + "learning_rate": 9.736647142744106e-06, + "loss": 0.8669, + "step": 3124 + }, + { + "epoch": 0.523944252331552, + "grad_norm": 0.5839582681655884, + "learning_rate": 9.731218481682973e-06, + "loss": 0.8992, + "step": 3125 + }, + { + "epoch": 0.524111914492298, + "grad_norm": 0.6649941802024841, + "learning_rate": 9.72578989988884e-06, + "loss": 0.8331, + "step": 3126 + }, + { + "epoch": 0.5242795766530441, + "grad_norm": 0.5015733242034912, + "learning_rate": 9.720361398962661e-06, + "loss": 0.8447, + "step": 3127 + }, + { + "epoch": 0.5244472388137902, + "grad_norm": 0.45080992579460144, + "learning_rate": 9.714932980505372e-06, + "loss": 0.8166, + "step": 3128 + }, + { + "epoch": 0.5246149009745363, + "grad_norm": 0.5960597395896912, + "learning_rate": 9.709504646117874e-06, + "loss": 0.8761, + "step": 3129 + }, + { + "epoch": 0.5247825631352824, + "grad_norm": 0.5214238166809082, + "learning_rate": 9.70407639740105e-06, + "loss": 0.8754, + "step": 3130 + }, + { + "epoch": 0.5249502252960285, + "grad_norm": 0.418317049741745, + "learning_rate": 9.698648235955764e-06, + "loss": 0.9651, + "step": 3131 + }, + { + "epoch": 0.5251178874567746, + "grad_norm": 0.40980616211891174, + "learning_rate": 9.693220163382844e-06, + "loss": 0.8648, + "step": 3132 + }, + { + "epoch": 0.5252855496175207, + "grad_norm": 0.5039193034172058, + "learning_rate": 9.687792181283098e-06, + "loss": 0.768, + "step": 3133 + }, + { + "epoch": 0.5254532117782668, + "grad_norm": 0.41915470361709595, + "learning_rate": 9.682364291257304e-06, + "loss": 0.8895, + "step": 3134 + }, + { + "epoch": 0.5256208739390129, + "grad_norm": 0.5267002582550049, + "learning_rate": 9.67693649490621e-06, + "loss": 0.7787, + "step": 3135 + }, + { + "epoch": 0.525788536099759, + "grad_norm": 0.43035128712654114, + "learning_rate": 9.671508793830548e-06, + "loss": 0.8968, + "step": 3136 + }, + { + "epoch": 0.5259561982605051, + "grad_norm": 0.4039972722530365, + "learning_rate": 9.666081189631007e-06, + "loss": 0.876, + "step": 3137 + }, + { + "epoch": 0.5261238604212511, + "grad_norm": 0.392192006111145, + "learning_rate": 9.660653683908263e-06, + "loss": 0.9633, + "step": 3138 + }, + { + "epoch": 0.5262915225819973, + "grad_norm": 0.4463840126991272, + "learning_rate": 9.655226278262945e-06, + "loss": 0.8594, + "step": 3139 + }, + { + "epoch": 0.5264591847427433, + "grad_norm": 0.8318145275115967, + "learning_rate": 9.649798974295668e-06, + "loss": 0.9015, + "step": 3140 + }, + { + "epoch": 0.5266268469034895, + "grad_norm": 0.42815402150154114, + "learning_rate": 9.644371773607008e-06, + "loss": 0.8858, + "step": 3141 + }, + { + "epoch": 0.5267945090642355, + "grad_norm": 0.41154927015304565, + "learning_rate": 9.638944677797522e-06, + "loss": 0.8129, + "step": 3142 + }, + { + "epoch": 0.5269621712249817, + "grad_norm": 1.375307559967041, + "learning_rate": 9.633517688467717e-06, + "loss": 0.8723, + "step": 3143 + }, + { + "epoch": 0.5271298333857277, + "grad_norm": 0.8266283869743347, + "learning_rate": 9.628090807218082e-06, + "loss": 0.9643, + "step": 3144 + }, + { + "epoch": 0.5272974955464739, + "grad_norm": 0.43826937675476074, + "learning_rate": 9.622664035649074e-06, + "loss": 1.0305, + "step": 3145 + }, + { + "epoch": 0.5274651577072199, + "grad_norm": 0.48454275727272034, + "learning_rate": 9.617237375361115e-06, + "loss": 0.892, + "step": 3146 + }, + { + "epoch": 0.5276328198679661, + "grad_norm": 0.7665025591850281, + "learning_rate": 9.6118108279546e-06, + "loss": 0.8997, + "step": 3147 + }, + { + "epoch": 0.5278004820287121, + "grad_norm": 0.5407666563987732, + "learning_rate": 9.60638439502987e-06, + "loss": 0.829, + "step": 3148 + }, + { + "epoch": 0.5279681441894583, + "grad_norm": 0.503190815448761, + "learning_rate": 9.600958078187258e-06, + "loss": 0.8621, + "step": 3149 + }, + { + "epoch": 0.5281358063502043, + "grad_norm": 0.5572085380554199, + "learning_rate": 9.595531879027052e-06, + "loss": 0.8968, + "step": 3150 + }, + { + "epoch": 0.5283034685109504, + "grad_norm": 0.42003175616264343, + "learning_rate": 9.590105799149502e-06, + "loss": 0.8576, + "step": 3151 + }, + { + "epoch": 0.5284711306716965, + "grad_norm": 0.44636985659599304, + "learning_rate": 9.584679840154829e-06, + "loss": 0.8329, + "step": 3152 + }, + { + "epoch": 0.5286387928324426, + "grad_norm": 0.5694844722747803, + "learning_rate": 9.579254003643213e-06, + "loss": 0.8488, + "step": 3153 + }, + { + "epoch": 0.5288064549931887, + "grad_norm": 0.5580739378929138, + "learning_rate": 9.573828291214799e-06, + "loss": 0.9743, + "step": 3154 + }, + { + "epoch": 0.5289741171539348, + "grad_norm": 0.4939458966255188, + "learning_rate": 9.568402704469701e-06, + "loss": 0.8983, + "step": 3155 + }, + { + "epoch": 0.5291417793146809, + "grad_norm": 0.7728688716888428, + "learning_rate": 9.562977245007994e-06, + "loss": 0.8119, + "step": 3156 + }, + { + "epoch": 0.529309441475427, + "grad_norm": 0.5365317463874817, + "learning_rate": 9.557551914429703e-06, + "loss": 0.8443, + "step": 3157 + }, + { + "epoch": 0.5294771036361731, + "grad_norm": 0.5773930549621582, + "learning_rate": 9.552126714334834e-06, + "loss": 0.8962, + "step": 3158 + }, + { + "epoch": 0.5296447657969192, + "grad_norm": 0.44260895252227783, + "learning_rate": 9.546701646323341e-06, + "loss": 0.8515, + "step": 3159 + }, + { + "epoch": 0.5298124279576653, + "grad_norm": 0.5546987652778625, + "learning_rate": 9.541276711995149e-06, + "loss": 0.9005, + "step": 3160 + }, + { + "epoch": 0.5299800901184114, + "grad_norm": 0.5294981598854065, + "learning_rate": 9.535851912950133e-06, + "loss": 0.8382, + "step": 3161 + }, + { + "epoch": 0.5301477522791574, + "grad_norm": 0.5314891934394836, + "learning_rate": 9.530427250788134e-06, + "loss": 0.8224, + "step": 3162 + }, + { + "epoch": 0.5303154144399036, + "grad_norm": 0.4622967839241028, + "learning_rate": 9.525002727108954e-06, + "loss": 0.8289, + "step": 3163 + }, + { + "epoch": 0.5304830766006496, + "grad_norm": 0.37820157408714294, + "learning_rate": 9.51957834351235e-06, + "loss": 0.8796, + "step": 3164 + }, + { + "epoch": 0.5306507387613958, + "grad_norm": 0.3496609330177307, + "learning_rate": 9.514154101598046e-06, + "loss": 0.8418, + "step": 3165 + }, + { + "epoch": 0.5308184009221418, + "grad_norm": 1.156503438949585, + "learning_rate": 9.508730002965711e-06, + "loss": 0.9668, + "step": 3166 + }, + { + "epoch": 0.530986063082888, + "grad_norm": 0.39240220189094543, + "learning_rate": 9.50330604921498e-06, + "loss": 0.9064, + "step": 3167 + }, + { + "epoch": 0.531153725243634, + "grad_norm": 0.4107801616191864, + "learning_rate": 9.497882241945446e-06, + "loss": 0.7483, + "step": 3168 + }, + { + "epoch": 0.5313213874043802, + "grad_norm": 0.3688758313655853, + "learning_rate": 9.492458582756658e-06, + "loss": 0.9135, + "step": 3169 + }, + { + "epoch": 0.5314890495651263, + "grad_norm": 0.5714778900146484, + "learning_rate": 9.487035073248123e-06, + "loss": 0.8872, + "step": 3170 + }, + { + "epoch": 0.5316567117258724, + "grad_norm": 0.6979575157165527, + "learning_rate": 9.481611715019287e-06, + "loss": 0.8417, + "step": 3171 + }, + { + "epoch": 0.5318243738866185, + "grad_norm": 0.41339144110679626, + "learning_rate": 9.476188509669582e-06, + "loss": 0.8654, + "step": 3172 + }, + { + "epoch": 0.5319920360473646, + "grad_norm": 0.3709242045879364, + "learning_rate": 9.470765458798369e-06, + "loss": 0.8673, + "step": 3173 + }, + { + "epoch": 0.5321596982081107, + "grad_norm": 0.5709183216094971, + "learning_rate": 9.46534256400498e-06, + "loss": 0.9746, + "step": 3174 + }, + { + "epoch": 0.5323273603688568, + "grad_norm": 0.9141378998756409, + "learning_rate": 9.459919826888685e-06, + "loss": 0.861, + "step": 3175 + }, + { + "epoch": 0.5324950225296029, + "grad_norm": 0.41457533836364746, + "learning_rate": 9.454497249048723e-06, + "loss": 0.8245, + "step": 3176 + }, + { + "epoch": 0.5326626846903489, + "grad_norm": 0.48134270310401917, + "learning_rate": 9.449074832084274e-06, + "loss": 0.8726, + "step": 3177 + }, + { + "epoch": 0.5328303468510951, + "grad_norm": 0.7749919891357422, + "learning_rate": 9.443652577594482e-06, + "loss": 0.7487, + "step": 3178 + }, + { + "epoch": 0.5329980090118411, + "grad_norm": 0.4469508230686188, + "learning_rate": 9.438230487178434e-06, + "loss": 0.7934, + "step": 3179 + }, + { + "epoch": 0.5331656711725873, + "grad_norm": 0.4624178111553192, + "learning_rate": 9.43280856243517e-06, + "loss": 0.9039, + "step": 3180 + }, + { + "epoch": 0.5333333333333333, + "grad_norm": 0.74716717004776, + "learning_rate": 9.427386804963684e-06, + "loss": 0.8049, + "step": 3181 + }, + { + "epoch": 0.5335009954940795, + "grad_norm": 0.8879395723342896, + "learning_rate": 9.421965216362921e-06, + "loss": 0.8456, + "step": 3182 + }, + { + "epoch": 0.5336686576548255, + "grad_norm": 0.3464726507663727, + "learning_rate": 9.416543798231777e-06, + "loss": 0.8753, + "step": 3183 + }, + { + "epoch": 0.5338363198155717, + "grad_norm": 0.3298185467720032, + "learning_rate": 9.411122552169086e-06, + "loss": 0.8975, + "step": 3184 + }, + { + "epoch": 0.5340039819763177, + "grad_norm": 0.5450412034988403, + "learning_rate": 9.405701479773649e-06, + "loss": 0.839, + "step": 3185 + }, + { + "epoch": 0.5341716441370639, + "grad_norm": 0.4394127428531647, + "learning_rate": 9.400280582644204e-06, + "loss": 0.8743, + "step": 3186 + }, + { + "epoch": 0.5343393062978099, + "grad_norm": 0.4711911976337433, + "learning_rate": 9.394859862379438e-06, + "loss": 0.8515, + "step": 3187 + }, + { + "epoch": 0.534506968458556, + "grad_norm": 0.5024020075798035, + "learning_rate": 9.389439320578e-06, + "loss": 0.9068, + "step": 3188 + }, + { + "epoch": 0.5346746306193021, + "grad_norm": 0.5180482864379883, + "learning_rate": 9.384018958838462e-06, + "loss": 0.884, + "step": 3189 + }, + { + "epoch": 0.5348422927800482, + "grad_norm": 0.3648953437805176, + "learning_rate": 9.378598778759357e-06, + "loss": 0.7798, + "step": 3190 + }, + { + "epoch": 0.5350099549407943, + "grad_norm": 0.503474235534668, + "learning_rate": 9.373178781939172e-06, + "loss": 0.8051, + "step": 3191 + }, + { + "epoch": 0.5351776171015404, + "grad_norm": 0.393228679895401, + "learning_rate": 9.367758969976327e-06, + "loss": 0.8734, + "step": 3192 + }, + { + "epoch": 0.5353452792622865, + "grad_norm": 0.4325978755950928, + "learning_rate": 9.362339344469189e-06, + "loss": 0.8687, + "step": 3193 + }, + { + "epoch": 0.5355129414230326, + "grad_norm": 0.5653034448623657, + "learning_rate": 9.35691990701607e-06, + "loss": 0.8447, + "step": 3194 + }, + { + "epoch": 0.5356806035837787, + "grad_norm": 0.4211442172527313, + "learning_rate": 9.351500659215238e-06, + "loss": 1.029, + "step": 3195 + }, + { + "epoch": 0.5358482657445248, + "grad_norm": 0.5952498912811279, + "learning_rate": 9.34608160266489e-06, + "loss": 0.9139, + "step": 3196 + }, + { + "epoch": 0.5360159279052709, + "grad_norm": 0.4283948838710785, + "learning_rate": 9.34066273896318e-06, + "loss": 0.8884, + "step": 3197 + }, + { + "epoch": 0.536183590066017, + "grad_norm": 0.4888134300708771, + "learning_rate": 9.335244069708188e-06, + "loss": 0.8698, + "step": 3198 + }, + { + "epoch": 0.536351252226763, + "grad_norm": 0.6299174427986145, + "learning_rate": 9.32982559649795e-06, + "loss": 0.8223, + "step": 3199 + }, + { + "epoch": 0.5365189143875092, + "grad_norm": 0.5382765531539917, + "learning_rate": 9.324407320930444e-06, + "loss": 0.8012, + "step": 3200 + }, + { + "epoch": 0.5366865765482552, + "grad_norm": 1.366763710975647, + "learning_rate": 9.318989244603586e-06, + "loss": 0.8934, + "step": 3201 + }, + { + "epoch": 0.5368542387090014, + "grad_norm": 0.5220536589622498, + "learning_rate": 9.313571369115231e-06, + "loss": 0.8822, + "step": 3202 + }, + { + "epoch": 0.5370219008697474, + "grad_norm": 0.5266855359077454, + "learning_rate": 9.30815369606318e-06, + "loss": 0.8345, + "step": 3203 + }, + { + "epoch": 0.5371895630304936, + "grad_norm": 0.5697585940361023, + "learning_rate": 9.30273622704517e-06, + "loss": 0.7852, + "step": 3204 + }, + { + "epoch": 0.5373572251912396, + "grad_norm": 0.5060375928878784, + "learning_rate": 9.297318963658882e-06, + "loss": 0.8918, + "step": 3205 + }, + { + "epoch": 0.5375248873519858, + "grad_norm": 0.3974671959877014, + "learning_rate": 9.291901907501934e-06, + "loss": 0.8648, + "step": 3206 + }, + { + "epoch": 0.5376925495127318, + "grad_norm": 0.36046454310417175, + "learning_rate": 9.286485060171881e-06, + "loss": 0.8569, + "step": 3207 + }, + { + "epoch": 0.537860211673478, + "grad_norm": 0.47915321588516235, + "learning_rate": 9.281068423266217e-06, + "loss": 0.9252, + "step": 3208 + }, + { + "epoch": 0.538027873834224, + "grad_norm": 0.44521069526672363, + "learning_rate": 9.275651998382377e-06, + "loss": 0.8365, + "step": 3209 + }, + { + "epoch": 0.5381955359949702, + "grad_norm": 1.0060954093933105, + "learning_rate": 9.270235787117732e-06, + "loss": 0.8021, + "step": 3210 + }, + { + "epoch": 0.5383631981557162, + "grad_norm": 0.6790063381195068, + "learning_rate": 9.264819791069596e-06, + "loss": 0.8377, + "step": 3211 + }, + { + "epoch": 0.5385308603164624, + "grad_norm": 0.45293375849723816, + "learning_rate": 9.259404011835203e-06, + "loss": 0.8844, + "step": 3212 + }, + { + "epoch": 0.5386985224772084, + "grad_norm": 0.37161368131637573, + "learning_rate": 9.253988451011734e-06, + "loss": 0.8102, + "step": 3213 + }, + { + "epoch": 0.5388661846379545, + "grad_norm": 0.5089200735092163, + "learning_rate": 9.24857311019631e-06, + "loss": 0.8451, + "step": 3214 + }, + { + "epoch": 0.5390338467987006, + "grad_norm": 0.4144771695137024, + "learning_rate": 9.243157990985989e-06, + "loss": 0.8866, + "step": 3215 + }, + { + "epoch": 0.5392015089594467, + "grad_norm": 0.48967111110687256, + "learning_rate": 9.237743094977737e-06, + "loss": 0.9099, + "step": 3216 + }, + { + "epoch": 0.5393691711201928, + "grad_norm": 0.4924265444278717, + "learning_rate": 9.232328423768491e-06, + "loss": 0.8574, + "step": 3217 + }, + { + "epoch": 0.5395368332809389, + "grad_norm": 0.6180545687675476, + "learning_rate": 9.226913978955097e-06, + "loss": 0.892, + "step": 3218 + }, + { + "epoch": 0.539704495441685, + "grad_norm": 1.3185502290725708, + "learning_rate": 9.221499762134344e-06, + "loss": 0.8855, + "step": 3219 + }, + { + "epoch": 0.5398721576024311, + "grad_norm": 0.36848878860473633, + "learning_rate": 9.216085774902952e-06, + "loss": 0.7998, + "step": 3220 + }, + { + "epoch": 0.5400398197631772, + "grad_norm": 0.5853696465492249, + "learning_rate": 9.210672018857572e-06, + "loss": 0.9138, + "step": 3221 + }, + { + "epoch": 0.5402074819239233, + "grad_norm": 0.46801823377609253, + "learning_rate": 9.205258495594786e-06, + "loss": 0.9831, + "step": 3222 + }, + { + "epoch": 0.5403751440846694, + "grad_norm": 0.8978372812271118, + "learning_rate": 9.199845206711112e-06, + "loss": 0.8875, + "step": 3223 + }, + { + "epoch": 0.5405428062454155, + "grad_norm": 0.6240953207015991, + "learning_rate": 9.194432153802996e-06, + "loss": 0.8235, + "step": 3224 + }, + { + "epoch": 0.5407104684061615, + "grad_norm": 0.7486779689788818, + "learning_rate": 9.189019338466812e-06, + "loss": 0.8672, + "step": 3225 + }, + { + "epoch": 0.5408781305669077, + "grad_norm": 0.4113537669181824, + "learning_rate": 9.183606762298867e-06, + "loss": 0.8666, + "step": 3226 + }, + { + "epoch": 0.5410457927276537, + "grad_norm": 0.4346449673175812, + "learning_rate": 9.178194426895396e-06, + "loss": 0.8889, + "step": 3227 + }, + { + "epoch": 0.5412134548883999, + "grad_norm": 0.6154700517654419, + "learning_rate": 9.172782333852566e-06, + "loss": 0.9337, + "step": 3228 + }, + { + "epoch": 0.5413811170491459, + "grad_norm": 0.9208604693412781, + "learning_rate": 9.16737048476647e-06, + "loss": 0.9082, + "step": 3229 + }, + { + "epoch": 0.5415487792098921, + "grad_norm": 0.72853684425354, + "learning_rate": 9.161958881233129e-06, + "loss": 0.9172, + "step": 3230 + }, + { + "epoch": 0.5417164413706381, + "grad_norm": 0.8726669549942017, + "learning_rate": 9.156547524848489e-06, + "loss": 0.7786, + "step": 3231 + }, + { + "epoch": 0.5418841035313843, + "grad_norm": 0.39041054248809814, + "learning_rate": 9.151136417208424e-06, + "loss": 0.9218, + "step": 3232 + }, + { + "epoch": 0.5420517656921303, + "grad_norm": 0.3628639280796051, + "learning_rate": 9.14572555990875e-06, + "loss": 0.8328, + "step": 3233 + }, + { + "epoch": 0.5422194278528765, + "grad_norm": 0.39143988490104675, + "learning_rate": 9.140314954545184e-06, + "loss": 0.874, + "step": 3234 + }, + { + "epoch": 0.5423870900136225, + "grad_norm": 0.8233510255813599, + "learning_rate": 9.13490460271338e-06, + "loss": 0.7631, + "step": 3235 + }, + { + "epoch": 0.5425547521743687, + "grad_norm": 0.5175696611404419, + "learning_rate": 9.12949450600892e-06, + "loss": 0.8708, + "step": 3236 + }, + { + "epoch": 0.5427224143351147, + "grad_norm": 0.4261394441127777, + "learning_rate": 9.124084666027311e-06, + "loss": 0.8978, + "step": 3237 + }, + { + "epoch": 0.5428900764958609, + "grad_norm": 0.48911020159721375, + "learning_rate": 9.118675084363986e-06, + "loss": 0.8587, + "step": 3238 + }, + { + "epoch": 0.5430577386566069, + "grad_norm": 0.4675133526325226, + "learning_rate": 9.113265762614285e-06, + "loss": 0.8338, + "step": 3239 + }, + { + "epoch": 0.543225400817353, + "grad_norm": 0.7315004467964172, + "learning_rate": 9.107856702373492e-06, + "loss": 0.8844, + "step": 3240 + }, + { + "epoch": 0.5433930629780991, + "grad_norm": 0.41969773173332214, + "learning_rate": 9.102447905236806e-06, + "loss": 0.8963, + "step": 3241 + }, + { + "epoch": 0.5435607251388452, + "grad_norm": 0.6964146494865417, + "learning_rate": 9.097039372799345e-06, + "loss": 0.7791, + "step": 3242 + }, + { + "epoch": 0.5437283872995913, + "grad_norm": 0.6561281681060791, + "learning_rate": 9.091631106656159e-06, + "loss": 0.8957, + "step": 3243 + }, + { + "epoch": 0.5438960494603374, + "grad_norm": 0.4261462986469269, + "learning_rate": 9.086223108402203e-06, + "loss": 0.7785, + "step": 3244 + }, + { + "epoch": 0.5440637116210835, + "grad_norm": 0.7788665890693665, + "learning_rate": 9.08081537963237e-06, + "loss": 0.8151, + "step": 3245 + }, + { + "epoch": 0.5442313737818296, + "grad_norm": 0.44058895111083984, + "learning_rate": 9.075407921941463e-06, + "loss": 0.8303, + "step": 3246 + }, + { + "epoch": 0.5443990359425757, + "grad_norm": 0.7544462084770203, + "learning_rate": 9.07000073692421e-06, + "loss": 0.8711, + "step": 3247 + }, + { + "epoch": 0.5445666981033218, + "grad_norm": 0.39317142963409424, + "learning_rate": 9.064593826175254e-06, + "loss": 0.7691, + "step": 3248 + }, + { + "epoch": 0.5447343602640679, + "grad_norm": 0.48124146461486816, + "learning_rate": 9.059187191289161e-06, + "loss": 0.8287, + "step": 3249 + }, + { + "epoch": 0.544902022424814, + "grad_norm": 0.4958220422267914, + "learning_rate": 9.053780833860416e-06, + "loss": 0.8171, + "step": 3250 + }, + { + "epoch": 0.54506968458556, + "grad_norm": 0.5220127701759338, + "learning_rate": 9.04837475548342e-06, + "loss": 0.857, + "step": 3251 + }, + { + "epoch": 0.5452373467463062, + "grad_norm": 0.37427911162376404, + "learning_rate": 9.042968957752493e-06, + "loss": 0.7819, + "step": 3252 + }, + { + "epoch": 0.5454050089070522, + "grad_norm": 0.3906075358390808, + "learning_rate": 9.037563442261872e-06, + "loss": 0.8551, + "step": 3253 + }, + { + "epoch": 0.5455726710677984, + "grad_norm": 0.3960884213447571, + "learning_rate": 9.032158210605708e-06, + "loss": 0.8849, + "step": 3254 + }, + { + "epoch": 0.5457403332285444, + "grad_norm": 0.651520311832428, + "learning_rate": 9.02675326437807e-06, + "loss": 0.8351, + "step": 3255 + }, + { + "epoch": 0.5459079953892906, + "grad_norm": 0.6072437763214111, + "learning_rate": 9.021348605172951e-06, + "loss": 0.9324, + "step": 3256 + }, + { + "epoch": 0.5460756575500367, + "grad_norm": 0.39354023337364197, + "learning_rate": 9.015944234584245e-06, + "loss": 0.8851, + "step": 3257 + }, + { + "epoch": 0.5462433197107828, + "grad_norm": 1.129841923713684, + "learning_rate": 9.010540154205767e-06, + "loss": 0.9725, + "step": 3258 + }, + { + "epoch": 0.5464109818715289, + "grad_norm": 0.7908410429954529, + "learning_rate": 9.00513636563125e-06, + "loss": 0.8154, + "step": 3259 + }, + { + "epoch": 0.546578644032275, + "grad_norm": 0.3996315002441406, + "learning_rate": 8.999732870454339e-06, + "loss": 0.8057, + "step": 3260 + }, + { + "epoch": 0.5467463061930211, + "grad_norm": 0.5092489719390869, + "learning_rate": 8.994329670268595e-06, + "loss": 0.8163, + "step": 3261 + }, + { + "epoch": 0.5469139683537672, + "grad_norm": 0.6034070253372192, + "learning_rate": 8.988926766667478e-06, + "loss": 0.8637, + "step": 3262 + }, + { + "epoch": 0.5470816305145133, + "grad_norm": 0.6034070253372192, + "learning_rate": 8.988926766667478e-06, + "loss": 0.929, + "step": 3263 + }, + { + "epoch": 0.5472492926752593, + "grad_norm": 0.531434953212738, + "learning_rate": 8.983524161244379e-06, + "loss": 0.8021, + "step": 3264 + }, + { + "epoch": 0.5474169548360055, + "grad_norm": 0.4655191898345947, + "learning_rate": 8.978121855592593e-06, + "loss": 0.8185, + "step": 3265 + }, + { + "epoch": 0.5475846169967515, + "grad_norm": 0.4762585163116455, + "learning_rate": 8.972719851305327e-06, + "loss": 0.8595, + "step": 3266 + }, + { + "epoch": 0.5477522791574977, + "grad_norm": 0.6660287380218506, + "learning_rate": 8.967318149975695e-06, + "loss": 0.8693, + "step": 3267 + }, + { + "epoch": 0.5479199413182437, + "grad_norm": 0.40586456656455994, + "learning_rate": 8.961916753196728e-06, + "loss": 0.8162, + "step": 3268 + }, + { + "epoch": 0.5480876034789899, + "grad_norm": 0.5270583629608154, + "learning_rate": 8.956515662561364e-06, + "loss": 0.9386, + "step": 3269 + }, + { + "epoch": 0.5482552656397359, + "grad_norm": 0.6502846479415894, + "learning_rate": 8.951114879662455e-06, + "loss": 0.9053, + "step": 3270 + }, + { + "epoch": 0.5484229278004821, + "grad_norm": 0.3641306757926941, + "learning_rate": 8.945714406092757e-06, + "loss": 0.7675, + "step": 3271 + }, + { + "epoch": 0.5485905899612281, + "grad_norm": 0.42551761865615845, + "learning_rate": 8.940314243444933e-06, + "loss": 0.8046, + "step": 3272 + }, + { + "epoch": 0.5487582521219743, + "grad_norm": 0.4767121374607086, + "learning_rate": 8.93491439331156e-06, + "loss": 0.8365, + "step": 3273 + }, + { + "epoch": 0.5489259142827203, + "grad_norm": 0.4678700864315033, + "learning_rate": 8.929514857285122e-06, + "loss": 0.8552, + "step": 3274 + }, + { + "epoch": 0.5490935764434665, + "grad_norm": 0.6308174133300781, + "learning_rate": 8.924115636958011e-06, + "loss": 0.7933, + "step": 3275 + }, + { + "epoch": 0.5492612386042125, + "grad_norm": 0.8057112097740173, + "learning_rate": 8.918716733922519e-06, + "loss": 0.9257, + "step": 3276 + }, + { + "epoch": 0.5494289007649586, + "grad_norm": 0.7312862277030945, + "learning_rate": 8.91331814977085e-06, + "loss": 0.8304, + "step": 3277 + }, + { + "epoch": 0.5495965629257047, + "grad_norm": 0.5104177594184875, + "learning_rate": 8.907919886095115e-06, + "loss": 0.8804, + "step": 3278 + }, + { + "epoch": 0.5497642250864508, + "grad_norm": 0.4285842180252075, + "learning_rate": 8.902521944487328e-06, + "loss": 0.9018, + "step": 3279 + }, + { + "epoch": 0.5499318872471969, + "grad_norm": 0.49149802327156067, + "learning_rate": 8.897124326539415e-06, + "loss": 0.7914, + "step": 3280 + }, + { + "epoch": 0.550099549407943, + "grad_norm": 0.46300008893013, + "learning_rate": 8.89172703384319e-06, + "loss": 0.9107, + "step": 3281 + }, + { + "epoch": 0.5502672115686891, + "grad_norm": 0.36543166637420654, + "learning_rate": 8.886330067990385e-06, + "loss": 0.8881, + "step": 3282 + }, + { + "epoch": 0.5504348737294352, + "grad_norm": 0.4292105436325073, + "learning_rate": 8.880933430572634e-06, + "loss": 0.8234, + "step": 3283 + }, + { + "epoch": 0.5506025358901813, + "grad_norm": 0.7258087396621704, + "learning_rate": 8.875537123181473e-06, + "loss": 0.7786, + "step": 3284 + }, + { + "epoch": 0.5507701980509274, + "grad_norm": 0.6154667735099792, + "learning_rate": 8.870141147408343e-06, + "loss": 0.8669, + "step": 3285 + }, + { + "epoch": 0.5509378602116735, + "grad_norm": 0.5430100560188293, + "learning_rate": 8.864745504844572e-06, + "loss": 0.8951, + "step": 3286 + }, + { + "epoch": 0.5511055223724196, + "grad_norm": 0.6223335862159729, + "learning_rate": 8.859350197081412e-06, + "loss": 0.8975, + "step": 3287 + }, + { + "epoch": 0.5512731845331656, + "grad_norm": 0.5001317262649536, + "learning_rate": 8.853955225710003e-06, + "loss": 0.8587, + "step": 3288 + }, + { + "epoch": 0.5514408466939118, + "grad_norm": 0.4287770390510559, + "learning_rate": 8.848560592321391e-06, + "loss": 0.8479, + "step": 3289 + }, + { + "epoch": 0.5516085088546578, + "grad_norm": 0.4979509711265564, + "learning_rate": 8.843166298506517e-06, + "loss": 0.7994, + "step": 3290 + }, + { + "epoch": 0.551776171015404, + "grad_norm": 0.6394463181495667, + "learning_rate": 8.837772345856226e-06, + "loss": 0.852, + "step": 3291 + }, + { + "epoch": 0.55194383317615, + "grad_norm": 0.4299682378768921, + "learning_rate": 8.832378735961263e-06, + "loss": 0.8318, + "step": 3292 + }, + { + "epoch": 0.5521114953368962, + "grad_norm": 0.4377739727497101, + "learning_rate": 8.826985470412267e-06, + "loss": 0.8591, + "step": 3293 + }, + { + "epoch": 0.5522791574976422, + "grad_norm": 0.3901938498020172, + "learning_rate": 8.821592550799784e-06, + "loss": 0.8548, + "step": 3294 + }, + { + "epoch": 0.5524468196583884, + "grad_norm": 0.46280205249786377, + "learning_rate": 8.816199978714249e-06, + "loss": 0.8837, + "step": 3295 + }, + { + "epoch": 0.5526144818191344, + "grad_norm": 0.44443899393081665, + "learning_rate": 8.810807755745997e-06, + "loss": 0.8733, + "step": 3296 + }, + { + "epoch": 0.5527821439798806, + "grad_norm": 0.6398127675056458, + "learning_rate": 8.805415883485264e-06, + "loss": 0.8055, + "step": 3297 + }, + { + "epoch": 0.5529498061406266, + "grad_norm": 0.5506210923194885, + "learning_rate": 8.800024363522181e-06, + "loss": 0.9219, + "step": 3298 + }, + { + "epoch": 0.5531174683013728, + "grad_norm": 0.4390423595905304, + "learning_rate": 8.79463319744677e-06, + "loss": 0.89, + "step": 3299 + }, + { + "epoch": 0.5532851304621188, + "grad_norm": 0.46664154529571533, + "learning_rate": 8.789242386848958e-06, + "loss": 0.906, + "step": 3300 + }, + { + "epoch": 0.553452792622865, + "grad_norm": 0.3888428211212158, + "learning_rate": 8.783851933318555e-06, + "loss": 0.7911, + "step": 3301 + }, + { + "epoch": 0.553620454783611, + "grad_norm": 0.6484602093696594, + "learning_rate": 8.778461838445277e-06, + "loss": 0.9439, + "step": 3302 + }, + { + "epoch": 0.5537881169443571, + "grad_norm": 0.3052983283996582, + "learning_rate": 8.773072103818734e-06, + "loss": 0.8427, + "step": 3303 + }, + { + "epoch": 0.5539557791051032, + "grad_norm": 0.42999911308288574, + "learning_rate": 8.767682731028415e-06, + "loss": 0.9214, + "step": 3304 + }, + { + "epoch": 0.5541234412658493, + "grad_norm": 0.36539483070373535, + "learning_rate": 8.762293721663718e-06, + "loss": 0.9242, + "step": 3305 + }, + { + "epoch": 0.5542911034265954, + "grad_norm": 0.5123153924942017, + "learning_rate": 8.75690507731393e-06, + "loss": 0.8203, + "step": 3306 + }, + { + "epoch": 0.5544587655873415, + "grad_norm": 0.38076192140579224, + "learning_rate": 8.75151679956823e-06, + "loss": 0.8908, + "step": 3307 + }, + { + "epoch": 0.5546264277480876, + "grad_norm": 0.5812100768089294, + "learning_rate": 8.746128890015684e-06, + "loss": 0.908, + "step": 3308 + }, + { + "epoch": 0.5547940899088337, + "grad_norm": 0.3563908636569977, + "learning_rate": 8.740741350245252e-06, + "loss": 0.8301, + "step": 3309 + }, + { + "epoch": 0.5549617520695798, + "grad_norm": 0.4139834940433502, + "learning_rate": 8.73535418184579e-06, + "loss": 0.8231, + "step": 3310 + }, + { + "epoch": 0.5551294142303259, + "grad_norm": 0.6863387823104858, + "learning_rate": 8.729967386406042e-06, + "loss": 0.7782, + "step": 3311 + }, + { + "epoch": 0.555297076391072, + "grad_norm": 0.774410605430603, + "learning_rate": 8.724580965514638e-06, + "loss": 0.8801, + "step": 3312 + }, + { + "epoch": 0.5554647385518181, + "grad_norm": 0.37416234612464905, + "learning_rate": 8.7191949207601e-06, + "loss": 0.8831, + "step": 3313 + }, + { + "epoch": 0.5556324007125641, + "grad_norm": 0.683139979839325, + "learning_rate": 8.71380925373084e-06, + "loss": 0.8318, + "step": 3314 + }, + { + "epoch": 0.5558000628733103, + "grad_norm": 0.41312792897224426, + "learning_rate": 8.708423966015156e-06, + "loss": 0.8583, + "step": 3315 + }, + { + "epoch": 0.5559677250340563, + "grad_norm": 0.41628390550613403, + "learning_rate": 8.70303905920124e-06, + "loss": 0.8398, + "step": 3316 + }, + { + "epoch": 0.5561353871948025, + "grad_norm": 0.540746808052063, + "learning_rate": 8.697654534877166e-06, + "loss": 0.863, + "step": 3317 + }, + { + "epoch": 0.5563030493555485, + "grad_norm": 0.398416668176651, + "learning_rate": 8.692270394630894e-06, + "loss": 0.8558, + "step": 3318 + }, + { + "epoch": 0.5564707115162947, + "grad_norm": 0.41070878505706787, + "learning_rate": 8.686886640050277e-06, + "loss": 0.9208, + "step": 3319 + }, + { + "epoch": 0.5566383736770407, + "grad_norm": 0.5760504603385925, + "learning_rate": 8.68150327272305e-06, + "loss": 0.8301, + "step": 3320 + }, + { + "epoch": 0.5568060358377869, + "grad_norm": 0.4808904230594635, + "learning_rate": 8.676120294236836e-06, + "loss": 0.9025, + "step": 3321 + }, + { + "epoch": 0.5569736979985329, + "grad_norm": 0.45318806171417236, + "learning_rate": 8.670737706179138e-06, + "loss": 0.8135, + "step": 3322 + }, + { + "epoch": 0.5571413601592791, + "grad_norm": 0.5390448570251465, + "learning_rate": 8.665355510137351e-06, + "loss": 0.9409, + "step": 3323 + }, + { + "epoch": 0.5573090223200251, + "grad_norm": 0.4017261266708374, + "learning_rate": 8.659973707698751e-06, + "loss": 0.8512, + "step": 3324 + }, + { + "epoch": 0.5574766844807713, + "grad_norm": 0.4574022889137268, + "learning_rate": 8.654592300450496e-06, + "loss": 0.8309, + "step": 3325 + }, + { + "epoch": 0.5576443466415173, + "grad_norm": 0.5664092302322388, + "learning_rate": 8.649211289979638e-06, + "loss": 0.9195, + "step": 3326 + }, + { + "epoch": 0.5578120088022634, + "grad_norm": 0.41941285133361816, + "learning_rate": 8.643830677873092e-06, + "loss": 0.8546, + "step": 3327 + }, + { + "epoch": 0.5579796709630095, + "grad_norm": 0.5523105263710022, + "learning_rate": 8.638450465717671e-06, + "loss": 0.82, + "step": 3328 + }, + { + "epoch": 0.5581473331237556, + "grad_norm": 0.4249464273452759, + "learning_rate": 8.633070655100069e-06, + "loss": 0.9138, + "step": 3329 + }, + { + "epoch": 0.5583149952845017, + "grad_norm": 0.6280999183654785, + "learning_rate": 8.627691247606862e-06, + "loss": 0.8309, + "step": 3330 + }, + { + "epoch": 0.5584826574452478, + "grad_norm": 0.41287532448768616, + "learning_rate": 8.622312244824492e-06, + "loss": 0.8317, + "step": 3331 + }, + { + "epoch": 0.5586503196059939, + "grad_norm": 0.5640285015106201, + "learning_rate": 8.616933648339303e-06, + "loss": 0.935, + "step": 3332 + }, + { + "epoch": 0.55881798176674, + "grad_norm": 0.38837930560112, + "learning_rate": 8.61155545973751e-06, + "loss": 0.8126, + "step": 3333 + }, + { + "epoch": 0.5589856439274861, + "grad_norm": 0.4220353364944458, + "learning_rate": 8.6061776806052e-06, + "loss": 0.8232, + "step": 3334 + }, + { + "epoch": 0.5591533060882322, + "grad_norm": 0.4612191617488861, + "learning_rate": 8.600800312528356e-06, + "loss": 0.8701, + "step": 3335 + }, + { + "epoch": 0.5593209682489783, + "grad_norm": 1.0260591506958008, + "learning_rate": 8.595423357092823e-06, + "loss": 0.9243, + "step": 3336 + }, + { + "epoch": 0.5594886304097244, + "grad_norm": 0.4499152600765228, + "learning_rate": 8.590046815884337e-06, + "loss": 0.8106, + "step": 3337 + }, + { + "epoch": 0.5596562925704704, + "grad_norm": 0.6131991147994995, + "learning_rate": 8.584670690488501e-06, + "loss": 0.8969, + "step": 3338 + }, + { + "epoch": 0.5598239547312166, + "grad_norm": 0.9590579271316528, + "learning_rate": 8.579294982490808e-06, + "loss": 0.9169, + "step": 3339 + }, + { + "epoch": 0.5599916168919626, + "grad_norm": 0.409368097782135, + "learning_rate": 8.573919693476614e-06, + "loss": 0.8295, + "step": 3340 + }, + { + "epoch": 0.5601592790527088, + "grad_norm": 0.4588213264942169, + "learning_rate": 8.568544825031163e-06, + "loss": 0.8349, + "step": 3341 + }, + { + "epoch": 0.5603269412134549, + "grad_norm": 0.6115896701812744, + "learning_rate": 8.563170378739567e-06, + "loss": 0.8692, + "step": 3342 + }, + { + "epoch": 0.560494603374201, + "grad_norm": 0.371785044670105, + "learning_rate": 8.557796356186818e-06, + "loss": 0.8572, + "step": 3343 + }, + { + "epoch": 0.5606622655349471, + "grad_norm": 0.38996487855911255, + "learning_rate": 8.552422758957785e-06, + "loss": 0.796, + "step": 3344 + }, + { + "epoch": 0.5608299276956932, + "grad_norm": 0.42374661564826965, + "learning_rate": 8.547049588637204e-06, + "loss": 0.963, + "step": 3345 + }, + { + "epoch": 0.5609975898564393, + "grad_norm": 0.37725475430488586, + "learning_rate": 8.54167684680969e-06, + "loss": 0.8658, + "step": 3346 + }, + { + "epoch": 0.5611652520171854, + "grad_norm": 0.5238369107246399, + "learning_rate": 8.536304535059731e-06, + "loss": 0.8589, + "step": 3347 + }, + { + "epoch": 0.5613329141779315, + "grad_norm": 0.4647602140903473, + "learning_rate": 8.530932654971693e-06, + "loss": 0.8395, + "step": 3348 + }, + { + "epoch": 0.5615005763386776, + "grad_norm": 0.3794912099838257, + "learning_rate": 8.525561208129805e-06, + "loss": 0.8179, + "step": 3349 + }, + { + "epoch": 0.5616682384994237, + "grad_norm": 0.4015001654624939, + "learning_rate": 8.520190196118173e-06, + "loss": 0.7695, + "step": 3350 + }, + { + "epoch": 0.5618359006601698, + "grad_norm": 0.4131832420825958, + "learning_rate": 8.514819620520773e-06, + "loss": 0.8353, + "step": 3351 + }, + { + "epoch": 0.5620035628209159, + "grad_norm": 0.44537413120269775, + "learning_rate": 8.509449482921462e-06, + "loss": 0.8112, + "step": 3352 + }, + { + "epoch": 0.5621712249816619, + "grad_norm": 0.5103172659873962, + "learning_rate": 8.50407978490396e-06, + "loss": 0.9033, + "step": 3353 + }, + { + "epoch": 0.5623388871424081, + "grad_norm": 0.4940226674079895, + "learning_rate": 8.498710528051845e-06, + "loss": 0.9388, + "step": 3354 + }, + { + "epoch": 0.5625065493031541, + "grad_norm": 0.5164597630500793, + "learning_rate": 8.493341713948588e-06, + "loss": 0.9151, + "step": 3355 + }, + { + "epoch": 0.5626742114639003, + "grad_norm": 0.41803959012031555, + "learning_rate": 8.487973344177517e-06, + "loss": 0.9144, + "step": 3356 + }, + { + "epoch": 0.5628418736246463, + "grad_norm": 0.4546486735343933, + "learning_rate": 8.482605420321829e-06, + "loss": 0.9372, + "step": 3357 + }, + { + "epoch": 0.5630095357853925, + "grad_norm": 0.3524572253227234, + "learning_rate": 8.477237943964596e-06, + "loss": 0.8226, + "step": 3358 + }, + { + "epoch": 0.5631771979461385, + "grad_norm": 0.5530046224594116, + "learning_rate": 8.471870916688745e-06, + "loss": 0.8778, + "step": 3359 + }, + { + "epoch": 0.5633448601068847, + "grad_norm": 0.45319312810897827, + "learning_rate": 8.466504340077084e-06, + "loss": 0.8568, + "step": 3360 + }, + { + "epoch": 0.5635125222676307, + "grad_norm": 0.6729868650436401, + "learning_rate": 8.461138215712283e-06, + "loss": 0.917, + "step": 3361 + }, + { + "epoch": 0.5636801844283769, + "grad_norm": 0.461455374956131, + "learning_rate": 8.45577254517688e-06, + "loss": 0.7528, + "step": 3362 + }, + { + "epoch": 0.5638478465891229, + "grad_norm": 0.9580578804016113, + "learning_rate": 8.450407330053271e-06, + "loss": 0.8159, + "step": 3363 + }, + { + "epoch": 0.564015508749869, + "grad_norm": 0.5438309907913208, + "learning_rate": 8.44504257192373e-06, + "loss": 0.7513, + "step": 3364 + }, + { + "epoch": 0.5641831709106151, + "grad_norm": 0.6372382044792175, + "learning_rate": 8.439678272370391e-06, + "loss": 0.9019, + "step": 3365 + }, + { + "epoch": 0.5643508330713612, + "grad_norm": 0.6940456628799438, + "learning_rate": 8.434314432975248e-06, + "loss": 0.8214, + "step": 3366 + }, + { + "epoch": 0.5645184952321073, + "grad_norm": 0.40165215730667114, + "learning_rate": 8.428951055320168e-06, + "loss": 0.8752, + "step": 3367 + }, + { + "epoch": 0.5646861573928534, + "grad_norm": 0.5314671993255615, + "learning_rate": 8.423588140986876e-06, + "loss": 0.7734, + "step": 3368 + }, + { + "epoch": 0.5648538195535995, + "grad_norm": 0.3801604211330414, + "learning_rate": 8.418225691556962e-06, + "loss": 0.8735, + "step": 3369 + }, + { + "epoch": 0.5650214817143456, + "grad_norm": 0.3730860948562622, + "learning_rate": 8.412863708611874e-06, + "loss": 0.8734, + "step": 3370 + }, + { + "epoch": 0.5651891438750917, + "grad_norm": 0.7912408709526062, + "learning_rate": 8.40750219373294e-06, + "loss": 0.9414, + "step": 3371 + }, + { + "epoch": 0.5653568060358378, + "grad_norm": 0.46992161870002747, + "learning_rate": 8.402141148501323e-06, + "loss": 0.9592, + "step": 3372 + }, + { + "epoch": 0.5655244681965839, + "grad_norm": 0.660154402256012, + "learning_rate": 8.396780574498066e-06, + "loss": 0.9717, + "step": 3373 + }, + { + "epoch": 0.56569213035733, + "grad_norm": 0.6320633292198181, + "learning_rate": 8.391420473304071e-06, + "loss": 0.8456, + "step": 3374 + }, + { + "epoch": 0.565859792518076, + "grad_norm": 0.5591331720352173, + "learning_rate": 8.386060846500097e-06, + "loss": 0.7859, + "step": 3375 + }, + { + "epoch": 0.5660274546788222, + "grad_norm": 0.43963560461997986, + "learning_rate": 8.38070169566677e-06, + "loss": 0.887, + "step": 3376 + }, + { + "epoch": 0.5661951168395682, + "grad_norm": 0.3942304849624634, + "learning_rate": 8.375343022384559e-06, + "loss": 0.8051, + "step": 3377 + }, + { + "epoch": 0.5663627790003144, + "grad_norm": 0.6130456924438477, + "learning_rate": 8.369984828233807e-06, + "loss": 0.8517, + "step": 3378 + }, + { + "epoch": 0.5665304411610604, + "grad_norm": 0.4271846413612366, + "learning_rate": 8.364627114794716e-06, + "loss": 0.8557, + "step": 3379 + }, + { + "epoch": 0.5666981033218066, + "grad_norm": 0.44319701194763184, + "learning_rate": 8.35926988364734e-06, + "loss": 0.8519, + "step": 3380 + }, + { + "epoch": 0.5668657654825526, + "grad_norm": 0.5144001245498657, + "learning_rate": 8.353913136371592e-06, + "loss": 0.9553, + "step": 3381 + }, + { + "epoch": 0.5670334276432988, + "grad_norm": 0.4724206030368805, + "learning_rate": 8.348556874547242e-06, + "loss": 0.8115, + "step": 3382 + }, + { + "epoch": 0.5672010898040448, + "grad_norm": 0.5161254405975342, + "learning_rate": 8.343201099753921e-06, + "loss": 0.9335, + "step": 3383 + }, + { + "epoch": 0.567368751964791, + "grad_norm": 0.7193112969398499, + "learning_rate": 8.337845813571109e-06, + "loss": 0.9, + "step": 3384 + }, + { + "epoch": 0.567536414125537, + "grad_norm": 0.6863871216773987, + "learning_rate": 8.332491017578152e-06, + "loss": 0.8844, + "step": 3385 + }, + { + "epoch": 0.5677040762862832, + "grad_norm": 0.5251758098602295, + "learning_rate": 8.32713671335424e-06, + "loss": 0.7751, + "step": 3386 + }, + { + "epoch": 0.5678717384470292, + "grad_norm": 0.5140281915664673, + "learning_rate": 8.321782902478426e-06, + "loss": 0.8754, + "step": 3387 + }, + { + "epoch": 0.5680394006077754, + "grad_norm": 0.6101822853088379, + "learning_rate": 8.316429586529616e-06, + "loss": 0.8659, + "step": 3388 + }, + { + "epoch": 0.5682070627685214, + "grad_norm": 0.4696868062019348, + "learning_rate": 8.311076767086566e-06, + "loss": 0.8208, + "step": 3389 + }, + { + "epoch": 0.5683747249292675, + "grad_norm": 0.6828814148902893, + "learning_rate": 8.305724445727896e-06, + "loss": 0.8719, + "step": 3390 + }, + { + "epoch": 0.5685423870900136, + "grad_norm": 0.3969288170337677, + "learning_rate": 8.300372624032062e-06, + "loss": 0.877, + "step": 3391 + }, + { + "epoch": 0.5687100492507597, + "grad_norm": 0.6055564284324646, + "learning_rate": 8.29502130357739e-06, + "loss": 0.8484, + "step": 3392 + }, + { + "epoch": 0.5688777114115058, + "grad_norm": 0.47719958424568176, + "learning_rate": 8.289670485942044e-06, + "loss": 1.0181, + "step": 3393 + }, + { + "epoch": 0.5690453735722519, + "grad_norm": 0.43543651700019836, + "learning_rate": 8.284320172704058e-06, + "loss": 0.835, + "step": 3394 + }, + { + "epoch": 0.569213035732998, + "grad_norm": 0.6597638130187988, + "learning_rate": 8.278970365441292e-06, + "loss": 0.8011, + "step": 3395 + }, + { + "epoch": 0.5693806978937441, + "grad_norm": 0.812634289264679, + "learning_rate": 8.273621065731479e-06, + "loss": 0.7814, + "step": 3396 + }, + { + "epoch": 0.5695483600544902, + "grad_norm": 0.45590394735336304, + "learning_rate": 8.268272275152187e-06, + "loss": 0.8903, + "step": 3397 + }, + { + "epoch": 0.5697160222152363, + "grad_norm": 0.4711604118347168, + "learning_rate": 8.262923995280847e-06, + "loss": 0.8244, + "step": 3398 + }, + { + "epoch": 0.5698836843759824, + "grad_norm": 0.6121546626091003, + "learning_rate": 8.257576227694736e-06, + "loss": 0.8225, + "step": 3399 + }, + { + "epoch": 0.5700513465367285, + "grad_norm": 0.9810781478881836, + "learning_rate": 8.252228973970964e-06, + "loss": 0.8684, + "step": 3400 + }, + { + "epoch": 0.5702190086974745, + "grad_norm": 0.39338693022727966, + "learning_rate": 8.246882235686512e-06, + "loss": 0.9007, + "step": 3401 + }, + { + "epoch": 0.5703866708582207, + "grad_norm": 0.448321133852005, + "learning_rate": 8.241536014418199e-06, + "loss": 0.8147, + "step": 3402 + }, + { + "epoch": 0.5705543330189667, + "grad_norm": 0.38478726148605347, + "learning_rate": 8.23619031174269e-06, + "loss": 0.793, + "step": 3403 + }, + { + "epoch": 0.5707219951797129, + "grad_norm": 0.4584026038646698, + "learning_rate": 8.230845129236498e-06, + "loss": 0.7901, + "step": 3404 + }, + { + "epoch": 0.5708896573404589, + "grad_norm": 0.41249966621398926, + "learning_rate": 8.225500468475984e-06, + "loss": 0.8622, + "step": 3405 + }, + { + "epoch": 0.5710573195012051, + "grad_norm": 0.7865109443664551, + "learning_rate": 8.220156331037355e-06, + "loss": 0.9252, + "step": 3406 + }, + { + "epoch": 0.5712249816619511, + "grad_norm": 0.4040682315826416, + "learning_rate": 8.214812718496666e-06, + "loss": 0.8741, + "step": 3407 + }, + { + "epoch": 0.5713926438226973, + "grad_norm": 0.6797685027122498, + "learning_rate": 8.209469632429811e-06, + "loss": 0.8442, + "step": 3408 + }, + { + "epoch": 0.5715603059834433, + "grad_norm": 0.40794631838798523, + "learning_rate": 8.204127074412531e-06, + "loss": 0.8695, + "step": 3409 + }, + { + "epoch": 0.5717279681441895, + "grad_norm": 0.3528442680835724, + "learning_rate": 8.198785046020417e-06, + "loss": 0.8551, + "step": 3410 + }, + { + "epoch": 0.5718956303049355, + "grad_norm": 0.4704998731613159, + "learning_rate": 8.193443548828896e-06, + "loss": 0.7734, + "step": 3411 + }, + { + "epoch": 0.5720632924656817, + "grad_norm": 0.4955574870109558, + "learning_rate": 8.188102584413244e-06, + "loss": 0.7795, + "step": 3412 + }, + { + "epoch": 0.5722309546264277, + "grad_norm": 0.5632057189941406, + "learning_rate": 8.182762154348575e-06, + "loss": 0.8672, + "step": 3413 + }, + { + "epoch": 0.5723986167871739, + "grad_norm": 0.4081422686576843, + "learning_rate": 8.177422260209849e-06, + "loss": 0.8319, + "step": 3414 + }, + { + "epoch": 0.5725662789479199, + "grad_norm": 0.44355064630508423, + "learning_rate": 8.172082903571868e-06, + "loss": 0.8316, + "step": 3415 + }, + { + "epoch": 0.572733941108666, + "grad_norm": 0.41933557391166687, + "learning_rate": 8.16674408600927e-06, + "loss": 0.904, + "step": 3416 + }, + { + "epoch": 0.5729016032694121, + "grad_norm": 0.47249361872673035, + "learning_rate": 8.161405809096547e-06, + "loss": 0.86, + "step": 3417 + }, + { + "epoch": 0.5730692654301582, + "grad_norm": 0.4084494113922119, + "learning_rate": 8.156068074408016e-06, + "loss": 0.9057, + "step": 3418 + }, + { + "epoch": 0.5732369275909043, + "grad_norm": 0.45873379707336426, + "learning_rate": 8.150730883517841e-06, + "loss": 0.802, + "step": 3419 + }, + { + "epoch": 0.5734045897516504, + "grad_norm": 0.8427038192749023, + "learning_rate": 8.145394238000028e-06, + "loss": 0.9047, + "step": 3420 + }, + { + "epoch": 0.5735722519123965, + "grad_norm": 0.33546656370162964, + "learning_rate": 8.140058139428425e-06, + "loss": 0.8843, + "step": 3421 + }, + { + "epoch": 0.5737399140731426, + "grad_norm": 0.40231165289878845, + "learning_rate": 8.134722589376704e-06, + "loss": 0.8627, + "step": 3422 + }, + { + "epoch": 0.5739075762338887, + "grad_norm": 0.40609556436538696, + "learning_rate": 8.129387589418385e-06, + "loss": 0.8711, + "step": 3423 + }, + { + "epoch": 0.5740752383946348, + "grad_norm": 0.4238656163215637, + "learning_rate": 8.124053141126833e-06, + "loss": 0.9049, + "step": 3424 + }, + { + "epoch": 0.5742429005553809, + "grad_norm": 0.9477406740188599, + "learning_rate": 8.118719246075239e-06, + "loss": 0.8209, + "step": 3425 + }, + { + "epoch": 0.574410562716127, + "grad_norm": 0.44262930750846863, + "learning_rate": 8.113385905836636e-06, + "loss": 0.8028, + "step": 3426 + }, + { + "epoch": 0.5745782248768732, + "grad_norm": 0.7265502214431763, + "learning_rate": 8.10805312198389e-06, + "loss": 0.8735, + "step": 3427 + }, + { + "epoch": 0.5747458870376192, + "grad_norm": 0.42299774289131165, + "learning_rate": 8.102720896089705e-06, + "loss": 0.8704, + "step": 3428 + }, + { + "epoch": 0.5749135491983653, + "grad_norm": 0.663567841053009, + "learning_rate": 8.097389229726623e-06, + "loss": 0.944, + "step": 3429 + }, + { + "epoch": 0.5750812113591114, + "grad_norm": 0.9425939917564392, + "learning_rate": 8.092058124467014e-06, + "loss": 0.8729, + "step": 3430 + }, + { + "epoch": 0.5752488735198575, + "grad_norm": 0.7026984691619873, + "learning_rate": 8.086727581883094e-06, + "loss": 0.8146, + "step": 3431 + }, + { + "epoch": 0.5754165356806036, + "grad_norm": 0.38072091341018677, + "learning_rate": 8.0813976035469e-06, + "loss": 0.8646, + "step": 3432 + }, + { + "epoch": 0.5755841978413497, + "grad_norm": 0.3900136947631836, + "learning_rate": 8.076068191030309e-06, + "loss": 0.8507, + "step": 3433 + }, + { + "epoch": 0.5757518600020958, + "grad_norm": 0.41140100359916687, + "learning_rate": 8.070739345905032e-06, + "loss": 0.8211, + "step": 3434 + }, + { + "epoch": 0.5759195221628419, + "grad_norm": 0.3828379809856415, + "learning_rate": 8.065411069742615e-06, + "loss": 0.8792, + "step": 3435 + }, + { + "epoch": 0.576087184323588, + "grad_norm": 0.5758510828018188, + "learning_rate": 8.060083364114427e-06, + "loss": 0.9086, + "step": 3436 + }, + { + "epoch": 0.5762548464843341, + "grad_norm": 0.6231465935707092, + "learning_rate": 8.054756230591676e-06, + "loss": 1.0062, + "step": 3437 + }, + { + "epoch": 0.5764225086450802, + "grad_norm": 0.42408302426338196, + "learning_rate": 8.0494296707454e-06, + "loss": 0.8886, + "step": 3438 + }, + { + "epoch": 0.5765901708058263, + "grad_norm": 0.39178594946861267, + "learning_rate": 8.044103686146465e-06, + "loss": 0.7908, + "step": 3439 + }, + { + "epoch": 0.5767578329665723, + "grad_norm": 0.35958993434906006, + "learning_rate": 8.03877827836558e-06, + "loss": 0.7988, + "step": 3440 + }, + { + "epoch": 0.5769254951273185, + "grad_norm": 0.5605311393737793, + "learning_rate": 8.033453448973261e-06, + "loss": 0.9095, + "step": 3441 + }, + { + "epoch": 0.5770931572880645, + "grad_norm": 0.571583092212677, + "learning_rate": 8.02812919953987e-06, + "loss": 0.8901, + "step": 3442 + }, + { + "epoch": 0.5772608194488107, + "grad_norm": 0.630528450012207, + "learning_rate": 8.022805531635599e-06, + "loss": 0.8026, + "step": 3443 + }, + { + "epoch": 0.5774284816095567, + "grad_norm": 0.4706905782222748, + "learning_rate": 8.01748244683046e-06, + "loss": 0.893, + "step": 3444 + }, + { + "epoch": 0.5775961437703029, + "grad_norm": 0.6315483450889587, + "learning_rate": 8.012159946694295e-06, + "loss": 0.8625, + "step": 3445 + }, + { + "epoch": 0.5777638059310489, + "grad_norm": 0.3732752799987793, + "learning_rate": 8.006838032796775e-06, + "loss": 0.8929, + "step": 3446 + }, + { + "epoch": 0.5779314680917951, + "grad_norm": 0.4668952524662018, + "learning_rate": 8.001516706707401e-06, + "loss": 0.8655, + "step": 3447 + }, + { + "epoch": 0.5780991302525411, + "grad_norm": 0.4400523900985718, + "learning_rate": 7.996195969995498e-06, + "loss": 0.8342, + "step": 3448 + }, + { + "epoch": 0.5782667924132873, + "grad_norm": 0.5376847386360168, + "learning_rate": 7.990875824230217e-06, + "loss": 0.8972, + "step": 3449 + }, + { + "epoch": 0.5784344545740333, + "grad_norm": 0.5515185594558716, + "learning_rate": 7.985556270980534e-06, + "loss": 0.8021, + "step": 3450 + }, + { + "epoch": 0.5786021167347795, + "grad_norm": 0.35805997252464294, + "learning_rate": 7.98023731181525e-06, + "loss": 0.8116, + "step": 3451 + }, + { + "epoch": 0.5787697788955255, + "grad_norm": 0.5249135494232178, + "learning_rate": 7.974918948302993e-06, + "loss": 0.9464, + "step": 3452 + }, + { + "epoch": 0.5789374410562717, + "grad_norm": 0.5082404017448425, + "learning_rate": 7.969601182012217e-06, + "loss": 0.8521, + "step": 3453 + }, + { + "epoch": 0.5791051032170177, + "grad_norm": 0.37490904331207275, + "learning_rate": 7.964284014511193e-06, + "loss": 0.7826, + "step": 3454 + }, + { + "epoch": 0.5792727653777638, + "grad_norm": 0.3770364224910736, + "learning_rate": 7.958967447368018e-06, + "loss": 0.8681, + "step": 3455 + }, + { + "epoch": 0.5794404275385099, + "grad_norm": 0.3770364224910736, + "learning_rate": 7.958967447368018e-06, + "loss": 0.7582, + "step": 3456 + }, + { + "epoch": 0.579608089699256, + "grad_norm": 0.8216555118560791, + "learning_rate": 7.953651482150619e-06, + "loss": 0.8643, + "step": 3457 + }, + { + "epoch": 0.5797757518600021, + "grad_norm": 0.4499664604663849, + "learning_rate": 7.948336120426736e-06, + "loss": 0.828, + "step": 3458 + }, + { + "epoch": 0.5799434140207482, + "grad_norm": 0.3453725278377533, + "learning_rate": 7.943021363763941e-06, + "loss": 0.8029, + "step": 3459 + }, + { + "epoch": 0.5801110761814943, + "grad_norm": 0.5152020454406738, + "learning_rate": 7.937707213729612e-06, + "loss": 0.8581, + "step": 3460 + }, + { + "epoch": 0.5802787383422404, + "grad_norm": 0.44359079003334045, + "learning_rate": 7.932393671890965e-06, + "loss": 0.847, + "step": 3461 + }, + { + "epoch": 0.5804464005029865, + "grad_norm": 0.6569787859916687, + "learning_rate": 7.927080739815025e-06, + "loss": 0.8359, + "step": 3462 + }, + { + "epoch": 0.5806140626637326, + "grad_norm": 0.49237895011901855, + "learning_rate": 7.921768419068643e-06, + "loss": 0.8377, + "step": 3463 + }, + { + "epoch": 0.5807817248244787, + "grad_norm": 0.34828248620033264, + "learning_rate": 7.916456711218494e-06, + "loss": 0.8696, + "step": 3464 + }, + { + "epoch": 0.5809493869852248, + "grad_norm": 0.6898155808448792, + "learning_rate": 7.911145617831057e-06, + "loss": 0.8389, + "step": 3465 + }, + { + "epoch": 0.5811170491459708, + "grad_norm": 0.5621565580368042, + "learning_rate": 7.90583514047264e-06, + "loss": 0.878, + "step": 3466 + }, + { + "epoch": 0.581284711306717, + "grad_norm": 0.7397379279136658, + "learning_rate": 7.900525280709375e-06, + "loss": 0.875, + "step": 3467 + }, + { + "epoch": 0.581452373467463, + "grad_norm": 0.39626482129096985, + "learning_rate": 7.895216040107207e-06, + "loss": 0.7933, + "step": 3468 + }, + { + "epoch": 0.5816200356282092, + "grad_norm": 0.573769211769104, + "learning_rate": 7.889907420231884e-06, + "loss": 0.8246, + "step": 3469 + }, + { + "epoch": 0.5817876977889552, + "grad_norm": 0.42165836691856384, + "learning_rate": 7.884599422648995e-06, + "loss": 0.8142, + "step": 3470 + }, + { + "epoch": 0.5819553599497014, + "grad_norm": 0.48622065782546997, + "learning_rate": 7.87929204892393e-06, + "loss": 0.8257, + "step": 3471 + }, + { + "epoch": 0.5821230221104474, + "grad_norm": 0.5852141380310059, + "learning_rate": 7.873985300621902e-06, + "loss": 0.8334, + "step": 3472 + }, + { + "epoch": 0.5822906842711936, + "grad_norm": 0.46930983662605286, + "learning_rate": 7.868679179307934e-06, + "loss": 0.7885, + "step": 3473 + }, + { + "epoch": 0.5824583464319396, + "grad_norm": 0.6163612604141235, + "learning_rate": 7.863373686546868e-06, + "loss": 0.8423, + "step": 3474 + }, + { + "epoch": 0.5826260085926858, + "grad_norm": 0.6135543584823608, + "learning_rate": 7.858068823903358e-06, + "loss": 0.8644, + "step": 3475 + }, + { + "epoch": 0.5827936707534318, + "grad_norm": 0.46974796056747437, + "learning_rate": 7.852764592941876e-06, + "loss": 0.8734, + "step": 3476 + }, + { + "epoch": 0.582961332914178, + "grad_norm": 0.4672534465789795, + "learning_rate": 7.847460995226706e-06, + "loss": 0.7436, + "step": 3477 + }, + { + "epoch": 0.583128995074924, + "grad_norm": 0.4216882586479187, + "learning_rate": 7.84215803232194e-06, + "loss": 0.922, + "step": 3478 + }, + { + "epoch": 0.5832966572356701, + "grad_norm": 0.6704283356666565, + "learning_rate": 7.836855705791492e-06, + "loss": 0.8957, + "step": 3479 + }, + { + "epoch": 0.5834643193964162, + "grad_norm": 0.46347349882125854, + "learning_rate": 7.83155401719908e-06, + "loss": 0.9365, + "step": 3480 + }, + { + "epoch": 0.5836319815571623, + "grad_norm": 0.32277539372444153, + "learning_rate": 7.826252968108242e-06, + "loss": 0.9403, + "step": 3481 + }, + { + "epoch": 0.5837996437179084, + "grad_norm": 0.4305226504802704, + "learning_rate": 7.820952560082322e-06, + "loss": 0.8092, + "step": 3482 + }, + { + "epoch": 0.5839673058786545, + "grad_norm": 0.7612534761428833, + "learning_rate": 7.815652794684472e-06, + "loss": 0.7739, + "step": 3483 + }, + { + "epoch": 0.5841349680394006, + "grad_norm": 0.5855340361595154, + "learning_rate": 7.810353673477664e-06, + "loss": 0.8841, + "step": 3484 + }, + { + "epoch": 0.5843026302001467, + "grad_norm": 0.5339702367782593, + "learning_rate": 7.805055198024668e-06, + "loss": 0.9109, + "step": 3485 + }, + { + "epoch": 0.5844702923608928, + "grad_norm": 0.39810431003570557, + "learning_rate": 7.799757369888079e-06, + "loss": 0.9111, + "step": 3486 + }, + { + "epoch": 0.5846379545216389, + "grad_norm": 0.46095889806747437, + "learning_rate": 7.794460190630283e-06, + "loss": 0.8265, + "step": 3487 + }, + { + "epoch": 0.584805616682385, + "grad_norm": 0.44814544916152954, + "learning_rate": 7.789163661813486e-06, + "loss": 0.8003, + "step": 3488 + }, + { + "epoch": 0.5849732788431311, + "grad_norm": 0.4918145537376404, + "learning_rate": 7.783867784999701e-06, + "loss": 0.8374, + "step": 3489 + }, + { + "epoch": 0.5851409410038771, + "grad_norm": 0.7340918183326721, + "learning_rate": 7.778572561750752e-06, + "loss": 0.8805, + "step": 3490 + }, + { + "epoch": 0.5853086031646233, + "grad_norm": 0.3280177414417267, + "learning_rate": 7.773277993628265e-06, + "loss": 0.8806, + "step": 3491 + }, + { + "epoch": 0.5854762653253693, + "grad_norm": 0.3526843786239624, + "learning_rate": 7.767984082193662e-06, + "loss": 0.7938, + "step": 3492 + }, + { + "epoch": 0.5856439274861155, + "grad_norm": 0.3464638292789459, + "learning_rate": 7.762690829008197e-06, + "loss": 0.9061, + "step": 3493 + }, + { + "epoch": 0.5858115896468615, + "grad_norm": 0.46480873227119446, + "learning_rate": 7.75739823563291e-06, + "loss": 0.7788, + "step": 3494 + }, + { + "epoch": 0.5859792518076077, + "grad_norm": 0.411076158285141, + "learning_rate": 7.752106303628656e-06, + "loss": 0.8505, + "step": 3495 + }, + { + "epoch": 0.5861469139683537, + "grad_norm": 0.5689971446990967, + "learning_rate": 7.746815034556084e-06, + "loss": 0.7777, + "step": 3496 + }, + { + "epoch": 0.5863145761290999, + "grad_norm": 0.9616680145263672, + "learning_rate": 7.74152442997566e-06, + "loss": 0.8638, + "step": 3497 + }, + { + "epoch": 0.5864822382898459, + "grad_norm": 1.2664830684661865, + "learning_rate": 7.736234491447649e-06, + "loss": 0.8451, + "step": 3498 + }, + { + "epoch": 0.5866499004505921, + "grad_norm": 1.1053142547607422, + "learning_rate": 7.730945220532116e-06, + "loss": 0.9524, + "step": 3499 + }, + { + "epoch": 0.5868175626113381, + "grad_norm": 0.40867578983306885, + "learning_rate": 7.725656618788938e-06, + "loss": 0.7747, + "step": 3500 + }, + { + "epoch": 0.5869852247720843, + "grad_norm": 0.6661640405654907, + "learning_rate": 7.720368687777784e-06, + "loss": 0.8648, + "step": 3501 + }, + { + "epoch": 0.5871528869328303, + "grad_norm": 0.39013051986694336, + "learning_rate": 7.715081429058132e-06, + "loss": 0.7754, + "step": 3502 + }, + { + "epoch": 0.5873205490935764, + "grad_norm": 0.47789233922958374, + "learning_rate": 7.709794844189262e-06, + "loss": 0.9175, + "step": 3503 + }, + { + "epoch": 0.5874882112543225, + "grad_norm": 0.6550090312957764, + "learning_rate": 7.70450893473025e-06, + "loss": 0.8314, + "step": 3504 + }, + { + "epoch": 0.5876558734150686, + "grad_norm": 0.5416508913040161, + "learning_rate": 7.69922370223998e-06, + "loss": 0.9176, + "step": 3505 + }, + { + "epoch": 0.5878235355758147, + "grad_norm": 0.5480464100837708, + "learning_rate": 7.693939148277128e-06, + "loss": 0.9007, + "step": 3506 + }, + { + "epoch": 0.5879911977365608, + "grad_norm": 0.6138718724250793, + "learning_rate": 7.688655274400176e-06, + "loss": 0.8431, + "step": 3507 + }, + { + "epoch": 0.5881588598973069, + "grad_norm": 0.39476701617240906, + "learning_rate": 7.683372082167404e-06, + "loss": 0.8014, + "step": 3508 + }, + { + "epoch": 0.588326522058053, + "grad_norm": 0.42859458923339844, + "learning_rate": 7.678089573136897e-06, + "loss": 0.876, + "step": 3509 + }, + { + "epoch": 0.5884941842187991, + "grad_norm": 0.5820524096488953, + "learning_rate": 7.672807748866523e-06, + "loss": 0.7965, + "step": 3510 + }, + { + "epoch": 0.5886618463795452, + "grad_norm": 0.41693684458732605, + "learning_rate": 7.667526610913958e-06, + "loss": 0.8259, + "step": 3511 + }, + { + "epoch": 0.5888295085402913, + "grad_norm": 0.5068554282188416, + "learning_rate": 7.66224616083668e-06, + "loss": 0.8, + "step": 3512 + }, + { + "epoch": 0.5889971707010374, + "grad_norm": 0.5913344621658325, + "learning_rate": 7.656966400191956e-06, + "loss": 0.7453, + "step": 3513 + }, + { + "epoch": 0.5891648328617836, + "grad_norm": 0.4259031414985657, + "learning_rate": 7.651687330536862e-06, + "loss": 0.8127, + "step": 3514 + }, + { + "epoch": 0.5893324950225296, + "grad_norm": 0.6040980815887451, + "learning_rate": 7.646408953428243e-06, + "loss": 0.9615, + "step": 3515 + }, + { + "epoch": 0.5895001571832758, + "grad_norm": 0.5717450976371765, + "learning_rate": 7.641131270422772e-06, + "loss": 0.8308, + "step": 3516 + }, + { + "epoch": 0.5896678193440218, + "grad_norm": 0.48589468002319336, + "learning_rate": 7.635854283076897e-06, + "loss": 0.7764, + "step": 3517 + }, + { + "epoch": 0.5898354815047679, + "grad_norm": 0.5082957744598389, + "learning_rate": 7.630577992946871e-06, + "loss": 0.9003, + "step": 3518 + }, + { + "epoch": 0.590003143665514, + "grad_norm": 0.47448721528053284, + "learning_rate": 7.625302401588735e-06, + "loss": 0.8517, + "step": 3519 + }, + { + "epoch": 0.5901708058262601, + "grad_norm": 0.5031949877738953, + "learning_rate": 7.620027510558324e-06, + "loss": 0.7788, + "step": 3520 + }, + { + "epoch": 0.5903384679870062, + "grad_norm": 0.601178765296936, + "learning_rate": 7.6147533214112725e-06, + "loss": 0.8609, + "step": 3521 + }, + { + "epoch": 0.5905061301477523, + "grad_norm": 0.46869635581970215, + "learning_rate": 7.609479835703002e-06, + "loss": 0.7583, + "step": 3522 + }, + { + "epoch": 0.5906737923084984, + "grad_norm": 0.6021032929420471, + "learning_rate": 7.60420705498873e-06, + "loss": 0.818, + "step": 3523 + }, + { + "epoch": 0.5908414544692445, + "grad_norm": 0.4280416667461395, + "learning_rate": 7.598934980823465e-06, + "loss": 0.7912, + "step": 3524 + }, + { + "epoch": 0.5910091166299906, + "grad_norm": 0.9477753639221191, + "learning_rate": 7.593663614762004e-06, + "loss": 0.8083, + "step": 3525 + }, + { + "epoch": 0.5911767787907367, + "grad_norm": 0.6156827211380005, + "learning_rate": 7.58839295835894e-06, + "loss": 0.8669, + "step": 3526 + }, + { + "epoch": 0.5913444409514828, + "grad_norm": 0.480912446975708, + "learning_rate": 7.5831230131686585e-06, + "loss": 0.7717, + "step": 3527 + }, + { + "epoch": 0.5915121031122289, + "grad_norm": 0.7437429428100586, + "learning_rate": 7.577853780745324e-06, + "loss": 0.8701, + "step": 3528 + }, + { + "epoch": 0.591679765272975, + "grad_norm": 0.5265527367591858, + "learning_rate": 7.572585262642904e-06, + "loss": 0.9543, + "step": 3529 + }, + { + "epoch": 0.5918474274337211, + "grad_norm": 0.5225236415863037, + "learning_rate": 7.567317460415149e-06, + "loss": 0.8445, + "step": 3530 + }, + { + "epoch": 0.5920150895944671, + "grad_norm": 0.5767743587493896, + "learning_rate": 7.562050375615595e-06, + "loss": 0.9037, + "step": 3531 + }, + { + "epoch": 0.5921827517552133, + "grad_norm": 0.3901894688606262, + "learning_rate": 7.556784009797581e-06, + "loss": 0.7714, + "step": 3532 + }, + { + "epoch": 0.5923504139159593, + "grad_norm": 0.39337241649627686, + "learning_rate": 7.551518364514212e-06, + "loss": 0.8793, + "step": 3533 + }, + { + "epoch": 0.5925180760767055, + "grad_norm": 0.4687483608722687, + "learning_rate": 7.546253441318396e-06, + "loss": 0.9443, + "step": 3534 + }, + { + "epoch": 0.5926857382374515, + "grad_norm": 0.43475142121315, + "learning_rate": 7.5409892417628215e-06, + "loss": 0.8109, + "step": 3535 + }, + { + "epoch": 0.5928534003981977, + "grad_norm": 0.40518566966056824, + "learning_rate": 7.535725767399973e-06, + "loss": 0.8149, + "step": 3536 + }, + { + "epoch": 0.5930210625589437, + "grad_norm": 0.5907869338989258, + "learning_rate": 7.530463019782114e-06, + "loss": 0.8444, + "step": 3537 + }, + { + "epoch": 0.5931887247196899, + "grad_norm": 0.5921319127082825, + "learning_rate": 7.525201000461284e-06, + "loss": 0.8426, + "step": 3538 + }, + { + "epoch": 0.5933563868804359, + "grad_norm": 0.5731843709945679, + "learning_rate": 7.519939710989326e-06, + "loss": 0.9189, + "step": 3539 + }, + { + "epoch": 0.5935240490411821, + "grad_norm": 0.5038250684738159, + "learning_rate": 7.514679152917857e-06, + "loss": 0.7586, + "step": 3540 + }, + { + "epoch": 0.5936917112019281, + "grad_norm": 0.5103881359100342, + "learning_rate": 7.509419327798283e-06, + "loss": 0.936, + "step": 3541 + }, + { + "epoch": 0.5938593733626742, + "grad_norm": 0.4073008596897125, + "learning_rate": 7.504160237181788e-06, + "loss": 0.8011, + "step": 3542 + }, + { + "epoch": 0.5940270355234203, + "grad_norm": 0.4870217740535736, + "learning_rate": 7.498901882619345e-06, + "loss": 0.8778, + "step": 3543 + }, + { + "epoch": 0.5941946976841664, + "grad_norm": 0.4272831380367279, + "learning_rate": 7.493644265661706e-06, + "loss": 0.8946, + "step": 3544 + }, + { + "epoch": 0.5943623598449125, + "grad_norm": 0.44826629757881165, + "learning_rate": 7.48838738785941e-06, + "loss": 0.8727, + "step": 3545 + }, + { + "epoch": 0.5945300220056586, + "grad_norm": 0.3500494658946991, + "learning_rate": 7.483131250762776e-06, + "loss": 0.8728, + "step": 3546 + }, + { + "epoch": 0.5946976841664047, + "grad_norm": 0.5210344195365906, + "learning_rate": 7.477875855921899e-06, + "loss": 0.8134, + "step": 3547 + }, + { + "epoch": 0.5948653463271508, + "grad_norm": 0.5424099564552307, + "learning_rate": 7.472621204886665e-06, + "loss": 0.9176, + "step": 3548 + }, + { + "epoch": 0.5950330084878969, + "grad_norm": 0.4739905893802643, + "learning_rate": 7.467367299206733e-06, + "loss": 0.8229, + "step": 3549 + }, + { + "epoch": 0.595200670648643, + "grad_norm": 0.7030133605003357, + "learning_rate": 7.4621141404315465e-06, + "loss": 0.833, + "step": 3550 + }, + { + "epoch": 0.5953683328093891, + "grad_norm": 0.36012545228004456, + "learning_rate": 7.456861730110326e-06, + "loss": 0.8498, + "step": 3551 + }, + { + "epoch": 0.5955359949701352, + "grad_norm": 0.4297751486301422, + "learning_rate": 7.45161006979207e-06, + "loss": 0.933, + "step": 3552 + }, + { + "epoch": 0.5957036571308812, + "grad_norm": 0.5189722776412964, + "learning_rate": 7.4463591610255626e-06, + "loss": 0.7511, + "step": 3553 + }, + { + "epoch": 0.5958713192916274, + "grad_norm": 0.8779076337814331, + "learning_rate": 7.441109005359357e-06, + "loss": 0.9207, + "step": 3554 + }, + { + "epoch": 0.5960389814523734, + "grad_norm": 0.35254567861557007, + "learning_rate": 7.4358596043418e-06, + "loss": 0.825, + "step": 3555 + }, + { + "epoch": 0.5962066436131196, + "grad_norm": 0.4413334131240845, + "learning_rate": 7.430610959520991e-06, + "loss": 0.8796, + "step": 3556 + }, + { + "epoch": 0.5963743057738656, + "grad_norm": 0.6897941827774048, + "learning_rate": 7.425363072444824e-06, + "loss": 0.8764, + "step": 3557 + }, + { + "epoch": 0.5965419679346118, + "grad_norm": 0.47298622131347656, + "learning_rate": 7.420115944660971e-06, + "loss": 0.8621, + "step": 3558 + }, + { + "epoch": 0.5967096300953578, + "grad_norm": 0.4050973057746887, + "learning_rate": 7.414869577716877e-06, + "loss": 0.9009, + "step": 3559 + }, + { + "epoch": 0.596877292256104, + "grad_norm": 0.4958822429180145, + "learning_rate": 7.409623973159749e-06, + "loss": 0.8612, + "step": 3560 + }, + { + "epoch": 0.59704495441685, + "grad_norm": 0.38055795431137085, + "learning_rate": 7.404379132536588e-06, + "loss": 0.7803, + "step": 3561 + }, + { + "epoch": 0.5972126165775962, + "grad_norm": 0.8478285670280457, + "learning_rate": 7.399135057394162e-06, + "loss": 0.8753, + "step": 3562 + }, + { + "epoch": 0.5973802787383422, + "grad_norm": 0.4878985285758972, + "learning_rate": 7.393891749279015e-06, + "loss": 0.8193, + "step": 3563 + }, + { + "epoch": 0.5975479408990884, + "grad_norm": 0.5586049556732178, + "learning_rate": 7.388649209737464e-06, + "loss": 0.8218, + "step": 3564 + }, + { + "epoch": 0.5977156030598344, + "grad_norm": 0.9109604358673096, + "learning_rate": 7.383407440315595e-06, + "loss": 0.8522, + "step": 3565 + }, + { + "epoch": 0.5978832652205806, + "grad_norm": 0.4115822911262512, + "learning_rate": 7.378166442559271e-06, + "loss": 0.799, + "step": 3566 + }, + { + "epoch": 0.5980509273813266, + "grad_norm": 0.5169524550437927, + "learning_rate": 7.372926218014131e-06, + "loss": 0.8786, + "step": 3567 + }, + { + "epoch": 0.5982185895420727, + "grad_norm": 1.4872629642486572, + "learning_rate": 7.367686768225582e-06, + "loss": 0.8903, + "step": 3568 + }, + { + "epoch": 0.5983862517028188, + "grad_norm": 0.5998322367668152, + "learning_rate": 7.362448094738797e-06, + "loss": 0.7769, + "step": 3569 + }, + { + "epoch": 0.5985539138635649, + "grad_norm": 0.4917464852333069, + "learning_rate": 7.35721019909873e-06, + "loss": 0.7684, + "step": 3570 + }, + { + "epoch": 0.598721576024311, + "grad_norm": 0.4330538213253021, + "learning_rate": 7.351973082850099e-06, + "loss": 0.8178, + "step": 3571 + }, + { + "epoch": 0.5988892381850571, + "grad_norm": 0.4778382182121277, + "learning_rate": 7.346736747537397e-06, + "loss": 0.9828, + "step": 3572 + }, + { + "epoch": 0.5990569003458032, + "grad_norm": 0.3780740201473236, + "learning_rate": 7.341501194704885e-06, + "loss": 0.8749, + "step": 3573 + }, + { + "epoch": 0.5992245625065493, + "grad_norm": 0.6358554363250732, + "learning_rate": 7.336266425896589e-06, + "loss": 0.8051, + "step": 3574 + }, + { + "epoch": 0.5993922246672954, + "grad_norm": 0.5639551281929016, + "learning_rate": 7.331032442656307e-06, + "loss": 0.8786, + "step": 3575 + }, + { + "epoch": 0.5995598868280415, + "grad_norm": 0.8509765267372131, + "learning_rate": 7.325799246527609e-06, + "loss": 0.9342, + "step": 3576 + }, + { + "epoch": 0.5997275489887876, + "grad_norm": 0.4847201406955719, + "learning_rate": 7.320566839053824e-06, + "loss": 0.8358, + "step": 3577 + }, + { + "epoch": 0.5998952111495337, + "grad_norm": 1.1400126218795776, + "learning_rate": 7.315335221778064e-06, + "loss": 0.9474, + "step": 3578 + }, + { + "epoch": 0.6000628733102797, + "grad_norm": 0.5179569125175476, + "learning_rate": 7.310104396243185e-06, + "loss": 0.7884, + "step": 3579 + }, + { + "epoch": 0.6002305354710259, + "grad_norm": 0.457474946975708, + "learning_rate": 7.304874363991828e-06, + "loss": 0.858, + "step": 3580 + }, + { + "epoch": 0.6003981976317719, + "grad_norm": 0.6846543550491333, + "learning_rate": 7.299645126566395e-06, + "loss": 0.8058, + "step": 3581 + }, + { + "epoch": 0.6005658597925181, + "grad_norm": 0.4435752034187317, + "learning_rate": 7.294416685509056e-06, + "loss": 0.8896, + "step": 3582 + }, + { + "epoch": 0.6007335219532641, + "grad_norm": 0.44891929626464844, + "learning_rate": 7.289189042361736e-06, + "loss": 0.8745, + "step": 3583 + }, + { + "epoch": 0.6009011841140103, + "grad_norm": 0.4819808602333069, + "learning_rate": 7.2839621986661325e-06, + "loss": 0.8753, + "step": 3584 + }, + { + "epoch": 0.6010688462747563, + "grad_norm": 0.5853043794631958, + "learning_rate": 7.27873615596371e-06, + "loss": 0.943, + "step": 3585 + }, + { + "epoch": 0.6012365084355025, + "grad_norm": 0.4091236889362335, + "learning_rate": 7.273510915795694e-06, + "loss": 0.7616, + "step": 3586 + }, + { + "epoch": 0.6014041705962485, + "grad_norm": 0.47817444801330566, + "learning_rate": 7.2682864797030705e-06, + "loss": 0.832, + "step": 3587 + }, + { + "epoch": 0.6015718327569947, + "grad_norm": 0.44186002016067505, + "learning_rate": 7.26306284922659e-06, + "loss": 0.8801, + "step": 3588 + }, + { + "epoch": 0.6017394949177407, + "grad_norm": 0.5388672351837158, + "learning_rate": 7.257840025906766e-06, + "loss": 0.9696, + "step": 3589 + }, + { + "epoch": 0.6019071570784869, + "grad_norm": 0.4427429139614105, + "learning_rate": 7.252618011283874e-06, + "loss": 0.8297, + "step": 3590 + }, + { + "epoch": 0.6020748192392329, + "grad_norm": 0.42489099502563477, + "learning_rate": 7.247396806897953e-06, + "loss": 0.8999, + "step": 3591 + }, + { + "epoch": 0.602242481399979, + "grad_norm": 0.41747936606407166, + "learning_rate": 7.242176414288795e-06, + "loss": 0.864, + "step": 3592 + }, + { + "epoch": 0.6024101435607251, + "grad_norm": 0.4210547208786011, + "learning_rate": 7.236956834995963e-06, + "loss": 0.8419, + "step": 3593 + }, + { + "epoch": 0.6025778057214712, + "grad_norm": 0.37546658515930176, + "learning_rate": 7.231738070558775e-06, + "loss": 0.7766, + "step": 3594 + }, + { + "epoch": 0.6027454678822173, + "grad_norm": 0.38674503564834595, + "learning_rate": 7.226520122516308e-06, + "loss": 0.9229, + "step": 3595 + }, + { + "epoch": 0.6029131300429634, + "grad_norm": 0.41017404198646545, + "learning_rate": 7.221302992407404e-06, + "loss": 0.9184, + "step": 3596 + }, + { + "epoch": 0.6030807922037095, + "grad_norm": 0.5656243562698364, + "learning_rate": 7.216086681770652e-06, + "loss": 0.9822, + "step": 3597 + }, + { + "epoch": 0.6032484543644556, + "grad_norm": 0.6094668507575989, + "learning_rate": 7.210871192144411e-06, + "loss": 0.8996, + "step": 3598 + }, + { + "epoch": 0.6034161165252018, + "grad_norm": 0.45814836025238037, + "learning_rate": 7.205656525066793e-06, + "loss": 0.8522, + "step": 3599 + }, + { + "epoch": 0.6035837786859478, + "grad_norm": 0.4150603115558624, + "learning_rate": 7.200442682075668e-06, + "loss": 0.8364, + "step": 3600 + }, + { + "epoch": 0.603751440846694, + "grad_norm": 0.500847578048706, + "learning_rate": 7.195229664708663e-06, + "loss": 0.9176, + "step": 3601 + }, + { + "epoch": 0.60391910300744, + "grad_norm": 0.47346231341362, + "learning_rate": 7.190017474503159e-06, + "loss": 0.8205, + "step": 3602 + }, + { + "epoch": 0.6040867651681862, + "grad_norm": 0.6502076983451843, + "learning_rate": 7.184806112996294e-06, + "loss": 0.9446, + "step": 3603 + }, + { + "epoch": 0.6042544273289322, + "grad_norm": 0.43551987409591675, + "learning_rate": 7.179595581724971e-06, + "loss": 0.7325, + "step": 3604 + }, + { + "epoch": 0.6044220894896783, + "grad_norm": 0.8503975868225098, + "learning_rate": 7.174385882225838e-06, + "loss": 0.7947, + "step": 3605 + }, + { + "epoch": 0.6045897516504244, + "grad_norm": 0.4926724135875702, + "learning_rate": 7.169177016035293e-06, + "loss": 0.8104, + "step": 3606 + }, + { + "epoch": 0.6047574138111705, + "grad_norm": 0.6093217730522156, + "learning_rate": 7.1639689846895e-06, + "loss": 0.8905, + "step": 3607 + }, + { + "epoch": 0.6049250759719166, + "grad_norm": 0.5507587194442749, + "learning_rate": 7.158761789724371e-06, + "loss": 0.984, + "step": 3608 + }, + { + "epoch": 0.6050927381326627, + "grad_norm": 0.5029594898223877, + "learning_rate": 7.153555432675574e-06, + "loss": 0.8459, + "step": 3609 + }, + { + "epoch": 0.6052604002934088, + "grad_norm": 0.4744623899459839, + "learning_rate": 7.148349915078528e-06, + "loss": 0.9074, + "step": 3610 + }, + { + "epoch": 0.6054280624541549, + "grad_norm": 0.4237988591194153, + "learning_rate": 7.143145238468403e-06, + "loss": 0.8329, + "step": 3611 + }, + { + "epoch": 0.605595724614901, + "grad_norm": 0.3881392776966095, + "learning_rate": 7.137941404380122e-06, + "loss": 0.771, + "step": 3612 + }, + { + "epoch": 0.6057633867756471, + "grad_norm": 0.4183195233345032, + "learning_rate": 7.132738414348361e-06, + "loss": 0.8318, + "step": 3613 + }, + { + "epoch": 0.6059310489363932, + "grad_norm": 0.5632359981536865, + "learning_rate": 7.127536269907549e-06, + "loss": 0.8639, + "step": 3614 + }, + { + "epoch": 0.6060987110971393, + "grad_norm": 0.40016984939575195, + "learning_rate": 7.122334972591858e-06, + "loss": 0.9338, + "step": 3615 + }, + { + "epoch": 0.6062663732578853, + "grad_norm": 0.5154279470443726, + "learning_rate": 7.117134523935217e-06, + "loss": 0.8404, + "step": 3616 + }, + { + "epoch": 0.6064340354186315, + "grad_norm": 0.5123127102851868, + "learning_rate": 7.111934925471302e-06, + "loss": 0.8732, + "step": 3617 + }, + { + "epoch": 0.6066016975793775, + "grad_norm": 0.45707687735557556, + "learning_rate": 7.10673617873354e-06, + "loss": 0.7817, + "step": 3618 + }, + { + "epoch": 0.6067693597401237, + "grad_norm": 0.4302932024002075, + "learning_rate": 7.101538285255108e-06, + "loss": 0.9229, + "step": 3619 + }, + { + "epoch": 0.6069370219008697, + "grad_norm": 0.4706227481365204, + "learning_rate": 7.096341246568925e-06, + "loss": 0.8589, + "step": 3620 + }, + { + "epoch": 0.6071046840616159, + "grad_norm": 0.9093366265296936, + "learning_rate": 7.091145064207665e-06, + "loss": 0.8966, + "step": 3621 + }, + { + "epoch": 0.6072723462223619, + "grad_norm": 0.4529713988304138, + "learning_rate": 7.0859497397037416e-06, + "loss": 0.7916, + "step": 3622 + }, + { + "epoch": 0.6074400083831081, + "grad_norm": 0.5407692790031433, + "learning_rate": 7.080755274589332e-06, + "loss": 0.8121, + "step": 3623 + }, + { + "epoch": 0.6076076705438541, + "grad_norm": 0.3993769884109497, + "learning_rate": 7.075561670396334e-06, + "loss": 0.8102, + "step": 3624 + }, + { + "epoch": 0.6077753327046003, + "grad_norm": 0.9855776429176331, + "learning_rate": 7.070368928656413e-06, + "loss": 0.9055, + "step": 3625 + }, + { + "epoch": 0.6079429948653463, + "grad_norm": 0.4427030384540558, + "learning_rate": 7.065177050900971e-06, + "loss": 0.8445, + "step": 3626 + }, + { + "epoch": 0.6081106570260925, + "grad_norm": 0.42641207575798035, + "learning_rate": 7.05998603866116e-06, + "loss": 0.8255, + "step": 3627 + }, + { + "epoch": 0.6082783191868385, + "grad_norm": 0.42815962433815, + "learning_rate": 7.054795893467875e-06, + "loss": 0.8354, + "step": 3628 + }, + { + "epoch": 0.6084459813475847, + "grad_norm": 0.4792901575565338, + "learning_rate": 7.0496066168517455e-06, + "loss": 0.8226, + "step": 3629 + }, + { + "epoch": 0.6086136435083307, + "grad_norm": 0.5730831027030945, + "learning_rate": 7.044418210343161e-06, + "loss": 0.8705, + "step": 3630 + }, + { + "epoch": 0.6087813056690768, + "grad_norm": 0.6771920919418335, + "learning_rate": 7.039230675472244e-06, + "loss": 0.8276, + "step": 3631 + }, + { + "epoch": 0.6089489678298229, + "grad_norm": 0.5605460405349731, + "learning_rate": 7.034044013768868e-06, + "loss": 0.7989, + "step": 3632 + }, + { + "epoch": 0.609116629990569, + "grad_norm": 0.7433153390884399, + "learning_rate": 7.028858226762632e-06, + "loss": 0.9028, + "step": 3633 + }, + { + "epoch": 0.6092842921513151, + "grad_norm": 0.5418516993522644, + "learning_rate": 7.0236733159829e-06, + "loss": 0.7459, + "step": 3634 + }, + { + "epoch": 0.6094519543120612, + "grad_norm": 0.5711153745651245, + "learning_rate": 7.0184892829587605e-06, + "loss": 0.7997, + "step": 3635 + }, + { + "epoch": 0.6096196164728073, + "grad_norm": 0.5344470143318176, + "learning_rate": 7.013306129219052e-06, + "loss": 0.823, + "step": 3636 + }, + { + "epoch": 0.6097872786335534, + "grad_norm": 0.9105053544044495, + "learning_rate": 7.008123856292352e-06, + "loss": 0.8645, + "step": 3637 + }, + { + "epoch": 0.6099549407942995, + "grad_norm": 0.33319270610809326, + "learning_rate": 7.002942465706972e-06, + "loss": 0.7663, + "step": 3638 + }, + { + "epoch": 0.6101226029550456, + "grad_norm": 0.6968861222267151, + "learning_rate": 6.9977619589909705e-06, + "loss": 0.7576, + "step": 3639 + }, + { + "epoch": 0.6102902651157917, + "grad_norm": 0.6255199313163757, + "learning_rate": 6.992582337672145e-06, + "loss": 0.886, + "step": 3640 + }, + { + "epoch": 0.6104579272765378, + "grad_norm": 0.6661763787269592, + "learning_rate": 6.987403603278029e-06, + "loss": 0.8666, + "step": 3641 + }, + { + "epoch": 0.6106255894372838, + "grad_norm": 0.4097493886947632, + "learning_rate": 6.982225757335892e-06, + "loss": 0.7663, + "step": 3642 + }, + { + "epoch": 0.61079325159803, + "grad_norm": 0.48773884773254395, + "learning_rate": 6.97704880137275e-06, + "loss": 0.8584, + "step": 3643 + }, + { + "epoch": 0.610960913758776, + "grad_norm": 0.520758867263794, + "learning_rate": 6.971872736915348e-06, + "loss": 0.8782, + "step": 3644 + }, + { + "epoch": 0.6111285759195222, + "grad_norm": 0.4427630603313446, + "learning_rate": 6.9666975654901704e-06, + "loss": 0.8038, + "step": 3645 + }, + { + "epoch": 0.6112962380802682, + "grad_norm": 0.8011310696601868, + "learning_rate": 6.9615232886234475e-06, + "loss": 0.8792, + "step": 3646 + }, + { + "epoch": 0.6114639002410144, + "grad_norm": 0.5017745494842529, + "learning_rate": 6.956349907841128e-06, + "loss": 0.9043, + "step": 3647 + }, + { + "epoch": 0.6116315624017604, + "grad_norm": 0.6061432957649231, + "learning_rate": 6.951177424668909e-06, + "loss": 0.8709, + "step": 3648 + }, + { + "epoch": 0.6117992245625066, + "grad_norm": 0.4249998927116394, + "learning_rate": 6.9460058406322195e-06, + "loss": 0.9202, + "step": 3649 + }, + { + "epoch": 0.6119668867232526, + "grad_norm": 0.47136473655700684, + "learning_rate": 6.940835157256226e-06, + "loss": 0.8131, + "step": 3650 + }, + { + "epoch": 0.6121345488839988, + "grad_norm": 0.39274460077285767, + "learning_rate": 6.93566537606583e-06, + "loss": 0.9147, + "step": 3651 + }, + { + "epoch": 0.6123022110447448, + "grad_norm": 0.613770067691803, + "learning_rate": 6.930496498585654e-06, + "loss": 0.7946, + "step": 3652 + }, + { + "epoch": 0.612469873205491, + "grad_norm": 0.4516802132129669, + "learning_rate": 6.925328526340072e-06, + "loss": 0.7832, + "step": 3653 + }, + { + "epoch": 0.612637535366237, + "grad_norm": 0.44192588329315186, + "learning_rate": 6.920161460853179e-06, + "loss": 0.9041, + "step": 3654 + }, + { + "epoch": 0.6128051975269831, + "grad_norm": 0.7549390196800232, + "learning_rate": 6.914995303648811e-06, + "loss": 0.8354, + "step": 3655 + }, + { + "epoch": 0.6129728596877292, + "grad_norm": 0.49047520756721497, + "learning_rate": 6.909830056250527e-06, + "loss": 0.9369, + "step": 3656 + }, + { + "epoch": 0.6131405218484753, + "grad_norm": 0.6662545800209045, + "learning_rate": 6.9046657201816245e-06, + "loss": 0.7678, + "step": 3657 + }, + { + "epoch": 0.6133081840092214, + "grad_norm": 0.5436139702796936, + "learning_rate": 6.89950229696513e-06, + "loss": 0.8306, + "step": 3658 + }, + { + "epoch": 0.6134758461699675, + "grad_norm": 0.40033626556396484, + "learning_rate": 6.894339788123801e-06, + "loss": 0.8623, + "step": 3659 + }, + { + "epoch": 0.6136435083307136, + "grad_norm": 0.43172773718833923, + "learning_rate": 6.889178195180128e-06, + "loss": 0.799, + "step": 3660 + }, + { + "epoch": 0.6138111704914597, + "grad_norm": 0.42334967851638794, + "learning_rate": 6.884017519656323e-06, + "loss": 0.8647, + "step": 3661 + }, + { + "epoch": 0.6139788326522058, + "grad_norm": 0.37546423077583313, + "learning_rate": 6.878857763074336e-06, + "loss": 0.7751, + "step": 3662 + }, + { + "epoch": 0.6141464948129519, + "grad_norm": 0.3862164318561554, + "learning_rate": 6.873698926955844e-06, + "loss": 0.8285, + "step": 3663 + }, + { + "epoch": 0.614314156973698, + "grad_norm": 0.5159278512001038, + "learning_rate": 6.868541012822253e-06, + "loss": 0.8971, + "step": 3664 + }, + { + "epoch": 0.6144818191344441, + "grad_norm": 0.7204200625419617, + "learning_rate": 6.86338402219469e-06, + "loss": 0.8071, + "step": 3665 + }, + { + "epoch": 0.6146494812951901, + "grad_norm": 0.4890683889389038, + "learning_rate": 6.858227956594022e-06, + "loss": 0.7644, + "step": 3666 + }, + { + "epoch": 0.6148171434559363, + "grad_norm": 0.544012725353241, + "learning_rate": 6.853072817540831e-06, + "loss": 0.8454, + "step": 3667 + }, + { + "epoch": 0.6149848056166823, + "grad_norm": 0.34836164116859436, + "learning_rate": 6.847918606555433e-06, + "loss": 0.8329, + "step": 3668 + }, + { + "epoch": 0.6151524677774285, + "grad_norm": 0.5558313727378845, + "learning_rate": 6.842765325157874e-06, + "loss": 0.8326, + "step": 3669 + }, + { + "epoch": 0.6153201299381745, + "grad_norm": 0.45246636867523193, + "learning_rate": 6.837612974867913e-06, + "loss": 0.849, + "step": 3670 + }, + { + "epoch": 0.6154877920989207, + "grad_norm": 0.6189624667167664, + "learning_rate": 6.832461557205043e-06, + "loss": 0.8273, + "step": 3671 + }, + { + "epoch": 0.6156554542596667, + "grad_norm": 0.4349825084209442, + "learning_rate": 6.827311073688485e-06, + "loss": 0.853, + "step": 3672 + }, + { + "epoch": 0.6158231164204129, + "grad_norm": 0.5392482876777649, + "learning_rate": 6.822161525837183e-06, + "loss": 0.8998, + "step": 3673 + }, + { + "epoch": 0.6159907785811589, + "grad_norm": 0.3823678195476532, + "learning_rate": 6.8170129151697925e-06, + "loss": 0.8844, + "step": 3674 + }, + { + "epoch": 0.6161584407419051, + "grad_norm": 0.6585565209388733, + "learning_rate": 6.811865243204706e-06, + "loss": 0.7969, + "step": 3675 + }, + { + "epoch": 0.6163261029026511, + "grad_norm": 0.4332069456577301, + "learning_rate": 6.80671851146004e-06, + "loss": 0.8772, + "step": 3676 + }, + { + "epoch": 0.6164937650633973, + "grad_norm": 0.5370450615882874, + "learning_rate": 6.801572721453625e-06, + "loss": 0.7749, + "step": 3677 + }, + { + "epoch": 0.6166614272241433, + "grad_norm": 0.3939927816390991, + "learning_rate": 6.796427874703025e-06, + "loss": 0.9287, + "step": 3678 + }, + { + "epoch": 0.6168290893848895, + "grad_norm": 0.5207778215408325, + "learning_rate": 6.791283972725509e-06, + "loss": 0.9161, + "step": 3679 + }, + { + "epoch": 0.6169967515456355, + "grad_norm": 0.8378856182098389, + "learning_rate": 6.786141017038084e-06, + "loss": 0.8162, + "step": 3680 + }, + { + "epoch": 0.6171644137063816, + "grad_norm": 0.5842326879501343, + "learning_rate": 6.780999009157469e-06, + "loss": 0.7822, + "step": 3681 + }, + { + "epoch": 0.6173320758671277, + "grad_norm": 0.4481719136238098, + "learning_rate": 6.775857950600107e-06, + "loss": 0.9661, + "step": 3682 + }, + { + "epoch": 0.6174997380278738, + "grad_norm": 0.37426868081092834, + "learning_rate": 6.77071784288216e-06, + "loss": 0.806, + "step": 3683 + }, + { + "epoch": 0.61766740018862, + "grad_norm": 0.48538267612457275, + "learning_rate": 6.765578687519508e-06, + "loss": 1.0161, + "step": 3684 + }, + { + "epoch": 0.617835062349366, + "grad_norm": 0.6142372488975525, + "learning_rate": 6.760440486027751e-06, + "loss": 0.878, + "step": 3685 + }, + { + "epoch": 0.6180027245101122, + "grad_norm": 0.38488391041755676, + "learning_rate": 6.7553032399222104e-06, + "loss": 0.8512, + "step": 3686 + }, + { + "epoch": 0.6181703866708582, + "grad_norm": 0.43986889719963074, + "learning_rate": 6.750166950717925e-06, + "loss": 0.8792, + "step": 3687 + }, + { + "epoch": 0.6183380488316044, + "grad_norm": 0.3855423331260681, + "learning_rate": 6.745031619929646e-06, + "loss": 0.8564, + "step": 3688 + }, + { + "epoch": 0.6185057109923504, + "grad_norm": 0.446071058511734, + "learning_rate": 6.739897249071847e-06, + "loss": 0.8781, + "step": 3689 + }, + { + "epoch": 0.6186733731530966, + "grad_norm": 0.5075111985206604, + "learning_rate": 6.73476383965872e-06, + "loss": 0.9628, + "step": 3690 + }, + { + "epoch": 0.6188410353138426, + "grad_norm": 0.3769055902957916, + "learning_rate": 6.729631393204167e-06, + "loss": 0.8846, + "step": 3691 + }, + { + "epoch": 0.6190086974745888, + "grad_norm": 0.9417046904563904, + "learning_rate": 6.7244999112218175e-06, + "loss": 0.9728, + "step": 3692 + }, + { + "epoch": 0.6191763596353348, + "grad_norm": 0.33916711807250977, + "learning_rate": 6.719369395225002e-06, + "loss": 0.7993, + "step": 3693 + }, + { + "epoch": 0.619344021796081, + "grad_norm": 0.5961330533027649, + "learning_rate": 6.714239846726771e-06, + "loss": 0.8623, + "step": 3694 + }, + { + "epoch": 0.619511683956827, + "grad_norm": 0.6619173884391785, + "learning_rate": 6.7091112672399e-06, + "loss": 0.8258, + "step": 3695 + }, + { + "epoch": 0.6196793461175731, + "grad_norm": 0.45241880416870117, + "learning_rate": 6.70398365827687e-06, + "loss": 0.8904, + "step": 3696 + }, + { + "epoch": 0.6198470082783192, + "grad_norm": 0.4222065806388855, + "learning_rate": 6.69885702134987e-06, + "loss": 0.9054, + "step": 3697 + }, + { + "epoch": 0.6200146704390653, + "grad_norm": 0.4871988594532013, + "learning_rate": 6.6937313579708074e-06, + "loss": 0.9105, + "step": 3698 + }, + { + "epoch": 0.6201823325998114, + "grad_norm": 0.4782107174396515, + "learning_rate": 6.688606669651312e-06, + "loss": 0.8084, + "step": 3699 + }, + { + "epoch": 0.6203499947605575, + "grad_norm": 0.3759131133556366, + "learning_rate": 6.683482957902714e-06, + "loss": 0.7746, + "step": 3700 + }, + { + "epoch": 0.6205176569213036, + "grad_norm": 0.46672704815864563, + "learning_rate": 6.678360224236062e-06, + "loss": 0.8601, + "step": 3701 + }, + { + "epoch": 0.6206853190820497, + "grad_norm": 0.46659931540489197, + "learning_rate": 6.673238470162109e-06, + "loss": 0.8369, + "step": 3702 + }, + { + "epoch": 0.6208529812427958, + "grad_norm": 0.5919094681739807, + "learning_rate": 6.668117697191325e-06, + "loss": 0.8311, + "step": 3703 + }, + { + "epoch": 0.6210206434035419, + "grad_norm": 0.430733323097229, + "learning_rate": 6.66299790683389e-06, + "loss": 0.8729, + "step": 3704 + }, + { + "epoch": 0.621188305564288, + "grad_norm": 0.38417741656303406, + "learning_rate": 6.657879100599697e-06, + "loss": 0.832, + "step": 3705 + }, + { + "epoch": 0.6213559677250341, + "grad_norm": 0.611831784248352, + "learning_rate": 6.652761279998338e-06, + "loss": 0.853, + "step": 3706 + }, + { + "epoch": 0.6215236298857801, + "grad_norm": 0.48803794384002686, + "learning_rate": 6.647644446539129e-06, + "loss": 0.805, + "step": 3707 + }, + { + "epoch": 0.6216912920465263, + "grad_norm": 0.9284753799438477, + "learning_rate": 6.642528601731082e-06, + "loss": 0.923, + "step": 3708 + }, + { + "epoch": 0.6218589542072723, + "grad_norm": 0.4466826915740967, + "learning_rate": 6.6374137470829256e-06, + "loss": 0.7906, + "step": 3709 + }, + { + "epoch": 0.6220266163680185, + "grad_norm": 0.6097007393836975, + "learning_rate": 6.632299884103096e-06, + "loss": 0.8859, + "step": 3710 + }, + { + "epoch": 0.6221942785287645, + "grad_norm": 0.5663206577301025, + "learning_rate": 6.62718701429973e-06, + "loss": 0.8152, + "step": 3711 + }, + { + "epoch": 0.6223619406895107, + "grad_norm": 0.42629796266555786, + "learning_rate": 6.622075139180678e-06, + "loss": 0.8714, + "step": 3712 + }, + { + "epoch": 0.6225296028502567, + "grad_norm": 0.548659086227417, + "learning_rate": 6.616964260253496e-06, + "loss": 0.9259, + "step": 3713 + }, + { + "epoch": 0.6226972650110029, + "grad_norm": 0.8277711272239685, + "learning_rate": 6.6118543790254465e-06, + "loss": 1.0303, + "step": 3714 + }, + { + "epoch": 0.6228649271717489, + "grad_norm": 0.3924505412578583, + "learning_rate": 6.606745497003495e-06, + "loss": 0.8514, + "step": 3715 + }, + { + "epoch": 0.6230325893324951, + "grad_norm": 0.4219715893268585, + "learning_rate": 6.601637615694313e-06, + "loss": 0.7878, + "step": 3716 + }, + { + "epoch": 0.6232002514932411, + "grad_norm": 0.4186965823173523, + "learning_rate": 6.596530736604278e-06, + "loss": 0.8878, + "step": 3717 + }, + { + "epoch": 0.6233679136539872, + "grad_norm": 1.0616917610168457, + "learning_rate": 6.591424861239475e-06, + "loss": 0.7697, + "step": 3718 + }, + { + "epoch": 0.6235355758147333, + "grad_norm": 0.6864936947822571, + "learning_rate": 6.586319991105692e-06, + "loss": 0.914, + "step": 3719 + }, + { + "epoch": 0.6237032379754794, + "grad_norm": 0.5054576992988586, + "learning_rate": 6.58121612770841e-06, + "loss": 0.8723, + "step": 3720 + }, + { + "epoch": 0.6238709001362255, + "grad_norm": 0.6317015886306763, + "learning_rate": 6.5761132725528265e-06, + "loss": 0.8447, + "step": 3721 + }, + { + "epoch": 0.6240385622969716, + "grad_norm": 0.5034641623497009, + "learning_rate": 6.571011427143837e-06, + "loss": 0.7845, + "step": 3722 + }, + { + "epoch": 0.6242062244577177, + "grad_norm": 0.5249319672584534, + "learning_rate": 6.565910592986038e-06, + "loss": 0.8479, + "step": 3723 + }, + { + "epoch": 0.6243738866184638, + "grad_norm": 0.47136643528938293, + "learning_rate": 6.5608107715837295e-06, + "loss": 0.8429, + "step": 3724 + }, + { + "epoch": 0.6245415487792099, + "grad_norm": 0.7187997698783875, + "learning_rate": 6.5557119644409115e-06, + "loss": 0.8155, + "step": 3725 + }, + { + "epoch": 0.624709210939956, + "grad_norm": 0.32335126399993896, + "learning_rate": 6.550614173061285e-06, + "loss": 0.7963, + "step": 3726 + }, + { + "epoch": 0.6248768731007021, + "grad_norm": 0.4105938673019409, + "learning_rate": 6.54551739894825e-06, + "loss": 0.8302, + "step": 3727 + }, + { + "epoch": 0.6250445352614482, + "grad_norm": 0.5323469042778015, + "learning_rate": 6.540421643604914e-06, + "loss": 0.8728, + "step": 3728 + }, + { + "epoch": 0.6252121974221942, + "grad_norm": 0.9208778142929077, + "learning_rate": 6.535326908534072e-06, + "loss": 0.838, + "step": 3729 + }, + { + "epoch": 0.6253798595829404, + "grad_norm": 0.5084686875343323, + "learning_rate": 6.530233195238227e-06, + "loss": 0.8912, + "step": 3730 + }, + { + "epoch": 0.6255475217436864, + "grad_norm": 0.5037923455238342, + "learning_rate": 6.525140505219578e-06, + "loss": 0.8224, + "step": 3731 + }, + { + "epoch": 0.6257151839044326, + "grad_norm": 0.6251398921012878, + "learning_rate": 6.520048839980023e-06, + "loss": 0.8781, + "step": 3732 + }, + { + "epoch": 0.6258828460651786, + "grad_norm": 0.4075562059879303, + "learning_rate": 6.5149582010211595e-06, + "loss": 0.9049, + "step": 3733 + }, + { + "epoch": 0.6260505082259248, + "grad_norm": 0.6155499219894409, + "learning_rate": 6.509868589844274e-06, + "loss": 0.8914, + "step": 3734 + }, + { + "epoch": 0.6262181703866708, + "grad_norm": 0.465683251619339, + "learning_rate": 6.50478000795036e-06, + "loss": 0.7889, + "step": 3735 + }, + { + "epoch": 0.626385832547417, + "grad_norm": 0.4047546684741974, + "learning_rate": 6.499692456840101e-06, + "loss": 0.8472, + "step": 3736 + }, + { + "epoch": 0.626553494708163, + "grad_norm": 0.5044263005256653, + "learning_rate": 6.494605938013883e-06, + "loss": 0.7842, + "step": 3737 + }, + { + "epoch": 0.6267211568689092, + "grad_norm": 0.39084577560424805, + "learning_rate": 6.489520452971778e-06, + "loss": 0.8353, + "step": 3738 + }, + { + "epoch": 0.6268888190296552, + "grad_norm": 0.495937705039978, + "learning_rate": 6.484436003213561e-06, + "loss": 0.7824, + "step": 3739 + }, + { + "epoch": 0.6270564811904014, + "grad_norm": 0.37145090103149414, + "learning_rate": 6.479352590238697e-06, + "loss": 0.8204, + "step": 3740 + }, + { + "epoch": 0.6272241433511474, + "grad_norm": 0.5771486163139343, + "learning_rate": 6.474270215546352e-06, + "loss": 0.9038, + "step": 3741 + }, + { + "epoch": 0.6273918055118936, + "grad_norm": 0.49100837111473083, + "learning_rate": 6.469188880635384e-06, + "loss": 0.8524, + "step": 3742 + }, + { + "epoch": 0.6275594676726396, + "grad_norm": 0.368540495634079, + "learning_rate": 6.46410858700433e-06, + "loss": 0.8622, + "step": 3743 + }, + { + "epoch": 0.6277271298333857, + "grad_norm": 0.5739620923995972, + "learning_rate": 6.4590293361514385e-06, + "loss": 0.7969, + "step": 3744 + }, + { + "epoch": 0.6278947919941318, + "grad_norm": 0.39057913422584534, + "learning_rate": 6.453951129574644e-06, + "loss": 0.8991, + "step": 3745 + }, + { + "epoch": 0.6280624541548779, + "grad_norm": 0.5076261162757874, + "learning_rate": 6.448873968771575e-06, + "loss": 0.8836, + "step": 3746 + }, + { + "epoch": 0.628230116315624, + "grad_norm": 0.5153020620346069, + "learning_rate": 6.44379785523954e-06, + "loss": 0.8959, + "step": 3747 + }, + { + "epoch": 0.6283977784763701, + "grad_norm": 0.4407910704612732, + "learning_rate": 6.438722790475555e-06, + "loss": 0.8904, + "step": 3748 + }, + { + "epoch": 0.6285654406371162, + "grad_norm": 0.40577349066734314, + "learning_rate": 6.433648775976317e-06, + "loss": 0.9416, + "step": 3749 + }, + { + "epoch": 0.6287331027978623, + "grad_norm": 0.7215719819068909, + "learning_rate": 6.428575813238218e-06, + "loss": 0.8255, + "step": 3750 + }, + { + "epoch": 0.6289007649586084, + "grad_norm": 0.5980567932128906, + "learning_rate": 6.423503903757338e-06, + "loss": 0.7864, + "step": 3751 + }, + { + "epoch": 0.6290684271193545, + "grad_norm": 0.9818123579025269, + "learning_rate": 6.418433049029441e-06, + "loss": 0.9186, + "step": 3752 + }, + { + "epoch": 0.6292360892801006, + "grad_norm": 0.935444712638855, + "learning_rate": 6.4133632505499876e-06, + "loss": 0.7366, + "step": 3753 + }, + { + "epoch": 0.6294037514408467, + "grad_norm": 0.4127931594848633, + "learning_rate": 6.4082945098141244e-06, + "loss": 0.8591, + "step": 3754 + }, + { + "epoch": 0.6295714136015927, + "grad_norm": 0.38492462038993835, + "learning_rate": 6.403226828316686e-06, + "loss": 0.8389, + "step": 3755 + }, + { + "epoch": 0.6297390757623389, + "grad_norm": 0.4961259365081787, + "learning_rate": 6.398160207552196e-06, + "loss": 0.8869, + "step": 3756 + }, + { + "epoch": 0.6299067379230849, + "grad_norm": 0.4814993739128113, + "learning_rate": 6.39309464901486e-06, + "loss": 0.8744, + "step": 3757 + }, + { + "epoch": 0.6300744000838311, + "grad_norm": 0.6155486702919006, + "learning_rate": 6.3880301541985746e-06, + "loss": 0.915, + "step": 3758 + }, + { + "epoch": 0.6302420622445771, + "grad_norm": 0.4084378778934479, + "learning_rate": 6.382966724596922e-06, + "loss": 0.8946, + "step": 3759 + }, + { + "epoch": 0.6304097244053233, + "grad_norm": 0.5810097455978394, + "learning_rate": 6.3779043617031775e-06, + "loss": 0.7778, + "step": 3760 + }, + { + "epoch": 0.6305773865660693, + "grad_norm": 0.6305485963821411, + "learning_rate": 6.372843067010284e-06, + "loss": 0.9236, + "step": 3761 + }, + { + "epoch": 0.6307450487268155, + "grad_norm": 0.7881393432617188, + "learning_rate": 6.367782842010885e-06, + "loss": 0.8842, + "step": 3762 + }, + { + "epoch": 0.6309127108875615, + "grad_norm": 0.6298486590385437, + "learning_rate": 6.3627236881973e-06, + "loss": 0.9203, + "step": 3763 + }, + { + "epoch": 0.6310803730483077, + "grad_norm": 0.5325604677200317, + "learning_rate": 6.357665607061542e-06, + "loss": 0.7429, + "step": 3764 + }, + { + "epoch": 0.6312480352090537, + "grad_norm": 0.4600087106227875, + "learning_rate": 6.352608600095302e-06, + "loss": 0.8481, + "step": 3765 + }, + { + "epoch": 0.6314156973697999, + "grad_norm": 0.5543720722198486, + "learning_rate": 6.3475526687899445e-06, + "loss": 0.9321, + "step": 3766 + }, + { + "epoch": 0.6315833595305459, + "grad_norm": 0.4753662049770355, + "learning_rate": 6.342497814636536e-06, + "loss": 0.8124, + "step": 3767 + }, + { + "epoch": 0.631751021691292, + "grad_norm": 0.42197152972221375, + "learning_rate": 6.3374440391258105e-06, + "loss": 0.8414, + "step": 3768 + }, + { + "epoch": 0.6319186838520381, + "grad_norm": 0.45187443494796753, + "learning_rate": 6.332391343748196e-06, + "loss": 0.8752, + "step": 3769 + }, + { + "epoch": 0.6320863460127842, + "grad_norm": 0.41934412717819214, + "learning_rate": 6.327339729993783e-06, + "loss": 0.9122, + "step": 3770 + }, + { + "epoch": 0.6322540081735304, + "grad_norm": 0.7022708654403687, + "learning_rate": 6.322289199352364e-06, + "loss": 0.918, + "step": 3771 + }, + { + "epoch": 0.6324216703342764, + "grad_norm": 0.5350022315979004, + "learning_rate": 6.3172397533134e-06, + "loss": 0.846, + "step": 3772 + }, + { + "epoch": 0.6325893324950226, + "grad_norm": 0.4491613805294037, + "learning_rate": 6.312191393366036e-06, + "loss": 0.8767, + "step": 3773 + }, + { + "epoch": 0.6327569946557686, + "grad_norm": 0.3387396037578583, + "learning_rate": 6.307144120999097e-06, + "loss": 0.7956, + "step": 3774 + }, + { + "epoch": 0.6329246568165148, + "grad_norm": 0.6551179885864258, + "learning_rate": 6.302097937701083e-06, + "loss": 0.867, + "step": 3775 + }, + { + "epoch": 0.6330923189772608, + "grad_norm": 0.467792809009552, + "learning_rate": 6.297052844960178e-06, + "loss": 0.8182, + "step": 3776 + }, + { + "epoch": 0.633259981138007, + "grad_norm": 0.5014797449111938, + "learning_rate": 6.292008844264242e-06, + "loss": 0.7121, + "step": 3777 + }, + { + "epoch": 0.633427643298753, + "grad_norm": 0.39362525939941406, + "learning_rate": 6.286965937100813e-06, + "loss": 0.8279, + "step": 3778 + }, + { + "epoch": 0.6335953054594992, + "grad_norm": 0.3900938630104065, + "learning_rate": 6.281924124957107e-06, + "loss": 0.7894, + "step": 3779 + }, + { + "epoch": 0.6337629676202452, + "grad_norm": 1.166131854057312, + "learning_rate": 6.276883409320016e-06, + "loss": 0.7686, + "step": 3780 + }, + { + "epoch": 0.6339306297809914, + "grad_norm": 0.668857753276825, + "learning_rate": 6.271843791676111e-06, + "loss": 0.8604, + "step": 3781 + }, + { + "epoch": 0.6340982919417374, + "grad_norm": 0.37729379534721375, + "learning_rate": 6.266805273511635e-06, + "loss": 0.9143, + "step": 3782 + }, + { + "epoch": 0.6342659541024835, + "grad_norm": 0.39083096385002136, + "learning_rate": 6.261767856312517e-06, + "loss": 0.9144, + "step": 3783 + }, + { + "epoch": 0.6344336162632296, + "grad_norm": 0.46704787015914917, + "learning_rate": 6.256731541564344e-06, + "loss": 0.9181, + "step": 3784 + }, + { + "epoch": 0.6346012784239757, + "grad_norm": 0.3930618166923523, + "learning_rate": 6.25169633075239e-06, + "loss": 0.7963, + "step": 3785 + }, + { + "epoch": 0.6347689405847218, + "grad_norm": 0.6432356834411621, + "learning_rate": 6.246662225361603e-06, + "loss": 0.9242, + "step": 3786 + }, + { + "epoch": 0.6349366027454679, + "grad_norm": 0.4379623830318451, + "learning_rate": 6.2416292268766075e-06, + "loss": 0.7977, + "step": 3787 + }, + { + "epoch": 0.635104264906214, + "grad_norm": 0.5398340225219727, + "learning_rate": 6.236597336781689e-06, + "loss": 0.8101, + "step": 3788 + }, + { + "epoch": 0.6352719270669601, + "grad_norm": 0.4199172854423523, + "learning_rate": 6.231566556560814e-06, + "loss": 0.7766, + "step": 3789 + }, + { + "epoch": 0.6354395892277062, + "grad_norm": 0.47107189893722534, + "learning_rate": 6.226536887697628e-06, + "loss": 0.8549, + "step": 3790 + }, + { + "epoch": 0.6356072513884523, + "grad_norm": 0.3926093578338623, + "learning_rate": 6.221508331675441e-06, + "loss": 0.8375, + "step": 3791 + }, + { + "epoch": 0.6357749135491984, + "grad_norm": 0.7772989869117737, + "learning_rate": 6.216480889977236e-06, + "loss": 0.8615, + "step": 3792 + }, + { + "epoch": 0.6359425757099445, + "grad_norm": 0.7368307709693909, + "learning_rate": 6.211454564085665e-06, + "loss": 0.9141, + "step": 3793 + }, + { + "epoch": 0.6361102378706905, + "grad_norm": 0.6034929752349854, + "learning_rate": 6.206429355483056e-06, + "loss": 0.871, + "step": 3794 + }, + { + "epoch": 0.6362779000314367, + "grad_norm": 0.6859830021858215, + "learning_rate": 6.201405265651406e-06, + "loss": 0.8354, + "step": 3795 + }, + { + "epoch": 0.6364455621921827, + "grad_norm": 0.5110700726509094, + "learning_rate": 6.1963822960723805e-06, + "loss": 0.8309, + "step": 3796 + }, + { + "epoch": 0.6366132243529289, + "grad_norm": 0.43675047159194946, + "learning_rate": 6.191360448227317e-06, + "loss": 0.7561, + "step": 3797 + }, + { + "epoch": 0.6367808865136749, + "grad_norm": 0.8488208055496216, + "learning_rate": 6.186339723597216e-06, + "loss": 0.8132, + "step": 3798 + }, + { + "epoch": 0.6369485486744211, + "grad_norm": 0.754406213760376, + "learning_rate": 6.181320123662755e-06, + "loss": 0.8205, + "step": 3799 + }, + { + "epoch": 0.6371162108351671, + "grad_norm": 0.6370114088058472, + "learning_rate": 6.176301649904276e-06, + "loss": 0.8229, + "step": 3800 + }, + { + "epoch": 0.6372838729959133, + "grad_norm": 0.6114148497581482, + "learning_rate": 6.17128430380179e-06, + "loss": 0.8499, + "step": 3801 + }, + { + "epoch": 0.6374515351566593, + "grad_norm": 0.4482009708881378, + "learning_rate": 6.166268086834972e-06, + "loss": 0.8606, + "step": 3802 + }, + { + "epoch": 0.6376191973174055, + "grad_norm": 0.4591228663921356, + "learning_rate": 6.1612530004831675e-06, + "loss": 0.87, + "step": 3803 + }, + { + "epoch": 0.6377868594781515, + "grad_norm": 0.37000200152397156, + "learning_rate": 6.156239046225386e-06, + "loss": 0.8392, + "step": 3804 + }, + { + "epoch": 0.6379545216388977, + "grad_norm": 0.417057067155838, + "learning_rate": 6.151226225540306e-06, + "loss": 0.8257, + "step": 3805 + }, + { + "epoch": 0.6381221837996437, + "grad_norm": 0.5155734419822693, + "learning_rate": 6.146214539906277e-06, + "loss": 0.8176, + "step": 3806 + }, + { + "epoch": 0.6382898459603898, + "grad_norm": 0.7226120829582214, + "learning_rate": 6.141203990801296e-06, + "loss": 0.8362, + "step": 3807 + }, + { + "epoch": 0.6384575081211359, + "grad_norm": 0.43286317586898804, + "learning_rate": 6.136194579703039e-06, + "loss": 0.8771, + "step": 3808 + }, + { + "epoch": 0.638625170281882, + "grad_norm": 0.6042872667312622, + "learning_rate": 6.131186308088846e-06, + "loss": 0.91, + "step": 3809 + }, + { + "epoch": 0.6387928324426281, + "grad_norm": 0.5188295841217041, + "learning_rate": 6.126179177435722e-06, + "loss": 0.8385, + "step": 3810 + }, + { + "epoch": 0.6389604946033742, + "grad_norm": 0.45364266633987427, + "learning_rate": 6.121173189220325e-06, + "loss": 0.7914, + "step": 3811 + }, + { + "epoch": 0.6391281567641203, + "grad_norm": 0.3534972369670868, + "learning_rate": 6.116168344918982e-06, + "loss": 0.857, + "step": 3812 + }, + { + "epoch": 0.6392958189248664, + "grad_norm": 0.8765437602996826, + "learning_rate": 6.11116464600769e-06, + "loss": 0.8746, + "step": 3813 + }, + { + "epoch": 0.6394634810856125, + "grad_norm": 0.4524255692958832, + "learning_rate": 6.106162093962098e-06, + "loss": 0.8807, + "step": 3814 + }, + { + "epoch": 0.6396311432463586, + "grad_norm": 0.5076004862785339, + "learning_rate": 6.101160690257523e-06, + "loss": 0.826, + "step": 3815 + }, + { + "epoch": 0.6397988054071047, + "grad_norm": 0.4958697557449341, + "learning_rate": 6.0961604363689364e-06, + "loss": 0.7956, + "step": 3816 + }, + { + "epoch": 0.6399664675678508, + "grad_norm": 0.4035363793373108, + "learning_rate": 6.091161333770978e-06, + "loss": 0.8481, + "step": 3817 + }, + { + "epoch": 0.6401341297285968, + "grad_norm": 0.5481789708137512, + "learning_rate": 6.086163383937946e-06, + "loss": 0.8822, + "step": 3818 + }, + { + "epoch": 0.640301791889343, + "grad_norm": 0.60627281665802, + "learning_rate": 6.081166588343798e-06, + "loss": 0.8073, + "step": 3819 + }, + { + "epoch": 0.640469454050089, + "grad_norm": 0.798624575138092, + "learning_rate": 6.076170948462146e-06, + "loss": 0.9205, + "step": 3820 + }, + { + "epoch": 0.6406371162108352, + "grad_norm": 0.4653504192829132, + "learning_rate": 6.07117646576627e-06, + "loss": 0.7914, + "step": 3821 + }, + { + "epoch": 0.6408047783715812, + "grad_norm": 0.7155263423919678, + "learning_rate": 6.066183141729104e-06, + "loss": 0.7911, + "step": 3822 + }, + { + "epoch": 0.6409724405323274, + "grad_norm": 1.396693468093872, + "learning_rate": 6.061190977823242e-06, + "loss": 0.8241, + "step": 3823 + }, + { + "epoch": 0.6411401026930734, + "grad_norm": 0.4433426558971405, + "learning_rate": 6.056199975520936e-06, + "loss": 0.8162, + "step": 3824 + }, + { + "epoch": 0.6413077648538196, + "grad_norm": 0.4936043620109558, + "learning_rate": 6.051210136294089e-06, + "loss": 0.8271, + "step": 3825 + }, + { + "epoch": 0.6414754270145656, + "grad_norm": 0.754987359046936, + "learning_rate": 6.046221461614271e-06, + "loss": 0.8933, + "step": 3826 + }, + { + "epoch": 0.6416430891753118, + "grad_norm": 0.6823899745941162, + "learning_rate": 6.041233952952703e-06, + "loss": 0.9588, + "step": 3827 + }, + { + "epoch": 0.6418107513360578, + "grad_norm": 0.3952237665653229, + "learning_rate": 6.036247611780265e-06, + "loss": 0.8251, + "step": 3828 + }, + { + "epoch": 0.641978413496804, + "grad_norm": 0.505317211151123, + "learning_rate": 6.031262439567487e-06, + "loss": 0.921, + "step": 3829 + }, + { + "epoch": 0.64214607565755, + "grad_norm": 0.3997824192047119, + "learning_rate": 6.026278437784558e-06, + "loss": 0.8279, + "step": 3830 + }, + { + "epoch": 0.6423137378182961, + "grad_norm": 0.39301612973213196, + "learning_rate": 6.021295607901323e-06, + "loss": 0.8972, + "step": 3831 + }, + { + "epoch": 0.6424813999790422, + "grad_norm": 0.3877316415309906, + "learning_rate": 6.016313951387283e-06, + "loss": 0.8126, + "step": 3832 + }, + { + "epoch": 0.6426490621397883, + "grad_norm": 0.8617894053459167, + "learning_rate": 6.011333469711592e-06, + "loss": 0.7885, + "step": 3833 + }, + { + "epoch": 0.6428167243005344, + "grad_norm": 0.39734694361686707, + "learning_rate": 6.006354164343047e-06, + "loss": 0.8481, + "step": 3834 + }, + { + "epoch": 0.6429843864612805, + "grad_norm": 0.529140293598175, + "learning_rate": 6.001376036750111e-06, + "loss": 0.8495, + "step": 3835 + }, + { + "epoch": 0.6431520486220266, + "grad_norm": 0.4331158399581909, + "learning_rate": 5.996399088400895e-06, + "loss": 0.8104, + "step": 3836 + }, + { + "epoch": 0.6433197107827727, + "grad_norm": 1.3587584495544434, + "learning_rate": 5.9914233207631655e-06, + "loss": 0.8862, + "step": 3837 + }, + { + "epoch": 0.6434873729435188, + "grad_norm": 0.4152337312698364, + "learning_rate": 5.986448735304339e-06, + "loss": 0.9072, + "step": 3838 + }, + { + "epoch": 0.6436550351042649, + "grad_norm": 0.4645063877105713, + "learning_rate": 5.981475333491477e-06, + "loss": 0.9365, + "step": 3839 + }, + { + "epoch": 0.643822697265011, + "grad_norm": 0.3764033019542694, + "learning_rate": 5.976503116791299e-06, + "loss": 0.78, + "step": 3840 + }, + { + "epoch": 0.6439903594257571, + "grad_norm": 0.5858711004257202, + "learning_rate": 5.971532086670174e-06, + "loss": 0.8986, + "step": 3841 + }, + { + "epoch": 0.6441580215865031, + "grad_norm": 0.5570114254951477, + "learning_rate": 5.966562244594123e-06, + "loss": 0.9575, + "step": 3842 + }, + { + "epoch": 0.6443256837472493, + "grad_norm": 0.3770409822463989, + "learning_rate": 5.9615935920288096e-06, + "loss": 0.7898, + "step": 3843 + }, + { + "epoch": 0.6444933459079953, + "grad_norm": 0.4317591190338135, + "learning_rate": 5.956626130439552e-06, + "loss": 0.8362, + "step": 3844 + }, + { + "epoch": 0.6446610080687415, + "grad_norm": 0.4174540042877197, + "learning_rate": 5.951659861291317e-06, + "loss": 0.9139, + "step": 3845 + }, + { + "epoch": 0.6448286702294875, + "grad_norm": 0.6102603673934937, + "learning_rate": 5.946694786048719e-06, + "loss": 0.7852, + "step": 3846 + }, + { + "epoch": 0.6449963323902337, + "grad_norm": 0.4747770428657532, + "learning_rate": 5.941730906176022e-06, + "loss": 0.7641, + "step": 3847 + }, + { + "epoch": 0.6451639945509797, + "grad_norm": 0.7990584373474121, + "learning_rate": 5.936768223137133e-06, + "loss": 0.7503, + "step": 3848 + }, + { + "epoch": 0.6453316567117259, + "grad_norm": 0.7170785665512085, + "learning_rate": 5.931806738395607e-06, + "loss": 0.8904, + "step": 3849 + }, + { + "epoch": 0.6454993188724719, + "grad_norm": 0.36233919858932495, + "learning_rate": 5.926846453414649e-06, + "loss": 0.7869, + "step": 3850 + }, + { + "epoch": 0.6456669810332181, + "grad_norm": 0.46254944801330566, + "learning_rate": 5.921887369657113e-06, + "loss": 0.8671, + "step": 3851 + }, + { + "epoch": 0.6458346431939641, + "grad_norm": 0.3812108039855957, + "learning_rate": 5.916929488585488e-06, + "loss": 0.8109, + "step": 3852 + }, + { + "epoch": 0.6460023053547103, + "grad_norm": 0.42729702591896057, + "learning_rate": 5.9119728116619165e-06, + "loss": 0.8988, + "step": 3853 + }, + { + "epoch": 0.6461699675154563, + "grad_norm": 0.40828776359558105, + "learning_rate": 5.907017340348182e-06, + "loss": 0.9616, + "step": 3854 + }, + { + "epoch": 0.6463376296762025, + "grad_norm": 0.7432594299316406, + "learning_rate": 5.9020630761057174e-06, + "loss": 0.9197, + "step": 3855 + }, + { + "epoch": 0.6465052918369486, + "grad_norm": 0.4486801028251648, + "learning_rate": 5.897110020395601e-06, + "loss": 0.8346, + "step": 3856 + }, + { + "epoch": 0.6466729539976946, + "grad_norm": 0.8108722567558289, + "learning_rate": 5.892158174678538e-06, + "loss": 0.841, + "step": 3857 + }, + { + "epoch": 0.6468406161584408, + "grad_norm": 0.5092005133628845, + "learning_rate": 5.887207540414899e-06, + "loss": 0.9681, + "step": 3858 + }, + { + "epoch": 0.6470082783191868, + "grad_norm": 0.4808082580566406, + "learning_rate": 5.8822581190646835e-06, + "loss": 0.8459, + "step": 3859 + }, + { + "epoch": 0.647175940479933, + "grad_norm": 0.5065873861312866, + "learning_rate": 5.877309912087542e-06, + "loss": 0.8047, + "step": 3860 + }, + { + "epoch": 0.647343602640679, + "grad_norm": 0.8511838316917419, + "learning_rate": 5.872362920942751e-06, + "loss": 0.861, + "step": 3861 + }, + { + "epoch": 0.6475112648014252, + "grad_norm": 0.4112839996814728, + "learning_rate": 5.8674171470892506e-06, + "loss": 0.8408, + "step": 3862 + }, + { + "epoch": 0.6476789269621712, + "grad_norm": 0.6125646829605103, + "learning_rate": 5.862472591985605e-06, + "loss": 0.833, + "step": 3863 + }, + { + "epoch": 0.6478465891229174, + "grad_norm": 0.378488689661026, + "learning_rate": 5.857529257090027e-06, + "loss": 0.844, + "step": 3864 + }, + { + "epoch": 0.6480142512836634, + "grad_norm": 0.8819833993911743, + "learning_rate": 5.852587143860368e-06, + "loss": 0.9019, + "step": 3865 + }, + { + "epoch": 0.6481819134444096, + "grad_norm": 0.43267491459846497, + "learning_rate": 5.847646253754115e-06, + "loss": 0.8093, + "step": 3866 + }, + { + "epoch": 0.6483495756051556, + "grad_norm": 0.5931116938591003, + "learning_rate": 5.8427065882284e-06, + "loss": 0.8261, + "step": 3867 + }, + { + "epoch": 0.6485172377659018, + "grad_norm": 0.6502438187599182, + "learning_rate": 5.837768148739989e-06, + "loss": 0.8822, + "step": 3868 + }, + { + "epoch": 0.6486848999266478, + "grad_norm": 0.4747580289840698, + "learning_rate": 5.832830936745294e-06, + "loss": 0.9101, + "step": 3869 + }, + { + "epoch": 0.648852562087394, + "grad_norm": 0.5370930433273315, + "learning_rate": 5.827894953700357e-06, + "loss": 0.9073, + "step": 3870 + }, + { + "epoch": 0.64902022424814, + "grad_norm": 0.6176109313964844, + "learning_rate": 5.822960201060858e-06, + "loss": 0.815, + "step": 3871 + }, + { + "epoch": 0.6491878864088861, + "grad_norm": 0.4676288068294525, + "learning_rate": 5.818026680282123e-06, + "loss": 0.8403, + "step": 3872 + }, + { + "epoch": 0.6493555485696322, + "grad_norm": 0.45306962728500366, + "learning_rate": 5.8130943928191e-06, + "loss": 0.7817, + "step": 3873 + }, + { + "epoch": 0.6495232107303783, + "grad_norm": 0.6352096199989319, + "learning_rate": 5.80816334012639e-06, + "loss": 0.7436, + "step": 3874 + }, + { + "epoch": 0.6496908728911244, + "grad_norm": 0.6790918111801147, + "learning_rate": 5.803233523658219e-06, + "loss": 0.847, + "step": 3875 + }, + { + "epoch": 0.6498585350518705, + "grad_norm": 0.43328049778938293, + "learning_rate": 5.7983049448684444e-06, + "loss": 0.8527, + "step": 3876 + }, + { + "epoch": 0.6500261972126166, + "grad_norm": 0.4411693513393402, + "learning_rate": 5.793377605210575e-06, + "loss": 0.7853, + "step": 3877 + }, + { + "epoch": 0.6501938593733627, + "grad_norm": 0.46323803067207336, + "learning_rate": 5.788451506137736e-06, + "loss": 0.8508, + "step": 3878 + }, + { + "epoch": 0.6503615215341088, + "grad_norm": 0.7709136009216309, + "learning_rate": 5.783526649102701e-06, + "loss": 0.8091, + "step": 3879 + }, + { + "epoch": 0.6505291836948549, + "grad_norm": 0.44342300295829773, + "learning_rate": 5.7786030355578704e-06, + "loss": 0.8796, + "step": 3880 + }, + { + "epoch": 0.650696845855601, + "grad_norm": 0.7442541718482971, + "learning_rate": 5.773680666955272e-06, + "loss": 0.8136, + "step": 3881 + }, + { + "epoch": 0.6508645080163471, + "grad_norm": 0.6671350002288818, + "learning_rate": 5.768759544746585e-06, + "loss": 0.8501, + "step": 3882 + }, + { + "epoch": 0.6510321701770931, + "grad_norm": 0.4241235554218292, + "learning_rate": 5.7638396703831e-06, + "loss": 0.8782, + "step": 3883 + }, + { + "epoch": 0.6511998323378393, + "grad_norm": 0.6073123812675476, + "learning_rate": 5.758921045315751e-06, + "loss": 0.7927, + "step": 3884 + }, + { + "epoch": 0.6513674944985853, + "grad_norm": 0.48959341645240784, + "learning_rate": 5.754003670995104e-06, + "loss": 0.8589, + "step": 3885 + }, + { + "epoch": 0.6515351566593315, + "grad_norm": 0.3820074200630188, + "learning_rate": 5.749087548871349e-06, + "loss": 0.8576, + "step": 3886 + }, + { + "epoch": 0.6517028188200775, + "grad_norm": 0.473310649394989, + "learning_rate": 5.744172680394313e-06, + "loss": 0.7948, + "step": 3887 + }, + { + "epoch": 0.6518704809808237, + "grad_norm": 0.4938386082649231, + "learning_rate": 5.73925906701346e-06, + "loss": 0.8689, + "step": 3888 + }, + { + "epoch": 0.6520381431415697, + "grad_norm": 0.46223780512809753, + "learning_rate": 5.734346710177863e-06, + "loss": 0.8193, + "step": 3889 + }, + { + "epoch": 0.6522058053023159, + "grad_norm": 0.34344786405563354, + "learning_rate": 5.729435611336239e-06, + "loss": 0.8681, + "step": 3890 + }, + { + "epoch": 0.6523734674630619, + "grad_norm": 0.4444260895252228, + "learning_rate": 5.724525771936941e-06, + "loss": 0.8012, + "step": 3891 + }, + { + "epoch": 0.6525411296238081, + "grad_norm": 0.4069541096687317, + "learning_rate": 5.719617193427933e-06, + "loss": 0.8493, + "step": 3892 + }, + { + "epoch": 0.6527087917845541, + "grad_norm": 0.3641412854194641, + "learning_rate": 5.714709877256817e-06, + "loss": 0.8897, + "step": 3893 + }, + { + "epoch": 0.6528764539453003, + "grad_norm": 0.5475788712501526, + "learning_rate": 5.709803824870823e-06, + "loss": 0.8944, + "step": 3894 + }, + { + "epoch": 0.6530441161060463, + "grad_norm": 0.8999494314193726, + "learning_rate": 5.704899037716803e-06, + "loss": 0.8227, + "step": 3895 + }, + { + "epoch": 0.6532117782667924, + "grad_norm": 0.5007481575012207, + "learning_rate": 5.699995517241245e-06, + "loss": 0.9822, + "step": 3896 + }, + { + "epoch": 0.6533794404275385, + "grad_norm": 0.46740853786468506, + "learning_rate": 5.695093264890258e-06, + "loss": 0.8919, + "step": 3897 + }, + { + "epoch": 0.6535471025882846, + "grad_norm": 0.4734141230583191, + "learning_rate": 5.690192282109568e-06, + "loss": 0.9623, + "step": 3898 + }, + { + "epoch": 0.6537147647490307, + "grad_norm": 0.4099631905555725, + "learning_rate": 5.685292570344547e-06, + "loss": 0.9021, + "step": 3899 + }, + { + "epoch": 0.6538824269097768, + "grad_norm": 0.44652417302131653, + "learning_rate": 5.68039413104017e-06, + "loss": 0.8015, + "step": 3900 + }, + { + "epoch": 0.6540500890705229, + "grad_norm": 0.3861241638660431, + "learning_rate": 5.675496965641057e-06, + "loss": 0.8545, + "step": 3901 + }, + { + "epoch": 0.654217751231269, + "grad_norm": 0.32507607340812683, + "learning_rate": 5.6706010755914375e-06, + "loss": 0.7932, + "step": 3902 + }, + { + "epoch": 0.6543854133920151, + "grad_norm": 0.403301864862442, + "learning_rate": 5.665706462335167e-06, + "loss": 0.864, + "step": 3903 + }, + { + "epoch": 0.6545530755527612, + "grad_norm": 0.34610313177108765, + "learning_rate": 5.6608131273157366e-06, + "loss": 0.8423, + "step": 3904 + }, + { + "epoch": 0.6547207377135073, + "grad_norm": 0.33153748512268066, + "learning_rate": 5.655921071976242e-06, + "loss": 0.8157, + "step": 3905 + }, + { + "epoch": 0.6548883998742534, + "grad_norm": 0.5666701793670654, + "learning_rate": 5.651030297759419e-06, + "loss": 0.8519, + "step": 3906 + }, + { + "epoch": 0.6550560620349994, + "grad_norm": 0.658186137676239, + "learning_rate": 5.646140806107614e-06, + "loss": 0.9748, + "step": 3907 + }, + { + "epoch": 0.6552237241957456, + "grad_norm": 0.5197761654853821, + "learning_rate": 5.641252598462794e-06, + "loss": 0.8331, + "step": 3908 + }, + { + "epoch": 0.6553913863564916, + "grad_norm": 0.38345766067504883, + "learning_rate": 5.636365676266556e-06, + "loss": 0.8573, + "step": 3909 + }, + { + "epoch": 0.6555590485172378, + "grad_norm": 0.5214496850967407, + "learning_rate": 5.6314800409601176e-06, + "loss": 0.7158, + "step": 3910 + }, + { + "epoch": 0.6557267106779838, + "grad_norm": 0.43563833832740784, + "learning_rate": 5.6265956939843116e-06, + "loss": 0.7852, + "step": 3911 + }, + { + "epoch": 0.65589437283873, + "grad_norm": 0.5575563907623291, + "learning_rate": 5.621712636779589e-06, + "loss": 0.8713, + "step": 3912 + }, + { + "epoch": 0.656062034999476, + "grad_norm": 0.6472848057746887, + "learning_rate": 5.616830870786022e-06, + "loss": 0.8153, + "step": 3913 + }, + { + "epoch": 0.6562296971602222, + "grad_norm": 0.49268069863319397, + "learning_rate": 5.611950397443307e-06, + "loss": 0.7555, + "step": 3914 + }, + { + "epoch": 0.6563973593209682, + "grad_norm": 0.4301689863204956, + "learning_rate": 5.607071218190764e-06, + "loss": 0.9521, + "step": 3915 + }, + { + "epoch": 0.6565650214817144, + "grad_norm": 0.3915613889694214, + "learning_rate": 5.602193334467307e-06, + "loss": 0.9636, + "step": 3916 + }, + { + "epoch": 0.6567326836424604, + "grad_norm": 0.8180561661720276, + "learning_rate": 5.597316747711495e-06, + "loss": 0.7843, + "step": 3917 + }, + { + "epoch": 0.6569003458032066, + "grad_norm": 0.4958726465702057, + "learning_rate": 5.592441459361493e-06, + "loss": 0.8674, + "step": 3918 + }, + { + "epoch": 0.6570680079639526, + "grad_norm": 0.3985406458377838, + "learning_rate": 5.587567470855079e-06, + "loss": 0.8227, + "step": 3919 + }, + { + "epoch": 0.6572356701246987, + "grad_norm": 0.6212599277496338, + "learning_rate": 5.582694783629663e-06, + "loss": 0.8025, + "step": 3920 + }, + { + "epoch": 0.6574033322854448, + "grad_norm": 0.5638007521629333, + "learning_rate": 5.577823399122251e-06, + "loss": 0.7337, + "step": 3921 + }, + { + "epoch": 0.6575709944461909, + "grad_norm": 0.3695148229598999, + "learning_rate": 5.572953318769475e-06, + "loss": 0.7863, + "step": 3922 + }, + { + "epoch": 0.657738656606937, + "grad_norm": 0.40442386269569397, + "learning_rate": 5.5680845440075885e-06, + "loss": 0.8887, + "step": 3923 + }, + { + "epoch": 0.6579063187676831, + "grad_norm": 0.8187882304191589, + "learning_rate": 5.563217076272449e-06, + "loss": 0.9607, + "step": 3924 + }, + { + "epoch": 0.6580739809284292, + "grad_norm": 0.4451625347137451, + "learning_rate": 5.55835091699953e-06, + "loss": 0.8935, + "step": 3925 + }, + { + "epoch": 0.6582416430891753, + "grad_norm": 0.6073238849639893, + "learning_rate": 5.55348606762393e-06, + "loss": 0.9452, + "step": 3926 + }, + { + "epoch": 0.6584093052499214, + "grad_norm": 0.8138495087623596, + "learning_rate": 5.548622529580343e-06, + "loss": 0.7612, + "step": 3927 + }, + { + "epoch": 0.6585769674106675, + "grad_norm": 0.3803512454032898, + "learning_rate": 5.543760304303098e-06, + "loss": 0.8742, + "step": 3928 + }, + { + "epoch": 0.6587446295714136, + "grad_norm": 0.4340585470199585, + "learning_rate": 5.538899393226122e-06, + "loss": 0.8225, + "step": 3929 + }, + { + "epoch": 0.6589122917321597, + "grad_norm": 0.35208043456077576, + "learning_rate": 5.534039797782948e-06, + "loss": 0.8354, + "step": 3930 + }, + { + "epoch": 0.6590799538929057, + "grad_norm": 0.41237229108810425, + "learning_rate": 5.529181519406745e-06, + "loss": 0.8451, + "step": 3931 + }, + { + "epoch": 0.6592476160536519, + "grad_norm": 0.7041605710983276, + "learning_rate": 5.524324559530269e-06, + "loss": 0.7869, + "step": 3932 + }, + { + "epoch": 0.6594152782143979, + "grad_norm": 0.416987806558609, + "learning_rate": 5.519468919585908e-06, + "loss": 0.8778, + "step": 3933 + }, + { + "epoch": 0.6595829403751441, + "grad_norm": 0.4939228594303131, + "learning_rate": 5.514614601005641e-06, + "loss": 0.8206, + "step": 3934 + }, + { + "epoch": 0.6597506025358901, + "grad_norm": 0.4769986867904663, + "learning_rate": 5.509761605221067e-06, + "loss": 0.8511, + "step": 3935 + }, + { + "epoch": 0.6599182646966363, + "grad_norm": 0.3630758225917816, + "learning_rate": 5.504909933663398e-06, + "loss": 0.8545, + "step": 3936 + }, + { + "epoch": 0.6600859268573823, + "grad_norm": 1.8517969846725464, + "learning_rate": 5.5000595877634555e-06, + "loss": 0.9223, + "step": 3937 + }, + { + "epoch": 0.6602535890181285, + "grad_norm": 0.6270850300788879, + "learning_rate": 5.495210568951661e-06, + "loss": 0.8476, + "step": 3938 + }, + { + "epoch": 0.6604212511788745, + "grad_norm": 0.7449013590812683, + "learning_rate": 5.490362878658049e-06, + "loss": 0.884, + "step": 3939 + }, + { + "epoch": 0.6605889133396207, + "grad_norm": 0.3724488317966461, + "learning_rate": 5.485516518312269e-06, + "loss": 0.8121, + "step": 3940 + }, + { + "epoch": 0.6607565755003668, + "grad_norm": 0.6408987641334534, + "learning_rate": 5.480671489343568e-06, + "loss": 0.8105, + "step": 3941 + }, + { + "epoch": 0.6609242376611129, + "grad_norm": 0.466304749250412, + "learning_rate": 5.475827793180808e-06, + "loss": 0.9077, + "step": 3942 + }, + { + "epoch": 0.661091899821859, + "grad_norm": 0.46313750743865967, + "learning_rate": 5.470985431252455e-06, + "loss": 0.8159, + "step": 3943 + }, + { + "epoch": 0.661259561982605, + "grad_norm": 0.723298192024231, + "learning_rate": 5.4661444049865755e-06, + "loss": 0.8625, + "step": 3944 + }, + { + "epoch": 0.6614272241433512, + "grad_norm": 0.4626912772655487, + "learning_rate": 5.461304715810858e-06, + "loss": 0.7983, + "step": 3945 + }, + { + "epoch": 0.6615948863040972, + "grad_norm": 0.626879870891571, + "learning_rate": 5.456466365152577e-06, + "loss": 0.7478, + "step": 3946 + }, + { + "epoch": 0.6617625484648434, + "grad_norm": 0.6091106534004211, + "learning_rate": 5.451629354438631e-06, + "loss": 0.7975, + "step": 3947 + }, + { + "epoch": 0.6619302106255894, + "grad_norm": 0.34452587366104126, + "learning_rate": 5.4467936850955105e-06, + "loss": 0.8352, + "step": 3948 + }, + { + "epoch": 0.6620978727863356, + "grad_norm": 0.43186381459236145, + "learning_rate": 5.441959358549309e-06, + "loss": 0.9035, + "step": 3949 + }, + { + "epoch": 0.6622655349470816, + "grad_norm": 0.46442583203315735, + "learning_rate": 5.437126376225738e-06, + "loss": 0.7824, + "step": 3950 + }, + { + "epoch": 0.6624331971078278, + "grad_norm": 0.40834471583366394, + "learning_rate": 5.4322947395500955e-06, + "loss": 0.8286, + "step": 3951 + }, + { + "epoch": 0.6626008592685738, + "grad_norm": 0.40990063548088074, + "learning_rate": 5.4274644499473e-06, + "loss": 0.8926, + "step": 3952 + }, + { + "epoch": 0.66276852142932, + "grad_norm": 0.335954487323761, + "learning_rate": 5.422635508841859e-06, + "loss": 0.9434, + "step": 3953 + }, + { + "epoch": 0.662936183590066, + "grad_norm": 0.3645780384540558, + "learning_rate": 5.4178079176578836e-06, + "loss": 0.9416, + "step": 3954 + }, + { + "epoch": 0.6631038457508122, + "grad_norm": 0.46499937772750854, + "learning_rate": 5.412981677819094e-06, + "loss": 0.7645, + "step": 3955 + }, + { + "epoch": 0.6632715079115582, + "grad_norm": 0.3888450264930725, + "learning_rate": 5.408156790748813e-06, + "loss": 0.8416, + "step": 3956 + }, + { + "epoch": 0.6634391700723044, + "grad_norm": 0.5346174240112305, + "learning_rate": 5.40333325786995e-06, + "loss": 0.8997, + "step": 3957 + }, + { + "epoch": 0.6636068322330504, + "grad_norm": 0.424798846244812, + "learning_rate": 5.398511080605027e-06, + "loss": 0.7671, + "step": 3958 + }, + { + "epoch": 0.6637744943937965, + "grad_norm": 0.45263102650642395, + "learning_rate": 5.39369026037617e-06, + "loss": 0.7754, + "step": 3959 + }, + { + "epoch": 0.6639421565545426, + "grad_norm": 0.6530085206031799, + "learning_rate": 5.388870798605089e-06, + "loss": 0.7891, + "step": 3960 + }, + { + "epoch": 0.6641098187152887, + "grad_norm": 0.6561893820762634, + "learning_rate": 5.384052696713118e-06, + "loss": 0.8694, + "step": 3961 + }, + { + "epoch": 0.6642774808760348, + "grad_norm": 0.594525158405304, + "learning_rate": 5.379235956121156e-06, + "loss": 0.7908, + "step": 3962 + }, + { + "epoch": 0.6644451430367809, + "grad_norm": 0.4702717661857605, + "learning_rate": 5.37442057824973e-06, + "loss": 0.8084, + "step": 3963 + }, + { + "epoch": 0.664612805197527, + "grad_norm": 0.5309458374977112, + "learning_rate": 5.369606564518957e-06, + "loss": 0.7848, + "step": 3964 + }, + { + "epoch": 0.6647804673582731, + "grad_norm": 0.38657906651496887, + "learning_rate": 5.364793916348544e-06, + "loss": 0.8487, + "step": 3965 + }, + { + "epoch": 0.6649481295190192, + "grad_norm": 0.560144305229187, + "learning_rate": 5.359982635157799e-06, + "loss": 0.8556, + "step": 3966 + }, + { + "epoch": 0.6651157916797653, + "grad_norm": 0.47209808230400085, + "learning_rate": 5.355172722365634e-06, + "loss": 0.8144, + "step": 3967 + }, + { + "epoch": 0.6652834538405114, + "grad_norm": 0.5120728015899658, + "learning_rate": 5.3503641793905485e-06, + "loss": 0.9506, + "step": 3968 + }, + { + "epoch": 0.6654511160012575, + "grad_norm": 0.38565343618392944, + "learning_rate": 5.345557007650642e-06, + "loss": 0.8712, + "step": 3969 + }, + { + "epoch": 0.6656187781620035, + "grad_norm": 0.5157579183578491, + "learning_rate": 5.340751208563611e-06, + "loss": 0.8789, + "step": 3970 + }, + { + "epoch": 0.6657864403227497, + "grad_norm": 0.5449783205986023, + "learning_rate": 5.3359467835467394e-06, + "loss": 0.809, + "step": 3971 + }, + { + "epoch": 0.6659541024834957, + "grad_norm": 0.4267489016056061, + "learning_rate": 5.33114373401692e-06, + "loss": 0.8642, + "step": 3972 + }, + { + "epoch": 0.6661217646442419, + "grad_norm": 0.38633015751838684, + "learning_rate": 5.326342061390621e-06, + "loss": 0.8654, + "step": 3973 + }, + { + "epoch": 0.6662894268049879, + "grad_norm": 0.5656725168228149, + "learning_rate": 5.321541767083927e-06, + "loss": 0.888, + "step": 3974 + }, + { + "epoch": 0.6664570889657341, + "grad_norm": 0.4438325762748718, + "learning_rate": 5.3167428525124976e-06, + "loss": 0.8804, + "step": 3975 + }, + { + "epoch": 0.6666247511264801, + "grad_norm": 0.4937484860420227, + "learning_rate": 5.311945319091589e-06, + "loss": 0.9184, + "step": 3976 + }, + { + "epoch": 0.6667924132872263, + "grad_norm": 0.4187779724597931, + "learning_rate": 5.307149168236061e-06, + "loss": 1.0089, + "step": 3977 + }, + { + "epoch": 0.6669600754479723, + "grad_norm": 0.4334414601325989, + "learning_rate": 5.30235440136035e-06, + "loss": 0.8644, + "step": 3978 + }, + { + "epoch": 0.6671277376087185, + "grad_norm": 0.6198210716247559, + "learning_rate": 5.297561019878501e-06, + "loss": 0.8508, + "step": 3979 + }, + { + "epoch": 0.6672953997694645, + "grad_norm": 0.6033726930618286, + "learning_rate": 5.292769025204137e-06, + "loss": 0.7934, + "step": 3980 + }, + { + "epoch": 0.6674630619302107, + "grad_norm": 0.4969884753227234, + "learning_rate": 5.28797841875047e-06, + "loss": 0.8088, + "step": 3981 + }, + { + "epoch": 0.6676307240909567, + "grad_norm": 0.35007086396217346, + "learning_rate": 5.283189201930316e-06, + "loss": 0.7962, + "step": 3982 + }, + { + "epoch": 0.6677983862517028, + "grad_norm": 0.4103473126888275, + "learning_rate": 5.2784013761560794e-06, + "loss": 0.8526, + "step": 3983 + }, + { + "epoch": 0.6679660484124489, + "grad_norm": 0.5877014398574829, + "learning_rate": 5.273614942839742e-06, + "loss": 0.8872, + "step": 3984 + }, + { + "epoch": 0.668133710573195, + "grad_norm": 0.39554399251937866, + "learning_rate": 5.26882990339288e-06, + "loss": 0.8546, + "step": 3985 + }, + { + "epoch": 0.6683013727339411, + "grad_norm": 0.8084924817085266, + "learning_rate": 5.2640462592266695e-06, + "loss": 0.7945, + "step": 3986 + }, + { + "epoch": 0.6684690348946872, + "grad_norm": 0.5444297790527344, + "learning_rate": 5.259264011751857e-06, + "loss": 0.9109, + "step": 3987 + }, + { + "epoch": 0.6686366970554333, + "grad_norm": 0.4030567705631256, + "learning_rate": 5.254483162378795e-06, + "loss": 0.8065, + "step": 3988 + }, + { + "epoch": 0.6688043592161794, + "grad_norm": 0.8205330967903137, + "learning_rate": 5.249703712517411e-06, + "loss": 0.7667, + "step": 3989 + }, + { + "epoch": 0.6689720213769255, + "grad_norm": 0.4742152690887451, + "learning_rate": 5.244925663577221e-06, + "loss": 0.8684, + "step": 3990 + }, + { + "epoch": 0.6691396835376716, + "grad_norm": 0.48614606261253357, + "learning_rate": 5.240149016967338e-06, + "loss": 0.8188, + "step": 3991 + }, + { + "epoch": 0.6693073456984177, + "grad_norm": 0.5022182464599609, + "learning_rate": 5.235373774096446e-06, + "loss": 0.8433, + "step": 3992 + }, + { + "epoch": 0.6694750078591638, + "grad_norm": 0.5748201608657837, + "learning_rate": 5.230599936372831e-06, + "loss": 0.9058, + "step": 3993 + }, + { + "epoch": 0.6696426700199098, + "grad_norm": 0.49480316042900085, + "learning_rate": 5.225827505204355e-06, + "loss": 0.8084, + "step": 3994 + }, + { + "epoch": 0.669810332180656, + "grad_norm": 0.4499543309211731, + "learning_rate": 5.221056481998459e-06, + "loss": 0.8476, + "step": 3995 + }, + { + "epoch": 0.669977994341402, + "grad_norm": 0.6074202656745911, + "learning_rate": 5.21628686816219e-06, + "loss": 0.8081, + "step": 3996 + }, + { + "epoch": 0.6701456565021482, + "grad_norm": 0.388506144285202, + "learning_rate": 5.2115186651021575e-06, + "loss": 0.8465, + "step": 3997 + }, + { + "epoch": 0.6703133186628942, + "grad_norm": 0.40389159321784973, + "learning_rate": 5.2067518742245615e-06, + "loss": 0.8688, + "step": 3998 + }, + { + "epoch": 0.6704809808236404, + "grad_norm": 0.5128644108772278, + "learning_rate": 5.201986496935196e-06, + "loss": 0.8765, + "step": 3999 + }, + { + "epoch": 0.6706486429843864, + "grad_norm": 0.38812342286109924, + "learning_rate": 5.1972225346394216e-06, + "loss": 0.8367, + "step": 4000 + }, + { + "epoch": 0.6708163051451326, + "grad_norm": 0.6672189831733704, + "learning_rate": 5.192459988742193e-06, + "loss": 0.841, + "step": 4001 + }, + { + "epoch": 0.6709839673058786, + "grad_norm": 0.39312562346458435, + "learning_rate": 5.187698860648054e-06, + "loss": 0.8424, + "step": 4002 + }, + { + "epoch": 0.6711516294666248, + "grad_norm": 0.7264593839645386, + "learning_rate": 5.182939151761102e-06, + "loss": 0.7477, + "step": 4003 + }, + { + "epoch": 0.6713192916273708, + "grad_norm": 0.5246336460113525, + "learning_rate": 5.178180863485043e-06, + "loss": 0.858, + "step": 4004 + }, + { + "epoch": 0.671486953788117, + "grad_norm": 0.6613538265228271, + "learning_rate": 5.173423997223159e-06, + "loss": 0.8167, + "step": 4005 + }, + { + "epoch": 0.671654615948863, + "grad_norm": 0.5246715545654297, + "learning_rate": 5.168668554378305e-06, + "loss": 0.8786, + "step": 4006 + }, + { + "epoch": 0.6718222781096092, + "grad_norm": 0.4578244686126709, + "learning_rate": 5.163914536352919e-06, + "loss": 0.7704, + "step": 4007 + }, + { + "epoch": 0.6719899402703552, + "grad_norm": 1.217862606048584, + "learning_rate": 5.159161944549017e-06, + "loss": 0.8864, + "step": 4008 + }, + { + "epoch": 0.6721576024311013, + "grad_norm": 0.46265730261802673, + "learning_rate": 5.154410780368202e-06, + "loss": 0.8769, + "step": 4009 + }, + { + "epoch": 0.6723252645918474, + "grad_norm": 0.49288302659988403, + "learning_rate": 5.1496610452116516e-06, + "loss": 0.886, + "step": 4010 + }, + { + "epoch": 0.6724929267525935, + "grad_norm": 0.4994318187236786, + "learning_rate": 5.144912740480122e-06, + "loss": 0.7578, + "step": 4011 + }, + { + "epoch": 0.6726605889133396, + "grad_norm": 0.6784866452217102, + "learning_rate": 5.14016586757394e-06, + "loss": 0.7887, + "step": 4012 + }, + { + "epoch": 0.6728282510740857, + "grad_norm": 0.40225499868392944, + "learning_rate": 5.135420427893026e-06, + "loss": 0.8414, + "step": 4013 + }, + { + "epoch": 0.6729959132348318, + "grad_norm": 0.4047643840312958, + "learning_rate": 5.13067642283686e-06, + "loss": 0.9051, + "step": 4014 + }, + { + "epoch": 0.6731635753955779, + "grad_norm": 0.3079306483268738, + "learning_rate": 5.125933853804519e-06, + "loss": 0.835, + "step": 4015 + }, + { + "epoch": 0.673331237556324, + "grad_norm": 0.41882896423339844, + "learning_rate": 5.121192722194637e-06, + "loss": 0.8247, + "step": 4016 + }, + { + "epoch": 0.6734988997170701, + "grad_norm": 0.8016254901885986, + "learning_rate": 5.116453029405431e-06, + "loss": 0.7974, + "step": 4017 + }, + { + "epoch": 0.6736665618778162, + "grad_norm": 0.4218415915966034, + "learning_rate": 5.111714776834704e-06, + "loss": 0.899, + "step": 4018 + }, + { + "epoch": 0.6738342240385623, + "grad_norm": 0.4429084062576294, + "learning_rate": 5.106977965879816e-06, + "loss": 0.824, + "step": 4019 + }, + { + "epoch": 0.6740018861993083, + "grad_norm": 0.5343514084815979, + "learning_rate": 5.1022425979377174e-06, + "loss": 0.874, + "step": 4020 + }, + { + "epoch": 0.6741695483600545, + "grad_norm": 0.37769246101379395, + "learning_rate": 5.0975086744049265e-06, + "loss": 0.8485, + "step": 4021 + }, + { + "epoch": 0.6743372105208005, + "grad_norm": 0.6687188148498535, + "learning_rate": 5.092776196677529e-06, + "loss": 0.8538, + "step": 4022 + }, + { + "epoch": 0.6745048726815467, + "grad_norm": 0.5082492232322693, + "learning_rate": 5.088045166151201e-06, + "loss": 0.8869, + "step": 4023 + }, + { + "epoch": 0.6746725348422927, + "grad_norm": 0.5272005200386047, + "learning_rate": 5.083315584221174e-06, + "loss": 0.7843, + "step": 4024 + }, + { + "epoch": 0.6748401970030389, + "grad_norm": 0.506908118724823, + "learning_rate": 5.078587452282267e-06, + "loss": 0.901, + "step": 4025 + }, + { + "epoch": 0.6750078591637849, + "grad_norm": 0.46326175332069397, + "learning_rate": 5.07386077172886e-06, + "loss": 0.8058, + "step": 4026 + }, + { + "epoch": 0.6751755213245311, + "grad_norm": 0.7596170902252197, + "learning_rate": 5.069135543954909e-06, + "loss": 0.8804, + "step": 4027 + }, + { + "epoch": 0.6753431834852772, + "grad_norm": 0.7367517352104187, + "learning_rate": 5.064411770353941e-06, + "loss": 0.9244, + "step": 4028 + }, + { + "epoch": 0.6755108456460233, + "grad_norm": 0.5703383684158325, + "learning_rate": 5.059689452319068e-06, + "loss": 0.9342, + "step": 4029 + }, + { + "epoch": 0.6756785078067694, + "grad_norm": 0.5530908703804016, + "learning_rate": 5.0549685912429434e-06, + "loss": 0.8298, + "step": 4030 + }, + { + "epoch": 0.6758461699675155, + "grad_norm": 0.530229389667511, + "learning_rate": 5.050249188517814e-06, + "loss": 0.8117, + "step": 4031 + }, + { + "epoch": 0.6760138321282616, + "grad_norm": 0.4607187807559967, + "learning_rate": 5.045531245535495e-06, + "loss": 0.9085, + "step": 4032 + }, + { + "epoch": 0.6761814942890076, + "grad_norm": 0.4605083167552948, + "learning_rate": 5.040814763687358e-06, + "loss": 0.8637, + "step": 4033 + }, + { + "epoch": 0.6763491564497538, + "grad_norm": 0.38844844698905945, + "learning_rate": 5.03609974436436e-06, + "loss": 0.8292, + "step": 4034 + }, + { + "epoch": 0.6765168186104998, + "grad_norm": 0.46526795625686646, + "learning_rate": 5.031386188957013e-06, + "loss": 0.8265, + "step": 4035 + }, + { + "epoch": 0.676684480771246, + "grad_norm": 0.39576253294944763, + "learning_rate": 5.026674098855402e-06, + "loss": 0.7683, + "step": 4036 + }, + { + "epoch": 0.676852142931992, + "grad_norm": 0.5031571388244629, + "learning_rate": 5.021963475449185e-06, + "loss": 0.9241, + "step": 4037 + }, + { + "epoch": 0.6770198050927382, + "grad_norm": 0.39039212465286255, + "learning_rate": 5.017254320127584e-06, + "loss": 0.8534, + "step": 4038 + }, + { + "epoch": 0.6771874672534842, + "grad_norm": 0.4956362545490265, + "learning_rate": 5.012546634279378e-06, + "loss": 0.7214, + "step": 4039 + }, + { + "epoch": 0.6773551294142304, + "grad_norm": 0.4101143181324005, + "learning_rate": 5.007840419292933e-06, + "loss": 0.8868, + "step": 4040 + }, + { + "epoch": 0.6775227915749764, + "grad_norm": 0.3842748701572418, + "learning_rate": 5.0031356765561625e-06, + "loss": 0.7652, + "step": 4041 + }, + { + "epoch": 0.6776904537357226, + "grad_norm": 0.43894824385643005, + "learning_rate": 4.998432407456559e-06, + "loss": 0.7629, + "step": 4042 + }, + { + "epoch": 0.6778581158964686, + "grad_norm": 0.6596818566322327, + "learning_rate": 4.993730613381173e-06, + "loss": 0.9031, + "step": 4043 + }, + { + "epoch": 0.6780257780572148, + "grad_norm": 0.3763595223426819, + "learning_rate": 4.989030295716615e-06, + "loss": 0.8784, + "step": 4044 + }, + { + "epoch": 0.6781934402179608, + "grad_norm": 0.4675236642360687, + "learning_rate": 4.984331455849077e-06, + "loss": 0.7289, + "step": 4045 + }, + { + "epoch": 0.678361102378707, + "grad_norm": 0.5311740040779114, + "learning_rate": 4.979634095164298e-06, + "loss": 0.8654, + "step": 4046 + }, + { + "epoch": 0.678528764539453, + "grad_norm": 0.3968749940395355, + "learning_rate": 4.974938215047595e-06, + "loss": 0.8619, + "step": 4047 + }, + { + "epoch": 0.6786964267001991, + "grad_norm": 0.4369450509548187, + "learning_rate": 4.970243816883835e-06, + "loss": 0.8361, + "step": 4048 + }, + { + "epoch": 0.6788640888609452, + "grad_norm": 0.723330020904541, + "learning_rate": 4.965550902057453e-06, + "loss": 0.9035, + "step": 4049 + }, + { + "epoch": 0.6790317510216913, + "grad_norm": 0.9839751720428467, + "learning_rate": 4.960859471952451e-06, + "loss": 0.8032, + "step": 4050 + }, + { + "epoch": 0.6791994131824374, + "grad_norm": 0.3737811744213104, + "learning_rate": 4.956169527952392e-06, + "loss": 0.8817, + "step": 4051 + }, + { + "epoch": 0.6793670753431835, + "grad_norm": 0.5440712571144104, + "learning_rate": 4.951481071440398e-06, + "loss": 0.7694, + "step": 4052 + }, + { + "epoch": 0.6795347375039296, + "grad_norm": 0.5463842749595642, + "learning_rate": 4.946794103799146e-06, + "loss": 0.8517, + "step": 4053 + }, + { + "epoch": 0.6797023996646757, + "grad_norm": 0.5319167375564575, + "learning_rate": 4.94210862641089e-06, + "loss": 0.8257, + "step": 4054 + }, + { + "epoch": 0.6798700618254218, + "grad_norm": 0.7051066756248474, + "learning_rate": 4.9374246406574254e-06, + "loss": 0.8603, + "step": 4055 + }, + { + "epoch": 0.6800377239861679, + "grad_norm": 0.4666949212551117, + "learning_rate": 4.932742147920128e-06, + "loss": 0.8018, + "step": 4056 + }, + { + "epoch": 0.680205386146914, + "grad_norm": 0.38911348581314087, + "learning_rate": 4.928061149579919e-06, + "loss": 0.8546, + "step": 4057 + }, + { + "epoch": 0.6803730483076601, + "grad_norm": 0.39490512013435364, + "learning_rate": 4.923381647017278e-06, + "loss": 0.8482, + "step": 4058 + }, + { + "epoch": 0.6805407104684061, + "grad_norm": 0.4537765085697174, + "learning_rate": 4.918703641612255e-06, + "loss": 0.7966, + "step": 4059 + }, + { + "epoch": 0.6807083726291523, + "grad_norm": 0.43742606043815613, + "learning_rate": 4.914027134744446e-06, + "loss": 0.778, + "step": 4060 + }, + { + "epoch": 0.6808760347898983, + "grad_norm": 0.4226575195789337, + "learning_rate": 4.909352127793016e-06, + "loss": 0.8533, + "step": 4061 + }, + { + "epoch": 0.6810436969506445, + "grad_norm": 0.479563444852829, + "learning_rate": 4.904678622136682e-06, + "loss": 0.8895, + "step": 4062 + }, + { + "epoch": 0.6812113591113905, + "grad_norm": 0.5353012084960938, + "learning_rate": 4.900006619153711e-06, + "loss": 0.8008, + "step": 4063 + }, + { + "epoch": 0.6813790212721367, + "grad_norm": 0.4960290193557739, + "learning_rate": 4.895336120221945e-06, + "loss": 0.8606, + "step": 4064 + }, + { + "epoch": 0.6815466834328827, + "grad_norm": 0.3402813971042633, + "learning_rate": 4.890667126718763e-06, + "loss": 0.7323, + "step": 4065 + }, + { + "epoch": 0.6817143455936289, + "grad_norm": 0.4431888163089752, + "learning_rate": 4.885999640021117e-06, + "loss": 0.8602, + "step": 4066 + }, + { + "epoch": 0.6818820077543749, + "grad_norm": 0.4127519726753235, + "learning_rate": 4.881333661505504e-06, + "loss": 0.8083, + "step": 4067 + }, + { + "epoch": 0.6820496699151211, + "grad_norm": 0.36416029930114746, + "learning_rate": 4.876669192547972e-06, + "loss": 0.8568, + "step": 4068 + }, + { + "epoch": 0.6822173320758671, + "grad_norm": 0.7066261768341064, + "learning_rate": 4.872006234524138e-06, + "loss": 0.957, + "step": 4069 + }, + { + "epoch": 0.6823849942366133, + "grad_norm": 0.5453369617462158, + "learning_rate": 4.867344788809172e-06, + "loss": 0.9183, + "step": 4070 + }, + { + "epoch": 0.6825526563973593, + "grad_norm": 0.5057035088539124, + "learning_rate": 4.862684856777776e-06, + "loss": 0.831, + "step": 4071 + }, + { + "epoch": 0.6827203185581054, + "grad_norm": 0.6210739016532898, + "learning_rate": 4.8580264398042355e-06, + "loss": 0.7656, + "step": 4072 + }, + { + "epoch": 0.6828879807188515, + "grad_norm": 0.39194726943969727, + "learning_rate": 4.853369539262365e-06, + "loss": 0.8204, + "step": 4073 + }, + { + "epoch": 0.6830556428795976, + "grad_norm": 0.39771005511283875, + "learning_rate": 4.848714156525547e-06, + "loss": 0.8624, + "step": 4074 + }, + { + "epoch": 0.6832233050403437, + "grad_norm": 0.5039075613021851, + "learning_rate": 4.844060292966719e-06, + "loss": 0.8323, + "step": 4075 + }, + { + "epoch": 0.6833909672010898, + "grad_norm": 0.5027855634689331, + "learning_rate": 4.839407949958349e-06, + "loss": 0.8624, + "step": 4076 + }, + { + "epoch": 0.6835586293618359, + "grad_norm": 0.46842116117477417, + "learning_rate": 4.834757128872477e-06, + "loss": 0.9685, + "step": 4077 + }, + { + "epoch": 0.683726291522582, + "grad_norm": 0.5276632308959961, + "learning_rate": 4.830107831080692e-06, + "loss": 0.8333, + "step": 4078 + }, + { + "epoch": 0.6838939536833281, + "grad_norm": 0.40876996517181396, + "learning_rate": 4.825460057954124e-06, + "loss": 0.8593, + "step": 4079 + }, + { + "epoch": 0.6840616158440742, + "grad_norm": 0.47190171480178833, + "learning_rate": 4.820813810863457e-06, + "loss": 0.8892, + "step": 4080 + }, + { + "epoch": 0.6842292780048203, + "grad_norm": 0.3707180321216583, + "learning_rate": 4.816169091178935e-06, + "loss": 0.8239, + "step": 4081 + }, + { + "epoch": 0.6843969401655664, + "grad_norm": 0.3884958326816559, + "learning_rate": 4.811525900270334e-06, + "loss": 0.8317, + "step": 4082 + }, + { + "epoch": 0.6845646023263124, + "grad_norm": 0.4536154866218567, + "learning_rate": 4.806884239506996e-06, + "loss": 0.7875, + "step": 4083 + }, + { + "epoch": 0.6847322644870586, + "grad_norm": 0.8998631238937378, + "learning_rate": 4.802244110257802e-06, + "loss": 0.9078, + "step": 4084 + }, + { + "epoch": 0.6848999266478046, + "grad_norm": 0.6511355042457581, + "learning_rate": 4.797605513891179e-06, + "loss": 0.8234, + "step": 4085 + }, + { + "epoch": 0.6850675888085508, + "grad_norm": 0.4585486054420471, + "learning_rate": 4.792968451775113e-06, + "loss": 0.8162, + "step": 4086 + }, + { + "epoch": 0.6852352509692968, + "grad_norm": 0.5549787282943726, + "learning_rate": 4.788332925277126e-06, + "loss": 0.8883, + "step": 4087 + }, + { + "epoch": 0.685402913130043, + "grad_norm": 0.4339296519756317, + "learning_rate": 4.783698935764297e-06, + "loss": 0.8856, + "step": 4088 + }, + { + "epoch": 0.685570575290789, + "grad_norm": 0.44029632210731506, + "learning_rate": 4.7790664846032455e-06, + "loss": 0.8069, + "step": 4089 + }, + { + "epoch": 0.6857382374515352, + "grad_norm": 0.4796565771102905, + "learning_rate": 4.774435573160132e-06, + "loss": 0.8275, + "step": 4090 + }, + { + "epoch": 0.6859058996122812, + "grad_norm": 0.6569730639457703, + "learning_rate": 4.769806202800681e-06, + "loss": 0.8657, + "step": 4091 + }, + { + "epoch": 0.6860735617730274, + "grad_norm": 0.43796613812446594, + "learning_rate": 4.76517837489014e-06, + "loss": 0.8583, + "step": 4092 + }, + { + "epoch": 0.6862412239337734, + "grad_norm": 0.41398483514785767, + "learning_rate": 4.760552090793322e-06, + "loss": 0.8822, + "step": 4093 + }, + { + "epoch": 0.6864088860945196, + "grad_norm": 0.5004035234451294, + "learning_rate": 4.755927351874572e-06, + "loss": 0.8834, + "step": 4094 + }, + { + "epoch": 0.6865765482552656, + "grad_norm": 0.5711424350738525, + "learning_rate": 4.7513041594977785e-06, + "loss": 0.9124, + "step": 4095 + }, + { + "epoch": 0.6867442104160117, + "grad_norm": 0.6565632820129395, + "learning_rate": 4.746682515026382e-06, + "loss": 0.8686, + "step": 4096 + }, + { + "epoch": 0.6869118725767578, + "grad_norm": 0.7613095641136169, + "learning_rate": 4.742062419823367e-06, + "loss": 0.8704, + "step": 4097 + }, + { + "epoch": 0.6870795347375039, + "grad_norm": 0.4163006842136383, + "learning_rate": 4.737443875251251e-06, + "loss": 0.7994, + "step": 4098 + }, + { + "epoch": 0.68724719689825, + "grad_norm": 0.4178316593170166, + "learning_rate": 4.732826882672099e-06, + "loss": 0.8177, + "step": 4099 + }, + { + "epoch": 0.6874148590589961, + "grad_norm": 0.6707078814506531, + "learning_rate": 4.728211443447525e-06, + "loss": 0.8434, + "step": 4100 + }, + { + "epoch": 0.6875825212197422, + "grad_norm": 0.8010030388832092, + "learning_rate": 4.7235975589386715e-06, + "loss": 0.9166, + "step": 4101 + }, + { + "epoch": 0.6877501833804883, + "grad_norm": 0.5210207104682922, + "learning_rate": 4.718985230506239e-06, + "loss": 0.9055, + "step": 4102 + }, + { + "epoch": 0.6879178455412344, + "grad_norm": 0.42187198996543884, + "learning_rate": 4.714374459510456e-06, + "loss": 0.8316, + "step": 4103 + }, + { + "epoch": 0.6880855077019805, + "grad_norm": 0.43324875831604004, + "learning_rate": 4.709765247311092e-06, + "loss": 0.8918, + "step": 4104 + }, + { + "epoch": 0.6882531698627266, + "grad_norm": 0.4681563973426819, + "learning_rate": 4.7051575952674675e-06, + "loss": 0.7918, + "step": 4105 + }, + { + "epoch": 0.6884208320234727, + "grad_norm": 0.7020363807678223, + "learning_rate": 4.700551504738431e-06, + "loss": 0.8408, + "step": 4106 + }, + { + "epoch": 0.6885884941842187, + "grad_norm": 0.4496000409126282, + "learning_rate": 4.695946977082381e-06, + "loss": 0.8581, + "step": 4107 + }, + { + "epoch": 0.6887561563449649, + "grad_norm": 0.4724990129470825, + "learning_rate": 4.691344013657245e-06, + "loss": 0.8753, + "step": 4108 + }, + { + "epoch": 0.6889238185057109, + "grad_norm": 0.39639273285865784, + "learning_rate": 4.686742615820493e-06, + "loss": 0.897, + "step": 4109 + }, + { + "epoch": 0.6890914806664571, + "grad_norm": 0.46177640557289124, + "learning_rate": 4.682142784929142e-06, + "loss": 0.7487, + "step": 4110 + }, + { + "epoch": 0.6892591428272031, + "grad_norm": 0.4963343143463135, + "learning_rate": 4.6775445223397306e-06, + "loss": 0.9372, + "step": 4111 + }, + { + "epoch": 0.6894268049879493, + "grad_norm": 0.4017559587955475, + "learning_rate": 4.672947829408346e-06, + "loss": 0.8145, + "step": 4112 + }, + { + "epoch": 0.6895944671486954, + "grad_norm": 0.5806766152381897, + "learning_rate": 4.668352707490611e-06, + "loss": 0.8044, + "step": 4113 + }, + { + "epoch": 0.6897621293094415, + "grad_norm": 0.4939943552017212, + "learning_rate": 4.663759157941682e-06, + "loss": 0.9683, + "step": 4114 + }, + { + "epoch": 0.6899297914701876, + "grad_norm": 0.3326122462749481, + "learning_rate": 4.6591671821162535e-06, + "loss": 0.7993, + "step": 4115 + }, + { + "epoch": 0.6900974536309337, + "grad_norm": 0.44053709506988525, + "learning_rate": 4.654576781368566e-06, + "loss": 0.81, + "step": 4116 + }, + { + "epoch": 0.6902651157916798, + "grad_norm": 0.6346563100814819, + "learning_rate": 4.6499879570523684e-06, + "loss": 0.8571, + "step": 4117 + }, + { + "epoch": 0.6904327779524259, + "grad_norm": 0.4332069754600525, + "learning_rate": 4.645400710520972e-06, + "loss": 0.7902, + "step": 4118 + }, + { + "epoch": 0.690600440113172, + "grad_norm": 0.4327547252178192, + "learning_rate": 4.6408150431272135e-06, + "loss": 0.8567, + "step": 4119 + }, + { + "epoch": 0.690768102273918, + "grad_norm": 0.5181637406349182, + "learning_rate": 4.636230956223462e-06, + "loss": 0.7896, + "step": 4120 + }, + { + "epoch": 0.6909357644346642, + "grad_norm": 0.5253118872642517, + "learning_rate": 4.6316484511616185e-06, + "loss": 0.818, + "step": 4121 + }, + { + "epoch": 0.6911034265954102, + "grad_norm": 0.3494114875793457, + "learning_rate": 4.627067529293119e-06, + "loss": 0.8766, + "step": 4122 + }, + { + "epoch": 0.6912710887561564, + "grad_norm": 0.45745769143104553, + "learning_rate": 4.622488191968938e-06, + "loss": 0.8666, + "step": 4123 + }, + { + "epoch": 0.6914387509169024, + "grad_norm": 0.5724239349365234, + "learning_rate": 4.61791044053958e-06, + "loss": 0.7707, + "step": 4124 + }, + { + "epoch": 0.6916064130776486, + "grad_norm": 0.41080933809280396, + "learning_rate": 4.613334276355079e-06, + "loss": 0.8641, + "step": 4125 + }, + { + "epoch": 0.6917740752383946, + "grad_norm": 0.4222384989261627, + "learning_rate": 4.608759700764998e-06, + "loss": 0.7909, + "step": 4126 + }, + { + "epoch": 0.6919417373991408, + "grad_norm": 0.4619828462600708, + "learning_rate": 4.604186715118445e-06, + "loss": 0.9484, + "step": 4127 + }, + { + "epoch": 0.6921093995598868, + "grad_norm": 0.36951783299446106, + "learning_rate": 4.599615320764041e-06, + "loss": 0.7432, + "step": 4128 + }, + { + "epoch": 0.692277061720633, + "grad_norm": 0.571994960308075, + "learning_rate": 4.595045519049954e-06, + "loss": 0.8783, + "step": 4129 + }, + { + "epoch": 0.692444723881379, + "grad_norm": 0.550298273563385, + "learning_rate": 4.590477311323872e-06, + "loss": 0.9082, + "step": 4130 + }, + { + "epoch": 0.6926123860421252, + "grad_norm": 0.39157381653785706, + "learning_rate": 4.585910698933013e-06, + "loss": 0.8105, + "step": 4131 + }, + { + "epoch": 0.6927800482028712, + "grad_norm": 0.7182436585426331, + "learning_rate": 4.581345683224133e-06, + "loss": 0.8682, + "step": 4132 + }, + { + "epoch": 0.6929477103636174, + "grad_norm": 0.5301082730293274, + "learning_rate": 4.576782265543507e-06, + "loss": 0.7852, + "step": 4133 + }, + { + "epoch": 0.6931153725243634, + "grad_norm": 0.6268473863601685, + "learning_rate": 4.572220447236948e-06, + "loss": 0.9147, + "step": 4134 + }, + { + "epoch": 0.6932830346851095, + "grad_norm": 0.41989865899086, + "learning_rate": 4.567660229649791e-06, + "loss": 0.832, + "step": 4135 + }, + { + "epoch": 0.6934506968458556, + "grad_norm": 0.4412514567375183, + "learning_rate": 4.563101614126895e-06, + "loss": 0.8326, + "step": 4136 + }, + { + "epoch": 0.6936183590066017, + "grad_norm": 0.6761359572410583, + "learning_rate": 4.5585446020126634e-06, + "loss": 0.8856, + "step": 4137 + }, + { + "epoch": 0.6937860211673478, + "grad_norm": 0.5193226933479309, + "learning_rate": 4.553989194651003e-06, + "loss": 0.8256, + "step": 4138 + }, + { + "epoch": 0.6939536833280939, + "grad_norm": 0.5478662252426147, + "learning_rate": 4.549435393385371e-06, + "loss": 0.8706, + "step": 4139 + }, + { + "epoch": 0.69412134548884, + "grad_norm": 0.559935450553894, + "learning_rate": 4.544883199558735e-06, + "loss": 0.8608, + "step": 4140 + }, + { + "epoch": 0.6942890076495861, + "grad_norm": 0.5725769996643066, + "learning_rate": 4.540332614513589e-06, + "loss": 0.7859, + "step": 4141 + }, + { + "epoch": 0.6944566698103322, + "grad_norm": 0.5817712545394897, + "learning_rate": 4.535783639591961e-06, + "loss": 0.9089, + "step": 4142 + }, + { + "epoch": 0.6946243319710783, + "grad_norm": 1.1822705268859863, + "learning_rate": 4.5312362761354065e-06, + "loss": 0.7498, + "step": 4143 + }, + { + "epoch": 0.6947919941318244, + "grad_norm": 0.6353855133056641, + "learning_rate": 4.526690525484987e-06, + "loss": 0.8337, + "step": 4144 + }, + { + "epoch": 0.6949596562925705, + "grad_norm": 0.3704085946083069, + "learning_rate": 4.522146388981306e-06, + "loss": 0.8336, + "step": 4145 + }, + { + "epoch": 0.6951273184533165, + "grad_norm": 0.5411508679389954, + "learning_rate": 4.51760386796449e-06, + "loss": 0.882, + "step": 4146 + }, + { + "epoch": 0.6952949806140627, + "grad_norm": 0.35425445437431335, + "learning_rate": 4.513062963774176e-06, + "loss": 0.8379, + "step": 4147 + }, + { + "epoch": 0.6954626427748087, + "grad_norm": 0.44852039217948914, + "learning_rate": 4.508523677749542e-06, + "loss": 0.8871, + "step": 4148 + }, + { + "epoch": 0.6956303049355549, + "grad_norm": 0.5101825594902039, + "learning_rate": 4.5039860112292755e-06, + "loss": 0.857, + "step": 4149 + }, + { + "epoch": 0.6957979670963009, + "grad_norm": 0.40634283423423767, + "learning_rate": 4.4994499655515865e-06, + "loss": 0.8711, + "step": 4150 + }, + { + "epoch": 0.6959656292570471, + "grad_norm": 0.7540550827980042, + "learning_rate": 4.494915542054219e-06, + "loss": 0.908, + "step": 4151 + }, + { + "epoch": 0.6961332914177931, + "grad_norm": 1.2536613941192627, + "learning_rate": 4.490382742074428e-06, + "loss": 0.7994, + "step": 4152 + }, + { + "epoch": 0.6963009535785393, + "grad_norm": 0.3635319471359253, + "learning_rate": 4.485851566948985e-06, + "loss": 0.7661, + "step": 4153 + }, + { + "epoch": 0.6964686157392853, + "grad_norm": 0.37484169006347656, + "learning_rate": 4.4813220180141995e-06, + "loss": 0.8714, + "step": 4154 + }, + { + "epoch": 0.6966362779000315, + "grad_norm": 0.3904739320278168, + "learning_rate": 4.476794096605884e-06, + "loss": 0.8176, + "step": 4155 + }, + { + "epoch": 0.6968039400607775, + "grad_norm": 0.4750972092151642, + "learning_rate": 4.472267804059387e-06, + "loss": 0.925, + "step": 4156 + }, + { + "epoch": 0.6969716022215237, + "grad_norm": 0.6135466694831848, + "learning_rate": 4.467743141709563e-06, + "loss": 0.781, + "step": 4157 + }, + { + "epoch": 0.6971392643822697, + "grad_norm": 0.38343942165374756, + "learning_rate": 4.463220110890787e-06, + "loss": 0.8202, + "step": 4158 + }, + { + "epoch": 0.6973069265430158, + "grad_norm": 1.104854702949524, + "learning_rate": 4.458698712936966e-06, + "loss": 0.8498, + "step": 4159 + }, + { + "epoch": 0.6974745887037619, + "grad_norm": 0.3883248269557953, + "learning_rate": 4.454178949181507e-06, + "loss": 0.9099, + "step": 4160 + }, + { + "epoch": 0.697642250864508, + "grad_norm": 1.0206170082092285, + "learning_rate": 4.4496608209573485e-06, + "loss": 0.8711, + "step": 4161 + }, + { + "epoch": 0.6978099130252541, + "grad_norm": 0.5924220085144043, + "learning_rate": 4.445144329596951e-06, + "loss": 0.828, + "step": 4162 + }, + { + "epoch": 0.6979775751860002, + "grad_norm": 0.349104642868042, + "learning_rate": 4.440629476432268e-06, + "loss": 0.7629, + "step": 4163 + }, + { + "epoch": 0.6981452373467463, + "grad_norm": 0.37381574511528015, + "learning_rate": 4.436116262794793e-06, + "loss": 0.9125, + "step": 4164 + }, + { + "epoch": 0.6983128995074924, + "grad_norm": 0.40002623200416565, + "learning_rate": 4.431604690015532e-06, + "loss": 0.8435, + "step": 4165 + }, + { + "epoch": 0.6984805616682385, + "grad_norm": 0.37091362476348877, + "learning_rate": 4.427094759425e-06, + "loss": 0.7839, + "step": 4166 + }, + { + "epoch": 0.6986482238289846, + "grad_norm": 0.6612892150878906, + "learning_rate": 4.422586472353229e-06, + "loss": 0.8337, + "step": 4167 + }, + { + "epoch": 0.6988158859897307, + "grad_norm": 0.4454594850540161, + "learning_rate": 4.418079830129773e-06, + "loss": 0.8311, + "step": 4168 + }, + { + "epoch": 0.6989835481504768, + "grad_norm": 0.6289452314376831, + "learning_rate": 4.413574834083694e-06, + "loss": 0.7649, + "step": 4169 + }, + { + "epoch": 0.6991512103112228, + "grad_norm": 0.39444994926452637, + "learning_rate": 4.4090714855435746e-06, + "loss": 0.7753, + "step": 4170 + }, + { + "epoch": 0.699318872471969, + "grad_norm": 0.7020822763442993, + "learning_rate": 4.404569785837505e-06, + "loss": 0.9139, + "step": 4171 + }, + { + "epoch": 0.699486534632715, + "grad_norm": 0.47531023621559143, + "learning_rate": 4.40006973629309e-06, + "loss": 0.8804, + "step": 4172 + }, + { + "epoch": 0.6996541967934612, + "grad_norm": 0.46519336104393005, + "learning_rate": 4.395571338237455e-06, + "loss": 0.837, + "step": 4173 + }, + { + "epoch": 0.6998218589542072, + "grad_norm": 0.5111847519874573, + "learning_rate": 4.391074592997228e-06, + "loss": 0.8368, + "step": 4174 + }, + { + "epoch": 0.6999895211149534, + "grad_norm": 0.40244993567466736, + "learning_rate": 4.386579501898563e-06, + "loss": 0.843, + "step": 4175 + }, + { + "epoch": 0.7001571832756994, + "grad_norm": 0.5221668481826782, + "learning_rate": 4.38208606626711e-06, + "loss": 0.8128, + "step": 4176 + }, + { + "epoch": 0.7003248454364456, + "grad_norm": 0.34330761432647705, + "learning_rate": 4.37759428742804e-06, + "loss": 0.8586, + "step": 4177 + }, + { + "epoch": 0.7004925075971916, + "grad_norm": 0.3652341365814209, + "learning_rate": 4.373104166706039e-06, + "loss": 0.8355, + "step": 4178 + }, + { + "epoch": 0.7006601697579378, + "grad_norm": 0.675369143486023, + "learning_rate": 4.368615705425292e-06, + "loss": 0.9091, + "step": 4179 + }, + { + "epoch": 0.7008278319186838, + "grad_norm": 0.40616875886917114, + "learning_rate": 4.364128904909509e-06, + "loss": 0.7958, + "step": 4180 + }, + { + "epoch": 0.70099549407943, + "grad_norm": 0.5879313945770264, + "learning_rate": 4.359643766481901e-06, + "loss": 0.8831, + "step": 4181 + }, + { + "epoch": 0.701163156240176, + "grad_norm": 0.49311167001724243, + "learning_rate": 4.355160291465186e-06, + "loss": 0.8027, + "step": 4182 + }, + { + "epoch": 0.7013308184009222, + "grad_norm": 0.38633251190185547, + "learning_rate": 4.3506784811815995e-06, + "loss": 0.8575, + "step": 4183 + }, + { + "epoch": 0.7014984805616682, + "grad_norm": 0.4957720637321472, + "learning_rate": 4.346198336952892e-06, + "loss": 0.801, + "step": 4184 + }, + { + "epoch": 0.7016661427224143, + "grad_norm": 0.368158221244812, + "learning_rate": 4.341719860100296e-06, + "loss": 0.9135, + "step": 4185 + }, + { + "epoch": 0.7018338048831604, + "grad_norm": 0.7882813215255737, + "learning_rate": 4.337243051944583e-06, + "loss": 0.7851, + "step": 4186 + }, + { + "epoch": 0.7020014670439065, + "grad_norm": 0.38310936093330383, + "learning_rate": 4.3327679138060105e-06, + "loss": 0.85, + "step": 4187 + }, + { + "epoch": 0.7021691292046526, + "grad_norm": 0.42763540148735046, + "learning_rate": 4.3282944470043576e-06, + "loss": 0.8013, + "step": 4188 + }, + { + "epoch": 0.7023367913653987, + "grad_norm": 0.359748899936676, + "learning_rate": 4.323822652858911e-06, + "loss": 0.7907, + "step": 4189 + }, + { + "epoch": 0.7025044535261448, + "grad_norm": 0.5136620998382568, + "learning_rate": 4.319352532688444e-06, + "loss": 0.8221, + "step": 4190 + }, + { + "epoch": 0.7026721156868909, + "grad_norm": 0.5542488098144531, + "learning_rate": 4.314884087811256e-06, + "loss": 0.8735, + "step": 4191 + }, + { + "epoch": 0.702839777847637, + "grad_norm": 0.5811720490455627, + "learning_rate": 4.310417319545154e-06, + "loss": 0.7878, + "step": 4192 + }, + { + "epoch": 0.7030074400083831, + "grad_norm": 0.5262954831123352, + "learning_rate": 4.305952229207436e-06, + "loss": 0.8474, + "step": 4193 + }, + { + "epoch": 0.7031751021691292, + "grad_norm": 0.459160178899765, + "learning_rate": 4.30148881811491e-06, + "loss": 0.9203, + "step": 4194 + }, + { + "epoch": 0.7033427643298753, + "grad_norm": 0.5392619967460632, + "learning_rate": 4.297027087583901e-06, + "loss": 0.8551, + "step": 4195 + }, + { + "epoch": 0.7035104264906213, + "grad_norm": 0.9503580927848816, + "learning_rate": 4.292567038930218e-06, + "loss": 0.8948, + "step": 4196 + }, + { + "epoch": 0.7036780886513675, + "grad_norm": 0.4359428584575653, + "learning_rate": 4.288108673469192e-06, + "loss": 0.8506, + "step": 4197 + }, + { + "epoch": 0.7038457508121136, + "grad_norm": 0.6888027191162109, + "learning_rate": 4.283651992515649e-06, + "loss": 0.8487, + "step": 4198 + }, + { + "epoch": 0.7040134129728597, + "grad_norm": 0.41162729263305664, + "learning_rate": 4.279196997383912e-06, + "loss": 0.8153, + "step": 4199 + }, + { + "epoch": 0.7041810751336058, + "grad_norm": 0.6298072934150696, + "learning_rate": 4.274743689387826e-06, + "loss": 0.8535, + "step": 4200 + }, + { + "epoch": 0.7043487372943519, + "grad_norm": 0.3720015585422516, + "learning_rate": 4.270292069840716e-06, + "loss": 0.9058, + "step": 4201 + }, + { + "epoch": 0.704516399455098, + "grad_norm": 0.4745865762233734, + "learning_rate": 4.265842140055428e-06, + "loss": 0.8507, + "step": 4202 + }, + { + "epoch": 0.7046840616158441, + "grad_norm": 0.6363827586174011, + "learning_rate": 4.261393901344299e-06, + "loss": 0.8739, + "step": 4203 + }, + { + "epoch": 0.7048517237765902, + "grad_norm": 0.695592999458313, + "learning_rate": 4.256947355019163e-06, + "loss": 0.9261, + "step": 4204 + }, + { + "epoch": 0.7050193859373363, + "grad_norm": 0.4004266560077667, + "learning_rate": 4.252502502391373e-06, + "loss": 0.8323, + "step": 4205 + }, + { + "epoch": 0.7051870480980824, + "grad_norm": 0.35616323351860046, + "learning_rate": 4.24805934477176e-06, + "loss": 0.8492, + "step": 4206 + }, + { + "epoch": 0.7053547102588285, + "grad_norm": 0.5288553237915039, + "learning_rate": 4.243617883470675e-06, + "loss": 0.8501, + "step": 4207 + }, + { + "epoch": 0.7055223724195746, + "grad_norm": 1.0049248933792114, + "learning_rate": 4.2391781197979575e-06, + "loss": 0.7618, + "step": 4208 + }, + { + "epoch": 0.7056900345803206, + "grad_norm": 1.5840026140213013, + "learning_rate": 4.2347400550629435e-06, + "loss": 0.8275, + "step": 4209 + }, + { + "epoch": 0.7058576967410668, + "grad_norm": 0.4762451648712158, + "learning_rate": 4.230303690574478e-06, + "loss": 0.8096, + "step": 4210 + }, + { + "epoch": 0.7060253589018128, + "grad_norm": 0.6156831383705139, + "learning_rate": 4.225869027640902e-06, + "loss": 0.9071, + "step": 4211 + }, + { + "epoch": 0.706193021062559, + "grad_norm": 0.5486253499984741, + "learning_rate": 4.22143606757005e-06, + "loss": 0.7992, + "step": 4212 + }, + { + "epoch": 0.706360683223305, + "grad_norm": 0.6755399703979492, + "learning_rate": 4.217004811669253e-06, + "loss": 1.0019, + "step": 4213 + }, + { + "epoch": 0.7065283453840512, + "grad_norm": 0.47275254130363464, + "learning_rate": 4.212575261245351e-06, + "loss": 0.8909, + "step": 4214 + }, + { + "epoch": 0.7066960075447972, + "grad_norm": 0.4999307692050934, + "learning_rate": 4.208147417604665e-06, + "loss": 0.9178, + "step": 4215 + }, + { + "epoch": 0.7068636697055434, + "grad_norm": 0.3870341181755066, + "learning_rate": 4.2037212820530285e-06, + "loss": 0.8844, + "step": 4216 + }, + { + "epoch": 0.7070313318662894, + "grad_norm": 0.45955073833465576, + "learning_rate": 4.19929685589576e-06, + "loss": 0.8724, + "step": 4217 + }, + { + "epoch": 0.7071989940270356, + "grad_norm": 0.4071888327598572, + "learning_rate": 4.194874140437675e-06, + "loss": 0.8095, + "step": 4218 + }, + { + "epoch": 0.7073666561877816, + "grad_norm": 0.4482811391353607, + "learning_rate": 4.190453136983091e-06, + "loss": 0.8302, + "step": 4219 + }, + { + "epoch": 0.7075343183485278, + "grad_norm": 0.46475750207901, + "learning_rate": 4.186033846835814e-06, + "loss": 0.8576, + "step": 4220 + }, + { + "epoch": 0.7077019805092738, + "grad_norm": 0.35893514752388, + "learning_rate": 4.181616271299153e-06, + "loss": 0.8915, + "step": 4221 + }, + { + "epoch": 0.70786964267002, + "grad_norm": 0.3959294557571411, + "learning_rate": 4.1772004116759004e-06, + "loss": 0.8427, + "step": 4222 + }, + { + "epoch": 0.708037304830766, + "grad_norm": 0.45966199040412903, + "learning_rate": 4.172786269268345e-06, + "loss": 0.8358, + "step": 4223 + }, + { + "epoch": 0.7082049669915121, + "grad_norm": 0.39270827174186707, + "learning_rate": 4.1683738453782795e-06, + "loss": 0.8227, + "step": 4224 + }, + { + "epoch": 0.7083726291522582, + "grad_norm": 0.3983052372932434, + "learning_rate": 4.16396314130698e-06, + "loss": 0.8065, + "step": 4225 + }, + { + "epoch": 0.7085402913130043, + "grad_norm": 0.5825944542884827, + "learning_rate": 4.1595541583552125e-06, + "loss": 0.8186, + "step": 4226 + }, + { + "epoch": 0.7087079534737504, + "grad_norm": 0.390227735042572, + "learning_rate": 4.155146897823249e-06, + "loss": 0.8417, + "step": 4227 + }, + { + "epoch": 0.7088756156344965, + "grad_norm": 0.4769403636455536, + "learning_rate": 4.150741361010837e-06, + "loss": 0.8028, + "step": 4228 + }, + { + "epoch": 0.7090432777952426, + "grad_norm": 0.6111634969711304, + "learning_rate": 4.1463375492172285e-06, + "loss": 0.8451, + "step": 4229 + }, + { + "epoch": 0.7092109399559887, + "grad_norm": 0.5125744938850403, + "learning_rate": 4.141935463741168e-06, + "loss": 0.892, + "step": 4230 + }, + { + "epoch": 0.7093786021167348, + "grad_norm": 0.37312835454940796, + "learning_rate": 4.137535105880872e-06, + "loss": 0.8185, + "step": 4231 + }, + { + "epoch": 0.7095462642774809, + "grad_norm": 0.4071962237358093, + "learning_rate": 4.1331364769340665e-06, + "loss": 0.8193, + "step": 4232 + }, + { + "epoch": 0.709713926438227, + "grad_norm": 0.3986750543117523, + "learning_rate": 4.128739578197965e-06, + "loss": 0.7764, + "step": 4233 + }, + { + "epoch": 0.7098815885989731, + "grad_norm": 0.3785371482372284, + "learning_rate": 4.124344410969263e-06, + "loss": 0.83, + "step": 4234 + }, + { + "epoch": 0.7100492507597191, + "grad_norm": 0.41927018761634827, + "learning_rate": 4.119950976544158e-06, + "loss": 0.8113, + "step": 4235 + }, + { + "epoch": 0.7102169129204653, + "grad_norm": 0.3834112584590912, + "learning_rate": 4.1155592762183135e-06, + "loss": 0.8667, + "step": 4236 + }, + { + "epoch": 0.7103845750812113, + "grad_norm": 0.5270161032676697, + "learning_rate": 4.111169311286905e-06, + "loss": 0.8546, + "step": 4237 + }, + { + "epoch": 0.7105522372419575, + "grad_norm": 0.3906877934932709, + "learning_rate": 4.106781083044591e-06, + "loss": 0.829, + "step": 4238 + }, + { + "epoch": 0.7107198994027035, + "grad_norm": 0.457753986120224, + "learning_rate": 4.102394592785509e-06, + "loss": 0.9014, + "step": 4239 + }, + { + "epoch": 0.7108875615634497, + "grad_norm": 0.6644868850708008, + "learning_rate": 4.098009841803288e-06, + "loss": 0.8609, + "step": 4240 + }, + { + "epoch": 0.7110552237241957, + "grad_norm": 0.3552396893501282, + "learning_rate": 4.093626831391051e-06, + "loss": 0.8647, + "step": 4241 + }, + { + "epoch": 0.7112228858849419, + "grad_norm": 0.4755293130874634, + "learning_rate": 4.089245562841396e-06, + "loss": 0.8338, + "step": 4242 + }, + { + "epoch": 0.7113905480456879, + "grad_norm": 0.7194167375564575, + "learning_rate": 4.0848660374464185e-06, + "loss": 0.8174, + "step": 4243 + }, + { + "epoch": 0.7115582102064341, + "grad_norm": 0.5331094264984131, + "learning_rate": 4.080488256497694e-06, + "loss": 0.9453, + "step": 4244 + }, + { + "epoch": 0.7117258723671801, + "grad_norm": 0.4913577139377594, + "learning_rate": 4.076112221286278e-06, + "loss": 0.9477, + "step": 4245 + }, + { + "epoch": 0.7118935345279263, + "grad_norm": 0.41921189427375793, + "learning_rate": 4.071737933102726e-06, + "loss": 0.9137, + "step": 4246 + }, + { + "epoch": 0.7120611966886723, + "grad_norm": 0.3868524432182312, + "learning_rate": 4.067365393237064e-06, + "loss": 0.7741, + "step": 4247 + }, + { + "epoch": 0.7122288588494184, + "grad_norm": 0.4459342658519745, + "learning_rate": 4.062994602978812e-06, + "loss": 0.8389, + "step": 4248 + }, + { + "epoch": 0.7123965210101645, + "grad_norm": 0.6608433127403259, + "learning_rate": 4.05862556361697e-06, + "loss": 0.8265, + "step": 4249 + }, + { + "epoch": 0.7125641831709106, + "grad_norm": 0.8078179955482483, + "learning_rate": 4.054258276440016e-06, + "loss": 0.893, + "step": 4250 + }, + { + "epoch": 0.7127318453316567, + "grad_norm": 0.3659352660179138, + "learning_rate": 4.049892742735924e-06, + "loss": 0.7951, + "step": 4251 + }, + { + "epoch": 0.7128995074924028, + "grad_norm": 0.6276419162750244, + "learning_rate": 4.045528963792139e-06, + "loss": 0.8599, + "step": 4252 + }, + { + "epoch": 0.7130671696531489, + "grad_norm": 0.4345807731151581, + "learning_rate": 4.041166940895599e-06, + "loss": 0.863, + "step": 4253 + }, + { + "epoch": 0.713234831813895, + "grad_norm": 0.5194419622421265, + "learning_rate": 4.036806675332715e-06, + "loss": 0.8563, + "step": 4254 + }, + { + "epoch": 0.7134024939746411, + "grad_norm": 0.44164738059043884, + "learning_rate": 4.03244816838938e-06, + "loss": 0.8806, + "step": 4255 + }, + { + "epoch": 0.7135701561353872, + "grad_norm": 0.40137559175491333, + "learning_rate": 4.028091421350975e-06, + "loss": 0.8699, + "step": 4256 + }, + { + "epoch": 0.7137378182961333, + "grad_norm": 0.4443762004375458, + "learning_rate": 4.023736435502367e-06, + "loss": 0.7674, + "step": 4257 + }, + { + "epoch": 0.7139054804568794, + "grad_norm": 0.5507704019546509, + "learning_rate": 4.019383212127879e-06, + "loss": 0.811, + "step": 4258 + }, + { + "epoch": 0.7140731426176254, + "grad_norm": 0.44957438111305237, + "learning_rate": 4.015031752511338e-06, + "loss": 0.8692, + "step": 4259 + }, + { + "epoch": 0.7142408047783716, + "grad_norm": 0.41376200318336487, + "learning_rate": 4.0106820579360485e-06, + "loss": 0.81, + "step": 4260 + }, + { + "epoch": 0.7144084669391176, + "grad_norm": 0.4778636693954468, + "learning_rate": 4.006334129684781e-06, + "loss": 0.7534, + "step": 4261 + }, + { + "epoch": 0.7145761290998638, + "grad_norm": 0.42495572566986084, + "learning_rate": 4.001987969039801e-06, + "loss": 0.8306, + "step": 4262 + }, + { + "epoch": 0.7147437912606098, + "grad_norm": 0.4066849946975708, + "learning_rate": 3.99764357728284e-06, + "loss": 0.8782, + "step": 4263 + }, + { + "epoch": 0.714911453421356, + "grad_norm": 0.5115940570831299, + "learning_rate": 3.993300955695112e-06, + "loss": 0.7987, + "step": 4264 + }, + { + "epoch": 0.715079115582102, + "grad_norm": 0.3866828680038452, + "learning_rate": 3.9889601055573145e-06, + "loss": 0.823, + "step": 4265 + }, + { + "epoch": 0.7152467777428482, + "grad_norm": 0.3616901636123657, + "learning_rate": 3.984621028149617e-06, + "loss": 0.9494, + "step": 4266 + }, + { + "epoch": 0.7154144399035942, + "grad_norm": 0.668507993221283, + "learning_rate": 3.98028372475166e-06, + "loss": 0.8685, + "step": 4267 + }, + { + "epoch": 0.7155821020643404, + "grad_norm": 0.5150197744369507, + "learning_rate": 3.975948196642578e-06, + "loss": 0.8761, + "step": 4268 + }, + { + "epoch": 0.7157497642250864, + "grad_norm": 0.3876942992210388, + "learning_rate": 3.9716144451009655e-06, + "loss": 0.8787, + "step": 4269 + }, + { + "epoch": 0.7159174263858326, + "grad_norm": 0.5321659445762634, + "learning_rate": 3.967282471404903e-06, + "loss": 0.8081, + "step": 4270 + }, + { + "epoch": 0.7160850885465786, + "grad_norm": 0.657336413860321, + "learning_rate": 3.962952276831943e-06, + "loss": 0.691, + "step": 4271 + }, + { + "epoch": 0.7162527507073247, + "grad_norm": 0.6628797650337219, + "learning_rate": 3.958623862659108e-06, + "loss": 0.8417, + "step": 4272 + }, + { + "epoch": 0.7164204128680708, + "grad_norm": 0.6671704649925232, + "learning_rate": 3.954297230162908e-06, + "loss": 0.7801, + "step": 4273 + }, + { + "epoch": 0.7165880750288169, + "grad_norm": 0.4789895713329315, + "learning_rate": 3.9499723806193145e-06, + "loss": 0.8519, + "step": 4274 + }, + { + "epoch": 0.716755737189563, + "grad_norm": 0.2626591920852661, + "learning_rate": 3.9456493153037835e-06, + "loss": 0.8363, + "step": 4275 + }, + { + "epoch": 0.7169233993503091, + "grad_norm": 0.5861395001411438, + "learning_rate": 3.941328035491245e-06, + "loss": 0.9238, + "step": 4276 + }, + { + "epoch": 0.7170910615110552, + "grad_norm": 0.40450188517570496, + "learning_rate": 3.937008542456085e-06, + "loss": 0.8712, + "step": 4277 + }, + { + "epoch": 0.7172587236718013, + "grad_norm": 0.41760146617889404, + "learning_rate": 3.932690837472184e-06, + "loss": 0.9621, + "step": 4278 + }, + { + "epoch": 0.7174263858325474, + "grad_norm": 0.42822742462158203, + "learning_rate": 3.9283749218128885e-06, + "loss": 0.8108, + "step": 4279 + }, + { + "epoch": 0.7175940479932935, + "grad_norm": 0.5657880902290344, + "learning_rate": 3.924060796751012e-06, + "loss": 0.7965, + "step": 4280 + }, + { + "epoch": 0.7177617101540396, + "grad_norm": 0.5241937637329102, + "learning_rate": 3.91974846355884e-06, + "loss": 0.8407, + "step": 4281 + }, + { + "epoch": 0.7179293723147857, + "grad_norm": 0.5473652482032776, + "learning_rate": 3.91543792350814e-06, + "loss": 0.9083, + "step": 4282 + }, + { + "epoch": 0.7180970344755317, + "grad_norm": 0.5616039633750916, + "learning_rate": 3.911129177870139e-06, + "loss": 0.8463, + "step": 4283 + }, + { + "epoch": 0.7182646966362779, + "grad_norm": 1.1539825201034546, + "learning_rate": 3.9068222279155425e-06, + "loss": 0.8511, + "step": 4284 + }, + { + "epoch": 0.718432358797024, + "grad_norm": 0.4508279860019684, + "learning_rate": 3.902517074914522e-06, + "loss": 0.8484, + "step": 4285 + }, + { + "epoch": 0.7186000209577701, + "grad_norm": 0.4086759388446808, + "learning_rate": 3.898213720136716e-06, + "loss": 0.8517, + "step": 4286 + }, + { + "epoch": 0.7187676831185162, + "grad_norm": 0.6924920678138733, + "learning_rate": 3.893912164851246e-06, + "loss": 0.8152, + "step": 4287 + }, + { + "epoch": 0.7189353452792623, + "grad_norm": 0.43707549571990967, + "learning_rate": 3.889612410326685e-06, + "loss": 0.8664, + "step": 4288 + }, + { + "epoch": 0.7191030074400084, + "grad_norm": 0.3894766569137573, + "learning_rate": 3.885314457831091e-06, + "loss": 0.8669, + "step": 4289 + }, + { + "epoch": 0.7192706696007545, + "grad_norm": 0.38495904207229614, + "learning_rate": 3.881018308631981e-06, + "loss": 0.8742, + "step": 4290 + }, + { + "epoch": 0.7194383317615006, + "grad_norm": 0.3835095167160034, + "learning_rate": 3.876723963996338e-06, + "loss": 0.9261, + "step": 4291 + }, + { + "epoch": 0.7196059939222467, + "grad_norm": 0.6569795608520508, + "learning_rate": 3.872431425190626e-06, + "loss": 0.9282, + "step": 4292 + }, + { + "epoch": 0.7197736560829928, + "grad_norm": 0.37913012504577637, + "learning_rate": 3.8681406934807585e-06, + "loss": 0.7464, + "step": 4293 + }, + { + "epoch": 0.7199413182437389, + "grad_norm": 0.44137638807296753, + "learning_rate": 3.863851770132134e-06, + "loss": 0.8153, + "step": 4294 + }, + { + "epoch": 0.720108980404485, + "grad_norm": 0.45716938376426697, + "learning_rate": 3.859564656409606e-06, + "loss": 0.899, + "step": 4295 + }, + { + "epoch": 0.720276642565231, + "grad_norm": 0.5398461818695068, + "learning_rate": 3.855279353577493e-06, + "loss": 0.9182, + "step": 4296 + }, + { + "epoch": 0.7204443047259772, + "grad_norm": 0.45532530546188354, + "learning_rate": 3.850995862899587e-06, + "loss": 0.8556, + "step": 4297 + }, + { + "epoch": 0.7206119668867232, + "grad_norm": 0.3884469270706177, + "learning_rate": 3.84671418563915e-06, + "loss": 0.9219, + "step": 4298 + }, + { + "epoch": 0.7207796290474694, + "grad_norm": 0.5447671413421631, + "learning_rate": 3.842434323058888e-06, + "loss": 0.8337, + "step": 4299 + }, + { + "epoch": 0.7209472912082154, + "grad_norm": 0.3661268949508667, + "learning_rate": 3.838156276420994e-06, + "loss": 0.8208, + "step": 4300 + }, + { + "epoch": 0.7211149533689616, + "grad_norm": 0.4238070249557495, + "learning_rate": 3.833880046987114e-06, + "loss": 0.797, + "step": 4301 + }, + { + "epoch": 0.7212826155297076, + "grad_norm": 0.6158653497695923, + "learning_rate": 3.8296056360183605e-06, + "loss": 0.8568, + "step": 4302 + }, + { + "epoch": 0.7214502776904538, + "grad_norm": 0.5459095239639282, + "learning_rate": 3.82533304477532e-06, + "loss": 0.8678, + "step": 4303 + }, + { + "epoch": 0.7216179398511998, + "grad_norm": 0.46029114723205566, + "learning_rate": 3.821062274518017e-06, + "loss": 0.7086, + "step": 4304 + }, + { + "epoch": 0.721785602011946, + "grad_norm": 0.6634854674339294, + "learning_rate": 3.816793326505961e-06, + "loss": 0.8317, + "step": 4305 + }, + { + "epoch": 0.721953264172692, + "grad_norm": 0.5513525605201721, + "learning_rate": 3.8125262019981224e-06, + "loss": 0.9066, + "step": 4306 + }, + { + "epoch": 0.7221209263334382, + "grad_norm": 0.9436096549034119, + "learning_rate": 3.8082609022529226e-06, + "loss": 0.8321, + "step": 4307 + }, + { + "epoch": 0.7222885884941842, + "grad_norm": 0.3836483359336853, + "learning_rate": 3.8039974285282567e-06, + "loss": 0.8915, + "step": 4308 + }, + { + "epoch": 0.7224562506549304, + "grad_norm": 0.6042637825012207, + "learning_rate": 3.799735782081473e-06, + "loss": 0.8074, + "step": 4309 + }, + { + "epoch": 0.7226239128156764, + "grad_norm": 0.795918345451355, + "learning_rate": 3.7954759641693795e-06, + "loss": 0.8419, + "step": 4310 + }, + { + "epoch": 0.7227915749764225, + "grad_norm": 0.49923935532569885, + "learning_rate": 3.7912179760482582e-06, + "loss": 0.9364, + "step": 4311 + }, + { + "epoch": 0.7229592371371686, + "grad_norm": 1.1601747274398804, + "learning_rate": 3.7869618189738377e-06, + "loss": 0.9209, + "step": 4312 + }, + { + "epoch": 0.7231268992979147, + "grad_norm": 0.46306419372558594, + "learning_rate": 3.7827074942013085e-06, + "loss": 0.8514, + "step": 4313 + }, + { + "epoch": 0.7232945614586608, + "grad_norm": 0.43269678950309753, + "learning_rate": 3.7784550029853297e-06, + "loss": 0.9167, + "step": 4314 + }, + { + "epoch": 0.7234622236194069, + "grad_norm": 0.36526837944984436, + "learning_rate": 3.774204346580008e-06, + "loss": 0.8636, + "step": 4315 + }, + { + "epoch": 0.723629885780153, + "grad_norm": 0.39668720960617065, + "learning_rate": 3.769955526238921e-06, + "loss": 0.8775, + "step": 4316 + }, + { + "epoch": 0.7237975479408991, + "grad_norm": 0.5399099588394165, + "learning_rate": 3.7657085432150942e-06, + "loss": 0.8603, + "step": 4317 + }, + { + "epoch": 0.7239652101016452, + "grad_norm": 0.4076732397079468, + "learning_rate": 3.7614633987610116e-06, + "loss": 0.8239, + "step": 4318 + }, + { + "epoch": 0.7241328722623913, + "grad_norm": 0.5806723237037659, + "learning_rate": 3.7572200941286284e-06, + "loss": 0.7681, + "step": 4319 + }, + { + "epoch": 0.7243005344231374, + "grad_norm": 0.38126981258392334, + "learning_rate": 3.7529786305693384e-06, + "loss": 0.7744, + "step": 4320 + }, + { + "epoch": 0.7244681965838835, + "grad_norm": 0.7048682570457458, + "learning_rate": 3.7487390093340102e-06, + "loss": 0.8332, + "step": 4321 + }, + { + "epoch": 0.7246358587446295, + "grad_norm": 0.37661606073379517, + "learning_rate": 3.7445012316729556e-06, + "loss": 0.8292, + "step": 4322 + }, + { + "epoch": 0.7248035209053757, + "grad_norm": 0.9588567018508911, + "learning_rate": 3.7402652988359454e-06, + "loss": 0.8884, + "step": 4323 + }, + { + "epoch": 0.7249711830661217, + "grad_norm": 0.4693271815776825, + "learning_rate": 3.736031212072211e-06, + "loss": 0.8441, + "step": 4324 + }, + { + "epoch": 0.7251388452268679, + "grad_norm": 0.45215103030204773, + "learning_rate": 3.7317989726304416e-06, + "loss": 0.8707, + "step": 4325 + }, + { + "epoch": 0.7253065073876139, + "grad_norm": 0.4870089292526245, + "learning_rate": 3.7275685817587738e-06, + "loss": 0.7248, + "step": 4326 + }, + { + "epoch": 0.7254741695483601, + "grad_norm": 0.49108240008354187, + "learning_rate": 3.723340040704797e-06, + "loss": 0.7689, + "step": 4327 + }, + { + "epoch": 0.7256418317091061, + "grad_norm": 0.4583317041397095, + "learning_rate": 3.7191133507155697e-06, + "loss": 0.8313, + "step": 4328 + }, + { + "epoch": 0.7258094938698523, + "grad_norm": 0.4599210321903229, + "learning_rate": 3.7148885130375858e-06, + "loss": 0.835, + "step": 4329 + }, + { + "epoch": 0.7259771560305983, + "grad_norm": 0.4726078510284424, + "learning_rate": 3.7106655289168105e-06, + "loss": 0.8012, + "step": 4330 + }, + { + "epoch": 0.7261448181913445, + "grad_norm": 0.507954478263855, + "learning_rate": 3.70644439959865e-06, + "loss": 0.8109, + "step": 4331 + }, + { + "epoch": 0.7263124803520905, + "grad_norm": 0.557049572467804, + "learning_rate": 3.702225126327965e-06, + "loss": 0.8935, + "step": 4332 + }, + { + "epoch": 0.7264801425128367, + "grad_norm": 0.3942612409591675, + "learning_rate": 3.6980077103490773e-06, + "loss": 0.7554, + "step": 4333 + }, + { + "epoch": 0.7266478046735827, + "grad_norm": 0.41343289613723755, + "learning_rate": 3.6937921529057497e-06, + "loss": 0.8367, + "step": 4334 + }, + { + "epoch": 0.7268154668343288, + "grad_norm": 0.36411136388778687, + "learning_rate": 3.6895784552412084e-06, + "loss": 0.7969, + "step": 4335 + }, + { + "epoch": 0.7269831289950749, + "grad_norm": 0.47542253136634827, + "learning_rate": 3.6853666185981206e-06, + "loss": 0.84, + "step": 4336 + }, + { + "epoch": 0.727150791155821, + "grad_norm": 0.37301474809646606, + "learning_rate": 3.6811566442186064e-06, + "loss": 0.8679, + "step": 4337 + }, + { + "epoch": 0.7273184533165671, + "grad_norm": 0.542221188545227, + "learning_rate": 3.6769485333442466e-06, + "loss": 0.9051, + "step": 4338 + }, + { + "epoch": 0.7274861154773132, + "grad_norm": 0.4234585464000702, + "learning_rate": 3.6727422872160614e-06, + "loss": 0.8635, + "step": 4339 + }, + { + "epoch": 0.7276537776380593, + "grad_norm": 0.31147825717926025, + "learning_rate": 3.668537907074522e-06, + "loss": 0.8151, + "step": 4340 + }, + { + "epoch": 0.7278214397988054, + "grad_norm": 0.7354328632354736, + "learning_rate": 3.6643353941595582e-06, + "loss": 0.8694, + "step": 4341 + }, + { + "epoch": 0.7279891019595515, + "grad_norm": 0.4674883782863617, + "learning_rate": 3.6601347497105366e-06, + "loss": 0.9029, + "step": 4342 + }, + { + "epoch": 0.7281567641202976, + "grad_norm": 0.4117675721645355, + "learning_rate": 3.6559359749662828e-06, + "loss": 0.8143, + "step": 4343 + }, + { + "epoch": 0.7283244262810437, + "grad_norm": 0.5582890510559082, + "learning_rate": 3.6517390711650737e-06, + "loss": 0.8927, + "step": 4344 + }, + { + "epoch": 0.7284920884417898, + "grad_norm": 0.5436880588531494, + "learning_rate": 3.647544039544615e-06, + "loss": 0.8263, + "step": 4345 + }, + { + "epoch": 0.7286597506025358, + "grad_norm": 0.4514172673225403, + "learning_rate": 3.643350881342085e-06, + "loss": 0.8709, + "step": 4346 + }, + { + "epoch": 0.728827412763282, + "grad_norm": 0.5195479989051819, + "learning_rate": 3.6391595977940884e-06, + "loss": 0.802, + "step": 4347 + }, + { + "epoch": 0.728995074924028, + "grad_norm": 0.40355563163757324, + "learning_rate": 3.6349701901366916e-06, + "loss": 0.8583, + "step": 4348 + }, + { + "epoch": 0.7291627370847742, + "grad_norm": 0.39967405796051025, + "learning_rate": 3.6307826596054108e-06, + "loss": 0.7492, + "step": 4349 + }, + { + "epoch": 0.7293303992455202, + "grad_norm": 0.37564578652381897, + "learning_rate": 3.6265970074351843e-06, + "loss": 0.8525, + "step": 4350 + }, + { + "epoch": 0.7294980614062664, + "grad_norm": 0.5045909285545349, + "learning_rate": 3.6224132348604213e-06, + "loss": 0.8313, + "step": 4351 + }, + { + "epoch": 0.7296657235670124, + "grad_norm": 0.35503095388412476, + "learning_rate": 3.618231343114972e-06, + "loss": 0.7584, + "step": 4352 + }, + { + "epoch": 0.7298333857277586, + "grad_norm": 0.3874315619468689, + "learning_rate": 3.614051333432124e-06, + "loss": 0.77, + "step": 4353 + }, + { + "epoch": 0.7300010478885046, + "grad_norm": 0.4769584834575653, + "learning_rate": 3.6098732070446084e-06, + "loss": 0.8299, + "step": 4354 + }, + { + "epoch": 0.7301687100492508, + "grad_norm": 0.39705613255500793, + "learning_rate": 3.6056969651846176e-06, + "loss": 0.8454, + "step": 4355 + }, + { + "epoch": 0.7303363722099968, + "grad_norm": 0.6885538697242737, + "learning_rate": 3.6015226090837664e-06, + "loss": 0.841, + "step": 4356 + }, + { + "epoch": 0.730504034370743, + "grad_norm": 0.6540099382400513, + "learning_rate": 3.5973501399731326e-06, + "loss": 0.8297, + "step": 4357 + }, + { + "epoch": 0.730671696531489, + "grad_norm": 0.4100639522075653, + "learning_rate": 3.593179559083225e-06, + "loss": 0.8448, + "step": 4358 + }, + { + "epoch": 0.7308393586922352, + "grad_norm": 0.6150904893875122, + "learning_rate": 3.589010867643997e-06, + "loss": 0.7983, + "step": 4359 + }, + { + "epoch": 0.7310070208529812, + "grad_norm": 0.4606917202472687, + "learning_rate": 3.5848440668848538e-06, + "loss": 0.8698, + "step": 4360 + }, + { + "epoch": 0.7311746830137273, + "grad_norm": 0.5340200066566467, + "learning_rate": 3.580679158034629e-06, + "loss": 0.788, + "step": 4361 + }, + { + "epoch": 0.7313423451744734, + "grad_norm": 0.5118436217308044, + "learning_rate": 3.5765161423216145e-06, + "loss": 0.8007, + "step": 4362 + }, + { + "epoch": 0.7315100073352195, + "grad_norm": 0.3920898139476776, + "learning_rate": 3.572355020973529e-06, + "loss": 0.8444, + "step": 4363 + }, + { + "epoch": 0.7316776694959656, + "grad_norm": 0.3829520344734192, + "learning_rate": 3.568195795217537e-06, + "loss": 0.8475, + "step": 4364 + }, + { + "epoch": 0.7318453316567117, + "grad_norm": 0.48329228162765503, + "learning_rate": 3.564038466280252e-06, + "loss": 0.9246, + "step": 4365 + }, + { + "epoch": 0.7320129938174578, + "grad_norm": 0.467337965965271, + "learning_rate": 3.5598830353877158e-06, + "loss": 0.8442, + "step": 4366 + }, + { + "epoch": 0.7321806559782039, + "grad_norm": 0.543903112411499, + "learning_rate": 3.5557295037654227e-06, + "loss": 0.8777, + "step": 4367 + }, + { + "epoch": 0.73234831813895, + "grad_norm": 0.4708911180496216, + "learning_rate": 3.5515778726382967e-06, + "loss": 0.8563, + "step": 4368 + }, + { + "epoch": 0.7325159802996961, + "grad_norm": 0.44531819224357605, + "learning_rate": 3.5474281432307024e-06, + "loss": 0.7835, + "step": 4369 + }, + { + "epoch": 0.7326836424604423, + "grad_norm": 0.9017184972763062, + "learning_rate": 3.54328031676645e-06, + "loss": 0.8309, + "step": 4370 + }, + { + "epoch": 0.7328513046211883, + "grad_norm": 0.39330801367759705, + "learning_rate": 3.5391343944687906e-06, + "loss": 0.8291, + "step": 4371 + }, + { + "epoch": 0.7330189667819345, + "grad_norm": 0.4601340889930725, + "learning_rate": 3.5349903775603955e-06, + "loss": 0.9007, + "step": 4372 + }, + { + "epoch": 0.7331866289426805, + "grad_norm": 0.3983735144138336, + "learning_rate": 3.530848267263394e-06, + "loss": 0.8141, + "step": 4373 + }, + { + "epoch": 0.7333542911034266, + "grad_norm": 0.3980402648448944, + "learning_rate": 3.5267080647993468e-06, + "loss": 0.8734, + "step": 4374 + }, + { + "epoch": 0.7335219532641727, + "grad_norm": 0.43856269121170044, + "learning_rate": 3.522569771389246e-06, + "loss": 0.8857, + "step": 4375 + }, + { + "epoch": 0.7336896154249188, + "grad_norm": 0.53123539686203, + "learning_rate": 3.5184333882535314e-06, + "loss": 0.8811, + "step": 4376 + }, + { + "epoch": 0.7338572775856649, + "grad_norm": 0.38238826394081116, + "learning_rate": 3.514298916612071e-06, + "loss": 0.8164, + "step": 4377 + }, + { + "epoch": 0.734024939746411, + "grad_norm": 0.551599383354187, + "learning_rate": 3.510166357684168e-06, + "loss": 0.896, + "step": 4378 + }, + { + "epoch": 0.7341926019071571, + "grad_norm": 1.1665607690811157, + "learning_rate": 3.5060357126885725e-06, + "loss": 0.8109, + "step": 4379 + }, + { + "epoch": 0.7343602640679032, + "grad_norm": 0.4617224633693695, + "learning_rate": 3.5019069828434547e-06, + "loss": 0.8853, + "step": 4380 + }, + { + "epoch": 0.7345279262286493, + "grad_norm": 0.4224250614643097, + "learning_rate": 3.497780169366436e-06, + "loss": 0.9396, + "step": 4381 + }, + { + "epoch": 0.7346955883893954, + "grad_norm": 0.36724886298179626, + "learning_rate": 3.493655273474561e-06, + "loss": 0.8428, + "step": 4382 + }, + { + "epoch": 0.7348632505501415, + "grad_norm": 0.47203654050827026, + "learning_rate": 3.489532296384309e-06, + "loss": 0.8778, + "step": 4383 + }, + { + "epoch": 0.7350309127108876, + "grad_norm": 0.4965214133262634, + "learning_rate": 3.4854112393116047e-06, + "loss": 0.8247, + "step": 4384 + }, + { + "epoch": 0.7351985748716336, + "grad_norm": 0.5500672459602356, + "learning_rate": 3.481292103471794e-06, + "loss": 0.9078, + "step": 4385 + }, + { + "epoch": 0.7353662370323798, + "grad_norm": 0.6066471338272095, + "learning_rate": 3.4771748900796597e-06, + "loss": 0.867, + "step": 4386 + }, + { + "epoch": 0.7355338991931258, + "grad_norm": 0.3829139173030853, + "learning_rate": 3.473059600349424e-06, + "loss": 0.8491, + "step": 4387 + }, + { + "epoch": 0.735701561353872, + "grad_norm": 0.77088463306427, + "learning_rate": 3.4689462354947312e-06, + "loss": 0.8976, + "step": 4388 + }, + { + "epoch": 0.735869223514618, + "grad_norm": 0.39039182662963867, + "learning_rate": 3.4648347967286655e-06, + "loss": 0.798, + "step": 4389 + }, + { + "epoch": 0.7360368856753642, + "grad_norm": 0.6210637092590332, + "learning_rate": 3.4607252852637494e-06, + "loss": 0.7954, + "step": 4390 + }, + { + "epoch": 0.7362045478361102, + "grad_norm": 0.3575737476348877, + "learning_rate": 3.456617702311914e-06, + "loss": 0.8076, + "step": 4391 + }, + { + "epoch": 0.7363722099968564, + "grad_norm": 0.41606733202934265, + "learning_rate": 3.4525120490845444e-06, + "loss": 0.9159, + "step": 4392 + }, + { + "epoch": 0.7365398721576024, + "grad_norm": 0.3249188959598541, + "learning_rate": 3.448408326792451e-06, + "loss": 0.7628, + "step": 4393 + }, + { + "epoch": 0.7367075343183486, + "grad_norm": 0.572081983089447, + "learning_rate": 3.4443065366458683e-06, + "loss": 0.8155, + "step": 4394 + }, + { + "epoch": 0.7368751964790946, + "grad_norm": 0.36839038133621216, + "learning_rate": 3.4402066798544664e-06, + "loss": 0.82, + "step": 4395 + }, + { + "epoch": 0.7370428586398408, + "grad_norm": 0.4063968360424042, + "learning_rate": 3.43610875762734e-06, + "loss": 0.8493, + "step": 4396 + }, + { + "epoch": 0.7372105208005868, + "grad_norm": 0.3296797573566437, + "learning_rate": 3.432012771173021e-06, + "loss": 0.8519, + "step": 4397 + }, + { + "epoch": 0.737378182961333, + "grad_norm": 0.39701390266418457, + "learning_rate": 3.427918721699469e-06, + "loss": 0.8686, + "step": 4398 + }, + { + "epoch": 0.737545845122079, + "grad_norm": 0.4578464925289154, + "learning_rate": 3.423826610414066e-06, + "loss": 0.7675, + "step": 4399 + }, + { + "epoch": 0.7377135072828251, + "grad_norm": 0.4097997546195984, + "learning_rate": 3.4197364385236255e-06, + "loss": 0.8792, + "step": 4400 + }, + { + "epoch": 0.7378811694435712, + "grad_norm": 0.36022815108299255, + "learning_rate": 3.415648207234393e-06, + "loss": 0.6904, + "step": 4401 + }, + { + "epoch": 0.7380488316043173, + "grad_norm": 0.5350514054298401, + "learning_rate": 3.4115619177520343e-06, + "loss": 0.8796, + "step": 4402 + }, + { + "epoch": 0.7382164937650634, + "grad_norm": 0.5072664618492126, + "learning_rate": 3.4074775712816533e-06, + "loss": 0.9138, + "step": 4403 + }, + { + "epoch": 0.7383841559258095, + "grad_norm": 0.36521509289741516, + "learning_rate": 3.40339516902777e-06, + "loss": 0.7506, + "step": 4404 + }, + { + "epoch": 0.7385518180865556, + "grad_norm": 0.3442452847957611, + "learning_rate": 3.3993147121943324e-06, + "loss": 0.8652, + "step": 4405 + }, + { + "epoch": 0.7387194802473017, + "grad_norm": 0.399734228849411, + "learning_rate": 3.395236201984725e-06, + "loss": 0.7461, + "step": 4406 + }, + { + "epoch": 0.7388871424080478, + "grad_norm": 0.35992172360420227, + "learning_rate": 3.3911596396017433e-06, + "loss": 0.8682, + "step": 4407 + }, + { + "epoch": 0.7390548045687939, + "grad_norm": 0.4022121727466583, + "learning_rate": 3.387085026247624e-06, + "loss": 0.8703, + "step": 4408 + }, + { + "epoch": 0.73922246672954, + "grad_norm": 0.4650803506374359, + "learning_rate": 3.3830123631240163e-06, + "loss": 0.817, + "step": 4409 + }, + { + "epoch": 0.7393901288902861, + "grad_norm": 0.4014977514743805, + "learning_rate": 3.378941651431996e-06, + "loss": 0.7755, + "step": 4410 + }, + { + "epoch": 0.7395577910510321, + "grad_norm": 0.5557853579521179, + "learning_rate": 3.374872892372074e-06, + "loss": 0.8462, + "step": 4411 + }, + { + "epoch": 0.7397254532117783, + "grad_norm": 0.380247563123703, + "learning_rate": 3.370806087144174e-06, + "loss": 0.8086, + "step": 4412 + }, + { + "epoch": 0.7398931153725243, + "grad_norm": 0.5533876419067383, + "learning_rate": 3.366741236947644e-06, + "loss": 0.814, + "step": 4413 + }, + { + "epoch": 0.7400607775332705, + "grad_norm": 0.8545756936073303, + "learning_rate": 3.3626783429812635e-06, + "loss": 0.8038, + "step": 4414 + }, + { + "epoch": 0.7402284396940165, + "grad_norm": 0.5220400094985962, + "learning_rate": 3.3586174064432262e-06, + "loss": 0.9231, + "step": 4415 + }, + { + "epoch": 0.7403961018547627, + "grad_norm": 0.4009403586387634, + "learning_rate": 3.354558428531154e-06, + "loss": 0.9235, + "step": 4416 + }, + { + "epoch": 0.7405637640155087, + "grad_norm": 0.426805704832077, + "learning_rate": 3.3505014104420976e-06, + "loss": 0.7895, + "step": 4417 + }, + { + "epoch": 0.7407314261762549, + "grad_norm": 0.3550259470939636, + "learning_rate": 3.346446353372508e-06, + "loss": 0.8136, + "step": 4418 + }, + { + "epoch": 0.7408990883370009, + "grad_norm": 0.38828107714653015, + "learning_rate": 3.342393258518278e-06, + "loss": 0.8093, + "step": 4419 + }, + { + "epoch": 0.7410667504977471, + "grad_norm": 1.0287100076675415, + "learning_rate": 3.338342127074721e-06, + "loss": 0.839, + "step": 4420 + }, + { + "epoch": 0.7412344126584931, + "grad_norm": 0.45076051354408264, + "learning_rate": 3.334292960236556e-06, + "loss": 0.8088, + "step": 4421 + }, + { + "epoch": 0.7414020748192393, + "grad_norm": 0.4348759949207306, + "learning_rate": 3.330245759197941e-06, + "loss": 0.8436, + "step": 4422 + }, + { + "epoch": 0.7415697369799853, + "grad_norm": 0.4137447774410248, + "learning_rate": 3.326200525152441e-06, + "loss": 0.8775, + "step": 4423 + }, + { + "epoch": 0.7417373991407314, + "grad_norm": 0.5322988629341125, + "learning_rate": 3.322157259293044e-06, + "loss": 0.8508, + "step": 4424 + }, + { + "epoch": 0.7419050613014775, + "grad_norm": 0.5858178734779358, + "learning_rate": 3.318115962812165e-06, + "loss": 0.8505, + "step": 4425 + }, + { + "epoch": 0.7420727234622236, + "grad_norm": 0.44442102313041687, + "learning_rate": 3.31407663690163e-06, + "loss": 0.7975, + "step": 4426 + }, + { + "epoch": 0.7422403856229697, + "grad_norm": 0.3603506088256836, + "learning_rate": 3.3100392827526807e-06, + "loss": 0.8318, + "step": 4427 + }, + { + "epoch": 0.7424080477837158, + "grad_norm": 0.4573531448841095, + "learning_rate": 3.306003901555991e-06, + "loss": 0.892, + "step": 4428 + }, + { + "epoch": 0.7425757099444619, + "grad_norm": 0.4490502178668976, + "learning_rate": 3.3019704945016385e-06, + "loss": 0.8911, + "step": 4429 + }, + { + "epoch": 0.742743372105208, + "grad_norm": 0.6446260809898376, + "learning_rate": 3.297939062779131e-06, + "loss": 0.9423, + "step": 4430 + }, + { + "epoch": 0.7429110342659541, + "grad_norm": 0.4458027780056, + "learning_rate": 3.2939096075773835e-06, + "loss": 0.8041, + "step": 4431 + }, + { + "epoch": 0.7430786964267002, + "grad_norm": 0.40491101145744324, + "learning_rate": 3.2898821300847294e-06, + "loss": 0.7675, + "step": 4432 + }, + { + "epoch": 0.7432463585874463, + "grad_norm": 0.650396466255188, + "learning_rate": 3.2858566314889295e-06, + "loss": 0.8963, + "step": 4433 + }, + { + "epoch": 0.7434140207481924, + "grad_norm": 0.4394979476928711, + "learning_rate": 3.281833112977145e-06, + "loss": 0.7744, + "step": 4434 + }, + { + "epoch": 0.7435816829089384, + "grad_norm": 0.488955557346344, + "learning_rate": 3.2778115757359695e-06, + "loss": 0.9099, + "step": 4435 + }, + { + "epoch": 0.7437493450696846, + "grad_norm": 0.4294459819793701, + "learning_rate": 3.2737920209513994e-06, + "loss": 0.7751, + "step": 4436 + }, + { + "epoch": 0.7439170072304306, + "grad_norm": 0.3901923596858978, + "learning_rate": 3.269774449808849e-06, + "loss": 0.8058, + "step": 4437 + }, + { + "epoch": 0.7440846693911768, + "grad_norm": 0.4745529294013977, + "learning_rate": 3.2657588634931516e-06, + "loss": 0.8483, + "step": 4438 + }, + { + "epoch": 0.7442523315519228, + "grad_norm": 0.42034587264060974, + "learning_rate": 3.261745263188557e-06, + "loss": 0.8383, + "step": 4439 + }, + { + "epoch": 0.744419993712669, + "grad_norm": 0.3854556977748871, + "learning_rate": 3.2577336500787237e-06, + "loss": 0.7848, + "step": 4440 + }, + { + "epoch": 0.744587655873415, + "grad_norm": 0.35779571533203125, + "learning_rate": 3.253724025346723e-06, + "loss": 0.8042, + "step": 4441 + }, + { + "epoch": 0.7447553180341612, + "grad_norm": 0.48553743958473206, + "learning_rate": 3.249716390175046e-06, + "loss": 0.8119, + "step": 4442 + }, + { + "epoch": 0.7449229801949072, + "grad_norm": 0.4442885220050812, + "learning_rate": 3.2457107457455915e-06, + "loss": 0.8189, + "step": 4443 + }, + { + "epoch": 0.7450906423556534, + "grad_norm": 0.3811306059360504, + "learning_rate": 3.241707093239681e-06, + "loss": 0.8197, + "step": 4444 + }, + { + "epoch": 0.7452583045163994, + "grad_norm": 0.5182856321334839, + "learning_rate": 3.23770543383803e-06, + "loss": 0.8627, + "step": 4445 + }, + { + "epoch": 0.7454259666771456, + "grad_norm": 0.5060182809829712, + "learning_rate": 3.2337057687207816e-06, + "loss": 0.8235, + "step": 4446 + }, + { + "epoch": 0.7455936288378916, + "grad_norm": 0.5154516100883484, + "learning_rate": 3.229708099067491e-06, + "loss": 0.893, + "step": 4447 + }, + { + "epoch": 0.7457612909986377, + "grad_norm": 0.40294402837753296, + "learning_rate": 3.225712426057115e-06, + "loss": 0.948, + "step": 4448 + }, + { + "epoch": 0.7459289531593838, + "grad_norm": 0.36278417706489563, + "learning_rate": 3.2217187508680314e-06, + "loss": 0.7185, + "step": 4449 + }, + { + "epoch": 0.7460966153201299, + "grad_norm": 0.4962402582168579, + "learning_rate": 3.217727074678023e-06, + "loss": 0.8488, + "step": 4450 + }, + { + "epoch": 0.746264277480876, + "grad_norm": 0.4561363160610199, + "learning_rate": 3.2137373986642807e-06, + "loss": 0.8673, + "step": 4451 + }, + { + "epoch": 0.7464319396416221, + "grad_norm": 0.5583760142326355, + "learning_rate": 3.2097497240034147e-06, + "loss": 0.841, + "step": 4452 + }, + { + "epoch": 0.7465996018023682, + "grad_norm": 0.4732387065887451, + "learning_rate": 3.205764051871435e-06, + "loss": 0.8238, + "step": 4453 + }, + { + "epoch": 0.7467672639631143, + "grad_norm": 0.4361114799976349, + "learning_rate": 3.201780383443772e-06, + "loss": 0.8675, + "step": 4454 + }, + { + "epoch": 0.7469349261238605, + "grad_norm": 0.6804105043411255, + "learning_rate": 3.1977987198952543e-06, + "loss": 0.8752, + "step": 4455 + }, + { + "epoch": 0.7471025882846065, + "grad_norm": 0.6557078957557678, + "learning_rate": 3.193819062400122e-06, + "loss": 0.7997, + "step": 4456 + }, + { + "epoch": 0.7472702504453527, + "grad_norm": 0.5991407036781311, + "learning_rate": 3.1898414121320277e-06, + "loss": 0.8506, + "step": 4457 + }, + { + "epoch": 0.7474379126060987, + "grad_norm": 0.5412923097610474, + "learning_rate": 3.185865770264035e-06, + "loss": 0.9054, + "step": 4458 + }, + { + "epoch": 0.7476055747668449, + "grad_norm": 0.48456665873527527, + "learning_rate": 3.1818921379685997e-06, + "loss": 0.9034, + "step": 4459 + }, + { + "epoch": 0.7477732369275909, + "grad_norm": 0.4137357771396637, + "learning_rate": 3.177920516417603e-06, + "loss": 0.8505, + "step": 4460 + }, + { + "epoch": 0.747940899088337, + "grad_norm": 0.48029109835624695, + "learning_rate": 3.1739509067823192e-06, + "loss": 0.8321, + "step": 4461 + }, + { + "epoch": 0.7481085612490831, + "grad_norm": 0.8238043785095215, + "learning_rate": 3.1699833102334397e-06, + "loss": 0.8374, + "step": 4462 + }, + { + "epoch": 0.7482762234098292, + "grad_norm": 0.42662978172302246, + "learning_rate": 3.166017727941062e-06, + "loss": 0.7012, + "step": 4463 + }, + { + "epoch": 0.7484438855705753, + "grad_norm": 0.42485249042510986, + "learning_rate": 3.1620541610746747e-06, + "loss": 0.8489, + "step": 4464 + }, + { + "epoch": 0.7486115477313214, + "grad_norm": 0.35580894351005554, + "learning_rate": 3.1580926108031884e-06, + "loss": 0.8409, + "step": 4465 + }, + { + "epoch": 0.7487792098920675, + "grad_norm": 0.3833909332752228, + "learning_rate": 3.154133078294915e-06, + "loss": 0.8809, + "step": 4466 + }, + { + "epoch": 0.7489468720528136, + "grad_norm": 0.4267159402370453, + "learning_rate": 3.1501755647175693e-06, + "loss": 0.7405, + "step": 4467 + }, + { + "epoch": 0.7491145342135597, + "grad_norm": 0.5588497519493103, + "learning_rate": 3.1462200712382662e-06, + "loss": 0.8341, + "step": 4468 + }, + { + "epoch": 0.7492821963743058, + "grad_norm": 0.48638758063316345, + "learning_rate": 3.142266599023537e-06, + "loss": 0.8086, + "step": 4469 + }, + { + "epoch": 0.7494498585350519, + "grad_norm": 0.37331661581993103, + "learning_rate": 3.1383151492393015e-06, + "loss": 0.8458, + "step": 4470 + }, + { + "epoch": 0.749617520695798, + "grad_norm": 0.5566949844360352, + "learning_rate": 3.1343657230508985e-06, + "loss": 0.8528, + "step": 4471 + }, + { + "epoch": 0.749785182856544, + "grad_norm": 0.38358446955680847, + "learning_rate": 3.130418321623061e-06, + "loss": 0.8063, + "step": 4472 + }, + { + "epoch": 0.7499528450172902, + "grad_norm": 0.5433733463287354, + "learning_rate": 3.1264729461199208e-06, + "loss": 0.845, + "step": 4473 + }, + { + "epoch": 0.7501205071780362, + "grad_norm": 0.46310821175575256, + "learning_rate": 3.1225295977050265e-06, + "loss": 0.8762, + "step": 4474 + }, + { + "epoch": 0.7502881693387824, + "grad_norm": 0.3718121647834778, + "learning_rate": 3.1185882775413123e-06, + "loss": 0.7871, + "step": 4475 + }, + { + "epoch": 0.7504558314995284, + "grad_norm": 0.44388100504875183, + "learning_rate": 3.114648986791129e-06, + "loss": 0.8149, + "step": 4476 + }, + { + "epoch": 0.7506234936602746, + "grad_norm": 0.4701133668422699, + "learning_rate": 3.11071172661622e-06, + "loss": 0.8766, + "step": 4477 + }, + { + "epoch": 0.7507911558210206, + "grad_norm": 0.4463254511356354, + "learning_rate": 3.1067764981777292e-06, + "loss": 0.8164, + "step": 4478 + }, + { + "epoch": 0.7509588179817668, + "grad_norm": 0.5636247396469116, + "learning_rate": 3.1028433026362093e-06, + "loss": 0.8465, + "step": 4479 + }, + { + "epoch": 0.7511264801425128, + "grad_norm": 1.0990864038467407, + "learning_rate": 3.098912141151602e-06, + "loss": 0.8136, + "step": 4480 + }, + { + "epoch": 0.751294142303259, + "grad_norm": 0.4750821888446808, + "learning_rate": 3.0949830148832627e-06, + "loss": 0.8361, + "step": 4481 + }, + { + "epoch": 0.751461804464005, + "grad_norm": 0.4538353383541107, + "learning_rate": 3.0910559249899364e-06, + "loss": 0.9258, + "step": 4482 + }, + { + "epoch": 0.7516294666247512, + "grad_norm": 0.408823162317276, + "learning_rate": 3.0871308726297667e-06, + "loss": 0.758, + "step": 4483 + }, + { + "epoch": 0.7517971287854972, + "grad_norm": 0.6528570055961609, + "learning_rate": 3.083207858960302e-06, + "loss": 0.8146, + "step": 4484 + }, + { + "epoch": 0.7519647909462434, + "grad_norm": 0.42876380681991577, + "learning_rate": 3.079286885138497e-06, + "loss": 0.8435, + "step": 4485 + }, + { + "epoch": 0.7521324531069894, + "grad_norm": 0.7172784805297852, + "learning_rate": 3.0753679523206803e-06, + "loss": 0.8946, + "step": 4486 + }, + { + "epoch": 0.7523001152677355, + "grad_norm": 0.5575166344642639, + "learning_rate": 3.0714510616626012e-06, + "loss": 0.8673, + "step": 4487 + }, + { + "epoch": 0.7524677774284816, + "grad_norm": 0.439159631729126, + "learning_rate": 3.067536214319402e-06, + "loss": 0.7876, + "step": 4488 + }, + { + "epoch": 0.7526354395892277, + "grad_norm": 0.3544710874557495, + "learning_rate": 3.0636234114456144e-06, + "loss": 0.7859, + "step": 4489 + }, + { + "epoch": 0.7528031017499738, + "grad_norm": 0.3452700972557068, + "learning_rate": 3.059712654195176e-06, + "loss": 0.8961, + "step": 4490 + }, + { + "epoch": 0.7529707639107199, + "grad_norm": 0.3834369480609894, + "learning_rate": 3.055803943721417e-06, + "loss": 0.8511, + "step": 4491 + }, + { + "epoch": 0.753138426071466, + "grad_norm": 0.3766850531101227, + "learning_rate": 3.051897281177061e-06, + "loss": 0.79, + "step": 4492 + }, + { + "epoch": 0.7533060882322121, + "grad_norm": 0.3521707057952881, + "learning_rate": 3.047992667714237e-06, + "loss": 0.6716, + "step": 4493 + }, + { + "epoch": 0.7534737503929582, + "grad_norm": 0.4038065969944, + "learning_rate": 3.044090104484456e-06, + "loss": 0.7564, + "step": 4494 + }, + { + "epoch": 0.7536414125537043, + "grad_norm": 0.4788320064544678, + "learning_rate": 3.0401895926386417e-06, + "loss": 0.8327, + "step": 4495 + }, + { + "epoch": 0.7538090747144504, + "grad_norm": 0.5993980765342712, + "learning_rate": 3.0362911333270983e-06, + "loss": 0.8236, + "step": 4496 + }, + { + "epoch": 0.7539767368751965, + "grad_norm": 0.538329005241394, + "learning_rate": 3.032394727699527e-06, + "loss": 0.8922, + "step": 4497 + }, + { + "epoch": 0.7541443990359425, + "grad_norm": 0.3130503296852112, + "learning_rate": 3.0285003769050314e-06, + "loss": 0.7434, + "step": 4498 + }, + { + "epoch": 0.7543120611966887, + "grad_norm": 0.4620181918144226, + "learning_rate": 3.024608082092102e-06, + "loss": 0.8075, + "step": 4499 + }, + { + "epoch": 0.7544797233574347, + "grad_norm": 0.40521252155303955, + "learning_rate": 3.0207178444086206e-06, + "loss": 0.881, + "step": 4500 + }, + { + "epoch": 0.7546473855181809, + "grad_norm": 0.39265090227127075, + "learning_rate": 3.0168296650018736e-06, + "loss": 0.8214, + "step": 4501 + }, + { + "epoch": 0.7548150476789269, + "grad_norm": 0.48276379704475403, + "learning_rate": 3.012943545018526e-06, + "loss": 0.8697, + "step": 4502 + }, + { + "epoch": 0.7549827098396731, + "grad_norm": 0.38085293769836426, + "learning_rate": 3.0090594856046473e-06, + "loss": 0.8828, + "step": 4503 + }, + { + "epoch": 0.7551503720004191, + "grad_norm": 0.3857152462005615, + "learning_rate": 3.0051774879056995e-06, + "loss": 0.8339, + "step": 4504 + }, + { + "epoch": 0.7553180341611653, + "grad_norm": 0.6330828070640564, + "learning_rate": 3.00129755306652e-06, + "loss": 0.6985, + "step": 4505 + }, + { + "epoch": 0.7554856963219113, + "grad_norm": 0.3679242432117462, + "learning_rate": 2.997419682231355e-06, + "loss": 0.8357, + "step": 4506 + }, + { + "epoch": 0.7556533584826575, + "grad_norm": 0.5782294273376465, + "learning_rate": 2.993543876543842e-06, + "loss": 0.8307, + "step": 4507 + }, + { + "epoch": 0.7558210206434035, + "grad_norm": 0.8186786770820618, + "learning_rate": 2.9896701371469983e-06, + "loss": 0.7709, + "step": 4508 + }, + { + "epoch": 0.7559886828041497, + "grad_norm": 0.46019449830055237, + "learning_rate": 2.9857984651832396e-06, + "loss": 0.8934, + "step": 4509 + }, + { + "epoch": 0.7561563449648957, + "grad_norm": 0.39549770951271057, + "learning_rate": 2.981928861794364e-06, + "loss": 0.8112, + "step": 4510 + }, + { + "epoch": 0.7563240071256419, + "grad_norm": 0.4370620548725128, + "learning_rate": 2.978061328121573e-06, + "loss": 0.8993, + "step": 4511 + }, + { + "epoch": 0.7564916692863879, + "grad_norm": 0.421627014875412, + "learning_rate": 2.9741958653054494e-06, + "loss": 0.841, + "step": 4512 + }, + { + "epoch": 0.756659331447134, + "grad_norm": 0.36653319001197815, + "learning_rate": 2.9703324744859653e-06, + "loss": 0.8663, + "step": 4513 + }, + { + "epoch": 0.7568269936078801, + "grad_norm": 0.38584211468696594, + "learning_rate": 2.966471156802477e-06, + "loss": 0.875, + "step": 4514 + }, + { + "epoch": 0.7569946557686262, + "grad_norm": 0.4006466269493103, + "learning_rate": 2.962611913393744e-06, + "loss": 0.8039, + "step": 4515 + }, + { + "epoch": 0.7571623179293723, + "grad_norm": 0.45578905940055847, + "learning_rate": 2.958754745397896e-06, + "loss": 0.7878, + "step": 4516 + }, + { + "epoch": 0.7573299800901184, + "grad_norm": 0.7102530598640442, + "learning_rate": 2.9548996539524676e-06, + "loss": 0.8346, + "step": 4517 + }, + { + "epoch": 0.7574976422508645, + "grad_norm": 0.4307156205177307, + "learning_rate": 2.9510466401943703e-06, + "loss": 0.8183, + "step": 4518 + }, + { + "epoch": 0.7576653044116106, + "grad_norm": 0.5517482161521912, + "learning_rate": 2.947195705259901e-06, + "loss": 0.8893, + "step": 4519 + }, + { + "epoch": 0.7578329665723567, + "grad_norm": 0.4402889311313629, + "learning_rate": 2.943346850284754e-06, + "loss": 0.7601, + "step": 4520 + }, + { + "epoch": 0.7580006287331028, + "grad_norm": 0.6082161068916321, + "learning_rate": 2.9395000764040007e-06, + "loss": 0.8256, + "step": 4521 + }, + { + "epoch": 0.7581682908938489, + "grad_norm": 0.43149957060813904, + "learning_rate": 2.935655384752105e-06, + "loss": 0.8834, + "step": 4522 + }, + { + "epoch": 0.758335953054595, + "grad_norm": 0.3590730130672455, + "learning_rate": 2.9318127764629135e-06, + "loss": 0.7763, + "step": 4523 + }, + { + "epoch": 0.758503615215341, + "grad_norm": 0.5444132089614868, + "learning_rate": 2.9279722526696552e-06, + "loss": 0.904, + "step": 4524 + }, + { + "epoch": 0.7586712773760872, + "grad_norm": 0.38523122668266296, + "learning_rate": 2.9241338145049546e-06, + "loss": 0.7918, + "step": 4525 + }, + { + "epoch": 0.7588389395368332, + "grad_norm": 0.9955673217773438, + "learning_rate": 2.9202974631008084e-06, + "loss": 0.8853, + "step": 4526 + }, + { + "epoch": 0.7590066016975794, + "grad_norm": 0.4135846197605133, + "learning_rate": 2.9164631995886095e-06, + "loss": 0.7754, + "step": 4527 + }, + { + "epoch": 0.7591742638583254, + "grad_norm": 0.5456740260124207, + "learning_rate": 2.9126310250991275e-06, + "loss": 0.8535, + "step": 4528 + }, + { + "epoch": 0.7593419260190716, + "grad_norm": 1.1154274940490723, + "learning_rate": 2.908800940762514e-06, + "loss": 0.9219, + "step": 4529 + }, + { + "epoch": 0.7595095881798176, + "grad_norm": 0.40234553813934326, + "learning_rate": 2.904972947708312e-06, + "loss": 0.8607, + "step": 4530 + }, + { + "epoch": 0.7596772503405638, + "grad_norm": 0.33538851141929626, + "learning_rate": 2.9011470470654512e-06, + "loss": 0.8112, + "step": 4531 + }, + { + "epoch": 0.7598449125013098, + "grad_norm": 0.42388421297073364, + "learning_rate": 2.8973232399622242e-06, + "loss": 0.8061, + "step": 4532 + }, + { + "epoch": 0.760012574662056, + "grad_norm": 0.6167278289794922, + "learning_rate": 2.8935015275263235e-06, + "loss": 0.8011, + "step": 4533 + }, + { + "epoch": 0.760180236822802, + "grad_norm": 0.473337858915329, + "learning_rate": 2.889681910884825e-06, + "loss": 0.9186, + "step": 4534 + }, + { + "epoch": 0.7603478989835482, + "grad_norm": 0.4435788691043854, + "learning_rate": 2.8858643911641726e-06, + "loss": 0.831, + "step": 4535 + }, + { + "epoch": 0.7605155611442942, + "grad_norm": 0.381301611661911, + "learning_rate": 2.882048969490208e-06, + "loss": 0.887, + "step": 4536 + }, + { + "epoch": 0.7606832233050403, + "grad_norm": 0.5315547585487366, + "learning_rate": 2.8782356469881424e-06, + "loss": 0.9597, + "step": 4537 + }, + { + "epoch": 0.7608508854657864, + "grad_norm": 0.43942558765411377, + "learning_rate": 2.874424424782569e-06, + "loss": 0.8002, + "step": 4538 + }, + { + "epoch": 0.7610185476265325, + "grad_norm": 0.35768264532089233, + "learning_rate": 2.8706153039974705e-06, + "loss": 0.8481, + "step": 4539 + }, + { + "epoch": 0.7611862097872786, + "grad_norm": 0.405945748090744, + "learning_rate": 2.8668082857562006e-06, + "loss": 0.7964, + "step": 4540 + }, + { + "epoch": 0.7613538719480247, + "grad_norm": 0.3525673449039459, + "learning_rate": 2.863003371181493e-06, + "loss": 0.8708, + "step": 4541 + }, + { + "epoch": 0.7615215341087709, + "grad_norm": 0.5565782189369202, + "learning_rate": 2.85920056139547e-06, + "loss": 0.8222, + "step": 4542 + }, + { + "epoch": 0.7616891962695169, + "grad_norm": 0.36413130164146423, + "learning_rate": 2.855399857519623e-06, + "loss": 0.894, + "step": 4543 + }, + { + "epoch": 0.7618568584302631, + "grad_norm": 0.4103677570819855, + "learning_rate": 2.85160126067483e-06, + "loss": 0.9292, + "step": 4544 + }, + { + "epoch": 0.7620245205910091, + "grad_norm": 0.4916219711303711, + "learning_rate": 2.847804771981343e-06, + "loss": 0.8147, + "step": 4545 + }, + { + "epoch": 0.7621921827517553, + "grad_norm": 0.4736553430557251, + "learning_rate": 2.8440103925587904e-06, + "loss": 0.7978, + "step": 4546 + }, + { + "epoch": 0.7623598449125013, + "grad_norm": 0.38820213079452515, + "learning_rate": 2.840218123526186e-06, + "loss": 0.9285, + "step": 4547 + }, + { + "epoch": 0.7625275070732475, + "grad_norm": 0.4562012553215027, + "learning_rate": 2.8364279660019135e-06, + "loss": 0.8501, + "step": 4548 + }, + { + "epoch": 0.7626951692339935, + "grad_norm": 0.4280639588832855, + "learning_rate": 2.8326399211037413e-06, + "loss": 0.8802, + "step": 4549 + }, + { + "epoch": 0.7628628313947396, + "grad_norm": 0.7219691872596741, + "learning_rate": 2.8288539899488086e-06, + "loss": 0.8217, + "step": 4550 + }, + { + "epoch": 0.7630304935554857, + "grad_norm": 0.4507511258125305, + "learning_rate": 2.825070173653631e-06, + "loss": 0.9204, + "step": 4551 + }, + { + "epoch": 0.7631981557162318, + "grad_norm": 0.47085505723953247, + "learning_rate": 2.821288473334104e-06, + "loss": 0.8355, + "step": 4552 + }, + { + "epoch": 0.7633658178769779, + "grad_norm": 0.38516542315483093, + "learning_rate": 2.8175088901055026e-06, + "loss": 0.83, + "step": 4553 + }, + { + "epoch": 0.763533480037724, + "grad_norm": 0.5478811264038086, + "learning_rate": 2.8137314250824686e-06, + "loss": 0.8329, + "step": 4554 + }, + { + "epoch": 0.7637011421984701, + "grad_norm": 0.4037216305732727, + "learning_rate": 2.8099560793790204e-06, + "loss": 0.9106, + "step": 4555 + }, + { + "epoch": 0.7638688043592162, + "grad_norm": 0.3449244499206543, + "learning_rate": 2.806182854108561e-06, + "loss": 0.7537, + "step": 4556 + }, + { + "epoch": 0.7640364665199623, + "grad_norm": 0.43161898851394653, + "learning_rate": 2.8024117503838543e-06, + "loss": 0.8963, + "step": 4557 + }, + { + "epoch": 0.7642041286807084, + "grad_norm": 0.3741050064563751, + "learning_rate": 2.798642769317055e-06, + "loss": 0.7326, + "step": 4558 + }, + { + "epoch": 0.7643717908414545, + "grad_norm": 0.4074038565158844, + "learning_rate": 2.794875912019669e-06, + "loss": 0.9326, + "step": 4559 + }, + { + "epoch": 0.7645394530022006, + "grad_norm": 0.5346018671989441, + "learning_rate": 2.791111179602597e-06, + "loss": 0.8441, + "step": 4560 + }, + { + "epoch": 0.7647071151629466, + "grad_norm": 0.2614295780658722, + "learning_rate": 2.7873485731761074e-06, + "loss": 0.7906, + "step": 4561 + }, + { + "epoch": 0.7648747773236928, + "grad_norm": 0.3567403256893158, + "learning_rate": 2.783588093849834e-06, + "loss": 0.8092, + "step": 4562 + }, + { + "epoch": 0.7650424394844388, + "grad_norm": 0.46598732471466064, + "learning_rate": 2.7798297427327938e-06, + "loss": 0.9027, + "step": 4563 + }, + { + "epoch": 0.765210101645185, + "grad_norm": 0.7750427722930908, + "learning_rate": 2.776073520933369e-06, + "loss": 0.8619, + "step": 4564 + }, + { + "epoch": 0.765377763805931, + "grad_norm": 0.3993176817893982, + "learning_rate": 2.7723194295593113e-06, + "loss": 0.854, + "step": 4565 + }, + { + "epoch": 0.7655454259666772, + "grad_norm": 0.40991339087486267, + "learning_rate": 2.7685674697177568e-06, + "loss": 0.9568, + "step": 4566 + }, + { + "epoch": 0.7657130881274232, + "grad_norm": 0.4112975001335144, + "learning_rate": 2.7648176425151972e-06, + "loss": 0.8523, + "step": 4567 + }, + { + "epoch": 0.7658807502881694, + "grad_norm": 0.35256874561309814, + "learning_rate": 2.76106994905751e-06, + "loss": 0.8159, + "step": 4568 + }, + { + "epoch": 0.7660484124489154, + "grad_norm": 0.3873903155326843, + "learning_rate": 2.7573243904499338e-06, + "loss": 0.8696, + "step": 4569 + }, + { + "epoch": 0.7662160746096616, + "grad_norm": 0.6696532368659973, + "learning_rate": 2.753580967797076e-06, + "loss": 0.8513, + "step": 4570 + }, + { + "epoch": 0.7663837367704076, + "grad_norm": 0.6126317977905273, + "learning_rate": 2.7498396822029216e-06, + "loss": 0.7612, + "step": 4571 + }, + { + "epoch": 0.7665513989311538, + "grad_norm": 0.5345373749732971, + "learning_rate": 2.746100534770829e-06, + "loss": 0.8084, + "step": 4572 + }, + { + "epoch": 0.7667190610918998, + "grad_norm": 0.3670309782028198, + "learning_rate": 2.7423635266035053e-06, + "loss": 0.7996, + "step": 4573 + }, + { + "epoch": 0.766886723252646, + "grad_norm": 0.5064563155174255, + "learning_rate": 2.738628658803052e-06, + "loss": 0.8099, + "step": 4574 + }, + { + "epoch": 0.767054385413392, + "grad_norm": 0.44875088334083557, + "learning_rate": 2.734895932470919e-06, + "loss": 0.8375, + "step": 4575 + }, + { + "epoch": 0.7672220475741381, + "grad_norm": 0.4273213744163513, + "learning_rate": 2.7311653487079394e-06, + "loss": 0.8003, + "step": 4576 + }, + { + "epoch": 0.7673897097348842, + "grad_norm": 0.42424747347831726, + "learning_rate": 2.7274369086143127e-06, + "loss": 0.8833, + "step": 4577 + }, + { + "epoch": 0.7675573718956303, + "grad_norm": 0.3960532546043396, + "learning_rate": 2.7237106132895897e-06, + "loss": 0.8208, + "step": 4578 + }, + { + "epoch": 0.7677250340563764, + "grad_norm": 0.38370004296302795, + "learning_rate": 2.719986463832708e-06, + "loss": 0.8362, + "step": 4579 + }, + { + "epoch": 0.7678926962171225, + "grad_norm": 0.478929728269577, + "learning_rate": 2.7162644613419687e-06, + "loss": 0.8642, + "step": 4580 + }, + { + "epoch": 0.7680603583778686, + "grad_norm": 0.806081235408783, + "learning_rate": 2.712544606915032e-06, + "loss": 0.8639, + "step": 4581 + }, + { + "epoch": 0.7682280205386147, + "grad_norm": 0.8366953134536743, + "learning_rate": 2.708826901648928e-06, + "loss": 0.8082, + "step": 4582 + }, + { + "epoch": 0.7683956826993608, + "grad_norm": 0.6663612723350525, + "learning_rate": 2.705111346640058e-06, + "loss": 0.8111, + "step": 4583 + }, + { + "epoch": 0.7685633448601069, + "grad_norm": 0.39448583126068115, + "learning_rate": 2.7013979429841807e-06, + "loss": 0.8399, + "step": 4584 + }, + { + "epoch": 0.768731007020853, + "grad_norm": 0.40021032094955444, + "learning_rate": 2.69768669177643e-06, + "loss": 0.7316, + "step": 4585 + }, + { + "epoch": 0.7688986691815991, + "grad_norm": 0.378376841545105, + "learning_rate": 2.6939775941112965e-06, + "loss": 0.7462, + "step": 4586 + }, + { + "epoch": 0.7690663313423451, + "grad_norm": 0.5967662334442139, + "learning_rate": 2.690270651082638e-06, + "loss": 0.8688, + "step": 4587 + }, + { + "epoch": 0.7692339935030913, + "grad_norm": 0.47656574845314026, + "learning_rate": 2.686565863783681e-06, + "loss": 0.8774, + "step": 4588 + }, + { + "epoch": 0.7694016556638373, + "grad_norm": 0.506759762763977, + "learning_rate": 2.68286323330701e-06, + "loss": 0.934, + "step": 4589 + }, + { + "epoch": 0.7695693178245835, + "grad_norm": 0.42268508672714233, + "learning_rate": 2.67916276074458e-06, + "loss": 0.8518, + "step": 4590 + }, + { + "epoch": 0.7697369799853295, + "grad_norm": 0.7668575048446655, + "learning_rate": 2.6754644471877054e-06, + "loss": 0.8149, + "step": 4591 + }, + { + "epoch": 0.7699046421460757, + "grad_norm": 0.6320164799690247, + "learning_rate": 2.6717682937270605e-06, + "loss": 0.8799, + "step": 4592 + }, + { + "epoch": 0.7700723043068217, + "grad_norm": 0.7159294486045837, + "learning_rate": 2.668074301452693e-06, + "loss": 0.8879, + "step": 4593 + }, + { + "epoch": 0.7702399664675679, + "grad_norm": 0.45723140239715576, + "learning_rate": 2.6643824714540002e-06, + "loss": 0.8416, + "step": 4594 + }, + { + "epoch": 0.7704076286283139, + "grad_norm": 0.3924510180950165, + "learning_rate": 2.6606928048197555e-06, + "loss": 0.8389, + "step": 4595 + }, + { + "epoch": 0.7705752907890601, + "grad_norm": 0.4730377197265625, + "learning_rate": 2.657005302638085e-06, + "loss": 0.8408, + "step": 4596 + }, + { + "epoch": 0.7707429529498061, + "grad_norm": 0.4658554196357727, + "learning_rate": 2.6533199659964737e-06, + "loss": 0.8246, + "step": 4597 + }, + { + "epoch": 0.7709106151105523, + "grad_norm": 0.7776197195053101, + "learning_rate": 2.649636795981777e-06, + "loss": 0.8719, + "step": 4598 + }, + { + "epoch": 0.7710782772712983, + "grad_norm": 0.5712535381317139, + "learning_rate": 2.6459557936802104e-06, + "loss": 0.8904, + "step": 4599 + }, + { + "epoch": 0.7712459394320444, + "grad_norm": 0.665768027305603, + "learning_rate": 2.6422769601773437e-06, + "loss": 0.7471, + "step": 4600 + }, + { + "epoch": 0.7714136015927905, + "grad_norm": 0.49118784070014954, + "learning_rate": 2.638600296558108e-06, + "loss": 0.8099, + "step": 4601 + }, + { + "epoch": 0.7715812637535366, + "grad_norm": 0.4034315347671509, + "learning_rate": 2.6349258039068026e-06, + "loss": 0.7578, + "step": 4602 + }, + { + "epoch": 0.7717489259142827, + "grad_norm": 0.6553141474723816, + "learning_rate": 2.6312534833070745e-06, + "loss": 0.8429, + "step": 4603 + }, + { + "epoch": 0.7719165880750288, + "grad_norm": 0.4088524878025055, + "learning_rate": 2.6275833358419447e-06, + "loss": 0.8106, + "step": 4604 + }, + { + "epoch": 0.7720842502357749, + "grad_norm": 0.533658504486084, + "learning_rate": 2.6239153625937786e-06, + "loss": 0.8721, + "step": 4605 + }, + { + "epoch": 0.772251912396521, + "grad_norm": 0.46169909834861755, + "learning_rate": 2.620249564644307e-06, + "loss": 0.8185, + "step": 4606 + }, + { + "epoch": 0.7724195745572671, + "grad_norm": 0.5614885687828064, + "learning_rate": 2.616585943074623e-06, + "loss": 0.8561, + "step": 4607 + }, + { + "epoch": 0.7725872367180132, + "grad_norm": 0.36912161111831665, + "learning_rate": 2.612924498965169e-06, + "loss": 0.9491, + "step": 4608 + }, + { + "epoch": 0.7727548988787593, + "grad_norm": 0.7435131072998047, + "learning_rate": 2.609265233395757e-06, + "loss": 0.778, + "step": 4609 + }, + { + "epoch": 0.7729225610395054, + "grad_norm": 0.4584323465824127, + "learning_rate": 2.605608147445544e-06, + "loss": 0.8414, + "step": 4610 + }, + { + "epoch": 0.7730902232002514, + "grad_norm": 0.5063695311546326, + "learning_rate": 2.6019532421930505e-06, + "loss": 0.747, + "step": 4611 + }, + { + "epoch": 0.7732578853609976, + "grad_norm": 0.40801185369491577, + "learning_rate": 2.598300518716156e-06, + "loss": 0.9518, + "step": 4612 + }, + { + "epoch": 0.7734255475217436, + "grad_norm": 0.4017718732357025, + "learning_rate": 2.594649978092093e-06, + "loss": 0.8631, + "step": 4613 + }, + { + "epoch": 0.7735932096824898, + "grad_norm": 0.37917450070381165, + "learning_rate": 2.5910016213974475e-06, + "loss": 0.8841, + "step": 4614 + }, + { + "epoch": 0.7737608718432358, + "grad_norm": 0.5232954621315002, + "learning_rate": 2.587355449708171e-06, + "loss": 0.7955, + "step": 4615 + }, + { + "epoch": 0.773928534003982, + "grad_norm": 0.45333942770957947, + "learning_rate": 2.58371146409956e-06, + "loss": 0.7909, + "step": 4616 + }, + { + "epoch": 0.774096196164728, + "grad_norm": 0.34033146500587463, + "learning_rate": 2.580069665646271e-06, + "loss": 0.7876, + "step": 4617 + }, + { + "epoch": 0.7742638583254742, + "grad_norm": 0.40777915716171265, + "learning_rate": 2.576430055422324e-06, + "loss": 0.934, + "step": 4618 + }, + { + "epoch": 0.7744315204862202, + "grad_norm": 0.3873259425163269, + "learning_rate": 2.5727926345010724e-06, + "loss": 0.8527, + "step": 4619 + }, + { + "epoch": 0.7745991826469664, + "grad_norm": 0.4247393012046814, + "learning_rate": 2.5691574039552423e-06, + "loss": 0.7864, + "step": 4620 + }, + { + "epoch": 0.7747668448077124, + "grad_norm": 0.9930641055107117, + "learning_rate": 2.5655243648569116e-06, + "loss": 0.7773, + "step": 4621 + }, + { + "epoch": 0.7749345069684586, + "grad_norm": 0.48081302642822266, + "learning_rate": 2.561893518277506e-06, + "loss": 0.8871, + "step": 4622 + }, + { + "epoch": 0.7751021691292046, + "grad_norm": 1.0073636770248413, + "learning_rate": 2.558264865287806e-06, + "loss": 0.9174, + "step": 4623 + }, + { + "epoch": 0.7752698312899508, + "grad_norm": 0.6616734266281128, + "learning_rate": 2.5546384069579446e-06, + "loss": 0.8713, + "step": 4624 + }, + { + "epoch": 0.7754374934506968, + "grad_norm": 0.43992239236831665, + "learning_rate": 2.55101414435741e-06, + "loss": 0.86, + "step": 4625 + }, + { + "epoch": 0.7756051556114429, + "grad_norm": 0.40123677253723145, + "learning_rate": 2.547392078555048e-06, + "loss": 0.8038, + "step": 4626 + }, + { + "epoch": 0.7757728177721891, + "grad_norm": 0.4202883243560791, + "learning_rate": 2.5437722106190454e-06, + "loss": 0.7753, + "step": 4627 + }, + { + "epoch": 0.7759404799329351, + "grad_norm": 0.6338961720466614, + "learning_rate": 2.540154541616945e-06, + "loss": 0.8622, + "step": 4628 + }, + { + "epoch": 0.7761081420936813, + "grad_norm": 0.7529990077018738, + "learning_rate": 2.5365390726156456e-06, + "loss": 0.9001, + "step": 4629 + }, + { + "epoch": 0.7762758042544273, + "grad_norm": 0.517359733581543, + "learning_rate": 2.532925804681391e-06, + "loss": 0.849, + "step": 4630 + }, + { + "epoch": 0.7764434664151735, + "grad_norm": 0.4139972925186157, + "learning_rate": 2.5293147388797813e-06, + "loss": 0.787, + "step": 4631 + }, + { + "epoch": 0.7766111285759195, + "grad_norm": 0.40291252732276917, + "learning_rate": 2.5257058762757647e-06, + "loss": 0.782, + "step": 4632 + }, + { + "epoch": 0.7767787907366657, + "grad_norm": 0.4477655589580536, + "learning_rate": 2.5220992179336344e-06, + "loss": 0.8255, + "step": 4633 + }, + { + "epoch": 0.7769464528974117, + "grad_norm": 0.5491387248039246, + "learning_rate": 2.5184947649170455e-06, + "loss": 0.7886, + "step": 4634 + }, + { + "epoch": 0.7771141150581579, + "grad_norm": 0.40593987703323364, + "learning_rate": 2.514892518288988e-06, + "loss": 0.8754, + "step": 4635 + }, + { + "epoch": 0.7772817772189039, + "grad_norm": 0.5853688716888428, + "learning_rate": 2.511292479111819e-06, + "loss": 0.8264, + "step": 4636 + }, + { + "epoch": 0.77744943937965, + "grad_norm": 0.3623380959033966, + "learning_rate": 2.5076946484472277e-06, + "loss": 0.8025, + "step": 4637 + }, + { + "epoch": 0.7776171015403961, + "grad_norm": 0.37506017088890076, + "learning_rate": 2.5040990273562572e-06, + "loss": 0.7539, + "step": 4638 + }, + { + "epoch": 0.7777847637011422, + "grad_norm": 0.36650562286376953, + "learning_rate": 2.500505616899307e-06, + "loss": 0.8193, + "step": 4639 + }, + { + "epoch": 0.7779524258618883, + "grad_norm": 0.5781775116920471, + "learning_rate": 2.496914418136113e-06, + "loss": 0.9064, + "step": 4640 + }, + { + "epoch": 0.7781200880226344, + "grad_norm": 0.46714088320732117, + "learning_rate": 2.4933254321257694e-06, + "loss": 0.8181, + "step": 4641 + }, + { + "epoch": 0.7782877501833805, + "grad_norm": 0.5534514784812927, + "learning_rate": 2.489738659926708e-06, + "loss": 0.9225, + "step": 4642 + }, + { + "epoch": 0.7784554123441266, + "grad_norm": 0.4399142861366272, + "learning_rate": 2.486154102596713e-06, + "loss": 0.8527, + "step": 4643 + }, + { + "epoch": 0.7786230745048727, + "grad_norm": 0.4053286910057068, + "learning_rate": 2.4825717611929144e-06, + "loss": 0.8012, + "step": 4644 + }, + { + "epoch": 0.7787907366656188, + "grad_norm": 0.5702162981033325, + "learning_rate": 2.478991636771797e-06, + "loss": 0.8388, + "step": 4645 + }, + { + "epoch": 0.7789583988263649, + "grad_norm": 0.45069995522499084, + "learning_rate": 2.475413730389171e-06, + "loss": 0.9109, + "step": 4646 + }, + { + "epoch": 0.779126060987111, + "grad_norm": 0.4615761935710907, + "learning_rate": 2.471838043100211e-06, + "loss": 0.8515, + "step": 4647 + }, + { + "epoch": 0.779293723147857, + "grad_norm": 0.44282227754592896, + "learning_rate": 2.468264575959436e-06, + "loss": 0.9149, + "step": 4648 + }, + { + "epoch": 0.7794613853086032, + "grad_norm": 0.45196306705474854, + "learning_rate": 2.4646933300206977e-06, + "loss": 0.8429, + "step": 4649 + }, + { + "epoch": 0.7796290474693492, + "grad_norm": 0.39488324522972107, + "learning_rate": 2.461124306337207e-06, + "loss": 0.8723, + "step": 4650 + }, + { + "epoch": 0.7797967096300954, + "grad_norm": 0.5315574407577515, + "learning_rate": 2.457557505961512e-06, + "loss": 0.9138, + "step": 4651 + }, + { + "epoch": 0.7799643717908414, + "grad_norm": 0.5758631229400635, + "learning_rate": 2.4539929299455024e-06, + "loss": 0.8382, + "step": 4652 + }, + { + "epoch": 0.7801320339515876, + "grad_norm": 0.6532036662101746, + "learning_rate": 2.4504305793404204e-06, + "loss": 0.8348, + "step": 4653 + }, + { + "epoch": 0.7802996961123336, + "grad_norm": 0.4632951021194458, + "learning_rate": 2.446870455196847e-06, + "loss": 0.9538, + "step": 4654 + }, + { + "epoch": 0.7804673582730798, + "grad_norm": 0.5184860825538635, + "learning_rate": 2.443312558564701e-06, + "loss": 0.8301, + "step": 4655 + }, + { + "epoch": 0.7806350204338258, + "grad_norm": 0.42855778336524963, + "learning_rate": 2.439756890493258e-06, + "loss": 0.9034, + "step": 4656 + }, + { + "epoch": 0.780802682594572, + "grad_norm": 0.6073377132415771, + "learning_rate": 2.4362034520311216e-06, + "loss": 0.8989, + "step": 4657 + }, + { + "epoch": 0.780970344755318, + "grad_norm": 0.5297231078147888, + "learning_rate": 2.4326522442262522e-06, + "loss": 0.9485, + "step": 4658 + }, + { + "epoch": 0.7811380069160642, + "grad_norm": 0.45657971501350403, + "learning_rate": 2.429103268125941e-06, + "loss": 0.8522, + "step": 4659 + }, + { + "epoch": 0.7813056690768102, + "grad_norm": 0.3763165771961212, + "learning_rate": 2.425556524776821e-06, + "loss": 0.8281, + "step": 4660 + }, + { + "epoch": 0.7814733312375564, + "grad_norm": 0.35260340571403503, + "learning_rate": 2.4220120152248783e-06, + "loss": 0.7833, + "step": 4661 + }, + { + "epoch": 0.7816409933983024, + "grad_norm": 0.8353078961372375, + "learning_rate": 2.418469740515427e-06, + "loss": 0.8843, + "step": 4662 + }, + { + "epoch": 0.7818086555590485, + "grad_norm": 0.4255785644054413, + "learning_rate": 2.4149297016931317e-06, + "loss": 0.8756, + "step": 4663 + }, + { + "epoch": 0.7819763177197946, + "grad_norm": 0.446634978055954, + "learning_rate": 2.4113918998019927e-06, + "loss": 0.8451, + "step": 4664 + }, + { + "epoch": 0.7821439798805407, + "grad_norm": 1.0861172676086426, + "learning_rate": 2.4078563358853467e-06, + "loss": 0.879, + "step": 4665 + }, + { + "epoch": 0.7823116420412868, + "grad_norm": 0.6283324360847473, + "learning_rate": 2.4043230109858805e-06, + "loss": 0.8991, + "step": 4666 + }, + { + "epoch": 0.7824793042020329, + "grad_norm": 0.3723627030849457, + "learning_rate": 2.400791926145617e-06, + "loss": 0.8549, + "step": 4667 + }, + { + "epoch": 0.782646966362779, + "grad_norm": 0.37841206789016724, + "learning_rate": 2.397263082405914e-06, + "loss": 0.74, + "step": 4668 + }, + { + "epoch": 0.7828146285235251, + "grad_norm": 0.41133737564086914, + "learning_rate": 2.3937364808074683e-06, + "loss": 0.8694, + "step": 4669 + }, + { + "epoch": 0.7829822906842712, + "grad_norm": 0.4194008409976959, + "learning_rate": 2.390212122390323e-06, + "loss": 0.8062, + "step": 4670 + }, + { + "epoch": 0.7831499528450173, + "grad_norm": 0.5233654379844666, + "learning_rate": 2.3866900081938517e-06, + "loss": 0.7899, + "step": 4671 + }, + { + "epoch": 0.7833176150057634, + "grad_norm": 0.48317602276802063, + "learning_rate": 2.383170139256774e-06, + "loss": 0.8615, + "step": 4672 + }, + { + "epoch": 0.7834852771665095, + "grad_norm": 0.7062588334083557, + "learning_rate": 2.37965251661714e-06, + "loss": 0.817, + "step": 4673 + }, + { + "epoch": 0.7836529393272555, + "grad_norm": 0.42143264412879944, + "learning_rate": 2.376137141312338e-06, + "loss": 0.7995, + "step": 4674 + }, + { + "epoch": 0.7838206014880017, + "grad_norm": 0.42928212881088257, + "learning_rate": 2.372624014379099e-06, + "loss": 0.8712, + "step": 4675 + }, + { + "epoch": 0.7839882636487477, + "grad_norm": 0.3896276652812958, + "learning_rate": 2.3691131368534858e-06, + "loss": 0.8254, + "step": 4676 + }, + { + "epoch": 0.7841559258094939, + "grad_norm": 0.4230552911758423, + "learning_rate": 2.365604509770901e-06, + "loss": 0.7947, + "step": 4677 + }, + { + "epoch": 0.7843235879702399, + "grad_norm": 0.41597506403923035, + "learning_rate": 2.3620981341660833e-06, + "loss": 0.8989, + "step": 4678 + }, + { + "epoch": 0.7844912501309861, + "grad_norm": 0.680502712726593, + "learning_rate": 2.3585940110731008e-06, + "loss": 0.8588, + "step": 4679 + }, + { + "epoch": 0.7846589122917321, + "grad_norm": 0.4649544954299927, + "learning_rate": 2.3550921415253704e-06, + "loss": 0.9314, + "step": 4680 + }, + { + "epoch": 0.7848265744524783, + "grad_norm": 0.4067036807537079, + "learning_rate": 2.3515925265556293e-06, + "loss": 0.7538, + "step": 4681 + }, + { + "epoch": 0.7849942366132243, + "grad_norm": 0.3940337002277374, + "learning_rate": 2.348095167195964e-06, + "loss": 0.83, + "step": 4682 + }, + { + "epoch": 0.7851618987739705, + "grad_norm": 0.502957820892334, + "learning_rate": 2.3446000644777856e-06, + "loss": 0.8394, + "step": 4683 + }, + { + "epoch": 0.7853295609347165, + "grad_norm": 0.5205345749855042, + "learning_rate": 2.341107219431842e-06, + "loss": 0.8014, + "step": 4684 + }, + { + "epoch": 0.7854972230954627, + "grad_norm": 0.4484470784664154, + "learning_rate": 2.337616633088219e-06, + "loss": 0.8719, + "step": 4685 + }, + { + "epoch": 0.7856648852562087, + "grad_norm": 0.4280824661254883, + "learning_rate": 2.3341283064763343e-06, + "loss": 0.8764, + "step": 4686 + }, + { + "epoch": 0.7858325474169549, + "grad_norm": 0.4743936061859131, + "learning_rate": 2.330642240624933e-06, + "loss": 0.8052, + "step": 4687 + }, + { + "epoch": 0.7860002095777009, + "grad_norm": 0.5149183869361877, + "learning_rate": 2.327158436562107e-06, + "loss": 0.826, + "step": 4688 + }, + { + "epoch": 0.786167871738447, + "grad_norm": 0.42814549803733826, + "learning_rate": 2.3236768953152665e-06, + "loss": 0.8015, + "step": 4689 + }, + { + "epoch": 0.7863355338991931, + "grad_norm": 0.49067017436027527, + "learning_rate": 2.320197617911163e-06, + "loss": 0.7903, + "step": 4690 + }, + { + "epoch": 0.7865031960599392, + "grad_norm": 0.3798750936985016, + "learning_rate": 2.316720605375886e-06, + "loss": 0.8719, + "step": 4691 + }, + { + "epoch": 0.7866708582206853, + "grad_norm": 1.520694613456726, + "learning_rate": 2.313245858734838e-06, + "loss": 0.8271, + "step": 4692 + }, + { + "epoch": 0.7868385203814314, + "grad_norm": 0.452315092086792, + "learning_rate": 2.30977337901277e-06, + "loss": 0.8206, + "step": 4693 + }, + { + "epoch": 0.7870061825421775, + "grad_norm": 0.3753838837146759, + "learning_rate": 2.306303167233761e-06, + "loss": 0.8666, + "step": 4694 + }, + { + "epoch": 0.7871738447029236, + "grad_norm": 0.4556758403778076, + "learning_rate": 2.3028352244212173e-06, + "loss": 0.8371, + "step": 4695 + }, + { + "epoch": 0.7873415068636697, + "grad_norm": 0.34695783257484436, + "learning_rate": 2.2993695515978767e-06, + "loss": 0.8478, + "step": 4696 + }, + { + "epoch": 0.7875091690244158, + "grad_norm": 0.3780613839626312, + "learning_rate": 2.2959061497858125e-06, + "loss": 0.7419, + "step": 4697 + }, + { + "epoch": 0.7876768311851619, + "grad_norm": 0.413053959608078, + "learning_rate": 2.292445020006422e-06, + "loss": 0.7941, + "step": 4698 + }, + { + "epoch": 0.787844493345908, + "grad_norm": 0.5531803965568542, + "learning_rate": 2.2889861632804377e-06, + "loss": 0.8418, + "step": 4699 + }, + { + "epoch": 0.788012155506654, + "grad_norm": 0.4788338541984558, + "learning_rate": 2.2855295806279188e-06, + "loss": 0.9246, + "step": 4700 + }, + { + "epoch": 0.7881798176674002, + "grad_norm": 0.42819738388061523, + "learning_rate": 2.2820752730682494e-06, + "loss": 0.8493, + "step": 4701 + }, + { + "epoch": 0.7883474798281462, + "grad_norm": 0.504607081413269, + "learning_rate": 2.2786232416201548e-06, + "loss": 0.8319, + "step": 4702 + }, + { + "epoch": 0.7885151419888924, + "grad_norm": 0.3961004912853241, + "learning_rate": 2.275173487301676e-06, + "loss": 0.7408, + "step": 4703 + }, + { + "epoch": 0.7886828041496384, + "grad_norm": 0.42878851294517517, + "learning_rate": 2.2717260111301943e-06, + "loss": 0.7563, + "step": 4704 + }, + { + "epoch": 0.7888504663103846, + "grad_norm": 0.5974430441856384, + "learning_rate": 2.268280814122409e-06, + "loss": 0.8772, + "step": 4705 + }, + { + "epoch": 0.7890181284711306, + "grad_norm": 0.3592509925365448, + "learning_rate": 2.2648378972943495e-06, + "loss": 0.8229, + "step": 4706 + }, + { + "epoch": 0.7891857906318768, + "grad_norm": 0.9069346189498901, + "learning_rate": 2.26139726166138e-06, + "loss": 0.9157, + "step": 4707 + }, + { + "epoch": 0.7893534527926228, + "grad_norm": 0.3409862220287323, + "learning_rate": 2.2579589082381813e-06, + "loss": 0.8402, + "step": 4708 + }, + { + "epoch": 0.789521114953369, + "grad_norm": 0.9539614915847778, + "learning_rate": 2.2545228380387706e-06, + "loss": 0.8716, + "step": 4709 + }, + { + "epoch": 0.789688777114115, + "grad_norm": 0.9350457787513733, + "learning_rate": 2.251089052076487e-06, + "loss": 0.822, + "step": 4710 + }, + { + "epoch": 0.7898564392748612, + "grad_norm": 0.6432152390480042, + "learning_rate": 2.247657551363992e-06, + "loss": 0.8366, + "step": 4711 + }, + { + "epoch": 0.7900241014356073, + "grad_norm": 0.6478852033615112, + "learning_rate": 2.24422833691328e-06, + "loss": 0.8435, + "step": 4712 + }, + { + "epoch": 0.7901917635963533, + "grad_norm": 0.4530707597732544, + "learning_rate": 2.2408014097356732e-06, + "loss": 0.8305, + "step": 4713 + }, + { + "epoch": 0.7903594257570995, + "grad_norm": 0.7520490288734436, + "learning_rate": 2.2373767708418116e-06, + "loss": 0.9746, + "step": 4714 + }, + { + "epoch": 0.7905270879178455, + "grad_norm": 0.40117794275283813, + "learning_rate": 2.233954421241661e-06, + "loss": 0.8815, + "step": 4715 + }, + { + "epoch": 0.7906947500785917, + "grad_norm": 0.5994216799736023, + "learning_rate": 2.2305343619445185e-06, + "loss": 0.9301, + "step": 4716 + }, + { + "epoch": 0.7908624122393377, + "grad_norm": 0.4644392430782318, + "learning_rate": 2.227116593958999e-06, + "loss": 0.7981, + "step": 4717 + }, + { + "epoch": 0.7910300744000839, + "grad_norm": 0.4093725085258484, + "learning_rate": 2.223701118293048e-06, + "loss": 0.8982, + "step": 4718 + }, + { + "epoch": 0.7911977365608299, + "grad_norm": 0.8823078870773315, + "learning_rate": 2.22028793595393e-06, + "loss": 0.9033, + "step": 4719 + }, + { + "epoch": 0.7913653987215761, + "grad_norm": 0.3445841372013092, + "learning_rate": 2.2168770479482315e-06, + "loss": 0.8072, + "step": 4720 + }, + { + "epoch": 0.7915330608823221, + "grad_norm": 0.3272508680820465, + "learning_rate": 2.213468455281872e-06, + "loss": 0.7164, + "step": 4721 + }, + { + "epoch": 0.7917007230430683, + "grad_norm": 0.411796510219574, + "learning_rate": 2.2100621589600813e-06, + "loss": 0.948, + "step": 4722 + }, + { + "epoch": 0.7918683852038143, + "grad_norm": 0.45838311314582825, + "learning_rate": 2.206658159987424e-06, + "loss": 0.8372, + "step": 4723 + }, + { + "epoch": 0.7920360473645605, + "grad_norm": 0.7174457907676697, + "learning_rate": 2.2032564593677773e-06, + "loss": 0.7418, + "step": 4724 + }, + { + "epoch": 0.7922037095253065, + "grad_norm": 0.8203630447387695, + "learning_rate": 2.1998570581043453e-06, + "loss": 0.8728, + "step": 4725 + }, + { + "epoch": 0.7923713716860527, + "grad_norm": 0.6109965443611145, + "learning_rate": 2.196459957199657e-06, + "loss": 0.7454, + "step": 4726 + }, + { + "epoch": 0.7925390338467987, + "grad_norm": 0.6313849687576294, + "learning_rate": 2.193065157655556e-06, + "loss": 0.8987, + "step": 4727 + }, + { + "epoch": 0.7927066960075448, + "grad_norm": 0.7500846982002258, + "learning_rate": 2.189672660473211e-06, + "loss": 0.8722, + "step": 4728 + }, + { + "epoch": 0.7928743581682909, + "grad_norm": 0.4414680302143097, + "learning_rate": 2.1862824666531135e-06, + "loss": 0.7453, + "step": 4729 + }, + { + "epoch": 0.793042020329037, + "grad_norm": 0.4681912958621979, + "learning_rate": 2.1828945771950704e-06, + "loss": 0.7655, + "step": 4730 + }, + { + "epoch": 0.7932096824897831, + "grad_norm": 0.4023488163948059, + "learning_rate": 2.179508993098214e-06, + "loss": 0.861, + "step": 4731 + }, + { + "epoch": 0.7933773446505292, + "grad_norm": 0.47425827383995056, + "learning_rate": 2.1761257153610004e-06, + "loss": 0.9054, + "step": 4732 + }, + { + "epoch": 0.7935450068112753, + "grad_norm": 0.3306763470172882, + "learning_rate": 2.17274474498119e-06, + "loss": 0.8637, + "step": 4733 + }, + { + "epoch": 0.7937126689720214, + "grad_norm": 0.3470577001571655, + "learning_rate": 2.1693660829558804e-06, + "loss": 0.819, + "step": 4734 + }, + { + "epoch": 0.7938803311327675, + "grad_norm": 0.4094432294368744, + "learning_rate": 2.165989730281475e-06, + "loss": 0.7996, + "step": 4735 + }, + { + "epoch": 0.7940479932935136, + "grad_norm": 0.4102506935596466, + "learning_rate": 2.1626156879537084e-06, + "loss": 0.7098, + "step": 4736 + }, + { + "epoch": 0.7942156554542597, + "grad_norm": 0.5132640600204468, + "learning_rate": 2.159243956967624e-06, + "loss": 0.8522, + "step": 4737 + }, + { + "epoch": 0.7943833176150058, + "grad_norm": 0.755061686038971, + "learning_rate": 2.155874538317585e-06, + "loss": 0.7889, + "step": 4738 + }, + { + "epoch": 0.7945509797757518, + "grad_norm": 0.6049810647964478, + "learning_rate": 2.152507432997277e-06, + "loss": 0.8063, + "step": 4739 + }, + { + "epoch": 0.794718641936498, + "grad_norm": 0.5153598785400391, + "learning_rate": 2.149142641999703e-06, + "loss": 0.7808, + "step": 4740 + }, + { + "epoch": 0.794886304097244, + "grad_norm": 0.3766288757324219, + "learning_rate": 2.1457801663171805e-06, + "loss": 0.827, + "step": 4741 + }, + { + "epoch": 0.7950539662579902, + "grad_norm": 0.6483139991760254, + "learning_rate": 2.1424200069413416e-06, + "loss": 0.8916, + "step": 4742 + }, + { + "epoch": 0.7952216284187362, + "grad_norm": 0.38120341300964355, + "learning_rate": 2.139062164863144e-06, + "loss": 0.768, + "step": 4743 + }, + { + "epoch": 0.7953892905794824, + "grad_norm": 0.5252415537834167, + "learning_rate": 2.135706641072851e-06, + "loss": 0.7788, + "step": 4744 + }, + { + "epoch": 0.7955569527402284, + "grad_norm": 0.3720047175884247, + "learning_rate": 2.132353436560055e-06, + "loss": 0.8189, + "step": 4745 + }, + { + "epoch": 0.7957246149009746, + "grad_norm": 0.329374760389328, + "learning_rate": 2.129002552313655e-06, + "loss": 0.7773, + "step": 4746 + }, + { + "epoch": 0.7958922770617206, + "grad_norm": 0.5027045011520386, + "learning_rate": 2.125653989321863e-06, + "loss": 0.8635, + "step": 4747 + }, + { + "epoch": 0.7960599392224668, + "grad_norm": 0.6417919993400574, + "learning_rate": 2.12230774857222e-06, + "loss": 0.8453, + "step": 4748 + }, + { + "epoch": 0.7962276013832128, + "grad_norm": 0.4022006690502167, + "learning_rate": 2.118963831051567e-06, + "loss": 0.8061, + "step": 4749 + }, + { + "epoch": 0.796395263543959, + "grad_norm": 0.4296365976333618, + "learning_rate": 2.1156222377460723e-06, + "loss": 0.8482, + "step": 4750 + }, + { + "epoch": 0.796562925704705, + "grad_norm": 0.28435319662094116, + "learning_rate": 2.1122829696412108e-06, + "loss": 0.8003, + "step": 4751 + }, + { + "epoch": 0.7967305878654511, + "grad_norm": 0.4156350791454315, + "learning_rate": 2.10894602772177e-06, + "loss": 0.7549, + "step": 4752 + }, + { + "epoch": 0.7968982500261972, + "grad_norm": 0.49609625339508057, + "learning_rate": 2.105611412971863e-06, + "loss": 0.8254, + "step": 4753 + }, + { + "epoch": 0.7970659121869433, + "grad_norm": 0.4338054358959198, + "learning_rate": 2.102279126374902e-06, + "loss": 0.8666, + "step": 4754 + }, + { + "epoch": 0.7972335743476894, + "grad_norm": 0.4368712604045868, + "learning_rate": 2.098949168913624e-06, + "loss": 0.8679, + "step": 4755 + }, + { + "epoch": 0.7974012365084355, + "grad_norm": 0.4860408306121826, + "learning_rate": 2.095621541570075e-06, + "loss": 0.8792, + "step": 4756 + }, + { + "epoch": 0.7975688986691816, + "grad_norm": 0.45234760642051697, + "learning_rate": 2.092296245325609e-06, + "loss": 0.7176, + "step": 4757 + }, + { + "epoch": 0.7977365608299277, + "grad_norm": 0.487379789352417, + "learning_rate": 2.0889732811608996e-06, + "loss": 0.8306, + "step": 4758 + }, + { + "epoch": 0.7979042229906738, + "grad_norm": 0.3871273100376129, + "learning_rate": 2.085652650055935e-06, + "loss": 0.7539, + "step": 4759 + }, + { + "epoch": 0.7980718851514199, + "grad_norm": 0.47201892733573914, + "learning_rate": 2.0823343529900007e-06, + "loss": 0.78, + "step": 4760 + }, + { + "epoch": 0.798239547312166, + "grad_norm": 0.5183556079864502, + "learning_rate": 2.0790183909417096e-06, + "loss": 0.9249, + "step": 4761 + }, + { + "epoch": 0.7984072094729121, + "grad_norm": 0.4703180491924286, + "learning_rate": 2.0757047648889816e-06, + "loss": 0.8947, + "step": 4762 + }, + { + "epoch": 0.7985748716336581, + "grad_norm": 0.3326199948787689, + "learning_rate": 2.07239347580904e-06, + "loss": 0.8183, + "step": 4763 + }, + { + "epoch": 0.7987425337944043, + "grad_norm": 0.4855404496192932, + "learning_rate": 2.069084524678432e-06, + "loss": 0.7859, + "step": 4764 + }, + { + "epoch": 0.7989101959551503, + "grad_norm": 0.5343421101570129, + "learning_rate": 2.0657779124730036e-06, + "loss": 0.8775, + "step": 4765 + }, + { + "epoch": 0.7990778581158965, + "grad_norm": 0.43089139461517334, + "learning_rate": 2.062473640167915e-06, + "loss": 0.8366, + "step": 4766 + }, + { + "epoch": 0.7992455202766425, + "grad_norm": 0.3698073923587799, + "learning_rate": 2.0591717087376416e-06, + "loss": 0.8894, + "step": 4767 + }, + { + "epoch": 0.7994131824373887, + "grad_norm": 0.48625025153160095, + "learning_rate": 2.0558721191559606e-06, + "loss": 0.854, + "step": 4768 + }, + { + "epoch": 0.7995808445981347, + "grad_norm": 0.6302156448364258, + "learning_rate": 2.0525748723959595e-06, + "loss": 0.826, + "step": 4769 + }, + { + "epoch": 0.7997485067588809, + "grad_norm": 0.4241481125354767, + "learning_rate": 2.049279969430044e-06, + "loss": 0.8336, + "step": 4770 + }, + { + "epoch": 0.7999161689196269, + "grad_norm": 0.3742135167121887, + "learning_rate": 2.045987411229913e-06, + "loss": 0.784, + "step": 4771 + }, + { + "epoch": 0.8000838310803731, + "grad_norm": 0.6190921664237976, + "learning_rate": 2.0426971987665912e-06, + "loss": 0.8335, + "step": 4772 + }, + { + "epoch": 0.8002514932411191, + "grad_norm": 0.8990462422370911, + "learning_rate": 2.0394093330103992e-06, + "loss": 0.9597, + "step": 4773 + }, + { + "epoch": 0.8004191554018653, + "grad_norm": 0.4658777117729187, + "learning_rate": 2.036123814930967e-06, + "loss": 0.8922, + "step": 4774 + }, + { + "epoch": 0.8005868175626113, + "grad_norm": 0.5698621869087219, + "learning_rate": 2.0328406454972395e-06, + "loss": 0.7925, + "step": 4775 + }, + { + "epoch": 0.8007544797233574, + "grad_norm": 0.49836301803588867, + "learning_rate": 2.0295598256774598e-06, + "loss": 0.8592, + "step": 4776 + }, + { + "epoch": 0.8009221418841035, + "grad_norm": 0.6130602359771729, + "learning_rate": 2.0262813564391846e-06, + "loss": 0.8279, + "step": 4777 + }, + { + "epoch": 0.8010898040448496, + "grad_norm": 0.39174386858940125, + "learning_rate": 2.023005238749276e-06, + "loss": 0.7482, + "step": 4778 + }, + { + "epoch": 0.8012574662055957, + "grad_norm": 0.5928496718406677, + "learning_rate": 2.019731473573897e-06, + "loss": 0.8215, + "step": 4779 + }, + { + "epoch": 0.8014251283663418, + "grad_norm": 0.38793522119522095, + "learning_rate": 2.0164600618785247e-06, + "loss": 0.7668, + "step": 4780 + }, + { + "epoch": 0.8015927905270879, + "grad_norm": 0.5155482888221741, + "learning_rate": 2.013191004627941e-06, + "loss": 0.9221, + "step": 4781 + }, + { + "epoch": 0.801760452687834, + "grad_norm": 0.5527405738830566, + "learning_rate": 2.0099243027862303e-06, + "loss": 0.8017, + "step": 4782 + }, + { + "epoch": 0.8019281148485801, + "grad_norm": 0.4220729470252991, + "learning_rate": 2.0066599573167824e-06, + "loss": 0.8613, + "step": 4783 + }, + { + "epoch": 0.8020957770093262, + "grad_norm": 0.37018314003944397, + "learning_rate": 2.003397969182289e-06, + "loss": 0.8766, + "step": 4784 + }, + { + "epoch": 0.8022634391700723, + "grad_norm": 0.5372514128684998, + "learning_rate": 2.000138339344756e-06, + "loss": 0.8321, + "step": 4785 + }, + { + "epoch": 0.8024311013308184, + "grad_norm": 0.4727790951728821, + "learning_rate": 1.9968810687654893e-06, + "loss": 0.8363, + "step": 4786 + }, + { + "epoch": 0.8025987634915644, + "grad_norm": 0.3786468207836151, + "learning_rate": 1.9936261584050974e-06, + "loss": 0.8643, + "step": 4787 + }, + { + "epoch": 0.8027664256523106, + "grad_norm": 0.5220190286636353, + "learning_rate": 1.9903736092234904e-06, + "loss": 0.8608, + "step": 4788 + }, + { + "epoch": 0.8029340878130566, + "grad_norm": 0.3876339793205261, + "learning_rate": 1.9871234221798895e-06, + "loss": 0.7836, + "step": 4789 + }, + { + "epoch": 0.8031017499738028, + "grad_norm": 0.37965089082717896, + "learning_rate": 1.9838755982328108e-06, + "loss": 0.81, + "step": 4790 + }, + { + "epoch": 0.8032694121345488, + "grad_norm": 0.7638852596282959, + "learning_rate": 1.9806301383400837e-06, + "loss": 0.9099, + "step": 4791 + }, + { + "epoch": 0.803437074295295, + "grad_norm": 0.9383553862571716, + "learning_rate": 1.9773870434588306e-06, + "loss": 0.8597, + "step": 4792 + }, + { + "epoch": 0.803604736456041, + "grad_norm": 0.48865485191345215, + "learning_rate": 1.9741463145454786e-06, + "loss": 0.8666, + "step": 4793 + }, + { + "epoch": 0.8037723986167872, + "grad_norm": 0.36044055223464966, + "learning_rate": 1.970907952555764e-06, + "loss": 0.8973, + "step": 4794 + }, + { + "epoch": 0.8039400607775332, + "grad_norm": 0.4043533504009247, + "learning_rate": 1.9676719584447134e-06, + "loss": 0.8613, + "step": 4795 + }, + { + "epoch": 0.8041077229382794, + "grad_norm": 0.3273265063762665, + "learning_rate": 1.9644383331666684e-06, + "loss": 0.8955, + "step": 4796 + }, + { + "epoch": 0.8042753850990254, + "grad_norm": 0.3665297329425812, + "learning_rate": 1.961207077675261e-06, + "loss": 0.8371, + "step": 4797 + }, + { + "epoch": 0.8044430472597716, + "grad_norm": 0.385436087846756, + "learning_rate": 1.9579781929234275e-06, + "loss": 0.8152, + "step": 4798 + }, + { + "epoch": 0.8046107094205177, + "grad_norm": 0.5350459814071655, + "learning_rate": 1.9547516798634104e-06, + "loss": 0.8141, + "step": 4799 + }, + { + "epoch": 0.8047783715812638, + "grad_norm": 0.8556716442108154, + "learning_rate": 1.9515275394467446e-06, + "loss": 0.825, + "step": 4800 + }, + { + "epoch": 0.8049460337420099, + "grad_norm": 0.3931393623352051, + "learning_rate": 1.9483057726242694e-06, + "loss": 0.8643, + "step": 4801 + }, + { + "epoch": 0.8051136959027559, + "grad_norm": 0.4177151024341583, + "learning_rate": 1.945086380346126e-06, + "loss": 0.8337, + "step": 4802 + }, + { + "epoch": 0.8052813580635021, + "grad_norm": 0.4200320839881897, + "learning_rate": 1.941869363561749e-06, + "loss": 0.8256, + "step": 4803 + }, + { + "epoch": 0.8054490202242481, + "grad_norm": 0.40648573637008667, + "learning_rate": 1.9386547232198795e-06, + "loss": 0.9559, + "step": 4804 + }, + { + "epoch": 0.8056166823849943, + "grad_norm": 0.3557724356651306, + "learning_rate": 1.9354424602685586e-06, + "loss": 0.8458, + "step": 4805 + }, + { + "epoch": 0.8057843445457403, + "grad_norm": 0.4115389883518219, + "learning_rate": 1.932232575655113e-06, + "loss": 0.8754, + "step": 4806 + }, + { + "epoch": 0.8059520067064865, + "grad_norm": 0.4639323353767395, + "learning_rate": 1.9290250703261824e-06, + "loss": 0.9602, + "step": 4807 + }, + { + "epoch": 0.8061196688672325, + "grad_norm": 0.42797258496284485, + "learning_rate": 1.9258199452277037e-06, + "loss": 0.8681, + "step": 4808 + }, + { + "epoch": 0.8062873310279787, + "grad_norm": 0.49095794558525085, + "learning_rate": 1.9226172013049027e-06, + "loss": 0.8429, + "step": 4809 + }, + { + "epoch": 0.8064549931887247, + "grad_norm": 0.4635695219039917, + "learning_rate": 1.9194168395023083e-06, + "loss": 0.7287, + "step": 4810 + }, + { + "epoch": 0.8066226553494709, + "grad_norm": 0.45930927991867065, + "learning_rate": 1.91621886076375e-06, + "loss": 0.8617, + "step": 4811 + }, + { + "epoch": 0.8067903175102169, + "grad_norm": 0.4171750247478485, + "learning_rate": 1.913023266032348e-06, + "loss": 0.8432, + "step": 4812 + }, + { + "epoch": 0.806957979670963, + "grad_norm": 0.6069956421852112, + "learning_rate": 1.9098300562505266e-06, + "loss": 0.8255, + "step": 4813 + }, + { + "epoch": 0.8071256418317091, + "grad_norm": 0.44642847776412964, + "learning_rate": 1.9066392323600003e-06, + "loss": 0.7873, + "step": 4814 + }, + { + "epoch": 0.8072933039924552, + "grad_norm": 0.37293142080307007, + "learning_rate": 1.903450795301781e-06, + "loss": 0.7853, + "step": 4815 + }, + { + "epoch": 0.8074609661532013, + "grad_norm": 0.5118458271026611, + "learning_rate": 1.9002647460161826e-06, + "loss": 0.8282, + "step": 4816 + }, + { + "epoch": 0.8076286283139474, + "grad_norm": 0.4068104326725006, + "learning_rate": 1.8970810854428046e-06, + "loss": 0.8237, + "step": 4817 + }, + { + "epoch": 0.8077962904746935, + "grad_norm": 0.4325851500034332, + "learning_rate": 1.8938998145205557e-06, + "loss": 0.8656, + "step": 4818 + }, + { + "epoch": 0.8079639526354396, + "grad_norm": 0.39056482911109924, + "learning_rate": 1.8907209341876266e-06, + "loss": 0.8158, + "step": 4819 + }, + { + "epoch": 0.8081316147961857, + "grad_norm": 0.4574761986732483, + "learning_rate": 1.8875444453815073e-06, + "loss": 0.7855, + "step": 4820 + }, + { + "epoch": 0.8082992769569318, + "grad_norm": 0.776763379573822, + "learning_rate": 1.8843703490389885e-06, + "loss": 0.8233, + "step": 4821 + }, + { + "epoch": 0.8084669391176779, + "grad_norm": 0.41815385222435, + "learning_rate": 1.8811986460961461e-06, + "loss": 0.7794, + "step": 4822 + }, + { + "epoch": 0.808634601278424, + "grad_norm": 0.3735414743423462, + "learning_rate": 1.8780293374883584e-06, + "loss": 0.7637, + "step": 4823 + }, + { + "epoch": 0.80880226343917, + "grad_norm": 0.39200180768966675, + "learning_rate": 1.8748624241502934e-06, + "loss": 0.7985, + "step": 4824 + }, + { + "epoch": 0.8089699255999162, + "grad_norm": 0.4573250114917755, + "learning_rate": 1.871697907015907e-06, + "loss": 0.7891, + "step": 4825 + }, + { + "epoch": 0.8091375877606622, + "grad_norm": 0.6612563729286194, + "learning_rate": 1.8685357870184605e-06, + "loss": 0.8541, + "step": 4826 + }, + { + "epoch": 0.8093052499214084, + "grad_norm": 0.4323849081993103, + "learning_rate": 1.8653760650905028e-06, + "loss": 0.8265, + "step": 4827 + }, + { + "epoch": 0.8094729120821544, + "grad_norm": 0.4029139280319214, + "learning_rate": 1.8622187421638737e-06, + "loss": 0.8441, + "step": 4828 + }, + { + "epoch": 0.8096405742429006, + "grad_norm": 0.8146257400512695, + "learning_rate": 1.8590638191697031e-06, + "loss": 0.8966, + "step": 4829 + }, + { + "epoch": 0.8098082364036466, + "grad_norm": 0.5496284365653992, + "learning_rate": 1.855911297038422e-06, + "loss": 0.8491, + "step": 4830 + }, + { + "epoch": 0.8099758985643928, + "grad_norm": 0.5226850509643555, + "learning_rate": 1.8527611766997444e-06, + "loss": 0.7946, + "step": 4831 + }, + { + "epoch": 0.8101435607251388, + "grad_norm": 0.4379395544528961, + "learning_rate": 1.8496134590826864e-06, + "loss": 0.8962, + "step": 4832 + }, + { + "epoch": 0.810311222885885, + "grad_norm": 0.5736410617828369, + "learning_rate": 1.8464681451155387e-06, + "loss": 0.89, + "step": 4833 + }, + { + "epoch": 0.810478885046631, + "grad_norm": 0.39776611328125, + "learning_rate": 1.8433252357258991e-06, + "loss": 0.8144, + "step": 4834 + }, + { + "epoch": 0.8106465472073772, + "grad_norm": 0.4503728151321411, + "learning_rate": 1.8401847318406518e-06, + "loss": 0.8192, + "step": 4835 + }, + { + "epoch": 0.8108142093681232, + "grad_norm": 0.42749375104904175, + "learning_rate": 1.8370466343859672e-06, + "loss": 0.8494, + "step": 4836 + }, + { + "epoch": 0.8109818715288694, + "grad_norm": 0.42171910405158997, + "learning_rate": 1.8339109442873115e-06, + "loss": 0.8293, + "step": 4837 + }, + { + "epoch": 0.8111495336896154, + "grad_norm": 0.49447721242904663, + "learning_rate": 1.8307776624694383e-06, + "loss": 0.8529, + "step": 4838 + }, + { + "epoch": 0.8113171958503615, + "grad_norm": 0.4726193845272064, + "learning_rate": 1.8276467898563887e-06, + "loss": 0.8647, + "step": 4839 + }, + { + "epoch": 0.8114848580111076, + "grad_norm": 0.9029093384742737, + "learning_rate": 1.8245183273714984e-06, + "loss": 0.8707, + "step": 4840 + }, + { + "epoch": 0.8116525201718537, + "grad_norm": 0.528593897819519, + "learning_rate": 1.8213922759373903e-06, + "loss": 0.8982, + "step": 4841 + }, + { + "epoch": 0.8118201823325998, + "grad_norm": 0.48823660612106323, + "learning_rate": 1.8182686364759704e-06, + "loss": 0.8779, + "step": 4842 + }, + { + "epoch": 0.8119878444933459, + "grad_norm": 0.7718713879585266, + "learning_rate": 1.8151474099084466e-06, + "loss": 0.8608, + "step": 4843 + }, + { + "epoch": 0.812155506654092, + "grad_norm": 0.4108054041862488, + "learning_rate": 1.8120285971552987e-06, + "loss": 0.8489, + "step": 4844 + }, + { + "epoch": 0.8123231688148381, + "grad_norm": 0.5947515368461609, + "learning_rate": 1.808912199136309e-06, + "loss": 0.8154, + "step": 4845 + }, + { + "epoch": 0.8124908309755842, + "grad_norm": 0.4202979505062103, + "learning_rate": 1.8057982167705446e-06, + "loss": 0.9065, + "step": 4846 + }, + { + "epoch": 0.8126584931363303, + "grad_norm": 0.3972373902797699, + "learning_rate": 1.8026866509763497e-06, + "loss": 0.7442, + "step": 4847 + }, + { + "epoch": 0.8128261552970764, + "grad_norm": 0.3535272479057312, + "learning_rate": 1.799577502671368e-06, + "loss": 0.8012, + "step": 4848 + }, + { + "epoch": 0.8129938174578225, + "grad_norm": 0.3755205571651459, + "learning_rate": 1.7964707727725238e-06, + "loss": 0.7669, + "step": 4849 + }, + { + "epoch": 0.8131614796185686, + "grad_norm": 0.6204691529273987, + "learning_rate": 1.7933664621960334e-06, + "loss": 0.8054, + "step": 4850 + }, + { + "epoch": 0.8133291417793147, + "grad_norm": 0.45454415678977966, + "learning_rate": 1.7902645718573953e-06, + "loss": 0.7948, + "step": 4851 + }, + { + "epoch": 0.8134968039400607, + "grad_norm": 0.6720665693283081, + "learning_rate": 1.787165102671391e-06, + "loss": 0.9038, + "step": 4852 + }, + { + "epoch": 0.8136644661008069, + "grad_norm": 0.34496569633483887, + "learning_rate": 1.7840680555520961e-06, + "loss": 0.8538, + "step": 4853 + }, + { + "epoch": 0.8138321282615529, + "grad_norm": 0.47709912061691284, + "learning_rate": 1.7809734314128712e-06, + "loss": 0.7695, + "step": 4854 + }, + { + "epoch": 0.8139997904222991, + "grad_norm": 0.4746073782444, + "learning_rate": 1.777881231166355e-06, + "loss": 0.8022, + "step": 4855 + }, + { + "epoch": 0.8141674525830451, + "grad_norm": 0.5646691918373108, + "learning_rate": 1.774791455724475e-06, + "loss": 0.7788, + "step": 4856 + }, + { + "epoch": 0.8143351147437913, + "grad_norm": 0.4032534062862396, + "learning_rate": 1.7717041059984486e-06, + "loss": 0.7922, + "step": 4857 + }, + { + "epoch": 0.8145027769045373, + "grad_norm": 1.140046238899231, + "learning_rate": 1.7686191828987676e-06, + "loss": 0.8125, + "step": 4858 + }, + { + "epoch": 0.8146704390652835, + "grad_norm": 0.3570913076400757, + "learning_rate": 1.7655366873352197e-06, + "loss": 0.816, + "step": 4859 + }, + { + "epoch": 0.8148381012260295, + "grad_norm": 0.36539798974990845, + "learning_rate": 1.762456620216869e-06, + "loss": 0.7864, + "step": 4860 + }, + { + "epoch": 0.8150057633867757, + "grad_norm": 0.3946821987628937, + "learning_rate": 1.7593789824520624e-06, + "loss": 0.8419, + "step": 4861 + }, + { + "epoch": 0.8151734255475217, + "grad_norm": 0.33802446722984314, + "learning_rate": 1.756303774948439e-06, + "loss": 0.8817, + "step": 4862 + }, + { + "epoch": 0.8153410877082679, + "grad_norm": 0.41984841227531433, + "learning_rate": 1.7532309986129092e-06, + "loss": 0.7901, + "step": 4863 + }, + { + "epoch": 0.8155087498690139, + "grad_norm": 0.416354775428772, + "learning_rate": 1.7501606543516792e-06, + "loss": 0.7435, + "step": 4864 + }, + { + "epoch": 0.81567641202976, + "grad_norm": 0.4285713732242584, + "learning_rate": 1.7470927430702277e-06, + "loss": 0.873, + "step": 4865 + }, + { + "epoch": 0.8158440741905061, + "grad_norm": 0.45582979917526245, + "learning_rate": 1.7440272656733182e-06, + "loss": 0.8108, + "step": 4866 + }, + { + "epoch": 0.8160117363512522, + "grad_norm": 0.42091938853263855, + "learning_rate": 1.7409642230650037e-06, + "loss": 0.79, + "step": 4867 + }, + { + "epoch": 0.8161793985119983, + "grad_norm": 0.6437473297119141, + "learning_rate": 1.7379036161486063e-06, + "loss": 0.8313, + "step": 4868 + }, + { + "epoch": 0.8163470606727444, + "grad_norm": 0.38362085819244385, + "learning_rate": 1.7348454458267428e-06, + "loss": 0.7879, + "step": 4869 + }, + { + "epoch": 0.8165147228334905, + "grad_norm": 0.4212976396083832, + "learning_rate": 1.731789713001304e-06, + "loss": 0.8341, + "step": 4870 + }, + { + "epoch": 0.8166823849942366, + "grad_norm": 0.47702792286872864, + "learning_rate": 1.7287364185734602e-06, + "loss": 0.9573, + "step": 4871 + }, + { + "epoch": 0.8168500471549827, + "grad_norm": 1.012007474899292, + "learning_rate": 1.7256855634436675e-06, + "loss": 0.7798, + "step": 4872 + }, + { + "epoch": 0.8170177093157288, + "grad_norm": 0.432649165391922, + "learning_rate": 1.7226371485116667e-06, + "loss": 0.8771, + "step": 4873 + }, + { + "epoch": 0.8171853714764749, + "grad_norm": 0.3531188368797302, + "learning_rate": 1.7195911746764627e-06, + "loss": 0.8903, + "step": 4874 + }, + { + "epoch": 0.817353033637221, + "grad_norm": 0.39750978350639343, + "learning_rate": 1.716547642836356e-06, + "loss": 0.8644, + "step": 4875 + }, + { + "epoch": 0.817520695797967, + "grad_norm": 0.4370575249195099, + "learning_rate": 1.713506553888924e-06, + "loss": 0.8534, + "step": 4876 + }, + { + "epoch": 0.8176883579587132, + "grad_norm": 1.1015609502792358, + "learning_rate": 1.7104679087310172e-06, + "loss": 0.8118, + "step": 4877 + }, + { + "epoch": 0.8178560201194592, + "grad_norm": 0.3910379707813263, + "learning_rate": 1.7074317082587755e-06, + "loss": 0.8986, + "step": 4878 + }, + { + "epoch": 0.8180236822802054, + "grad_norm": 0.3427804708480835, + "learning_rate": 1.704397953367607e-06, + "loss": 0.8055, + "step": 4879 + }, + { + "epoch": 0.8181913444409514, + "grad_norm": 0.42192915081977844, + "learning_rate": 1.7013666449522025e-06, + "loss": 0.8696, + "step": 4880 + }, + { + "epoch": 0.8183590066016976, + "grad_norm": 0.518771767616272, + "learning_rate": 1.6983377839065373e-06, + "loss": 0.8898, + "step": 4881 + }, + { + "epoch": 0.8185266687624436, + "grad_norm": 0.3843253254890442, + "learning_rate": 1.6953113711238578e-06, + "loss": 0.8434, + "step": 4882 + }, + { + "epoch": 0.8186943309231898, + "grad_norm": 0.38649222254753113, + "learning_rate": 1.692287407496689e-06, + "loss": 0.8963, + "step": 4883 + }, + { + "epoch": 0.8188619930839359, + "grad_norm": 0.5236232876777649, + "learning_rate": 1.689265893916837e-06, + "loss": 0.8112, + "step": 4884 + }, + { + "epoch": 0.819029655244682, + "grad_norm": 0.434641033411026, + "learning_rate": 1.686246831275382e-06, + "loss": 0.8081, + "step": 4885 + }, + { + "epoch": 0.8191973174054281, + "grad_norm": 0.7000330686569214, + "learning_rate": 1.683230220462686e-06, + "loss": 0.9179, + "step": 4886 + }, + { + "epoch": 0.8193649795661742, + "grad_norm": 0.4006652534008026, + "learning_rate": 1.6802160623683838e-06, + "loss": 0.7776, + "step": 4887 + }, + { + "epoch": 0.8195326417269203, + "grad_norm": 0.3367191553115845, + "learning_rate": 1.6772043578813846e-06, + "loss": 0.7792, + "step": 4888 + }, + { + "epoch": 0.8197003038876663, + "grad_norm": 0.8314791321754456, + "learning_rate": 1.674195107889881e-06, + "loss": 0.9016, + "step": 4889 + }, + { + "epoch": 0.8198679660484125, + "grad_norm": 0.6004918217658997, + "learning_rate": 1.6711883132813355e-06, + "loss": 0.87, + "step": 4890 + }, + { + "epoch": 0.8200356282091585, + "grad_norm": 0.5409256815910339, + "learning_rate": 1.668183974942491e-06, + "loss": 0.872, + "step": 4891 + }, + { + "epoch": 0.8202032903699047, + "grad_norm": 0.7900858521461487, + "learning_rate": 1.6651820937593677e-06, + "loss": 0.8623, + "step": 4892 + }, + { + "epoch": 0.8203709525306507, + "grad_norm": 0.3986113667488098, + "learning_rate": 1.6621826706172494e-06, + "loss": 0.8713, + "step": 4893 + }, + { + "epoch": 0.8205386146913969, + "grad_norm": 0.35505178570747375, + "learning_rate": 1.6591857064007067e-06, + "loss": 0.8332, + "step": 4894 + }, + { + "epoch": 0.8207062768521429, + "grad_norm": 0.3726397752761841, + "learning_rate": 1.6561912019935843e-06, + "loss": 0.7985, + "step": 4895 + }, + { + "epoch": 0.8208739390128891, + "grad_norm": 0.3969880938529968, + "learning_rate": 1.6531991582789964e-06, + "loss": 0.8429, + "step": 4896 + }, + { + "epoch": 0.8210416011736351, + "grad_norm": 0.4723566472530365, + "learning_rate": 1.6502095761393332e-06, + "loss": 0.7666, + "step": 4897 + }, + { + "epoch": 0.8212092633343813, + "grad_norm": 0.8912177085876465, + "learning_rate": 1.6472224564562577e-06, + "loss": 0.7355, + "step": 4898 + }, + { + "epoch": 0.8213769254951273, + "grad_norm": 0.3965393602848053, + "learning_rate": 1.6442378001107106e-06, + "loss": 0.8948, + "step": 4899 + }, + { + "epoch": 0.8215445876558735, + "grad_norm": 0.45103752613067627, + "learning_rate": 1.6412556079829066e-06, + "loss": 0.7278, + "step": 4900 + }, + { + "epoch": 0.8217122498166195, + "grad_norm": 0.4880988895893097, + "learning_rate": 1.638275880952328e-06, + "loss": 0.7972, + "step": 4901 + }, + { + "epoch": 0.8218799119773657, + "grad_norm": 0.4846517741680145, + "learning_rate": 1.6352986198977327e-06, + "loss": 0.8895, + "step": 4902 + }, + { + "epoch": 0.8220475741381117, + "grad_norm": 0.6601195335388184, + "learning_rate": 1.632323825697154e-06, + "loss": 0.8519, + "step": 4903 + }, + { + "epoch": 0.8222152362988578, + "grad_norm": 0.4242129921913147, + "learning_rate": 1.6293514992278935e-06, + "loss": 0.8136, + "step": 4904 + }, + { + "epoch": 0.8223828984596039, + "grad_norm": 0.3869190514087677, + "learning_rate": 1.626381641366529e-06, + "loss": 0.8755, + "step": 4905 + }, + { + "epoch": 0.82255056062035, + "grad_norm": 0.393439918756485, + "learning_rate": 1.6234142529889084e-06, + "loss": 0.8926, + "step": 4906 + }, + { + "epoch": 0.8227182227810961, + "grad_norm": 0.42509788274765015, + "learning_rate": 1.6204493349701477e-06, + "loss": 0.8543, + "step": 4907 + }, + { + "epoch": 0.8228858849418422, + "grad_norm": 0.6783158779144287, + "learning_rate": 1.6174868881846429e-06, + "loss": 0.7661, + "step": 4908 + }, + { + "epoch": 0.8230535471025883, + "grad_norm": 0.41522926092147827, + "learning_rate": 1.6145269135060514e-06, + "loss": 0.9209, + "step": 4909 + }, + { + "epoch": 0.8232212092633344, + "grad_norm": 0.5016360282897949, + "learning_rate": 1.6115694118073112e-06, + "loss": 0.7873, + "step": 4910 + }, + { + "epoch": 0.8233888714240805, + "grad_norm": 0.6372166872024536, + "learning_rate": 1.6086143839606238e-06, + "loss": 0.8311, + "step": 4911 + }, + { + "epoch": 0.8235565335848266, + "grad_norm": 0.44918274879455566, + "learning_rate": 1.6056618308374606e-06, + "loss": 0.8858, + "step": 4912 + }, + { + "epoch": 0.8237241957455727, + "grad_norm": 0.36326393485069275, + "learning_rate": 1.6027117533085723e-06, + "loss": 0.7643, + "step": 4913 + }, + { + "epoch": 0.8238918579063188, + "grad_norm": 0.49712425470352173, + "learning_rate": 1.59976415224397e-06, + "loss": 0.8362, + "step": 4914 + }, + { + "epoch": 0.8240595200670648, + "grad_norm": 0.5948045253753662, + "learning_rate": 1.5968190285129349e-06, + "loss": 0.8402, + "step": 4915 + }, + { + "epoch": 0.824227182227811, + "grad_norm": 0.39514023065567017, + "learning_rate": 1.5938763829840265e-06, + "loss": 0.8359, + "step": 4916 + }, + { + "epoch": 0.824394844388557, + "grad_norm": 0.5166681408882141, + "learning_rate": 1.5909362165250609e-06, + "loss": 0.7518, + "step": 4917 + }, + { + "epoch": 0.8245625065493032, + "grad_norm": 1.076876163482666, + "learning_rate": 1.587998530003133e-06, + "loss": 0.8548, + "step": 4918 + }, + { + "epoch": 0.8247301687100492, + "grad_norm": 0.4052722454071045, + "learning_rate": 1.5850633242846082e-06, + "loss": 0.7875, + "step": 4919 + }, + { + "epoch": 0.8248978308707954, + "grad_norm": 0.4035169184207916, + "learning_rate": 1.5821306002351045e-06, + "loss": 0.8195, + "step": 4920 + }, + { + "epoch": 0.8250654930315414, + "grad_norm": 0.49509620666503906, + "learning_rate": 1.5792003587195237e-06, + "loss": 0.8904, + "step": 4921 + }, + { + "epoch": 0.8252331551922876, + "grad_norm": 0.5392580628395081, + "learning_rate": 1.5762726006020325e-06, + "loss": 0.8599, + "step": 4922 + }, + { + "epoch": 0.8254008173530336, + "grad_norm": 0.5295587182044983, + "learning_rate": 1.5733473267460609e-06, + "loss": 0.8348, + "step": 4923 + }, + { + "epoch": 0.8255684795137798, + "grad_norm": 0.625406801700592, + "learning_rate": 1.570424538014307e-06, + "loss": 0.873, + "step": 4924 + }, + { + "epoch": 0.8257361416745258, + "grad_norm": 0.7624419331550598, + "learning_rate": 1.5675042352687397e-06, + "loss": 0.8689, + "step": 4925 + }, + { + "epoch": 0.825903803835272, + "grad_norm": 0.3165587782859802, + "learning_rate": 1.5645864193705896e-06, + "loss": 0.8504, + "step": 4926 + }, + { + "epoch": 0.826071465996018, + "grad_norm": 0.377168744802475, + "learning_rate": 1.5616710911803611e-06, + "loss": 0.7757, + "step": 4927 + }, + { + "epoch": 0.8262391281567641, + "grad_norm": 0.35758042335510254, + "learning_rate": 1.558758251557817e-06, + "loss": 0.8374, + "step": 4928 + }, + { + "epoch": 0.8264067903175102, + "grad_norm": 0.3596714735031128, + "learning_rate": 1.55584790136199e-06, + "loss": 0.8467, + "step": 4929 + }, + { + "epoch": 0.8265744524782563, + "grad_norm": 0.3461354970932007, + "learning_rate": 1.5529400414511809e-06, + "loss": 0.7623, + "step": 4930 + }, + { + "epoch": 0.8267421146390024, + "grad_norm": 0.4955442547798157, + "learning_rate": 1.5500346726829495e-06, + "loss": 0.7475, + "step": 4931 + }, + { + "epoch": 0.8269097767997485, + "grad_norm": 0.3839476406574249, + "learning_rate": 1.5471317959141296e-06, + "loss": 0.7481, + "step": 4932 + }, + { + "epoch": 0.8270774389604946, + "grad_norm": 0.5946776866912842, + "learning_rate": 1.544231412000814e-06, + "loss": 0.806, + "step": 4933 + }, + { + "epoch": 0.8272451011212407, + "grad_norm": 0.48449745774269104, + "learning_rate": 1.5413335217983594e-06, + "loss": 0.7931, + "step": 4934 + }, + { + "epoch": 0.8274127632819868, + "grad_norm": 0.36142489314079285, + "learning_rate": 1.5384381261613924e-06, + "loss": 0.8365, + "step": 4935 + }, + { + "epoch": 0.8275804254427329, + "grad_norm": 0.5161846876144409, + "learning_rate": 1.5355452259437986e-06, + "loss": 0.9174, + "step": 4936 + }, + { + "epoch": 0.827748087603479, + "grad_norm": 0.5528901815414429, + "learning_rate": 1.5326548219987326e-06, + "loss": 0.8181, + "step": 4937 + }, + { + "epoch": 0.8279157497642251, + "grad_norm": 0.45687246322631836, + "learning_rate": 1.52976691517861e-06, + "loss": 0.8257, + "step": 4938 + }, + { + "epoch": 0.8280834119249711, + "grad_norm": 0.3636484146118164, + "learning_rate": 1.5268815063351072e-06, + "loss": 0.7596, + "step": 4939 + }, + { + "epoch": 0.8282510740857173, + "grad_norm": 0.36633291840553284, + "learning_rate": 1.5239985963191683e-06, + "loss": 0.7881, + "step": 4940 + }, + { + "epoch": 0.8284187362464633, + "grad_norm": 0.41070568561553955, + "learning_rate": 1.5211181859810032e-06, + "loss": 0.8486, + "step": 4941 + }, + { + "epoch": 0.8285863984072095, + "grad_norm": 0.8401875495910645, + "learning_rate": 1.5182402761700776e-06, + "loss": 0.8714, + "step": 4942 + }, + { + "epoch": 0.8287540605679555, + "grad_norm": 0.3784967064857483, + "learning_rate": 1.5153648677351196e-06, + "loss": 0.7949, + "step": 4943 + }, + { + "epoch": 0.8289217227287017, + "grad_norm": 0.40644749999046326, + "learning_rate": 1.5124919615241284e-06, + "loss": 0.7011, + "step": 4944 + }, + { + "epoch": 0.8290893848894477, + "grad_norm": 0.4508950412273407, + "learning_rate": 1.5096215583843554e-06, + "loss": 0.8217, + "step": 4945 + }, + { + "epoch": 0.8292570470501939, + "grad_norm": 0.3945903182029724, + "learning_rate": 1.5067536591623233e-06, + "loss": 0.74, + "step": 4946 + }, + { + "epoch": 0.8294247092109399, + "grad_norm": 0.3296574652194977, + "learning_rate": 1.5038882647038034e-06, + "loss": 0.8235, + "step": 4947 + }, + { + "epoch": 0.8295923713716861, + "grad_norm": 0.3594135344028473, + "learning_rate": 1.5010253758538396e-06, + "loss": 0.8322, + "step": 4948 + }, + { + "epoch": 0.8297600335324321, + "grad_norm": 0.6399166584014893, + "learning_rate": 1.4981649934567365e-06, + "loss": 0.8605, + "step": 4949 + }, + { + "epoch": 0.8299276956931783, + "grad_norm": 0.3736465275287628, + "learning_rate": 1.4953071183560508e-06, + "loss": 0.7765, + "step": 4950 + }, + { + "epoch": 0.8300953578539243, + "grad_norm": 0.37793681025505066, + "learning_rate": 1.49245175139461e-06, + "loss": 0.8175, + "step": 4951 + }, + { + "epoch": 0.8302630200146704, + "grad_norm": 0.5853179693222046, + "learning_rate": 1.4895988934144956e-06, + "loss": 0.7904, + "step": 4952 + }, + { + "epoch": 0.8304306821754165, + "grad_norm": 0.5771636962890625, + "learning_rate": 1.4867485452570473e-06, + "loss": 0.7803, + "step": 4953 + }, + { + "epoch": 0.8305983443361626, + "grad_norm": 0.35129693150520325, + "learning_rate": 1.483900707762873e-06, + "loss": 0.8745, + "step": 4954 + }, + { + "epoch": 0.8307660064969087, + "grad_norm": 0.5694350600242615, + "learning_rate": 1.4810553817718343e-06, + "loss": 0.8235, + "step": 4955 + }, + { + "epoch": 0.8309336686576548, + "grad_norm": 0.41765889525413513, + "learning_rate": 1.4782125681230497e-06, + "loss": 0.9147, + "step": 4956 + }, + { + "epoch": 0.8311013308184009, + "grad_norm": 0.6250484585762024, + "learning_rate": 1.4753722676549042e-06, + "loss": 0.8673, + "step": 4957 + }, + { + "epoch": 0.831268992979147, + "grad_norm": 0.33489978313446045, + "learning_rate": 1.4725344812050336e-06, + "loss": 0.7793, + "step": 4958 + }, + { + "epoch": 0.8314366551398931, + "grad_norm": 0.39565497636795044, + "learning_rate": 1.469699209610338e-06, + "loss": 0.8873, + "step": 4959 + }, + { + "epoch": 0.8316043173006392, + "grad_norm": 0.41611769795417786, + "learning_rate": 1.46686645370698e-06, + "loss": 0.7699, + "step": 4960 + }, + { + "epoch": 0.8317719794613853, + "grad_norm": 0.3994881212711334, + "learning_rate": 1.4640362143303644e-06, + "loss": 0.935, + "step": 4961 + }, + { + "epoch": 0.8319396416221314, + "grad_norm": 0.7628393173217773, + "learning_rate": 1.4612084923151703e-06, + "loss": 0.8898, + "step": 4962 + }, + { + "epoch": 0.8321073037828774, + "grad_norm": 0.445016473531723, + "learning_rate": 1.4583832884953241e-06, + "loss": 0.8472, + "step": 4963 + }, + { + "epoch": 0.8322749659436236, + "grad_norm": 0.46770772337913513, + "learning_rate": 1.4555606037040159e-06, + "loss": 0.8566, + "step": 4964 + }, + { + "epoch": 0.8324426281043696, + "grad_norm": 0.35536736249923706, + "learning_rate": 1.4527404387736943e-06, + "loss": 0.8784, + "step": 4965 + }, + { + "epoch": 0.8326102902651158, + "grad_norm": 0.3618987202644348, + "learning_rate": 1.4499227945360505e-06, + "loss": 0.834, + "step": 4966 + }, + { + "epoch": 0.8327779524258618, + "grad_norm": 0.46326589584350586, + "learning_rate": 1.44710767182205e-06, + "loss": 0.8092, + "step": 4967 + }, + { + "epoch": 0.832945614586608, + "grad_norm": 0.3517701029777527, + "learning_rate": 1.4442950714619075e-06, + "loss": 0.7963, + "step": 4968 + }, + { + "epoch": 0.8331132767473541, + "grad_norm": 0.5149338245391846, + "learning_rate": 1.4414849942850927e-06, + "loss": 0.8521, + "step": 4969 + }, + { + "epoch": 0.8332809389081002, + "grad_norm": 0.5947842597961426, + "learning_rate": 1.4386774411203285e-06, + "loss": 0.8438, + "step": 4970 + }, + { + "epoch": 0.8334486010688463, + "grad_norm": 0.46693262457847595, + "learning_rate": 1.4358724127956036e-06, + "loss": 0.9143, + "step": 4971 + }, + { + "epoch": 0.8336162632295924, + "grad_norm": 0.3993178904056549, + "learning_rate": 1.4330699101381496e-06, + "loss": 0.8364, + "step": 4972 + }, + { + "epoch": 0.8337839253903385, + "grad_norm": 0.44750991463661194, + "learning_rate": 1.430269933974463e-06, + "loss": 0.7832, + "step": 4973 + }, + { + "epoch": 0.8339515875510846, + "grad_norm": 0.40281417965888977, + "learning_rate": 1.4274724851302912e-06, + "loss": 0.9152, + "step": 4974 + }, + { + "epoch": 0.8341192497118307, + "grad_norm": 0.34603404998779297, + "learning_rate": 1.4246775644306333e-06, + "loss": 0.7777, + "step": 4975 + }, + { + "epoch": 0.8342869118725768, + "grad_norm": 0.4452677071094513, + "learning_rate": 1.4218851726997507e-06, + "loss": 0.8059, + "step": 4976 + }, + { + "epoch": 0.8344545740333229, + "grad_norm": 0.3128507137298584, + "learning_rate": 1.4190953107611483e-06, + "loss": 0.8199, + "step": 4977 + }, + { + "epoch": 0.834622236194069, + "grad_norm": 0.40192288160324097, + "learning_rate": 1.4163079794375977e-06, + "loss": 0.7547, + "step": 4978 + }, + { + "epoch": 0.8347898983548151, + "grad_norm": 0.3185981214046478, + "learning_rate": 1.413523179551115e-06, + "loss": 0.7361, + "step": 4979 + }, + { + "epoch": 0.8349575605155611, + "grad_norm": 0.5001709461212158, + "learning_rate": 1.4107409119229675e-06, + "loss": 0.8404, + "step": 4980 + }, + { + "epoch": 0.8351252226763073, + "grad_norm": 0.5474051237106323, + "learning_rate": 1.4079611773736879e-06, + "loss": 0.8478, + "step": 4981 + }, + { + "epoch": 0.8352928848370533, + "grad_norm": 0.4860477149486542, + "learning_rate": 1.4051839767230479e-06, + "loss": 0.787, + "step": 4982 + }, + { + "epoch": 0.8354605469977995, + "grad_norm": 0.3890592157840729, + "learning_rate": 1.4024093107900838e-06, + "loss": 0.7679, + "step": 4983 + }, + { + "epoch": 0.8356282091585455, + "grad_norm": 0.4080706536769867, + "learning_rate": 1.3996371803930753e-06, + "loss": 0.9019, + "step": 4984 + }, + { + "epoch": 0.8357958713192917, + "grad_norm": 0.36382532119750977, + "learning_rate": 1.3968675863495572e-06, + "loss": 0.878, + "step": 4985 + }, + { + "epoch": 0.8359635334800377, + "grad_norm": 0.40647241473197937, + "learning_rate": 1.39410052947632e-06, + "loss": 0.8264, + "step": 4986 + }, + { + "epoch": 0.8361311956407839, + "grad_norm": 0.36876314878463745, + "learning_rate": 1.3913360105894048e-06, + "loss": 0.8902, + "step": 4987 + }, + { + "epoch": 0.8362988578015299, + "grad_norm": 0.402743935585022, + "learning_rate": 1.3885740305040962e-06, + "loss": 0.9009, + "step": 4988 + }, + { + "epoch": 0.8364665199622761, + "grad_norm": 0.4157020151615143, + "learning_rate": 1.3858145900349385e-06, + "loss": 0.7778, + "step": 4989 + }, + { + "epoch": 0.8366341821230221, + "grad_norm": 0.9957639575004578, + "learning_rate": 1.3830576899957292e-06, + "loss": 0.8508, + "step": 4990 + }, + { + "epoch": 0.8368018442837682, + "grad_norm": 0.4136386215686798, + "learning_rate": 1.3803033311995072e-06, + "loss": 0.8265, + "step": 4991 + }, + { + "epoch": 0.8369695064445143, + "grad_norm": 0.4569907486438751, + "learning_rate": 1.3775515144585705e-06, + "loss": 0.8268, + "step": 4992 + }, + { + "epoch": 0.8371371686052604, + "grad_norm": 0.35812056064605713, + "learning_rate": 1.3748022405844619e-06, + "loss": 0.807, + "step": 4993 + }, + { + "epoch": 0.8373048307660065, + "grad_norm": 0.4523765444755554, + "learning_rate": 1.3720555103879751e-06, + "loss": 0.9474, + "step": 4994 + }, + { + "epoch": 0.8374724929267526, + "grad_norm": 0.3831980228424072, + "learning_rate": 1.369311324679159e-06, + "loss": 0.8371, + "step": 4995 + }, + { + "epoch": 0.8376401550874987, + "grad_norm": 0.4144824147224426, + "learning_rate": 1.366569684267306e-06, + "loss": 0.7924, + "step": 4996 + }, + { + "epoch": 0.8378078172482448, + "grad_norm": 0.40071797370910645, + "learning_rate": 1.363830589960956e-06, + "loss": 0.7691, + "step": 4997 + }, + { + "epoch": 0.8379754794089909, + "grad_norm": 0.46362945437431335, + "learning_rate": 1.3610940425679076e-06, + "loss": 0.8519, + "step": 4998 + }, + { + "epoch": 0.838143141569737, + "grad_norm": 0.3925900161266327, + "learning_rate": 1.358360042895198e-06, + "loss": 0.795, + "step": 4999 + }, + { + "epoch": 0.8383108037304831, + "grad_norm": 0.5459030270576477, + "learning_rate": 1.355628591749122e-06, + "loss": 0.8003, + "step": 5000 + }, + { + "epoch": 0.8384784658912292, + "grad_norm": 0.3537822663784027, + "learning_rate": 1.3528996899352155e-06, + "loss": 0.849, + "step": 5001 + }, + { + "epoch": 0.8386461280519752, + "grad_norm": 0.42230555415153503, + "learning_rate": 1.3501733382582627e-06, + "loss": 0.771, + "step": 5002 + }, + { + "epoch": 0.8388137902127214, + "grad_norm": 0.36975836753845215, + "learning_rate": 1.347449537522305e-06, + "loss": 0.8506, + "step": 5003 + }, + { + "epoch": 0.8389814523734674, + "grad_norm": 0.4725952446460724, + "learning_rate": 1.3447282885306201e-06, + "loss": 0.7526, + "step": 5004 + }, + { + "epoch": 0.8391491145342136, + "grad_norm": 0.6361253261566162, + "learning_rate": 1.3420095920857378e-06, + "loss": 0.7488, + "step": 5005 + }, + { + "epoch": 0.8393167766949596, + "grad_norm": 0.3656197786331177, + "learning_rate": 1.3392934489894427e-06, + "loss": 0.7931, + "step": 5006 + }, + { + "epoch": 0.8394844388557058, + "grad_norm": 0.3905761241912842, + "learning_rate": 1.336579860042748e-06, + "loss": 0.7973, + "step": 5007 + }, + { + "epoch": 0.8396521010164518, + "grad_norm": 0.35232487320899963, + "learning_rate": 1.333868826045932e-06, + "loss": 0.7842, + "step": 5008 + }, + { + "epoch": 0.839819763177198, + "grad_norm": 0.5195075273513794, + "learning_rate": 1.3311603477985101e-06, + "loss": 0.8098, + "step": 5009 + }, + { + "epoch": 0.839987425337944, + "grad_norm": 0.5655362606048584, + "learning_rate": 1.3284544260992483e-06, + "loss": 0.7666, + "step": 5010 + }, + { + "epoch": 0.8401550874986902, + "grad_norm": 0.38470450043678284, + "learning_rate": 1.3257510617461533e-06, + "loss": 0.7918, + "step": 5011 + }, + { + "epoch": 0.8403227496594362, + "grad_norm": 0.43328046798706055, + "learning_rate": 1.32305025553648e-06, + "loss": 0.7772, + "step": 5012 + }, + { + "epoch": 0.8404904118201824, + "grad_norm": 0.38678938150405884, + "learning_rate": 1.320352008266731e-06, + "loss": 0.7945, + "step": 5013 + }, + { + "epoch": 0.8406580739809284, + "grad_norm": 0.4375285506248474, + "learning_rate": 1.3176563207326554e-06, + "loss": 0.8519, + "step": 5014 + }, + { + "epoch": 0.8408257361416746, + "grad_norm": 0.5520859360694885, + "learning_rate": 1.314963193729243e-06, + "loss": 0.7942, + "step": 5015 + }, + { + "epoch": 0.8409933983024206, + "grad_norm": 0.5742729902267456, + "learning_rate": 1.312272628050728e-06, + "loss": 0.7296, + "step": 5016 + }, + { + "epoch": 0.8411610604631667, + "grad_norm": 0.35533037781715393, + "learning_rate": 1.3095846244905952e-06, + "loss": 0.8134, + "step": 5017 + }, + { + "epoch": 0.8413287226239128, + "grad_norm": 0.4646261930465698, + "learning_rate": 1.3068991838415669e-06, + "loss": 0.8755, + "step": 5018 + }, + { + "epoch": 0.8414963847846589, + "grad_norm": 0.56504225730896, + "learning_rate": 1.3042163068956147e-06, + "loss": 0.745, + "step": 5019 + }, + { + "epoch": 0.841664046945405, + "grad_norm": 0.38503456115722656, + "learning_rate": 1.3015359944439544e-06, + "loss": 0.8063, + "step": 5020 + }, + { + "epoch": 0.8418317091061511, + "grad_norm": 0.3094675540924072, + "learning_rate": 1.2988582472770372e-06, + "loss": 0.806, + "step": 5021 + }, + { + "epoch": 0.8419993712668972, + "grad_norm": 0.4315580129623413, + "learning_rate": 1.2961830661845697e-06, + "loss": 0.8175, + "step": 5022 + }, + { + "epoch": 0.8421670334276433, + "grad_norm": 0.3926454484462738, + "learning_rate": 1.293510451955492e-06, + "loss": 0.8225, + "step": 5023 + }, + { + "epoch": 0.8423346955883894, + "grad_norm": 0.7259623408317566, + "learning_rate": 1.2908404053779943e-06, + "loss": 0.8553, + "step": 5024 + }, + { + "epoch": 0.8425023577491355, + "grad_norm": 0.3977755606174469, + "learning_rate": 1.2881729272395049e-06, + "loss": 0.7511, + "step": 5025 + }, + { + "epoch": 0.8426700199098816, + "grad_norm": 0.37917256355285645, + "learning_rate": 1.2855080183266954e-06, + "loss": 0.8226, + "step": 5026 + }, + { + "epoch": 0.8428376820706277, + "grad_norm": 0.3832605481147766, + "learning_rate": 1.2828456794254818e-06, + "loss": 0.8326, + "step": 5027 + }, + { + "epoch": 0.8430053442313737, + "grad_norm": 0.3946949243545532, + "learning_rate": 1.2801859113210213e-06, + "loss": 0.7871, + "step": 5028 + }, + { + "epoch": 0.8431730063921199, + "grad_norm": 0.38526833057403564, + "learning_rate": 1.277528714797709e-06, + "loss": 0.8081, + "step": 5029 + }, + { + "epoch": 0.8433406685528659, + "grad_norm": 0.32758602499961853, + "learning_rate": 1.2748740906391887e-06, + "loss": 0.7913, + "step": 5030 + }, + { + "epoch": 0.8435083307136121, + "grad_norm": 0.5229128003120422, + "learning_rate": 1.2722220396283401e-06, + "loss": 0.7908, + "step": 5031 + }, + { + "epoch": 0.8436759928743581, + "grad_norm": 0.46804341673851013, + "learning_rate": 1.2695725625472854e-06, + "loss": 0.7986, + "step": 5032 + }, + { + "epoch": 0.8438436550351043, + "grad_norm": 0.5670343637466431, + "learning_rate": 1.266925660177394e-06, + "loss": 0.8722, + "step": 5033 + }, + { + "epoch": 0.8440113171958503, + "grad_norm": 0.3946131765842438, + "learning_rate": 1.264281333299261e-06, + "loss": 0.7333, + "step": 5034 + }, + { + "epoch": 0.8441789793565965, + "grad_norm": 1.2765463590621948, + "learning_rate": 1.2616395826927352e-06, + "loss": 0.9144, + "step": 5035 + }, + { + "epoch": 0.8443466415173425, + "grad_norm": 0.39245614409446716, + "learning_rate": 1.2590004091369045e-06, + "loss": 0.8414, + "step": 5036 + }, + { + "epoch": 0.8445143036780887, + "grad_norm": 0.3988438844680786, + "learning_rate": 1.2563638134100898e-06, + "loss": 0.8746, + "step": 5037 + }, + { + "epoch": 0.8446819658388347, + "grad_norm": 0.7531177997589111, + "learning_rate": 1.2537297962898587e-06, + "loss": 0.8994, + "step": 5038 + }, + { + "epoch": 0.8448496279995809, + "grad_norm": 0.3667586147785187, + "learning_rate": 1.2510983585530134e-06, + "loss": 0.8386, + "step": 5039 + }, + { + "epoch": 0.8450172901603269, + "grad_norm": 0.4034627377986908, + "learning_rate": 1.2484695009755976e-06, + "loss": 0.8495, + "step": 5040 + }, + { + "epoch": 0.845184952321073, + "grad_norm": 0.39673033356666565, + "learning_rate": 1.2458432243328955e-06, + "loss": 0.8627, + "step": 5041 + }, + { + "epoch": 0.8453526144818191, + "grad_norm": 0.4349612295627594, + "learning_rate": 1.2432195293994288e-06, + "loss": 0.8247, + "step": 5042 + }, + { + "epoch": 0.8455202766425652, + "grad_norm": 0.34705671668052673, + "learning_rate": 1.2405984169489526e-06, + "loss": 0.8052, + "step": 5043 + }, + { + "epoch": 0.8456879388033113, + "grad_norm": 0.39677438139915466, + "learning_rate": 1.2379798877544735e-06, + "loss": 0.8462, + "step": 5044 + }, + { + "epoch": 0.8458556009640574, + "grad_norm": 0.42162394523620605, + "learning_rate": 1.2353639425882203e-06, + "loss": 0.8864, + "step": 5045 + }, + { + "epoch": 0.8460232631248035, + "grad_norm": 0.40522149205207825, + "learning_rate": 1.2327505822216734e-06, + "loss": 0.8179, + "step": 5046 + }, + { + "epoch": 0.8461909252855496, + "grad_norm": 0.45286887884140015, + "learning_rate": 1.2301398074255444e-06, + "loss": 0.7911, + "step": 5047 + }, + { + "epoch": 0.8463585874462957, + "grad_norm": 0.5825637578964233, + "learning_rate": 1.2275316189697784e-06, + "loss": 0.8652, + "step": 5048 + }, + { + "epoch": 0.8465262496070418, + "grad_norm": 0.49502861499786377, + "learning_rate": 1.2249260176235678e-06, + "loss": 0.8698, + "step": 5049 + }, + { + "epoch": 0.8466939117677879, + "grad_norm": 0.41778823733329773, + "learning_rate": 1.2223230041553335e-06, + "loss": 0.7378, + "step": 5050 + }, + { + "epoch": 0.846861573928534, + "grad_norm": 0.38942602276802063, + "learning_rate": 1.2197225793327393e-06, + "loss": 0.8643, + "step": 5051 + }, + { + "epoch": 0.84702923608928, + "grad_norm": 0.4558364748954773, + "learning_rate": 1.2171247439226809e-06, + "loss": 0.8634, + "step": 5052 + }, + { + "epoch": 0.8471968982500262, + "grad_norm": 0.7153915762901306, + "learning_rate": 1.2145294986912903e-06, + "loss": 0.7978, + "step": 5053 + }, + { + "epoch": 0.8473645604107722, + "grad_norm": 0.3906574547290802, + "learning_rate": 1.2119368444039392e-06, + "loss": 0.765, + "step": 5054 + }, + { + "epoch": 0.8475322225715184, + "grad_norm": 0.412419855594635, + "learning_rate": 1.2093467818252358e-06, + "loss": 0.8181, + "step": 5055 + }, + { + "epoch": 0.8476998847322645, + "grad_norm": 0.3883296549320221, + "learning_rate": 1.206759311719019e-06, + "loss": 0.8006, + "step": 5056 + }, + { + "epoch": 0.8478675468930106, + "grad_norm": 0.39480140805244446, + "learning_rate": 1.2041744348483652e-06, + "loss": 0.8763, + "step": 5057 + }, + { + "epoch": 0.8480352090537567, + "grad_norm": 0.6128484606742859, + "learning_rate": 1.2015921519755891e-06, + "loss": 0.7826, + "step": 5058 + }, + { + "epoch": 0.8482028712145028, + "grad_norm": 0.6531548500061035, + "learning_rate": 1.1990124638622337e-06, + "loss": 0.8052, + "step": 5059 + }, + { + "epoch": 0.8483705333752489, + "grad_norm": 0.3930145502090454, + "learning_rate": 1.196435371269089e-06, + "loss": 0.8634, + "step": 5060 + }, + { + "epoch": 0.848538195535995, + "grad_norm": 0.5058462619781494, + "learning_rate": 1.193860874956162e-06, + "loss": 0.8377, + "step": 5061 + }, + { + "epoch": 0.8487058576967411, + "grad_norm": 0.3774513602256775, + "learning_rate": 1.1912889756827073e-06, + "loss": 0.7747, + "step": 5062 + }, + { + "epoch": 0.8488735198574872, + "grad_norm": 0.5487313270568848, + "learning_rate": 1.1887196742072115e-06, + "loss": 0.9396, + "step": 5063 + }, + { + "epoch": 0.8490411820182333, + "grad_norm": 0.40379512310028076, + "learning_rate": 1.186152971287392e-06, + "loss": 0.875, + "step": 5064 + }, + { + "epoch": 0.8492088441789793, + "grad_norm": 0.45634934306144714, + "learning_rate": 1.1835888676802021e-06, + "loss": 0.8128, + "step": 5065 + }, + { + "epoch": 0.8493765063397255, + "grad_norm": 0.4219779074192047, + "learning_rate": 1.181027364141828e-06, + "loss": 0.854, + "step": 5066 + }, + { + "epoch": 0.8495441685004715, + "grad_norm": 0.4084298014640808, + "learning_rate": 1.1784684614276854e-06, + "loss": 0.8018, + "step": 5067 + }, + { + "epoch": 0.8497118306612177, + "grad_norm": 0.40283524990081787, + "learning_rate": 1.1759121602924306e-06, + "loss": 0.8024, + "step": 5068 + }, + { + "epoch": 0.8498794928219637, + "grad_norm": 0.718620240688324, + "learning_rate": 1.1733584614899474e-06, + "loss": 0.8947, + "step": 5069 + }, + { + "epoch": 0.8500471549827099, + "grad_norm": 0.4366646111011505, + "learning_rate": 1.1708073657733498e-06, + "loss": 0.9105, + "step": 5070 + }, + { + "epoch": 0.8502148171434559, + "grad_norm": 0.6225143074989319, + "learning_rate": 1.168258873894993e-06, + "loss": 0.8594, + "step": 5071 + }, + { + "epoch": 0.8503824793042021, + "grad_norm": 0.3575076758861542, + "learning_rate": 1.1657129866064542e-06, + "loss": 0.7517, + "step": 5072 + }, + { + "epoch": 0.8505501414649481, + "grad_norm": 0.38253408670425415, + "learning_rate": 1.1631697046585511e-06, + "loss": 0.8772, + "step": 5073 + }, + { + "epoch": 0.8507178036256943, + "grad_norm": 0.4497597813606262, + "learning_rate": 1.1606290288013266e-06, + "loss": 0.8285, + "step": 5074 + }, + { + "epoch": 0.8508854657864403, + "grad_norm": 0.45971205830574036, + "learning_rate": 1.158090959784055e-06, + "loss": 0.8196, + "step": 5075 + }, + { + "epoch": 0.8510531279471865, + "grad_norm": 0.3878073990345001, + "learning_rate": 1.1555554983552498e-06, + "loss": 0.8788, + "step": 5076 + }, + { + "epoch": 0.8512207901079325, + "grad_norm": 0.7124265432357788, + "learning_rate": 1.1530226452626457e-06, + "loss": 0.7709, + "step": 5077 + }, + { + "epoch": 0.8513884522686787, + "grad_norm": 0.7837972044944763, + "learning_rate": 1.1504924012532136e-06, + "loss": 0.8354, + "step": 5078 + }, + { + "epoch": 0.8515561144294247, + "grad_norm": 0.4087355136871338, + "learning_rate": 1.1479647670731574e-06, + "loss": 0.7846, + "step": 5079 + }, + { + "epoch": 0.8517237765901708, + "grad_norm": 0.3593689799308777, + "learning_rate": 1.1454397434679022e-06, + "loss": 0.786, + "step": 5080 + }, + { + "epoch": 0.8518914387509169, + "grad_norm": 0.836773693561554, + "learning_rate": 1.1429173311821095e-06, + "loss": 0.9083, + "step": 5081 + }, + { + "epoch": 0.852059100911663, + "grad_norm": 0.3732576370239258, + "learning_rate": 1.1403975309596738e-06, + "loss": 0.8702, + "step": 5082 + }, + { + "epoch": 0.8522267630724091, + "grad_norm": 0.6807574033737183, + "learning_rate": 1.1378803435437135e-06, + "loss": 0.8596, + "step": 5083 + }, + { + "epoch": 0.8523944252331552, + "grad_norm": 0.4082207977771759, + "learning_rate": 1.1353657696765752e-06, + "loss": 0.8888, + "step": 5084 + }, + { + "epoch": 0.8525620873939013, + "grad_norm": 0.406039297580719, + "learning_rate": 1.132853810099843e-06, + "loss": 0.7492, + "step": 5085 + }, + { + "epoch": 0.8527297495546474, + "grad_norm": 0.3803439736366272, + "learning_rate": 1.1303444655543206e-06, + "loss": 0.8707, + "step": 5086 + }, + { + "epoch": 0.8528974117153935, + "grad_norm": 0.5561349987983704, + "learning_rate": 1.1278377367800486e-06, + "loss": 0.8445, + "step": 5087 + }, + { + "epoch": 0.8530650738761396, + "grad_norm": 0.5199757218360901, + "learning_rate": 1.1253336245162893e-06, + "loss": 0.8566, + "step": 5088 + }, + { + "epoch": 0.8532327360368857, + "grad_norm": 0.39440688490867615, + "learning_rate": 1.1228321295015365e-06, + "loss": 0.7599, + "step": 5089 + }, + { + "epoch": 0.8534003981976318, + "grad_norm": 0.6887969374656677, + "learning_rate": 1.1203332524735145e-06, + "loss": 0.816, + "step": 5090 + }, + { + "epoch": 0.8535680603583778, + "grad_norm": 0.3930720090866089, + "learning_rate": 1.1178369941691703e-06, + "loss": 0.7775, + "step": 5091 + }, + { + "epoch": 0.853735722519124, + "grad_norm": 0.5686238408088684, + "learning_rate": 1.115343355324684e-06, + "loss": 0.7328, + "step": 5092 + }, + { + "epoch": 0.85390338467987, + "grad_norm": 0.4349891245365143, + "learning_rate": 1.1128523366754595e-06, + "loss": 0.8757, + "step": 5093 + }, + { + "epoch": 0.8540710468406162, + "grad_norm": 0.5157162547111511, + "learning_rate": 1.1103639389561272e-06, + "loss": 0.8234, + "step": 5094 + }, + { + "epoch": 0.8542387090013622, + "grad_norm": 0.45852211117744446, + "learning_rate": 1.1078781629005498e-06, + "loss": 0.7912, + "step": 5095 + }, + { + "epoch": 0.8544063711621084, + "grad_norm": 0.3618316054344177, + "learning_rate": 1.1053950092418097e-06, + "loss": 0.8335, + "step": 5096 + }, + { + "epoch": 0.8545740333228544, + "grad_norm": 0.35856497287750244, + "learning_rate": 1.1029144787122226e-06, + "loss": 0.8641, + "step": 5097 + }, + { + "epoch": 0.8547416954836006, + "grad_norm": 0.4254955053329468, + "learning_rate": 1.1004365720433274e-06, + "loss": 0.8358, + "step": 5098 + }, + { + "epoch": 0.8549093576443466, + "grad_norm": 0.4587903320789337, + "learning_rate": 1.0979612899658875e-06, + "loss": 0.8598, + "step": 5099 + }, + { + "epoch": 0.8550770198050928, + "grad_norm": 0.5157368779182434, + "learning_rate": 1.0954886332098946e-06, + "loss": 0.7303, + "step": 5100 + }, + { + "epoch": 0.8552446819658388, + "grad_norm": 0.6669063568115234, + "learning_rate": 1.0930186025045708e-06, + "loss": 0.8335, + "step": 5101 + }, + { + "epoch": 0.855412344126585, + "grad_norm": 0.39919060468673706, + "learning_rate": 1.0905511985783513e-06, + "loss": 0.8401, + "step": 5102 + }, + { + "epoch": 0.855580006287331, + "grad_norm": 0.46347081661224365, + "learning_rate": 1.088086422158906e-06, + "loss": 0.8404, + "step": 5103 + }, + { + "epoch": 0.8557476684480771, + "grad_norm": 0.466889888048172, + "learning_rate": 1.0856242739731325e-06, + "loss": 0.9507, + "step": 5104 + }, + { + "epoch": 0.8559153306088232, + "grad_norm": 0.39966902136802673, + "learning_rate": 1.0831647547471435e-06, + "loss": 0.8386, + "step": 5105 + }, + { + "epoch": 0.8560829927695693, + "grad_norm": 0.4732997715473175, + "learning_rate": 1.0807078652062874e-06, + "loss": 0.8487, + "step": 5106 + }, + { + "epoch": 0.8562506549303154, + "grad_norm": 0.5787397027015686, + "learning_rate": 1.0782536060751237e-06, + "loss": 0.8471, + "step": 5107 + }, + { + "epoch": 0.8564183170910615, + "grad_norm": 0.40982586145401, + "learning_rate": 1.0758019780774488e-06, + "loss": 0.7979, + "step": 5108 + }, + { + "epoch": 0.8565859792518076, + "grad_norm": 0.43968313932418823, + "learning_rate": 1.0733529819362786e-06, + "loss": 0.7994, + "step": 5109 + }, + { + "epoch": 0.8567536414125537, + "grad_norm": 0.4219721853733063, + "learning_rate": 1.0709066183738493e-06, + "loss": 0.9183, + "step": 5110 + }, + { + "epoch": 0.8569213035732998, + "grad_norm": 0.3750663101673126, + "learning_rate": 1.0684628881116265e-06, + "loss": 0.9232, + "step": 5111 + }, + { + "epoch": 0.8570889657340459, + "grad_norm": 0.4194062054157257, + "learning_rate": 1.0660217918702965e-06, + "loss": 0.8087, + "step": 5112 + }, + { + "epoch": 0.857256627894792, + "grad_norm": 0.38401052355766296, + "learning_rate": 1.0635833303697663e-06, + "loss": 0.8462, + "step": 5113 + }, + { + "epoch": 0.8574242900555381, + "grad_norm": 0.38535383343696594, + "learning_rate": 1.0611475043291708e-06, + "loss": 0.8383, + "step": 5114 + }, + { + "epoch": 0.8575919522162841, + "grad_norm": 0.4055083394050598, + "learning_rate": 1.0587143144668655e-06, + "loss": 0.8271, + "step": 5115 + }, + { + "epoch": 0.8577596143770303, + "grad_norm": 0.5805023908615112, + "learning_rate": 1.0562837615004252e-06, + "loss": 0.8211, + "step": 5116 + }, + { + "epoch": 0.8579272765377763, + "grad_norm": 0.44198617339134216, + "learning_rate": 1.053855846146653e-06, + "loss": 0.9089, + "step": 5117 + }, + { + "epoch": 0.8580949386985225, + "grad_norm": 0.6872033476829529, + "learning_rate": 1.0514305691215687e-06, + "loss": 0.8235, + "step": 5118 + }, + { + "epoch": 0.8582626008592685, + "grad_norm": 0.3935192823410034, + "learning_rate": 1.0490079311404178e-06, + "loss": 0.9665, + "step": 5119 + }, + { + "epoch": 0.8584302630200147, + "grad_norm": 0.42438197135925293, + "learning_rate": 1.046587932917671e-06, + "loss": 0.8355, + "step": 5120 + }, + { + "epoch": 0.8585979251807607, + "grad_norm": 0.48464399576187134, + "learning_rate": 1.0441705751670073e-06, + "loss": 0.8473, + "step": 5121 + }, + { + "epoch": 0.8587655873415069, + "grad_norm": 0.6867997646331787, + "learning_rate": 1.041755858601341e-06, + "loss": 0.933, + "step": 5122 + }, + { + "epoch": 0.8589332495022529, + "grad_norm": 0.36188092827796936, + "learning_rate": 1.039343783932798e-06, + "loss": 0.8585, + "step": 5123 + }, + { + "epoch": 0.8591009116629991, + "grad_norm": 0.45864078402519226, + "learning_rate": 1.0369343518727338e-06, + "loss": 0.8093, + "step": 5124 + }, + { + "epoch": 0.8592685738237451, + "grad_norm": 0.5093013644218445, + "learning_rate": 1.0345275631317165e-06, + "loss": 0.8529, + "step": 5125 + }, + { + "epoch": 0.8594362359844913, + "grad_norm": 0.4407593309879303, + "learning_rate": 1.0321234184195372e-06, + "loss": 0.859, + "step": 5126 + }, + { + "epoch": 0.8596038981452373, + "grad_norm": 0.6633378267288208, + "learning_rate": 1.029721918445209e-06, + "loss": 0.9038, + "step": 5127 + }, + { + "epoch": 0.8597715603059835, + "grad_norm": 0.38820528984069824, + "learning_rate": 1.0273230639169663e-06, + "loss": 0.8278, + "step": 5128 + }, + { + "epoch": 0.8599392224667295, + "grad_norm": 0.3281702399253845, + "learning_rate": 1.0249268555422597e-06, + "loss": 0.8391, + "step": 5129 + }, + { + "epoch": 0.8601068846274756, + "grad_norm": 0.3887026906013489, + "learning_rate": 1.0225332940277577e-06, + "loss": 0.7809, + "step": 5130 + }, + { + "epoch": 0.8602745467882217, + "grad_norm": 0.40416014194488525, + "learning_rate": 1.0201423800793553e-06, + "loss": 0.8301, + "step": 5131 + }, + { + "epoch": 0.8604422089489678, + "grad_norm": 0.5501821637153625, + "learning_rate": 1.0177541144021596e-06, + "loss": 0.9014, + "step": 5132 + }, + { + "epoch": 0.8606098711097139, + "grad_norm": 0.4721451997756958, + "learning_rate": 1.0153684977005018e-06, + "loss": 0.8609, + "step": 5133 + }, + { + "epoch": 0.86077753327046, + "grad_norm": 0.5385779738426208, + "learning_rate": 1.0129855306779301e-06, + "loss": 0.7811, + "step": 5134 + }, + { + "epoch": 0.8609451954312061, + "grad_norm": 0.39139533042907715, + "learning_rate": 1.010605214037208e-06, + "loss": 0.8845, + "step": 5135 + }, + { + "epoch": 0.8611128575919522, + "grad_norm": 0.34860363602638245, + "learning_rate": 1.008227548480324e-06, + "loss": 0.8886, + "step": 5136 + }, + { + "epoch": 0.8612805197526983, + "grad_norm": 0.3745764195919037, + "learning_rate": 1.005852534708477e-06, + "loss": 0.7613, + "step": 5137 + }, + { + "epoch": 0.8614481819134444, + "grad_norm": 0.33172425627708435, + "learning_rate": 1.0034801734220922e-06, + "loss": 0.7259, + "step": 5138 + }, + { + "epoch": 0.8616158440741905, + "grad_norm": 0.4795641005039215, + "learning_rate": 1.001110465320807e-06, + "loss": 0.8696, + "step": 5139 + }, + { + "epoch": 0.8617835062349366, + "grad_norm": 0.363203227519989, + "learning_rate": 9.987434111034756e-07, + "loss": 0.8035, + "step": 5140 + }, + { + "epoch": 0.8619511683956828, + "grad_norm": 0.44719621539115906, + "learning_rate": 9.96379011468175e-07, + "loss": 0.8841, + "step": 5141 + }, + { + "epoch": 0.8621188305564288, + "grad_norm": 0.3977639079093933, + "learning_rate": 9.940172671121929e-07, + "loss": 0.8901, + "step": 5142 + }, + { + "epoch": 0.862286492717175, + "grad_norm": 0.4082188010215759, + "learning_rate": 9.91658178732038e-07, + "loss": 0.7791, + "step": 5143 + }, + { + "epoch": 0.862454154877921, + "grad_norm": 0.4015042185783386, + "learning_rate": 9.893017470234356e-07, + "loss": 0.9356, + "step": 5144 + }, + { + "epoch": 0.8626218170386671, + "grad_norm": 0.5281703472137451, + "learning_rate": 9.869479726813236e-07, + "loss": 0.7711, + "step": 5145 + }, + { + "epoch": 0.8627894791994132, + "grad_norm": 0.8039502501487732, + "learning_rate": 9.845968563998621e-07, + "loss": 0.8713, + "step": 5146 + }, + { + "epoch": 0.8629571413601593, + "grad_norm": 0.4611329436302185, + "learning_rate": 9.822483988724273e-07, + "loss": 0.8137, + "step": 5147 + }, + { + "epoch": 0.8631248035209054, + "grad_norm": 0.5646185278892517, + "learning_rate": 9.799026007916012e-07, + "loss": 0.7827, + "step": 5148 + }, + { + "epoch": 0.8632924656816515, + "grad_norm": 0.5700774788856506, + "learning_rate": 9.775594628491913e-07, + "loss": 0.904, + "step": 5149 + }, + { + "epoch": 0.8634601278423976, + "grad_norm": 0.3906707167625427, + "learning_rate": 9.752189857362226e-07, + "loss": 0.8337, + "step": 5150 + }, + { + "epoch": 0.8636277900031437, + "grad_norm": 0.41302984952926636, + "learning_rate": 9.728811701429242e-07, + "loss": 0.8204, + "step": 5151 + }, + { + "epoch": 0.8637954521638898, + "grad_norm": 0.49759745597839355, + "learning_rate": 9.705460167587521e-07, + "loss": 0.7336, + "step": 5152 + }, + { + "epoch": 0.8639631143246359, + "grad_norm": 0.39098086953163147, + "learning_rate": 9.68213526272369e-07, + "loss": 0.8366, + "step": 5153 + }, + { + "epoch": 0.864130776485382, + "grad_norm": 0.3385332226753235, + "learning_rate": 9.658836993716548e-07, + "loss": 0.7663, + "step": 5154 + }, + { + "epoch": 0.8642984386461281, + "grad_norm": 0.3982413709163666, + "learning_rate": 9.635565367437073e-07, + "loss": 0.9586, + "step": 5155 + }, + { + "epoch": 0.8644661008068741, + "grad_norm": 0.4666019082069397, + "learning_rate": 9.612320390748331e-07, + "loss": 0.7692, + "step": 5156 + }, + { + "epoch": 0.8646337629676203, + "grad_norm": 0.9053720235824585, + "learning_rate": 9.58910207050555e-07, + "loss": 0.8789, + "step": 5157 + }, + { + "epoch": 0.8648014251283663, + "grad_norm": 0.3826894760131836, + "learning_rate": 9.565910413556123e-07, + "loss": 0.848, + "step": 5158 + }, + { + "epoch": 0.8649690872891125, + "grad_norm": 0.3934033513069153, + "learning_rate": 9.542745426739518e-07, + "loss": 0.8458, + "step": 5159 + }, + { + "epoch": 0.8651367494498585, + "grad_norm": 0.4008258581161499, + "learning_rate": 9.519607116887431e-07, + "loss": 0.8578, + "step": 5160 + }, + { + "epoch": 0.8653044116106047, + "grad_norm": 0.44706180691719055, + "learning_rate": 9.49649549082361e-07, + "loss": 0.8339, + "step": 5161 + }, + { + "epoch": 0.8654720737713507, + "grad_norm": 0.7951359748840332, + "learning_rate": 9.473410555363927e-07, + "loss": 0.8574, + "step": 5162 + }, + { + "epoch": 0.8656397359320969, + "grad_norm": 0.4611322581768036, + "learning_rate": 9.450352317316469e-07, + "loss": 0.8483, + "step": 5163 + }, + { + "epoch": 0.8658073980928429, + "grad_norm": 0.4625532925128937, + "learning_rate": 9.427320783481353e-07, + "loss": 0.8936, + "step": 5164 + }, + { + "epoch": 0.8659750602535891, + "grad_norm": 0.37857306003570557, + "learning_rate": 9.4043159606509e-07, + "loss": 0.8119, + "step": 5165 + }, + { + "epoch": 0.8661427224143351, + "grad_norm": 0.5109477639198303, + "learning_rate": 9.381337855609507e-07, + "loss": 0.8802, + "step": 5166 + }, + { + "epoch": 0.8663103845750812, + "grad_norm": 1.7759164571762085, + "learning_rate": 9.358386475133674e-07, + "loss": 0.9155, + "step": 5167 + }, + { + "epoch": 0.8664780467358273, + "grad_norm": 0.3659452497959137, + "learning_rate": 9.335461825992054e-07, + "loss": 0.8481, + "step": 5168 + }, + { + "epoch": 0.8666457088965734, + "grad_norm": 0.4770681858062744, + "learning_rate": 9.312563914945461e-07, + "loss": 0.7682, + "step": 5169 + }, + { + "epoch": 0.8668133710573195, + "grad_norm": 0.8329793214797974, + "learning_rate": 9.289692748746726e-07, + "loss": 0.9175, + "step": 5170 + }, + { + "epoch": 0.8669810332180656, + "grad_norm": 0.441036581993103, + "learning_rate": 9.266848334140843e-07, + "loss": 0.8101, + "step": 5171 + }, + { + "epoch": 0.8671486953788117, + "grad_norm": 0.4591298997402191, + "learning_rate": 9.244030677864901e-07, + "loss": 0.9234, + "step": 5172 + }, + { + "epoch": 0.8673163575395578, + "grad_norm": 0.39525070786476135, + "learning_rate": 9.221239786648139e-07, + "loss": 0.7818, + "step": 5173 + }, + { + "epoch": 0.8674840197003039, + "grad_norm": 0.3860989511013031, + "learning_rate": 9.198475667211881e-07, + "loss": 0.851, + "step": 5174 + }, + { + "epoch": 0.86765168186105, + "grad_norm": 0.4074042737483978, + "learning_rate": 9.175738326269501e-07, + "loss": 0.8905, + "step": 5175 + }, + { + "epoch": 0.8678193440217961, + "grad_norm": 0.3486952483654022, + "learning_rate": 9.153027770526546e-07, + "loss": 0.8109, + "step": 5176 + }, + { + "epoch": 0.8679870061825422, + "grad_norm": 0.3882848918437958, + "learning_rate": 9.130344006680658e-07, + "loss": 0.7736, + "step": 5177 + }, + { + "epoch": 0.8681546683432882, + "grad_norm": 0.3613647520542145, + "learning_rate": 9.107687041421531e-07, + "loss": 0.765, + "step": 5178 + }, + { + "epoch": 0.8683223305040344, + "grad_norm": 0.3715355396270752, + "learning_rate": 9.085056881431009e-07, + "loss": 0.9118, + "step": 5179 + }, + { + "epoch": 0.8684899926647804, + "grad_norm": 0.5110697150230408, + "learning_rate": 9.062453533383009e-07, + "loss": 0.8445, + "step": 5180 + }, + { + "epoch": 0.8686576548255266, + "grad_norm": 0.8367433547973633, + "learning_rate": 9.039877003943497e-07, + "loss": 0.8523, + "step": 5181 + }, + { + "epoch": 0.8688253169862726, + "grad_norm": 0.4580248296260834, + "learning_rate": 9.017327299770617e-07, + "loss": 0.868, + "step": 5182 + }, + { + "epoch": 0.8689929791470188, + "grad_norm": 0.9230180978775024, + "learning_rate": 8.994804427514514e-07, + "loss": 0.7957, + "step": 5183 + }, + { + "epoch": 0.8691606413077648, + "grad_norm": 0.37561511993408203, + "learning_rate": 8.972308393817519e-07, + "loss": 0.9475, + "step": 5184 + }, + { + "epoch": 0.869328303468511, + "grad_norm": 0.35868212580680847, + "learning_rate": 8.949839205313948e-07, + "loss": 0.7533, + "step": 5185 + }, + { + "epoch": 0.869495965629257, + "grad_norm": 0.6453197598457336, + "learning_rate": 8.927396868630234e-07, + "loss": 0.9276, + "step": 5186 + }, + { + "epoch": 0.8696636277900032, + "grad_norm": 0.5974776744842529, + "learning_rate": 8.904981390384937e-07, + "loss": 0.8797, + "step": 5187 + }, + { + "epoch": 0.8698312899507492, + "grad_norm": 0.7530732154846191, + "learning_rate": 8.882592777188636e-07, + "loss": 0.7779, + "step": 5188 + }, + { + "epoch": 0.8699989521114954, + "grad_norm": 0.4910556674003601, + "learning_rate": 8.860231035644007e-07, + "loss": 0.8568, + "step": 5189 + }, + { + "epoch": 0.8701666142722414, + "grad_norm": 0.39665699005126953, + "learning_rate": 8.837896172345827e-07, + "loss": 0.8578, + "step": 5190 + }, + { + "epoch": 0.8703342764329876, + "grad_norm": 0.3718913793563843, + "learning_rate": 8.815588193880887e-07, + "loss": 0.7943, + "step": 5191 + }, + { + "epoch": 0.8705019385937336, + "grad_norm": 0.3665577471256256, + "learning_rate": 8.793307106828119e-07, + "loss": 0.903, + "step": 5192 + }, + { + "epoch": 0.8706696007544797, + "grad_norm": 0.38223451375961304, + "learning_rate": 8.771052917758505e-07, + "loss": 0.8133, + "step": 5193 + }, + { + "epoch": 0.8708372629152258, + "grad_norm": 0.45462703704833984, + "learning_rate": 8.748825633235026e-07, + "loss": 0.8094, + "step": 5194 + }, + { + "epoch": 0.8710049250759719, + "grad_norm": 0.6003639101982117, + "learning_rate": 8.726625259812816e-07, + "loss": 0.8195, + "step": 5195 + }, + { + "epoch": 0.871172587236718, + "grad_norm": 0.39782842993736267, + "learning_rate": 8.70445180403905e-07, + "loss": 0.8417, + "step": 5196 + }, + { + "epoch": 0.8713402493974641, + "grad_norm": 0.399222731590271, + "learning_rate": 8.68230527245294e-07, + "loss": 0.844, + "step": 5197 + }, + { + "epoch": 0.8715079115582102, + "grad_norm": 0.4528278112411499, + "learning_rate": 8.660185671585764e-07, + "loss": 0.8549, + "step": 5198 + }, + { + "epoch": 0.8716755737189563, + "grad_norm": 0.3690025210380554, + "learning_rate": 8.638093007960891e-07, + "loss": 0.8686, + "step": 5199 + }, + { + "epoch": 0.8718432358797024, + "grad_norm": 0.43686869740486145, + "learning_rate": 8.616027288093698e-07, + "loss": 0.7774, + "step": 5200 + }, + { + "epoch": 0.8720108980404485, + "grad_norm": 0.6555036902427673, + "learning_rate": 8.593988518491658e-07, + "loss": 0.9075, + "step": 5201 + }, + { + "epoch": 0.8721785602011946, + "grad_norm": 0.32533329725265503, + "learning_rate": 8.571976705654273e-07, + "loss": 0.7884, + "step": 5202 + }, + { + "epoch": 0.8723462223619407, + "grad_norm": 0.7068662643432617, + "learning_rate": 8.54999185607307e-07, + "loss": 0.7727, + "step": 5203 + }, + { + "epoch": 0.8725138845226867, + "grad_norm": 0.6542850732803345, + "learning_rate": 8.528033976231698e-07, + "loss": 0.8242, + "step": 5204 + }, + { + "epoch": 0.8726815466834329, + "grad_norm": 0.411618709564209, + "learning_rate": 8.506103072605776e-07, + "loss": 0.998, + "step": 5205 + }, + { + "epoch": 0.8728492088441789, + "grad_norm": 0.3620193600654602, + "learning_rate": 8.484199151663019e-07, + "loss": 0.7989, + "step": 5206 + }, + { + "epoch": 0.8730168710049251, + "grad_norm": 0.5024863481521606, + "learning_rate": 8.462322219863151e-07, + "loss": 0.8511, + "step": 5207 + }, + { + "epoch": 0.8731845331656711, + "grad_norm": 0.4129505157470703, + "learning_rate": 8.440472283657941e-07, + "loss": 0.715, + "step": 5208 + }, + { + "epoch": 0.8733521953264173, + "grad_norm": 0.3400903642177582, + "learning_rate": 8.418649349491248e-07, + "loss": 0.852, + "step": 5209 + }, + { + "epoch": 0.8735198574871633, + "grad_norm": 0.607019305229187, + "learning_rate": 8.396853423798867e-07, + "loss": 0.7937, + "step": 5210 + }, + { + "epoch": 0.8736875196479095, + "grad_norm": 0.3749978840351105, + "learning_rate": 8.375084513008736e-07, + "loss": 0.8734, + "step": 5211 + }, + { + "epoch": 0.8738551818086555, + "grad_norm": 0.3694119155406952, + "learning_rate": 8.35334262354075e-07, + "loss": 0.8821, + "step": 5212 + }, + { + "epoch": 0.8740228439694017, + "grad_norm": 0.5530465841293335, + "learning_rate": 8.33162776180686e-07, + "loss": 0.7952, + "step": 5213 + }, + { + "epoch": 0.8741905061301477, + "grad_norm": 0.3724326491355896, + "learning_rate": 8.309939934211053e-07, + "loss": 0.8398, + "step": 5214 + }, + { + "epoch": 0.8743581682908939, + "grad_norm": 0.4362427890300751, + "learning_rate": 8.288279147149369e-07, + "loss": 0.8927, + "step": 5215 + }, + { + "epoch": 0.8745258304516399, + "grad_norm": 0.42852744460105896, + "learning_rate": 8.266645407009788e-07, + "loss": 0.9007, + "step": 5216 + }, + { + "epoch": 0.874693492612386, + "grad_norm": 0.43490931391716003, + "learning_rate": 8.245038720172393e-07, + "loss": 0.8715, + "step": 5217 + }, + { + "epoch": 0.8748611547731321, + "grad_norm": 0.8107596039772034, + "learning_rate": 8.223459093009279e-07, + "loss": 0.8452, + "step": 5218 + }, + { + "epoch": 0.8750288169338782, + "grad_norm": 0.42709115147590637, + "learning_rate": 8.201906531884507e-07, + "loss": 0.8322, + "step": 5219 + }, + { + "epoch": 0.8751964790946243, + "grad_norm": 0.4896807074546814, + "learning_rate": 8.180381043154262e-07, + "loss": 0.9062, + "step": 5220 + }, + { + "epoch": 0.8753641412553704, + "grad_norm": 0.4519088566303253, + "learning_rate": 8.158882633166587e-07, + "loss": 0.7793, + "step": 5221 + }, + { + "epoch": 0.8755318034161165, + "grad_norm": 0.36894258856773376, + "learning_rate": 8.137411308261689e-07, + "loss": 0.8589, + "step": 5222 + }, + { + "epoch": 0.8756994655768626, + "grad_norm": 0.41007664799690247, + "learning_rate": 8.115967074771714e-07, + "loss": 0.8599, + "step": 5223 + }, + { + "epoch": 0.8758671277376087, + "grad_norm": 0.366462767124176, + "learning_rate": 8.094549939020824e-07, + "loss": 0.8493, + "step": 5224 + }, + { + "epoch": 0.8760347898983548, + "grad_norm": 0.3870672583580017, + "learning_rate": 8.073159907325223e-07, + "loss": 0.8376, + "step": 5225 + }, + { + "epoch": 0.876202452059101, + "grad_norm": 0.3918464779853821, + "learning_rate": 8.051796985993077e-07, + "loss": 0.8635, + "step": 5226 + }, + { + "epoch": 0.876370114219847, + "grad_norm": 0.39704617857933044, + "learning_rate": 8.030461181324567e-07, + "loss": 0.8127, + "step": 5227 + }, + { + "epoch": 0.8765377763805932, + "grad_norm": 0.9166821241378784, + "learning_rate": 8.009152499611905e-07, + "loss": 0.892, + "step": 5228 + }, + { + "epoch": 0.8767054385413392, + "grad_norm": 0.6163578629493713, + "learning_rate": 7.987870947139276e-07, + "loss": 0.8442, + "step": 5229 + }, + { + "epoch": 0.8768731007020854, + "grad_norm": 0.8241729140281677, + "learning_rate": 7.966616530182858e-07, + "loss": 0.9117, + "step": 5230 + }, + { + "epoch": 0.8770407628628314, + "grad_norm": 0.7489495277404785, + "learning_rate": 7.945389255010872e-07, + "loss": 0.9063, + "step": 5231 + }, + { + "epoch": 0.8772084250235775, + "grad_norm": 1.046480655670166, + "learning_rate": 7.924189127883475e-07, + "loss": 0.8623, + "step": 5232 + }, + { + "epoch": 0.8773760871843236, + "grad_norm": 0.3621228039264679, + "learning_rate": 7.903016155052867e-07, + "loss": 0.8169, + "step": 5233 + }, + { + "epoch": 0.8775437493450697, + "grad_norm": 0.36051076650619507, + "learning_rate": 7.881870342763243e-07, + "loss": 0.8182, + "step": 5234 + }, + { + "epoch": 0.8777114115058158, + "grad_norm": 0.3524509072303772, + "learning_rate": 7.860751697250702e-07, + "loss": 0.7873, + "step": 5235 + }, + { + "epoch": 0.8778790736665619, + "grad_norm": 0.3441102206707001, + "learning_rate": 7.83966022474345e-07, + "loss": 0.7736, + "step": 5236 + }, + { + "epoch": 0.878046735827308, + "grad_norm": 0.3441102206707001, + "learning_rate": 7.83966022474345e-07, + "loss": 0.859, + "step": 5237 + }, + { + "epoch": 0.8782143979880541, + "grad_norm": 0.5010706186294556, + "learning_rate": 7.818595931461581e-07, + "loss": 0.8187, + "step": 5238 + }, + { + "epoch": 0.8783820601488002, + "grad_norm": 0.40332692861557007, + "learning_rate": 7.797558823617257e-07, + "loss": 0.9172, + "step": 5239 + }, + { + "epoch": 0.8785497223095463, + "grad_norm": 0.476007342338562, + "learning_rate": 7.77654890741456e-07, + "loss": 0.7967, + "step": 5240 + }, + { + "epoch": 0.8787173844702924, + "grad_norm": 0.3826432526111603, + "learning_rate": 7.755566189049557e-07, + "loss": 0.859, + "step": 5241 + }, + { + "epoch": 0.8788850466310385, + "grad_norm": 0.35931214690208435, + "learning_rate": 7.73461067471033e-07, + "loss": 0.7708, + "step": 5242 + }, + { + "epoch": 0.8790527087917845, + "grad_norm": 0.3573903441429138, + "learning_rate": 7.713682370576947e-07, + "loss": 0.9159, + "step": 5243 + }, + { + "epoch": 0.8792203709525307, + "grad_norm": 0.4550726115703583, + "learning_rate": 7.692781282821383e-07, + "loss": 0.8461, + "step": 5244 + }, + { + "epoch": 0.8793880331132767, + "grad_norm": 0.6071157455444336, + "learning_rate": 7.671907417607638e-07, + "loss": 0.8955, + "step": 5245 + }, + { + "epoch": 0.8795556952740229, + "grad_norm": 0.4022810757160187, + "learning_rate": 7.65106078109168e-07, + "loss": 0.8044, + "step": 5246 + }, + { + "epoch": 0.8797233574347689, + "grad_norm": 0.37888088822364807, + "learning_rate": 7.630241379421432e-07, + "loss": 0.8642, + "step": 5247 + }, + { + "epoch": 0.8798910195955151, + "grad_norm": 0.3679821789264679, + "learning_rate": 7.609449218736808e-07, + "loss": 0.8851, + "step": 5248 + }, + { + "epoch": 0.8800586817562611, + "grad_norm": 0.47906655073165894, + "learning_rate": 7.588684305169669e-07, + "loss": 0.9026, + "step": 5249 + }, + { + "epoch": 0.8802263439170073, + "grad_norm": 0.4536413848400116, + "learning_rate": 7.567946644843827e-07, + "loss": 0.9075, + "step": 5250 + }, + { + "epoch": 0.8803940060777533, + "grad_norm": 0.3367442190647125, + "learning_rate": 7.5472362438751e-07, + "loss": 0.7654, + "step": 5251 + }, + { + "epoch": 0.8805616682384995, + "grad_norm": 0.43579649925231934, + "learning_rate": 7.526553108371215e-07, + "loss": 0.8099, + "step": 5252 + }, + { + "epoch": 0.8807293303992455, + "grad_norm": 0.8711832165718079, + "learning_rate": 7.505897244431914e-07, + "loss": 0.9072, + "step": 5253 + }, + { + "epoch": 0.8808969925599917, + "grad_norm": 0.38806793093681335, + "learning_rate": 7.485268658148859e-07, + "loss": 0.7709, + "step": 5254 + }, + { + "epoch": 0.8810646547207377, + "grad_norm": 0.4234929382801056, + "learning_rate": 7.464667355605648e-07, + "loss": 0.8378, + "step": 5255 + }, + { + "epoch": 0.8812323168814838, + "grad_norm": 0.5764008164405823, + "learning_rate": 7.4440933428779e-07, + "loss": 0.7818, + "step": 5256 + }, + { + "epoch": 0.8813999790422299, + "grad_norm": 0.4254293143749237, + "learning_rate": 7.423546626033129e-07, + "loss": 0.837, + "step": 5257 + }, + { + "epoch": 0.881567641202976, + "grad_norm": 0.336610347032547, + "learning_rate": 7.403027211130786e-07, + "loss": 0.7539, + "step": 5258 + }, + { + "epoch": 0.8817353033637221, + "grad_norm": 0.377946674823761, + "learning_rate": 7.382535104222366e-07, + "loss": 0.8741, + "step": 5259 + }, + { + "epoch": 0.8819029655244682, + "grad_norm": 0.42889639735221863, + "learning_rate": 7.362070311351177e-07, + "loss": 0.8826, + "step": 5260 + }, + { + "epoch": 0.8820706276852143, + "grad_norm": 0.5174009203910828, + "learning_rate": 7.341632838552582e-07, + "loss": 0.8708, + "step": 5261 + }, + { + "epoch": 0.8822382898459604, + "grad_norm": 0.4477638900279999, + "learning_rate": 7.321222691853858e-07, + "loss": 0.872, + "step": 5262 + }, + { + "epoch": 0.8824059520067065, + "grad_norm": 0.3257504999637604, + "learning_rate": 7.300839877274157e-07, + "loss": 0.8529, + "step": 5263 + }, + { + "epoch": 0.8825736141674526, + "grad_norm": 0.3661295175552368, + "learning_rate": 7.28048440082465e-07, + "loss": 0.8091, + "step": 5264 + }, + { + "epoch": 0.8827412763281987, + "grad_norm": 0.49933600425720215, + "learning_rate": 7.260156268508444e-07, + "loss": 0.8484, + "step": 5265 + }, + { + "epoch": 0.8829089384889448, + "grad_norm": 0.33709481358528137, + "learning_rate": 7.239855486320513e-07, + "loss": 0.8874, + "step": 5266 + }, + { + "epoch": 0.8830766006496908, + "grad_norm": 0.5059892535209656, + "learning_rate": 7.219582060247854e-07, + "loss": 0.8608, + "step": 5267 + }, + { + "epoch": 0.883244262810437, + "grad_norm": 0.43796518445014954, + "learning_rate": 7.199335996269319e-07, + "loss": 0.9711, + "step": 5268 + }, + { + "epoch": 0.883411924971183, + "grad_norm": 0.47821682691574097, + "learning_rate": 7.17911730035572e-07, + "loss": 0.8459, + "step": 5269 + }, + { + "epoch": 0.8835795871319292, + "grad_norm": 0.34628474712371826, + "learning_rate": 7.158925978469833e-07, + "loss": 0.8287, + "step": 5270 + }, + { + "epoch": 0.8837472492926752, + "grad_norm": 0.429559588432312, + "learning_rate": 7.138762036566305e-07, + "loss": 0.8746, + "step": 5271 + }, + { + "epoch": 0.8839149114534214, + "grad_norm": 0.35698437690734863, + "learning_rate": 7.118625480591712e-07, + "loss": 0.8658, + "step": 5272 + }, + { + "epoch": 0.8840825736141674, + "grad_norm": 0.37137287855148315, + "learning_rate": 7.098516316484628e-07, + "loss": 0.8086, + "step": 5273 + }, + { + "epoch": 0.8842502357749136, + "grad_norm": 0.6553245186805725, + "learning_rate": 7.07843455017544e-07, + "loss": 0.8455, + "step": 5274 + }, + { + "epoch": 0.8844178979356596, + "grad_norm": 0.4681013226509094, + "learning_rate": 7.058380187586544e-07, + "loss": 0.7681, + "step": 5275 + }, + { + "epoch": 0.8845855600964058, + "grad_norm": 0.3756442666053772, + "learning_rate": 7.038353234632223e-07, + "loss": 0.7792, + "step": 5276 + }, + { + "epoch": 0.8847532222571518, + "grad_norm": 0.49385350942611694, + "learning_rate": 7.018353697218638e-07, + "loss": 0.8486, + "step": 5277 + }, + { + "epoch": 0.884920884417898, + "grad_norm": 0.6206707954406738, + "learning_rate": 6.998381581243929e-07, + "loss": 0.8994, + "step": 5278 + }, + { + "epoch": 0.885088546578644, + "grad_norm": 0.31013917922973633, + "learning_rate": 6.978436892598106e-07, + "loss": 0.8784, + "step": 5279 + }, + { + "epoch": 0.8852562087393901, + "grad_norm": 0.34381526708602905, + "learning_rate": 6.95851963716313e-07, + "loss": 0.7866, + "step": 5280 + }, + { + "epoch": 0.8854238709001362, + "grad_norm": 0.45185163617134094, + "learning_rate": 6.938629820812826e-07, + "loss": 0.7698, + "step": 5281 + }, + { + "epoch": 0.8855915330608823, + "grad_norm": 0.4260341227054596, + "learning_rate": 6.918767449412933e-07, + "loss": 0.8308, + "step": 5282 + }, + { + "epoch": 0.8857591952216284, + "grad_norm": 0.8337491154670715, + "learning_rate": 6.898932528821123e-07, + "loss": 0.9048, + "step": 5283 + }, + { + "epoch": 0.8859268573823745, + "grad_norm": 0.43806082010269165, + "learning_rate": 6.879125064886993e-07, + "loss": 0.8681, + "step": 5284 + }, + { + "epoch": 0.8860945195431206, + "grad_norm": 0.3748376667499542, + "learning_rate": 6.859345063451961e-07, + "loss": 0.8701, + "step": 5285 + }, + { + "epoch": 0.8862621817038667, + "grad_norm": 0.3831605613231659, + "learning_rate": 6.839592530349426e-07, + "loss": 0.8497, + "step": 5286 + }, + { + "epoch": 0.8864298438646128, + "grad_norm": 0.41587305068969727, + "learning_rate": 6.819867471404628e-07, + "loss": 0.7885, + "step": 5287 + }, + { + "epoch": 0.8865975060253589, + "grad_norm": 0.43317797780036926, + "learning_rate": 6.800169892434739e-07, + "loss": 0.7672, + "step": 5288 + }, + { + "epoch": 0.886765168186105, + "grad_norm": 0.399836003780365, + "learning_rate": 6.780499799248852e-07, + "loss": 0.7892, + "step": 5289 + }, + { + "epoch": 0.8869328303468511, + "grad_norm": 0.3666892349720001, + "learning_rate": 6.760857197647863e-07, + "loss": 0.8582, + "step": 5290 + }, + { + "epoch": 0.8871004925075971, + "grad_norm": 0.681106448173523, + "learning_rate": 6.74124209342466e-07, + "loss": 0.8437, + "step": 5291 + }, + { + "epoch": 0.8872681546683433, + "grad_norm": 0.3994339108467102, + "learning_rate": 6.721654492363972e-07, + "loss": 0.8039, + "step": 5292 + }, + { + "epoch": 0.8874358168290893, + "grad_norm": 0.4830002188682556, + "learning_rate": 6.702094400242421e-07, + "loss": 0.7977, + "step": 5293 + }, + { + "epoch": 0.8876034789898355, + "grad_norm": 0.6147891879081726, + "learning_rate": 6.682561822828526e-07, + "loss": 0.8956, + "step": 5294 + }, + { + "epoch": 0.8877711411505815, + "grad_norm": 0.5718827843666077, + "learning_rate": 6.663056765882692e-07, + "loss": 0.8331, + "step": 5295 + }, + { + "epoch": 0.8879388033113277, + "grad_norm": 0.8253036737442017, + "learning_rate": 6.64357923515716e-07, + "loss": 0.8295, + "step": 5296 + }, + { + "epoch": 0.8881064654720737, + "grad_norm": 0.36979398131370544, + "learning_rate": 6.624129236396159e-07, + "loss": 0.7818, + "step": 5297 + }, + { + "epoch": 0.8882741276328199, + "grad_norm": 0.36641526222229004, + "learning_rate": 6.604706775335679e-07, + "loss": 0.7572, + "step": 5298 + }, + { + "epoch": 0.8884417897935659, + "grad_norm": 0.4561346471309662, + "learning_rate": 6.58531185770368e-07, + "loss": 0.84, + "step": 5299 + }, + { + "epoch": 0.8886094519543121, + "grad_norm": 0.3623506426811218, + "learning_rate": 6.565944489219955e-07, + "loss": 0.884, + "step": 5300 + }, + { + "epoch": 0.8887771141150581, + "grad_norm": 0.4001149833202362, + "learning_rate": 6.546604675596169e-07, + "loss": 0.8588, + "step": 5301 + }, + { + "epoch": 0.8889447762758043, + "grad_norm": 0.38912656903266907, + "learning_rate": 6.52729242253588e-07, + "loss": 0.8358, + "step": 5302 + }, + { + "epoch": 0.8891124384365503, + "grad_norm": 0.4054897129535675, + "learning_rate": 6.508007735734522e-07, + "loss": 0.7658, + "step": 5303 + }, + { + "epoch": 0.8892801005972965, + "grad_norm": 0.4189501106739044, + "learning_rate": 6.488750620879358e-07, + "loss": 0.883, + "step": 5304 + }, + { + "epoch": 0.8894477627580425, + "grad_norm": 0.36771664023399353, + "learning_rate": 6.469521083649588e-07, + "loss": 0.8027, + "step": 5305 + }, + { + "epoch": 0.8896154249187886, + "grad_norm": 0.37720581889152527, + "learning_rate": 6.450319129716198e-07, + "loss": 0.7863, + "step": 5306 + }, + { + "epoch": 0.8897830870795347, + "grad_norm": 0.35849684476852417, + "learning_rate": 6.431144764742104e-07, + "loss": 0.825, + "step": 5307 + }, + { + "epoch": 0.8899507492402808, + "grad_norm": 0.7659775614738464, + "learning_rate": 6.411997994382102e-07, + "loss": 0.8178, + "step": 5308 + }, + { + "epoch": 0.8901184114010269, + "grad_norm": 0.41936665773391724, + "learning_rate": 6.392878824282744e-07, + "loss": 0.8437, + "step": 5309 + }, + { + "epoch": 0.890286073561773, + "grad_norm": 0.5967978835105896, + "learning_rate": 6.373787260082531e-07, + "loss": 0.864, + "step": 5310 + }, + { + "epoch": 0.8904537357225191, + "grad_norm": 0.4905378520488739, + "learning_rate": 6.354723307411837e-07, + "loss": 0.7812, + "step": 5311 + }, + { + "epoch": 0.8906213978832652, + "grad_norm": 0.34936320781707764, + "learning_rate": 6.33568697189283e-07, + "loss": 0.8333, + "step": 5312 + }, + { + "epoch": 0.8907890600440114, + "grad_norm": 0.760906994342804, + "learning_rate": 6.316678259139542e-07, + "loss": 0.8622, + "step": 5313 + }, + { + "epoch": 0.8909567222047574, + "grad_norm": 0.44719940423965454, + "learning_rate": 6.29769717475791e-07, + "loss": 0.8796, + "step": 5314 + }, + { + "epoch": 0.8911243843655036, + "grad_norm": 0.4683433771133423, + "learning_rate": 6.278743724345682e-07, + "loss": 0.8847, + "step": 5315 + }, + { + "epoch": 0.8912920465262496, + "grad_norm": 0.40432867407798767, + "learning_rate": 6.259817913492461e-07, + "loss": 0.8508, + "step": 5316 + }, + { + "epoch": 0.8914597086869958, + "grad_norm": 0.4725228548049927, + "learning_rate": 6.240919747779705e-07, + "loss": 0.8185, + "step": 5317 + }, + { + "epoch": 0.8916273708477418, + "grad_norm": 0.3801216185092926, + "learning_rate": 6.222049232780714e-07, + "loss": 0.7974, + "step": 5318 + }, + { + "epoch": 0.891795033008488, + "grad_norm": 0.3568783402442932, + "learning_rate": 6.203206374060644e-07, + "loss": 0.8628, + "step": 5319 + }, + { + "epoch": 0.891962695169234, + "grad_norm": 0.33219611644744873, + "learning_rate": 6.184391177176474e-07, + "loss": 0.866, + "step": 5320 + }, + { + "epoch": 0.8921303573299801, + "grad_norm": 0.4283650517463684, + "learning_rate": 6.165603647677054e-07, + "loss": 0.7798, + "step": 5321 + }, + { + "epoch": 0.8922980194907262, + "grad_norm": 0.43948283791542053, + "learning_rate": 6.146843791103052e-07, + "loss": 0.7352, + "step": 5322 + }, + { + "epoch": 0.8924656816514723, + "grad_norm": 0.39937087893486023, + "learning_rate": 6.128111612986975e-07, + "loss": 0.8576, + "step": 5323 + }, + { + "epoch": 0.8926333438122184, + "grad_norm": 0.4143848717212677, + "learning_rate": 6.109407118853195e-07, + "loss": 0.8693, + "step": 5324 + }, + { + "epoch": 0.8928010059729645, + "grad_norm": 1.0609921216964722, + "learning_rate": 6.090730314217863e-07, + "loss": 0.872, + "step": 5325 + }, + { + "epoch": 0.8929686681337106, + "grad_norm": 0.4229665994644165, + "learning_rate": 6.072081204589042e-07, + "loss": 0.828, + "step": 5326 + }, + { + "epoch": 0.8931363302944567, + "grad_norm": 0.3355676233768463, + "learning_rate": 6.053459795466565e-07, + "loss": 0.8271, + "step": 5327 + }, + { + "epoch": 0.8933039924552028, + "grad_norm": 0.48312902450561523, + "learning_rate": 6.034866092342107e-07, + "loss": 0.7578, + "step": 5328 + }, + { + "epoch": 0.8934716546159489, + "grad_norm": 0.46982505917549133, + "learning_rate": 6.016300100699191e-07, + "loss": 0.8265, + "step": 5329 + }, + { + "epoch": 0.893639316776695, + "grad_norm": 0.40215837955474854, + "learning_rate": 5.997761826013182e-07, + "loss": 0.8408, + "step": 5330 + }, + { + "epoch": 0.8938069789374411, + "grad_norm": 0.48121073842048645, + "learning_rate": 5.979251273751197e-07, + "loss": 0.8354, + "step": 5331 + }, + { + "epoch": 0.8939746410981871, + "grad_norm": 0.3733910918235779, + "learning_rate": 5.960768449372246e-07, + "loss": 0.8275, + "step": 5332 + }, + { + "epoch": 0.8941423032589333, + "grad_norm": 0.5000764727592468, + "learning_rate": 5.942313358327167e-07, + "loss": 0.7803, + "step": 5333 + }, + { + "epoch": 0.8943099654196793, + "grad_norm": 0.4887655973434448, + "learning_rate": 5.923886006058566e-07, + "loss": 0.8851, + "step": 5334 + }, + { + "epoch": 0.8944776275804255, + "grad_norm": 0.3644387722015381, + "learning_rate": 5.905486398000937e-07, + "loss": 0.8665, + "step": 5335 + }, + { + "epoch": 0.8946452897411715, + "grad_norm": 0.412713885307312, + "learning_rate": 5.887114539580497e-07, + "loss": 0.7574, + "step": 5336 + }, + { + "epoch": 0.8948129519019177, + "grad_norm": 0.3497019112110138, + "learning_rate": 5.868770436215366e-07, + "loss": 0.8163, + "step": 5337 + }, + { + "epoch": 0.8949806140626637, + "grad_norm": 0.35792189836502075, + "learning_rate": 5.85045409331545e-07, + "loss": 0.8441, + "step": 5338 + }, + { + "epoch": 0.8951482762234099, + "grad_norm": 0.4269263446331024, + "learning_rate": 5.83216551628244e-07, + "loss": 0.8261, + "step": 5339 + }, + { + "epoch": 0.8953159383841559, + "grad_norm": 0.4859539270401001, + "learning_rate": 5.813904710509899e-07, + "loss": 0.8215, + "step": 5340 + }, + { + "epoch": 0.8954836005449021, + "grad_norm": 0.36638912558555603, + "learning_rate": 5.795671681383152e-07, + "loss": 0.8916, + "step": 5341 + }, + { + "epoch": 0.8956512627056481, + "grad_norm": 0.4190085530281067, + "learning_rate": 5.777466434279322e-07, + "loss": 0.8154, + "step": 5342 + }, + { + "epoch": 0.8958189248663943, + "grad_norm": 0.41331014037132263, + "learning_rate": 5.7592889745674e-07, + "loss": 0.8592, + "step": 5343 + }, + { + "epoch": 0.8959865870271403, + "grad_norm": 0.36141493916511536, + "learning_rate": 5.741139307608124e-07, + "loss": 0.8003, + "step": 5344 + }, + { + "epoch": 0.8961542491878864, + "grad_norm": 0.5100393891334534, + "learning_rate": 5.723017438754042e-07, + "loss": 0.8431, + "step": 5345 + }, + { + "epoch": 0.8963219113486325, + "grad_norm": 0.3829497694969177, + "learning_rate": 5.704923373349536e-07, + "loss": 0.8047, + "step": 5346 + }, + { + "epoch": 0.8964895735093786, + "grad_norm": 0.5301305651664734, + "learning_rate": 5.68685711673076e-07, + "loss": 0.7256, + "step": 5347 + }, + { + "epoch": 0.8966572356701247, + "grad_norm": 0.3523816168308258, + "learning_rate": 5.668818674225684e-07, + "loss": 0.8074, + "step": 5348 + }, + { + "epoch": 0.8968248978308708, + "grad_norm": 0.41455909609794617, + "learning_rate": 5.650808051154089e-07, + "loss": 0.8822, + "step": 5349 + }, + { + "epoch": 0.8969925599916169, + "grad_norm": 0.4645654559135437, + "learning_rate": 5.632825252827479e-07, + "loss": 0.934, + "step": 5350 + }, + { + "epoch": 0.897160222152363, + "grad_norm": 0.5317776799201965, + "learning_rate": 5.614870284549256e-07, + "loss": 0.8398, + "step": 5351 + }, + { + "epoch": 0.8973278843131091, + "grad_norm": 0.5853843092918396, + "learning_rate": 5.59694315161452e-07, + "loss": 0.8806, + "step": 5352 + }, + { + "epoch": 0.8974955464738552, + "grad_norm": 0.39755651354789734, + "learning_rate": 5.57904385931024e-07, + "loss": 0.8073, + "step": 5353 + }, + { + "epoch": 0.8976632086346013, + "grad_norm": 0.3885393440723419, + "learning_rate": 5.561172412915117e-07, + "loss": 0.8263, + "step": 5354 + }, + { + "epoch": 0.8978308707953474, + "grad_norm": 0.6345336437225342, + "learning_rate": 5.543328817699667e-07, + "loss": 0.8009, + "step": 5355 + }, + { + "epoch": 0.8979985329560934, + "grad_norm": 0.5248864889144897, + "learning_rate": 5.525513078926171e-07, + "loss": 0.8164, + "step": 5356 + }, + { + "epoch": 0.8981661951168396, + "grad_norm": 0.49112460017204285, + "learning_rate": 5.507725201848757e-07, + "loss": 0.7499, + "step": 5357 + }, + { + "epoch": 0.8983338572775856, + "grad_norm": 0.4366603493690491, + "learning_rate": 5.489965191713254e-07, + "loss": 0.8424, + "step": 5358 + }, + { + "epoch": 0.8985015194383318, + "grad_norm": 0.4593455195426941, + "learning_rate": 5.472233053757303e-07, + "loss": 0.8159, + "step": 5359 + }, + { + "epoch": 0.8986691815990778, + "grad_norm": 0.4607357680797577, + "learning_rate": 5.454528793210356e-07, + "loss": 0.8151, + "step": 5360 + }, + { + "epoch": 0.898836843759824, + "grad_norm": 0.4741382598876953, + "learning_rate": 5.436852415293592e-07, + "loss": 0.8323, + "step": 5361 + }, + { + "epoch": 0.89900450592057, + "grad_norm": 0.5529037714004517, + "learning_rate": 5.419203925220029e-07, + "loss": 0.9415, + "step": 5362 + }, + { + "epoch": 0.8991721680813162, + "grad_norm": 0.5256397128105164, + "learning_rate": 5.401583328194404e-07, + "loss": 0.8095, + "step": 5363 + }, + { + "epoch": 0.8993398302420622, + "grad_norm": 0.3243965804576874, + "learning_rate": 5.383990629413227e-07, + "loss": 0.8096, + "step": 5364 + }, + { + "epoch": 0.8995074924028084, + "grad_norm": 1.1402196884155273, + "learning_rate": 5.366425834064848e-07, + "loss": 0.7679, + "step": 5365 + }, + { + "epoch": 0.8996751545635544, + "grad_norm": 0.37329888343811035, + "learning_rate": 5.348888947329312e-07, + "loss": 0.8054, + "step": 5366 + }, + { + "epoch": 0.8998428167243006, + "grad_norm": 0.4582318961620331, + "learning_rate": 5.33137997437847e-07, + "loss": 0.8313, + "step": 5367 + }, + { + "epoch": 0.9000104788850466, + "grad_norm": 0.39775338768959045, + "learning_rate": 5.313898920375948e-07, + "loss": 0.8463, + "step": 5368 + }, + { + "epoch": 0.9001781410457927, + "grad_norm": 0.4152122437953949, + "learning_rate": 5.296445790477101e-07, + "loss": 0.8729, + "step": 5369 + }, + { + "epoch": 0.9003458032065388, + "grad_norm": 0.7189455628395081, + "learning_rate": 5.279020589829099e-07, + "loss": 0.8718, + "step": 5370 + }, + { + "epoch": 0.9005134653672849, + "grad_norm": 0.3881506621837616, + "learning_rate": 5.26162332357083e-07, + "loss": 0.7515, + "step": 5371 + }, + { + "epoch": 0.900681127528031, + "grad_norm": 0.407922625541687, + "learning_rate": 5.244253996832971e-07, + "loss": 0.9079, + "step": 5372 + }, + { + "epoch": 0.9008487896887771, + "grad_norm": 0.3899659812450409, + "learning_rate": 5.226912614737956e-07, + "loss": 0.8661, + "step": 5373 + }, + { + "epoch": 0.9010164518495232, + "grad_norm": 0.42218416929244995, + "learning_rate": 5.20959918239996e-07, + "loss": 0.7457, + "step": 5374 + }, + { + "epoch": 0.9011841140102693, + "grad_norm": 0.5291997790336609, + "learning_rate": 5.192313704924933e-07, + "loss": 0.8461, + "step": 5375 + }, + { + "epoch": 0.9013517761710154, + "grad_norm": 0.43528470396995544, + "learning_rate": 5.175056187410609e-07, + "loss": 0.8246, + "step": 5376 + }, + { + "epoch": 0.9015194383317615, + "grad_norm": 0.42045772075653076, + "learning_rate": 5.157826634946395e-07, + "loss": 0.8547, + "step": 5377 + }, + { + "epoch": 0.9016871004925076, + "grad_norm": 0.40543699264526367, + "learning_rate": 5.140625052613501e-07, + "loss": 0.8254, + "step": 5378 + }, + { + "epoch": 0.9018547626532537, + "grad_norm": 0.3896552622318268, + "learning_rate": 5.12345144548494e-07, + "loss": 0.8132, + "step": 5379 + }, + { + "epoch": 0.9020224248139997, + "grad_norm": 0.47298580408096313, + "learning_rate": 5.106305818625357e-07, + "loss": 0.8961, + "step": 5380 + }, + { + "epoch": 0.9021900869747459, + "grad_norm": 0.6200173497200012, + "learning_rate": 5.089188177091254e-07, + "loss": 0.921, + "step": 5381 + }, + { + "epoch": 0.9023577491354919, + "grad_norm": 0.5027073621749878, + "learning_rate": 5.072098525930825e-07, + "loss": 0.774, + "step": 5382 + }, + { + "epoch": 0.9025254112962381, + "grad_norm": 0.49361899495124817, + "learning_rate": 5.055036870184005e-07, + "loss": 0.8256, + "step": 5383 + }, + { + "epoch": 0.9026930734569841, + "grad_norm": 0.390128493309021, + "learning_rate": 5.038003214882503e-07, + "loss": 0.8341, + "step": 5384 + }, + { + "epoch": 0.9028607356177303, + "grad_norm": 0.550040602684021, + "learning_rate": 5.020997565049756e-07, + "loss": 0.7736, + "step": 5385 + }, + { + "epoch": 0.9030283977784763, + "grad_norm": 0.3496764898300171, + "learning_rate": 5.004019925700921e-07, + "loss": 0.8466, + "step": 5386 + }, + { + "epoch": 0.9031960599392225, + "grad_norm": 0.3841930031776428, + "learning_rate": 4.987070301842934e-07, + "loss": 0.8168, + "step": 5387 + }, + { + "epoch": 0.9033637220999685, + "grad_norm": 0.4360392987728119, + "learning_rate": 4.97014869847443e-07, + "loss": 0.9347, + "step": 5388 + }, + { + "epoch": 0.9035313842607147, + "grad_norm": 1.048956036567688, + "learning_rate": 4.953255120585832e-07, + "loss": 0.8688, + "step": 5389 + }, + { + "epoch": 0.9036990464214607, + "grad_norm": 0.3388780951499939, + "learning_rate": 4.936389573159239e-07, + "loss": 0.8564, + "step": 5390 + }, + { + "epoch": 0.9038667085822069, + "grad_norm": 0.5225685238838196, + "learning_rate": 4.919552061168509e-07, + "loss": 0.7863, + "step": 5391 + }, + { + "epoch": 0.9040343707429529, + "grad_norm": 0.5430662631988525, + "learning_rate": 4.902742589579246e-07, + "loss": 0.8832, + "step": 5392 + }, + { + "epoch": 0.904202032903699, + "grad_norm": 0.44941407442092896, + "learning_rate": 4.885961163348762e-07, + "loss": 0.776, + "step": 5393 + }, + { + "epoch": 0.9043696950644451, + "grad_norm": 0.46763765811920166, + "learning_rate": 4.869207787426111e-07, + "loss": 0.938, + "step": 5394 + }, + { + "epoch": 0.9045373572251912, + "grad_norm": 0.3220028579235077, + "learning_rate": 4.852482466752074e-07, + "loss": 0.7516, + "step": 5395 + }, + { + "epoch": 0.9047050193859373, + "grad_norm": 1.0635429620742798, + "learning_rate": 4.835785206259147e-07, + "loss": 0.8318, + "step": 5396 + }, + { + "epoch": 0.9048726815466834, + "grad_norm": 0.43439677357673645, + "learning_rate": 4.81911601087155e-07, + "loss": 0.8004, + "step": 5397 + }, + { + "epoch": 0.9050403437074296, + "grad_norm": 0.4261639714241028, + "learning_rate": 4.80247488550527e-07, + "loss": 0.7668, + "step": 5398 + }, + { + "epoch": 0.9052080058681756, + "grad_norm": 0.46629682183265686, + "learning_rate": 4.785861835067962e-07, + "loss": 0.8571, + "step": 5399 + }, + { + "epoch": 0.9053756680289218, + "grad_norm": 0.3818049728870392, + "learning_rate": 4.769276864459016e-07, + "loss": 0.8136, + "step": 5400 + }, + { + "epoch": 0.9055433301896678, + "grad_norm": 0.39135274291038513, + "learning_rate": 4.752719978569531e-07, + "loss": 0.7031, + "step": 5401 + }, + { + "epoch": 0.905710992350414, + "grad_norm": 0.40635475516319275, + "learning_rate": 4.7361911822823457e-07, + "loss": 0.8831, + "step": 5402 + }, + { + "epoch": 0.90587865451116, + "grad_norm": 0.5607383847236633, + "learning_rate": 4.71969048047205e-07, + "loss": 0.8562, + "step": 5403 + }, + { + "epoch": 0.9060463166719062, + "grad_norm": 0.7370794415473938, + "learning_rate": 4.703217878004851e-07, + "loss": 0.8472, + "step": 5404 + }, + { + "epoch": 0.9062139788326522, + "grad_norm": 0.5216053128242493, + "learning_rate": 4.686773379738729e-07, + "loss": 0.8061, + "step": 5405 + }, + { + "epoch": 0.9063816409933984, + "grad_norm": 0.43987199664115906, + "learning_rate": 4.670356990523406e-07, + "loss": 0.803, + "step": 5406 + }, + { + "epoch": 0.9065493031541444, + "grad_norm": 0.4011021554470062, + "learning_rate": 4.65396871520023e-07, + "loss": 0.8302, + "step": 5407 + }, + { + "epoch": 0.9067169653148905, + "grad_norm": 0.40155351161956787, + "learning_rate": 4.637608558602358e-07, + "loss": 0.8402, + "step": 5408 + }, + { + "epoch": 0.9068846274756366, + "grad_norm": 0.4337129592895508, + "learning_rate": 4.6212765255545745e-07, + "loss": 0.8696, + "step": 5409 + }, + { + "epoch": 0.9070522896363827, + "grad_norm": 0.390887975692749, + "learning_rate": 4.6049726208733825e-07, + "loss": 0.7668, + "step": 5410 + }, + { + "epoch": 0.9072199517971288, + "grad_norm": 0.3842809498310089, + "learning_rate": 4.588696849367036e-07, + "loss": 0.8776, + "step": 5411 + }, + { + "epoch": 0.9073876139578749, + "grad_norm": 0.43689432740211487, + "learning_rate": 4.5724492158354397e-07, + "loss": 0.8334, + "step": 5412 + }, + { + "epoch": 0.907555276118621, + "grad_norm": 0.4424873888492584, + "learning_rate": 4.556229725070249e-07, + "loss": 0.8537, + "step": 5413 + }, + { + "epoch": 0.9077229382793671, + "grad_norm": 0.6077720522880554, + "learning_rate": 4.540038381854772e-07, + "loss": 0.8586, + "step": 5414 + }, + { + "epoch": 0.9078906004401132, + "grad_norm": 0.4896509349346161, + "learning_rate": 4.5238751909640224e-07, + "loss": 0.8448, + "step": 5415 + }, + { + "epoch": 0.9080582626008593, + "grad_norm": 0.440701425075531, + "learning_rate": 4.5077401571647527e-07, + "loss": 0.8318, + "step": 5416 + }, + { + "epoch": 0.9082259247616054, + "grad_norm": 0.5645205974578857, + "learning_rate": 4.4916332852153797e-07, + "loss": 0.948, + "step": 5417 + }, + { + "epoch": 0.9083935869223515, + "grad_norm": 0.4756021499633789, + "learning_rate": 4.4755545798659926e-07, + "loss": 0.8636, + "step": 5418 + }, + { + "epoch": 0.9085612490830975, + "grad_norm": 0.5430362820625305, + "learning_rate": 4.459504045858443e-07, + "loss": 0.8831, + "step": 5419 + }, + { + "epoch": 0.9087289112438437, + "grad_norm": 0.3706998825073242, + "learning_rate": 4.4434816879261987e-07, + "loss": 0.7866, + "step": 5420 + }, + { + "epoch": 0.9088965734045897, + "grad_norm": 0.3644903302192688, + "learning_rate": 4.427487510794459e-07, + "loss": 0.8777, + "step": 5421 + }, + { + "epoch": 0.9090642355653359, + "grad_norm": 0.38833266496658325, + "learning_rate": 4.4115215191801285e-07, + "loss": 0.8398, + "step": 5422 + }, + { + "epoch": 0.9092318977260819, + "grad_norm": 0.46523725986480713, + "learning_rate": 4.3955837177917403e-07, + "loss": 0.8199, + "step": 5423 + }, + { + "epoch": 0.9093995598868281, + "grad_norm": 0.7484700083732605, + "learning_rate": 4.379674111329557e-07, + "loss": 0.81, + "step": 5424 + }, + { + "epoch": 0.9095672220475741, + "grad_norm": 0.4971903860569, + "learning_rate": 4.3637927044855476e-07, + "loss": 0.901, + "step": 5425 + }, + { + "epoch": 0.9097348842083203, + "grad_norm": 0.4184390604496002, + "learning_rate": 4.3479395019433104e-07, + "loss": 0.8638, + "step": 5426 + }, + { + "epoch": 0.9099025463690663, + "grad_norm": 0.6759586334228516, + "learning_rate": 4.332114508378138e-07, + "loss": 0.768, + "step": 5427 + }, + { + "epoch": 0.9100702085298125, + "grad_norm": 0.4398367404937744, + "learning_rate": 4.3163177284570536e-07, + "loss": 0.8337, + "step": 5428 + }, + { + "epoch": 0.9102378706905585, + "grad_norm": 0.6298412084579468, + "learning_rate": 4.300549166838697e-07, + "loss": 0.8829, + "step": 5429 + }, + { + "epoch": 0.9104055328513047, + "grad_norm": 0.5195366144180298, + "learning_rate": 4.284808828173426e-07, + "loss": 0.8517, + "step": 5430 + }, + { + "epoch": 0.9105731950120507, + "grad_norm": 0.44466158747673035, + "learning_rate": 4.2690967171032604e-07, + "loss": 0.8467, + "step": 5431 + }, + { + "epoch": 0.9107408571727968, + "grad_norm": 0.4637994170188904, + "learning_rate": 4.253412838261872e-07, + "loss": 0.8496, + "step": 5432 + }, + { + "epoch": 0.9109085193335429, + "grad_norm": 0.42862823605537415, + "learning_rate": 4.237757196274661e-07, + "loss": 0.7244, + "step": 5433 + }, + { + "epoch": 0.911076181494289, + "grad_norm": 0.479573518037796, + "learning_rate": 4.222129795758645e-07, + "loss": 0.9259, + "step": 5434 + }, + { + "epoch": 0.9112438436550351, + "grad_norm": 0.47090259194374084, + "learning_rate": 4.2065306413225704e-07, + "loss": 0.9028, + "step": 5435 + }, + { + "epoch": 0.9114115058157812, + "grad_norm": 0.4738175570964813, + "learning_rate": 4.190959737566802e-07, + "loss": 0.8315, + "step": 5436 + }, + { + "epoch": 0.9115791679765273, + "grad_norm": 0.6840828657150269, + "learning_rate": 4.1754170890833777e-07, + "loss": 0.8972, + "step": 5437 + }, + { + "epoch": 0.9117468301372734, + "grad_norm": 0.4690623879432678, + "learning_rate": 4.1599027004560535e-07, + "loss": 0.8862, + "step": 5438 + }, + { + "epoch": 0.9119144922980195, + "grad_norm": 0.4256584346294403, + "learning_rate": 4.144416576260168e-07, + "loss": 0.8003, + "step": 5439 + }, + { + "epoch": 0.9120821544587656, + "grad_norm": 0.4423993229866028, + "learning_rate": 4.1289587210628146e-07, + "loss": 0.8348, + "step": 5440 + }, + { + "epoch": 0.9122498166195117, + "grad_norm": 1.1761881113052368, + "learning_rate": 4.1135291394226785e-07, + "loss": 0.8472, + "step": 5441 + }, + { + "epoch": 0.9124174787802578, + "grad_norm": 0.41688793897628784, + "learning_rate": 4.0981278358901424e-07, + "loss": 0.891, + "step": 5442 + }, + { + "epoch": 0.9125851409410038, + "grad_norm": 0.43712949752807617, + "learning_rate": 4.082754815007239e-07, + "loss": 0.8237, + "step": 5443 + }, + { + "epoch": 0.91275280310175, + "grad_norm": 0.4862794578075409, + "learning_rate": 4.0674100813076765e-07, + "loss": 0.8626, + "step": 5444 + }, + { + "epoch": 0.912920465262496, + "grad_norm": 0.42175430059432983, + "learning_rate": 4.052093639316801e-07, + "loss": 0.7836, + "step": 5445 + }, + { + "epoch": 0.9130881274232422, + "grad_norm": 0.7146934866905212, + "learning_rate": 4.0368054935516097e-07, + "loss": 0.8276, + "step": 5446 + }, + { + "epoch": 0.9132557895839882, + "grad_norm": 0.5035149455070496, + "learning_rate": 4.0215456485207636e-07, + "loss": 0.8425, + "step": 5447 + }, + { + "epoch": 0.9134234517447344, + "grad_norm": 0.48977479338645935, + "learning_rate": 4.006314108724585e-07, + "loss": 0.8014, + "step": 5448 + }, + { + "epoch": 0.9135911139054804, + "grad_norm": 0.3573587238788605, + "learning_rate": 3.99111087865508e-07, + "loss": 0.8272, + "step": 5449 + }, + { + "epoch": 0.9137587760662266, + "grad_norm": 0.3598082363605499, + "learning_rate": 3.975935962795807e-07, + "loss": 0.8045, + "step": 5450 + }, + { + "epoch": 0.9139264382269726, + "grad_norm": 0.4267180860042572, + "learning_rate": 3.960789365622075e-07, + "loss": 0.7992, + "step": 5451 + }, + { + "epoch": 0.9140941003877188, + "grad_norm": 0.5257976651191711, + "learning_rate": 3.9456710916008003e-07, + "loss": 0.7921, + "step": 5452 + }, + { + "epoch": 0.9142617625484648, + "grad_norm": 0.3591015934944153, + "learning_rate": 3.930581145190526e-07, + "loss": 0.8088, + "step": 5453 + }, + { + "epoch": 0.914429424709211, + "grad_norm": 0.4040519893169403, + "learning_rate": 3.915519530841505e-07, + "loss": 0.9061, + "step": 5454 + }, + { + "epoch": 0.914597086869957, + "grad_norm": 0.41480520367622375, + "learning_rate": 3.9004862529955613e-07, + "loss": 0.7608, + "step": 5455 + }, + { + "epoch": 0.9147647490307032, + "grad_norm": 0.4387241303920746, + "learning_rate": 3.8854813160861924e-07, + "loss": 0.8797, + "step": 5456 + }, + { + "epoch": 0.9149324111914492, + "grad_norm": 0.48141932487487793, + "learning_rate": 3.8705047245385707e-07, + "loss": 0.8078, + "step": 5457 + }, + { + "epoch": 0.9151000733521953, + "grad_norm": 0.518697202205658, + "learning_rate": 3.8555564827694627e-07, + "loss": 0.7451, + "step": 5458 + }, + { + "epoch": 0.9152677355129414, + "grad_norm": 0.4883868396282196, + "learning_rate": 3.8406365951872636e-07, + "loss": 0.8408, + "step": 5459 + }, + { + "epoch": 0.9154353976736875, + "grad_norm": 0.5068694949150085, + "learning_rate": 3.825745066192066e-07, + "loss": 0.8275, + "step": 5460 + }, + { + "epoch": 0.9156030598344336, + "grad_norm": 0.4290688633918762, + "learning_rate": 3.810881900175556e-07, + "loss": 0.8499, + "step": 5461 + }, + { + "epoch": 0.9157707219951797, + "grad_norm": 0.5387886762619019, + "learning_rate": 3.796047101521072e-07, + "loss": 0.8061, + "step": 5462 + }, + { + "epoch": 0.9159383841559258, + "grad_norm": 0.5438194870948792, + "learning_rate": 3.78124067460357e-07, + "loss": 0.8132, + "step": 5463 + }, + { + "epoch": 0.9161060463166719, + "grad_norm": 0.33834096789360046, + "learning_rate": 3.766462623789646e-07, + "loss": 0.7614, + "step": 5464 + }, + { + "epoch": 0.916273708477418, + "grad_norm": 0.47196847200393677, + "learning_rate": 3.7517129534375475e-07, + "loss": 0.9023, + "step": 5465 + }, + { + "epoch": 0.9164413706381641, + "grad_norm": 0.4197845160961151, + "learning_rate": 3.7369916678971053e-07, + "loss": 0.8842, + "step": 5466 + }, + { + "epoch": 0.9166090327989102, + "grad_norm": 0.43537554144859314, + "learning_rate": 3.722298771509847e-07, + "loss": 0.8246, + "step": 5467 + }, + { + "epoch": 0.9167766949596563, + "grad_norm": 0.4930468201637268, + "learning_rate": 3.707634268608862e-07, + "loss": 0.9056, + "step": 5468 + }, + { + "epoch": 0.9169443571204023, + "grad_norm": 0.4002600908279419, + "learning_rate": 3.692998163518902e-07, + "loss": 0.8087, + "step": 5469 + }, + { + "epoch": 0.9171120192811485, + "grad_norm": 1.5046316385269165, + "learning_rate": 3.6783904605563246e-07, + "loss": 0.8554, + "step": 5470 + }, + { + "epoch": 0.9172796814418945, + "grad_norm": 0.4091032147407532, + "learning_rate": 3.663811164029152e-07, + "loss": 0.8805, + "step": 5471 + }, + { + "epoch": 0.9174473436026407, + "grad_norm": 0.35924676060676575, + "learning_rate": 3.6492602782369657e-07, + "loss": 0.8437, + "step": 5472 + }, + { + "epoch": 0.9176150057633867, + "grad_norm": 0.5918196439743042, + "learning_rate": 3.634737807471023e-07, + "loss": 0.9303, + "step": 5473 + }, + { + "epoch": 0.9177826679241329, + "grad_norm": 0.3623254895210266, + "learning_rate": 3.620243756014175e-07, + "loss": 0.8151, + "step": 5474 + }, + { + "epoch": 0.9179503300848789, + "grad_norm": 0.5924551486968994, + "learning_rate": 3.605778128140891e-07, + "loss": 0.8822, + "step": 5475 + }, + { + "epoch": 0.9181179922456251, + "grad_norm": 0.3630390465259552, + "learning_rate": 3.5913409281172706e-07, + "loss": 0.8507, + "step": 5476 + }, + { + "epoch": 0.9182856544063711, + "grad_norm": 1.2100481986999512, + "learning_rate": 3.57693216020103e-07, + "loss": 0.7789, + "step": 5477 + }, + { + "epoch": 0.9184533165671173, + "grad_norm": 0.4164574444293976, + "learning_rate": 3.5625518286414696e-07, + "loss": 0.83, + "step": 5478 + }, + { + "epoch": 0.9186209787278633, + "grad_norm": 0.4077569842338562, + "learning_rate": 3.548199937679553e-07, + "loss": 0.9101, + "step": 5479 + }, + { + "epoch": 0.9187886408886095, + "grad_norm": 0.6429985761642456, + "learning_rate": 3.533876491547805e-07, + "loss": 0.868, + "step": 5480 + }, + { + "epoch": 0.9189563030493555, + "grad_norm": 0.4084544777870178, + "learning_rate": 3.5195814944704143e-07, + "loss": 0.9758, + "step": 5481 + }, + { + "epoch": 0.9191239652101016, + "grad_norm": 0.43578848242759705, + "learning_rate": 3.5053149506631413e-07, + "loss": 0.8348, + "step": 5482 + }, + { + "epoch": 0.9192916273708478, + "grad_norm": 0.4982907176017761, + "learning_rate": 3.491076864333354e-07, + "loss": 0.844, + "step": 5483 + }, + { + "epoch": 0.9194592895315938, + "grad_norm": 0.4257267415523529, + "learning_rate": 3.476867239680071e-07, + "loss": 0.7825, + "step": 5484 + }, + { + "epoch": 0.91962695169234, + "grad_norm": 0.4639023244380951, + "learning_rate": 3.46268608089384e-07, + "loss": 0.8848, + "step": 5485 + }, + { + "epoch": 0.919794613853086, + "grad_norm": 0.5799000263214111, + "learning_rate": 3.4485333921569166e-07, + "loss": 0.8532, + "step": 5486 + }, + { + "epoch": 0.9199622760138322, + "grad_norm": 0.532958447933197, + "learning_rate": 3.434409177643061e-07, + "loss": 0.8224, + "step": 5487 + }, + { + "epoch": 0.9201299381745782, + "grad_norm": 0.31039291620254517, + "learning_rate": 3.420313441517686e-07, + "loss": 0.8132, + "step": 5488 + }, + { + "epoch": 0.9202976003353244, + "grad_norm": 0.41978487372398376, + "learning_rate": 3.4062461879378005e-07, + "loss": 0.8231, + "step": 5489 + }, + { + "epoch": 0.9204652624960704, + "grad_norm": 0.3548107147216797, + "learning_rate": 3.3922074210520407e-07, + "loss": 0.9061, + "step": 5490 + }, + { + "epoch": 0.9206329246568166, + "grad_norm": 0.4682113826274872, + "learning_rate": 3.3781971450005724e-07, + "loss": 0.8242, + "step": 5491 + }, + { + "epoch": 0.9208005868175626, + "grad_norm": 0.46550726890563965, + "learning_rate": 3.364215363915213e-07, + "loss": 0.9678, + "step": 5492 + }, + { + "epoch": 0.9209682489783088, + "grad_norm": 0.46769365668296814, + "learning_rate": 3.350262081919375e-07, + "loss": 0.8332, + "step": 5493 + }, + { + "epoch": 0.9211359111390548, + "grad_norm": 0.4007333219051361, + "learning_rate": 3.336337303128034e-07, + "loss": 0.8389, + "step": 5494 + }, + { + "epoch": 0.921303573299801, + "grad_norm": 0.4930897057056427, + "learning_rate": 3.322441031647816e-07, + "loss": 0.7938, + "step": 5495 + }, + { + "epoch": 0.921471235460547, + "grad_norm": 0.4777367115020752, + "learning_rate": 3.308573271576854e-07, + "loss": 0.7612, + "step": 5496 + }, + { + "epoch": 0.9216388976212931, + "grad_norm": 1.0620161294937134, + "learning_rate": 3.2947340270049424e-07, + "loss": 0.8541, + "step": 5497 + }, + { + "epoch": 0.9218065597820392, + "grad_norm": 0.462362140417099, + "learning_rate": 3.2809233020134614e-07, + "loss": 0.8281, + "step": 5498 + }, + { + "epoch": 0.9219742219427853, + "grad_norm": 0.651817262172699, + "learning_rate": 3.267141100675353e-07, + "loss": 0.8666, + "step": 5499 + }, + { + "epoch": 0.9221418841035314, + "grad_norm": 0.3979651927947998, + "learning_rate": 3.2533874270551546e-07, + "loss": 0.8032, + "step": 5500 + }, + { + "epoch": 0.9223095462642775, + "grad_norm": 0.4143291413784027, + "learning_rate": 3.239662285209e-07, + "loss": 0.8331, + "step": 5501 + }, + { + "epoch": 0.9224772084250236, + "grad_norm": 0.3905138373374939, + "learning_rate": 3.225965679184595e-07, + "loss": 0.8526, + "step": 5502 + }, + { + "epoch": 0.9226448705857697, + "grad_norm": 0.4227253198623657, + "learning_rate": 3.2122976130212644e-07, + "loss": 0.7858, + "step": 5503 + }, + { + "epoch": 0.9228125327465158, + "grad_norm": 0.37597405910491943, + "learning_rate": 3.198658090749862e-07, + "loss": 0.8548, + "step": 5504 + }, + { + "epoch": 0.9229801949072619, + "grad_norm": 0.611361026763916, + "learning_rate": 3.185047116392859e-07, + "loss": 0.8773, + "step": 5505 + }, + { + "epoch": 0.923147857068008, + "grad_norm": 0.38209253549575806, + "learning_rate": 3.17146469396431e-07, + "loss": 0.7404, + "step": 5506 + }, + { + "epoch": 0.9233155192287541, + "grad_norm": 0.8735145330429077, + "learning_rate": 3.157910827469823e-07, + "loss": 0.8744, + "step": 5507 + }, + { + "epoch": 0.9234831813895001, + "grad_norm": 0.35140228271484375, + "learning_rate": 3.144385520906623e-07, + "loss": 0.8837, + "step": 5508 + }, + { + "epoch": 0.9236508435502463, + "grad_norm": 0.38805145025253296, + "learning_rate": 3.1308887782634746e-07, + "loss": 0.7759, + "step": 5509 + }, + { + "epoch": 0.9238185057109923, + "grad_norm": 0.4310365617275238, + "learning_rate": 3.1174206035207265e-07, + "loss": 0.7953, + "step": 5510 + }, + { + "epoch": 0.9239861678717385, + "grad_norm": 0.367244154214859, + "learning_rate": 3.1039810006503357e-07, + "loss": 0.811, + "step": 5511 + }, + { + "epoch": 0.9241538300324845, + "grad_norm": 1.0172940492630005, + "learning_rate": 3.0905699736157756e-07, + "loss": 0.9186, + "step": 5512 + }, + { + "epoch": 0.9243214921932307, + "grad_norm": 0.47557225823402405, + "learning_rate": 3.077187526372161e-07, + "loss": 0.8462, + "step": 5513 + }, + { + "epoch": 0.9244891543539767, + "grad_norm": 0.42626944184303284, + "learning_rate": 3.063833662866111e-07, + "loss": 0.8948, + "step": 5514 + }, + { + "epoch": 0.9246568165147229, + "grad_norm": 0.6726216077804565, + "learning_rate": 3.0505083870358555e-07, + "loss": 0.8881, + "step": 5515 + }, + { + "epoch": 0.9248244786754689, + "grad_norm": 1.152274250984192, + "learning_rate": 3.0372117028111825e-07, + "loss": 0.7966, + "step": 5516 + }, + { + "epoch": 0.9249921408362151, + "grad_norm": 0.4575171172618866, + "learning_rate": 3.023943614113456e-07, + "loss": 0.7819, + "step": 5517 + }, + { + "epoch": 0.9251598029969611, + "grad_norm": 0.4207251965999603, + "learning_rate": 3.0107041248556015e-07, + "loss": 0.7604, + "step": 5518 + }, + { + "epoch": 0.9253274651577073, + "grad_norm": 0.4027195870876312, + "learning_rate": 2.997493238942084e-07, + "loss": 0.8249, + "step": 5519 + }, + { + "epoch": 0.9254951273184533, + "grad_norm": 0.32407307624816895, + "learning_rate": 2.984310960268988e-07, + "loss": 0.7537, + "step": 5520 + }, + { + "epoch": 0.9256627894791994, + "grad_norm": 0.5370124578475952, + "learning_rate": 2.971157292723903e-07, + "loss": 0.8758, + "step": 5521 + }, + { + "epoch": 0.9258304516399455, + "grad_norm": 0.4623180329799652, + "learning_rate": 2.9580322401860374e-07, + "loss": 0.8286, + "step": 5522 + }, + { + "epoch": 0.9259981138006916, + "grad_norm": 0.6158974766731262, + "learning_rate": 2.944935806526117e-07, + "loss": 0.732, + "step": 5523 + }, + { + "epoch": 0.9261657759614377, + "grad_norm": 0.4130534529685974, + "learning_rate": 2.93186799560643e-07, + "loss": 0.8576, + "step": 5524 + }, + { + "epoch": 0.9263334381221838, + "grad_norm": 0.979594349861145, + "learning_rate": 2.9188288112808606e-07, + "loss": 0.8856, + "step": 5525 + }, + { + "epoch": 0.9265011002829299, + "grad_norm": 0.4869593679904938, + "learning_rate": 2.905818257394799e-07, + "loss": 0.8958, + "step": 5526 + }, + { + "epoch": 0.926668762443676, + "grad_norm": 0.43282878398895264, + "learning_rate": 2.892836337785243e-07, + "loss": 0.816, + "step": 5527 + }, + { + "epoch": 0.9268364246044221, + "grad_norm": 0.5063544511795044, + "learning_rate": 2.8798830562807077e-07, + "loss": 0.8357, + "step": 5528 + }, + { + "epoch": 0.9270040867651682, + "grad_norm": 0.3956616520881653, + "learning_rate": 2.866958416701271e-07, + "loss": 0.8457, + "step": 5529 + }, + { + "epoch": 0.9271717489259143, + "grad_norm": 0.46155110001564026, + "learning_rate": 2.8540624228585947e-07, + "loss": 0.9228, + "step": 5530 + }, + { + "epoch": 0.9273394110866604, + "grad_norm": 0.3728257417678833, + "learning_rate": 2.8411950785558474e-07, + "loss": 0.8098, + "step": 5531 + }, + { + "epoch": 0.9275070732474064, + "grad_norm": 0.4718465805053711, + "learning_rate": 2.8283563875877494e-07, + "loss": 0.7529, + "step": 5532 + }, + { + "epoch": 0.9276747354081526, + "grad_norm": 0.4334772825241089, + "learning_rate": 2.815546353740639e-07, + "loss": 0.83, + "step": 5533 + }, + { + "epoch": 0.9278423975688986, + "grad_norm": 0.7013729810714722, + "learning_rate": 2.8027649807923053e-07, + "loss": 0.8759, + "step": 5534 + }, + { + "epoch": 0.9280100597296448, + "grad_norm": 0.4334397614002228, + "learning_rate": 2.7900122725121547e-07, + "loss": 0.8663, + "step": 5535 + }, + { + "epoch": 0.9281777218903908, + "grad_norm": 0.4532167613506317, + "learning_rate": 2.777288232661135e-07, + "loss": 0.7878, + "step": 5536 + }, + { + "epoch": 0.928345384051137, + "grad_norm": 0.37148404121398926, + "learning_rate": 2.764592864991689e-07, + "loss": 0.8114, + "step": 5537 + }, + { + "epoch": 0.928513046211883, + "grad_norm": 0.43077895045280457, + "learning_rate": 2.751926173247865e-07, + "loss": 0.7621, + "step": 5538 + }, + { + "epoch": 0.9286807083726292, + "grad_norm": 0.5970816016197205, + "learning_rate": 2.73928816116521e-07, + "loss": 0.8411, + "step": 5539 + }, + { + "epoch": 0.9288483705333752, + "grad_norm": 0.4724122881889343, + "learning_rate": 2.7266788324708524e-07, + "loss": 0.8413, + "step": 5540 + }, + { + "epoch": 0.9290160326941214, + "grad_norm": 0.4347897171974182, + "learning_rate": 2.714098190883407e-07, + "loss": 0.9437, + "step": 5541 + }, + { + "epoch": 0.9291836948548674, + "grad_norm": 0.42247769236564636, + "learning_rate": 2.7015462401130843e-07, + "loss": 0.7411, + "step": 5542 + }, + { + "epoch": 0.9293513570156136, + "grad_norm": 0.5921533107757568, + "learning_rate": 2.6890229838616e-07, + "loss": 0.8147, + "step": 5543 + }, + { + "epoch": 0.9295190191763596, + "grad_norm": 0.28696200251579285, + "learning_rate": 2.6765284258222337e-07, + "loss": 0.8451, + "step": 5544 + }, + { + "epoch": 0.9296866813371057, + "grad_norm": 0.8700627684593201, + "learning_rate": 2.6640625696797706e-07, + "loss": 0.87, + "step": 5545 + }, + { + "epoch": 0.9298543434978518, + "grad_norm": 0.308586448431015, + "learning_rate": 2.6516254191105374e-07, + "loss": 0.7779, + "step": 5546 + }, + { + "epoch": 0.9300220056585979, + "grad_norm": 0.4023432433605194, + "learning_rate": 2.639216977782433e-07, + "loss": 0.8421, + "step": 5547 + }, + { + "epoch": 0.930189667819344, + "grad_norm": 0.502148449420929, + "learning_rate": 2.626837249354819e-07, + "loss": 0.8936, + "step": 5548 + }, + { + "epoch": 0.9303573299800901, + "grad_norm": 0.9387050271034241, + "learning_rate": 2.614486237478675e-07, + "loss": 0.8268, + "step": 5549 + }, + { + "epoch": 0.9305249921408362, + "grad_norm": 0.6043473482131958, + "learning_rate": 2.602163945796432e-07, + "loss": 0.8885, + "step": 5550 + }, + { + "epoch": 0.9306926543015823, + "grad_norm": 0.4056151509284973, + "learning_rate": 2.589870377942105e-07, + "loss": 0.7908, + "step": 5551 + }, + { + "epoch": 0.9308603164623284, + "grad_norm": 0.41021332144737244, + "learning_rate": 2.577605537541217e-07, + "loss": 0.8381, + "step": 5552 + }, + { + "epoch": 0.9310279786230745, + "grad_norm": 0.447323739528656, + "learning_rate": 2.565369428210818e-07, + "loss": 0.8611, + "step": 5553 + }, + { + "epoch": 0.9311956407838206, + "grad_norm": 0.4121238887310028, + "learning_rate": 2.553162053559488e-07, + "loss": 0.7491, + "step": 5554 + }, + { + "epoch": 0.9313633029445667, + "grad_norm": 0.35634592175483704, + "learning_rate": 2.540983417187348e-07, + "loss": 0.8236, + "step": 5555 + }, + { + "epoch": 0.9315309651053127, + "grad_norm": 0.3375703990459442, + "learning_rate": 2.528833522685992e-07, + "loss": 0.8379, + "step": 5556 + }, + { + "epoch": 0.9316986272660589, + "grad_norm": 0.524369478225708, + "learning_rate": 2.516712373638608e-07, + "loss": 0.8166, + "step": 5557 + }, + { + "epoch": 0.9318662894268049, + "grad_norm": 0.43301674723625183, + "learning_rate": 2.5046199736198596e-07, + "loss": 0.8631, + "step": 5558 + }, + { + "epoch": 0.9320339515875511, + "grad_norm": 0.47489795088768005, + "learning_rate": 2.4925563261959606e-07, + "loss": 0.7435, + "step": 5559 + }, + { + "epoch": 0.9322016137482971, + "grad_norm": 0.41612544655799866, + "learning_rate": 2.48052143492461e-07, + "loss": 0.8816, + "step": 5560 + }, + { + "epoch": 0.9323692759090433, + "grad_norm": 0.43906569480895996, + "learning_rate": 2.4685153033550457e-07, + "loss": 0.7912, + "step": 5561 + }, + { + "epoch": 0.9325369380697893, + "grad_norm": 0.4497454762458801, + "learning_rate": 2.456537935028036e-07, + "loss": 0.8735, + "step": 5562 + }, + { + "epoch": 0.9327046002305355, + "grad_norm": 0.3805798590183258, + "learning_rate": 2.4445893334758664e-07, + "loss": 0.7605, + "step": 5563 + }, + { + "epoch": 0.9328722623912815, + "grad_norm": 0.4280644357204437, + "learning_rate": 2.4326695022222956e-07, + "loss": 0.7823, + "step": 5564 + }, + { + "epoch": 0.9330399245520277, + "grad_norm": 0.47225022315979004, + "learning_rate": 2.420778444782645e-07, + "loss": 0.9112, + "step": 5565 + }, + { + "epoch": 0.9332075867127737, + "grad_norm": 0.3943757116794586, + "learning_rate": 2.408916164663755e-07, + "loss": 0.8492, + "step": 5566 + }, + { + "epoch": 0.9333752488735199, + "grad_norm": 0.3952292799949646, + "learning_rate": 2.397082665363915e-07, + "loss": 0.8742, + "step": 5567 + }, + { + "epoch": 0.9335429110342659, + "grad_norm": 0.3373555541038513, + "learning_rate": 2.3852779503730217e-07, + "loss": 0.7162, + "step": 5568 + }, + { + "epoch": 0.933710573195012, + "grad_norm": 0.46765002608299255, + "learning_rate": 2.3735020231723915e-07, + "loss": 0.8617, + "step": 5569 + }, + { + "epoch": 0.9338782353557582, + "grad_norm": 0.44101208448410034, + "learning_rate": 2.3617548872349015e-07, + "loss": 0.8869, + "step": 5570 + }, + { + "epoch": 0.9340458975165042, + "grad_norm": 0.3950546383857727, + "learning_rate": 2.3500365460249474e-07, + "loss": 0.7568, + "step": 5571 + }, + { + "epoch": 0.9342135596772504, + "grad_norm": 0.38100630044937134, + "learning_rate": 2.3383470029983868e-07, + "loss": 0.8651, + "step": 5572 + }, + { + "epoch": 0.9343812218379964, + "grad_norm": 0.33927208185195923, + "learning_rate": 2.3266862616026175e-07, + "loss": 0.7564, + "step": 5573 + }, + { + "epoch": 0.9345488839987426, + "grad_norm": 0.31735849380493164, + "learning_rate": 2.3150543252765556e-07, + "loss": 0.7747, + "step": 5574 + }, + { + "epoch": 0.9347165461594886, + "grad_norm": 0.5408655405044556, + "learning_rate": 2.303451197450568e-07, + "loss": 0.767, + "step": 5575 + }, + { + "epoch": 0.9348842083202348, + "grad_norm": 0.45583081245422363, + "learning_rate": 2.2918768815465951e-07, + "loss": 0.7727, + "step": 5576 + }, + { + "epoch": 0.9350518704809808, + "grad_norm": 0.8801291584968567, + "learning_rate": 2.280331380978029e-07, + "loss": 0.7964, + "step": 5577 + }, + { + "epoch": 0.935219532641727, + "grad_norm": 0.2814652621746063, + "learning_rate": 2.2688146991497683e-07, + "loss": 0.8358, + "step": 5578 + }, + { + "epoch": 0.935387194802473, + "grad_norm": 0.532590925693512, + "learning_rate": 2.2573268394582625e-07, + "loss": 0.8855, + "step": 5579 + }, + { + "epoch": 0.9355548569632192, + "grad_norm": 0.34196722507476807, + "learning_rate": 2.2458678052913796e-07, + "loss": 0.8477, + "step": 5580 + }, + { + "epoch": 0.9357225191239652, + "grad_norm": 0.39009302854537964, + "learning_rate": 2.2344376000285606e-07, + "loss": 0.8902, + "step": 5581 + }, + { + "epoch": 0.9358901812847114, + "grad_norm": 0.4790273904800415, + "learning_rate": 2.2230362270407203e-07, + "loss": 0.8542, + "step": 5582 + }, + { + "epoch": 0.9360578434454574, + "grad_norm": 0.45219048857688904, + "learning_rate": 2.2116636896902244e-07, + "loss": 0.8752, + "step": 5583 + }, + { + "epoch": 0.9362255056062035, + "grad_norm": 0.35386043787002563, + "learning_rate": 2.2003199913310124e-07, + "loss": 0.9029, + "step": 5584 + }, + { + "epoch": 0.9363931677669496, + "grad_norm": 0.43155017495155334, + "learning_rate": 2.189005135308475e-07, + "loss": 0.8639, + "step": 5585 + }, + { + "epoch": 0.9365608299276957, + "grad_norm": 0.40214449167251587, + "learning_rate": 2.1777191249594986e-07, + "loss": 0.8996, + "step": 5586 + }, + { + "epoch": 0.9367284920884418, + "grad_norm": 0.33067846298217773, + "learning_rate": 2.1664619636124652e-07, + "loss": 0.781, + "step": 5587 + }, + { + "epoch": 0.9368961542491879, + "grad_norm": 0.6097775101661682, + "learning_rate": 2.1552336545872522e-07, + "loss": 0.8176, + "step": 5588 + }, + { + "epoch": 0.937063816409934, + "grad_norm": 0.36495113372802734, + "learning_rate": 2.144034201195222e-07, + "loss": 0.9063, + "step": 5589 + }, + { + "epoch": 0.9372314785706801, + "grad_norm": 0.3748246133327484, + "learning_rate": 2.132863606739255e-07, + "loss": 0.8615, + "step": 5590 + }, + { + "epoch": 0.9373991407314262, + "grad_norm": 0.3747648596763611, + "learning_rate": 2.1217218745136715e-07, + "loss": 0.8227, + "step": 5591 + }, + { + "epoch": 0.9375668028921723, + "grad_norm": 0.4228968322277069, + "learning_rate": 2.1106090078043095e-07, + "loss": 0.7919, + "step": 5592 + }, + { + "epoch": 0.9377344650529184, + "grad_norm": 0.43143996596336365, + "learning_rate": 2.0995250098885034e-07, + "loss": 0.7377, + "step": 5593 + }, + { + "epoch": 0.9379021272136645, + "grad_norm": 0.611870527267456, + "learning_rate": 2.0884698840350492e-07, + "loss": 0.8197, + "step": 5594 + }, + { + "epoch": 0.9380697893744105, + "grad_norm": 0.44937756657600403, + "learning_rate": 2.0774436335042614e-07, + "loss": 0.9146, + "step": 5595 + }, + { + "epoch": 0.9382374515351567, + "grad_norm": 0.40494683384895325, + "learning_rate": 2.0664462615478942e-07, + "loss": 0.8464, + "step": 5596 + }, + { + "epoch": 0.9384051136959027, + "grad_norm": 0.3957933485507965, + "learning_rate": 2.055477771409209e-07, + "loss": 0.7419, + "step": 5597 + }, + { + "epoch": 0.9385727758566489, + "grad_norm": 0.6733267903327942, + "learning_rate": 2.0445381663229734e-07, + "loss": 0.9049, + "step": 5598 + }, + { + "epoch": 0.9387404380173949, + "grad_norm": 0.42263564467430115, + "learning_rate": 2.033627449515374e-07, + "loss": 0.8662, + "step": 5599 + }, + { + "epoch": 0.9389081001781411, + "grad_norm": 0.5451824069023132, + "learning_rate": 2.0227456242041589e-07, + "loss": 0.8229, + "step": 5600 + }, + { + "epoch": 0.9390757623388871, + "grad_norm": 0.36930420994758606, + "learning_rate": 2.0118926935984939e-07, + "loss": 0.8783, + "step": 5601 + }, + { + "epoch": 0.9392434244996333, + "grad_norm": 0.36576372385025024, + "learning_rate": 2.0010686608990304e-07, + "loss": 0.8387, + "step": 5602 + }, + { + "epoch": 0.9394110866603793, + "grad_norm": 0.32187724113464355, + "learning_rate": 1.9902735292979257e-07, + "loss": 0.8154, + "step": 5603 + }, + { + "epoch": 0.9395787488211255, + "grad_norm": 0.4188230037689209, + "learning_rate": 1.9795073019788002e-07, + "loss": 0.8459, + "step": 5604 + }, + { + "epoch": 0.9397464109818715, + "grad_norm": 0.33690860867500305, + "learning_rate": 1.968769982116736e-07, + "loss": 0.8161, + "step": 5605 + }, + { + "epoch": 0.9399140731426177, + "grad_norm": 0.3758140504360199, + "learning_rate": 1.9580615728783004e-07, + "loss": 0.7501, + "step": 5606 + }, + { + "epoch": 0.9400817353033637, + "grad_norm": 0.9181646108627319, + "learning_rate": 1.9473820774215557e-07, + "loss": 0.8525, + "step": 5607 + }, + { + "epoch": 0.9402493974641098, + "grad_norm": 0.44023045897483826, + "learning_rate": 1.9367314988959829e-07, + "loss": 0.7978, + "step": 5608 + }, + { + "epoch": 0.9404170596248559, + "grad_norm": 0.5621471405029297, + "learning_rate": 1.9261098404426136e-07, + "loss": 0.8633, + "step": 5609 + }, + { + "epoch": 0.940584721785602, + "grad_norm": 0.7647391557693481, + "learning_rate": 1.9155171051938758e-07, + "loss": 0.7623, + "step": 5610 + }, + { + "epoch": 0.9407523839463481, + "grad_norm": 0.34152573347091675, + "learning_rate": 1.9049532962736926e-07, + "loss": 0.7409, + "step": 5611 + }, + { + "epoch": 0.9409200461070942, + "grad_norm": 0.4390283226966858, + "learning_rate": 1.8944184167974832e-07, + "loss": 0.8357, + "step": 5612 + }, + { + "epoch": 0.9410877082678403, + "grad_norm": 0.3948175609111786, + "learning_rate": 1.883912469872118e-07, + "loss": 0.8497, + "step": 5613 + }, + { + "epoch": 0.9412553704285864, + "grad_norm": 0.4613439440727234, + "learning_rate": 1.8734354585958958e-07, + "loss": 0.8871, + "step": 5614 + }, + { + "epoch": 0.9414230325893325, + "grad_norm": 0.4463028609752655, + "learning_rate": 1.8629873860586567e-07, + "loss": 0.8103, + "step": 5615 + }, + { + "epoch": 0.9415906947500786, + "grad_norm": 0.3706364929676056, + "learning_rate": 1.8525682553416468e-07, + "loss": 0.8776, + "step": 5616 + }, + { + "epoch": 0.9417583569108247, + "grad_norm": 0.5275152325630188, + "learning_rate": 1.8421780695175973e-07, + "loss": 0.8077, + "step": 5617 + }, + { + "epoch": 0.9419260190715708, + "grad_norm": 0.5429855585098267, + "learning_rate": 1.8318168316507123e-07, + "loss": 0.8583, + "step": 5618 + }, + { + "epoch": 0.9420936812323168, + "grad_norm": 0.3907036781311035, + "learning_rate": 1.8214845447966256e-07, + "loss": 0.8475, + "step": 5619 + }, + { + "epoch": 0.942261343393063, + "grad_norm": 0.3771839141845703, + "learning_rate": 1.8111812120024884e-07, + "loss": 0.8338, + "step": 5620 + }, + { + "epoch": 0.942429005553809, + "grad_norm": 0.38360580801963806, + "learning_rate": 1.8009068363068595e-07, + "loss": 0.8439, + "step": 5621 + }, + { + "epoch": 0.9425966677145552, + "grad_norm": 0.6593756675720215, + "learning_rate": 1.790661420739781e-07, + "loss": 0.8713, + "step": 5622 + }, + { + "epoch": 0.9427643298753012, + "grad_norm": 0.35056811571121216, + "learning_rate": 1.7804449683227697e-07, + "loss": 0.8065, + "step": 5623 + }, + { + "epoch": 0.9429319920360474, + "grad_norm": 0.5193197131156921, + "learning_rate": 1.7702574820687602e-07, + "loss": 0.9337, + "step": 5624 + }, + { + "epoch": 0.9430996541967934, + "grad_norm": 0.4338991045951843, + "learning_rate": 1.7600989649821932e-07, + "loss": 0.7891, + "step": 5625 + }, + { + "epoch": 0.9432673163575396, + "grad_norm": 0.4451599717140198, + "learning_rate": 1.7499694200589057e-07, + "loss": 0.9064, + "step": 5626 + }, + { + "epoch": 0.9434349785182856, + "grad_norm": 0.3576316833496094, + "learning_rate": 1.7398688502862748e-07, + "loss": 0.8033, + "step": 5627 + }, + { + "epoch": 0.9436026406790318, + "grad_norm": 0.3730756938457489, + "learning_rate": 1.7297972586430401e-07, + "loss": 0.882, + "step": 5628 + }, + { + "epoch": 0.9437703028397778, + "grad_norm": 0.4072226583957672, + "learning_rate": 1.7197546480994587e-07, + "loss": 0.8271, + "step": 5629 + }, + { + "epoch": 0.943937965000524, + "grad_norm": 0.3262406289577484, + "learning_rate": 1.7097410216172062e-07, + "loss": 0.7549, + "step": 5630 + }, + { + "epoch": 0.94410562716127, + "grad_norm": 0.47185268998146057, + "learning_rate": 1.6997563821494533e-07, + "loss": 0.9325, + "step": 5631 + }, + { + "epoch": 0.9442732893220162, + "grad_norm": 0.4424075186252594, + "learning_rate": 1.6898007326407784e-07, + "loss": 0.7956, + "step": 5632 + }, + { + "epoch": 0.9444409514827622, + "grad_norm": 0.5058478713035583, + "learning_rate": 1.6798740760272104e-07, + "loss": 0.9044, + "step": 5633 + }, + { + "epoch": 0.9446086136435083, + "grad_norm": 0.6497793197631836, + "learning_rate": 1.6699764152362628e-07, + "loss": 0.7954, + "step": 5634 + }, + { + "epoch": 0.9447762758042544, + "grad_norm": 0.38459107279777527, + "learning_rate": 1.6601077531868794e-07, + "loss": 0.8354, + "step": 5635 + }, + { + "epoch": 0.9449439379650005, + "grad_norm": 0.40192103385925293, + "learning_rate": 1.650268092789442e-07, + "loss": 0.8224, + "step": 5636 + }, + { + "epoch": 0.9451116001257466, + "grad_norm": 0.40187108516693115, + "learning_rate": 1.640457436945786e-07, + "loss": 0.9093, + "step": 5637 + }, + { + "epoch": 0.9452792622864927, + "grad_norm": 0.42827141284942627, + "learning_rate": 1.6306757885491964e-07, + "loss": 0.817, + "step": 5638 + }, + { + "epoch": 0.9454469244472388, + "grad_norm": 0.37759557366371155, + "learning_rate": 1.6209231504844102e-07, + "loss": 0.7827, + "step": 5639 + }, + { + "epoch": 0.9456145866079849, + "grad_norm": 0.3767712712287903, + "learning_rate": 1.6111995256275936e-07, + "loss": 0.8499, + "step": 5640 + }, + { + "epoch": 0.945782248768731, + "grad_norm": 0.3564743995666504, + "learning_rate": 1.6015049168463748e-07, + "loss": 0.758, + "step": 5641 + }, + { + "epoch": 0.9459499109294771, + "grad_norm": 0.6888882517814636, + "learning_rate": 1.5918393269998112e-07, + "loss": 0.8285, + "step": 5642 + }, + { + "epoch": 0.9461175730902232, + "grad_norm": 0.46881723403930664, + "learning_rate": 1.582202758938378e-07, + "loss": 0.9183, + "step": 5643 + }, + { + "epoch": 0.9462852352509693, + "grad_norm": 0.8559983372688293, + "learning_rate": 1.572595215504058e-07, + "loss": 0.7672, + "step": 5644 + }, + { + "epoch": 0.9464528974117153, + "grad_norm": 0.9411035180091858, + "learning_rate": 1.5630166995302177e-07, + "loss": 0.8157, + "step": 5645 + }, + { + "epoch": 0.9466205595724615, + "grad_norm": 0.4257885217666626, + "learning_rate": 1.553467213841664e-07, + "loss": 0.8358, + "step": 5646 + }, + { + "epoch": 0.9467882217332075, + "grad_norm": 0.697925865650177, + "learning_rate": 1.543946761254689e-07, + "loss": 0.8611, + "step": 5647 + }, + { + "epoch": 0.9469558838939537, + "grad_norm": 0.5628939867019653, + "learning_rate": 1.5344553445769683e-07, + "loss": 0.8576, + "step": 5648 + }, + { + "epoch": 0.9471235460546997, + "grad_norm": 0.6172940731048584, + "learning_rate": 1.5249929666076525e-07, + "loss": 0.8426, + "step": 5649 + }, + { + "epoch": 0.9472912082154459, + "grad_norm": 0.4763745963573456, + "learning_rate": 1.51555963013732e-07, + "loss": 0.8808, + "step": 5650 + }, + { + "epoch": 0.9474588703761919, + "grad_norm": 0.3848874866962433, + "learning_rate": 1.5061553379479455e-07, + "loss": 0.8111, + "step": 5651 + }, + { + "epoch": 0.9476265325369381, + "grad_norm": 0.39039984345436096, + "learning_rate": 1.4967800928129994e-07, + "loss": 0.8439, + "step": 5652 + }, + { + "epoch": 0.9477941946976841, + "grad_norm": 0.3452758193016052, + "learning_rate": 1.4874338974973478e-07, + "loss": 0.8758, + "step": 5653 + }, + { + "epoch": 0.9479618568584303, + "grad_norm": 0.4906795024871826, + "learning_rate": 1.4781167547573194e-07, + "loss": 0.8095, + "step": 5654 + }, + { + "epoch": 0.9481295190191764, + "grad_norm": 0.37396731972694397, + "learning_rate": 1.468828667340616e-07, + "loss": 0.7964, + "step": 5655 + }, + { + "epoch": 0.9482971811799225, + "grad_norm": 1.0198975801467896, + "learning_rate": 1.4595696379864467e-07, + "loss": 0.9225, + "step": 5656 + }, + { + "epoch": 0.9484648433406686, + "grad_norm": 0.47543761134147644, + "learning_rate": 1.4503396694253936e-07, + "loss": 0.8874, + "step": 5657 + }, + { + "epoch": 0.9486325055014146, + "grad_norm": 0.3990827202796936, + "learning_rate": 1.4411387643795017e-07, + "loss": 0.815, + "step": 5658 + }, + { + "epoch": 0.9488001676621608, + "grad_norm": 0.3973783850669861, + "learning_rate": 1.4319669255622115e-07, + "loss": 0.7715, + "step": 5659 + }, + { + "epoch": 0.9489678298229068, + "grad_norm": 0.35283103585243225, + "learning_rate": 1.422824155678426e-07, + "loss": 0.8014, + "step": 5660 + }, + { + "epoch": 0.949135491983653, + "grad_norm": 0.5298373103141785, + "learning_rate": 1.4137104574244665e-07, + "loss": 0.9061, + "step": 5661 + }, + { + "epoch": 0.949303154144399, + "grad_norm": 0.3985663056373596, + "learning_rate": 1.4046258334880714e-07, + "loss": 0.8046, + "step": 5662 + }, + { + "epoch": 0.9494708163051452, + "grad_norm": 0.5122137069702148, + "learning_rate": 1.3955702865483977e-07, + "loss": 0.9043, + "step": 5663 + }, + { + "epoch": 0.9496384784658912, + "grad_norm": 0.5534654855728149, + "learning_rate": 1.386543819276054e-07, + "loss": 0.8845, + "step": 5664 + }, + { + "epoch": 0.9498061406266374, + "grad_norm": 0.38909614086151123, + "learning_rate": 1.3775464343330437e-07, + "loss": 0.8299, + "step": 5665 + }, + { + "epoch": 0.9499738027873834, + "grad_norm": 0.5959657430648804, + "learning_rate": 1.3685781343728222e-07, + "loss": 0.8008, + "step": 5666 + }, + { + "epoch": 0.9501414649481296, + "grad_norm": 0.4258873760700226, + "learning_rate": 1.3596389220402296e-07, + "loss": 0.794, + "step": 5667 + }, + { + "epoch": 0.9503091271088756, + "grad_norm": 0.46945369243621826, + "learning_rate": 1.350728799971579e-07, + "loss": 0.7653, + "step": 5668 + }, + { + "epoch": 0.9504767892696218, + "grad_norm": 0.3949854373931885, + "learning_rate": 1.3418477707945576e-07, + "loss": 0.8105, + "step": 5669 + }, + { + "epoch": 0.9506444514303678, + "grad_norm": 0.4090568721294403, + "learning_rate": 1.3329958371282815e-07, + "loss": 0.9001, + "step": 5670 + }, + { + "epoch": 0.950812113591114, + "grad_norm": 0.4272141456604004, + "learning_rate": 1.3241730015833176e-07, + "loss": 0.8375, + "step": 5671 + }, + { + "epoch": 0.95097977575186, + "grad_norm": 0.46921345591545105, + "learning_rate": 1.3153792667616183e-07, + "loss": 0.8256, + "step": 5672 + }, + { + "epoch": 0.9511474379126061, + "grad_norm": 0.4632411599159241, + "learning_rate": 1.3066146352565755e-07, + "loss": 0.8708, + "step": 5673 + }, + { + "epoch": 0.9513151000733522, + "grad_norm": 0.8034499883651733, + "learning_rate": 1.2978791096529663e-07, + "loss": 0.8364, + "step": 5674 + }, + { + "epoch": 0.9514827622340983, + "grad_norm": 0.4491528570652008, + "learning_rate": 1.2891726925270186e-07, + "loss": 0.8402, + "step": 5675 + }, + { + "epoch": 0.9516504243948444, + "grad_norm": 0.4233703911304474, + "learning_rate": 1.2804953864463676e-07, + "loss": 0.7798, + "step": 5676 + }, + { + "epoch": 0.9518180865555905, + "grad_norm": 0.37281307578086853, + "learning_rate": 1.271847193970055e-07, + "loss": 0.8267, + "step": 5677 + }, + { + "epoch": 0.9519857487163366, + "grad_norm": 0.43550345301628113, + "learning_rate": 1.2632281176485295e-07, + "loss": 0.813, + "step": 5678 + }, + { + "epoch": 0.9521534108770827, + "grad_norm": 0.6797903180122375, + "learning_rate": 1.254638160023669e-07, + "loss": 0.8983, + "step": 5679 + }, + { + "epoch": 0.9523210730378288, + "grad_norm": 0.43068262934684753, + "learning_rate": 1.2460773236287694e-07, + "loss": 0.8743, + "step": 5680 + }, + { + "epoch": 0.9524887351985749, + "grad_norm": 0.36471620202064514, + "learning_rate": 1.2375456109885108e-07, + "loss": 0.7999, + "step": 5681 + }, + { + "epoch": 0.952656397359321, + "grad_norm": 0.3796907067298889, + "learning_rate": 1.2290430246190145e-07, + "loss": 0.8054, + "step": 5682 + }, + { + "epoch": 0.9528240595200671, + "grad_norm": 0.6740773320198059, + "learning_rate": 1.2205695670277963e-07, + "loss": 0.8228, + "step": 5683 + }, + { + "epoch": 0.9529917216808131, + "grad_norm": 0.4136049747467041, + "learning_rate": 1.2121252407137685e-07, + "loss": 0.7726, + "step": 5684 + }, + { + "epoch": 0.9531593838415593, + "grad_norm": 0.45437702536582947, + "learning_rate": 1.2037100481672836e-07, + "loss": 0.8748, + "step": 5685 + }, + { + "epoch": 0.9533270460023053, + "grad_norm": 0.4190893769264221, + "learning_rate": 1.1953239918700897e-07, + "loss": 0.8073, + "step": 5686 + }, + { + "epoch": 0.9534947081630515, + "grad_norm": 1.0747568607330322, + "learning_rate": 1.1869670742953199e-07, + "loss": 0.865, + "step": 5687 + }, + { + "epoch": 0.9536623703237975, + "grad_norm": 0.5825490951538086, + "learning_rate": 1.1786392979075467e-07, + "loss": 0.8079, + "step": 5688 + }, + { + "epoch": 0.9538300324845437, + "grad_norm": 0.44110995531082153, + "learning_rate": 1.170340665162728e-07, + "loss": 0.8474, + "step": 5689 + }, + { + "epoch": 0.9539976946452897, + "grad_norm": 0.34802770614624023, + "learning_rate": 1.1620711785082395e-07, + "loss": 0.761, + "step": 5690 + }, + { + "epoch": 0.9541653568060359, + "grad_norm": 0.4393647313117981, + "learning_rate": 1.1538308403828524e-07, + "loss": 0.801, + "step": 5691 + }, + { + "epoch": 0.9543330189667819, + "grad_norm": 0.3807205259799957, + "learning_rate": 1.1456196532167452e-07, + "loss": 0.8628, + "step": 5692 + }, + { + "epoch": 0.9545006811275281, + "grad_norm": 0.4649239778518677, + "learning_rate": 1.1374376194315028e-07, + "loss": 0.834, + "step": 5693 + }, + { + "epoch": 0.9546683432882741, + "grad_norm": 0.47582659125328064, + "learning_rate": 1.1292847414400843e-07, + "loss": 0.9082, + "step": 5694 + }, + { + "epoch": 0.9548360054490203, + "grad_norm": 0.4253336489200592, + "learning_rate": 1.1211610216469105e-07, + "loss": 0.7836, + "step": 5695 + }, + { + "epoch": 0.9550036676097663, + "grad_norm": 0.4425368010997772, + "learning_rate": 1.1130664624477539e-07, + "loss": 0.8245, + "step": 5696 + }, + { + "epoch": 0.9551713297705124, + "grad_norm": 0.4651656746864319, + "learning_rate": 1.1050010662297827e-07, + "loss": 0.7881, + "step": 5697 + }, + { + "epoch": 0.9553389919312585, + "grad_norm": 0.6690202355384827, + "learning_rate": 1.0969648353715945e-07, + "loss": 0.846, + "step": 5698 + }, + { + "epoch": 0.9555066540920046, + "grad_norm": 0.4050973653793335, + "learning_rate": 1.0889577722431821e-07, + "loss": 0.8111, + "step": 5699 + }, + { + "epoch": 0.9556743162527507, + "grad_norm": 0.4939645528793335, + "learning_rate": 1.0809798792059234e-07, + "loss": 0.7909, + "step": 5700 + }, + { + "epoch": 0.9558419784134968, + "grad_norm": 0.36605989933013916, + "learning_rate": 1.0730311586125808e-07, + "loss": 0.8195, + "step": 5701 + }, + { + "epoch": 0.9560096405742429, + "grad_norm": 0.4095785319805145, + "learning_rate": 1.0651116128073568e-07, + "loss": 0.8232, + "step": 5702 + }, + { + "epoch": 0.956177302734989, + "grad_norm": 0.40485769510269165, + "learning_rate": 1.0572212441257945e-07, + "loss": 0.757, + "step": 5703 + }, + { + "epoch": 0.9563449648957351, + "grad_norm": 0.5039658546447754, + "learning_rate": 1.0493600548948879e-07, + "loss": 0.7698, + "step": 5704 + }, + { + "epoch": 0.9565126270564812, + "grad_norm": 0.42550402879714966, + "learning_rate": 1.0415280474329936e-07, + "loss": 0.825, + "step": 5705 + }, + { + "epoch": 0.9566802892172273, + "grad_norm": 0.425540566444397, + "learning_rate": 1.033725224049853e-07, + "loss": 0.868, + "step": 5706 + }, + { + "epoch": 0.9568479513779734, + "grad_norm": 0.5083727836608887, + "learning_rate": 1.0259515870466363e-07, + "loss": 0.7654, + "step": 5707 + }, + { + "epoch": 0.9570156135387194, + "grad_norm": 0.3627629578113556, + "learning_rate": 1.0182071387158543e-07, + "loss": 0.88, + "step": 5708 + }, + { + "epoch": 0.9571832756994656, + "grad_norm": 0.7076917886734009, + "learning_rate": 1.01049188134148e-07, + "loss": 0.831, + "step": 5709 + }, + { + "epoch": 0.9573509378602116, + "grad_norm": 0.502890408039093, + "learning_rate": 1.0028058171988153e-07, + "loss": 0.8811, + "step": 5710 + }, + { + "epoch": 0.9575186000209578, + "grad_norm": 0.7879444360733032, + "learning_rate": 9.951489485545696e-08, + "loss": 0.836, + "step": 5711 + }, + { + "epoch": 0.9576862621817038, + "grad_norm": 0.5187103152275085, + "learning_rate": 9.875212776668586e-08, + "loss": 0.8376, + "step": 5712 + }, + { + "epoch": 0.95785392434245, + "grad_norm": 0.4570627808570862, + "learning_rate": 9.799228067851829e-08, + "loss": 0.8397, + "step": 5713 + }, + { + "epoch": 0.958021586503196, + "grad_norm": 0.41080382466316223, + "learning_rate": 9.723535381504167e-08, + "loss": 0.8595, + "step": 5714 + }, + { + "epoch": 0.9581892486639422, + "grad_norm": 0.7080820202827454, + "learning_rate": 9.6481347399483e-08, + "loss": 0.8357, + "step": 5715 + }, + { + "epoch": 0.9583569108246882, + "grad_norm": 0.41043514013290405, + "learning_rate": 9.573026165420773e-08, + "loss": 0.8083, + "step": 5716 + }, + { + "epoch": 0.9585245729854344, + "grad_norm": 0.48574718832969666, + "learning_rate": 9.498209680072202e-08, + "loss": 0.8493, + "step": 5717 + }, + { + "epoch": 0.9586922351461804, + "grad_norm": 0.8123177886009216, + "learning_rate": 9.423685305966824e-08, + "loss": 0.8967, + "step": 5718 + }, + { + "epoch": 0.9588598973069266, + "grad_norm": 0.4755695164203644, + "learning_rate": 9.349453065082725e-08, + "loss": 0.7297, + "step": 5719 + }, + { + "epoch": 0.9590275594676726, + "grad_norm": 0.3649205267429352, + "learning_rate": 9.275512979311952e-08, + "loss": 0.8195, + "step": 5720 + }, + { + "epoch": 0.9591952216284187, + "grad_norm": 0.7999193668365479, + "learning_rate": 9.201865070460392e-08, + "loss": 0.8144, + "step": 5721 + }, + { + "epoch": 0.9593628837891648, + "grad_norm": 0.3566318452358246, + "learning_rate": 9.128509360247562e-08, + "loss": 0.7668, + "step": 5722 + }, + { + "epoch": 0.9595305459499109, + "grad_norm": 0.5108819007873535, + "learning_rate": 9.055445870307267e-08, + "loss": 0.7538, + "step": 5723 + }, + { + "epoch": 0.959698208110657, + "grad_norm": 0.4764617681503296, + "learning_rate": 8.982674622186605e-08, + "loss": 0.843, + "step": 5724 + }, + { + "epoch": 0.9598658702714031, + "grad_norm": 0.35026487708091736, + "learning_rate": 8.910195637346741e-08, + "loss": 0.8462, + "step": 5725 + }, + { + "epoch": 0.9600335324321492, + "grad_norm": 0.3958798050880432, + "learning_rate": 8.83800893716269e-08, + "loss": 0.8281, + "step": 5726 + }, + { + "epoch": 0.9602011945928953, + "grad_norm": 0.6569202542304993, + "learning_rate": 8.76611454292331e-08, + "loss": 0.7438, + "step": 5727 + }, + { + "epoch": 0.9603688567536414, + "grad_norm": 1.2969248294830322, + "learning_rate": 8.694512475830863e-08, + "loss": 0.76, + "step": 5728 + }, + { + "epoch": 0.9605365189143875, + "grad_norm": 0.44533786177635193, + "learning_rate": 8.623202757001792e-08, + "loss": 0.8143, + "step": 5729 + }, + { + "epoch": 0.9607041810751336, + "grad_norm": 0.5025610327720642, + "learning_rate": 8.552185407466384e-08, + "loss": 0.8676, + "step": 5730 + }, + { + "epoch": 0.9608718432358797, + "grad_norm": 0.46003156900405884, + "learning_rate": 8.481460448168333e-08, + "loss": 0.8146, + "step": 5731 + }, + { + "epoch": 0.9610395053966257, + "grad_norm": 0.3955024480819702, + "learning_rate": 8.411027899965507e-08, + "loss": 0.8441, + "step": 5732 + }, + { + "epoch": 0.9612071675573719, + "grad_norm": 0.4125998020172119, + "learning_rate": 8.34088778362907e-08, + "loss": 0.7981, + "step": 5733 + }, + { + "epoch": 0.9613748297181179, + "grad_norm": 0.509367823600769, + "learning_rate": 8.271040119844476e-08, + "loss": 0.8146, + "step": 5734 + }, + { + "epoch": 0.9615424918788641, + "grad_norm": 0.5742055773735046, + "learning_rate": 8.201484929210357e-08, + "loss": 0.8677, + "step": 5735 + }, + { + "epoch": 0.9617101540396101, + "grad_norm": 0.5999758243560791, + "learning_rate": 8.13222223223975e-08, + "loss": 0.8455, + "step": 5736 + }, + { + "epoch": 0.9618778162003563, + "grad_norm": 0.3258301615715027, + "learning_rate": 8.063252049358983e-08, + "loss": 0.8064, + "step": 5737 + }, + { + "epoch": 0.9620454783611023, + "grad_norm": 0.5386455059051514, + "learning_rate": 7.994574400908007e-08, + "loss": 0.8215, + "step": 5738 + }, + { + "epoch": 0.9622131405218485, + "grad_norm": 0.452926903963089, + "learning_rate": 7.926189307140952e-08, + "loss": 0.8158, + "step": 5739 + }, + { + "epoch": 0.9623808026825946, + "grad_norm": 0.45257359743118286, + "learning_rate": 7.858096788225245e-08, + "loss": 0.9341, + "step": 5740 + }, + { + "epoch": 0.9625484648433407, + "grad_norm": 0.7201059460639954, + "learning_rate": 7.790296864242374e-08, + "loss": 0.8401, + "step": 5741 + }, + { + "epoch": 0.9627161270040868, + "grad_norm": 0.4358721077442169, + "learning_rate": 7.722789555187238e-08, + "loss": 0.7329, + "step": 5742 + }, + { + "epoch": 0.9628837891648329, + "grad_norm": 0.42482608556747437, + "learning_rate": 7.655574880968685e-08, + "loss": 0.8294, + "step": 5743 + }, + { + "epoch": 0.963051451325579, + "grad_norm": 0.38124197721481323, + "learning_rate": 7.588652861409084e-08, + "loss": 0.7149, + "step": 5744 + }, + { + "epoch": 0.963219113486325, + "grad_norm": 0.4425771236419678, + "learning_rate": 7.522023516244648e-08, + "loss": 0.759, + "step": 5745 + }, + { + "epoch": 0.9633867756470712, + "grad_norm": 0.567328929901123, + "learning_rate": 7.455686865125212e-08, + "loss": 0.8754, + "step": 5746 + }, + { + "epoch": 0.9635544378078172, + "grad_norm": 0.35585835576057434, + "learning_rate": 7.389642927614127e-08, + "loss": 0.8482, + "step": 5747 + }, + { + "epoch": 0.9637220999685634, + "grad_norm": 0.3198508024215698, + "learning_rate": 7.323891723188592e-08, + "loss": 0.829, + "step": 5748 + }, + { + "epoch": 0.9638897621293094, + "grad_norm": 0.4167691171169281, + "learning_rate": 7.258433271239651e-08, + "loss": 0.8377, + "step": 5749 + }, + { + "epoch": 0.9640574242900556, + "grad_norm": 0.5370296835899353, + "learning_rate": 7.193267591071529e-08, + "loss": 0.8478, + "step": 5750 + }, + { + "epoch": 0.9642250864508016, + "grad_norm": 0.8849428296089172, + "learning_rate": 7.12839470190263e-08, + "loss": 0.952, + "step": 5751 + }, + { + "epoch": 0.9643927486115478, + "grad_norm": 0.40987876057624817, + "learning_rate": 7.063814622864651e-08, + "loss": 0.8097, + "step": 5752 + }, + { + "epoch": 0.9645604107722938, + "grad_norm": 0.4275193512439728, + "learning_rate": 6.999527373003134e-08, + "loss": 0.7957, + "step": 5753 + }, + { + "epoch": 0.96472807293304, + "grad_norm": 0.49265700578689575, + "learning_rate": 6.935532971277026e-08, + "loss": 0.8611, + "step": 5754 + }, + { + "epoch": 0.964895735093786, + "grad_norm": 0.4670751094818115, + "learning_rate": 6.871831436559339e-08, + "loss": 0.8502, + "step": 5755 + }, + { + "epoch": 0.9650633972545322, + "grad_norm": 0.419053852558136, + "learning_rate": 6.80842278763616e-08, + "loss": 0.8133, + "step": 5756 + }, + { + "epoch": 0.9652310594152782, + "grad_norm": 0.3705775737762451, + "learning_rate": 6.745307043207638e-08, + "loss": 0.7635, + "step": 5757 + }, + { + "epoch": 0.9653987215760244, + "grad_norm": 0.40138015151023865, + "learning_rate": 6.682484221887331e-08, + "loss": 0.8773, + "step": 5758 + }, + { + "epoch": 0.9655663837367704, + "grad_norm": 0.4074966013431549, + "learning_rate": 6.619954342202528e-08, + "loss": 0.8126, + "step": 5759 + }, + { + "epoch": 0.9657340458975165, + "grad_norm": 0.5536928772926331, + "learning_rate": 6.557717422594035e-08, + "loss": 0.8109, + "step": 5760 + }, + { + "epoch": 0.9659017080582626, + "grad_norm": 0.4149467647075653, + "learning_rate": 6.495773481416278e-08, + "loss": 0.7873, + "step": 5761 + }, + { + "epoch": 0.9660693702190087, + "grad_norm": 0.4304526448249817, + "learning_rate": 6.434122536937204e-08, + "loss": 0.8521, + "step": 5762 + }, + { + "epoch": 0.9662370323797548, + "grad_norm": 0.6651942729949951, + "learning_rate": 6.3727646073386e-08, + "loss": 0.8533, + "step": 5763 + }, + { + "epoch": 0.9664046945405009, + "grad_norm": 0.415158212184906, + "learning_rate": 6.311699710715657e-08, + "loss": 0.8647, + "step": 5764 + }, + { + "epoch": 0.966572356701247, + "grad_norm": 0.38120999932289124, + "learning_rate": 6.250927865077083e-08, + "loss": 0.7783, + "step": 5765 + }, + { + "epoch": 0.9667400188619931, + "grad_norm": 1.4348599910736084, + "learning_rate": 6.190449088345318e-08, + "loss": 0.836, + "step": 5766 + }, + { + "epoch": 0.9669076810227392, + "grad_norm": 0.3521517217159271, + "learning_rate": 6.130263398356318e-08, + "loss": 0.8043, + "step": 5767 + }, + { + "epoch": 0.9670753431834853, + "grad_norm": 0.3433744013309479, + "learning_rate": 6.070370812859438e-08, + "loss": 0.8868, + "step": 5768 + }, + { + "epoch": 0.9672430053442314, + "grad_norm": 0.46505022048950195, + "learning_rate": 6.010771349517885e-08, + "loss": 1.0081, + "step": 5769 + }, + { + "epoch": 0.9674106675049775, + "grad_norm": 0.29260390996932983, + "learning_rate": 5.951465025908265e-08, + "loss": 0.7565, + "step": 5770 + }, + { + "epoch": 0.9675783296657235, + "grad_norm": 0.5390307903289795, + "learning_rate": 5.892451859520698e-08, + "loss": 0.8892, + "step": 5771 + }, + { + "epoch": 0.9677459918264697, + "grad_norm": 0.5621128678321838, + "learning_rate": 5.83373186775904e-08, + "loss": 0.8231, + "step": 5772 + }, + { + "epoch": 0.9679136539872157, + "grad_norm": 0.41262245178222656, + "learning_rate": 5.775305067940329e-08, + "loss": 0.8835, + "step": 5773 + }, + { + "epoch": 0.9680813161479619, + "grad_norm": 0.5214173197746277, + "learning_rate": 5.717171477295447e-08, + "loss": 0.9254, + "step": 5774 + }, + { + "epoch": 0.9682489783087079, + "grad_norm": 0.5759170651435852, + "learning_rate": 5.659331112968791e-08, + "loss": 0.9011, + "step": 5775 + }, + { + "epoch": 0.9684166404694541, + "grad_norm": 0.5135636925697327, + "learning_rate": 5.6017839920180506e-08, + "loss": 0.8083, + "step": 5776 + }, + { + "epoch": 0.9685843026302001, + "grad_norm": 0.3557618260383606, + "learning_rate": 5.544530131414649e-08, + "loss": 0.9134, + "step": 5777 + }, + { + "epoch": 0.9687519647909463, + "grad_norm": 0.5884382724761963, + "learning_rate": 5.4875695480435255e-08, + "loss": 0.8587, + "step": 5778 + }, + { + "epoch": 0.9689196269516923, + "grad_norm": 0.4821484088897705, + "learning_rate": 5.4309022587029084e-08, + "loss": 0.8129, + "step": 5779 + }, + { + "epoch": 0.9690872891124385, + "grad_norm": 0.6065845489501953, + "learning_rate": 5.374528280104874e-08, + "loss": 0.8444, + "step": 5780 + }, + { + "epoch": 0.9692549512731845, + "grad_norm": 0.4409104287624359, + "learning_rate": 5.318447628874679e-08, + "loss": 0.903, + "step": 5781 + }, + { + "epoch": 0.9694226134339307, + "grad_norm": 0.3960440456867218, + "learning_rate": 5.262660321551316e-08, + "loss": 0.7283, + "step": 5782 + }, + { + "epoch": 0.9695902755946767, + "grad_norm": 0.40227869153022766, + "learning_rate": 5.2071663745869585e-08, + "loss": 0.7678, + "step": 5783 + }, + { + "epoch": 0.9697579377554228, + "grad_norm": 0.4411965012550354, + "learning_rate": 5.151965804347625e-08, + "loss": 0.8097, + "step": 5784 + }, + { + "epoch": 0.9699255999161689, + "grad_norm": 0.5686452984809875, + "learning_rate": 5.0970586271126274e-08, + "loss": 0.8054, + "step": 5785 + }, + { + "epoch": 0.970093262076915, + "grad_norm": 0.39256587624549866, + "learning_rate": 5.0424448590746796e-08, + "loss": 0.8827, + "step": 5786 + }, + { + "epoch": 0.9702609242376611, + "grad_norm": 0.3850398659706116, + "learning_rate": 4.98812451634012e-08, + "loss": 0.8567, + "step": 5787 + }, + { + "epoch": 0.9704285863984072, + "grad_norm": 0.35744708776474, + "learning_rate": 4.93409761492869e-08, + "loss": 0.8392, + "step": 5788 + }, + { + "epoch": 0.9705962485591533, + "grad_norm": 0.4973679780960083, + "learning_rate": 4.880364170773533e-08, + "loss": 0.8695, + "step": 5789 + }, + { + "epoch": 0.9707639107198994, + "grad_norm": 0.3606968820095062, + "learning_rate": 4.8269241997214166e-08, + "loss": 0.8918, + "step": 5790 + }, + { + "epoch": 0.9709315728806455, + "grad_norm": 0.4637780487537384, + "learning_rate": 4.773777717532513e-08, + "loss": 0.8227, + "step": 5791 + }, + { + "epoch": 0.9710992350413916, + "grad_norm": 0.9876329898834229, + "learning_rate": 4.720924739880062e-08, + "loss": 0.7865, + "step": 5792 + }, + { + "epoch": 0.9712668972021377, + "grad_norm": 0.5332596302032471, + "learning_rate": 4.6683652823513725e-08, + "loss": 0.8898, + "step": 5793 + }, + { + "epoch": 0.9714345593628838, + "grad_norm": 0.47035887837409973, + "learning_rate": 4.616099360446713e-08, + "loss": 0.8151, + "step": 5794 + }, + { + "epoch": 0.9716022215236298, + "grad_norm": 0.40923890471458435, + "learning_rate": 4.5641269895800864e-08, + "loss": 0.9865, + "step": 5795 + }, + { + "epoch": 0.971769883684376, + "grad_norm": 0.6284613013267517, + "learning_rate": 4.512448185078677e-08, + "loss": 0.9175, + "step": 5796 + }, + { + "epoch": 0.971937545845122, + "grad_norm": 0.4579513967037201, + "learning_rate": 4.461062962183183e-08, + "loss": 0.8939, + "step": 5797 + }, + { + "epoch": 0.9721052080058682, + "grad_norm": 0.396099328994751, + "learning_rate": 4.4099713360478133e-08, + "loss": 0.8508, + "step": 5798 + }, + { + "epoch": 0.9722728701666142, + "grad_norm": 0.42241689562797546, + "learning_rate": 4.359173321740184e-08, + "loss": 0.7861, + "step": 5799 + }, + { + "epoch": 0.9724405323273604, + "grad_norm": 0.35261476039886475, + "learning_rate": 4.308668934241089e-08, + "loss": 0.7617, + "step": 5800 + }, + { + "epoch": 0.9726081944881064, + "grad_norm": 0.387504518032074, + "learning_rate": 4.2584581884450584e-08, + "loss": 0.8291, + "step": 5801 + }, + { + "epoch": 0.9727758566488526, + "grad_norm": 0.4456618130207062, + "learning_rate": 4.208541099159691e-08, + "loss": 0.8707, + "step": 5802 + }, + { + "epoch": 0.9729435188095986, + "grad_norm": 0.5461599826812744, + "learning_rate": 4.158917681106322e-08, + "loss": 0.8829, + "step": 5803 + }, + { + "epoch": 0.9731111809703448, + "grad_norm": 0.4105444848537445, + "learning_rate": 4.109587948919469e-08, + "loss": 0.858, + "step": 5804 + }, + { + "epoch": 0.9732788431310908, + "grad_norm": 0.937830924987793, + "learning_rate": 4.0605519171469375e-08, + "loss": 0.7827, + "step": 5805 + }, + { + "epoch": 0.973446505291837, + "grad_norm": 0.5319908857345581, + "learning_rate": 4.011809600250272e-08, + "loss": 0.7846, + "step": 5806 + }, + { + "epoch": 0.973614167452583, + "grad_norm": 0.3267498314380646, + "learning_rate": 3.9633610126040836e-08, + "loss": 0.8895, + "step": 5807 + }, + { + "epoch": 0.9737818296133292, + "grad_norm": 0.8419097065925598, + "learning_rate": 3.9152061684964993e-08, + "loss": 0.8483, + "step": 5808 + }, + { + "epoch": 0.9739494917740752, + "grad_norm": 0.5133106112480164, + "learning_rate": 3.8673450821288264e-08, + "loss": 0.8842, + "step": 5809 + }, + { + "epoch": 0.9741171539348213, + "grad_norm": 0.8418963551521301, + "learning_rate": 3.8197777676161064e-08, + "loss": 0.8217, + "step": 5810 + }, + { + "epoch": 0.9742848160955674, + "grad_norm": 0.37620463967323303, + "learning_rate": 3.772504238986452e-08, + "loss": 0.7553, + "step": 5811 + }, + { + "epoch": 0.9744524782563135, + "grad_norm": 0.4085863530635834, + "learning_rate": 3.725524510181378e-08, + "loss": 0.8598, + "step": 5812 + }, + { + "epoch": 0.9746201404170596, + "grad_norm": 0.3842092752456665, + "learning_rate": 3.6788385950558e-08, + "loss": 0.9393, + "step": 5813 + }, + { + "epoch": 0.9747878025778057, + "grad_norm": 0.4535366892814636, + "learning_rate": 3.63244650737804e-08, + "loss": 0.8294, + "step": 5814 + }, + { + "epoch": 0.9749554647385518, + "grad_norm": 0.5589142441749573, + "learning_rate": 3.586348260829486e-08, + "loss": 0.8977, + "step": 5815 + }, + { + "epoch": 0.9751231268992979, + "grad_norm": 0.4944474399089813, + "learning_rate": 3.540543869005375e-08, + "loss": 0.8625, + "step": 5816 + }, + { + "epoch": 0.975290789060044, + "grad_norm": 0.3996366858482361, + "learning_rate": 3.4950333454137895e-08, + "loss": 0.8303, + "step": 5817 + }, + { + "epoch": 0.9754584512207901, + "grad_norm": 0.44617748260498047, + "learning_rate": 3.449816703476438e-08, + "loss": 0.892, + "step": 5818 + }, + { + "epoch": 0.9756261133815362, + "grad_norm": 0.45459893345832825, + "learning_rate": 3.40489395652821e-08, + "loss": 0.8965, + "step": 5819 + }, + { + "epoch": 0.9757937755422823, + "grad_norm": 0.3187973201274872, + "learning_rate": 3.360265117817396e-08, + "loss": 0.7541, + "step": 5820 + }, + { + "epoch": 0.9759614377030283, + "grad_norm": 0.38300496339797974, + "learning_rate": 3.315930200505579e-08, + "loss": 0.8877, + "step": 5821 + }, + { + "epoch": 0.9761290998637745, + "grad_norm": 0.5122457146644592, + "learning_rate": 3.271889217667745e-08, + "loss": 0.8554, + "step": 5822 + }, + { + "epoch": 0.9762967620245205, + "grad_norm": 0.3872210383415222, + "learning_rate": 3.22814218229206e-08, + "loss": 0.7912, + "step": 5823 + }, + { + "epoch": 0.9764644241852667, + "grad_norm": 0.48558467626571655, + "learning_rate": 3.1846891072802034e-08, + "loss": 0.8185, + "step": 5824 + }, + { + "epoch": 0.9766320863460127, + "grad_norm": 0.4107375144958496, + "learning_rate": 3.141530005446702e-08, + "loss": 0.8521, + "step": 5825 + }, + { + "epoch": 0.9767997485067589, + "grad_norm": 0.5492359399795532, + "learning_rate": 3.098664889520042e-08, + "loss": 0.8031, + "step": 5826 + }, + { + "epoch": 0.976967410667505, + "grad_norm": 0.3595879077911377, + "learning_rate": 3.0560937721415554e-08, + "loss": 0.9382, + "step": 5827 + }, + { + "epoch": 0.9771350728282511, + "grad_norm": 0.6469846367835999, + "learning_rate": 3.013816665865976e-08, + "loss": 0.8775, + "step": 5828 + }, + { + "epoch": 0.9773027349889972, + "grad_norm": 0.45741432905197144, + "learning_rate": 2.971833583161332e-08, + "loss": 0.8431, + "step": 5829 + }, + { + "epoch": 0.9774703971497433, + "grad_norm": 0.37668368220329285, + "learning_rate": 2.9301445364089408e-08, + "loss": 0.8204, + "step": 5830 + }, + { + "epoch": 0.9776380593104894, + "grad_norm": 0.3521329164505005, + "learning_rate": 2.8887495379035235e-08, + "loss": 0.8736, + "step": 5831 + }, + { + "epoch": 0.9778057214712355, + "grad_norm": 0.42450135946273804, + "learning_rate": 2.8476485998528703e-08, + "loss": 0.8061, + "step": 5832 + }, + { + "epoch": 0.9779733836319816, + "grad_norm": 0.43218979239463806, + "learning_rate": 2.806841734378174e-08, + "loss": 0.8153, + "step": 5833 + }, + { + "epoch": 0.9781410457927276, + "grad_norm": 0.4481380581855774, + "learning_rate": 2.7663289535140302e-08, + "loss": 0.7531, + "step": 5834 + }, + { + "epoch": 0.9783087079534738, + "grad_norm": 0.41322800517082214, + "learning_rate": 2.7261102692078822e-08, + "loss": 0.8489, + "step": 5835 + }, + { + "epoch": 0.9784763701142198, + "grad_norm": 0.45851191878318787, + "learning_rate": 2.6861856933209086e-08, + "loss": 0.7952, + "step": 5836 + }, + { + "epoch": 0.978644032274966, + "grad_norm": 0.5079235434532166, + "learning_rate": 2.6465552376274683e-08, + "loss": 0.8555, + "step": 5837 + }, + { + "epoch": 0.978811694435712, + "grad_norm": 0.38337695598602295, + "learning_rate": 2.6072189138147686e-08, + "loss": 0.7607, + "step": 5838 + }, + { + "epoch": 0.9789793565964582, + "grad_norm": 0.40837207436561584, + "learning_rate": 2.5681767334837514e-08, + "loss": 0.8458, + "step": 5839 + }, + { + "epoch": 0.9791470187572042, + "grad_norm": 0.35640430450439453, + "learning_rate": 2.5294287081485404e-08, + "loss": 0.8652, + "step": 5840 + }, + { + "epoch": 0.9793146809179504, + "grad_norm": 0.8105623722076416, + "learning_rate": 2.4909748492362162e-08, + "loss": 0.873, + "step": 5841 + }, + { + "epoch": 0.9794823430786964, + "grad_norm": 0.4976637661457062, + "learning_rate": 2.4528151680873748e-08, + "loss": 0.9144, + "step": 5842 + }, + { + "epoch": 0.9796500052394426, + "grad_norm": 0.5497066378593445, + "learning_rate": 2.4149496759557912e-08, + "loss": 0.8361, + "step": 5843 + }, + { + "epoch": 0.9798176674001886, + "grad_norm": 0.5602542757987976, + "learning_rate": 2.377378384008422e-08, + "loss": 0.8222, + "step": 5844 + }, + { + "epoch": 0.9799853295609348, + "grad_norm": 0.3840111494064331, + "learning_rate": 2.340101303325626e-08, + "loss": 0.8269, + "step": 5845 + }, + { + "epoch": 0.9801529917216808, + "grad_norm": 0.5249937176704407, + "learning_rate": 2.3031184449007204e-08, + "loss": 0.8154, + "step": 5846 + }, + { + "epoch": 0.980320653882427, + "grad_norm": 0.8526900410652161, + "learning_rate": 2.2664298196405367e-08, + "loss": 0.8309, + "step": 5847 + }, + { + "epoch": 0.980488316043173, + "grad_norm": 1.0609358549118042, + "learning_rate": 2.2300354383648636e-08, + "loss": 0.771, + "step": 5848 + }, + { + "epoch": 0.9806559782039191, + "grad_norm": 0.36970868706703186, + "learning_rate": 2.193935311807005e-08, + "loss": 0.8063, + "step": 5849 + }, + { + "epoch": 0.9808236403646652, + "grad_norm": 0.4812318682670593, + "learning_rate": 2.1581294506132223e-08, + "loss": 0.8783, + "step": 5850 + }, + { + "epoch": 0.9809913025254113, + "grad_norm": 0.39542368054389954, + "learning_rate": 2.1226178653432905e-08, + "loss": 0.8123, + "step": 5851 + }, + { + "epoch": 0.9811589646861574, + "grad_norm": 0.4277333617210388, + "learning_rate": 2.0874005664697218e-08, + "loss": 0.8418, + "step": 5852 + }, + { + "epoch": 0.9813266268469035, + "grad_norm": 0.5052667856216431, + "learning_rate": 2.052477564378763e-08, + "loss": 0.7711, + "step": 5853 + }, + { + "epoch": 0.9814942890076496, + "grad_norm": 0.5133655071258545, + "learning_rate": 2.0178488693695096e-08, + "loss": 0.9144, + "step": 5854 + }, + { + "epoch": 0.9816619511683957, + "grad_norm": 0.4648545980453491, + "learning_rate": 1.98351449165457e-08, + "loss": 0.8852, + "step": 5855 + }, + { + "epoch": 0.9818296133291418, + "grad_norm": 1.2075058221817017, + "learning_rate": 1.9494744413594003e-08, + "loss": 0.7866, + "step": 5856 + }, + { + "epoch": 0.9819972754898879, + "grad_norm": 0.4560854434967041, + "learning_rate": 1.9157287285228587e-08, + "loss": 0.8429, + "step": 5857 + }, + { + "epoch": 0.982164937650634, + "grad_norm": 0.3814972937107086, + "learning_rate": 1.8822773630969847e-08, + "loss": 0.8514, + "step": 5858 + }, + { + "epoch": 0.9823325998113801, + "grad_norm": 0.38170596957206726, + "learning_rate": 1.8491203549471092e-08, + "loss": 0.7808, + "step": 5859 + }, + { + "epoch": 0.9825002619721261, + "grad_norm": 0.40041080117225647, + "learning_rate": 1.8162577138515215e-08, + "loss": 0.8326, + "step": 5860 + }, + { + "epoch": 0.9826679241328723, + "grad_norm": 0.37729009985923767, + "learning_rate": 1.783689449501802e-08, + "loss": 0.8323, + "step": 5861 + }, + { + "epoch": 0.9828355862936183, + "grad_norm": 0.4086211919784546, + "learning_rate": 1.7514155715028236e-08, + "loss": 0.8604, + "step": 5862 + }, + { + "epoch": 0.9830032484543645, + "grad_norm": 0.3598542809486389, + "learning_rate": 1.7194360893725282e-08, + "loss": 0.8003, + "step": 5863 + }, + { + "epoch": 0.9831709106151105, + "grad_norm": 0.5794903039932251, + "learning_rate": 1.6877510125420384e-08, + "loss": 0.9411, + "step": 5864 + }, + { + "epoch": 0.9833385727758567, + "grad_norm": 0.390186071395874, + "learning_rate": 1.656360350355768e-08, + "loss": 0.7728, + "step": 5865 + }, + { + "epoch": 0.9835062349366027, + "grad_norm": 0.4856814444065094, + "learning_rate": 1.6252641120710898e-08, + "loss": 0.9004, + "step": 5866 + }, + { + "epoch": 0.9836738970973489, + "grad_norm": 0.5970719456672668, + "learning_rate": 1.5944623068586683e-08, + "loss": 0.872, + "step": 5867 + }, + { + "epoch": 0.9838415592580949, + "grad_norm": 0.3844272494316101, + "learning_rate": 1.5639549438023483e-08, + "loss": 0.6954, + "step": 5868 + }, + { + "epoch": 0.9840092214188411, + "grad_norm": 0.3902965188026428, + "learning_rate": 1.5337420318992656e-08, + "loss": 0.8182, + "step": 5869 + }, + { + "epoch": 0.9841768835795871, + "grad_norm": 0.42631059885025024, + "learning_rate": 1.5038235800595158e-08, + "loss": 0.8431, + "step": 5870 + }, + { + "epoch": 0.9843445457403333, + "grad_norm": 0.4387223720550537, + "learning_rate": 1.4741995971062627e-08, + "loss": 0.7704, + "step": 5871 + }, + { + "epoch": 0.9845122079010793, + "grad_norm": 0.4834582805633545, + "learning_rate": 1.4448700917761849e-08, + "loss": 0.9173, + "step": 5872 + }, + { + "epoch": 0.9846798700618254, + "grad_norm": 0.4982442557811737, + "learning_rate": 1.4158350727188075e-08, + "loss": 0.7943, + "step": 5873 + }, + { + "epoch": 0.9848475322225715, + "grad_norm": 0.37011417746543884, + "learning_rate": 1.3870945484969478e-08, + "loss": 0.9477, + "step": 5874 + }, + { + "epoch": 0.9850151943833176, + "grad_norm": 0.49638018012046814, + "learning_rate": 1.3586485275864925e-08, + "loss": 0.7223, + "step": 5875 + }, + { + "epoch": 0.9851828565440637, + "grad_norm": 0.33733877539634705, + "learning_rate": 1.3304970183766198e-08, + "loss": 0.8092, + "step": 5876 + }, + { + "epoch": 0.9853505187048098, + "grad_norm": 0.3381383717060089, + "learning_rate": 1.3026400291695774e-08, + "loss": 0.8701, + "step": 5877 + }, + { + "epoch": 0.9855181808655559, + "grad_norm": 0.35699698328971863, + "learning_rate": 1.2750775681805717e-08, + "loss": 0.8487, + "step": 5878 + }, + { + "epoch": 0.985685843026302, + "grad_norm": 0.4587336480617523, + "learning_rate": 1.2478096435383225e-08, + "loss": 0.9753, + "step": 5879 + }, + { + "epoch": 0.9858535051870481, + "grad_norm": 0.5707885026931763, + "learning_rate": 1.2208362632842863e-08, + "loss": 0.8661, + "step": 5880 + }, + { + "epoch": 0.9860211673477942, + "grad_norm": 0.40471991896629333, + "learning_rate": 1.1941574353733221e-08, + "loss": 0.9193, + "step": 5881 + }, + { + "epoch": 0.9861888295085403, + "grad_norm": 0.4383912980556488, + "learning_rate": 1.1677731676733584e-08, + "loss": 0.8396, + "step": 5882 + }, + { + "epoch": 0.9863564916692864, + "grad_norm": 0.4767834544181824, + "learning_rate": 1.1416834679655042e-08, + "loss": 0.8994, + "step": 5883 + }, + { + "epoch": 0.9865241538300324, + "grad_norm": 0.593654215335846, + "learning_rate": 1.1158883439438273e-08, + "loss": 0.7644, + "step": 5884 + }, + { + "epoch": 0.9866918159907786, + "grad_norm": 0.38353127241134644, + "learning_rate": 1.0903878032156867e-08, + "loss": 0.8433, + "step": 5885 + }, + { + "epoch": 0.9868594781515246, + "grad_norm": 0.43884798884391785, + "learning_rate": 1.0651818533015113e-08, + "loss": 0.7387, + "step": 5886 + }, + { + "epoch": 0.9870271403122708, + "grad_norm": 0.3942774534225464, + "learning_rate": 1.0402705016349102e-08, + "loss": 0.8686, + "step": 5887 + }, + { + "epoch": 0.9871948024730168, + "grad_norm": 0.36644846200942993, + "learning_rate": 1.0156537555623402e-08, + "loss": 0.7794, + "step": 5888 + }, + { + "epoch": 0.987362464633763, + "grad_norm": 0.4375738203525543, + "learning_rate": 9.913316223437718e-09, + "loss": 0.831, + "step": 5889 + }, + { + "epoch": 0.987530126794509, + "grad_norm": 0.49118849635124207, + "learning_rate": 9.67304109152023e-09, + "loss": 0.8345, + "step": 5890 + }, + { + "epoch": 0.9876977889552552, + "grad_norm": 0.4413753151893616, + "learning_rate": 9.43571223073092e-09, + "loss": 0.8346, + "step": 5891 + }, + { + "epoch": 0.9878654511160012, + "grad_norm": 0.4078384339809418, + "learning_rate": 9.20132971106269e-09, + "loss": 0.805, + "step": 5892 + }, + { + "epoch": 0.9880331132767474, + "grad_norm": 0.42576390504837036, + "learning_rate": 8.969893601634694e-09, + "loss": 0.8319, + "step": 5893 + }, + { + "epoch": 0.9882007754374934, + "grad_norm": 0.46708157658576965, + "learning_rate": 8.741403970703443e-09, + "loss": 0.8839, + "step": 5894 + }, + { + "epoch": 0.9883684375982396, + "grad_norm": 0.34723997116088867, + "learning_rate": 8.515860885651706e-09, + "loss": 0.798, + "step": 5895 + }, + { + "epoch": 0.9885360997589856, + "grad_norm": 0.37282437086105347, + "learning_rate": 8.293264412995161e-09, + "loss": 0.8861, + "step": 5896 + }, + { + "epoch": 0.9887037619197317, + "grad_norm": 0.42206084728240967, + "learning_rate": 8.073614618380188e-09, + "loss": 0.8341, + "step": 5897 + }, + { + "epoch": 0.9888714240804778, + "grad_norm": 0.4986555278301239, + "learning_rate": 7.856911566584968e-09, + "loss": 0.7804, + "step": 5898 + }, + { + "epoch": 0.9890390862412239, + "grad_norm": 1.1532371044158936, + "learning_rate": 7.64315532151727e-09, + "loss": 0.7834, + "step": 5899 + }, + { + "epoch": 0.98920674840197, + "grad_norm": 1.007047414779663, + "learning_rate": 7.432345946216668e-09, + "loss": 0.837, + "step": 5900 + }, + { + "epoch": 0.9893744105627161, + "grad_norm": 0.46277493238449097, + "learning_rate": 7.22448350285343e-09, + "loss": 0.8525, + "step": 5901 + }, + { + "epoch": 0.9895420727234622, + "grad_norm": 0.38527214527130127, + "learning_rate": 7.01956805272852e-09, + "loss": 0.7731, + "step": 5902 + }, + { + "epoch": 0.9897097348842083, + "grad_norm": 0.45211055874824524, + "learning_rate": 6.817599656273599e-09, + "loss": 0.9063, + "step": 5903 + }, + { + "epoch": 0.9898773970449544, + "grad_norm": 0.42291173338890076, + "learning_rate": 6.6185783730521315e-09, + "loss": 0.7942, + "step": 5904 + }, + { + "epoch": 0.9900450592057005, + "grad_norm": 0.35738232731819153, + "learning_rate": 6.422504261759388e-09, + "loss": 0.7929, + "step": 5905 + }, + { + "epoch": 0.9902127213664466, + "grad_norm": 0.5838137865066528, + "learning_rate": 6.229377380218005e-09, + "loss": 0.845, + "step": 5906 + }, + { + "epoch": 0.9903803835271927, + "grad_norm": 0.4058472812175751, + "learning_rate": 6.039197785383533e-09, + "loss": 0.8274, + "step": 5907 + }, + { + "epoch": 0.9905480456879387, + "grad_norm": 0.44134339690208435, + "learning_rate": 5.8519655333444394e-09, + "loss": 0.8507, + "step": 5908 + }, + { + "epoch": 0.9907157078486849, + "grad_norm": 0.34918412566185, + "learning_rate": 5.667680679315446e-09, + "loss": 0.7411, + "step": 5909 + }, + { + "epoch": 0.9908833700094309, + "grad_norm": 0.3565990924835205, + "learning_rate": 5.4863432776453006e-09, + "loss": 0.8605, + "step": 5910 + }, + { + "epoch": 0.9910510321701771, + "grad_norm": 0.7799057960510254, + "learning_rate": 5.307953381813447e-09, + "loss": 0.793, + "step": 5911 + }, + { + "epoch": 0.9912186943309232, + "grad_norm": 0.412158340215683, + "learning_rate": 5.132511044427801e-09, + "loss": 0.8211, + "step": 5912 + }, + { + "epoch": 0.9913863564916693, + "grad_norm": 0.4002961218357086, + "learning_rate": 4.960016317229199e-09, + "loss": 0.8245, + "step": 5913 + }, + { + "epoch": 0.9915540186524154, + "grad_norm": 0.4357675611972809, + "learning_rate": 4.7904692510891695e-09, + "loss": 0.7889, + "step": 5914 + }, + { + "epoch": 0.9917216808131615, + "grad_norm": 0.4863226115703583, + "learning_rate": 4.623869896008826e-09, + "loss": 0.8934, + "step": 5915 + }, + { + "epoch": 0.9918893429739076, + "grad_norm": 0.41823381185531616, + "learning_rate": 4.4602183011188685e-09, + "loss": 0.9074, + "step": 5916 + }, + { + "epoch": 0.9920570051346537, + "grad_norm": 0.3756341338157654, + "learning_rate": 4.299514514684022e-09, + "loss": 0.7179, + "step": 5917 + }, + { + "epoch": 0.9922246672953998, + "grad_norm": 0.34542134404182434, + "learning_rate": 4.141758584097488e-09, + "loss": 0.7625, + "step": 5918 + }, + { + "epoch": 0.9923923294561459, + "grad_norm": 0.43031448125839233, + "learning_rate": 3.986950555883162e-09, + "loss": 0.7759, + "step": 5919 + }, + { + "epoch": 0.992559991616892, + "grad_norm": 0.38881802558898926, + "learning_rate": 3.835090475695635e-09, + "loss": 0.8077, + "step": 5920 + }, + { + "epoch": 0.992727653777638, + "grad_norm": 0.3611317574977875, + "learning_rate": 3.686178388321304e-09, + "loss": 0.7856, + "step": 5921 + }, + { + "epoch": 0.9928953159383842, + "grad_norm": 0.47091519832611084, + "learning_rate": 3.540214337675041e-09, + "loss": 0.7925, + "step": 5922 + }, + { + "epoch": 0.9930629780991302, + "grad_norm": 0.46980783343315125, + "learning_rate": 3.3971983668035224e-09, + "loss": 0.8761, + "step": 5923 + }, + { + "epoch": 0.9932306402598764, + "grad_norm": 0.3938907980918884, + "learning_rate": 3.2571305178841218e-09, + "loss": 0.775, + "step": 5924 + }, + { + "epoch": 0.9933983024206224, + "grad_norm": 0.42583513259887695, + "learning_rate": 3.1200108322249066e-09, + "loss": 0.8806, + "step": 5925 + }, + { + "epoch": 0.9935659645813686, + "grad_norm": 0.4112643897533417, + "learning_rate": 2.985839350264641e-09, + "loss": 0.8161, + "step": 5926 + }, + { + "epoch": 0.9937336267421146, + "grad_norm": 0.4502149820327759, + "learning_rate": 2.854616111570563e-09, + "loss": 0.8438, + "step": 5927 + }, + { + "epoch": 0.9939012889028608, + "grad_norm": 0.5014587044715881, + "learning_rate": 2.726341154843937e-09, + "loss": 0.8073, + "step": 5928 + }, + { + "epoch": 0.9940689510636068, + "grad_norm": 0.42569249868392944, + "learning_rate": 2.6010145179133916e-09, + "loss": 0.8726, + "step": 5929 + }, + { + "epoch": 0.994236613224353, + "grad_norm": 0.5752376914024353, + "learning_rate": 2.4786362377393624e-09, + "loss": 0.7958, + "step": 5930 + }, + { + "epoch": 0.994404275385099, + "grad_norm": 0.439833402633667, + "learning_rate": 2.3592063504129793e-09, + "loss": 0.8239, + "step": 5931 + }, + { + "epoch": 0.9945719375458452, + "grad_norm": 0.3875184655189514, + "learning_rate": 2.242724891156067e-09, + "loss": 0.792, + "step": 5932 + }, + { + "epoch": 0.9947395997065912, + "grad_norm": 0.42492565512657166, + "learning_rate": 2.1291918943189272e-09, + "loss": 0.8352, + "step": 5933 + }, + { + "epoch": 0.9949072618673374, + "grad_norm": 0.5234774947166443, + "learning_rate": 2.0186073933858852e-09, + "loss": 0.7744, + "step": 5934 + }, + { + "epoch": 0.9950749240280834, + "grad_norm": 0.40094274282455444, + "learning_rate": 1.9109714209675223e-09, + "loss": 0.8669, + "step": 5935 + }, + { + "epoch": 0.9952425861888295, + "grad_norm": 0.5302832126617432, + "learning_rate": 1.8062840088095558e-09, + "loss": 0.8151, + "step": 5936 + }, + { + "epoch": 0.9954102483495756, + "grad_norm": 0.438462495803833, + "learning_rate": 1.7045451877828467e-09, + "loss": 0.843, + "step": 5937 + }, + { + "epoch": 0.9955779105103217, + "grad_norm": 0.489299476146698, + "learning_rate": 1.6057549878933931e-09, + "loss": 0.9438, + "step": 5938 + }, + { + "epoch": 0.9957455726710678, + "grad_norm": 0.4687711298465729, + "learning_rate": 1.5099134382756676e-09, + "loss": 0.8566, + "step": 5939 + }, + { + "epoch": 0.9959132348318139, + "grad_norm": 0.43907466530799866, + "learning_rate": 1.4170205671937275e-09, + "loss": 0.7816, + "step": 5940 + }, + { + "epoch": 0.99608089699256, + "grad_norm": 0.3608318865299225, + "learning_rate": 1.3270764020423265e-09, + "loss": 0.8377, + "step": 5941 + }, + { + "epoch": 0.9962485591533061, + "grad_norm": 0.4274871349334717, + "learning_rate": 1.2400809693480232e-09, + "loss": 0.8222, + "step": 5942 + }, + { + "epoch": 0.9964162213140522, + "grad_norm": 0.4970013201236725, + "learning_rate": 1.156034294766961e-09, + "loss": 0.8703, + "step": 5943 + }, + { + "epoch": 0.9965838834747983, + "grad_norm": 0.4293815791606903, + "learning_rate": 1.0749364030848697e-09, + "loss": 0.8397, + "step": 5944 + }, + { + "epoch": 0.9967515456355444, + "grad_norm": 0.451774537563324, + "learning_rate": 9.96787318218173e-10, + "loss": 0.8459, + "step": 5945 + }, + { + "epoch": 0.9969192077962905, + "grad_norm": 0.5851801037788391, + "learning_rate": 9.215870632162117e-10, + "loss": 0.873, + "step": 5946 + }, + { + "epoch": 0.9970868699570365, + "grad_norm": 0.6409519910812378, + "learning_rate": 8.493356602545799e-10, + "loss": 0.8277, + "step": 5947 + }, + { + "epoch": 0.9972545321177827, + "grad_norm": 0.4242671728134155, + "learning_rate": 7.800331306406783e-10, + "loss": 0.8626, + "step": 5948 + }, + { + "epoch": 0.9974221942785287, + "grad_norm": 0.5252429246902466, + "learning_rate": 7.136794948137127e-10, + "loss": 0.8759, + "step": 5949 + }, + { + "epoch": 0.9975898564392749, + "grad_norm": 0.4827357232570648, + "learning_rate": 6.502747723424741e-10, + "loss": 0.9101, + "step": 5950 + }, + { + "epoch": 0.9977575186000209, + "grad_norm": 0.45835551619529724, + "learning_rate": 5.898189819253386e-10, + "loss": 0.7571, + "step": 5951 + }, + { + "epoch": 0.9979251807607671, + "grad_norm": 0.4265470802783966, + "learning_rate": 5.323121413902677e-10, + "loss": 0.7681, + "step": 5952 + }, + { + "epoch": 0.9980928429215131, + "grad_norm": 0.3536740839481354, + "learning_rate": 4.777542676992486e-10, + "loss": 0.794, + "step": 5953 + }, + { + "epoch": 0.9982605050822593, + "grad_norm": 1.0968900918960571, + "learning_rate": 4.2614537693941304e-10, + "loss": 0.7514, + "step": 5954 + }, + { + "epoch": 0.9984281672430053, + "grad_norm": 0.3914303183555603, + "learning_rate": 3.7748548433302887e-10, + "loss": 0.7851, + "step": 5955 + }, + { + "epoch": 0.9985958294037515, + "grad_norm": 0.39772719144821167, + "learning_rate": 3.317746042297287e-10, + "loss": 0.8082, + "step": 5956 + }, + { + "epoch": 0.9987634915644975, + "grad_norm": 0.5874171853065491, + "learning_rate": 2.8901275010984055e-10, + "loss": 0.8401, + "step": 5957 + }, + { + "epoch": 0.9989311537252437, + "grad_norm": 0.42458993196487427, + "learning_rate": 2.4919993458549783e-10, + "loss": 0.8856, + "step": 5958 + }, + { + "epoch": 0.9990988158859897, + "grad_norm": 0.622805655002594, + "learning_rate": 2.1233616939619896e-10, + "loss": 0.775, + "step": 5959 + }, + { + "epoch": 0.9992664780467359, + "grad_norm": 0.4151398837566376, + "learning_rate": 1.784214654143579e-10, + "loss": 0.8683, + "step": 5960 + }, + { + "epoch": 0.9994341402074819, + "grad_norm": 0.42758768796920776, + "learning_rate": 1.4745583264197394e-10, + "loss": 0.8124, + "step": 5961 + }, + { + "epoch": 0.999601802368228, + "grad_norm": 0.5456928014755249, + "learning_rate": 1.1943928021174167e-10, + "loss": 0.8261, + "step": 5962 + }, + { + "epoch": 0.9997694645289741, + "grad_norm": 0.9677426218986511, + "learning_rate": 9.437181638483061e-11, + "loss": 0.8238, + "step": 5963 + }, + { + "epoch": 0.9999371266897202, + "grad_norm": 0.5349898934364319, + "learning_rate": 7.225344855532612e-11, + "loss": 0.8254, + "step": 5964 + }, + { + "epoch": 0.9999371266897202, + "step": 5964, + "total_flos": 5.467049278255923e+18, + "train_loss": 0.8923786733971595, + "train_runtime": 124734.865, + "train_samples_per_second": 6.12, + "train_steps_per_second": 0.048 + } + ], + "logging_steps": 1.0, + "max_steps": 5964, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 5.467049278255923e+18, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}