{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 0, "global_step": 461, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0021691973969631237, "grad_norm": 2.11081600189209, "learning_rate": 1e-05, "loss": 2.6121, "step": 1 }, { "epoch": 0.004338394793926247, "grad_norm": 1.9794774055480957, "learning_rate": 9.97830802603037e-06, "loss": 2.5875, "step": 2 }, { "epoch": 0.006507592190889371, "grad_norm": 1.8341801166534424, "learning_rate": 9.956616052060738e-06, "loss": 2.6118, "step": 3 }, { "epoch": 0.008676789587852495, "grad_norm": 1.5984541177749634, "learning_rate": 9.934924078091108e-06, "loss": 2.5194, "step": 4 }, { "epoch": 0.010845986984815618, "grad_norm": 1.3061403036117554, "learning_rate": 9.913232104121477e-06, "loss": 2.4602, "step": 5 }, { "epoch": 0.013015184381778741, "grad_norm": 1.0785109996795654, "learning_rate": 9.891540130151845e-06, "loss": 2.4176, "step": 6 }, { "epoch": 0.015184381778741865, "grad_norm": 0.8170204758644104, "learning_rate": 9.869848156182214e-06, "loss": 2.315, "step": 7 }, { "epoch": 0.01735357917570499, "grad_norm": 0.7127103209495544, "learning_rate": 9.848156182212582e-06, "loss": 2.2658, "step": 8 }, { "epoch": 0.019522776572668113, "grad_norm": 0.7081116437911987, "learning_rate": 9.82646420824295e-06, "loss": 2.281, "step": 9 }, { "epoch": 0.021691973969631236, "grad_norm": 0.768051028251648, "learning_rate": 9.80477223427332e-06, "loss": 2.2928, "step": 10 }, { "epoch": 0.02386117136659436, "grad_norm": 0.8635964393615723, "learning_rate": 9.783080260303689e-06, "loss": 2.2291, "step": 11 }, { "epoch": 0.026030368763557483, "grad_norm": 0.8809370398521423, "learning_rate": 9.761388286334057e-06, "loss": 2.2048, "step": 12 }, { "epoch": 0.028199566160520606, "grad_norm": 0.908088207244873, "learning_rate": 9.739696312364426e-06, "loss": 2.1956, "step": 13 }, { "epoch": 0.03036876355748373, "grad_norm": 1.0039963722229004, "learning_rate": 9.718004338394794e-06, "loss": 2.1148, "step": 14 }, { "epoch": 0.03253796095444685, "grad_norm": 0.907769501209259, "learning_rate": 9.696312364425163e-06, "loss": 2.0764, "step": 15 }, { "epoch": 0.03470715835140998, "grad_norm": 0.8381527066230774, "learning_rate": 9.674620390455533e-06, "loss": 2.0732, "step": 16 }, { "epoch": 0.0368763557483731, "grad_norm": 0.7855611443519592, "learning_rate": 9.6529284164859e-06, "loss": 1.9714, "step": 17 }, { "epoch": 0.039045553145336226, "grad_norm": 0.7366162538528442, "learning_rate": 9.63123644251627e-06, "loss": 1.965, "step": 18 }, { "epoch": 0.04121475054229935, "grad_norm": 0.6809542179107666, "learning_rate": 9.60954446854664e-06, "loss": 2.0019, "step": 19 }, { "epoch": 0.04338394793926247, "grad_norm": 0.6486644744873047, "learning_rate": 9.587852494577007e-06, "loss": 1.9383, "step": 20 }, { "epoch": 0.0455531453362256, "grad_norm": 0.6221187710762024, "learning_rate": 9.566160520607377e-06, "loss": 1.883, "step": 21 }, { "epoch": 0.04772234273318872, "grad_norm": 0.5827253460884094, "learning_rate": 9.544468546637745e-06, "loss": 1.8966, "step": 22 }, { "epoch": 0.049891540130151846, "grad_norm": 0.5862624049186707, "learning_rate": 9.522776572668114e-06, "loss": 1.8187, "step": 23 }, { "epoch": 0.052060737527114966, "grad_norm": 0.6064738035202026, "learning_rate": 9.501084598698482e-06, "loss": 1.8233, "step": 24 }, { "epoch": 0.05422993492407809, "grad_norm": 0.6191762685775757, "learning_rate": 9.479392624728851e-06, "loss": 1.7323, "step": 25 }, { "epoch": 0.05639913232104121, "grad_norm": 0.6317607164382935, "learning_rate": 9.457700650759219e-06, "loss": 1.7398, "step": 26 }, { "epoch": 0.05856832971800434, "grad_norm": 0.6248988509178162, "learning_rate": 9.436008676789589e-06, "loss": 1.7171, "step": 27 }, { "epoch": 0.06073752711496746, "grad_norm": 0.6342098712921143, "learning_rate": 9.414316702819958e-06, "loss": 1.6891, "step": 28 }, { "epoch": 0.06290672451193059, "grad_norm": 0.6058349609375, "learning_rate": 9.392624728850326e-06, "loss": 1.6624, "step": 29 }, { "epoch": 0.0650759219088937, "grad_norm": 0.5988097786903381, "learning_rate": 9.370932754880695e-06, "loss": 1.6438, "step": 30 }, { "epoch": 0.06724511930585683, "grad_norm": 0.6042107939720154, "learning_rate": 9.349240780911065e-06, "loss": 1.5956, "step": 31 }, { "epoch": 0.06941431670281996, "grad_norm": 0.5972896218299866, "learning_rate": 9.327548806941433e-06, "loss": 1.5142, "step": 32 }, { "epoch": 0.07158351409978309, "grad_norm": 0.5832527875900269, "learning_rate": 9.305856832971802e-06, "loss": 1.5384, "step": 33 }, { "epoch": 0.0737527114967462, "grad_norm": 0.5889501571655273, "learning_rate": 9.28416485900217e-06, "loss": 1.5183, "step": 34 }, { "epoch": 0.07592190889370933, "grad_norm": 0.622042715549469, "learning_rate": 9.26247288503254e-06, "loss": 1.4945, "step": 35 }, { "epoch": 0.07809110629067245, "grad_norm": 0.630580723285675, "learning_rate": 9.240780911062907e-06, "loss": 1.4925, "step": 36 }, { "epoch": 0.08026030368763558, "grad_norm": 0.5893872976303101, "learning_rate": 9.219088937093276e-06, "loss": 1.4306, "step": 37 }, { "epoch": 0.0824295010845987, "grad_norm": 0.5707374811172485, "learning_rate": 9.197396963123644e-06, "loss": 1.418, "step": 38 }, { "epoch": 0.08459869848156182, "grad_norm": 0.5506007075309753, "learning_rate": 9.175704989154014e-06, "loss": 1.3998, "step": 39 }, { "epoch": 0.08676789587852494, "grad_norm": 0.5373205542564392, "learning_rate": 9.154013015184382e-06, "loss": 1.3508, "step": 40 }, { "epoch": 0.08893709327548807, "grad_norm": 0.5845463871955872, "learning_rate": 9.132321041214751e-06, "loss": 1.3387, "step": 41 }, { "epoch": 0.0911062906724512, "grad_norm": 0.6560288071632385, "learning_rate": 9.11062906724512e-06, "loss": 1.3078, "step": 42 }, { "epoch": 0.09327548806941431, "grad_norm": 0.580771267414093, "learning_rate": 9.088937093275488e-06, "loss": 1.309, "step": 43 }, { "epoch": 0.09544468546637744, "grad_norm": 0.5993460416793823, "learning_rate": 9.067245119305858e-06, "loss": 1.3005, "step": 44 }, { "epoch": 0.09761388286334056, "grad_norm": 0.6303313374519348, "learning_rate": 9.045553145336227e-06, "loss": 1.2776, "step": 45 }, { "epoch": 0.09978308026030369, "grad_norm": 0.5471832156181335, "learning_rate": 9.023861171366595e-06, "loss": 1.2588, "step": 46 }, { "epoch": 0.1019522776572668, "grad_norm": 0.5459536910057068, "learning_rate": 9.002169197396964e-06, "loss": 1.2094, "step": 47 }, { "epoch": 0.10412147505422993, "grad_norm": 0.5408421754837036, "learning_rate": 8.980477223427332e-06, "loss": 1.2355, "step": 48 }, { "epoch": 0.10629067245119306, "grad_norm": 0.48613736033439636, "learning_rate": 8.958785249457702e-06, "loss": 1.1594, "step": 49 }, { "epoch": 0.10845986984815618, "grad_norm": 0.4992137849330902, "learning_rate": 8.93709327548807e-06, "loss": 1.1853, "step": 50 }, { "epoch": 0.11062906724511931, "grad_norm": 0.4883258044719696, "learning_rate": 8.915401301518439e-06, "loss": 1.1467, "step": 51 }, { "epoch": 0.11279826464208242, "grad_norm": 0.48137181997299194, "learning_rate": 8.893709327548807e-06, "loss": 1.1541, "step": 52 }, { "epoch": 0.11496746203904555, "grad_norm": 0.4401426911354065, "learning_rate": 8.872017353579176e-06, "loss": 1.1112, "step": 53 }, { "epoch": 0.11713665943600868, "grad_norm": 0.4389711320400238, "learning_rate": 8.850325379609546e-06, "loss": 1.0961, "step": 54 }, { "epoch": 0.1193058568329718, "grad_norm": 0.439217746257782, "learning_rate": 8.828633405639913e-06, "loss": 1.0838, "step": 55 }, { "epoch": 0.12147505422993492, "grad_norm": 0.521327555179596, "learning_rate": 8.806941431670283e-06, "loss": 1.0813, "step": 56 }, { "epoch": 0.12364425162689804, "grad_norm": 0.4604964256286621, "learning_rate": 8.785249457700652e-06, "loss": 1.0857, "step": 57 }, { "epoch": 0.12581344902386118, "grad_norm": 0.42932406067848206, "learning_rate": 8.76355748373102e-06, "loss": 1.0513, "step": 58 }, { "epoch": 0.1279826464208243, "grad_norm": 0.43107494711875916, "learning_rate": 8.74186550976139e-06, "loss": 1.0488, "step": 59 }, { "epoch": 0.1301518438177874, "grad_norm": 0.42236247658729553, "learning_rate": 8.720173535791757e-06, "loss": 1.028, "step": 60 }, { "epoch": 0.13232104121475055, "grad_norm": 0.43343520164489746, "learning_rate": 8.698481561822127e-06, "loss": 1.0272, "step": 61 }, { "epoch": 0.13449023861171366, "grad_norm": 0.48012059926986694, "learning_rate": 8.676789587852495e-06, "loss": 1.0099, "step": 62 }, { "epoch": 0.13665943600867678, "grad_norm": 0.42098668217658997, "learning_rate": 8.655097613882864e-06, "loss": 1.0158, "step": 63 }, { "epoch": 0.13882863340563992, "grad_norm": 0.4227031171321869, "learning_rate": 8.633405639913232e-06, "loss": 0.9815, "step": 64 }, { "epoch": 0.14099783080260303, "grad_norm": 0.41489773988723755, "learning_rate": 8.611713665943601e-06, "loss": 0.9635, "step": 65 }, { "epoch": 0.14316702819956617, "grad_norm": 0.40091952681541443, "learning_rate": 8.59002169197397e-06, "loss": 0.9528, "step": 66 }, { "epoch": 0.14533622559652928, "grad_norm": 0.4547802805900574, "learning_rate": 8.568329718004339e-06, "loss": 0.9726, "step": 67 }, { "epoch": 0.1475054229934924, "grad_norm": 0.4234392046928406, "learning_rate": 8.546637744034708e-06, "loss": 0.962, "step": 68 }, { "epoch": 0.14967462039045554, "grad_norm": 0.38503268361091614, "learning_rate": 8.524945770065076e-06, "loss": 0.9463, "step": 69 }, { "epoch": 0.15184381778741865, "grad_norm": 0.4524126350879669, "learning_rate": 8.503253796095445e-06, "loss": 0.9061, "step": 70 }, { "epoch": 0.1540130151843818, "grad_norm": 0.40082183480262756, "learning_rate": 8.481561822125815e-06, "loss": 0.9034, "step": 71 }, { "epoch": 0.1561822125813449, "grad_norm": 0.41073668003082275, "learning_rate": 8.459869848156183e-06, "loss": 0.8961, "step": 72 }, { "epoch": 0.15835140997830802, "grad_norm": 0.5651035308837891, "learning_rate": 8.438177874186552e-06, "loss": 0.8851, "step": 73 }, { "epoch": 0.16052060737527116, "grad_norm": 0.44745904207229614, "learning_rate": 8.416485900216922e-06, "loss": 0.9203, "step": 74 }, { "epoch": 0.16268980477223427, "grad_norm": 0.4688916504383087, "learning_rate": 8.39479392624729e-06, "loss": 0.9449, "step": 75 }, { "epoch": 0.1648590021691974, "grad_norm": 0.386109322309494, "learning_rate": 8.373101952277657e-06, "loss": 0.868, "step": 76 }, { "epoch": 0.16702819956616052, "grad_norm": 0.47646722197532654, "learning_rate": 8.351409978308027e-06, "loss": 0.8853, "step": 77 }, { "epoch": 0.16919739696312364, "grad_norm": 0.38544556498527527, "learning_rate": 8.329718004338394e-06, "loss": 0.8703, "step": 78 }, { "epoch": 0.17136659436008678, "grad_norm": 0.5066248178482056, "learning_rate": 8.308026030368764e-06, "loss": 0.8963, "step": 79 }, { "epoch": 0.1735357917570499, "grad_norm": 0.471044659614563, "learning_rate": 8.286334056399133e-06, "loss": 0.8547, "step": 80 }, { "epoch": 0.175704989154013, "grad_norm": 0.4496097266674042, "learning_rate": 8.264642082429501e-06, "loss": 0.8556, "step": 81 }, { "epoch": 0.17787418655097614, "grad_norm": 0.40527230501174927, "learning_rate": 8.24295010845987e-06, "loss": 0.8756, "step": 82 }, { "epoch": 0.18004338394793926, "grad_norm": 0.40289977192878723, "learning_rate": 8.22125813449024e-06, "loss": 0.8412, "step": 83 }, { "epoch": 0.1822125813449024, "grad_norm": 0.3807680606842041, "learning_rate": 8.199566160520608e-06, "loss": 0.8241, "step": 84 }, { "epoch": 0.1843817787418655, "grad_norm": 0.4011147618293762, "learning_rate": 8.177874186550977e-06, "loss": 0.8308, "step": 85 }, { "epoch": 0.18655097613882862, "grad_norm": 0.6492615938186646, "learning_rate": 8.156182212581345e-06, "loss": 0.8091, "step": 86 }, { "epoch": 0.18872017353579176, "grad_norm": 0.45752060413360596, "learning_rate": 8.134490238611715e-06, "loss": 0.86, "step": 87 }, { "epoch": 0.19088937093275488, "grad_norm": 0.39894232153892517, "learning_rate": 8.112798264642084e-06, "loss": 0.8087, "step": 88 }, { "epoch": 0.19305856832971802, "grad_norm": 0.4044681787490845, "learning_rate": 8.091106290672452e-06, "loss": 0.8287, "step": 89 }, { "epoch": 0.19522776572668113, "grad_norm": 0.5479378700256348, "learning_rate": 8.06941431670282e-06, "loss": 0.7788, "step": 90 }, { "epoch": 0.19739696312364424, "grad_norm": 0.5609095096588135, "learning_rate": 8.04772234273319e-06, "loss": 0.8196, "step": 91 }, { "epoch": 0.19956616052060738, "grad_norm": 0.3801177144050598, "learning_rate": 8.026030368763557e-06, "loss": 0.7683, "step": 92 }, { "epoch": 0.2017353579175705, "grad_norm": 0.38417235016822815, "learning_rate": 8.004338394793926e-06, "loss": 0.7955, "step": 93 }, { "epoch": 0.2039045553145336, "grad_norm": 0.6401964426040649, "learning_rate": 7.982646420824296e-06, "loss": 0.7861, "step": 94 }, { "epoch": 0.20607375271149675, "grad_norm": 0.4698239266872406, "learning_rate": 7.960954446854664e-06, "loss": 0.7872, "step": 95 }, { "epoch": 0.20824295010845986, "grad_norm": 0.45705363154411316, "learning_rate": 7.939262472885033e-06, "loss": 0.7865, "step": 96 }, { "epoch": 0.210412147505423, "grad_norm": 0.4010639488697052, "learning_rate": 7.917570498915403e-06, "loss": 0.7641, "step": 97 }, { "epoch": 0.21258134490238612, "grad_norm": 0.40273797512054443, "learning_rate": 7.89587852494577e-06, "loss": 0.7719, "step": 98 }, { "epoch": 0.21475054229934923, "grad_norm": 0.4527304470539093, "learning_rate": 7.87418655097614e-06, "loss": 0.8008, "step": 99 }, { "epoch": 0.21691973969631237, "grad_norm": 0.432674765586853, "learning_rate": 7.85249457700651e-06, "loss": 0.7426, "step": 100 }, { "epoch": 0.21908893709327548, "grad_norm": 0.4265141785144806, "learning_rate": 7.830802603036877e-06, "loss": 0.7713, "step": 101 }, { "epoch": 0.22125813449023862, "grad_norm": 0.4253065288066864, "learning_rate": 7.809110629067247e-06, "loss": 0.7484, "step": 102 }, { "epoch": 0.22342733188720174, "grad_norm": 0.44442278146743774, "learning_rate": 7.787418655097614e-06, "loss": 0.763, "step": 103 }, { "epoch": 0.22559652928416485, "grad_norm": 0.47105884552001953, "learning_rate": 7.765726681127982e-06, "loss": 0.7703, "step": 104 }, { "epoch": 0.227765726681128, "grad_norm": 0.41451483964920044, "learning_rate": 7.744034707158352e-06, "loss": 0.7558, "step": 105 }, { "epoch": 0.2299349240780911, "grad_norm": 0.4369511604309082, "learning_rate": 7.722342733188721e-06, "loss": 0.7592, "step": 106 }, { "epoch": 0.23210412147505424, "grad_norm": 0.4827646017074585, "learning_rate": 7.700650759219089e-06, "loss": 0.753, "step": 107 }, { "epoch": 0.23427331887201736, "grad_norm": 0.39963749051094055, "learning_rate": 7.678958785249458e-06, "loss": 0.7333, "step": 108 }, { "epoch": 0.23644251626898047, "grad_norm": 0.4362649917602539, "learning_rate": 7.657266811279828e-06, "loss": 0.7392, "step": 109 }, { "epoch": 0.2386117136659436, "grad_norm": 0.4753969609737396, "learning_rate": 7.635574837310196e-06, "loss": 0.7411, "step": 110 }, { "epoch": 0.24078091106290672, "grad_norm": 0.3958611488342285, "learning_rate": 7.613882863340565e-06, "loss": 0.7125, "step": 111 }, { "epoch": 0.24295010845986983, "grad_norm": 0.4026221036911011, "learning_rate": 7.592190889370933e-06, "loss": 0.7434, "step": 112 }, { "epoch": 0.24511930585683298, "grad_norm": 0.672747790813446, "learning_rate": 7.570498915401302e-06, "loss": 0.702, "step": 113 }, { "epoch": 0.2472885032537961, "grad_norm": 0.39787012338638306, "learning_rate": 7.548806941431671e-06, "loss": 0.7329, "step": 114 }, { "epoch": 0.24945770065075923, "grad_norm": 0.4380010664463043, "learning_rate": 7.527114967462039e-06, "loss": 0.7281, "step": 115 }, { "epoch": 0.25162689804772237, "grad_norm": 0.41002675890922546, "learning_rate": 7.505422993492408e-06, "loss": 0.7229, "step": 116 }, { "epoch": 0.25379609544468545, "grad_norm": 0.41404595971107483, "learning_rate": 7.483731019522778e-06, "loss": 0.7177, "step": 117 }, { "epoch": 0.2559652928416486, "grad_norm": 0.4458067715167999, "learning_rate": 7.4620390455531455e-06, "loss": 0.7226, "step": 118 }, { "epoch": 0.25813449023861174, "grad_norm": 0.3937883973121643, "learning_rate": 7.440347071583515e-06, "loss": 0.7109, "step": 119 }, { "epoch": 0.2603036876355748, "grad_norm": 0.42459914088249207, "learning_rate": 7.418655097613884e-06, "loss": 0.7292, "step": 120 }, { "epoch": 0.26247288503253796, "grad_norm": 0.47104501724243164, "learning_rate": 7.396963123644252e-06, "loss": 0.7018, "step": 121 }, { "epoch": 0.2646420824295011, "grad_norm": 0.413194477558136, "learning_rate": 7.375271149674621e-06, "loss": 0.6862, "step": 122 }, { "epoch": 0.2668112798264642, "grad_norm": 0.44947072863578796, "learning_rate": 7.35357917570499e-06, "loss": 0.7134, "step": 123 }, { "epoch": 0.26898047722342733, "grad_norm": 0.4308297038078308, "learning_rate": 7.331887201735358e-06, "loss": 0.7214, "step": 124 }, { "epoch": 0.27114967462039047, "grad_norm": 0.48938828706741333, "learning_rate": 7.310195227765728e-06, "loss": 0.7065, "step": 125 }, { "epoch": 0.27331887201735355, "grad_norm": 0.4659271538257599, "learning_rate": 7.288503253796096e-06, "loss": 0.717, "step": 126 }, { "epoch": 0.2754880694143167, "grad_norm": 0.43395182490348816, "learning_rate": 7.266811279826465e-06, "loss": 0.7137, "step": 127 }, { "epoch": 0.27765726681127983, "grad_norm": 0.4359440505504608, "learning_rate": 7.2451193058568335e-06, "loss": 0.6858, "step": 128 }, { "epoch": 0.279826464208243, "grad_norm": 0.6049463152885437, "learning_rate": 7.223427331887203e-06, "loss": 0.7262, "step": 129 }, { "epoch": 0.28199566160520606, "grad_norm": 0.9202204942703247, "learning_rate": 7.201735357917571e-06, "loss": 0.7012, "step": 130 }, { "epoch": 0.2841648590021692, "grad_norm": 0.4885154664516449, "learning_rate": 7.18004338394794e-06, "loss": 0.7016, "step": 131 }, { "epoch": 0.28633405639913234, "grad_norm": 0.5771850943565369, "learning_rate": 7.158351409978309e-06, "loss": 0.7069, "step": 132 }, { "epoch": 0.2885032537960954, "grad_norm": 0.4814640283584595, "learning_rate": 7.1366594360086775e-06, "loss": 0.7129, "step": 133 }, { "epoch": 0.29067245119305857, "grad_norm": 0.4861976206302643, "learning_rate": 7.114967462039046e-06, "loss": 0.6926, "step": 134 }, { "epoch": 0.2928416485900217, "grad_norm": 0.43924835324287415, "learning_rate": 7.093275488069415e-06, "loss": 0.6729, "step": 135 }, { "epoch": 0.2950108459869848, "grad_norm": 0.44776543974876404, "learning_rate": 7.071583514099783e-06, "loss": 0.69, "step": 136 }, { "epoch": 0.29718004338394793, "grad_norm": 0.48698529601097107, "learning_rate": 7.049891540130153e-06, "loss": 0.6707, "step": 137 }, { "epoch": 0.2993492407809111, "grad_norm": 0.49502941966056824, "learning_rate": 7.028199566160521e-06, "loss": 0.6705, "step": 138 }, { "epoch": 0.30151843817787416, "grad_norm": 0.43487516045570374, "learning_rate": 7.00650759219089e-06, "loss": 0.6468, "step": 139 }, { "epoch": 0.3036876355748373, "grad_norm": 0.45082563161849976, "learning_rate": 6.984815618221259e-06, "loss": 0.6882, "step": 140 }, { "epoch": 0.30585683297180044, "grad_norm": 0.5704577565193176, "learning_rate": 6.963123644251627e-06, "loss": 0.6546, "step": 141 }, { "epoch": 0.3080260303687636, "grad_norm": 0.5590646266937256, "learning_rate": 6.941431670281996e-06, "loss": 0.691, "step": 142 }, { "epoch": 0.31019522776572667, "grad_norm": 0.4791410565376282, "learning_rate": 6.9197396963123654e-06, "loss": 0.6327, "step": 143 }, { "epoch": 0.3123644251626898, "grad_norm": 0.435273140668869, "learning_rate": 6.898047722342733e-06, "loss": 0.675, "step": 144 }, { "epoch": 0.31453362255965295, "grad_norm": 0.5272928476333618, "learning_rate": 6.876355748373103e-06, "loss": 0.6683, "step": 145 }, { "epoch": 0.31670281995661603, "grad_norm": 0.5004708766937256, "learning_rate": 6.854663774403471e-06, "loss": 0.6518, "step": 146 }, { "epoch": 0.3188720173535792, "grad_norm": 0.45495614409446716, "learning_rate": 6.83297180043384e-06, "loss": 0.6576, "step": 147 }, { "epoch": 0.3210412147505423, "grad_norm": 0.48196735978126526, "learning_rate": 6.8112798264642086e-06, "loss": 0.6389, "step": 148 }, { "epoch": 0.3232104121475054, "grad_norm": 0.4797559082508087, "learning_rate": 6.789587852494578e-06, "loss": 0.6677, "step": 149 }, { "epoch": 0.32537960954446854, "grad_norm": 0.4930654764175415, "learning_rate": 6.767895878524946e-06, "loss": 0.6437, "step": 150 }, { "epoch": 0.3275488069414317, "grad_norm": 0.45326152443885803, "learning_rate": 6.746203904555315e-06, "loss": 0.6616, "step": 151 }, { "epoch": 0.3297180043383948, "grad_norm": 0.48853611946105957, "learning_rate": 6.724511930585684e-06, "loss": 0.6473, "step": 152 }, { "epoch": 0.3318872017353579, "grad_norm": 0.4732632040977478, "learning_rate": 6.7028199566160526e-06, "loss": 0.6396, "step": 153 }, { "epoch": 0.33405639913232105, "grad_norm": 0.4409032464027405, "learning_rate": 6.681127982646421e-06, "loss": 0.6507, "step": 154 }, { "epoch": 0.3362255965292842, "grad_norm": 0.4853736162185669, "learning_rate": 6.659436008676791e-06, "loss": 0.6495, "step": 155 }, { "epoch": 0.3383947939262473, "grad_norm": 0.5413075685501099, "learning_rate": 6.6377440347071584e-06, "loss": 0.6719, "step": 156 }, { "epoch": 0.3405639913232104, "grad_norm": 0.4533112943172455, "learning_rate": 6.616052060737528e-06, "loss": 0.6358, "step": 157 }, { "epoch": 0.34273318872017355, "grad_norm": 0.5843056440353394, "learning_rate": 6.594360086767897e-06, "loss": 0.6724, "step": 158 }, { "epoch": 0.34490238611713664, "grad_norm": 0.48672884702682495, "learning_rate": 6.572668112798265e-06, "loss": 0.6449, "step": 159 }, { "epoch": 0.3470715835140998, "grad_norm": 0.6113792657852173, "learning_rate": 6.550976138828634e-06, "loss": 0.6297, "step": 160 }, { "epoch": 0.3492407809110629, "grad_norm": 0.5860591530799866, "learning_rate": 6.529284164859002e-06, "loss": 0.6617, "step": 161 }, { "epoch": 0.351409978308026, "grad_norm": 0.5638704299926758, "learning_rate": 6.507592190889371e-06, "loss": 0.6225, "step": 162 }, { "epoch": 0.35357917570498915, "grad_norm": 0.5348539352416992, "learning_rate": 6.4859002169197405e-06, "loss": 0.6351, "step": 163 }, { "epoch": 0.3557483731019523, "grad_norm": 0.4822864830493927, "learning_rate": 6.464208242950108e-06, "loss": 0.6315, "step": 164 }, { "epoch": 0.3579175704989154, "grad_norm": 0.5151847004890442, "learning_rate": 6.442516268980478e-06, "loss": 0.6311, "step": 165 }, { "epoch": 0.3600867678958785, "grad_norm": 0.4632371962070465, "learning_rate": 6.420824295010846e-06, "loss": 0.6515, "step": 166 }, { "epoch": 0.36225596529284165, "grad_norm": 0.4700559675693512, "learning_rate": 6.399132321041215e-06, "loss": 0.6218, "step": 167 }, { "epoch": 0.3644251626898048, "grad_norm": 0.5487787127494812, "learning_rate": 6.377440347071584e-06, "loss": 0.6485, "step": 168 }, { "epoch": 0.3665943600867679, "grad_norm": 0.5952288508415222, "learning_rate": 6.355748373101953e-06, "loss": 0.6263, "step": 169 }, { "epoch": 0.368763557483731, "grad_norm": 0.4692787230014801, "learning_rate": 6.334056399132321e-06, "loss": 0.637, "step": 170 }, { "epoch": 0.37093275488069416, "grad_norm": 0.48934659361839294, "learning_rate": 6.31236442516269e-06, "loss": 0.6177, "step": 171 }, { "epoch": 0.37310195227765725, "grad_norm": 0.6284322142601013, "learning_rate": 6.29067245119306e-06, "loss": 0.6533, "step": 172 }, { "epoch": 0.3752711496746204, "grad_norm": 0.583229660987854, "learning_rate": 6.268980477223428e-06, "loss": 0.6468, "step": 173 }, { "epoch": 0.3774403470715835, "grad_norm": 0.5373253226280212, "learning_rate": 6.247288503253796e-06, "loss": 0.6219, "step": 174 }, { "epoch": 0.3796095444685466, "grad_norm": 0.5065279603004456, "learning_rate": 6.225596529284166e-06, "loss": 0.609, "step": 175 }, { "epoch": 0.38177874186550975, "grad_norm": 0.4818476438522339, "learning_rate": 6.2039045553145335e-06, "loss": 0.6203, "step": 176 }, { "epoch": 0.3839479392624729, "grad_norm": 0.47112759947776794, "learning_rate": 6.182212581344903e-06, "loss": 0.6315, "step": 177 }, { "epoch": 0.38611713665943603, "grad_norm": 0.5504185557365417, "learning_rate": 6.1605206073752725e-06, "loss": 0.6318, "step": 178 }, { "epoch": 0.3882863340563991, "grad_norm": 0.5567037463188171, "learning_rate": 6.13882863340564e-06, "loss": 0.6104, "step": 179 }, { "epoch": 0.39045553145336226, "grad_norm": 0.4654327630996704, "learning_rate": 6.11713665943601e-06, "loss": 0.619, "step": 180 }, { "epoch": 0.3926247288503254, "grad_norm": 0.47351258993148804, "learning_rate": 6.095444685466378e-06, "loss": 0.6206, "step": 181 }, { "epoch": 0.3947939262472885, "grad_norm": 0.47146716713905334, "learning_rate": 6.073752711496746e-06, "loss": 0.6111, "step": 182 }, { "epoch": 0.3969631236442516, "grad_norm": 0.49498575925827026, "learning_rate": 6.052060737527116e-06, "loss": 0.6179, "step": 183 }, { "epoch": 0.39913232104121477, "grad_norm": 0.5393364429473877, "learning_rate": 6.030368763557483e-06, "loss": 0.6331, "step": 184 }, { "epoch": 0.40130151843817785, "grad_norm": 0.5126506686210632, "learning_rate": 6.008676789587853e-06, "loss": 0.6143, "step": 185 }, { "epoch": 0.403470715835141, "grad_norm": 0.6334784626960754, "learning_rate": 5.986984815618222e-06, "loss": 0.6322, "step": 186 }, { "epoch": 0.40563991323210413, "grad_norm": 0.5737828016281128, "learning_rate": 5.96529284164859e-06, "loss": 0.6206, "step": 187 }, { "epoch": 0.4078091106290672, "grad_norm": 0.5050855875015259, "learning_rate": 5.943600867678959e-06, "loss": 0.6262, "step": 188 }, { "epoch": 0.40997830802603036, "grad_norm": 0.4655757546424866, "learning_rate": 5.921908893709328e-06, "loss": 0.6062, "step": 189 }, { "epoch": 0.4121475054229935, "grad_norm": 0.7657625675201416, "learning_rate": 5.900216919739696e-06, "loss": 0.6278, "step": 190 }, { "epoch": 0.41431670281995664, "grad_norm": 0.4740203619003296, "learning_rate": 5.8785249457700655e-06, "loss": 0.6125, "step": 191 }, { "epoch": 0.4164859002169197, "grad_norm": 0.47104138135910034, "learning_rate": 5.856832971800435e-06, "loss": 0.6181, "step": 192 }, { "epoch": 0.41865509761388287, "grad_norm": 0.6184363961219788, "learning_rate": 5.835140997830803e-06, "loss": 0.6227, "step": 193 }, { "epoch": 0.420824295010846, "grad_norm": 0.6750501990318298, "learning_rate": 5.813449023861172e-06, "loss": 0.6407, "step": 194 }, { "epoch": 0.4229934924078091, "grad_norm": 0.6018952131271362, "learning_rate": 5.791757049891541e-06, "loss": 0.5914, "step": 195 }, { "epoch": 0.42516268980477223, "grad_norm": 0.5619893670082092, "learning_rate": 5.770065075921909e-06, "loss": 0.6222, "step": 196 }, { "epoch": 0.42733188720173537, "grad_norm": 0.4996378719806671, "learning_rate": 5.748373101952278e-06, "loss": 0.6076, "step": 197 }, { "epoch": 0.42950108459869846, "grad_norm": 0.47961318492889404, "learning_rate": 5.7266811279826476e-06, "loss": 0.6102, "step": 198 }, { "epoch": 0.4316702819956616, "grad_norm": 0.6786702871322632, "learning_rate": 5.704989154013015e-06, "loss": 0.6353, "step": 199 }, { "epoch": 0.43383947939262474, "grad_norm": 0.6170738935470581, "learning_rate": 5.683297180043385e-06, "loss": 0.6038, "step": 200 }, { "epoch": 0.4360086767895879, "grad_norm": 0.5487593412399292, "learning_rate": 5.6616052060737535e-06, "loss": 0.5929, "step": 201 }, { "epoch": 0.43817787418655096, "grad_norm": 0.593626856803894, "learning_rate": 5.639913232104122e-06, "loss": 0.5977, "step": 202 }, { "epoch": 0.4403470715835141, "grad_norm": 0.7302510738372803, "learning_rate": 5.618221258134491e-06, "loss": 0.5996, "step": 203 }, { "epoch": 0.44251626898047725, "grad_norm": 0.5379276275634766, "learning_rate": 5.59652928416486e-06, "loss": 0.6119, "step": 204 }, { "epoch": 0.44468546637744033, "grad_norm": 0.6861233711242676, "learning_rate": 5.574837310195228e-06, "loss": 0.6119, "step": 205 }, { "epoch": 0.44685466377440347, "grad_norm": 0.5536229014396667, "learning_rate": 5.5531453362255974e-06, "loss": 0.5926, "step": 206 }, { "epoch": 0.4490238611713666, "grad_norm": 0.5160278677940369, "learning_rate": 5.531453362255966e-06, "loss": 0.6013, "step": 207 }, { "epoch": 0.4511930585683297, "grad_norm": 0.5844294428825378, "learning_rate": 5.509761388286335e-06, "loss": 0.6044, "step": 208 }, { "epoch": 0.45336225596529284, "grad_norm": 0.6622424721717834, "learning_rate": 5.488069414316703e-06, "loss": 0.6005, "step": 209 }, { "epoch": 0.455531453362256, "grad_norm": 0.5025097727775574, "learning_rate": 5.466377440347071e-06, "loss": 0.5802, "step": 210 }, { "epoch": 0.45770065075921906, "grad_norm": 0.5540115237236023, "learning_rate": 5.444685466377441e-06, "loss": 0.5932, "step": 211 }, { "epoch": 0.4598698481561822, "grad_norm": 0.5804458856582642, "learning_rate": 5.42299349240781e-06, "loss": 0.5987, "step": 212 }, { "epoch": 0.46203904555314534, "grad_norm": 0.5122389793395996, "learning_rate": 5.401301518438178e-06, "loss": 0.5916, "step": 213 }, { "epoch": 0.4642082429501085, "grad_norm": 0.5717864632606506, "learning_rate": 5.379609544468547e-06, "loss": 0.5948, "step": 214 }, { "epoch": 0.46637744034707157, "grad_norm": 0.6112474799156189, "learning_rate": 5.357917570498916e-06, "loss": 0.6016, "step": 215 }, { "epoch": 0.4685466377440347, "grad_norm": 0.58445805311203, "learning_rate": 5.3362255965292846e-06, "loss": 0.5943, "step": 216 }, { "epoch": 0.47071583514099785, "grad_norm": 0.5126850605010986, "learning_rate": 5.314533622559653e-06, "loss": 0.6058, "step": 217 }, { "epoch": 0.47288503253796094, "grad_norm": 0.498695969581604, "learning_rate": 5.292841648590023e-06, "loss": 0.5867, "step": 218 }, { "epoch": 0.4750542299349241, "grad_norm": 0.5060663223266602, "learning_rate": 5.2711496746203904e-06, "loss": 0.6161, "step": 219 }, { "epoch": 0.4772234273318872, "grad_norm": 0.49926185607910156, "learning_rate": 5.24945770065076e-06, "loss": 0.594, "step": 220 }, { "epoch": 0.4793926247288503, "grad_norm": 0.7231149673461914, "learning_rate": 5.2277657266811285e-06, "loss": 0.6506, "step": 221 }, { "epoch": 0.48156182212581344, "grad_norm": 0.501641571521759, "learning_rate": 5.206073752711497e-06, "loss": 0.6031, "step": 222 }, { "epoch": 0.4837310195227766, "grad_norm": 0.5351401567459106, "learning_rate": 5.184381778741866e-06, "loss": 0.5957, "step": 223 }, { "epoch": 0.48590021691973967, "grad_norm": 0.4861268997192383, "learning_rate": 5.162689804772235e-06, "loss": 0.6109, "step": 224 }, { "epoch": 0.4880694143167028, "grad_norm": 0.4920121133327484, "learning_rate": 5.140997830802603e-06, "loss": 0.5931, "step": 225 }, { "epoch": 0.49023861171366595, "grad_norm": 0.5210174322128296, "learning_rate": 5.1193058568329725e-06, "loss": 0.6122, "step": 226 }, { "epoch": 0.4924078091106291, "grad_norm": 0.7002459764480591, "learning_rate": 5.097613882863341e-06, "loss": 0.603, "step": 227 }, { "epoch": 0.4945770065075922, "grad_norm": 0.4967700242996216, "learning_rate": 5.07592190889371e-06, "loss": 0.5675, "step": 228 }, { "epoch": 0.4967462039045553, "grad_norm": 0.49998047947883606, "learning_rate": 5.054229934924078e-06, "loss": 0.5885, "step": 229 }, { "epoch": 0.49891540130151846, "grad_norm": 0.46569979190826416, "learning_rate": 5.032537960954448e-06, "loss": 0.588, "step": 230 }, { "epoch": 0.5010845986984815, "grad_norm": 0.518514096736908, "learning_rate": 5.010845986984816e-06, "loss": 0.5795, "step": 231 }, { "epoch": 0.5032537960954447, "grad_norm": 0.5044274926185608, "learning_rate": 4.989154013015185e-06, "loss": 0.5777, "step": 232 }, { "epoch": 0.5054229934924078, "grad_norm": 0.5333325862884521, "learning_rate": 4.967462039045554e-06, "loss": 0.5902, "step": 233 }, { "epoch": 0.5075921908893709, "grad_norm": 0.666131854057312, "learning_rate": 4.945770065075922e-06, "loss": 0.5735, "step": 234 }, { "epoch": 0.5097613882863341, "grad_norm": 0.5070677995681763, "learning_rate": 4.924078091106291e-06, "loss": 0.5677, "step": 235 }, { "epoch": 0.5119305856832972, "grad_norm": 0.5434484481811523, "learning_rate": 4.90238611713666e-06, "loss": 0.5984, "step": 236 }, { "epoch": 0.5140997830802603, "grad_norm": 0.5604612231254578, "learning_rate": 4.880694143167028e-06, "loss": 0.5915, "step": 237 }, { "epoch": 0.5162689804772235, "grad_norm": 0.5144227147102356, "learning_rate": 4.859002169197397e-06, "loss": 0.5765, "step": 238 }, { "epoch": 0.5184381778741866, "grad_norm": 0.4857323169708252, "learning_rate": 4.837310195227766e-06, "loss": 0.5722, "step": 239 }, { "epoch": 0.5206073752711496, "grad_norm": 0.4755032956600189, "learning_rate": 4.815618221258135e-06, "loss": 0.5749, "step": 240 }, { "epoch": 0.5227765726681128, "grad_norm": 0.587493896484375, "learning_rate": 4.793926247288504e-06, "loss": 0.5981, "step": 241 }, { "epoch": 0.5249457700650759, "grad_norm": 0.5679769515991211, "learning_rate": 4.772234273318872e-06, "loss": 0.5884, "step": 242 }, { "epoch": 0.527114967462039, "grad_norm": 0.498691588640213, "learning_rate": 4.750542299349241e-06, "loss": 0.5809, "step": 243 }, { "epoch": 0.5292841648590022, "grad_norm": 0.675168514251709, "learning_rate": 4.7288503253796095e-06, "loss": 0.5874, "step": 244 }, { "epoch": 0.5314533622559653, "grad_norm": 0.5249897241592407, "learning_rate": 4.707158351409979e-06, "loss": 0.5878, "step": 245 }, { "epoch": 0.5336225596529284, "grad_norm": 0.48491179943084717, "learning_rate": 4.685466377440348e-06, "loss": 0.5743, "step": 246 }, { "epoch": 0.5357917570498916, "grad_norm": 0.5798050761222839, "learning_rate": 4.663774403470716e-06, "loss": 0.5682, "step": 247 }, { "epoch": 0.5379609544468547, "grad_norm": 0.5193563103675842, "learning_rate": 4.642082429501085e-06, "loss": 0.5743, "step": 248 }, { "epoch": 0.5401301518438177, "grad_norm": 0.5050124526023865, "learning_rate": 4.6203904555314535e-06, "loss": 0.5823, "step": 249 }, { "epoch": 0.5422993492407809, "grad_norm": 0.5084236264228821, "learning_rate": 4.598698481561822e-06, "loss": 0.5865, "step": 250 }, { "epoch": 0.544468546637744, "grad_norm": 0.5621696710586548, "learning_rate": 4.577006507592191e-06, "loss": 0.6023, "step": 251 }, { "epoch": 0.5466377440347071, "grad_norm": 0.4698992669582367, "learning_rate": 4.55531453362256e-06, "loss": 0.5835, "step": 252 }, { "epoch": 0.5488069414316703, "grad_norm": 0.5616809129714966, "learning_rate": 4.533622559652929e-06, "loss": 0.559, "step": 253 }, { "epoch": 0.5509761388286334, "grad_norm": 0.5029488801956177, "learning_rate": 4.5119305856832975e-06, "loss": 0.5693, "step": 254 }, { "epoch": 0.5531453362255966, "grad_norm": 0.47164446115493774, "learning_rate": 4.490238611713666e-06, "loss": 0.564, "step": 255 }, { "epoch": 0.5553145336225597, "grad_norm": 0.49710217118263245, "learning_rate": 4.468546637744035e-06, "loss": 0.5826, "step": 256 }, { "epoch": 0.5574837310195228, "grad_norm": 0.48587456345558167, "learning_rate": 4.446854663774403e-06, "loss": 0.5816, "step": 257 }, { "epoch": 0.559652928416486, "grad_norm": 0.5498842597007751, "learning_rate": 4.425162689804773e-06, "loss": 0.5736, "step": 258 }, { "epoch": 0.561822125813449, "grad_norm": 0.4747694432735443, "learning_rate": 4.4034707158351415e-06, "loss": 0.574, "step": 259 }, { "epoch": 0.5639913232104121, "grad_norm": 0.9344505667686462, "learning_rate": 4.38177874186551e-06, "loss": 0.5957, "step": 260 }, { "epoch": 0.5661605206073753, "grad_norm": 0.5000290274620056, "learning_rate": 4.360086767895879e-06, "loss": 0.5774, "step": 261 }, { "epoch": 0.5683297180043384, "grad_norm": 0.5080724358558655, "learning_rate": 4.338394793926247e-06, "loss": 0.5688, "step": 262 }, { "epoch": 0.5704989154013015, "grad_norm": 0.5054919123649597, "learning_rate": 4.316702819956616e-06, "loss": 0.5834, "step": 263 }, { "epoch": 0.5726681127982647, "grad_norm": 0.5046693682670593, "learning_rate": 4.295010845986985e-06, "loss": 0.5687, "step": 264 }, { "epoch": 0.5748373101952278, "grad_norm": 0.5170451998710632, "learning_rate": 4.273318872017354e-06, "loss": 0.5497, "step": 265 }, { "epoch": 0.5770065075921909, "grad_norm": 0.5297961235046387, "learning_rate": 4.251626898047723e-06, "loss": 0.6013, "step": 266 }, { "epoch": 0.579175704989154, "grad_norm": 0.4973108172416687, "learning_rate": 4.229934924078091e-06, "loss": 0.5544, "step": 267 }, { "epoch": 0.5813449023861171, "grad_norm": 0.6242494583129883, "learning_rate": 4.208242950108461e-06, "loss": 0.5917, "step": 268 }, { "epoch": 0.5835140997830802, "grad_norm": 0.5227379202842712, "learning_rate": 4.186550976138829e-06, "loss": 0.5709, "step": 269 }, { "epoch": 0.5856832971800434, "grad_norm": 0.527258574962616, "learning_rate": 4.164859002169197e-06, "loss": 0.5746, "step": 270 }, { "epoch": 0.5878524945770065, "grad_norm": 0.5078297257423401, "learning_rate": 4.143167028199567e-06, "loss": 0.5569, "step": 271 }, { "epoch": 0.5900216919739696, "grad_norm": 0.7073683738708496, "learning_rate": 4.121475054229935e-06, "loss": 0.5806, "step": 272 }, { "epoch": 0.5921908893709328, "grad_norm": 0.6542203426361084, "learning_rate": 4.099783080260304e-06, "loss": 0.5963, "step": 273 }, { "epoch": 0.5943600867678959, "grad_norm": 0.4986463189125061, "learning_rate": 4.078091106290673e-06, "loss": 0.5786, "step": 274 }, { "epoch": 0.596529284164859, "grad_norm": 0.4862741231918335, "learning_rate": 4.056399132321042e-06, "loss": 0.5508, "step": 275 }, { "epoch": 0.5986984815618221, "grad_norm": 0.5407434105873108, "learning_rate": 4.03470715835141e-06, "loss": 0.5586, "step": 276 }, { "epoch": 0.6008676789587852, "grad_norm": 0.513647198677063, "learning_rate": 4.0130151843817785e-06, "loss": 0.5653, "step": 277 }, { "epoch": 0.6030368763557483, "grad_norm": 0.4941583573818207, "learning_rate": 3.991323210412148e-06, "loss": 0.5724, "step": 278 }, { "epoch": 0.6052060737527115, "grad_norm": 0.5022282600402832, "learning_rate": 3.9696312364425166e-06, "loss": 0.5597, "step": 279 }, { "epoch": 0.6073752711496746, "grad_norm": 0.5344776511192322, "learning_rate": 3.947939262472885e-06, "loss": 0.576, "step": 280 }, { "epoch": 0.6095444685466378, "grad_norm": 0.8711801171302795, "learning_rate": 3.926247288503255e-06, "loss": 0.5653, "step": 281 }, { "epoch": 0.6117136659436009, "grad_norm": 0.6406508684158325, "learning_rate": 3.904555314533623e-06, "loss": 0.5612, "step": 282 }, { "epoch": 0.613882863340564, "grad_norm": 0.5289674997329712, "learning_rate": 3.882863340563991e-06, "loss": 0.5553, "step": 283 }, { "epoch": 0.6160520607375272, "grad_norm": 0.49930331110954285, "learning_rate": 3.8611713665943606e-06, "loss": 0.5661, "step": 284 }, { "epoch": 0.6182212581344902, "grad_norm": 0.49601882696151733, "learning_rate": 3.839479392624729e-06, "loss": 0.5613, "step": 285 }, { "epoch": 0.6203904555314533, "grad_norm": 0.5551280975341797, "learning_rate": 3.817787418655098e-06, "loss": 0.5692, "step": 286 }, { "epoch": 0.6225596529284165, "grad_norm": 0.5268258452415466, "learning_rate": 3.7960954446854664e-06, "loss": 0.5653, "step": 287 }, { "epoch": 0.6247288503253796, "grad_norm": 0.5202245712280273, "learning_rate": 3.7744034707158355e-06, "loss": 0.5623, "step": 288 }, { "epoch": 0.6268980477223427, "grad_norm": 0.7391339540481567, "learning_rate": 3.752711496746204e-06, "loss": 0.5485, "step": 289 }, { "epoch": 0.6290672451193059, "grad_norm": 0.7942306995391846, "learning_rate": 3.7310195227765728e-06, "loss": 0.5693, "step": 290 }, { "epoch": 0.631236442516269, "grad_norm": 0.5168442130088806, "learning_rate": 3.709327548806942e-06, "loss": 0.563, "step": 291 }, { "epoch": 0.6334056399132321, "grad_norm": 0.5481964349746704, "learning_rate": 3.6876355748373104e-06, "loss": 0.5694, "step": 292 }, { "epoch": 0.6355748373101953, "grad_norm": 0.7699911594390869, "learning_rate": 3.665943600867679e-06, "loss": 0.602, "step": 293 }, { "epoch": 0.6377440347071583, "grad_norm": 0.5012184381484985, "learning_rate": 3.644251626898048e-06, "loss": 0.5566, "step": 294 }, { "epoch": 0.6399132321041214, "grad_norm": 0.5150504112243652, "learning_rate": 3.6225596529284167e-06, "loss": 0.5575, "step": 295 }, { "epoch": 0.6420824295010846, "grad_norm": 0.4825197160243988, "learning_rate": 3.6008676789587854e-06, "loss": 0.578, "step": 296 }, { "epoch": 0.6442516268980477, "grad_norm": 0.5708716511726379, "learning_rate": 3.5791757049891544e-06, "loss": 0.5748, "step": 297 }, { "epoch": 0.6464208242950108, "grad_norm": 0.49083611369132996, "learning_rate": 3.557483731019523e-06, "loss": 0.5656, "step": 298 }, { "epoch": 0.648590021691974, "grad_norm": 0.5466710329055786, "learning_rate": 3.5357917570498917e-06, "loss": 0.5808, "step": 299 }, { "epoch": 0.6507592190889371, "grad_norm": 0.6208142042160034, "learning_rate": 3.5140997830802603e-06, "loss": 0.5665, "step": 300 }, { "epoch": 0.6529284164859002, "grad_norm": 0.4832100570201874, "learning_rate": 3.4924078091106293e-06, "loss": 0.572, "step": 301 }, { "epoch": 0.6550976138828634, "grad_norm": 0.5140089392662048, "learning_rate": 3.470715835140998e-06, "loss": 0.5565, "step": 302 }, { "epoch": 0.6572668112798264, "grad_norm": 0.5033791661262512, "learning_rate": 3.4490238611713666e-06, "loss": 0.5522, "step": 303 }, { "epoch": 0.6594360086767896, "grad_norm": 0.4576267600059509, "learning_rate": 3.4273318872017357e-06, "loss": 0.5511, "step": 304 }, { "epoch": 0.6616052060737527, "grad_norm": 0.5113222002983093, "learning_rate": 3.4056399132321043e-06, "loss": 0.5736, "step": 305 }, { "epoch": 0.6637744034707158, "grad_norm": 0.528408408164978, "learning_rate": 3.383947939262473e-06, "loss": 0.5679, "step": 306 }, { "epoch": 0.665943600867679, "grad_norm": 0.4761233925819397, "learning_rate": 3.362255965292842e-06, "loss": 0.535, "step": 307 }, { "epoch": 0.6681127982646421, "grad_norm": 0.4666268527507782, "learning_rate": 3.3405639913232106e-06, "loss": 0.5731, "step": 308 }, { "epoch": 0.6702819956616052, "grad_norm": 0.5518378615379333, "learning_rate": 3.3188720173535792e-06, "loss": 0.5679, "step": 309 }, { "epoch": 0.6724511930585684, "grad_norm": 0.4833706021308899, "learning_rate": 3.2971800433839487e-06, "loss": 0.5615, "step": 310 }, { "epoch": 0.6746203904555315, "grad_norm": 0.5420160889625549, "learning_rate": 3.275488069414317e-06, "loss": 0.5529, "step": 311 }, { "epoch": 0.6767895878524945, "grad_norm": 0.6026318669319153, "learning_rate": 3.2537960954446855e-06, "loss": 0.5478, "step": 312 }, { "epoch": 0.6789587852494577, "grad_norm": 0.5486297607421875, "learning_rate": 3.232104121475054e-06, "loss": 0.5619, "step": 313 }, { "epoch": 0.6811279826464208, "grad_norm": 0.49950018525123596, "learning_rate": 3.210412147505423e-06, "loss": 0.5539, "step": 314 }, { "epoch": 0.6832971800433839, "grad_norm": 0.4778575301170349, "learning_rate": 3.188720173535792e-06, "loss": 0.546, "step": 315 }, { "epoch": 0.6854663774403471, "grad_norm": 0.4857572317123413, "learning_rate": 3.1670281995661605e-06, "loss": 0.5502, "step": 316 }, { "epoch": 0.6876355748373102, "grad_norm": 0.5437939167022705, "learning_rate": 3.14533622559653e-06, "loss": 0.5448, "step": 317 }, { "epoch": 0.6898047722342733, "grad_norm": 0.6015022397041321, "learning_rate": 3.123644251626898e-06, "loss": 0.5562, "step": 318 }, { "epoch": 0.6919739696312365, "grad_norm": 0.5425083041191101, "learning_rate": 3.1019522776572668e-06, "loss": 0.5496, "step": 319 }, { "epoch": 0.6941431670281996, "grad_norm": 0.4970496594905853, "learning_rate": 3.0802603036876362e-06, "loss": 0.5577, "step": 320 }, { "epoch": 0.6963123644251626, "grad_norm": 0.5393553972244263, "learning_rate": 3.058568329718005e-06, "loss": 0.5665, "step": 321 }, { "epoch": 0.6984815618221258, "grad_norm": 0.5788228511810303, "learning_rate": 3.036876355748373e-06, "loss": 0.5525, "step": 322 }, { "epoch": 0.7006507592190889, "grad_norm": 0.582302987575531, "learning_rate": 3.0151843817787417e-06, "loss": 0.5337, "step": 323 }, { "epoch": 0.702819956616052, "grad_norm": 0.5455271601676941, "learning_rate": 2.993492407809111e-06, "loss": 0.567, "step": 324 }, { "epoch": 0.7049891540130152, "grad_norm": 0.5106908679008484, "learning_rate": 2.9718004338394794e-06, "loss": 0.5572, "step": 325 }, { "epoch": 0.7071583514099783, "grad_norm": 0.5830085277557373, "learning_rate": 2.950108459869848e-06, "loss": 0.5799, "step": 326 }, { "epoch": 0.7093275488069414, "grad_norm": 0.543565571308136, "learning_rate": 2.9284164859002175e-06, "loss": 0.5486, "step": 327 }, { "epoch": 0.7114967462039046, "grad_norm": 0.6363551616668701, "learning_rate": 2.906724511930586e-06, "loss": 0.5589, "step": 328 }, { "epoch": 0.7136659436008677, "grad_norm": 0.5492388606071472, "learning_rate": 2.8850325379609543e-06, "loss": 0.5439, "step": 329 }, { "epoch": 0.7158351409978309, "grad_norm": 0.48855558037757874, "learning_rate": 2.8633405639913238e-06, "loss": 0.5342, "step": 330 }, { "epoch": 0.7180043383947939, "grad_norm": 0.5338431000709534, "learning_rate": 2.8416485900216924e-06, "loss": 0.5313, "step": 331 }, { "epoch": 0.720173535791757, "grad_norm": 0.5648824572563171, "learning_rate": 2.819956616052061e-06, "loss": 0.5521, "step": 332 }, { "epoch": 0.7223427331887202, "grad_norm": 0.5221781730651855, "learning_rate": 2.79826464208243e-06, "loss": 0.5445, "step": 333 }, { "epoch": 0.7245119305856833, "grad_norm": 0.4889802634716034, "learning_rate": 2.7765726681127987e-06, "loss": 0.5588, "step": 334 }, { "epoch": 0.7266811279826464, "grad_norm": 0.4981290400028229, "learning_rate": 2.7548806941431673e-06, "loss": 0.5709, "step": 335 }, { "epoch": 0.7288503253796096, "grad_norm": 0.5129498243331909, "learning_rate": 2.7331887201735356e-06, "loss": 0.5592, "step": 336 }, { "epoch": 0.7310195227765727, "grad_norm": 0.5097779631614685, "learning_rate": 2.711496746203905e-06, "loss": 0.5608, "step": 337 }, { "epoch": 0.7331887201735358, "grad_norm": 0.49256396293640137, "learning_rate": 2.6898047722342737e-06, "loss": 0.547, "step": 338 }, { "epoch": 0.735357917570499, "grad_norm": 0.49484625458717346, "learning_rate": 2.6681127982646423e-06, "loss": 0.5605, "step": 339 }, { "epoch": 0.737527114967462, "grad_norm": 0.4865284264087677, "learning_rate": 2.6464208242950113e-06, "loss": 0.552, "step": 340 }, { "epoch": 0.7396963123644251, "grad_norm": 0.5712233185768127, "learning_rate": 2.62472885032538e-06, "loss": 0.5229, "step": 341 }, { "epoch": 0.7418655097613883, "grad_norm": 0.4782271981239319, "learning_rate": 2.6030368763557486e-06, "loss": 0.5409, "step": 342 }, { "epoch": 0.7440347071583514, "grad_norm": 0.48679468035697937, "learning_rate": 2.5813449023861176e-06, "loss": 0.5502, "step": 343 }, { "epoch": 0.7462039045553145, "grad_norm": 0.49135392904281616, "learning_rate": 2.5596529284164863e-06, "loss": 0.552, "step": 344 }, { "epoch": 0.7483731019522777, "grad_norm": 0.4905093014240265, "learning_rate": 2.537960954446855e-06, "loss": 0.5388, "step": 345 }, { "epoch": 0.7505422993492408, "grad_norm": 0.4936610758304596, "learning_rate": 2.516268980477224e-06, "loss": 0.5408, "step": 346 }, { "epoch": 0.7527114967462039, "grad_norm": 0.4581211805343628, "learning_rate": 2.4945770065075926e-06, "loss": 0.5438, "step": 347 }, { "epoch": 0.754880694143167, "grad_norm": 0.9855303764343262, "learning_rate": 2.472885032537961e-06, "loss": 0.5422, "step": 348 }, { "epoch": 0.7570498915401301, "grad_norm": 0.4950404465198517, "learning_rate": 2.45119305856833e-06, "loss": 0.5488, "step": 349 }, { "epoch": 0.7592190889370932, "grad_norm": 0.5436333417892456, "learning_rate": 2.4295010845986985e-06, "loss": 0.529, "step": 350 }, { "epoch": 0.7613882863340564, "grad_norm": 0.6923372149467468, "learning_rate": 2.4078091106290675e-06, "loss": 0.566, "step": 351 }, { "epoch": 0.7635574837310195, "grad_norm": 0.4927901327610016, "learning_rate": 2.386117136659436e-06, "loss": 0.5445, "step": 352 }, { "epoch": 0.7657266811279827, "grad_norm": 0.6745809316635132, "learning_rate": 2.3644251626898048e-06, "loss": 0.5644, "step": 353 }, { "epoch": 0.7678958785249458, "grad_norm": 0.5441464185714722, "learning_rate": 2.342733188720174e-06, "loss": 0.5481, "step": 354 }, { "epoch": 0.7700650759219089, "grad_norm": 0.5015382170677185, "learning_rate": 2.3210412147505424e-06, "loss": 0.5632, "step": 355 }, { "epoch": 0.7722342733188721, "grad_norm": 0.48177745938301086, "learning_rate": 2.299349240780911e-06, "loss": 0.5435, "step": 356 }, { "epoch": 0.7744034707158352, "grad_norm": 0.6477603316307068, "learning_rate": 2.27765726681128e-06, "loss": 0.5488, "step": 357 }, { "epoch": 0.7765726681127982, "grad_norm": 0.49363404512405396, "learning_rate": 2.2559652928416487e-06, "loss": 0.5359, "step": 358 }, { "epoch": 0.7787418655097614, "grad_norm": 0.49064791202545166, "learning_rate": 2.2342733188720174e-06, "loss": 0.5708, "step": 359 }, { "epoch": 0.7809110629067245, "grad_norm": 0.5506601929664612, "learning_rate": 2.2125813449023864e-06, "loss": 0.5638, "step": 360 }, { "epoch": 0.7830802603036876, "grad_norm": 0.49600258469581604, "learning_rate": 2.190889370932755e-06, "loss": 0.5578, "step": 361 }, { "epoch": 0.7852494577006508, "grad_norm": 0.6045665740966797, "learning_rate": 2.1691973969631237e-06, "loss": 0.5387, "step": 362 }, { "epoch": 0.7874186550976139, "grad_norm": 0.4785294234752655, "learning_rate": 2.1475054229934923e-06, "loss": 0.5461, "step": 363 }, { "epoch": 0.789587852494577, "grad_norm": 0.505807101726532, "learning_rate": 2.1258134490238614e-06, "loss": 0.5421, "step": 364 }, { "epoch": 0.7917570498915402, "grad_norm": 0.5481471419334412, "learning_rate": 2.1041214750542304e-06, "loss": 0.5359, "step": 365 }, { "epoch": 0.7939262472885033, "grad_norm": 0.5260441899299622, "learning_rate": 2.0824295010845986e-06, "loss": 0.5801, "step": 366 }, { "epoch": 0.7960954446854663, "grad_norm": 0.5684512257575989, "learning_rate": 2.0607375271149677e-06, "loss": 0.5317, "step": 367 }, { "epoch": 0.7982646420824295, "grad_norm": 0.5323051810264587, "learning_rate": 2.0390455531453363e-06, "loss": 0.5669, "step": 368 }, { "epoch": 0.8004338394793926, "grad_norm": 0.6611142158508301, "learning_rate": 2.017353579175705e-06, "loss": 0.5582, "step": 369 }, { "epoch": 0.8026030368763557, "grad_norm": 0.47256988286972046, "learning_rate": 1.995661605206074e-06, "loss": 0.5276, "step": 370 }, { "epoch": 0.8047722342733189, "grad_norm": 0.6162890195846558, "learning_rate": 1.9739696312364426e-06, "loss": 0.5517, "step": 371 }, { "epoch": 0.806941431670282, "grad_norm": 0.5171302556991577, "learning_rate": 1.9522776572668117e-06, "loss": 0.5467, "step": 372 }, { "epoch": 0.8091106290672451, "grad_norm": 0.5168177485466003, "learning_rate": 1.9305856832971803e-06, "loss": 0.537, "step": 373 }, { "epoch": 0.8112798264642083, "grad_norm": 0.49433693289756775, "learning_rate": 1.908893709327549e-06, "loss": 0.5441, "step": 374 }, { "epoch": 0.8134490238611713, "grad_norm": 0.6254696249961853, "learning_rate": 1.8872017353579177e-06, "loss": 0.568, "step": 375 }, { "epoch": 0.8156182212581344, "grad_norm": 0.4846223294734955, "learning_rate": 1.8655097613882864e-06, "loss": 0.5314, "step": 376 }, { "epoch": 0.8177874186550976, "grad_norm": 0.5695744156837463, "learning_rate": 1.8438177874186552e-06, "loss": 0.5556, "step": 377 }, { "epoch": 0.8199566160520607, "grad_norm": 0.5135949850082397, "learning_rate": 1.822125813449024e-06, "loss": 0.5513, "step": 378 }, { "epoch": 0.8221258134490239, "grad_norm": 0.5039001107215881, "learning_rate": 1.8004338394793927e-06, "loss": 0.5395, "step": 379 }, { "epoch": 0.824295010845987, "grad_norm": 0.5153144598007202, "learning_rate": 1.7787418655097615e-06, "loss": 0.5499, "step": 380 }, { "epoch": 0.8264642082429501, "grad_norm": 0.6042090058326721, "learning_rate": 1.7570498915401301e-06, "loss": 0.561, "step": 381 }, { "epoch": 0.8286334056399133, "grad_norm": 0.5201963186264038, "learning_rate": 1.735357917570499e-06, "loss": 0.5514, "step": 382 }, { "epoch": 0.8308026030368764, "grad_norm": 0.7226424217224121, "learning_rate": 1.7136659436008678e-06, "loss": 0.5481, "step": 383 }, { "epoch": 0.8329718004338394, "grad_norm": 0.49008411169052124, "learning_rate": 1.6919739696312365e-06, "loss": 0.5577, "step": 384 }, { "epoch": 0.8351409978308026, "grad_norm": 0.5384876728057861, "learning_rate": 1.6702819956616053e-06, "loss": 0.5744, "step": 385 }, { "epoch": 0.8373101952277657, "grad_norm": 0.6859086751937866, "learning_rate": 1.6485900216919743e-06, "loss": 0.5495, "step": 386 }, { "epoch": 0.8394793926247288, "grad_norm": 0.5295915007591248, "learning_rate": 1.6268980477223428e-06, "loss": 0.5555, "step": 387 }, { "epoch": 0.841648590021692, "grad_norm": 0.49512577056884766, "learning_rate": 1.6052060737527116e-06, "loss": 0.5336, "step": 388 }, { "epoch": 0.8438177874186551, "grad_norm": 0.5301417112350464, "learning_rate": 1.5835140997830802e-06, "loss": 0.5366, "step": 389 }, { "epoch": 0.8459869848156182, "grad_norm": 0.5286414623260498, "learning_rate": 1.561822125813449e-06, "loss": 0.5301, "step": 390 }, { "epoch": 0.8481561822125814, "grad_norm": 0.4798189699649811, "learning_rate": 1.5401301518438181e-06, "loss": 0.5529, "step": 391 }, { "epoch": 0.8503253796095445, "grad_norm": 0.5175362825393677, "learning_rate": 1.5184381778741865e-06, "loss": 0.5418, "step": 392 }, { "epoch": 0.8524945770065075, "grad_norm": 0.5054755210876465, "learning_rate": 1.4967462039045556e-06, "loss": 0.5471, "step": 393 }, { "epoch": 0.8546637744034707, "grad_norm": 0.48426923155784607, "learning_rate": 1.475054229934924e-06, "loss": 0.543, "step": 394 }, { "epoch": 0.8568329718004338, "grad_norm": 0.49632081389427185, "learning_rate": 1.453362255965293e-06, "loss": 0.5336, "step": 395 }, { "epoch": 0.8590021691973969, "grad_norm": 0.5040681958198547, "learning_rate": 1.4316702819956619e-06, "loss": 0.5334, "step": 396 }, { "epoch": 0.8611713665943601, "grad_norm": 0.5700734853744507, "learning_rate": 1.4099783080260305e-06, "loss": 0.5365, "step": 397 }, { "epoch": 0.8633405639913232, "grad_norm": 0.5232958197593689, "learning_rate": 1.3882863340563994e-06, "loss": 0.5372, "step": 398 }, { "epoch": 0.8655097613882863, "grad_norm": 0.5089038610458374, "learning_rate": 1.3665943600867678e-06, "loss": 0.5431, "step": 399 }, { "epoch": 0.8676789587852495, "grad_norm": 0.5343663692474365, "learning_rate": 1.3449023861171368e-06, "loss": 0.5505, "step": 400 }, { "epoch": 0.8698481561822126, "grad_norm": 0.7061499953269958, "learning_rate": 1.3232104121475057e-06, "loss": 0.5394, "step": 401 }, { "epoch": 0.8720173535791758, "grad_norm": 0.5129175186157227, "learning_rate": 1.3015184381778743e-06, "loss": 0.5515, "step": 402 }, { "epoch": 0.8741865509761388, "grad_norm": 0.6809354424476624, "learning_rate": 1.2798264642082431e-06, "loss": 0.5453, "step": 403 }, { "epoch": 0.8763557483731019, "grad_norm": 0.5095650553703308, "learning_rate": 1.258134490238612e-06, "loss": 0.5559, "step": 404 }, { "epoch": 0.8785249457700651, "grad_norm": 0.707577645778656, "learning_rate": 1.2364425162689806e-06, "loss": 0.5318, "step": 405 }, { "epoch": 0.8806941431670282, "grad_norm": 0.5874403119087219, "learning_rate": 1.2147505422993492e-06, "loss": 0.5481, "step": 406 }, { "epoch": 0.8828633405639913, "grad_norm": 0.4959114193916321, "learning_rate": 1.193058568329718e-06, "loss": 0.5394, "step": 407 }, { "epoch": 0.8850325379609545, "grad_norm": 0.5033082365989685, "learning_rate": 1.171366594360087e-06, "loss": 0.5644, "step": 408 }, { "epoch": 0.8872017353579176, "grad_norm": 0.5417726635932922, "learning_rate": 1.1496746203904555e-06, "loss": 0.5209, "step": 409 }, { "epoch": 0.8893709327548807, "grad_norm": 0.488228440284729, "learning_rate": 1.1279826464208244e-06, "loss": 0.5406, "step": 410 }, { "epoch": 0.8915401301518439, "grad_norm": 0.9152227640151978, "learning_rate": 1.1062906724511932e-06, "loss": 0.532, "step": 411 }, { "epoch": 0.8937093275488069, "grad_norm": 0.5113642811775208, "learning_rate": 1.0845986984815618e-06, "loss": 0.5476, "step": 412 }, { "epoch": 0.89587852494577, "grad_norm": 0.5312101244926453, "learning_rate": 1.0629067245119307e-06, "loss": 0.5368, "step": 413 }, { "epoch": 0.8980477223427332, "grad_norm": 0.589953601360321, "learning_rate": 1.0412147505422993e-06, "loss": 0.5381, "step": 414 }, { "epoch": 0.9002169197396963, "grad_norm": 0.5381547212600708, "learning_rate": 1.0195227765726681e-06, "loss": 0.5492, "step": 415 }, { "epoch": 0.9023861171366594, "grad_norm": 0.5344692468643188, "learning_rate": 9.97830802603037e-07, "loss": 0.5507, "step": 416 }, { "epoch": 0.9045553145336226, "grad_norm": 0.48827818036079407, "learning_rate": 9.761388286334058e-07, "loss": 0.5365, "step": 417 }, { "epoch": 0.9067245119305857, "grad_norm": 0.5503534078598022, "learning_rate": 9.544468546637745e-07, "loss": 0.531, "step": 418 }, { "epoch": 0.9088937093275488, "grad_norm": 0.5214491486549377, "learning_rate": 9.327548806941432e-07, "loss": 0.5492, "step": 419 }, { "epoch": 0.911062906724512, "grad_norm": 0.5597646236419678, "learning_rate": 9.11062906724512e-07, "loss": 0.5273, "step": 420 }, { "epoch": 0.913232104121475, "grad_norm": 0.4814266264438629, "learning_rate": 8.893709327548808e-07, "loss": 0.5467, "step": 421 }, { "epoch": 0.9154013015184381, "grad_norm": 0.5103883147239685, "learning_rate": 8.676789587852495e-07, "loss": 0.5514, "step": 422 }, { "epoch": 0.9175704989154013, "grad_norm": 0.4807824492454529, "learning_rate": 8.459869848156182e-07, "loss": 0.5377, "step": 423 }, { "epoch": 0.9197396963123644, "grad_norm": 0.49137192964553833, "learning_rate": 8.242950108459872e-07, "loss": 0.5384, "step": 424 }, { "epoch": 0.9219088937093276, "grad_norm": 0.47830575704574585, "learning_rate": 8.026030368763558e-07, "loss": 0.5458, "step": 425 }, { "epoch": 0.9240780911062907, "grad_norm": 0.5331500768661499, "learning_rate": 7.809110629067245e-07, "loss": 0.514, "step": 426 }, { "epoch": 0.9262472885032538, "grad_norm": 0.4643237888813019, "learning_rate": 7.592190889370933e-07, "loss": 0.5404, "step": 427 }, { "epoch": 0.928416485900217, "grad_norm": 0.49474000930786133, "learning_rate": 7.37527114967462e-07, "loss": 0.5687, "step": 428 }, { "epoch": 0.93058568329718, "grad_norm": 0.551901638507843, "learning_rate": 7.158351409978309e-07, "loss": 0.5485, "step": 429 }, { "epoch": 0.9327548806941431, "grad_norm": 0.5091038346290588, "learning_rate": 6.941431670281997e-07, "loss": 0.545, "step": 430 }, { "epoch": 0.9349240780911063, "grad_norm": 0.4894218444824219, "learning_rate": 6.724511930585684e-07, "loss": 0.5391, "step": 431 }, { "epoch": 0.9370932754880694, "grad_norm": 0.5032333731651306, "learning_rate": 6.507592190889371e-07, "loss": 0.5404, "step": 432 }, { "epoch": 0.9392624728850325, "grad_norm": 0.5365532040596008, "learning_rate": 6.29067245119306e-07, "loss": 0.524, "step": 433 }, { "epoch": 0.9414316702819957, "grad_norm": 0.47491738200187683, "learning_rate": 6.073752711496746e-07, "loss": 0.5451, "step": 434 }, { "epoch": 0.9436008676789588, "grad_norm": 0.4964900314807892, "learning_rate": 5.856832971800435e-07, "loss": 0.5465, "step": 435 }, { "epoch": 0.9457700650759219, "grad_norm": 0.4745555520057678, "learning_rate": 5.639913232104122e-07, "loss": 0.5432, "step": 436 }, { "epoch": 0.9479392624728851, "grad_norm": 0.6818556785583496, "learning_rate": 5.422993492407809e-07, "loss": 0.5382, "step": 437 }, { "epoch": 0.9501084598698482, "grad_norm": 0.5324418544769287, "learning_rate": 5.206073752711497e-07, "loss": 0.5733, "step": 438 }, { "epoch": 0.9522776572668112, "grad_norm": 0.5001943111419678, "learning_rate": 4.989154013015185e-07, "loss": 0.5331, "step": 439 }, { "epoch": 0.9544468546637744, "grad_norm": 0.5757445096969604, "learning_rate": 4.772234273318872e-07, "loss": 0.5383, "step": 440 }, { "epoch": 0.9566160520607375, "grad_norm": 0.6072180271148682, "learning_rate": 4.55531453362256e-07, "loss": 0.5509, "step": 441 }, { "epoch": 0.9587852494577006, "grad_norm": 0.5225549936294556, "learning_rate": 4.3383947939262475e-07, "loss": 0.5279, "step": 442 }, { "epoch": 0.9609544468546638, "grad_norm": 0.4668637216091156, "learning_rate": 4.121475054229936e-07, "loss": 0.5488, "step": 443 }, { "epoch": 0.9631236442516269, "grad_norm": 0.46520882844924927, "learning_rate": 3.9045553145336227e-07, "loss": 0.5272, "step": 444 }, { "epoch": 0.96529284164859, "grad_norm": 0.45933347940444946, "learning_rate": 3.68763557483731e-07, "loss": 0.5526, "step": 445 }, { "epoch": 0.9674620390455532, "grad_norm": 0.6567378044128418, "learning_rate": 3.4707158351409984e-07, "loss": 0.5387, "step": 446 }, { "epoch": 0.9696312364425163, "grad_norm": 0.49450600147247314, "learning_rate": 3.2537960954446857e-07, "loss": 0.5329, "step": 447 }, { "epoch": 0.9718004338394793, "grad_norm": 0.5677459239959717, "learning_rate": 3.036876355748373e-07, "loss": 0.5459, "step": 448 }, { "epoch": 0.9739696312364425, "grad_norm": 0.4713030755519867, "learning_rate": 2.819956616052061e-07, "loss": 0.5394, "step": 449 }, { "epoch": 0.9761388286334056, "grad_norm": 0.5460206866264343, "learning_rate": 2.6030368763557483e-07, "loss": 0.5318, "step": 450 }, { "epoch": 0.9783080260303688, "grad_norm": 0.4780767560005188, "learning_rate": 2.386117136659436e-07, "loss": 0.5407, "step": 451 }, { "epoch": 0.9804772234273319, "grad_norm": 0.4803195297718048, "learning_rate": 2.1691973969631237e-07, "loss": 0.5199, "step": 452 }, { "epoch": 0.982646420824295, "grad_norm": 0.5350447297096252, "learning_rate": 1.9522776572668113e-07, "loss": 0.5434, "step": 453 }, { "epoch": 0.9848156182212582, "grad_norm": 0.5030990839004517, "learning_rate": 1.7353579175704992e-07, "loss": 0.5368, "step": 454 }, { "epoch": 0.9869848156182213, "grad_norm": 0.4834834933280945, "learning_rate": 1.5184381778741865e-07, "loss": 0.5386, "step": 455 }, { "epoch": 0.9891540130151844, "grad_norm": 0.5929272174835205, "learning_rate": 1.3015184381778741e-07, "loss": 0.5452, "step": 456 }, { "epoch": 0.9913232104121475, "grad_norm": 0.4731636047363281, "learning_rate": 1.0845986984815619e-07, "loss": 0.5294, "step": 457 }, { "epoch": 0.9934924078091106, "grad_norm": 0.48876383900642395, "learning_rate": 8.676789587852496e-08, "loss": 0.539, "step": 458 }, { "epoch": 0.9956616052060737, "grad_norm": 0.4793936610221863, "learning_rate": 6.507592190889371e-08, "loss": 0.5496, "step": 459 }, { "epoch": 0.9978308026030369, "grad_norm": 0.46681129932403564, "learning_rate": 4.338394793926248e-08, "loss": 0.5283, "step": 460 }, { "epoch": 1.0, "grad_norm": 0.5591140389442444, "learning_rate": 2.169197396963124e-08, "loss": 0.5435, "step": 461 } ], "logging_steps": 1.0, "max_steps": 461, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 0, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.281141529490227e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }