diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,14582 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0, + "eval_steps": 500, + "global_step": 2077, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00048152160828217165, + "grad_norm": 1.4487173557281494, + "learning_rate": 0.0, + "loss": 1.7129, + "step": 1 + }, + { + "epoch": 0.0009630432165643433, + "grad_norm": 2.391146183013916, + "learning_rate": 3.1746031746031746e-06, + "loss": 2.3104, + "step": 2 + }, + { + "epoch": 0.001444564824846515, + "grad_norm": 0.6623585820198059, + "learning_rate": 6.349206349206349e-06, + "loss": 1.2074, + "step": 3 + }, + { + "epoch": 0.0019260864331286866, + "grad_norm": 2.9669229984283447, + "learning_rate": 9.523809523809523e-06, + "loss": 1.7921, + "step": 4 + }, + { + "epoch": 0.002407608041410858, + "grad_norm": 1.4202985763549805, + "learning_rate": 1.2698412698412699e-05, + "loss": 1.5058, + "step": 5 + }, + { + "epoch": 0.00288912964969303, + "grad_norm": 0.908497154712677, + "learning_rate": 1.5873015873015872e-05, + "loss": 1.3187, + "step": 6 + }, + { + "epoch": 0.0033706512579752016, + "grad_norm": 4.113708019256592, + "learning_rate": 1.9047619047619046e-05, + "loss": 2.1155, + "step": 7 + }, + { + "epoch": 0.003852172866257373, + "grad_norm": 2.203523635864258, + "learning_rate": 2.2222222222222223e-05, + "loss": 1.577, + "step": 8 + }, + { + "epoch": 0.004333694474539545, + "grad_norm": 3.0024302005767822, + "learning_rate": 2.5396825396825397e-05, + "loss": 1.7602, + "step": 9 + }, + { + "epoch": 0.004815216082821716, + "grad_norm": 2.271422863006592, + "learning_rate": 2.857142857142857e-05, + "loss": 1.6359, + "step": 10 + }, + { + "epoch": 0.005296737691103889, + "grad_norm": 1.6946759223937988, + "learning_rate": 3.1746031746031745e-05, + "loss": 1.4716, + "step": 11 + }, + { + "epoch": 0.00577825929938606, + "grad_norm": 1.5421451330184937, + "learning_rate": 3.492063492063492e-05, + "loss": 0.9894, + "step": 12 + }, + { + "epoch": 0.006259780907668232, + "grad_norm": 2.4348111152648926, + "learning_rate": 3.809523809523809e-05, + "loss": 1.326, + "step": 13 + }, + { + "epoch": 0.006741302515950403, + "grad_norm": 2.774260997772217, + "learning_rate": 4.126984126984127e-05, + "loss": 1.3238, + "step": 14 + }, + { + "epoch": 0.007222824124232575, + "grad_norm": 1.2383824586868286, + "learning_rate": 4.4444444444444447e-05, + "loss": 1.3033, + "step": 15 + }, + { + "epoch": 0.007704345732514746, + "grad_norm": 2.3136370182037354, + "learning_rate": 4.761904761904762e-05, + "loss": 1.9822, + "step": 16 + }, + { + "epoch": 0.008185867340796918, + "grad_norm": 0.5560957193374634, + "learning_rate": 5.0793650793650794e-05, + "loss": 1.3964, + "step": 17 + }, + { + "epoch": 0.00866738894907909, + "grad_norm": 1.1794217824935913, + "learning_rate": 5.396825396825397e-05, + "loss": 1.2084, + "step": 18 + }, + { + "epoch": 0.009148910557361261, + "grad_norm": 0.8785364627838135, + "learning_rate": 5.714285714285714e-05, + "loss": 1.1667, + "step": 19 + }, + { + "epoch": 0.009630432165643433, + "grad_norm": 2.39178466796875, + "learning_rate": 6.0317460317460316e-05, + "loss": 1.8301, + "step": 20 + }, + { + "epoch": 0.010111953773925604, + "grad_norm": 1.0043989419937134, + "learning_rate": 6.349206349206349e-05, + "loss": 1.346, + "step": 21 + }, + { + "epoch": 0.010593475382207777, + "grad_norm": 1.1557499170303345, + "learning_rate": 6.666666666666667e-05, + "loss": 1.4663, + "step": 22 + }, + { + "epoch": 0.011074996990489949, + "grad_norm": 0.5397685170173645, + "learning_rate": 6.984126984126984e-05, + "loss": 0.9156, + "step": 23 + }, + { + "epoch": 0.01155651859877212, + "grad_norm": 0.86281418800354, + "learning_rate": 7.301587301587302e-05, + "loss": 0.9897, + "step": 24 + }, + { + "epoch": 0.012038040207054292, + "grad_norm": 2.705343723297119, + "learning_rate": 7.619047619047618e-05, + "loss": 1.6137, + "step": 25 + }, + { + "epoch": 0.012519561815336464, + "grad_norm": 0.8081772327423096, + "learning_rate": 7.936507936507937e-05, + "loss": 0.9733, + "step": 26 + }, + { + "epoch": 0.013001083423618635, + "grad_norm": 0.5619446039199829, + "learning_rate": 8.253968253968255e-05, + "loss": 1.0055, + "step": 27 + }, + { + "epoch": 0.013482605031900807, + "grad_norm": 0.9184542298316956, + "learning_rate": 8.571428571428571e-05, + "loss": 0.7692, + "step": 28 + }, + { + "epoch": 0.013964126640182978, + "grad_norm": 0.7189937829971313, + "learning_rate": 8.888888888888889e-05, + "loss": 0.9745, + "step": 29 + }, + { + "epoch": 0.01444564824846515, + "grad_norm": 1.1548259258270264, + "learning_rate": 9.206349206349206e-05, + "loss": 1.0737, + "step": 30 + }, + { + "epoch": 0.014927169856747321, + "grad_norm": 1.7942982912063599, + "learning_rate": 9.523809523809524e-05, + "loss": 0.8184, + "step": 31 + }, + { + "epoch": 0.015408691465029493, + "grad_norm": 0.9638906121253967, + "learning_rate": 9.841269841269841e-05, + "loss": 1.2128, + "step": 32 + }, + { + "epoch": 0.015890213073311666, + "grad_norm": 0.9639189839363098, + "learning_rate": 0.00010158730158730159, + "loss": 1.0256, + "step": 33 + }, + { + "epoch": 0.016371734681593836, + "grad_norm": 0.6873184442520142, + "learning_rate": 0.00010476190476190477, + "loss": 1.048, + "step": 34 + }, + { + "epoch": 0.01685325628987601, + "grad_norm": 1.1122071743011475, + "learning_rate": 0.00010793650793650794, + "loss": 0.6637, + "step": 35 + }, + { + "epoch": 0.01733477789815818, + "grad_norm": 0.5951284766197205, + "learning_rate": 0.00011111111111111112, + "loss": 1.3751, + "step": 36 + }, + { + "epoch": 0.017816299506440352, + "grad_norm": 0.6848040819168091, + "learning_rate": 0.00011428571428571428, + "loss": 0.6586, + "step": 37 + }, + { + "epoch": 0.018297821114722522, + "grad_norm": 0.9294832944869995, + "learning_rate": 0.00011746031746031746, + "loss": 1.2054, + "step": 38 + }, + { + "epoch": 0.018779342723004695, + "grad_norm": 0.853848934173584, + "learning_rate": 0.00012063492063492063, + "loss": 0.6379, + "step": 39 + }, + { + "epoch": 0.019260864331286865, + "grad_norm": 1.1381187438964844, + "learning_rate": 0.0001238095238095238, + "loss": 1.2677, + "step": 40 + }, + { + "epoch": 0.01974238593956904, + "grad_norm": 0.812929093837738, + "learning_rate": 0.00012698412698412698, + "loss": 1.2011, + "step": 41 + }, + { + "epoch": 0.020223907547851208, + "grad_norm": 0.9746362566947937, + "learning_rate": 0.00013015873015873017, + "loss": 1.1164, + "step": 42 + }, + { + "epoch": 0.02070542915613338, + "grad_norm": 0.5625665783882141, + "learning_rate": 0.00013333333333333334, + "loss": 0.9418, + "step": 43 + }, + { + "epoch": 0.021186950764415555, + "grad_norm": 0.7534303665161133, + "learning_rate": 0.0001365079365079365, + "loss": 0.8973, + "step": 44 + }, + { + "epoch": 0.021668472372697724, + "grad_norm": 0.8771131038665771, + "learning_rate": 0.00013968253968253967, + "loss": 0.9622, + "step": 45 + }, + { + "epoch": 0.022149993980979898, + "grad_norm": 1.096232295036316, + "learning_rate": 0.00014285714285714287, + "loss": 1.3677, + "step": 46 + }, + { + "epoch": 0.022631515589262068, + "grad_norm": 2.257794141769409, + "learning_rate": 0.00014603174603174603, + "loss": 1.0777, + "step": 47 + }, + { + "epoch": 0.02311303719754424, + "grad_norm": 0.9336938858032227, + "learning_rate": 0.00014920634920634923, + "loss": 1.215, + "step": 48 + }, + { + "epoch": 0.02359455880582641, + "grad_norm": 2.0201683044433594, + "learning_rate": 0.00015238095238095237, + "loss": 0.974, + "step": 49 + }, + { + "epoch": 0.024076080414108584, + "grad_norm": 0.36621397733688354, + "learning_rate": 0.00015555555555555556, + "loss": 0.5777, + "step": 50 + }, + { + "epoch": 0.024557602022390754, + "grad_norm": 0.9629436135292053, + "learning_rate": 0.00015873015873015873, + "loss": 0.9346, + "step": 51 + }, + { + "epoch": 0.025039123630672927, + "grad_norm": 1.607643485069275, + "learning_rate": 0.00016190476190476192, + "loss": 1.8116, + "step": 52 + }, + { + "epoch": 0.025520645238955097, + "grad_norm": 0.7304220199584961, + "learning_rate": 0.0001650793650793651, + "loss": 0.7118, + "step": 53 + }, + { + "epoch": 0.02600216684723727, + "grad_norm": 0.6517884135246277, + "learning_rate": 0.00016825396825396826, + "loss": 0.6906, + "step": 54 + }, + { + "epoch": 0.02648368845551944, + "grad_norm": 0.833574652671814, + "learning_rate": 0.00017142857142857143, + "loss": 1.3414, + "step": 55 + }, + { + "epoch": 0.026965210063801613, + "grad_norm": 1.3700557947158813, + "learning_rate": 0.00017460317460317462, + "loss": 0.7933, + "step": 56 + }, + { + "epoch": 0.027446731672083786, + "grad_norm": 1.6920475959777832, + "learning_rate": 0.00017777777777777779, + "loss": 1.2288, + "step": 57 + }, + { + "epoch": 0.027928253280365956, + "grad_norm": 0.6903666853904724, + "learning_rate": 0.00018095238095238095, + "loss": 0.7095, + "step": 58 + }, + { + "epoch": 0.02840977488864813, + "grad_norm": 1.1906098127365112, + "learning_rate": 0.00018412698412698412, + "loss": 0.9009, + "step": 59 + }, + { + "epoch": 0.0288912964969303, + "grad_norm": 1.0149602890014648, + "learning_rate": 0.00018730158730158731, + "loss": 0.6004, + "step": 60 + }, + { + "epoch": 0.029372818105212473, + "grad_norm": 0.5224968791007996, + "learning_rate": 0.00019047619047619048, + "loss": 0.9209, + "step": 61 + }, + { + "epoch": 0.029854339713494642, + "grad_norm": 1.0166723728179932, + "learning_rate": 0.00019365079365079365, + "loss": 0.7085, + "step": 62 + }, + { + "epoch": 0.030335861321776816, + "grad_norm": 0.8379011750221252, + "learning_rate": 0.00019682539682539682, + "loss": 0.7444, + "step": 63 + }, + { + "epoch": 0.030817382930058985, + "grad_norm": 1.1669416427612305, + "learning_rate": 0.0002, + "loss": 1.12, + "step": 64 + }, + { + "epoch": 0.03129890453834116, + "grad_norm": 1.1994937658309937, + "learning_rate": 0.00019999987833918285, + "loss": 1.0237, + "step": 65 + }, + { + "epoch": 0.03178042614662333, + "grad_norm": 0.520425021648407, + "learning_rate": 0.00019999951335702735, + "loss": 1.0711, + "step": 66 + }, + { + "epoch": 0.0322619477549055, + "grad_norm": 1.4933056831359863, + "learning_rate": 0.00019999890505442158, + "loss": 0.9168, + "step": 67 + }, + { + "epoch": 0.03274346936318767, + "grad_norm": 1.2318795919418335, + "learning_rate": 0.0001999980534328457, + "loss": 1.1426, + "step": 68 + }, + { + "epoch": 0.033224990971469845, + "grad_norm": 1.2501200437545776, + "learning_rate": 0.0001999969584943719, + "loss": 1.14, + "step": 69 + }, + { + "epoch": 0.03370651257975202, + "grad_norm": 1.9055863618850708, + "learning_rate": 0.00019999562024166438, + "loss": 0.8023, + "step": 70 + }, + { + "epoch": 0.03418803418803419, + "grad_norm": 0.5718879699707031, + "learning_rate": 0.0001999940386779794, + "loss": 0.3126, + "step": 71 + }, + { + "epoch": 0.03466955579631636, + "grad_norm": 0.7291958928108215, + "learning_rate": 0.00019999221380716527, + "loss": 0.7172, + "step": 72 + }, + { + "epoch": 0.03515107740459853, + "grad_norm": 1.2963175773620605, + "learning_rate": 0.00019999014563366226, + "loss": 0.9218, + "step": 73 + }, + { + "epoch": 0.035632599012880704, + "grad_norm": 1.3575979471206665, + "learning_rate": 0.00019998783416250268, + "loss": 1.0466, + "step": 74 + }, + { + "epoch": 0.03611412062116288, + "grad_norm": 1.1832976341247559, + "learning_rate": 0.0001999852793993109, + "loss": 0.8877, + "step": 75 + }, + { + "epoch": 0.036595642229445044, + "grad_norm": 0.5200212597846985, + "learning_rate": 0.00019998248135030315, + "loss": 0.9763, + "step": 76 + }, + { + "epoch": 0.03707716383772722, + "grad_norm": 0.32084816694259644, + "learning_rate": 0.00019997944002228774, + "loss": 0.8581, + "step": 77 + }, + { + "epoch": 0.03755868544600939, + "grad_norm": 1.1948219537734985, + "learning_rate": 0.00019997615542266482, + "loss": 0.709, + "step": 78 + }, + { + "epoch": 0.038040207054291564, + "grad_norm": 0.6765559911727905, + "learning_rate": 0.00019997262755942655, + "loss": 0.6078, + "step": 79 + }, + { + "epoch": 0.03852172866257373, + "grad_norm": 1.0067569017410278, + "learning_rate": 0.000199968856441157, + "loss": 0.7802, + "step": 80 + }, + { + "epoch": 0.0390032502708559, + "grad_norm": 0.9474171996116638, + "learning_rate": 0.0001999648420770321, + "loss": 0.748, + "step": 81 + }, + { + "epoch": 0.03948477187913808, + "grad_norm": 0.7215222716331482, + "learning_rate": 0.0001999605844768197, + "loss": 1.2064, + "step": 82 + }, + { + "epoch": 0.03996629348742025, + "grad_norm": 0.8327376246452332, + "learning_rate": 0.00019995608365087946, + "loss": 1.214, + "step": 83 + }, + { + "epoch": 0.040447815095702416, + "grad_norm": 0.7590318918228149, + "learning_rate": 0.0001999513396101628, + "loss": 0.9192, + "step": 84 + }, + { + "epoch": 0.04092933670398459, + "grad_norm": 1.0768215656280518, + "learning_rate": 0.00019994635236621306, + "loss": 0.509, + "step": 85 + }, + { + "epoch": 0.04141085831226676, + "grad_norm": 0.596147358417511, + "learning_rate": 0.00019994112193116528, + "loss": 0.9004, + "step": 86 + }, + { + "epoch": 0.041892379920548936, + "grad_norm": 0.8125873804092407, + "learning_rate": 0.00019993564831774618, + "loss": 1.2089, + "step": 87 + }, + { + "epoch": 0.04237390152883111, + "grad_norm": 2.323214054107666, + "learning_rate": 0.00019992993153927432, + "loss": 0.9856, + "step": 88 + }, + { + "epoch": 0.042855423137113276, + "grad_norm": 1.0737050771713257, + "learning_rate": 0.00019992397160965982, + "loss": 0.5664, + "step": 89 + }, + { + "epoch": 0.04333694474539545, + "grad_norm": 0.5349381566047668, + "learning_rate": 0.0001999177685434045, + "loss": 1.1484, + "step": 90 + }, + { + "epoch": 0.04381846635367762, + "grad_norm": 1.065415382385254, + "learning_rate": 0.00019991132235560176, + "loss": 1.3118, + "step": 91 + }, + { + "epoch": 0.044299987961959796, + "grad_norm": 0.8875418305397034, + "learning_rate": 0.00019990463306193652, + "loss": 1.0626, + "step": 92 + }, + { + "epoch": 0.04478150957024196, + "grad_norm": 0.6135897040367126, + "learning_rate": 0.00019989770067868533, + "loss": 0.7007, + "step": 93 + }, + { + "epoch": 0.045263031178524135, + "grad_norm": 1.3404655456542969, + "learning_rate": 0.00019989052522271622, + "loss": 1.2922, + "step": 94 + }, + { + "epoch": 0.04574455278680631, + "grad_norm": 0.8707893490791321, + "learning_rate": 0.00019988310671148848, + "loss": 0.9461, + "step": 95 + }, + { + "epoch": 0.04622607439508848, + "grad_norm": 0.8725842833518982, + "learning_rate": 0.00019987544516305311, + "loss": 1.2371, + "step": 96 + }, + { + "epoch": 0.04670759600337065, + "grad_norm": 0.83966064453125, + "learning_rate": 0.00019986754059605222, + "loss": 0.6515, + "step": 97 + }, + { + "epoch": 0.04718911761165282, + "grad_norm": 0.6398138403892517, + "learning_rate": 0.00019985939302971938, + "loss": 0.8191, + "step": 98 + }, + { + "epoch": 0.047670639219934995, + "grad_norm": 0.7741346955299377, + "learning_rate": 0.00019985100248387933, + "loss": 1.1037, + "step": 99 + }, + { + "epoch": 0.04815216082821717, + "grad_norm": 1.0956766605377197, + "learning_rate": 0.00019984236897894816, + "loss": 0.5513, + "step": 100 + }, + { + "epoch": 0.04863368243649934, + "grad_norm": 0.7802196741104126, + "learning_rate": 0.000199833492535933, + "loss": 0.5401, + "step": 101 + }, + { + "epoch": 0.04911520404478151, + "grad_norm": 0.8452202677726746, + "learning_rate": 0.00019982437317643217, + "loss": 0.9785, + "step": 102 + }, + { + "epoch": 0.04959672565306368, + "grad_norm": 1.0651404857635498, + "learning_rate": 0.00019981501092263503, + "loss": 1.1675, + "step": 103 + }, + { + "epoch": 0.050078247261345854, + "grad_norm": 1.3210868835449219, + "learning_rate": 0.00019980540579732196, + "loss": 0.7879, + "step": 104 + }, + { + "epoch": 0.05055976886962803, + "grad_norm": 1.8965524435043335, + "learning_rate": 0.00019979555782386434, + "loss": 1.0122, + "step": 105 + }, + { + "epoch": 0.051041290477910194, + "grad_norm": 0.71902996301651, + "learning_rate": 0.00019978546702622443, + "loss": 1.0812, + "step": 106 + }, + { + "epoch": 0.05152281208619237, + "grad_norm": 0.6849554181098938, + "learning_rate": 0.00019977513342895532, + "loss": 0.3471, + "step": 107 + }, + { + "epoch": 0.05200433369447454, + "grad_norm": 1.114531397819519, + "learning_rate": 0.00019976455705720083, + "loss": 0.8091, + "step": 108 + }, + { + "epoch": 0.05248585530275671, + "grad_norm": 0.9422350525856018, + "learning_rate": 0.0001997537379366956, + "loss": 1.328, + "step": 109 + }, + { + "epoch": 0.05296737691103888, + "grad_norm": 0.8205842971801758, + "learning_rate": 0.00019974267609376494, + "loss": 1.325, + "step": 110 + }, + { + "epoch": 0.05344889851932105, + "grad_norm": 1.2657097578048706, + "learning_rate": 0.00019973137155532462, + "loss": 0.8736, + "step": 111 + }, + { + "epoch": 0.053930420127603226, + "grad_norm": 1.1228113174438477, + "learning_rate": 0.00019971982434888107, + "loss": 1.0789, + "step": 112 + }, + { + "epoch": 0.0544119417358854, + "grad_norm": 1.5230176448822021, + "learning_rate": 0.00019970803450253114, + "loss": 1.3641, + "step": 113 + }, + { + "epoch": 0.05489346334416757, + "grad_norm": 1.4824843406677246, + "learning_rate": 0.0001996960020449621, + "loss": 1.023, + "step": 114 + }, + { + "epoch": 0.05537498495244974, + "grad_norm": 0.8390150666236877, + "learning_rate": 0.00019968372700545145, + "loss": 1.0051, + "step": 115 + }, + { + "epoch": 0.05585650656073191, + "grad_norm": 0.5034816265106201, + "learning_rate": 0.00019967120941386709, + "loss": 0.8851, + "step": 116 + }, + { + "epoch": 0.056338028169014086, + "grad_norm": 0.5197919011116028, + "learning_rate": 0.000199658449300667, + "loss": 0.9949, + "step": 117 + }, + { + "epoch": 0.05681954977729626, + "grad_norm": 0.7896221280097961, + "learning_rate": 0.00019964544669689928, + "loss": 1.3479, + "step": 118 + }, + { + "epoch": 0.057301071385578425, + "grad_norm": 0.9110464453697205, + "learning_rate": 0.00019963220163420214, + "loss": 1.4343, + "step": 119 + }, + { + "epoch": 0.0577825929938606, + "grad_norm": 0.8352183699607849, + "learning_rate": 0.0001996187141448036, + "loss": 0.5157, + "step": 120 + }, + { + "epoch": 0.05826411460214277, + "grad_norm": 0.6002696752548218, + "learning_rate": 0.0001996049842615217, + "loss": 0.8888, + "step": 121 + }, + { + "epoch": 0.058745636210424945, + "grad_norm": 1.1759757995605469, + "learning_rate": 0.0001995910120177642, + "loss": 1.1579, + "step": 122 + }, + { + "epoch": 0.05922715781870711, + "grad_norm": 0.6822336912155151, + "learning_rate": 0.00019957679744752859, + "loss": 0.8854, + "step": 123 + }, + { + "epoch": 0.059708679426989285, + "grad_norm": 1.1303515434265137, + "learning_rate": 0.00019956234058540195, + "loss": 0.9317, + "step": 124 + }, + { + "epoch": 0.06019020103527146, + "grad_norm": 1.6187382936477661, + "learning_rate": 0.00019954764146656105, + "loss": 0.6916, + "step": 125 + }, + { + "epoch": 0.06067172264355363, + "grad_norm": 0.7719541788101196, + "learning_rate": 0.00019953270012677195, + "loss": 0.6124, + "step": 126 + }, + { + "epoch": 0.0611532442518358, + "grad_norm": 0.6049287915229797, + "learning_rate": 0.00019951751660239015, + "loss": 0.8589, + "step": 127 + }, + { + "epoch": 0.06163476586011797, + "grad_norm": 1.2466943264007568, + "learning_rate": 0.00019950209093036052, + "loss": 0.7661, + "step": 128 + }, + { + "epoch": 0.062116287468400144, + "grad_norm": 0.6226364970207214, + "learning_rate": 0.000199486423148217, + "loss": 0.7367, + "step": 129 + }, + { + "epoch": 0.06259780907668232, + "grad_norm": 0.714819610118866, + "learning_rate": 0.00019947051329408276, + "loss": 0.6814, + "step": 130 + }, + { + "epoch": 0.06307933068496449, + "grad_norm": 0.8032063841819763, + "learning_rate": 0.00019945436140666981, + "loss": 1.4286, + "step": 131 + }, + { + "epoch": 0.06356085229324666, + "grad_norm": 0.940105676651001, + "learning_rate": 0.0001994379675252793, + "loss": 1.2501, + "step": 132 + }, + { + "epoch": 0.06404237390152884, + "grad_norm": 1.491990327835083, + "learning_rate": 0.00019942133168980103, + "loss": 1.5854, + "step": 133 + }, + { + "epoch": 0.064523895509811, + "grad_norm": 0.8070606589317322, + "learning_rate": 0.00019940445394071355, + "loss": 0.9551, + "step": 134 + }, + { + "epoch": 0.06500541711809317, + "grad_norm": 0.8355477452278137, + "learning_rate": 0.0001993873343190842, + "loss": 0.7561, + "step": 135 + }, + { + "epoch": 0.06548693872637534, + "grad_norm": 0.6875390410423279, + "learning_rate": 0.00019936997286656855, + "loss": 1.025, + "step": 136 + }, + { + "epoch": 0.06596846033465752, + "grad_norm": 0.6399025321006775, + "learning_rate": 0.00019935236962541092, + "loss": 0.9071, + "step": 137 + }, + { + "epoch": 0.06644998194293969, + "grad_norm": 0.6599445343017578, + "learning_rate": 0.00019933452463844376, + "loss": 0.9648, + "step": 138 + }, + { + "epoch": 0.06693150355122186, + "grad_norm": 1.112597942352295, + "learning_rate": 0.00019931643794908772, + "loss": 0.7063, + "step": 139 + }, + { + "epoch": 0.06741302515950404, + "grad_norm": 0.9359776973724365, + "learning_rate": 0.00019929810960135172, + "loss": 0.9689, + "step": 140 + }, + { + "epoch": 0.06789454676778621, + "grad_norm": 1.7165433168411255, + "learning_rate": 0.00019927953963983254, + "loss": 1.0222, + "step": 141 + }, + { + "epoch": 0.06837606837606838, + "grad_norm": 1.3653531074523926, + "learning_rate": 0.00019926072810971492, + "loss": 1.159, + "step": 142 + }, + { + "epoch": 0.06885758998435054, + "grad_norm": 1.2677444219589233, + "learning_rate": 0.00019924167505677137, + "loss": 1.4072, + "step": 143 + }, + { + "epoch": 0.06933911159263272, + "grad_norm": 2.713906764984131, + "learning_rate": 0.00019922238052736215, + "loss": 0.8661, + "step": 144 + }, + { + "epoch": 0.06982063320091489, + "grad_norm": 1.47153902053833, + "learning_rate": 0.00019920284456843498, + "loss": 1.2206, + "step": 145 + }, + { + "epoch": 0.07030215480919706, + "grad_norm": 1.4640045166015625, + "learning_rate": 0.00019918306722752505, + "loss": 1.3151, + "step": 146 + }, + { + "epoch": 0.07078367641747924, + "grad_norm": 1.0816256999969482, + "learning_rate": 0.00019916304855275497, + "loss": 0.8234, + "step": 147 + }, + { + "epoch": 0.07126519802576141, + "grad_norm": 1.3785395622253418, + "learning_rate": 0.00019914278859283445, + "loss": 0.9928, + "step": 148 + }, + { + "epoch": 0.07174671963404358, + "grad_norm": 1.7482737302780151, + "learning_rate": 0.0001991222873970604, + "loss": 0.3996, + "step": 149 + }, + { + "epoch": 0.07222824124232576, + "grad_norm": 0.5450143814086914, + "learning_rate": 0.00019910154501531663, + "loss": 0.6338, + "step": 150 + }, + { + "epoch": 0.07270976285060791, + "grad_norm": 0.9865407943725586, + "learning_rate": 0.0001990805614980739, + "loss": 0.803, + "step": 151 + }, + { + "epoch": 0.07319128445889009, + "grad_norm": 0.5290705561637878, + "learning_rate": 0.00019905933689638955, + "loss": 0.8526, + "step": 152 + }, + { + "epoch": 0.07367280606717226, + "grad_norm": 0.8891838788986206, + "learning_rate": 0.00019903787126190772, + "loss": 0.7215, + "step": 153 + }, + { + "epoch": 0.07415432767545443, + "grad_norm": 1.012518048286438, + "learning_rate": 0.00019901616464685888, + "loss": 0.8488, + "step": 154 + }, + { + "epoch": 0.07463584928373661, + "grad_norm": 1.3221955299377441, + "learning_rate": 0.00019899421710405996, + "loss": 1.1629, + "step": 155 + }, + { + "epoch": 0.07511737089201878, + "grad_norm": 1.0022908449172974, + "learning_rate": 0.00019897202868691407, + "loss": 0.9271, + "step": 156 + }, + { + "epoch": 0.07559889250030095, + "grad_norm": 1.7466130256652832, + "learning_rate": 0.00019894959944941038, + "loss": 0.703, + "step": 157 + }, + { + "epoch": 0.07608041410858313, + "grad_norm": 0.8413887619972229, + "learning_rate": 0.0001989269294461242, + "loss": 0.8749, + "step": 158 + }, + { + "epoch": 0.0765619357168653, + "grad_norm": 0.904451310634613, + "learning_rate": 0.0001989040187322164, + "loss": 0.9528, + "step": 159 + }, + { + "epoch": 0.07704345732514746, + "grad_norm": 0.7697594165802002, + "learning_rate": 0.00019888086736343384, + "loss": 0.9407, + "step": 160 + }, + { + "epoch": 0.07752497893342963, + "grad_norm": 1.4493168592453003, + "learning_rate": 0.0001988574753961087, + "loss": 0.7736, + "step": 161 + }, + { + "epoch": 0.0780065005417118, + "grad_norm": 0.8093467354774475, + "learning_rate": 0.00019883384288715874, + "loss": 0.7555, + "step": 162 + }, + { + "epoch": 0.07848802214999398, + "grad_norm": 1.269743800163269, + "learning_rate": 0.000198809969894087, + "loss": 1.3101, + "step": 163 + }, + { + "epoch": 0.07896954375827615, + "grad_norm": 0.961480438709259, + "learning_rate": 0.0001987858564749816, + "loss": 1.049, + "step": 164 + }, + { + "epoch": 0.07945106536655833, + "grad_norm": 0.8676347136497498, + "learning_rate": 0.00019876150268851572, + "loss": 1.0772, + "step": 165 + }, + { + "epoch": 0.0799325869748405, + "grad_norm": 1.016195297241211, + "learning_rate": 0.00019873690859394737, + "loss": 0.971, + "step": 166 + }, + { + "epoch": 0.08041410858312267, + "grad_norm": 1.0543973445892334, + "learning_rate": 0.0001987120742511193, + "loss": 0.8215, + "step": 167 + }, + { + "epoch": 0.08089563019140483, + "grad_norm": 0.9103195071220398, + "learning_rate": 0.0001986869997204589, + "loss": 0.4004, + "step": 168 + }, + { + "epoch": 0.081377151799687, + "grad_norm": 1.200688362121582, + "learning_rate": 0.00019866168506297788, + "loss": 0.855, + "step": 169 + }, + { + "epoch": 0.08185867340796918, + "grad_norm": 0.5053849816322327, + "learning_rate": 0.00019863613034027224, + "loss": 0.7865, + "step": 170 + }, + { + "epoch": 0.08234019501625135, + "grad_norm": 2.139782190322876, + "learning_rate": 0.00019861033561452223, + "loss": 0.7017, + "step": 171 + }, + { + "epoch": 0.08282171662453353, + "grad_norm": 1.6163843870162964, + "learning_rate": 0.00019858430094849195, + "loss": 0.9894, + "step": 172 + }, + { + "epoch": 0.0833032382328157, + "grad_norm": 0.6523648500442505, + "learning_rate": 0.0001985580264055294, + "loss": 0.5497, + "step": 173 + }, + { + "epoch": 0.08378475984109787, + "grad_norm": 0.8324865102767944, + "learning_rate": 0.00019853151204956616, + "loss": 0.7934, + "step": 174 + }, + { + "epoch": 0.08426628144938005, + "grad_norm": 1.4201933145523071, + "learning_rate": 0.00019850475794511749, + "loss": 0.3068, + "step": 175 + }, + { + "epoch": 0.08474780305766222, + "grad_norm": 0.9500930309295654, + "learning_rate": 0.00019847776415728185, + "loss": 0.6451, + "step": 176 + }, + { + "epoch": 0.08522932466594438, + "grad_norm": 0.7073138356208801, + "learning_rate": 0.000198450530751741, + "loss": 1.1907, + "step": 177 + }, + { + "epoch": 0.08571084627422655, + "grad_norm": 0.6987308859825134, + "learning_rate": 0.00019842305779475968, + "loss": 0.5881, + "step": 178 + }, + { + "epoch": 0.08619236788250872, + "grad_norm": 0.8138691186904907, + "learning_rate": 0.00019839534535318558, + "loss": 1.0889, + "step": 179 + }, + { + "epoch": 0.0866738894907909, + "grad_norm": 0.7772369980812073, + "learning_rate": 0.00019836739349444899, + "loss": 0.7811, + "step": 180 + }, + { + "epoch": 0.08715541109907307, + "grad_norm": 0.8350958228111267, + "learning_rate": 0.00019833920228656292, + "loss": 1.0841, + "step": 181 + }, + { + "epoch": 0.08763693270735524, + "grad_norm": 1.1595239639282227, + "learning_rate": 0.0001983107717981226, + "loss": 1.0653, + "step": 182 + }, + { + "epoch": 0.08811845431563742, + "grad_norm": 0.6263972520828247, + "learning_rate": 0.00019828210209830562, + "loss": 0.5907, + "step": 183 + }, + { + "epoch": 0.08859997592391959, + "grad_norm": 0.4528619349002838, + "learning_rate": 0.00019825319325687154, + "loss": 0.6648, + "step": 184 + }, + { + "epoch": 0.08908149753220176, + "grad_norm": 1.5184047222137451, + "learning_rate": 0.00019822404534416182, + "loss": 0.7916, + "step": 185 + }, + { + "epoch": 0.08956301914048392, + "grad_norm": 0.7816428542137146, + "learning_rate": 0.00019819465843109963, + "loss": 0.9215, + "step": 186 + }, + { + "epoch": 0.0900445407487661, + "grad_norm": 0.7619443535804749, + "learning_rate": 0.00019816503258918969, + "loss": 0.447, + "step": 187 + }, + { + "epoch": 0.09052606235704827, + "grad_norm": 1.0862969160079956, + "learning_rate": 0.00019813516789051808, + "loss": 0.6545, + "step": 188 + }, + { + "epoch": 0.09100758396533044, + "grad_norm": 1.1096842288970947, + "learning_rate": 0.0001981050644077521, + "loss": 0.9364, + "step": 189 + }, + { + "epoch": 0.09148910557361262, + "grad_norm": 1.102784514427185, + "learning_rate": 0.00019807472221414002, + "loss": 0.5648, + "step": 190 + }, + { + "epoch": 0.09197062718189479, + "grad_norm": 0.9051916599273682, + "learning_rate": 0.00019804414138351094, + "loss": 1.0847, + "step": 191 + }, + { + "epoch": 0.09245214879017696, + "grad_norm": 1.1457304954528809, + "learning_rate": 0.00019801332199027467, + "loss": 1.2033, + "step": 192 + }, + { + "epoch": 0.09293367039845914, + "grad_norm": 0.9964216351509094, + "learning_rate": 0.00019798226410942146, + "loss": 0.8802, + "step": 193 + }, + { + "epoch": 0.0934151920067413, + "grad_norm": 1.1561510562896729, + "learning_rate": 0.00019795096781652182, + "loss": 1.1658, + "step": 194 + }, + { + "epoch": 0.09389671361502347, + "grad_norm": 1.1975256204605103, + "learning_rate": 0.00019791943318772643, + "loss": 1.4838, + "step": 195 + }, + { + "epoch": 0.09437823522330564, + "grad_norm": 0.8118493556976318, + "learning_rate": 0.00019788766029976587, + "loss": 0.8335, + "step": 196 + }, + { + "epoch": 0.09485975683158782, + "grad_norm": 0.6374335885047913, + "learning_rate": 0.0001978556492299504, + "loss": 0.6217, + "step": 197 + }, + { + "epoch": 0.09534127843986999, + "grad_norm": 0.48891177773475647, + "learning_rate": 0.00019782340005616996, + "loss": 0.5794, + "step": 198 + }, + { + "epoch": 0.09582280004815216, + "grad_norm": 0.6944652199745178, + "learning_rate": 0.0001977909128568937, + "loss": 0.6489, + "step": 199 + }, + { + "epoch": 0.09630432165643434, + "grad_norm": 1.1781611442565918, + "learning_rate": 0.00019775818771117, + "loss": 0.7355, + "step": 200 + }, + { + "epoch": 0.09678584326471651, + "grad_norm": 0.6658651232719421, + "learning_rate": 0.00019772522469862626, + "loss": 0.6897, + "step": 201 + }, + { + "epoch": 0.09726736487299868, + "grad_norm": 0.9195631146430969, + "learning_rate": 0.00019769202389946863, + "loss": 0.8157, + "step": 202 + }, + { + "epoch": 0.09774888648128084, + "grad_norm": 0.7591468095779419, + "learning_rate": 0.0001976585853944818, + "loss": 0.8502, + "step": 203 + }, + { + "epoch": 0.09823040808956301, + "grad_norm": 0.5556705594062805, + "learning_rate": 0.0001976249092650289, + "loss": 0.6078, + "step": 204 + }, + { + "epoch": 0.09871192969784519, + "grad_norm": 0.5372620820999146, + "learning_rate": 0.00019759099559305124, + "loss": 0.6814, + "step": 205 + }, + { + "epoch": 0.09919345130612736, + "grad_norm": 0.6747313141822815, + "learning_rate": 0.00019755684446106812, + "loss": 0.693, + "step": 206 + }, + { + "epoch": 0.09967497291440953, + "grad_norm": 0.9310837388038635, + "learning_rate": 0.00019752245595217662, + "loss": 1.0744, + "step": 207 + }, + { + "epoch": 0.10015649452269171, + "grad_norm": 0.7573024034500122, + "learning_rate": 0.00019748783015005144, + "loss": 0.7726, + "step": 208 + }, + { + "epoch": 0.10063801613097388, + "grad_norm": 0.8931111693382263, + "learning_rate": 0.00019745296713894465, + "loss": 0.542, + "step": 209 + }, + { + "epoch": 0.10111953773925605, + "grad_norm": 0.8912571668624878, + "learning_rate": 0.00019741786700368548, + "loss": 0.4853, + "step": 210 + }, + { + "epoch": 0.10160105934753821, + "grad_norm": 0.8952215313911438, + "learning_rate": 0.00019738252982968017, + "loss": 0.6625, + "step": 211 + }, + { + "epoch": 0.10208258095582039, + "grad_norm": 0.5370394587516785, + "learning_rate": 0.00019734695570291168, + "loss": 0.6617, + "step": 212 + }, + { + "epoch": 0.10256410256410256, + "grad_norm": 0.8716063499450684, + "learning_rate": 0.00019731114470993962, + "loss": 0.7222, + "step": 213 + }, + { + "epoch": 0.10304562417238473, + "grad_norm": 1.1294581890106201, + "learning_rate": 0.0001972750969378998, + "loss": 0.8233, + "step": 214 + }, + { + "epoch": 0.10352714578066691, + "grad_norm": 0.9793927669525146, + "learning_rate": 0.00019723881247450434, + "loss": 0.7701, + "step": 215 + }, + { + "epoch": 0.10400866738894908, + "grad_norm": 1.0255509614944458, + "learning_rate": 0.0001972022914080411, + "loss": 1.3642, + "step": 216 + }, + { + "epoch": 0.10449018899723125, + "grad_norm": 0.6845418214797974, + "learning_rate": 0.00019716553382737379, + "loss": 0.3581, + "step": 217 + }, + { + "epoch": 0.10497171060551343, + "grad_norm": 1.35993492603302, + "learning_rate": 0.00019712853982194152, + "loss": 0.996, + "step": 218 + }, + { + "epoch": 0.1054532322137956, + "grad_norm": 0.62435382604599, + "learning_rate": 0.00019709130948175876, + "loss": 0.6824, + "step": 219 + }, + { + "epoch": 0.10593475382207776, + "grad_norm": 0.5951329469680786, + "learning_rate": 0.0001970538428974149, + "loss": 0.8215, + "step": 220 + }, + { + "epoch": 0.10641627543035993, + "grad_norm": 0.7611098885536194, + "learning_rate": 0.00019701614016007436, + "loss": 0.9625, + "step": 221 + }, + { + "epoch": 0.1068977970386421, + "grad_norm": 1.2601478099822998, + "learning_rate": 0.00019697820136147597, + "loss": 0.8209, + "step": 222 + }, + { + "epoch": 0.10737931864692428, + "grad_norm": 1.0291566848754883, + "learning_rate": 0.00019694002659393305, + "loss": 1.0147, + "step": 223 + }, + { + "epoch": 0.10786084025520645, + "grad_norm": 0.6460703015327454, + "learning_rate": 0.0001969016159503331, + "loss": 1.0697, + "step": 224 + }, + { + "epoch": 0.10834236186348863, + "grad_norm": 0.6381105780601501, + "learning_rate": 0.00019686296952413747, + "loss": 0.4803, + "step": 225 + }, + { + "epoch": 0.1088238834717708, + "grad_norm": 1.399173378944397, + "learning_rate": 0.0001968240874093813, + "loss": 0.654, + "step": 226 + }, + { + "epoch": 0.10930540508005297, + "grad_norm": 1.1444861888885498, + "learning_rate": 0.00019678496970067325, + "loss": 0.4746, + "step": 227 + }, + { + "epoch": 0.10978692668833515, + "grad_norm": 0.8985688090324402, + "learning_rate": 0.0001967456164931951, + "loss": 0.4273, + "step": 228 + }, + { + "epoch": 0.1102684482966173, + "grad_norm": 0.747041642665863, + "learning_rate": 0.0001967060278827017, + "loss": 0.4249, + "step": 229 + }, + { + "epoch": 0.11074996990489948, + "grad_norm": 1.1315629482269287, + "learning_rate": 0.00019666620396552076, + "loss": 0.7375, + "step": 230 + }, + { + "epoch": 0.11123149151318165, + "grad_norm": 1.8923050165176392, + "learning_rate": 0.00019662614483855246, + "loss": 0.9888, + "step": 231 + }, + { + "epoch": 0.11171301312146383, + "grad_norm": 2.3420629501342773, + "learning_rate": 0.00019658585059926934, + "loss": 0.5667, + "step": 232 + }, + { + "epoch": 0.112194534729746, + "grad_norm": 0.8595489263534546, + "learning_rate": 0.00019654532134571594, + "loss": 0.8782, + "step": 233 + }, + { + "epoch": 0.11267605633802817, + "grad_norm": 0.9815763235092163, + "learning_rate": 0.00019650455717650878, + "loss": 0.5611, + "step": 234 + }, + { + "epoch": 0.11315757794631034, + "grad_norm": 0.5129497647285461, + "learning_rate": 0.00019646355819083589, + "loss": 0.5094, + "step": 235 + }, + { + "epoch": 0.11363909955459252, + "grad_norm": 1.0263252258300781, + "learning_rate": 0.0001964223244884566, + "loss": 0.634, + "step": 236 + }, + { + "epoch": 0.11412062116287468, + "grad_norm": 1.8378629684448242, + "learning_rate": 0.00019638085616970153, + "loss": 0.7657, + "step": 237 + }, + { + "epoch": 0.11460214277115685, + "grad_norm": 1.0862857103347778, + "learning_rate": 0.00019633915333547202, + "loss": 0.8037, + "step": 238 + }, + { + "epoch": 0.11508366437943902, + "grad_norm": 0.7772231101989746, + "learning_rate": 0.00019629721608724004, + "loss": 0.6409, + "step": 239 + }, + { + "epoch": 0.1155651859877212, + "grad_norm": 0.6412681341171265, + "learning_rate": 0.0001962550445270481, + "loss": 0.4673, + "step": 240 + }, + { + "epoch": 0.11604670759600337, + "grad_norm": 1.4209880828857422, + "learning_rate": 0.00019621263875750864, + "loss": 0.5923, + "step": 241 + }, + { + "epoch": 0.11652822920428554, + "grad_norm": 0.7568668127059937, + "learning_rate": 0.00019616999888180406, + "loss": 0.3972, + "step": 242 + }, + { + "epoch": 0.11700975081256772, + "grad_norm": 1.2214607000350952, + "learning_rate": 0.0001961271250036865, + "loss": 0.8049, + "step": 243 + }, + { + "epoch": 0.11749127242084989, + "grad_norm": 0.8679167032241821, + "learning_rate": 0.0001960840172274773, + "loss": 0.5002, + "step": 244 + }, + { + "epoch": 0.11797279402913206, + "grad_norm": 0.5739406943321228, + "learning_rate": 0.00019604067565806704, + "loss": 0.4761, + "step": 245 + }, + { + "epoch": 0.11845431563741422, + "grad_norm": 0.822028636932373, + "learning_rate": 0.00019599710040091512, + "loss": 0.6141, + "step": 246 + }, + { + "epoch": 0.1189358372456964, + "grad_norm": 0.6882205009460449, + "learning_rate": 0.00019595329156204955, + "loss": 0.174, + "step": 247 + }, + { + "epoch": 0.11941735885397857, + "grad_norm": 1.7359886169433594, + "learning_rate": 0.00019590924924806676, + "loss": 0.777, + "step": 248 + }, + { + "epoch": 0.11989888046226074, + "grad_norm": 0.9156647324562073, + "learning_rate": 0.0001958649735661312, + "loss": 0.9562, + "step": 249 + }, + { + "epoch": 0.12038040207054292, + "grad_norm": 0.8275269865989685, + "learning_rate": 0.00019582046462397515, + "loss": 0.6719, + "step": 250 + }, + { + "epoch": 0.12086192367882509, + "grad_norm": 1.1353651285171509, + "learning_rate": 0.00019577572252989854, + "loss": 1.0057, + "step": 251 + }, + { + "epoch": 0.12134344528710726, + "grad_norm": 1.2231826782226562, + "learning_rate": 0.00019573074739276858, + "loss": 1.0401, + "step": 252 + }, + { + "epoch": 0.12182496689538944, + "grad_norm": 1.7041336297988892, + "learning_rate": 0.00019568553932201947, + "loss": 0.631, + "step": 253 + }, + { + "epoch": 0.1223064885036716, + "grad_norm": 0.949766218662262, + "learning_rate": 0.00019564009842765225, + "loss": 1.0427, + "step": 254 + }, + { + "epoch": 0.12278801011195377, + "grad_norm": 0.9756664633750916, + "learning_rate": 0.00019559442482023444, + "loss": 1.0868, + "step": 255 + }, + { + "epoch": 0.12326953172023594, + "grad_norm": 2.625432014465332, + "learning_rate": 0.0001955485186108998, + "loss": 0.9394, + "step": 256 + }, + { + "epoch": 0.12375105332851812, + "grad_norm": 0.9411224722862244, + "learning_rate": 0.00019550237991134805, + "loss": 0.5891, + "step": 257 + }, + { + "epoch": 0.12423257493680029, + "grad_norm": 0.8636229634284973, + "learning_rate": 0.00019545600883384467, + "loss": 0.9391, + "step": 258 + }, + { + "epoch": 0.12471409654508246, + "grad_norm": 0.5954817533493042, + "learning_rate": 0.0001954094054912205, + "loss": 0.7277, + "step": 259 + }, + { + "epoch": 0.12519561815336464, + "grad_norm": 0.6612427234649658, + "learning_rate": 0.00019536256999687157, + "loss": 0.4814, + "step": 260 + }, + { + "epoch": 0.1256771397616468, + "grad_norm": 0.7988054752349854, + "learning_rate": 0.00019531550246475876, + "loss": 0.5545, + "step": 261 + }, + { + "epoch": 0.12615866136992898, + "grad_norm": 1.6089789867401123, + "learning_rate": 0.00019526820300940756, + "loss": 1.0021, + "step": 262 + }, + { + "epoch": 0.12664018297821114, + "grad_norm": 1.084684133529663, + "learning_rate": 0.00019522067174590778, + "loss": 1.0537, + "step": 263 + }, + { + "epoch": 0.12712170458649333, + "grad_norm": 1.3705826997756958, + "learning_rate": 0.00019517290878991324, + "loss": 0.7344, + "step": 264 + }, + { + "epoch": 0.1276032261947755, + "grad_norm": 1.0224491357803345, + "learning_rate": 0.0001951249142576416, + "loss": 1.4591, + "step": 265 + }, + { + "epoch": 0.12808474780305767, + "grad_norm": 1.1117572784423828, + "learning_rate": 0.00019507668826587387, + "loss": 0.9157, + "step": 266 + }, + { + "epoch": 0.12856626941133983, + "grad_norm": 1.0106638669967651, + "learning_rate": 0.0001950282309319544, + "loss": 0.998, + "step": 267 + }, + { + "epoch": 0.129047791019622, + "grad_norm": 1.1685928106307983, + "learning_rate": 0.0001949795423737903, + "loss": 1.0733, + "step": 268 + }, + { + "epoch": 0.12952931262790418, + "grad_norm": 0.8869622349739075, + "learning_rate": 0.00019493062270985144, + "loss": 0.9227, + "step": 269 + }, + { + "epoch": 0.13001083423618634, + "grad_norm": 0.8877792358398438, + "learning_rate": 0.00019488147205916985, + "loss": 0.7803, + "step": 270 + }, + { + "epoch": 0.13049235584446853, + "grad_norm": 0.8910313844680786, + "learning_rate": 0.00019483209054133976, + "loss": 0.7102, + "step": 271 + }, + { + "epoch": 0.1309738774527507, + "grad_norm": 1.707604169845581, + "learning_rate": 0.00019478247827651708, + "loss": 0.6851, + "step": 272 + }, + { + "epoch": 0.13145539906103287, + "grad_norm": 1.1673096418380737, + "learning_rate": 0.00019473263538541914, + "loss": 0.5218, + "step": 273 + }, + { + "epoch": 0.13193692066931503, + "grad_norm": 1.01649808883667, + "learning_rate": 0.00019468256198932455, + "loss": 0.6644, + "step": 274 + }, + { + "epoch": 0.13241844227759722, + "grad_norm": 0.7683111429214478, + "learning_rate": 0.00019463225821007268, + "loss": 0.979, + "step": 275 + }, + { + "epoch": 0.13289996388587938, + "grad_norm": 0.6378679871559143, + "learning_rate": 0.00019458172417006347, + "loss": 0.6165, + "step": 276 + }, + { + "epoch": 0.13338148549416154, + "grad_norm": 0.6471127271652222, + "learning_rate": 0.00019453095999225726, + "loss": 0.5204, + "step": 277 + }, + { + "epoch": 0.13386300710244373, + "grad_norm": 0.7345598936080933, + "learning_rate": 0.0001944799658001742, + "loss": 0.4905, + "step": 278 + }, + { + "epoch": 0.13434452871072589, + "grad_norm": 0.670501172542572, + "learning_rate": 0.00019442874171789418, + "loss": 0.4426, + "step": 279 + }, + { + "epoch": 0.13482605031900807, + "grad_norm": 0.7248306274414062, + "learning_rate": 0.00019437728787005657, + "loss": 1.3919, + "step": 280 + }, + { + "epoch": 0.13530757192729023, + "grad_norm": 1.8311445713043213, + "learning_rate": 0.00019432560438185963, + "loss": 0.5238, + "step": 281 + }, + { + "epoch": 0.13578909353557242, + "grad_norm": 0.7017165422439575, + "learning_rate": 0.00019427369137906046, + "loss": 0.8092, + "step": 282 + }, + { + "epoch": 0.13627061514385458, + "grad_norm": 1.079423427581787, + "learning_rate": 0.00019422154898797472, + "loss": 0.6991, + "step": 283 + }, + { + "epoch": 0.13675213675213677, + "grad_norm": 0.9701410531997681, + "learning_rate": 0.00019416917733547603, + "loss": 1.4933, + "step": 284 + }, + { + "epoch": 0.13723365836041893, + "grad_norm": 1.552991509437561, + "learning_rate": 0.00019411657654899597, + "loss": 0.7102, + "step": 285 + }, + { + "epoch": 0.13771517996870108, + "grad_norm": 0.9199215769767761, + "learning_rate": 0.0001940637467565237, + "loss": 0.7313, + "step": 286 + }, + { + "epoch": 0.13819670157698327, + "grad_norm": 1.1101661920547485, + "learning_rate": 0.00019401068808660546, + "loss": 0.8939, + "step": 287 + }, + { + "epoch": 0.13867822318526543, + "grad_norm": 0.817594051361084, + "learning_rate": 0.0001939574006683445, + "loss": 0.7946, + "step": 288 + }, + { + "epoch": 0.13915974479354762, + "grad_norm": 0.8862625360488892, + "learning_rate": 0.00019390388463140065, + "loss": 0.4525, + "step": 289 + }, + { + "epoch": 0.13964126640182978, + "grad_norm": 0.799997091293335, + "learning_rate": 0.00019385014010598998, + "loss": 0.4477, + "step": 290 + }, + { + "epoch": 0.14012278801011196, + "grad_norm": 0.8338310718536377, + "learning_rate": 0.00019379616722288456, + "loss": 0.787, + "step": 291 + }, + { + "epoch": 0.14060430961839412, + "grad_norm": 0.9072525501251221, + "learning_rate": 0.0001937419661134121, + "loss": 0.8017, + "step": 292 + }, + { + "epoch": 0.14108583122667628, + "grad_norm": 0.46390998363494873, + "learning_rate": 0.0001936875369094556, + "loss": 0.6444, + "step": 293 + }, + { + "epoch": 0.14156735283495847, + "grad_norm": 0.9819871187210083, + "learning_rate": 0.0001936328797434531, + "loss": 0.533, + "step": 294 + }, + { + "epoch": 0.14204887444324063, + "grad_norm": 0.9907609224319458, + "learning_rate": 0.00019357799474839735, + "loss": 1.1361, + "step": 295 + }, + { + "epoch": 0.14253039605152282, + "grad_norm": 1.1553535461425781, + "learning_rate": 0.00019352288205783536, + "loss": 0.5737, + "step": 296 + }, + { + "epoch": 0.14301191765980498, + "grad_norm": 0.8127950429916382, + "learning_rate": 0.00019346754180586825, + "loss": 0.4669, + "step": 297 + }, + { + "epoch": 0.14349343926808716, + "grad_norm": 0.869855523109436, + "learning_rate": 0.00019341197412715082, + "loss": 0.9356, + "step": 298 + }, + { + "epoch": 0.14397496087636932, + "grad_norm": 0.6495091319084167, + "learning_rate": 0.00019335617915689128, + "loss": 0.6174, + "step": 299 + }, + { + "epoch": 0.1444564824846515, + "grad_norm": 0.7985929846763611, + "learning_rate": 0.00019330015703085082, + "loss": 0.5252, + "step": 300 + }, + { + "epoch": 0.14493800409293367, + "grad_norm": 0.5880833864212036, + "learning_rate": 0.00019324390788534343, + "loss": 0.3703, + "step": 301 + }, + { + "epoch": 0.14541952570121583, + "grad_norm": 1.1846781969070435, + "learning_rate": 0.00019318743185723546, + "loss": 0.6423, + "step": 302 + }, + { + "epoch": 0.14590104730949802, + "grad_norm": 1.47828209400177, + "learning_rate": 0.00019313072908394525, + "loss": 1.1003, + "step": 303 + }, + { + "epoch": 0.14638256891778018, + "grad_norm": 1.1608823537826538, + "learning_rate": 0.00019307379970344294, + "loss": 0.8453, + "step": 304 + }, + { + "epoch": 0.14686409052606236, + "grad_norm": 0.9130098819732666, + "learning_rate": 0.00019301664385425004, + "loss": 0.6183, + "step": 305 + }, + { + "epoch": 0.14734561213434452, + "grad_norm": 0.8722681999206543, + "learning_rate": 0.0001929592616754391, + "loss": 0.6744, + "step": 306 + }, + { + "epoch": 0.1478271337426267, + "grad_norm": 0.7916250228881836, + "learning_rate": 0.00019290165330663336, + "loss": 0.7351, + "step": 307 + }, + { + "epoch": 0.14830865535090887, + "grad_norm": 0.604591965675354, + "learning_rate": 0.00019284381888800647, + "loss": 0.7715, + "step": 308 + }, + { + "epoch": 0.14879017695919106, + "grad_norm": 0.7270418405532837, + "learning_rate": 0.00019278575856028206, + "loss": 0.94, + "step": 309 + }, + { + "epoch": 0.14927169856747322, + "grad_norm": 0.7614173293113708, + "learning_rate": 0.00019272747246473345, + "loss": 0.4351, + "step": 310 + }, + { + "epoch": 0.14975322017575537, + "grad_norm": 0.7510772347450256, + "learning_rate": 0.00019266896074318334, + "loss": 0.6094, + "step": 311 + }, + { + "epoch": 0.15023474178403756, + "grad_norm": 0.6468477845191956, + "learning_rate": 0.00019261022353800344, + "loss": 0.61, + "step": 312 + }, + { + "epoch": 0.15071626339231972, + "grad_norm": 1.057340145111084, + "learning_rate": 0.00019255126099211402, + "loss": 0.7782, + "step": 313 + }, + { + "epoch": 0.1511977850006019, + "grad_norm": 0.5399413704872131, + "learning_rate": 0.00019249207324898376, + "loss": 0.6912, + "step": 314 + }, + { + "epoch": 0.15167930660888407, + "grad_norm": 1.0437463521957397, + "learning_rate": 0.0001924326604526292, + "loss": 0.7274, + "step": 315 + }, + { + "epoch": 0.15216082821716626, + "grad_norm": 0.7240994572639465, + "learning_rate": 0.00019237302274761458, + "loss": 0.505, + "step": 316 + }, + { + "epoch": 0.15264234982544841, + "grad_norm": 1.8743011951446533, + "learning_rate": 0.0001923131602790513, + "loss": 0.9457, + "step": 317 + }, + { + "epoch": 0.1531238714337306, + "grad_norm": 1.084415078163147, + "learning_rate": 0.00019225307319259768, + "loss": 1.2172, + "step": 318 + }, + { + "epoch": 0.15360539304201276, + "grad_norm": 0.8688251972198486, + "learning_rate": 0.00019219276163445862, + "loss": 0.6531, + "step": 319 + }, + { + "epoch": 0.15408691465029492, + "grad_norm": 0.9400373101234436, + "learning_rate": 0.00019213222575138522, + "loss": 1.4234, + "step": 320 + }, + { + "epoch": 0.1545684362585771, + "grad_norm": 1.1395434141159058, + "learning_rate": 0.00019207146569067435, + "loss": 0.45, + "step": 321 + }, + { + "epoch": 0.15504995786685927, + "grad_norm": 1.3613590002059937, + "learning_rate": 0.00019201048160016838, + "loss": 0.9647, + "step": 322 + }, + { + "epoch": 0.15553147947514145, + "grad_norm": 0.7267064452171326, + "learning_rate": 0.00019194927362825478, + "loss": 0.9339, + "step": 323 + }, + { + "epoch": 0.1560130010834236, + "grad_norm": 1.1009801626205444, + "learning_rate": 0.00019188784192386587, + "loss": 0.9611, + "step": 324 + }, + { + "epoch": 0.1564945226917058, + "grad_norm": 0.6520840525627136, + "learning_rate": 0.00019182618663647817, + "loss": 0.3283, + "step": 325 + }, + { + "epoch": 0.15697604429998796, + "grad_norm": 1.0760483741760254, + "learning_rate": 0.0001917643079161124, + "loss": 0.6973, + "step": 326 + }, + { + "epoch": 0.15745756590827015, + "grad_norm": 1.0936267375946045, + "learning_rate": 0.00019170220591333283, + "loss": 0.3996, + "step": 327 + }, + { + "epoch": 0.1579390875165523, + "grad_norm": 1.3938161134719849, + "learning_rate": 0.00019163988077924713, + "loss": 1.0365, + "step": 328 + }, + { + "epoch": 0.15842060912483447, + "grad_norm": 0.8608887195587158, + "learning_rate": 0.00019157733266550575, + "loss": 0.6497, + "step": 329 + }, + { + "epoch": 0.15890213073311665, + "grad_norm": 0.8093037605285645, + "learning_rate": 0.00019151456172430183, + "loss": 0.4926, + "step": 330 + }, + { + "epoch": 0.1593836523413988, + "grad_norm": 1.156674861907959, + "learning_rate": 0.0001914515681083707, + "loss": 0.9478, + "step": 331 + }, + { + "epoch": 0.159865173949681, + "grad_norm": 0.639973521232605, + "learning_rate": 0.00019138835197098937, + "loss": 0.7327, + "step": 332 + }, + { + "epoch": 0.16034669555796316, + "grad_norm": 1.1579989194869995, + "learning_rate": 0.00019132491346597643, + "loss": 0.8106, + "step": 333 + }, + { + "epoch": 0.16082821716624535, + "grad_norm": 0.8499979376792908, + "learning_rate": 0.00019126125274769145, + "loss": 0.3414, + "step": 334 + }, + { + "epoch": 0.1613097387745275, + "grad_norm": 1.4322779178619385, + "learning_rate": 0.00019119736997103476, + "loss": 0.6139, + "step": 335 + }, + { + "epoch": 0.16179126038280967, + "grad_norm": 0.696955144405365, + "learning_rate": 0.000191133265291447, + "loss": 0.615, + "step": 336 + }, + { + "epoch": 0.16227278199109185, + "grad_norm": 0.7599828243255615, + "learning_rate": 0.00019106893886490864, + "loss": 0.7996, + "step": 337 + }, + { + "epoch": 0.162754303599374, + "grad_norm": 0.5892730951309204, + "learning_rate": 0.00019100439084793989, + "loss": 0.7155, + "step": 338 + }, + { + "epoch": 0.1632358252076562, + "grad_norm": 1.2862459421157837, + "learning_rate": 0.00019093962139759998, + "loss": 1.0595, + "step": 339 + }, + { + "epoch": 0.16371734681593836, + "grad_norm": 0.6903579235076904, + "learning_rate": 0.000190874630671487, + "loss": 0.6998, + "step": 340 + }, + { + "epoch": 0.16419886842422055, + "grad_norm": 0.6340907216072083, + "learning_rate": 0.00019080941882773745, + "loss": 0.8251, + "step": 341 + }, + { + "epoch": 0.1646803900325027, + "grad_norm": 0.7512362003326416, + "learning_rate": 0.00019074398602502584, + "loss": 0.7369, + "step": 342 + }, + { + "epoch": 0.1651619116407849, + "grad_norm": 1.1983510255813599, + "learning_rate": 0.00019067833242256442, + "loss": 0.8641, + "step": 343 + }, + { + "epoch": 0.16564343324906705, + "grad_norm": 0.7607635259628296, + "learning_rate": 0.0001906124581801025, + "loss": 0.406, + "step": 344 + }, + { + "epoch": 0.1661249548573492, + "grad_norm": 0.5438935160636902, + "learning_rate": 0.0001905463634579264, + "loss": 0.5055, + "step": 345 + }, + { + "epoch": 0.1666064764656314, + "grad_norm": 1.1291062831878662, + "learning_rate": 0.00019048004841685888, + "loss": 1.1312, + "step": 346 + }, + { + "epoch": 0.16708799807391356, + "grad_norm": 1.3860900402069092, + "learning_rate": 0.00019041351321825883, + "loss": 1.0954, + "step": 347 + }, + { + "epoch": 0.16756951968219574, + "grad_norm": 1.0058181285858154, + "learning_rate": 0.00019034675802402068, + "loss": 1.1283, + "step": 348 + }, + { + "epoch": 0.1680510412904779, + "grad_norm": 0.8692428469657898, + "learning_rate": 0.00019027978299657436, + "loss": 0.3789, + "step": 349 + }, + { + "epoch": 0.1685325628987601, + "grad_norm": 1.0767672061920166, + "learning_rate": 0.00019021258829888456, + "loss": 1.1003, + "step": 350 + }, + { + "epoch": 0.16901408450704225, + "grad_norm": 0.8247532248497009, + "learning_rate": 0.00019014517409445052, + "loss": 0.3231, + "step": 351 + }, + { + "epoch": 0.16949560611532444, + "grad_norm": 1.0215187072753906, + "learning_rate": 0.00019007754054730554, + "loss": 0.88, + "step": 352 + }, + { + "epoch": 0.1699771277236066, + "grad_norm": 0.8507590889930725, + "learning_rate": 0.00019000968782201675, + "loss": 0.37, + "step": 353 + }, + { + "epoch": 0.17045864933188876, + "grad_norm": 0.7044029831886292, + "learning_rate": 0.00018994161608368448, + "loss": 1.2317, + "step": 354 + }, + { + "epoch": 0.17094017094017094, + "grad_norm": 0.7668285369873047, + "learning_rate": 0.00018987332549794196, + "loss": 0.9292, + "step": 355 + }, + { + "epoch": 0.1714216925484531, + "grad_norm": 0.5195323824882507, + "learning_rate": 0.00018980481623095502, + "loss": 0.6176, + "step": 356 + }, + { + "epoch": 0.1719032141567353, + "grad_norm": 0.7647833824157715, + "learning_rate": 0.00018973608844942148, + "loss": 0.5043, + "step": 357 + }, + { + "epoch": 0.17238473576501745, + "grad_norm": 0.7875663638114929, + "learning_rate": 0.00018966714232057094, + "loss": 0.726, + "step": 358 + }, + { + "epoch": 0.17286625737329964, + "grad_norm": 0.6289336681365967, + "learning_rate": 0.00018959797801216418, + "loss": 0.2639, + "step": 359 + }, + { + "epoch": 0.1733477789815818, + "grad_norm": 1.9089537858963013, + "learning_rate": 0.000189528595692493, + "loss": 0.7418, + "step": 360 + }, + { + "epoch": 0.17382930058986398, + "grad_norm": 0.8167737722396851, + "learning_rate": 0.00018945899553037956, + "loss": 0.7737, + "step": 361 + }, + { + "epoch": 0.17431082219814614, + "grad_norm": 0.7543114423751831, + "learning_rate": 0.00018938917769517613, + "loss": 0.5564, + "step": 362 + }, + { + "epoch": 0.1747923438064283, + "grad_norm": 1.0456900596618652, + "learning_rate": 0.00018931914235676458, + "loss": 0.98, + "step": 363 + }, + { + "epoch": 0.1752738654147105, + "grad_norm": 0.49631309509277344, + "learning_rate": 0.00018924888968555606, + "loss": 0.6173, + "step": 364 + }, + { + "epoch": 0.17575538702299265, + "grad_norm": 0.449246883392334, + "learning_rate": 0.00018917841985249055, + "loss": 0.3526, + "step": 365 + }, + { + "epoch": 0.17623690863127484, + "grad_norm": 1.2787824869155884, + "learning_rate": 0.0001891077330290363, + "loss": 1.2029, + "step": 366 + }, + { + "epoch": 0.176718430239557, + "grad_norm": 1.1600645780563354, + "learning_rate": 0.00018903682938718977, + "loss": 1.1068, + "step": 367 + }, + { + "epoch": 0.17719995184783918, + "grad_norm": 0.5357136130332947, + "learning_rate": 0.00018896570909947475, + "loss": 0.5062, + "step": 368 + }, + { + "epoch": 0.17768147345612134, + "grad_norm": 0.872021496295929, + "learning_rate": 0.00018889437233894234, + "loss": 0.6622, + "step": 369 + }, + { + "epoch": 0.17816299506440353, + "grad_norm": 1.1199880838394165, + "learning_rate": 0.0001888228192791703, + "loss": 1.1557, + "step": 370 + }, + { + "epoch": 0.1786445166726857, + "grad_norm": 1.332701563835144, + "learning_rate": 0.00018875105009426272, + "loss": 0.6267, + "step": 371 + }, + { + "epoch": 0.17912603828096785, + "grad_norm": 1.0211436748504639, + "learning_rate": 0.00018867906495884955, + "loss": 1.06, + "step": 372 + }, + { + "epoch": 0.17960755988925003, + "grad_norm": 1.0804567337036133, + "learning_rate": 0.0001886068640480862, + "loss": 0.8207, + "step": 373 + }, + { + "epoch": 0.1800890814975322, + "grad_norm": 1.1362806558609009, + "learning_rate": 0.00018853444753765306, + "loss": 0.8978, + "step": 374 + }, + { + "epoch": 0.18057060310581438, + "grad_norm": 0.8540496826171875, + "learning_rate": 0.00018846181560375525, + "loss": 0.7343, + "step": 375 + }, + { + "epoch": 0.18105212471409654, + "grad_norm": 0.9461526274681091, + "learning_rate": 0.0001883889684231219, + "loss": 0.7977, + "step": 376 + }, + { + "epoch": 0.18153364632237873, + "grad_norm": 1.1165428161621094, + "learning_rate": 0.000188315906173006, + "loss": 0.8063, + "step": 377 + }, + { + "epoch": 0.1820151679306609, + "grad_norm": 0.5946139097213745, + "learning_rate": 0.0001882426290311838, + "loss": 0.4324, + "step": 378 + }, + { + "epoch": 0.18249668953894305, + "grad_norm": 1.2596697807312012, + "learning_rate": 0.00018816913717595445, + "loss": 1.0515, + "step": 379 + }, + { + "epoch": 0.18297821114722523, + "grad_norm": 0.8724560737609863, + "learning_rate": 0.00018809543078613953, + "loss": 0.4745, + "step": 380 + }, + { + "epoch": 0.1834597327555074, + "grad_norm": 1.4185947179794312, + "learning_rate": 0.00018802151004108263, + "loss": 0.3437, + "step": 381 + }, + { + "epoch": 0.18394125436378958, + "grad_norm": 1.0995190143585205, + "learning_rate": 0.0001879473751206489, + "loss": 1.0071, + "step": 382 + }, + { + "epoch": 0.18442277597207174, + "grad_norm": 0.5417447090148926, + "learning_rate": 0.00018787302620522467, + "loss": 0.3357, + "step": 383 + }, + { + "epoch": 0.18490429758035393, + "grad_norm": 0.9899541735649109, + "learning_rate": 0.00018779846347571693, + "loss": 0.1654, + "step": 384 + }, + { + "epoch": 0.1853858191886361, + "grad_norm": 0.5713137984275818, + "learning_rate": 0.0001877236871135529, + "loss": 0.3284, + "step": 385 + }, + { + "epoch": 0.18586734079691827, + "grad_norm": 1.2448984384536743, + "learning_rate": 0.00018764869730067968, + "loss": 1.2335, + "step": 386 + }, + { + "epoch": 0.18634886240520043, + "grad_norm": 0.8601797223091125, + "learning_rate": 0.0001875734942195637, + "loss": 0.4766, + "step": 387 + }, + { + "epoch": 0.1868303840134826, + "grad_norm": 0.8154586553573608, + "learning_rate": 0.0001874980780531903, + "loss": 0.7463, + "step": 388 + }, + { + "epoch": 0.18731190562176478, + "grad_norm": 0.9013132452964783, + "learning_rate": 0.00018742244898506337, + "loss": 0.6605, + "step": 389 + }, + { + "epoch": 0.18779342723004694, + "grad_norm": 1.5045801401138306, + "learning_rate": 0.00018734660719920475, + "loss": 1.1213, + "step": 390 + }, + { + "epoch": 0.18827494883832913, + "grad_norm": 1.017081379890442, + "learning_rate": 0.00018727055288015397, + "loss": 1.1526, + "step": 391 + }, + { + "epoch": 0.18875647044661129, + "grad_norm": 0.8374575972557068, + "learning_rate": 0.00018719428621296764, + "loss": 1.0737, + "step": 392 + }, + { + "epoch": 0.18923799205489347, + "grad_norm": 1.0993142127990723, + "learning_rate": 0.00018711780738321897, + "loss": 0.3277, + "step": 393 + }, + { + "epoch": 0.18971951366317563, + "grad_norm": 0.9795992970466614, + "learning_rate": 0.00018704111657699758, + "loss": 0.3837, + "step": 394 + }, + { + "epoch": 0.19020103527145782, + "grad_norm": 0.5359293818473816, + "learning_rate": 0.0001869642139809088, + "loss": 0.85, + "step": 395 + }, + { + "epoch": 0.19068255687973998, + "grad_norm": 0.6480598449707031, + "learning_rate": 0.00018688709978207323, + "loss": 0.7276, + "step": 396 + }, + { + "epoch": 0.19116407848802214, + "grad_norm": 0.9133846759796143, + "learning_rate": 0.00018680977416812644, + "loss": 1.0814, + "step": 397 + }, + { + "epoch": 0.19164560009630432, + "grad_norm": 0.8055482506752014, + "learning_rate": 0.00018673223732721837, + "loss": 0.5734, + "step": 398 + }, + { + "epoch": 0.19212712170458648, + "grad_norm": 0.83376145362854, + "learning_rate": 0.0001866544894480129, + "loss": 1.0693, + "step": 399 + }, + { + "epoch": 0.19260864331286867, + "grad_norm": 0.6174197196960449, + "learning_rate": 0.00018657653071968747, + "loss": 0.8912, + "step": 400 + }, + { + "epoch": 0.19309016492115083, + "grad_norm": 0.9090850353240967, + "learning_rate": 0.00018649836133193253, + "loss": 0.6061, + "step": 401 + }, + { + "epoch": 0.19357168652943302, + "grad_norm": 0.7542705535888672, + "learning_rate": 0.00018641998147495112, + "loss": 0.3132, + "step": 402 + }, + { + "epoch": 0.19405320813771518, + "grad_norm": 0.8012394309043884, + "learning_rate": 0.00018634139133945837, + "loss": 0.4904, + "step": 403 + }, + { + "epoch": 0.19453472974599736, + "grad_norm": 1.5305569171905518, + "learning_rate": 0.00018626259111668105, + "loss": 1.2355, + "step": 404 + }, + { + "epoch": 0.19501625135427952, + "grad_norm": 0.6317727565765381, + "learning_rate": 0.00018618358099835723, + "loss": 0.7488, + "step": 405 + }, + { + "epoch": 0.19549777296256168, + "grad_norm": 1.183491826057434, + "learning_rate": 0.00018610436117673555, + "loss": 0.484, + "step": 406 + }, + { + "epoch": 0.19597929457084387, + "grad_norm": 1.0415927171707153, + "learning_rate": 0.00018602493184457505, + "loss": 0.4665, + "step": 407 + }, + { + "epoch": 0.19646081617912603, + "grad_norm": 0.8337236046791077, + "learning_rate": 0.00018594529319514437, + "loss": 0.3315, + "step": 408 + }, + { + "epoch": 0.19694233778740822, + "grad_norm": 1.061383605003357, + "learning_rate": 0.00018586544542222169, + "loss": 1.101, + "step": 409 + }, + { + "epoch": 0.19742385939569038, + "grad_norm": 1.9225060939788818, + "learning_rate": 0.00018578538872009384, + "loss": 0.6274, + "step": 410 + }, + { + "epoch": 0.19790538100397256, + "grad_norm": 0.5955725312232971, + "learning_rate": 0.00018570512328355612, + "loss": 0.5071, + "step": 411 + }, + { + "epoch": 0.19838690261225472, + "grad_norm": 1.295961618423462, + "learning_rate": 0.00018562464930791167, + "loss": 0.9883, + "step": 412 + }, + { + "epoch": 0.1988684242205369, + "grad_norm": 1.068457007408142, + "learning_rate": 0.00018554396698897116, + "loss": 0.5021, + "step": 413 + }, + { + "epoch": 0.19934994582881907, + "grad_norm": 0.5233125686645508, + "learning_rate": 0.00018546307652305205, + "loss": 0.6233, + "step": 414 + }, + { + "epoch": 0.19983146743710123, + "grad_norm": 0.9043152928352356, + "learning_rate": 0.00018538197810697842, + "loss": 0.5477, + "step": 415 + }, + { + "epoch": 0.20031298904538342, + "grad_norm": 0.6947689056396484, + "learning_rate": 0.0001853006719380802, + "loss": 0.7851, + "step": 416 + }, + { + "epoch": 0.20079451065366558, + "grad_norm": 1.0613230466842651, + "learning_rate": 0.00018521915821419284, + "loss": 0.6935, + "step": 417 + }, + { + "epoch": 0.20127603226194776, + "grad_norm": 0.9804971218109131, + "learning_rate": 0.00018513743713365698, + "loss": 0.7748, + "step": 418 + }, + { + "epoch": 0.20175755387022992, + "grad_norm": 0.9837954640388489, + "learning_rate": 0.00018505550889531765, + "loss": 0.4582, + "step": 419 + }, + { + "epoch": 0.2022390754785121, + "grad_norm": 0.8446076512336731, + "learning_rate": 0.00018497337369852395, + "loss": 0.1956, + "step": 420 + }, + { + "epoch": 0.20272059708679427, + "grad_norm": 0.9201738238334656, + "learning_rate": 0.0001848910317431286, + "loss": 0.6636, + "step": 421 + }, + { + "epoch": 0.20320211869507643, + "grad_norm": 1.523264765739441, + "learning_rate": 0.00018480848322948739, + "loss": 0.7119, + "step": 422 + }, + { + "epoch": 0.20368364030335862, + "grad_norm": 1.1211214065551758, + "learning_rate": 0.00018472572835845873, + "loss": 0.8146, + "step": 423 + }, + { + "epoch": 0.20416516191164077, + "grad_norm": 0.6903687715530396, + "learning_rate": 0.00018464276733140306, + "loss": 0.6589, + "step": 424 + }, + { + "epoch": 0.20464668351992296, + "grad_norm": 1.2048624753952026, + "learning_rate": 0.0001845596003501826, + "loss": 0.7029, + "step": 425 + }, + { + "epoch": 0.20512820512820512, + "grad_norm": 0.6592060327529907, + "learning_rate": 0.00018447622761716057, + "loss": 0.492, + "step": 426 + }, + { + "epoch": 0.2056097267364873, + "grad_norm": 0.8045766353607178, + "learning_rate": 0.00018439264933520084, + "loss": 0.9407, + "step": 427 + }, + { + "epoch": 0.20609124834476947, + "grad_norm": 0.957320511341095, + "learning_rate": 0.00018430886570766747, + "loss": 0.9598, + "step": 428 + }, + { + "epoch": 0.20657276995305165, + "grad_norm": 0.9459810256958008, + "learning_rate": 0.0001842248769384242, + "loss": 0.8971, + "step": 429 + }, + { + "epoch": 0.20705429156133381, + "grad_norm": 1.4347730875015259, + "learning_rate": 0.00018414068323183375, + "loss": 0.737, + "step": 430 + }, + { + "epoch": 0.20753581316961597, + "grad_norm": 1.069244623184204, + "learning_rate": 0.00018405628479275775, + "loss": 0.5817, + "step": 431 + }, + { + "epoch": 0.20801733477789816, + "grad_norm": 0.9500924348831177, + "learning_rate": 0.00018397168182655583, + "loss": 0.4447, + "step": 432 + }, + { + "epoch": 0.20849885638618032, + "grad_norm": 1.1861311197280884, + "learning_rate": 0.00018388687453908527, + "loss": 0.4656, + "step": 433 + }, + { + "epoch": 0.2089803779944625, + "grad_norm": 1.3185081481933594, + "learning_rate": 0.00018380186313670058, + "loss": 0.9314, + "step": 434 + }, + { + "epoch": 0.20946189960274467, + "grad_norm": 1.0046930313110352, + "learning_rate": 0.00018371664782625287, + "loss": 0.3958, + "step": 435 + }, + { + "epoch": 0.20994342121102685, + "grad_norm": 1.3395354747772217, + "learning_rate": 0.00018363122881508945, + "loss": 0.8353, + "step": 436 + }, + { + "epoch": 0.210424942819309, + "grad_norm": 1.279977798461914, + "learning_rate": 0.00018354560631105328, + "loss": 0.2531, + "step": 437 + }, + { + "epoch": 0.2109064644275912, + "grad_norm": 1.3670711517333984, + "learning_rate": 0.00018345978052248233, + "loss": 0.6959, + "step": 438 + }, + { + "epoch": 0.21138798603587336, + "grad_norm": 1.0435469150543213, + "learning_rate": 0.00018337375165820944, + "loss": 1.0629, + "step": 439 + }, + { + "epoch": 0.21186950764415552, + "grad_norm": 1.0792940855026245, + "learning_rate": 0.00018328751992756137, + "loss": 0.7146, + "step": 440 + }, + { + "epoch": 0.2123510292524377, + "grad_norm": 1.0037693977355957, + "learning_rate": 0.0001832010855403586, + "loss": 0.5551, + "step": 441 + }, + { + "epoch": 0.21283255086071987, + "grad_norm": 0.7013088464736938, + "learning_rate": 0.0001831144487069147, + "loss": 0.7306, + "step": 442 + }, + { + "epoch": 0.21331407246900205, + "grad_norm": 1.636138916015625, + "learning_rate": 0.0001830276096380358, + "loss": 0.3383, + "step": 443 + }, + { + "epoch": 0.2137955940772842, + "grad_norm": 1.0589203834533691, + "learning_rate": 0.0001829405685450202, + "loss": 0.965, + "step": 444 + }, + { + "epoch": 0.2142771156855664, + "grad_norm": 1.0804965496063232, + "learning_rate": 0.00018285332563965765, + "loss": 1.0993, + "step": 445 + }, + { + "epoch": 0.21475863729384856, + "grad_norm": 0.742942214012146, + "learning_rate": 0.00018276588113422905, + "loss": 0.5167, + "step": 446 + }, + { + "epoch": 0.21524015890213075, + "grad_norm": 0.7951419353485107, + "learning_rate": 0.00018267823524150575, + "loss": 0.9815, + "step": 447 + }, + { + "epoch": 0.2157216805104129, + "grad_norm": 0.8592374920845032, + "learning_rate": 0.00018259038817474923, + "loss": 0.9137, + "step": 448 + }, + { + "epoch": 0.21620320211869506, + "grad_norm": 0.5096921920776367, + "learning_rate": 0.0001825023401477104, + "loss": 0.4994, + "step": 449 + }, + { + "epoch": 0.21668472372697725, + "grad_norm": 0.9474820494651794, + "learning_rate": 0.0001824140913746291, + "loss": 0.7438, + "step": 450 + }, + { + "epoch": 0.2171662453352594, + "grad_norm": 0.4747283160686493, + "learning_rate": 0.00018232564207023376, + "loss": 0.3613, + "step": 451 + }, + { + "epoch": 0.2176477669435416, + "grad_norm": 1.2512322664260864, + "learning_rate": 0.00018223699244974064, + "loss": 0.5614, + "step": 452 + }, + { + "epoch": 0.21812928855182376, + "grad_norm": 0.7003946900367737, + "learning_rate": 0.00018214814272885343, + "loss": 0.7942, + "step": 453 + }, + { + "epoch": 0.21861081016010595, + "grad_norm": 0.7767881155014038, + "learning_rate": 0.00018205909312376276, + "loss": 0.8337, + "step": 454 + }, + { + "epoch": 0.2190923317683881, + "grad_norm": 1.2129909992218018, + "learning_rate": 0.00018196984385114554, + "loss": 0.6863, + "step": 455 + }, + { + "epoch": 0.2195738533766703, + "grad_norm": 0.5771981477737427, + "learning_rate": 0.0001818803951281646, + "loss": 0.5401, + "step": 456 + }, + { + "epoch": 0.22005537498495245, + "grad_norm": 1.2848843336105347, + "learning_rate": 0.000181790747172468, + "loss": 0.7414, + "step": 457 + }, + { + "epoch": 0.2205368965932346, + "grad_norm": 0.7299485206604004, + "learning_rate": 0.00018170090020218864, + "loss": 0.5679, + "step": 458 + }, + { + "epoch": 0.2210184182015168, + "grad_norm": 0.48027756810188293, + "learning_rate": 0.00018161085443594365, + "loss": 0.2938, + "step": 459 + }, + { + "epoch": 0.22149993980979896, + "grad_norm": 0.7984429001808167, + "learning_rate": 0.00018152061009283382, + "loss": 0.3071, + "step": 460 + }, + { + "epoch": 0.22198146141808114, + "grad_norm": 1.3941313028335571, + "learning_rate": 0.00018143016739244314, + "loss": 0.9465, + "step": 461 + }, + { + "epoch": 0.2224629830263633, + "grad_norm": 1.314907193183899, + "learning_rate": 0.0001813395265548383, + "loss": 0.7413, + "step": 462 + }, + { + "epoch": 0.2229445046346455, + "grad_norm": 0.863312304019928, + "learning_rate": 0.00018124868780056814, + "loss": 0.7547, + "step": 463 + }, + { + "epoch": 0.22342602624292765, + "grad_norm": 1.0457203388214111, + "learning_rate": 0.0001811576513506629, + "loss": 0.6511, + "step": 464 + }, + { + "epoch": 0.2239075478512098, + "grad_norm": 1.8019721508026123, + "learning_rate": 0.00018106641742663397, + "loss": 0.8317, + "step": 465 + }, + { + "epoch": 0.224389069459492, + "grad_norm": 0.7878347039222717, + "learning_rate": 0.00018097498625047328, + "loss": 0.3196, + "step": 466 + }, + { + "epoch": 0.22487059106777416, + "grad_norm": 1.159590721130371, + "learning_rate": 0.00018088335804465258, + "loss": 0.6533, + "step": 467 + }, + { + "epoch": 0.22535211267605634, + "grad_norm": 0.7787841558456421, + "learning_rate": 0.00018079153303212318, + "loss": 0.4519, + "step": 468 + }, + { + "epoch": 0.2258336342843385, + "grad_norm": 1.8181616067886353, + "learning_rate": 0.0001806995114363152, + "loss": 0.8562, + "step": 469 + }, + { + "epoch": 0.2263151558926207, + "grad_norm": 0.6238754987716675, + "learning_rate": 0.00018060729348113707, + "loss": 0.4416, + "step": 470 + }, + { + "epoch": 0.22679667750090285, + "grad_norm": 0.5132167339324951, + "learning_rate": 0.00018051487939097505, + "loss": 0.4001, + "step": 471 + }, + { + "epoch": 0.22727819910918504, + "grad_norm": 0.9573964476585388, + "learning_rate": 0.00018042226939069255, + "loss": 0.7387, + "step": 472 + }, + { + "epoch": 0.2277597207174672, + "grad_norm": 0.6595173478126526, + "learning_rate": 0.00018032946370562982, + "loss": 0.6063, + "step": 473 + }, + { + "epoch": 0.22824124232574936, + "grad_norm": 0.8158476948738098, + "learning_rate": 0.00018023646256160313, + "loss": 0.5302, + "step": 474 + }, + { + "epoch": 0.22872276393403154, + "grad_norm": 0.6084806323051453, + "learning_rate": 0.00018014326618490437, + "loss": 0.394, + "step": 475 + }, + { + "epoch": 0.2292042855423137, + "grad_norm": 0.9932984113693237, + "learning_rate": 0.0001800498748023005, + "loss": 0.8495, + "step": 476 + }, + { + "epoch": 0.2296858071505959, + "grad_norm": 0.49222004413604736, + "learning_rate": 0.000179956288641033, + "loss": 0.4108, + "step": 477 + }, + { + "epoch": 0.23016732875887805, + "grad_norm": 0.6978711485862732, + "learning_rate": 0.00017986250792881718, + "loss": 0.3108, + "step": 478 + }, + { + "epoch": 0.23064885036716024, + "grad_norm": 1.0181289911270142, + "learning_rate": 0.00017976853289384184, + "loss": 1.0249, + "step": 479 + }, + { + "epoch": 0.2311303719754424, + "grad_norm": 1.5912185907363892, + "learning_rate": 0.00017967436376476855, + "loss": 1.1765, + "step": 480 + }, + { + "epoch": 0.23161189358372458, + "grad_norm": 1.0102556943893433, + "learning_rate": 0.0001795800007707312, + "loss": 0.4972, + "step": 481 + }, + { + "epoch": 0.23209341519200674, + "grad_norm": 0.4210749566555023, + "learning_rate": 0.00017948544414133534, + "loss": 0.289, + "step": 482 + }, + { + "epoch": 0.2325749368002889, + "grad_norm": 0.4447524845600128, + "learning_rate": 0.00017939069410665773, + "loss": 0.2973, + "step": 483 + }, + { + "epoch": 0.2330564584085711, + "grad_norm": 1.2053577899932861, + "learning_rate": 0.0001792957508972457, + "loss": 0.52, + "step": 484 + }, + { + "epoch": 0.23353798001685325, + "grad_norm": 1.1216098070144653, + "learning_rate": 0.00017920061474411658, + "loss": 0.4747, + "step": 485 + }, + { + "epoch": 0.23401950162513543, + "grad_norm": 2.1695163249969482, + "learning_rate": 0.00017910528587875729, + "loss": 0.7822, + "step": 486 + }, + { + "epoch": 0.2345010232334176, + "grad_norm": 0.5691484212875366, + "learning_rate": 0.00017900976453312352, + "loss": 0.3004, + "step": 487 + }, + { + "epoch": 0.23498254484169978, + "grad_norm": 1.2666643857955933, + "learning_rate": 0.00017891405093963938, + "loss": 0.5522, + "step": 488 + }, + { + "epoch": 0.23546406644998194, + "grad_norm": 1.0269070863723755, + "learning_rate": 0.00017881814533119675, + "loss": 0.9501, + "step": 489 + }, + { + "epoch": 0.23594558805826413, + "grad_norm": 1.3407925367355347, + "learning_rate": 0.00017872204794115474, + "loss": 0.498, + "step": 490 + }, + { + "epoch": 0.2364271096665463, + "grad_norm": 0.6913367509841919, + "learning_rate": 0.0001786257590033391, + "loss": 0.46, + "step": 491 + }, + { + "epoch": 0.23690863127482845, + "grad_norm": 1.1869349479675293, + "learning_rate": 0.00017852927875204163, + "loss": 0.7829, + "step": 492 + }, + { + "epoch": 0.23739015288311063, + "grad_norm": 0.6723718047142029, + "learning_rate": 0.00017843260742201963, + "loss": 0.3372, + "step": 493 + }, + { + "epoch": 0.2378716744913928, + "grad_norm": 0.679757297039032, + "learning_rate": 0.00017833574524849535, + "loss": 0.9048, + "step": 494 + }, + { + "epoch": 0.23835319609967498, + "grad_norm": 1.0547089576721191, + "learning_rate": 0.00017823869246715553, + "loss": 0.6047, + "step": 495 + }, + { + "epoch": 0.23883471770795714, + "grad_norm": 1.2114295959472656, + "learning_rate": 0.00017814144931415043, + "loss": 0.9538, + "step": 496 + }, + { + "epoch": 0.23931623931623933, + "grad_norm": 1.396855354309082, + "learning_rate": 0.0001780440160260938, + "loss": 1.0792, + "step": 497 + }, + { + "epoch": 0.23979776092452149, + "grad_norm": 0.9748196601867676, + "learning_rate": 0.00017794639284006184, + "loss": 1.0788, + "step": 498 + }, + { + "epoch": 0.24027928253280367, + "grad_norm": 1.2064623832702637, + "learning_rate": 0.0001778485799935929, + "loss": 0.7759, + "step": 499 + }, + { + "epoch": 0.24076080414108583, + "grad_norm": 1.0788531303405762, + "learning_rate": 0.00017775057772468679, + "loss": 0.9132, + "step": 500 + }, + { + "epoch": 0.241242325749368, + "grad_norm": 0.8036178350448608, + "learning_rate": 0.00017765238627180424, + "loss": 0.6564, + "step": 501 + }, + { + "epoch": 0.24172384735765018, + "grad_norm": 0.7760926485061646, + "learning_rate": 0.00017755400587386632, + "loss": 0.638, + "step": 502 + }, + { + "epoch": 0.24220536896593234, + "grad_norm": 0.6797692775726318, + "learning_rate": 0.00017745543677025378, + "loss": 0.4243, + "step": 503 + }, + { + "epoch": 0.24268689057421453, + "grad_norm": 0.9898531436920166, + "learning_rate": 0.00017735667920080661, + "loss": 0.6076, + "step": 504 + }, + { + "epoch": 0.24316841218249668, + "grad_norm": 0.7974802255630493, + "learning_rate": 0.0001772577334058233, + "loss": 0.6662, + "step": 505 + }, + { + "epoch": 0.24364993379077887, + "grad_norm": 1.0043370723724365, + "learning_rate": 0.00017715859962606043, + "loss": 0.865, + "step": 506 + }, + { + "epoch": 0.24413145539906103, + "grad_norm": 0.6671229600906372, + "learning_rate": 0.00017705927810273187, + "loss": 0.2871, + "step": 507 + }, + { + "epoch": 0.2446129770073432, + "grad_norm": 1.1656029224395752, + "learning_rate": 0.00017695976907750844, + "loss": 0.5763, + "step": 508 + }, + { + "epoch": 0.24509449861562538, + "grad_norm": 1.1035945415496826, + "learning_rate": 0.00017686007279251706, + "loss": 0.3747, + "step": 509 + }, + { + "epoch": 0.24557602022390754, + "grad_norm": 0.8782248497009277, + "learning_rate": 0.00017676018949034045, + "loss": 0.7433, + "step": 510 + }, + { + "epoch": 0.24605754183218972, + "grad_norm": 0.902968168258667, + "learning_rate": 0.0001766601194140162, + "loss": 0.6472, + "step": 511 + }, + { + "epoch": 0.24653906344047188, + "grad_norm": 1.9581364393234253, + "learning_rate": 0.0001765598628070365, + "loss": 0.7606, + "step": 512 + }, + { + "epoch": 0.24702058504875407, + "grad_norm": 0.7049884796142578, + "learning_rate": 0.00017645941991334732, + "loss": 0.6779, + "step": 513 + }, + { + "epoch": 0.24750210665703623, + "grad_norm": 0.9359114170074463, + "learning_rate": 0.00017635879097734804, + "loss": 0.5744, + "step": 514 + }, + { + "epoch": 0.24798362826531842, + "grad_norm": 1.0945175886154175, + "learning_rate": 0.00017625797624389055, + "loss": 1.2865, + "step": 515 + }, + { + "epoch": 0.24846514987360058, + "grad_norm": 0.6059966087341309, + "learning_rate": 0.00017615697595827897, + "loss": 0.5536, + "step": 516 + }, + { + "epoch": 0.24894667148188274, + "grad_norm": 0.6404644846916199, + "learning_rate": 0.0001760557903662688, + "loss": 0.454, + "step": 517 + }, + { + "epoch": 0.24942819309016492, + "grad_norm": 1.5128916501998901, + "learning_rate": 0.00017595441971406648, + "loss": 1.4014, + "step": 518 + }, + { + "epoch": 0.24990971469844708, + "grad_norm": 0.662467896938324, + "learning_rate": 0.00017585286424832874, + "loss": 0.4347, + "step": 519 + }, + { + "epoch": 0.25039123630672927, + "grad_norm": 0.6688117980957031, + "learning_rate": 0.00017575112421616202, + "loss": 0.7102, + "step": 520 + }, + { + "epoch": 0.25087275791501146, + "grad_norm": 0.5434272885322571, + "learning_rate": 0.0001756491998651218, + "loss": 0.4888, + "step": 521 + }, + { + "epoch": 0.2513542795232936, + "grad_norm": 0.6772214770317078, + "learning_rate": 0.0001755470914432121, + "loss": 0.7237, + "step": 522 + }, + { + "epoch": 0.2518358011315758, + "grad_norm": 0.49257853627204895, + "learning_rate": 0.0001754447991988848, + "loss": 0.3216, + "step": 523 + }, + { + "epoch": 0.25231732273985796, + "grad_norm": 0.9488145709037781, + "learning_rate": 0.00017534232338103903, + "loss": 1.2589, + "step": 524 + }, + { + "epoch": 0.25279884434814015, + "grad_norm": 0.6488147377967834, + "learning_rate": 0.0001752396642390207, + "loss": 0.845, + "step": 525 + }, + { + "epoch": 0.2532803659564223, + "grad_norm": 0.9418458342552185, + "learning_rate": 0.00017513682202262163, + "loss": 0.9113, + "step": 526 + }, + { + "epoch": 0.25376188756470447, + "grad_norm": 0.8072599768638611, + "learning_rate": 0.00017503379698207918, + "loss": 0.4421, + "step": 527 + }, + { + "epoch": 0.25424340917298666, + "grad_norm": 2.4496564865112305, + "learning_rate": 0.00017493058936807562, + "loss": 0.7209, + "step": 528 + }, + { + "epoch": 0.2547249307812688, + "grad_norm": 0.8409088850021362, + "learning_rate": 0.00017482719943173739, + "loss": 0.5874, + "step": 529 + }, + { + "epoch": 0.255206452389551, + "grad_norm": 0.7664375901222229, + "learning_rate": 0.00017472362742463455, + "loss": 0.5683, + "step": 530 + }, + { + "epoch": 0.25568797399783316, + "grad_norm": 1.121518850326538, + "learning_rate": 0.0001746198735987802, + "loss": 0.957, + "step": 531 + }, + { + "epoch": 0.25616949560611535, + "grad_norm": 0.6444967985153198, + "learning_rate": 0.00017451593820662988, + "loss": 0.2528, + "step": 532 + }, + { + "epoch": 0.2566510172143975, + "grad_norm": 1.1670219898223877, + "learning_rate": 0.00017441182150108086, + "loss": 0.8635, + "step": 533 + }, + { + "epoch": 0.25713253882267967, + "grad_norm": 0.790825605392456, + "learning_rate": 0.0001743075237354716, + "loss": 0.4617, + "step": 534 + }, + { + "epoch": 0.25761406043096186, + "grad_norm": 0.7336540222167969, + "learning_rate": 0.00017420304516358113, + "loss": 0.6611, + "step": 535 + }, + { + "epoch": 0.258095582039244, + "grad_norm": 1.0129753351211548, + "learning_rate": 0.00017409838603962843, + "loss": 0.868, + "step": 536 + }, + { + "epoch": 0.2585771036475262, + "grad_norm": 0.9752351641654968, + "learning_rate": 0.00017399354661827178, + "loss": 0.6151, + "step": 537 + }, + { + "epoch": 0.25905862525580836, + "grad_norm": 2.983139753341675, + "learning_rate": 0.00017388852715460819, + "loss": 0.5691, + "step": 538 + }, + { + "epoch": 0.25954014686409055, + "grad_norm": 0.6391310095787048, + "learning_rate": 0.00017378332790417273, + "loss": 0.4331, + "step": 539 + }, + { + "epoch": 0.2600216684723727, + "grad_norm": 0.9141309261322021, + "learning_rate": 0.00017367794912293794, + "loss": 0.8207, + "step": 540 + }, + { + "epoch": 0.26050319008065487, + "grad_norm": 0.5318431854248047, + "learning_rate": 0.00017357239106731317, + "loss": 0.3676, + "step": 541 + }, + { + "epoch": 0.26098471168893705, + "grad_norm": 1.2545955181121826, + "learning_rate": 0.00017346665399414405, + "loss": 0.9302, + "step": 542 + }, + { + "epoch": 0.2614662332972192, + "grad_norm": 0.7107487320899963, + "learning_rate": 0.00017336073816071168, + "loss": 0.3593, + "step": 543 + }, + { + "epoch": 0.2619477549055014, + "grad_norm": 3.153602123260498, + "learning_rate": 0.00017325464382473226, + "loss": 0.6631, + "step": 544 + }, + { + "epoch": 0.26242927651378356, + "grad_norm": 0.750834584236145, + "learning_rate": 0.00017314837124435622, + "loss": 0.7142, + "step": 545 + }, + { + "epoch": 0.26291079812206575, + "grad_norm": 0.7795765995979309, + "learning_rate": 0.00017304192067816782, + "loss": 0.3292, + "step": 546 + }, + { + "epoch": 0.2633923197303479, + "grad_norm": 0.5292965173721313, + "learning_rate": 0.00017293529238518422, + "loss": 0.2838, + "step": 547 + }, + { + "epoch": 0.26387384133863007, + "grad_norm": 0.7363600134849548, + "learning_rate": 0.0001728284866248552, + "loss": 0.7192, + "step": 548 + }, + { + "epoch": 0.26435536294691225, + "grad_norm": 1.5265337228775024, + "learning_rate": 0.00017272150365706224, + "loss": 0.4173, + "step": 549 + }, + { + "epoch": 0.26483688455519444, + "grad_norm": 0.7974181771278381, + "learning_rate": 0.00017261434374211802, + "loss": 0.5197, + "step": 550 + }, + { + "epoch": 0.2653184061634766, + "grad_norm": 0.7714865803718567, + "learning_rate": 0.00017250700714076586, + "loss": 0.5794, + "step": 551 + }, + { + "epoch": 0.26579992777175876, + "grad_norm": 0.587791919708252, + "learning_rate": 0.00017239949411417888, + "loss": 0.6678, + "step": 552 + }, + { + "epoch": 0.26628144938004095, + "grad_norm": 1.730128288269043, + "learning_rate": 0.0001722918049239596, + "loss": 1.2, + "step": 553 + }, + { + "epoch": 0.2667629709883231, + "grad_norm": 0.7081428170204163, + "learning_rate": 0.00017218393983213902, + "loss": 0.3536, + "step": 554 + }, + { + "epoch": 0.26724449259660527, + "grad_norm": 0.8215841054916382, + "learning_rate": 0.00017207589910117634, + "loss": 0.4589, + "step": 555 + }, + { + "epoch": 0.26772601420488745, + "grad_norm": 1.0128780603408813, + "learning_rate": 0.00017196768299395797, + "loss": 0.7296, + "step": 556 + }, + { + "epoch": 0.26820753581316964, + "grad_norm": 1.796431303024292, + "learning_rate": 0.00017185929177379714, + "loss": 0.7464, + "step": 557 + }, + { + "epoch": 0.26868905742145177, + "grad_norm": 1.8734339475631714, + "learning_rate": 0.00017175072570443312, + "loss": 1.0069, + "step": 558 + }, + { + "epoch": 0.26917057902973396, + "grad_norm": 0.860616147518158, + "learning_rate": 0.00017164198505003066, + "loss": 0.5769, + "step": 559 + }, + { + "epoch": 0.26965210063801615, + "grad_norm": 0.49917736649513245, + "learning_rate": 0.0001715330700751793, + "loss": 0.2335, + "step": 560 + }, + { + "epoch": 0.2701336222462983, + "grad_norm": 0.6714743375778198, + "learning_rate": 0.00017142398104489273, + "loss": 0.4088, + "step": 561 + }, + { + "epoch": 0.27061514385458046, + "grad_norm": 0.6164880990982056, + "learning_rate": 0.00017131471822460814, + "loss": 0.7426, + "step": 562 + }, + { + "epoch": 0.27109666546286265, + "grad_norm": 0.5979893803596497, + "learning_rate": 0.00017120528188018565, + "loss": 0.2952, + "step": 563 + }, + { + "epoch": 0.27157818707114484, + "grad_norm": 0.6977973580360413, + "learning_rate": 0.00017109567227790754, + "loss": 0.4368, + "step": 564 + }, + { + "epoch": 0.27205970867942697, + "grad_norm": 0.7051186561584473, + "learning_rate": 0.00017098588968447766, + "loss": 0.8569, + "step": 565 + }, + { + "epoch": 0.27254123028770916, + "grad_norm": 0.5577298998832703, + "learning_rate": 0.00017087593436702084, + "loss": 0.4247, + "step": 566 + }, + { + "epoch": 0.27302275189599134, + "grad_norm": 0.9759942293167114, + "learning_rate": 0.00017076580659308222, + "loss": 1.0734, + "step": 567 + }, + { + "epoch": 0.27350427350427353, + "grad_norm": 0.5310359001159668, + "learning_rate": 0.00017065550663062634, + "loss": 0.3319, + "step": 568 + }, + { + "epoch": 0.27398579511255566, + "grad_norm": 0.8164540529251099, + "learning_rate": 0.00017054503474803702, + "loss": 0.4834, + "step": 569 + }, + { + "epoch": 0.27446731672083785, + "grad_norm": 0.9512844085693359, + "learning_rate": 0.00017043439121411618, + "loss": 0.4703, + "step": 570 + }, + { + "epoch": 0.27494883832912004, + "grad_norm": 1.0269114971160889, + "learning_rate": 0.0001703235762980835, + "loss": 0.3504, + "step": 571 + }, + { + "epoch": 0.27543035993740217, + "grad_norm": 0.6352307200431824, + "learning_rate": 0.00017021259026957567, + "loss": 0.7479, + "step": 572 + }, + { + "epoch": 0.27591188154568436, + "grad_norm": 1.2709698677062988, + "learning_rate": 0.00017010143339864562, + "loss": 0.8125, + "step": 573 + }, + { + "epoch": 0.27639340315396654, + "grad_norm": 1.1885913610458374, + "learning_rate": 0.0001699901059557621, + "loss": 0.7269, + "step": 574 + }, + { + "epoch": 0.27687492476224873, + "grad_norm": 1.061423420906067, + "learning_rate": 0.00016987860821180895, + "loss": 0.8678, + "step": 575 + }, + { + "epoch": 0.27735644637053086, + "grad_norm": 1.2436217069625854, + "learning_rate": 0.00016976694043808416, + "loss": 1.0843, + "step": 576 + }, + { + "epoch": 0.27783796797881305, + "grad_norm": 1.5231688022613525, + "learning_rate": 0.00016965510290629972, + "loss": 0.6489, + "step": 577 + }, + { + "epoch": 0.27831948958709524, + "grad_norm": 0.8685100674629211, + "learning_rate": 0.00016954309588858044, + "loss": 0.5554, + "step": 578 + }, + { + "epoch": 0.27880101119537737, + "grad_norm": 0.8666234612464905, + "learning_rate": 0.00016943091965746366, + "loss": 0.6095, + "step": 579 + }, + { + "epoch": 0.27928253280365956, + "grad_norm": 1.7214361429214478, + "learning_rate": 0.00016931857448589845, + "loss": 1.1025, + "step": 580 + }, + { + "epoch": 0.27976405441194174, + "grad_norm": 0.4393554627895355, + "learning_rate": 0.00016920606064724488, + "loss": 0.4294, + "step": 581 + }, + { + "epoch": 0.28024557602022393, + "grad_norm": 0.4375576972961426, + "learning_rate": 0.00016909337841527344, + "loss": 0.5308, + "step": 582 + }, + { + "epoch": 0.28072709762850606, + "grad_norm": 0.9495024085044861, + "learning_rate": 0.00016898052806416444, + "loss": 0.6907, + "step": 583 + }, + { + "epoch": 0.28120861923678825, + "grad_norm": 0.7171463370323181, + "learning_rate": 0.00016886750986850718, + "loss": 0.3165, + "step": 584 + }, + { + "epoch": 0.28169014084507044, + "grad_norm": 0.6095327138900757, + "learning_rate": 0.00016875432410329934, + "loss": 0.7329, + "step": 585 + }, + { + "epoch": 0.28217166245335257, + "grad_norm": 0.9908806085586548, + "learning_rate": 0.0001686409710439464, + "loss": 0.5553, + "step": 586 + }, + { + "epoch": 0.28265318406163475, + "grad_norm": 0.9707655906677246, + "learning_rate": 0.00016852745096626088, + "loss": 0.8345, + "step": 587 + }, + { + "epoch": 0.28313470566991694, + "grad_norm": 0.6714180707931519, + "learning_rate": 0.0001684137641464617, + "loss": 0.6241, + "step": 588 + }, + { + "epoch": 0.28361622727819913, + "grad_norm": 1.1676234006881714, + "learning_rate": 0.0001682999108611735, + "loss": 0.3852, + "step": 589 + }, + { + "epoch": 0.28409774888648126, + "grad_norm": 0.5526348948478699, + "learning_rate": 0.00016818589138742587, + "loss": 0.5523, + "step": 590 + }, + { + "epoch": 0.28457927049476345, + "grad_norm": 0.7403177618980408, + "learning_rate": 0.00016807170600265296, + "loss": 0.5595, + "step": 591 + }, + { + "epoch": 0.28506079210304563, + "grad_norm": 1.745147705078125, + "learning_rate": 0.00016795735498469246, + "loss": 0.3464, + "step": 592 + }, + { + "epoch": 0.2855423137113278, + "grad_norm": 1.051491141319275, + "learning_rate": 0.00016784283861178513, + "loss": 0.6082, + "step": 593 + }, + { + "epoch": 0.28602383531960995, + "grad_norm": 1.1357232332229614, + "learning_rate": 0.00016772815716257412, + "loss": 0.4084, + "step": 594 + }, + { + "epoch": 0.28650535692789214, + "grad_norm": 1.4268995523452759, + "learning_rate": 0.00016761331091610416, + "loss": 0.8438, + "step": 595 + }, + { + "epoch": 0.28698687853617433, + "grad_norm": 0.7449113130569458, + "learning_rate": 0.00016749830015182107, + "loss": 0.5502, + "step": 596 + }, + { + "epoch": 0.28746840014445646, + "grad_norm": 0.5473132729530334, + "learning_rate": 0.00016738312514957086, + "loss": 0.2627, + "step": 597 + }, + { + "epoch": 0.28794992175273865, + "grad_norm": 0.5442512035369873, + "learning_rate": 0.00016726778618959926, + "loss": 0.4934, + "step": 598 + }, + { + "epoch": 0.28843144336102083, + "grad_norm": 0.7760728001594543, + "learning_rate": 0.00016715228355255093, + "loss": 0.6557, + "step": 599 + }, + { + "epoch": 0.288912964969303, + "grad_norm": 2.4835331439971924, + "learning_rate": 0.00016703661751946874, + "loss": 1.2639, + "step": 600 + }, + { + "epoch": 0.28939448657758515, + "grad_norm": 0.7905292510986328, + "learning_rate": 0.00016692078837179318, + "loss": 0.778, + "step": 601 + }, + { + "epoch": 0.28987600818586734, + "grad_norm": 0.8698615431785583, + "learning_rate": 0.00016680479639136163, + "loss": 0.7836, + "step": 602 + }, + { + "epoch": 0.2903575297941495, + "grad_norm": 0.9689038395881653, + "learning_rate": 0.0001666886418604077, + "loss": 0.3091, + "step": 603 + }, + { + "epoch": 0.29083905140243166, + "grad_norm": 0.6081171035766602, + "learning_rate": 0.0001665723250615604, + "loss": 0.5046, + "step": 604 + }, + { + "epoch": 0.29132057301071385, + "grad_norm": 1.1759047508239746, + "learning_rate": 0.00016645584627784381, + "loss": 0.8711, + "step": 605 + }, + { + "epoch": 0.29180209461899603, + "grad_norm": 0.8777217268943787, + "learning_rate": 0.0001663392057926759, + "loss": 0.6645, + "step": 606 + }, + { + "epoch": 0.2922836162272782, + "grad_norm": 0.5921443700790405, + "learning_rate": 0.00016622240388986824, + "loss": 0.3308, + "step": 607 + }, + { + "epoch": 0.29276513783556035, + "grad_norm": 1.0305765867233276, + "learning_rate": 0.0001661054408536251, + "loss": 1.1618, + "step": 608 + }, + { + "epoch": 0.29324665944384254, + "grad_norm": 0.7973262667655945, + "learning_rate": 0.00016598831696854288, + "loss": 0.5259, + "step": 609 + }, + { + "epoch": 0.2937281810521247, + "grad_norm": 0.7054857015609741, + "learning_rate": 0.00016587103251960937, + "loss": 0.6195, + "step": 610 + }, + { + "epoch": 0.2942097026604069, + "grad_norm": 0.850437581539154, + "learning_rate": 0.00016575358779220294, + "loss": 0.8113, + "step": 611 + }, + { + "epoch": 0.29469122426868904, + "grad_norm": 1.504550576210022, + "learning_rate": 0.00016563598307209204, + "loss": 1.0547, + "step": 612 + }, + { + "epoch": 0.29517274587697123, + "grad_norm": 0.4776376485824585, + "learning_rate": 0.0001655182186454344, + "loss": 0.2646, + "step": 613 + }, + { + "epoch": 0.2956542674852534, + "grad_norm": 0.8611229062080383, + "learning_rate": 0.00016540029479877638, + "loss": 1.0374, + "step": 614 + }, + { + "epoch": 0.29613578909353555, + "grad_norm": 0.8100734353065491, + "learning_rate": 0.00016528221181905217, + "loss": 0.2605, + "step": 615 + }, + { + "epoch": 0.29661731070181774, + "grad_norm": 1.2284023761749268, + "learning_rate": 0.00016516396999358322, + "loss": 0.8769, + "step": 616 + }, + { + "epoch": 0.2970988323100999, + "grad_norm": 0.7563827037811279, + "learning_rate": 0.00016504556961007748, + "loss": 0.6459, + "step": 617 + }, + { + "epoch": 0.2975803539183821, + "grad_norm": 0.7285965085029602, + "learning_rate": 0.00016492701095662866, + "loss": 0.4741, + "step": 618 + }, + { + "epoch": 0.29806187552666424, + "grad_norm": 0.8569052219390869, + "learning_rate": 0.00016480829432171564, + "loss": 0.6236, + "step": 619 + }, + { + "epoch": 0.29854339713494643, + "grad_norm": 0.6136265397071838, + "learning_rate": 0.0001646894199942017, + "loss": 0.6324, + "step": 620 + }, + { + "epoch": 0.2990249187432286, + "grad_norm": 0.8613694310188293, + "learning_rate": 0.0001645703882633338, + "loss": 0.8755, + "step": 621 + }, + { + "epoch": 0.29950644035151075, + "grad_norm": 0.5661609768867493, + "learning_rate": 0.00016445119941874183, + "loss": 0.3262, + "step": 622 + }, + { + "epoch": 0.29998796195979294, + "grad_norm": 0.8413470983505249, + "learning_rate": 0.00016433185375043809, + "loss": 0.6209, + "step": 623 + }, + { + "epoch": 0.3004694835680751, + "grad_norm": 1.3993409872055054, + "learning_rate": 0.00016421235154881638, + "loss": 0.9886, + "step": 624 + }, + { + "epoch": 0.3009510051763573, + "grad_norm": 0.6332157850265503, + "learning_rate": 0.00016409269310465146, + "loss": 0.3571, + "step": 625 + }, + { + "epoch": 0.30143252678463944, + "grad_norm": 1.0084850788116455, + "learning_rate": 0.00016397287870909813, + "loss": 1.2828, + "step": 626 + }, + { + "epoch": 0.30191404839292163, + "grad_norm": 0.9696491360664368, + "learning_rate": 0.00016385290865369079, + "loss": 1.0937, + "step": 627 + }, + { + "epoch": 0.3023955700012038, + "grad_norm": 1.2192230224609375, + "learning_rate": 0.00016373278323034255, + "loss": 0.4202, + "step": 628 + }, + { + "epoch": 0.30287709160948595, + "grad_norm": 0.7007105350494385, + "learning_rate": 0.0001636125027313445, + "loss": 0.4055, + "step": 629 + }, + { + "epoch": 0.30335861321776814, + "grad_norm": 1.2077027559280396, + "learning_rate": 0.00016349206744936518, + "loss": 0.8379, + "step": 630 + }, + { + "epoch": 0.3038401348260503, + "grad_norm": 1.1231609582901, + "learning_rate": 0.00016337147767744967, + "loss": 1.1825, + "step": 631 + }, + { + "epoch": 0.3043216564343325, + "grad_norm": 1.050195336341858, + "learning_rate": 0.0001632507337090189, + "loss": 0.6219, + "step": 632 + }, + { + "epoch": 0.30480317804261464, + "grad_norm": 0.8502358198165894, + "learning_rate": 0.0001631298358378692, + "loss": 0.7966, + "step": 633 + }, + { + "epoch": 0.30528469965089683, + "grad_norm": 0.923303484916687, + "learning_rate": 0.00016300878435817113, + "loss": 0.8952, + "step": 634 + }, + { + "epoch": 0.305766221259179, + "grad_norm": 0.8884153366088867, + "learning_rate": 0.00016288757956446918, + "loss": 0.6296, + "step": 635 + }, + { + "epoch": 0.3062477428674612, + "grad_norm": 1.4516477584838867, + "learning_rate": 0.00016276622175168083, + "loss": 0.4871, + "step": 636 + }, + { + "epoch": 0.30672926447574334, + "grad_norm": 1.6598066091537476, + "learning_rate": 0.0001626447112150959, + "loss": 1.208, + "step": 637 + }, + { + "epoch": 0.3072107860840255, + "grad_norm": 0.6673110723495483, + "learning_rate": 0.00016252304825037576, + "loss": 0.7643, + "step": 638 + }, + { + "epoch": 0.3076923076923077, + "grad_norm": 1.602836012840271, + "learning_rate": 0.0001624012331535528, + "loss": 1.1832, + "step": 639 + }, + { + "epoch": 0.30817382930058984, + "grad_norm": 0.8961763978004456, + "learning_rate": 0.00016227926622102947, + "loss": 0.5228, + "step": 640 + }, + { + "epoch": 0.30865535090887203, + "grad_norm": 0.9909448623657227, + "learning_rate": 0.00016215714774957772, + "loss": 0.9294, + "step": 641 + }, + { + "epoch": 0.3091368725171542, + "grad_norm": 0.9084739685058594, + "learning_rate": 0.00016203487803633822, + "loss": 0.6734, + "step": 642 + }, + { + "epoch": 0.3096183941254364, + "grad_norm": 0.7972694635391235, + "learning_rate": 0.00016191245737881956, + "loss": 0.4454, + "step": 643 + }, + { + "epoch": 0.31009991573371853, + "grad_norm": 0.909772515296936, + "learning_rate": 0.00016178988607489777, + "loss": 0.5753, + "step": 644 + }, + { + "epoch": 0.3105814373420007, + "grad_norm": 1.612001895904541, + "learning_rate": 0.00016166716442281528, + "loss": 1.2121, + "step": 645 + }, + { + "epoch": 0.3110629589502829, + "grad_norm": 1.3951966762542725, + "learning_rate": 0.0001615442927211805, + "loss": 0.6069, + "step": 646 + }, + { + "epoch": 0.31154448055856504, + "grad_norm": 0.5090633630752563, + "learning_rate": 0.0001614212712689668, + "loss": 0.4586, + "step": 647 + }, + { + "epoch": 0.3120260021668472, + "grad_norm": 1.3730465173721313, + "learning_rate": 0.00016129810036551198, + "loss": 1.0391, + "step": 648 + }, + { + "epoch": 0.3125075237751294, + "grad_norm": 1.2121727466583252, + "learning_rate": 0.00016117478031051755, + "loss": 1.0782, + "step": 649 + }, + { + "epoch": 0.3129890453834116, + "grad_norm": 1.1098741292953491, + "learning_rate": 0.00016105131140404787, + "loss": 0.642, + "step": 650 + }, + { + "epoch": 0.31347056699169373, + "grad_norm": 1.337183952331543, + "learning_rate": 0.00016092769394652947, + "loss": 0.6911, + "step": 651 + }, + { + "epoch": 0.3139520885999759, + "grad_norm": 0.5842844247817993, + "learning_rate": 0.0001608039282387504, + "loss": 0.2357, + "step": 652 + }, + { + "epoch": 0.3144336102082581, + "grad_norm": 1.0377110242843628, + "learning_rate": 0.00016068001458185936, + "loss": 0.6098, + "step": 653 + }, + { + "epoch": 0.3149151318165403, + "grad_norm": 0.868000328540802, + "learning_rate": 0.0001605559532773651, + "loss": 0.689, + "step": 654 + }, + { + "epoch": 0.3153966534248224, + "grad_norm": 0.9237671494483948, + "learning_rate": 0.00016043174462713566, + "loss": 0.2829, + "step": 655 + }, + { + "epoch": 0.3158781750331046, + "grad_norm": 1.0178114175796509, + "learning_rate": 0.00016030738893339753, + "loss": 0.5194, + "step": 656 + }, + { + "epoch": 0.3163596966413868, + "grad_norm": 0.6489741802215576, + "learning_rate": 0.00016018288649873497, + "loss": 0.6068, + "step": 657 + }, + { + "epoch": 0.31684121824966893, + "grad_norm": 0.8735011219978333, + "learning_rate": 0.0001600582376260894, + "loss": 0.3923, + "step": 658 + }, + { + "epoch": 0.3173227398579511, + "grad_norm": 0.7427851557731628, + "learning_rate": 0.00015993344261875847, + "loss": 0.6916, + "step": 659 + }, + { + "epoch": 0.3178042614662333, + "grad_norm": 0.6694474220275879, + "learning_rate": 0.00015980850178039547, + "loss": 0.6211, + "step": 660 + }, + { + "epoch": 0.3182857830745155, + "grad_norm": 0.7267779111862183, + "learning_rate": 0.00015968341541500842, + "loss": 0.8095, + "step": 661 + }, + { + "epoch": 0.3187673046827976, + "grad_norm": 0.7212939262390137, + "learning_rate": 0.00015955818382695953, + "loss": 0.4024, + "step": 662 + }, + { + "epoch": 0.3192488262910798, + "grad_norm": 1.4665344953536987, + "learning_rate": 0.00015943280732096438, + "loss": 1.3903, + "step": 663 + }, + { + "epoch": 0.319730347899362, + "grad_norm": 0.7396234273910522, + "learning_rate": 0.00015930728620209113, + "loss": 0.6563, + "step": 664 + }, + { + "epoch": 0.32021186950764413, + "grad_norm": 1.8860375881195068, + "learning_rate": 0.00015918162077575976, + "loss": 0.8506, + "step": 665 + }, + { + "epoch": 0.3206933911159263, + "grad_norm": 0.8444084525108337, + "learning_rate": 0.00015905581134774153, + "loss": 0.5888, + "step": 666 + }, + { + "epoch": 0.3211749127242085, + "grad_norm": 0.8889416456222534, + "learning_rate": 0.0001589298582241579, + "loss": 0.9857, + "step": 667 + }, + { + "epoch": 0.3216564343324907, + "grad_norm": 0.9507708549499512, + "learning_rate": 0.00015880376171148014, + "loss": 0.4902, + "step": 668 + }, + { + "epoch": 0.3221379559407728, + "grad_norm": 0.7778554558753967, + "learning_rate": 0.00015867752211652831, + "loss": 0.6368, + "step": 669 + }, + { + "epoch": 0.322619477549055, + "grad_norm": 0.9680811762809753, + "learning_rate": 0.00015855113974647068, + "loss": 1.2206, + "step": 670 + }, + { + "epoch": 0.3231009991573372, + "grad_norm": 1.222985029220581, + "learning_rate": 0.0001584246149088229, + "loss": 0.5694, + "step": 671 + }, + { + "epoch": 0.32358252076561933, + "grad_norm": 0.7617347240447998, + "learning_rate": 0.0001582979479114472, + "loss": 0.6872, + "step": 672 + }, + { + "epoch": 0.3240640423739015, + "grad_norm": 0.8455665111541748, + "learning_rate": 0.0001581711390625519, + "loss": 0.4931, + "step": 673 + }, + { + "epoch": 0.3245455639821837, + "grad_norm": 1.143609642982483, + "learning_rate": 0.0001580441886706903, + "loss": 0.8739, + "step": 674 + }, + { + "epoch": 0.3250270855904659, + "grad_norm": 0.8333857655525208, + "learning_rate": 0.00015791709704476015, + "loss": 0.6653, + "step": 675 + }, + { + "epoch": 0.325508607198748, + "grad_norm": 0.41765791177749634, + "learning_rate": 0.00015778986449400292, + "loss": 0.3178, + "step": 676 + }, + { + "epoch": 0.3259901288070302, + "grad_norm": 0.5796922445297241, + "learning_rate": 0.00015766249132800292, + "loss": 0.5722, + "step": 677 + }, + { + "epoch": 0.3264716504153124, + "grad_norm": 0.763243556022644, + "learning_rate": 0.00015753497785668663, + "loss": 0.3656, + "step": 678 + }, + { + "epoch": 0.3269531720235946, + "grad_norm": 1.2675962448120117, + "learning_rate": 0.00015740732439032187, + "loss": 0.4063, + "step": 679 + }, + { + "epoch": 0.3274346936318767, + "grad_norm": 0.7724534273147583, + "learning_rate": 0.00015727953123951716, + "loss": 0.9757, + "step": 680 + }, + { + "epoch": 0.3279162152401589, + "grad_norm": 0.8847770690917969, + "learning_rate": 0.00015715159871522086, + "loss": 0.5965, + "step": 681 + }, + { + "epoch": 0.3283977368484411, + "grad_norm": 0.6064730882644653, + "learning_rate": 0.00015702352712872056, + "loss": 0.4475, + "step": 682 + }, + { + "epoch": 0.3288792584567232, + "grad_norm": 0.9881625175476074, + "learning_rate": 0.00015689531679164204, + "loss": 0.8257, + "step": 683 + }, + { + "epoch": 0.3293607800650054, + "grad_norm": 1.2793527841567993, + "learning_rate": 0.00015676696801594886, + "loss": 0.9258, + "step": 684 + }, + { + "epoch": 0.3298423016732876, + "grad_norm": 0.9854462146759033, + "learning_rate": 0.00015663848111394132, + "loss": 1.0733, + "step": 685 + }, + { + "epoch": 0.3303238232815698, + "grad_norm": 1.0031025409698486, + "learning_rate": 0.00015650985639825585, + "loss": 1.077, + "step": 686 + }, + { + "epoch": 0.3308053448898519, + "grad_norm": 0.9065315127372742, + "learning_rate": 0.00015638109418186424, + "loss": 0.2222, + "step": 687 + }, + { + "epoch": 0.3312868664981341, + "grad_norm": 0.8304358124732971, + "learning_rate": 0.00015625219477807277, + "loss": 0.7335, + "step": 688 + }, + { + "epoch": 0.3317683881064163, + "grad_norm": 0.6953612565994263, + "learning_rate": 0.00015612315850052166, + "loss": 0.8908, + "step": 689 + }, + { + "epoch": 0.3322499097146984, + "grad_norm": 1.183565616607666, + "learning_rate": 0.00015599398566318396, + "loss": 0.6687, + "step": 690 + }, + { + "epoch": 0.3327314313229806, + "grad_norm": 0.7502872943878174, + "learning_rate": 0.00015586467658036524, + "loss": 0.4798, + "step": 691 + }, + { + "epoch": 0.3332129529312628, + "grad_norm": 1.2281885147094727, + "learning_rate": 0.00015573523156670244, + "loss": 1.0472, + "step": 692 + }, + { + "epoch": 0.333694474539545, + "grad_norm": 2.2051644325256348, + "learning_rate": 0.0001556056509371633, + "loss": 0.9149, + "step": 693 + }, + { + "epoch": 0.3341759961478271, + "grad_norm": 0.5925647616386414, + "learning_rate": 0.00015547593500704547, + "loss": 0.8149, + "step": 694 + }, + { + "epoch": 0.3346575177561093, + "grad_norm": 0.5184139013290405, + "learning_rate": 0.00015534608409197592, + "loss": 0.3174, + "step": 695 + }, + { + "epoch": 0.3351390393643915, + "grad_norm": 0.702053964138031, + "learning_rate": 0.00015521609850791004, + "loss": 0.4663, + "step": 696 + }, + { + "epoch": 0.3356205609726737, + "grad_norm": 0.6669120788574219, + "learning_rate": 0.0001550859785711308, + "loss": 0.4877, + "step": 697 + }, + { + "epoch": 0.3361020825809558, + "grad_norm": 1.3044205904006958, + "learning_rate": 0.0001549557245982482, + "loss": 0.9579, + "step": 698 + }, + { + "epoch": 0.336583604189238, + "grad_norm": 0.6670746207237244, + "learning_rate": 0.00015482533690619837, + "loss": 0.1644, + "step": 699 + }, + { + "epoch": 0.3370651257975202, + "grad_norm": 1.218701720237732, + "learning_rate": 0.00015469481581224272, + "loss": 0.517, + "step": 700 + }, + { + "epoch": 0.3375466474058023, + "grad_norm": 0.6596776843070984, + "learning_rate": 0.0001545641616339673, + "loss": 0.3598, + "step": 701 + }, + { + "epoch": 0.3380281690140845, + "grad_norm": 0.7192044854164124, + "learning_rate": 0.00015443337468928206, + "loss": 0.3021, + "step": 702 + }, + { + "epoch": 0.3385096906223667, + "grad_norm": 0.8185523152351379, + "learning_rate": 0.00015430245529641986, + "loss": 0.331, + "step": 703 + }, + { + "epoch": 0.3389912122306489, + "grad_norm": 0.8126874566078186, + "learning_rate": 0.00015417140377393596, + "loss": 0.7754, + "step": 704 + }, + { + "epoch": 0.339472733838931, + "grad_norm": 0.8206513524055481, + "learning_rate": 0.00015404022044070704, + "loss": 0.4213, + "step": 705 + }, + { + "epoch": 0.3399542554472132, + "grad_norm": 0.9104949831962585, + "learning_rate": 0.00015390890561593052, + "loss": 0.6292, + "step": 706 + }, + { + "epoch": 0.3404357770554954, + "grad_norm": 0.8065645694732666, + "learning_rate": 0.0001537774596191238, + "loss": 0.5776, + "step": 707 + }, + { + "epoch": 0.3409172986637775, + "grad_norm": 0.8759814500808716, + "learning_rate": 0.00015364588277012344, + "loss": 0.6147, + "step": 708 + }, + { + "epoch": 0.3413988202720597, + "grad_norm": 0.8460003137588501, + "learning_rate": 0.00015351417538908435, + "loss": 1.1585, + "step": 709 + }, + { + "epoch": 0.3418803418803419, + "grad_norm": 0.5036705136299133, + "learning_rate": 0.0001533823377964791, + "loss": 0.5074, + "step": 710 + }, + { + "epoch": 0.3423618634886241, + "grad_norm": 0.9990852475166321, + "learning_rate": 0.00015325037031309704, + "loss": 0.817, + "step": 711 + }, + { + "epoch": 0.3428433850969062, + "grad_norm": 1.0572148561477661, + "learning_rate": 0.00015311827326004363, + "loss": 0.5744, + "step": 712 + }, + { + "epoch": 0.3433249067051884, + "grad_norm": 0.6696701049804688, + "learning_rate": 0.0001529860469587396, + "loss": 0.5898, + "step": 713 + }, + { + "epoch": 0.3438064283134706, + "grad_norm": 1.029279351234436, + "learning_rate": 0.00015285369173092015, + "loss": 0.4406, + "step": 714 + }, + { + "epoch": 0.3442879499217527, + "grad_norm": 1.0727728605270386, + "learning_rate": 0.00015272120789863413, + "loss": 0.7744, + "step": 715 + }, + { + "epoch": 0.3447694715300349, + "grad_norm": 0.7367604374885559, + "learning_rate": 0.00015258859578424342, + "loss": 0.7764, + "step": 716 + }, + { + "epoch": 0.3452509931383171, + "grad_norm": 0.877487063407898, + "learning_rate": 0.00015245585571042194, + "loss": 0.3476, + "step": 717 + }, + { + "epoch": 0.3457325147465993, + "grad_norm": 0.7964932918548584, + "learning_rate": 0.00015232298800015506, + "loss": 0.399, + "step": 718 + }, + { + "epoch": 0.3462140363548814, + "grad_norm": 0.7974869608879089, + "learning_rate": 0.00015218999297673862, + "loss": 0.3338, + "step": 719 + }, + { + "epoch": 0.3466955579631636, + "grad_norm": 0.8630387783050537, + "learning_rate": 0.0001520568709637783, + "loss": 0.7174, + "step": 720 + }, + { + "epoch": 0.3471770795714458, + "grad_norm": 0.797203540802002, + "learning_rate": 0.00015192362228518875, + "loss": 0.4063, + "step": 721 + }, + { + "epoch": 0.34765860117972797, + "grad_norm": 0.870797336101532, + "learning_rate": 0.00015179024726519284, + "loss": 0.8056, + "step": 722 + }, + { + "epoch": 0.3481401227880101, + "grad_norm": 0.7476208806037903, + "learning_rate": 0.00015165674622832085, + "loss": 0.4459, + "step": 723 + }, + { + "epoch": 0.3486216443962923, + "grad_norm": 1.1456966400146484, + "learning_rate": 0.0001515231194994097, + "loss": 1.2802, + "step": 724 + }, + { + "epoch": 0.34910316600457447, + "grad_norm": 0.7962654232978821, + "learning_rate": 0.00015138936740360207, + "loss": 0.8159, + "step": 725 + }, + { + "epoch": 0.3495846876128566, + "grad_norm": 0.5400972366333008, + "learning_rate": 0.00015125549026634585, + "loss": 0.3333, + "step": 726 + }, + { + "epoch": 0.3500662092211388, + "grad_norm": 0.683711051940918, + "learning_rate": 0.00015112148841339295, + "loss": 0.648, + "step": 727 + }, + { + "epoch": 0.350547730829421, + "grad_norm": 1.0952285528182983, + "learning_rate": 0.000150987362170799, + "loss": 0.8084, + "step": 728 + }, + { + "epoch": 0.35102925243770317, + "grad_norm": 1.6250587701797485, + "learning_rate": 0.00015085311186492206, + "loss": 0.8142, + "step": 729 + }, + { + "epoch": 0.3515107740459853, + "grad_norm": 1.0683465003967285, + "learning_rate": 0.00015071873782242223, + "loss": 0.6077, + "step": 730 + }, + { + "epoch": 0.3519922956542675, + "grad_norm": 1.0285414457321167, + "learning_rate": 0.0001505842403702606, + "loss": 0.7088, + "step": 731 + }, + { + "epoch": 0.35247381726254967, + "grad_norm": 0.675561785697937, + "learning_rate": 0.00015044961983569856, + "loss": 0.4079, + "step": 732 + }, + { + "epoch": 0.3529553388708318, + "grad_norm": 0.5926306247711182, + "learning_rate": 0.00015031487654629702, + "loss": 0.249, + "step": 733 + }, + { + "epoch": 0.353436860479114, + "grad_norm": 0.6946805119514465, + "learning_rate": 0.00015018001082991553, + "loss": 0.2278, + "step": 734 + }, + { + "epoch": 0.3539183820873962, + "grad_norm": 1.2133913040161133, + "learning_rate": 0.0001500450230147116, + "loss": 0.6548, + "step": 735 + }, + { + "epoch": 0.35439990369567836, + "grad_norm": 1.1130623817443848, + "learning_rate": 0.00014990991342913974, + "loss": 0.87, + "step": 736 + }, + { + "epoch": 0.3548814253039605, + "grad_norm": 1.3685212135314941, + "learning_rate": 0.00014977468240195084, + "loss": 0.8236, + "step": 737 + }, + { + "epoch": 0.3553629469122427, + "grad_norm": 1.4713952541351318, + "learning_rate": 0.0001496393302621912, + "loss": 1.1213, + "step": 738 + }, + { + "epoch": 0.35584446852052487, + "grad_norm": 0.9908416867256165, + "learning_rate": 0.00014950385733920188, + "loss": 0.5959, + "step": 739 + }, + { + "epoch": 0.35632599012880706, + "grad_norm": 1.0449920892715454, + "learning_rate": 0.00014936826396261783, + "loss": 0.735, + "step": 740 + }, + { + "epoch": 0.3568075117370892, + "grad_norm": 1.2540510892868042, + "learning_rate": 0.00014923255046236705, + "loss": 0.6025, + "step": 741 + }, + { + "epoch": 0.3572890333453714, + "grad_norm": 0.9532543420791626, + "learning_rate": 0.00014909671716866984, + "loss": 0.622, + "step": 742 + }, + { + "epoch": 0.35777055495365356, + "grad_norm": 1.1329494714736938, + "learning_rate": 0.00014896076441203802, + "loss": 1.3104, + "step": 743 + }, + { + "epoch": 0.3582520765619357, + "grad_norm": 0.6802012324333191, + "learning_rate": 0.000148824692523274, + "loss": 0.7314, + "step": 744 + }, + { + "epoch": 0.3587335981702179, + "grad_norm": 0.8869780898094177, + "learning_rate": 0.0001486885018334702, + "loss": 0.9224, + "step": 745 + }, + { + "epoch": 0.35921511977850007, + "grad_norm": 1.5476834774017334, + "learning_rate": 0.00014855219267400797, + "loss": 0.7926, + "step": 746 + }, + { + "epoch": 0.35969664138678226, + "grad_norm": 0.7869364023208618, + "learning_rate": 0.00014841576537655705, + "loss": 0.4486, + "step": 747 + }, + { + "epoch": 0.3601781629950644, + "grad_norm": 0.6606510281562805, + "learning_rate": 0.00014827922027307451, + "loss": 0.4667, + "step": 748 + }, + { + "epoch": 0.3606596846033466, + "grad_norm": 0.6294459700584412, + "learning_rate": 0.00014814255769580415, + "loss": 0.4268, + "step": 749 + }, + { + "epoch": 0.36114120621162876, + "grad_norm": 1.0464389324188232, + "learning_rate": 0.00014800577797727558, + "loss": 0.8922, + "step": 750 + }, + { + "epoch": 0.3616227278199109, + "grad_norm": 0.8546883463859558, + "learning_rate": 0.00014786888145030343, + "loss": 1.1779, + "step": 751 + }, + { + "epoch": 0.3621042494281931, + "grad_norm": 0.7854865789413452, + "learning_rate": 0.0001477318684479866, + "loss": 0.1934, + "step": 752 + }, + { + "epoch": 0.36258577103647527, + "grad_norm": 2.1393556594848633, + "learning_rate": 0.00014759473930370736, + "loss": 0.7357, + "step": 753 + }, + { + "epoch": 0.36306729264475746, + "grad_norm": 0.7535027861595154, + "learning_rate": 0.0001474574943511306, + "loss": 0.6884, + "step": 754 + }, + { + "epoch": 0.3635488142530396, + "grad_norm": 1.1644819974899292, + "learning_rate": 0.0001473201339242029, + "loss": 1.1809, + "step": 755 + }, + { + "epoch": 0.3640303358613218, + "grad_norm": 1.5013142824172974, + "learning_rate": 0.000147182658357152, + "loss": 0.6, + "step": 756 + }, + { + "epoch": 0.36451185746960396, + "grad_norm": 0.8900628685951233, + "learning_rate": 0.00014704506798448566, + "loss": 0.5416, + "step": 757 + }, + { + "epoch": 0.3649933790778861, + "grad_norm": 0.8392864465713501, + "learning_rate": 0.00014690736314099101, + "loss": 0.711, + "step": 758 + }, + { + "epoch": 0.3654749006861683, + "grad_norm": 0.5832529067993164, + "learning_rate": 0.00014676954416173373, + "loss": 0.5333, + "step": 759 + }, + { + "epoch": 0.36595642229445047, + "grad_norm": 1.227352261543274, + "learning_rate": 0.00014663161138205724, + "loss": 1.1198, + "step": 760 + }, + { + "epoch": 0.36643794390273265, + "grad_norm": 0.8971624970436096, + "learning_rate": 0.00014649356513758176, + "loss": 0.429, + "step": 761 + }, + { + "epoch": 0.3669194655110148, + "grad_norm": 0.860392153263092, + "learning_rate": 0.00014635540576420374, + "loss": 0.6527, + "step": 762 + }, + { + "epoch": 0.367400987119297, + "grad_norm": 0.34480923414230347, + "learning_rate": 0.0001462171335980948, + "loss": 0.3004, + "step": 763 + }, + { + "epoch": 0.36788250872757916, + "grad_norm": 1.0565292835235596, + "learning_rate": 0.00014607874897570105, + "loss": 0.3944, + "step": 764 + }, + { + "epoch": 0.36836403033586135, + "grad_norm": 0.7156208753585815, + "learning_rate": 0.0001459402522337422, + "loss": 0.1887, + "step": 765 + }, + { + "epoch": 0.3688455519441435, + "grad_norm": 1.2437664270401, + "learning_rate": 0.00014580164370921078, + "loss": 0.4478, + "step": 766 + }, + { + "epoch": 0.36932707355242567, + "grad_norm": 0.6799991726875305, + "learning_rate": 0.0001456629237393713, + "loss": 0.4307, + "step": 767 + }, + { + "epoch": 0.36980859516070785, + "grad_norm": 0.487223744392395, + "learning_rate": 0.00014552409266175952, + "loss": 0.1922, + "step": 768 + }, + { + "epoch": 0.37029011676899, + "grad_norm": 0.73237544298172, + "learning_rate": 0.00014538515081418142, + "loss": 0.6541, + "step": 769 + }, + { + "epoch": 0.3707716383772722, + "grad_norm": 1.2179851531982422, + "learning_rate": 0.00014524609853471264, + "loss": 0.8882, + "step": 770 + }, + { + "epoch": 0.37125315998555436, + "grad_norm": 1.2429336309432983, + "learning_rate": 0.00014510693616169741, + "loss": 0.6963, + "step": 771 + }, + { + "epoch": 0.37173468159383655, + "grad_norm": 0.764161229133606, + "learning_rate": 0.0001449676640337479, + "loss": 0.8483, + "step": 772 + }, + { + "epoch": 0.3722162032021187, + "grad_norm": 0.7411964535713196, + "learning_rate": 0.00014482828248974335, + "loss": 0.3896, + "step": 773 + }, + { + "epoch": 0.37269772481040087, + "grad_norm": 0.7775552868843079, + "learning_rate": 0.00014468879186882916, + "loss": 0.7667, + "step": 774 + }, + { + "epoch": 0.37317924641868305, + "grad_norm": 0.5666738748550415, + "learning_rate": 0.00014454919251041622, + "loss": 0.3198, + "step": 775 + }, + { + "epoch": 0.3736607680269652, + "grad_norm": 1.0107135772705078, + "learning_rate": 0.00014440948475418, + "loss": 0.8955, + "step": 776 + }, + { + "epoch": 0.37414228963524737, + "grad_norm": 0.7883855104446411, + "learning_rate": 0.00014426966894005966, + "loss": 0.5721, + "step": 777 + }, + { + "epoch": 0.37462381124352956, + "grad_norm": 0.8659805059432983, + "learning_rate": 0.0001441297454082573, + "loss": 1.0827, + "step": 778 + }, + { + "epoch": 0.37510533285181175, + "grad_norm": 0.8902775049209595, + "learning_rate": 0.00014398971449923722, + "loss": 0.3986, + "step": 779 + }, + { + "epoch": 0.3755868544600939, + "grad_norm": 0.9566343426704407, + "learning_rate": 0.00014384957655372483, + "loss": 0.622, + "step": 780 + }, + { + "epoch": 0.37606837606837606, + "grad_norm": 0.7155655026435852, + "learning_rate": 0.00014370933191270617, + "loss": 0.5707, + "step": 781 + }, + { + "epoch": 0.37654989767665825, + "grad_norm": 1.2331191301345825, + "learning_rate": 0.0001435689809174267, + "loss": 0.8049, + "step": 782 + }, + { + "epoch": 0.37703141928494044, + "grad_norm": 1.137040615081787, + "learning_rate": 0.0001434285239093908, + "loss": 1.2335, + "step": 783 + }, + { + "epoch": 0.37751294089322257, + "grad_norm": 1.0161875486373901, + "learning_rate": 0.00014328796123036071, + "loss": 0.665, + "step": 784 + }, + { + "epoch": 0.37799446250150476, + "grad_norm": 0.5084791779518127, + "learning_rate": 0.0001431472932223559, + "loss": 0.3809, + "step": 785 + }, + { + "epoch": 0.37847598410978694, + "grad_norm": 0.9150707721710205, + "learning_rate": 0.00014300652022765207, + "loss": 0.4742, + "step": 786 + }, + { + "epoch": 0.3789575057180691, + "grad_norm": 0.849323034286499, + "learning_rate": 0.00014286564258878033, + "loss": 0.7684, + "step": 787 + }, + { + "epoch": 0.37943902732635126, + "grad_norm": 1.380608320236206, + "learning_rate": 0.00014272466064852644, + "loss": 0.522, + "step": 788 + }, + { + "epoch": 0.37992054893463345, + "grad_norm": 1.035540223121643, + "learning_rate": 0.00014258357474993, + "loss": 1.0147, + "step": 789 + }, + { + "epoch": 0.38040207054291564, + "grad_norm": 1.7030549049377441, + "learning_rate": 0.0001424423852362835, + "loss": 0.541, + "step": 790 + }, + { + "epoch": 0.38088359215119777, + "grad_norm": 0.6918346285820007, + "learning_rate": 0.00014230109245113158, + "loss": 0.7923, + "step": 791 + }, + { + "epoch": 0.38136511375947996, + "grad_norm": 0.7446008324623108, + "learning_rate": 0.00014215969673827018, + "loss": 0.3233, + "step": 792 + }, + { + "epoch": 0.38184663536776214, + "grad_norm": 0.4108048975467682, + "learning_rate": 0.00014201819844174564, + "loss": 0.1844, + "step": 793 + }, + { + "epoch": 0.3823281569760443, + "grad_norm": 1.1727381944656372, + "learning_rate": 0.0001418765979058539, + "loss": 0.8246, + "step": 794 + }, + { + "epoch": 0.38280967858432646, + "grad_norm": 0.8327621221542358, + "learning_rate": 0.00014173489547513973, + "loss": 0.9647, + "step": 795 + }, + { + "epoch": 0.38329120019260865, + "grad_norm": 0.9883455634117126, + "learning_rate": 0.00014159309149439582, + "loss": 0.6689, + "step": 796 + }, + { + "epoch": 0.38377272180089084, + "grad_norm": 0.9439548254013062, + "learning_rate": 0.00014145118630866187, + "loss": 0.3843, + "step": 797 + }, + { + "epoch": 0.38425424340917297, + "grad_norm": 1.062584638595581, + "learning_rate": 0.000141309180263224, + "loss": 0.2714, + "step": 798 + }, + { + "epoch": 0.38473576501745516, + "grad_norm": 1.0796810388565063, + "learning_rate": 0.0001411670737036135, + "loss": 0.4017, + "step": 799 + }, + { + "epoch": 0.38521728662573734, + "grad_norm": 0.6542953252792358, + "learning_rate": 0.0001410248669756065, + "loss": 0.3069, + "step": 800 + }, + { + "epoch": 0.3856988082340195, + "grad_norm": 1.2995156049728394, + "learning_rate": 0.00014088256042522264, + "loss": 0.7217, + "step": 801 + }, + { + "epoch": 0.38618032984230166, + "grad_norm": 0.58650803565979, + "learning_rate": 0.00014074015439872458, + "loss": 0.4809, + "step": 802 + }, + { + "epoch": 0.38666185145058385, + "grad_norm": 0.7780939340591431, + "learning_rate": 0.00014059764924261703, + "loss": 0.3629, + "step": 803 + }, + { + "epoch": 0.38714337305886604, + "grad_norm": 1.2356970310211182, + "learning_rate": 0.00014045504530364584, + "loss": 0.3341, + "step": 804 + }, + { + "epoch": 0.38762489466714817, + "grad_norm": 1.130676507949829, + "learning_rate": 0.00014031234292879725, + "loss": 0.5205, + "step": 805 + }, + { + "epoch": 0.38810641627543035, + "grad_norm": 0.8669336438179016, + "learning_rate": 0.00014016954246529696, + "loss": 0.7471, + "step": 806 + }, + { + "epoch": 0.38858793788371254, + "grad_norm": 1.1832457780838013, + "learning_rate": 0.00014002664426060942, + "loss": 0.4838, + "step": 807 + }, + { + "epoch": 0.38906945949199473, + "grad_norm": 0.7935713529586792, + "learning_rate": 0.00013988364866243693, + "loss": 0.883, + "step": 808 + }, + { + "epoch": 0.38955098110027686, + "grad_norm": 0.6437901258468628, + "learning_rate": 0.00013974055601871868, + "loss": 0.4664, + "step": 809 + }, + { + "epoch": 0.39003250270855905, + "grad_norm": 1.122780680656433, + "learning_rate": 0.00013959736667762998, + "loss": 0.7229, + "step": 810 + }, + { + "epoch": 0.39051402431684124, + "grad_norm": 0.9974272847175598, + "learning_rate": 0.00013945408098758156, + "loss": 0.5652, + "step": 811 + }, + { + "epoch": 0.39099554592512337, + "grad_norm": 0.5831164121627808, + "learning_rate": 0.0001393106992972184, + "loss": 0.4848, + "step": 812 + }, + { + "epoch": 0.39147706753340555, + "grad_norm": 0.643200159072876, + "learning_rate": 0.00013916722195541926, + "loss": 0.3277, + "step": 813 + }, + { + "epoch": 0.39195858914168774, + "grad_norm": 1.4757206439971924, + "learning_rate": 0.00013902364931129557, + "loss": 0.6434, + "step": 814 + }, + { + "epoch": 0.39244011074996993, + "grad_norm": 2.269386053085327, + "learning_rate": 0.00013887998171419058, + "loss": 0.9765, + "step": 815 + }, + { + "epoch": 0.39292163235825206, + "grad_norm": 1.381133794784546, + "learning_rate": 0.00013873621951367862, + "loss": 1.3941, + "step": 816 + }, + { + "epoch": 0.39340315396653425, + "grad_norm": 0.5003638863563538, + "learning_rate": 0.00013859236305956425, + "loss": 0.5137, + "step": 817 + }, + { + "epoch": 0.39388467557481643, + "grad_norm": 0.9194219708442688, + "learning_rate": 0.00013844841270188132, + "loss": 0.7313, + "step": 818 + }, + { + "epoch": 0.39436619718309857, + "grad_norm": 0.6109256148338318, + "learning_rate": 0.00013830436879089228, + "loss": 0.7035, + "step": 819 + }, + { + "epoch": 0.39484771879138075, + "grad_norm": 0.5414918065071106, + "learning_rate": 0.00013816023167708704, + "loss": 0.241, + "step": 820 + }, + { + "epoch": 0.39532924039966294, + "grad_norm": 0.564078688621521, + "learning_rate": 0.00013801600171118244, + "loss": 0.2047, + "step": 821 + }, + { + "epoch": 0.3958107620079451, + "grad_norm": 1.235252857208252, + "learning_rate": 0.00013787167924412112, + "loss": 0.7474, + "step": 822 + }, + { + "epoch": 0.39629228361622726, + "grad_norm": 0.5138280987739563, + "learning_rate": 0.0001377272646270709, + "loss": 0.2388, + "step": 823 + }, + { + "epoch": 0.39677380522450945, + "grad_norm": 0.6855552196502686, + "learning_rate": 0.00013758275821142382, + "loss": 0.3218, + "step": 824 + }, + { + "epoch": 0.39725532683279163, + "grad_norm": 1.0833673477172852, + "learning_rate": 0.00013743816034879523, + "loss": 0.3818, + "step": 825 + }, + { + "epoch": 0.3977368484410738, + "grad_norm": 0.8923762440681458, + "learning_rate": 0.000137293471391023, + "loss": 0.8322, + "step": 826 + }, + { + "epoch": 0.39821837004935595, + "grad_norm": 1.838834524154663, + "learning_rate": 0.00013714869169016667, + "loss": 0.3636, + "step": 827 + }, + { + "epoch": 0.39869989165763814, + "grad_norm": 0.7193182706832886, + "learning_rate": 0.00013700382159850656, + "loss": 0.3181, + "step": 828 + }, + { + "epoch": 0.3991814132659203, + "grad_norm": 0.867734432220459, + "learning_rate": 0.00013685886146854297, + "loss": 0.3646, + "step": 829 + }, + { + "epoch": 0.39966293487420246, + "grad_norm": 1.1769030094146729, + "learning_rate": 0.00013671381165299525, + "loss": 0.6471, + "step": 830 + }, + { + "epoch": 0.40014445648248465, + "grad_norm": 2.466710329055786, + "learning_rate": 0.00013656867250480098, + "loss": 0.5789, + "step": 831 + }, + { + "epoch": 0.40062597809076683, + "grad_norm": 0.5317124724388123, + "learning_rate": 0.00013642344437711512, + "loss": 0.2964, + "step": 832 + }, + { + "epoch": 0.401107499699049, + "grad_norm": 1.1287859678268433, + "learning_rate": 0.00013627812762330912, + "loss": 0.8241, + "step": 833 + }, + { + "epoch": 0.40158902130733115, + "grad_norm": 0.7147384285926819, + "learning_rate": 0.00013613272259697007, + "loss": 0.5656, + "step": 834 + }, + { + "epoch": 0.40207054291561334, + "grad_norm": 1.6891028881072998, + "learning_rate": 0.00013598722965189986, + "loss": 0.9899, + "step": 835 + }, + { + "epoch": 0.4025520645238955, + "grad_norm": 0.943759560585022, + "learning_rate": 0.0001358416491421143, + "loss": 0.5832, + "step": 836 + }, + { + "epoch": 0.40303358613217766, + "grad_norm": 0.8546479344367981, + "learning_rate": 0.00013569598142184225, + "loss": 0.7991, + "step": 837 + }, + { + "epoch": 0.40351510774045984, + "grad_norm": 0.9846085906028748, + "learning_rate": 0.00013555022684552483, + "loss": 0.9692, + "step": 838 + }, + { + "epoch": 0.40399662934874203, + "grad_norm": 1.969257116317749, + "learning_rate": 0.00013540438576781441, + "loss": 0.8052, + "step": 839 + }, + { + "epoch": 0.4044781509570242, + "grad_norm": 1.115907907485962, + "learning_rate": 0.0001352584585435739, + "loss": 1.2649, + "step": 840 + }, + { + "epoch": 0.40495967256530635, + "grad_norm": 0.6951058506965637, + "learning_rate": 0.00013511244552787583, + "loss": 0.2709, + "step": 841 + }, + { + "epoch": 0.40544119417358854, + "grad_norm": 1.5416226387023926, + "learning_rate": 0.00013496634707600147, + "loss": 0.7605, + "step": 842 + }, + { + "epoch": 0.4059227157818707, + "grad_norm": 1.0881750583648682, + "learning_rate": 0.0001348201635434399, + "loss": 0.5618, + "step": 843 + }, + { + "epoch": 0.40640423739015286, + "grad_norm": 1.3643512725830078, + "learning_rate": 0.0001346738952858873, + "loss": 1.0733, + "step": 844 + }, + { + "epoch": 0.40688575899843504, + "grad_norm": 0.7877455353736877, + "learning_rate": 0.000134527542659246, + "loss": 1.6297, + "step": 845 + }, + { + "epoch": 0.40736728060671723, + "grad_norm": 0.5141893029212952, + "learning_rate": 0.00013438110601962362, + "loss": 0.1867, + "step": 846 + }, + { + "epoch": 0.4078488022149994, + "grad_norm": 0.5263735055923462, + "learning_rate": 0.00013423458572333214, + "loss": 0.7638, + "step": 847 + }, + { + "epoch": 0.40833032382328155, + "grad_norm": 0.5297187566757202, + "learning_rate": 0.0001340879821268872, + "loss": 0.4511, + "step": 848 + }, + { + "epoch": 0.40881184543156374, + "grad_norm": 1.2766536474227905, + "learning_rate": 0.000133941295587007, + "loss": 0.4998, + "step": 849 + }, + { + "epoch": 0.4092933670398459, + "grad_norm": 0.8083428740501404, + "learning_rate": 0.00013379452646061164, + "loss": 0.1786, + "step": 850 + }, + { + "epoch": 0.4097748886481281, + "grad_norm": 0.8990474343299866, + "learning_rate": 0.0001336476751048222, + "loss": 0.99, + "step": 851 + }, + { + "epoch": 0.41025641025641024, + "grad_norm": 0.793449878692627, + "learning_rate": 0.00013350074187695979, + "loss": 0.5605, + "step": 852 + }, + { + "epoch": 0.41073793186469243, + "grad_norm": 0.7886202931404114, + "learning_rate": 0.00013335372713454467, + "loss": 0.294, + "step": 853 + }, + { + "epoch": 0.4112194534729746, + "grad_norm": 0.8802816867828369, + "learning_rate": 0.0001332066312352956, + "loss": 0.6326, + "step": 854 + }, + { + "epoch": 0.41170097508125675, + "grad_norm": 0.8170877695083618, + "learning_rate": 0.00013305945453712868, + "loss": 0.6033, + "step": 855 + }, + { + "epoch": 0.41218249668953894, + "grad_norm": 0.5515619516372681, + "learning_rate": 0.0001329121973981567, + "loss": 0.5479, + "step": 856 + }, + { + "epoch": 0.4126640182978211, + "grad_norm": 1.2902063131332397, + "learning_rate": 0.00013276486017668807, + "loss": 0.6442, + "step": 857 + }, + { + "epoch": 0.4131455399061033, + "grad_norm": 0.8257691860198975, + "learning_rate": 0.0001326174432312262, + "loss": 1.1458, + "step": 858 + }, + { + "epoch": 0.41362706151438544, + "grad_norm": 0.5979056358337402, + "learning_rate": 0.00013246994692046836, + "loss": 0.5462, + "step": 859 + }, + { + "epoch": 0.41410858312266763, + "grad_norm": 0.6814475655555725, + "learning_rate": 0.000132322371603305, + "loss": 0.6534, + "step": 860 + }, + { + "epoch": 0.4145901047309498, + "grad_norm": 0.6793029308319092, + "learning_rate": 0.0001321747176388188, + "loss": 0.2851, + "step": 861 + }, + { + "epoch": 0.41507162633923195, + "grad_norm": 0.5598545670509338, + "learning_rate": 0.00013202698538628376, + "loss": 0.7152, + "step": 862 + }, + { + "epoch": 0.41555314794751413, + "grad_norm": 0.7994054555892944, + "learning_rate": 0.00013187917520516448, + "loss": 0.6246, + "step": 863 + }, + { + "epoch": 0.4160346695557963, + "grad_norm": 0.9414712190628052, + "learning_rate": 0.00013173128745511508, + "loss": 1.0983, + "step": 864 + }, + { + "epoch": 0.4165161911640785, + "grad_norm": 0.8112841844558716, + "learning_rate": 0.0001315833224959784, + "loss": 0.5866, + "step": 865 + }, + { + "epoch": 0.41699771277236064, + "grad_norm": 0.9308454990386963, + "learning_rate": 0.00013143528068778525, + "loss": 1.0087, + "step": 866 + }, + { + "epoch": 0.4174792343806428, + "grad_norm": 0.592212438583374, + "learning_rate": 0.00013128716239075338, + "loss": 0.3319, + "step": 867 + }, + { + "epoch": 0.417960755988925, + "grad_norm": 0.8218088150024414, + "learning_rate": 0.00013113896796528664, + "loss": 0.8615, + "step": 868 + }, + { + "epoch": 0.4184422775972072, + "grad_norm": 0.7835381627082825, + "learning_rate": 0.00013099069777197412, + "loss": 0.8362, + "step": 869 + }, + { + "epoch": 0.41892379920548933, + "grad_norm": 0.8324255347251892, + "learning_rate": 0.0001308423521715893, + "loss": 0.9373, + "step": 870 + }, + { + "epoch": 0.4194053208137715, + "grad_norm": 0.7907895445823669, + "learning_rate": 0.00013069393152508906, + "loss": 0.5857, + "step": 871 + }, + { + "epoch": 0.4198868424220537, + "grad_norm": 0.7639227509498596, + "learning_rate": 0.00013054543619361303, + "loss": 0.3794, + "step": 872 + }, + { + "epoch": 0.42036836403033584, + "grad_norm": 1.056784749031067, + "learning_rate": 0.0001303968665384824, + "loss": 1.1836, + "step": 873 + }, + { + "epoch": 0.420849885638618, + "grad_norm": 0.6656973958015442, + "learning_rate": 0.00013024822292119934, + "loss": 0.2145, + "step": 874 + }, + { + "epoch": 0.4213314072469002, + "grad_norm": 0.7051798105239868, + "learning_rate": 0.0001300995057034459, + "loss": 0.5826, + "step": 875 + }, + { + "epoch": 0.4218129288551824, + "grad_norm": 0.9368335604667664, + "learning_rate": 0.00012995071524708325, + "loss": 0.5162, + "step": 876 + }, + { + "epoch": 0.42229445046346453, + "grad_norm": 0.608950138092041, + "learning_rate": 0.00012980185191415074, + "loss": 0.2055, + "step": 877 + }, + { + "epoch": 0.4227759720717467, + "grad_norm": 0.8486785292625427, + "learning_rate": 0.0001296529160668651, + "loss": 0.1395, + "step": 878 + }, + { + "epoch": 0.4232574936800289, + "grad_norm": 0.6813993453979492, + "learning_rate": 0.00012950390806761944, + "loss": 0.8436, + "step": 879 + }, + { + "epoch": 0.42373901528831104, + "grad_norm": 0.5258020758628845, + "learning_rate": 0.0001293548282789825, + "loss": 0.1877, + "step": 880 + }, + { + "epoch": 0.4242205368965932, + "grad_norm": 0.9190642833709717, + "learning_rate": 0.00012920567706369758, + "loss": 1.1206, + "step": 881 + }, + { + "epoch": 0.4247020585048754, + "grad_norm": 0.545403003692627, + "learning_rate": 0.00012905645478468192, + "loss": 0.3437, + "step": 882 + }, + { + "epoch": 0.4251835801131576, + "grad_norm": 0.7194792032241821, + "learning_rate": 0.00012890716180502564, + "loss": 0.3147, + "step": 883 + }, + { + "epoch": 0.42566510172143973, + "grad_norm": 0.661845862865448, + "learning_rate": 0.00012875779848799078, + "loss": 1.0886, + "step": 884 + }, + { + "epoch": 0.4261466233297219, + "grad_norm": 1.2993658781051636, + "learning_rate": 0.00012860836519701063, + "loss": 0.9139, + "step": 885 + }, + { + "epoch": 0.4266281449380041, + "grad_norm": 0.7186025381088257, + "learning_rate": 0.00012845886229568873, + "loss": 0.5141, + "step": 886 + }, + { + "epoch": 0.42710966654628624, + "grad_norm": 0.9918057322502136, + "learning_rate": 0.00012830929014779797, + "loss": 1.1286, + "step": 887 + }, + { + "epoch": 0.4275911881545684, + "grad_norm": 1.213967204093933, + "learning_rate": 0.0001281596491172797, + "loss": 0.3991, + "step": 888 + }, + { + "epoch": 0.4280727097628506, + "grad_norm": 1.1254137754440308, + "learning_rate": 0.00012800993956824303, + "loss": 1.0619, + "step": 889 + }, + { + "epoch": 0.4285542313711328, + "grad_norm": 0.8088738322257996, + "learning_rate": 0.00012786016186496358, + "loss": 0.4409, + "step": 890 + }, + { + "epoch": 0.42903575297941493, + "grad_norm": 0.8131324648857117, + "learning_rate": 0.000127710316371883, + "loss": 0.4324, + "step": 891 + }, + { + "epoch": 0.4295172745876971, + "grad_norm": 0.5402470231056213, + "learning_rate": 0.0001275604034536077, + "loss": 0.2812, + "step": 892 + }, + { + "epoch": 0.4299987961959793, + "grad_norm": 0.4589332938194275, + "learning_rate": 0.0001274104234749083, + "loss": 0.2145, + "step": 893 + }, + { + "epoch": 0.4304803178042615, + "grad_norm": 1.0071089267730713, + "learning_rate": 0.00012726037680071853, + "loss": 0.6196, + "step": 894 + }, + { + "epoch": 0.4309618394125436, + "grad_norm": 0.6019487380981445, + "learning_rate": 0.00012711026379613434, + "loss": 0.4071, + "step": 895 + }, + { + "epoch": 0.4314433610208258, + "grad_norm": 0.4118281602859497, + "learning_rate": 0.00012696008482641325, + "loss": 0.3439, + "step": 896 + }, + { + "epoch": 0.431924882629108, + "grad_norm": 0.6263596415519714, + "learning_rate": 0.00012680984025697313, + "loss": 0.3853, + "step": 897 + }, + { + "epoch": 0.43240640423739013, + "grad_norm": 1.7972133159637451, + "learning_rate": 0.00012665953045339152, + "loss": 0.6076, + "step": 898 + }, + { + "epoch": 0.4328879258456723, + "grad_norm": 0.615397572517395, + "learning_rate": 0.0001265091557814047, + "loss": 0.6319, + "step": 899 + }, + { + "epoch": 0.4333694474539545, + "grad_norm": 1.5493266582489014, + "learning_rate": 0.00012635871660690676, + "loss": 0.4057, + "step": 900 + }, + { + "epoch": 0.4338509690622367, + "grad_norm": 0.7660396099090576, + "learning_rate": 0.0001262082132959488, + "loss": 0.5197, + "step": 901 + }, + { + "epoch": 0.4343324906705188, + "grad_norm": 0.9497624635696411, + "learning_rate": 0.00012605764621473792, + "loss": 0.6527, + "step": 902 + }, + { + "epoch": 0.434814012278801, + "grad_norm": 0.5582836866378784, + "learning_rate": 0.00012590701572963642, + "loss": 0.5082, + "step": 903 + }, + { + "epoch": 0.4352955338870832, + "grad_norm": 1.3667628765106201, + "learning_rate": 0.00012575632220716078, + "loss": 0.8073, + "step": 904 + }, + { + "epoch": 0.43577705549536533, + "grad_norm": 0.6490867733955383, + "learning_rate": 0.000125605566013981, + "loss": 0.7873, + "step": 905 + }, + { + "epoch": 0.4362585771036475, + "grad_norm": 0.7977167963981628, + "learning_rate": 0.00012545474751691953, + "loss": 0.8709, + "step": 906 + }, + { + "epoch": 0.4367400987119297, + "grad_norm": 0.9856142401695251, + "learning_rate": 0.00012530386708295036, + "loss": 0.9742, + "step": 907 + }, + { + "epoch": 0.4372216203202119, + "grad_norm": 0.665690541267395, + "learning_rate": 0.00012515292507919829, + "loss": 0.4871, + "step": 908 + }, + { + "epoch": 0.437703141928494, + "grad_norm": 0.4781898558139801, + "learning_rate": 0.0001250019218729378, + "loss": 0.4098, + "step": 909 + }, + { + "epoch": 0.4381846635367762, + "grad_norm": 0.7730495929718018, + "learning_rate": 0.00012485085783159238, + "loss": 0.2263, + "step": 910 + }, + { + "epoch": 0.4386661851450584, + "grad_norm": 0.9093911647796631, + "learning_rate": 0.00012469973332273354, + "loss": 0.4428, + "step": 911 + }, + { + "epoch": 0.4391477067533406, + "grad_norm": 0.7920442819595337, + "learning_rate": 0.00012454854871407994, + "loss": 0.4661, + "step": 912 + }, + { + "epoch": 0.4396292283616227, + "grad_norm": 0.8402988910675049, + "learning_rate": 0.00012439730437349635, + "loss": 0.4356, + "step": 913 + }, + { + "epoch": 0.4401107499699049, + "grad_norm": 0.9629238247871399, + "learning_rate": 0.00012424600066899302, + "loss": 0.6917, + "step": 914 + }, + { + "epoch": 0.4405922715781871, + "grad_norm": 1.3128741979599, + "learning_rate": 0.00012409463796872464, + "loss": 0.6547, + "step": 915 + }, + { + "epoch": 0.4410737931864692, + "grad_norm": 0.8441032767295837, + "learning_rate": 0.0001239432166409893, + "loss": 0.8729, + "step": 916 + }, + { + "epoch": 0.4415553147947514, + "grad_norm": 0.6667619347572327, + "learning_rate": 0.00012379173705422795, + "loss": 0.5423, + "step": 917 + }, + { + "epoch": 0.4420368364030336, + "grad_norm": 0.8423894643783569, + "learning_rate": 0.00012364019957702315, + "loss": 0.7282, + "step": 918 + }, + { + "epoch": 0.4425183580113158, + "grad_norm": 2.0424094200134277, + "learning_rate": 0.00012348860457809838, + "loss": 1.1008, + "step": 919 + }, + { + "epoch": 0.4429998796195979, + "grad_norm": 0.6044926047325134, + "learning_rate": 0.00012333695242631705, + "loss": 0.8077, + "step": 920 + }, + { + "epoch": 0.4434814012278801, + "grad_norm": 0.8025707602500916, + "learning_rate": 0.0001231852434906817, + "loss": 0.7705, + "step": 921 + }, + { + "epoch": 0.4439629228361623, + "grad_norm": 0.8117222785949707, + "learning_rate": 0.00012303347814033292, + "loss": 0.4655, + "step": 922 + }, + { + "epoch": 0.4444444444444444, + "grad_norm": 0.715910017490387, + "learning_rate": 0.0001228816567445487, + "loss": 0.6408, + "step": 923 + }, + { + "epoch": 0.4449259660527266, + "grad_norm": 0.7318637371063232, + "learning_rate": 0.0001227297796727433, + "loss": 0.4922, + "step": 924 + }, + { + "epoch": 0.4454074876610088, + "grad_norm": 0.8235260844230652, + "learning_rate": 0.00012257784729446656, + "loss": 0.7587, + "step": 925 + }, + { + "epoch": 0.445889009269291, + "grad_norm": 0.7627338171005249, + "learning_rate": 0.00012242585997940275, + "loss": 0.391, + "step": 926 + }, + { + "epoch": 0.4463705308775731, + "grad_norm": 0.5861409306526184, + "learning_rate": 0.0001222738180973699, + "loss": 0.5866, + "step": 927 + }, + { + "epoch": 0.4468520524858553, + "grad_norm": 0.6915172934532166, + "learning_rate": 0.00012212172201831885, + "loss": 0.3472, + "step": 928 + }, + { + "epoch": 0.4473335740941375, + "grad_norm": 0.6583700776100159, + "learning_rate": 0.00012196957211233222, + "loss": 0.5164, + "step": 929 + }, + { + "epoch": 0.4478150957024196, + "grad_norm": 0.8904280662536621, + "learning_rate": 0.00012181736874962371, + "loss": 1.3786, + "step": 930 + }, + { + "epoch": 0.4482966173107018, + "grad_norm": 0.9074341058731079, + "learning_rate": 0.00012166511230053696, + "loss": 0.5478, + "step": 931 + }, + { + "epoch": 0.448778138918984, + "grad_norm": 0.6648260354995728, + "learning_rate": 0.00012151280313554486, + "loss": 1.1953, + "step": 932 + }, + { + "epoch": 0.4492596605272662, + "grad_norm": 1.5189319849014282, + "learning_rate": 0.00012136044162524858, + "loss": 1.042, + "step": 933 + }, + { + "epoch": 0.4497411821355483, + "grad_norm": 0.5307234525680542, + "learning_rate": 0.00012120802814037663, + "loss": 0.3387, + "step": 934 + }, + { + "epoch": 0.4502227037438305, + "grad_norm": 0.6986709237098694, + "learning_rate": 0.00012105556305178399, + "loss": 0.3857, + "step": 935 + }, + { + "epoch": 0.4507042253521127, + "grad_norm": 0.5282310843467712, + "learning_rate": 0.00012090304673045123, + "loss": 0.3448, + "step": 936 + }, + { + "epoch": 0.4511857469603949, + "grad_norm": 1.0674173831939697, + "learning_rate": 0.00012075047954748353, + "loss": 0.9065, + "step": 937 + }, + { + "epoch": 0.451667268568677, + "grad_norm": 1.6909977197647095, + "learning_rate": 0.00012059786187410984, + "loss": 0.7831, + "step": 938 + }, + { + "epoch": 0.4521487901769592, + "grad_norm": 1.0591840744018555, + "learning_rate": 0.000120445194081682, + "loss": 0.5693, + "step": 939 + }, + { + "epoch": 0.4526303117852414, + "grad_norm": 0.5961440801620483, + "learning_rate": 0.00012029247654167379, + "loss": 0.2809, + "step": 940 + }, + { + "epoch": 0.4531118333935235, + "grad_norm": 0.8247968554496765, + "learning_rate": 0.00012013970962568002, + "loss": 0.4512, + "step": 941 + }, + { + "epoch": 0.4535933550018057, + "grad_norm": 1.0450423955917358, + "learning_rate": 0.00011998689370541562, + "loss": 0.8942, + "step": 942 + }, + { + "epoch": 0.4540748766100879, + "grad_norm": 1.201209545135498, + "learning_rate": 0.00011983402915271478, + "loss": 0.4332, + "step": 943 + }, + { + "epoch": 0.4545563982183701, + "grad_norm": 0.8509264588356018, + "learning_rate": 0.00011968111633953007, + "loss": 0.3682, + "step": 944 + }, + { + "epoch": 0.4550379198266522, + "grad_norm": 0.7999116778373718, + "learning_rate": 0.0001195281556379314, + "loss": 0.5891, + "step": 945 + }, + { + "epoch": 0.4555194414349344, + "grad_norm": 0.8322030305862427, + "learning_rate": 0.0001193751474201053, + "loss": 0.2077, + "step": 946 + }, + { + "epoch": 0.4560009630432166, + "grad_norm": 1.6226719617843628, + "learning_rate": 0.00011922209205835382, + "loss": 0.2934, + "step": 947 + }, + { + "epoch": 0.4564824846514987, + "grad_norm": 1.2562118768692017, + "learning_rate": 0.0001190689899250938, + "loss": 0.7436, + "step": 948 + }, + { + "epoch": 0.4569640062597809, + "grad_norm": 0.7772266864776611, + "learning_rate": 0.00011891584139285582, + "loss": 0.8018, + "step": 949 + }, + { + "epoch": 0.4574455278680631, + "grad_norm": 0.8317371606826782, + "learning_rate": 0.00011876264683428344, + "loss": 0.4488, + "step": 950 + }, + { + "epoch": 0.45792704947634527, + "grad_norm": 0.3852919042110443, + "learning_rate": 0.00011860940662213211, + "loss": 0.1071, + "step": 951 + }, + { + "epoch": 0.4584085710846274, + "grad_norm": 1.2257070541381836, + "learning_rate": 0.00011845612112926843, + "loss": 0.8347, + "step": 952 + }, + { + "epoch": 0.4588900926929096, + "grad_norm": 0.7787636518478394, + "learning_rate": 0.00011830279072866921, + "loss": 0.6742, + "step": 953 + }, + { + "epoch": 0.4593716143011918, + "grad_norm": 1.6153544187545776, + "learning_rate": 0.00011814941579342044, + "loss": 0.432, + "step": 954 + }, + { + "epoch": 0.45985313590947396, + "grad_norm": 0.6438875198364258, + "learning_rate": 0.00011799599669671654, + "loss": 0.2599, + "step": 955 + }, + { + "epoch": 0.4603346575177561, + "grad_norm": 1.2926790714263916, + "learning_rate": 0.00011784253381185937, + "loss": 0.811, + "step": 956 + }, + { + "epoch": 0.4608161791260383, + "grad_norm": 0.5644267797470093, + "learning_rate": 0.0001176890275122573, + "loss": 0.4718, + "step": 957 + }, + { + "epoch": 0.46129770073432047, + "grad_norm": 0.6094907522201538, + "learning_rate": 0.0001175354781714244, + "loss": 0.216, + "step": 958 + }, + { + "epoch": 0.4617792223426026, + "grad_norm": 0.8137814402580261, + "learning_rate": 0.0001173818861629794, + "loss": 0.9217, + "step": 959 + }, + { + "epoch": 0.4622607439508848, + "grad_norm": 0.5632073879241943, + "learning_rate": 0.00011722825186064494, + "loss": 0.4477, + "step": 960 + }, + { + "epoch": 0.462742265559167, + "grad_norm": 0.6808025240898132, + "learning_rate": 0.00011707457563824646, + "loss": 0.3814, + "step": 961 + }, + { + "epoch": 0.46322378716744916, + "grad_norm": 0.7273874878883362, + "learning_rate": 0.00011692085786971149, + "loss": 0.3307, + "step": 962 + }, + { + "epoch": 0.4637053087757313, + "grad_norm": 1.1415742635726929, + "learning_rate": 0.00011676709892906858, + "loss": 0.3101, + "step": 963 + }, + { + "epoch": 0.4641868303840135, + "grad_norm": 0.8693737983703613, + "learning_rate": 0.00011661329919044656, + "loss": 0.7078, + "step": 964 + }, + { + "epoch": 0.46466835199229567, + "grad_norm": 0.4996722638607025, + "learning_rate": 0.00011645945902807341, + "loss": 0.1454, + "step": 965 + }, + { + "epoch": 0.4651498736005778, + "grad_norm": 0.9881248474121094, + "learning_rate": 0.00011630557881627553, + "loss": 0.7298, + "step": 966 + }, + { + "epoch": 0.46563139520886, + "grad_norm": 0.7234969139099121, + "learning_rate": 0.0001161516589294768, + "loss": 0.7709, + "step": 967 + }, + { + "epoch": 0.4661129168171422, + "grad_norm": 1.0357717275619507, + "learning_rate": 0.00011599769974219757, + "loss": 0.5054, + "step": 968 + }, + { + "epoch": 0.46659443842542436, + "grad_norm": 1.1411643028259277, + "learning_rate": 0.0001158437016290539, + "loss": 1.0258, + "step": 969 + }, + { + "epoch": 0.4670759600337065, + "grad_norm": 0.5702272653579712, + "learning_rate": 0.00011568966496475649, + "loss": 0.3137, + "step": 970 + }, + { + "epoch": 0.4675574816419887, + "grad_norm": 0.838179886341095, + "learning_rate": 0.00011553559012410984, + "loss": 0.4455, + "step": 971 + }, + { + "epoch": 0.46803900325027087, + "grad_norm": 0.9521105885505676, + "learning_rate": 0.00011538147748201138, + "loss": 0.3627, + "step": 972 + }, + { + "epoch": 0.468520524858553, + "grad_norm": 0.6624534130096436, + "learning_rate": 0.00011522732741345053, + "loss": 0.6825, + "step": 973 + }, + { + "epoch": 0.4690020464668352, + "grad_norm": 0.5236545205116272, + "learning_rate": 0.00011507314029350776, + "loss": 0.3712, + "step": 974 + }, + { + "epoch": 0.4694835680751174, + "grad_norm": 0.49228930473327637, + "learning_rate": 0.00011491891649735366, + "loss": 0.5079, + "step": 975 + }, + { + "epoch": 0.46996508968339956, + "grad_norm": 1.3148273229599, + "learning_rate": 0.00011476465640024814, + "loss": 1.0293, + "step": 976 + }, + { + "epoch": 0.4704466112916817, + "grad_norm": 0.534582257270813, + "learning_rate": 0.00011461036037753934, + "loss": 0.2893, + "step": 977 + }, + { + "epoch": 0.4709281328999639, + "grad_norm": 0.7397232055664062, + "learning_rate": 0.00011445602880466288, + "loss": 0.2611, + "step": 978 + }, + { + "epoch": 0.47140965450824607, + "grad_norm": 0.5304082036018372, + "learning_rate": 0.00011430166205714088, + "loss": 0.7689, + "step": 979 + }, + { + "epoch": 0.47189117611652825, + "grad_norm": 0.6593488454818726, + "learning_rate": 0.00011414726051058102, + "loss": 0.4602, + "step": 980 + }, + { + "epoch": 0.4723726977248104, + "grad_norm": 1.089796543121338, + "learning_rate": 0.0001139928245406757, + "loss": 0.4331, + "step": 981 + }, + { + "epoch": 0.4728542193330926, + "grad_norm": 0.6848332285881042, + "learning_rate": 0.00011383835452320097, + "loss": 0.4966, + "step": 982 + }, + { + "epoch": 0.47333574094137476, + "grad_norm": 0.784115195274353, + "learning_rate": 0.00011368385083401585, + "loss": 1.1322, + "step": 983 + }, + { + "epoch": 0.4738172625496569, + "grad_norm": 1.0432567596435547, + "learning_rate": 0.00011352931384906125, + "loss": 0.578, + "step": 984 + }, + { + "epoch": 0.4742987841579391, + "grad_norm": 0.6884013414382935, + "learning_rate": 0.00011337474394435908, + "loss": 0.8005, + "step": 985 + }, + { + "epoch": 0.47478030576622127, + "grad_norm": 0.7809309959411621, + "learning_rate": 0.00011322014149601136, + "loss": 0.656, + "step": 986 + }, + { + "epoch": 0.47526182737450345, + "grad_norm": 0.5823318958282471, + "learning_rate": 0.00011306550688019926, + "loss": 0.6732, + "step": 987 + }, + { + "epoch": 0.4757433489827856, + "grad_norm": 0.6410465240478516, + "learning_rate": 0.0001129108404731823, + "loss": 0.6143, + "step": 988 + }, + { + "epoch": 0.4762248705910678, + "grad_norm": 1.332196593284607, + "learning_rate": 0.0001127561426512973, + "loss": 1.0384, + "step": 989 + }, + { + "epoch": 0.47670639219934996, + "grad_norm": 0.7239651679992676, + "learning_rate": 0.0001126014137909575, + "loss": 0.378, + "step": 990 + }, + { + "epoch": 0.4771879138076321, + "grad_norm": 0.5169118046760559, + "learning_rate": 0.00011244665426865174, + "loss": 0.5249, + "step": 991 + }, + { + "epoch": 0.4776694354159143, + "grad_norm": 1.1747828722000122, + "learning_rate": 0.00011229186446094338, + "loss": 0.5218, + "step": 992 + }, + { + "epoch": 0.47815095702419647, + "grad_norm": 0.5271717309951782, + "learning_rate": 0.00011213704474446951, + "loss": 0.6091, + "step": 993 + }, + { + "epoch": 0.47863247863247865, + "grad_norm": 1.4207522869110107, + "learning_rate": 0.00011198219549594, + "loss": 0.8363, + "step": 994 + }, + { + "epoch": 0.4791140002407608, + "grad_norm": 0.704620897769928, + "learning_rate": 0.00011182731709213659, + "loss": 0.3071, + "step": 995 + }, + { + "epoch": 0.47959552184904297, + "grad_norm": 0.775363564491272, + "learning_rate": 0.00011167240990991192, + "loss": 0.248, + "step": 996 + }, + { + "epoch": 0.48007704345732516, + "grad_norm": 0.6383013725280762, + "learning_rate": 0.00011151747432618871, + "loss": 0.6298, + "step": 997 + }, + { + "epoch": 0.48055856506560735, + "grad_norm": 0.826794445514679, + "learning_rate": 0.00011136251071795871, + "loss": 0.6077, + "step": 998 + }, + { + "epoch": 0.4810400866738895, + "grad_norm": 1.2073599100112915, + "learning_rate": 0.00011120751946228197, + "loss": 0.7025, + "step": 999 + }, + { + "epoch": 0.48152160828217166, + "grad_norm": 0.6839677691459656, + "learning_rate": 0.00011105250093628565, + "loss": 0.608, + "step": 1000 + }, + { + "epoch": 0.48200312989045385, + "grad_norm": 0.6721844673156738, + "learning_rate": 0.00011089745551716344, + "loss": 0.6139, + "step": 1001 + }, + { + "epoch": 0.482484651498736, + "grad_norm": 1.1474583148956299, + "learning_rate": 0.00011074238358217437, + "loss": 1.0103, + "step": 1002 + }, + { + "epoch": 0.48296617310701817, + "grad_norm": 0.670769989490509, + "learning_rate": 0.00011058728550864197, + "loss": 0.2139, + "step": 1003 + }, + { + "epoch": 0.48344769471530036, + "grad_norm": 0.5323492884635925, + "learning_rate": 0.00011043216167395344, + "loss": 0.3595, + "step": 1004 + }, + { + "epoch": 0.48392921632358255, + "grad_norm": 0.8688573837280273, + "learning_rate": 0.00011027701245555865, + "loss": 0.6961, + "step": 1005 + }, + { + "epoch": 0.4844107379318647, + "grad_norm": 1.572704553604126, + "learning_rate": 0.00011012183823096917, + "loss": 0.2322, + "step": 1006 + }, + { + "epoch": 0.48489225954014686, + "grad_norm": 0.6855084896087646, + "learning_rate": 0.00010996663937775751, + "loss": 0.2161, + "step": 1007 + }, + { + "epoch": 0.48537378114842905, + "grad_norm": 0.722327470779419, + "learning_rate": 0.000109811416273556, + "loss": 1.2421, + "step": 1008 + }, + { + "epoch": 0.4858553027567112, + "grad_norm": 0.6337584257125854, + "learning_rate": 0.00010965616929605609, + "loss": 0.8193, + "step": 1009 + }, + { + "epoch": 0.48633682436499337, + "grad_norm": 0.7821305394172668, + "learning_rate": 0.0001095008988230072, + "loss": 0.4506, + "step": 1010 + }, + { + "epoch": 0.48681834597327556, + "grad_norm": 1.3698606491088867, + "learning_rate": 0.00010934560523221602, + "loss": 0.7479, + "step": 1011 + }, + { + "epoch": 0.48729986758155774, + "grad_norm": 0.9158796072006226, + "learning_rate": 0.00010919028890154543, + "loss": 0.5681, + "step": 1012 + }, + { + "epoch": 0.4877813891898399, + "grad_norm": 0.808804452419281, + "learning_rate": 0.00010903495020891375, + "loss": 0.7309, + "step": 1013 + }, + { + "epoch": 0.48826291079812206, + "grad_norm": 0.6452478170394897, + "learning_rate": 0.00010887958953229349, + "loss": 1.1787, + "step": 1014 + }, + { + "epoch": 0.48874443240640425, + "grad_norm": 0.7447563409805298, + "learning_rate": 0.00010872420724971088, + "loss": 0.5845, + "step": 1015 + }, + { + "epoch": 0.4892259540146864, + "grad_norm": 0.6775797009468079, + "learning_rate": 0.0001085688037392446, + "loss": 0.3645, + "step": 1016 + }, + { + "epoch": 0.48970747562296857, + "grad_norm": 0.8377703428268433, + "learning_rate": 0.000108413379379025, + "loss": 0.4498, + "step": 1017 + }, + { + "epoch": 0.49018899723125076, + "grad_norm": 0.3953412175178528, + "learning_rate": 0.00010825793454723325, + "loss": 0.4478, + "step": 1018 + }, + { + "epoch": 0.49067051883953294, + "grad_norm": 0.7678600549697876, + "learning_rate": 0.00010810246962210018, + "loss": 0.978, + "step": 1019 + }, + { + "epoch": 0.4911520404478151, + "grad_norm": 1.129772424697876, + "learning_rate": 0.00010794698498190557, + "loss": 0.5298, + "step": 1020 + }, + { + "epoch": 0.49163356205609726, + "grad_norm": 0.7634359002113342, + "learning_rate": 0.00010779148100497722, + "loss": 0.5905, + "step": 1021 + }, + { + "epoch": 0.49211508366437945, + "grad_norm": 0.6913365125656128, + "learning_rate": 0.00010763595806968996, + "loss": 0.9734, + "step": 1022 + }, + { + "epoch": 0.49259660527266164, + "grad_norm": 1.0433826446533203, + "learning_rate": 0.00010748041655446473, + "loss": 0.9034, + "step": 1023 + }, + { + "epoch": 0.49307812688094377, + "grad_norm": 0.9241974949836731, + "learning_rate": 0.00010732485683776768, + "loss": 0.9534, + "step": 1024 + }, + { + "epoch": 0.49355964848922596, + "grad_norm": 0.8278858661651611, + "learning_rate": 0.00010716927929810925, + "loss": 0.6806, + "step": 1025 + }, + { + "epoch": 0.49404117009750814, + "grad_norm": 1.0138304233551025, + "learning_rate": 0.00010701368431404326, + "loss": 0.5995, + "step": 1026 + }, + { + "epoch": 0.4945226917057903, + "grad_norm": 0.5914368629455566, + "learning_rate": 0.00010685807226416598, + "loss": 0.3246, + "step": 1027 + }, + { + "epoch": 0.49500421331407246, + "grad_norm": 0.81244295835495, + "learning_rate": 0.00010670244352711518, + "loss": 0.4668, + "step": 1028 + }, + { + "epoch": 0.49548573492235465, + "grad_norm": 0.7371711134910583, + "learning_rate": 0.00010654679848156925, + "loss": 0.3986, + "step": 1029 + }, + { + "epoch": 0.49596725653063684, + "grad_norm": 0.7539998292922974, + "learning_rate": 0.00010639113750624625, + "loss": 0.3003, + "step": 1030 + }, + { + "epoch": 0.49644877813891897, + "grad_norm": 0.879912793636322, + "learning_rate": 0.00010623546097990303, + "loss": 0.844, + "step": 1031 + }, + { + "epoch": 0.49693029974720115, + "grad_norm": 0.8178533315658569, + "learning_rate": 0.00010607976928133423, + "loss": 0.253, + "step": 1032 + }, + { + "epoch": 0.49741182135548334, + "grad_norm": 0.8457581400871277, + "learning_rate": 0.00010592406278937144, + "loss": 0.4384, + "step": 1033 + }, + { + "epoch": 0.4978933429637655, + "grad_norm": 1.146519660949707, + "learning_rate": 0.00010576834188288226, + "loss": 0.4484, + "step": 1034 + }, + { + "epoch": 0.49837486457204766, + "grad_norm": 0.8387516736984253, + "learning_rate": 0.00010561260694076935, + "loss": 0.6428, + "step": 1035 + }, + { + "epoch": 0.49885638618032985, + "grad_norm": 0.5453752279281616, + "learning_rate": 0.00010545685834196948, + "loss": 0.4481, + "step": 1036 + }, + { + "epoch": 0.49933790778861203, + "grad_norm": 1.3885741233825684, + "learning_rate": 0.00010530109646545272, + "loss": 0.5443, + "step": 1037 + }, + { + "epoch": 0.49981942939689417, + "grad_norm": 0.8055309653282166, + "learning_rate": 0.0001051453216902214, + "loss": 0.5387, + "step": 1038 + }, + { + "epoch": 0.5003009510051764, + "grad_norm": 0.8343227505683899, + "learning_rate": 0.00010498953439530925, + "loss": 0.7728, + "step": 1039 + }, + { + "epoch": 0.5007824726134585, + "grad_norm": 0.5583564639091492, + "learning_rate": 0.00010483373495978046, + "loss": 0.4572, + "step": 1040 + }, + { + "epoch": 0.5012639942217407, + "grad_norm": 1.2023993730545044, + "learning_rate": 0.00010467792376272877, + "loss": 0.8647, + "step": 1041 + }, + { + "epoch": 0.5017455158300229, + "grad_norm": 0.6801705360412598, + "learning_rate": 0.00010452210118327652, + "loss": 0.4475, + "step": 1042 + }, + { + "epoch": 0.502227037438305, + "grad_norm": 1.6109269857406616, + "learning_rate": 0.00010436626760057378, + "loss": 0.6505, + "step": 1043 + }, + { + "epoch": 0.5027085590465872, + "grad_norm": 0.7384167313575745, + "learning_rate": 0.00010421042339379732, + "loss": 0.2938, + "step": 1044 + }, + { + "epoch": 0.5031900806548694, + "grad_norm": 0.7244082689285278, + "learning_rate": 0.00010405456894214987, + "loss": 0.1328, + "step": 1045 + }, + { + "epoch": 0.5036716022631516, + "grad_norm": 1.1768090724945068, + "learning_rate": 0.00010389870462485902, + "loss": 1.5518, + "step": 1046 + }, + { + "epoch": 0.5041531238714337, + "grad_norm": 0.6785743832588196, + "learning_rate": 0.00010374283082117635, + "loss": 0.2975, + "step": 1047 + }, + { + "epoch": 0.5046346454797159, + "grad_norm": 0.9195048213005066, + "learning_rate": 0.00010358694791037653, + "loss": 0.6257, + "step": 1048 + }, + { + "epoch": 0.5051161670879981, + "grad_norm": 1.71181058883667, + "learning_rate": 0.00010343105627175644, + "loss": 0.6551, + "step": 1049 + }, + { + "epoch": 0.5055976886962803, + "grad_norm": 0.998651385307312, + "learning_rate": 0.00010327515628463415, + "loss": 0.823, + "step": 1050 + }, + { + "epoch": 0.5060792103045624, + "grad_norm": 0.828630805015564, + "learning_rate": 0.00010311924832834808, + "loss": 0.9087, + "step": 1051 + }, + { + "epoch": 0.5065607319128446, + "grad_norm": 1.2712898254394531, + "learning_rate": 0.00010296333278225599, + "loss": 0.4819, + "step": 1052 + }, + { + "epoch": 0.5070422535211268, + "grad_norm": 1.1045798063278198, + "learning_rate": 0.00010280741002573413, + "loss": 0.2973, + "step": 1053 + }, + { + "epoch": 0.5075237751294089, + "grad_norm": 0.6555344462394714, + "learning_rate": 0.00010265148043817632, + "loss": 0.4144, + "step": 1054 + }, + { + "epoch": 0.5080052967376911, + "grad_norm": 0.92887943983078, + "learning_rate": 0.00010249554439899298, + "loss": 0.3468, + "step": 1055 + }, + { + "epoch": 0.5084868183459733, + "grad_norm": 0.6355475187301636, + "learning_rate": 0.00010233960228761022, + "loss": 0.5131, + "step": 1056 + }, + { + "epoch": 0.5089683399542555, + "grad_norm": 0.5725042223930359, + "learning_rate": 0.00010218365448346893, + "loss": 0.5927, + "step": 1057 + }, + { + "epoch": 0.5094498615625376, + "grad_norm": 1.2848678827285767, + "learning_rate": 0.00010202770136602388, + "loss": 0.7266, + "step": 1058 + }, + { + "epoch": 0.5099313831708198, + "grad_norm": 0.6311907768249512, + "learning_rate": 0.00010187174331474271, + "loss": 0.2647, + "step": 1059 + }, + { + "epoch": 0.510412904779102, + "grad_norm": 0.8230004906654358, + "learning_rate": 0.00010171578070910512, + "loss": 0.4257, + "step": 1060 + }, + { + "epoch": 0.5108944263873841, + "grad_norm": 0.9461697340011597, + "learning_rate": 0.00010155981392860185, + "loss": 0.3434, + "step": 1061 + }, + { + "epoch": 0.5113759479956663, + "grad_norm": 0.7990442514419556, + "learning_rate": 0.00010140384335273386, + "loss": 0.6928, + "step": 1062 + }, + { + "epoch": 0.5118574696039485, + "grad_norm": 0.9711220860481262, + "learning_rate": 0.00010124786936101127, + "loss": 0.4676, + "step": 1063 + }, + { + "epoch": 0.5123389912122307, + "grad_norm": 0.7467138171195984, + "learning_rate": 0.00010109189233295255, + "loss": 0.7543, + "step": 1064 + }, + { + "epoch": 0.5128205128205128, + "grad_norm": 1.2974790334701538, + "learning_rate": 0.00010093591264808358, + "loss": 0.6458, + "step": 1065 + }, + { + "epoch": 0.513302034428795, + "grad_norm": 0.8935269117355347, + "learning_rate": 0.00010077993068593663, + "loss": 0.581, + "step": 1066 + }, + { + "epoch": 0.5137835560370771, + "grad_norm": 1.0466784238815308, + "learning_rate": 0.00010062394682604963, + "loss": 0.5093, + "step": 1067 + }, + { + "epoch": 0.5142650776453593, + "grad_norm": 0.6737858057022095, + "learning_rate": 0.00010046796144796497, + "loss": 0.5222, + "step": 1068 + }, + { + "epoch": 0.5147465992536415, + "grad_norm": 0.957423746585846, + "learning_rate": 0.0001003119749312289, + "loss": 0.506, + "step": 1069 + }, + { + "epoch": 0.5152281208619237, + "grad_norm": 0.963703989982605, + "learning_rate": 0.00010015598765539031, + "loss": 0.6979, + "step": 1070 + }, + { + "epoch": 0.5157096424702059, + "grad_norm": 0.8813979029655457, + "learning_rate": 0.0001, + "loss": 0.5174, + "step": 1071 + }, + { + "epoch": 0.516191164078488, + "grad_norm": 0.6864508390426636, + "learning_rate": 9.984401234460971e-05, + "loss": 0.5415, + "step": 1072 + }, + { + "epoch": 0.5166726856867702, + "grad_norm": 1.0436968803405762, + "learning_rate": 9.968802506877111e-05, + "loss": 1.0143, + "step": 1073 + }, + { + "epoch": 0.5171542072950523, + "grad_norm": 0.4878181517124176, + "learning_rate": 9.953203855203504e-05, + "loss": 0.266, + "step": 1074 + }, + { + "epoch": 0.5176357289033345, + "grad_norm": 0.9048622846603394, + "learning_rate": 9.93760531739504e-05, + "loss": 0.7538, + "step": 1075 + }, + { + "epoch": 0.5181172505116167, + "grad_norm": 1.1710916757583618, + "learning_rate": 9.922006931406338e-05, + "loss": 0.8256, + "step": 1076 + }, + { + "epoch": 0.5185987721198989, + "grad_norm": 0.8252659440040588, + "learning_rate": 9.906408735191643e-05, + "loss": 0.2878, + "step": 1077 + }, + { + "epoch": 0.5190802937281811, + "grad_norm": 0.8352766633033752, + "learning_rate": 9.890810766704745e-05, + "loss": 0.6413, + "step": 1078 + }, + { + "epoch": 0.5195618153364632, + "grad_norm": 1.7258776426315308, + "learning_rate": 9.875213063898875e-05, + "loss": 0.6046, + "step": 1079 + }, + { + "epoch": 0.5200433369447454, + "grad_norm": 0.6900729537010193, + "learning_rate": 9.859615664726615e-05, + "loss": 0.6857, + "step": 1080 + }, + { + "epoch": 0.5205248585530275, + "grad_norm": 1.3947724103927612, + "learning_rate": 9.844018607139818e-05, + "loss": 0.6629, + "step": 1081 + }, + { + "epoch": 0.5210063801613097, + "grad_norm": 0.6548985242843628, + "learning_rate": 9.828421929089493e-05, + "loss": 0.2404, + "step": 1082 + }, + { + "epoch": 0.5214879017695919, + "grad_norm": 1.1727840900421143, + "learning_rate": 9.812825668525733e-05, + "loss": 0.668, + "step": 1083 + }, + { + "epoch": 0.5219694233778741, + "grad_norm": 0.7604357004165649, + "learning_rate": 9.797229863397615e-05, + "loss": 0.5043, + "step": 1084 + }, + { + "epoch": 0.5224509449861563, + "grad_norm": 0.792493999004364, + "learning_rate": 9.781634551653108e-05, + "loss": 0.4905, + "step": 1085 + }, + { + "epoch": 0.5229324665944384, + "grad_norm": 0.8167349696159363, + "learning_rate": 9.766039771238982e-05, + "loss": 0.9703, + "step": 1086 + }, + { + "epoch": 0.5234139882027206, + "grad_norm": 0.9371551275253296, + "learning_rate": 9.750445560100706e-05, + "loss": 0.7786, + "step": 1087 + }, + { + "epoch": 0.5238955098110027, + "grad_norm": 1.1863070726394653, + "learning_rate": 9.73485195618237e-05, + "loss": 1.097, + "step": 1088 + }, + { + "epoch": 0.5243770314192849, + "grad_norm": 0.9151334166526794, + "learning_rate": 9.719258997426588e-05, + "loss": 0.6039, + "step": 1089 + }, + { + "epoch": 0.5248585530275671, + "grad_norm": 0.6455950140953064, + "learning_rate": 9.703666721774402e-05, + "loss": 0.2652, + "step": 1090 + }, + { + "epoch": 0.5253400746358493, + "grad_norm": 0.9584352374076843, + "learning_rate": 9.688075167165194e-05, + "loss": 0.5744, + "step": 1091 + }, + { + "epoch": 0.5258215962441315, + "grad_norm": 0.9051257967948914, + "learning_rate": 9.672484371536586e-05, + "loss": 0.4059, + "step": 1092 + }, + { + "epoch": 0.5263031178524137, + "grad_norm": 0.840788722038269, + "learning_rate": 9.656894372824358e-05, + "loss": 1.0904, + "step": 1093 + }, + { + "epoch": 0.5267846394606958, + "grad_norm": 0.787898063659668, + "learning_rate": 9.64130520896235e-05, + "loss": 0.4965, + "step": 1094 + }, + { + "epoch": 0.527266161068978, + "grad_norm": 0.5363286733627319, + "learning_rate": 9.625716917882367e-05, + "loss": 0.4016, + "step": 1095 + }, + { + "epoch": 0.5277476826772601, + "grad_norm": 0.6209604740142822, + "learning_rate": 9.6101295375141e-05, + "loss": 0.9046, + "step": 1096 + }, + { + "epoch": 0.5282292042855423, + "grad_norm": 0.7012305855751038, + "learning_rate": 9.594543105785013e-05, + "loss": 0.6374, + "step": 1097 + }, + { + "epoch": 0.5287107258938245, + "grad_norm": 0.6587700247764587, + "learning_rate": 9.578957660620267e-05, + "loss": 0.3757, + "step": 1098 + }, + { + "epoch": 0.5291922475021067, + "grad_norm": 1.1536179780960083, + "learning_rate": 9.563373239942623e-05, + "loss": 0.3881, + "step": 1099 + }, + { + "epoch": 0.5296737691103889, + "grad_norm": 0.8401791453361511, + "learning_rate": 9.547789881672348e-05, + "loss": 0.5374, + "step": 1100 + }, + { + "epoch": 0.530155290718671, + "grad_norm": 1.215343713760376, + "learning_rate": 9.532207623727126e-05, + "loss": 0.9501, + "step": 1101 + }, + { + "epoch": 0.5306368123269531, + "grad_norm": 0.7829989194869995, + "learning_rate": 9.516626504021957e-05, + "loss": 0.5429, + "step": 1102 + }, + { + "epoch": 0.5311183339352353, + "grad_norm": 1.4323217868804932, + "learning_rate": 9.501046560469079e-05, + "loss": 0.7029, + "step": 1103 + }, + { + "epoch": 0.5315998555435175, + "grad_norm": 0.8599506616592407, + "learning_rate": 9.485467830977864e-05, + "loss": 1.1013, + "step": 1104 + }, + { + "epoch": 0.5320813771517997, + "grad_norm": 1.4327293634414673, + "learning_rate": 9.469890353454732e-05, + "loss": 0.2591, + "step": 1105 + }, + { + "epoch": 0.5325628987600819, + "grad_norm": 0.6792353987693787, + "learning_rate": 9.454314165803054e-05, + "loss": 0.2762, + "step": 1106 + }, + { + "epoch": 0.5330444203683641, + "grad_norm": 0.7597439885139465, + "learning_rate": 9.438739305923067e-05, + "loss": 0.5865, + "step": 1107 + }, + { + "epoch": 0.5335259419766462, + "grad_norm": 1.0125542879104614, + "learning_rate": 9.423165811711777e-05, + "loss": 0.9981, + "step": 1108 + }, + { + "epoch": 0.5340074635849283, + "grad_norm": 1.0826612710952759, + "learning_rate": 9.407593721062859e-05, + "loss": 0.4732, + "step": 1109 + }, + { + "epoch": 0.5344889851932105, + "grad_norm": 0.7020871639251709, + "learning_rate": 9.39202307186658e-05, + "loss": 0.3723, + "step": 1110 + }, + { + "epoch": 0.5349705068014927, + "grad_norm": 0.6123396754264832, + "learning_rate": 9.3764539020097e-05, + "loss": 0.5444, + "step": 1111 + }, + { + "epoch": 0.5354520284097749, + "grad_norm": 0.9989874958992004, + "learning_rate": 9.360886249375376e-05, + "loss": 0.9759, + "step": 1112 + }, + { + "epoch": 0.5359335500180571, + "grad_norm": 0.4249228239059448, + "learning_rate": 9.345320151843078e-05, + "loss": 0.3224, + "step": 1113 + }, + { + "epoch": 0.5364150716263393, + "grad_norm": 1.9821633100509644, + "learning_rate": 9.329755647288485e-05, + "loss": 1.0542, + "step": 1114 + }, + { + "epoch": 0.5368965932346214, + "grad_norm": 0.9637462496757507, + "learning_rate": 9.314192773583403e-05, + "loss": 0.5969, + "step": 1115 + }, + { + "epoch": 0.5373781148429035, + "grad_norm": 0.8755899667739868, + "learning_rate": 9.298631568595674e-05, + "loss": 0.5683, + "step": 1116 + }, + { + "epoch": 0.5378596364511857, + "grad_norm": 0.6031029224395752, + "learning_rate": 9.283072070189075e-05, + "loss": 0.624, + "step": 1117 + }, + { + "epoch": 0.5383411580594679, + "grad_norm": 0.8484808802604675, + "learning_rate": 9.267514316223234e-05, + "loss": 0.5871, + "step": 1118 + }, + { + "epoch": 0.5388226796677501, + "grad_norm": 0.7090550065040588, + "learning_rate": 9.251958344553528e-05, + "loss": 0.3945, + "step": 1119 + }, + { + "epoch": 0.5393042012760323, + "grad_norm": 0.906046450138092, + "learning_rate": 9.23640419303101e-05, + "loss": 0.572, + "step": 1120 + }, + { + "epoch": 0.5397857228843145, + "grad_norm": 0.7741056084632874, + "learning_rate": 9.220851899502283e-05, + "loss": 1.0599, + "step": 1121 + }, + { + "epoch": 0.5402672444925966, + "grad_norm": 0.6731898784637451, + "learning_rate": 9.205301501809448e-05, + "loss": 0.4439, + "step": 1122 + }, + { + "epoch": 0.5407487661008787, + "grad_norm": 0.40701019763946533, + "learning_rate": 9.189753037789987e-05, + "loss": 0.5468, + "step": 1123 + }, + { + "epoch": 0.5412302877091609, + "grad_norm": 0.5402950048446655, + "learning_rate": 9.174206545276677e-05, + "loss": 0.7252, + "step": 1124 + }, + { + "epoch": 0.5417118093174431, + "grad_norm": 0.8427053093910217, + "learning_rate": 9.158662062097501e-05, + "loss": 1.004, + "step": 1125 + }, + { + "epoch": 0.5421933309257253, + "grad_norm": 0.36185023188591003, + "learning_rate": 9.143119626075542e-05, + "loss": 0.137, + "step": 1126 + }, + { + "epoch": 0.5426748525340075, + "grad_norm": 1.0534988641738892, + "learning_rate": 9.127579275028914e-05, + "loss": 1.3426, + "step": 1127 + }, + { + "epoch": 0.5431563741422897, + "grad_norm": 0.47852763533592224, + "learning_rate": 9.112041046770653e-05, + "loss": 0.5912, + "step": 1128 + }, + { + "epoch": 0.5436378957505718, + "grad_norm": 1.3245254755020142, + "learning_rate": 9.096504979108629e-05, + "loss": 0.5478, + "step": 1129 + }, + { + "epoch": 0.5441194173588539, + "grad_norm": 0.6596415638923645, + "learning_rate": 9.080971109845458e-05, + "loss": 0.665, + "step": 1130 + }, + { + "epoch": 0.5446009389671361, + "grad_norm": 1.0128172636032104, + "learning_rate": 9.0654394767784e-05, + "loss": 0.523, + "step": 1131 + }, + { + "epoch": 0.5450824605754183, + "grad_norm": 0.9200664758682251, + "learning_rate": 9.049910117699281e-05, + "loss": 0.4824, + "step": 1132 + }, + { + "epoch": 0.5455639821837005, + "grad_norm": 0.8327468633651733, + "learning_rate": 9.034383070394393e-05, + "loss": 0.5497, + "step": 1133 + }, + { + "epoch": 0.5460455037919827, + "grad_norm": 0.7922326326370239, + "learning_rate": 9.0188583726444e-05, + "loss": 1.0697, + "step": 1134 + }, + { + "epoch": 0.5465270254002649, + "grad_norm": 0.9038323760032654, + "learning_rate": 9.00333606222425e-05, + "loss": 0.572, + "step": 1135 + }, + { + "epoch": 0.5470085470085471, + "grad_norm": 0.9835309386253357, + "learning_rate": 8.987816176903082e-05, + "loss": 0.5335, + "step": 1136 + }, + { + "epoch": 0.5474900686168291, + "grad_norm": 0.7787536978721619, + "learning_rate": 8.972298754444136e-05, + "loss": 0.7036, + "step": 1137 + }, + { + "epoch": 0.5479715902251113, + "grad_norm": 0.6886593699455261, + "learning_rate": 8.956783832604654e-05, + "loss": 0.2535, + "step": 1138 + }, + { + "epoch": 0.5484531118333935, + "grad_norm": 1.0839778184890747, + "learning_rate": 8.941271449135806e-05, + "loss": 1.2139, + "step": 1139 + }, + { + "epoch": 0.5489346334416757, + "grad_norm": 1.1864848136901855, + "learning_rate": 8.925761641782567e-05, + "loss": 1.1105, + "step": 1140 + }, + { + "epoch": 0.5494161550499579, + "grad_norm": 0.4651387631893158, + "learning_rate": 8.910254448283659e-05, + "loss": 0.2631, + "step": 1141 + }, + { + "epoch": 0.5498976766582401, + "grad_norm": 1.0268776416778564, + "learning_rate": 8.894749906371439e-05, + "loss": 0.6797, + "step": 1142 + }, + { + "epoch": 0.5503791982665223, + "grad_norm": 0.5834671258926392, + "learning_rate": 8.87924805377181e-05, + "loss": 0.3967, + "step": 1143 + }, + { + "epoch": 0.5508607198748043, + "grad_norm": 0.7293856143951416, + "learning_rate": 8.863748928204131e-05, + "loss": 0.2256, + "step": 1144 + }, + { + "epoch": 0.5513422414830865, + "grad_norm": 1.627150058746338, + "learning_rate": 8.848252567381131e-05, + "loss": 0.6562, + "step": 1145 + }, + { + "epoch": 0.5518237630913687, + "grad_norm": 0.8662706613540649, + "learning_rate": 8.83275900900881e-05, + "loss": 0.4498, + "step": 1146 + }, + { + "epoch": 0.5523052846996509, + "grad_norm": 0.8959720730781555, + "learning_rate": 8.817268290786343e-05, + "loss": 0.6327, + "step": 1147 + }, + { + "epoch": 0.5527868063079331, + "grad_norm": 0.780279815196991, + "learning_rate": 8.801780450406002e-05, + "loss": 0.3141, + "step": 1148 + }, + { + "epoch": 0.5532683279162153, + "grad_norm": 0.8042910099029541, + "learning_rate": 8.786295525553053e-05, + "loss": 0.1242, + "step": 1149 + }, + { + "epoch": 0.5537498495244975, + "grad_norm": 0.8482663631439209, + "learning_rate": 8.770813553905664e-05, + "loss": 0.2934, + "step": 1150 + }, + { + "epoch": 0.5542313711327795, + "grad_norm": 1.3287039995193481, + "learning_rate": 8.755334573134829e-05, + "loss": 0.5532, + "step": 1151 + }, + { + "epoch": 0.5547128927410617, + "grad_norm": 0.7641162872314453, + "learning_rate": 8.739858620904251e-05, + "loss": 0.512, + "step": 1152 + }, + { + "epoch": 0.5551944143493439, + "grad_norm": 0.44970107078552246, + "learning_rate": 8.724385734870271e-05, + "loss": 0.4683, + "step": 1153 + }, + { + "epoch": 0.5556759359576261, + "grad_norm": 0.8389862179756165, + "learning_rate": 8.708915952681769e-05, + "loss": 0.353, + "step": 1154 + }, + { + "epoch": 0.5561574575659083, + "grad_norm": 0.7373841404914856, + "learning_rate": 8.693449311980074e-05, + "loss": 0.7819, + "step": 1155 + }, + { + "epoch": 0.5566389791741905, + "grad_norm": 1.1287013292312622, + "learning_rate": 8.677985850398866e-05, + "loss": 0.5839, + "step": 1156 + }, + { + "epoch": 0.5571205007824727, + "grad_norm": 0.9452033638954163, + "learning_rate": 8.662525605564093e-05, + "loss": 0.534, + "step": 1157 + }, + { + "epoch": 0.5576020223907547, + "grad_norm": 0.9100724458694458, + "learning_rate": 8.647068615093875e-05, + "loss": 1.1361, + "step": 1158 + }, + { + "epoch": 0.5580835439990369, + "grad_norm": 1.0564773082733154, + "learning_rate": 8.631614916598419e-05, + "loss": 0.6164, + "step": 1159 + }, + { + "epoch": 0.5585650656073191, + "grad_norm": 1.761771559715271, + "learning_rate": 8.616164547679906e-05, + "loss": 0.3506, + "step": 1160 + }, + { + "epoch": 0.5590465872156013, + "grad_norm": 1.167173147201538, + "learning_rate": 8.600717545932435e-05, + "loss": 1.0166, + "step": 1161 + }, + { + "epoch": 0.5595281088238835, + "grad_norm": 0.8327739238739014, + "learning_rate": 8.5852739489419e-05, + "loss": 0.1438, + "step": 1162 + }, + { + "epoch": 0.5600096304321657, + "grad_norm": 0.8960033655166626, + "learning_rate": 8.569833794285915e-05, + "loss": 0.8653, + "step": 1163 + }, + { + "epoch": 0.5604911520404479, + "grad_norm": 0.8796095848083496, + "learning_rate": 8.554397119533714e-05, + "loss": 1.2771, + "step": 1164 + }, + { + "epoch": 0.5609726736487299, + "grad_norm": 0.594007670879364, + "learning_rate": 8.538963962246069e-05, + "loss": 0.3621, + "step": 1165 + }, + { + "epoch": 0.5614541952570121, + "grad_norm": 0.8897623419761658, + "learning_rate": 8.523534359975189e-05, + "loss": 0.5407, + "step": 1166 + }, + { + "epoch": 0.5619357168652943, + "grad_norm": 0.7416290044784546, + "learning_rate": 8.508108350264635e-05, + "loss": 0.3464, + "step": 1167 + }, + { + "epoch": 0.5624172384735765, + "grad_norm": 0.941420316696167, + "learning_rate": 8.492685970649228e-05, + "loss": 0.432, + "step": 1168 + }, + { + "epoch": 0.5628987600818587, + "grad_norm": 1.5289616584777832, + "learning_rate": 8.477267258654949e-05, + "loss": 0.7616, + "step": 1169 + }, + { + "epoch": 0.5633802816901409, + "grad_norm": 1.0448668003082275, + "learning_rate": 8.461852251798866e-05, + "loss": 0.9855, + "step": 1170 + }, + { + "epoch": 0.5638618032984231, + "grad_norm": 0.48141905665397644, + "learning_rate": 8.44644098758902e-05, + "loss": 0.2598, + "step": 1171 + }, + { + "epoch": 0.5643433249067051, + "grad_norm": 0.6879386305809021, + "learning_rate": 8.431033503524354e-05, + "loss": 0.3566, + "step": 1172 + }, + { + "epoch": 0.5648248465149873, + "grad_norm": 1.2092243432998657, + "learning_rate": 8.415629837094611e-05, + "loss": 0.8088, + "step": 1173 + }, + { + "epoch": 0.5653063681232695, + "grad_norm": 0.7214942574501038, + "learning_rate": 8.400230025780243e-05, + "loss": 0.395, + "step": 1174 + }, + { + "epoch": 0.5657878897315517, + "grad_norm": 1.370095133781433, + "learning_rate": 8.384834107052321e-05, + "loss": 0.6134, + "step": 1175 + }, + { + "epoch": 0.5662694113398339, + "grad_norm": 1.073525071144104, + "learning_rate": 8.369442118372447e-05, + "loss": 0.8196, + "step": 1176 + }, + { + "epoch": 0.5667509329481161, + "grad_norm": 0.705037534236908, + "learning_rate": 8.35405409719266e-05, + "loss": 0.6361, + "step": 1177 + }, + { + "epoch": 0.5672324545563983, + "grad_norm": 0.7596774697303772, + "learning_rate": 8.338670080955349e-05, + "loss": 0.438, + "step": 1178 + }, + { + "epoch": 0.5677139761646804, + "grad_norm": 0.9082719087600708, + "learning_rate": 8.323290107093143e-05, + "loss": 0.6048, + "step": 1179 + }, + { + "epoch": 0.5681954977729625, + "grad_norm": 0.7221473455429077, + "learning_rate": 8.307914213028856e-05, + "loss": 0.6149, + "step": 1180 + }, + { + "epoch": 0.5686770193812447, + "grad_norm": 0.6236855983734131, + "learning_rate": 8.292542436175356e-05, + "loss": 0.5865, + "step": 1181 + }, + { + "epoch": 0.5691585409895269, + "grad_norm": 1.6320685148239136, + "learning_rate": 8.277174813935508e-05, + "loss": 1.1466, + "step": 1182 + }, + { + "epoch": 0.5696400625978091, + "grad_norm": 0.8854456543922424, + "learning_rate": 8.261811383702061e-05, + "loss": 0.5219, + "step": 1183 + }, + { + "epoch": 0.5701215842060913, + "grad_norm": 0.8134288787841797, + "learning_rate": 8.246452182857562e-05, + "loss": 0.4383, + "step": 1184 + }, + { + "epoch": 0.5706031058143735, + "grad_norm": 1.0532677173614502, + "learning_rate": 8.231097248774274e-05, + "loss": 0.9061, + "step": 1185 + }, + { + "epoch": 0.5710846274226556, + "grad_norm": 0.7403393983840942, + "learning_rate": 8.215746618814067e-05, + "loss": 0.2132, + "step": 1186 + }, + { + "epoch": 0.5715661490309377, + "grad_norm": 1.1494709253311157, + "learning_rate": 8.200400330328348e-05, + "loss": 0.866, + "step": 1187 + }, + { + "epoch": 0.5720476706392199, + "grad_norm": 1.0485625267028809, + "learning_rate": 8.185058420657957e-05, + "loss": 1.1284, + "step": 1188 + }, + { + "epoch": 0.5725291922475021, + "grad_norm": 1.1191109418869019, + "learning_rate": 8.16972092713308e-05, + "loss": 0.6806, + "step": 1189 + }, + { + "epoch": 0.5730107138557843, + "grad_norm": 1.6901990175247192, + "learning_rate": 8.154387887073158e-05, + "loss": 0.4902, + "step": 1190 + }, + { + "epoch": 0.5734922354640665, + "grad_norm": 0.799044132232666, + "learning_rate": 8.139059337786792e-05, + "loss": 0.5245, + "step": 1191 + }, + { + "epoch": 0.5739737570723487, + "grad_norm": 0.6634275913238525, + "learning_rate": 8.12373531657166e-05, + "loss": 0.7991, + "step": 1192 + }, + { + "epoch": 0.5744552786806308, + "grad_norm": 0.5595707893371582, + "learning_rate": 8.108415860714418e-05, + "loss": 0.2768, + "step": 1193 + }, + { + "epoch": 0.5749368002889129, + "grad_norm": 1.764378547668457, + "learning_rate": 8.093101007490622e-05, + "loss": 1.0182, + "step": 1194 + }, + { + "epoch": 0.5754183218971951, + "grad_norm": 0.7943925261497498, + "learning_rate": 8.077790794164619e-05, + "loss": 0.4515, + "step": 1195 + }, + { + "epoch": 0.5758998435054773, + "grad_norm": 0.6726824045181274, + "learning_rate": 8.062485257989471e-05, + "loss": 0.6998, + "step": 1196 + }, + { + "epoch": 0.5763813651137595, + "grad_norm": 0.791667103767395, + "learning_rate": 8.047184436206864e-05, + "loss": 0.7586, + "step": 1197 + }, + { + "epoch": 0.5768628867220417, + "grad_norm": 0.7900064587593079, + "learning_rate": 8.031888366046998e-05, + "loss": 0.5003, + "step": 1198 + }, + { + "epoch": 0.5773444083303239, + "grad_norm": 0.6237706542015076, + "learning_rate": 8.016597084728526e-05, + "loss": 0.6402, + "step": 1199 + }, + { + "epoch": 0.577825929938606, + "grad_norm": 0.612813413143158, + "learning_rate": 8.001310629458443e-05, + "loss": 0.8667, + "step": 1200 + }, + { + "epoch": 0.5783074515468881, + "grad_norm": 0.8505640625953674, + "learning_rate": 7.986029037432002e-05, + "loss": 0.9555, + "step": 1201 + }, + { + "epoch": 0.5787889731551703, + "grad_norm": 0.801115095615387, + "learning_rate": 7.970752345832623e-05, + "loss": 0.4727, + "step": 1202 + }, + { + "epoch": 0.5792704947634525, + "grad_norm": 0.49146944284439087, + "learning_rate": 7.9554805918318e-05, + "loss": 0.2439, + "step": 1203 + }, + { + "epoch": 0.5797520163717347, + "grad_norm": 0.7376623749732971, + "learning_rate": 7.940213812589018e-05, + "loss": 0.807, + "step": 1204 + }, + { + "epoch": 0.5802335379800169, + "grad_norm": 0.5820401906967163, + "learning_rate": 7.92495204525165e-05, + "loss": 0.4257, + "step": 1205 + }, + { + "epoch": 0.580715059588299, + "grad_norm": 0.8996889591217041, + "learning_rate": 7.909695326954878e-05, + "loss": 0.5803, + "step": 1206 + }, + { + "epoch": 0.5811965811965812, + "grad_norm": 0.5629736185073853, + "learning_rate": 7.894443694821602e-05, + "loss": 0.3024, + "step": 1207 + }, + { + "epoch": 0.5816781028048633, + "grad_norm": 0.8386406302452087, + "learning_rate": 7.879197185962339e-05, + "loss": 0.5131, + "step": 1208 + }, + { + "epoch": 0.5821596244131455, + "grad_norm": 1.4311405420303345, + "learning_rate": 7.863955837475144e-05, + "loss": 0.5596, + "step": 1209 + }, + { + "epoch": 0.5826411460214277, + "grad_norm": 0.671789824962616, + "learning_rate": 7.848719686445515e-05, + "loss": 0.374, + "step": 1210 + }, + { + "epoch": 0.5831226676297099, + "grad_norm": 0.8818476796150208, + "learning_rate": 7.833488769946306e-05, + "loss": 0.6616, + "step": 1211 + }, + { + "epoch": 0.5836041892379921, + "grad_norm": 0.9540754556655884, + "learning_rate": 7.818263125037633e-05, + "loss": 0.4891, + "step": 1212 + }, + { + "epoch": 0.5840857108462743, + "grad_norm": 0.7437126040458679, + "learning_rate": 7.803042788766777e-05, + "loss": 0.3433, + "step": 1213 + }, + { + "epoch": 0.5845672324545564, + "grad_norm": 0.3874172568321228, + "learning_rate": 7.787827798168115e-05, + "loss": 0.2215, + "step": 1214 + }, + { + "epoch": 0.5850487540628385, + "grad_norm": 0.6588778495788574, + "learning_rate": 7.772618190263009e-05, + "loss": 0.5484, + "step": 1215 + }, + { + "epoch": 0.5855302756711207, + "grad_norm": 0.536056399345398, + "learning_rate": 7.757414002059726e-05, + "loss": 0.4488, + "step": 1216 + }, + { + "epoch": 0.5860117972794029, + "grad_norm": 1.0876915454864502, + "learning_rate": 7.742215270553349e-05, + "loss": 0.4387, + "step": 1217 + }, + { + "epoch": 0.5864933188876851, + "grad_norm": 1.0314831733703613, + "learning_rate": 7.727022032725672e-05, + "loss": 0.6554, + "step": 1218 + }, + { + "epoch": 0.5869748404959673, + "grad_norm": 0.8363491296768188, + "learning_rate": 7.711834325545135e-05, + "loss": 0.483, + "step": 1219 + }, + { + "epoch": 0.5874563621042495, + "grad_norm": 0.8176609873771667, + "learning_rate": 7.696652185966711e-05, + "loss": 0.6103, + "step": 1220 + }, + { + "epoch": 0.5879378837125316, + "grad_norm": 0.7678027153015137, + "learning_rate": 7.681475650931834e-05, + "loss": 0.5092, + "step": 1221 + }, + { + "epoch": 0.5884194053208138, + "grad_norm": 0.6872639656066895, + "learning_rate": 7.666304757368297e-05, + "loss": 0.7324, + "step": 1222 + }, + { + "epoch": 0.5889009269290959, + "grad_norm": 0.47040292620658875, + "learning_rate": 7.651139542190164e-05, + "loss": 0.6611, + "step": 1223 + }, + { + "epoch": 0.5893824485373781, + "grad_norm": 0.6442480087280273, + "learning_rate": 7.635980042297687e-05, + "loss": 0.3392, + "step": 1224 + }, + { + "epoch": 0.5898639701456603, + "grad_norm": 0.7973917126655579, + "learning_rate": 7.620826294577208e-05, + "loss": 0.4293, + "step": 1225 + }, + { + "epoch": 0.5903454917539425, + "grad_norm": 1.671976923942566, + "learning_rate": 7.605678335901071e-05, + "loss": 0.5356, + "step": 1226 + }, + { + "epoch": 0.5908270133622247, + "grad_norm": 1.2067997455596924, + "learning_rate": 7.59053620312754e-05, + "loss": 0.8919, + "step": 1227 + }, + { + "epoch": 0.5913085349705068, + "grad_norm": 0.5278024077415466, + "learning_rate": 7.575399933100697e-05, + "loss": 0.4646, + "step": 1228 + }, + { + "epoch": 0.591790056578789, + "grad_norm": 0.8876713514328003, + "learning_rate": 7.560269562650368e-05, + "loss": 0.6154, + "step": 1229 + }, + { + "epoch": 0.5922715781870711, + "grad_norm": 0.7381996512413025, + "learning_rate": 7.54514512859201e-05, + "loss": 0.3608, + "step": 1230 + }, + { + "epoch": 0.5927530997953533, + "grad_norm": 0.8265383839607239, + "learning_rate": 7.530026667726645e-05, + "loss": 0.4787, + "step": 1231 + }, + { + "epoch": 0.5932346214036355, + "grad_norm": 0.8717719316482544, + "learning_rate": 7.51491421684076e-05, + "loss": 0.9259, + "step": 1232 + }, + { + "epoch": 0.5937161430119177, + "grad_norm": 1.1018342971801758, + "learning_rate": 7.49980781270622e-05, + "loss": 0.5792, + "step": 1233 + }, + { + "epoch": 0.5941976646201999, + "grad_norm": 0.6729315519332886, + "learning_rate": 7.484707492080172e-05, + "loss": 0.7996, + "step": 1234 + }, + { + "epoch": 0.594679186228482, + "grad_norm": 0.9117011427879333, + "learning_rate": 7.469613291704962e-05, + "loss": 0.9377, + "step": 1235 + }, + { + "epoch": 0.5951607078367642, + "grad_norm": 0.7894290089607239, + "learning_rate": 7.45452524830805e-05, + "loss": 0.7289, + "step": 1236 + }, + { + "epoch": 0.5956422294450463, + "grad_norm": 0.4750572443008423, + "learning_rate": 7.439443398601903e-05, + "loss": 0.5718, + "step": 1237 + }, + { + "epoch": 0.5961237510533285, + "grad_norm": 0.7789412140846252, + "learning_rate": 7.424367779283926e-05, + "loss": 0.9233, + "step": 1238 + }, + { + "epoch": 0.5966052726616107, + "grad_norm": 0.7270252704620361, + "learning_rate": 7.409298427036364e-05, + "loss": 0.3033, + "step": 1239 + }, + { + "epoch": 0.5970867942698929, + "grad_norm": 1.0719108581542969, + "learning_rate": 7.39423537852621e-05, + "loss": 0.4617, + "step": 1240 + }, + { + "epoch": 0.597568315878175, + "grad_norm": 0.5281317234039307, + "learning_rate": 7.379178670405123e-05, + "loss": 0.3342, + "step": 1241 + }, + { + "epoch": 0.5980498374864572, + "grad_norm": 0.40090054273605347, + "learning_rate": 7.364128339309326e-05, + "loss": 0.3611, + "step": 1242 + }, + { + "epoch": 0.5985313590947394, + "grad_norm": 0.7400622367858887, + "learning_rate": 7.349084421859533e-05, + "loss": 0.3046, + "step": 1243 + }, + { + "epoch": 0.5990128807030215, + "grad_norm": 0.7203524708747864, + "learning_rate": 7.334046954660852e-05, + "loss": 0.34, + "step": 1244 + }, + { + "epoch": 0.5994944023113037, + "grad_norm": 0.7717897891998291, + "learning_rate": 7.31901597430269e-05, + "loss": 0.7379, + "step": 1245 + }, + { + "epoch": 0.5999759239195859, + "grad_norm": 0.8210568428039551, + "learning_rate": 7.303991517358678e-05, + "loss": 0.8208, + "step": 1246 + }, + { + "epoch": 0.6004574455278681, + "grad_norm": 0.8457664251327515, + "learning_rate": 7.288973620386568e-05, + "loss": 0.8886, + "step": 1247 + }, + { + "epoch": 0.6009389671361502, + "grad_norm": 0.5993969440460205, + "learning_rate": 7.273962319928151e-05, + "loss": 0.488, + "step": 1248 + }, + { + "epoch": 0.6014204887444324, + "grad_norm": 0.5078853964805603, + "learning_rate": 7.258957652509171e-05, + "loss": 0.4295, + "step": 1249 + }, + { + "epoch": 0.6019020103527146, + "grad_norm": 0.6125525236129761, + "learning_rate": 7.24395965463923e-05, + "loss": 0.151, + "step": 1250 + }, + { + "epoch": 0.6023835319609967, + "grad_norm": 0.6394007802009583, + "learning_rate": 7.228968362811702e-05, + "loss": 0.4192, + "step": 1251 + }, + { + "epoch": 0.6028650535692789, + "grad_norm": 0.8358025550842285, + "learning_rate": 7.21398381350364e-05, + "loss": 0.6631, + "step": 1252 + }, + { + "epoch": 0.6033465751775611, + "grad_norm": 0.5983412265777588, + "learning_rate": 7.199006043175698e-05, + "loss": 0.6831, + "step": 1253 + }, + { + "epoch": 0.6038280967858433, + "grad_norm": 0.5821589231491089, + "learning_rate": 7.184035088272028e-05, + "loss": 0.4347, + "step": 1254 + }, + { + "epoch": 0.6043096183941254, + "grad_norm": 0.6282423734664917, + "learning_rate": 7.169070985220208e-05, + "loss": 0.2918, + "step": 1255 + }, + { + "epoch": 0.6047911400024076, + "grad_norm": 0.7748804092407227, + "learning_rate": 7.154113770431132e-05, + "loss": 0.5475, + "step": 1256 + }, + { + "epoch": 0.6052726616106898, + "grad_norm": 0.6467021107673645, + "learning_rate": 7.13916348029894e-05, + "loss": 0.6028, + "step": 1257 + }, + { + "epoch": 0.6057541832189719, + "grad_norm": 0.7658000588417053, + "learning_rate": 7.124220151200926e-05, + "loss": 0.4896, + "step": 1258 + }, + { + "epoch": 0.6062357048272541, + "grad_norm": 0.5645423531532288, + "learning_rate": 7.10928381949744e-05, + "loss": 0.3842, + "step": 1259 + }, + { + "epoch": 0.6067172264355363, + "grad_norm": 1.0179091691970825, + "learning_rate": 7.094354521531807e-05, + "loss": 0.8221, + "step": 1260 + }, + { + "epoch": 0.6071987480438185, + "grad_norm": 0.9218524098396301, + "learning_rate": 7.079432293630244e-05, + "loss": 1.0428, + "step": 1261 + }, + { + "epoch": 0.6076802696521006, + "grad_norm": 1.1272860765457153, + "learning_rate": 7.064517172101753e-05, + "loss": 1.1758, + "step": 1262 + }, + { + "epoch": 0.6081617912603828, + "grad_norm": 1.2278001308441162, + "learning_rate": 7.04960919323806e-05, + "loss": 0.5739, + "step": 1263 + }, + { + "epoch": 0.608643312868665, + "grad_norm": 0.9171412587165833, + "learning_rate": 7.034708393313493e-05, + "loss": 1.0, + "step": 1264 + }, + { + "epoch": 0.6091248344769472, + "grad_norm": 1.4739869832992554, + "learning_rate": 7.019814808584928e-05, + "loss": 1.0325, + "step": 1265 + }, + { + "epoch": 0.6096063560852293, + "grad_norm": 0.9023488163948059, + "learning_rate": 7.004928475291678e-05, + "loss": 0.527, + "step": 1266 + }, + { + "epoch": 0.6100878776935115, + "grad_norm": 1.3271821737289429, + "learning_rate": 6.990049429655412e-05, + "loss": 0.6692, + "step": 1267 + }, + { + "epoch": 0.6105693993017937, + "grad_norm": 0.9032428860664368, + "learning_rate": 6.97517770788007e-05, + "loss": 0.7722, + "step": 1268 + }, + { + "epoch": 0.6110509209100758, + "grad_norm": 0.7868967652320862, + "learning_rate": 6.960313346151761e-05, + "loss": 0.6579, + "step": 1269 + }, + { + "epoch": 0.611532442518358, + "grad_norm": 1.430017113685608, + "learning_rate": 6.9454563806387e-05, + "loss": 1.0602, + "step": 1270 + }, + { + "epoch": 0.6120139641266402, + "grad_norm": 0.6199007630348206, + "learning_rate": 6.930606847491094e-05, + "loss": 0.5871, + "step": 1271 + }, + { + "epoch": 0.6124954857349224, + "grad_norm": 0.9371126294136047, + "learning_rate": 6.915764782841072e-05, + "loss": 0.8398, + "step": 1272 + }, + { + "epoch": 0.6129770073432045, + "grad_norm": 0.8416410684585571, + "learning_rate": 6.900930222802588e-05, + "loss": 0.5942, + "step": 1273 + }, + { + "epoch": 0.6134585289514867, + "grad_norm": 0.6848739385604858, + "learning_rate": 6.886103203471337e-05, + "loss": 0.7256, + "step": 1274 + }, + { + "epoch": 0.6139400505597689, + "grad_norm": 0.899040162563324, + "learning_rate": 6.871283760924665e-05, + "loss": 0.5853, + "step": 1275 + }, + { + "epoch": 0.614421572168051, + "grad_norm": 0.7222660183906555, + "learning_rate": 6.856471931221478e-05, + "loss": 0.7026, + "step": 1276 + }, + { + "epoch": 0.6149030937763332, + "grad_norm": 0.5651343464851379, + "learning_rate": 6.841667750402162e-05, + "loss": 0.3303, + "step": 1277 + }, + { + "epoch": 0.6153846153846154, + "grad_norm": 0.5582093000411987, + "learning_rate": 6.826871254488496e-05, + "loss": 0.3963, + "step": 1278 + }, + { + "epoch": 0.6158661369928976, + "grad_norm": 0.6273393034934998, + "learning_rate": 6.812082479483553e-05, + "loss": 0.1722, + "step": 1279 + }, + { + "epoch": 0.6163476586011797, + "grad_norm": 0.5864158868789673, + "learning_rate": 6.797301461371625e-05, + "loss": 0.2222, + "step": 1280 + }, + { + "epoch": 0.6168291802094619, + "grad_norm": 0.4145785868167877, + "learning_rate": 6.782528236118124e-05, + "loss": 0.1757, + "step": 1281 + }, + { + "epoch": 0.6173107018177441, + "grad_norm": 0.8258092403411865, + "learning_rate": 6.767762839669503e-05, + "loss": 0.4112, + "step": 1282 + }, + { + "epoch": 0.6177922234260262, + "grad_norm": 0.7576352953910828, + "learning_rate": 6.753005307953167e-05, + "loss": 0.4947, + "step": 1283 + }, + { + "epoch": 0.6182737450343084, + "grad_norm": 3.720597743988037, + "learning_rate": 6.738255676877381e-05, + "loss": 0.3222, + "step": 1284 + }, + { + "epoch": 0.6187552666425906, + "grad_norm": 0.43584874272346497, + "learning_rate": 6.723513982331195e-05, + "loss": 0.3454, + "step": 1285 + }, + { + "epoch": 0.6192367882508728, + "grad_norm": 1.3589578866958618, + "learning_rate": 6.708780260184333e-05, + "loss": 0.4681, + "step": 1286 + }, + { + "epoch": 0.6197183098591549, + "grad_norm": 0.5593907237052917, + "learning_rate": 6.694054546287132e-05, + "loss": 0.4205, + "step": 1287 + }, + { + "epoch": 0.6201998314674371, + "grad_norm": 0.9470499157905579, + "learning_rate": 6.679336876470441e-05, + "loss": 0.7325, + "step": 1288 + }, + { + "epoch": 0.6206813530757193, + "grad_norm": 1.2402819395065308, + "learning_rate": 6.664627286545535e-05, + "loss": 0.5052, + "step": 1289 + }, + { + "epoch": 0.6211628746840014, + "grad_norm": 0.6698610782623291, + "learning_rate": 6.649925812304025e-05, + "loss": 0.4292, + "step": 1290 + }, + { + "epoch": 0.6216443962922836, + "grad_norm": 0.7191325426101685, + "learning_rate": 6.635232489517782e-05, + "loss": 0.669, + "step": 1291 + }, + { + "epoch": 0.6221259179005658, + "grad_norm": 0.5654041171073914, + "learning_rate": 6.620547353938836e-05, + "loss": 0.4625, + "step": 1292 + }, + { + "epoch": 0.622607439508848, + "grad_norm": 0.7504422068595886, + "learning_rate": 6.605870441299302e-05, + "loss": 0.3405, + "step": 1293 + }, + { + "epoch": 0.6230889611171301, + "grad_norm": 1.1788744926452637, + "learning_rate": 6.591201787311285e-05, + "loss": 1.0483, + "step": 1294 + }, + { + "epoch": 0.6235704827254123, + "grad_norm": 0.8092876076698303, + "learning_rate": 6.57654142766679e-05, + "loss": 0.8794, + "step": 1295 + }, + { + "epoch": 0.6240520043336945, + "grad_norm": 0.7599140405654907, + "learning_rate": 6.561889398037643e-05, + "loss": 0.3576, + "step": 1296 + }, + { + "epoch": 0.6245335259419766, + "grad_norm": 1.1310917139053345, + "learning_rate": 6.547245734075403e-05, + "loss": 0.4158, + "step": 1297 + }, + { + "epoch": 0.6250150475502588, + "grad_norm": 0.5297826528549194, + "learning_rate": 6.532610471411274e-05, + "loss": 0.3243, + "step": 1298 + }, + { + "epoch": 0.625496569158541, + "grad_norm": 0.8771138191223145, + "learning_rate": 6.517983645656014e-05, + "loss": 0.3289, + "step": 1299 + }, + { + "epoch": 0.6259780907668232, + "grad_norm": 0.9111340641975403, + "learning_rate": 6.503365292399857e-05, + "loss": 0.8497, + "step": 1300 + }, + { + "epoch": 0.6264596123751053, + "grad_norm": 0.7726771235466003, + "learning_rate": 6.488755447212418e-05, + "loss": 0.6634, + "step": 1301 + }, + { + "epoch": 0.6269411339833875, + "grad_norm": 0.7417446970939636, + "learning_rate": 6.474154145642612e-05, + "loss": 1.17, + "step": 1302 + }, + { + "epoch": 0.6274226555916697, + "grad_norm": 0.7755337953567505, + "learning_rate": 6.459561423218561e-05, + "loss": 1.199, + "step": 1303 + }, + { + "epoch": 0.6279041771999518, + "grad_norm": 0.4433456361293793, + "learning_rate": 6.444977315447521e-05, + "loss": 0.3119, + "step": 1304 + }, + { + "epoch": 0.628385698808234, + "grad_norm": 0.6270627975463867, + "learning_rate": 6.430401857815776e-05, + "loss": 0.415, + "step": 1305 + }, + { + "epoch": 0.6288672204165162, + "grad_norm": 0.9268920421600342, + "learning_rate": 6.415835085788575e-05, + "loss": 0.6387, + "step": 1306 + }, + { + "epoch": 0.6293487420247984, + "grad_norm": 0.47612449526786804, + "learning_rate": 6.401277034810017e-05, + "loss": 0.1866, + "step": 1307 + }, + { + "epoch": 0.6298302636330806, + "grad_norm": 0.8117198348045349, + "learning_rate": 6.386727740302994e-05, + "loss": 0.6716, + "step": 1308 + }, + { + "epoch": 0.6303117852413627, + "grad_norm": 1.017600417137146, + "learning_rate": 6.37218723766909e-05, + "loss": 0.2563, + "step": 1309 + }, + { + "epoch": 0.6307933068496449, + "grad_norm": 1.1226799488067627, + "learning_rate": 6.357655562288488e-05, + "loss": 0.8567, + "step": 1310 + }, + { + "epoch": 0.631274828457927, + "grad_norm": 1.2738351821899414, + "learning_rate": 6.343132749519902e-05, + "loss": 0.9565, + "step": 1311 + }, + { + "epoch": 0.6317563500662092, + "grad_norm": 1.0473432540893555, + "learning_rate": 6.328618834700474e-05, + "loss": 0.8701, + "step": 1312 + }, + { + "epoch": 0.6322378716744914, + "grad_norm": 1.9390569925308228, + "learning_rate": 6.314113853145703e-05, + "loss": 0.4867, + "step": 1313 + }, + { + "epoch": 0.6327193932827736, + "grad_norm": 0.5643951296806335, + "learning_rate": 6.299617840149349e-05, + "loss": 0.184, + "step": 1314 + }, + { + "epoch": 0.6332009148910558, + "grad_norm": 0.5072335004806519, + "learning_rate": 6.285130830983339e-05, + "loss": 0.1524, + "step": 1315 + }, + { + "epoch": 0.6336824364993379, + "grad_norm": 0.8295088410377502, + "learning_rate": 6.270652860897704e-05, + "loss": 0.6265, + "step": 1316 + }, + { + "epoch": 0.63416395810762, + "grad_norm": 0.7741579413414001, + "learning_rate": 6.25618396512048e-05, + "loss": 0.5177, + "step": 1317 + }, + { + "epoch": 0.6346454797159022, + "grad_norm": 0.6545950770378113, + "learning_rate": 6.24172417885762e-05, + "loss": 0.7957, + "step": 1318 + }, + { + "epoch": 0.6351270013241844, + "grad_norm": 0.7442547082901001, + "learning_rate": 6.227273537292911e-05, + "loss": 0.8306, + "step": 1319 + }, + { + "epoch": 0.6356085229324666, + "grad_norm": 0.6095592975616455, + "learning_rate": 6.212832075587891e-05, + "loss": 0.442, + "step": 1320 + }, + { + "epoch": 0.6360900445407488, + "grad_norm": 1.4066251516342163, + "learning_rate": 6.19839982888176e-05, + "loss": 0.5071, + "step": 1321 + }, + { + "epoch": 0.636571566149031, + "grad_norm": 0.7725120782852173, + "learning_rate": 6.183976832291296e-05, + "loss": 0.3988, + "step": 1322 + }, + { + "epoch": 0.6370530877573131, + "grad_norm": 0.5813517570495605, + "learning_rate": 6.169563120910775e-05, + "loss": 0.5359, + "step": 1323 + }, + { + "epoch": 0.6375346093655953, + "grad_norm": 0.9773880839347839, + "learning_rate": 6.155158729811867e-05, + "loss": 0.5878, + "step": 1324 + }, + { + "epoch": 0.6380161309738774, + "grad_norm": 0.6472037434577942, + "learning_rate": 6.140763694043578e-05, + "loss": 0.3894, + "step": 1325 + }, + { + "epoch": 0.6384976525821596, + "grad_norm": 0.5015087723731995, + "learning_rate": 6.126378048632139e-05, + "loss": 0.5544, + "step": 1326 + }, + { + "epoch": 0.6389791741904418, + "grad_norm": 0.7416954636573792, + "learning_rate": 6.112001828580944e-05, + "loss": 0.8504, + "step": 1327 + }, + { + "epoch": 0.639460695798724, + "grad_norm": 0.4432564973831177, + "learning_rate": 6.0976350688704455e-05, + "loss": 0.1723, + "step": 1328 + }, + { + "epoch": 0.6399422174070062, + "grad_norm": 0.6768189072608948, + "learning_rate": 6.083277804458072e-05, + "loss": 0.7732, + "step": 1329 + }, + { + "epoch": 0.6404237390152883, + "grad_norm": 1.28372061252594, + "learning_rate": 6.068930070278159e-05, + "loss": 0.7025, + "step": 1330 + }, + { + "epoch": 0.6409052606235704, + "grad_norm": 0.7883098125457764, + "learning_rate": 6.054591901241846e-05, + "loss": 0.4218, + "step": 1331 + }, + { + "epoch": 0.6413867822318526, + "grad_norm": 1.1983596086502075, + "learning_rate": 6.040263332237002e-05, + "loss": 0.9044, + "step": 1332 + }, + { + "epoch": 0.6418683038401348, + "grad_norm": 0.5475990772247314, + "learning_rate": 6.025944398128137e-05, + "loss": 0.5043, + "step": 1333 + }, + { + "epoch": 0.642349825448417, + "grad_norm": 1.6600192785263062, + "learning_rate": 6.011635133756309e-05, + "loss": 0.5005, + "step": 1334 + }, + { + "epoch": 0.6428313470566992, + "grad_norm": 0.6498377919197083, + "learning_rate": 5.99733557393906e-05, + "loss": 0.7945, + "step": 1335 + }, + { + "epoch": 0.6433128686649814, + "grad_norm": 0.9499074816703796, + "learning_rate": 5.983045753470308e-05, + "loss": 0.9912, + "step": 1336 + }, + { + "epoch": 0.6437943902732635, + "grad_norm": 0.6833046078681946, + "learning_rate": 5.96876570712028e-05, + "loss": 0.985, + "step": 1337 + }, + { + "epoch": 0.6442759118815456, + "grad_norm": 0.7049830555915833, + "learning_rate": 5.954495469635417e-05, + "loss": 0.5277, + "step": 1338 + }, + { + "epoch": 0.6447574334898278, + "grad_norm": 0.9056922197341919, + "learning_rate": 5.940235075738296e-05, + "loss": 0.6231, + "step": 1339 + }, + { + "epoch": 0.64523895509811, + "grad_norm": 0.8221293091773987, + "learning_rate": 5.925984560127542e-05, + "loss": 0.8753, + "step": 1340 + }, + { + "epoch": 0.6457204767063922, + "grad_norm": 0.7707056999206543, + "learning_rate": 5.911743957477739e-05, + "loss": 0.8813, + "step": 1341 + }, + { + "epoch": 0.6462019983146744, + "grad_norm": 0.6258026957511902, + "learning_rate": 5.897513302439355e-05, + "loss": 0.4881, + "step": 1342 + }, + { + "epoch": 0.6466835199229566, + "grad_norm": 0.7180628776550293, + "learning_rate": 5.883292629638651e-05, + "loss": 0.7696, + "step": 1343 + }, + { + "epoch": 0.6471650415312387, + "grad_norm": 0.7205024361610413, + "learning_rate": 5.869081973677604e-05, + "loss": 0.4843, + "step": 1344 + }, + { + "epoch": 0.6476465631395208, + "grad_norm": 0.9404179453849792, + "learning_rate": 5.8548813691338134e-05, + "loss": 0.8674, + "step": 1345 + }, + { + "epoch": 0.648128084747803, + "grad_norm": 0.9139581322669983, + "learning_rate": 5.84069085056042e-05, + "loss": 0.9279, + "step": 1346 + }, + { + "epoch": 0.6486096063560852, + "grad_norm": 0.902281641960144, + "learning_rate": 5.826510452486027e-05, + "loss": 0.6238, + "step": 1347 + }, + { + "epoch": 0.6490911279643674, + "grad_norm": 0.8244741559028625, + "learning_rate": 5.81234020941461e-05, + "loss": 0.4828, + "step": 1348 + }, + { + "epoch": 0.6495726495726496, + "grad_norm": 0.6052896976470947, + "learning_rate": 5.798180155825437e-05, + "loss": 0.5735, + "step": 1349 + }, + { + "epoch": 0.6500541711809318, + "grad_norm": 0.6693307757377625, + "learning_rate": 5.784030326172981e-05, + "loss": 0.5781, + "step": 1350 + }, + { + "epoch": 0.650535692789214, + "grad_norm": 0.7676788568496704, + "learning_rate": 5.7698907548868395e-05, + "loss": 0.5507, + "step": 1351 + }, + { + "epoch": 0.651017214397496, + "grad_norm": 0.5379273295402527, + "learning_rate": 5.755761476371653e-05, + "loss": 0.5197, + "step": 1352 + }, + { + "epoch": 0.6514987360057782, + "grad_norm": 0.7525188326835632, + "learning_rate": 5.741642525007003e-05, + "loss": 0.9615, + "step": 1353 + }, + { + "epoch": 0.6519802576140604, + "grad_norm": 1.112884521484375, + "learning_rate": 5.727533935147359e-05, + "loss": 0.7473, + "step": 1354 + }, + { + "epoch": 0.6524617792223426, + "grad_norm": 1.1446647644042969, + "learning_rate": 5.713435741121975e-05, + "loss": 0.3788, + "step": 1355 + }, + { + "epoch": 0.6529433008306248, + "grad_norm": 1.2720409631729126, + "learning_rate": 5.699347977234799e-05, + "loss": 0.616, + "step": 1356 + }, + { + "epoch": 0.653424822438907, + "grad_norm": 1.1350654363632202, + "learning_rate": 5.685270677764412e-05, + "loss": 0.6315, + "step": 1357 + }, + { + "epoch": 0.6539063440471892, + "grad_norm": 0.6351329684257507, + "learning_rate": 5.671203876963931e-05, + "loss": 0.6546, + "step": 1358 + }, + { + "epoch": 0.6543878656554712, + "grad_norm": 0.8298540711402893, + "learning_rate": 5.657147609060924e-05, + "loss": 0.7425, + "step": 1359 + }, + { + "epoch": 0.6548693872637534, + "grad_norm": 0.930432915687561, + "learning_rate": 5.643101908257333e-05, + "loss": 0.6273, + "step": 1360 + }, + { + "epoch": 0.6553509088720356, + "grad_norm": 0.4794197082519531, + "learning_rate": 5.629066808729385e-05, + "loss": 0.1168, + "step": 1361 + }, + { + "epoch": 0.6558324304803178, + "grad_norm": 0.9514833092689514, + "learning_rate": 5.6150423446275144e-05, + "loss": 0.3497, + "step": 1362 + }, + { + "epoch": 0.6563139520886, + "grad_norm": 0.6880574822425842, + "learning_rate": 5.601028550076277e-05, + "loss": 0.1992, + "step": 1363 + }, + { + "epoch": 0.6567954736968822, + "grad_norm": 0.6361804008483887, + "learning_rate": 5.587025459174271e-05, + "loss": 0.5376, + "step": 1364 + }, + { + "epoch": 0.6572769953051644, + "grad_norm": 0.7879760265350342, + "learning_rate": 5.573033105994038e-05, + "loss": 0.5568, + "step": 1365 + }, + { + "epoch": 0.6577585169134464, + "grad_norm": 0.8786200284957886, + "learning_rate": 5.559051524582002e-05, + "loss": 0.7973, + "step": 1366 + }, + { + "epoch": 0.6582400385217286, + "grad_norm": 0.602081835269928, + "learning_rate": 5.5450807489583777e-05, + "loss": 0.9517, + "step": 1367 + }, + { + "epoch": 0.6587215601300108, + "grad_norm": 0.4655710458755493, + "learning_rate": 5.531120813117085e-05, + "loss": 0.3658, + "step": 1368 + }, + { + "epoch": 0.659203081738293, + "grad_norm": 0.7563959956169128, + "learning_rate": 5.517171751025667e-05, + "loss": 0.6331, + "step": 1369 + }, + { + "epoch": 0.6596846033465752, + "grad_norm": 0.630258321762085, + "learning_rate": 5.5032335966252103e-05, + "loss": 0.3159, + "step": 1370 + }, + { + "epoch": 0.6601661249548574, + "grad_norm": 0.7362279295921326, + "learning_rate": 5.489306383830258e-05, + "loss": 0.2588, + "step": 1371 + }, + { + "epoch": 0.6606476465631396, + "grad_norm": 0.608984649181366, + "learning_rate": 5.475390146528738e-05, + "loss": 0.2641, + "step": 1372 + }, + { + "epoch": 0.6611291681714216, + "grad_norm": 0.6702209115028381, + "learning_rate": 5.461484918581858e-05, + "loss": 0.3605, + "step": 1373 + }, + { + "epoch": 0.6616106897797038, + "grad_norm": 1.0780071020126343, + "learning_rate": 5.4475907338240494e-05, + "loss": 0.3744, + "step": 1374 + }, + { + "epoch": 0.662092211387986, + "grad_norm": 1.0941399335861206, + "learning_rate": 5.43370762606287e-05, + "loss": 0.6251, + "step": 1375 + }, + { + "epoch": 0.6625737329962682, + "grad_norm": 0.6190235018730164, + "learning_rate": 5.4198356290789276e-05, + "loss": 0.6111, + "step": 1376 + }, + { + "epoch": 0.6630552546045504, + "grad_norm": 1.441415548324585, + "learning_rate": 5.405974776625785e-05, + "loss": 0.4304, + "step": 1377 + }, + { + "epoch": 0.6635367762128326, + "grad_norm": 0.496945858001709, + "learning_rate": 5.392125102429899e-05, + "loss": 0.52, + "step": 1378 + }, + { + "epoch": 0.6640182978211148, + "grad_norm": 0.6923816800117493, + "learning_rate": 5.378286640190522e-05, + "loss": 0.4308, + "step": 1379 + }, + { + "epoch": 0.6644998194293968, + "grad_norm": 0.938606858253479, + "learning_rate": 5.364459423579629e-05, + "loss": 0.3205, + "step": 1380 + }, + { + "epoch": 0.664981341037679, + "grad_norm": 0.6108879446983337, + "learning_rate": 5.350643486241825e-05, + "loss": 0.2478, + "step": 1381 + }, + { + "epoch": 0.6654628626459612, + "grad_norm": 0.5518016815185547, + "learning_rate": 5.33683886179428e-05, + "loss": 0.4628, + "step": 1382 + }, + { + "epoch": 0.6659443842542434, + "grad_norm": 0.5168072581291199, + "learning_rate": 5.3230455838266266e-05, + "loss": 0.2749, + "step": 1383 + }, + { + "epoch": 0.6664259058625256, + "grad_norm": 0.7263458967208862, + "learning_rate": 5.309263685900898e-05, + "loss": 0.2829, + "step": 1384 + }, + { + "epoch": 0.6669074274708078, + "grad_norm": 0.574898898601532, + "learning_rate": 5.295493201551433e-05, + "loss": 0.2551, + "step": 1385 + }, + { + "epoch": 0.66738894907909, + "grad_norm": 0.9553223848342896, + "learning_rate": 5.281734164284802e-05, + "loss": 0.6584, + "step": 1386 + }, + { + "epoch": 0.667870470687372, + "grad_norm": 1.0642074346542358, + "learning_rate": 5.26798660757971e-05, + "loss": 0.9263, + "step": 1387 + }, + { + "epoch": 0.6683519922956542, + "grad_norm": 0.9214754104614258, + "learning_rate": 5.2542505648869434e-05, + "loss": 0.6928, + "step": 1388 + }, + { + "epoch": 0.6688335139039364, + "grad_norm": 0.7935864329338074, + "learning_rate": 5.240526069629265e-05, + "loss": 0.7917, + "step": 1389 + }, + { + "epoch": 0.6693150355122186, + "grad_norm": 0.8348840475082397, + "learning_rate": 5.22681315520134e-05, + "loss": 1.1228, + "step": 1390 + }, + { + "epoch": 0.6697965571205008, + "grad_norm": 1.2822461128234863, + "learning_rate": 5.213111854969661e-05, + "loss": 0.4916, + "step": 1391 + }, + { + "epoch": 0.670278078728783, + "grad_norm": 0.6782784461975098, + "learning_rate": 5.199422202272448e-05, + "loss": 0.3989, + "step": 1392 + }, + { + "epoch": 0.6707596003370652, + "grad_norm": 1.1376097202301025, + "learning_rate": 5.185744230419589e-05, + "loss": 0.6257, + "step": 1393 + }, + { + "epoch": 0.6712411219453474, + "grad_norm": 1.124526858329773, + "learning_rate": 5.172077972692553e-05, + "loss": 0.7778, + "step": 1394 + }, + { + "epoch": 0.6717226435536294, + "grad_norm": 0.7318700551986694, + "learning_rate": 5.1584234623442974e-05, + "loss": 0.7091, + "step": 1395 + }, + { + "epoch": 0.6722041651619116, + "grad_norm": 0.3620985448360443, + "learning_rate": 5.1447807325992025e-05, + "loss": 0.1584, + "step": 1396 + }, + { + "epoch": 0.6726856867701938, + "grad_norm": 0.6356948614120483, + "learning_rate": 5.13114981665298e-05, + "loss": 0.9857, + "step": 1397 + }, + { + "epoch": 0.673167208378476, + "grad_norm": 0.6141039729118347, + "learning_rate": 5.117530747672603e-05, + "loss": 0.7072, + "step": 1398 + }, + { + "epoch": 0.6736487299867582, + "grad_norm": 0.4410860240459442, + "learning_rate": 5.103923558796203e-05, + "loss": 0.5948, + "step": 1399 + }, + { + "epoch": 0.6741302515950404, + "grad_norm": 0.7041406631469727, + "learning_rate": 5.090328283133019e-05, + "loss": 0.3836, + "step": 1400 + }, + { + "epoch": 0.6746117732033226, + "grad_norm": 0.596027135848999, + "learning_rate": 5.0767449537632986e-05, + "loss": 0.2445, + "step": 1401 + }, + { + "epoch": 0.6750932948116046, + "grad_norm": 0.6667171716690063, + "learning_rate": 5.06317360373822e-05, + "loss": 0.4582, + "step": 1402 + }, + { + "epoch": 0.6755748164198868, + "grad_norm": 0.6862946152687073, + "learning_rate": 5.049614266079813e-05, + "loss": 0.3222, + "step": 1403 + }, + { + "epoch": 0.676056338028169, + "grad_norm": 0.5457684397697449, + "learning_rate": 5.036066973780882e-05, + "loss": 0.5482, + "step": 1404 + }, + { + "epoch": 0.6765378596364512, + "grad_norm": 0.7717635631561279, + "learning_rate": 5.022531759804918e-05, + "loss": 0.5445, + "step": 1405 + }, + { + "epoch": 0.6770193812447334, + "grad_norm": 0.994358241558075, + "learning_rate": 5.009008657086025e-05, + "loss": 0.826, + "step": 1406 + }, + { + "epoch": 0.6775009028530156, + "grad_norm": 0.8234153985977173, + "learning_rate": 4.9954976985288395e-05, + "loss": 0.3035, + "step": 1407 + }, + { + "epoch": 0.6779824244612978, + "grad_norm": 1.1973003149032593, + "learning_rate": 4.981998917008448e-05, + "loss": 0.406, + "step": 1408 + }, + { + "epoch": 0.6784639460695798, + "grad_norm": 0.8367493748664856, + "learning_rate": 4.9685123453703e-05, + "loss": 0.7669, + "step": 1409 + }, + { + "epoch": 0.678945467677862, + "grad_norm": 0.4213549494743347, + "learning_rate": 4.955038016430149e-05, + "loss": 0.1896, + "step": 1410 + }, + { + "epoch": 0.6794269892861442, + "grad_norm": 0.8429536819458008, + "learning_rate": 4.9415759629739455e-05, + "loss": 0.532, + "step": 1411 + }, + { + "epoch": 0.6799085108944264, + "grad_norm": 1.1326082944869995, + "learning_rate": 4.928126217757782e-05, + "loss": 1.2761, + "step": 1412 + }, + { + "epoch": 0.6803900325027086, + "grad_norm": 0.42613062262535095, + "learning_rate": 4.914688813507797e-05, + "loss": 0.6146, + "step": 1413 + }, + { + "epoch": 0.6808715541109908, + "grad_norm": 0.6152105927467346, + "learning_rate": 4.901263782920105e-05, + "loss": 0.6455, + "step": 1414 + }, + { + "epoch": 0.681353075719273, + "grad_norm": 0.9997090101242065, + "learning_rate": 4.887851158660706e-05, + "loss": 0.307, + "step": 1415 + }, + { + "epoch": 0.681834597327555, + "grad_norm": 0.8020671606063843, + "learning_rate": 4.8744509733654184e-05, + "loss": 0.7692, + "step": 1416 + }, + { + "epoch": 0.6823161189358372, + "grad_norm": 1.184843897819519, + "learning_rate": 4.861063259639793e-05, + "loss": 0.714, + "step": 1417 + }, + { + "epoch": 0.6827976405441194, + "grad_norm": 0.5545733571052551, + "learning_rate": 4.847688050059033e-05, + "loss": 0.3846, + "step": 1418 + }, + { + "epoch": 0.6832791621524016, + "grad_norm": 0.7227141857147217, + "learning_rate": 4.8343253771679155e-05, + "loss": 0.4455, + "step": 1419 + }, + { + "epoch": 0.6837606837606838, + "grad_norm": 0.6209794878959656, + "learning_rate": 4.82097527348072e-05, + "loss": 1.0001, + "step": 1420 + }, + { + "epoch": 0.684242205368966, + "grad_norm": 0.47615864872932434, + "learning_rate": 4.8076377714811284e-05, + "loss": 0.5838, + "step": 1421 + }, + { + "epoch": 0.6847237269772481, + "grad_norm": 0.4865884780883789, + "learning_rate": 4.7943129036221735e-05, + "loss": 0.419, + "step": 1422 + }, + { + "epoch": 0.6852052485855302, + "grad_norm": 0.5061058402061462, + "learning_rate": 4.781000702326142e-05, + "loss": 0.2144, + "step": 1423 + }, + { + "epoch": 0.6856867701938124, + "grad_norm": 0.9223092794418335, + "learning_rate": 4.767701199984497e-05, + "loss": 0.4185, + "step": 1424 + }, + { + "epoch": 0.6861682918020946, + "grad_norm": 0.6437268853187561, + "learning_rate": 4.7544144289578066e-05, + "loss": 0.4223, + "step": 1425 + }, + { + "epoch": 0.6866498134103768, + "grad_norm": 0.6153153777122498, + "learning_rate": 4.7411404215756594e-05, + "loss": 0.335, + "step": 1426 + }, + { + "epoch": 0.687131335018659, + "grad_norm": 1.2971034049987793, + "learning_rate": 4.7278792101365866e-05, + "loss": 0.7627, + "step": 1427 + }, + { + "epoch": 0.6876128566269412, + "grad_norm": 0.7124690413475037, + "learning_rate": 4.714630826907985e-05, + "loss": 0.6546, + "step": 1428 + }, + { + "epoch": 0.6880943782352233, + "grad_norm": 1.0929278135299683, + "learning_rate": 4.701395304126038e-05, + "loss": 0.8798, + "step": 1429 + }, + { + "epoch": 0.6885758998435054, + "grad_norm": 0.6752956509590149, + "learning_rate": 4.6881726739956375e-05, + "loss": 0.5457, + "step": 1430 + }, + { + "epoch": 0.6890574214517876, + "grad_norm": 0.9237945675849915, + "learning_rate": 4.6749629686902984e-05, + "loss": 1.2754, + "step": 1431 + }, + { + "epoch": 0.6895389430600698, + "grad_norm": 0.6456115245819092, + "learning_rate": 4.661766220352097e-05, + "loss": 0.2924, + "step": 1432 + }, + { + "epoch": 0.690020464668352, + "grad_norm": 1.0163204669952393, + "learning_rate": 4.64858246109157e-05, + "loss": 1.2191, + "step": 1433 + }, + { + "epoch": 0.6905019862766342, + "grad_norm": 0.9082103967666626, + "learning_rate": 4.63541172298766e-05, + "loss": 0.7933, + "step": 1434 + }, + { + "epoch": 0.6909835078849164, + "grad_norm": 0.5109544992446899, + "learning_rate": 4.622254038087622e-05, + "loss": 0.3987, + "step": 1435 + }, + { + "epoch": 0.6914650294931985, + "grad_norm": 0.88107830286026, + "learning_rate": 4.60910943840695e-05, + "loss": 0.586, + "step": 1436 + }, + { + "epoch": 0.6919465511014807, + "grad_norm": 0.6991325616836548, + "learning_rate": 4.5959779559292985e-05, + "loss": 0.3088, + "step": 1437 + }, + { + "epoch": 0.6924280727097628, + "grad_norm": 0.7649497985839844, + "learning_rate": 4.582859622606406e-05, + "loss": 0.5081, + "step": 1438 + }, + { + "epoch": 0.692909594318045, + "grad_norm": 1.080644965171814, + "learning_rate": 4.569754470358014e-05, + "loss": 0.8645, + "step": 1439 + }, + { + "epoch": 0.6933911159263272, + "grad_norm": 0.8446236252784729, + "learning_rate": 4.556662531071796e-05, + "loss": 0.3782, + "step": 1440 + }, + { + "epoch": 0.6938726375346094, + "grad_norm": 0.8029199242591858, + "learning_rate": 4.54358383660327e-05, + "loss": 0.6237, + "step": 1441 + }, + { + "epoch": 0.6943541591428916, + "grad_norm": 0.8086987137794495, + "learning_rate": 4.530518418775733e-05, + "loss": 0.4498, + "step": 1442 + }, + { + "epoch": 0.6948356807511737, + "grad_norm": 1.1346222162246704, + "learning_rate": 4.5174663093801674e-05, + "loss": 0.7626, + "step": 1443 + }, + { + "epoch": 0.6953172023594559, + "grad_norm": 0.5477440357208252, + "learning_rate": 4.504427540175181e-05, + "loss": 0.3377, + "step": 1444 + }, + { + "epoch": 0.695798723967738, + "grad_norm": 0.37069225311279297, + "learning_rate": 4.491402142886922e-05, + "loss": 0.2361, + "step": 1445 + }, + { + "epoch": 0.6962802455760202, + "grad_norm": 1.2140967845916748, + "learning_rate": 4.4783901492089984e-05, + "loss": 0.7769, + "step": 1446 + }, + { + "epoch": 0.6967617671843024, + "grad_norm": 0.42369186878204346, + "learning_rate": 4.465391590802407e-05, + "loss": 0.403, + "step": 1447 + }, + { + "epoch": 0.6972432887925846, + "grad_norm": 1.4413726329803467, + "learning_rate": 4.4524064992954516e-05, + "loss": 0.9888, + "step": 1448 + }, + { + "epoch": 0.6977248104008668, + "grad_norm": 0.7362139821052551, + "learning_rate": 4.4394349062836736e-05, + "loss": 0.7057, + "step": 1449 + }, + { + "epoch": 0.6982063320091489, + "grad_norm": 0.700434148311615, + "learning_rate": 4.4264768433297565e-05, + "loss": 0.385, + "step": 1450 + }, + { + "epoch": 0.6986878536174311, + "grad_norm": 1.0879504680633545, + "learning_rate": 4.4135323419634766e-05, + "loss": 1.1829, + "step": 1451 + }, + { + "epoch": 0.6991693752257132, + "grad_norm": 1.2308809757232666, + "learning_rate": 4.4006014336816035e-05, + "loss": 1.0978, + "step": 1452 + }, + { + "epoch": 0.6996508968339954, + "grad_norm": 0.6011996865272522, + "learning_rate": 4.387684149947837e-05, + "loss": 0.362, + "step": 1453 + }, + { + "epoch": 0.7001324184422776, + "grad_norm": 1.436378002166748, + "learning_rate": 4.374780522192726e-05, + "loss": 1.3376, + "step": 1454 + }, + { + "epoch": 0.7006139400505598, + "grad_norm": 0.7376883625984192, + "learning_rate": 4.3618905818135805e-05, + "loss": 0.3268, + "step": 1455 + }, + { + "epoch": 0.701095461658842, + "grad_norm": 0.6043660044670105, + "learning_rate": 4.349014360174417e-05, + "loss": 0.6669, + "step": 1456 + }, + { + "epoch": 0.7015769832671241, + "grad_norm": 0.7776713371276855, + "learning_rate": 4.336151888605871e-05, + "loss": 0.4488, + "step": 1457 + }, + { + "epoch": 0.7020585048754063, + "grad_norm": 0.7803006172180176, + "learning_rate": 4.323303198405117e-05, + "loss": 0.8023, + "step": 1458 + }, + { + "epoch": 0.7025400264836884, + "grad_norm": 0.7615090608596802, + "learning_rate": 4.310468320835796e-05, + "loss": 0.5186, + "step": 1459 + }, + { + "epoch": 0.7030215480919706, + "grad_norm": 0.9116225242614746, + "learning_rate": 4.297647287127946e-05, + "loss": 0.3674, + "step": 1460 + }, + { + "epoch": 0.7035030697002528, + "grad_norm": 0.9148691892623901, + "learning_rate": 4.284840128477913e-05, + "loss": 0.5605, + "step": 1461 + }, + { + "epoch": 0.703984591308535, + "grad_norm": 0.7424203157424927, + "learning_rate": 4.2720468760482854e-05, + "loss": 0.4504, + "step": 1462 + }, + { + "epoch": 0.7044661129168172, + "grad_norm": 0.6610391139984131, + "learning_rate": 4.2592675609678135e-05, + "loss": 0.3636, + "step": 1463 + }, + { + "epoch": 0.7049476345250993, + "grad_norm": 0.6278050541877747, + "learning_rate": 4.24650221433134e-05, + "loss": 0.529, + "step": 1464 + }, + { + "epoch": 0.7054291561333815, + "grad_norm": 0.5832677483558655, + "learning_rate": 4.2337508671997086e-05, + "loss": 0.3057, + "step": 1465 + }, + { + "epoch": 0.7059106777416636, + "grad_norm": 0.7446826696395874, + "learning_rate": 4.221013550599707e-05, + "loss": 0.3268, + "step": 1466 + }, + { + "epoch": 0.7063921993499458, + "grad_norm": 0.9066232442855835, + "learning_rate": 4.208290295523984e-05, + "loss": 0.624, + "step": 1467 + }, + { + "epoch": 0.706873720958228, + "grad_norm": 1.0170692205429077, + "learning_rate": 4.1955811329309746e-05, + "loss": 0.6293, + "step": 1468 + }, + { + "epoch": 0.7073552425665102, + "grad_norm": 0.7351876497268677, + "learning_rate": 4.182886093744813e-05, + "loss": 0.9277, + "step": 1469 + }, + { + "epoch": 0.7078367641747924, + "grad_norm": 0.7814710140228271, + "learning_rate": 4.170205208855281e-05, + "loss": 0.4777, + "step": 1470 + }, + { + "epoch": 0.7083182857830745, + "grad_norm": 0.8738664984703064, + "learning_rate": 4.157538509117714e-05, + "loss": 0.7923, + "step": 1471 + }, + { + "epoch": 0.7087998073913567, + "grad_norm": 0.9012139439582825, + "learning_rate": 4.144886025352934e-05, + "loss": 0.6325, + "step": 1472 + }, + { + "epoch": 0.7092813289996388, + "grad_norm": 1.0267746448516846, + "learning_rate": 4.13224778834717e-05, + "loss": 0.5129, + "step": 1473 + }, + { + "epoch": 0.709762850607921, + "grad_norm": 0.795587956905365, + "learning_rate": 4.1196238288519874e-05, + "loss": 0.4198, + "step": 1474 + }, + { + "epoch": 0.7102443722162032, + "grad_norm": 0.9339183568954468, + "learning_rate": 4.107014177584211e-05, + "loss": 0.4369, + "step": 1475 + }, + { + "epoch": 0.7107258938244854, + "grad_norm": 0.6808327436447144, + "learning_rate": 4.094418865225853e-05, + "loss": 0.3405, + "step": 1476 + }, + { + "epoch": 0.7112074154327676, + "grad_norm": 0.9060590863227844, + "learning_rate": 4.081837922424027e-05, + "loss": 0.43, + "step": 1477 + }, + { + "epoch": 0.7116889370410497, + "grad_norm": 0.4485381543636322, + "learning_rate": 4.069271379790891e-05, + "loss": 0.2424, + "step": 1478 + }, + { + "epoch": 0.7121704586493319, + "grad_norm": 0.6939108967781067, + "learning_rate": 4.0567192679035636e-05, + "loss": 0.4309, + "step": 1479 + }, + { + "epoch": 0.7126519802576141, + "grad_norm": 0.8871076107025146, + "learning_rate": 4.044181617304048e-05, + "loss": 0.3657, + "step": 1480 + }, + { + "epoch": 0.7131335018658962, + "grad_norm": 0.6109139323234558, + "learning_rate": 4.03165845849916e-05, + "loss": 0.3523, + "step": 1481 + }, + { + "epoch": 0.7136150234741784, + "grad_norm": 0.755155622959137, + "learning_rate": 4.019149821960455e-05, + "loss": 0.5116, + "step": 1482 + }, + { + "epoch": 0.7140965450824606, + "grad_norm": 0.684121310710907, + "learning_rate": 4.006655738124152e-05, + "loss": 0.4508, + "step": 1483 + }, + { + "epoch": 0.7145780666907428, + "grad_norm": 0.7652056813240051, + "learning_rate": 3.9941762373910586e-05, + "loss": 0.4199, + "step": 1484 + }, + { + "epoch": 0.7150595882990249, + "grad_norm": 0.8583548069000244, + "learning_rate": 3.9817113501265016e-05, + "loss": 0.6258, + "step": 1485 + }, + { + "epoch": 0.7155411099073071, + "grad_norm": 0.8159501552581787, + "learning_rate": 3.9692611066602516e-05, + "loss": 0.6959, + "step": 1486 + }, + { + "epoch": 0.7160226315155893, + "grad_norm": 0.7192055583000183, + "learning_rate": 3.956825537286436e-05, + "loss": 0.2499, + "step": 1487 + }, + { + "epoch": 0.7165041531238714, + "grad_norm": 0.5423712134361267, + "learning_rate": 3.944404672263494e-05, + "loss": 0.4917, + "step": 1488 + }, + { + "epoch": 0.7169856747321536, + "grad_norm": 0.9170364737510681, + "learning_rate": 3.931998541814069e-05, + "loss": 0.898, + "step": 1489 + }, + { + "epoch": 0.7174671963404358, + "grad_norm": 1.1975452899932861, + "learning_rate": 3.919607176124966e-05, + "loss": 1.2809, + "step": 1490 + }, + { + "epoch": 0.717948717948718, + "grad_norm": 0.5959516167640686, + "learning_rate": 3.9072306053470566e-05, + "loss": 0.3478, + "step": 1491 + }, + { + "epoch": 0.7184302395570001, + "grad_norm": 0.7946900129318237, + "learning_rate": 3.8948688595952164e-05, + "loss": 0.1685, + "step": 1492 + }, + { + "epoch": 0.7189117611652823, + "grad_norm": 1.519437551498413, + "learning_rate": 3.882521968948246e-05, + "loss": 0.7702, + "step": 1493 + }, + { + "epoch": 0.7193932827735645, + "grad_norm": 0.7796205878257751, + "learning_rate": 3.8701899634488014e-05, + "loss": 0.6952, + "step": 1494 + }, + { + "epoch": 0.7198748043818466, + "grad_norm": 0.7571779489517212, + "learning_rate": 3.857872873103322e-05, + "loss": 0.3655, + "step": 1495 + }, + { + "epoch": 0.7203563259901288, + "grad_norm": 0.4898691177368164, + "learning_rate": 3.8455707278819507e-05, + "loss": 0.371, + "step": 1496 + }, + { + "epoch": 0.720837847598411, + "grad_norm": 0.4889354407787323, + "learning_rate": 3.833283557718471e-05, + "loss": 0.2633, + "step": 1497 + }, + { + "epoch": 0.7213193692066932, + "grad_norm": 0.5309864282608032, + "learning_rate": 3.821011392510228e-05, + "loss": 0.2295, + "step": 1498 + }, + { + "epoch": 0.7218008908149753, + "grad_norm": 0.5448015928268433, + "learning_rate": 3.808754262118046e-05, + "loss": 0.2642, + "step": 1499 + }, + { + "epoch": 0.7222824124232575, + "grad_norm": 0.9109936952590942, + "learning_rate": 3.796512196366182e-05, + "loss": 0.3738, + "step": 1500 + }, + { + "epoch": 0.7227639340315397, + "grad_norm": 0.8303773999214172, + "learning_rate": 3.784285225042229e-05, + "loss": 0.7826, + "step": 1501 + }, + { + "epoch": 0.7232454556398218, + "grad_norm": 0.9534183144569397, + "learning_rate": 3.772073377897052e-05, + "loss": 0.6641, + "step": 1502 + }, + { + "epoch": 0.723726977248104, + "grad_norm": 1.6140865087509155, + "learning_rate": 3.7598766846447184e-05, + "loss": 0.5608, + "step": 1503 + }, + { + "epoch": 0.7242084988563862, + "grad_norm": 0.47576579451560974, + "learning_rate": 3.747695174962423e-05, + "loss": 0.4865, + "step": 1504 + }, + { + "epoch": 0.7246900204646683, + "grad_norm": 0.8235217928886414, + "learning_rate": 3.7355288784904116e-05, + "loss": 0.4982, + "step": 1505 + }, + { + "epoch": 0.7251715420729505, + "grad_norm": 0.9987104535102844, + "learning_rate": 3.7233778248319176e-05, + "loss": 0.4883, + "step": 1506 + }, + { + "epoch": 0.7256530636812327, + "grad_norm": 0.8933877348899841, + "learning_rate": 3.7112420435530845e-05, + "loss": 0.511, + "step": 1507 + }, + { + "epoch": 0.7261345852895149, + "grad_norm": 0.5606909394264221, + "learning_rate": 3.69912156418289e-05, + "loss": 0.2032, + "step": 1508 + }, + { + "epoch": 0.726616106897797, + "grad_norm": 0.7086384296417236, + "learning_rate": 3.687016416213084e-05, + "loss": 0.5728, + "step": 1509 + }, + { + "epoch": 0.7270976285060792, + "grad_norm": 0.6372523903846741, + "learning_rate": 3.674926629098113e-05, + "loss": 0.1722, + "step": 1510 + }, + { + "epoch": 0.7275791501143614, + "grad_norm": 0.7619569301605225, + "learning_rate": 3.6628522322550394e-05, + "loss": 0.2178, + "step": 1511 + }, + { + "epoch": 0.7280606717226435, + "grad_norm": 0.811999499797821, + "learning_rate": 3.6507932550634846e-05, + "loss": 0.5692, + "step": 1512 + }, + { + "epoch": 0.7285421933309257, + "grad_norm": 0.6715248227119446, + "learning_rate": 3.638749726865552e-05, + "loss": 0.1933, + "step": 1513 + }, + { + "epoch": 0.7290237149392079, + "grad_norm": 0.4963141679763794, + "learning_rate": 3.6267216769657485e-05, + "loss": 0.4058, + "step": 1514 + }, + { + "epoch": 0.7295052365474901, + "grad_norm": 0.7091789245605469, + "learning_rate": 3.6147091346309224e-05, + "loss": 0.846, + "step": 1515 + }, + { + "epoch": 0.7299867581557722, + "grad_norm": 0.6115385293960571, + "learning_rate": 3.602712129090189e-05, + "loss": 0.3185, + "step": 1516 + }, + { + "epoch": 0.7304682797640544, + "grad_norm": 0.5440830588340759, + "learning_rate": 3.590730689534857e-05, + "loss": 0.3759, + "step": 1517 + }, + { + "epoch": 0.7309498013723366, + "grad_norm": 0.5895819664001465, + "learning_rate": 3.578764845118362e-05, + "loss": 0.3927, + "step": 1518 + }, + { + "epoch": 0.7314313229806187, + "grad_norm": 1.0050193071365356, + "learning_rate": 3.566814624956194e-05, + "loss": 0.4598, + "step": 1519 + }, + { + "epoch": 0.7319128445889009, + "grad_norm": 0.67363440990448, + "learning_rate": 3.554880058125819e-05, + "loss": 0.5574, + "step": 1520 + }, + { + "epoch": 0.7323943661971831, + "grad_norm": 1.1256099939346313, + "learning_rate": 3.5429611736666235e-05, + "loss": 0.6292, + "step": 1521 + }, + { + "epoch": 0.7328758878054653, + "grad_norm": 0.5500608086585999, + "learning_rate": 3.53105800057983e-05, + "loss": 0.2614, + "step": 1522 + }, + { + "epoch": 0.7333574094137475, + "grad_norm": 0.6335020065307617, + "learning_rate": 3.519170567828435e-05, + "loss": 0.5855, + "step": 1523 + }, + { + "epoch": 0.7338389310220296, + "grad_norm": 1.045906901359558, + "learning_rate": 3.507298904337134e-05, + "loss": 0.763, + "step": 1524 + }, + { + "epoch": 0.7343204526303118, + "grad_norm": 0.7786596417427063, + "learning_rate": 3.495443038992253e-05, + "loss": 0.4323, + "step": 1525 + }, + { + "epoch": 0.734801974238594, + "grad_norm": 0.8979135155677795, + "learning_rate": 3.4836030006416775e-05, + "loss": 0.424, + "step": 1526 + }, + { + "epoch": 0.7352834958468761, + "grad_norm": 0.8464924097061157, + "learning_rate": 3.471778818094785e-05, + "loss": 0.6309, + "step": 1527 + }, + { + "epoch": 0.7357650174551583, + "grad_norm": 0.38327470421791077, + "learning_rate": 3.459970520122364e-05, + "loss": 0.2277, + "step": 1528 + }, + { + "epoch": 0.7362465390634405, + "grad_norm": 1.0083869695663452, + "learning_rate": 3.44817813545656e-05, + "loss": 0.4736, + "step": 1529 + }, + { + "epoch": 0.7367280606717227, + "grad_norm": 0.7795249819755554, + "learning_rate": 3.4364016927907974e-05, + "loss": 0.329, + "step": 1530 + }, + { + "epoch": 0.7372095822800048, + "grad_norm": 0.6777287125587463, + "learning_rate": 3.424641220779711e-05, + "loss": 0.5657, + "step": 1531 + }, + { + "epoch": 0.737691103888287, + "grad_norm": 0.6228256821632385, + "learning_rate": 3.412896748039067e-05, + "loss": 0.4603, + "step": 1532 + }, + { + "epoch": 0.7381726254965691, + "grad_norm": 0.5707159042358398, + "learning_rate": 3.401168303145713e-05, + "loss": 0.4364, + "step": 1533 + }, + { + "epoch": 0.7386541471048513, + "grad_norm": 0.6516470909118652, + "learning_rate": 3.3894559146374924e-05, + "loss": 0.3653, + "step": 1534 + }, + { + "epoch": 0.7391356687131335, + "grad_norm": 0.8824191093444824, + "learning_rate": 3.37775961101318e-05, + "loss": 0.8059, + "step": 1535 + }, + { + "epoch": 0.7396171903214157, + "grad_norm": 0.6240821480751038, + "learning_rate": 3.366079420732413e-05, + "loss": 0.5367, + "step": 1536 + }, + { + "epoch": 0.7400987119296979, + "grad_norm": 1.1601349115371704, + "learning_rate": 3.3544153722156216e-05, + "loss": 0.5936, + "step": 1537 + }, + { + "epoch": 0.74058023353798, + "grad_norm": 0.4957769811153412, + "learning_rate": 3.3427674938439594e-05, + "loss": 0.5746, + "step": 1538 + }, + { + "epoch": 0.7410617551462622, + "grad_norm": 0.3493504226207733, + "learning_rate": 3.3311358139592317e-05, + "loss": 0.3965, + "step": 1539 + }, + { + "epoch": 0.7415432767545443, + "grad_norm": 0.7727856636047363, + "learning_rate": 3.319520360863837e-05, + "loss": 0.5924, + "step": 1540 + }, + { + "epoch": 0.7420247983628265, + "grad_norm": 0.8370104432106018, + "learning_rate": 3.3079211628206854e-05, + "loss": 0.3911, + "step": 1541 + }, + { + "epoch": 0.7425063199711087, + "grad_norm": 1.2629445791244507, + "learning_rate": 3.296338248053129e-05, + "loss": 0.3831, + "step": 1542 + }, + { + "epoch": 0.7429878415793909, + "grad_norm": 0.7771987915039062, + "learning_rate": 3.2847716447449096e-05, + "loss": 0.5615, + "step": 1543 + }, + { + "epoch": 0.7434693631876731, + "grad_norm": 0.7107818722724915, + "learning_rate": 3.2732213810400745e-05, + "loss": 0.4814, + "step": 1544 + }, + { + "epoch": 0.7439508847959552, + "grad_norm": 0.7308077812194824, + "learning_rate": 3.261687485042915e-05, + "loss": 0.4592, + "step": 1545 + }, + { + "epoch": 0.7444324064042374, + "grad_norm": 1.0506657361984253, + "learning_rate": 3.250169984817897e-05, + "loss": 0.5339, + "step": 1546 + }, + { + "epoch": 0.7449139280125195, + "grad_norm": 1.354854702949524, + "learning_rate": 3.238668908389586e-05, + "loss": 0.7969, + "step": 1547 + }, + { + "epoch": 0.7453954496208017, + "grad_norm": 0.7732512354850769, + "learning_rate": 3.227184283742591e-05, + "loss": 0.786, + "step": 1548 + }, + { + "epoch": 0.7458769712290839, + "grad_norm": 1.486000895500183, + "learning_rate": 3.215716138821488e-05, + "loss": 0.3588, + "step": 1549 + }, + { + "epoch": 0.7463584928373661, + "grad_norm": 0.7014599442481995, + "learning_rate": 3.204264501530756e-05, + "loss": 0.5279, + "step": 1550 + }, + { + "epoch": 0.7468400144456483, + "grad_norm": 1.0592234134674072, + "learning_rate": 3.192829399734706e-05, + "loss": 0.7584, + "step": 1551 + }, + { + "epoch": 0.7473215360539304, + "grad_norm": 0.7656548023223877, + "learning_rate": 3.181410861257413e-05, + "loss": 0.8036, + "step": 1552 + }, + { + "epoch": 0.7478030576622126, + "grad_norm": 2.293597936630249, + "learning_rate": 3.170008913882656e-05, + "loss": 0.6918, + "step": 1553 + }, + { + "epoch": 0.7482845792704947, + "grad_norm": 0.6488149762153625, + "learning_rate": 3.1586235853538325e-05, + "loss": 0.7967, + "step": 1554 + }, + { + "epoch": 0.7487661008787769, + "grad_norm": 1.1204566955566406, + "learning_rate": 3.1472549033739126e-05, + "loss": 0.4146, + "step": 1555 + }, + { + "epoch": 0.7492476224870591, + "grad_norm": 0.6003812551498413, + "learning_rate": 3.1359028956053615e-05, + "loss": 0.3162, + "step": 1556 + }, + { + "epoch": 0.7497291440953413, + "grad_norm": 1.8639508485794067, + "learning_rate": 3.1245675896700685e-05, + "loss": 1.1739, + "step": 1557 + }, + { + "epoch": 0.7502106657036235, + "grad_norm": 0.7486677169799805, + "learning_rate": 3.113249013149284e-05, + "loss": 0.1952, + "step": 1558 + }, + { + "epoch": 0.7506921873119056, + "grad_norm": 1.239945888519287, + "learning_rate": 3.101947193583557e-05, + "loss": 0.754, + "step": 1559 + }, + { + "epoch": 0.7511737089201878, + "grad_norm": 1.0496931076049805, + "learning_rate": 3.0906621584726546e-05, + "loss": 0.7861, + "step": 1560 + }, + { + "epoch": 0.7516552305284699, + "grad_norm": 0.981391191482544, + "learning_rate": 3.079393935275513e-05, + "loss": 0.4328, + "step": 1561 + }, + { + "epoch": 0.7521367521367521, + "grad_norm": 1.4165148735046387, + "learning_rate": 3.068142551410155e-05, + "loss": 0.398, + "step": 1562 + }, + { + "epoch": 0.7526182737450343, + "grad_norm": 0.9278653860092163, + "learning_rate": 3.0569080342536347e-05, + "loss": 0.6272, + "step": 1563 + }, + { + "epoch": 0.7530997953533165, + "grad_norm": 1.465111255645752, + "learning_rate": 3.0456904111419572e-05, + "loss": 1.0547, + "step": 1564 + }, + { + "epoch": 0.7535813169615987, + "grad_norm": 0.7694311738014221, + "learning_rate": 3.034489709370033e-05, + "loss": 0.7232, + "step": 1565 + }, + { + "epoch": 0.7540628385698809, + "grad_norm": 0.9846645593643188, + "learning_rate": 3.0233059561915855e-05, + "loss": 0.5923, + "step": 1566 + }, + { + "epoch": 0.754544360178163, + "grad_norm": 0.9460967779159546, + "learning_rate": 3.01213917881911e-05, + "loss": 0.8232, + "step": 1567 + }, + { + "epoch": 0.7550258817864451, + "grad_norm": 0.761855959892273, + "learning_rate": 3.0009894044237907e-05, + "loss": 0.6175, + "step": 1568 + }, + { + "epoch": 0.7555074033947273, + "grad_norm": 0.5922994017601013, + "learning_rate": 2.9898566601354418e-05, + "loss": 0.6428, + "step": 1569 + }, + { + "epoch": 0.7559889250030095, + "grad_norm": 0.5083173513412476, + "learning_rate": 2.9787409730424374e-05, + "loss": 0.4254, + "step": 1570 + }, + { + "epoch": 0.7564704466112917, + "grad_norm": 0.6428760886192322, + "learning_rate": 2.96764237019165e-05, + "loss": 0.4674, + "step": 1571 + }, + { + "epoch": 0.7569519682195739, + "grad_norm": 0.906851053237915, + "learning_rate": 2.9565608785883815e-05, + "loss": 0.9182, + "step": 1572 + }, + { + "epoch": 0.7574334898278561, + "grad_norm": 0.6919524073600769, + "learning_rate": 2.9454965251962973e-05, + "loss": 0.7283, + "step": 1573 + }, + { + "epoch": 0.7579150114361382, + "grad_norm": 0.5480821132659912, + "learning_rate": 2.9344493369373637e-05, + "loss": 0.4671, + "step": 1574 + }, + { + "epoch": 0.7583965330444203, + "grad_norm": 0.929789125919342, + "learning_rate": 2.9234193406917833e-05, + "loss": 0.5857, + "step": 1575 + }, + { + "epoch": 0.7588780546527025, + "grad_norm": 1.0483143329620361, + "learning_rate": 2.912406563297916e-05, + "loss": 0.7638, + "step": 1576 + }, + { + "epoch": 0.7593595762609847, + "grad_norm": 0.7521671056747437, + "learning_rate": 2.901411031552236e-05, + "loss": 0.5069, + "step": 1577 + }, + { + "epoch": 0.7598410978692669, + "grad_norm": 0.6959591507911682, + "learning_rate": 2.8904327722092495e-05, + "loss": 0.898, + "step": 1578 + }, + { + "epoch": 0.7603226194775491, + "grad_norm": 0.9434571862220764, + "learning_rate": 2.879471811981437e-05, + "loss": 0.8069, + "step": 1579 + }, + { + "epoch": 0.7608041410858313, + "grad_norm": 0.9961397051811218, + "learning_rate": 2.868528177539187e-05, + "loss": 0.8701, + "step": 1580 + }, + { + "epoch": 0.7612856626941134, + "grad_norm": 0.8834352493286133, + "learning_rate": 2.8576018955107285e-05, + "loss": 0.7611, + "step": 1581 + }, + { + "epoch": 0.7617671843023955, + "grad_norm": 0.7454970479011536, + "learning_rate": 2.8466929924820705e-05, + "loss": 0.9264, + "step": 1582 + }, + { + "epoch": 0.7622487059106777, + "grad_norm": 0.5560579299926758, + "learning_rate": 2.8358014949969334e-05, + "loss": 0.6513, + "step": 1583 + }, + { + "epoch": 0.7627302275189599, + "grad_norm": 0.793477475643158, + "learning_rate": 2.8249274295566864e-05, + "loss": 0.4191, + "step": 1584 + }, + { + "epoch": 0.7632117491272421, + "grad_norm": 0.7971614003181458, + "learning_rate": 2.8140708226202884e-05, + "loss": 0.607, + "step": 1585 + }, + { + "epoch": 0.7636932707355243, + "grad_norm": 0.9233664870262146, + "learning_rate": 2.803231700604204e-05, + "loss": 0.3923, + "step": 1586 + }, + { + "epoch": 0.7641747923438065, + "grad_norm": 0.8491899967193604, + "learning_rate": 2.7924100898823702e-05, + "loss": 0.6279, + "step": 1587 + }, + { + "epoch": 0.7646563139520886, + "grad_norm": 0.3680364787578583, + "learning_rate": 2.7816060167861002e-05, + "loss": 0.4774, + "step": 1588 + }, + { + "epoch": 0.7651378355603707, + "grad_norm": 0.42586857080459595, + "learning_rate": 2.7708195076040445e-05, + "loss": 0.234, + "step": 1589 + }, + { + "epoch": 0.7656193571686529, + "grad_norm": 0.4964386820793152, + "learning_rate": 2.760050588582114e-05, + "loss": 0.7201, + "step": 1590 + }, + { + "epoch": 0.7661008787769351, + "grad_norm": 0.886875569820404, + "learning_rate": 2.749299285923417e-05, + "loss": 0.6147, + "step": 1591 + }, + { + "epoch": 0.7665824003852173, + "grad_norm": 0.7478306293487549, + "learning_rate": 2.7385656257881997e-05, + "loss": 0.5431, + "step": 1592 + }, + { + "epoch": 0.7670639219934995, + "grad_norm": 0.4899425208568573, + "learning_rate": 2.7278496342937788e-05, + "loss": 0.1444, + "step": 1593 + }, + { + "epoch": 0.7675454436017817, + "grad_norm": 1.0289299488067627, + "learning_rate": 2.717151337514482e-05, + "loss": 0.5972, + "step": 1594 + }, + { + "epoch": 0.7680269652100638, + "grad_norm": 0.7863545417785645, + "learning_rate": 2.7064707614815776e-05, + "loss": 0.4758, + "step": 1595 + }, + { + "epoch": 0.7685084868183459, + "grad_norm": 0.6140004992485046, + "learning_rate": 2.6958079321832185e-05, + "loss": 0.2617, + "step": 1596 + }, + { + "epoch": 0.7689900084266281, + "grad_norm": 0.5038211941719055, + "learning_rate": 2.6851628755643776e-05, + "loss": 0.1513, + "step": 1597 + }, + { + "epoch": 0.7694715300349103, + "grad_norm": 0.7092880010604858, + "learning_rate": 2.6745356175267765e-05, + "loss": 0.3773, + "step": 1598 + }, + { + "epoch": 0.7699530516431925, + "grad_norm": 0.9827279448509216, + "learning_rate": 2.6639261839288343e-05, + "loss": 0.5516, + "step": 1599 + }, + { + "epoch": 0.7704345732514747, + "grad_norm": 0.9941007494926453, + "learning_rate": 2.6533346005855987e-05, + "loss": 1.1843, + "step": 1600 + }, + { + "epoch": 0.7709160948597569, + "grad_norm": 0.5808009505271912, + "learning_rate": 2.6427608932686843e-05, + "loss": 0.3892, + "step": 1601 + }, + { + "epoch": 0.771397616468039, + "grad_norm": 0.8638581037521362, + "learning_rate": 2.6322050877062064e-05, + "loss": 0.3914, + "step": 1602 + }, + { + "epoch": 0.7718791380763211, + "grad_norm": 0.8344993591308594, + "learning_rate": 2.6216672095827266e-05, + "loss": 0.6955, + "step": 1603 + }, + { + "epoch": 0.7723606596846033, + "grad_norm": 0.7511135339736938, + "learning_rate": 2.6111472845391827e-05, + "loss": 0.6755, + "step": 1604 + }, + { + "epoch": 0.7728421812928855, + "grad_norm": 0.6375028491020203, + "learning_rate": 2.6006453381728236e-05, + "loss": 0.1404, + "step": 1605 + }, + { + "epoch": 0.7733237029011677, + "grad_norm": 0.7611637711524963, + "learning_rate": 2.5901613960371585e-05, + "loss": 0.377, + "step": 1606 + }, + { + "epoch": 0.7738052245094499, + "grad_norm": 2.2889697551727295, + "learning_rate": 2.5796954836418884e-05, + "loss": 0.8719, + "step": 1607 + }, + { + "epoch": 0.7742867461177321, + "grad_norm": 0.8444758057594299, + "learning_rate": 2.569247626452842e-05, + "loss": 0.5739, + "step": 1608 + }, + { + "epoch": 0.7747682677260143, + "grad_norm": 1.26003897190094, + "learning_rate": 2.558817849891918e-05, + "loss": 0.433, + "step": 1609 + }, + { + "epoch": 0.7752497893342963, + "grad_norm": 0.647497296333313, + "learning_rate": 2.548406179337015e-05, + "loss": 0.547, + "step": 1610 + }, + { + "epoch": 0.7757313109425785, + "grad_norm": 0.31829091906547546, + "learning_rate": 2.5380126401219807e-05, + "loss": 0.2283, + "step": 1611 + }, + { + "epoch": 0.7762128325508607, + "grad_norm": 0.9248780012130737, + "learning_rate": 2.527637257536547e-05, + "loss": 0.3302, + "step": 1612 + }, + { + "epoch": 0.7766943541591429, + "grad_norm": 0.8120497465133667, + "learning_rate": 2.517280056826262e-05, + "loss": 0.4713, + "step": 1613 + }, + { + "epoch": 0.7771758757674251, + "grad_norm": 0.47866806387901306, + "learning_rate": 2.5069410631924385e-05, + "loss": 0.2786, + "step": 1614 + }, + { + "epoch": 0.7776573973757073, + "grad_norm": 0.48474082350730896, + "learning_rate": 2.4966203017920818e-05, + "loss": 0.2915, + "step": 1615 + }, + { + "epoch": 0.7781389189839895, + "grad_norm": 0.8543782830238342, + "learning_rate": 2.4863177977378392e-05, + "loss": 0.795, + "step": 1616 + }, + { + "epoch": 0.7786204405922715, + "grad_norm": 0.4673691987991333, + "learning_rate": 2.4760335760979312e-05, + "loss": 0.6174, + "step": 1617 + }, + { + "epoch": 0.7791019622005537, + "grad_norm": 0.7312822937965393, + "learning_rate": 2.4657676618960944e-05, + "loss": 0.2589, + "step": 1618 + }, + { + "epoch": 0.7795834838088359, + "grad_norm": 0.8223809003829956, + "learning_rate": 2.455520080111522e-05, + "loss": 0.7372, + "step": 1619 + }, + { + "epoch": 0.7800650054171181, + "grad_norm": 0.8796128034591675, + "learning_rate": 2.4452908556787912e-05, + "loss": 0.5219, + "step": 1620 + }, + { + "epoch": 0.7805465270254003, + "grad_norm": 0.8800025582313538, + "learning_rate": 2.4350800134878203e-05, + "loss": 0.5718, + "step": 1621 + }, + { + "epoch": 0.7810280486336825, + "grad_norm": 0.3794774115085602, + "learning_rate": 2.4248875783837987e-05, + "loss": 0.1713, + "step": 1622 + }, + { + "epoch": 0.7815095702419647, + "grad_norm": 1.717668890953064, + "learning_rate": 2.414713575167129e-05, + "loss": 0.2789, + "step": 1623 + }, + { + "epoch": 0.7819910918502467, + "grad_norm": 0.5892308354377747, + "learning_rate": 2.4045580285933557e-05, + "loss": 0.3826, + "step": 1624 + }, + { + "epoch": 0.7824726134585289, + "grad_norm": 1.4412480592727661, + "learning_rate": 2.3944209633731242e-05, + "loss": 0.9213, + "step": 1625 + }, + { + "epoch": 0.7829541350668111, + "grad_norm": 0.9920421838760376, + "learning_rate": 2.3843024041721053e-05, + "loss": 0.2247, + "step": 1626 + }, + { + "epoch": 0.7834356566750933, + "grad_norm": 1.3514714241027832, + "learning_rate": 2.3742023756109456e-05, + "loss": 1.0385, + "step": 1627 + }, + { + "epoch": 0.7839171782833755, + "grad_norm": 0.6163105368614197, + "learning_rate": 2.3641209022651976e-05, + "loss": 0.6583, + "step": 1628 + }, + { + "epoch": 0.7843986998916577, + "grad_norm": 0.7857782244682312, + "learning_rate": 2.3540580086652675e-05, + "loss": 0.4282, + "step": 1629 + }, + { + "epoch": 0.7848802214999399, + "grad_norm": 0.8993891477584839, + "learning_rate": 2.344013719296353e-05, + "loss": 0.35, + "step": 1630 + }, + { + "epoch": 0.7853617431082219, + "grad_norm": 0.7018194794654846, + "learning_rate": 2.3339880585983842e-05, + "loss": 0.3266, + "step": 1631 + }, + { + "epoch": 0.7858432647165041, + "grad_norm": 0.35083359479904175, + "learning_rate": 2.3239810509659597e-05, + "loss": 0.3474, + "step": 1632 + }, + { + "epoch": 0.7863247863247863, + "grad_norm": 0.9153875708580017, + "learning_rate": 2.313992720748295e-05, + "loss": 0.5287, + "step": 1633 + }, + { + "epoch": 0.7868063079330685, + "grad_norm": 0.8885396718978882, + "learning_rate": 2.304023092249159e-05, + "loss": 0.7783, + "step": 1634 + }, + { + "epoch": 0.7872878295413507, + "grad_norm": 1.1704275608062744, + "learning_rate": 2.2940721897268136e-05, + "loss": 0.6887, + "step": 1635 + }, + { + "epoch": 0.7877693511496329, + "grad_norm": 0.5215083360671997, + "learning_rate": 2.2841400373939592e-05, + "loss": 0.4906, + "step": 1636 + }, + { + "epoch": 0.788250872757915, + "grad_norm": 0.9289473295211792, + "learning_rate": 2.274226659417671e-05, + "loss": 1.1291, + "step": 1637 + }, + { + "epoch": 0.7887323943661971, + "grad_norm": 0.5596996545791626, + "learning_rate": 2.2643320799193402e-05, + "loss": 0.2618, + "step": 1638 + }, + { + "epoch": 0.7892139159744793, + "grad_norm": 0.8898619413375854, + "learning_rate": 2.2544563229746218e-05, + "loss": 0.6691, + "step": 1639 + }, + { + "epoch": 0.7896954375827615, + "grad_norm": 0.943631649017334, + "learning_rate": 2.2445994126133708e-05, + "loss": 1.1104, + "step": 1640 + }, + { + "epoch": 0.7901769591910437, + "grad_norm": 0.8219139575958252, + "learning_rate": 2.234761372819577e-05, + "loss": 0.7676, + "step": 1641 + }, + { + "epoch": 0.7906584807993259, + "grad_norm": 0.8116326928138733, + "learning_rate": 2.2249422275313214e-05, + "loss": 0.434, + "step": 1642 + }, + { + "epoch": 0.7911400024076081, + "grad_norm": 0.5445429682731628, + "learning_rate": 2.215142000640714e-05, + "loss": 0.252, + "step": 1643 + }, + { + "epoch": 0.7916215240158903, + "grad_norm": 0.751610279083252, + "learning_rate": 2.2053607159938195e-05, + "loss": 0.5864, + "step": 1644 + }, + { + "epoch": 0.7921030456241723, + "grad_norm": 0.7384713292121887, + "learning_rate": 2.1955983973906236e-05, + "loss": 0.3833, + "step": 1645 + }, + { + "epoch": 0.7925845672324545, + "grad_norm": 0.6013317108154297, + "learning_rate": 2.1858550685849578e-05, + "loss": 0.5739, + "step": 1646 + }, + { + "epoch": 0.7930660888407367, + "grad_norm": 1.1576766967773438, + "learning_rate": 2.17613075328445e-05, + "loss": 0.425, + "step": 1647 + }, + { + "epoch": 0.7935476104490189, + "grad_norm": 1.0560667514801025, + "learning_rate": 2.1664254751504642e-05, + "loss": 0.5623, + "step": 1648 + }, + { + "epoch": 0.7940291320573011, + "grad_norm": 0.37843796610832214, + "learning_rate": 2.1567392577980393e-05, + "loss": 0.1924, + "step": 1649 + }, + { + "epoch": 0.7945106536655833, + "grad_norm": 1.0658668279647827, + "learning_rate": 2.1470721247958404e-05, + "loss": 1.2051, + "step": 1650 + }, + { + "epoch": 0.7949921752738655, + "grad_norm": 1.119051456451416, + "learning_rate": 2.137424099666091e-05, + "loss": 0.7042, + "step": 1651 + }, + { + "epoch": 0.7954736968821476, + "grad_norm": 0.3540979027748108, + "learning_rate": 2.1277952058845284e-05, + "loss": 0.1893, + "step": 1652 + }, + { + "epoch": 0.7959552184904297, + "grad_norm": 0.6795416474342346, + "learning_rate": 2.118185466880327e-05, + "loss": 1.0059, + "step": 1653 + }, + { + "epoch": 0.7964367400987119, + "grad_norm": 0.9295594692230225, + "learning_rate": 2.1085949060360654e-05, + "loss": 0.6773, + "step": 1654 + }, + { + "epoch": 0.7969182617069941, + "grad_norm": 0.669072687625885, + "learning_rate": 2.0990235466876517e-05, + "loss": 0.5336, + "step": 1655 + }, + { + "epoch": 0.7973997833152763, + "grad_norm": 0.5881422162055969, + "learning_rate": 2.089471412124274e-05, + "loss": 0.5242, + "step": 1656 + }, + { + "epoch": 0.7978813049235585, + "grad_norm": 0.7635432481765747, + "learning_rate": 2.079938525588342e-05, + "loss": 0.2237, + "step": 1657 + }, + { + "epoch": 0.7983628265318407, + "grad_norm": 0.6804701089859009, + "learning_rate": 2.0704249102754324e-05, + "loss": 0.6416, + "step": 1658 + }, + { + "epoch": 0.7988443481401228, + "grad_norm": 0.4592357575893402, + "learning_rate": 2.0609305893342278e-05, + "loss": 0.4071, + "step": 1659 + }, + { + "epoch": 0.7993258697484049, + "grad_norm": 0.7150890827178955, + "learning_rate": 2.0514555858664663e-05, + "loss": 0.9323, + "step": 1660 + }, + { + "epoch": 0.7998073913566871, + "grad_norm": 0.5522563457489014, + "learning_rate": 2.0419999229268805e-05, + "loss": 0.3048, + "step": 1661 + }, + { + "epoch": 0.8002889129649693, + "grad_norm": 1.1032781600952148, + "learning_rate": 2.032563623523147e-05, + "loss": 0.6891, + "step": 1662 + }, + { + "epoch": 0.8007704345732515, + "grad_norm": 0.8238122463226318, + "learning_rate": 2.0231467106158186e-05, + "loss": 0.281, + "step": 1663 + }, + { + "epoch": 0.8012519561815337, + "grad_norm": 0.7528010010719299, + "learning_rate": 2.0137492071182863e-05, + "loss": 0.6063, + "step": 1664 + }, + { + "epoch": 0.8017334777898159, + "grad_norm": 1.0722914934158325, + "learning_rate": 2.0043711358967043e-05, + "loss": 1.0133, + "step": 1665 + }, + { + "epoch": 0.802214999398098, + "grad_norm": 0.44027334451675415, + "learning_rate": 1.9950125197699508e-05, + "loss": 0.2833, + "step": 1666 + }, + { + "epoch": 0.8026965210063801, + "grad_norm": 0.9600662589073181, + "learning_rate": 1.985673381509565e-05, + "loss": 0.3938, + "step": 1667 + }, + { + "epoch": 0.8031780426146623, + "grad_norm": 0.6282616853713989, + "learning_rate": 1.9763537438396894e-05, + "loss": 0.6348, + "step": 1668 + }, + { + "epoch": 0.8036595642229445, + "grad_norm": 0.5281164646148682, + "learning_rate": 1.96705362943702e-05, + "loss": 0.5014, + "step": 1669 + }, + { + "epoch": 0.8041410858312267, + "grad_norm": 0.6858304142951965, + "learning_rate": 1.9577730609307454e-05, + "loss": 0.5066, + "step": 1670 + }, + { + "epoch": 0.8046226074395089, + "grad_norm": 0.6477140784263611, + "learning_rate": 1.9485120609024975e-05, + "loss": 0.445, + "step": 1671 + }, + { + "epoch": 0.805104129047791, + "grad_norm": 0.6046128273010254, + "learning_rate": 1.9392706518862935e-05, + "loss": 0.2247, + "step": 1672 + }, + { + "epoch": 0.8055856506560732, + "grad_norm": 0.8289013504981995, + "learning_rate": 1.9300488563684804e-05, + "loss": 0.2918, + "step": 1673 + }, + { + "epoch": 0.8060671722643553, + "grad_norm": 0.46251946687698364, + "learning_rate": 1.920846696787684e-05, + "loss": 0.6147, + "step": 1674 + }, + { + "epoch": 0.8065486938726375, + "grad_norm": 0.5980027914047241, + "learning_rate": 1.9116641955347446e-05, + "loss": 0.3733, + "step": 1675 + }, + { + "epoch": 0.8070302154809197, + "grad_norm": 0.6853753924369812, + "learning_rate": 1.9025013749526767e-05, + "loss": 0.5122, + "step": 1676 + }, + { + "epoch": 0.8075117370892019, + "grad_norm": 1.0705738067626953, + "learning_rate": 1.8933582573366036e-05, + "loss": 0.9123, + "step": 1677 + }, + { + "epoch": 0.8079932586974841, + "grad_norm": 0.9407148957252502, + "learning_rate": 1.8842348649337116e-05, + "loss": 0.9415, + "step": 1678 + }, + { + "epoch": 0.8084747803057662, + "grad_norm": 1.0631650686264038, + "learning_rate": 1.875131219943187e-05, + "loss": 1.1339, + "step": 1679 + }, + { + "epoch": 0.8089563019140484, + "grad_norm": 0.6693783402442932, + "learning_rate": 1.8660473445161663e-05, + "loss": 0.2972, + "step": 1680 + }, + { + "epoch": 0.8094378235223305, + "grad_norm": 0.8969826102256775, + "learning_rate": 1.856983260755686e-05, + "loss": 0.5933, + "step": 1681 + }, + { + "epoch": 0.8099193451306127, + "grad_norm": 1.0234383344650269, + "learning_rate": 1.8479389907166223e-05, + "loss": 0.3956, + "step": 1682 + }, + { + "epoch": 0.8104008667388949, + "grad_norm": 0.9167569875717163, + "learning_rate": 1.8389145564056387e-05, + "loss": 0.4642, + "step": 1683 + }, + { + "epoch": 0.8108823883471771, + "grad_norm": 0.5638496279716492, + "learning_rate": 1.829909979781137e-05, + "loss": 0.3361, + "step": 1684 + }, + { + "epoch": 0.8113639099554593, + "grad_norm": 0.7476749420166016, + "learning_rate": 1.820925282753201e-05, + "loss": 0.5014, + "step": 1685 + }, + { + "epoch": 0.8118454315637414, + "grad_norm": 0.7180486917495728, + "learning_rate": 1.8119604871835437e-05, + "loss": 0.5145, + "step": 1686 + }, + { + "epoch": 0.8123269531720236, + "grad_norm": 0.5081984996795654, + "learning_rate": 1.8030156148854492e-05, + "loss": 0.3872, + "step": 1687 + }, + { + "epoch": 0.8128084747803057, + "grad_norm": 0.9229518175125122, + "learning_rate": 1.7940906876237284e-05, + "loss": 0.3678, + "step": 1688 + }, + { + "epoch": 0.8132899963885879, + "grad_norm": 0.8328425288200378, + "learning_rate": 1.78518572711466e-05, + "loss": 0.5457, + "step": 1689 + }, + { + "epoch": 0.8137715179968701, + "grad_norm": 1.213712215423584, + "learning_rate": 1.776300755025939e-05, + "loss": 1.061, + "step": 1690 + }, + { + "epoch": 0.8142530396051523, + "grad_norm": 1.0858510732650757, + "learning_rate": 1.767435792976626e-05, + "loss": 0.7261, + "step": 1691 + }, + { + "epoch": 0.8147345612134345, + "grad_norm": 0.6739503145217896, + "learning_rate": 1.7585908625370905e-05, + "loss": 0.5279, + "step": 1692 + }, + { + "epoch": 0.8152160828217166, + "grad_norm": 0.8865543007850647, + "learning_rate": 1.749765985228963e-05, + "loss": 0.9244, + "step": 1693 + }, + { + "epoch": 0.8156976044299988, + "grad_norm": 0.7478240728378296, + "learning_rate": 1.740961182525077e-05, + "loss": 0.2868, + "step": 1694 + }, + { + "epoch": 0.816179126038281, + "grad_norm": 1.0323337316513062, + "learning_rate": 1.7321764758494252e-05, + "loss": 1.0011, + "step": 1695 + }, + { + "epoch": 0.8166606476465631, + "grad_norm": 0.9194225668907166, + "learning_rate": 1.7234118865770987e-05, + "loss": 0.9113, + "step": 1696 + }, + { + "epoch": 0.8171421692548453, + "grad_norm": 0.8026208877563477, + "learning_rate": 1.7146674360342373e-05, + "loss": 0.4512, + "step": 1697 + }, + { + "epoch": 0.8176236908631275, + "grad_norm": 0.4755818545818329, + "learning_rate": 1.7059431454979824e-05, + "loss": 0.694, + "step": 1698 + }, + { + "epoch": 0.8181052124714097, + "grad_norm": 1.2281359434127808, + "learning_rate": 1.6972390361964195e-05, + "loss": 1.1019, + "step": 1699 + }, + { + "epoch": 0.8185867340796918, + "grad_norm": 0.7719329595565796, + "learning_rate": 1.688555129308531e-05, + "loss": 0.5681, + "step": 1700 + }, + { + "epoch": 0.819068255687974, + "grad_norm": 1.0498366355895996, + "learning_rate": 1.6798914459641434e-05, + "loss": 0.7296, + "step": 1701 + }, + { + "epoch": 0.8195497772962562, + "grad_norm": 0.6994045376777649, + "learning_rate": 1.6712480072438662e-05, + "loss": 0.9328, + "step": 1702 + }, + { + "epoch": 0.8200312989045383, + "grad_norm": 0.9230305552482605, + "learning_rate": 1.6626248341790596e-05, + "loss": 1.1862, + "step": 1703 + }, + { + "epoch": 0.8205128205128205, + "grad_norm": 0.6558325886726379, + "learning_rate": 1.6540219477517684e-05, + "loss": 0.2061, + "step": 1704 + }, + { + "epoch": 0.8209943421211027, + "grad_norm": 0.6145660877227783, + "learning_rate": 1.6454393688946767e-05, + "loss": 0.4734, + "step": 1705 + }, + { + "epoch": 0.8214758637293849, + "grad_norm": 0.8060041666030884, + "learning_rate": 1.6368771184910557e-05, + "loss": 0.6261, + "step": 1706 + }, + { + "epoch": 0.821957385337667, + "grad_norm": 0.7207950949668884, + "learning_rate": 1.6283352173747145e-05, + "loss": 0.7611, + "step": 1707 + }, + { + "epoch": 0.8224389069459492, + "grad_norm": 0.5801020860671997, + "learning_rate": 1.619813686329946e-05, + "loss": 0.7966, + "step": 1708 + }, + { + "epoch": 0.8229204285542314, + "grad_norm": 0.5637111067771912, + "learning_rate": 1.611312546091476e-05, + "loss": 0.7279, + "step": 1709 + }, + { + "epoch": 0.8234019501625135, + "grad_norm": 1.224278211593628, + "learning_rate": 1.6028318173444202e-05, + "loss": 1.3064, + "step": 1710 + }, + { + "epoch": 0.8238834717707957, + "grad_norm": 0.832797646522522, + "learning_rate": 1.594371520724226e-05, + "loss": 0.6784, + "step": 1711 + }, + { + "epoch": 0.8243649933790779, + "grad_norm": 0.7406448721885681, + "learning_rate": 1.5859316768166244e-05, + "loss": 0.5418, + "step": 1712 + }, + { + "epoch": 0.8248465149873601, + "grad_norm": 0.9135259985923767, + "learning_rate": 1.5775123061575836e-05, + "loss": 1.03, + "step": 1713 + }, + { + "epoch": 0.8253280365956422, + "grad_norm": 0.486418217420578, + "learning_rate": 1.569113429233252e-05, + "loss": 0.2692, + "step": 1714 + }, + { + "epoch": 0.8258095582039244, + "grad_norm": 0.7394134998321533, + "learning_rate": 1.5607350664799157e-05, + "loss": 0.6892, + "step": 1715 + }, + { + "epoch": 0.8262910798122066, + "grad_norm": 0.7706437110900879, + "learning_rate": 1.552377238283943e-05, + "loss": 0.2832, + "step": 1716 + }, + { + "epoch": 0.8267726014204887, + "grad_norm": 0.9024636745452881, + "learning_rate": 1.5440399649817385e-05, + "loss": 0.3443, + "step": 1717 + }, + { + "epoch": 0.8272541230287709, + "grad_norm": 0.6336109042167664, + "learning_rate": 1.5357232668596933e-05, + "loss": 0.6889, + "step": 1718 + }, + { + "epoch": 0.8277356446370531, + "grad_norm": 0.3393593728542328, + "learning_rate": 1.5274271641541295e-05, + "loss": 0.1545, + "step": 1719 + }, + { + "epoch": 0.8282171662453353, + "grad_norm": 0.6772252321243286, + "learning_rate": 1.5191516770512649e-05, + "loss": 0.7456, + "step": 1720 + }, + { + "epoch": 0.8286986878536174, + "grad_norm": 0.7937234044075012, + "learning_rate": 1.5108968256871437e-05, + "loss": 0.349, + "step": 1721 + }, + { + "epoch": 0.8291802094618996, + "grad_norm": 1.5913159847259521, + "learning_rate": 1.5026626301476087e-05, + "loss": 0.9463, + "step": 1722 + }, + { + "epoch": 0.8296617310701818, + "grad_norm": 0.6769959330558777, + "learning_rate": 1.4944491104682379e-05, + "loss": 0.5529, + "step": 1723 + }, + { + "epoch": 0.8301432526784639, + "grad_norm": 1.4310240745544434, + "learning_rate": 1.4862562866343034e-05, + "loss": 0.5971, + "step": 1724 + }, + { + "epoch": 0.8306247742867461, + "grad_norm": 0.9801859259605408, + "learning_rate": 1.4780841785807164e-05, + "loss": 0.7973, + "step": 1725 + }, + { + "epoch": 0.8311062958950283, + "grad_norm": 0.6424405574798584, + "learning_rate": 1.4699328061919848e-05, + "loss": 0.2604, + "step": 1726 + }, + { + "epoch": 0.8315878175033105, + "grad_norm": 1.0231726169586182, + "learning_rate": 1.4618021893021605e-05, + "loss": 0.3021, + "step": 1727 + }, + { + "epoch": 0.8320693391115926, + "grad_norm": 0.6480558514595032, + "learning_rate": 1.453692347694794e-05, + "loss": 0.4345, + "step": 1728 + }, + { + "epoch": 0.8325508607198748, + "grad_norm": 0.8369768857955933, + "learning_rate": 1.4456033011028835e-05, + "loss": 0.3847, + "step": 1729 + }, + { + "epoch": 0.833032382328157, + "grad_norm": 0.6749899983406067, + "learning_rate": 1.437535069208833e-05, + "loss": 0.2968, + "step": 1730 + }, + { + "epoch": 0.8335139039364391, + "grad_norm": 0.6067473888397217, + "learning_rate": 1.4294876716443906e-05, + "loss": 0.3409, + "step": 1731 + }, + { + "epoch": 0.8339954255447213, + "grad_norm": 0.641484797000885, + "learning_rate": 1.4214611279906187e-05, + "loss": 0.2532, + "step": 1732 + }, + { + "epoch": 0.8344769471530035, + "grad_norm": 0.698528528213501, + "learning_rate": 1.4134554577778337e-05, + "loss": 0.3922, + "step": 1733 + }, + { + "epoch": 0.8349584687612857, + "grad_norm": 0.823616087436676, + "learning_rate": 1.4054706804855634e-05, + "loss": 0.8489, + "step": 1734 + }, + { + "epoch": 0.8354399903695678, + "grad_norm": 0.7121361494064331, + "learning_rate": 1.3975068155424976e-05, + "loss": 0.6374, + "step": 1735 + }, + { + "epoch": 0.83592151197785, + "grad_norm": 1.6407108306884766, + "learning_rate": 1.3895638823264446e-05, + "loss": 0.5144, + "step": 1736 + }, + { + "epoch": 0.8364030335861322, + "grad_norm": 1.073464035987854, + "learning_rate": 1.3816419001642777e-05, + "loss": 0.6168, + "step": 1737 + }, + { + "epoch": 0.8368845551944144, + "grad_norm": 1.1071691513061523, + "learning_rate": 1.3737408883318948e-05, + "loss": 0.5041, + "step": 1738 + }, + { + "epoch": 0.8373660768026965, + "grad_norm": 0.9471048712730408, + "learning_rate": 1.365860866054165e-05, + "loss": 0.3752, + "step": 1739 + }, + { + "epoch": 0.8378475984109787, + "grad_norm": 0.42994487285614014, + "learning_rate": 1.358001852504891e-05, + "loss": 0.2467, + "step": 1740 + }, + { + "epoch": 0.8383291200192609, + "grad_norm": 1.1875547170639038, + "learning_rate": 1.3501638668067485e-05, + "loss": 0.8259, + "step": 1741 + }, + { + "epoch": 0.838810641627543, + "grad_norm": 1.273987889289856, + "learning_rate": 1.3423469280312562e-05, + "loss": 0.478, + "step": 1742 + }, + { + "epoch": 0.8392921632358252, + "grad_norm": 0.571315586566925, + "learning_rate": 1.3345510551987128e-05, + "loss": 0.2148, + "step": 1743 + }, + { + "epoch": 0.8397736848441074, + "grad_norm": 0.632285475730896, + "learning_rate": 1.326776267278167e-05, + "loss": 0.3969, + "step": 1744 + }, + { + "epoch": 0.8402552064523896, + "grad_norm": 1.3262178897857666, + "learning_rate": 1.3190225831873581e-05, + "loss": 0.4584, + "step": 1745 + }, + { + "epoch": 0.8407367280606717, + "grad_norm": 0.6027417778968811, + "learning_rate": 1.3112900217926782e-05, + "loss": 0.6441, + "step": 1746 + }, + { + "epoch": 0.8412182496689539, + "grad_norm": 0.8074085712432861, + "learning_rate": 1.3035786019091223e-05, + "loss": 0.5024, + "step": 1747 + }, + { + "epoch": 0.841699771277236, + "grad_norm": 0.7299503684043884, + "learning_rate": 1.2958883423002422e-05, + "loss": 1.0229, + "step": 1748 + }, + { + "epoch": 0.8421812928855182, + "grad_norm": 0.6064135432243347, + "learning_rate": 1.288219261678103e-05, + "loss": 0.3824, + "step": 1749 + }, + { + "epoch": 0.8426628144938004, + "grad_norm": 0.9795793294906616, + "learning_rate": 1.2805713787032381e-05, + "loss": 0.3306, + "step": 1750 + }, + { + "epoch": 0.8431443361020826, + "grad_norm": 0.3210099935531616, + "learning_rate": 1.2729447119846016e-05, + "loss": 0.2856, + "step": 1751 + }, + { + "epoch": 0.8436258577103648, + "grad_norm": 1.0421223640441895, + "learning_rate": 1.265339280079525e-05, + "loss": 0.7998, + "step": 1752 + }, + { + "epoch": 0.8441073793186469, + "grad_norm": 0.8377938270568848, + "learning_rate": 1.257755101493665e-05, + "loss": 0.6302, + "step": 1753 + }, + { + "epoch": 0.8445889009269291, + "grad_norm": 0.3239783048629761, + "learning_rate": 1.2501921946809714e-05, + "loss": 0.1868, + "step": 1754 + }, + { + "epoch": 0.8450704225352113, + "grad_norm": 0.5518014430999756, + "learning_rate": 1.2426505780436326e-05, + "loss": 0.3083, + "step": 1755 + }, + { + "epoch": 0.8455519441434934, + "grad_norm": 0.7590487003326416, + "learning_rate": 1.2351302699320332e-05, + "loss": 0.7756, + "step": 1756 + }, + { + "epoch": 0.8460334657517756, + "grad_norm": 0.8196136951446533, + "learning_rate": 1.2276312886447106e-05, + "loss": 0.734, + "step": 1757 + }, + { + "epoch": 0.8465149873600578, + "grad_norm": 0.9058529138565063, + "learning_rate": 1.2201536524283074e-05, + "loss": 0.5072, + "step": 1758 + }, + { + "epoch": 0.84699650896834, + "grad_norm": 0.389612078666687, + "learning_rate": 1.2126973794775343e-05, + "loss": 0.308, + "step": 1759 + }, + { + "epoch": 0.8474780305766221, + "grad_norm": 0.9301180243492126, + "learning_rate": 1.2052624879351104e-05, + "loss": 0.3945, + "step": 1760 + }, + { + "epoch": 0.8479595521849043, + "grad_norm": 0.7932952046394348, + "learning_rate": 1.1978489958917382e-05, + "loss": 0.651, + "step": 1761 + }, + { + "epoch": 0.8484410737931865, + "grad_norm": 0.8034958839416504, + "learning_rate": 1.1904569213860472e-05, + "loss": 0.7516, + "step": 1762 + }, + { + "epoch": 0.8489225954014686, + "grad_norm": 0.584695041179657, + "learning_rate": 1.1830862824045552e-05, + "loss": 0.8093, + "step": 1763 + }, + { + "epoch": 0.8494041170097508, + "grad_norm": 1.2081027030944824, + "learning_rate": 1.1757370968816217e-05, + "loss": 0.6173, + "step": 1764 + }, + { + "epoch": 0.849885638618033, + "grad_norm": 0.9123238921165466, + "learning_rate": 1.1684093826994024e-05, + "loss": 0.2586, + "step": 1765 + }, + { + "epoch": 0.8503671602263152, + "grad_norm": 0.805109977722168, + "learning_rate": 1.1611031576878117e-05, + "loss": 0.519, + "step": 1766 + }, + { + "epoch": 0.8508486818345973, + "grad_norm": 0.49961966276168823, + "learning_rate": 1.1538184396244778e-05, + "loss": 0.2066, + "step": 1767 + }, + { + "epoch": 0.8513302034428795, + "grad_norm": 0.7591211199760437, + "learning_rate": 1.146555246234694e-05, + "loss": 0.3607, + "step": 1768 + }, + { + "epoch": 0.8518117250511617, + "grad_norm": 1.0066025257110596, + "learning_rate": 1.1393135951913824e-05, + "loss": 0.6076, + "step": 1769 + }, + { + "epoch": 0.8522932466594438, + "grad_norm": 0.8413494229316711, + "learning_rate": 1.132093504115046e-05, + "loss": 0.5581, + "step": 1770 + }, + { + "epoch": 0.852774768267726, + "grad_norm": 0.9132838249206543, + "learning_rate": 1.1248949905737283e-05, + "loss": 0.3262, + "step": 1771 + }, + { + "epoch": 0.8532562898760082, + "grad_norm": 0.4182647466659546, + "learning_rate": 1.1177180720829694e-05, + "loss": 0.1514, + "step": 1772 + }, + { + "epoch": 0.8537378114842904, + "grad_norm": 0.4618658125400543, + "learning_rate": 1.1105627661057671e-05, + "loss": 0.4407, + "step": 1773 + }, + { + "epoch": 0.8542193330925725, + "grad_norm": 0.7735852003097534, + "learning_rate": 1.103429090052528e-05, + "loss": 0.7609, + "step": 1774 + }, + { + "epoch": 0.8547008547008547, + "grad_norm": 0.5562105178833008, + "learning_rate": 1.096317061281027e-05, + "loss": 0.5554, + "step": 1775 + }, + { + "epoch": 0.8551823763091368, + "grad_norm": 1.005146861076355, + "learning_rate": 1.0892266970963704e-05, + "loss": 0.8242, + "step": 1776 + }, + { + "epoch": 0.855663897917419, + "grad_norm": 0.9059540033340454, + "learning_rate": 1.082158014750948e-05, + "loss": 0.8464, + "step": 1777 + }, + { + "epoch": 0.8561454195257012, + "grad_norm": 0.5801354646682739, + "learning_rate": 1.0751110314443958e-05, + "loss": 0.4619, + "step": 1778 + }, + { + "epoch": 0.8566269411339834, + "grad_norm": 0.5400471091270447, + "learning_rate": 1.0680857643235431e-05, + "loss": 0.2479, + "step": 1779 + }, + { + "epoch": 0.8571084627422656, + "grad_norm": 0.6659761071205139, + "learning_rate": 1.0610822304823887e-05, + "loss": 0.5576, + "step": 1780 + }, + { + "epoch": 0.8575899843505478, + "grad_norm": 0.9168156981468201, + "learning_rate": 1.0541004469620452e-05, + "loss": 0.6976, + "step": 1781 + }, + { + "epoch": 0.8580715059588299, + "grad_norm": 0.9955235719680786, + "learning_rate": 1.0471404307507016e-05, + "loss": 1.502, + "step": 1782 + }, + { + "epoch": 0.858553027567112, + "grad_norm": 0.669767439365387, + "learning_rate": 1.0402021987835831e-05, + "loss": 0.2673, + "step": 1783 + }, + { + "epoch": 0.8590345491753942, + "grad_norm": 0.3703511953353882, + "learning_rate": 1.0332857679429098e-05, + "loss": 0.212, + "step": 1784 + }, + { + "epoch": 0.8595160707836764, + "grad_norm": 1.1823803186416626, + "learning_rate": 1.0263911550578531e-05, + "loss": 0.6145, + "step": 1785 + }, + { + "epoch": 0.8599975923919586, + "grad_norm": 0.8142738938331604, + "learning_rate": 1.0195183769045013e-05, + "loss": 0.2403, + "step": 1786 + }, + { + "epoch": 0.8604791140002408, + "grad_norm": 0.8147161602973938, + "learning_rate": 1.0126674502058054e-05, + "loss": 0.8877, + "step": 1787 + }, + { + "epoch": 0.860960635608523, + "grad_norm": 0.6185226440429688, + "learning_rate": 1.005838391631555e-05, + "loss": 0.3974, + "step": 1788 + }, + { + "epoch": 0.8614421572168051, + "grad_norm": 0.7300569415092468, + "learning_rate": 9.990312177983263e-06, + "loss": 0.4777, + "step": 1789 + }, + { + "epoch": 0.8619236788250872, + "grad_norm": 0.8982576131820679, + "learning_rate": 9.922459452694466e-06, + "loss": 0.6434, + "step": 1790 + }, + { + "epoch": 0.8624052004333694, + "grad_norm": 0.5035396218299866, + "learning_rate": 9.854825905549503e-06, + "loss": 0.5651, + "step": 1791 + }, + { + "epoch": 0.8628867220416516, + "grad_norm": 0.6236551403999329, + "learning_rate": 9.787411701115456e-06, + "loss": 0.2783, + "step": 1792 + }, + { + "epoch": 0.8633682436499338, + "grad_norm": 1.1999423503875732, + "learning_rate": 9.720217003425647e-06, + "loss": 1.0484, + "step": 1793 + }, + { + "epoch": 0.863849765258216, + "grad_norm": 0.6424204111099243, + "learning_rate": 9.65324197597931e-06, + "loss": 0.294, + "step": 1794 + }, + { + "epoch": 0.8643312868664982, + "grad_norm": 0.724156379699707, + "learning_rate": 9.58648678174121e-06, + "loss": 0.3884, + "step": 1795 + }, + { + "epoch": 0.8648128084747803, + "grad_norm": 0.47926265001296997, + "learning_rate": 9.51995158314113e-06, + "loss": 0.5631, + "step": 1796 + }, + { + "epoch": 0.8652943300830624, + "grad_norm": 1.520438551902771, + "learning_rate": 9.45363654207363e-06, + "loss": 0.8433, + "step": 1797 + }, + { + "epoch": 0.8657758516913446, + "grad_norm": 0.3574250340461731, + "learning_rate": 9.387541819897549e-06, + "loss": 0.4286, + "step": 1798 + }, + { + "epoch": 0.8662573732996268, + "grad_norm": 1.6910595893859863, + "learning_rate": 9.321667577435634e-06, + "loss": 0.691, + "step": 1799 + }, + { + "epoch": 0.866738894907909, + "grad_norm": 0.5806576013565063, + "learning_rate": 9.256013974974175e-06, + "loss": 0.3654, + "step": 1800 + }, + { + "epoch": 0.8672204165161912, + "grad_norm": 0.661484956741333, + "learning_rate": 9.19058117226258e-06, + "loss": 0.6508, + "step": 1801 + }, + { + "epoch": 0.8677019381244734, + "grad_norm": 0.8972189426422119, + "learning_rate": 9.125369328513034e-06, + "loss": 0.9418, + "step": 1802 + }, + { + "epoch": 0.8681834597327555, + "grad_norm": 1.4898693561553955, + "learning_rate": 9.060378602400054e-06, + "loss": 0.374, + "step": 1803 + }, + { + "epoch": 0.8686649813410376, + "grad_norm": 1.0133569240570068, + "learning_rate": 8.995609152060136e-06, + "loss": 0.8903, + "step": 1804 + }, + { + "epoch": 0.8691465029493198, + "grad_norm": 0.5878236889839172, + "learning_rate": 8.931061135091357e-06, + "loss": 0.7779, + "step": 1805 + }, + { + "epoch": 0.869628024557602, + "grad_norm": 0.7630800008773804, + "learning_rate": 8.866734708553015e-06, + "loss": 0.8601, + "step": 1806 + }, + { + "epoch": 0.8701095461658842, + "grad_norm": 0.45022618770599365, + "learning_rate": 8.802630028965242e-06, + "loss": 0.2447, + "step": 1807 + }, + { + "epoch": 0.8705910677741664, + "grad_norm": 0.34775686264038086, + "learning_rate": 8.738747252308555e-06, + "loss": 0.1954, + "step": 1808 + }, + { + "epoch": 0.8710725893824486, + "grad_norm": 0.6343087553977966, + "learning_rate": 8.675086534023591e-06, + "loss": 0.3324, + "step": 1809 + }, + { + "epoch": 0.8715541109907307, + "grad_norm": 1.0207873582839966, + "learning_rate": 8.611648029010643e-06, + "loss": 0.3045, + "step": 1810 + }, + { + "epoch": 0.8720356325990128, + "grad_norm": 0.7967151403427124, + "learning_rate": 8.548431891629316e-06, + "loss": 0.2367, + "step": 1811 + }, + { + "epoch": 0.872517154207295, + "grad_norm": 0.6096175909042358, + "learning_rate": 8.485438275698154e-06, + "loss": 0.1747, + "step": 1812 + }, + { + "epoch": 0.8729986758155772, + "grad_norm": 1.0279371738433838, + "learning_rate": 8.422667334494249e-06, + "loss": 0.6618, + "step": 1813 + }, + { + "epoch": 0.8734801974238594, + "grad_norm": 0.5840936899185181, + "learning_rate": 8.360119220752893e-06, + "loss": 0.1349, + "step": 1814 + }, + { + "epoch": 0.8739617190321416, + "grad_norm": 0.6559387445449829, + "learning_rate": 8.297794086667165e-06, + "loss": 0.6161, + "step": 1815 + }, + { + "epoch": 0.8744432406404238, + "grad_norm": 0.7474731802940369, + "learning_rate": 8.235692083887613e-06, + "loss": 0.4046, + "step": 1816 + }, + { + "epoch": 0.8749247622487059, + "grad_norm": 1.1205002069473267, + "learning_rate": 8.173813363521843e-06, + "loss": 0.8628, + "step": 1817 + }, + { + "epoch": 0.875406283856988, + "grad_norm": 0.8745394349098206, + "learning_rate": 8.112158076134157e-06, + "loss": 0.6641, + "step": 1818 + }, + { + "epoch": 0.8758878054652702, + "grad_norm": 0.773175060749054, + "learning_rate": 8.05072637174522e-06, + "loss": 0.5763, + "step": 1819 + }, + { + "epoch": 0.8763693270735524, + "grad_norm": 0.4917106032371521, + "learning_rate": 7.989518399831641e-06, + "loss": 0.2699, + "step": 1820 + }, + { + "epoch": 0.8768508486818346, + "grad_norm": 0.6174960732460022, + "learning_rate": 7.928534309325675e-06, + "loss": 0.6247, + "step": 1821 + }, + { + "epoch": 0.8773323702901168, + "grad_norm": 1.1026489734649658, + "learning_rate": 7.8677742486148e-06, + "loss": 0.7072, + "step": 1822 + }, + { + "epoch": 0.877813891898399, + "grad_norm": 0.716529130935669, + "learning_rate": 7.807238365541391e-06, + "loss": 0.2818, + "step": 1823 + }, + { + "epoch": 0.8782954135066812, + "grad_norm": 0.7560042142868042, + "learning_rate": 7.746926807402344e-06, + "loss": 0.5247, + "step": 1824 + }, + { + "epoch": 0.8787769351149632, + "grad_norm": 0.9436668753623962, + "learning_rate": 7.686839720948736e-06, + "loss": 1.0009, + "step": 1825 + }, + { + "epoch": 0.8792584567232454, + "grad_norm": 1.0390268564224243, + "learning_rate": 7.6269772523854365e-06, + "loss": 0.2738, + "step": 1826 + }, + { + "epoch": 0.8797399783315276, + "grad_norm": 0.6893677115440369, + "learning_rate": 7.567339547370789e-06, + "loss": 0.3845, + "step": 1827 + }, + { + "epoch": 0.8802214999398098, + "grad_norm": 1.3264459371566772, + "learning_rate": 7.507926751016248e-06, + "loss": 0.4929, + "step": 1828 + }, + { + "epoch": 0.880703021548092, + "grad_norm": 0.6041299700737, + "learning_rate": 7.4487390078859855e-06, + "loss": 0.3404, + "step": 1829 + }, + { + "epoch": 0.8811845431563742, + "grad_norm": 0.7319620251655579, + "learning_rate": 7.389776461996578e-06, + "loss": 0.6722, + "step": 1830 + }, + { + "epoch": 0.8816660647646564, + "grad_norm": 0.9752634167671204, + "learning_rate": 7.331039256816663e-06, + "loss": 0.9134, + "step": 1831 + }, + { + "epoch": 0.8821475863729384, + "grad_norm": 1.0027194023132324, + "learning_rate": 7.27252753526656e-06, + "loss": 0.8346, + "step": 1832 + }, + { + "epoch": 0.8826291079812206, + "grad_norm": 0.8595365285873413, + "learning_rate": 7.214241439717962e-06, + "loss": 0.4265, + "step": 1833 + }, + { + "epoch": 0.8831106295895028, + "grad_norm": 0.7600406408309937, + "learning_rate": 7.1561811119935425e-06, + "loss": 0.3207, + "step": 1834 + }, + { + "epoch": 0.883592151197785, + "grad_norm": 0.5692286491394043, + "learning_rate": 7.098346693366642e-06, + "loss": 0.3267, + "step": 1835 + }, + { + "epoch": 0.8840736728060672, + "grad_norm": 0.8591275215148926, + "learning_rate": 7.0407383245609136e-06, + "loss": 0.3607, + "step": 1836 + }, + { + "epoch": 0.8845551944143494, + "grad_norm": 0.6609878540039062, + "learning_rate": 6.983356145749975e-06, + "loss": 0.2951, + "step": 1837 + }, + { + "epoch": 0.8850367160226316, + "grad_norm": 0.565071702003479, + "learning_rate": 6.9262002965570835e-06, + "loss": 0.1949, + "step": 1838 + }, + { + "epoch": 0.8855182376309136, + "grad_norm": 0.515185534954071, + "learning_rate": 6.869270916054782e-06, + "loss": 0.4361, + "step": 1839 + }, + { + "epoch": 0.8859997592391958, + "grad_norm": 0.5699068903923035, + "learning_rate": 6.812568142764575e-06, + "loss": 0.3235, + "step": 1840 + }, + { + "epoch": 0.886481280847478, + "grad_norm": 0.7445323467254639, + "learning_rate": 6.756092114656587e-06, + "loss": 0.622, + "step": 1841 + }, + { + "epoch": 0.8869628024557602, + "grad_norm": 0.9684603214263916, + "learning_rate": 6.699842969149195e-06, + "loss": 0.3984, + "step": 1842 + }, + { + "epoch": 0.8874443240640424, + "grad_norm": 0.6813137531280518, + "learning_rate": 6.64382084310875e-06, + "loss": 0.3141, + "step": 1843 + }, + { + "epoch": 0.8879258456723246, + "grad_norm": 1.0037721395492554, + "learning_rate": 6.5880258728491905e-06, + "loss": 0.4377, + "step": 1844 + }, + { + "epoch": 0.8884073672806068, + "grad_norm": 1.097926139831543, + "learning_rate": 6.532458194131763e-06, + "loss": 0.8514, + "step": 1845 + }, + { + "epoch": 0.8888888888888888, + "grad_norm": 0.8100180625915527, + "learning_rate": 6.477117942164657e-06, + "loss": 0.3044, + "step": 1846 + }, + { + "epoch": 0.889370410497171, + "grad_norm": 0.5506505370140076, + "learning_rate": 6.422005251602658e-06, + "loss": 0.2629, + "step": 1847 + }, + { + "epoch": 0.8898519321054532, + "grad_norm": 0.5061823725700378, + "learning_rate": 6.367120256546888e-06, + "loss": 0.2579, + "step": 1848 + }, + { + "epoch": 0.8903334537137354, + "grad_norm": 0.8128954768180847, + "learning_rate": 6.312463090544396e-06, + "loss": 0.4331, + "step": 1849 + }, + { + "epoch": 0.8908149753220176, + "grad_norm": 0.6516621708869934, + "learning_rate": 6.258033886587911e-06, + "loss": 0.8575, + "step": 1850 + }, + { + "epoch": 0.8912964969302998, + "grad_norm": 0.6090648174285889, + "learning_rate": 6.2038327771154485e-06, + "loss": 0.3023, + "step": 1851 + }, + { + "epoch": 0.891778018538582, + "grad_norm": 0.5383626818656921, + "learning_rate": 6.1498598940100346e-06, + "loss": 0.4444, + "step": 1852 + }, + { + "epoch": 0.892259540146864, + "grad_norm": 0.432041198015213, + "learning_rate": 6.0961153685993646e-06, + "loss": 0.2102, + "step": 1853 + }, + { + "epoch": 0.8927410617551462, + "grad_norm": 0.5301342010498047, + "learning_rate": 6.0425993316554965e-06, + "loss": 0.2474, + "step": 1854 + }, + { + "epoch": 0.8932225833634284, + "grad_norm": 0.5947489738464355, + "learning_rate": 5.989311913394546e-06, + "loss": 0.294, + "step": 1855 + }, + { + "epoch": 0.8937041049717106, + "grad_norm": 0.46525564789772034, + "learning_rate": 5.93625324347632e-06, + "loss": 0.1681, + "step": 1856 + }, + { + "epoch": 0.8941856265799928, + "grad_norm": 1.2389193773269653, + "learning_rate": 5.8834234510040335e-06, + "loss": 0.7665, + "step": 1857 + }, + { + "epoch": 0.894667148188275, + "grad_norm": 0.48213279247283936, + "learning_rate": 5.830822664523994e-06, + "loss": 0.0887, + "step": 1858 + }, + { + "epoch": 0.8951486697965572, + "grad_norm": 0.721508800983429, + "learning_rate": 5.77845101202531e-06, + "loss": 0.3518, + "step": 1859 + }, + { + "epoch": 0.8956301914048392, + "grad_norm": 0.5365184545516968, + "learning_rate": 5.726308620939536e-06, + "loss": 0.1631, + "step": 1860 + }, + { + "epoch": 0.8961117130131214, + "grad_norm": 0.500117838382721, + "learning_rate": 5.674395618140393e-06, + "loss": 0.2508, + "step": 1861 + }, + { + "epoch": 0.8965932346214036, + "grad_norm": 0.418396919965744, + "learning_rate": 5.622712129943453e-06, + "loss": 0.1585, + "step": 1862 + }, + { + "epoch": 0.8970747562296858, + "grad_norm": 0.5959372520446777, + "learning_rate": 5.571258282105829e-06, + "loss": 0.7838, + "step": 1863 + }, + { + "epoch": 0.897556277837968, + "grad_norm": 0.8515798449516296, + "learning_rate": 5.520034199825841e-06, + "loss": 1.1867, + "step": 1864 + }, + { + "epoch": 0.8980377994462502, + "grad_norm": 0.6227186918258667, + "learning_rate": 5.469040007742776e-06, + "loss": 0.6046, + "step": 1865 + }, + { + "epoch": 0.8985193210545324, + "grad_norm": 1.0233310461044312, + "learning_rate": 5.418275829936537e-06, + "loss": 1.1337, + "step": 1866 + }, + { + "epoch": 0.8990008426628145, + "grad_norm": 1.1459534168243408, + "learning_rate": 5.36774178992735e-06, + "loss": 0.5849, + "step": 1867 + }, + { + "epoch": 0.8994823642710966, + "grad_norm": 0.5045539736747742, + "learning_rate": 5.317438010675469e-06, + "loss": 0.3194, + "step": 1868 + }, + { + "epoch": 0.8999638858793788, + "grad_norm": 0.5089043974876404, + "learning_rate": 5.267364614580861e-06, + "loss": 0.1604, + "step": 1869 + }, + { + "epoch": 0.900445407487661, + "grad_norm": 0.706655740737915, + "learning_rate": 5.217521723482943e-06, + "loss": 0.3677, + "step": 1870 + }, + { + "epoch": 0.9009269290959432, + "grad_norm": 0.5909689664840698, + "learning_rate": 5.167909458660258e-06, + "loss": 0.5146, + "step": 1871 + }, + { + "epoch": 0.9014084507042254, + "grad_norm": 0.9625548720359802, + "learning_rate": 5.118527940830165e-06, + "loss": 0.3767, + "step": 1872 + }, + { + "epoch": 0.9018899723125076, + "grad_norm": 1.1841224431991577, + "learning_rate": 5.069377290148602e-06, + "loss": 0.4, + "step": 1873 + }, + { + "epoch": 0.9023714939207897, + "grad_norm": 0.5126852989196777, + "learning_rate": 5.020457626209707e-06, + "loss": 0.2379, + "step": 1874 + }, + { + "epoch": 0.9028530155290718, + "grad_norm": 0.9060919284820557, + "learning_rate": 4.971769068045628e-06, + "loss": 1.1507, + "step": 1875 + }, + { + "epoch": 0.903334537137354, + "grad_norm": 0.5319393873214722, + "learning_rate": 4.923311734126135e-06, + "loss": 0.4051, + "step": 1876 + }, + { + "epoch": 0.9038160587456362, + "grad_norm": 0.4319634735584259, + "learning_rate": 4.875085742358432e-06, + "loss": 0.5979, + "step": 1877 + }, + { + "epoch": 0.9042975803539184, + "grad_norm": 0.5022990107536316, + "learning_rate": 4.827091210086776e-06, + "loss": 0.2293, + "step": 1878 + }, + { + "epoch": 0.9047791019622006, + "grad_norm": 0.6433436870574951, + "learning_rate": 4.779328254092252e-06, + "loss": 0.7774, + "step": 1879 + }, + { + "epoch": 0.9052606235704828, + "grad_norm": 0.859475314617157, + "learning_rate": 4.731796990592452e-06, + "loss": 0.4518, + "step": 1880 + }, + { + "epoch": 0.905742145178765, + "grad_norm": 0.5635149478912354, + "learning_rate": 4.68449753524125e-06, + "loss": 0.6374, + "step": 1881 + }, + { + "epoch": 0.906223666787047, + "grad_norm": 0.5391207933425903, + "learning_rate": 4.637430003128429e-06, + "loss": 0.5627, + "step": 1882 + }, + { + "epoch": 0.9067051883953292, + "grad_norm": 1.4014676809310913, + "learning_rate": 4.5905945087794996e-06, + "loss": 0.6463, + "step": 1883 + }, + { + "epoch": 0.9071867100036114, + "grad_norm": 0.8849884271621704, + "learning_rate": 4.543991166155337e-06, + "loss": 0.8515, + "step": 1884 + }, + { + "epoch": 0.9076682316118936, + "grad_norm": 0.7712661623954773, + "learning_rate": 4.497620088651966e-06, + "loss": 0.3442, + "step": 1885 + }, + { + "epoch": 0.9081497532201758, + "grad_norm": 0.4782671332359314, + "learning_rate": 4.451481389100232e-06, + "loss": 0.5196, + "step": 1886 + }, + { + "epoch": 0.908631274828458, + "grad_norm": 0.9418804049491882, + "learning_rate": 4.405575179765586e-06, + "loss": 0.4203, + "step": 1887 + }, + { + "epoch": 0.9091127964367401, + "grad_norm": 0.6803520321846008, + "learning_rate": 4.359901572347758e-06, + "loss": 0.9799, + "step": 1888 + }, + { + "epoch": 0.9095943180450222, + "grad_norm": 0.8574094772338867, + "learning_rate": 4.314460677980537e-06, + "loss": 0.4231, + "step": 1889 + }, + { + "epoch": 0.9100758396533044, + "grad_norm": 0.4506451487541199, + "learning_rate": 4.269252607231422e-06, + "loss": 0.269, + "step": 1890 + }, + { + "epoch": 0.9105573612615866, + "grad_norm": 0.6377048492431641, + "learning_rate": 4.224277470101445e-06, + "loss": 0.447, + "step": 1891 + }, + { + "epoch": 0.9110388828698688, + "grad_norm": 0.6049963235855103, + "learning_rate": 4.179535376024857e-06, + "loss": 0.4994, + "step": 1892 + }, + { + "epoch": 0.911520404478151, + "grad_norm": 0.8465968370437622, + "learning_rate": 4.135026433868827e-06, + "loss": 0.6467, + "step": 1893 + }, + { + "epoch": 0.9120019260864332, + "grad_norm": 0.8001445531845093, + "learning_rate": 4.090750751933248e-06, + "loss": 0.9881, + "step": 1894 + }, + { + "epoch": 0.9124834476947153, + "grad_norm": 0.8922566175460815, + "learning_rate": 4.046708437950464e-06, + "loss": 0.5415, + "step": 1895 + }, + { + "epoch": 0.9129649693029974, + "grad_norm": 1.0354160070419312, + "learning_rate": 4.0028995990849084e-06, + "loss": 0.4927, + "step": 1896 + }, + { + "epoch": 0.9134464909112796, + "grad_norm": 0.4661247432231903, + "learning_rate": 3.95932434193299e-06, + "loss": 0.1731, + "step": 1897 + }, + { + "epoch": 0.9139280125195618, + "grad_norm": 0.596748948097229, + "learning_rate": 3.915982772522719e-06, + "loss": 0.4255, + "step": 1898 + }, + { + "epoch": 0.914409534127844, + "grad_norm": 0.9924038648605347, + "learning_rate": 3.872874996313513e-06, + "loss": 1.0724, + "step": 1899 + }, + { + "epoch": 0.9148910557361262, + "grad_norm": 0.7841641306877136, + "learning_rate": 3.830001118195936e-06, + "loss": 0.3032, + "step": 1900 + }, + { + "epoch": 0.9153725773444084, + "grad_norm": 0.9350888729095459, + "learning_rate": 3.787361242491394e-06, + "loss": 0.4085, + "step": 1901 + }, + { + "epoch": 0.9158540989526905, + "grad_norm": 0.3341303765773773, + "learning_rate": 3.744955472951928e-06, + "loss": 0.1407, + "step": 1902 + }, + { + "epoch": 0.9163356205609726, + "grad_norm": 1.7002257108688354, + "learning_rate": 3.702783912759955e-06, + "loss": 0.5921, + "step": 1903 + }, + { + "epoch": 0.9168171421692548, + "grad_norm": 0.5047416687011719, + "learning_rate": 3.660846664528006e-06, + "loss": 0.3457, + "step": 1904 + }, + { + "epoch": 0.917298663777537, + "grad_norm": 2.0711960792541504, + "learning_rate": 3.6191438302984772e-06, + "loss": 0.947, + "step": 1905 + }, + { + "epoch": 0.9177801853858192, + "grad_norm": 0.5945473909378052, + "learning_rate": 3.577675511543388e-06, + "loss": 0.4444, + "step": 1906 + }, + { + "epoch": 0.9182617069941014, + "grad_norm": 0.8059927225112915, + "learning_rate": 3.5364418091641373e-06, + "loss": 1.0539, + "step": 1907 + }, + { + "epoch": 0.9187432286023836, + "grad_norm": 0.6776530146598816, + "learning_rate": 3.495442823491224e-06, + "loss": 0.5971, + "step": 1908 + }, + { + "epoch": 0.9192247502106657, + "grad_norm": 1.1483100652694702, + "learning_rate": 3.4546786542840605e-06, + "loss": 0.6773, + "step": 1909 + }, + { + "epoch": 0.9197062718189479, + "grad_norm": 0.8443659543991089, + "learning_rate": 3.4141494007306816e-06, + "loss": 0.3389, + "step": 1910 + }, + { + "epoch": 0.92018779342723, + "grad_norm": 0.8820102214813232, + "learning_rate": 3.373855161447548e-06, + "loss": 1.1183, + "step": 1911 + }, + { + "epoch": 0.9206693150355122, + "grad_norm": 0.750338613986969, + "learning_rate": 3.333796034479242e-06, + "loss": 0.3541, + "step": 1912 + }, + { + "epoch": 0.9211508366437944, + "grad_norm": 0.4099019169807434, + "learning_rate": 3.293972117298294e-06, + "loss": 0.1908, + "step": 1913 + }, + { + "epoch": 0.9216323582520766, + "grad_norm": 0.8014377951622009, + "learning_rate": 3.2543835068049255e-06, + "loss": 0.3878, + "step": 1914 + }, + { + "epoch": 0.9221138798603588, + "grad_norm": 0.7586486339569092, + "learning_rate": 3.21503029932676e-06, + "loss": 0.4422, + "step": 1915 + }, + { + "epoch": 0.9225954014686409, + "grad_norm": 0.32165953516960144, + "learning_rate": 3.1759125906186793e-06, + "loss": 0.1944, + "step": 1916 + }, + { + "epoch": 0.9230769230769231, + "grad_norm": 0.6342649459838867, + "learning_rate": 3.137030475862535e-06, + "loss": 0.7593, + "step": 1917 + }, + { + "epoch": 0.9235584446852052, + "grad_norm": 0.6746546030044556, + "learning_rate": 3.098384049666925e-06, + "loss": 0.2116, + "step": 1918 + }, + { + "epoch": 0.9240399662934874, + "grad_norm": 0.5435047149658203, + "learning_rate": 3.059973406066963e-06, + "loss": 0.3394, + "step": 1919 + }, + { + "epoch": 0.9245214879017696, + "grad_norm": 0.5165433883666992, + "learning_rate": 3.0217986385240537e-06, + "loss": 0.2825, + "step": 1920 + }, + { + "epoch": 0.9250030095100518, + "grad_norm": 0.8822990655899048, + "learning_rate": 2.983859839925662e-06, + "loss": 0.6913, + "step": 1921 + }, + { + "epoch": 0.925484531118334, + "grad_norm": 1.065269112586975, + "learning_rate": 2.94615710258509e-06, + "loss": 0.9278, + "step": 1922 + }, + { + "epoch": 0.9259660527266161, + "grad_norm": 1.2459629774093628, + "learning_rate": 2.908690518241275e-06, + "loss": 0.6308, + "step": 1923 + }, + { + "epoch": 0.9264475743348983, + "grad_norm": 0.7520962953567505, + "learning_rate": 2.8714601780584937e-06, + "loss": 0.3369, + "step": 1924 + }, + { + "epoch": 0.9269290959431804, + "grad_norm": 0.5629241466522217, + "learning_rate": 2.834466172626238e-06, + "loss": 0.2532, + "step": 1925 + }, + { + "epoch": 0.9274106175514626, + "grad_norm": 0.9873923659324646, + "learning_rate": 2.7977085919589254e-06, + "loss": 0.4126, + "step": 1926 + }, + { + "epoch": 0.9278921391597448, + "grad_norm": 0.7369255423545837, + "learning_rate": 2.76118752549569e-06, + "loss": 0.169, + "step": 1927 + }, + { + "epoch": 0.928373660768027, + "grad_norm": 0.8374475240707397, + "learning_rate": 2.7249030621001924e-06, + "loss": 0.7864, + "step": 1928 + }, + { + "epoch": 0.9288551823763092, + "grad_norm": 0.8158120512962341, + "learning_rate": 2.688855290060399e-06, + "loss": 0.6308, + "step": 1929 + }, + { + "epoch": 0.9293367039845913, + "grad_norm": 0.8472880721092224, + "learning_rate": 2.653044297088314e-06, + "loss": 0.9682, + "step": 1930 + }, + { + "epoch": 0.9298182255928735, + "grad_norm": 0.49577999114990234, + "learning_rate": 2.6174701703198468e-06, + "loss": 0.3384, + "step": 1931 + }, + { + "epoch": 0.9302997472011556, + "grad_norm": 0.7291809916496277, + "learning_rate": 2.5821329963145347e-06, + "loss": 0.6925, + "step": 1932 + }, + { + "epoch": 0.9307812688094378, + "grad_norm": 0.7352073788642883, + "learning_rate": 2.547032861055376e-06, + "loss": 0.3493, + "step": 1933 + }, + { + "epoch": 0.93126279041772, + "grad_norm": 0.6518670320510864, + "learning_rate": 2.5121698499485757e-06, + "loss": 0.7209, + "step": 1934 + }, + { + "epoch": 0.9317443120260022, + "grad_norm": 0.9821415543556213, + "learning_rate": 2.4775440478233993e-06, + "loss": 0.8259, + "step": 1935 + }, + { + "epoch": 0.9322258336342844, + "grad_norm": 0.6522502899169922, + "learning_rate": 2.4431555389319074e-06, + "loss": 0.4079, + "step": 1936 + }, + { + "epoch": 0.9327073552425665, + "grad_norm": 0.7061333060264587, + "learning_rate": 2.4090044069487784e-06, + "loss": 0.7641, + "step": 1937 + }, + { + "epoch": 0.9331888768508487, + "grad_norm": 0.49332860112190247, + "learning_rate": 2.3750907349711084e-06, + "loss": 0.2268, + "step": 1938 + }, + { + "epoch": 0.9336703984591308, + "grad_norm": 1.0159000158309937, + "learning_rate": 2.3414146055182106e-06, + "loss": 0.785, + "step": 1939 + }, + { + "epoch": 0.934151920067413, + "grad_norm": 0.6537801027297974, + "learning_rate": 2.307976100531384e-06, + "loss": 0.2548, + "step": 1940 + }, + { + "epoch": 0.9346334416756952, + "grad_norm": 0.5711776614189148, + "learning_rate": 2.274775301373744e-06, + "loss": 0.4644, + "step": 1941 + }, + { + "epoch": 0.9351149632839774, + "grad_norm": 0.6775410175323486, + "learning_rate": 2.241812288830003e-06, + "loss": 0.8405, + "step": 1942 + }, + { + "epoch": 0.9355964848922596, + "grad_norm": 0.4718037247657776, + "learning_rate": 2.2090871431063253e-06, + "loss": 0.2021, + "step": 1943 + }, + { + "epoch": 0.9360780065005417, + "grad_norm": 0.9879113435745239, + "learning_rate": 2.176599943830071e-06, + "loss": 0.5937, + "step": 1944 + }, + { + "epoch": 0.9365595281088239, + "grad_norm": 0.8017491698265076, + "learning_rate": 2.144350770049597e-06, + "loss": 0.3747, + "step": 1945 + }, + { + "epoch": 0.937041049717106, + "grad_norm": 0.752362847328186, + "learning_rate": 2.112339700234156e-06, + "loss": 0.7747, + "step": 1946 + }, + { + "epoch": 0.9375225713253882, + "grad_norm": 0.8939694166183472, + "learning_rate": 2.0805668122735767e-06, + "loss": 0.3778, + "step": 1947 + }, + { + "epoch": 0.9380040929336704, + "grad_norm": 0.6203170418739319, + "learning_rate": 2.0490321834781833e-06, + "loss": 0.3018, + "step": 1948 + }, + { + "epoch": 0.9384856145419526, + "grad_norm": 0.8364834785461426, + "learning_rate": 2.0177358905785537e-06, + "loss": 0.5277, + "step": 1949 + }, + { + "epoch": 0.9389671361502347, + "grad_norm": 0.6893821358680725, + "learning_rate": 1.986678009725329e-06, + "loss": 0.5612, + "step": 1950 + }, + { + "epoch": 0.9394486577585169, + "grad_norm": 0.5284529328346252, + "learning_rate": 1.955858616489059e-06, + "loss": 0.7037, + "step": 1951 + }, + { + "epoch": 0.9399301793667991, + "grad_norm": 0.9045463800430298, + "learning_rate": 1.9252777858599915e-06, + "loss": 0.7651, + "step": 1952 + }, + { + "epoch": 0.9404117009750813, + "grad_norm": 0.7330710291862488, + "learning_rate": 1.8949355922479151e-06, + "loss": 0.4345, + "step": 1953 + }, + { + "epoch": 0.9408932225833634, + "grad_norm": 0.9409630298614502, + "learning_rate": 1.8648321094819287e-06, + "loss": 0.7097, + "step": 1954 + }, + { + "epoch": 0.9413747441916456, + "grad_norm": 0.9952296018600464, + "learning_rate": 1.8349674108103288e-06, + "loss": 0.5895, + "step": 1955 + }, + { + "epoch": 0.9418562657999278, + "grad_norm": 0.7514768838882446, + "learning_rate": 1.8053415689003872e-06, + "loss": 0.566, + "step": 1956 + }, + { + "epoch": 0.94233778740821, + "grad_norm": 0.7092472314834595, + "learning_rate": 1.7759546558381967e-06, + "loss": 0.2969, + "step": 1957 + }, + { + "epoch": 0.9428193090164921, + "grad_norm": 0.7995561361312866, + "learning_rate": 1.7468067431284707e-06, + "loss": 0.286, + "step": 1958 + }, + { + "epoch": 0.9433008306247743, + "grad_norm": 1.1525226831436157, + "learning_rate": 1.7178979016943764e-06, + "loss": 0.5039, + "step": 1959 + }, + { + "epoch": 0.9437823522330565, + "grad_norm": 0.40452930331230164, + "learning_rate": 1.6892282018773908e-06, + "loss": 0.2762, + "step": 1960 + }, + { + "epoch": 0.9442638738413386, + "grad_norm": 0.7908893823623657, + "learning_rate": 1.6607977134370789e-06, + "loss": 0.6573, + "step": 1961 + }, + { + "epoch": 0.9447453954496208, + "grad_norm": 0.7641875743865967, + "learning_rate": 1.6326065055510043e-06, + "loss": 0.5544, + "step": 1962 + }, + { + "epoch": 0.945226917057903, + "grad_norm": 0.8250647187232971, + "learning_rate": 1.6046546468144407e-06, + "loss": 0.6286, + "step": 1963 + }, + { + "epoch": 0.9457084386661851, + "grad_norm": 0.6536379456520081, + "learning_rate": 1.576942205240317e-06, + "loss": 0.361, + "step": 1964 + }, + { + "epoch": 0.9461899602744673, + "grad_norm": 0.6074386835098267, + "learning_rate": 1.5494692482590057e-06, + "loss": 0.4025, + "step": 1965 + }, + { + "epoch": 0.9466714818827495, + "grad_norm": 0.6296261548995972, + "learning_rate": 1.522235842718156e-06, + "loss": 0.5711, + "step": 1966 + }, + { + "epoch": 0.9471530034910317, + "grad_norm": 0.8962925672531128, + "learning_rate": 1.4952420548825285e-06, + "loss": 0.4009, + "step": 1967 + }, + { + "epoch": 0.9476345250993138, + "grad_norm": 0.6276614665985107, + "learning_rate": 1.468487950433839e-06, + "loss": 0.4574, + "step": 1968 + }, + { + "epoch": 0.948116046707596, + "grad_norm": 0.5954809188842773, + "learning_rate": 1.441973594470636e-06, + "loss": 0.4849, + "step": 1969 + }, + { + "epoch": 0.9485975683158782, + "grad_norm": 1.2679002285003662, + "learning_rate": 1.415699051508068e-06, + "loss": 0.5457, + "step": 1970 + }, + { + "epoch": 0.9490790899241603, + "grad_norm": 0.5073424577713013, + "learning_rate": 1.3896643854777847e-06, + "loss": 0.4817, + "step": 1971 + }, + { + "epoch": 0.9495606115324425, + "grad_norm": 0.6933119893074036, + "learning_rate": 1.3638696597277679e-06, + "loss": 0.3137, + "step": 1972 + }, + { + "epoch": 0.9500421331407247, + "grad_norm": 0.6285628080368042, + "learning_rate": 1.3383149370221449e-06, + "loss": 0.5385, + "step": 1973 + }, + { + "epoch": 0.9505236547490069, + "grad_norm": 1.3554258346557617, + "learning_rate": 1.313000279541121e-06, + "loss": 0.4391, + "step": 1974 + }, + { + "epoch": 0.951005176357289, + "grad_norm": 0.6554133892059326, + "learning_rate": 1.287925748880703e-06, + "loss": 0.4671, + "step": 1975 + }, + { + "epoch": 0.9514866979655712, + "grad_norm": 1.4406400918960571, + "learning_rate": 1.2630914060526522e-06, + "loss": 0.8691, + "step": 1976 + }, + { + "epoch": 0.9519682195738534, + "grad_norm": 0.5234692692756653, + "learning_rate": 1.2384973114843101e-06, + "loss": 0.5122, + "step": 1977 + }, + { + "epoch": 0.9524497411821355, + "grad_norm": 0.3509373664855957, + "learning_rate": 1.2141435250184185e-06, + "loss": 0.0802, + "step": 1978 + }, + { + "epoch": 0.9529312627904177, + "grad_norm": 0.504723072052002, + "learning_rate": 1.1900301059130093e-06, + "loss": 0.5345, + "step": 1979 + }, + { + "epoch": 0.9534127843986999, + "grad_norm": 0.5742740631103516, + "learning_rate": 1.1661571128412596e-06, + "loss": 0.4383, + "step": 1980 + }, + { + "epoch": 0.9538943060069821, + "grad_norm": 0.576581597328186, + "learning_rate": 1.142524603891315e-06, + "loss": 0.5606, + "step": 1981 + }, + { + "epoch": 0.9543758276152642, + "grad_norm": 0.7667593955993652, + "learning_rate": 1.1191326365661892e-06, + "loss": 0.5136, + "step": 1982 + }, + { + "epoch": 0.9548573492235464, + "grad_norm": 0.4355958104133606, + "learning_rate": 1.0959812677835968e-06, + "loss": 0.2748, + "step": 1983 + }, + { + "epoch": 0.9553388708318286, + "grad_norm": 0.7036843299865723, + "learning_rate": 1.0730705538758322e-06, + "loss": 0.5994, + "step": 1984 + }, + { + "epoch": 0.9558203924401107, + "grad_norm": 0.5516580939292908, + "learning_rate": 1.0504005505896141e-06, + "loss": 0.4582, + "step": 1985 + }, + { + "epoch": 0.9563019140483929, + "grad_norm": 0.5592047572135925, + "learning_rate": 1.0279713130859514e-06, + "loss": 0.5129, + "step": 1986 + }, + { + "epoch": 0.9567834356566751, + "grad_norm": 0.5601054430007935, + "learning_rate": 1.005782895940055e-06, + "loss": 0.8754, + "step": 1987 + }, + { + "epoch": 0.9572649572649573, + "grad_norm": 0.7053203582763672, + "learning_rate": 9.838353531411272e-07, + "loss": 0.4834, + "step": 1988 + }, + { + "epoch": 0.9577464788732394, + "grad_norm": 0.4365988075733185, + "learning_rate": 9.62128738092294e-07, + "loss": 0.5507, + "step": 1989 + }, + { + "epoch": 0.9582280004815216, + "grad_norm": 0.4267100989818573, + "learning_rate": 9.406631036104508e-07, + "loss": 0.2282, + "step": 1990 + }, + { + "epoch": 0.9587095220898038, + "grad_norm": 0.9049471020698547, + "learning_rate": 9.194385019261287e-07, + "loss": 0.7224, + "step": 1991 + }, + { + "epoch": 0.9591910436980859, + "grad_norm": 0.8407077789306641, + "learning_rate": 8.984549846833612e-07, + "loss": 0.407, + "step": 1992 + }, + { + "epoch": 0.9596725653063681, + "grad_norm": 0.5468326210975647, + "learning_rate": 8.777126029396065e-07, + "loss": 0.2352, + "step": 1993 + }, + { + "epoch": 0.9601540869146503, + "grad_norm": 0.5015347003936768, + "learning_rate": 8.572114071655479e-07, + "loss": 0.2969, + "step": 1994 + }, + { + "epoch": 0.9606356085229325, + "grad_norm": 0.6514408588409424, + "learning_rate": 8.369514472450379e-07, + "loss": 0.2341, + "step": 1995 + }, + { + "epoch": 0.9611171301312147, + "grad_norm": 0.6202306151390076, + "learning_rate": 8.169327724749543e-07, + "loss": 0.6092, + "step": 1996 + }, + { + "epoch": 0.9615986517394968, + "grad_norm": 0.596602737903595, + "learning_rate": 7.971554315650442e-07, + "loss": 0.5926, + "step": 1997 + }, + { + "epoch": 0.962080173347779, + "grad_norm": 1.011766791343689, + "learning_rate": 7.776194726378583e-07, + "loss": 0.8328, + "step": 1998 + }, + { + "epoch": 0.9625616949560611, + "grad_norm": 0.7206372022628784, + "learning_rate": 7.583249432286277e-07, + "loss": 0.4498, + "step": 1999 + }, + { + "epoch": 0.9630432165643433, + "grad_norm": 0.48031672835350037, + "learning_rate": 7.392718902850981e-07, + "loss": 0.306, + "step": 2000 + }, + { + "epoch": 0.9635247381726255, + "grad_norm": 0.7958242297172546, + "learning_rate": 7.204603601674853e-07, + "loss": 0.3053, + "step": 2001 + }, + { + "epoch": 0.9640062597809077, + "grad_norm": 1.0383106470108032, + "learning_rate": 7.018903986483083e-07, + "loss": 0.7788, + "step": 2002 + }, + { + "epoch": 0.9644877813891899, + "grad_norm": 0.9940798282623291, + "learning_rate": 6.835620509122897e-07, + "loss": 0.5206, + "step": 2003 + }, + { + "epoch": 0.964969302997472, + "grad_norm": 0.8290696144104004, + "learning_rate": 6.65475361556267e-07, + "loss": 0.452, + "step": 2004 + }, + { + "epoch": 0.9654508246057542, + "grad_norm": 0.6911272406578064, + "learning_rate": 6.47630374589081e-07, + "loss": 0.5752, + "step": 2005 + }, + { + "epoch": 0.9659323462140363, + "grad_norm": 0.6400201320648193, + "learning_rate": 6.300271334314434e-07, + "loss": 0.3883, + "step": 2006 + }, + { + "epoch": 0.9664138678223185, + "grad_norm": 0.5815731287002563, + "learning_rate": 6.126656809158359e-07, + "loss": 0.6059, + "step": 2007 + }, + { + "epoch": 0.9668953894306007, + "grad_norm": 0.5034971237182617, + "learning_rate": 5.955460592864337e-07, + "loss": 0.5949, + "step": 2008 + }, + { + "epoch": 0.9673769110388829, + "grad_norm": 0.64183509349823, + "learning_rate": 5.78668310198982e-07, + "loss": 0.7758, + "step": 2009 + }, + { + "epoch": 0.9678584326471651, + "grad_norm": 0.5774924159049988, + "learning_rate": 5.620324747207084e-07, + "loss": 0.2339, + "step": 2010 + }, + { + "epoch": 0.9683399542554472, + "grad_norm": 0.6001867055892944, + "learning_rate": 5.456385933301777e-07, + "loss": 0.2502, + "step": 2011 + }, + { + "epoch": 0.9688214758637294, + "grad_norm": 0.5865710377693176, + "learning_rate": 5.294867059172592e-07, + "loss": 0.2, + "step": 2012 + }, + { + "epoch": 0.9693029974720115, + "grad_norm": 0.47389328479766846, + "learning_rate": 5.135768517829819e-07, + "loss": 0.4203, + "step": 2013 + }, + { + "epoch": 0.9697845190802937, + "grad_norm": 0.6576399803161621, + "learning_rate": 4.979090696394795e-07, + "loss": 0.7084, + "step": 2014 + }, + { + "epoch": 0.9702660406885759, + "grad_norm": 0.9264883995056152, + "learning_rate": 4.824833976098453e-07, + "loss": 0.3988, + "step": 2015 + }, + { + "epoch": 0.9707475622968581, + "grad_norm": 1.5536925792694092, + "learning_rate": 4.6729987322807757e-07, + "loss": 0.9253, + "step": 2016 + }, + { + "epoch": 0.9712290839051403, + "grad_norm": 0.6507259607315063, + "learning_rate": 4.523585334389679e-07, + "loss": 0.806, + "step": 2017 + }, + { + "epoch": 0.9717106055134224, + "grad_norm": 0.9468237161636353, + "learning_rate": 4.3765941459804614e-07, + "loss": 0.4927, + "step": 2018 + }, + { + "epoch": 0.9721921271217046, + "grad_norm": 0.6918371915817261, + "learning_rate": 4.232025524714356e-07, + "loss": 0.1446, + "step": 2019 + }, + { + "epoch": 0.9726736487299867, + "grad_norm": 0.9984461665153503, + "learning_rate": 4.0898798223582e-07, + "loss": 0.5403, + "step": 2020 + }, + { + "epoch": 0.9731551703382689, + "grad_norm": 0.6529189348220825, + "learning_rate": 3.950157384783104e-07, + "loss": 0.5878, + "step": 2021 + }, + { + "epoch": 0.9736366919465511, + "grad_norm": 0.6561586260795593, + "learning_rate": 3.8128585519640046e-07, + "loss": 0.3575, + "step": 2022 + }, + { + "epoch": 0.9741182135548333, + "grad_norm": 0.8341250419616699, + "learning_rate": 3.677983657978779e-07, + "loss": 0.7522, + "step": 2023 + }, + { + "epoch": 0.9745997351631155, + "grad_norm": 0.7429258823394775, + "learning_rate": 3.545533031007131e-07, + "loss": 0.4519, + "step": 2024 + }, + { + "epoch": 0.9750812567713976, + "grad_norm": 0.771535336971283, + "learning_rate": 3.415506993330153e-07, + "loss": 0.2805, + "step": 2025 + }, + { + "epoch": 0.9755627783796798, + "grad_norm": 1.0194131135940552, + "learning_rate": 3.2879058613292105e-07, + "loss": 0.2996, + "step": 2026 + }, + { + "epoch": 0.9760442999879619, + "grad_norm": 0.5492636561393738, + "learning_rate": 3.1627299454856095e-07, + "loss": 0.3451, + "step": 2027 + }, + { + "epoch": 0.9765258215962441, + "grad_norm": 0.6983237862586975, + "learning_rate": 3.0399795503793793e-07, + "loss": 0.8118, + "step": 2028 + }, + { + "epoch": 0.9770073432045263, + "grad_norm": 0.5727863311767578, + "learning_rate": 2.9196549746888235e-07, + "loss": 0.1593, + "step": 2029 + }, + { + "epoch": 0.9774888648128085, + "grad_norm": 0.5290704965591431, + "learning_rate": 2.801756511189524e-07, + "loss": 0.1473, + "step": 2030 + }, + { + "epoch": 0.9779703864210907, + "grad_norm": 0.647200882434845, + "learning_rate": 2.686284446754006e-07, + "loss": 0.2473, + "step": 2031 + }, + { + "epoch": 0.9784519080293728, + "grad_norm": 0.8799815773963928, + "learning_rate": 2.573239062350963e-07, + "loss": 0.9664, + "step": 2032 + }, + { + "epoch": 0.978933429637655, + "grad_norm": 0.8440240025520325, + "learning_rate": 2.4626206330440326e-07, + "loss": 1.0026, + "step": 2033 + }, + { + "epoch": 0.9794149512459371, + "grad_norm": 0.7759560942649841, + "learning_rate": 2.3544294279918e-07, + "loss": 0.4478, + "step": 2034 + }, + { + "epoch": 0.9798964728542193, + "grad_norm": 0.4280017614364624, + "learning_rate": 2.2486657104471286e-07, + "loss": 0.6288, + "step": 2035 + }, + { + "epoch": 0.9803779944625015, + "grad_norm": 0.7536220550537109, + "learning_rate": 2.1453297377557191e-07, + "loss": 0.3241, + "step": 2036 + }, + { + "epoch": 0.9808595160707837, + "grad_norm": 0.8434854745864868, + "learning_rate": 2.044421761356552e-07, + "loss": 0.345, + "step": 2037 + }, + { + "epoch": 0.9813410376790659, + "grad_norm": 1.095892310142517, + "learning_rate": 1.9459420267804452e-07, + "loss": 1.0977, + "step": 2038 + }, + { + "epoch": 0.9818225592873481, + "grad_norm": 1.2428241968154907, + "learning_rate": 1.8498907736499426e-07, + "loss": 0.7439, + "step": 2039 + }, + { + "epoch": 0.9823040808956301, + "grad_norm": 1.0324878692626953, + "learning_rate": 1.7562682356786487e-07, + "loss": 0.4408, + "step": 2040 + }, + { + "epoch": 0.9827856025039123, + "grad_norm": 0.7506169676780701, + "learning_rate": 1.665074640670228e-07, + "loss": 0.8056, + "step": 2041 + }, + { + "epoch": 0.9832671241121945, + "grad_norm": 0.48109811544418335, + "learning_rate": 1.576310210518517e-07, + "loss": 0.2994, + "step": 2042 + }, + { + "epoch": 0.9837486457204767, + "grad_norm": 0.7194755673408508, + "learning_rate": 1.489975161206636e-07, + "loss": 0.572, + "step": 2043 + }, + { + "epoch": 0.9842301673287589, + "grad_norm": 0.6078580617904663, + "learning_rate": 1.406069702806323e-07, + "loss": 0.2252, + "step": 2044 + }, + { + "epoch": 0.9847116889370411, + "grad_norm": 1.0418319702148438, + "learning_rate": 1.324594039477822e-07, + "loss": 0.9186, + "step": 2045 + }, + { + "epoch": 0.9851932105453233, + "grad_norm": 0.6754590272903442, + "learning_rate": 1.2455483694689962e-07, + "loss": 0.9269, + "step": 2046 + }, + { + "epoch": 0.9856747321536053, + "grad_norm": 0.754641592502594, + "learning_rate": 1.1689328851151038e-07, + "loss": 0.9815, + "step": 2047 + }, + { + "epoch": 0.9861562537618875, + "grad_norm": 0.5206847786903381, + "learning_rate": 1.0947477728381339e-07, + "loss": 0.8868, + "step": 2048 + }, + { + "epoch": 0.9866377753701697, + "grad_norm": 0.599271297454834, + "learning_rate": 1.0229932131465836e-07, + "loss": 0.3564, + "step": 2049 + }, + { + "epoch": 0.9871192969784519, + "grad_norm": 0.6615093350410461, + "learning_rate": 9.536693806347919e-08, + "loss": 0.5842, + "step": 2050 + }, + { + "epoch": 0.9876008185867341, + "grad_norm": 0.693345308303833, + "learning_rate": 8.867764439826065e-08, + "loss": 0.3433, + "step": 2051 + }, + { + "epoch": 0.9880823401950163, + "grad_norm": 0.5857682824134827, + "learning_rate": 8.223145659550513e-08, + "loss": 0.6968, + "step": 2052 + }, + { + "epoch": 0.9885638618032985, + "grad_norm": 0.8526471853256226, + "learning_rate": 7.602839034017706e-08, + "loss": 0.6844, + "step": 2053 + }, + { + "epoch": 0.9890453834115805, + "grad_norm": 0.7888714671134949, + "learning_rate": 7.006846072568074e-08, + "loss": 0.3751, + "step": 2054 + }, + { + "epoch": 0.9895269050198627, + "grad_norm": 0.969866156578064, + "learning_rate": 6.435168225381594e-08, + "loss": 0.5924, + "step": 2055 + }, + { + "epoch": 0.9900084266281449, + "grad_norm": 0.3491519093513489, + "learning_rate": 5.887806883474456e-08, + "loss": 0.1858, + "step": 2056 + }, + { + "epoch": 0.9904899482364271, + "grad_norm": 0.8504678606987, + "learning_rate": 5.364763378694626e-08, + "loss": 0.4558, + "step": 2057 + }, + { + "epoch": 0.9909714698447093, + "grad_norm": 0.9942469596862793, + "learning_rate": 4.8660389837207334e-08, + "loss": 0.8354, + "step": 2058 + }, + { + "epoch": 0.9914529914529915, + "grad_norm": 0.7020037770271301, + "learning_rate": 4.391634912056519e-08, + "loss": 1.1202, + "step": 2059 + }, + { + "epoch": 0.9919345130612737, + "grad_norm": 0.4268772602081299, + "learning_rate": 3.9415523180297286e-08, + "loss": 0.1628, + "step": 2060 + }, + { + "epoch": 0.9924160346695557, + "grad_norm": 0.5923699140548706, + "learning_rate": 3.515792296789888e-08, + "loss": 0.1115, + "step": 2061 + }, + { + "epoch": 0.9928975562778379, + "grad_norm": 0.8240038156509399, + "learning_rate": 3.114355884301645e-08, + "loss": 0.7954, + "step": 2062 + }, + { + "epoch": 0.9933790778861201, + "grad_norm": 0.8708813190460205, + "learning_rate": 2.7372440573469883e-08, + "loss": 0.4003, + "step": 2063 + }, + { + "epoch": 0.9938605994944023, + "grad_norm": 0.6724766492843628, + "learning_rate": 2.384457733520806e-08, + "loss": 0.225, + "step": 2064 + }, + { + "epoch": 0.9943421211026845, + "grad_norm": 0.7021605372428894, + "learning_rate": 2.0559977712297785e-08, + "loss": 0.2706, + "step": 2065 + }, + { + "epoch": 0.9948236427109667, + "grad_norm": 0.7431803941726685, + "learning_rate": 1.7518649696857126e-08, + "loss": 0.6236, + "step": 2066 + }, + { + "epoch": 0.9953051643192489, + "grad_norm": 0.4876454174518585, + "learning_rate": 1.4720600689110963e-08, + "loss": 0.596, + "step": 2067 + }, + { + "epoch": 0.995786685927531, + "grad_norm": 0.9122818112373352, + "learning_rate": 1.216583749731326e-08, + "loss": 0.5965, + "step": 2068 + }, + { + "epoch": 0.9962682075358131, + "grad_norm": 0.8722514510154724, + "learning_rate": 9.854366337758159e-09, + "loss": 0.8851, + "step": 2069 + }, + { + "epoch": 0.9967497291440953, + "grad_norm": 0.9117795825004578, + "learning_rate": 7.786192834746686e-09, + "loss": 0.8438, + "step": 2070 + }, + { + "epoch": 0.9972312507523775, + "grad_norm": 0.7254170179367065, + "learning_rate": 5.961322020608951e-09, + "loss": 0.4586, + "step": 2071 + }, + { + "epoch": 0.9977127723606597, + "grad_norm": 0.433788001537323, + "learning_rate": 4.3797583356264275e-09, + "loss": 0.1281, + "step": 2072 + }, + { + "epoch": 0.9981942939689419, + "grad_norm": 0.6030900478363037, + "learning_rate": 3.0415056281096755e-09, + "loss": 0.371, + "step": 2073 + }, + { + "epoch": 0.9986758155772241, + "grad_norm": 0.9900925755500793, + "learning_rate": 1.9465671543095197e-09, + "loss": 0.7417, + "step": 2074 + }, + { + "epoch": 0.9991573371855061, + "grad_norm": 0.774526834487915, + "learning_rate": 1.094945578439255e-09, + "loss": 0.6423, + "step": 2075 + }, + { + "epoch": 0.9996388587937883, + "grad_norm": 0.49781298637390137, + "learning_rate": 4.866429726857469e-10, + "loss": 0.582, + "step": 2076 + }, + { + "epoch": 1.0, + "grad_norm": 0.6113485097885132, + "learning_rate": 1.2166081717612797e-10, + "loss": 0.4426, + "step": 2077 + }, + { + "epoch": 1.0, + "step": 2077, + "total_flos": 4.4863387190181914e+17, + "train_loss": 0.6395983156794302, + "train_runtime": 14822.4278, + "train_samples_per_second": 2.242, + "train_steps_per_second": 0.14 + } + ], + "logging_steps": 1, + "max_steps": 2077, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 2400000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 4.4863387190181914e+17, + "train_batch_size": 4, + "trial_name": null, + "trial_params": null +}