diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,7033 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.1381406271584473, + "eval_steps": 500, + "global_step": 1000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0001381406271584473, + "grad_norm": 0.13134899735450745, + "learning_rate": 0.0, + "loss": 0.2546, + "step": 1 + }, + { + "epoch": 0.0002762812543168946, + "grad_norm": 0.1231406033039093, + "learning_rate": 8.695768706980801e-06, + "loss": 0.2051, + "step": 2 + }, + { + "epoch": 0.0004144218814753419, + "grad_norm": 0.08561521023511887, + "learning_rate": 1.3782467315509068e-05, + "loss": 0.1455, + "step": 3 + }, + { + "epoch": 0.0005525625086337892, + "grad_norm": 0.1449182778596878, + "learning_rate": 1.7391537413961602e-05, + "loss": 0.3489, + "step": 4 + }, + { + "epoch": 0.0006907031357922365, + "grad_norm": 0.07311359792947769, + "learning_rate": 2.019094966738107e-05, + "loss": 0.1115, + "step": 5 + }, + { + "epoch": 0.0008288437629506838, + "grad_norm": 0.09480362385511398, + "learning_rate": 2.247823602248987e-05, + "loss": 0.1602, + "step": 6 + }, + { + "epoch": 0.0009669843901091311, + "grad_norm": 0.08785154670476913, + "learning_rate": 2.441210908061704e-05, + "loss": 0.1448, + "step": 7 + }, + { + "epoch": 0.0011051250172675784, + "grad_norm": 0.10113295167684555, + "learning_rate": 2.60873061209424e-05, + "loss": 0.1726, + "step": 8 + }, + { + "epoch": 0.0012432656444260257, + "grad_norm": 0.10489944368600845, + "learning_rate": 2.7564934631018137e-05, + "loss": 0.1343, + "step": 9 + }, + { + "epoch": 0.001381406271584473, + "grad_norm": 0.07018060237169266, + "learning_rate": 2.8886718374361883e-05, + "loss": 0.0986, + "step": 10 + }, + { + "epoch": 0.0015195468987429202, + "grad_norm": 0.07492844760417938, + "learning_rate": 3.008241721328615e-05, + "loss": 0.1243, + "step": 11 + }, + { + "epoch": 0.0016576875259013675, + "grad_norm": 0.02947733923792839, + "learning_rate": 3.117400472947067e-05, + "loss": 0.0427, + "step": 12 + }, + { + "epoch": 0.0017958281530598148, + "grad_norm": 0.07098696380853653, + "learning_rate": 3.217816790308016e-05, + "loss": 0.074, + "step": 13 + }, + { + "epoch": 0.0019339687802182623, + "grad_norm": 0.06729952991008759, + "learning_rate": 3.3107877787597844e-05, + "loss": 0.0756, + "step": 14 + }, + { + "epoch": 0.0020721094073767096, + "grad_norm": 0.047634709626436234, + "learning_rate": 3.397341698289014e-05, + "loss": 0.0335, + "step": 15 + }, + { + "epoch": 0.002210250034535157, + "grad_norm": 0.01928865537047386, + "learning_rate": 3.4783074827923204e-05, + "loss": 0.0253, + "step": 16 + }, + { + "epoch": 0.002348390661693604, + "grad_norm": 0.10047394782304764, + "learning_rate": 3.5543631465891534e-05, + "loss": 0.118, + "step": 17 + }, + { + "epoch": 0.0024865312888520514, + "grad_norm": 0.05856269970536232, + "learning_rate": 3.626070333799893e-05, + "loss": 0.0372, + "step": 18 + }, + { + "epoch": 0.0026246719160104987, + "grad_norm": 0.011952841654419899, + "learning_rate": 3.69389951409255e-05, + "loss": 0.0123, + "step": 19 + }, + { + "epoch": 0.002762812543168946, + "grad_norm": 0.0623110830783844, + "learning_rate": 3.7582487081342674e-05, + "loss": 0.0615, + "step": 20 + }, + { + "epoch": 0.002900953170327393, + "grad_norm": 0.020977556705474854, + "learning_rate": 3.819457639612611e-05, + "loss": 0.0215, + "step": 21 + }, + { + "epoch": 0.0030390937974858405, + "grad_norm": 0.02789384499192238, + "learning_rate": 3.8778185920266955e-05, + "loss": 0.0263, + "step": 22 + }, + { + "epoch": 0.0031772344246442877, + "grad_norm": 0.04567106068134308, + "learning_rate": 3.9335848501569437e-05, + "loss": 0.0298, + "step": 23 + }, + { + "epoch": 0.003315375051802735, + "grad_norm": 0.09989040344953537, + "learning_rate": 3.9869773436451475e-05, + "loss": 0.0894, + "step": 24 + }, + { + "epoch": 0.0034535156789611823, + "grad_norm": 0.07369614392518997, + "learning_rate": 4.038189933476214e-05, + "loss": 0.0671, + "step": 25 + }, + { + "epoch": 0.0035916563061196296, + "grad_norm": 0.1597849428653717, + "learning_rate": 4.087393661006097e-05, + "loss": 0.0934, + "step": 26 + }, + { + "epoch": 0.0037297969332780773, + "grad_norm": 0.06050405278801918, + "learning_rate": 4.13474019465272e-05, + "loss": 0.0418, + "step": 27 + }, + { + "epoch": 0.0038679375604365246, + "grad_norm": 0.12362563610076904, + "learning_rate": 4.180364649457864e-05, + "loss": 0.1035, + "step": 28 + }, + { + "epoch": 0.004006078187594971, + "grad_norm": 0.0370742604136467, + "learning_rate": 4.22438791165378e-05, + "loss": 0.0298, + "step": 29 + }, + { + "epoch": 0.004144218814753419, + "grad_norm": 0.13234247267246246, + "learning_rate": 4.2669185689870944e-05, + "loss": 0.1149, + "step": 30 + }, + { + "epoch": 0.004282359441911866, + "grad_norm": 0.11454839259386063, + "learning_rate": 4.308054524410194e-05, + "loss": 0.0825, + "step": 31 + }, + { + "epoch": 0.004420500069070314, + "grad_norm": 0.055333029478788376, + "learning_rate": 4.347884353490401e-05, + "loss": 0.0309, + "step": 32 + }, + { + "epoch": 0.0045586406962287605, + "grad_norm": 0.010015130043029785, + "learning_rate": 4.386488452879522e-05, + "loss": 0.0112, + "step": 33 + }, + { + "epoch": 0.004696781323387208, + "grad_norm": 0.02108495496213436, + "learning_rate": 4.423940017287234e-05, + "loss": 0.0144, + "step": 34 + }, + { + "epoch": 0.004834921950545655, + "grad_norm": 0.045554034411907196, + "learning_rate": 4.460305874799812e-05, + "loss": 0.0325, + "step": 35 + }, + { + "epoch": 0.004973062577704103, + "grad_norm": 0.05177578702569008, + "learning_rate": 4.495647204497974e-05, + "loss": 0.0333, + "step": 36 + }, + { + "epoch": 0.0051112032048625505, + "grad_norm": 0.08447037637233734, + "learning_rate": 4.5300201557312036e-05, + "loss": 0.0489, + "step": 37 + }, + { + "epoch": 0.005249343832020997, + "grad_norm": 0.03207926079630852, + "learning_rate": 4.56347638479063e-05, + "loss": 0.0217, + "step": 38 + }, + { + "epoch": 0.005387484459179445, + "grad_norm": 0.04534975811839104, + "learning_rate": 4.596063521858923e-05, + "loss": 0.0332, + "step": 39 + }, + { + "epoch": 0.005525625086337892, + "grad_norm": 0.04216322675347328, + "learning_rate": 4.627825578832348e-05, + "loss": 0.0363, + "step": 40 + }, + { + "epoch": 0.0056637657134963396, + "grad_norm": 0.04155502840876579, + "learning_rate": 4.65880330677802e-05, + "loss": 0.0238, + "step": 41 + }, + { + "epoch": 0.005801906340654786, + "grad_norm": 0.15111327171325684, + "learning_rate": 4.6890345103106905e-05, + "loss": 0.1467, + "step": 42 + }, + { + "epoch": 0.005940046967813234, + "grad_norm": 0.057506952434778214, + "learning_rate": 4.7185543249731356e-05, + "loss": 0.0366, + "step": 43 + }, + { + "epoch": 0.006078187594971681, + "grad_norm": 0.15154612064361572, + "learning_rate": 4.747395462724775e-05, + "loss": 0.2312, + "step": 44 + }, + { + "epoch": 0.006216328222130129, + "grad_norm": 0.1836182028055191, + "learning_rate": 4.775588429839921e-05, + "loss": 0.0679, + "step": 45 + }, + { + "epoch": 0.0063544688492885755, + "grad_norm": 0.04606263339519501, + "learning_rate": 4.803161720855024e-05, + "loss": 0.0216, + "step": 46 + }, + { + "epoch": 0.006492609476447023, + "grad_norm": 0.07287360727787018, + "learning_rate": 4.830141991656282e-05, + "loss": 0.0355, + "step": 47 + }, + { + "epoch": 0.00663075010360547, + "grad_norm": 0.10376694798469543, + "learning_rate": 4.856554214343227e-05, + "loss": 0.0384, + "step": 48 + }, + { + "epoch": 0.006768890730763918, + "grad_norm": 0.0253737885504961, + "learning_rate": 4.882421816123408e-05, + "loss": 0.0145, + "step": 49 + }, + { + "epoch": 0.006907031357922365, + "grad_norm": 0.045460280030965805, + "learning_rate": 4.9077668041742954e-05, + "loss": 0.0272, + "step": 50 + }, + { + "epoch": 0.007045171985080812, + "grad_norm": 0.00954690482467413, + "learning_rate": 4.93260987814006e-05, + "loss": 0.0086, + "step": 51 + }, + { + "epoch": 0.007183312612239259, + "grad_norm": 0.12355463951826096, + "learning_rate": 4.956970531704177e-05, + "loss": 0.0398, + "step": 52 + }, + { + "epoch": 0.007321453239397707, + "grad_norm": 0.028379764407873154, + "learning_rate": 4.9808671444865916e-05, + "loss": 0.0166, + "step": 53 + }, + { + "epoch": 0.0074595938665561546, + "grad_norm": 0.029451319947838783, + "learning_rate": 5.0043170653507996e-05, + "loss": 0.0285, + "step": 54 + }, + { + "epoch": 0.007597734493714601, + "grad_norm": 0.1718253642320633, + "learning_rate": 5.027336688066723e-05, + "loss": 0.1021, + "step": 55 + }, + { + "epoch": 0.007735875120873049, + "grad_norm": 0.02496183291077614, + "learning_rate": 5.049941520155945e-05, + "loss": 0.0167, + "step": 56 + }, + { + "epoch": 0.007874015748031496, + "grad_norm": 0.14813463389873505, + "learning_rate": 5.072146245643457e-05, + "loss": 0.0571, + "step": 57 + }, + { + "epoch": 0.008012156375189943, + "grad_norm": 0.004978721961379051, + "learning_rate": 5.09396478235186e-05, + "loss": 0.0063, + "step": 58 + }, + { + "epoch": 0.008150297002348391, + "grad_norm": 0.027830282226204872, + "learning_rate": 5.115410334297882e-05, + "loss": 0.0209, + "step": 59 + }, + { + "epoch": 0.008288437629506838, + "grad_norm": 0.02240496128797531, + "learning_rate": 5.136495439685174e-05, + "loss": 0.0166, + "step": 60 + }, + { + "epoch": 0.008426578256665285, + "grad_norm": 0.004362566862255335, + "learning_rate": 5.157232014930198e-05, + "loss": 0.005, + "step": 61 + }, + { + "epoch": 0.008564718883823732, + "grad_norm": 0.31787729263305664, + "learning_rate": 5.1776313951082735e-05, + "loss": 0.1389, + "step": 62 + }, + { + "epoch": 0.00870285951098218, + "grad_norm": 0.009066242724657059, + "learning_rate": 5.1977043711635176e-05, + "loss": 0.0081, + "step": 63 + }, + { + "epoch": 0.008841000138140627, + "grad_norm": 0.006892406847327948, + "learning_rate": 5.21746122418848e-05, + "loss": 0.007, + "step": 64 + }, + { + "epoch": 0.008979140765299074, + "grad_norm": 0.06632635742425919, + "learning_rate": 5.236911757046123e-05, + "loss": 0.0397, + "step": 65 + }, + { + "epoch": 0.009117281392457521, + "grad_norm": 0.03089607134461403, + "learning_rate": 5.2560653235776016e-05, + "loss": 0.0194, + "step": 66 + }, + { + "epoch": 0.00925542201961597, + "grad_norm": 0.010297919623553753, + "learning_rate": 5.27493085561372e-05, + "loss": 0.0094, + "step": 67 + }, + { + "epoch": 0.009393562646774416, + "grad_norm": 0.00596292968839407, + "learning_rate": 5.293516887985315e-05, + "loss": 0.0056, + "step": 68 + }, + { + "epoch": 0.009531703273932863, + "grad_norm": 0.03583955764770508, + "learning_rate": 5.31183158170785e-05, + "loss": 0.034, + "step": 69 + }, + { + "epoch": 0.00966984390109131, + "grad_norm": 0.022406477481126785, + "learning_rate": 5.3298827454978915e-05, + "loss": 0.0177, + "step": 70 + }, + { + "epoch": 0.009807984528249759, + "grad_norm": 0.02521800622344017, + "learning_rate": 5.347677855763414e-05, + "loss": 0.0137, + "step": 71 + }, + { + "epoch": 0.009946125155408206, + "grad_norm": 0.023860285058617592, + "learning_rate": 5.3652240751960537e-05, + "loss": 0.0223, + "step": 72 + }, + { + "epoch": 0.010084265782566652, + "grad_norm": 0.044385574758052826, + "learning_rate": 5.3825282700810095e-05, + "loss": 0.023, + "step": 73 + }, + { + "epoch": 0.010222406409725101, + "grad_norm": 0.011455412954092026, + "learning_rate": 5.399597026429285e-05, + "loss": 0.0115, + "step": 74 + }, + { + "epoch": 0.010360547036883548, + "grad_norm": 0.04733498394489288, + "learning_rate": 5.416436665027121e-05, + "loss": 0.0301, + "step": 75 + }, + { + "epoch": 0.010498687664041995, + "grad_norm": 0.00907475408166647, + "learning_rate": 5.43305325548871e-05, + "loss": 0.0068, + "step": 76 + }, + { + "epoch": 0.010636828291200441, + "grad_norm": 0.14968900382518768, + "learning_rate": 5.449452629390319e-05, + "loss": 0.0619, + "step": 77 + }, + { + "epoch": 0.01077496891835889, + "grad_norm": 0.15661145746707916, + "learning_rate": 5.4656403925570034e-05, + "loss": 0.0722, + "step": 78 + }, + { + "epoch": 0.010913109545517337, + "grad_norm": 0.004802990239113569, + "learning_rate": 5.481621936566648e-05, + "loss": 0.0054, + "step": 79 + }, + { + "epoch": 0.011051250172675784, + "grad_norm": 0.0075927311554551125, + "learning_rate": 5.4974024495304276e-05, + "loss": 0.0071, + "step": 80 + }, + { + "epoch": 0.01118939079983423, + "grad_norm": 0.11258488148450851, + "learning_rate": 5.5129869262036274e-05, + "loss": 0.0985, + "step": 81 + }, + { + "epoch": 0.011327531426992679, + "grad_norm": 0.052467264235019684, + "learning_rate": 5.5283801774761e-05, + "loss": 0.0298, + "step": 82 + }, + { + "epoch": 0.011465672054151126, + "grad_norm": 0.05677814409136772, + "learning_rate": 5.543586839287528e-05, + "loss": 0.0272, + "step": 83 + }, + { + "epoch": 0.011603812681309573, + "grad_norm": 0.11758122593164444, + "learning_rate": 5.55861138100877e-05, + "loss": 0.0555, + "step": 84 + }, + { + "epoch": 0.01174195330846802, + "grad_norm": 0.021579308435320854, + "learning_rate": 5.573458113327261e-05, + "loss": 0.0141, + "step": 85 + }, + { + "epoch": 0.011880093935626468, + "grad_norm": 0.018838461488485336, + "learning_rate": 5.5881311956712153e-05, + "loss": 0.016, + "step": 86 + }, + { + "epoch": 0.012018234562784915, + "grad_norm": 0.07577263563871384, + "learning_rate": 5.602634643204686e-05, + "loss": 0.0568, + "step": 87 + }, + { + "epoch": 0.012156375189943362, + "grad_norm": 0.16609540581703186, + "learning_rate": 5.6169723334228564e-05, + "loss": 0.2139, + "step": 88 + }, + { + "epoch": 0.01229451581710181, + "grad_norm": 0.008787001483142376, + "learning_rate": 5.631148012374702e-05, + "loss": 0.0088, + "step": 89 + }, + { + "epoch": 0.012432656444260257, + "grad_norm": 0.005372604820877314, + "learning_rate": 5.645165300538001e-05, + "loss": 0.0064, + "step": 90 + }, + { + "epoch": 0.012570797071418704, + "grad_norm": 0.020076818764209747, + "learning_rate": 5.65902769836972e-05, + "loss": 0.0146, + "step": 91 + }, + { + "epoch": 0.012708937698577151, + "grad_norm": 0.02066107466816902, + "learning_rate": 5.672738591553104e-05, + "loss": 0.0245, + "step": 92 + }, + { + "epoch": 0.0128470783257356, + "grad_norm": 0.03774566575884819, + "learning_rate": 5.6863012559611006e-05, + "loss": 0.0254, + "step": 93 + }, + { + "epoch": 0.012985218952894046, + "grad_norm": 0.0144899757578969, + "learning_rate": 5.699718862354363e-05, + "loss": 0.0153, + "step": 94 + }, + { + "epoch": 0.013123359580052493, + "grad_norm": 0.06870149821043015, + "learning_rate": 5.712994480830657e-05, + "loss": 0.0499, + "step": 95 + }, + { + "epoch": 0.01326150020721094, + "grad_norm": 0.02667887881398201, + "learning_rate": 5.726131085041308e-05, + "loss": 0.0202, + "step": 96 + }, + { + "epoch": 0.013399640834369389, + "grad_norm": 0.012837301008403301, + "learning_rate": 5.7391315561891545e-05, + "loss": 0.0113, + "step": 97 + }, + { + "epoch": 0.013537781461527836, + "grad_norm": 0.052629996091127396, + "learning_rate": 5.751998686821489e-05, + "loss": 0.0505, + "step": 98 + }, + { + "epoch": 0.013675922088686282, + "grad_norm": 0.020552391186356544, + "learning_rate": 5.764735184430429e-05, + "loss": 0.0186, + "step": 99 + }, + { + "epoch": 0.01381406271584473, + "grad_norm": 0.019090712070465088, + "learning_rate": 5.7773436748723766e-05, + "loss": 0.0126, + "step": 100 + }, + { + "epoch": 0.013952203343003178, + "grad_norm": 0.05564524605870247, + "learning_rate": 5.7898267056173305e-05, + "loss": 0.0311, + "step": 101 + }, + { + "epoch": 0.014090343970161625, + "grad_norm": 0.003988645505160093, + "learning_rate": 5.80218674883814e-05, + "loss": 0.0043, + "step": 102 + }, + { + "epoch": 0.014228484597320071, + "grad_norm": 0.014693093486130238, + "learning_rate": 5.814426204349046e-05, + "loss": 0.0135, + "step": 103 + }, + { + "epoch": 0.014366625224478518, + "grad_norm": 0.010103048756718636, + "learning_rate": 5.826547402402257e-05, + "loss": 0.0088, + "step": 104 + }, + { + "epoch": 0.014504765851636967, + "grad_norm": 0.06549032032489777, + "learning_rate": 5.838552606350718e-05, + "loss": 0.0285, + "step": 105 + }, + { + "epoch": 0.014642906478795414, + "grad_norm": 0.014358027838170528, + "learning_rate": 5.850444015184672e-05, + "loss": 0.0131, + "step": 106 + }, + { + "epoch": 0.01478104710595386, + "grad_norm": 0.009209529496729374, + "learning_rate": 5.862223765949125e-05, + "loss": 0.0084, + "step": 107 + }, + { + "epoch": 0.014919187733112309, + "grad_norm": 0.004671560134738684, + "learning_rate": 5.873893936048881e-05, + "loss": 0.0058, + "step": 108 + }, + { + "epoch": 0.015057328360270756, + "grad_norm": 0.018950628116726875, + "learning_rate": 5.885456545447318e-05, + "loss": 0.0122, + "step": 109 + }, + { + "epoch": 0.015195468987429203, + "grad_norm": 0.009532412514090538, + "learning_rate": 5.8969135587648026e-05, + "loss": 0.0081, + "step": 110 + }, + { + "epoch": 0.01533360961458765, + "grad_norm": 0.11297975480556488, + "learning_rate": 5.90826688728211e-05, + "loss": 0.0985, + "step": 111 + }, + { + "epoch": 0.015471750241746098, + "grad_norm": 0.021297922357916832, + "learning_rate": 5.9195183908540244e-05, + "loss": 0.0172, + "step": 112 + }, + { + "epoch": 0.015609890868904545, + "grad_norm": 0.011318448930978775, + "learning_rate": 5.930669879737879e-05, + "loss": 0.0106, + "step": 113 + }, + { + "epoch": 0.015748031496062992, + "grad_norm": 0.01391973253339529, + "learning_rate": 5.9417231163415365e-05, + "loss": 0.0135, + "step": 114 + }, + { + "epoch": 0.01588617212322144, + "grad_norm": 0.004168018698692322, + "learning_rate": 5.9526798168950515e-05, + "loss": 0.0048, + "step": 115 + }, + { + "epoch": 0.016024312750379886, + "grad_norm": 0.08173353970050812, + "learning_rate": 5.96354165304994e-05, + "loss": 0.0481, + "step": 116 + }, + { + "epoch": 0.016162453377538334, + "grad_norm": 0.03505641967058182, + "learning_rate": 5.9743102534098305e-05, + "loss": 0.0212, + "step": 117 + }, + { + "epoch": 0.016300594004696783, + "grad_norm": 0.020770832896232605, + "learning_rate": 5.9849872049959624e-05, + "loss": 0.0154, + "step": 118 + }, + { + "epoch": 0.016438734631855228, + "grad_norm": 0.013789334334433079, + "learning_rate": 5.9955740546508573e-05, + "loss": 0.0138, + "step": 119 + }, + { + "epoch": 0.016576875259013676, + "grad_norm": 0.004777516704052687, + "learning_rate": 6.006072310383254e-05, + "loss": 0.0051, + "step": 120 + }, + { + "epoch": 0.01671501588617212, + "grad_norm": 0.09303977340459824, + "learning_rate": 6.01648344265723e-05, + "loss": 0.0718, + "step": 121 + }, + { + "epoch": 0.01685315651333057, + "grad_norm": 0.006040315609425306, + "learning_rate": 6.026808885628279e-05, + "loss": 0.006, + "step": 122 + }, + { + "epoch": 0.01699129714048902, + "grad_norm": 0.0029070712625980377, + "learning_rate": 6.0370500383289253e-05, + "loss": 0.0041, + "step": 123 + }, + { + "epoch": 0.017129437767647464, + "grad_norm": 0.018031740561127663, + "learning_rate": 6.0472082658063546e-05, + "loss": 0.0118, + "step": 124 + }, + { + "epoch": 0.017267578394805912, + "grad_norm": 0.0054465411230921745, + "learning_rate": 6.057284900214323e-05, + "loss": 0.006, + "step": 125 + }, + { + "epoch": 0.01740571902196436, + "grad_norm": 0.024758221581578255, + "learning_rate": 6.0672812418615974e-05, + "loss": 0.0181, + "step": 126 + }, + { + "epoch": 0.017543859649122806, + "grad_norm": 0.060118090361356735, + "learning_rate": 6.077198560218932e-05, + "loss": 0.0314, + "step": 127 + }, + { + "epoch": 0.017682000276281255, + "grad_norm": 0.0032233481761068106, + "learning_rate": 6.087038094886561e-05, + "loss": 0.0044, + "step": 128 + }, + { + "epoch": 0.017820140903439703, + "grad_norm": 0.023020045831799507, + "learning_rate": 6.096801056524042e-05, + "loss": 0.0205, + "step": 129 + }, + { + "epoch": 0.01795828153059815, + "grad_norm": 0.012697757221758366, + "learning_rate": 6.106488627744203e-05, + "loss": 0.0126, + "step": 130 + }, + { + "epoch": 0.018096422157756597, + "grad_norm": 0.023354003205895424, + "learning_rate": 6.116101963972834e-05, + "loss": 0.012, + "step": 131 + }, + { + "epoch": 0.018234562784915042, + "grad_norm": 0.020162884145975113, + "learning_rate": 6.125642194275681e-05, + "loss": 0.0168, + "step": 132 + }, + { + "epoch": 0.01837270341207349, + "grad_norm": 0.02640388533473015, + "learning_rate": 6.135110422154255e-05, + "loss": 0.0148, + "step": 133 + }, + { + "epoch": 0.01851084403923194, + "grad_norm": 0.004422938451170921, + "learning_rate": 6.1445077263118e-05, + "loss": 0.0053, + "step": 134 + }, + { + "epoch": 0.018648984666390384, + "grad_norm": 0.0029348707757890224, + "learning_rate": 6.153835161390827e-05, + "loss": 0.0041, + "step": 135 + }, + { + "epoch": 0.018787125293548833, + "grad_norm": 0.008122744038701057, + "learning_rate": 6.163093758683395e-05, + "loss": 0.0072, + "step": 136 + }, + { + "epoch": 0.01892526592070728, + "grad_norm": 0.012685508467257023, + "learning_rate": 6.172284526815391e-05, + "loss": 0.0108, + "step": 137 + }, + { + "epoch": 0.019063406547865726, + "grad_norm": 0.07143422216176987, + "learning_rate": 6.181408452405932e-05, + "loss": 0.0354, + "step": 138 + }, + { + "epoch": 0.019201547175024175, + "grad_norm": 0.013023455627262592, + "learning_rate": 6.190466500702942e-05, + "loss": 0.0121, + "step": 139 + }, + { + "epoch": 0.01933968780218262, + "grad_norm": 0.11455456912517548, + "learning_rate": 6.199459616195971e-05, + "loss": 0.0916, + "step": 140 + }, + { + "epoch": 0.01947782842934107, + "grad_norm": 0.004115304443985224, + "learning_rate": 6.208388723207189e-05, + "loss": 0.005, + "step": 141 + }, + { + "epoch": 0.019615969056499517, + "grad_norm": 0.006390172056853771, + "learning_rate": 6.217254726461495e-05, + "loss": 0.0056, + "step": 142 + }, + { + "epoch": 0.019754109683657962, + "grad_norm": 0.0032778135500848293, + "learning_rate": 6.22605851163663e-05, + "loss": 0.0043, + "step": 143 + }, + { + "epoch": 0.01989225031081641, + "grad_norm": 0.005337041802704334, + "learning_rate": 6.234800945894134e-05, + "loss": 0.0054, + "step": 144 + }, + { + "epoch": 0.02003039093797486, + "grad_norm": 0.004124751314520836, + "learning_rate": 6.243482878391886e-05, + "loss": 0.0051, + "step": 145 + }, + { + "epoch": 0.020168531565133305, + "grad_norm": 0.005773474462330341, + "learning_rate": 6.25210514077909e-05, + "loss": 0.006, + "step": 146 + }, + { + "epoch": 0.020306672192291753, + "grad_norm": 0.003976911772042513, + "learning_rate": 6.260668547674315e-05, + "loss": 0.005, + "step": 147 + }, + { + "epoch": 0.020444812819450202, + "grad_norm": 0.015533742494881153, + "learning_rate": 6.269173897127364e-05, + "loss": 0.0115, + "step": 148 + }, + { + "epoch": 0.020582953446608647, + "grad_norm": 0.007217871956527233, + "learning_rate": 6.277621971065575e-05, + "loss": 0.007, + "step": 149 + }, + { + "epoch": 0.020721094073767096, + "grad_norm": 0.02741631306707859, + "learning_rate": 6.286013535725201e-05, + "loss": 0.0262, + "step": 150 + }, + { + "epoch": 0.02085923470092554, + "grad_norm": 0.009277957491576672, + "learning_rate": 6.294349342068456e-05, + "loss": 0.0074, + "step": 151 + }, + { + "epoch": 0.02099737532808399, + "grad_norm": 0.03387785702943802, + "learning_rate": 6.30263012618679e-05, + "loss": 0.02, + "step": 152 + }, + { + "epoch": 0.021135515955242438, + "grad_norm": 0.02018630877137184, + "learning_rate": 6.310856609690967e-05, + "loss": 0.0135, + "step": 153 + }, + { + "epoch": 0.021273656582400883, + "grad_norm": 0.0030231664422899485, + "learning_rate": 6.3190295000884e-05, + "loss": 0.0039, + "step": 154 + }, + { + "epoch": 0.02141179720955933, + "grad_norm": 0.038589298725128174, + "learning_rate": 6.327149491148301e-05, + "loss": 0.0358, + "step": 155 + }, + { + "epoch": 0.02154993783671778, + "grad_norm": 0.11658922582864761, + "learning_rate": 6.335217263255083e-05, + "loss": 0.0583, + "step": 156 + }, + { + "epoch": 0.021688078463876225, + "grad_norm": 0.015740083530545235, + "learning_rate": 6.343233483750467e-05, + "loss": 0.0134, + "step": 157 + }, + { + "epoch": 0.021826219091034674, + "grad_norm": 0.04937628656625748, + "learning_rate": 6.351198807264728e-05, + "loss": 0.0245, + "step": 158 + }, + { + "epoch": 0.021964359718193122, + "grad_norm": 0.03167060390114784, + "learning_rate": 6.359113876037498e-05, + "loss": 0.0142, + "step": 159 + }, + { + "epoch": 0.022102500345351567, + "grad_norm": 0.0016946723917499185, + "learning_rate": 6.366979320228508e-05, + "loss": 0.0025, + "step": 160 + }, + { + "epoch": 0.022240640972510016, + "grad_norm": 0.017945002764463425, + "learning_rate": 6.374795758218648e-05, + "loss": 0.0114, + "step": 161 + }, + { + "epoch": 0.02237878159966846, + "grad_norm": 0.015089811757206917, + "learning_rate": 6.382563796901706e-05, + "loss": 0.0104, + "step": 162 + }, + { + "epoch": 0.02251692222682691, + "grad_norm": 0.037023257464170456, + "learning_rate": 6.390284031967139e-05, + "loss": 0.0212, + "step": 163 + }, + { + "epoch": 0.022655062853985358, + "grad_norm": 0.00312859402038157, + "learning_rate": 6.397957048174181e-05, + "loss": 0.0041, + "step": 164 + }, + { + "epoch": 0.022793203481143803, + "grad_norm": 0.016800010576844215, + "learning_rate": 6.405583419617628e-05, + "loss": 0.0117, + "step": 165 + }, + { + "epoch": 0.022931344108302252, + "grad_norm": 0.007050058338791132, + "learning_rate": 6.413163709985608e-05, + "loss": 0.0077, + "step": 166 + }, + { + "epoch": 0.0230694847354607, + "grad_norm": 0.010559717193245888, + "learning_rate": 6.420698472809568e-05, + "loss": 0.0101, + "step": 167 + }, + { + "epoch": 0.023207625362619146, + "grad_norm": 0.007447375915944576, + "learning_rate": 6.428188251706851e-05, + "loss": 0.0063, + "step": 168 + }, + { + "epoch": 0.023345765989777594, + "grad_norm": 0.016936203464865685, + "learning_rate": 6.435633580616032e-05, + "loss": 0.0123, + "step": 169 + }, + { + "epoch": 0.02348390661693604, + "grad_norm": 0.012545132078230381, + "learning_rate": 6.443034984025341e-05, + "loss": 0.0106, + "step": 170 + }, + { + "epoch": 0.023622047244094488, + "grad_norm": 0.004759025294333696, + "learning_rate": 6.450392977194364e-05, + "loss": 0.0055, + "step": 171 + }, + { + "epoch": 0.023760187871252936, + "grad_norm": 0.004177019465714693, + "learning_rate": 6.457708066369296e-05, + "loss": 0.005, + "step": 172 + }, + { + "epoch": 0.02389832849841138, + "grad_norm": 0.005387986544519663, + "learning_rate": 6.464980748991957e-05, + "loss": 0.0052, + "step": 173 + }, + { + "epoch": 0.02403646912556983, + "grad_norm": 0.014265534467995167, + "learning_rate": 6.472211513902766e-05, + "loss": 0.0105, + "step": 174 + }, + { + "epoch": 0.02417460975272828, + "grad_norm": 0.009414114989340305, + "learning_rate": 6.47940084153792e-05, + "loss": 0.0082, + "step": 175 + }, + { + "epoch": 0.024312750379886724, + "grad_norm": 0.006094431504607201, + "learning_rate": 6.486549204120935e-05, + "loss": 0.007, + "step": 176 + }, + { + "epoch": 0.024450891007045172, + "grad_norm": 0.01183071918785572, + "learning_rate": 6.493657065848789e-05, + "loss": 0.0051, + "step": 177 + }, + { + "epoch": 0.02458903163420362, + "grad_norm": 0.01014156173914671, + "learning_rate": 6.500724883072781e-05, + "loss": 0.008, + "step": 178 + }, + { + "epoch": 0.024727172261362066, + "grad_norm": 0.011028112843632698, + "learning_rate": 6.507753104474373e-05, + "loss": 0.0117, + "step": 179 + }, + { + "epoch": 0.024865312888520515, + "grad_norm": 0.09993032366037369, + "learning_rate": 6.514742171236081e-05, + "loss": 0.0368, + "step": 180 + }, + { + "epoch": 0.02500345351567896, + "grad_norm": 0.00730351684615016, + "learning_rate": 6.521692517207681e-05, + "loss": 0.0055, + "step": 181 + }, + { + "epoch": 0.02514159414283741, + "grad_norm": 0.012081836350262165, + "learning_rate": 6.5286045690678e-05, + "loss": 0.0096, + "step": 182 + }, + { + "epoch": 0.025279734769995857, + "grad_norm": 0.03789973631501198, + "learning_rate": 6.535478746481105e-05, + "loss": 0.0214, + "step": 183 + }, + { + "epoch": 0.025417875397154302, + "grad_norm": 0.15203222632408142, + "learning_rate": 6.542315462251184e-05, + "loss": 0.173, + "step": 184 + }, + { + "epoch": 0.02555601602431275, + "grad_norm": 0.01032046414911747, + "learning_rate": 6.54911512246931e-05, + "loss": 0.0096, + "step": 185 + }, + { + "epoch": 0.0256941566514712, + "grad_norm": 0.028638096526265144, + "learning_rate": 6.555878126659181e-05, + "loss": 0.0145, + "step": 186 + }, + { + "epoch": 0.025832297278629644, + "grad_norm": 0.003815449308604002, + "learning_rate": 6.562604867917768e-05, + "loss": 0.005, + "step": 187 + }, + { + "epoch": 0.025970437905788093, + "grad_norm": 0.02196453884243965, + "learning_rate": 6.569295733052443e-05, + "loss": 0.0163, + "step": 188 + }, + { + "epoch": 0.026108578532946538, + "grad_norm": 0.027504578232765198, + "learning_rate": 6.575951102714424e-05, + "loss": 0.0236, + "step": 189 + }, + { + "epoch": 0.026246719160104987, + "grad_norm": 0.01594018191099167, + "learning_rate": 6.582571351528738e-05, + "loss": 0.0151, + "step": 190 + }, + { + "epoch": 0.026384859787263435, + "grad_norm": 0.014933554455637932, + "learning_rate": 6.589156848220747e-05, + "loss": 0.0081, + "step": 191 + }, + { + "epoch": 0.02652300041442188, + "grad_norm": 0.0055963508784770966, + "learning_rate": 6.595707955739388e-05, + "loss": 0.0062, + "step": 192 + }, + { + "epoch": 0.02666114104158033, + "grad_norm": 0.050148047506809235, + "learning_rate": 6.602225031377195e-05, + "loss": 0.0194, + "step": 193 + }, + { + "epoch": 0.026799281668738777, + "grad_norm": 0.029830805957317352, + "learning_rate": 6.608708426887235e-05, + "loss": 0.0205, + "step": 194 + }, + { + "epoch": 0.026937422295897222, + "grad_norm": 0.011922827921807766, + "learning_rate": 6.61515848859703e-05, + "loss": 0.0096, + "step": 195 + }, + { + "epoch": 0.02707556292305567, + "grad_norm": 0.009799070656299591, + "learning_rate": 6.621575557519569e-05, + "loss": 0.01, + "step": 196 + }, + { + "epoch": 0.02721370355021412, + "grad_norm": 0.01243631262332201, + "learning_rate": 6.627959969461484e-05, + "loss": 0.0099, + "step": 197 + }, + { + "epoch": 0.027351844177372565, + "grad_norm": 0.07121865451335907, + "learning_rate": 6.634312055128508e-05, + "loss": 0.0304, + "step": 198 + }, + { + "epoch": 0.027489984804531013, + "grad_norm": 0.017733663320541382, + "learning_rate": 6.64063214022826e-05, + "loss": 0.0137, + "step": 199 + }, + { + "epoch": 0.02762812543168946, + "grad_norm": 0.012368605472147465, + "learning_rate": 6.646920545570455e-05, + "loss": 0.0084, + "step": 200 + }, + { + "epoch": 0.027766266058847907, + "grad_norm": 0.02112652361392975, + "learning_rate": 6.653177587164626e-05, + "loss": 0.0179, + "step": 201 + }, + { + "epoch": 0.027904406686006356, + "grad_norm": 0.02623148076236248, + "learning_rate": 6.659403576315411e-05, + "loss": 0.0197, + "step": 202 + }, + { + "epoch": 0.0280425473131648, + "grad_norm": 0.03918753191828728, + "learning_rate": 6.665598819715483e-05, + "loss": 0.0258, + "step": 203 + }, + { + "epoch": 0.02818068794032325, + "grad_norm": 0.013951225206255913, + "learning_rate": 6.67176361953622e-05, + "loss": 0.0135, + "step": 204 + }, + { + "epoch": 0.028318828567481698, + "grad_norm": 0.007952759973704815, + "learning_rate": 6.677898273516128e-05, + "loss": 0.0089, + "step": 205 + }, + { + "epoch": 0.028456969194640143, + "grad_norm": 0.018430588766932487, + "learning_rate": 6.684003075047128e-05, + "loss": 0.0149, + "step": 206 + }, + { + "epoch": 0.02859510982179859, + "grad_norm": 0.023949814960360527, + "learning_rate": 6.690078313258756e-05, + "loss": 0.0132, + "step": 207 + }, + { + "epoch": 0.028733250448957037, + "grad_norm": 0.024928990751504898, + "learning_rate": 6.696124273100337e-05, + "loss": 0.0167, + "step": 208 + }, + { + "epoch": 0.028871391076115485, + "grad_norm": 0.030003704130649567, + "learning_rate": 6.702141235421166e-05, + "loss": 0.0172, + "step": 209 + }, + { + "epoch": 0.029009531703273934, + "grad_norm": 0.0721101239323616, + "learning_rate": 6.708129477048798e-05, + "loss": 0.0368, + "step": 210 + }, + { + "epoch": 0.02914767233043238, + "grad_norm": 0.08490179479122162, + "learning_rate": 6.71408927086548e-05, + "loss": 0.0536, + "step": 211 + }, + { + "epoch": 0.029285812957590827, + "grad_norm": 0.03323817253112793, + "learning_rate": 6.720020885882752e-05, + "loss": 0.0178, + "step": 212 + }, + { + "epoch": 0.029423953584749276, + "grad_norm": 0.1837650090456009, + "learning_rate": 6.72592458731432e-05, + "loss": 0.1725, + "step": 213 + }, + { + "epoch": 0.02956209421190772, + "grad_norm": 0.019839327782392502, + "learning_rate": 6.731800636647207e-05, + "loss": 0.0181, + "step": 214 + }, + { + "epoch": 0.02970023483906617, + "grad_norm": 0.1045767292380333, + "learning_rate": 6.737649291711243e-05, + "loss": 0.0833, + "step": 215 + }, + { + "epoch": 0.029838375466224618, + "grad_norm": 0.006606792565435171, + "learning_rate": 6.743470806746961e-05, + "loss": 0.0075, + "step": 216 + }, + { + "epoch": 0.029976516093383063, + "grad_norm": 0.03129244223237038, + "learning_rate": 6.749265432471897e-05, + "loss": 0.0193, + "step": 217 + }, + { + "epoch": 0.030114656720541512, + "grad_norm": 0.045653145760297775, + "learning_rate": 6.755033416145399e-05, + "loss": 0.0199, + "step": 218 + }, + { + "epoch": 0.030252797347699957, + "grad_norm": 0.010944471694529057, + "learning_rate": 6.760775001631917e-05, + "loss": 0.0095, + "step": 219 + }, + { + "epoch": 0.030390937974858406, + "grad_norm": 0.01569080539047718, + "learning_rate": 6.766490429462884e-05, + "loss": 0.0104, + "step": 220 + }, + { + "epoch": 0.030529078602016854, + "grad_norm": 0.021444043144583702, + "learning_rate": 6.772179936897169e-05, + "loss": 0.0142, + "step": 221 + }, + { + "epoch": 0.0306672192291753, + "grad_norm": 0.28277868032455444, + "learning_rate": 6.77784375798019e-05, + "loss": 0.2374, + "step": 222 + }, + { + "epoch": 0.030805359856333748, + "grad_norm": 0.45913952589035034, + "learning_rate": 6.783482123601665e-05, + "loss": 0.4551, + "step": 223 + }, + { + "epoch": 0.030943500483492196, + "grad_norm": 0.020372901111841202, + "learning_rate": 6.789095261552104e-05, + "loss": 0.0168, + "step": 224 + }, + { + "epoch": 0.03108164111065064, + "grad_norm": 0.004885848145931959, + "learning_rate": 6.794683396578028e-05, + "loss": 0.005, + "step": 225 + }, + { + "epoch": 0.03121978173780909, + "grad_norm": 0.01693105325102806, + "learning_rate": 6.80024675043596e-05, + "loss": 0.0163, + "step": 226 + }, + { + "epoch": 0.03135792236496754, + "grad_norm": 0.11799801141023636, + "learning_rate": 6.805785541945228e-05, + "loss": 0.0717, + "step": 227 + }, + { + "epoch": 0.031496062992125984, + "grad_norm": 0.036976467818021774, + "learning_rate": 6.811299987039618e-05, + "loss": 0.0245, + "step": 228 + }, + { + "epoch": 0.03163420361928443, + "grad_norm": 0.02155802771449089, + "learning_rate": 6.816790298817876e-05, + "loss": 0.0177, + "step": 229 + }, + { + "epoch": 0.03177234424644288, + "grad_norm": 0.07842877507209778, + "learning_rate": 6.822256687593131e-05, + "loss": 0.0581, + "step": 230 + }, + { + "epoch": 0.031910484873601326, + "grad_norm": 0.03338398411870003, + "learning_rate": 6.827699360941226e-05, + "loss": 0.0325, + "step": 231 + }, + { + "epoch": 0.03204862550075977, + "grad_norm": 0.032553404569625854, + "learning_rate": 6.83311852374802e-05, + "loss": 0.0322, + "step": 232 + }, + { + "epoch": 0.03218676612791822, + "grad_norm": 0.04135012626647949, + "learning_rate": 6.838514378255668e-05, + "loss": 0.0309, + "step": 233 + }, + { + "epoch": 0.03232490675507667, + "grad_norm": 0.015881696715950966, + "learning_rate": 6.84388712410791e-05, + "loss": 0.0122, + "step": 234 + }, + { + "epoch": 0.03246304738223511, + "grad_norm": 0.007995386607944965, + "learning_rate": 6.849236958394389e-05, + "loss": 0.0064, + "step": 235 + }, + { + "epoch": 0.032601188009393565, + "grad_norm": 0.02736758068203926, + "learning_rate": 6.854564075694041e-05, + "loss": 0.0235, + "step": 236 + }, + { + "epoch": 0.03273932863655201, + "grad_norm": 0.002618682337924838, + "learning_rate": 6.859868668117555e-05, + "loss": 0.0042, + "step": 237 + }, + { + "epoch": 0.032877469263710456, + "grad_norm": 0.005231305491179228, + "learning_rate": 6.865150925348938e-05, + "loss": 0.0068, + "step": 238 + }, + { + "epoch": 0.03301560989086891, + "grad_norm": 0.09725295752286911, + "learning_rate": 6.870411034686229e-05, + "loss": 0.0792, + "step": 239 + }, + { + "epoch": 0.03315375051802735, + "grad_norm": 0.050293877720832825, + "learning_rate": 6.875649181081335e-05, + "loss": 0.0283, + "step": 240 + }, + { + "epoch": 0.0332918911451858, + "grad_norm": 0.028621939942240715, + "learning_rate": 6.880865547179058e-05, + "loss": 0.0281, + "step": 241 + }, + { + "epoch": 0.03343003177234424, + "grad_norm": 0.008942426182329655, + "learning_rate": 6.886060313355311e-05, + "loss": 0.0102, + "step": 242 + }, + { + "epoch": 0.033568172399502695, + "grad_norm": 0.017691925168037415, + "learning_rate": 6.891233657754533e-05, + "loss": 0.0144, + "step": 243 + }, + { + "epoch": 0.03370631302666114, + "grad_norm": 0.06508173048496246, + "learning_rate": 6.896385756326359e-05, + "loss": 0.0417, + "step": 244 + }, + { + "epoch": 0.033844453653819585, + "grad_norm": 0.04150944948196411, + "learning_rate": 6.901516782861516e-05, + "loss": 0.0291, + "step": 245 + }, + { + "epoch": 0.03398259428097804, + "grad_norm": 0.03793828561902046, + "learning_rate": 6.906626909027006e-05, + "loss": 0.0233, + "step": 246 + }, + { + "epoch": 0.03412073490813648, + "grad_norm": 0.010130356065928936, + "learning_rate": 6.911716304400567e-05, + "loss": 0.0099, + "step": 247 + }, + { + "epoch": 0.03425887553529493, + "grad_norm": 0.13118235766887665, + "learning_rate": 6.916785136504435e-05, + "loss": 0.0867, + "step": 248 + }, + { + "epoch": 0.03439701616245338, + "grad_norm": 0.01260526105761528, + "learning_rate": 6.921833570838434e-05, + "loss": 0.0125, + "step": 249 + }, + { + "epoch": 0.034535156789611825, + "grad_norm": 0.008999558165669441, + "learning_rate": 6.926861770912402e-05, + "loss": 0.0073, + "step": 250 + }, + { + "epoch": 0.03467329741677027, + "grad_norm": 0.00830650795251131, + "learning_rate": 6.931869898277965e-05, + "loss": 0.0091, + "step": 251 + }, + { + "epoch": 0.03481143804392872, + "grad_norm": 0.008377288468182087, + "learning_rate": 6.936858112559677e-05, + "loss": 0.0075, + "step": 252 + }, + { + "epoch": 0.03494957867108717, + "grad_norm": 0.053149525076150894, + "learning_rate": 6.941826571485559e-05, + "loss": 0.0127, + "step": 253 + }, + { + "epoch": 0.03508771929824561, + "grad_norm": 0.00812042597681284, + "learning_rate": 6.946775430917013e-05, + "loss": 0.0089, + "step": 254 + }, + { + "epoch": 0.035225859925404064, + "grad_norm": 0.11070810258388519, + "learning_rate": 6.951704844878168e-05, + "loss": 0.0831, + "step": 255 + }, + { + "epoch": 0.03536400055256251, + "grad_norm": 0.012675175443291664, + "learning_rate": 6.956614965584641e-05, + "loss": 0.0114, + "step": 256 + }, + { + "epoch": 0.035502141179720954, + "grad_norm": 0.04479183256626129, + "learning_rate": 6.96150594347174e-05, + "loss": 0.0212, + "step": 257 + }, + { + "epoch": 0.035640281806879406, + "grad_norm": 0.0070395260117948055, + "learning_rate": 6.966377927222123e-05, + "loss": 0.0085, + "step": 258 + }, + { + "epoch": 0.03577842243403785, + "grad_norm": 0.0065799071453511715, + "learning_rate": 6.971231063792908e-05, + "loss": 0.0077, + "step": 259 + }, + { + "epoch": 0.0359165630611963, + "grad_norm": 0.011478329077363014, + "learning_rate": 6.976065498442284e-05, + "loss": 0.0096, + "step": 260 + }, + { + "epoch": 0.03605470368835474, + "grad_norm": 0.005571560934185982, + "learning_rate": 6.980881374755593e-05, + "loss": 0.0058, + "step": 261 + }, + { + "epoch": 0.036192844315513194, + "grad_norm": 0.15769460797309875, + "learning_rate": 6.985678834670915e-05, + "loss": 0.057, + "step": 262 + }, + { + "epoch": 0.03633098494267164, + "grad_norm": 0.010631869547069073, + "learning_rate": 6.990458018504173e-05, + "loss": 0.0106, + "step": 263 + }, + { + "epoch": 0.036469125569830084, + "grad_norm": 0.020581720396876335, + "learning_rate": 6.995219064973763e-05, + "loss": 0.0151, + "step": 264 + }, + { + "epoch": 0.036607266196988536, + "grad_norm": 0.15721233189105988, + "learning_rate": 6.9999621112247e-05, + "loss": 0.1997, + "step": 265 + }, + { + "epoch": 0.03674540682414698, + "grad_norm": 0.08665527403354645, + "learning_rate": 7.004687292852334e-05, + "loss": 0.0567, + "step": 266 + }, + { + "epoch": 0.036883547451305426, + "grad_norm": 0.01629851944744587, + "learning_rate": 7.009394743925609e-05, + "loss": 0.0129, + "step": 267 + }, + { + "epoch": 0.03702168807846388, + "grad_norm": 0.017079714685678482, + "learning_rate": 7.01408459700988e-05, + "loss": 0.0122, + "step": 268 + }, + { + "epoch": 0.03715982870562232, + "grad_norm": 0.011573737487196922, + "learning_rate": 7.018756983189322e-05, + "loss": 0.0111, + "step": 269 + }, + { + "epoch": 0.03729796933278077, + "grad_norm": 0.057232815772295, + "learning_rate": 7.023412032088907e-05, + "loss": 0.0334, + "step": 270 + }, + { + "epoch": 0.03743610995993922, + "grad_norm": 0.01176871731877327, + "learning_rate": 7.028049871895989e-05, + "loss": 0.0109, + "step": 271 + }, + { + "epoch": 0.037574250587097666, + "grad_norm": 0.005038474686443806, + "learning_rate": 7.032670629381475e-05, + "loss": 0.0057, + "step": 272 + }, + { + "epoch": 0.03771239121425611, + "grad_norm": 0.02611648477613926, + "learning_rate": 7.037274429920628e-05, + "loss": 0.0212, + "step": 273 + }, + { + "epoch": 0.03785053184141456, + "grad_norm": 0.04503704607486725, + "learning_rate": 7.04186139751347e-05, + "loss": 0.0316, + "step": 274 + }, + { + "epoch": 0.03798867246857301, + "grad_norm": 0.02312229387462139, + "learning_rate": 7.046431654804831e-05, + "loss": 0.0195, + "step": 275 + }, + { + "epoch": 0.03812681309573145, + "grad_norm": 0.07315024733543396, + "learning_rate": 7.05098532310401e-05, + "loss": 0.0422, + "step": 276 + }, + { + "epoch": 0.038264953722889905, + "grad_norm": 0.01006357092410326, + "learning_rate": 7.055522522404115e-05, + "loss": 0.0083, + "step": 277 + }, + { + "epoch": 0.03840309435004835, + "grad_norm": 0.02410217933356762, + "learning_rate": 7.060043371401022e-05, + "loss": 0.0164, + "step": 278 + }, + { + "epoch": 0.038541234977206795, + "grad_norm": 0.029703989624977112, + "learning_rate": 7.064547987512007e-05, + "loss": 0.0232, + "step": 279 + }, + { + "epoch": 0.03867937560436524, + "grad_norm": 0.03558236360549927, + "learning_rate": 7.069036486894051e-05, + "loss": 0.0194, + "step": 280 + }, + { + "epoch": 0.03881751623152369, + "grad_norm": 0.004356220830231905, + "learning_rate": 7.073508984461812e-05, + "loss": 0.006, + "step": 281 + }, + { + "epoch": 0.03895565685868214, + "grad_norm": 0.06183575093746185, + "learning_rate": 7.077965593905269e-05, + "loss": 0.0374, + "step": 282 + }, + { + "epoch": 0.03909379748584058, + "grad_norm": 0.03367290645837784, + "learning_rate": 7.082406427707072e-05, + "loss": 0.0147, + "step": 283 + }, + { + "epoch": 0.039231938112999035, + "grad_norm": 0.05731538310647011, + "learning_rate": 7.086831597159574e-05, + "loss": 0.0267, + "step": 284 + }, + { + "epoch": 0.03937007874015748, + "grad_norm": 0.013960711658000946, + "learning_rate": 7.091241212381564e-05, + "loss": 0.0115, + "step": 285 + }, + { + "epoch": 0.039508219367315925, + "grad_norm": 0.04287987947463989, + "learning_rate": 7.095635382334712e-05, + "loss": 0.0186, + "step": 286 + }, + { + "epoch": 0.03964635999447438, + "grad_norm": 0.02721596322953701, + "learning_rate": 7.100014214839724e-05, + "loss": 0.0153, + "step": 287 + }, + { + "epoch": 0.03978450062163282, + "grad_norm": 0.09616994112730026, + "learning_rate": 7.104377816592214e-05, + "loss": 0.0479, + "step": 288 + }, + { + "epoch": 0.03992264124879127, + "grad_norm": 0.01681303046643734, + "learning_rate": 7.108726293178307e-05, + "loss": 0.0099, + "step": 289 + }, + { + "epoch": 0.04006078187594972, + "grad_norm": 0.007514542900025845, + "learning_rate": 7.113059749089967e-05, + "loss": 0.007, + "step": 290 + }, + { + "epoch": 0.040198922503108164, + "grad_norm": 0.00759028410539031, + "learning_rate": 7.117378287740062e-05, + "loss": 0.0093, + "step": 291 + }, + { + "epoch": 0.04033706313026661, + "grad_norm": 0.04445767030119896, + "learning_rate": 7.12168201147717e-05, + "loss": 0.0192, + "step": 292 + }, + { + "epoch": 0.04047520375742506, + "grad_norm": 0.004815852735191584, + "learning_rate": 7.125971021600142e-05, + "loss": 0.0058, + "step": 293 + }, + { + "epoch": 0.04061334438458351, + "grad_norm": 0.01803855411708355, + "learning_rate": 7.130245418372394e-05, + "loss": 0.0189, + "step": 294 + }, + { + "epoch": 0.04075148501174195, + "grad_norm": 0.022579418495297432, + "learning_rate": 7.13450530103599e-05, + "loss": 0.0124, + "step": 295 + }, + { + "epoch": 0.040889625638900404, + "grad_norm": 0.006007243413478136, + "learning_rate": 7.138750767825444e-05, + "loss": 0.0067, + "step": 296 + }, + { + "epoch": 0.04102776626605885, + "grad_norm": 0.01632005348801613, + "learning_rate": 7.142981915981336e-05, + "loss": 0.0115, + "step": 297 + }, + { + "epoch": 0.041165906893217294, + "grad_norm": 0.009706880897283554, + "learning_rate": 7.147198841763656e-05, + "loss": 0.0096, + "step": 298 + }, + { + "epoch": 0.041304047520375746, + "grad_norm": 0.004515258129686117, + "learning_rate": 7.15140164046496e-05, + "loss": 0.0058, + "step": 299 + }, + { + "epoch": 0.04144218814753419, + "grad_norm": 0.05710865557193756, + "learning_rate": 7.155590406423282e-05, + "loss": 0.0388, + "step": 300 + }, + { + "epoch": 0.041580328774692636, + "grad_norm": 0.012932428158819675, + "learning_rate": 7.15976523303484e-05, + "loss": 0.0075, + "step": 301 + }, + { + "epoch": 0.04171846940185108, + "grad_norm": 0.08358635753393173, + "learning_rate": 7.163926212766535e-05, + "loss": 0.0512, + "step": 302 + }, + { + "epoch": 0.04185661002900953, + "grad_norm": 0.00845788512378931, + "learning_rate": 7.168073437168235e-05, + "loss": 0.0085, + "step": 303 + }, + { + "epoch": 0.04199475065616798, + "grad_norm": 0.008571326732635498, + "learning_rate": 7.172206996884871e-05, + "loss": 0.0074, + "step": 304 + }, + { + "epoch": 0.042132891283326424, + "grad_norm": 0.1380782574415207, + "learning_rate": 7.176326981668306e-05, + "loss": 0.0893, + "step": 305 + }, + { + "epoch": 0.042271031910484876, + "grad_norm": 0.10377588123083115, + "learning_rate": 7.180433480389047e-05, + "loss": 0.1076, + "step": 306 + }, + { + "epoch": 0.04240917253764332, + "grad_norm": 0.005449495278298855, + "learning_rate": 7.184526581047739e-05, + "loss": 0.0056, + "step": 307 + }, + { + "epoch": 0.042547313164801766, + "grad_norm": 0.016402093693614006, + "learning_rate": 7.18860637078648e-05, + "loss": 0.0144, + "step": 308 + }, + { + "epoch": 0.04268545379196022, + "grad_norm": 0.057461198419332504, + "learning_rate": 7.192672935899954e-05, + "loss": 0.0335, + "step": 309 + }, + { + "epoch": 0.04282359441911866, + "grad_norm": 0.018814753741025925, + "learning_rate": 7.196726361846382e-05, + "loss": 0.019, + "step": 310 + }, + { + "epoch": 0.04296173504627711, + "grad_norm": 0.021789349615573883, + "learning_rate": 7.200766733258301e-05, + "loss": 0.0174, + "step": 311 + }, + { + "epoch": 0.04309987567343556, + "grad_norm": 0.05062391981482506, + "learning_rate": 7.204794133953164e-05, + "loss": 0.0256, + "step": 312 + }, + { + "epoch": 0.043238016300594005, + "grad_norm": 0.05671432614326477, + "learning_rate": 7.208808646943773e-05, + "loss": 0.0402, + "step": 313 + }, + { + "epoch": 0.04337615692775245, + "grad_norm": 0.02019777148962021, + "learning_rate": 7.212810354448547e-05, + "loss": 0.0152, + "step": 314 + }, + { + "epoch": 0.0435142975549109, + "grad_norm": 0.020292505621910095, + "learning_rate": 7.216799337901625e-05, + "loss": 0.0189, + "step": 315 + }, + { + "epoch": 0.04365243818206935, + "grad_norm": 0.03928740695118904, + "learning_rate": 7.220775677962808e-05, + "loss": 0.0383, + "step": 316 + }, + { + "epoch": 0.04379057880922779, + "grad_norm": 0.005962614435702562, + "learning_rate": 7.224739454527347e-05, + "loss": 0.0069, + "step": 317 + }, + { + "epoch": 0.043928719436386245, + "grad_norm": 0.011562081053853035, + "learning_rate": 7.228690746735578e-05, + "loss": 0.0113, + "step": 318 + }, + { + "epoch": 0.04406686006354469, + "grad_norm": 0.009331891313195229, + "learning_rate": 7.232629632982394e-05, + "loss": 0.0118, + "step": 319 + }, + { + "epoch": 0.044205000690703135, + "grad_norm": 0.07156947255134583, + "learning_rate": 7.236556190926588e-05, + "loss": 0.0474, + "step": 320 + }, + { + "epoch": 0.04434314131786158, + "grad_norm": 0.0070722345262765884, + "learning_rate": 7.240470497500033e-05, + "loss": 0.0071, + "step": 321 + }, + { + "epoch": 0.04448128194502003, + "grad_norm": 0.24819540977478027, + "learning_rate": 7.244372628916727e-05, + "loss": 0.0668, + "step": 322 + }, + { + "epoch": 0.04461942257217848, + "grad_norm": 0.012170674279332161, + "learning_rate": 7.248262660681704e-05, + "loss": 0.0139, + "step": 323 + }, + { + "epoch": 0.04475756319933692, + "grad_norm": 0.01726563833653927, + "learning_rate": 7.252140667599786e-05, + "loss": 0.0148, + "step": 324 + }, + { + "epoch": 0.044895703826495374, + "grad_norm": 0.006273672450333834, + "learning_rate": 7.256006723784231e-05, + "loss": 0.0073, + "step": 325 + }, + { + "epoch": 0.04503384445365382, + "grad_norm": 0.02530139684677124, + "learning_rate": 7.25986090266522e-05, + "loss": 0.0145, + "step": 326 + }, + { + "epoch": 0.045171985080812264, + "grad_norm": 0.043121714144945145, + "learning_rate": 7.263703276998225e-05, + "loss": 0.0374, + "step": 327 + }, + { + "epoch": 0.045310125707970716, + "grad_norm": 0.011094048619270325, + "learning_rate": 7.26753391887226e-05, + "loss": 0.0098, + "step": 328 + }, + { + "epoch": 0.04544826633512916, + "grad_norm": 0.05844907462596893, + "learning_rate": 7.271352899717988e-05, + "loss": 0.0314, + "step": 329 + }, + { + "epoch": 0.04558640696228761, + "grad_norm": 0.008745652623474598, + "learning_rate": 7.27516029031571e-05, + "loss": 0.0107, + "step": 330 + }, + { + "epoch": 0.04572454758944606, + "grad_norm": 0.12875714898109436, + "learning_rate": 7.278956160803247e-05, + "loss": 0.0946, + "step": 331 + }, + { + "epoch": 0.045862688216604504, + "grad_norm": 0.09134670346975327, + "learning_rate": 7.282740580683686e-05, + "loss": 0.0767, + "step": 332 + }, + { + "epoch": 0.04600082884376295, + "grad_norm": 0.01726178452372551, + "learning_rate": 7.286513618833017e-05, + "loss": 0.0131, + "step": 333 + }, + { + "epoch": 0.0461389694709214, + "grad_norm": 0.006260738708078861, + "learning_rate": 7.290275343507648e-05, + "loss": 0.0069, + "step": 334 + }, + { + "epoch": 0.046277110098079846, + "grad_norm": 0.02264183759689331, + "learning_rate": 7.294025822351828e-05, + "loss": 0.0157, + "step": 335 + }, + { + "epoch": 0.04641525072523829, + "grad_norm": 0.033730894327163696, + "learning_rate": 7.297765122404931e-05, + "loss": 0.0338, + "step": 336 + }, + { + "epoch": 0.04655339135239674, + "grad_norm": 0.010735332034528255, + "learning_rate": 7.301493310108658e-05, + "loss": 0.0117, + "step": 337 + }, + { + "epoch": 0.04669153197955519, + "grad_norm": 0.13428980112075806, + "learning_rate": 7.305210451314113e-05, + "loss": 0.0662, + "step": 338 + }, + { + "epoch": 0.04682967260671363, + "grad_norm": 0.007416439242660999, + "learning_rate": 7.308916611288784e-05, + "loss": 0.0093, + "step": 339 + }, + { + "epoch": 0.04696781323387208, + "grad_norm": 0.029140817001461983, + "learning_rate": 7.312611854723422e-05, + "loss": 0.0268, + "step": 340 + }, + { + "epoch": 0.04710595386103053, + "grad_norm": 0.002688678679987788, + "learning_rate": 7.316296245738808e-05, + "loss": 0.0041, + "step": 341 + }, + { + "epoch": 0.047244094488188976, + "grad_norm": 0.09536877274513245, + "learning_rate": 7.319969847892443e-05, + "loss": 0.0513, + "step": 342 + }, + { + "epoch": 0.04738223511534742, + "grad_norm": 0.0194696132093668, + "learning_rate": 7.323632724185112e-05, + "loss": 0.0156, + "step": 343 + }, + { + "epoch": 0.04752037574250587, + "grad_norm": 0.012019694782793522, + "learning_rate": 7.327284937067376e-05, + "loss": 0.0128, + "step": 344 + }, + { + "epoch": 0.04765851636966432, + "grad_norm": 0.07920549809932709, + "learning_rate": 7.330926548445958e-05, + "loss": 0.0496, + "step": 345 + }, + { + "epoch": 0.04779665699682276, + "grad_norm": 0.012154892086982727, + "learning_rate": 7.334557619690038e-05, + "loss": 0.0163, + "step": 346 + }, + { + "epoch": 0.047934797623981215, + "grad_norm": 0.02246297337114811, + "learning_rate": 7.338178211637459e-05, + "loss": 0.016, + "step": 347 + }, + { + "epoch": 0.04807293825113966, + "grad_norm": 0.1408204585313797, + "learning_rate": 7.341788384600846e-05, + "loss": 0.1492, + "step": 348 + }, + { + "epoch": 0.048211078878298105, + "grad_norm": 0.03998423367738724, + "learning_rate": 7.345388198373633e-05, + "loss": 0.0386, + "step": 349 + }, + { + "epoch": 0.04834921950545656, + "grad_norm": 0.0037037734873592854, + "learning_rate": 7.348977712236e-05, + "loss": 0.0052, + "step": 350 + }, + { + "epoch": 0.048487360132615, + "grad_norm": 0.01490323431789875, + "learning_rate": 7.352556984960736e-05, + "loss": 0.0154, + "step": 351 + }, + { + "epoch": 0.04862550075977345, + "grad_norm": 0.022191043943166733, + "learning_rate": 7.356126074819015e-05, + "loss": 0.0173, + "step": 352 + }, + { + "epoch": 0.0487636413869319, + "grad_norm": 0.016748666763305664, + "learning_rate": 7.359685039586083e-05, + "loss": 0.0147, + "step": 353 + }, + { + "epoch": 0.048901782014090345, + "grad_norm": 0.018724625930190086, + "learning_rate": 7.363233936546869e-05, + "loss": 0.0196, + "step": 354 + }, + { + "epoch": 0.04903992264124879, + "grad_norm": 0.04513927921652794, + "learning_rate": 7.366772822501523e-05, + "loss": 0.024, + "step": 355 + }, + { + "epoch": 0.04917806326840724, + "grad_norm": 0.017063314095139503, + "learning_rate": 7.370301753770863e-05, + "loss": 0.0164, + "step": 356 + }, + { + "epoch": 0.04931620389556569, + "grad_norm": 0.01662248931825161, + "learning_rate": 7.373820786201764e-05, + "loss": 0.0148, + "step": 357 + }, + { + "epoch": 0.04945434452272413, + "grad_norm": 0.023914910852909088, + "learning_rate": 7.377329975172453e-05, + "loss": 0.0205, + "step": 358 + }, + { + "epoch": 0.04959248514988258, + "grad_norm": 0.01144400890916586, + "learning_rate": 7.380829375597736e-05, + "loss": 0.0135, + "step": 359 + }, + { + "epoch": 0.04973062577704103, + "grad_norm": 0.04367682710289955, + "learning_rate": 7.384319041934161e-05, + "loss": 0.0376, + "step": 360 + }, + { + "epoch": 0.049868766404199474, + "grad_norm": 0.05894111096858978, + "learning_rate": 7.3877990281851e-05, + "loss": 0.0239, + "step": 361 + }, + { + "epoch": 0.05000690703135792, + "grad_norm": 0.008600580506026745, + "learning_rate": 7.391269387905761e-05, + "loss": 0.01, + "step": 362 + }, + { + "epoch": 0.05014504765851637, + "grad_norm": 0.012402276508510113, + "learning_rate": 7.394730174208137e-05, + "loss": 0.0165, + "step": 363 + }, + { + "epoch": 0.05028318828567482, + "grad_norm": 0.017745958641171455, + "learning_rate": 7.398181439765882e-05, + "loss": 0.0218, + "step": 364 + }, + { + "epoch": 0.05042132891283326, + "grad_norm": 0.05549019202589989, + "learning_rate": 7.401623236819118e-05, + "loss": 0.0429, + "step": 365 + }, + { + "epoch": 0.050559469539991714, + "grad_norm": 0.0038389600813388824, + "learning_rate": 7.405055617179185e-05, + "loss": 0.0056, + "step": 366 + }, + { + "epoch": 0.05069761016715016, + "grad_norm": 0.14552558958530426, + "learning_rate": 7.408478632233319e-05, + "loss": 0.0786, + "step": 367 + }, + { + "epoch": 0.050835750794308604, + "grad_norm": 0.011212456971406937, + "learning_rate": 7.411892332949265e-05, + "loss": 0.008, + "step": 368 + }, + { + "epoch": 0.050973891421467056, + "grad_norm": 0.09112522006034851, + "learning_rate": 7.415296769879832e-05, + "loss": 0.0501, + "step": 369 + }, + { + "epoch": 0.0511120320486255, + "grad_norm": 0.0417109876871109, + "learning_rate": 7.418691993167391e-05, + "loss": 0.0289, + "step": 370 + }, + { + "epoch": 0.051250172675783946, + "grad_norm": 0.03118273988366127, + "learning_rate": 7.422078052548295e-05, + "loss": 0.023, + "step": 371 + }, + { + "epoch": 0.0513883133029424, + "grad_norm": 0.014401586726307869, + "learning_rate": 7.42545499735726e-05, + "loss": 0.0159, + "step": 372 + }, + { + "epoch": 0.05152645393010084, + "grad_norm": 0.0387452095746994, + "learning_rate": 7.428822876531674e-05, + "loss": 0.0359, + "step": 373 + }, + { + "epoch": 0.05166459455725929, + "grad_norm": 0.0016639974201098084, + "learning_rate": 7.43218173861585e-05, + "loss": 0.0028, + "step": 374 + }, + { + "epoch": 0.05180273518441774, + "grad_norm": 0.007370566017925739, + "learning_rate": 7.435531631765229e-05, + "loss": 0.0091, + "step": 375 + }, + { + "epoch": 0.051940875811576186, + "grad_norm": 0.008531290106475353, + "learning_rate": 7.438872603750523e-05, + "loss": 0.0091, + "step": 376 + }, + { + "epoch": 0.05207901643873463, + "grad_norm": 0.02294515073299408, + "learning_rate": 7.442204701961796e-05, + "loss": 0.0254, + "step": 377 + }, + { + "epoch": 0.052217157065893076, + "grad_norm": 0.02039843425154686, + "learning_rate": 7.445527973412506e-05, + "loss": 0.0208, + "step": 378 + }, + { + "epoch": 0.05235529769305153, + "grad_norm": 0.050735872238874435, + "learning_rate": 7.448842464743471e-05, + "loss": 0.0319, + "step": 379 + }, + { + "epoch": 0.05249343832020997, + "grad_norm": 0.016932478174567223, + "learning_rate": 7.452148222226818e-05, + "loss": 0.0109, + "step": 380 + }, + { + "epoch": 0.05263157894736842, + "grad_norm": 0.01010705903172493, + "learning_rate": 7.455445291769839e-05, + "loss": 0.0106, + "step": 381 + }, + { + "epoch": 0.05276971957452687, + "grad_norm": 0.2523377239704132, + "learning_rate": 7.458733718918828e-05, + "loss": 0.1156, + "step": 382 + }, + { + "epoch": 0.052907860201685315, + "grad_norm": 0.006449633743613958, + "learning_rate": 7.462013548862848e-05, + "loss": 0.0086, + "step": 383 + }, + { + "epoch": 0.05304600082884376, + "grad_norm": 0.0557723194360733, + "learning_rate": 7.465284826437468e-05, + "loss": 0.0106, + "step": 384 + }, + { + "epoch": 0.05318414145600221, + "grad_norm": 0.0064123859629035, + "learning_rate": 7.468547596128427e-05, + "loss": 0.0085, + "step": 385 + }, + { + "epoch": 0.05332228208316066, + "grad_norm": 0.009457213804125786, + "learning_rate": 7.471801902075274e-05, + "loss": 0.0117, + "step": 386 + }, + { + "epoch": 0.0534604227103191, + "grad_norm": 0.0031539765186607838, + "learning_rate": 7.475047788074949e-05, + "loss": 0.0047, + "step": 387 + }, + { + "epoch": 0.053598563337477555, + "grad_norm": 0.023614773526787758, + "learning_rate": 7.478285297585316e-05, + "loss": 0.0219, + "step": 388 + }, + { + "epoch": 0.053736703964636, + "grad_norm": 0.10714581608772278, + "learning_rate": 7.481514473728654e-05, + "loss": 0.0712, + "step": 389 + }, + { + "epoch": 0.053874844591794445, + "grad_norm": 0.014973816461861134, + "learning_rate": 7.484735359295111e-05, + "loss": 0.0115, + "step": 390 + }, + { + "epoch": 0.0540129852189529, + "grad_norm": 0.035133246332407, + "learning_rate": 7.487947996746097e-05, + "loss": 0.0189, + "step": 391 + }, + { + "epoch": 0.05415112584611134, + "grad_norm": 0.01953275129199028, + "learning_rate": 7.491152428217649e-05, + "loss": 0.0255, + "step": 392 + }, + { + "epoch": 0.05428926647326979, + "grad_norm": 0.012475092895328999, + "learning_rate": 7.494348695523741e-05, + "loss": 0.0117, + "step": 393 + }, + { + "epoch": 0.05442740710042824, + "grad_norm": 0.018982090055942535, + "learning_rate": 7.497536840159565e-05, + "loss": 0.0186, + "step": 394 + }, + { + "epoch": 0.054565547727586684, + "grad_norm": 0.0650404840707779, + "learning_rate": 7.500716903304755e-05, + "loss": 0.0505, + "step": 395 + }, + { + "epoch": 0.05470368835474513, + "grad_norm": 0.02308955229818821, + "learning_rate": 7.503888925826588e-05, + "loss": 0.0208, + "step": 396 + }, + { + "epoch": 0.054841828981903575, + "grad_norm": 0.006316799204796553, + "learning_rate": 7.50705294828313e-05, + "loss": 0.009, + "step": 397 + }, + { + "epoch": 0.05497996960906203, + "grad_norm": 0.03284195438027382, + "learning_rate": 7.510209010926341e-05, + "loss": 0.0286, + "step": 398 + }, + { + "epoch": 0.05511811023622047, + "grad_norm": 0.03874586522579193, + "learning_rate": 7.51335715370516e-05, + "loss": 0.0282, + "step": 399 + }, + { + "epoch": 0.05525625086337892, + "grad_norm": 0.017570259049534798, + "learning_rate": 7.516497416268535e-05, + "loss": 0.0159, + "step": 400 + }, + { + "epoch": 0.05539439149053737, + "grad_norm": 0.1028120145201683, + "learning_rate": 7.519629837968414e-05, + "loss": 0.0638, + "step": 401 + }, + { + "epoch": 0.055532532117695814, + "grad_norm": 0.04638221859931946, + "learning_rate": 7.522754457862707e-05, + "loss": 0.0261, + "step": 402 + }, + { + "epoch": 0.05567067274485426, + "grad_norm": 0.006671587936580181, + "learning_rate": 7.52587131471821e-05, + "loss": 0.0098, + "step": 403 + }, + { + "epoch": 0.05580881337201271, + "grad_norm": 0.0016172940377146006, + "learning_rate": 7.528980447013491e-05, + "loss": 0.0032, + "step": 404 + }, + { + "epoch": 0.055946953999171156, + "grad_norm": 0.009872229769825935, + "learning_rate": 7.532081892941734e-05, + "loss": 0.0102, + "step": 405 + }, + { + "epoch": 0.0560850946263296, + "grad_norm": 0.014002018608152866, + "learning_rate": 7.535175690413565e-05, + "loss": 0.0126, + "step": 406 + }, + { + "epoch": 0.05622323525348805, + "grad_norm": 0.01186671108007431, + "learning_rate": 7.538261877059817e-05, + "loss": 0.0151, + "step": 407 + }, + { + "epoch": 0.0563613758806465, + "grad_norm": 0.010572116822004318, + "learning_rate": 7.541340490234301e-05, + "loss": 0.0128, + "step": 408 + }, + { + "epoch": 0.056499516507804944, + "grad_norm": 0.006026785355061293, + "learning_rate": 7.544411567016487e-05, + "loss": 0.009, + "step": 409 + }, + { + "epoch": 0.056637657134963396, + "grad_norm": 0.0200533214956522, + "learning_rate": 7.547475144214207e-05, + "loss": 0.0171, + "step": 410 + }, + { + "epoch": 0.05677579776212184, + "grad_norm": 0.016931114718317986, + "learning_rate": 7.550531258366296e-05, + "loss": 0.015, + "step": 411 + }, + { + "epoch": 0.056913938389280286, + "grad_norm": 0.011858226731419563, + "learning_rate": 7.553579945745208e-05, + "loss": 0.0113, + "step": 412 + }, + { + "epoch": 0.05705207901643874, + "grad_norm": 0.02081177569925785, + "learning_rate": 7.556621242359586e-05, + "loss": 0.0161, + "step": 413 + }, + { + "epoch": 0.05719021964359718, + "grad_norm": 0.008650501258671284, + "learning_rate": 7.559655183956836e-05, + "loss": 0.0096, + "step": 414 + }, + { + "epoch": 0.05732836027075563, + "grad_norm": 0.008785598911345005, + "learning_rate": 7.562681806025635e-05, + "loss": 0.0096, + "step": 415 + }, + { + "epoch": 0.05746650089791407, + "grad_norm": 0.004337130580097437, + "learning_rate": 7.565701143798417e-05, + "loss": 0.006, + "step": 416 + }, + { + "epoch": 0.057604641525072525, + "grad_norm": 0.01971902698278427, + "learning_rate": 7.568713232253847e-05, + "loss": 0.0144, + "step": 417 + }, + { + "epoch": 0.05774278215223097, + "grad_norm": 0.003371615894138813, + "learning_rate": 7.571718106119245e-05, + "loss": 0.004, + "step": 418 + }, + { + "epoch": 0.057880922779389415, + "grad_norm": 0.08240024745464325, + "learning_rate": 7.574715799872985e-05, + "loss": 0.0443, + "step": 419 + }, + { + "epoch": 0.05801906340654787, + "grad_norm": 0.013157228007912636, + "learning_rate": 7.577706347746878e-05, + "loss": 0.01, + "step": 420 + }, + { + "epoch": 0.05815720403370631, + "grad_norm": 0.08836307376623154, + "learning_rate": 7.58068978372851e-05, + "loss": 0.053, + "step": 421 + }, + { + "epoch": 0.05829534466086476, + "grad_norm": 0.008075368590652943, + "learning_rate": 7.58366614156356e-05, + "loss": 0.0103, + "step": 422 + }, + { + "epoch": 0.05843348528802321, + "grad_norm": 0.013115230947732925, + "learning_rate": 7.586635454758096e-05, + "loss": 0.0145, + "step": 423 + }, + { + "epoch": 0.058571625915181655, + "grad_norm": 0.041858140379190445, + "learning_rate": 7.589597756580832e-05, + "loss": 0.0266, + "step": 424 + }, + { + "epoch": 0.0587097665423401, + "grad_norm": 0.006484493613243103, + "learning_rate": 7.592553080065369e-05, + "loss": 0.007, + "step": 425 + }, + { + "epoch": 0.05884790716949855, + "grad_norm": 0.08854290097951889, + "learning_rate": 7.5955014580124e-05, + "loss": 0.0769, + "step": 426 + }, + { + "epoch": 0.058986047796657, + "grad_norm": 0.03287418931722641, + "learning_rate": 7.598442922991903e-05, + "loss": 0.0359, + "step": 427 + }, + { + "epoch": 0.05912418842381544, + "grad_norm": 0.0041122897528111935, + "learning_rate": 7.601377507345287e-05, + "loss": 0.0056, + "step": 428 + }, + { + "epoch": 0.059262329050973894, + "grad_norm": 0.01932491548359394, + "learning_rate": 7.604305243187539e-05, + "loss": 0.0185, + "step": 429 + }, + { + "epoch": 0.05940046967813234, + "grad_norm": 0.034386295825242996, + "learning_rate": 7.607226162409324e-05, + "loss": 0.0259, + "step": 430 + }, + { + "epoch": 0.059538610305290784, + "grad_norm": 0.014645845629274845, + "learning_rate": 7.610140296679071e-05, + "loss": 0.0159, + "step": 431 + }, + { + "epoch": 0.059676750932449237, + "grad_norm": 0.025981949642300606, + "learning_rate": 7.613047677445041e-05, + "loss": 0.0199, + "step": 432 + }, + { + "epoch": 0.05981489155960768, + "grad_norm": 0.00898673851042986, + "learning_rate": 7.615948335937346e-05, + "loss": 0.0087, + "step": 433 + }, + { + "epoch": 0.05995303218676613, + "grad_norm": 0.005743580870330334, + "learning_rate": 7.618842303169978e-05, + "loss": 0.0069, + "step": 434 + }, + { + "epoch": 0.06009117281392457, + "grad_norm": 0.00816387590020895, + "learning_rate": 7.621729609942794e-05, + "loss": 0.0101, + "step": 435 + }, + { + "epoch": 0.060229313441083024, + "grad_norm": 0.021145803853869438, + "learning_rate": 7.624610286843479e-05, + "loss": 0.0145, + "step": 436 + }, + { + "epoch": 0.06036745406824147, + "grad_norm": 0.01337206270545721, + "learning_rate": 7.627484364249493e-05, + "loss": 0.0123, + "step": 437 + }, + { + "epoch": 0.060505594695399914, + "grad_norm": 0.03297748416662216, + "learning_rate": 7.630351872329997e-05, + "loss": 0.0201, + "step": 438 + }, + { + "epoch": 0.060643735322558366, + "grad_norm": 0.05304405465722084, + "learning_rate": 7.633212841047743e-05, + "loss": 0.026, + "step": 439 + }, + { + "epoch": 0.06078187594971681, + "grad_norm": 0.009639502502977848, + "learning_rate": 7.636067300160962e-05, + "loss": 0.0098, + "step": 440 + }, + { + "epoch": 0.060920016576875256, + "grad_norm": 0.08749550580978394, + "learning_rate": 7.638915279225222e-05, + "loss": 0.0489, + "step": 441 + }, + { + "epoch": 0.06105815720403371, + "grad_norm": 0.0455004945397377, + "learning_rate": 7.64175680759525e-05, + "loss": 0.0276, + "step": 442 + }, + { + "epoch": 0.061196297831192153, + "grad_norm": 0.07244568318128586, + "learning_rate": 7.644591914426769e-05, + "loss": 0.0554, + "step": 443 + }, + { + "epoch": 0.0613344384583506, + "grad_norm": 0.007952029816806316, + "learning_rate": 7.64742062867827e-05, + "loss": 0.0101, + "step": 444 + }, + { + "epoch": 0.06147257908550905, + "grad_norm": 0.045144565403461456, + "learning_rate": 7.65024297911281e-05, + "loss": 0.0249, + "step": 445 + }, + { + "epoch": 0.061610719712667496, + "grad_norm": 0.00880199670791626, + "learning_rate": 7.653058994299745e-05, + "loss": 0.0077, + "step": 446 + }, + { + "epoch": 0.06174886033982594, + "grad_norm": 0.015727581456303596, + "learning_rate": 7.655868702616483e-05, + "loss": 0.0163, + "step": 447 + }, + { + "epoch": 0.06188700096698439, + "grad_norm": 0.002647354966029525, + "learning_rate": 7.658672132250185e-05, + "loss": 0.0046, + "step": 448 + }, + { + "epoch": 0.06202514159414284, + "grad_norm": 0.0048969099298119545, + "learning_rate": 7.661469311199475e-05, + "loss": 0.0063, + "step": 449 + }, + { + "epoch": 0.06216328222130128, + "grad_norm": 0.010342615656554699, + "learning_rate": 7.664260267276109e-05, + "loss": 0.011, + "step": 450 + }, + { + "epoch": 0.062301422848459735, + "grad_norm": 0.004208502359688282, + "learning_rate": 7.667045028106635e-05, + "loss": 0.0059, + "step": 451 + }, + { + "epoch": 0.06243956347561818, + "grad_norm": 0.012316681444644928, + "learning_rate": 7.66982362113404e-05, + "loss": 0.0106, + "step": 452 + }, + { + "epoch": 0.06257770410277663, + "grad_norm": 0.006314276251941919, + "learning_rate": 7.672596073619361e-05, + "loss": 0.0086, + "step": 453 + }, + { + "epoch": 0.06271584472993508, + "grad_norm": 0.004086961969733238, + "learning_rate": 7.675362412643307e-05, + "loss": 0.0052, + "step": 454 + }, + { + "epoch": 0.06285398535709352, + "grad_norm": 0.048082876950502396, + "learning_rate": 7.678122665107829e-05, + "loss": 0.0333, + "step": 455 + }, + { + "epoch": 0.06299212598425197, + "grad_norm": 0.01329890824854374, + "learning_rate": 7.680876857737698e-05, + "loss": 0.0115, + "step": 456 + }, + { + "epoch": 0.06313026661141041, + "grad_norm": 0.02435714565217495, + "learning_rate": 7.683625017082056e-05, + "loss": 0.0225, + "step": 457 + }, + { + "epoch": 0.06326840723856886, + "grad_norm": 0.09511169046163559, + "learning_rate": 7.686367169515956e-05, + "loss": 0.084, + "step": 458 + }, + { + "epoch": 0.06340654786572732, + "grad_norm": 0.01172893587499857, + "learning_rate": 7.689103341241874e-05, + "loss": 0.0115, + "step": 459 + }, + { + "epoch": 0.06354468849288576, + "grad_norm": 0.022892113775014877, + "learning_rate": 7.691833558291211e-05, + "loss": 0.0145, + "step": 460 + }, + { + "epoch": 0.06368282912004421, + "grad_norm": 0.005384576041251421, + "learning_rate": 7.694557846525789e-05, + "loss": 0.0055, + "step": 461 + }, + { + "epoch": 0.06382096974720265, + "grad_norm": 0.007982897572219372, + "learning_rate": 7.697276231639306e-05, + "loss": 0.009, + "step": 462 + }, + { + "epoch": 0.0639591103743611, + "grad_norm": 0.014571700245141983, + "learning_rate": 7.699988739158804e-05, + "loss": 0.0134, + "step": 463 + }, + { + "epoch": 0.06409725100151954, + "grad_norm": 0.048437412828207016, + "learning_rate": 7.7026953944461e-05, + "loss": 0.0285, + "step": 464 + }, + { + "epoch": 0.06423539162867799, + "grad_norm": 0.1020222008228302, + "learning_rate": 7.705396222699208e-05, + "loss": 0.0604, + "step": 465 + }, + { + "epoch": 0.06437353225583645, + "grad_norm": 0.002950117690488696, + "learning_rate": 7.708091248953748e-05, + "loss": 0.0039, + "step": 466 + }, + { + "epoch": 0.06451167288299489, + "grad_norm": 0.011465366929769516, + "learning_rate": 7.710780498084345e-05, + "loss": 0.0105, + "step": 467 + }, + { + "epoch": 0.06464981351015334, + "grad_norm": 0.009519292041659355, + "learning_rate": 7.71346399480599e-05, + "loss": 0.0091, + "step": 468 + }, + { + "epoch": 0.06478795413731178, + "grad_norm": 0.009986692108213902, + "learning_rate": 7.716141763675424e-05, + "loss": 0.0111, + "step": 469 + }, + { + "epoch": 0.06492609476447023, + "grad_norm": 0.01423732005059719, + "learning_rate": 7.718813829092471e-05, + "loss": 0.0126, + "step": 470 + }, + { + "epoch": 0.06506423539162867, + "grad_norm": 0.008685347624123096, + "learning_rate": 7.721480215301373e-05, + "loss": 0.0114, + "step": 471 + }, + { + "epoch": 0.06520237601878713, + "grad_norm": 0.003593148896470666, + "learning_rate": 7.724140946392123e-05, + "loss": 0.0048, + "step": 472 + }, + { + "epoch": 0.06534051664594558, + "grad_norm": 0.019208243116736412, + "learning_rate": 7.726796046301751e-05, + "loss": 0.0131, + "step": 473 + }, + { + "epoch": 0.06547865727310402, + "grad_norm": 0.0028865146450698376, + "learning_rate": 7.729445538815635e-05, + "loss": 0.004, + "step": 474 + }, + { + "epoch": 0.06561679790026247, + "grad_norm": 0.006914031691849232, + "learning_rate": 7.732089447568765e-05, + "loss": 0.0073, + "step": 475 + }, + { + "epoch": 0.06575493852742091, + "grad_norm": 0.005969736259430647, + "learning_rate": 7.734727796047018e-05, + "loss": 0.0061, + "step": 476 + }, + { + "epoch": 0.06589307915457936, + "grad_norm": 0.009544942528009415, + "learning_rate": 7.737360607588405e-05, + "loss": 0.0109, + "step": 477 + }, + { + "epoch": 0.06603121978173782, + "grad_norm": 0.017079075798392296, + "learning_rate": 7.73998790538431e-05, + "loss": 0.0119, + "step": 478 + }, + { + "epoch": 0.06616936040889626, + "grad_norm": 0.006762394681572914, + "learning_rate": 7.742609712480713e-05, + "loss": 0.006, + "step": 479 + }, + { + "epoch": 0.0663075010360547, + "grad_norm": 0.006828032899647951, + "learning_rate": 7.745226051779415e-05, + "loss": 0.0069, + "step": 480 + }, + { + "epoch": 0.06644564166321315, + "grad_norm": 0.008298908360302448, + "learning_rate": 7.74783694603922e-05, + "loss": 0.0083, + "step": 481 + }, + { + "epoch": 0.0665837822903716, + "grad_norm": 0.15002965927124023, + "learning_rate": 7.750442417877138e-05, + "loss": 0.0388, + "step": 482 + }, + { + "epoch": 0.06672192291753004, + "grad_norm": 0.00954232458025217, + "learning_rate": 7.753042489769554e-05, + "loss": 0.0088, + "step": 483 + }, + { + "epoch": 0.06686006354468849, + "grad_norm": 0.10171414166688919, + "learning_rate": 7.755637184053391e-05, + "loss": 0.058, + "step": 484 + }, + { + "epoch": 0.06699820417184695, + "grad_norm": 0.0051827989518642426, + "learning_rate": 7.758226522927262e-05, + "loss": 0.0043, + "step": 485 + }, + { + "epoch": 0.06713634479900539, + "grad_norm": 0.00380541174672544, + "learning_rate": 7.760810528452614e-05, + "loss": 0.0052, + "step": 486 + }, + { + "epoch": 0.06727448542616384, + "grad_norm": 0.0701836422085762, + "learning_rate": 7.763389222554847e-05, + "loss": 0.0342, + "step": 487 + }, + { + "epoch": 0.06741262605332228, + "grad_norm": 0.0025863810442388058, + "learning_rate": 7.765962627024439e-05, + "loss": 0.0041, + "step": 488 + }, + { + "epoch": 0.06755076668048073, + "grad_norm": 0.004156921990215778, + "learning_rate": 7.768530763518046e-05, + "loss": 0.0057, + "step": 489 + }, + { + "epoch": 0.06768890730763917, + "grad_norm": 0.020250679925084114, + "learning_rate": 7.771093653559595e-05, + "loss": 0.0098, + "step": 490 + }, + { + "epoch": 0.06782704793479763, + "grad_norm": 0.008420642465353012, + "learning_rate": 7.773651318541372e-05, + "loss": 0.0073, + "step": 491 + }, + { + "epoch": 0.06796518856195607, + "grad_norm": 0.021448403596878052, + "learning_rate": 7.776203779725086e-05, + "loss": 0.0201, + "step": 492 + }, + { + "epoch": 0.06810332918911452, + "grad_norm": 0.00401209807023406, + "learning_rate": 7.778751058242933e-05, + "loss": 0.0044, + "step": 493 + }, + { + "epoch": 0.06824146981627296, + "grad_norm": 0.004071755334734917, + "learning_rate": 7.781293175098647e-05, + "loss": 0.0048, + "step": 494 + }, + { + "epoch": 0.06837961044343141, + "grad_norm": 0.07812987267971039, + "learning_rate": 7.783830151168537e-05, + "loss": 0.0596, + "step": 495 + }, + { + "epoch": 0.06851775107058986, + "grad_norm": 0.006245663855224848, + "learning_rate": 7.786362007202515e-05, + "loss": 0.0084, + "step": 496 + }, + { + "epoch": 0.06865589169774831, + "grad_norm": 0.029952548444271088, + "learning_rate": 7.788888763825119e-05, + "loss": 0.0229, + "step": 497 + }, + { + "epoch": 0.06879403232490676, + "grad_norm": 0.00528394291177392, + "learning_rate": 7.791410441536515e-05, + "loss": 0.0072, + "step": 498 + }, + { + "epoch": 0.0689321729520652, + "grad_norm": 0.004051877185702324, + "learning_rate": 7.793927060713498e-05, + "loss": 0.0049, + "step": 499 + }, + { + "epoch": 0.06907031357922365, + "grad_norm": 0.036640387028455734, + "learning_rate": 7.796438641610483e-05, + "loss": 0.0186, + "step": 500 + }, + { + "epoch": 0.0692084542063821, + "grad_norm": 0.004171683453023434, + "learning_rate": 7.798945204360475e-05, + "loss": 0.0067, + "step": 501 + }, + { + "epoch": 0.06934659483354054, + "grad_norm": 0.012148253619670868, + "learning_rate": 7.801446768976046e-05, + "loss": 0.0156, + "step": 502 + }, + { + "epoch": 0.06948473546069898, + "grad_norm": 0.0024077165871858597, + "learning_rate": 7.803943355350285e-05, + "loss": 0.0042, + "step": 503 + }, + { + "epoch": 0.06962287608785744, + "grad_norm": 0.018709257245063782, + "learning_rate": 7.806434983257758e-05, + "loss": 0.0147, + "step": 504 + }, + { + "epoch": 0.06976101671501589, + "grad_norm": 0.004529232159256935, + "learning_rate": 7.808921672355438e-05, + "loss": 0.0051, + "step": 505 + }, + { + "epoch": 0.06989915734217433, + "grad_norm": 0.08686839044094086, + "learning_rate": 7.811403442183638e-05, + "loss": 0.0485, + "step": 506 + }, + { + "epoch": 0.07003729796933278, + "grad_norm": 0.009920633397996426, + "learning_rate": 7.81388031216694e-05, + "loss": 0.0093, + "step": 507 + }, + { + "epoch": 0.07017543859649122, + "grad_norm": 0.11127305775880814, + "learning_rate": 7.816352301615093e-05, + "loss": 0.0544, + "step": 508 + }, + { + "epoch": 0.07031357922364967, + "grad_norm": 0.010723556391894817, + "learning_rate": 7.818819429723929e-05, + "loss": 0.0121, + "step": 509 + }, + { + "epoch": 0.07045171985080813, + "grad_norm": 0.009187200106680393, + "learning_rate": 7.821281715576248e-05, + "loss": 0.0091, + "step": 510 + }, + { + "epoch": 0.07058986047796657, + "grad_norm": 0.012122333981096745, + "learning_rate": 7.823739178142714e-05, + "loss": 0.0117, + "step": 511 + }, + { + "epoch": 0.07072800110512502, + "grad_norm": 0.008281617425382137, + "learning_rate": 7.826191836282722e-05, + "loss": 0.0073, + "step": 512 + }, + { + "epoch": 0.07086614173228346, + "grad_norm": 0.018169116228818893, + "learning_rate": 7.82863970874527e-05, + "loss": 0.0199, + "step": 513 + }, + { + "epoch": 0.07100428235944191, + "grad_norm": 0.02657570131123066, + "learning_rate": 7.831082814169822e-05, + "loss": 0.0206, + "step": 514 + }, + { + "epoch": 0.07114242298660035, + "grad_norm": 0.006763003766536713, + "learning_rate": 7.833521171087153e-05, + "loss": 0.0081, + "step": 515 + }, + { + "epoch": 0.07128056361375881, + "grad_norm": 0.14291508495807648, + "learning_rate": 7.835954797920203e-05, + "loss": 0.1127, + "step": 516 + }, + { + "epoch": 0.07141870424091726, + "grad_norm": 0.015182922594249249, + "learning_rate": 7.838383712984899e-05, + "loss": 0.0183, + "step": 517 + }, + { + "epoch": 0.0715568448680757, + "grad_norm": 0.006437589880079031, + "learning_rate": 7.840807934490987e-05, + "loss": 0.007, + "step": 518 + }, + { + "epoch": 0.07169498549523415, + "grad_norm": 0.006644203793257475, + "learning_rate": 7.843227480542863e-05, + "loss": 0.007, + "step": 519 + }, + { + "epoch": 0.0718331261223926, + "grad_norm": 0.0024770349264144897, + "learning_rate": 7.845642369140365e-05, + "loss": 0.0037, + "step": 520 + }, + { + "epoch": 0.07197126674955104, + "grad_norm": 0.007362083997577429, + "learning_rate": 7.848052618179587e-05, + "loss": 0.0083, + "step": 521 + }, + { + "epoch": 0.07210940737670948, + "grad_norm": 0.019205942749977112, + "learning_rate": 7.850458245453673e-05, + "loss": 0.0222, + "step": 522 + }, + { + "epoch": 0.07224754800386794, + "grad_norm": 0.009427196346223354, + "learning_rate": 7.852859268653608e-05, + "loss": 0.0113, + "step": 523 + }, + { + "epoch": 0.07238568863102639, + "grad_norm": 0.006181191653013229, + "learning_rate": 7.855255705368995e-05, + "loss": 0.006, + "step": 524 + }, + { + "epoch": 0.07252382925818483, + "grad_norm": 0.0038815487641841173, + "learning_rate": 7.857647573088825e-05, + "loss": 0.006, + "step": 525 + }, + { + "epoch": 0.07266196988534328, + "grad_norm": 0.002488507889211178, + "learning_rate": 7.860034889202254e-05, + "loss": 0.0038, + "step": 526 + }, + { + "epoch": 0.07280011051250172, + "grad_norm": 0.007900571450591087, + "learning_rate": 7.862417670999348e-05, + "loss": 0.0086, + "step": 527 + }, + { + "epoch": 0.07293825113966017, + "grad_norm": 0.025396516546607018, + "learning_rate": 7.864795935671842e-05, + "loss": 0.0152, + "step": 528 + }, + { + "epoch": 0.07307639176681863, + "grad_norm": 0.03293461725115776, + "learning_rate": 7.867169700313887e-05, + "loss": 0.0226, + "step": 529 + }, + { + "epoch": 0.07321453239397707, + "grad_norm": 0.017437491565942764, + "learning_rate": 7.869538981922779e-05, + "loss": 0.0159, + "step": 530 + }, + { + "epoch": 0.07335267302113552, + "grad_norm": 0.005919649265706539, + "learning_rate": 7.871903797399694e-05, + "loss": 0.0073, + "step": 531 + }, + { + "epoch": 0.07349081364829396, + "grad_norm": 0.009071402251720428, + "learning_rate": 7.874264163550414e-05, + "loss": 0.0095, + "step": 532 + }, + { + "epoch": 0.07362895427545241, + "grad_norm": 0.005668788682669401, + "learning_rate": 7.876620097086035e-05, + "loss": 0.0073, + "step": 533 + }, + { + "epoch": 0.07376709490261085, + "grad_norm": 0.02198665961623192, + "learning_rate": 7.87897161462369e-05, + "loss": 0.0133, + "step": 534 + }, + { + "epoch": 0.07390523552976931, + "grad_norm": 0.00984006654471159, + "learning_rate": 7.881318732687234e-05, + "loss": 0.008, + "step": 535 + }, + { + "epoch": 0.07404337615692776, + "grad_norm": 0.006679135840386152, + "learning_rate": 7.88366146770796e-05, + "loss": 0.0078, + "step": 536 + }, + { + "epoch": 0.0741815167840862, + "grad_norm": 0.016889598220586777, + "learning_rate": 7.88599983602528e-05, + "loss": 0.0151, + "step": 537 + }, + { + "epoch": 0.07431965741124465, + "grad_norm": 0.09604030102491379, + "learning_rate": 7.888333853887401e-05, + "loss": 0.0759, + "step": 538 + }, + { + "epoch": 0.07445779803840309, + "grad_norm": 0.01921008713543415, + "learning_rate": 7.890663537452024e-05, + "loss": 0.0166, + "step": 539 + }, + { + "epoch": 0.07459593866556154, + "grad_norm": 0.008759930729866028, + "learning_rate": 7.892988902786988e-05, + "loss": 0.01, + "step": 540 + }, + { + "epoch": 0.07473407929271998, + "grad_norm": 0.0019715612288564444, + "learning_rate": 7.895309965870956e-05, + "loss": 0.0036, + "step": 541 + }, + { + "epoch": 0.07487221991987844, + "grad_norm": 0.012198393233120441, + "learning_rate": 7.897626742594068e-05, + "loss": 0.0103, + "step": 542 + }, + { + "epoch": 0.07501036054703689, + "grad_norm": 0.009106654673814774, + "learning_rate": 7.899939248758587e-05, + "loss": 0.0087, + "step": 543 + }, + { + "epoch": 0.07514850117419533, + "grad_norm": 0.010135483928024769, + "learning_rate": 7.902247500079554e-05, + "loss": 0.0102, + "step": 544 + }, + { + "epoch": 0.07528664180135378, + "grad_norm": 0.03742365166544914, + "learning_rate": 7.904551512185426e-05, + "loss": 0.016, + "step": 545 + }, + { + "epoch": 0.07542478242851222, + "grad_norm": 0.004954047035425901, + "learning_rate": 7.906851300618707e-05, + "loss": 0.0065, + "step": 546 + }, + { + "epoch": 0.07556292305567067, + "grad_norm": 0.02959989383816719, + "learning_rate": 7.909146880836586e-05, + "loss": 0.0156, + "step": 547 + }, + { + "epoch": 0.07570106368282913, + "grad_norm": 0.016155797988176346, + "learning_rate": 7.911438268211552e-05, + "loss": 0.0175, + "step": 548 + }, + { + "epoch": 0.07583920430998757, + "grad_norm": 0.007207171525806189, + "learning_rate": 7.913725478032012e-05, + "loss": 0.0074, + "step": 549 + }, + { + "epoch": 0.07597734493714602, + "grad_norm": 0.004485739395022392, + "learning_rate": 7.916008525502909e-05, + "loss": 0.0054, + "step": 550 + }, + { + "epoch": 0.07611548556430446, + "grad_norm": 0.02434372715651989, + "learning_rate": 7.91828742574633e-05, + "loss": 0.0146, + "step": 551 + }, + { + "epoch": 0.0762536261914629, + "grad_norm": 0.01013586763292551, + "learning_rate": 7.92056219380209e-05, + "loss": 0.0116, + "step": 552 + }, + { + "epoch": 0.07639176681862135, + "grad_norm": 0.0093398317694664, + "learning_rate": 7.922832844628351e-05, + "loss": 0.0085, + "step": 553 + }, + { + "epoch": 0.07652990744577981, + "grad_norm": 0.018571248278021812, + "learning_rate": 7.925099393102196e-05, + "loss": 0.0153, + "step": 554 + }, + { + "epoch": 0.07666804807293826, + "grad_norm": 0.006554738152772188, + "learning_rate": 7.927361854020218e-05, + "loss": 0.0079, + "step": 555 + }, + { + "epoch": 0.0768061887000967, + "grad_norm": 0.004420983139425516, + "learning_rate": 7.929620242099101e-05, + "loss": 0.0057, + "step": 556 + }, + { + "epoch": 0.07694432932725515, + "grad_norm": 0.0557420551776886, + "learning_rate": 7.931874571976197e-05, + "loss": 0.0338, + "step": 557 + }, + { + "epoch": 0.07708246995441359, + "grad_norm": 0.031253259629011154, + "learning_rate": 7.934124858210087e-05, + "loss": 0.028, + "step": 558 + }, + { + "epoch": 0.07722061058157204, + "grad_norm": 0.005481123924255371, + "learning_rate": 7.936371115281153e-05, + "loss": 0.0075, + "step": 559 + }, + { + "epoch": 0.07735875120873048, + "grad_norm": 0.013984930701553822, + "learning_rate": 7.938613357592132e-05, + "loss": 0.0168, + "step": 560 + }, + { + "epoch": 0.07749689183588894, + "grad_norm": 0.052795786410570145, + "learning_rate": 7.940851599468675e-05, + "loss": 0.0454, + "step": 561 + }, + { + "epoch": 0.07763503246304738, + "grad_norm": 0.02815697155892849, + "learning_rate": 7.943085855159892e-05, + "loss": 0.0204, + "step": 562 + }, + { + "epoch": 0.07777317309020583, + "grad_norm": 0.0057924659922719, + "learning_rate": 7.945316138838898e-05, + "loss": 0.0071, + "step": 563 + }, + { + "epoch": 0.07791131371736428, + "grad_norm": 0.015068231150507927, + "learning_rate": 7.94754246460335e-05, + "loss": 0.0121, + "step": 564 + }, + { + "epoch": 0.07804945434452272, + "grad_norm": 0.006565915886312723, + "learning_rate": 7.949764846475986e-05, + "loss": 0.006, + "step": 565 + }, + { + "epoch": 0.07818759497168117, + "grad_norm": 0.026122109964489937, + "learning_rate": 7.951983298405152e-05, + "loss": 0.0229, + "step": 566 + }, + { + "epoch": 0.07832573559883962, + "grad_norm": 0.009172670543193817, + "learning_rate": 7.95419783426533e-05, + "loss": 0.0109, + "step": 567 + }, + { + "epoch": 0.07846387622599807, + "grad_norm": 0.012615257874131203, + "learning_rate": 7.956408467857654e-05, + "loss": 0.0137, + "step": 568 + }, + { + "epoch": 0.07860201685315651, + "grad_norm": 0.015784192830324173, + "learning_rate": 7.95861521291043e-05, + "loss": 0.0144, + "step": 569 + }, + { + "epoch": 0.07874015748031496, + "grad_norm": 0.009694431908428669, + "learning_rate": 7.960818083079644e-05, + "loss": 0.0091, + "step": 570 + }, + { + "epoch": 0.0788782981074734, + "grad_norm": 0.010170293040573597, + "learning_rate": 7.963017091949477e-05, + "loss": 0.0113, + "step": 571 + }, + { + "epoch": 0.07901643873463185, + "grad_norm": 0.016057651489973068, + "learning_rate": 7.965212253032793e-05, + "loss": 0.0123, + "step": 572 + }, + { + "epoch": 0.07915457936179031, + "grad_norm": 0.002096309559419751, + "learning_rate": 7.967403579771654e-05, + "loss": 0.0037, + "step": 573 + }, + { + "epoch": 0.07929271998894875, + "grad_norm": 0.002688502660021186, + "learning_rate": 7.969591085537804e-05, + "loss": 0.0038, + "step": 574 + }, + { + "epoch": 0.0794308606161072, + "grad_norm": 0.019958576187491417, + "learning_rate": 7.97177478363316e-05, + "loss": 0.0188, + "step": 575 + }, + { + "epoch": 0.07956900124326564, + "grad_norm": 0.03227323293685913, + "learning_rate": 7.973954687290295e-05, + "loss": 0.0295, + "step": 576 + }, + { + "epoch": 0.07970714187042409, + "grad_norm": 0.0040475050918757915, + "learning_rate": 7.976130809672927e-05, + "loss": 0.0051, + "step": 577 + }, + { + "epoch": 0.07984528249758253, + "grad_norm": 0.029720835387706757, + "learning_rate": 7.978303163876388e-05, + "loss": 0.012, + "step": 578 + }, + { + "epoch": 0.079983423124741, + "grad_norm": 0.02928655594587326, + "learning_rate": 7.980471762928102e-05, + "loss": 0.019, + "step": 579 + }, + { + "epoch": 0.08012156375189944, + "grad_norm": 0.0031133065931499004, + "learning_rate": 7.982636619788048e-05, + "loss": 0.0047, + "step": 580 + }, + { + "epoch": 0.08025970437905788, + "grad_norm": 0.009925676509737968, + "learning_rate": 7.984797747349231e-05, + "loss": 0.0104, + "step": 581 + }, + { + "epoch": 0.08039784500621633, + "grad_norm": 0.013947058469057083, + "learning_rate": 7.986955158438142e-05, + "loss": 0.0148, + "step": 582 + }, + { + "epoch": 0.08053598563337477, + "grad_norm": 0.07294765114784241, + "learning_rate": 7.989108865815207e-05, + "loss": 0.0709, + "step": 583 + }, + { + "epoch": 0.08067412626053322, + "grad_norm": 0.021420951932668686, + "learning_rate": 7.99125888217525e-05, + "loss": 0.0134, + "step": 584 + }, + { + "epoch": 0.08081226688769166, + "grad_norm": 0.002957011340186, + "learning_rate": 7.993405220147938e-05, + "loss": 0.0042, + "step": 585 + }, + { + "epoch": 0.08095040751485012, + "grad_norm": 0.0029216560069471598, + "learning_rate": 7.995547892298221e-05, + "loss": 0.0036, + "step": 586 + }, + { + "epoch": 0.08108854814200857, + "grad_norm": 0.010200831107795238, + "learning_rate": 7.997686911126785e-05, + "loss": 0.011, + "step": 587 + }, + { + "epoch": 0.08122668876916701, + "grad_norm": 0.009586147964000702, + "learning_rate": 7.999822289070476e-05, + "loss": 0.0102, + "step": 588 + }, + { + "epoch": 0.08136482939632546, + "grad_norm": 0.024438267573714256, + "learning_rate": 8.001954038502744e-05, + "loss": 0.0247, + "step": 589 + }, + { + "epoch": 0.0815029700234839, + "grad_norm": 0.01673005148768425, + "learning_rate": 8.004082171734068e-05, + "loss": 0.0135, + "step": 590 + }, + { + "epoch": 0.08164111065064235, + "grad_norm": 0.03330973908305168, + "learning_rate": 8.006206701012391e-05, + "loss": 0.0231, + "step": 591 + }, + { + "epoch": 0.08177925127780081, + "grad_norm": 0.015957612544298172, + "learning_rate": 8.008327638523524e-05, + "loss": 0.0117, + "step": 592 + }, + { + "epoch": 0.08191739190495925, + "grad_norm": 0.0030719093047082424, + "learning_rate": 8.010444996391589e-05, + "loss": 0.005, + "step": 593 + }, + { + "epoch": 0.0820555325321177, + "grad_norm": 0.029894646257162094, + "learning_rate": 8.012558786679415e-05, + "loss": 0.0171, + "step": 594 + }, + { + "epoch": 0.08219367315927614, + "grad_norm": 0.0035989086609333754, + "learning_rate": 8.014669021388965e-05, + "loss": 0.0043, + "step": 595 + }, + { + "epoch": 0.08233181378643459, + "grad_norm": 0.0025325831957161427, + "learning_rate": 8.016775712461736e-05, + "loss": 0.0041, + "step": 596 + }, + { + "epoch": 0.08246995441359303, + "grad_norm": 0.023065388202667236, + "learning_rate": 8.018878871779166e-05, + "loss": 0.0185, + "step": 597 + }, + { + "epoch": 0.08260809504075149, + "grad_norm": 0.003533895593136549, + "learning_rate": 8.02097851116304e-05, + "loss": 0.0057, + "step": 598 + }, + { + "epoch": 0.08274623566790994, + "grad_norm": 0.0017970139160752296, + "learning_rate": 8.023074642375884e-05, + "loss": 0.0034, + "step": 599 + }, + { + "epoch": 0.08288437629506838, + "grad_norm": 0.009932472370564938, + "learning_rate": 8.025167277121362e-05, + "loss": 0.0103, + "step": 600 + }, + { + "epoch": 0.08302251692222683, + "grad_norm": 0.010555654764175415, + "learning_rate": 8.02725642704467e-05, + "loss": 0.009, + "step": 601 + }, + { + "epoch": 0.08316065754938527, + "grad_norm": 0.008634793572127819, + "learning_rate": 8.02934210373292e-05, + "loss": 0.0087, + "step": 602 + }, + { + "epoch": 0.08329879817654372, + "grad_norm": 0.010200063697993755, + "learning_rate": 8.031424318715534e-05, + "loss": 0.0088, + "step": 603 + }, + { + "epoch": 0.08343693880370216, + "grad_norm": 0.012205686420202255, + "learning_rate": 8.033503083464615e-05, + "loss": 0.0106, + "step": 604 + }, + { + "epoch": 0.08357507943086062, + "grad_norm": 0.010989578440785408, + "learning_rate": 8.035578409395338e-05, + "loss": 0.0085, + "step": 605 + }, + { + "epoch": 0.08371322005801907, + "grad_norm": 0.028562815859913826, + "learning_rate": 8.037650307866316e-05, + "loss": 0.0207, + "step": 606 + }, + { + "epoch": 0.08385136068517751, + "grad_norm": 0.011450926773250103, + "learning_rate": 8.039718790179982e-05, + "loss": 0.0141, + "step": 607 + }, + { + "epoch": 0.08398950131233596, + "grad_norm": 0.0031172465533018112, + "learning_rate": 8.041783867582952e-05, + "loss": 0.0047, + "step": 608 + }, + { + "epoch": 0.0841276419394944, + "grad_norm": 0.033424630761146545, + "learning_rate": 8.04384555126639e-05, + "loss": 0.0168, + "step": 609 + }, + { + "epoch": 0.08426578256665285, + "grad_norm": 0.033151671290397644, + "learning_rate": 8.045903852366387e-05, + "loss": 0.0197, + "step": 610 + }, + { + "epoch": 0.0844039231938113, + "grad_norm": 0.00440793065354228, + "learning_rate": 8.0479587819643e-05, + "loss": 0.0046, + "step": 611 + }, + { + "epoch": 0.08454206382096975, + "grad_norm": 0.019544310867786407, + "learning_rate": 8.050010351087127e-05, + "loss": 0.0152, + "step": 612 + }, + { + "epoch": 0.0846802044481282, + "grad_norm": 0.012192309834063053, + "learning_rate": 8.052058570707858e-05, + "loss": 0.0107, + "step": 613 + }, + { + "epoch": 0.08481834507528664, + "grad_norm": 0.00563014717772603, + "learning_rate": 8.054103451745819e-05, + "loss": 0.0059, + "step": 614 + }, + { + "epoch": 0.08495648570244509, + "grad_norm": 0.006608523894101381, + "learning_rate": 8.056145005067033e-05, + "loss": 0.0072, + "step": 615 + }, + { + "epoch": 0.08509462632960353, + "grad_norm": 0.006565593648701906, + "learning_rate": 8.05818324148456e-05, + "loss": 0.007, + "step": 616 + }, + { + "epoch": 0.08523276695676199, + "grad_norm": 0.0052340151742100716, + "learning_rate": 8.060218171758839e-05, + "loss": 0.0051, + "step": 617 + }, + { + "epoch": 0.08537090758392044, + "grad_norm": 0.0058824713341891766, + "learning_rate": 8.062249806598032e-05, + "loss": 0.0068, + "step": 618 + }, + { + "epoch": 0.08550904821107888, + "grad_norm": 0.008156881667673588, + "learning_rate": 8.064278156658369e-05, + "loss": 0.0073, + "step": 619 + }, + { + "epoch": 0.08564718883823733, + "grad_norm": 0.003906435798853636, + "learning_rate": 8.066303232544462e-05, + "loss": 0.0044, + "step": 620 + }, + { + "epoch": 0.08578532946539577, + "grad_norm": 0.004190659616142511, + "learning_rate": 8.068325044809664e-05, + "loss": 0.0045, + "step": 621 + }, + { + "epoch": 0.08592347009255422, + "grad_norm": 0.004072085488587618, + "learning_rate": 8.070343603956381e-05, + "loss": 0.0048, + "step": 622 + }, + { + "epoch": 0.08606161071971266, + "grad_norm": 0.08516935259103775, + "learning_rate": 8.072358920436406e-05, + "loss": 0.065, + "step": 623 + }, + { + "epoch": 0.08619975134687112, + "grad_norm": 0.12318930774927139, + "learning_rate": 8.074371004651244e-05, + "loss": 0.0462, + "step": 624 + }, + { + "epoch": 0.08633789197402957, + "grad_norm": 0.026846202090382576, + "learning_rate": 8.076379866952429e-05, + "loss": 0.0188, + "step": 625 + }, + { + "epoch": 0.08647603260118801, + "grad_norm": 0.036953698843717575, + "learning_rate": 8.078385517641853e-05, + "loss": 0.0196, + "step": 626 + }, + { + "epoch": 0.08661417322834646, + "grad_norm": 0.07914821058511734, + "learning_rate": 8.080387966972071e-05, + "loss": 0.0722, + "step": 627 + }, + { + "epoch": 0.0867523138555049, + "grad_norm": 0.05335409194231033, + "learning_rate": 8.082387225146627e-05, + "loss": 0.043, + "step": 628 + }, + { + "epoch": 0.08689045448266335, + "grad_norm": 0.008553317748010159, + "learning_rate": 8.084383302320358e-05, + "loss": 0.0077, + "step": 629 + }, + { + "epoch": 0.0870285951098218, + "grad_norm": 0.005956161767244339, + "learning_rate": 8.086376208599705e-05, + "loss": 0.0067, + "step": 630 + }, + { + "epoch": 0.08716673573698025, + "grad_norm": 0.03032430075109005, + "learning_rate": 8.088365954043025e-05, + "loss": 0.021, + "step": 631 + }, + { + "epoch": 0.0873048763641387, + "grad_norm": 0.0037184508983045816, + "learning_rate": 8.090352548660889e-05, + "loss": 0.0053, + "step": 632 + }, + { + "epoch": 0.08744301699129714, + "grad_norm": 0.014337614178657532, + "learning_rate": 8.092336002416385e-05, + "loss": 0.0122, + "step": 633 + }, + { + "epoch": 0.08758115761845559, + "grad_norm": 0.007453493308275938, + "learning_rate": 8.094316325225429e-05, + "loss": 0.0097, + "step": 634 + }, + { + "epoch": 0.08771929824561403, + "grad_norm": 0.019370267167687416, + "learning_rate": 8.096293526957041e-05, + "loss": 0.0165, + "step": 635 + }, + { + "epoch": 0.08785743887277249, + "grad_norm": 0.01445276290178299, + "learning_rate": 8.098267617433659e-05, + "loss": 0.016, + "step": 636 + }, + { + "epoch": 0.08799557949993093, + "grad_norm": 0.017797252163290977, + "learning_rate": 8.100238606431425e-05, + "loss": 0.0122, + "step": 637 + }, + { + "epoch": 0.08813372012708938, + "grad_norm": 0.014362436719238758, + "learning_rate": 8.102206503680476e-05, + "loss": 0.0126, + "step": 638 + }, + { + "epoch": 0.08827186075424782, + "grad_norm": 0.10174919664859772, + "learning_rate": 8.104171318865227e-05, + "loss": 0.0851, + "step": 639 + }, + { + "epoch": 0.08841000138140627, + "grad_norm": 0.015217412263154984, + "learning_rate": 8.106133061624669e-05, + "loss": 0.0163, + "step": 640 + }, + { + "epoch": 0.08854814200856471, + "grad_norm": 0.002396307187154889, + "learning_rate": 8.108091741552639e-05, + "loss": 0.0039, + "step": 641 + }, + { + "epoch": 0.08868628263572316, + "grad_norm": 0.002735432470217347, + "learning_rate": 8.110047368198112e-05, + "loss": 0.0045, + "step": 642 + }, + { + "epoch": 0.08882442326288162, + "grad_norm": 0.007928053848445415, + "learning_rate": 8.111999951065476e-05, + "loss": 0.0095, + "step": 643 + }, + { + "epoch": 0.08896256389004006, + "grad_norm": 0.004873833619058132, + "learning_rate": 8.113949499614808e-05, + "loss": 0.0054, + "step": 644 + }, + { + "epoch": 0.08910070451719851, + "grad_norm": 0.005578899290412664, + "learning_rate": 8.11589602326215e-05, + "loss": 0.0066, + "step": 645 + }, + { + "epoch": 0.08923884514435695, + "grad_norm": 0.004303886089473963, + "learning_rate": 8.117839531379784e-05, + "loss": 0.0062, + "step": 646 + }, + { + "epoch": 0.0893769857715154, + "grad_norm": 0.038032252341508865, + "learning_rate": 8.1197800332965e-05, + "loss": 0.0369, + "step": 647 + }, + { + "epoch": 0.08951512639867384, + "grad_norm": 0.06186262145638466, + "learning_rate": 8.121717538297867e-05, + "loss": 0.031, + "step": 648 + }, + { + "epoch": 0.0896532670258323, + "grad_norm": 0.029956836253404617, + "learning_rate": 8.123652055626497e-05, + "loss": 0.0184, + "step": 649 + }, + { + "epoch": 0.08979140765299075, + "grad_norm": 0.005330318119376898, + "learning_rate": 8.125583594482312e-05, + "loss": 0.0073, + "step": 650 + }, + { + "epoch": 0.0899295482801492, + "grad_norm": 0.006597398314625025, + "learning_rate": 8.127512164022804e-05, + "loss": 0.0071, + "step": 651 + }, + { + "epoch": 0.09006768890730764, + "grad_norm": 0.012712563388049603, + "learning_rate": 8.1294377733633e-05, + "loss": 0.0152, + "step": 652 + }, + { + "epoch": 0.09020582953446608, + "grad_norm": 0.0039291055873036385, + "learning_rate": 8.131360431577212e-05, + "loss": 0.0049, + "step": 653 + }, + { + "epoch": 0.09034397016162453, + "grad_norm": 0.009396882727742195, + "learning_rate": 8.133280147696304e-05, + "loss": 0.0122, + "step": 654 + }, + { + "epoch": 0.09048211078878299, + "grad_norm": 0.003607046790421009, + "learning_rate": 8.13519693071094e-05, + "loss": 0.0062, + "step": 655 + }, + { + "epoch": 0.09062025141594143, + "grad_norm": 0.060046661645174026, + "learning_rate": 8.137110789570339e-05, + "loss": 0.0501, + "step": 656 + }, + { + "epoch": 0.09075839204309988, + "grad_norm": 0.020894410088658333, + "learning_rate": 8.139021733182823e-05, + "loss": 0.0164, + "step": 657 + }, + { + "epoch": 0.09089653267025832, + "grad_norm": 0.01064255740493536, + "learning_rate": 8.140929770416067e-05, + "loss": 0.0113, + "step": 658 + }, + { + "epoch": 0.09103467329741677, + "grad_norm": 0.01144749391824007, + "learning_rate": 8.142834910097348e-05, + "loss": 0.0116, + "step": 659 + }, + { + "epoch": 0.09117281392457521, + "grad_norm": 0.0037209605798125267, + "learning_rate": 8.144737161013789e-05, + "loss": 0.0042, + "step": 660 + }, + { + "epoch": 0.09131095455173366, + "grad_norm": 0.014418653212487698, + "learning_rate": 8.146636531912602e-05, + "loss": 0.019, + "step": 661 + }, + { + "epoch": 0.09144909517889212, + "grad_norm": 0.004248477052897215, + "learning_rate": 8.148533031501328e-05, + "loss": 0.004, + "step": 662 + }, + { + "epoch": 0.09158723580605056, + "grad_norm": 0.023498015478253365, + "learning_rate": 8.150426668448077e-05, + "loss": 0.0156, + "step": 663 + }, + { + "epoch": 0.09172537643320901, + "grad_norm": 0.02856908179819584, + "learning_rate": 8.152317451381767e-05, + "loss": 0.021, + "step": 664 + }, + { + "epoch": 0.09186351706036745, + "grad_norm": 0.03920527175068855, + "learning_rate": 8.154205388892362e-05, + "loss": 0.041, + "step": 665 + }, + { + "epoch": 0.0920016576875259, + "grad_norm": 0.006108124274760485, + "learning_rate": 8.156090489531097e-05, + "loss": 0.0053, + "step": 666 + }, + { + "epoch": 0.09213979831468434, + "grad_norm": 0.040040936321020126, + "learning_rate": 8.157972761810723e-05, + "loss": 0.0089, + "step": 667 + }, + { + "epoch": 0.0922779389418428, + "grad_norm": 0.008494898676872253, + "learning_rate": 8.159852214205728e-05, + "loss": 0.0117, + "step": 668 + }, + { + "epoch": 0.09241607956900125, + "grad_norm": 0.07821296900510788, + "learning_rate": 8.161728855152573e-05, + "loss": 0.0578, + "step": 669 + }, + { + "epoch": 0.09255422019615969, + "grad_norm": 0.027581755071878433, + "learning_rate": 8.163602693049908e-05, + "loss": 0.0198, + "step": 670 + }, + { + "epoch": 0.09269236082331814, + "grad_norm": 0.007122738752514124, + "learning_rate": 8.165473736258814e-05, + "loss": 0.009, + "step": 671 + }, + { + "epoch": 0.09283050145047658, + "grad_norm": 0.028153732419013977, + "learning_rate": 8.167341993103011e-05, + "loss": 0.0241, + "step": 672 + }, + { + "epoch": 0.09296864207763503, + "grad_norm": 0.05543315038084984, + "learning_rate": 8.169207471869094e-05, + "loss": 0.0428, + "step": 673 + }, + { + "epoch": 0.09310678270479349, + "grad_norm": 0.013588045723736286, + "learning_rate": 8.171070180806739e-05, + "loss": 0.0097, + "step": 674 + }, + { + "epoch": 0.09324492333195193, + "grad_norm": 0.009861117228865623, + "learning_rate": 8.172930128128935e-05, + "loss": 0.009, + "step": 675 + }, + { + "epoch": 0.09338306395911038, + "grad_norm": 0.004336031153798103, + "learning_rate": 8.174787322012195e-05, + "loss": 0.0061, + "step": 676 + }, + { + "epoch": 0.09352120458626882, + "grad_norm": 0.008924839086830616, + "learning_rate": 8.176641770596768e-05, + "loss": 0.0101, + "step": 677 + }, + { + "epoch": 0.09365934521342727, + "grad_norm": 0.005986085161566734, + "learning_rate": 8.178493481986864e-05, + "loss": 0.0083, + "step": 678 + }, + { + "epoch": 0.09379748584058571, + "grad_norm": 0.006522186566144228, + "learning_rate": 8.180342464250859e-05, + "loss": 0.0075, + "step": 679 + }, + { + "epoch": 0.09393562646774416, + "grad_norm": 0.007981918752193451, + "learning_rate": 8.182188725421502e-05, + "loss": 0.0058, + "step": 680 + }, + { + "epoch": 0.09407376709490262, + "grad_norm": 0.018126454204320908, + "learning_rate": 8.184032273496133e-05, + "loss": 0.013, + "step": 681 + }, + { + "epoch": 0.09421190772206106, + "grad_norm": 0.0011805971153080463, + "learning_rate": 8.185873116436889e-05, + "loss": 0.0023, + "step": 682 + }, + { + "epoch": 0.0943500483492195, + "grad_norm": 0.016573583707213402, + "learning_rate": 8.187711262170905e-05, + "loss": 0.0152, + "step": 683 + }, + { + "epoch": 0.09448818897637795, + "grad_norm": 0.0179180596023798, + "learning_rate": 8.189546718590522e-05, + "loss": 0.0141, + "step": 684 + }, + { + "epoch": 0.0946263296035364, + "grad_norm": 0.012282337993383408, + "learning_rate": 8.191379493553499e-05, + "loss": 0.0115, + "step": 685 + }, + { + "epoch": 0.09476447023069484, + "grad_norm": 0.01766025647521019, + "learning_rate": 8.193209594883192e-05, + "loss": 0.0159, + "step": 686 + }, + { + "epoch": 0.0949026108578533, + "grad_norm": 0.002552238991484046, + "learning_rate": 8.195037030368783e-05, + "loss": 0.0041, + "step": 687 + }, + { + "epoch": 0.09504075148501175, + "grad_norm": 0.004465501289814711, + "learning_rate": 8.196861807765457e-05, + "loss": 0.0051, + "step": 688 + }, + { + "epoch": 0.09517889211217019, + "grad_norm": 0.012139561586081982, + "learning_rate": 8.198683934794608e-05, + "loss": 0.0082, + "step": 689 + }, + { + "epoch": 0.09531703273932864, + "grad_norm": 0.010561689734458923, + "learning_rate": 8.200503419144038e-05, + "loss": 0.0114, + "step": 690 + }, + { + "epoch": 0.09545517336648708, + "grad_norm": 0.025389693677425385, + "learning_rate": 8.202320268468144e-05, + "loss": 0.0187, + "step": 691 + }, + { + "epoch": 0.09559331399364553, + "grad_norm": 0.0038718711584806442, + "learning_rate": 8.204134490388117e-05, + "loss": 0.0048, + "step": 692 + }, + { + "epoch": 0.09573145462080399, + "grad_norm": 0.007652644068002701, + "learning_rate": 8.205946092492133e-05, + "loss": 0.0082, + "step": 693 + }, + { + "epoch": 0.09586959524796243, + "grad_norm": 0.006432169582694769, + "learning_rate": 8.207755082335539e-05, + "loss": 0.007, + "step": 694 + }, + { + "epoch": 0.09600773587512088, + "grad_norm": 0.0037059388123452663, + "learning_rate": 8.20956146744105e-05, + "loss": 0.0063, + "step": 695 + }, + { + "epoch": 0.09614587650227932, + "grad_norm": 0.0046118441969156265, + "learning_rate": 8.211365255298927e-05, + "loss": 0.0059, + "step": 696 + }, + { + "epoch": 0.09628401712943777, + "grad_norm": 0.0039950888603925705, + "learning_rate": 8.213166453367173e-05, + "loss": 0.0057, + "step": 697 + }, + { + "epoch": 0.09642215775659621, + "grad_norm": 0.010843550786376, + "learning_rate": 8.214965069071713e-05, + "loss": 0.0119, + "step": 698 + }, + { + "epoch": 0.09656029838375466, + "grad_norm": 0.0019335473189130425, + "learning_rate": 8.216761109806576e-05, + "loss": 0.0034, + "step": 699 + }, + { + "epoch": 0.09669843901091311, + "grad_norm": 0.031584322452545166, + "learning_rate": 8.218554582934079e-05, + "loss": 0.0217, + "step": 700 + }, + { + "epoch": 0.09683657963807156, + "grad_norm": 0.001754386001266539, + "learning_rate": 8.220345495785013e-05, + "loss": 0.0031, + "step": 701 + }, + { + "epoch": 0.09697472026523, + "grad_norm": 0.02052963897585869, + "learning_rate": 8.222133855658817e-05, + "loss": 0.0121, + "step": 702 + }, + { + "epoch": 0.09711286089238845, + "grad_norm": 0.02404898777604103, + "learning_rate": 8.223919669823754e-05, + "loss": 0.0218, + "step": 703 + }, + { + "epoch": 0.0972510015195469, + "grad_norm": 0.004977188538759947, + "learning_rate": 8.225702945517095e-05, + "loss": 0.0062, + "step": 704 + }, + { + "epoch": 0.09738914214670534, + "grad_norm": 0.007673116400837898, + "learning_rate": 8.227483689945297e-05, + "loss": 0.0083, + "step": 705 + }, + { + "epoch": 0.0975272827738638, + "grad_norm": 0.016122104600071907, + "learning_rate": 8.229261910284163e-05, + "loss": 0.0114, + "step": 706 + }, + { + "epoch": 0.09766542340102224, + "grad_norm": 0.0029692344833165407, + "learning_rate": 8.231037613679035e-05, + "loss": 0.0032, + "step": 707 + }, + { + "epoch": 0.09780356402818069, + "grad_norm": 0.016377883031964302, + "learning_rate": 8.232810807244948e-05, + "loss": 0.0107, + "step": 708 + }, + { + "epoch": 0.09794170465533913, + "grad_norm": 0.0028426761273294687, + "learning_rate": 8.23458149806682e-05, + "loss": 0.005, + "step": 709 + }, + { + "epoch": 0.09807984528249758, + "grad_norm": 0.004359726328402758, + "learning_rate": 8.236349693199602e-05, + "loss": 0.0051, + "step": 710 + }, + { + "epoch": 0.09821798590965602, + "grad_norm": 0.014035112224519253, + "learning_rate": 8.23811539966846e-05, + "loss": 0.013, + "step": 711 + }, + { + "epoch": 0.09835612653681448, + "grad_norm": 0.0015871514333412051, + "learning_rate": 8.239878624468942e-05, + "loss": 0.0029, + "step": 712 + }, + { + "epoch": 0.09849426716397293, + "grad_norm": 0.0035029149148613214, + "learning_rate": 8.241639374567137e-05, + "loss": 0.0044, + "step": 713 + }, + { + "epoch": 0.09863240779113137, + "grad_norm": 0.008812729269266129, + "learning_rate": 8.243397656899844e-05, + "loss": 0.0065, + "step": 714 + }, + { + "epoch": 0.09877054841828982, + "grad_norm": 0.03456171602010727, + "learning_rate": 8.245153478374738e-05, + "loss": 0.025, + "step": 715 + }, + { + "epoch": 0.09890868904544826, + "grad_norm": 0.0055639478377997875, + "learning_rate": 8.246906845870534e-05, + "loss": 0.0058, + "step": 716 + }, + { + "epoch": 0.09904682967260671, + "grad_norm": 0.013393756933510303, + "learning_rate": 8.248657766237134e-05, + "loss": 0.011, + "step": 717 + }, + { + "epoch": 0.09918497029976515, + "grad_norm": 0.006130352150648832, + "learning_rate": 8.250406246295815e-05, + "loss": 0.0049, + "step": 718 + }, + { + "epoch": 0.09932311092692361, + "grad_norm": 0.008264213800430298, + "learning_rate": 8.252152292839364e-05, + "loss": 0.0096, + "step": 719 + }, + { + "epoch": 0.09946125155408206, + "grad_norm": 0.05160053074359894, + "learning_rate": 8.253895912632242e-05, + "loss": 0.0319, + "step": 720 + }, + { + "epoch": 0.0995993921812405, + "grad_norm": 0.026709305122494698, + "learning_rate": 8.25563711241075e-05, + "loss": 0.018, + "step": 721 + }, + { + "epoch": 0.09973753280839895, + "grad_norm": 0.04814797267317772, + "learning_rate": 8.25737589888318e-05, + "loss": 0.0403, + "step": 722 + }, + { + "epoch": 0.0998756734355574, + "grad_norm": 0.009802248328924179, + "learning_rate": 8.259112278729965e-05, + "loss": 0.0069, + "step": 723 + }, + { + "epoch": 0.10001381406271584, + "grad_norm": 0.004918436054140329, + "learning_rate": 8.260846258603841e-05, + "loss": 0.0061, + "step": 724 + }, + { + "epoch": 0.1001519546898743, + "grad_norm": 0.00288878520950675, + "learning_rate": 8.262577845129994e-05, + "loss": 0.0035, + "step": 725 + }, + { + "epoch": 0.10029009531703274, + "grad_norm": 0.0023648052010685205, + "learning_rate": 8.264307044906217e-05, + "loss": 0.0035, + "step": 726 + }, + { + "epoch": 0.10042823594419119, + "grad_norm": 0.001214659190736711, + "learning_rate": 8.266033864503055e-05, + "loss": 0.0023, + "step": 727 + }, + { + "epoch": 0.10056637657134963, + "grad_norm": 0.00870110746473074, + "learning_rate": 8.267758310463961e-05, + "loss": 0.0082, + "step": 728 + }, + { + "epoch": 0.10070451719850808, + "grad_norm": 0.001736936392262578, + "learning_rate": 8.26948038930544e-05, + "loss": 0.0032, + "step": 729 + }, + { + "epoch": 0.10084265782566652, + "grad_norm": 0.00857970304787159, + "learning_rate": 8.271200107517198e-05, + "loss": 0.0084, + "step": 730 + }, + { + "epoch": 0.10098079845282498, + "grad_norm": 0.0043541984632611275, + "learning_rate": 8.27291747156229e-05, + "loss": 0.0059, + "step": 731 + }, + { + "epoch": 0.10111893907998343, + "grad_norm": 0.017549846321344376, + "learning_rate": 8.274632487877266e-05, + "loss": 0.0156, + "step": 732 + }, + { + "epoch": 0.10125707970714187, + "grad_norm": 0.02662530541419983, + "learning_rate": 8.276345162872311e-05, + "loss": 0.0148, + "step": 733 + }, + { + "epoch": 0.10139522033430032, + "grad_norm": 0.016454922035336494, + "learning_rate": 8.278055502931398e-05, + "loss": 0.0126, + "step": 734 + }, + { + "epoch": 0.10153336096145876, + "grad_norm": 0.0410042479634285, + "learning_rate": 8.279763514412423e-05, + "loss": 0.0273, + "step": 735 + }, + { + "epoch": 0.10167150158861721, + "grad_norm": 0.03968954086303711, + "learning_rate": 8.281469203647345e-05, + "loss": 0.0259, + "step": 736 + }, + { + "epoch": 0.10180964221577565, + "grad_norm": 0.014899387024343014, + "learning_rate": 8.283172576942335e-05, + "loss": 0.0072, + "step": 737 + }, + { + "epoch": 0.10194778284293411, + "grad_norm": 0.012564278207719326, + "learning_rate": 8.284873640577912e-05, + "loss": 0.0108, + "step": 738 + }, + { + "epoch": 0.10208592347009256, + "grad_norm": 0.0013456420274451375, + "learning_rate": 8.286572400809081e-05, + "loss": 0.0028, + "step": 739 + }, + { + "epoch": 0.102224064097251, + "grad_norm": 0.014120531268417835, + "learning_rate": 8.288268863865471e-05, + "loss": 0.0127, + "step": 740 + }, + { + "epoch": 0.10236220472440945, + "grad_norm": 0.0025778827257454395, + "learning_rate": 8.289963035951473e-05, + "loss": 0.0037, + "step": 741 + }, + { + "epoch": 0.10250034535156789, + "grad_norm": 0.06253144890069962, + "learning_rate": 8.291654923246375e-05, + "loss": 0.0239, + "step": 742 + }, + { + "epoch": 0.10263848597872634, + "grad_norm": 0.0070442091673612595, + "learning_rate": 8.293344531904503e-05, + "loss": 0.0078, + "step": 743 + }, + { + "epoch": 0.1027766266058848, + "grad_norm": 0.02459808997809887, + "learning_rate": 8.29503186805534e-05, + "loss": 0.0195, + "step": 744 + }, + { + "epoch": 0.10291476723304324, + "grad_norm": 0.0076606678776443005, + "learning_rate": 8.296716937803683e-05, + "loss": 0.0096, + "step": 745 + }, + { + "epoch": 0.10305290786020169, + "grad_norm": 0.011984504759311676, + "learning_rate": 8.298399747229752e-05, + "loss": 0.0139, + "step": 746 + }, + { + "epoch": 0.10319104848736013, + "grad_norm": 0.00821708794683218, + "learning_rate": 8.30008030238934e-05, + "loss": 0.0089, + "step": 747 + }, + { + "epoch": 0.10332918911451858, + "grad_norm": 0.03523290902376175, + "learning_rate": 8.30175860931393e-05, + "loss": 0.0264, + "step": 748 + }, + { + "epoch": 0.10346732974167702, + "grad_norm": 0.018691513687372208, + "learning_rate": 8.303434674010831e-05, + "loss": 0.0121, + "step": 749 + }, + { + "epoch": 0.10360547036883548, + "grad_norm": 0.004369445610791445, + "learning_rate": 8.305108502463309e-05, + "loss": 0.0072, + "step": 750 + }, + { + "epoch": 0.10374361099599393, + "grad_norm": 0.034502919763326645, + "learning_rate": 8.306780100630714e-05, + "loss": 0.0242, + "step": 751 + }, + { + "epoch": 0.10388175162315237, + "grad_norm": 0.0035339398309588432, + "learning_rate": 8.308449474448603e-05, + "loss": 0.0039, + "step": 752 + }, + { + "epoch": 0.10401989225031082, + "grad_norm": 0.0015408458421006799, + "learning_rate": 8.310116629828871e-05, + "loss": 0.0027, + "step": 753 + }, + { + "epoch": 0.10415803287746926, + "grad_norm": 0.09164151549339294, + "learning_rate": 8.311781572659877e-05, + "loss": 0.0502, + "step": 754 + }, + { + "epoch": 0.1042961735046277, + "grad_norm": 0.009650464169681072, + "learning_rate": 8.313444308806562e-05, + "loss": 0.0087, + "step": 755 + }, + { + "epoch": 0.10443431413178615, + "grad_norm": 0.038072582334280014, + "learning_rate": 8.315104844110585e-05, + "loss": 0.0405, + "step": 756 + }, + { + "epoch": 0.10457245475894461, + "grad_norm": 0.019803661853075027, + "learning_rate": 8.316763184390432e-05, + "loss": 0.0156, + "step": 757 + }, + { + "epoch": 0.10471059538610306, + "grad_norm": 0.02868686057627201, + "learning_rate": 8.318419335441551e-05, + "loss": 0.0259, + "step": 758 + }, + { + "epoch": 0.1048487360132615, + "grad_norm": 0.04778165742754936, + "learning_rate": 8.320073303036467e-05, + "loss": 0.0396, + "step": 759 + }, + { + "epoch": 0.10498687664041995, + "grad_norm": 0.009861230850219727, + "learning_rate": 8.321725092924898e-05, + "loss": 0.0119, + "step": 760 + }, + { + "epoch": 0.10512501726757839, + "grad_norm": 0.014904690906405449, + "learning_rate": 8.32337471083389e-05, + "loss": 0.0121, + "step": 761 + }, + { + "epoch": 0.10526315789473684, + "grad_norm": 0.022462379187345505, + "learning_rate": 8.32502216246792e-05, + "loss": 0.0252, + "step": 762 + }, + { + "epoch": 0.1054012985218953, + "grad_norm": 0.03091195411980152, + "learning_rate": 8.326667453509023e-05, + "loss": 0.0193, + "step": 763 + }, + { + "epoch": 0.10553943914905374, + "grad_norm": 0.00843064021319151, + "learning_rate": 8.328310589616908e-05, + "loss": 0.0073, + "step": 764 + }, + { + "epoch": 0.10567757977621219, + "grad_norm": 0.00403135921806097, + "learning_rate": 8.329951576429075e-05, + "loss": 0.0061, + "step": 765 + }, + { + "epoch": 0.10581572040337063, + "grad_norm": 0.024939602240920067, + "learning_rate": 8.331590419560928e-05, + "loss": 0.0218, + "step": 766 + }, + { + "epoch": 0.10595386103052908, + "grad_norm": 0.00784094538539648, + "learning_rate": 8.333227124605897e-05, + "loss": 0.0069, + "step": 767 + }, + { + "epoch": 0.10609200165768752, + "grad_norm": 0.014025907963514328, + "learning_rate": 8.334861697135548e-05, + "loss": 0.0129, + "step": 768 + }, + { + "epoch": 0.10623014228484598, + "grad_norm": 0.004523341543972492, + "learning_rate": 8.336494142699693e-05, + "loss": 0.0049, + "step": 769 + }, + { + "epoch": 0.10636828291200442, + "grad_norm": 0.09783921390771866, + "learning_rate": 8.338124466826507e-05, + "loss": 0.0225, + "step": 770 + }, + { + "epoch": 0.10650642353916287, + "grad_norm": 0.0024887770414352417, + "learning_rate": 8.339752675022647e-05, + "loss": 0.0039, + "step": 771 + }, + { + "epoch": 0.10664456416632132, + "grad_norm": 0.022068385034799576, + "learning_rate": 8.341378772773356e-05, + "loss": 0.0169, + "step": 772 + }, + { + "epoch": 0.10678270479347976, + "grad_norm": 0.02197118103504181, + "learning_rate": 8.343002765542567e-05, + "loss": 0.0161, + "step": 773 + }, + { + "epoch": 0.1069208454206382, + "grad_norm": 0.004046064335852861, + "learning_rate": 8.34462465877303e-05, + "loss": 0.0058, + "step": 774 + }, + { + "epoch": 0.10705898604779665, + "grad_norm": 0.013067352585494518, + "learning_rate": 8.346244457886409e-05, + "loss": 0.0119, + "step": 775 + }, + { + "epoch": 0.10719712667495511, + "grad_norm": 0.017200065776705742, + "learning_rate": 8.347862168283396e-05, + "loss": 0.0146, + "step": 776 + }, + { + "epoch": 0.10733526730211355, + "grad_norm": 0.007307767402380705, + "learning_rate": 8.349477795343814e-05, + "loss": 0.0089, + "step": 777 + }, + { + "epoch": 0.107473407929272, + "grad_norm": 0.06693489849567413, + "learning_rate": 8.351091344426733e-05, + "loss": 0.0603, + "step": 778 + }, + { + "epoch": 0.10761154855643044, + "grad_norm": 0.00934355054050684, + "learning_rate": 8.352702820870569e-05, + "loss": 0.0108, + "step": 779 + }, + { + "epoch": 0.10774968918358889, + "grad_norm": 0.0020181015133857727, + "learning_rate": 8.354312229993191e-05, + "loss": 0.0036, + "step": 780 + }, + { + "epoch": 0.10788782981074733, + "grad_norm": 0.018208837136626244, + "learning_rate": 8.35591957709203e-05, + "loss": 0.0162, + "step": 781 + }, + { + "epoch": 0.1080259704379058, + "grad_norm": 0.026561802253127098, + "learning_rate": 8.357524867444177e-05, + "loss": 0.0139, + "step": 782 + }, + { + "epoch": 0.10816411106506424, + "grad_norm": 0.03126946836709976, + "learning_rate": 8.359128106306499e-05, + "loss": 0.0312, + "step": 783 + }, + { + "epoch": 0.10830225169222268, + "grad_norm": 0.002232447499409318, + "learning_rate": 8.360729298915728e-05, + "loss": 0.0038, + "step": 784 + }, + { + "epoch": 0.10844039231938113, + "grad_norm": 0.012769825756549835, + "learning_rate": 8.362328450488575e-05, + "loss": 0.0112, + "step": 785 + }, + { + "epoch": 0.10857853294653957, + "grad_norm": 0.005585674196481705, + "learning_rate": 8.36392556622182e-05, + "loss": 0.0063, + "step": 786 + }, + { + "epoch": 0.10871667357369802, + "grad_norm": 0.0049601225182414055, + "learning_rate": 8.365520651292431e-05, + "loss": 0.0055, + "step": 787 + }, + { + "epoch": 0.10885481420085648, + "grad_norm": 0.008773828856647015, + "learning_rate": 8.367113710857645e-05, + "loss": 0.0102, + "step": 788 + }, + { + "epoch": 0.10899295482801492, + "grad_norm": 0.009176612831652164, + "learning_rate": 8.36870475005508e-05, + "loss": 0.0098, + "step": 789 + }, + { + "epoch": 0.10913109545517337, + "grad_norm": 0.04428212344646454, + "learning_rate": 8.370293774002835e-05, + "loss": 0.0264, + "step": 790 + }, + { + "epoch": 0.10926923608233181, + "grad_norm": 0.0029104934073984623, + "learning_rate": 8.371880787799582e-05, + "loss": 0.0039, + "step": 791 + }, + { + "epoch": 0.10940737670949026, + "grad_norm": 0.01047749724239111, + "learning_rate": 8.37346579652467e-05, + "loss": 0.0093, + "step": 792 + }, + { + "epoch": 0.1095455173366487, + "grad_norm": 0.01725156232714653, + "learning_rate": 8.375048805238215e-05, + "loss": 0.0174, + "step": 793 + }, + { + "epoch": 0.10968365796380715, + "grad_norm": 0.0077973040752112865, + "learning_rate": 8.37662981898121e-05, + "loss": 0.0073, + "step": 794 + }, + { + "epoch": 0.10982179859096561, + "grad_norm": 0.0034406129270792007, + "learning_rate": 8.378208842775606e-05, + "loss": 0.0047, + "step": 795 + }, + { + "epoch": 0.10995993921812405, + "grad_norm": 0.004409421235322952, + "learning_rate": 8.37978588162442e-05, + "loss": 0.0052, + "step": 796 + }, + { + "epoch": 0.1100980798452825, + "grad_norm": 0.020356660708785057, + "learning_rate": 8.381360940511824e-05, + "loss": 0.0189, + "step": 797 + }, + { + "epoch": 0.11023622047244094, + "grad_norm": 0.009040805511176586, + "learning_rate": 8.382934024403241e-05, + "loss": 0.0075, + "step": 798 + }, + { + "epoch": 0.11037436109959939, + "grad_norm": 0.004525200929492712, + "learning_rate": 8.384505138245437e-05, + "loss": 0.0055, + "step": 799 + }, + { + "epoch": 0.11051250172675783, + "grad_norm": 0.0028125497046858072, + "learning_rate": 8.386074286966616e-05, + "loss": 0.0045, + "step": 800 + }, + { + "epoch": 0.11065064235391629, + "grad_norm": 0.007167664356529713, + "learning_rate": 8.387641475476517e-05, + "loss": 0.0089, + "step": 801 + }, + { + "epoch": 0.11078878298107474, + "grad_norm": 0.02145918644964695, + "learning_rate": 8.389206708666495e-05, + "loss": 0.0109, + "step": 802 + }, + { + "epoch": 0.11092692360823318, + "grad_norm": 0.07421177625656128, + "learning_rate": 8.390769991409625e-05, + "loss": 0.0339, + "step": 803 + }, + { + "epoch": 0.11106506423539163, + "grad_norm": 0.016273144632577896, + "learning_rate": 8.392331328560788e-05, + "loss": 0.0138, + "step": 804 + }, + { + "epoch": 0.11120320486255007, + "grad_norm": 0.006689921021461487, + "learning_rate": 8.393890724956755e-05, + "loss": 0.0051, + "step": 805 + }, + { + "epoch": 0.11134134548970852, + "grad_norm": 0.013465960510075092, + "learning_rate": 8.39544818541629e-05, + "loss": 0.0142, + "step": 806 + }, + { + "epoch": 0.11147948611686698, + "grad_norm": 0.003067552577704191, + "learning_rate": 8.39700371474023e-05, + "loss": 0.0035, + "step": 807 + }, + { + "epoch": 0.11161762674402542, + "grad_norm": 0.002693150658160448, + "learning_rate": 8.398557317711569e-05, + "loss": 0.0042, + "step": 808 + }, + { + "epoch": 0.11175576737118387, + "grad_norm": 0.0046765501610934734, + "learning_rate": 8.400108999095567e-05, + "loss": 0.0049, + "step": 809 + }, + { + "epoch": 0.11189390799834231, + "grad_norm": 0.004729992710053921, + "learning_rate": 8.401658763639814e-05, + "loss": 0.0059, + "step": 810 + }, + { + "epoch": 0.11203204862550076, + "grad_norm": 0.011979191564023495, + "learning_rate": 8.403206616074328e-05, + "loss": 0.015, + "step": 811 + }, + { + "epoch": 0.1121701892526592, + "grad_norm": 0.002755200956016779, + "learning_rate": 8.404752561111644e-05, + "loss": 0.0038, + "step": 812 + }, + { + "epoch": 0.11230832987981765, + "grad_norm": 0.010400881990790367, + "learning_rate": 8.406296603446894e-05, + "loss": 0.0117, + "step": 813 + }, + { + "epoch": 0.1124464705069761, + "grad_norm": 0.006119609344750643, + "learning_rate": 8.407838747757898e-05, + "loss": 0.0071, + "step": 814 + }, + { + "epoch": 0.11258461113413455, + "grad_norm": 0.005961157381534576, + "learning_rate": 8.409378998705247e-05, + "loss": 0.0072, + "step": 815 + }, + { + "epoch": 0.112722751761293, + "grad_norm": 0.0021235195454210043, + "learning_rate": 8.41091736093238e-05, + "loss": 0.0033, + "step": 816 + }, + { + "epoch": 0.11286089238845144, + "grad_norm": 0.004867125768214464, + "learning_rate": 8.412453839065685e-05, + "loss": 0.0047, + "step": 817 + }, + { + "epoch": 0.11299903301560989, + "grad_norm": 0.021859489381313324, + "learning_rate": 8.413988437714566e-05, + "loss": 0.0218, + "step": 818 + }, + { + "epoch": 0.11313717364276833, + "grad_norm": 0.004767335951328278, + "learning_rate": 8.415521161471534e-05, + "loss": 0.005, + "step": 819 + }, + { + "epoch": 0.11327531426992679, + "grad_norm": 0.0023131791967898607, + "learning_rate": 8.417052014912287e-05, + "loss": 0.003, + "step": 820 + }, + { + "epoch": 0.11341345489708524, + "grad_norm": 0.001652119099162519, + "learning_rate": 8.418581002595795e-05, + "loss": 0.003, + "step": 821 + }, + { + "epoch": 0.11355159552424368, + "grad_norm": 0.06345450133085251, + "learning_rate": 8.420108129064377e-05, + "loss": 0.0511, + "step": 822 + }, + { + "epoch": 0.11368973615140213, + "grad_norm": 0.001477993791922927, + "learning_rate": 8.421633398843785e-05, + "loss": 0.0027, + "step": 823 + }, + { + "epoch": 0.11382787677856057, + "grad_norm": 0.010723377577960491, + "learning_rate": 8.423156816443287e-05, + "loss": 0.0119, + "step": 824 + }, + { + "epoch": 0.11396601740571902, + "grad_norm": 0.005179021507501602, + "learning_rate": 8.424678386355736e-05, + "loss": 0.0044, + "step": 825 + }, + { + "epoch": 0.11410415803287748, + "grad_norm": 0.004340772982686758, + "learning_rate": 8.426198113057666e-05, + "loss": 0.0051, + "step": 826 + }, + { + "epoch": 0.11424229866003592, + "grad_norm": 0.011915227398276329, + "learning_rate": 8.427716001009356e-05, + "loss": 0.0123, + "step": 827 + }, + { + "epoch": 0.11438043928719437, + "grad_norm": 0.006282525602728128, + "learning_rate": 8.429232054654917e-05, + "loss": 0.0064, + "step": 828 + }, + { + "epoch": 0.11451857991435281, + "grad_norm": 0.10944747924804688, + "learning_rate": 8.43074627842237e-05, + "loss": 0.0614, + "step": 829 + }, + { + "epoch": 0.11465672054151126, + "grad_norm": 0.002959243021905422, + "learning_rate": 8.432258676723714e-05, + "loss": 0.0042, + "step": 830 + }, + { + "epoch": 0.1147948611686697, + "grad_norm": 0.010976544581353664, + "learning_rate": 8.433769253955023e-05, + "loss": 0.0088, + "step": 831 + }, + { + "epoch": 0.11493300179582815, + "grad_norm": 0.015172791667282581, + "learning_rate": 8.435278014496498e-05, + "loss": 0.0099, + "step": 832 + }, + { + "epoch": 0.1150711424229866, + "grad_norm": 0.0273448433727026, + "learning_rate": 8.436784962712563e-05, + "loss": 0.0216, + "step": 833 + }, + { + "epoch": 0.11520928305014505, + "grad_norm": 0.05054100230336189, + "learning_rate": 8.438290102951929e-05, + "loss": 0.0301, + "step": 834 + }, + { + "epoch": 0.1153474236773035, + "grad_norm": 0.006080084480345249, + "learning_rate": 8.439793439547675e-05, + "loss": 0.0086, + "step": 835 + }, + { + "epoch": 0.11548556430446194, + "grad_norm": 0.008105085231363773, + "learning_rate": 8.441294976817325e-05, + "loss": 0.0088, + "step": 836 + }, + { + "epoch": 0.11562370493162039, + "grad_norm": 0.030704544857144356, + "learning_rate": 8.442794719062914e-05, + "loss": 0.029, + "step": 837 + }, + { + "epoch": 0.11576184555877883, + "grad_norm": 0.09055589139461517, + "learning_rate": 8.444292670571065e-05, + "loss": 0.0448, + "step": 838 + }, + { + "epoch": 0.11589998618593729, + "grad_norm": 0.03190995752811432, + "learning_rate": 8.445788835613072e-05, + "loss": 0.0238, + "step": 839 + }, + { + "epoch": 0.11603812681309573, + "grad_norm": 0.0042504677549004555, + "learning_rate": 8.447283218444959e-05, + "loss": 0.0047, + "step": 840 + }, + { + "epoch": 0.11617626744025418, + "grad_norm": 0.006505970843136311, + "learning_rate": 8.44877582330756e-05, + "loss": 0.0067, + "step": 841 + }, + { + "epoch": 0.11631440806741263, + "grad_norm": 0.0037945907097309828, + "learning_rate": 8.450266654426589e-05, + "loss": 0.0043, + "step": 842 + }, + { + "epoch": 0.11645254869457107, + "grad_norm": 0.14548099040985107, + "learning_rate": 8.451755716012719e-05, + "loss": 0.2458, + "step": 843 + }, + { + "epoch": 0.11659068932172952, + "grad_norm": 0.01771988905966282, + "learning_rate": 8.453243012261641e-05, + "loss": 0.0138, + "step": 844 + }, + { + "epoch": 0.11672882994888797, + "grad_norm": 0.01083680521696806, + "learning_rate": 8.454728547354141e-05, + "loss": 0.0101, + "step": 845 + }, + { + "epoch": 0.11686697057604642, + "grad_norm": 0.006191121879965067, + "learning_rate": 8.456212325456176e-05, + "loss": 0.007, + "step": 846 + }, + { + "epoch": 0.11700511120320486, + "grad_norm": 0.035239238291978836, + "learning_rate": 8.457694350718935e-05, + "loss": 0.0178, + "step": 847 + }, + { + "epoch": 0.11714325183036331, + "grad_norm": 0.020794779062271118, + "learning_rate": 8.459174627278912e-05, + "loss": 0.0143, + "step": 848 + }, + { + "epoch": 0.11728139245752175, + "grad_norm": 0.0027690723072737455, + "learning_rate": 8.460653159257979e-05, + "loss": 0.0043, + "step": 849 + }, + { + "epoch": 0.1174195330846802, + "grad_norm": 0.011040976271033287, + "learning_rate": 8.46212995076345e-05, + "loss": 0.0097, + "step": 850 + }, + { + "epoch": 0.11755767371183865, + "grad_norm": 0.05898001417517662, + "learning_rate": 8.463605005888147e-05, + "loss": 0.0432, + "step": 851 + }, + { + "epoch": 0.1176958143389971, + "grad_norm": 0.03841552138328552, + "learning_rate": 8.465078328710481e-05, + "loss": 0.0265, + "step": 852 + }, + { + "epoch": 0.11783395496615555, + "grad_norm": 0.09727146476507187, + "learning_rate": 8.466549923294504e-05, + "loss": 0.062, + "step": 853 + }, + { + "epoch": 0.117972095593314, + "grad_norm": 0.006218411959707737, + "learning_rate": 8.468019793689983e-05, + "loss": 0.0076, + "step": 854 + }, + { + "epoch": 0.11811023622047244, + "grad_norm": 0.008461462333798409, + "learning_rate": 8.469487943932471e-05, + "loss": 0.0089, + "step": 855 + }, + { + "epoch": 0.11824837684763088, + "grad_norm": 0.011293224059045315, + "learning_rate": 8.470954378043366e-05, + "loss": 0.0106, + "step": 856 + }, + { + "epoch": 0.11838651747478933, + "grad_norm": 0.001236687763594091, + "learning_rate": 8.472419100029984e-05, + "loss": 0.0026, + "step": 857 + }, + { + "epoch": 0.11852465810194779, + "grad_norm": 0.017730748280882835, + "learning_rate": 8.473882113885618e-05, + "loss": 0.0171, + "step": 858 + }, + { + "epoch": 0.11866279872910623, + "grad_norm": 0.017721619457006454, + "learning_rate": 8.475343423589608e-05, + "loss": 0.0151, + "step": 859 + }, + { + "epoch": 0.11880093935626468, + "grad_norm": 0.005204800516366959, + "learning_rate": 8.476803033107404e-05, + "loss": 0.0048, + "step": 860 + }, + { + "epoch": 0.11893907998342312, + "grad_norm": 0.013895371928811073, + "learning_rate": 8.47826094639063e-05, + "loss": 0.0091, + "step": 861 + }, + { + "epoch": 0.11907722061058157, + "grad_norm": 0.046134259551763535, + "learning_rate": 8.479717167377152e-05, + "loss": 0.0408, + "step": 862 + }, + { + "epoch": 0.11921536123774001, + "grad_norm": 0.00713316909968853, + "learning_rate": 8.481171699991137e-05, + "loss": 0.0094, + "step": 863 + }, + { + "epoch": 0.11935350186489847, + "grad_norm": 0.08703207969665527, + "learning_rate": 8.48262454814312e-05, + "loss": 0.0581, + "step": 864 + }, + { + "epoch": 0.11949164249205692, + "grad_norm": 0.0035085680428892374, + "learning_rate": 8.484075715730064e-05, + "loss": 0.0053, + "step": 865 + }, + { + "epoch": 0.11962978311921536, + "grad_norm": 0.01007845625281334, + "learning_rate": 8.485525206635425e-05, + "loss": 0.0089, + "step": 866 + }, + { + "epoch": 0.11976792374637381, + "grad_norm": 0.02679300867021084, + "learning_rate": 8.486973024729214e-05, + "loss": 0.0161, + "step": 867 + }, + { + "epoch": 0.11990606437353225, + "grad_norm": 0.006040680222213268, + "learning_rate": 8.488419173868058e-05, + "loss": 0.0079, + "step": 868 + }, + { + "epoch": 0.1200442050006907, + "grad_norm": 0.004664016887545586, + "learning_rate": 8.489863657895262e-05, + "loss": 0.006, + "step": 869 + }, + { + "epoch": 0.12018234562784914, + "grad_norm": 0.01343533769249916, + "learning_rate": 8.491306480640874e-05, + "loss": 0.0146, + "step": 870 + }, + { + "epoch": 0.1203204862550076, + "grad_norm": 0.019741838797926903, + "learning_rate": 8.492747645921737e-05, + "loss": 0.0166, + "step": 871 + }, + { + "epoch": 0.12045862688216605, + "grad_norm": 0.01905057206749916, + "learning_rate": 8.494187157541558e-05, + "loss": 0.0118, + "step": 872 + }, + { + "epoch": 0.12059676750932449, + "grad_norm": 0.02122565358877182, + "learning_rate": 8.495625019290968e-05, + "loss": 0.0171, + "step": 873 + }, + { + "epoch": 0.12073490813648294, + "grad_norm": 0.006704521831125021, + "learning_rate": 8.497061234947573e-05, + "loss": 0.0061, + "step": 874 + }, + { + "epoch": 0.12087304876364138, + "grad_norm": 0.0038077067583799362, + "learning_rate": 8.498495808276026e-05, + "loss": 0.005, + "step": 875 + }, + { + "epoch": 0.12101118939079983, + "grad_norm": 0.036112312227487564, + "learning_rate": 8.499928743028077e-05, + "loss": 0.0333, + "step": 876 + }, + { + "epoch": 0.12114933001795829, + "grad_norm": 0.005154281854629517, + "learning_rate": 8.501360042942633e-05, + "loss": 0.0075, + "step": 877 + }, + { + "epoch": 0.12128747064511673, + "grad_norm": 0.0023703796323388815, + "learning_rate": 8.502789711745823e-05, + "loss": 0.0042, + "step": 878 + }, + { + "epoch": 0.12142561127227518, + "grad_norm": 0.03203393518924713, + "learning_rate": 8.504217753151046e-05, + "loss": 0.0235, + "step": 879 + }, + { + "epoch": 0.12156375189943362, + "grad_norm": 0.02325097844004631, + "learning_rate": 8.505644170859043e-05, + "loss": 0.0168, + "step": 880 + }, + { + "epoch": 0.12170189252659207, + "grad_norm": 0.04458438232541084, + "learning_rate": 8.507068968557937e-05, + "loss": 0.0298, + "step": 881 + }, + { + "epoch": 0.12184003315375051, + "grad_norm": 0.022350402548909187, + "learning_rate": 8.508492149923301e-05, + "loss": 0.0173, + "step": 882 + }, + { + "epoch": 0.12197817378090897, + "grad_norm": 0.0038661733269691467, + "learning_rate": 8.509913718618219e-05, + "loss": 0.0058, + "step": 883 + }, + { + "epoch": 0.12211631440806742, + "grad_norm": 0.03452976420521736, + "learning_rate": 8.511333678293331e-05, + "loss": 0.0153, + "step": 884 + }, + { + "epoch": 0.12225445503522586, + "grad_norm": 0.00885214377194643, + "learning_rate": 8.512752032586895e-05, + "loss": 0.0092, + "step": 885 + }, + { + "epoch": 0.12239259566238431, + "grad_norm": 0.01287474762648344, + "learning_rate": 8.514168785124848e-05, + "loss": 0.0101, + "step": 886 + }, + { + "epoch": 0.12253073628954275, + "grad_norm": 0.0044240327551960945, + "learning_rate": 8.51558393952085e-05, + "loss": 0.004, + "step": 887 + }, + { + "epoch": 0.1226688769167012, + "grad_norm": 0.0011661675525829196, + "learning_rate": 8.516997499376351e-05, + "loss": 0.0023, + "step": 888 + }, + { + "epoch": 0.12280701754385964, + "grad_norm": 0.014828033745288849, + "learning_rate": 8.518409468280637e-05, + "loss": 0.0093, + "step": 889 + }, + { + "epoch": 0.1229451581710181, + "grad_norm": 0.006530273240059614, + "learning_rate": 8.51981984981089e-05, + "loss": 0.0072, + "step": 890 + }, + { + "epoch": 0.12308329879817655, + "grad_norm": 0.010840199887752533, + "learning_rate": 8.521228647532241e-05, + "loss": 0.0099, + "step": 891 + }, + { + "epoch": 0.12322143942533499, + "grad_norm": 0.012677641585469246, + "learning_rate": 8.522635864997826e-05, + "loss": 0.0098, + "step": 892 + }, + { + "epoch": 0.12335958005249344, + "grad_norm": 0.011761125177145004, + "learning_rate": 8.524041505748831e-05, + "loss": 0.0112, + "step": 893 + }, + { + "epoch": 0.12349772067965188, + "grad_norm": 0.05986681953072548, + "learning_rate": 8.525445573314562e-05, + "loss": 0.0406, + "step": 894 + }, + { + "epoch": 0.12363586130681033, + "grad_norm": 0.015386831946671009, + "learning_rate": 8.52684807121248e-05, + "loss": 0.0148, + "step": 895 + }, + { + "epoch": 0.12377400193396879, + "grad_norm": 0.011633604764938354, + "learning_rate": 8.528249002948265e-05, + "loss": 0.0101, + "step": 896 + }, + { + "epoch": 0.12391214256112723, + "grad_norm": 0.008506370708346367, + "learning_rate": 8.529648372015867e-05, + "loss": 0.0066, + "step": 897 + }, + { + "epoch": 0.12405028318828568, + "grad_norm": 0.05086902156472206, + "learning_rate": 8.531046181897555e-05, + "loss": 0.023, + "step": 898 + }, + { + "epoch": 0.12418842381544412, + "grad_norm": 0.001341451657935977, + "learning_rate": 8.532442436063973e-05, + "loss": 0.0021, + "step": 899 + }, + { + "epoch": 0.12432656444260257, + "grad_norm": 0.05210532248020172, + "learning_rate": 8.533837137974189e-05, + "loss": 0.0323, + "step": 900 + }, + { + "epoch": 0.12446470506976101, + "grad_norm": 0.007349935360252857, + "learning_rate": 8.535230291075745e-05, + "loss": 0.0087, + "step": 901 + }, + { + "epoch": 0.12460284569691947, + "grad_norm": 0.004401295445859432, + "learning_rate": 8.536621898804715e-05, + "loss": 0.0056, + "step": 902 + }, + { + "epoch": 0.12474098632407792, + "grad_norm": 0.016419509425759315, + "learning_rate": 8.538011964585747e-05, + "loss": 0.0146, + "step": 903 + }, + { + "epoch": 0.12487912695123636, + "grad_norm": 0.012662366963922977, + "learning_rate": 8.539400491832118e-05, + "loss": 0.0102, + "step": 904 + }, + { + "epoch": 0.12501726757839482, + "grad_norm": 0.0041525522246956825, + "learning_rate": 8.540787483945788e-05, + "loss": 0.0056, + "step": 905 + }, + { + "epoch": 0.12515540820555326, + "grad_norm": 0.001923445495776832, + "learning_rate": 8.542172944317442e-05, + "loss": 0.0033, + "step": 906 + }, + { + "epoch": 0.1252935488327117, + "grad_norm": 0.039593107998371124, + "learning_rate": 8.543556876326543e-05, + "loss": 0.0238, + "step": 907 + }, + { + "epoch": 0.12543168945987015, + "grad_norm": 0.004473451524972916, + "learning_rate": 8.544939283341387e-05, + "loss": 0.0058, + "step": 908 + }, + { + "epoch": 0.1255698300870286, + "grad_norm": 0.004258911591023207, + "learning_rate": 8.546320168719143e-05, + "loss": 0.005, + "step": 909 + }, + { + "epoch": 0.12570797071418705, + "grad_norm": 0.026720302179455757, + "learning_rate": 8.547699535805908e-05, + "loss": 0.022, + "step": 910 + }, + { + "epoch": 0.1258461113413455, + "grad_norm": 0.005456454586237669, + "learning_rate": 8.549077387936755e-05, + "loss": 0.0052, + "step": 911 + }, + { + "epoch": 0.12598425196850394, + "grad_norm": 0.01937148906290531, + "learning_rate": 8.550453728435777e-05, + "loss": 0.0169, + "step": 912 + }, + { + "epoch": 0.12612239259566238, + "grad_norm": 0.14971759915351868, + "learning_rate": 8.551828560616142e-05, + "loss": 0.2231, + "step": 913 + }, + { + "epoch": 0.12626053322282083, + "grad_norm": 0.003774692304432392, + "learning_rate": 8.553201887780136e-05, + "loss": 0.0051, + "step": 914 + }, + { + "epoch": 0.12639867384997927, + "grad_norm": 0.012158801779150963, + "learning_rate": 8.554573713219211e-05, + "loss": 0.0086, + "step": 915 + }, + { + "epoch": 0.12653681447713772, + "grad_norm": 0.0024195557925850153, + "learning_rate": 8.555944040214037e-05, + "loss": 0.004, + "step": 916 + }, + { + "epoch": 0.12667495510429616, + "grad_norm": 0.012309069745242596, + "learning_rate": 8.557312872034538e-05, + "loss": 0.0113, + "step": 917 + }, + { + "epoch": 0.12681309573145463, + "grad_norm": 0.017295846715569496, + "learning_rate": 8.558680211939955e-05, + "loss": 0.0169, + "step": 918 + }, + { + "epoch": 0.12695123635861308, + "grad_norm": 0.013580954633653164, + "learning_rate": 8.560046063178874e-05, + "loss": 0.014, + "step": 919 + }, + { + "epoch": 0.12708937698577152, + "grad_norm": 0.03825406730175018, + "learning_rate": 8.561410428989292e-05, + "loss": 0.0345, + "step": 920 + }, + { + "epoch": 0.12722751761292997, + "grad_norm": 0.010689822025597095, + "learning_rate": 8.562773312598646e-05, + "loss": 0.0128, + "step": 921 + }, + { + "epoch": 0.12736565824008841, + "grad_norm": 0.005025971680879593, + "learning_rate": 8.564134717223869e-05, + "loss": 0.0053, + "step": 922 + }, + { + "epoch": 0.12750379886724686, + "grad_norm": 0.005272683221846819, + "learning_rate": 8.56549464607143e-05, + "loss": 0.0048, + "step": 923 + }, + { + "epoch": 0.1276419394944053, + "grad_norm": 0.006757293362170458, + "learning_rate": 8.566853102337387e-05, + "loss": 0.0071, + "step": 924 + }, + { + "epoch": 0.12778008012156375, + "grad_norm": 0.013046073727309704, + "learning_rate": 8.568210089207418e-05, + "loss": 0.0075, + "step": 925 + }, + { + "epoch": 0.1279182207487222, + "grad_norm": 0.06407853960990906, + "learning_rate": 8.569565609856885e-05, + "loss": 0.0422, + "step": 926 + }, + { + "epoch": 0.12805636137588064, + "grad_norm": 0.02519175596535206, + "learning_rate": 8.570919667450859e-05, + "loss": 0.0159, + "step": 927 + }, + { + "epoch": 0.12819450200303908, + "grad_norm": 0.0054557062685489655, + "learning_rate": 8.57227226514418e-05, + "loss": 0.0065, + "step": 928 + }, + { + "epoch": 0.12833264263019753, + "grad_norm": 0.023621072992682457, + "learning_rate": 8.573623406081491e-05, + "loss": 0.0177, + "step": 929 + }, + { + "epoch": 0.12847078325735597, + "grad_norm": 0.0037180378567427397, + "learning_rate": 8.574973093397288e-05, + "loss": 0.0041, + "step": 930 + }, + { + "epoch": 0.12860892388451445, + "grad_norm": 0.020301776006817818, + "learning_rate": 8.576321330215958e-05, + "loss": 0.0146, + "step": 931 + }, + { + "epoch": 0.1287470645116729, + "grad_norm": 0.004708379507064819, + "learning_rate": 8.577668119651829e-05, + "loss": 0.0063, + "step": 932 + }, + { + "epoch": 0.12888520513883134, + "grad_norm": 0.0036832920741289854, + "learning_rate": 8.579013464809206e-05, + "loss": 0.0043, + "step": 933 + }, + { + "epoch": 0.12902334576598978, + "grad_norm": 0.016041239723563194, + "learning_rate": 8.580357368782425e-05, + "loss": 0.0126, + "step": 934 + }, + { + "epoch": 0.12916148639314823, + "grad_norm": 0.07763788849115372, + "learning_rate": 8.581699834655876e-05, + "loss": 0.0569, + "step": 935 + }, + { + "epoch": 0.12929962702030667, + "grad_norm": 0.04748247563838959, + "learning_rate": 8.58304086550407e-05, + "loss": 0.0277, + "step": 936 + }, + { + "epoch": 0.12943776764746512, + "grad_norm": 0.006311520934104919, + "learning_rate": 8.584380464391663e-05, + "loss": 0.0056, + "step": 937 + }, + { + "epoch": 0.12957590827462356, + "grad_norm": 0.004489540588110685, + "learning_rate": 8.585718634373504e-05, + "loss": 0.0057, + "step": 938 + }, + { + "epoch": 0.129714048901782, + "grad_norm": 0.01480119675397873, + "learning_rate": 8.587055378494679e-05, + "loss": 0.0148, + "step": 939 + }, + { + "epoch": 0.12985218952894045, + "grad_norm": 0.025000635534524918, + "learning_rate": 8.588390699790551e-05, + "loss": 0.0173, + "step": 940 + }, + { + "epoch": 0.1299903301560989, + "grad_norm": 0.0023975621443241835, + "learning_rate": 8.589724601286795e-05, + "loss": 0.0031, + "step": 941 + }, + { + "epoch": 0.13012847078325734, + "grad_norm": 0.036758214235305786, + "learning_rate": 8.591057085999454e-05, + "loss": 0.0375, + "step": 942 + }, + { + "epoch": 0.13026661141041582, + "grad_norm": 0.02752687968313694, + "learning_rate": 8.592388156934964e-05, + "loss": 0.0153, + "step": 943 + }, + { + "epoch": 0.13040475203757426, + "grad_norm": 0.0026282810140401125, + "learning_rate": 8.593717817090202e-05, + "loss": 0.004, + "step": 944 + }, + { + "epoch": 0.1305428926647327, + "grad_norm": 0.00695823086425662, + "learning_rate": 8.595046069452531e-05, + "loss": 0.0067, + "step": 945 + }, + { + "epoch": 0.13068103329189115, + "grad_norm": 0.005962789058685303, + "learning_rate": 8.596372916999831e-05, + "loss": 0.007, + "step": 946 + }, + { + "epoch": 0.1308191739190496, + "grad_norm": 0.19016483426094055, + "learning_rate": 8.597698362700545e-05, + "loss": 0.1038, + "step": 947 + }, + { + "epoch": 0.13095731454620804, + "grad_norm": 0.009827204048633575, + "learning_rate": 8.599022409513714e-05, + "loss": 0.0071, + "step": 948 + }, + { + "epoch": 0.1310954551733665, + "grad_norm": 0.0029148852918297052, + "learning_rate": 8.600345060389027e-05, + "loss": 0.0041, + "step": 949 + }, + { + "epoch": 0.13123359580052493, + "grad_norm": 0.04867846518754959, + "learning_rate": 8.601666318266846e-05, + "loss": 0.0244, + "step": 950 + }, + { + "epoch": 0.13137173642768338, + "grad_norm": 0.005912700667977333, + "learning_rate": 8.602986186078254e-05, + "loss": 0.0086, + "step": 951 + }, + { + "epoch": 0.13150987705484182, + "grad_norm": 0.006628013215959072, + "learning_rate": 8.604304666745098e-05, + "loss": 0.0081, + "step": 952 + }, + { + "epoch": 0.13164801768200027, + "grad_norm": 0.008250492624938488, + "learning_rate": 8.605621763180016e-05, + "loss": 0.0074, + "step": 953 + }, + { + "epoch": 0.1317861583091587, + "grad_norm": 0.013309494592249393, + "learning_rate": 8.606937478286485e-05, + "loss": 0.0143, + "step": 954 + }, + { + "epoch": 0.13192429893631716, + "grad_norm": 0.0022110934369266033, + "learning_rate": 8.608251814958855e-05, + "loss": 0.0041, + "step": 955 + }, + { + "epoch": 0.13206243956347563, + "grad_norm": 0.009347507730126381, + "learning_rate": 8.609564776082388e-05, + "loss": 0.0117, + "step": 956 + }, + { + "epoch": 0.13220058019063408, + "grad_norm": 0.007068824488669634, + "learning_rate": 8.610876364533301e-05, + "loss": 0.0107, + "step": 957 + }, + { + "epoch": 0.13233872081779252, + "grad_norm": 0.02662699855864048, + "learning_rate": 8.612186583178794e-05, + "loss": 0.0315, + "step": 958 + }, + { + "epoch": 0.13247686144495097, + "grad_norm": 0.015548110008239746, + "learning_rate": 8.613495434877095e-05, + "loss": 0.016, + "step": 959 + }, + { + "epoch": 0.1326150020721094, + "grad_norm": 0.0028302946593612432, + "learning_rate": 8.614802922477495e-05, + "loss": 0.0037, + "step": 960 + }, + { + "epoch": 0.13275314269926786, + "grad_norm": 0.0047238729894161224, + "learning_rate": 8.616109048820388e-05, + "loss": 0.006, + "step": 961 + }, + { + "epoch": 0.1328912833264263, + "grad_norm": 0.016504082828760147, + "learning_rate": 8.617413816737302e-05, + "loss": 0.014, + "step": 962 + }, + { + "epoch": 0.13302942395358475, + "grad_norm": 0.03382395952939987, + "learning_rate": 8.61871722905094e-05, + "loss": 0.0266, + "step": 963 + }, + { + "epoch": 0.1331675645807432, + "grad_norm": 0.008775300346314907, + "learning_rate": 8.620019288575218e-05, + "loss": 0.01, + "step": 964 + }, + { + "epoch": 0.13330570520790164, + "grad_norm": 0.0034813378006219864, + "learning_rate": 8.621319998115303e-05, + "loss": 0.0044, + "step": 965 + }, + { + "epoch": 0.13344384583506008, + "grad_norm": 0.007844419218599796, + "learning_rate": 8.622619360467634e-05, + "loss": 0.0125, + "step": 966 + }, + { + "epoch": 0.13358198646221853, + "grad_norm": 0.0025214802008122206, + "learning_rate": 8.623917378419984e-05, + "loss": 0.0046, + "step": 967 + }, + { + "epoch": 0.13372012708937697, + "grad_norm": 0.02048412896692753, + "learning_rate": 8.625214054751471e-05, + "loss": 0.0141, + "step": 968 + }, + { + "epoch": 0.13385826771653545, + "grad_norm": 0.008773239329457283, + "learning_rate": 8.62650939223261e-05, + "loss": 0.0126, + "step": 969 + }, + { + "epoch": 0.1339964083436939, + "grad_norm": 0.07702983915805817, + "learning_rate": 8.627803393625343e-05, + "loss": 0.0595, + "step": 970 + }, + { + "epoch": 0.13413454897085234, + "grad_norm": 0.012115233577787876, + "learning_rate": 8.629096061683071e-05, + "loss": 0.0109, + "step": 971 + }, + { + "epoch": 0.13427268959801078, + "grad_norm": 0.013093401677906513, + "learning_rate": 8.630387399150694e-05, + "loss": 0.0094, + "step": 972 + }, + { + "epoch": 0.13441083022516923, + "grad_norm": 0.006639427505433559, + "learning_rate": 8.631677408764646e-05, + "loss": 0.0076, + "step": 973 + }, + { + "epoch": 0.13454897085232767, + "grad_norm": 0.04810924455523491, + "learning_rate": 8.632966093252927e-05, + "loss": 0.0368, + "step": 974 + }, + { + "epoch": 0.13468711147948612, + "grad_norm": 0.007319958880543709, + "learning_rate": 8.634253455335138e-05, + "loss": 0.0072, + "step": 975 + }, + { + "epoch": 0.13482525210664456, + "grad_norm": 0.05134734511375427, + "learning_rate": 8.635539497722518e-05, + "loss": 0.0367, + "step": 976 + }, + { + "epoch": 0.134963392733803, + "grad_norm": 0.010004381649196148, + "learning_rate": 8.636824223117976e-05, + "loss": 0.0104, + "step": 977 + }, + { + "epoch": 0.13510153336096145, + "grad_norm": 0.00346842291764915, + "learning_rate": 8.638107634216126e-05, + "loss": 0.0052, + "step": 978 + }, + { + "epoch": 0.1352396739881199, + "grad_norm": 0.01166492234915495, + "learning_rate": 8.639389733703317e-05, + "loss": 0.0085, + "step": 979 + }, + { + "epoch": 0.13537781461527834, + "grad_norm": 0.02950691431760788, + "learning_rate": 8.640670524257675e-05, + "loss": 0.0222, + "step": 980 + }, + { + "epoch": 0.13551595524243681, + "grad_norm": 0.010350341908633709, + "learning_rate": 8.641950008549133e-05, + "loss": 0.0077, + "step": 981 + }, + { + "epoch": 0.13565409586959526, + "grad_norm": 0.0018771348986774683, + "learning_rate": 8.643228189239453e-05, + "loss": 0.003, + "step": 982 + }, + { + "epoch": 0.1357922364967537, + "grad_norm": 0.012851797975599766, + "learning_rate": 8.644505068982283e-05, + "loss": 0.011, + "step": 983 + }, + { + "epoch": 0.13593037712391215, + "grad_norm": 0.004215499386191368, + "learning_rate": 8.645780650423166e-05, + "loss": 0.0054, + "step": 984 + }, + { + "epoch": 0.1360685177510706, + "grad_norm": 0.13906902074813843, + "learning_rate": 8.647054936199591e-05, + "loss": 0.0696, + "step": 985 + }, + { + "epoch": 0.13620665837822904, + "grad_norm": 0.007836158387362957, + "learning_rate": 8.648327928941014e-05, + "loss": 0.0095, + "step": 986 + }, + { + "epoch": 0.13634479900538748, + "grad_norm": 0.0017098481766879559, + "learning_rate": 8.649599631268893e-05, + "loss": 0.0028, + "step": 987 + }, + { + "epoch": 0.13648293963254593, + "grad_norm": 0.004099471494555473, + "learning_rate": 8.650870045796727e-05, + "loss": 0.0046, + "step": 988 + }, + { + "epoch": 0.13662108025970437, + "grad_norm": 0.0024917221162468195, + "learning_rate": 8.65213917513008e-05, + "loss": 0.0032, + "step": 989 + }, + { + "epoch": 0.13675922088686282, + "grad_norm": 0.006787384394556284, + "learning_rate": 8.653407021866616e-05, + "loss": 0.0052, + "step": 990 + }, + { + "epoch": 0.13689736151402127, + "grad_norm": 0.07070998102426529, + "learning_rate": 8.654673588596133e-05, + "loss": 0.0427, + "step": 991 + }, + { + "epoch": 0.1370355021411797, + "grad_norm": 0.02854343131184578, + "learning_rate": 8.655938877900593e-05, + "loss": 0.0225, + "step": 992 + }, + { + "epoch": 0.13717364276833816, + "grad_norm": 0.014134407043457031, + "learning_rate": 8.657202892354155e-05, + "loss": 0.0097, + "step": 993 + }, + { + "epoch": 0.13731178339549663, + "grad_norm": 0.006957308389246464, + "learning_rate": 8.658465634523199e-05, + "loss": 0.0075, + "step": 994 + }, + { + "epoch": 0.13744992402265507, + "grad_norm": 0.008567973971366882, + "learning_rate": 8.659727106966367e-05, + "loss": 0.0095, + "step": 995 + }, + { + "epoch": 0.13758806464981352, + "grad_norm": 0.010392943397164345, + "learning_rate": 8.660987312234595e-05, + "loss": 0.0079, + "step": 996 + }, + { + "epoch": 0.13772620527697196, + "grad_norm": 0.09674349427223206, + "learning_rate": 8.662246252871131e-05, + "loss": 0.0385, + "step": 997 + }, + { + "epoch": 0.1378643459041304, + "grad_norm": 0.002456868067383766, + "learning_rate": 8.663503931411578e-05, + "loss": 0.0032, + "step": 998 + }, + { + "epoch": 0.13800248653128885, + "grad_norm": 0.07229039818048477, + "learning_rate": 8.664760350383924e-05, + "loss": 0.0491, + "step": 999 + }, + { + "epoch": 0.1381406271584473, + "grad_norm": 0.00896090641617775, + "learning_rate": 8.666015512308563e-05, + "loss": 0.0083, + "step": 1000 + } + ], + "logging_steps": 1.0, + "max_steps": 28956, + "num_input_tokens_seen": 0, + "num_train_epochs": 4, + "save_steps": 1000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 0.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}