| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 4.999242079733212, |
| "eval_steps": 500, |
| "global_step": 1030, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.004850689707442777, |
| "grad_norm": 6.339065858846749, |
| "learning_rate": 7.766990291262136e-07, |
| "loss": 1.0147, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.009701379414885554, |
| "grad_norm": 6.370901874406726, |
| "learning_rate": 1.5533980582524272e-06, |
| "loss": 1.0217, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.01455206912232833, |
| "grad_norm": 6.253076167725343, |
| "learning_rate": 2.330097087378641e-06, |
| "loss": 1.0059, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.01940275882977111, |
| "grad_norm": 5.8627331903052005, |
| "learning_rate": 3.1067961165048544e-06, |
| "loss": 0.9987, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.024253448537213885, |
| "grad_norm": 4.624724616975588, |
| "learning_rate": 3.883495145631068e-06, |
| "loss": 0.9654, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.02910413824465666, |
| "grad_norm": 2.701973440331149, |
| "learning_rate": 4.660194174757282e-06, |
| "loss": 0.9221, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.03395482795209944, |
| "grad_norm": 2.5821400220833683, |
| "learning_rate": 5.436893203883496e-06, |
| "loss": 0.9118, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.03880551765954222, |
| "grad_norm": 3.1473551148693146, |
| "learning_rate": 6.213592233009709e-06, |
| "loss": 0.8818, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.04365620736698499, |
| "grad_norm": 3.8230150062051638, |
| "learning_rate": 6.990291262135923e-06, |
| "loss": 0.8912, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.04850689707442777, |
| "grad_norm": 3.4669236063777715, |
| "learning_rate": 7.766990291262136e-06, |
| "loss": 0.8779, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.053357586781870546, |
| "grad_norm": 2.712055876575345, |
| "learning_rate": 8.54368932038835e-06, |
| "loss": 0.8328, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.05820827648931332, |
| "grad_norm": 2.574263966000136, |
| "learning_rate": 9.320388349514565e-06, |
| "loss": 0.8149, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.0630589661967561, |
| "grad_norm": 1.8016385817876701, |
| "learning_rate": 1.0097087378640778e-05, |
| "loss": 0.7941, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.06790965590419888, |
| "grad_norm": 1.16110696712433, |
| "learning_rate": 1.0873786407766991e-05, |
| "loss": 0.7751, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.07276034561164166, |
| "grad_norm": 1.5441894400920566, |
| "learning_rate": 1.1650485436893204e-05, |
| "loss": 0.7603, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.07761103531908443, |
| "grad_norm": 1.3570174190036193, |
| "learning_rate": 1.2427184466019418e-05, |
| "loss": 0.7456, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.08246172502652721, |
| "grad_norm": 1.0160879152766609, |
| "learning_rate": 1.3203883495145633e-05, |
| "loss": 0.74, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.08731241473396999, |
| "grad_norm": 1.195923679791525, |
| "learning_rate": 1.3980582524271846e-05, |
| "loss": 0.7223, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.09216310444141276, |
| "grad_norm": 1.0381307779091873, |
| "learning_rate": 1.475728155339806e-05, |
| "loss": 0.7149, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.09701379414885554, |
| "grad_norm": 0.9191697728302082, |
| "learning_rate": 1.5533980582524273e-05, |
| "loss": 0.7032, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.10186448385629832, |
| "grad_norm": 1.0389109685950821, |
| "learning_rate": 1.6310679611650486e-05, |
| "loss": 0.697, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.10671517356374109, |
| "grad_norm": 0.8528569833940303, |
| "learning_rate": 1.70873786407767e-05, |
| "loss": 0.6913, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.11156586327118387, |
| "grad_norm": 0.7397574673832126, |
| "learning_rate": 1.7864077669902916e-05, |
| "loss": 0.6844, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.11641655297862664, |
| "grad_norm": 0.6762376097915315, |
| "learning_rate": 1.864077669902913e-05, |
| "loss": 0.6807, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.12126724268606942, |
| "grad_norm": 0.6801312007046909, |
| "learning_rate": 1.9417475728155343e-05, |
| "loss": 0.6651, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.1261179323935122, |
| "grad_norm": 0.5373129321939298, |
| "learning_rate": 2.0194174757281556e-05, |
| "loss": 0.6661, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.13096862210095497, |
| "grad_norm": 0.6155691125010336, |
| "learning_rate": 2.097087378640777e-05, |
| "loss": 0.6655, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.13581931180839776, |
| "grad_norm": 0.5373412410981904, |
| "learning_rate": 2.1747572815533982e-05, |
| "loss": 0.6541, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.14067000151584053, |
| "grad_norm": 0.5722094683121568, |
| "learning_rate": 2.2524271844660196e-05, |
| "loss": 0.6534, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.14552069122328332, |
| "grad_norm": 0.6369873796903149, |
| "learning_rate": 2.330097087378641e-05, |
| "loss": 0.6536, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.15037138093072608, |
| "grad_norm": 0.5246684440675834, |
| "learning_rate": 2.4077669902912622e-05, |
| "loss": 0.6545, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.15522207063816887, |
| "grad_norm": 0.504247506683658, |
| "learning_rate": 2.4854368932038836e-05, |
| "loss": 0.6314, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.16007276034561163, |
| "grad_norm": 0.5566944063536889, |
| "learning_rate": 2.5631067961165052e-05, |
| "loss": 0.6373, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.16492345005305442, |
| "grad_norm": 0.8570989167580252, |
| "learning_rate": 2.6407766990291266e-05, |
| "loss": 0.63, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.16977413976049718, |
| "grad_norm": 1.8667995731915865, |
| "learning_rate": 2.718446601941748e-05, |
| "loss": 0.6456, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.17462482946793997, |
| "grad_norm": 0.5164194413667431, |
| "learning_rate": 2.7961165048543692e-05, |
| "loss": 0.6308, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.17947551917538274, |
| "grad_norm": 1.6098330943831782, |
| "learning_rate": 2.8737864077669905e-05, |
| "loss": 0.6398, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.18432620888282553, |
| "grad_norm": 1.1492129920694993, |
| "learning_rate": 2.951456310679612e-05, |
| "loss": 0.6276, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.18917689859026832, |
| "grad_norm": 0.7127614761088336, |
| "learning_rate": 3.0291262135922332e-05, |
| "loss": 0.6218, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.19402758829771108, |
| "grad_norm": 1.0312378150228299, |
| "learning_rate": 3.1067961165048545e-05, |
| "loss": 0.6253, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.19887827800515387, |
| "grad_norm": 1.7574387168351864, |
| "learning_rate": 3.184466019417476e-05, |
| "loss": 0.6281, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.20372896771259663, |
| "grad_norm": 0.9416599187328968, |
| "learning_rate": 3.262135922330097e-05, |
| "loss": 0.6235, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.20857965742003942, |
| "grad_norm": 1.8301787236679616, |
| "learning_rate": 3.339805825242719e-05, |
| "loss": 0.622, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.21343034712748218, |
| "grad_norm": 0.9292078210446757, |
| "learning_rate": 3.41747572815534e-05, |
| "loss": 0.6107, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.21828103683492497, |
| "grad_norm": 1.994214362456412, |
| "learning_rate": 3.4951456310679615e-05, |
| "loss": 0.6151, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.22313172654236774, |
| "grad_norm": 1.3619344540131681, |
| "learning_rate": 3.572815533980583e-05, |
| "loss": 0.6089, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.22798241624981053, |
| "grad_norm": 1.6323821476629805, |
| "learning_rate": 3.650485436893204e-05, |
| "loss": 0.6165, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.2328331059572533, |
| "grad_norm": 1.1474300438640261, |
| "learning_rate": 3.728155339805826e-05, |
| "loss": 0.6104, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.23768379566469608, |
| "grad_norm": 1.1936059623728144, |
| "learning_rate": 3.805825242718447e-05, |
| "loss": 0.6086, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.24253448537213884, |
| "grad_norm": 1.4126137333521573, |
| "learning_rate": 3.8834951456310685e-05, |
| "loss": 0.6123, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.24738517507958163, |
| "grad_norm": 0.7835607330331523, |
| "learning_rate": 3.9611650485436895e-05, |
| "loss": 0.5979, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.2522358647870244, |
| "grad_norm": 1.3575257469759314, |
| "learning_rate": 4.038834951456311e-05, |
| "loss": 0.6134, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.25708655449446716, |
| "grad_norm": 0.9580505107131282, |
| "learning_rate": 4.116504854368932e-05, |
| "loss": 0.5987, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.26193724420190995, |
| "grad_norm": 1.1647956860260527, |
| "learning_rate": 4.194174757281554e-05, |
| "loss": 0.6134, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.26678793390935274, |
| "grad_norm": 0.9133078407230598, |
| "learning_rate": 4.271844660194175e-05, |
| "loss": 0.5978, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.2716386236167955, |
| "grad_norm": 1.3505836368180404, |
| "learning_rate": 4.3495145631067965e-05, |
| "loss": 0.5944, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.2764893133242383, |
| "grad_norm": 1.1628512424723636, |
| "learning_rate": 4.4271844660194175e-05, |
| "loss": 0.6052, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.28134000303168105, |
| "grad_norm": 1.6421048813051027, |
| "learning_rate": 4.504854368932039e-05, |
| "loss": 0.6008, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.28619069273912384, |
| "grad_norm": 1.3019802198524983, |
| "learning_rate": 4.58252427184466e-05, |
| "loss": 0.594, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.29104138244656663, |
| "grad_norm": 1.2774402772060065, |
| "learning_rate": 4.660194174757282e-05, |
| "loss": 0.5892, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.2958920721540094, |
| "grad_norm": 1.4435670386305743, |
| "learning_rate": 4.737864077669903e-05, |
| "loss": 0.5931, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.30074276186145216, |
| "grad_norm": 1.0284857540916943, |
| "learning_rate": 4.8155339805825245e-05, |
| "loss": 0.5932, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.30559345156889495, |
| "grad_norm": 1.1698103896183938, |
| "learning_rate": 4.8932038834951454e-05, |
| "loss": 0.5948, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.31044414127633774, |
| "grad_norm": 1.434142725222452, |
| "learning_rate": 4.970873786407767e-05, |
| "loss": 0.5897, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.3152948309837805, |
| "grad_norm": 1.3482179068151203, |
| "learning_rate": 5.0485436893203895e-05, |
| "loss": 0.5824, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.32014552069122326, |
| "grad_norm": 1.3069227496554443, |
| "learning_rate": 5.1262135922330105e-05, |
| "loss": 0.5802, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.32499621039866605, |
| "grad_norm": 1.1366236056516827, |
| "learning_rate": 5.203883495145632e-05, |
| "loss": 0.5804, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.32984690010610884, |
| "grad_norm": 1.6243350603336242, |
| "learning_rate": 5.281553398058253e-05, |
| "loss": 0.5812, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.33469758981355163, |
| "grad_norm": 1.0822457196908746, |
| "learning_rate": 5.359223300970875e-05, |
| "loss": 0.5862, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.33954827952099437, |
| "grad_norm": 1.6360957248140573, |
| "learning_rate": 5.436893203883496e-05, |
| "loss": 0.5912, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.34439896922843716, |
| "grad_norm": 2.3477845490488813, |
| "learning_rate": 5.5145631067961174e-05, |
| "loss": 0.5906, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.34924965893587995, |
| "grad_norm": 1.105543560593242, |
| "learning_rate": 5.5922330097087384e-05, |
| "loss": 0.5824, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.35410034864332274, |
| "grad_norm": 4.137098681881185, |
| "learning_rate": 5.66990291262136e-05, |
| "loss": 0.6359, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.3589510383507655, |
| "grad_norm": 3.8847979837997033, |
| "learning_rate": 5.747572815533981e-05, |
| "loss": 0.6486, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.36380172805820826, |
| "grad_norm": 1.1747128429519862, |
| "learning_rate": 5.825242718446603e-05, |
| "loss": 0.595, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.36865241776565105, |
| "grad_norm": 3.008245900701061, |
| "learning_rate": 5.902912621359224e-05, |
| "loss": 0.6387, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.37350310747309384, |
| "grad_norm": 2.3713075132931554, |
| "learning_rate": 5.9805825242718454e-05, |
| "loss": 0.6344, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.37835379718053663, |
| "grad_norm": 1.6213053074921984, |
| "learning_rate": 6.0582524271844664e-05, |
| "loss": 0.6048, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.38320448688797937, |
| "grad_norm": 1.6024154837501339, |
| "learning_rate": 6.135922330097087e-05, |
| "loss": 0.6203, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.38805517659542216, |
| "grad_norm": 1.1825030677591377, |
| "learning_rate": 6.213592233009709e-05, |
| "loss": 0.6052, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.39290586630286495, |
| "grad_norm": 1.3964525731128163, |
| "learning_rate": 6.291262135922331e-05, |
| "loss": 0.6112, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.39775655601030774, |
| "grad_norm": 1.1205074621871551, |
| "learning_rate": 6.368932038834952e-05, |
| "loss": 0.5977, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.4026072457177505, |
| "grad_norm": 1.045620374565707, |
| "learning_rate": 6.446601941747573e-05, |
| "loss": 0.5914, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.40745793542519326, |
| "grad_norm": 1.3974586249408472, |
| "learning_rate": 6.524271844660194e-05, |
| "loss": 0.5918, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.41230862513263605, |
| "grad_norm": 1.0818483302602913, |
| "learning_rate": 6.601941747572816e-05, |
| "loss": 0.5948, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.41715931484007884, |
| "grad_norm": 0.9808456957793906, |
| "learning_rate": 6.679611650485438e-05, |
| "loss": 0.5839, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.4220100045475216, |
| "grad_norm": 1.2035779456517084, |
| "learning_rate": 6.757281553398058e-05, |
| "loss": 0.5833, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.42686069425496437, |
| "grad_norm": 1.6887623926979713, |
| "learning_rate": 6.83495145631068e-05, |
| "loss": 0.5818, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.43171138396240716, |
| "grad_norm": 0.8023218391013366, |
| "learning_rate": 6.912621359223301e-05, |
| "loss": 0.5863, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.43656207366984995, |
| "grad_norm": 1.574900348178855, |
| "learning_rate": 6.990291262135923e-05, |
| "loss": 0.5821, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.4414127633772927, |
| "grad_norm": 0.9288518542917786, |
| "learning_rate": 7.067961165048545e-05, |
| "loss": 0.5814, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.4462634530847355, |
| "grad_norm": 1.6871845487045471, |
| "learning_rate": 7.145631067961166e-05, |
| "loss": 0.5819, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.45111414279217826, |
| "grad_norm": 1.147490028185953, |
| "learning_rate": 7.223300970873787e-05, |
| "loss": 0.5752, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.45596483249962105, |
| "grad_norm": 1.0734179177901382, |
| "learning_rate": 7.300970873786408e-05, |
| "loss": 0.5786, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.46081552220706384, |
| "grad_norm": 1.2263367009960806, |
| "learning_rate": 7.37864077669903e-05, |
| "loss": 0.5789, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.4656662119145066, |
| "grad_norm": 1.4570032389620742, |
| "learning_rate": 7.456310679611652e-05, |
| "loss": 0.5745, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.47051690162194937, |
| "grad_norm": 1.3246870726440927, |
| "learning_rate": 7.533980582524272e-05, |
| "loss": 0.5775, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.47536759132939216, |
| "grad_norm": 0.9415857506868542, |
| "learning_rate": 7.611650485436894e-05, |
| "loss": 0.5699, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.48021828103683495, |
| "grad_norm": 1.2384384474151087, |
| "learning_rate": 7.689320388349515e-05, |
| "loss": 0.5733, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.4850689707442777, |
| "grad_norm": 1.5627749991572353, |
| "learning_rate": 7.766990291262137e-05, |
| "loss": 0.5735, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.4899196604517205, |
| "grad_norm": 1.0078484211944914, |
| "learning_rate": 7.844660194174757e-05, |
| "loss": 0.5733, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.49477035015916326, |
| "grad_norm": 1.6421211712488573, |
| "learning_rate": 7.922330097087379e-05, |
| "loss": 0.576, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.49962103986660605, |
| "grad_norm": 0.8416126904816602, |
| "learning_rate": 8e-05, |
| "loss": 0.5697, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.5044717295740488, |
| "grad_norm": 1.7540458195736903, |
| "learning_rate": 7.999977029531286e-05, |
| "loss": 0.5799, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.5093224192814916, |
| "grad_norm": 1.050447231844734, |
| "learning_rate": 7.999908118388965e-05, |
| "loss": 0.5756, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.5141731089889343, |
| "grad_norm": 1.394330877960354, |
| "learning_rate": 7.999793267364497e-05, |
| "loss": 0.5713, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.5190237986963772, |
| "grad_norm": 1.5873678777006228, |
| "learning_rate": 7.999632477776974e-05, |
| "loss": 0.5733, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.5238744884038199, |
| "grad_norm": 0.8115732123781836, |
| "learning_rate": 7.9994257514731e-05, |
| "loss": 0.5661, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.5287251781112627, |
| "grad_norm": 1.1857754936411384, |
| "learning_rate": 7.999173090827177e-05, |
| "loss": 0.5719, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.5335758678187055, |
| "grad_norm": 0.8066157800150973, |
| "learning_rate": 7.998874498741072e-05, |
| "loss": 0.5695, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.5384265575261482, |
| "grad_norm": 1.4739854945603235, |
| "learning_rate": 7.998529978644183e-05, |
| "loss": 0.5712, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.543277247233591, |
| "grad_norm": 0.778875512416499, |
| "learning_rate": 7.998139534493407e-05, |
| "loss": 0.5609, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.5481279369410338, |
| "grad_norm": 0.9919152170469479, |
| "learning_rate": 7.997703170773084e-05, |
| "loss": 0.5648, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.5529786266484766, |
| "grad_norm": 1.3093959851041357, |
| "learning_rate": 7.997220892494955e-05, |
| "loss": 0.5757, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.5578293163559194, |
| "grad_norm": 0.9066117120369992, |
| "learning_rate": 7.996692705198097e-05, |
| "loss": 0.566, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.5626800060633621, |
| "grad_norm": 1.2498381901490132, |
| "learning_rate": 7.996118614948869e-05, |
| "loss": 0.5757, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.567530695770805, |
| "grad_norm": 1.0385189418340641, |
| "learning_rate": 7.995498628340827e-05, |
| "loss": 0.5697, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.5723813854782477, |
| "grad_norm": 1.3065564110858372, |
| "learning_rate": 7.994832752494667e-05, |
| "loss": 0.5672, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.5772320751856904, |
| "grad_norm": 0.9524061899396458, |
| "learning_rate": 7.994120995058127e-05, |
| "loss": 0.5624, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.5820827648931333, |
| "grad_norm": 1.018979775317657, |
| "learning_rate": 7.993363364205907e-05, |
| "loss": 0.554, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.586933454600576, |
| "grad_norm": 0.9276898951496916, |
| "learning_rate": 7.992559868639576e-05, |
| "loss": 0.556, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.5917841443080188, |
| "grad_norm": 1.1930439188526805, |
| "learning_rate": 7.99171051758747e-05, |
| "loss": 0.5526, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.5966348340154616, |
| "grad_norm": 0.6826863389197481, |
| "learning_rate": 7.990815320804583e-05, |
| "loss": 0.5609, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.6014855237229043, |
| "grad_norm": 0.8061980858751947, |
| "learning_rate": 7.98987428857246e-05, |
| "loss": 0.5586, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.6063362134303472, |
| "grad_norm": 0.9239948149603757, |
| "learning_rate": 7.988887431699079e-05, |
| "loss": 0.5507, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.6111869031377899, |
| "grad_norm": 1.0465957889843347, |
| "learning_rate": 7.987854761518719e-05, |
| "loss": 0.5568, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.6160375928452326, |
| "grad_norm": 0.8308130357408615, |
| "learning_rate": 7.986776289891842e-05, |
| "loss": 0.5591, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.6208882825526755, |
| "grad_norm": 1.0246744766037437, |
| "learning_rate": 7.985652029204946e-05, |
| "loss": 0.5563, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.6257389722601182, |
| "grad_norm": 1.3238612418839921, |
| "learning_rate": 7.984481992370429e-05, |
| "loss": 0.5491, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.630589661967561, |
| "grad_norm": 0.6948258976249133, |
| "learning_rate": 7.983266192826437e-05, |
| "loss": 0.5418, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.6354403516750038, |
| "grad_norm": 0.7288790333090353, |
| "learning_rate": 7.982004644536716e-05, |
| "loss": 0.5441, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.6402910413824465, |
| "grad_norm": 1.0943204811023435, |
| "learning_rate": 7.98069736199044e-05, |
| "loss": 0.5493, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.6451417310898894, |
| "grad_norm": 0.8706827515570799, |
| "learning_rate": 7.979344360202055e-05, |
| "loss": 0.5465, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.6499924207973321, |
| "grad_norm": 0.9127380693761118, |
| "learning_rate": 7.977945654711108e-05, |
| "loss": 0.5475, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.654843110504775, |
| "grad_norm": 0.9510188119086359, |
| "learning_rate": 7.976501261582056e-05, |
| "loss": 0.543, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.6596938002122177, |
| "grad_norm": 0.9350726421156861, |
| "learning_rate": 7.975011197404092e-05, |
| "loss": 0.5525, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.6645444899196604, |
| "grad_norm": 1.272258773056705, |
| "learning_rate": 7.973475479290956e-05, |
| "loss": 0.5518, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.6693951796271033, |
| "grad_norm": 0.9765265325518906, |
| "learning_rate": 7.971894124880727e-05, |
| "loss": 0.5417, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.674245869334546, |
| "grad_norm": 1.1917501609756302, |
| "learning_rate": 7.970267152335632e-05, |
| "loss": 0.5464, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.6790965590419887, |
| "grad_norm": 0.6719257463868904, |
| "learning_rate": 7.968594580341832e-05, |
| "loss": 0.544, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.6839472487494316, |
| "grad_norm": 0.7137208028607956, |
| "learning_rate": 7.966876428109209e-05, |
| "loss": 0.5351, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.6887979384568743, |
| "grad_norm": 0.7449688014358767, |
| "learning_rate": 7.965112715371144e-05, |
| "loss": 0.5397, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.6936486281643172, |
| "grad_norm": 0.5377305872108858, |
| "learning_rate": 7.96330346238429e-05, |
| "loss": 0.5346, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.6984993178717599, |
| "grad_norm": 0.5856757070627496, |
| "learning_rate": 7.961448689928341e-05, |
| "loss": 0.5395, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.7033500075792026, |
| "grad_norm": 0.6000325566823206, |
| "learning_rate": 7.959548419305796e-05, |
| "loss": 0.5447, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.7082006972866455, |
| "grad_norm": 0.7819963257560868, |
| "learning_rate": 7.957602672341707e-05, |
| "loss": 0.5364, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.7130513869940882, |
| "grad_norm": 1.1223449508846108, |
| "learning_rate": 7.955611471383433e-05, |
| "loss": 0.5381, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.717902076701531, |
| "grad_norm": 1.1418981667975974, |
| "learning_rate": 7.953574839300385e-05, |
| "loss": 0.5381, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.7227527664089738, |
| "grad_norm": 0.758286759296052, |
| "learning_rate": 7.95149279948376e-05, |
| "loss": 0.5398, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.7276034561164165, |
| "grad_norm": 0.7637204957772546, |
| "learning_rate": 7.949365375846271e-05, |
| "loss": 0.5386, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.7324541458238594, |
| "grad_norm": 0.6982030938329856, |
| "learning_rate": 7.94719259282188e-05, |
| "loss": 0.5328, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.7373048355313021, |
| "grad_norm": 0.7115887055025976, |
| "learning_rate": 7.944974475365506e-05, |
| "loss": 0.5406, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.7421555252387448, |
| "grad_norm": 0.6914212445412167, |
| "learning_rate": 7.94271104895275e-05, |
| "loss": 0.5375, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.7470062149461877, |
| "grad_norm": 0.6376946136665823, |
| "learning_rate": 7.940402339579596e-05, |
| "loss": 0.5322, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.7518569046536304, |
| "grad_norm": 0.662468788270689, |
| "learning_rate": 7.93804837376211e-05, |
| "loss": 0.5312, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.7567075943610733, |
| "grad_norm": 0.8116591959883654, |
| "learning_rate": 7.935649178536142e-05, |
| "loss": 0.5362, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.761558284068516, |
| "grad_norm": 1.2251954995336705, |
| "learning_rate": 7.93320478145701e-05, |
| "loss": 0.5454, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.7664089737759587, |
| "grad_norm": 1.1753293382340935, |
| "learning_rate": 7.93071521059919e-05, |
| "loss": 0.5369, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.7712596634834016, |
| "grad_norm": 0.5797209510428332, |
| "learning_rate": 7.928180494555983e-05, |
| "loss": 0.5255, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.7761103531908443, |
| "grad_norm": 0.9260629876609666, |
| "learning_rate": 7.925600662439201e-05, |
| "loss": 0.535, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.780961042898287, |
| "grad_norm": 1.2828815170548864, |
| "learning_rate": 7.922975743878817e-05, |
| "loss": 0.5293, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.7858117326057299, |
| "grad_norm": 0.5171701751512903, |
| "learning_rate": 7.92030576902264e-05, |
| "loss": 0.5298, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.7906624223131726, |
| "grad_norm": 0.9531992162184508, |
| "learning_rate": 7.917590768535952e-05, |
| "loss": 0.5315, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.7955131120206155, |
| "grad_norm": 1.1106333536493764, |
| "learning_rate": 7.914830773601173e-05, |
| "loss": 0.5279, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.8003638017280582, |
| "grad_norm": 0.45675301699833715, |
| "learning_rate": 7.912025815917489e-05, |
| "loss": 0.5372, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.805214491435501, |
| "grad_norm": 0.7189143844280501, |
| "learning_rate": 7.909175927700499e-05, |
| "loss": 0.535, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.8100651811429438, |
| "grad_norm": 0.8686823215812186, |
| "learning_rate": 7.906281141681839e-05, |
| "loss": 0.5333, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.8149158708503865, |
| "grad_norm": 0.588440050652225, |
| "learning_rate": 7.903341491108798e-05, |
| "loss": 0.5289, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.8197665605578294, |
| "grad_norm": 0.563873615495661, |
| "learning_rate": 7.900357009743958e-05, |
| "loss": 0.5331, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.8246172502652721, |
| "grad_norm": 0.545022127050129, |
| "learning_rate": 7.897327731864784e-05, |
| "loss": 0.5266, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.8294679399727148, |
| "grad_norm": 0.5592040867673563, |
| "learning_rate": 7.894253692263244e-05, |
| "loss": 0.522, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.8343186296801577, |
| "grad_norm": 0.6324827822327501, |
| "learning_rate": 7.891134926245402e-05, |
| "loss": 0.5297, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.8391693193876004, |
| "grad_norm": 0.8244822385641454, |
| "learning_rate": 7.887971469631016e-05, |
| "loss": 0.5319, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.8440200090950432, |
| "grad_norm": 1.2087031591715138, |
| "learning_rate": 7.884763358753129e-05, |
| "loss": 0.5408, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.848870698802486, |
| "grad_norm": 0.7427926592130248, |
| "learning_rate": 7.881510630457643e-05, |
| "loss": 0.5326, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.8537213885099287, |
| "grad_norm": 0.5074413186000275, |
| "learning_rate": 7.878213322102908e-05, |
| "loss": 0.5281, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.8585720782173716, |
| "grad_norm": 0.7191422226961306, |
| "learning_rate": 7.874871471559282e-05, |
| "loss": 0.5269, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.8634227679248143, |
| "grad_norm": 0.7351841845171684, |
| "learning_rate": 7.8714851172087e-05, |
| "loss": 0.5326, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.868273457632257, |
| "grad_norm": 0.661668481852005, |
| "learning_rate": 7.868054297944237e-05, |
| "loss": 0.5312, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.8731241473396999, |
| "grad_norm": 0.6257843185205204, |
| "learning_rate": 7.864579053169657e-05, |
| "loss": 0.5265, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.8779748370471426, |
| "grad_norm": 0.6272508186215946, |
| "learning_rate": 7.86105942279896e-05, |
| "loss": 0.5242, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.8828255267545854, |
| "grad_norm": 0.5859524441087289, |
| "learning_rate": 7.857495447255925e-05, |
| "loss": 0.5117, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.8876762164620282, |
| "grad_norm": 0.5314342493506371, |
| "learning_rate": 7.853887167473646e-05, |
| "loss": 0.5275, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.892526906169471, |
| "grad_norm": 0.5894396606474899, |
| "learning_rate": 7.850234624894064e-05, |
| "loss": 0.5236, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.8973775958769138, |
| "grad_norm": 0.7789309413273731, |
| "learning_rate": 7.846537861467485e-05, |
| "loss": 0.5269, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.9022282855843565, |
| "grad_norm": 0.9749399305462054, |
| "learning_rate": 7.842796919652104e-05, |
| "loss": 0.5177, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.9070789752917993, |
| "grad_norm": 0.8944031183061603, |
| "learning_rate": 7.839011842413514e-05, |
| "loss": 0.5236, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.9119296649992421, |
| "grad_norm": 0.6072136307213187, |
| "learning_rate": 7.835182673224212e-05, |
| "loss": 0.5237, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.9167803547066848, |
| "grad_norm": 0.44815995922956803, |
| "learning_rate": 7.831309456063107e-05, |
| "loss": 0.5193, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.9216310444141277, |
| "grad_norm": 0.5362363058315522, |
| "learning_rate": 7.827392235415005e-05, |
| "loss": 0.5242, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.9264817341215704, |
| "grad_norm": 0.5675209533538658, |
| "learning_rate": 7.823431056270103e-05, |
| "loss": 0.5223, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.9313324238290132, |
| "grad_norm": 0.5087214285921436, |
| "learning_rate": 7.81942596412347e-05, |
| "loss": 0.522, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.936183113536456, |
| "grad_norm": 0.46687413642954234, |
| "learning_rate": 7.815377004974532e-05, |
| "loss": 0.509, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.9410338032438987, |
| "grad_norm": 0.43127006261715695, |
| "learning_rate": 7.811284225326529e-05, |
| "loss": 0.522, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.9458844929513415, |
| "grad_norm": 0.43502529088952246, |
| "learning_rate": 7.807147672185996e-05, |
| "loss": 0.5258, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.9507351826587843, |
| "grad_norm": 0.5247431038277521, |
| "learning_rate": 7.802967393062219e-05, |
| "loss": 0.524, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.955585872366227, |
| "grad_norm": 0.6873590048364063, |
| "learning_rate": 7.798743435966676e-05, |
| "loss": 0.5227, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.9604365620736699, |
| "grad_norm": 0.8978763261543932, |
| "learning_rate": 7.794475849412512e-05, |
| "loss": 0.5143, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.9652872517811126, |
| "grad_norm": 0.8491329782406589, |
| "learning_rate": 7.790164682413954e-05, |
| "loss": 0.5186, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.9701379414885554, |
| "grad_norm": 0.5004070361840594, |
| "learning_rate": 7.785809984485765e-05, |
| "loss": 0.5185, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.9749886311959982, |
| "grad_norm": 0.4620691718653308, |
| "learning_rate": 7.781411805642675e-05, |
| "loss": 0.5179, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.979839320903441, |
| "grad_norm": 0.6078145800613444, |
| "learning_rate": 7.776970196398795e-05, |
| "loss": 0.5185, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.9846900106108837, |
| "grad_norm": 0.6744449960922595, |
| "learning_rate": 7.77248520776705e-05, |
| "loss": 0.5223, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.9895407003183265, |
| "grad_norm": 0.769858159356013, |
| "learning_rate": 7.767956891258585e-05, |
| "loss": 0.514, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.9943913900257693, |
| "grad_norm": 0.9703823352556248, |
| "learning_rate": 7.763385298882177e-05, |
| "loss": 0.5227, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.9992420797332121, |
| "grad_norm": 1.0790500827883502, |
| "learning_rate": 7.758770483143634e-05, |
| "loss": 0.5161, |
| "step": 206 |
| }, |
| { |
| "epoch": 1.0048506897074427, |
| "grad_norm": 0.979399858326228, |
| "learning_rate": 7.754112497045198e-05, |
| "loss": 0.5126, |
| "step": 207 |
| }, |
| { |
| "epoch": 1.0097013794148855, |
| "grad_norm": 0.8061955074549858, |
| "learning_rate": 7.749411394084931e-05, |
| "loss": 0.5093, |
| "step": 208 |
| }, |
| { |
| "epoch": 1.0145520691223284, |
| "grad_norm": 0.769561002371883, |
| "learning_rate": 7.744667228256102e-05, |
| "loss": 0.5129, |
| "step": 209 |
| }, |
| { |
| "epoch": 1.0194027588297712, |
| "grad_norm": 0.7563794216226459, |
| "learning_rate": 7.739880054046567e-05, |
| "loss": 0.504, |
| "step": 210 |
| }, |
| { |
| "epoch": 1.024253448537214, |
| "grad_norm": 0.4861095042439798, |
| "learning_rate": 7.735049926438143e-05, |
| "loss": 0.5008, |
| "step": 211 |
| }, |
| { |
| "epoch": 1.0291041382446566, |
| "grad_norm": 0.521763707621908, |
| "learning_rate": 7.730176900905978e-05, |
| "loss": 0.5039, |
| "step": 212 |
| }, |
| { |
| "epoch": 1.0339548279520994, |
| "grad_norm": 0.6847654097095645, |
| "learning_rate": 7.725261033417914e-05, |
| "loss": 0.4987, |
| "step": 213 |
| }, |
| { |
| "epoch": 1.0388055176595423, |
| "grad_norm": 0.45617035587869154, |
| "learning_rate": 7.720302380433838e-05, |
| "loss": 0.5082, |
| "step": 214 |
| }, |
| { |
| "epoch": 1.043656207366985, |
| "grad_norm": 0.5136841747695677, |
| "learning_rate": 7.715300998905045e-05, |
| "loss": 0.4903, |
| "step": 215 |
| }, |
| { |
| "epoch": 1.0485068970744278, |
| "grad_norm": 0.5201029165395914, |
| "learning_rate": 7.710256946273572e-05, |
| "loss": 0.5061, |
| "step": 216 |
| }, |
| { |
| "epoch": 1.0533575867818705, |
| "grad_norm": 0.5369451831716502, |
| "learning_rate": 7.705170280471546e-05, |
| "loss": 0.4923, |
| "step": 217 |
| }, |
| { |
| "epoch": 1.0582082764893133, |
| "grad_norm": 0.5292650633659572, |
| "learning_rate": 7.700041059920516e-05, |
| "loss": 0.4958, |
| "step": 218 |
| }, |
| { |
| "epoch": 1.0630589661967562, |
| "grad_norm": 0.38347803199441816, |
| "learning_rate": 7.694869343530781e-05, |
| "loss": 0.4949, |
| "step": 219 |
| }, |
| { |
| "epoch": 1.067909655904199, |
| "grad_norm": 0.42205008813365347, |
| "learning_rate": 7.689655190700719e-05, |
| "loss": 0.4958, |
| "step": 220 |
| }, |
| { |
| "epoch": 1.0727603456116417, |
| "grad_norm": 0.43326896920779806, |
| "learning_rate": 7.684398661316092e-05, |
| "loss": 0.5034, |
| "step": 221 |
| }, |
| { |
| "epoch": 1.0776110353190844, |
| "grad_norm": 0.3610486041319715, |
| "learning_rate": 7.679099815749377e-05, |
| "loss": 0.508, |
| "step": 222 |
| }, |
| { |
| "epoch": 1.0824617250265272, |
| "grad_norm": 0.4631910613487445, |
| "learning_rate": 7.673758714859052e-05, |
| "loss": 0.5012, |
| "step": 223 |
| }, |
| { |
| "epoch": 1.08731241473397, |
| "grad_norm": 0.5933537887890846, |
| "learning_rate": 7.668375419988918e-05, |
| "loss": 0.5003, |
| "step": 224 |
| }, |
| { |
| "epoch": 1.0921631044414128, |
| "grad_norm": 0.7536616463697116, |
| "learning_rate": 7.662949992967375e-05, |
| "loss": 0.4984, |
| "step": 225 |
| }, |
| { |
| "epoch": 1.0970137941488556, |
| "grad_norm": 0.8908783317227219, |
| "learning_rate": 7.657482496106725e-05, |
| "loss": 0.499, |
| "step": 226 |
| }, |
| { |
| "epoch": 1.1018644838562983, |
| "grad_norm": 0.8949365262495667, |
| "learning_rate": 7.651972992202449e-05, |
| "loss": 0.4964, |
| "step": 227 |
| }, |
| { |
| "epoch": 1.106715173563741, |
| "grad_norm": 0.7666080530207662, |
| "learning_rate": 7.646421544532492e-05, |
| "loss": 0.501, |
| "step": 228 |
| }, |
| { |
| "epoch": 1.1115658632711838, |
| "grad_norm": 0.5974548399149405, |
| "learning_rate": 7.640828216856532e-05, |
| "loss": 0.5019, |
| "step": 229 |
| }, |
| { |
| "epoch": 1.1164165529786267, |
| "grad_norm": 0.5208576739553936, |
| "learning_rate": 7.635193073415246e-05, |
| "loss": 0.4954, |
| "step": 230 |
| }, |
| { |
| "epoch": 1.1212672426860695, |
| "grad_norm": 0.4449218126901089, |
| "learning_rate": 7.62951617892958e-05, |
| "loss": 0.4966, |
| "step": 231 |
| }, |
| { |
| "epoch": 1.1261179323935122, |
| "grad_norm": 0.3413546355163314, |
| "learning_rate": 7.623797598599995e-05, |
| "loss": 0.4869, |
| "step": 232 |
| }, |
| { |
| "epoch": 1.130968622100955, |
| "grad_norm": 0.4021494906350838, |
| "learning_rate": 7.618037398105728e-05, |
| "loss": 0.4876, |
| "step": 233 |
| }, |
| { |
| "epoch": 1.1358193118083977, |
| "grad_norm": 0.5401093810240136, |
| "learning_rate": 7.612235643604031e-05, |
| "loss": 0.495, |
| "step": 234 |
| }, |
| { |
| "epoch": 1.1406700015158404, |
| "grad_norm": 0.6450843987922678, |
| "learning_rate": 7.606392401729415e-05, |
| "loss": 0.4953, |
| "step": 235 |
| }, |
| { |
| "epoch": 1.1455206912232834, |
| "grad_norm": 0.6413460171528268, |
| "learning_rate": 7.600507739592879e-05, |
| "loss": 0.4972, |
| "step": 236 |
| }, |
| { |
| "epoch": 1.150371380930726, |
| "grad_norm": 0.6276158556341717, |
| "learning_rate": 7.594581724781152e-05, |
| "loss": 0.4957, |
| "step": 237 |
| }, |
| { |
| "epoch": 1.1552220706381688, |
| "grad_norm": 0.7553018024510589, |
| "learning_rate": 7.588614425355898e-05, |
| "loss": 0.4955, |
| "step": 238 |
| }, |
| { |
| "epoch": 1.1600727603456116, |
| "grad_norm": 0.8968837770832118, |
| "learning_rate": 7.582605909852951e-05, |
| "loss": 0.4937, |
| "step": 239 |
| }, |
| { |
| "epoch": 1.1649234500530543, |
| "grad_norm": 0.7618870978199445, |
| "learning_rate": 7.576556247281522e-05, |
| "loss": 0.4969, |
| "step": 240 |
| }, |
| { |
| "epoch": 1.1697741397604973, |
| "grad_norm": 0.6531717186517569, |
| "learning_rate": 7.570465507123401e-05, |
| "loss": 0.5042, |
| "step": 241 |
| }, |
| { |
| "epoch": 1.17462482946794, |
| "grad_norm": 0.591515239054212, |
| "learning_rate": 7.564333759332167e-05, |
| "loss": 0.4906, |
| "step": 242 |
| }, |
| { |
| "epoch": 1.1794755191753827, |
| "grad_norm": 0.49797581923871925, |
| "learning_rate": 7.558161074332379e-05, |
| "loss": 0.4966, |
| "step": 243 |
| }, |
| { |
| "epoch": 1.1843262088828255, |
| "grad_norm": 0.5865910810734263, |
| "learning_rate": 7.551947523018774e-05, |
| "loss": 0.4997, |
| "step": 244 |
| }, |
| { |
| "epoch": 1.1891768985902682, |
| "grad_norm": 0.5003536808662635, |
| "learning_rate": 7.54569317675544e-05, |
| "loss": 0.4954, |
| "step": 245 |
| }, |
| { |
| "epoch": 1.1940275882977112, |
| "grad_norm": 0.4328856052217075, |
| "learning_rate": 7.539398107375015e-05, |
| "loss": 0.4979, |
| "step": 246 |
| }, |
| { |
| "epoch": 1.198878278005154, |
| "grad_norm": 0.5623661241326378, |
| "learning_rate": 7.533062387177843e-05, |
| "loss": 0.4982, |
| "step": 247 |
| }, |
| { |
| "epoch": 1.2037289677125966, |
| "grad_norm": 0.542587238982675, |
| "learning_rate": 7.526686088931156e-05, |
| "loss": 0.4991, |
| "step": 248 |
| }, |
| { |
| "epoch": 1.2085796574200394, |
| "grad_norm": 0.4949553127282243, |
| "learning_rate": 7.520269285868235e-05, |
| "loss": 0.4908, |
| "step": 249 |
| }, |
| { |
| "epoch": 1.213430347127482, |
| "grad_norm": 0.39917468694971237, |
| "learning_rate": 7.513812051687564e-05, |
| "loss": 0.4917, |
| "step": 250 |
| }, |
| { |
| "epoch": 1.218281036834925, |
| "grad_norm": 0.44831825611716425, |
| "learning_rate": 7.507314460551993e-05, |
| "loss": 0.4898, |
| "step": 251 |
| }, |
| { |
| "epoch": 1.2231317265423678, |
| "grad_norm": 0.4694728642442923, |
| "learning_rate": 7.500776587087878e-05, |
| "loss": 0.4929, |
| "step": 252 |
| }, |
| { |
| "epoch": 1.2279824162498105, |
| "grad_norm": 0.4548469944052975, |
| "learning_rate": 7.494198506384229e-05, |
| "loss": 0.4826, |
| "step": 253 |
| }, |
| { |
| "epoch": 1.2328331059572533, |
| "grad_norm": 0.6666754202357293, |
| "learning_rate": 7.487580293991844e-05, |
| "loss": 0.5021, |
| "step": 254 |
| }, |
| { |
| "epoch": 1.237683795664696, |
| "grad_norm": 0.6791995766586792, |
| "learning_rate": 7.480922025922443e-05, |
| "loss": 0.4974, |
| "step": 255 |
| }, |
| { |
| "epoch": 1.242534485372139, |
| "grad_norm": 0.48795487898431833, |
| "learning_rate": 7.474223778647796e-05, |
| "loss": 0.4934, |
| "step": 256 |
| }, |
| { |
| "epoch": 1.2473851750795817, |
| "grad_norm": 0.32007742081193336, |
| "learning_rate": 7.467485629098842e-05, |
| "loss": 0.4907, |
| "step": 257 |
| }, |
| { |
| "epoch": 1.2522358647870244, |
| "grad_norm": 0.2948779105297471, |
| "learning_rate": 7.460707654664807e-05, |
| "loss": 0.4974, |
| "step": 258 |
| }, |
| { |
| "epoch": 1.2570865544944672, |
| "grad_norm": 0.2954711132741566, |
| "learning_rate": 7.453889933192316e-05, |
| "loss": 0.4893, |
| "step": 259 |
| }, |
| { |
| "epoch": 1.26193724420191, |
| "grad_norm": 0.3234156565448516, |
| "learning_rate": 7.447032542984502e-05, |
| "loss": 0.4882, |
| "step": 260 |
| }, |
| { |
| "epoch": 1.2667879339093528, |
| "grad_norm": 0.37736907562809757, |
| "learning_rate": 7.440135562800093e-05, |
| "loss": 0.487, |
| "step": 261 |
| }, |
| { |
| "epoch": 1.2716386236167956, |
| "grad_norm": 0.4380964102542513, |
| "learning_rate": 7.433199071852526e-05, |
| "loss": 0.4965, |
| "step": 262 |
| }, |
| { |
| "epoch": 1.2764893133242383, |
| "grad_norm": 0.5377640268616285, |
| "learning_rate": 7.426223149809023e-05, |
| "loss": 0.4922, |
| "step": 263 |
| }, |
| { |
| "epoch": 1.281340003031681, |
| "grad_norm": 0.5447192325063156, |
| "learning_rate": 7.419207876789685e-05, |
| "loss": 0.4844, |
| "step": 264 |
| }, |
| { |
| "epoch": 1.2861906927391238, |
| "grad_norm": 0.5569387883590096, |
| "learning_rate": 7.412153333366567e-05, |
| "loss": 0.4887, |
| "step": 265 |
| }, |
| { |
| "epoch": 1.2910413824465667, |
| "grad_norm": 0.6127559785319273, |
| "learning_rate": 7.405059600562751e-05, |
| "loss": 0.4974, |
| "step": 266 |
| }, |
| { |
| "epoch": 1.2958920721540095, |
| "grad_norm": 0.5372861059475705, |
| "learning_rate": 7.397926759851425e-05, |
| "loss": 0.4946, |
| "step": 267 |
| }, |
| { |
| "epoch": 1.3007427618614522, |
| "grad_norm": 0.3715441206816088, |
| "learning_rate": 7.390754893154933e-05, |
| "loss": 0.4914, |
| "step": 268 |
| }, |
| { |
| "epoch": 1.305593451568895, |
| "grad_norm": 0.3468457624853246, |
| "learning_rate": 7.383544082843846e-05, |
| "loss": 0.4906, |
| "step": 269 |
| }, |
| { |
| "epoch": 1.3104441412763377, |
| "grad_norm": 0.37917435116867076, |
| "learning_rate": 7.376294411736009e-05, |
| "loss": 0.4877, |
| "step": 270 |
| }, |
| { |
| "epoch": 1.3152948309837806, |
| "grad_norm": 0.44133896758740837, |
| "learning_rate": 7.369005963095596e-05, |
| "loss": 0.4962, |
| "step": 271 |
| }, |
| { |
| "epoch": 1.3201455206912232, |
| "grad_norm": 0.5453886771483799, |
| "learning_rate": 7.361678820632145e-05, |
| "loss": 0.4918, |
| "step": 272 |
| }, |
| { |
| "epoch": 1.324996210398666, |
| "grad_norm": 0.6036321305780495, |
| "learning_rate": 7.354313068499607e-05, |
| "loss": 0.4892, |
| "step": 273 |
| }, |
| { |
| "epoch": 1.3298469001061088, |
| "grad_norm": 0.620838391242342, |
| "learning_rate": 7.346908791295369e-05, |
| "loss": 0.495, |
| "step": 274 |
| }, |
| { |
| "epoch": 1.3346975898135516, |
| "grad_norm": 0.5848284090531766, |
| "learning_rate": 7.339466074059292e-05, |
| "loss": 0.4862, |
| "step": 275 |
| }, |
| { |
| "epoch": 1.3395482795209943, |
| "grad_norm": 0.5915616847361488, |
| "learning_rate": 7.331985002272726e-05, |
| "loss": 0.4986, |
| "step": 276 |
| }, |
| { |
| "epoch": 1.344398969228437, |
| "grad_norm": 0.5871009859314463, |
| "learning_rate": 7.324465661857534e-05, |
| "loss": 0.4868, |
| "step": 277 |
| }, |
| { |
| "epoch": 1.34924965893588, |
| "grad_norm": 0.4325881770640798, |
| "learning_rate": 7.316908139175105e-05, |
| "loss": 0.4886, |
| "step": 278 |
| }, |
| { |
| "epoch": 1.3541003486433227, |
| "grad_norm": 0.342133338929163, |
| "learning_rate": 7.309312521025356e-05, |
| "loss": 0.4909, |
| "step": 279 |
| }, |
| { |
| "epoch": 1.3589510383507655, |
| "grad_norm": 0.4415991996544984, |
| "learning_rate": 7.301678894645742e-05, |
| "loss": 0.4915, |
| "step": 280 |
| }, |
| { |
| "epoch": 1.3638017280582082, |
| "grad_norm": 0.510599069072842, |
| "learning_rate": 7.294007347710251e-05, |
| "loss": 0.4935, |
| "step": 281 |
| }, |
| { |
| "epoch": 1.368652417765651, |
| "grad_norm": 0.5220646005978747, |
| "learning_rate": 7.286297968328397e-05, |
| "loss": 0.4834, |
| "step": 282 |
| }, |
| { |
| "epoch": 1.373503107473094, |
| "grad_norm": 0.4836929320122624, |
| "learning_rate": 7.27855084504421e-05, |
| "loss": 0.4953, |
| "step": 283 |
| }, |
| { |
| "epoch": 1.3783537971805366, |
| "grad_norm": 0.4678184328008887, |
| "learning_rate": 7.270766066835217e-05, |
| "loss": 0.4872, |
| "step": 284 |
| }, |
| { |
| "epoch": 1.3832044868879794, |
| "grad_norm": 0.4479819209137706, |
| "learning_rate": 7.262943723111419e-05, |
| "loss": 0.4916, |
| "step": 285 |
| }, |
| { |
| "epoch": 1.388055176595422, |
| "grad_norm": 0.44661460126650243, |
| "learning_rate": 7.255083903714266e-05, |
| "loss": 0.4866, |
| "step": 286 |
| }, |
| { |
| "epoch": 1.3929058663028648, |
| "grad_norm": 0.46700839086853646, |
| "learning_rate": 7.247186698915625e-05, |
| "loss": 0.4879, |
| "step": 287 |
| }, |
| { |
| "epoch": 1.3977565560103078, |
| "grad_norm": 0.5111356334340853, |
| "learning_rate": 7.239252199416749e-05, |
| "loss": 0.4812, |
| "step": 288 |
| }, |
| { |
| "epoch": 1.4026072457177505, |
| "grad_norm": 0.5198819509993116, |
| "learning_rate": 7.23128049634722e-05, |
| "loss": 0.4809, |
| "step": 289 |
| }, |
| { |
| "epoch": 1.4074579354251933, |
| "grad_norm": 0.5071119674882796, |
| "learning_rate": 7.223271681263916e-05, |
| "loss": 0.4839, |
| "step": 290 |
| }, |
| { |
| "epoch": 1.412308625132636, |
| "grad_norm": 0.4871736388487796, |
| "learning_rate": 7.215225846149957e-05, |
| "loss": 0.4899, |
| "step": 291 |
| }, |
| { |
| "epoch": 1.4171593148400787, |
| "grad_norm": 0.4644477177967194, |
| "learning_rate": 7.207143083413643e-05, |
| "loss": 0.4865, |
| "step": 292 |
| }, |
| { |
| "epoch": 1.4220100045475217, |
| "grad_norm": 0.5305010693129898, |
| "learning_rate": 7.1990234858874e-05, |
| "loss": 0.4876, |
| "step": 293 |
| }, |
| { |
| "epoch": 1.4268606942549644, |
| "grad_norm": 0.5381812012245146, |
| "learning_rate": 7.190867146826707e-05, |
| "loss": 0.4936, |
| "step": 294 |
| }, |
| { |
| "epoch": 1.4317113839624072, |
| "grad_norm": 0.42926452605809334, |
| "learning_rate": 7.182674159909031e-05, |
| "loss": 0.4845, |
| "step": 295 |
| }, |
| { |
| "epoch": 1.43656207366985, |
| "grad_norm": 0.42959757228135126, |
| "learning_rate": 7.174444619232745e-05, |
| "loss": 0.4952, |
| "step": 296 |
| }, |
| { |
| "epoch": 1.4414127633772926, |
| "grad_norm": 0.4994771109955089, |
| "learning_rate": 7.166178619316056e-05, |
| "loss": 0.4912, |
| "step": 297 |
| }, |
| { |
| "epoch": 1.4462634530847356, |
| "grad_norm": 0.4676921077663137, |
| "learning_rate": 7.157876255095906e-05, |
| "loss": 0.4875, |
| "step": 298 |
| }, |
| { |
| "epoch": 1.4511141427921783, |
| "grad_norm": 0.430209471088095, |
| "learning_rate": 7.149537621926895e-05, |
| "loss": 0.4862, |
| "step": 299 |
| }, |
| { |
| "epoch": 1.455964832499621, |
| "grad_norm": 0.5267214852499816, |
| "learning_rate": 7.14116281558018e-05, |
| "loss": 0.4879, |
| "step": 300 |
| }, |
| { |
| "epoch": 1.4608155222070638, |
| "grad_norm": 0.5169308603441447, |
| "learning_rate": 7.132751932242376e-05, |
| "loss": 0.4984, |
| "step": 301 |
| }, |
| { |
| "epoch": 1.4656662119145065, |
| "grad_norm": 0.3435076728807633, |
| "learning_rate": 7.124305068514444e-05, |
| "loss": 0.487, |
| "step": 302 |
| }, |
| { |
| "epoch": 1.4705169016219495, |
| "grad_norm": 0.2662958717194974, |
| "learning_rate": 7.1158223214106e-05, |
| "loss": 0.4878, |
| "step": 303 |
| }, |
| { |
| "epoch": 1.4753675913293922, |
| "grad_norm": 0.31770648802942325, |
| "learning_rate": 7.107303788357177e-05, |
| "loss": 0.4819, |
| "step": 304 |
| }, |
| { |
| "epoch": 1.480218281036835, |
| "grad_norm": 0.29599579651368213, |
| "learning_rate": 7.098749567191527e-05, |
| "loss": 0.4852, |
| "step": 305 |
| }, |
| { |
| "epoch": 1.4850689707442777, |
| "grad_norm": 0.3093179833213525, |
| "learning_rate": 7.090159756160886e-05, |
| "loss": 0.4877, |
| "step": 306 |
| }, |
| { |
| "epoch": 1.4899196604517204, |
| "grad_norm": 0.38971280696993216, |
| "learning_rate": 7.081534453921242e-05, |
| "loss": 0.4852, |
| "step": 307 |
| }, |
| { |
| "epoch": 1.4947703501591634, |
| "grad_norm": 0.3825128078756503, |
| "learning_rate": 7.072873759536217e-05, |
| "loss": 0.4913, |
| "step": 308 |
| }, |
| { |
| "epoch": 1.499621039866606, |
| "grad_norm": 0.2886624346515238, |
| "learning_rate": 7.064177772475912e-05, |
| "loss": 0.4798, |
| "step": 309 |
| }, |
| { |
| "epoch": 1.5044717295740488, |
| "grad_norm": 0.3712678777863381, |
| "learning_rate": 7.05544659261578e-05, |
| "loss": 0.4867, |
| "step": 310 |
| }, |
| { |
| "epoch": 1.5093224192814916, |
| "grad_norm": 0.42331161544955054, |
| "learning_rate": 7.046680320235466e-05, |
| "loss": 0.4871, |
| "step": 311 |
| }, |
| { |
| "epoch": 1.5141731089889343, |
| "grad_norm": 0.4018226592697959, |
| "learning_rate": 7.037879056017663e-05, |
| "loss": 0.4842, |
| "step": 312 |
| }, |
| { |
| "epoch": 1.5190237986963773, |
| "grad_norm": 0.3959692369182176, |
| "learning_rate": 7.029042901046952e-05, |
| "loss": 0.4802, |
| "step": 313 |
| }, |
| { |
| "epoch": 1.5238744884038198, |
| "grad_norm": 0.4052319536586186, |
| "learning_rate": 7.020171956808645e-05, |
| "loss": 0.4859, |
| "step": 314 |
| }, |
| { |
| "epoch": 1.5287251781112627, |
| "grad_norm": 0.40673400305291324, |
| "learning_rate": 7.011266325187615e-05, |
| "loss": 0.496, |
| "step": 315 |
| }, |
| { |
| "epoch": 1.5335758678187055, |
| "grad_norm": 0.44258718402573904, |
| "learning_rate": 7.002326108467129e-05, |
| "loss": 0.4864, |
| "step": 316 |
| }, |
| { |
| "epoch": 1.5384265575261482, |
| "grad_norm": 0.4022422201093314, |
| "learning_rate": 6.993351409327672e-05, |
| "loss": 0.4763, |
| "step": 317 |
| }, |
| { |
| "epoch": 1.5432772472335912, |
| "grad_norm": 0.39107692009497685, |
| "learning_rate": 6.984342330845764e-05, |
| "loss": 0.4952, |
| "step": 318 |
| }, |
| { |
| "epoch": 1.5481279369410337, |
| "grad_norm": 0.36401667233363455, |
| "learning_rate": 6.975298976492785e-05, |
| "loss": 0.4952, |
| "step": 319 |
| }, |
| { |
| "epoch": 1.5529786266484766, |
| "grad_norm": 0.3964209144895704, |
| "learning_rate": 6.966221450133779e-05, |
| "loss": 0.4901, |
| "step": 320 |
| }, |
| { |
| "epoch": 1.5578293163559194, |
| "grad_norm": 0.41002635948711413, |
| "learning_rate": 6.957109856026261e-05, |
| "loss": 0.4917, |
| "step": 321 |
| }, |
| { |
| "epoch": 1.562680006063362, |
| "grad_norm": 0.33043042887382146, |
| "learning_rate": 6.94796429881903e-05, |
| "loss": 0.4771, |
| "step": 322 |
| }, |
| { |
| "epoch": 1.567530695770805, |
| "grad_norm": 0.3258167342621945, |
| "learning_rate": 6.938784883550948e-05, |
| "loss": 0.4889, |
| "step": 323 |
| }, |
| { |
| "epoch": 1.5723813854782476, |
| "grad_norm": 0.33864970777234993, |
| "learning_rate": 6.929571715649755e-05, |
| "loss": 0.4866, |
| "step": 324 |
| }, |
| { |
| "epoch": 1.5772320751856905, |
| "grad_norm": 0.3799335436470155, |
| "learning_rate": 6.920324900930842e-05, |
| "loss": 0.4907, |
| "step": 325 |
| }, |
| { |
| "epoch": 1.5820827648931333, |
| "grad_norm": 0.38649667346458383, |
| "learning_rate": 6.911044545596042e-05, |
| "loss": 0.4854, |
| "step": 326 |
| }, |
| { |
| "epoch": 1.586933454600576, |
| "grad_norm": 0.3332457174533195, |
| "learning_rate": 6.901730756232411e-05, |
| "loss": 0.4895, |
| "step": 327 |
| }, |
| { |
| "epoch": 1.591784144308019, |
| "grad_norm": 0.3688581464993942, |
| "learning_rate": 6.892383639811005e-05, |
| "loss": 0.4958, |
| "step": 328 |
| }, |
| { |
| "epoch": 1.5966348340154615, |
| "grad_norm": 0.43127701691860393, |
| "learning_rate": 6.883003303685644e-05, |
| "loss": 0.4844, |
| "step": 329 |
| }, |
| { |
| "epoch": 1.6014855237229044, |
| "grad_norm": 0.5098788849460419, |
| "learning_rate": 6.87358985559169e-05, |
| "loss": 0.489, |
| "step": 330 |
| }, |
| { |
| "epoch": 1.6063362134303472, |
| "grad_norm": 0.531466303384909, |
| "learning_rate": 6.864143403644797e-05, |
| "loss": 0.4945, |
| "step": 331 |
| }, |
| { |
| "epoch": 1.61118690313779, |
| "grad_norm": 0.4575057116450561, |
| "learning_rate": 6.85466405633968e-05, |
| "loss": 0.4855, |
| "step": 332 |
| }, |
| { |
| "epoch": 1.6160375928452326, |
| "grad_norm": 0.42418027914564915, |
| "learning_rate": 6.845151922548865e-05, |
| "loss": 0.4783, |
| "step": 333 |
| }, |
| { |
| "epoch": 1.6208882825526754, |
| "grad_norm": 0.38431703300530295, |
| "learning_rate": 6.835607111521439e-05, |
| "loss": 0.4796, |
| "step": 334 |
| }, |
| { |
| "epoch": 1.6257389722601183, |
| "grad_norm": 0.325044588754403, |
| "learning_rate": 6.826029732881793e-05, |
| "loss": 0.4928, |
| "step": 335 |
| }, |
| { |
| "epoch": 1.630589661967561, |
| "grad_norm": 0.2723847494193817, |
| "learning_rate": 6.816419896628363e-05, |
| "loss": 0.4851, |
| "step": 336 |
| }, |
| { |
| "epoch": 1.6354403516750038, |
| "grad_norm": 0.25325170865464947, |
| "learning_rate": 6.806777713132374e-05, |
| "loss": 0.4826, |
| "step": 337 |
| }, |
| { |
| "epoch": 1.6402910413824465, |
| "grad_norm": 0.2793387163645126, |
| "learning_rate": 6.79710329313656e-05, |
| "loss": 0.4873, |
| "step": 338 |
| }, |
| { |
| "epoch": 1.6451417310898893, |
| "grad_norm": 0.38835730960557174, |
| "learning_rate": 6.787396747753903e-05, |
| "loss": 0.4744, |
| "step": 339 |
| }, |
| { |
| "epoch": 1.6499924207973322, |
| "grad_norm": 0.5896470389457479, |
| "learning_rate": 6.777658188466354e-05, |
| "loss": 0.4765, |
| "step": 340 |
| }, |
| { |
| "epoch": 1.654843110504775, |
| "grad_norm": 0.7717199781637745, |
| "learning_rate": 6.767887727123544e-05, |
| "loss": 0.4931, |
| "step": 341 |
| }, |
| { |
| "epoch": 1.6596938002122177, |
| "grad_norm": 1.0491183054565791, |
| "learning_rate": 6.758085475941516e-05, |
| "loss": 0.4875, |
| "step": 342 |
| }, |
| { |
| "epoch": 1.6645444899196604, |
| "grad_norm": 1.0963514331569248, |
| "learning_rate": 6.748251547501418e-05, |
| "loss": 0.4783, |
| "step": 343 |
| }, |
| { |
| "epoch": 1.6693951796271032, |
| "grad_norm": 0.6008540810802777, |
| "learning_rate": 6.738386054748226e-05, |
| "loss": 0.4836, |
| "step": 344 |
| }, |
| { |
| "epoch": 1.674245869334546, |
| "grad_norm": 0.4448291346545642, |
| "learning_rate": 6.728489110989434e-05, |
| "loss": 0.4883, |
| "step": 345 |
| }, |
| { |
| "epoch": 1.6790965590419886, |
| "grad_norm": 0.4942617693141126, |
| "learning_rate": 6.718560829893762e-05, |
| "loss": 0.4799, |
| "step": 346 |
| }, |
| { |
| "epoch": 1.6839472487494316, |
| "grad_norm": 0.4996733158915889, |
| "learning_rate": 6.708601325489844e-05, |
| "loss": 0.4872, |
| "step": 347 |
| }, |
| { |
| "epoch": 1.6887979384568743, |
| "grad_norm": 0.3715020675799419, |
| "learning_rate": 6.698610712164924e-05, |
| "loss": 0.4864, |
| "step": 348 |
| }, |
| { |
| "epoch": 1.693648628164317, |
| "grad_norm": 0.27977348496462506, |
| "learning_rate": 6.688589104663536e-05, |
| "loss": 0.4731, |
| "step": 349 |
| }, |
| { |
| "epoch": 1.69849931787176, |
| "grad_norm": 0.3329757061053683, |
| "learning_rate": 6.67853661808619e-05, |
| "loss": 0.4771, |
| "step": 350 |
| }, |
| { |
| "epoch": 1.7033500075792025, |
| "grad_norm": 0.35849519050184514, |
| "learning_rate": 6.668453367888052e-05, |
| "loss": 0.4867, |
| "step": 351 |
| }, |
| { |
| "epoch": 1.7082006972866455, |
| "grad_norm": 0.3093064728603477, |
| "learning_rate": 6.658339469877613e-05, |
| "loss": 0.478, |
| "step": 352 |
| }, |
| { |
| "epoch": 1.7130513869940882, |
| "grad_norm": 0.2983932071893944, |
| "learning_rate": 6.64819504021536e-05, |
| "loss": 0.4814, |
| "step": 353 |
| }, |
| { |
| "epoch": 1.717902076701531, |
| "grad_norm": 0.32895472335219694, |
| "learning_rate": 6.638020195412448e-05, |
| "loss": 0.4771, |
| "step": 354 |
| }, |
| { |
| "epoch": 1.722752766408974, |
| "grad_norm": 0.4277599821762847, |
| "learning_rate": 6.627815052329354e-05, |
| "loss": 0.4925, |
| "step": 355 |
| }, |
| { |
| "epoch": 1.7276034561164164, |
| "grad_norm": 0.4690652488456737, |
| "learning_rate": 6.617579728174535e-05, |
| "loss": 0.4854, |
| "step": 356 |
| }, |
| { |
| "epoch": 1.7324541458238594, |
| "grad_norm": 0.43487992599318925, |
| "learning_rate": 6.60731434050309e-05, |
| "loss": 0.4777, |
| "step": 357 |
| }, |
| { |
| "epoch": 1.737304835531302, |
| "grad_norm": 0.4398837483674929, |
| "learning_rate": 6.597019007215401e-05, |
| "loss": 0.4783, |
| "step": 358 |
| }, |
| { |
| "epoch": 1.7421555252387448, |
| "grad_norm": 0.43050412759021583, |
| "learning_rate": 6.586693846555788e-05, |
| "loss": 0.4743, |
| "step": 359 |
| }, |
| { |
| "epoch": 1.7470062149461878, |
| "grad_norm": 0.40202174280112624, |
| "learning_rate": 6.576338977111134e-05, |
| "loss": 0.48, |
| "step": 360 |
| }, |
| { |
| "epoch": 1.7518569046536303, |
| "grad_norm": 0.36259944082265505, |
| "learning_rate": 6.565954517809543e-05, |
| "loss": 0.4747, |
| "step": 361 |
| }, |
| { |
| "epoch": 1.7567075943610733, |
| "grad_norm": 0.28300782787905476, |
| "learning_rate": 6.555540587918968e-05, |
| "loss": 0.4778, |
| "step": 362 |
| }, |
| { |
| "epoch": 1.761558284068516, |
| "grad_norm": 0.27286805994349533, |
| "learning_rate": 6.545097307045831e-05, |
| "loss": 0.4795, |
| "step": 363 |
| }, |
| { |
| "epoch": 1.7664089737759587, |
| "grad_norm": 0.31934491384929364, |
| "learning_rate": 6.534624795133662e-05, |
| "loss": 0.4851, |
| "step": 364 |
| }, |
| { |
| "epoch": 1.7712596634834017, |
| "grad_norm": 0.35245692182661065, |
| "learning_rate": 6.524123172461711e-05, |
| "loss": 0.4794, |
| "step": 365 |
| }, |
| { |
| "epoch": 1.7761103531908442, |
| "grad_norm": 0.3396944161317505, |
| "learning_rate": 6.51359255964358e-05, |
| "loss": 0.4774, |
| "step": 366 |
| }, |
| { |
| "epoch": 1.7809610428982872, |
| "grad_norm": 0.2877754553279699, |
| "learning_rate": 6.503033077625824e-05, |
| "loss": 0.4746, |
| "step": 367 |
| }, |
| { |
| "epoch": 1.78581173260573, |
| "grad_norm": 0.3326775146217252, |
| "learning_rate": 6.492444847686566e-05, |
| "loss": 0.4849, |
| "step": 368 |
| }, |
| { |
| "epoch": 1.7906624223131726, |
| "grad_norm": 0.419702080624426, |
| "learning_rate": 6.481827991434111e-05, |
| "loss": 0.4814, |
| "step": 369 |
| }, |
| { |
| "epoch": 1.7955131120206156, |
| "grad_norm": 0.43780909545325103, |
| "learning_rate": 6.471182630805538e-05, |
| "loss": 0.4813, |
| "step": 370 |
| }, |
| { |
| "epoch": 1.800363801728058, |
| "grad_norm": 0.5080371772111751, |
| "learning_rate": 6.460508888065314e-05, |
| "loss": 0.4865, |
| "step": 371 |
| }, |
| { |
| "epoch": 1.805214491435501, |
| "grad_norm": 0.5541771901573003, |
| "learning_rate": 6.449806885803873e-05, |
| "loss": 0.4752, |
| "step": 372 |
| }, |
| { |
| "epoch": 1.8100651811429438, |
| "grad_norm": 0.5077930985424878, |
| "learning_rate": 6.439076746936219e-05, |
| "loss": 0.4776, |
| "step": 373 |
| }, |
| { |
| "epoch": 1.8149158708503865, |
| "grad_norm": 0.3974055597915673, |
| "learning_rate": 6.428318594700509e-05, |
| "loss": 0.4833, |
| "step": 374 |
| }, |
| { |
| "epoch": 1.8197665605578295, |
| "grad_norm": 0.27838793632255576, |
| "learning_rate": 6.417532552656647e-05, |
| "loss": 0.4808, |
| "step": 375 |
| }, |
| { |
| "epoch": 1.824617250265272, |
| "grad_norm": 0.25531453372215107, |
| "learning_rate": 6.406718744684851e-05, |
| "loss": 0.475, |
| "step": 376 |
| }, |
| { |
| "epoch": 1.829467939972715, |
| "grad_norm": 0.37619306197372576, |
| "learning_rate": 6.395877294984241e-05, |
| "loss": 0.4718, |
| "step": 377 |
| }, |
| { |
| "epoch": 1.8343186296801577, |
| "grad_norm": 0.45314622671289495, |
| "learning_rate": 6.385008328071406e-05, |
| "loss": 0.4858, |
| "step": 378 |
| }, |
| { |
| "epoch": 1.8391693193876004, |
| "grad_norm": 0.4091697946782105, |
| "learning_rate": 6.374111968778982e-05, |
| "loss": 0.4797, |
| "step": 379 |
| }, |
| { |
| "epoch": 1.8440200090950432, |
| "grad_norm": 0.31229108750141443, |
| "learning_rate": 6.363188342254206e-05, |
| "loss": 0.4819, |
| "step": 380 |
| }, |
| { |
| "epoch": 1.848870698802486, |
| "grad_norm": 0.29389799920114323, |
| "learning_rate": 6.352237573957488e-05, |
| "loss": 0.4777, |
| "step": 381 |
| }, |
| { |
| "epoch": 1.8537213885099288, |
| "grad_norm": 0.31830405845533455, |
| "learning_rate": 6.341259789660969e-05, |
| "loss": 0.4812, |
| "step": 382 |
| }, |
| { |
| "epoch": 1.8585720782173716, |
| "grad_norm": 0.2783419658252336, |
| "learning_rate": 6.330255115447076e-05, |
| "loss": 0.4722, |
| "step": 383 |
| }, |
| { |
| "epoch": 1.8634227679248143, |
| "grad_norm": 0.252195902565345, |
| "learning_rate": 6.319223677707069e-05, |
| "loss": 0.4786, |
| "step": 384 |
| }, |
| { |
| "epoch": 1.868273457632257, |
| "grad_norm": 0.24931429544692238, |
| "learning_rate": 6.308165603139598e-05, |
| "loss": 0.4766, |
| "step": 385 |
| }, |
| { |
| "epoch": 1.8731241473396998, |
| "grad_norm": 0.343056469344131, |
| "learning_rate": 6.29708101874924e-05, |
| "loss": 0.4746, |
| "step": 386 |
| }, |
| { |
| "epoch": 1.8779748370471427, |
| "grad_norm": 0.39163026769861153, |
| "learning_rate": 6.285970051845045e-05, |
| "loss": 0.4777, |
| "step": 387 |
| }, |
| { |
| "epoch": 1.8828255267545853, |
| "grad_norm": 0.3925083138992103, |
| "learning_rate": 6.274832830039071e-05, |
| "loss": 0.4762, |
| "step": 388 |
| }, |
| { |
| "epoch": 1.8876762164620282, |
| "grad_norm": 0.3908015146094751, |
| "learning_rate": 6.26366948124492e-05, |
| "loss": 0.4882, |
| "step": 389 |
| }, |
| { |
| "epoch": 1.892526906169471, |
| "grad_norm": 0.38768255703534454, |
| "learning_rate": 6.25248013367627e-05, |
| "loss": 0.4746, |
| "step": 390 |
| }, |
| { |
| "epoch": 1.8973775958769137, |
| "grad_norm": 0.34797592572865116, |
| "learning_rate": 6.241264915845401e-05, |
| "loss": 0.4863, |
| "step": 391 |
| }, |
| { |
| "epoch": 1.9022282855843566, |
| "grad_norm": 0.2939167601152598, |
| "learning_rate": 6.230023956561716e-05, |
| "loss": 0.4803, |
| "step": 392 |
| }, |
| { |
| "epoch": 1.9070789752917991, |
| "grad_norm": 0.2683598696324213, |
| "learning_rate": 6.218757384930268e-05, |
| "loss": 0.4769, |
| "step": 393 |
| }, |
| { |
| "epoch": 1.911929664999242, |
| "grad_norm": 0.2963955328357841, |
| "learning_rate": 6.207465330350273e-05, |
| "loss": 0.4798, |
| "step": 394 |
| }, |
| { |
| "epoch": 1.9167803547066848, |
| "grad_norm": 0.34936564317737695, |
| "learning_rate": 6.196147922513623e-05, |
| "loss": 0.479, |
| "step": 395 |
| }, |
| { |
| "epoch": 1.9216310444141276, |
| "grad_norm": 0.3023947037005716, |
| "learning_rate": 6.184805291403402e-05, |
| "loss": 0.4776, |
| "step": 396 |
| }, |
| { |
| "epoch": 1.9264817341215705, |
| "grad_norm": 0.23472105439779495, |
| "learning_rate": 6.173437567292383e-05, |
| "loss": 0.475, |
| "step": 397 |
| }, |
| { |
| "epoch": 1.931332423829013, |
| "grad_norm": 0.30151702935680424, |
| "learning_rate": 6.162044880741544e-05, |
| "loss": 0.4719, |
| "step": 398 |
| }, |
| { |
| "epoch": 1.936183113536456, |
| "grad_norm": 0.4380813843564537, |
| "learning_rate": 6.150627362598557e-05, |
| "loss": 0.4871, |
| "step": 399 |
| }, |
| { |
| "epoch": 1.9410338032438987, |
| "grad_norm": 0.5034940901090467, |
| "learning_rate": 6.139185143996298e-05, |
| "loss": 0.4806, |
| "step": 400 |
| }, |
| { |
| "epoch": 1.9458844929513415, |
| "grad_norm": 0.44831770290541656, |
| "learning_rate": 6.127718356351326e-05, |
| "loss": 0.478, |
| "step": 401 |
| }, |
| { |
| "epoch": 1.9507351826587844, |
| "grad_norm": 0.3881020850233725, |
| "learning_rate": 6.116227131362385e-05, |
| "loss": 0.4714, |
| "step": 402 |
| }, |
| { |
| "epoch": 1.955585872366227, |
| "grad_norm": 0.3382903208345561, |
| "learning_rate": 6.104711601008888e-05, |
| "loss": 0.4779, |
| "step": 403 |
| }, |
| { |
| "epoch": 1.96043656207367, |
| "grad_norm": 0.2830781275191087, |
| "learning_rate": 6.0931718975493985e-05, |
| "loss": 0.4846, |
| "step": 404 |
| }, |
| { |
| "epoch": 1.9652872517811126, |
| "grad_norm": 0.27767803820547865, |
| "learning_rate": 6.081608153520117e-05, |
| "loss": 0.4691, |
| "step": 405 |
| }, |
| { |
| "epoch": 1.9701379414885554, |
| "grad_norm": 0.35527234014372044, |
| "learning_rate": 6.0700205017333525e-05, |
| "loss": 0.4787, |
| "step": 406 |
| }, |
| { |
| "epoch": 1.9749886311959983, |
| "grad_norm": 0.34300192919407774, |
| "learning_rate": 6.058409075276002e-05, |
| "loss": 0.4689, |
| "step": 407 |
| }, |
| { |
| "epoch": 1.9798393209034408, |
| "grad_norm": 0.2964726842271146, |
| "learning_rate": 6.046774007508019e-05, |
| "loss": 0.475, |
| "step": 408 |
| }, |
| { |
| "epoch": 1.9846900106108838, |
| "grad_norm": 0.2502113324820329, |
| "learning_rate": 6.035115432060883e-05, |
| "loss": 0.4747, |
| "step": 409 |
| }, |
| { |
| "epoch": 1.9895407003183265, |
| "grad_norm": 0.23471156048166733, |
| "learning_rate": 6.0234334828360655e-05, |
| "loss": 0.4786, |
| "step": 410 |
| }, |
| { |
| "epoch": 1.9943913900257693, |
| "grad_norm": 0.27137410563019304, |
| "learning_rate": 6.011728294003494e-05, |
| "loss": 0.4802, |
| "step": 411 |
| }, |
| { |
| "epoch": 1.9992420797332122, |
| "grad_norm": 0.2960970008273601, |
| "learning_rate": 6.000000000000001e-05, |
| "loss": 0.4768, |
| "step": 412 |
| }, |
| { |
| "epoch": 2.004850689707443, |
| "grad_norm": 0.33644070901215145, |
| "learning_rate": 5.988248735527793e-05, |
| "loss": 0.4473, |
| "step": 413 |
| }, |
| { |
| "epoch": 2.0097013794148855, |
| "grad_norm": 0.39546749871672404, |
| "learning_rate": 5.9764746355528994e-05, |
| "loss": 0.4501, |
| "step": 414 |
| }, |
| { |
| "epoch": 2.0145520691223284, |
| "grad_norm": 0.47967552460987467, |
| "learning_rate": 5.964677835303615e-05, |
| "loss": 0.4483, |
| "step": 415 |
| }, |
| { |
| "epoch": 2.019402758829771, |
| "grad_norm": 0.43081443925289625, |
| "learning_rate": 5.952858470268955e-05, |
| "loss": 0.4468, |
| "step": 416 |
| }, |
| { |
| "epoch": 2.024253448537214, |
| "grad_norm": 0.42692273173821377, |
| "learning_rate": 5.941016676197098e-05, |
| "loss": 0.4499, |
| "step": 417 |
| }, |
| { |
| "epoch": 2.029104138244657, |
| "grad_norm": 0.5203095933335755, |
| "learning_rate": 5.929152589093825e-05, |
| "loss": 0.4498, |
| "step": 418 |
| }, |
| { |
| "epoch": 2.0339548279520994, |
| "grad_norm": 0.5390774215298352, |
| "learning_rate": 5.9172663452209554e-05, |
| "loss": 0.449, |
| "step": 419 |
| }, |
| { |
| "epoch": 2.0388055176595423, |
| "grad_norm": 0.5618164866931807, |
| "learning_rate": 5.9053580810947845e-05, |
| "loss": 0.4555, |
| "step": 420 |
| }, |
| { |
| "epoch": 2.043656207366985, |
| "grad_norm": 0.6473211411897097, |
| "learning_rate": 5.89342793348452e-05, |
| "loss": 0.4537, |
| "step": 421 |
| }, |
| { |
| "epoch": 2.048506897074428, |
| "grad_norm": 0.7641174380933736, |
| "learning_rate": 5.881476039410699e-05, |
| "loss": 0.4579, |
| "step": 422 |
| }, |
| { |
| "epoch": 2.0533575867818707, |
| "grad_norm": 0.6589276143277975, |
| "learning_rate": 5.869502536143629e-05, |
| "loss": 0.4478, |
| "step": 423 |
| }, |
| { |
| "epoch": 2.0582082764893133, |
| "grad_norm": 0.4055360078631115, |
| "learning_rate": 5.857507561201802e-05, |
| "loss": 0.4501, |
| "step": 424 |
| }, |
| { |
| "epoch": 2.063058966196756, |
| "grad_norm": 0.3576976481492298, |
| "learning_rate": 5.845491252350312e-05, |
| "loss": 0.4479, |
| "step": 425 |
| }, |
| { |
| "epoch": 2.0679096559041987, |
| "grad_norm": 0.444390387775824, |
| "learning_rate": 5.833453747599286e-05, |
| "loss": 0.4466, |
| "step": 426 |
| }, |
| { |
| "epoch": 2.0727603456116417, |
| "grad_norm": 0.38706062123939283, |
| "learning_rate": 5.821395185202285e-05, |
| "loss": 0.449, |
| "step": 427 |
| }, |
| { |
| "epoch": 2.0776110353190846, |
| "grad_norm": 0.3243650769978332, |
| "learning_rate": 5.809315703654726e-05, |
| "loss": 0.4581, |
| "step": 428 |
| }, |
| { |
| "epoch": 2.082461725026527, |
| "grad_norm": 0.35862489527477903, |
| "learning_rate": 5.797215441692284e-05, |
| "loss": 0.4534, |
| "step": 429 |
| }, |
| { |
| "epoch": 2.08731241473397, |
| "grad_norm": 0.3353344776943914, |
| "learning_rate": 5.785094538289304e-05, |
| "loss": 0.4537, |
| "step": 430 |
| }, |
| { |
| "epoch": 2.0921631044414126, |
| "grad_norm": 0.3620686259692887, |
| "learning_rate": 5.772953132657202e-05, |
| "loss": 0.4553, |
| "step": 431 |
| }, |
| { |
| "epoch": 2.0970137941488556, |
| "grad_norm": 0.307403523851727, |
| "learning_rate": 5.7607913642428666e-05, |
| "loss": 0.4424, |
| "step": 432 |
| }, |
| { |
| "epoch": 2.101864483856298, |
| "grad_norm": 0.28710201149568576, |
| "learning_rate": 5.7486093727270606e-05, |
| "loss": 0.4462, |
| "step": 433 |
| }, |
| { |
| "epoch": 2.106715173563741, |
| "grad_norm": 0.31451691469735704, |
| "learning_rate": 5.736407298022809e-05, |
| "loss": 0.4434, |
| "step": 434 |
| }, |
| { |
| "epoch": 2.111565863271184, |
| "grad_norm": 0.3105010146819863, |
| "learning_rate": 5.7241852802738e-05, |
| "loss": 0.4533, |
| "step": 435 |
| }, |
| { |
| "epoch": 2.1164165529786265, |
| "grad_norm": 0.2878408032383936, |
| "learning_rate": 5.711943459852772e-05, |
| "loss": 0.4427, |
| "step": 436 |
| }, |
| { |
| "epoch": 2.1212672426860695, |
| "grad_norm": 0.3026608573456508, |
| "learning_rate": 5.699681977359902e-05, |
| "loss": 0.4385, |
| "step": 437 |
| }, |
| { |
| "epoch": 2.1261179323935124, |
| "grad_norm": 0.2857208984253648, |
| "learning_rate": 5.6874009736211896e-05, |
| "loss": 0.4465, |
| "step": 438 |
| }, |
| { |
| "epoch": 2.130968622100955, |
| "grad_norm": 0.28786201316205207, |
| "learning_rate": 5.675100589686839e-05, |
| "loss": 0.4472, |
| "step": 439 |
| }, |
| { |
| "epoch": 2.135819311808398, |
| "grad_norm": 0.3397455432854385, |
| "learning_rate": 5.662780966829646e-05, |
| "loss": 0.4486, |
| "step": 440 |
| }, |
| { |
| "epoch": 2.1406700015158404, |
| "grad_norm": 0.3060713829784068, |
| "learning_rate": 5.650442246543364e-05, |
| "loss": 0.4525, |
| "step": 441 |
| }, |
| { |
| "epoch": 2.1455206912232834, |
| "grad_norm": 0.2554715156912105, |
| "learning_rate": 5.638084570541088e-05, |
| "loss": 0.4451, |
| "step": 442 |
| }, |
| { |
| "epoch": 2.150371380930726, |
| "grad_norm": 0.31763468046624377, |
| "learning_rate": 5.625708080753621e-05, |
| "loss": 0.455, |
| "step": 443 |
| }, |
| { |
| "epoch": 2.155222070638169, |
| "grad_norm": 0.3288505103534938, |
| "learning_rate": 5.6133129193278525e-05, |
| "loss": 0.4453, |
| "step": 444 |
| }, |
| { |
| "epoch": 2.160072760345612, |
| "grad_norm": 0.27302062189682574, |
| "learning_rate": 5.600899228625112e-05, |
| "loss": 0.4523, |
| "step": 445 |
| }, |
| { |
| "epoch": 2.1649234500530543, |
| "grad_norm": 0.2367917713116079, |
| "learning_rate": 5.588467151219549e-05, |
| "loss": 0.4481, |
| "step": 446 |
| }, |
| { |
| "epoch": 2.1697741397604973, |
| "grad_norm": 0.27892881990044693, |
| "learning_rate": 5.5760168298964874e-05, |
| "loss": 0.4397, |
| "step": 447 |
| }, |
| { |
| "epoch": 2.17462482946794, |
| "grad_norm": 0.3659363982261399, |
| "learning_rate": 5.563548407650782e-05, |
| "loss": 0.4464, |
| "step": 448 |
| }, |
| { |
| "epoch": 2.1794755191753827, |
| "grad_norm": 0.3139974928184246, |
| "learning_rate": 5.551062027685187e-05, |
| "loss": 0.4487, |
| "step": 449 |
| }, |
| { |
| "epoch": 2.1843262088828257, |
| "grad_norm": 0.25894377780084493, |
| "learning_rate": 5.5385578334087006e-05, |
| "loss": 0.4481, |
| "step": 450 |
| }, |
| { |
| "epoch": 2.189176898590268, |
| "grad_norm": 0.17951964246489394, |
| "learning_rate": 5.526035968434927e-05, |
| "loss": 0.4469, |
| "step": 451 |
| }, |
| { |
| "epoch": 2.194027588297711, |
| "grad_norm": 0.21554515294483917, |
| "learning_rate": 5.513496576580418e-05, |
| "loss": 0.4573, |
| "step": 452 |
| }, |
| { |
| "epoch": 2.1988782780051537, |
| "grad_norm": 0.24799983544152385, |
| "learning_rate": 5.5009398018630276e-05, |
| "loss": 0.4498, |
| "step": 453 |
| }, |
| { |
| "epoch": 2.2037289677125966, |
| "grad_norm": 0.186973375547793, |
| "learning_rate": 5.4883657885002575e-05, |
| "loss": 0.4449, |
| "step": 454 |
| }, |
| { |
| "epoch": 2.2085796574200396, |
| "grad_norm": 0.1888483688795653, |
| "learning_rate": 5.475774680907597e-05, |
| "loss": 0.443, |
| "step": 455 |
| }, |
| { |
| "epoch": 2.213430347127482, |
| "grad_norm": 0.18914695692387654, |
| "learning_rate": 5.463166623696868e-05, |
| "loss": 0.4434, |
| "step": 456 |
| }, |
| { |
| "epoch": 2.218281036834925, |
| "grad_norm": 0.19070268199404652, |
| "learning_rate": 5.450541761674562e-05, |
| "loss": 0.4445, |
| "step": 457 |
| }, |
| { |
| "epoch": 2.2231317265423676, |
| "grad_norm": 0.2259939851217927, |
| "learning_rate": 5.437900239840179e-05, |
| "loss": 0.4465, |
| "step": 458 |
| }, |
| { |
| "epoch": 2.2279824162498105, |
| "grad_norm": 0.18023421500987896, |
| "learning_rate": 5.42524220338456e-05, |
| "loss": 0.4453, |
| "step": 459 |
| }, |
| { |
| "epoch": 2.2328331059572535, |
| "grad_norm": 0.16598518940751159, |
| "learning_rate": 5.412567797688219e-05, |
| "loss": 0.4498, |
| "step": 460 |
| }, |
| { |
| "epoch": 2.237683795664696, |
| "grad_norm": 0.16004882092407235, |
| "learning_rate": 5.3998771683196754e-05, |
| "loss": 0.4527, |
| "step": 461 |
| }, |
| { |
| "epoch": 2.242534485372139, |
| "grad_norm": 0.17890202722776521, |
| "learning_rate": 5.3871704610337836e-05, |
| "loss": 0.4444, |
| "step": 462 |
| }, |
| { |
| "epoch": 2.2473851750795815, |
| "grad_norm": 0.182744670257566, |
| "learning_rate": 5.374447821770053e-05, |
| "loss": 0.4431, |
| "step": 463 |
| }, |
| { |
| "epoch": 2.2522358647870244, |
| "grad_norm": 0.19342699163866056, |
| "learning_rate": 5.361709396650977e-05, |
| "loss": 0.4404, |
| "step": 464 |
| }, |
| { |
| "epoch": 2.2570865544944674, |
| "grad_norm": 0.20442779398031627, |
| "learning_rate": 5.3489553319803566e-05, |
| "loss": 0.4496, |
| "step": 465 |
| }, |
| { |
| "epoch": 2.26193724420191, |
| "grad_norm": 0.23359410284964036, |
| "learning_rate": 5.336185774241609e-05, |
| "loss": 0.4469, |
| "step": 466 |
| }, |
| { |
| "epoch": 2.266787933909353, |
| "grad_norm": 0.23295417695606166, |
| "learning_rate": 5.3234008700961e-05, |
| "loss": 0.4505, |
| "step": 467 |
| }, |
| { |
| "epoch": 2.2716386236167954, |
| "grad_norm": 0.20207094824496044, |
| "learning_rate": 5.3106007663814505e-05, |
| "loss": 0.4406, |
| "step": 468 |
| }, |
| { |
| "epoch": 2.2764893133242383, |
| "grad_norm": 0.1850252157104855, |
| "learning_rate": 5.2977856101098484e-05, |
| "loss": 0.4525, |
| "step": 469 |
| }, |
| { |
| "epoch": 2.281340003031681, |
| "grad_norm": 0.1821206965545461, |
| "learning_rate": 5.284955548466371e-05, |
| "loss": 0.4592, |
| "step": 470 |
| }, |
| { |
| "epoch": 2.286190692739124, |
| "grad_norm": 0.1913435003815255, |
| "learning_rate": 5.272110728807279e-05, |
| "loss": 0.4459, |
| "step": 471 |
| }, |
| { |
| "epoch": 2.2910413824465667, |
| "grad_norm": 0.17908151714339782, |
| "learning_rate": 5.25925129865834e-05, |
| "loss": 0.4523, |
| "step": 472 |
| }, |
| { |
| "epoch": 2.2958920721540093, |
| "grad_norm": 0.17796456682985312, |
| "learning_rate": 5.246377405713121e-05, |
| "loss": 0.4426, |
| "step": 473 |
| }, |
| { |
| "epoch": 2.300742761861452, |
| "grad_norm": 0.16950491734508644, |
| "learning_rate": 5.2334891978313006e-05, |
| "loss": 0.4426, |
| "step": 474 |
| }, |
| { |
| "epoch": 2.305593451568895, |
| "grad_norm": 0.18036359667208995, |
| "learning_rate": 5.220586823036966e-05, |
| "loss": 0.4458, |
| "step": 475 |
| }, |
| { |
| "epoch": 2.3104441412763377, |
| "grad_norm": 0.1873060682555774, |
| "learning_rate": 5.207670429516915e-05, |
| "loss": 0.4433, |
| "step": 476 |
| }, |
| { |
| "epoch": 2.3152948309837806, |
| "grad_norm": 0.1986452939709168, |
| "learning_rate": 5.1947401656189546e-05, |
| "loss": 0.4593, |
| "step": 477 |
| }, |
| { |
| "epoch": 2.320145520691223, |
| "grad_norm": 0.20494449288937291, |
| "learning_rate": 5.181796179850197e-05, |
| "loss": 0.4424, |
| "step": 478 |
| }, |
| { |
| "epoch": 2.324996210398666, |
| "grad_norm": 0.17952910191793728, |
| "learning_rate": 5.168838620875352e-05, |
| "loss": 0.4503, |
| "step": 479 |
| }, |
| { |
| "epoch": 2.3298469001061086, |
| "grad_norm": 0.1909701609269039, |
| "learning_rate": 5.155867637515019e-05, |
| "loss": 0.4506, |
| "step": 480 |
| }, |
| { |
| "epoch": 2.3346975898135516, |
| "grad_norm": 0.1838298898954926, |
| "learning_rate": 5.142883378743984e-05, |
| "loss": 0.4513, |
| "step": 481 |
| }, |
| { |
| "epoch": 2.3395482795209945, |
| "grad_norm": 0.1818267641683358, |
| "learning_rate": 5.129885993689502e-05, |
| "loss": 0.4488, |
| "step": 482 |
| }, |
| { |
| "epoch": 2.344398969228437, |
| "grad_norm": 0.21197375880432345, |
| "learning_rate": 5.116875631629585e-05, |
| "loss": 0.4456, |
| "step": 483 |
| }, |
| { |
| "epoch": 2.34924965893588, |
| "grad_norm": 0.21240893965447508, |
| "learning_rate": 5.10385244199129e-05, |
| "loss": 0.4386, |
| "step": 484 |
| }, |
| { |
| "epoch": 2.354100348643323, |
| "grad_norm": 0.19244612255162405, |
| "learning_rate": 5.0908165743490047e-05, |
| "loss": 0.4482, |
| "step": 485 |
| }, |
| { |
| "epoch": 2.3589510383507655, |
| "grad_norm": 0.22440529731925618, |
| "learning_rate": 5.0777681784227224e-05, |
| "loss": 0.4496, |
| "step": 486 |
| }, |
| { |
| "epoch": 2.3638017280582084, |
| "grad_norm": 0.249440062974833, |
| "learning_rate": 5.064707404076327e-05, |
| "loss": 0.4502, |
| "step": 487 |
| }, |
| { |
| "epoch": 2.368652417765651, |
| "grad_norm": 0.2374206142112278, |
| "learning_rate": 5.051634401315875e-05, |
| "loss": 0.448, |
| "step": 488 |
| }, |
| { |
| "epoch": 2.373503107473094, |
| "grad_norm": 0.21044332969367502, |
| "learning_rate": 5.0385493202878656e-05, |
| "loss": 0.4416, |
| "step": 489 |
| }, |
| { |
| "epoch": 2.3783537971805364, |
| "grad_norm": 0.15343545111269605, |
| "learning_rate": 5.025452311277522e-05, |
| "loss": 0.4413, |
| "step": 490 |
| }, |
| { |
| "epoch": 2.3832044868879794, |
| "grad_norm": 0.17472771019103053, |
| "learning_rate": 5.01234352470706e-05, |
| "loss": 0.4472, |
| "step": 491 |
| }, |
| { |
| "epoch": 2.3880551765954223, |
| "grad_norm": 0.2225509747823868, |
| "learning_rate": 4.999223111133968e-05, |
| "loss": 0.4405, |
| "step": 492 |
| }, |
| { |
| "epoch": 2.392905866302865, |
| "grad_norm": 0.27110633436791925, |
| "learning_rate": 4.986091221249269e-05, |
| "loss": 0.44, |
| "step": 493 |
| }, |
| { |
| "epoch": 2.397756556010308, |
| "grad_norm": 0.24255464597168586, |
| "learning_rate": 4.972948005875796e-05, |
| "loss": 0.4432, |
| "step": 494 |
| }, |
| { |
| "epoch": 2.4026072457177503, |
| "grad_norm": 0.2496648842091371, |
| "learning_rate": 4.959793615966459e-05, |
| "loss": 0.4401, |
| "step": 495 |
| }, |
| { |
| "epoch": 2.4074579354251933, |
| "grad_norm": 0.24806426439634907, |
| "learning_rate": 4.946628202602508e-05, |
| "loss": 0.4526, |
| "step": 496 |
| }, |
| { |
| "epoch": 2.412308625132636, |
| "grad_norm": 0.21808090914084832, |
| "learning_rate": 4.933451916991802e-05, |
| "loss": 0.4474, |
| "step": 497 |
| }, |
| { |
| "epoch": 2.4171593148400787, |
| "grad_norm": 0.19833835766366836, |
| "learning_rate": 4.920264910467066e-05, |
| "loss": 0.4485, |
| "step": 498 |
| }, |
| { |
| "epoch": 2.4220100045475217, |
| "grad_norm": 0.1904056029579938, |
| "learning_rate": 4.9070673344841645e-05, |
| "loss": 0.4471, |
| "step": 499 |
| }, |
| { |
| "epoch": 2.426860694254964, |
| "grad_norm": 0.17821880940044135, |
| "learning_rate": 4.893859340620348e-05, |
| "loss": 0.4518, |
| "step": 500 |
| }, |
| { |
| "epoch": 2.431711383962407, |
| "grad_norm": 0.16242846601925154, |
| "learning_rate": 4.880641080572522e-05, |
| "loss": 0.4426, |
| "step": 501 |
| }, |
| { |
| "epoch": 2.43656207366985, |
| "grad_norm": 0.16230843192633562, |
| "learning_rate": 4.8674127061555025e-05, |
| "loss": 0.4492, |
| "step": 502 |
| }, |
| { |
| "epoch": 2.4414127633772926, |
| "grad_norm": 0.18692985874064466, |
| "learning_rate": 4.8541743693002676e-05, |
| "loss": 0.4576, |
| "step": 503 |
| }, |
| { |
| "epoch": 2.4462634530847356, |
| "grad_norm": 0.17489245993778632, |
| "learning_rate": 4.8409262220522196e-05, |
| "loss": 0.4476, |
| "step": 504 |
| }, |
| { |
| "epoch": 2.451114142792178, |
| "grad_norm": 0.14061959670906948, |
| "learning_rate": 4.8276684165694336e-05, |
| "loss": 0.4479, |
| "step": 505 |
| }, |
| { |
| "epoch": 2.455964832499621, |
| "grad_norm": 0.17289206898304424, |
| "learning_rate": 4.814401105120914e-05, |
| "loss": 0.4479, |
| "step": 506 |
| }, |
| { |
| "epoch": 2.460815522207064, |
| "grad_norm": 0.21063084112901795, |
| "learning_rate": 4.8011244400848414e-05, |
| "loss": 0.4466, |
| "step": 507 |
| }, |
| { |
| "epoch": 2.4656662119145065, |
| "grad_norm": 0.2134287283629687, |
| "learning_rate": 4.787838573946825e-05, |
| "loss": 0.4503, |
| "step": 508 |
| }, |
| { |
| "epoch": 2.4705169016219495, |
| "grad_norm": 0.19387557882251144, |
| "learning_rate": 4.774543659298152e-05, |
| "loss": 0.4419, |
| "step": 509 |
| }, |
| { |
| "epoch": 2.475367591329392, |
| "grad_norm": 0.1690053079886072, |
| "learning_rate": 4.761239848834031e-05, |
| "loss": 0.4443, |
| "step": 510 |
| }, |
| { |
| "epoch": 2.480218281036835, |
| "grad_norm": 0.17406180900609755, |
| "learning_rate": 4.747927295351845e-05, |
| "loss": 0.4474, |
| "step": 511 |
| }, |
| { |
| "epoch": 2.485068970744278, |
| "grad_norm": 0.2024050850623432, |
| "learning_rate": 4.734606151749389e-05, |
| "loss": 0.4473, |
| "step": 512 |
| }, |
| { |
| "epoch": 2.4899196604517204, |
| "grad_norm": 0.22659357369802574, |
| "learning_rate": 4.7212765710231204e-05, |
| "loss": 0.4481, |
| "step": 513 |
| }, |
| { |
| "epoch": 2.4947703501591634, |
| "grad_norm": 0.23595309939097722, |
| "learning_rate": 4.707938706266397e-05, |
| "loss": 0.4484, |
| "step": 514 |
| }, |
| { |
| "epoch": 2.499621039866606, |
| "grad_norm": 0.19918584163751257, |
| "learning_rate": 4.694592710667723e-05, |
| "loss": 0.444, |
| "step": 515 |
| }, |
| { |
| "epoch": 2.504471729574049, |
| "grad_norm": 0.18418670752131802, |
| "learning_rate": 4.681238737508983e-05, |
| "loss": 0.4424, |
| "step": 516 |
| }, |
| { |
| "epoch": 2.5093224192814914, |
| "grad_norm": 0.1926237495649244, |
| "learning_rate": 4.6678769401636894e-05, |
| "loss": 0.4444, |
| "step": 517 |
| }, |
| { |
| "epoch": 2.5141731089889343, |
| "grad_norm": 0.20706125086296728, |
| "learning_rate": 4.6545074720952166e-05, |
| "loss": 0.456, |
| "step": 518 |
| }, |
| { |
| "epoch": 2.5190237986963773, |
| "grad_norm": 0.17699664563372686, |
| "learning_rate": 4.641130486855038e-05, |
| "loss": 0.4396, |
| "step": 519 |
| }, |
| { |
| "epoch": 2.52387448840382, |
| "grad_norm": 0.18317752602670304, |
| "learning_rate": 4.627746138080966e-05, |
| "loss": 0.4432, |
| "step": 520 |
| }, |
| { |
| "epoch": 2.5287251781112627, |
| "grad_norm": 0.2190424482227647, |
| "learning_rate": 4.614354579495379e-05, |
| "loss": 0.4448, |
| "step": 521 |
| }, |
| { |
| "epoch": 2.5335758678187057, |
| "grad_norm": 0.20135719119048615, |
| "learning_rate": 4.6009559649034695e-05, |
| "loss": 0.4432, |
| "step": 522 |
| }, |
| { |
| "epoch": 2.538426557526148, |
| "grad_norm": 0.20782420010728125, |
| "learning_rate": 4.587550448191465e-05, |
| "loss": 0.4474, |
| "step": 523 |
| }, |
| { |
| "epoch": 2.543277247233591, |
| "grad_norm": 0.18668296726800496, |
| "learning_rate": 4.5741381833248655e-05, |
| "loss": 0.455, |
| "step": 524 |
| }, |
| { |
| "epoch": 2.5481279369410337, |
| "grad_norm": 0.17935132627421838, |
| "learning_rate": 4.560719324346677e-05, |
| "loss": 0.4457, |
| "step": 525 |
| }, |
| { |
| "epoch": 2.5529786266484766, |
| "grad_norm": 0.16835981784522308, |
| "learning_rate": 4.547294025375641e-05, |
| "loss": 0.4478, |
| "step": 526 |
| }, |
| { |
| "epoch": 2.557829316355919, |
| "grad_norm": 0.18084487512355504, |
| "learning_rate": 4.533862440604461e-05, |
| "loss": 0.447, |
| "step": 527 |
| }, |
| { |
| "epoch": 2.562680006063362, |
| "grad_norm": 0.17384784743298828, |
| "learning_rate": 4.520424724298036e-05, |
| "loss": 0.4408, |
| "step": 528 |
| }, |
| { |
| "epoch": 2.567530695770805, |
| "grad_norm": 0.20150460275113774, |
| "learning_rate": 4.5069810307916874e-05, |
| "loss": 0.4441, |
| "step": 529 |
| }, |
| { |
| "epoch": 2.5723813854782476, |
| "grad_norm": 0.2240004858996321, |
| "learning_rate": 4.493531514489385e-05, |
| "loss": 0.4425, |
| "step": 530 |
| }, |
| { |
| "epoch": 2.5772320751856905, |
| "grad_norm": 0.2286831099325836, |
| "learning_rate": 4.480076329861977e-05, |
| "loss": 0.4433, |
| "step": 531 |
| }, |
| { |
| "epoch": 2.5820827648931335, |
| "grad_norm": 0.2048648110357608, |
| "learning_rate": 4.46661563144541e-05, |
| "loss": 0.4487, |
| "step": 532 |
| }, |
| { |
| "epoch": 2.586933454600576, |
| "grad_norm": 0.2121488428415987, |
| "learning_rate": 4.453149573838962e-05, |
| "loss": 0.4445, |
| "step": 533 |
| }, |
| { |
| "epoch": 2.591784144308019, |
| "grad_norm": 0.18541474138380978, |
| "learning_rate": 4.43967831170346e-05, |
| "loss": 0.4494, |
| "step": 534 |
| }, |
| { |
| "epoch": 2.5966348340154615, |
| "grad_norm": 0.17452762694525445, |
| "learning_rate": 4.426201999759505e-05, |
| "loss": 0.4484, |
| "step": 535 |
| }, |
| { |
| "epoch": 2.6014855237229044, |
| "grad_norm": 0.21944224068377363, |
| "learning_rate": 4.4127207927857e-05, |
| "loss": 0.4419, |
| "step": 536 |
| }, |
| { |
| "epoch": 2.606336213430347, |
| "grad_norm": 0.17656272332454842, |
| "learning_rate": 4.3992348456168666e-05, |
| "loss": 0.4568, |
| "step": 537 |
| }, |
| { |
| "epoch": 2.61118690313779, |
| "grad_norm": 0.1892562653364182, |
| "learning_rate": 4.385744313142267e-05, |
| "loss": 0.4427, |
| "step": 538 |
| }, |
| { |
| "epoch": 2.616037592845233, |
| "grad_norm": 0.21611454670373548, |
| "learning_rate": 4.372249350303828e-05, |
| "loss": 0.4418, |
| "step": 539 |
| }, |
| { |
| "epoch": 2.6208882825526754, |
| "grad_norm": 0.17168747953224547, |
| "learning_rate": 4.358750112094363e-05, |
| "loss": 0.4544, |
| "step": 540 |
| }, |
| { |
| "epoch": 2.6257389722601183, |
| "grad_norm": 0.17941819138400728, |
| "learning_rate": 4.3452467535557846e-05, |
| "loss": 0.4372, |
| "step": 541 |
| }, |
| { |
| "epoch": 2.6305896619675613, |
| "grad_norm": 0.2025265834742146, |
| "learning_rate": 4.3317394297773304e-05, |
| "loss": 0.4517, |
| "step": 542 |
| }, |
| { |
| "epoch": 2.635440351675004, |
| "grad_norm": 0.20441246530938206, |
| "learning_rate": 4.3182282958937816e-05, |
| "loss": 0.4333, |
| "step": 543 |
| }, |
| { |
| "epoch": 2.6402910413824463, |
| "grad_norm": 0.2334105452950634, |
| "learning_rate": 4.304713507083673e-05, |
| "loss": 0.4481, |
| "step": 544 |
| }, |
| { |
| "epoch": 2.6451417310898893, |
| "grad_norm": 0.26291969340773214, |
| "learning_rate": 4.291195218567523e-05, |
| "loss": 0.4466, |
| "step": 545 |
| }, |
| { |
| "epoch": 2.649992420797332, |
| "grad_norm": 0.1863631298156993, |
| "learning_rate": 4.277673585606046e-05, |
| "loss": 0.4405, |
| "step": 546 |
| }, |
| { |
| "epoch": 2.6548431105047747, |
| "grad_norm": 0.23226855973797117, |
| "learning_rate": 4.264148763498364e-05, |
| "loss": 0.4566, |
| "step": 547 |
| }, |
| { |
| "epoch": 2.6596938002122177, |
| "grad_norm": 0.30482274820740174, |
| "learning_rate": 4.250620907580226e-05, |
| "loss": 0.4407, |
| "step": 548 |
| }, |
| { |
| "epoch": 2.6645444899196606, |
| "grad_norm": 0.23781311620065457, |
| "learning_rate": 4.237090173222231e-05, |
| "loss": 0.4493, |
| "step": 549 |
| }, |
| { |
| "epoch": 2.669395179627103, |
| "grad_norm": 0.1808214801234254, |
| "learning_rate": 4.223556715828033e-05, |
| "loss": 0.4511, |
| "step": 550 |
| }, |
| { |
| "epoch": 2.674245869334546, |
| "grad_norm": 0.26315804734468673, |
| "learning_rate": 4.2100206908325603e-05, |
| "loss": 0.447, |
| "step": 551 |
| }, |
| { |
| "epoch": 2.6790965590419886, |
| "grad_norm": 0.25781234163394623, |
| "learning_rate": 4.196482253700235e-05, |
| "loss": 0.4415, |
| "step": 552 |
| }, |
| { |
| "epoch": 2.6839472487494316, |
| "grad_norm": 0.17133762584152984, |
| "learning_rate": 4.182941559923179e-05, |
| "loss": 0.4457, |
| "step": 553 |
| }, |
| { |
| "epoch": 2.688797938456874, |
| "grad_norm": 0.2266803612041648, |
| "learning_rate": 4.169398765019433e-05, |
| "loss": 0.4422, |
| "step": 554 |
| }, |
| { |
| "epoch": 2.693648628164317, |
| "grad_norm": 0.23286738752123257, |
| "learning_rate": 4.15585402453117e-05, |
| "loss": 0.4429, |
| "step": 555 |
| }, |
| { |
| "epoch": 2.69849931787176, |
| "grad_norm": 0.20226496811604636, |
| "learning_rate": 4.14230749402291e-05, |
| "loss": 0.4421, |
| "step": 556 |
| }, |
| { |
| "epoch": 2.7033500075792025, |
| "grad_norm": 0.21746634743317236, |
| "learning_rate": 4.128759329079732e-05, |
| "loss": 0.4318, |
| "step": 557 |
| }, |
| { |
| "epoch": 2.7082006972866455, |
| "grad_norm": 0.24285493960537577, |
| "learning_rate": 4.115209685305482e-05, |
| "loss": 0.4374, |
| "step": 558 |
| }, |
| { |
| "epoch": 2.7130513869940884, |
| "grad_norm": 0.20035101285126697, |
| "learning_rate": 4.101658718320998e-05, |
| "loss": 0.4429, |
| "step": 559 |
| }, |
| { |
| "epoch": 2.717902076701531, |
| "grad_norm": 0.1733102653989901, |
| "learning_rate": 4.088106583762309e-05, |
| "loss": 0.4456, |
| "step": 560 |
| }, |
| { |
| "epoch": 2.722752766408974, |
| "grad_norm": 0.25116764609287723, |
| "learning_rate": 4.074553437278857e-05, |
| "loss": 0.4494, |
| "step": 561 |
| }, |
| { |
| "epoch": 2.7276034561164164, |
| "grad_norm": 0.19896329775589092, |
| "learning_rate": 4.060999434531704e-05, |
| "loss": 0.4449, |
| "step": 562 |
| }, |
| { |
| "epoch": 2.7324541458238594, |
| "grad_norm": 0.16517536428811208, |
| "learning_rate": 4.047444731191751e-05, |
| "loss": 0.4426, |
| "step": 563 |
| }, |
| { |
| "epoch": 2.737304835531302, |
| "grad_norm": 0.1656807626865065, |
| "learning_rate": 4.033889482937943e-05, |
| "loss": 0.4445, |
| "step": 564 |
| }, |
| { |
| "epoch": 2.742155525238745, |
| "grad_norm": 0.15103159619749504, |
| "learning_rate": 4.020333845455478e-05, |
| "loss": 0.4565, |
| "step": 565 |
| }, |
| { |
| "epoch": 2.747006214946188, |
| "grad_norm": 0.16996820086522443, |
| "learning_rate": 4.0067779744340345e-05, |
| "loss": 0.4459, |
| "step": 566 |
| }, |
| { |
| "epoch": 2.7518569046536303, |
| "grad_norm": 0.1495970266083701, |
| "learning_rate": 3.993222025565966e-05, |
| "loss": 0.4447, |
| "step": 567 |
| }, |
| { |
| "epoch": 2.7567075943610733, |
| "grad_norm": 0.15458974892236554, |
| "learning_rate": 3.979666154544522e-05, |
| "loss": 0.4452, |
| "step": 568 |
| }, |
| { |
| "epoch": 2.7615582840685162, |
| "grad_norm": 0.1741093401099396, |
| "learning_rate": 3.96611051706206e-05, |
| "loss": 0.4421, |
| "step": 569 |
| }, |
| { |
| "epoch": 2.7664089737759587, |
| "grad_norm": 0.1819530197226333, |
| "learning_rate": 3.9525552688082494e-05, |
| "loss": 0.4509, |
| "step": 570 |
| }, |
| { |
| "epoch": 2.7712596634834017, |
| "grad_norm": 0.14996389947080183, |
| "learning_rate": 3.939000565468297e-05, |
| "loss": 0.4442, |
| "step": 571 |
| }, |
| { |
| "epoch": 2.776110353190844, |
| "grad_norm": 0.19015205226216172, |
| "learning_rate": 3.9254465627211444e-05, |
| "loss": 0.4458, |
| "step": 572 |
| }, |
| { |
| "epoch": 2.780961042898287, |
| "grad_norm": 0.2147271939320094, |
| "learning_rate": 3.911893416237693e-05, |
| "loss": 0.4423, |
| "step": 573 |
| }, |
| { |
| "epoch": 2.7858117326057297, |
| "grad_norm": 0.17368493707493848, |
| "learning_rate": 3.8983412816790045e-05, |
| "loss": 0.4415, |
| "step": 574 |
| }, |
| { |
| "epoch": 2.7906624223131726, |
| "grad_norm": 0.18366782229193682, |
| "learning_rate": 3.8847903146945186e-05, |
| "loss": 0.4419, |
| "step": 575 |
| }, |
| { |
| "epoch": 2.7955131120206156, |
| "grad_norm": 0.1770373720928735, |
| "learning_rate": 3.871240670920269e-05, |
| "loss": 0.4477, |
| "step": 576 |
| }, |
| { |
| "epoch": 2.800363801728058, |
| "grad_norm": 0.15508137985676013, |
| "learning_rate": 3.85769250597709e-05, |
| "loss": 0.4458, |
| "step": 577 |
| }, |
| { |
| "epoch": 2.805214491435501, |
| "grad_norm": 0.15518418619016236, |
| "learning_rate": 3.844145975468832e-05, |
| "loss": 0.4403, |
| "step": 578 |
| }, |
| { |
| "epoch": 2.810065181142944, |
| "grad_norm": 0.13016021632650948, |
| "learning_rate": 3.830601234980569e-05, |
| "loss": 0.4509, |
| "step": 579 |
| }, |
| { |
| "epoch": 2.8149158708503865, |
| "grad_norm": 0.17125198843453068, |
| "learning_rate": 3.8170584400768224e-05, |
| "loss": 0.4492, |
| "step": 580 |
| }, |
| { |
| "epoch": 2.8197665605578295, |
| "grad_norm": 0.18182599605587274, |
| "learning_rate": 3.8035177462997664e-05, |
| "loss": 0.4475, |
| "step": 581 |
| }, |
| { |
| "epoch": 2.824617250265272, |
| "grad_norm": 0.16612208849559923, |
| "learning_rate": 3.7899793091674396e-05, |
| "loss": 0.4419, |
| "step": 582 |
| }, |
| { |
| "epoch": 2.829467939972715, |
| "grad_norm": 0.15504382692612345, |
| "learning_rate": 3.776443284171969e-05, |
| "loss": 0.4421, |
| "step": 583 |
| }, |
| { |
| "epoch": 2.8343186296801575, |
| "grad_norm": 0.14158493240403466, |
| "learning_rate": 3.7629098267777706e-05, |
| "loss": 0.4399, |
| "step": 584 |
| }, |
| { |
| "epoch": 2.8391693193876004, |
| "grad_norm": 0.14521790840725082, |
| "learning_rate": 3.7493790924197746e-05, |
| "loss": 0.4328, |
| "step": 585 |
| }, |
| { |
| "epoch": 2.8440200090950434, |
| "grad_norm": 0.14223117063886642, |
| "learning_rate": 3.735851236501637e-05, |
| "loss": 0.4403, |
| "step": 586 |
| }, |
| { |
| "epoch": 2.848870698802486, |
| "grad_norm": 0.13716153813556975, |
| "learning_rate": 3.722326414393954e-05, |
| "loss": 0.4375, |
| "step": 587 |
| }, |
| { |
| "epoch": 2.853721388509929, |
| "grad_norm": 0.13481835773066367, |
| "learning_rate": 3.708804781432478e-05, |
| "loss": 0.4465, |
| "step": 588 |
| }, |
| { |
| "epoch": 2.858572078217372, |
| "grad_norm": 0.15098861155914894, |
| "learning_rate": 3.6952864929163286e-05, |
| "loss": 0.4478, |
| "step": 589 |
| }, |
| { |
| "epoch": 2.8634227679248143, |
| "grad_norm": 0.14428029096945075, |
| "learning_rate": 3.6817717041062204e-05, |
| "loss": 0.4433, |
| "step": 590 |
| }, |
| { |
| "epoch": 2.868273457632257, |
| "grad_norm": 0.14941685817838865, |
| "learning_rate": 3.66826057022267e-05, |
| "loss": 0.4426, |
| "step": 591 |
| }, |
| { |
| "epoch": 2.8731241473397, |
| "grad_norm": 0.15814915640950794, |
| "learning_rate": 3.654753246444217e-05, |
| "loss": 0.437, |
| "step": 592 |
| }, |
| { |
| "epoch": 2.8779748370471427, |
| "grad_norm": 0.1666350361712344, |
| "learning_rate": 3.641249887905638e-05, |
| "loss": 0.4404, |
| "step": 593 |
| }, |
| { |
| "epoch": 2.8828255267545853, |
| "grad_norm": 0.14761581189993908, |
| "learning_rate": 3.627750649696173e-05, |
| "loss": 0.4418, |
| "step": 594 |
| }, |
| { |
| "epoch": 2.887676216462028, |
| "grad_norm": 0.16433140815176697, |
| "learning_rate": 3.614255686857734e-05, |
| "loss": 0.4482, |
| "step": 595 |
| }, |
| { |
| "epoch": 2.892526906169471, |
| "grad_norm": 0.14898117787478918, |
| "learning_rate": 3.600765154383134e-05, |
| "loss": 0.4407, |
| "step": 596 |
| }, |
| { |
| "epoch": 2.8973775958769137, |
| "grad_norm": 0.1576311183401583, |
| "learning_rate": 3.587279207214301e-05, |
| "loss": 0.4502, |
| "step": 597 |
| }, |
| { |
| "epoch": 2.9022282855843566, |
| "grad_norm": 0.12713739329535242, |
| "learning_rate": 3.5737980002404965e-05, |
| "loss": 0.4504, |
| "step": 598 |
| }, |
| { |
| "epoch": 2.907078975291799, |
| "grad_norm": 0.1407162797770253, |
| "learning_rate": 3.5603216882965415e-05, |
| "loss": 0.4444, |
| "step": 599 |
| }, |
| { |
| "epoch": 2.911929664999242, |
| "grad_norm": 0.15096239269549241, |
| "learning_rate": 3.5468504261610387e-05, |
| "loss": 0.4497, |
| "step": 600 |
| }, |
| { |
| "epoch": 2.9167803547066846, |
| "grad_norm": 0.1341187523707445, |
| "learning_rate": 3.5333843685545914e-05, |
| "loss": 0.4449, |
| "step": 601 |
| }, |
| { |
| "epoch": 2.9216310444141276, |
| "grad_norm": 0.1619335943958044, |
| "learning_rate": 3.519923670138025e-05, |
| "loss": 0.4434, |
| "step": 602 |
| }, |
| { |
| "epoch": 2.9264817341215705, |
| "grad_norm": 0.16211928956809254, |
| "learning_rate": 3.506468485510616e-05, |
| "loss": 0.4394, |
| "step": 603 |
| }, |
| { |
| "epoch": 2.931332423829013, |
| "grad_norm": 0.13565985896740923, |
| "learning_rate": 3.493018969208314e-05, |
| "loss": 0.4513, |
| "step": 604 |
| }, |
| { |
| "epoch": 2.936183113536456, |
| "grad_norm": 0.14551792578562125, |
| "learning_rate": 3.479575275701965e-05, |
| "loss": 0.4425, |
| "step": 605 |
| }, |
| { |
| "epoch": 2.941033803243899, |
| "grad_norm": 0.17680002225930672, |
| "learning_rate": 3.4661375593955405e-05, |
| "loss": 0.4384, |
| "step": 606 |
| }, |
| { |
| "epoch": 2.9458844929513415, |
| "grad_norm": 0.16447736465809054, |
| "learning_rate": 3.45270597462436e-05, |
| "loss": 0.4451, |
| "step": 607 |
| }, |
| { |
| "epoch": 2.9507351826587844, |
| "grad_norm": 0.15299942108774958, |
| "learning_rate": 3.4392806756533233e-05, |
| "loss": 0.4383, |
| "step": 608 |
| }, |
| { |
| "epoch": 2.955585872366227, |
| "grad_norm": 0.1536623395643003, |
| "learning_rate": 3.425861816675135e-05, |
| "loss": 0.4453, |
| "step": 609 |
| }, |
| { |
| "epoch": 2.96043656207367, |
| "grad_norm": 0.15727509499916784, |
| "learning_rate": 3.4124495518085366e-05, |
| "loss": 0.436, |
| "step": 610 |
| }, |
| { |
| "epoch": 2.9652872517811124, |
| "grad_norm": 0.17906600408772821, |
| "learning_rate": 3.399044035096532e-05, |
| "loss": 0.4467, |
| "step": 611 |
| }, |
| { |
| "epoch": 2.9701379414885554, |
| "grad_norm": 0.16243556143131102, |
| "learning_rate": 3.3856454205046223e-05, |
| "loss": 0.4364, |
| "step": 612 |
| }, |
| { |
| "epoch": 2.9749886311959983, |
| "grad_norm": 0.1608114743453715, |
| "learning_rate": 3.372253861919036e-05, |
| "loss": 0.4517, |
| "step": 613 |
| }, |
| { |
| "epoch": 2.979839320903441, |
| "grad_norm": 0.16249676695986184, |
| "learning_rate": 3.3588695131449626e-05, |
| "loss": 0.4464, |
| "step": 614 |
| }, |
| { |
| "epoch": 2.984690010610884, |
| "grad_norm": 0.14682205641070967, |
| "learning_rate": 3.3454925279047854e-05, |
| "loss": 0.4446, |
| "step": 615 |
| }, |
| { |
| "epoch": 2.9895407003183267, |
| "grad_norm": 0.158750060737996, |
| "learning_rate": 3.3321230598363126e-05, |
| "loss": 0.4449, |
| "step": 616 |
| }, |
| { |
| "epoch": 2.9943913900257693, |
| "grad_norm": 0.15103569523913019, |
| "learning_rate": 3.3187612624910185e-05, |
| "loss": 0.4457, |
| "step": 617 |
| }, |
| { |
| "epoch": 2.999242079733212, |
| "grad_norm": 0.13279281722008607, |
| "learning_rate": 3.305407289332279e-05, |
| "loss": 0.4524, |
| "step": 618 |
| }, |
| { |
| "epoch": 3.004850689707443, |
| "grad_norm": 0.21786343264375474, |
| "learning_rate": 3.2920612937336035e-05, |
| "loss": 0.4186, |
| "step": 619 |
| }, |
| { |
| "epoch": 3.0097013794148855, |
| "grad_norm": 0.1940657522375173, |
| "learning_rate": 3.2787234289768816e-05, |
| "loss": 0.4198, |
| "step": 620 |
| }, |
| { |
| "epoch": 3.0145520691223284, |
| "grad_norm": 0.21773031084989108, |
| "learning_rate": 3.2653938482506125e-05, |
| "loss": 0.4257, |
| "step": 621 |
| }, |
| { |
| "epoch": 3.019402758829771, |
| "grad_norm": 0.20326547030411232, |
| "learning_rate": 3.252072704648157e-05, |
| "loss": 0.4165, |
| "step": 622 |
| }, |
| { |
| "epoch": 3.024253448537214, |
| "grad_norm": 0.1952882443077655, |
| "learning_rate": 3.2387601511659695e-05, |
| "loss": 0.4099, |
| "step": 623 |
| }, |
| { |
| "epoch": 3.029104138244657, |
| "grad_norm": 0.1918042514566916, |
| "learning_rate": 3.22545634070185e-05, |
| "loss": 0.4152, |
| "step": 624 |
| }, |
| { |
| "epoch": 3.0339548279520994, |
| "grad_norm": 0.19705116304049744, |
| "learning_rate": 3.212161426053177e-05, |
| "loss": 0.4128, |
| "step": 625 |
| }, |
| { |
| "epoch": 3.0388055176595423, |
| "grad_norm": 0.19649678060258355, |
| "learning_rate": 3.19887555991516e-05, |
| "loss": 0.4129, |
| "step": 626 |
| }, |
| { |
| "epoch": 3.043656207366985, |
| "grad_norm": 0.21708556979200094, |
| "learning_rate": 3.1855988948790866e-05, |
| "loss": 0.419, |
| "step": 627 |
| }, |
| { |
| "epoch": 3.048506897074428, |
| "grad_norm": 0.1975715363358727, |
| "learning_rate": 3.172331583430567e-05, |
| "loss": 0.4179, |
| "step": 628 |
| }, |
| { |
| "epoch": 3.0533575867818707, |
| "grad_norm": 0.23335868145268443, |
| "learning_rate": 3.1590737779477825e-05, |
| "loss": 0.4187, |
| "step": 629 |
| }, |
| { |
| "epoch": 3.0582082764893133, |
| "grad_norm": 0.16312780113553543, |
| "learning_rate": 3.145825630699734e-05, |
| "loss": 0.4145, |
| "step": 630 |
| }, |
| { |
| "epoch": 3.063058966196756, |
| "grad_norm": 0.1995031684584169, |
| "learning_rate": 3.1325872938444995e-05, |
| "loss": 0.4269, |
| "step": 631 |
| }, |
| { |
| "epoch": 3.0679096559041987, |
| "grad_norm": 0.16321617015370155, |
| "learning_rate": 3.119358919427478e-05, |
| "loss": 0.4189, |
| "step": 632 |
| }, |
| { |
| "epoch": 3.0727603456116417, |
| "grad_norm": 0.18410001775256316, |
| "learning_rate": 3.106140659379652e-05, |
| "loss": 0.4218, |
| "step": 633 |
| }, |
| { |
| "epoch": 3.0776110353190846, |
| "grad_norm": 0.15016856412332835, |
| "learning_rate": 3.092932665515837e-05, |
| "loss": 0.4142, |
| "step": 634 |
| }, |
| { |
| "epoch": 3.082461725026527, |
| "grad_norm": 0.1502499895895468, |
| "learning_rate": 3.079735089532935e-05, |
| "loss": 0.4092, |
| "step": 635 |
| }, |
| { |
| "epoch": 3.08731241473397, |
| "grad_norm": 0.1564011362374121, |
| "learning_rate": 3.0665480830082e-05, |
| "loss": 0.4176, |
| "step": 636 |
| }, |
| { |
| "epoch": 3.0921631044414126, |
| "grad_norm": 0.15490177035198396, |
| "learning_rate": 3.0533717973974924e-05, |
| "loss": 0.4171, |
| "step": 637 |
| }, |
| { |
| "epoch": 3.0970137941488556, |
| "grad_norm": 0.16842243538179394, |
| "learning_rate": 3.040206384033542e-05, |
| "loss": 0.4142, |
| "step": 638 |
| }, |
| { |
| "epoch": 3.101864483856298, |
| "grad_norm": 0.15647723657119908, |
| "learning_rate": 3.0270519941242052e-05, |
| "loss": 0.4045, |
| "step": 639 |
| }, |
| { |
| "epoch": 3.106715173563741, |
| "grad_norm": 0.14187950473293476, |
| "learning_rate": 3.0139087787507323e-05, |
| "loss": 0.4162, |
| "step": 640 |
| }, |
| { |
| "epoch": 3.111565863271184, |
| "grad_norm": 0.1486994798612613, |
| "learning_rate": 3.0007768888660337e-05, |
| "loss": 0.4162, |
| "step": 641 |
| }, |
| { |
| "epoch": 3.1164165529786265, |
| "grad_norm": 0.14045982542928215, |
| "learning_rate": 2.9876564752929406e-05, |
| "loss": 0.423, |
| "step": 642 |
| }, |
| { |
| "epoch": 3.1212672426860695, |
| "grad_norm": 0.14911944405394412, |
| "learning_rate": 2.9745476887224806e-05, |
| "loss": 0.4186, |
| "step": 643 |
| }, |
| { |
| "epoch": 3.1261179323935124, |
| "grad_norm": 0.130768356118423, |
| "learning_rate": 2.961450679712135e-05, |
| "loss": 0.4149, |
| "step": 644 |
| }, |
| { |
| "epoch": 3.130968622100955, |
| "grad_norm": 0.15344876789210227, |
| "learning_rate": 2.9483655986841265e-05, |
| "loss": 0.4185, |
| "step": 645 |
| }, |
| { |
| "epoch": 3.135819311808398, |
| "grad_norm": 0.16373681347456412, |
| "learning_rate": 2.9352925959236732e-05, |
| "loss": 0.4199, |
| "step": 646 |
| }, |
| { |
| "epoch": 3.1406700015158404, |
| "grad_norm": 0.13681917154319687, |
| "learning_rate": 2.92223182157728e-05, |
| "loss": 0.4187, |
| "step": 647 |
| }, |
| { |
| "epoch": 3.1455206912232834, |
| "grad_norm": 0.16376916195733163, |
| "learning_rate": 2.909183425650996e-05, |
| "loss": 0.4144, |
| "step": 648 |
| }, |
| { |
| "epoch": 3.150371380930726, |
| "grad_norm": 0.15231760580025508, |
| "learning_rate": 2.8961475580087108e-05, |
| "loss": 0.4065, |
| "step": 649 |
| }, |
| { |
| "epoch": 3.155222070638169, |
| "grad_norm": 0.1482250629623317, |
| "learning_rate": 2.8831243683704162e-05, |
| "loss": 0.4167, |
| "step": 650 |
| }, |
| { |
| "epoch": 3.160072760345612, |
| "grad_norm": 0.1279306947970543, |
| "learning_rate": 2.8701140063104996e-05, |
| "loss": 0.4163, |
| "step": 651 |
| }, |
| { |
| "epoch": 3.1649234500530543, |
| "grad_norm": 0.14839352063841615, |
| "learning_rate": 2.857116621256018e-05, |
| "loss": 0.4066, |
| "step": 652 |
| }, |
| { |
| "epoch": 3.1697741397604973, |
| "grad_norm": 0.14182099741632095, |
| "learning_rate": 2.8441323624849827e-05, |
| "loss": 0.4073, |
| "step": 653 |
| }, |
| { |
| "epoch": 3.17462482946794, |
| "grad_norm": 0.14062989686711488, |
| "learning_rate": 2.83116137912465e-05, |
| "loss": 0.4148, |
| "step": 654 |
| }, |
| { |
| "epoch": 3.1794755191753827, |
| "grad_norm": 0.13350262763245205, |
| "learning_rate": 2.8182038201498038e-05, |
| "loss": 0.4185, |
| "step": 655 |
| }, |
| { |
| "epoch": 3.1843262088828257, |
| "grad_norm": 0.14000623741594295, |
| "learning_rate": 2.8052598343810474e-05, |
| "loss": 0.4084, |
| "step": 656 |
| }, |
| { |
| "epoch": 3.189176898590268, |
| "grad_norm": 0.13922618624994615, |
| "learning_rate": 2.7923295704830868e-05, |
| "loss": 0.4209, |
| "step": 657 |
| }, |
| { |
| "epoch": 3.194027588297711, |
| "grad_norm": 0.13424360868636123, |
| "learning_rate": 2.7794131769630355e-05, |
| "loss": 0.4203, |
| "step": 658 |
| }, |
| { |
| "epoch": 3.1988782780051537, |
| "grad_norm": 0.1410971557670894, |
| "learning_rate": 2.7665108021687007e-05, |
| "loss": 0.4229, |
| "step": 659 |
| }, |
| { |
| "epoch": 3.2037289677125966, |
| "grad_norm": 0.15216131334925778, |
| "learning_rate": 2.753622594286879e-05, |
| "loss": 0.4145, |
| "step": 660 |
| }, |
| { |
| "epoch": 3.2085796574200396, |
| "grad_norm": 0.1261567069468345, |
| "learning_rate": 2.7407487013416615e-05, |
| "loss": 0.4083, |
| "step": 661 |
| }, |
| { |
| "epoch": 3.213430347127482, |
| "grad_norm": 0.15853568579534694, |
| "learning_rate": 2.727889271192722e-05, |
| "loss": 0.4187, |
| "step": 662 |
| }, |
| { |
| "epoch": 3.218281036834925, |
| "grad_norm": 0.1151093763460037, |
| "learning_rate": 2.715044451533631e-05, |
| "loss": 0.4164, |
| "step": 663 |
| }, |
| { |
| "epoch": 3.2231317265423676, |
| "grad_norm": 0.14296777545925654, |
| "learning_rate": 2.702214389890152e-05, |
| "loss": 0.413, |
| "step": 664 |
| }, |
| { |
| "epoch": 3.2279824162498105, |
| "grad_norm": 0.12844538251966497, |
| "learning_rate": 2.6893992336185512e-05, |
| "loss": 0.4035, |
| "step": 665 |
| }, |
| { |
| "epoch": 3.2328331059572535, |
| "grad_norm": 0.13036276585432338, |
| "learning_rate": 2.6765991299039025e-05, |
| "loss": 0.4145, |
| "step": 666 |
| }, |
| { |
| "epoch": 3.237683795664696, |
| "grad_norm": 0.12243507743746514, |
| "learning_rate": 2.663814225758393e-05, |
| "loss": 0.4117, |
| "step": 667 |
| }, |
| { |
| "epoch": 3.242534485372139, |
| "grad_norm": 0.14245972122975337, |
| "learning_rate": 2.6510446680196448e-05, |
| "loss": 0.4195, |
| "step": 668 |
| }, |
| { |
| "epoch": 3.2473851750795815, |
| "grad_norm": 0.12799039988824565, |
| "learning_rate": 2.638290603349023e-05, |
| "loss": 0.4203, |
| "step": 669 |
| }, |
| { |
| "epoch": 3.2522358647870244, |
| "grad_norm": 0.12961756855915293, |
| "learning_rate": 2.625552178229949e-05, |
| "loss": 0.4159, |
| "step": 670 |
| }, |
| { |
| "epoch": 3.2570865544944674, |
| "grad_norm": 0.1296944618260485, |
| "learning_rate": 2.612829538966218e-05, |
| "loss": 0.4111, |
| "step": 671 |
| }, |
| { |
| "epoch": 3.26193724420191, |
| "grad_norm": 0.14508116679367689, |
| "learning_rate": 2.6001228316803256e-05, |
| "loss": 0.4196, |
| "step": 672 |
| }, |
| { |
| "epoch": 3.266787933909353, |
| "grad_norm": 0.1216509600449271, |
| "learning_rate": 2.5874322023117824e-05, |
| "loss": 0.4162, |
| "step": 673 |
| }, |
| { |
| "epoch": 3.2716386236167954, |
| "grad_norm": 0.1407698696044201, |
| "learning_rate": 2.5747577966154404e-05, |
| "loss": 0.4165, |
| "step": 674 |
| }, |
| { |
| "epoch": 3.2764893133242383, |
| "grad_norm": 0.1142728140236527, |
| "learning_rate": 2.5620997601598215e-05, |
| "loss": 0.4076, |
| "step": 675 |
| }, |
| { |
| "epoch": 3.281340003031681, |
| "grad_norm": 0.14212475123175447, |
| "learning_rate": 2.5494582383254388e-05, |
| "loss": 0.4174, |
| "step": 676 |
| }, |
| { |
| "epoch": 3.286190692739124, |
| "grad_norm": 0.11798282133422631, |
| "learning_rate": 2.5368333763031324e-05, |
| "loss": 0.4131, |
| "step": 677 |
| }, |
| { |
| "epoch": 3.2910413824465667, |
| "grad_norm": 0.14618443390836333, |
| "learning_rate": 2.5242253190924034e-05, |
| "loss": 0.4092, |
| "step": 678 |
| }, |
| { |
| "epoch": 3.2958920721540093, |
| "grad_norm": 0.12815235524330332, |
| "learning_rate": 2.5116342114997442e-05, |
| "loss": 0.409, |
| "step": 679 |
| }, |
| { |
| "epoch": 3.300742761861452, |
| "grad_norm": 0.1346079173937725, |
| "learning_rate": 2.4990601981369737e-05, |
| "loss": 0.4201, |
| "step": 680 |
| }, |
| { |
| "epoch": 3.305593451568895, |
| "grad_norm": 0.12020401678661803, |
| "learning_rate": 2.4865034234195834e-05, |
| "loss": 0.4107, |
| "step": 681 |
| }, |
| { |
| "epoch": 3.3104441412763377, |
| "grad_norm": 0.1342860022847603, |
| "learning_rate": 2.4739640315650747e-05, |
| "loss": 0.4145, |
| "step": 682 |
| }, |
| { |
| "epoch": 3.3152948309837806, |
| "grad_norm": 0.11394559661374248, |
| "learning_rate": 2.4614421665912997e-05, |
| "loss": 0.4213, |
| "step": 683 |
| }, |
| { |
| "epoch": 3.320145520691223, |
| "grad_norm": 0.126805017438777, |
| "learning_rate": 2.4489379723148147e-05, |
| "loss": 0.4129, |
| "step": 684 |
| }, |
| { |
| "epoch": 3.324996210398666, |
| "grad_norm": 0.11817811298525939, |
| "learning_rate": 2.4364515923492187e-05, |
| "loss": 0.4193, |
| "step": 685 |
| }, |
| { |
| "epoch": 3.3298469001061086, |
| "grad_norm": 0.1342332761111468, |
| "learning_rate": 2.4239831701035143e-05, |
| "loss": 0.418, |
| "step": 686 |
| }, |
| { |
| "epoch": 3.3346975898135516, |
| "grad_norm": 0.12474527460295737, |
| "learning_rate": 2.411532848780451e-05, |
| "loss": 0.4166, |
| "step": 687 |
| }, |
| { |
| "epoch": 3.3395482795209945, |
| "grad_norm": 0.1329059048345405, |
| "learning_rate": 2.399100771374888e-05, |
| "loss": 0.4138, |
| "step": 688 |
| }, |
| { |
| "epoch": 3.344398969228437, |
| "grad_norm": 0.11628441384991241, |
| "learning_rate": 2.3866870806721495e-05, |
| "loss": 0.4111, |
| "step": 689 |
| }, |
| { |
| "epoch": 3.34924965893588, |
| "grad_norm": 0.1453864432398833, |
| "learning_rate": 2.37429191924638e-05, |
| "loss": 0.42, |
| "step": 690 |
| }, |
| { |
| "epoch": 3.354100348643323, |
| "grad_norm": 0.11264715608612798, |
| "learning_rate": 2.361915429458913e-05, |
| "loss": 0.417, |
| "step": 691 |
| }, |
| { |
| "epoch": 3.3589510383507655, |
| "grad_norm": 0.14248256005839127, |
| "learning_rate": 2.349557753456637e-05, |
| "loss": 0.4168, |
| "step": 692 |
| }, |
| { |
| "epoch": 3.3638017280582084, |
| "grad_norm": 0.12398727462550883, |
| "learning_rate": 2.3372190331703556e-05, |
| "loss": 0.4189, |
| "step": 693 |
| }, |
| { |
| "epoch": 3.368652417765651, |
| "grad_norm": 0.11869463547338541, |
| "learning_rate": 2.324899410313161e-05, |
| "loss": 0.4125, |
| "step": 694 |
| }, |
| { |
| "epoch": 3.373503107473094, |
| "grad_norm": 0.12377753865053726, |
| "learning_rate": 2.3125990263788118e-05, |
| "loss": 0.4186, |
| "step": 695 |
| }, |
| { |
| "epoch": 3.3783537971805364, |
| "grad_norm": 0.13141561076446406, |
| "learning_rate": 2.3003180226400986e-05, |
| "loss": 0.4123, |
| "step": 696 |
| }, |
| { |
| "epoch": 3.3832044868879794, |
| "grad_norm": 0.11816327910326484, |
| "learning_rate": 2.288056540147229e-05, |
| "loss": 0.4129, |
| "step": 697 |
| }, |
| { |
| "epoch": 3.3880551765954223, |
| "grad_norm": 0.12260669143866527, |
| "learning_rate": 2.275814719726201e-05, |
| "loss": 0.4133, |
| "step": 698 |
| }, |
| { |
| "epoch": 3.392905866302865, |
| "grad_norm": 0.133084483131333, |
| "learning_rate": 2.263592701977193e-05, |
| "loss": 0.4219, |
| "step": 699 |
| }, |
| { |
| "epoch": 3.397756556010308, |
| "grad_norm": 0.11848073628628028, |
| "learning_rate": 2.2513906272729397e-05, |
| "loss": 0.4143, |
| "step": 700 |
| }, |
| { |
| "epoch": 3.4026072457177503, |
| "grad_norm": 0.12858950370510128, |
| "learning_rate": 2.239208635757133e-05, |
| "loss": 0.4166, |
| "step": 701 |
| }, |
| { |
| "epoch": 3.4074579354251933, |
| "grad_norm": 0.11722692816596028, |
| "learning_rate": 2.2270468673428004e-05, |
| "loss": 0.4259, |
| "step": 702 |
| }, |
| { |
| "epoch": 3.412308625132636, |
| "grad_norm": 0.11830608786302087, |
| "learning_rate": 2.2149054617106974e-05, |
| "loss": 0.407, |
| "step": 703 |
| }, |
| { |
| "epoch": 3.4171593148400787, |
| "grad_norm": 0.12285165179780408, |
| "learning_rate": 2.2027845583077175e-05, |
| "loss": 0.4231, |
| "step": 704 |
| }, |
| { |
| "epoch": 3.4220100045475217, |
| "grad_norm": 0.12884220459818324, |
| "learning_rate": 2.1906842963452757e-05, |
| "loss": 0.4069, |
| "step": 705 |
| }, |
| { |
| "epoch": 3.426860694254964, |
| "grad_norm": 0.10700415399515635, |
| "learning_rate": 2.178604814797715e-05, |
| "loss": 0.4149, |
| "step": 706 |
| }, |
| { |
| "epoch": 3.431711383962407, |
| "grad_norm": 0.11853515304365536, |
| "learning_rate": 2.1665462524007162e-05, |
| "loss": 0.4125, |
| "step": 707 |
| }, |
| { |
| "epoch": 3.43656207366985, |
| "grad_norm": 0.10772813217204756, |
| "learning_rate": 2.1545087476496903e-05, |
| "loss": 0.4216, |
| "step": 708 |
| }, |
| { |
| "epoch": 3.4414127633772926, |
| "grad_norm": 0.12946073793938545, |
| "learning_rate": 2.1424924387981996e-05, |
| "loss": 0.4227, |
| "step": 709 |
| }, |
| { |
| "epoch": 3.4462634530847356, |
| "grad_norm": 0.11256888467579416, |
| "learning_rate": 2.1304974638563715e-05, |
| "loss": 0.4116, |
| "step": 710 |
| }, |
| { |
| "epoch": 3.451114142792178, |
| "grad_norm": 0.12334394272902899, |
| "learning_rate": 2.1185239605893013e-05, |
| "loss": 0.4217, |
| "step": 711 |
| }, |
| { |
| "epoch": 3.455964832499621, |
| "grad_norm": 0.11467056354627854, |
| "learning_rate": 2.106572066515482e-05, |
| "loss": 0.4193, |
| "step": 712 |
| }, |
| { |
| "epoch": 3.460815522207064, |
| "grad_norm": 0.11248104529388973, |
| "learning_rate": 2.0946419189052162e-05, |
| "loss": 0.4147, |
| "step": 713 |
| }, |
| { |
| "epoch": 3.4656662119145065, |
| "grad_norm": 0.11839620237634582, |
| "learning_rate": 2.0827336547790452e-05, |
| "loss": 0.4214, |
| "step": 714 |
| }, |
| { |
| "epoch": 3.4705169016219495, |
| "grad_norm": 0.11350680266976185, |
| "learning_rate": 2.0708474109061752e-05, |
| "loss": 0.416, |
| "step": 715 |
| }, |
| { |
| "epoch": 3.475367591329392, |
| "grad_norm": 0.11310394993568851, |
| "learning_rate": 2.0589833238029032e-05, |
| "loss": 0.4001, |
| "step": 716 |
| }, |
| { |
| "epoch": 3.480218281036835, |
| "grad_norm": 0.1142115014102716, |
| "learning_rate": 2.0471415297310455e-05, |
| "loss": 0.4158, |
| "step": 717 |
| }, |
| { |
| "epoch": 3.485068970744278, |
| "grad_norm": 0.12114855144347077, |
| "learning_rate": 2.0353221646963864e-05, |
| "loss": 0.41, |
| "step": 718 |
| }, |
| { |
| "epoch": 3.4899196604517204, |
| "grad_norm": 0.11155618948058961, |
| "learning_rate": 2.0235253644471012e-05, |
| "loss": 0.4226, |
| "step": 719 |
| }, |
| { |
| "epoch": 3.4947703501591634, |
| "grad_norm": 0.12346483161916652, |
| "learning_rate": 2.011751264472206e-05, |
| "loss": 0.4163, |
| "step": 720 |
| }, |
| { |
| "epoch": 3.499621039866606, |
| "grad_norm": 0.1314504891806893, |
| "learning_rate": 2.0000000000000012e-05, |
| "loss": 0.4185, |
| "step": 721 |
| }, |
| { |
| "epoch": 3.504471729574049, |
| "grad_norm": 0.11587490363769114, |
| "learning_rate": 1.9882717059965086e-05, |
| "loss": 0.4243, |
| "step": 722 |
| }, |
| { |
| "epoch": 3.5093224192814914, |
| "grad_norm": 0.147492515392888, |
| "learning_rate": 1.9765665171639345e-05, |
| "loss": 0.4139, |
| "step": 723 |
| }, |
| { |
| "epoch": 3.5141731089889343, |
| "grad_norm": 0.13161876983358525, |
| "learning_rate": 1.964884567939118e-05, |
| "loss": 0.4023, |
| "step": 724 |
| }, |
| { |
| "epoch": 3.5190237986963773, |
| "grad_norm": 0.1288012773595963, |
| "learning_rate": 1.9532259924919823e-05, |
| "loss": 0.416, |
| "step": 725 |
| }, |
| { |
| "epoch": 3.52387448840382, |
| "grad_norm": 0.10979902623938809, |
| "learning_rate": 1.9415909247239996e-05, |
| "loss": 0.4198, |
| "step": 726 |
| }, |
| { |
| "epoch": 3.5287251781112627, |
| "grad_norm": 0.12273321372911124, |
| "learning_rate": 1.9299794982666485e-05, |
| "loss": 0.4084, |
| "step": 727 |
| }, |
| { |
| "epoch": 3.5335758678187057, |
| "grad_norm": 0.11485189001172974, |
| "learning_rate": 1.9183918464798837e-05, |
| "loss": 0.408, |
| "step": 728 |
| }, |
| { |
| "epoch": 3.538426557526148, |
| "grad_norm": 0.1142263107632135, |
| "learning_rate": 1.906828102450601e-05, |
| "loss": 0.4131, |
| "step": 729 |
| }, |
| { |
| "epoch": 3.543277247233591, |
| "grad_norm": 0.11235720950088848, |
| "learning_rate": 1.895288398991114e-05, |
| "loss": 0.4102, |
| "step": 730 |
| }, |
| { |
| "epoch": 3.5481279369410337, |
| "grad_norm": 0.11476446551397697, |
| "learning_rate": 1.8837728686376158e-05, |
| "loss": 0.4176, |
| "step": 731 |
| }, |
| { |
| "epoch": 3.5529786266484766, |
| "grad_norm": 0.12252689326129268, |
| "learning_rate": 1.8722816436486754e-05, |
| "loss": 0.4238, |
| "step": 732 |
| }, |
| { |
| "epoch": 3.557829316355919, |
| "grad_norm": 0.10742367387279236, |
| "learning_rate": 1.8608148560037036e-05, |
| "loss": 0.4248, |
| "step": 733 |
| }, |
| { |
| "epoch": 3.562680006063362, |
| "grad_norm": 0.11202485484827082, |
| "learning_rate": 1.8493726374014442e-05, |
| "loss": 0.4047, |
| "step": 734 |
| }, |
| { |
| "epoch": 3.567530695770805, |
| "grad_norm": 0.10619927645163178, |
| "learning_rate": 1.8379551192584588e-05, |
| "loss": 0.4105, |
| "step": 735 |
| }, |
| { |
| "epoch": 3.5723813854782476, |
| "grad_norm": 0.1104031551739802, |
| "learning_rate": 1.826562432707619e-05, |
| "loss": 0.4158, |
| "step": 736 |
| }, |
| { |
| "epoch": 3.5772320751856905, |
| "grad_norm": 0.10920848831138874, |
| "learning_rate": 1.8151947085965994e-05, |
| "loss": 0.4157, |
| "step": 737 |
| }, |
| { |
| "epoch": 3.5820827648931335, |
| "grad_norm": 0.10703865180962287, |
| "learning_rate": 1.803852077486377e-05, |
| "loss": 0.4144, |
| "step": 738 |
| }, |
| { |
| "epoch": 3.586933454600576, |
| "grad_norm": 0.1043548569854331, |
| "learning_rate": 1.7925346696497295e-05, |
| "loss": 0.4082, |
| "step": 739 |
| }, |
| { |
| "epoch": 3.591784144308019, |
| "grad_norm": 0.11138455070453787, |
| "learning_rate": 1.781242615069733e-05, |
| "loss": 0.4137, |
| "step": 740 |
| }, |
| { |
| "epoch": 3.5966348340154615, |
| "grad_norm": 0.11820322046978973, |
| "learning_rate": 1.7699760434382853e-05, |
| "loss": 0.4108, |
| "step": 741 |
| }, |
| { |
| "epoch": 3.6014855237229044, |
| "grad_norm": 0.11278678223744326, |
| "learning_rate": 1.758735084154601e-05, |
| "loss": 0.4189, |
| "step": 742 |
| }, |
| { |
| "epoch": 3.606336213430347, |
| "grad_norm": 0.11323272041806605, |
| "learning_rate": 1.7475198663237297e-05, |
| "loss": 0.4123, |
| "step": 743 |
| }, |
| { |
| "epoch": 3.61118690313779, |
| "grad_norm": 0.10904989882061365, |
| "learning_rate": 1.736330518755082e-05, |
| "loss": 0.4158, |
| "step": 744 |
| }, |
| { |
| "epoch": 3.616037592845233, |
| "grad_norm": 0.11229772775053595, |
| "learning_rate": 1.7251671699609313e-05, |
| "loss": 0.4182, |
| "step": 745 |
| }, |
| { |
| "epoch": 3.6208882825526754, |
| "grad_norm": 0.10951334908001022, |
| "learning_rate": 1.7140299481549557e-05, |
| "loss": 0.4213, |
| "step": 746 |
| }, |
| { |
| "epoch": 3.6257389722601183, |
| "grad_norm": 0.11434493342191, |
| "learning_rate": 1.7029189812507603e-05, |
| "loss": 0.4224, |
| "step": 747 |
| }, |
| { |
| "epoch": 3.6305896619675613, |
| "grad_norm": 0.10344385368037658, |
| "learning_rate": 1.6918343968604027e-05, |
| "loss": 0.4106, |
| "step": 748 |
| }, |
| { |
| "epoch": 3.635440351675004, |
| "grad_norm": 0.11168549168378746, |
| "learning_rate": 1.6807763222929315e-05, |
| "loss": 0.408, |
| "step": 749 |
| }, |
| { |
| "epoch": 3.6402910413824463, |
| "grad_norm": 0.11594040802990377, |
| "learning_rate": 1.669744884552926e-05, |
| "loss": 0.4169, |
| "step": 750 |
| }, |
| { |
| "epoch": 3.6451417310898893, |
| "grad_norm": 0.11452998735643415, |
| "learning_rate": 1.6587402103390314e-05, |
| "loss": 0.4162, |
| "step": 751 |
| }, |
| { |
| "epoch": 3.649992420797332, |
| "grad_norm": 0.11033578052460767, |
| "learning_rate": 1.6477624260425137e-05, |
| "loss": 0.4179, |
| "step": 752 |
| }, |
| { |
| "epoch": 3.6548431105047747, |
| "grad_norm": 0.1198668006175803, |
| "learning_rate": 1.6368116577457973e-05, |
| "loss": 0.4124, |
| "step": 753 |
| }, |
| { |
| "epoch": 3.6596938002122177, |
| "grad_norm": 0.10884921001547737, |
| "learning_rate": 1.6258880312210195e-05, |
| "loss": 0.4152, |
| "step": 754 |
| }, |
| { |
| "epoch": 3.6645444899196606, |
| "grad_norm": 0.12009675101549108, |
| "learning_rate": 1.6149916719285942e-05, |
| "loss": 0.4147, |
| "step": 755 |
| }, |
| { |
| "epoch": 3.669395179627103, |
| "grad_norm": 0.11742205741912104, |
| "learning_rate": 1.6041227050157607e-05, |
| "loss": 0.4096, |
| "step": 756 |
| }, |
| { |
| "epoch": 3.674245869334546, |
| "grad_norm": 0.1123251375365744, |
| "learning_rate": 1.5932812553151506e-05, |
| "loss": 0.4128, |
| "step": 757 |
| }, |
| { |
| "epoch": 3.6790965590419886, |
| "grad_norm": 0.12009728599563303, |
| "learning_rate": 1.582467447343355e-05, |
| "loss": 0.4179, |
| "step": 758 |
| }, |
| { |
| "epoch": 3.6839472487494316, |
| "grad_norm": 0.10679356522304362, |
| "learning_rate": 1.5716814052994928e-05, |
| "loss": 0.4153, |
| "step": 759 |
| }, |
| { |
| "epoch": 3.688797938456874, |
| "grad_norm": 0.11287526417521035, |
| "learning_rate": 1.5609232530637827e-05, |
| "loss": 0.4065, |
| "step": 760 |
| }, |
| { |
| "epoch": 3.693648628164317, |
| "grad_norm": 0.10604174975295146, |
| "learning_rate": 1.5501931141961278e-05, |
| "loss": 0.4135, |
| "step": 761 |
| }, |
| { |
| "epoch": 3.69849931787176, |
| "grad_norm": 0.10895447717123838, |
| "learning_rate": 1.539491111934686e-05, |
| "loss": 0.4102, |
| "step": 762 |
| }, |
| { |
| "epoch": 3.7033500075792025, |
| "grad_norm": 0.11522106117016057, |
| "learning_rate": 1.5288173691944613e-05, |
| "loss": 0.4193, |
| "step": 763 |
| }, |
| { |
| "epoch": 3.7082006972866455, |
| "grad_norm": 0.11668799730823959, |
| "learning_rate": 1.5181720085658906e-05, |
| "loss": 0.4131, |
| "step": 764 |
| }, |
| { |
| "epoch": 3.7130513869940884, |
| "grad_norm": 0.1179757070897769, |
| "learning_rate": 1.5075551523134358e-05, |
| "loss": 0.4107, |
| "step": 765 |
| }, |
| { |
| "epoch": 3.717902076701531, |
| "grad_norm": 0.11352197320512793, |
| "learning_rate": 1.4969669223741771e-05, |
| "loss": 0.4093, |
| "step": 766 |
| }, |
| { |
| "epoch": 3.722752766408974, |
| "grad_norm": 0.11945491424386492, |
| "learning_rate": 1.4864074403564216e-05, |
| "loss": 0.4142, |
| "step": 767 |
| }, |
| { |
| "epoch": 3.7276034561164164, |
| "grad_norm": 0.10892285815783607, |
| "learning_rate": 1.4758768275382887e-05, |
| "loss": 0.4205, |
| "step": 768 |
| }, |
| { |
| "epoch": 3.7324541458238594, |
| "grad_norm": 0.12765157069597566, |
| "learning_rate": 1.4653752048663394e-05, |
| "loss": 0.412, |
| "step": 769 |
| }, |
| { |
| "epoch": 3.737304835531302, |
| "grad_norm": 0.10956587187939422, |
| "learning_rate": 1.4549026929541693e-05, |
| "loss": 0.4148, |
| "step": 770 |
| }, |
| { |
| "epoch": 3.742155525238745, |
| "grad_norm": 0.12138544117757244, |
| "learning_rate": 1.4444594120810326e-05, |
| "loss": 0.4115, |
| "step": 771 |
| }, |
| { |
| "epoch": 3.747006214946188, |
| "grad_norm": 0.10697996518645103, |
| "learning_rate": 1.4340454821904573e-05, |
| "loss": 0.4194, |
| "step": 772 |
| }, |
| { |
| "epoch": 3.7518569046536303, |
| "grad_norm": 0.10998535306961314, |
| "learning_rate": 1.4236610228888683e-05, |
| "loss": 0.4143, |
| "step": 773 |
| }, |
| { |
| "epoch": 3.7567075943610733, |
| "grad_norm": 0.10998459629638649, |
| "learning_rate": 1.4133061534442133e-05, |
| "loss": 0.4121, |
| "step": 774 |
| }, |
| { |
| "epoch": 3.7615582840685162, |
| "grad_norm": 0.1056241184630642, |
| "learning_rate": 1.4029809927845981e-05, |
| "loss": 0.4146, |
| "step": 775 |
| }, |
| { |
| "epoch": 3.7664089737759587, |
| "grad_norm": 0.11617390368347923, |
| "learning_rate": 1.3926856594969115e-05, |
| "loss": 0.4125, |
| "step": 776 |
| }, |
| { |
| "epoch": 3.7712596634834017, |
| "grad_norm": 0.10105741678809248, |
| "learning_rate": 1.3824202718254655e-05, |
| "loss": 0.4081, |
| "step": 777 |
| }, |
| { |
| "epoch": 3.776110353190844, |
| "grad_norm": 0.12303653861990478, |
| "learning_rate": 1.3721849476706477e-05, |
| "loss": 0.413, |
| "step": 778 |
| }, |
| { |
| "epoch": 3.780961042898287, |
| "grad_norm": 0.10128855312982177, |
| "learning_rate": 1.3619798045875529e-05, |
| "loss": 0.4117, |
| "step": 779 |
| }, |
| { |
| "epoch": 3.7858117326057297, |
| "grad_norm": 0.1118697000457396, |
| "learning_rate": 1.3518049597846412e-05, |
| "loss": 0.4097, |
| "step": 780 |
| }, |
| { |
| "epoch": 3.7906624223131726, |
| "grad_norm": 0.10689143155371568, |
| "learning_rate": 1.3416605301223893e-05, |
| "loss": 0.4146, |
| "step": 781 |
| }, |
| { |
| "epoch": 3.7955131120206156, |
| "grad_norm": 0.09942573069367382, |
| "learning_rate": 1.3315466321119486e-05, |
| "loss": 0.415, |
| "step": 782 |
| }, |
| { |
| "epoch": 3.800363801728058, |
| "grad_norm": 0.112223020708749, |
| "learning_rate": 1.3214633819138105e-05, |
| "loss": 0.4187, |
| "step": 783 |
| }, |
| { |
| "epoch": 3.805214491435501, |
| "grad_norm": 0.09509376603334437, |
| "learning_rate": 1.3114108953364655e-05, |
| "loss": 0.4083, |
| "step": 784 |
| }, |
| { |
| "epoch": 3.810065181142944, |
| "grad_norm": 0.1039694903593983, |
| "learning_rate": 1.3013892878350771e-05, |
| "loss": 0.415, |
| "step": 785 |
| }, |
| { |
| "epoch": 3.8149158708503865, |
| "grad_norm": 0.09799457517015436, |
| "learning_rate": 1.2913986745101567e-05, |
| "loss": 0.4082, |
| "step": 786 |
| }, |
| { |
| "epoch": 3.8197665605578295, |
| "grad_norm": 0.0994744466895985, |
| "learning_rate": 1.2814391701062392e-05, |
| "loss": 0.416, |
| "step": 787 |
| }, |
| { |
| "epoch": 3.824617250265272, |
| "grad_norm": 0.09937882084294014, |
| "learning_rate": 1.2715108890105663e-05, |
| "loss": 0.4118, |
| "step": 788 |
| }, |
| { |
| "epoch": 3.829467939972715, |
| "grad_norm": 0.09366936781008306, |
| "learning_rate": 1.2616139452517748e-05, |
| "loss": 0.4202, |
| "step": 789 |
| }, |
| { |
| "epoch": 3.8343186296801575, |
| "grad_norm": 0.10103247526201467, |
| "learning_rate": 1.2517484524985836e-05, |
| "loss": 0.414, |
| "step": 790 |
| }, |
| { |
| "epoch": 3.8391693193876004, |
| "grad_norm": 0.10154281976161658, |
| "learning_rate": 1.2419145240584856e-05, |
| "loss": 0.4169, |
| "step": 791 |
| }, |
| { |
| "epoch": 3.8440200090950434, |
| "grad_norm": 0.09917419975099584, |
| "learning_rate": 1.2321122728764566e-05, |
| "loss": 0.4121, |
| "step": 792 |
| }, |
| { |
| "epoch": 3.848870698802486, |
| "grad_norm": 0.10067017561953691, |
| "learning_rate": 1.222341811533648e-05, |
| "loss": 0.4177, |
| "step": 793 |
| }, |
| { |
| "epoch": 3.853721388509929, |
| "grad_norm": 0.10318548830281854, |
| "learning_rate": 1.2126032522460975e-05, |
| "loss": 0.4211, |
| "step": 794 |
| }, |
| { |
| "epoch": 3.858572078217372, |
| "grad_norm": 0.09616344349182201, |
| "learning_rate": 1.2028967068634417e-05, |
| "loss": 0.4204, |
| "step": 795 |
| }, |
| { |
| "epoch": 3.8634227679248143, |
| "grad_norm": 0.10489574786705688, |
| "learning_rate": 1.193222286867628e-05, |
| "loss": 0.4119, |
| "step": 796 |
| }, |
| { |
| "epoch": 3.868273457632257, |
| "grad_norm": 0.1008569871750799, |
| "learning_rate": 1.1835801033716372e-05, |
| "loss": 0.4086, |
| "step": 797 |
| }, |
| { |
| "epoch": 3.8731241473397, |
| "grad_norm": 0.09122317803788128, |
| "learning_rate": 1.1739702671182083e-05, |
| "loss": 0.4214, |
| "step": 798 |
| }, |
| { |
| "epoch": 3.8779748370471427, |
| "grad_norm": 0.0991904412555539, |
| "learning_rate": 1.1643928884785618e-05, |
| "loss": 0.4095, |
| "step": 799 |
| }, |
| { |
| "epoch": 3.8828255267545853, |
| "grad_norm": 0.0959491624589887, |
| "learning_rate": 1.1548480774511353e-05, |
| "loss": 0.4218, |
| "step": 800 |
| }, |
| { |
| "epoch": 3.887676216462028, |
| "grad_norm": 0.09541302821779325, |
| "learning_rate": 1.1453359436603213e-05, |
| "loss": 0.4218, |
| "step": 801 |
| }, |
| { |
| "epoch": 3.892526906169471, |
| "grad_norm": 0.09054297930670004, |
| "learning_rate": 1.1358565963552039e-05, |
| "loss": 0.421, |
| "step": 802 |
| }, |
| { |
| "epoch": 3.8973775958769137, |
| "grad_norm": 0.09175434223796135, |
| "learning_rate": 1.126410144408312e-05, |
| "loss": 0.4088, |
| "step": 803 |
| }, |
| { |
| "epoch": 3.9022282855843566, |
| "grad_norm": 0.0933968969265613, |
| "learning_rate": 1.1169966963143568e-05, |
| "loss": 0.4105, |
| "step": 804 |
| }, |
| { |
| "epoch": 3.907078975291799, |
| "grad_norm": 0.09827723734859882, |
| "learning_rate": 1.1076163601889953e-05, |
| "loss": 0.4114, |
| "step": 805 |
| }, |
| { |
| "epoch": 3.911929664999242, |
| "grad_norm": 0.09185288445531155, |
| "learning_rate": 1.098269243767589e-05, |
| "loss": 0.4099, |
| "step": 806 |
| }, |
| { |
| "epoch": 3.9167803547066846, |
| "grad_norm": 0.10344811490009215, |
| "learning_rate": 1.0889554544039593e-05, |
| "loss": 0.417, |
| "step": 807 |
| }, |
| { |
| "epoch": 3.9216310444141276, |
| "grad_norm": 0.09584223413813131, |
| "learning_rate": 1.0796750990691596e-05, |
| "loss": 0.4092, |
| "step": 808 |
| }, |
| { |
| "epoch": 3.9264817341215705, |
| "grad_norm": 0.10225202685968936, |
| "learning_rate": 1.0704282843502459e-05, |
| "loss": 0.4156, |
| "step": 809 |
| }, |
| { |
| "epoch": 3.931332423829013, |
| "grad_norm": 0.10087206295646214, |
| "learning_rate": 1.0612151164490525e-05, |
| "loss": 0.4209, |
| "step": 810 |
| }, |
| { |
| "epoch": 3.936183113536456, |
| "grad_norm": 0.09495839512641314, |
| "learning_rate": 1.0520357011809707e-05, |
| "loss": 0.4193, |
| "step": 811 |
| }, |
| { |
| "epoch": 3.941033803243899, |
| "grad_norm": 0.10115287872844174, |
| "learning_rate": 1.0428901439737387e-05, |
| "loss": 0.415, |
| "step": 812 |
| }, |
| { |
| "epoch": 3.9458844929513415, |
| "grad_norm": 0.09538434619439141, |
| "learning_rate": 1.0337785498662223e-05, |
| "loss": 0.4152, |
| "step": 813 |
| }, |
| { |
| "epoch": 3.9507351826587844, |
| "grad_norm": 0.0955931932252973, |
| "learning_rate": 1.024701023507216e-05, |
| "loss": 0.4153, |
| "step": 814 |
| }, |
| { |
| "epoch": 3.955585872366227, |
| "grad_norm": 0.10045702643945939, |
| "learning_rate": 1.015657669154237e-05, |
| "loss": 0.4156, |
| "step": 815 |
| }, |
| { |
| "epoch": 3.96043656207367, |
| "grad_norm": 0.09335473216661304, |
| "learning_rate": 1.00664859067233e-05, |
| "loss": 0.4109, |
| "step": 816 |
| }, |
| { |
| "epoch": 3.9652872517811124, |
| "grad_norm": 0.09645722262367523, |
| "learning_rate": 9.976738915328719e-06, |
| "loss": 0.4107, |
| "step": 817 |
| }, |
| { |
| "epoch": 3.9701379414885554, |
| "grad_norm": 0.09446841411423582, |
| "learning_rate": 9.887336748123864e-06, |
| "loss": 0.4178, |
| "step": 818 |
| }, |
| { |
| "epoch": 3.9749886311959983, |
| "grad_norm": 0.0903235693705833, |
| "learning_rate": 9.798280431913558e-06, |
| "loss": 0.4202, |
| "step": 819 |
| }, |
| { |
| "epoch": 3.979839320903441, |
| "grad_norm": 0.0948091272844266, |
| "learning_rate": 9.709570989530493e-06, |
| "loss": 0.4123, |
| "step": 820 |
| }, |
| { |
| "epoch": 3.984690010610884, |
| "grad_norm": 0.10103242493534337, |
| "learning_rate": 9.621209439823388e-06, |
| "loss": 0.4132, |
| "step": 821 |
| }, |
| { |
| "epoch": 3.9895407003183267, |
| "grad_norm": 0.09199808648298305, |
| "learning_rate": 9.533196797645354e-06, |
| "loss": 0.4101, |
| "step": 822 |
| }, |
| { |
| "epoch": 3.9943913900257693, |
| "grad_norm": 0.09866422487619428, |
| "learning_rate": 9.44553407384221e-06, |
| "loss": 0.412, |
| "step": 823 |
| }, |
| { |
| "epoch": 3.999242079733212, |
| "grad_norm": 0.09950215009808663, |
| "learning_rate": 9.358222275240884e-06, |
| "loss": 0.4113, |
| "step": 824 |
| }, |
| { |
| "epoch": 4.0048506897074425, |
| "grad_norm": 0.16227668213643984, |
| "learning_rate": 9.271262404637835e-06, |
| "loss": 0.4032, |
| "step": 825 |
| }, |
| { |
| "epoch": 4.009701379414886, |
| "grad_norm": 0.11430363283166092, |
| "learning_rate": 9.184655460787591e-06, |
| "loss": 0.3988, |
| "step": 826 |
| }, |
| { |
| "epoch": 4.014552069122328, |
| "grad_norm": 0.11751760908320912, |
| "learning_rate": 9.098402438391161e-06, |
| "loss": 0.3943, |
| "step": 827 |
| }, |
| { |
| "epoch": 4.019402758829771, |
| "grad_norm": 0.13225951402360853, |
| "learning_rate": 9.012504328084724e-06, |
| "loss": 0.4024, |
| "step": 828 |
| }, |
| { |
| "epoch": 4.0242534485372135, |
| "grad_norm": 0.1269725744791709, |
| "learning_rate": 8.926962116428228e-06, |
| "loss": 0.4, |
| "step": 829 |
| }, |
| { |
| "epoch": 4.029104138244657, |
| "grad_norm": 0.12436207204015083, |
| "learning_rate": 8.841776785894014e-06, |
| "loss": 0.3994, |
| "step": 830 |
| }, |
| { |
| "epoch": 4.033954827952099, |
| "grad_norm": 0.12882961013897004, |
| "learning_rate": 8.756949314855565e-06, |
| "loss": 0.3977, |
| "step": 831 |
| }, |
| { |
| "epoch": 4.038805517659542, |
| "grad_norm": 0.11685196489455994, |
| "learning_rate": 8.672480677576267e-06, |
| "loss": 0.3906, |
| "step": 832 |
| }, |
| { |
| "epoch": 4.043656207366985, |
| "grad_norm": 0.106568471380297, |
| "learning_rate": 8.58837184419821e-06, |
| "loss": 0.388, |
| "step": 833 |
| }, |
| { |
| "epoch": 4.048506897074428, |
| "grad_norm": 0.1134401036269532, |
| "learning_rate": 8.504623780731056e-06, |
| "loss": 0.3918, |
| "step": 834 |
| }, |
| { |
| "epoch": 4.05335758678187, |
| "grad_norm": 0.12093411095557142, |
| "learning_rate": 8.421237449040962e-06, |
| "loss": 0.397, |
| "step": 835 |
| }, |
| { |
| "epoch": 4.058208276489314, |
| "grad_norm": 0.11674197768223235, |
| "learning_rate": 8.338213806839453e-06, |
| "loss": 0.393, |
| "step": 836 |
| }, |
| { |
| "epoch": 4.063058966196756, |
| "grad_norm": 0.10082248897884766, |
| "learning_rate": 8.255553807672547e-06, |
| "loss": 0.3897, |
| "step": 837 |
| }, |
| { |
| "epoch": 4.067909655904199, |
| "grad_norm": 0.11069126863357161, |
| "learning_rate": 8.1732584009097e-06, |
| "loss": 0.3882, |
| "step": 838 |
| }, |
| { |
| "epoch": 4.072760345611641, |
| "grad_norm": 0.11738423041846735, |
| "learning_rate": 8.091328531732925e-06, |
| "loss": 0.3959, |
| "step": 839 |
| }, |
| { |
| "epoch": 4.077611035319085, |
| "grad_norm": 0.10387395387726432, |
| "learning_rate": 8.009765141126014e-06, |
| "loss": 0.3891, |
| "step": 840 |
| }, |
| { |
| "epoch": 4.082461725026527, |
| "grad_norm": 0.10003982048853301, |
| "learning_rate": 7.928569165863584e-06, |
| "loss": 0.3909, |
| "step": 841 |
| }, |
| { |
| "epoch": 4.08731241473397, |
| "grad_norm": 0.10549262492820612, |
| "learning_rate": 7.847741538500439e-06, |
| "loss": 0.3875, |
| "step": 842 |
| }, |
| { |
| "epoch": 4.092163104441413, |
| "grad_norm": 0.10986857014715488, |
| "learning_rate": 7.767283187360846e-06, |
| "loss": 0.3929, |
| "step": 843 |
| }, |
| { |
| "epoch": 4.097013794148856, |
| "grad_norm": 0.09674503227632227, |
| "learning_rate": 7.687195036527813e-06, |
| "loss": 0.3928, |
| "step": 844 |
| }, |
| { |
| "epoch": 4.101864483856298, |
| "grad_norm": 0.10305617355184506, |
| "learning_rate": 7.60747800583252e-06, |
| "loss": 0.4002, |
| "step": 845 |
| }, |
| { |
| "epoch": 4.1067151735637415, |
| "grad_norm": 0.09835145173574618, |
| "learning_rate": 7.52813301084375e-06, |
| "loss": 0.3961, |
| "step": 846 |
| }, |
| { |
| "epoch": 4.111565863271184, |
| "grad_norm": 0.09426867245890704, |
| "learning_rate": 7.449160962857358e-06, |
| "loss": 0.3946, |
| "step": 847 |
| }, |
| { |
| "epoch": 4.1164165529786265, |
| "grad_norm": 0.0993123791236732, |
| "learning_rate": 7.370562768885823e-06, |
| "loss": 0.3937, |
| "step": 848 |
| }, |
| { |
| "epoch": 4.121267242686069, |
| "grad_norm": 0.09692548506896993, |
| "learning_rate": 7.292339331647848e-06, |
| "loss": 0.3957, |
| "step": 849 |
| }, |
| { |
| "epoch": 4.126117932393512, |
| "grad_norm": 0.0977639150148905, |
| "learning_rate": 7.214491549557898e-06, |
| "loss": 0.3969, |
| "step": 850 |
| }, |
| { |
| "epoch": 4.130968622100955, |
| "grad_norm": 0.10100358672328608, |
| "learning_rate": 7.1370203167160326e-06, |
| "loss": 0.3937, |
| "step": 851 |
| }, |
| { |
| "epoch": 4.1358193118083975, |
| "grad_norm": 0.0961508995875077, |
| "learning_rate": 7.0599265228975e-06, |
| "loss": 0.3965, |
| "step": 852 |
| }, |
| { |
| "epoch": 4.140670001515841, |
| "grad_norm": 0.09450992963252156, |
| "learning_rate": 6.983211053542591e-06, |
| "loss": 0.4008, |
| "step": 853 |
| }, |
| { |
| "epoch": 4.145520691223283, |
| "grad_norm": 0.10411768031595499, |
| "learning_rate": 6.9068747897464535e-06, |
| "loss": 0.4032, |
| "step": 854 |
| }, |
| { |
| "epoch": 4.150371380930726, |
| "grad_norm": 0.09981896937364662, |
| "learning_rate": 6.830918608248964e-06, |
| "loss": 0.4002, |
| "step": 855 |
| }, |
| { |
| "epoch": 4.155222070638169, |
| "grad_norm": 0.08754517232013051, |
| "learning_rate": 6.755343381424659e-06, |
| "loss": 0.3976, |
| "step": 856 |
| }, |
| { |
| "epoch": 4.160072760345612, |
| "grad_norm": 0.09520671649907132, |
| "learning_rate": 6.68014997727275e-06, |
| "loss": 0.3921, |
| "step": 857 |
| }, |
| { |
| "epoch": 4.164923450053054, |
| "grad_norm": 0.09719121632502135, |
| "learning_rate": 6.605339259407104e-06, |
| "loss": 0.3852, |
| "step": 858 |
| }, |
| { |
| "epoch": 4.169774139760497, |
| "grad_norm": 0.08870928687089744, |
| "learning_rate": 6.530912087046317e-06, |
| "loss": 0.395, |
| "step": 859 |
| }, |
| { |
| "epoch": 4.17462482946794, |
| "grad_norm": 0.08685694037296439, |
| "learning_rate": 6.456869315003946e-06, |
| "loss": 0.3941, |
| "step": 860 |
| }, |
| { |
| "epoch": 4.179475519175383, |
| "grad_norm": 0.08904997313439429, |
| "learning_rate": 6.3832117936785564e-06, |
| "loss": 0.3997, |
| "step": 861 |
| }, |
| { |
| "epoch": 4.184326208882825, |
| "grad_norm": 0.0910482544729166, |
| "learning_rate": 6.309940369044047e-06, |
| "loss": 0.3945, |
| "step": 862 |
| }, |
| { |
| "epoch": 4.189176898590269, |
| "grad_norm": 0.0887134263309384, |
| "learning_rate": 6.23705588263992e-06, |
| "loss": 0.3881, |
| "step": 863 |
| }, |
| { |
| "epoch": 4.194027588297711, |
| "grad_norm": 0.08977043998387933, |
| "learning_rate": 6.164559171561553e-06, |
| "loss": 0.3957, |
| "step": 864 |
| }, |
| { |
| "epoch": 4.198878278005154, |
| "grad_norm": 0.09216637353138708, |
| "learning_rate": 6.092451068450671e-06, |
| "loss": 0.3969, |
| "step": 865 |
| }, |
| { |
| "epoch": 4.203728967712596, |
| "grad_norm": 0.08464526503746563, |
| "learning_rate": 6.020732401485751e-06, |
| "loss": 0.3896, |
| "step": 866 |
| }, |
| { |
| "epoch": 4.20857965742004, |
| "grad_norm": 0.08983668958241023, |
| "learning_rate": 5.9494039943724845e-06, |
| "loss": 0.4025, |
| "step": 867 |
| }, |
| { |
| "epoch": 4.213430347127482, |
| "grad_norm": 0.08753368121579265, |
| "learning_rate": 5.878466666334341e-06, |
| "loss": 0.3954, |
| "step": 868 |
| }, |
| { |
| "epoch": 4.218281036834925, |
| "grad_norm": 0.08687937990016806, |
| "learning_rate": 5.80792123210316e-06, |
| "loss": 0.391, |
| "step": 869 |
| }, |
| { |
| "epoch": 4.223131726542368, |
| "grad_norm": 0.08701962255264487, |
| "learning_rate": 5.737768501909773e-06, |
| "loss": 0.3965, |
| "step": 870 |
| }, |
| { |
| "epoch": 4.2279824162498105, |
| "grad_norm": 0.08811499746627446, |
| "learning_rate": 5.668009281474751e-06, |
| "loss": 0.3846, |
| "step": 871 |
| }, |
| { |
| "epoch": 4.232833105957253, |
| "grad_norm": 0.08561633996061273, |
| "learning_rate": 5.598644371999085e-06, |
| "loss": 0.3919, |
| "step": 872 |
| }, |
| { |
| "epoch": 4.237683795664696, |
| "grad_norm": 0.08321419945310787, |
| "learning_rate": 5.5296745701549906e-06, |
| "loss": 0.394, |
| "step": 873 |
| }, |
| { |
| "epoch": 4.242534485372139, |
| "grad_norm": 0.08724643182123827, |
| "learning_rate": 5.4611006680768305e-06, |
| "loss": 0.3994, |
| "step": 874 |
| }, |
| { |
| "epoch": 4.2473851750795815, |
| "grad_norm": 0.08361301460174729, |
| "learning_rate": 5.3929234533519345e-06, |
| "loss": 0.388, |
| "step": 875 |
| }, |
| { |
| "epoch": 4.252235864787025, |
| "grad_norm": 0.0862576560444418, |
| "learning_rate": 5.325143709011587e-06, |
| "loss": 0.3942, |
| "step": 876 |
| }, |
| { |
| "epoch": 4.257086554494467, |
| "grad_norm": 0.082606577377262, |
| "learning_rate": 5.257762213522055e-06, |
| "loss": 0.3907, |
| "step": 877 |
| }, |
| { |
| "epoch": 4.26193724420191, |
| "grad_norm": 0.08601520095290771, |
| "learning_rate": 5.19077974077558e-06, |
| "loss": 0.3927, |
| "step": 878 |
| }, |
| { |
| "epoch": 4.266787933909352, |
| "grad_norm": 0.08144847498379731, |
| "learning_rate": 5.124197060081564e-06, |
| "loss": 0.3981, |
| "step": 879 |
| }, |
| { |
| "epoch": 4.271638623616796, |
| "grad_norm": 0.08369648223405692, |
| "learning_rate": 5.058014936157714e-06, |
| "loss": 0.3987, |
| "step": 880 |
| }, |
| { |
| "epoch": 4.276489313324238, |
| "grad_norm": 0.08487450387282452, |
| "learning_rate": 4.992234129121225e-06, |
| "loss": 0.3993, |
| "step": 881 |
| }, |
| { |
| "epoch": 4.281340003031681, |
| "grad_norm": 0.0835928002976722, |
| "learning_rate": 4.926855394480079e-06, |
| "loss": 0.394, |
| "step": 882 |
| }, |
| { |
| "epoch": 4.286190692739124, |
| "grad_norm": 0.0877663910781713, |
| "learning_rate": 4.861879483124372e-06, |
| "loss": 0.3888, |
| "step": 883 |
| }, |
| { |
| "epoch": 4.291041382446567, |
| "grad_norm": 0.08201902566316537, |
| "learning_rate": 4.797307141317666e-06, |
| "loss": 0.3978, |
| "step": 884 |
| }, |
| { |
| "epoch": 4.295892072154009, |
| "grad_norm": 0.08252391090570044, |
| "learning_rate": 4.7331391106884364e-06, |
| "loss": 0.3949, |
| "step": 885 |
| }, |
| { |
| "epoch": 4.300742761861452, |
| "grad_norm": 0.08121494056766716, |
| "learning_rate": 4.6693761282215766e-06, |
| "loss": 0.3922, |
| "step": 886 |
| }, |
| { |
| "epoch": 4.305593451568895, |
| "grad_norm": 0.08414265476669228, |
| "learning_rate": 4.606018926249851e-06, |
| "loss": 0.389, |
| "step": 887 |
| }, |
| { |
| "epoch": 4.310444141276338, |
| "grad_norm": 0.08528892500629966, |
| "learning_rate": 4.543068232445596e-06, |
| "loss": 0.3956, |
| "step": 888 |
| }, |
| { |
| "epoch": 4.31529483098378, |
| "grad_norm": 0.08280199210655267, |
| "learning_rate": 4.480524769812276e-06, |
| "loss": 0.3938, |
| "step": 889 |
| }, |
| { |
| "epoch": 4.320145520691224, |
| "grad_norm": 0.08884076809121294, |
| "learning_rate": 4.418389256676206e-06, |
| "loss": 0.3947, |
| "step": 890 |
| }, |
| { |
| "epoch": 4.324996210398666, |
| "grad_norm": 0.08902762031211295, |
| "learning_rate": 4.35666240667834e-06, |
| "loss": 0.3907, |
| "step": 891 |
| }, |
| { |
| "epoch": 4.329846900106109, |
| "grad_norm": 0.08464410835355753, |
| "learning_rate": 4.295344928765999e-06, |
| "loss": 0.3939, |
| "step": 892 |
| }, |
| { |
| "epoch": 4.334697589813552, |
| "grad_norm": 0.08347421805827179, |
| "learning_rate": 4.234437527184785e-06, |
| "loss": 0.3985, |
| "step": 893 |
| }, |
| { |
| "epoch": 4.3395482795209945, |
| "grad_norm": 0.08770395898608876, |
| "learning_rate": 4.173940901470488e-06, |
| "loss": 0.395, |
| "step": 894 |
| }, |
| { |
| "epoch": 4.344398969228437, |
| "grad_norm": 0.08139356257791996, |
| "learning_rate": 4.11385574644104e-06, |
| "loss": 0.39, |
| "step": 895 |
| }, |
| { |
| "epoch": 4.34924965893588, |
| "grad_norm": 0.0818047074224124, |
| "learning_rate": 4.054182752188501e-06, |
| "loss": 0.3948, |
| "step": 896 |
| }, |
| { |
| "epoch": 4.354100348643323, |
| "grad_norm": 0.08370905382784635, |
| "learning_rate": 3.994922604071217e-06, |
| "loss": 0.3941, |
| "step": 897 |
| }, |
| { |
| "epoch": 4.3589510383507655, |
| "grad_norm": 0.08543140680473625, |
| "learning_rate": 3.936075982705871e-06, |
| "loss": 0.3982, |
| "step": 898 |
| }, |
| { |
| "epoch": 4.363801728058208, |
| "grad_norm": 0.08330546263387466, |
| "learning_rate": 3.877643563959694e-06, |
| "loss": 0.3986, |
| "step": 899 |
| }, |
| { |
| "epoch": 4.368652417765651, |
| "grad_norm": 0.08181359063571672, |
| "learning_rate": 3.819626018942732e-06, |
| "loss": 0.3962, |
| "step": 900 |
| }, |
| { |
| "epoch": 4.373503107473094, |
| "grad_norm": 0.08458157848919085, |
| "learning_rate": 3.762024014000054e-06, |
| "loss": 0.3974, |
| "step": 901 |
| }, |
| { |
| "epoch": 4.378353797180536, |
| "grad_norm": 0.0797245693813596, |
| "learning_rate": 3.7048382107042113e-06, |
| "loss": 0.3849, |
| "step": 902 |
| }, |
| { |
| "epoch": 4.38320448688798, |
| "grad_norm": 0.08552325047075819, |
| "learning_rate": 3.6480692658475446e-06, |
| "loss": 0.3908, |
| "step": 903 |
| }, |
| { |
| "epoch": 4.388055176595422, |
| "grad_norm": 0.08191614537028945, |
| "learning_rate": 3.5917178314346955e-06, |
| "loss": 0.398, |
| "step": 904 |
| }, |
| { |
| "epoch": 4.392905866302865, |
| "grad_norm": 0.08324472646577967, |
| "learning_rate": 3.535784554675088e-06, |
| "loss": 0.3941, |
| "step": 905 |
| }, |
| { |
| "epoch": 4.397756556010307, |
| "grad_norm": 0.08386476178169076, |
| "learning_rate": 3.480270077975525e-06, |
| "loss": 0.395, |
| "step": 906 |
| }, |
| { |
| "epoch": 4.402607245717751, |
| "grad_norm": 0.08759720428686872, |
| "learning_rate": 3.42517503893276e-06, |
| "loss": 0.3879, |
| "step": 907 |
| }, |
| { |
| "epoch": 4.407457935425193, |
| "grad_norm": 0.0834695533465509, |
| "learning_rate": 3.370500070326257e-06, |
| "loss": 0.3832, |
| "step": 908 |
| }, |
| { |
| "epoch": 4.412308625132636, |
| "grad_norm": 0.080547419306159, |
| "learning_rate": 3.3162458001108332e-06, |
| "loss": 0.3858, |
| "step": 909 |
| }, |
| { |
| "epoch": 4.417159314840079, |
| "grad_norm": 0.08406690651169581, |
| "learning_rate": 3.2624128514094778e-06, |
| "loss": 0.3923, |
| "step": 910 |
| }, |
| { |
| "epoch": 4.422010004547522, |
| "grad_norm": 0.08417275645743631, |
| "learning_rate": 3.20900184250625e-06, |
| "loss": 0.3933, |
| "step": 911 |
| }, |
| { |
| "epoch": 4.426860694254964, |
| "grad_norm": 0.08104238775596306, |
| "learning_rate": 3.1560133868390895e-06, |
| "loss": 0.4023, |
| "step": 912 |
| }, |
| { |
| "epoch": 4.431711383962407, |
| "grad_norm": 0.08376361425082632, |
| "learning_rate": 3.1034480929928333e-06, |
| "loss": 0.399, |
| "step": 913 |
| }, |
| { |
| "epoch": 4.43656207366985, |
| "grad_norm": 0.08058350379685782, |
| "learning_rate": 3.0513065646921957e-06, |
| "loss": 0.3946, |
| "step": 914 |
| }, |
| { |
| "epoch": 4.441412763377293, |
| "grad_norm": 0.08013760556674378, |
| "learning_rate": 2.999589400794851e-06, |
| "loss": 0.392, |
| "step": 915 |
| }, |
| { |
| "epoch": 4.446263453084735, |
| "grad_norm": 0.08082216615145961, |
| "learning_rate": 2.948297195284546e-06, |
| "loss": 0.3916, |
| "step": 916 |
| }, |
| { |
| "epoch": 4.4511141427921785, |
| "grad_norm": 0.09007717527915819, |
| "learning_rate": 2.897430537264283e-06, |
| "loss": 0.3947, |
| "step": 917 |
| }, |
| { |
| "epoch": 4.455964832499621, |
| "grad_norm": 0.0866390993026153, |
| "learning_rate": 2.8469900109495553e-06, |
| "loss": 0.3942, |
| "step": 918 |
| }, |
| { |
| "epoch": 4.460815522207064, |
| "grad_norm": 0.08034663815912857, |
| "learning_rate": 2.79697619566162e-06, |
| "loss": 0.3888, |
| "step": 919 |
| }, |
| { |
| "epoch": 4.465666211914507, |
| "grad_norm": 0.08423465438674188, |
| "learning_rate": 2.7473896658208743e-06, |
| "loss": 0.391, |
| "step": 920 |
| }, |
| { |
| "epoch": 4.4705169016219495, |
| "grad_norm": 0.08769412223500794, |
| "learning_rate": 2.6982309909402293e-06, |
| "loss": 0.3936, |
| "step": 921 |
| }, |
| { |
| "epoch": 4.475367591329392, |
| "grad_norm": 0.0794092004599977, |
| "learning_rate": 2.649500735618582e-06, |
| "loss": 0.3993, |
| "step": 922 |
| }, |
| { |
| "epoch": 4.480218281036835, |
| "grad_norm": 0.08321318361659834, |
| "learning_rate": 2.6011994595343516e-06, |
| "loss": 0.3965, |
| "step": 923 |
| }, |
| { |
| "epoch": 4.485068970744278, |
| "grad_norm": 0.08400144661884966, |
| "learning_rate": 2.5533277174389916e-06, |
| "loss": 0.3927, |
| "step": 924 |
| }, |
| { |
| "epoch": 4.48991966045172, |
| "grad_norm": 0.08039773750467258, |
| "learning_rate": 2.5058860591506973e-06, |
| "loss": 0.3927, |
| "step": 925 |
| }, |
| { |
| "epoch": 4.494770350159163, |
| "grad_norm": 0.08132175209524008, |
| "learning_rate": 2.4588750295480246e-06, |
| "loss": 0.3888, |
| "step": 926 |
| }, |
| { |
| "epoch": 4.499621039866606, |
| "grad_norm": 0.07920562392368859, |
| "learning_rate": 2.4122951685636674e-06, |
| "loss": 0.3896, |
| "step": 927 |
| }, |
| { |
| "epoch": 4.504471729574049, |
| "grad_norm": 0.07833571502263627, |
| "learning_rate": 2.366147011178246e-06, |
| "loss": 0.398, |
| "step": 928 |
| }, |
| { |
| "epoch": 4.509322419281491, |
| "grad_norm": 0.08232920047142565, |
| "learning_rate": 2.320431087414159e-06, |
| "loss": 0.3838, |
| "step": 929 |
| }, |
| { |
| "epoch": 4.514173108988935, |
| "grad_norm": 0.08136833742803433, |
| "learning_rate": 2.275147922329506e-06, |
| "loss": 0.3935, |
| "step": 930 |
| }, |
| { |
| "epoch": 4.519023798696377, |
| "grad_norm": 0.08383396678636225, |
| "learning_rate": 2.230298036012055e-06, |
| "loss": 0.3913, |
| "step": 931 |
| }, |
| { |
| "epoch": 4.52387448840382, |
| "grad_norm": 0.08000391439211, |
| "learning_rate": 2.1858819435732583e-06, |
| "loss": 0.395, |
| "step": 932 |
| }, |
| { |
| "epoch": 4.528725178111262, |
| "grad_norm": 0.080147749223563, |
| "learning_rate": 2.141900155142351e-06, |
| "loss": 0.3951, |
| "step": 933 |
| }, |
| { |
| "epoch": 4.533575867818706, |
| "grad_norm": 0.07978655829531454, |
| "learning_rate": 2.0983531758604726e-06, |
| "loss": 0.4011, |
| "step": 934 |
| }, |
| { |
| "epoch": 4.538426557526148, |
| "grad_norm": 0.08386421814730027, |
| "learning_rate": 2.055241505874892e-06, |
| "loss": 0.3999, |
| "step": 935 |
| }, |
| { |
| "epoch": 4.543277247233591, |
| "grad_norm": 0.08174904946083562, |
| "learning_rate": 2.0125656403332396e-06, |
| "loss": 0.3968, |
| "step": 936 |
| }, |
| { |
| "epoch": 4.548127936941034, |
| "grad_norm": 0.07927025601942647, |
| "learning_rate": 1.970326069377828e-06, |
| "loss": 0.4001, |
| "step": 937 |
| }, |
| { |
| "epoch": 4.552978626648477, |
| "grad_norm": 0.0783638905343862, |
| "learning_rate": 1.928523278140033e-06, |
| "loss": 0.3943, |
| "step": 938 |
| }, |
| { |
| "epoch": 4.557829316355919, |
| "grad_norm": 0.08012821453349714, |
| "learning_rate": 1.887157746734718e-06, |
| "loss": 0.3972, |
| "step": 939 |
| }, |
| { |
| "epoch": 4.562680006063362, |
| "grad_norm": 0.08022389886437123, |
| "learning_rate": 1.846229950254692e-06, |
| "loss": 0.3888, |
| "step": 940 |
| }, |
| { |
| "epoch": 4.567530695770805, |
| "grad_norm": 0.07938938767749412, |
| "learning_rate": 1.8057403587652977e-06, |
| "loss": 0.3892, |
| "step": 941 |
| }, |
| { |
| "epoch": 4.572381385478248, |
| "grad_norm": 0.07979264917062728, |
| "learning_rate": 1.7656894372989785e-06, |
| "loss": 0.3991, |
| "step": 942 |
| }, |
| { |
| "epoch": 4.57723207518569, |
| "grad_norm": 0.07900416574598519, |
| "learning_rate": 1.726077645849955e-06, |
| "loss": 0.4033, |
| "step": 943 |
| }, |
| { |
| "epoch": 4.5820827648931335, |
| "grad_norm": 0.0816244017270053, |
| "learning_rate": 1.6869054393689265e-06, |
| "loss": 0.394, |
| "step": 944 |
| }, |
| { |
| "epoch": 4.586933454600576, |
| "grad_norm": 0.07678354405433441, |
| "learning_rate": 1.6481732677578798e-06, |
| "loss": 0.4026, |
| "step": 945 |
| }, |
| { |
| "epoch": 4.5917841443080185, |
| "grad_norm": 0.07913436605637802, |
| "learning_rate": 1.60988157586488e-06, |
| "loss": 0.378, |
| "step": 946 |
| }, |
| { |
| "epoch": 4.596634834015462, |
| "grad_norm": 0.0793297266554538, |
| "learning_rate": 1.5720308034789721e-06, |
| "loss": 0.391, |
| "step": 947 |
| }, |
| { |
| "epoch": 4.601485523722904, |
| "grad_norm": 0.07767980761029898, |
| "learning_rate": 1.5346213853251546e-06, |
| "loss": 0.3978, |
| "step": 948 |
| }, |
| { |
| "epoch": 4.606336213430347, |
| "grad_norm": 0.08494507410525762, |
| "learning_rate": 1.4976537510593646e-06, |
| "loss": 0.3995, |
| "step": 949 |
| }, |
| { |
| "epoch": 4.61118690313779, |
| "grad_norm": 0.08237577571806884, |
| "learning_rate": 1.4611283252635412e-06, |
| "loss": 0.4038, |
| "step": 950 |
| }, |
| { |
| "epoch": 4.616037592845233, |
| "grad_norm": 0.07590194142960284, |
| "learning_rate": 1.425045527440756e-06, |
| "loss": 0.3956, |
| "step": 951 |
| }, |
| { |
| "epoch": 4.620888282552675, |
| "grad_norm": 0.07954982455169565, |
| "learning_rate": 1.3894057720104104e-06, |
| "loss": 0.399, |
| "step": 952 |
| }, |
| { |
| "epoch": 4.625738972260118, |
| "grad_norm": 0.07731849776688487, |
| "learning_rate": 1.354209468303429e-06, |
| "loss": 0.3828, |
| "step": 953 |
| }, |
| { |
| "epoch": 4.630589661967561, |
| "grad_norm": 0.07970299041030604, |
| "learning_rate": 1.3194570205576284e-06, |
| "loss": 0.3954, |
| "step": 954 |
| }, |
| { |
| "epoch": 4.635440351675004, |
| "grad_norm": 0.07874124868840192, |
| "learning_rate": 1.2851488279130053e-06, |
| "loss": 0.3876, |
| "step": 955 |
| }, |
| { |
| "epoch": 4.640291041382446, |
| "grad_norm": 0.07691612719760402, |
| "learning_rate": 1.2512852844071933e-06, |
| "loss": 0.3949, |
| "step": 956 |
| }, |
| { |
| "epoch": 4.64514173108989, |
| "grad_norm": 0.08126016720695953, |
| "learning_rate": 1.2178667789709287e-06, |
| "loss": 0.3919, |
| "step": 957 |
| }, |
| { |
| "epoch": 4.649992420797332, |
| "grad_norm": 0.08002643782323322, |
| "learning_rate": 1.1848936954235702e-06, |
| "loss": 0.395, |
| "step": 958 |
| }, |
| { |
| "epoch": 4.654843110504775, |
| "grad_norm": 0.08026194265591635, |
| "learning_rate": 1.1523664124687284e-06, |
| "loss": 0.3997, |
| "step": 959 |
| }, |
| { |
| "epoch": 4.659693800212217, |
| "grad_norm": 0.07660352848559206, |
| "learning_rate": 1.1202853036898476e-06, |
| "loss": 0.3974, |
| "step": 960 |
| }, |
| { |
| "epoch": 4.664544489919661, |
| "grad_norm": 0.08113236962163348, |
| "learning_rate": 1.0886507375459908e-06, |
| "loss": 0.3981, |
| "step": 961 |
| }, |
| { |
| "epoch": 4.669395179627103, |
| "grad_norm": 0.07701309756204706, |
| "learning_rate": 1.0574630773675687e-06, |
| "loss": 0.3839, |
| "step": 962 |
| }, |
| { |
| "epoch": 4.674245869334546, |
| "grad_norm": 0.07891978078549244, |
| "learning_rate": 1.0267226813521635e-06, |
| "loss": 0.3877, |
| "step": 963 |
| }, |
| { |
| "epoch": 4.679096559041989, |
| "grad_norm": 0.07852321868608765, |
| "learning_rate": 9.964299025604274e-07, |
| "loss": 0.3921, |
| "step": 964 |
| }, |
| { |
| "epoch": 4.683947248749432, |
| "grad_norm": 0.07818107376466682, |
| "learning_rate": 9.66585088912022e-07, |
| "loss": 0.3967, |
| "step": 965 |
| }, |
| { |
| "epoch": 4.688797938456874, |
| "grad_norm": 0.076600770628096, |
| "learning_rate": 9.371885831816319e-07, |
| "loss": 0.395, |
| "step": 966 |
| }, |
| { |
| "epoch": 4.6936486281643175, |
| "grad_norm": 0.07627250485843899, |
| "learning_rate": 9.082407229950018e-07, |
| "loss": 0.3976, |
| "step": 967 |
| }, |
| { |
| "epoch": 4.69849931787176, |
| "grad_norm": 0.07905168555654667, |
| "learning_rate": 8.797418408251101e-07, |
| "loss": 0.3918, |
| "step": 968 |
| }, |
| { |
| "epoch": 4.7033500075792025, |
| "grad_norm": 0.07625038396328788, |
| "learning_rate": 8.516922639882819e-07, |
| "loss": 0.3897, |
| "step": 969 |
| }, |
| { |
| "epoch": 4.708200697286646, |
| "grad_norm": 0.08055085830124646, |
| "learning_rate": 8.2409231464049e-07, |
| "loss": 0.3901, |
| "step": 970 |
| }, |
| { |
| "epoch": 4.713051386994088, |
| "grad_norm": 0.07599467665769233, |
| "learning_rate": 7.969423097736162e-07, |
| "loss": 0.3931, |
| "step": 971 |
| }, |
| { |
| "epoch": 4.717902076701531, |
| "grad_norm": 0.08049982038815076, |
| "learning_rate": 7.702425612118269e-07, |
| "loss": 0.3962, |
| "step": 972 |
| }, |
| { |
| "epoch": 4.7227527664089735, |
| "grad_norm": 0.07817439712988589, |
| "learning_rate": 7.439933756079942e-07, |
| "loss": 0.3913, |
| "step": 973 |
| }, |
| { |
| "epoch": 4.727603456116417, |
| "grad_norm": 0.08001209525722262, |
| "learning_rate": 7.181950544401695e-07, |
| "loss": 0.3935, |
| "step": 974 |
| }, |
| { |
| "epoch": 4.732454145823859, |
| "grad_norm": 0.07544329449164154, |
| "learning_rate": 6.928478940081107e-07, |
| "loss": 0.3984, |
| "step": 975 |
| }, |
| { |
| "epoch": 4.737304835531302, |
| "grad_norm": 0.07792056514995602, |
| "learning_rate": 6.679521854299032e-07, |
| "loss": 0.3914, |
| "step": 976 |
| }, |
| { |
| "epoch": 4.742155525238745, |
| "grad_norm": 0.07777159770613211, |
| "learning_rate": 6.435082146385885e-07, |
| "loss": 0.3901, |
| "step": 977 |
| }, |
| { |
| "epoch": 4.747006214946188, |
| "grad_norm": 0.07684966968716175, |
| "learning_rate": 6.195162623789052e-07, |
| "loss": 0.393, |
| "step": 978 |
| }, |
| { |
| "epoch": 4.75185690465363, |
| "grad_norm": 0.07663626226452194, |
| "learning_rate": 5.959766042040426e-07, |
| "loss": 0.3879, |
| "step": 979 |
| }, |
| { |
| "epoch": 4.756707594361073, |
| "grad_norm": 0.07653692258850123, |
| "learning_rate": 5.728895104724963e-07, |
| "loss": 0.3959, |
| "step": 980 |
| }, |
| { |
| "epoch": 4.761558284068516, |
| "grad_norm": 0.07731590848782262, |
| "learning_rate": 5.502552463449418e-07, |
| "loss": 0.3984, |
| "step": 981 |
| }, |
| { |
| "epoch": 4.766408973775959, |
| "grad_norm": 0.07715742201671594, |
| "learning_rate": 5.280740717812149e-07, |
| "loss": 0.3977, |
| "step": 982 |
| }, |
| { |
| "epoch": 4.771259663483401, |
| "grad_norm": 0.07733908081214975, |
| "learning_rate": 5.063462415372967e-07, |
| "loss": 0.3948, |
| "step": 983 |
| }, |
| { |
| "epoch": 4.776110353190845, |
| "grad_norm": 0.07797117300936077, |
| "learning_rate": 4.850720051624124e-07, |
| "loss": 0.3914, |
| "step": 984 |
| }, |
| { |
| "epoch": 4.780961042898287, |
| "grad_norm": 0.07721734718928677, |
| "learning_rate": 4.642516069961556e-07, |
| "loss": 0.392, |
| "step": 985 |
| }, |
| { |
| "epoch": 4.78581173260573, |
| "grad_norm": 0.07648551851063208, |
| "learning_rate": 4.438852861656751e-07, |
| "loss": 0.3951, |
| "step": 986 |
| }, |
| { |
| "epoch": 4.790662422313172, |
| "grad_norm": 0.07521119037010907, |
| "learning_rate": 4.2397327658294076e-07, |
| "loss": 0.3899, |
| "step": 987 |
| }, |
| { |
| "epoch": 4.795513112020616, |
| "grad_norm": 0.07655879860272995, |
| "learning_rate": 4.045158069420474e-07, |
| "loss": 0.3963, |
| "step": 988 |
| }, |
| { |
| "epoch": 4.800363801728058, |
| "grad_norm": 0.07622760847546149, |
| "learning_rate": 3.8551310071659023e-07, |
| "loss": 0.3975, |
| "step": 989 |
| }, |
| { |
| "epoch": 4.805214491435501, |
| "grad_norm": 0.0750428400819878, |
| "learning_rate": 3.6696537615711124e-07, |
| "loss": 0.3968, |
| "step": 990 |
| }, |
| { |
| "epoch": 4.810065181142944, |
| "grad_norm": 0.07852243689578567, |
| "learning_rate": 3.4887284628857266e-07, |
| "loss": 0.3932, |
| "step": 991 |
| }, |
| { |
| "epoch": 4.8149158708503865, |
| "grad_norm": 0.0771348195411035, |
| "learning_rate": 3.3123571890791405e-07, |
| "loss": 0.3887, |
| "step": 992 |
| }, |
| { |
| "epoch": 4.819766560557829, |
| "grad_norm": 0.07614331045752255, |
| "learning_rate": 3.1405419658168125e-07, |
| "loss": 0.394, |
| "step": 993 |
| }, |
| { |
| "epoch": 4.824617250265272, |
| "grad_norm": 0.07563944264490313, |
| "learning_rate": 2.973284766436857e-07, |
| "loss": 0.3917, |
| "step": 994 |
| }, |
| { |
| "epoch": 4.829467939972715, |
| "grad_norm": 0.07503164664083585, |
| "learning_rate": 2.810587511927354e-07, |
| "loss": 0.3901, |
| "step": 995 |
| }, |
| { |
| "epoch": 4.8343186296801575, |
| "grad_norm": 0.07663595734772072, |
| "learning_rate": 2.652452070904499e-07, |
| "loss": 0.3923, |
| "step": 996 |
| }, |
| { |
| "epoch": 4.839169319387601, |
| "grad_norm": 0.07606961706547127, |
| "learning_rate": 2.498880259590797e-07, |
| "loss": 0.3944, |
| "step": 997 |
| }, |
| { |
| "epoch": 4.844020009095043, |
| "grad_norm": 0.0761182148197196, |
| "learning_rate": 2.3498738417945034e-07, |
| "loss": 0.3975, |
| "step": 998 |
| }, |
| { |
| "epoch": 4.848870698802486, |
| "grad_norm": 0.08039104985634041, |
| "learning_rate": 2.205434528889283e-07, |
| "loss": 0.3971, |
| "step": 999 |
| }, |
| { |
| "epoch": 4.853721388509928, |
| "grad_norm": 0.07775606427920397, |
| "learning_rate": 2.0655639797944937e-07, |
| "loss": 0.3903, |
| "step": 1000 |
| }, |
| { |
| "epoch": 4.858572078217372, |
| "grad_norm": 0.07610940999590161, |
| "learning_rate": 1.9302638009561782e-07, |
| "loss": 0.396, |
| "step": 1001 |
| }, |
| { |
| "epoch": 4.863422767924814, |
| "grad_norm": 0.07569450912275462, |
| "learning_rate": 1.7995355463285457e-07, |
| "loss": 0.3965, |
| "step": 1002 |
| }, |
| { |
| "epoch": 4.868273457632257, |
| "grad_norm": 0.07643720956773195, |
| "learning_rate": 1.6733807173562988e-07, |
| "loss": 0.3913, |
| "step": 1003 |
| }, |
| { |
| "epoch": 4.8731241473397, |
| "grad_norm": 0.07571887859535044, |
| "learning_rate": 1.5518007629571342e-07, |
| "loss": 0.3995, |
| "step": 1004 |
| }, |
| { |
| "epoch": 4.877974837047143, |
| "grad_norm": 0.07657090064859913, |
| "learning_rate": 1.4347970795054456e-07, |
| "loss": 0.3967, |
| "step": 1005 |
| }, |
| { |
| "epoch": 4.882825526754585, |
| "grad_norm": 0.07475277823538463, |
| "learning_rate": 1.3223710108158483e-07, |
| "loss": 0.3965, |
| "step": 1006 |
| }, |
| { |
| "epoch": 4.887676216462028, |
| "grad_norm": 0.07465351831333361, |
| "learning_rate": 1.214523848128124e-07, |
| "loss": 0.3971, |
| "step": 1007 |
| }, |
| { |
| "epoch": 4.892526906169471, |
| "grad_norm": 0.07637388169909817, |
| "learning_rate": 1.111256830092211e-07, |
| "loss": 0.3996, |
| "step": 1008 |
| }, |
| { |
| "epoch": 4.897377595876914, |
| "grad_norm": 0.07639665552172381, |
| "learning_rate": 1.0125711427540374e-07, |
| "loss": 0.3949, |
| "step": 1009 |
| }, |
| { |
| "epoch": 4.902228285584356, |
| "grad_norm": 0.07373092012101537, |
| "learning_rate": 9.184679195417989e-08, |
| "loss": 0.3889, |
| "step": 1010 |
| }, |
| { |
| "epoch": 4.9070789752918, |
| "grad_norm": 0.07583698163385665, |
| "learning_rate": 8.289482412531246e-08, |
| "loss": 0.3984, |
| "step": 1011 |
| }, |
| { |
| "epoch": 4.911929664999242, |
| "grad_norm": 0.07647787343235872, |
| "learning_rate": 7.440131360424652e-08, |
| "loss": 0.3887, |
| "step": 1012 |
| }, |
| { |
| "epoch": 4.916780354706685, |
| "grad_norm": 0.07751142464425709, |
| "learning_rate": 6.636635794094126e-08, |
| "loss": 0.3908, |
| "step": 1013 |
| }, |
| { |
| "epoch": 4.921631044414128, |
| "grad_norm": 0.07884353160777417, |
| "learning_rate": 5.879004941874655e-08, |
| "loss": 0.3951, |
| "step": 1014 |
| }, |
| { |
| "epoch": 4.9264817341215705, |
| "grad_norm": 0.07661585400143268, |
| "learning_rate": 5.16724750533415e-08, |
| "loss": 0.398, |
| "step": 1015 |
| }, |
| { |
| "epoch": 4.931332423829013, |
| "grad_norm": 0.07506936516416932, |
| "learning_rate": 4.5013716591730815e-08, |
| "loss": 0.3944, |
| "step": 1016 |
| }, |
| { |
| "epoch": 4.9361831135364564, |
| "grad_norm": 0.07625120541429038, |
| "learning_rate": 3.881385051132114e-08, |
| "loss": 0.3997, |
| "step": 1017 |
| }, |
| { |
| "epoch": 4.941033803243899, |
| "grad_norm": 0.0754548388472603, |
| "learning_rate": 3.307294801902838e-08, |
| "loss": 0.3968, |
| "step": 1018 |
| }, |
| { |
| "epoch": 4.9458844929513415, |
| "grad_norm": 0.07563871272528049, |
| "learning_rate": 2.7791075050460636e-08, |
| "loss": 0.3977, |
| "step": 1019 |
| }, |
| { |
| "epoch": 4.950735182658784, |
| "grad_norm": 0.0760975342061979, |
| "learning_rate": 2.2968292269167637e-08, |
| "loss": 0.3978, |
| "step": 1020 |
| }, |
| { |
| "epoch": 4.955585872366227, |
| "grad_norm": 0.07484229156933196, |
| "learning_rate": 1.8604655065939116e-08, |
| "loss": 0.3888, |
| "step": 1021 |
| }, |
| { |
| "epoch": 4.96043656207367, |
| "grad_norm": 0.07489129192172529, |
| "learning_rate": 1.470021355816975e-08, |
| "loss": 0.3952, |
| "step": 1022 |
| }, |
| { |
| "epoch": 4.965287251781112, |
| "grad_norm": 0.07561234415729748, |
| "learning_rate": 1.1255012589286297e-08, |
| "loss": 0.3988, |
| "step": 1023 |
| }, |
| { |
| "epoch": 4.970137941488556, |
| "grad_norm": 0.07589712959511802, |
| "learning_rate": 8.269091728232426e-09, |
| "loss": 0.3858, |
| "step": 1024 |
| }, |
| { |
| "epoch": 4.974988631195998, |
| "grad_norm": 0.07684966203710471, |
| "learning_rate": 5.742485269006892e-09, |
| "loss": 0.3976, |
| "step": 1025 |
| }, |
| { |
| "epoch": 4.979839320903441, |
| "grad_norm": 0.07450707523151534, |
| "learning_rate": 3.6752222302727238e-09, |
| "loss": 0.3922, |
| "step": 1026 |
| }, |
| { |
| "epoch": 4.984690010610883, |
| "grad_norm": 0.0754634711281639, |
| "learning_rate": 2.06732635503748e-09, |
| "loss": 0.3894, |
| "step": 1027 |
| }, |
| { |
| "epoch": 4.989540700318327, |
| "grad_norm": 0.07460842756020492, |
| "learning_rate": 9.188161103557136e-10, |
| "loss": 0.3893, |
| "step": 1028 |
| }, |
| { |
| "epoch": 4.994391390025769, |
| "grad_norm": 0.07411367814575705, |
| "learning_rate": 2.2970468714245132e-10, |
| "loss": 0.3934, |
| "step": 1029 |
| }, |
| { |
| "epoch": 4.999242079733212, |
| "grad_norm": 0.07533391242414708, |
| "learning_rate": 0.0, |
| "loss": 0.3877, |
| "step": 1030 |
| }, |
| { |
| "epoch": 4.999242079733212, |
| "step": 1030, |
| "total_flos": 2.739131934768418e+19, |
| "train_loss": 0.07882811409755817, |
| "train_runtime": 48155.7887, |
| "train_samples_per_second": 10.958, |
| "train_steps_per_second": 0.021 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 1030, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 5, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 2.739131934768418e+19, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|