{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.9970788704965923, "eval_steps": 500, "global_step": 1026, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0029211295034079843, "grad_norm": 6.789194746397651, "learning_rate": 9.70873786407767e-08, "loss": 1.0797, "step": 1 }, { "epoch": 0.005842259006815969, "grad_norm": 6.623744256841628, "learning_rate": 1.941747572815534e-07, "loss": 1.0936, "step": 2 }, { "epoch": 0.008763388510223954, "grad_norm": 6.814087231706784, "learning_rate": 2.9126213592233014e-07, "loss": 1.0915, "step": 3 }, { "epoch": 0.011684518013631937, "grad_norm": 6.979864404527602, "learning_rate": 3.883495145631068e-07, "loss": 1.113, "step": 4 }, { "epoch": 0.014605647517039922, "grad_norm": 7.130506329535252, "learning_rate": 4.854368932038835e-07, "loss": 1.1201, "step": 5 }, { "epoch": 0.017526777020447908, "grad_norm": 6.457316225305797, "learning_rate": 5.825242718446603e-07, "loss": 1.0589, "step": 6 }, { "epoch": 0.02044790652385589, "grad_norm": 6.634803833935598, "learning_rate": 6.79611650485437e-07, "loss": 1.0948, "step": 7 }, { "epoch": 0.023369036027263874, "grad_norm": 6.51924054198785, "learning_rate": 7.766990291262136e-07, "loss": 1.1033, "step": 8 }, { "epoch": 0.02629016553067186, "grad_norm": 6.172807303107381, "learning_rate": 8.737864077669904e-07, "loss": 1.1025, "step": 9 }, { "epoch": 0.029211295034079845, "grad_norm": 6.123481829151969, "learning_rate": 9.70873786407767e-07, "loss": 1.0593, "step": 10 }, { "epoch": 0.03213242453748783, "grad_norm": 5.067066736988861, "learning_rate": 1.0679611650485437e-06, "loss": 1.0514, "step": 11 }, { "epoch": 0.035053554040895815, "grad_norm": 5.316963931526159, "learning_rate": 1.1650485436893206e-06, "loss": 1.0649, "step": 12 }, { "epoch": 0.0379746835443038, "grad_norm": 4.8023708760386326, "learning_rate": 1.2621359223300972e-06, "loss": 1.0595, "step": 13 }, { "epoch": 0.04089581304771178, "grad_norm": 4.307477266760391, "learning_rate": 1.359223300970874e-06, "loss": 1.0045, "step": 14 }, { "epoch": 0.043816942551119765, "grad_norm": 2.9737946154856254, "learning_rate": 1.4563106796116506e-06, "loss": 0.9864, "step": 15 }, { "epoch": 0.04673807205452775, "grad_norm": 2.969891758892917, "learning_rate": 1.5533980582524272e-06, "loss": 1.0038, "step": 16 }, { "epoch": 0.04965920155793573, "grad_norm": 2.6410254985919264, "learning_rate": 1.650485436893204e-06, "loss": 0.9688, "step": 17 }, { "epoch": 0.05258033106134372, "grad_norm": 2.6767839969522385, "learning_rate": 1.7475728155339808e-06, "loss": 0.9777, "step": 18 }, { "epoch": 0.055501460564751706, "grad_norm": 2.3845837305117867, "learning_rate": 1.8446601941747574e-06, "loss": 0.9803, "step": 19 }, { "epoch": 0.05842259006815969, "grad_norm": 2.485668257022799, "learning_rate": 1.941747572815534e-06, "loss": 0.9512, "step": 20 }, { "epoch": 0.06134371957156767, "grad_norm": 2.9203150216318057, "learning_rate": 2.0388349514563107e-06, "loss": 0.93, "step": 21 }, { "epoch": 0.06426484907497566, "grad_norm": 2.868722547204036, "learning_rate": 2.1359223300970874e-06, "loss": 0.9171, "step": 22 }, { "epoch": 0.06718597857838364, "grad_norm": 2.679648431038279, "learning_rate": 2.2330097087378645e-06, "loss": 0.9199, "step": 23 }, { "epoch": 0.07010710808179163, "grad_norm": 2.459073576199394, "learning_rate": 2.330097087378641e-06, "loss": 0.9157, "step": 24 }, { "epoch": 0.0730282375851996, "grad_norm": 2.014055908464458, "learning_rate": 2.427184466019418e-06, "loss": 0.9164, "step": 25 }, { "epoch": 0.0759493670886076, "grad_norm": 1.5955860458303692, "learning_rate": 2.5242718446601945e-06, "loss": 0.8909, "step": 26 }, { "epoch": 0.07887049659201557, "grad_norm": 1.3790379745407235, "learning_rate": 2.621359223300971e-06, "loss": 0.8631, "step": 27 }, { "epoch": 0.08179162609542356, "grad_norm": 1.6462939916147095, "learning_rate": 2.718446601941748e-06, "loss": 0.8384, "step": 28 }, { "epoch": 0.08471275559883155, "grad_norm": 1.7416274692927092, "learning_rate": 2.8155339805825245e-06, "loss": 0.8643, "step": 29 }, { "epoch": 0.08763388510223953, "grad_norm": 1.4689264445022938, "learning_rate": 2.912621359223301e-06, "loss": 0.8401, "step": 30 }, { "epoch": 0.09055501460564752, "grad_norm": 1.2970139226424346, "learning_rate": 3.0097087378640778e-06, "loss": 0.8423, "step": 31 }, { "epoch": 0.0934761441090555, "grad_norm": 1.1674121947058942, "learning_rate": 3.1067961165048544e-06, "loss": 0.837, "step": 32 }, { "epoch": 0.09639727361246349, "grad_norm": 1.1908632755730892, "learning_rate": 3.2038834951456315e-06, "loss": 0.8203, "step": 33 }, { "epoch": 0.09931840311587146, "grad_norm": 1.1042928737436872, "learning_rate": 3.300970873786408e-06, "loss": 0.799, "step": 34 }, { "epoch": 0.10223953261927946, "grad_norm": 1.1603936762022113, "learning_rate": 3.398058252427185e-06, "loss": 0.8303, "step": 35 }, { "epoch": 0.10516066212268745, "grad_norm": 1.0681312398436076, "learning_rate": 3.4951456310679615e-06, "loss": 0.8052, "step": 36 }, { "epoch": 0.10808179162609542, "grad_norm": 0.8925175456575719, "learning_rate": 3.592233009708738e-06, "loss": 0.8069, "step": 37 }, { "epoch": 0.11100292112950341, "grad_norm": 0.8822021824942768, "learning_rate": 3.689320388349515e-06, "loss": 0.7812, "step": 38 }, { "epoch": 0.11392405063291139, "grad_norm": 0.8935693976115221, "learning_rate": 3.7864077669902915e-06, "loss": 0.7787, "step": 39 }, { "epoch": 0.11684518013631938, "grad_norm": 0.9243397791705507, "learning_rate": 3.883495145631068e-06, "loss": 0.7936, "step": 40 }, { "epoch": 0.11976630963972736, "grad_norm": 0.8181188534202998, "learning_rate": 3.980582524271845e-06, "loss": 0.7752, "step": 41 }, { "epoch": 0.12268743914313535, "grad_norm": 0.723717052367605, "learning_rate": 4.0776699029126215e-06, "loss": 0.7544, "step": 42 }, { "epoch": 0.12560856864654332, "grad_norm": 0.7268642480625847, "learning_rate": 4.1747572815533986e-06, "loss": 0.7871, "step": 43 }, { "epoch": 0.12852969814995133, "grad_norm": 0.7154200711509474, "learning_rate": 4.271844660194175e-06, "loss": 0.7617, "step": 44 }, { "epoch": 0.1314508276533593, "grad_norm": 0.7425189888635864, "learning_rate": 4.368932038834952e-06, "loss": 0.7661, "step": 45 }, { "epoch": 0.13437195715676728, "grad_norm": 0.6964525874804132, "learning_rate": 4.466019417475729e-06, "loss": 0.7578, "step": 46 }, { "epoch": 0.13729308666017526, "grad_norm": 0.7077568254698756, "learning_rate": 4.563106796116505e-06, "loss": 0.7466, "step": 47 }, { "epoch": 0.14021421616358326, "grad_norm": 0.728109048064247, "learning_rate": 4.660194174757282e-06, "loss": 0.7536, "step": 48 }, { "epoch": 0.14313534566699124, "grad_norm": 0.7073551382013681, "learning_rate": 4.7572815533980585e-06, "loss": 0.7386, "step": 49 }, { "epoch": 0.1460564751703992, "grad_norm": 0.6810942703152736, "learning_rate": 4.854368932038836e-06, "loss": 0.7771, "step": 50 }, { "epoch": 0.14897760467380722, "grad_norm": 0.7280450266974076, "learning_rate": 4.951456310679612e-06, "loss": 0.739, "step": 51 }, { "epoch": 0.1518987341772152, "grad_norm": 0.7021483972343961, "learning_rate": 5.048543689320389e-06, "loss": 0.7437, "step": 52 }, { "epoch": 0.15481986368062317, "grad_norm": 0.7167957676962433, "learning_rate": 5.145631067961165e-06, "loss": 0.764, "step": 53 }, { "epoch": 0.15774099318403115, "grad_norm": 0.6140023964252928, "learning_rate": 5.242718446601942e-06, "loss": 0.7338, "step": 54 }, { "epoch": 0.16066212268743915, "grad_norm": 0.7304557287827438, "learning_rate": 5.3398058252427185e-06, "loss": 0.7815, "step": 55 }, { "epoch": 0.16358325219084713, "grad_norm": 0.6396185255120146, "learning_rate": 5.436893203883496e-06, "loss": 0.7349, "step": 56 }, { "epoch": 0.1665043816942551, "grad_norm": 0.6558732255969539, "learning_rate": 5.533980582524272e-06, "loss": 0.7212, "step": 57 }, { "epoch": 0.1694255111976631, "grad_norm": 0.5940933645304805, "learning_rate": 5.631067961165049e-06, "loss": 0.7577, "step": 58 }, { "epoch": 0.17234664070107109, "grad_norm": 0.6218585913117457, "learning_rate": 5.728155339805825e-06, "loss": 0.7431, "step": 59 }, { "epoch": 0.17526777020447906, "grad_norm": 0.6908136674947283, "learning_rate": 5.825242718446602e-06, "loss": 0.7368, "step": 60 }, { "epoch": 0.17818889970788704, "grad_norm": 0.6522752947974526, "learning_rate": 5.9223300970873785e-06, "loss": 0.7126, "step": 61 }, { "epoch": 0.18111002921129504, "grad_norm": 0.6780346336214896, "learning_rate": 6.0194174757281556e-06, "loss": 0.7367, "step": 62 }, { "epoch": 0.18403115871470302, "grad_norm": 0.7722253573433183, "learning_rate": 6.116504854368932e-06, "loss": 0.7228, "step": 63 }, { "epoch": 0.186952288218111, "grad_norm": 0.6399768773236866, "learning_rate": 6.213592233009709e-06, "loss": 0.7244, "step": 64 }, { "epoch": 0.189873417721519, "grad_norm": 0.6674128228125541, "learning_rate": 6.310679611650487e-06, "loss": 0.7376, "step": 65 }, { "epoch": 0.19279454722492698, "grad_norm": 0.6775138710637573, "learning_rate": 6.407766990291263e-06, "loss": 0.7286, "step": 66 }, { "epoch": 0.19571567672833495, "grad_norm": 0.7443299390748634, "learning_rate": 6.50485436893204e-06, "loss": 0.7123, "step": 67 }, { "epoch": 0.19863680623174293, "grad_norm": 0.6712217876186718, "learning_rate": 6.601941747572816e-06, "loss": 0.7118, "step": 68 }, { "epoch": 0.20155793573515093, "grad_norm": 0.6173482204736721, "learning_rate": 6.6990291262135935e-06, "loss": 0.7141, "step": 69 }, { "epoch": 0.2044790652385589, "grad_norm": 0.7586706797857404, "learning_rate": 6.79611650485437e-06, "loss": 0.7182, "step": 70 }, { "epoch": 0.2074001947419669, "grad_norm": 0.6711328351671544, "learning_rate": 6.893203883495147e-06, "loss": 0.7182, "step": 71 }, { "epoch": 0.2103213242453749, "grad_norm": 0.6423529972707454, "learning_rate": 6.990291262135923e-06, "loss": 0.7147, "step": 72 }, { "epoch": 0.21324245374878287, "grad_norm": 0.7175157009799245, "learning_rate": 7.0873786407767e-06, "loss": 0.7078, "step": 73 }, { "epoch": 0.21616358325219084, "grad_norm": 0.7156449836663106, "learning_rate": 7.184466019417476e-06, "loss": 0.7026, "step": 74 }, { "epoch": 0.21908471275559882, "grad_norm": 0.591120785534527, "learning_rate": 7.2815533980582534e-06, "loss": 0.6996, "step": 75 }, { "epoch": 0.22200584225900682, "grad_norm": 0.6568235675952798, "learning_rate": 7.37864077669903e-06, "loss": 0.7098, "step": 76 }, { "epoch": 0.2249269717624148, "grad_norm": 0.6969907394816692, "learning_rate": 7.475728155339807e-06, "loss": 0.7112, "step": 77 }, { "epoch": 0.22784810126582278, "grad_norm": 0.674624972312595, "learning_rate": 7.572815533980583e-06, "loss": 0.6905, "step": 78 }, { "epoch": 0.23076923076923078, "grad_norm": 0.6004655479528318, "learning_rate": 7.66990291262136e-06, "loss": 0.6987, "step": 79 }, { "epoch": 0.23369036027263876, "grad_norm": 0.672439140786889, "learning_rate": 7.766990291262136e-06, "loss": 0.7059, "step": 80 }, { "epoch": 0.23661148977604674, "grad_norm": 0.6379167803971234, "learning_rate": 7.864077669902913e-06, "loss": 0.6993, "step": 81 }, { "epoch": 0.2395326192794547, "grad_norm": 0.624979376741808, "learning_rate": 7.96116504854369e-06, "loss": 0.6961, "step": 82 }, { "epoch": 0.24245374878286272, "grad_norm": 0.5967840653189634, "learning_rate": 8.058252427184466e-06, "loss": 0.7115, "step": 83 }, { "epoch": 0.2453748782862707, "grad_norm": 0.6448351126797235, "learning_rate": 8.155339805825243e-06, "loss": 0.7058, "step": 84 }, { "epoch": 0.24829600778967867, "grad_norm": 0.5456246974361787, "learning_rate": 8.25242718446602e-06, "loss": 0.6971, "step": 85 }, { "epoch": 0.25121713729308665, "grad_norm": 0.5981523190285354, "learning_rate": 8.349514563106797e-06, "loss": 0.6644, "step": 86 }, { "epoch": 0.25413826679649465, "grad_norm": 0.6094281430676193, "learning_rate": 8.446601941747573e-06, "loss": 0.6822, "step": 87 }, { "epoch": 0.25705939629990265, "grad_norm": 0.645486530934357, "learning_rate": 8.54368932038835e-06, "loss": 0.6767, "step": 88 }, { "epoch": 0.2599805258033106, "grad_norm": 0.6225273389721123, "learning_rate": 8.640776699029127e-06, "loss": 0.7152, "step": 89 }, { "epoch": 0.2629016553067186, "grad_norm": 0.6432259026110182, "learning_rate": 8.737864077669904e-06, "loss": 0.7159, "step": 90 }, { "epoch": 0.26582278481012656, "grad_norm": 0.6162731782600254, "learning_rate": 8.834951456310681e-06, "loss": 0.7132, "step": 91 }, { "epoch": 0.26874391431353456, "grad_norm": 0.6542304058964258, "learning_rate": 8.932038834951458e-06, "loss": 0.6809, "step": 92 }, { "epoch": 0.27166504381694256, "grad_norm": 0.6079172030969984, "learning_rate": 9.029126213592233e-06, "loss": 0.6824, "step": 93 }, { "epoch": 0.2745861733203505, "grad_norm": 0.6519383743853913, "learning_rate": 9.12621359223301e-06, "loss": 0.6899, "step": 94 }, { "epoch": 0.2775073028237585, "grad_norm": 0.7282902638094394, "learning_rate": 9.223300970873788e-06, "loss": 0.6922, "step": 95 }, { "epoch": 0.2804284323271665, "grad_norm": 0.5989786182176935, "learning_rate": 9.320388349514565e-06, "loss": 0.6734, "step": 96 }, { "epoch": 0.28334956183057447, "grad_norm": 0.7160709546948157, "learning_rate": 9.41747572815534e-06, "loss": 0.6815, "step": 97 }, { "epoch": 0.2862706913339825, "grad_norm": 0.6813096412081009, "learning_rate": 9.514563106796117e-06, "loss": 0.6885, "step": 98 }, { "epoch": 0.2891918208373905, "grad_norm": 0.6540975722149734, "learning_rate": 9.611650485436894e-06, "loss": 0.69, "step": 99 }, { "epoch": 0.2921129503407984, "grad_norm": 0.8050571281257926, "learning_rate": 9.708737864077671e-06, "loss": 0.691, "step": 100 }, { "epoch": 0.29503407984420643, "grad_norm": 0.6433559989032654, "learning_rate": 9.805825242718447e-06, "loss": 0.6986, "step": 101 }, { "epoch": 0.29795520934761444, "grad_norm": 0.8412736504392738, "learning_rate": 9.902912621359224e-06, "loss": 0.6926, "step": 102 }, { "epoch": 0.3008763388510224, "grad_norm": 0.7443726086923518, "learning_rate": 1e-05, "loss": 0.6906, "step": 103 }, { "epoch": 0.3037974683544304, "grad_norm": 0.6597015006842325, "learning_rate": 9.999971037507608e-06, "loss": 0.675, "step": 104 }, { "epoch": 0.30671859785783834, "grad_norm": 0.5931435595661035, "learning_rate": 9.99988415036596e-06, "loss": 0.6802, "step": 105 }, { "epoch": 0.30963972736124634, "grad_norm": 0.7214217523040783, "learning_rate": 9.99973933958164e-06, "loss": 0.7041, "step": 106 }, { "epoch": 0.31256085686465435, "grad_norm": 0.7234513096207073, "learning_rate": 9.999536606832288e-06, "loss": 0.6872, "step": 107 }, { "epoch": 0.3154819863680623, "grad_norm": 0.7879752038918911, "learning_rate": 9.999275954466555e-06, "loss": 0.6873, "step": 108 }, { "epoch": 0.3184031158714703, "grad_norm": 0.6655845938433153, "learning_rate": 9.998957385504103e-06, "loss": 0.6976, "step": 109 }, { "epoch": 0.3213242453748783, "grad_norm": 0.8522730059744493, "learning_rate": 9.99858090363555e-06, "loss": 0.6719, "step": 110 }, { "epoch": 0.32424537487828625, "grad_norm": 0.6292219731062122, "learning_rate": 9.998146513222436e-06, "loss": 0.6993, "step": 111 }, { "epoch": 0.32716650438169426, "grad_norm": 0.7181176135878521, "learning_rate": 9.997654219297176e-06, "loss": 0.6901, "step": 112 }, { "epoch": 0.33008763388510226, "grad_norm": 0.6962928810640735, "learning_rate": 9.997104027562991e-06, "loss": 0.6951, "step": 113 }, { "epoch": 0.3330087633885102, "grad_norm": 0.6849230790401417, "learning_rate": 9.996495944393853e-06, "loss": 0.6828, "step": 114 }, { "epoch": 0.3359298928919182, "grad_norm": 0.7135171270031142, "learning_rate": 9.995829976834402e-06, "loss": 0.6737, "step": 115 }, { "epoch": 0.3388510223953262, "grad_norm": 0.6814570642632325, "learning_rate": 9.995106132599869e-06, "loss": 0.6875, "step": 116 }, { "epoch": 0.34177215189873417, "grad_norm": 0.8248995841087691, "learning_rate": 9.99432442007599e-06, "loss": 0.6873, "step": 117 }, { "epoch": 0.34469328140214217, "grad_norm": 0.6919031766051941, "learning_rate": 9.993484848318899e-06, "loss": 0.6835, "step": 118 }, { "epoch": 0.3476144109055501, "grad_norm": 0.7748697122331325, "learning_rate": 9.992587427055036e-06, "loss": 0.6933, "step": 119 }, { "epoch": 0.3505355404089581, "grad_norm": 0.6232356121853884, "learning_rate": 9.99163216668102e-06, "loss": 0.6634, "step": 120 }, { "epoch": 0.35345666991236613, "grad_norm": 0.8152682417581196, "learning_rate": 9.990619078263543e-06, "loss": 0.6833, "step": 121 }, { "epoch": 0.3563777994157741, "grad_norm": 0.6829212246748637, "learning_rate": 9.989548173539229e-06, "loss": 0.6904, "step": 122 }, { "epoch": 0.3592989289191821, "grad_norm": 0.7764702428398512, "learning_rate": 9.988419464914505e-06, "loss": 0.6911, "step": 123 }, { "epoch": 0.3622200584225901, "grad_norm": 0.792332480824063, "learning_rate": 9.98723296546546e-06, "loss": 0.6817, "step": 124 }, { "epoch": 0.36514118792599803, "grad_norm": 0.6594623726005864, "learning_rate": 9.985988688937684e-06, "loss": 0.6873, "step": 125 }, { "epoch": 0.36806231742940604, "grad_norm": 0.8269823477396988, "learning_rate": 9.984686649746119e-06, "loss": 0.693, "step": 126 }, { "epoch": 0.37098344693281404, "grad_norm": 0.6422126697095933, "learning_rate": 9.983326862974882e-06, "loss": 0.6576, "step": 127 }, { "epoch": 0.373904576436222, "grad_norm": 0.8200568586438982, "learning_rate": 9.981909344377101e-06, "loss": 0.6929, "step": 128 }, { "epoch": 0.37682570593963, "grad_norm": 0.7400126500706381, "learning_rate": 9.980434110374725e-06, "loss": 0.6557, "step": 129 }, { "epoch": 0.379746835443038, "grad_norm": 0.713854878991036, "learning_rate": 9.978901178058333e-06, "loss": 0.6942, "step": 130 }, { "epoch": 0.38266796494644595, "grad_norm": 0.7483541113835968, "learning_rate": 9.977310565186945e-06, "loss": 0.6781, "step": 131 }, { "epoch": 0.38558909444985395, "grad_norm": 0.7131870241688308, "learning_rate": 9.975662290187802e-06, "loss": 0.6564, "step": 132 }, { "epoch": 0.3885102239532619, "grad_norm": 0.7880618680074065, "learning_rate": 9.973956372156166e-06, "loss": 0.6752, "step": 133 }, { "epoch": 0.3914313534566699, "grad_norm": 0.5977156256261835, "learning_rate": 9.972192830855095e-06, "loss": 0.6763, "step": 134 }, { "epoch": 0.3943524829600779, "grad_norm": 0.76186235107972, "learning_rate": 9.970371686715205e-06, "loss": 0.7014, "step": 135 }, { "epoch": 0.39727361246348586, "grad_norm": 0.8438134698760479, "learning_rate": 9.96849296083445e-06, "loss": 0.6902, "step": 136 }, { "epoch": 0.40019474196689386, "grad_norm": 0.6120514167187466, "learning_rate": 9.966556674977864e-06, "loss": 0.663, "step": 137 }, { "epoch": 0.40311587147030187, "grad_norm": 0.7891889540366791, "learning_rate": 9.964562851577307e-06, "loss": 0.6458, "step": 138 }, { "epoch": 0.4060370009737098, "grad_norm": 0.8381022052177086, "learning_rate": 9.962511513731219e-06, "loss": 0.6728, "step": 139 }, { "epoch": 0.4089581304771178, "grad_norm": 0.7489332182994131, "learning_rate": 9.960402685204347e-06, "loss": 0.6585, "step": 140 }, { "epoch": 0.4118792599805258, "grad_norm": 0.804869829411989, "learning_rate": 9.958236390427458e-06, "loss": 0.6784, "step": 141 }, { "epoch": 0.4148003894839338, "grad_norm": 0.7474713341428297, "learning_rate": 9.956012654497073e-06, "loss": 0.6488, "step": 142 }, { "epoch": 0.4177215189873418, "grad_norm": 0.8115239389486957, "learning_rate": 9.953731503175166e-06, "loss": 0.6894, "step": 143 }, { "epoch": 0.4206426484907498, "grad_norm": 0.781168520393115, "learning_rate": 9.951392962888868e-06, "loss": 0.6534, "step": 144 }, { "epoch": 0.42356377799415773, "grad_norm": 0.7722795439014734, "learning_rate": 9.948997060730161e-06, "loss": 0.6504, "step": 145 }, { "epoch": 0.42648490749756574, "grad_norm": 0.8668435801796398, "learning_rate": 9.946543824455563e-06, "loss": 0.6507, "step": 146 }, { "epoch": 0.4294060370009737, "grad_norm": 0.8391877814865175, "learning_rate": 9.94403328248581e-06, "loss": 0.6702, "step": 147 }, { "epoch": 0.4323271665043817, "grad_norm": 0.6805575537389376, "learning_rate": 9.941465463905522e-06, "loss": 0.6744, "step": 148 }, { "epoch": 0.4352482960077897, "grad_norm": 0.7033559759136386, "learning_rate": 9.938840398462872e-06, "loss": 0.6732, "step": 149 }, { "epoch": 0.43816942551119764, "grad_norm": 0.8071774421297472, "learning_rate": 9.936158116569231e-06, "loss": 0.6704, "step": 150 }, { "epoch": 0.44109055501460565, "grad_norm": 0.7788309365778397, "learning_rate": 9.933418649298831e-06, "loss": 0.6697, "step": 151 }, { "epoch": 0.44401168451801365, "grad_norm": 0.633484330960411, "learning_rate": 9.930622028388388e-06, "loss": 0.6533, "step": 152 }, { "epoch": 0.4469328140214216, "grad_norm": 0.6677671353777382, "learning_rate": 9.92776828623675e-06, "loss": 0.6323, "step": 153 }, { "epoch": 0.4498539435248296, "grad_norm": 0.6580113405415786, "learning_rate": 9.924857455904511e-06, "loss": 0.6569, "step": 154 }, { "epoch": 0.4527750730282376, "grad_norm": 0.7153967279296722, "learning_rate": 9.921889571113629e-06, "loss": 0.6651, "step": 155 }, { "epoch": 0.45569620253164556, "grad_norm": 0.6507727168616313, "learning_rate": 9.918864666247042e-06, "loss": 0.6709, "step": 156 }, { "epoch": 0.45861733203505356, "grad_norm": 0.6614595067208825, "learning_rate": 9.915782776348263e-06, "loss": 0.6558, "step": 157 }, { "epoch": 0.46153846153846156, "grad_norm": 0.8137858393206525, "learning_rate": 9.912643937120978e-06, "loss": 0.6756, "step": 158 }, { "epoch": 0.4644595910418695, "grad_norm": 0.6306212983973334, "learning_rate": 9.909448184928629e-06, "loss": 0.6416, "step": 159 }, { "epoch": 0.4673807205452775, "grad_norm": 0.6836208681318199, "learning_rate": 9.906195556793996e-06, "loss": 0.6662, "step": 160 }, { "epoch": 0.47030185004868547, "grad_norm": 0.6141476817991299, "learning_rate": 9.902886090398764e-06, "loss": 0.6774, "step": 161 }, { "epoch": 0.47322297955209347, "grad_norm": 0.7052661650958192, "learning_rate": 9.899519824083095e-06, "loss": 0.6618, "step": 162 }, { "epoch": 0.4761441090555015, "grad_norm": 0.5986401947237227, "learning_rate": 9.896096796845172e-06, "loss": 0.6738, "step": 163 }, { "epoch": 0.4790652385589094, "grad_norm": 0.6916970139226505, "learning_rate": 9.892617048340754e-06, "loss": 0.6588, "step": 164 }, { "epoch": 0.4819863680623174, "grad_norm": 0.6805676670949773, "learning_rate": 9.889080618882719e-06, "loss": 0.6826, "step": 165 }, { "epoch": 0.48490749756572543, "grad_norm": 0.6658431348883423, "learning_rate": 9.88548754944059e-06, "loss": 0.6702, "step": 166 }, { "epoch": 0.4878286270691334, "grad_norm": 0.682259295410329, "learning_rate": 9.881837881640064e-06, "loss": 0.6735, "step": 167 }, { "epoch": 0.4907497565725414, "grad_norm": 0.5993509164653038, "learning_rate": 9.878131657762535e-06, "loss": 0.6545, "step": 168 }, { "epoch": 0.4936708860759494, "grad_norm": 0.7534948952289838, "learning_rate": 9.874368920744594e-06, "loss": 0.6812, "step": 169 }, { "epoch": 0.49659201557935734, "grad_norm": 0.707512873305316, "learning_rate": 9.870549714177538e-06, "loss": 0.6513, "step": 170 }, { "epoch": 0.49951314508276534, "grad_norm": 0.7200355139629895, "learning_rate": 9.866674082306861e-06, "loss": 0.6438, "step": 171 }, { "epoch": 0.5024342745861733, "grad_norm": 0.8156295486958224, "learning_rate": 9.86274207003175e-06, "loss": 0.6564, "step": 172 }, { "epoch": 0.5053554040895814, "grad_norm": 0.6117511012510686, "learning_rate": 9.858753722904552e-06, "loss": 0.6827, "step": 173 }, { "epoch": 0.5082765335929893, "grad_norm": 0.8219493047877929, "learning_rate": 9.854709087130261e-06, "loss": 0.6718, "step": 174 }, { "epoch": 0.5111976630963972, "grad_norm": 0.6662487126424073, "learning_rate": 9.850608209565967e-06, "loss": 0.6388, "step": 175 }, { "epoch": 0.5141187925998053, "grad_norm": 0.7125438710540766, "learning_rate": 9.84645113772032e-06, "loss": 0.6589, "step": 176 }, { "epoch": 0.5170399221032133, "grad_norm": 0.7487609228616714, "learning_rate": 9.842237919752994e-06, "loss": 0.6544, "step": 177 }, { "epoch": 0.5199610516066212, "grad_norm": 0.6468420309416394, "learning_rate": 9.8379686044741e-06, "loss": 0.6565, "step": 178 }, { "epoch": 0.5228821811100293, "grad_norm": 0.7546649500098744, "learning_rate": 9.833643241343642e-06, "loss": 0.6647, "step": 179 }, { "epoch": 0.5258033106134372, "grad_norm": 0.6258186572488958, "learning_rate": 9.829261880470941e-06, "loss": 0.6392, "step": 180 }, { "epoch": 0.5287244401168452, "grad_norm": 0.5997447293335689, "learning_rate": 9.82482457261405e-06, "loss": 0.6398, "step": 181 }, { "epoch": 0.5316455696202531, "grad_norm": 0.6452320212378018, "learning_rate": 9.820331369179166e-06, "loss": 0.6611, "step": 182 }, { "epoch": 0.5345666991236612, "grad_norm": 0.5580735581912285, "learning_rate": 9.815782322220036e-06, "loss": 0.6548, "step": 183 }, { "epoch": 0.5374878286270691, "grad_norm": 0.6058080142971995, "learning_rate": 9.811177484437357e-06, "loss": 0.6664, "step": 184 }, { "epoch": 0.5404089581304771, "grad_norm": 0.7487147947448509, "learning_rate": 9.806516909178161e-06, "loss": 0.665, "step": 185 }, { "epoch": 0.5433300876338851, "grad_norm": 0.5994532711002538, "learning_rate": 9.801800650435194e-06, "loss": 0.6345, "step": 186 }, { "epoch": 0.5462512171372931, "grad_norm": 0.6402638834025774, "learning_rate": 9.797028762846305e-06, "loss": 0.6689, "step": 187 }, { "epoch": 0.549172346640701, "grad_norm": 0.8110120277002857, "learning_rate": 9.792201301693793e-06, "loss": 0.6623, "step": 188 }, { "epoch": 0.5520934761441091, "grad_norm": 0.6022982451173915, "learning_rate": 9.787318322903784e-06, "loss": 0.642, "step": 189 }, { "epoch": 0.555014605647517, "grad_norm": 0.8359964918822578, "learning_rate": 9.78237988304557e-06, "loss": 0.6828, "step": 190 }, { "epoch": 0.557935735150925, "grad_norm": 0.5971974835914099, "learning_rate": 9.77738603933096e-06, "loss": 0.6637, "step": 191 }, { "epoch": 0.560856864654333, "grad_norm": 0.9007553762322157, "learning_rate": 9.772336849613624e-06, "loss": 0.6489, "step": 192 }, { "epoch": 0.563777994157741, "grad_norm": 0.6097333173160772, "learning_rate": 9.767232372388406e-06, "loss": 0.6195, "step": 193 }, { "epoch": 0.5666991236611489, "grad_norm": 0.7430323349181741, "learning_rate": 9.762072666790658e-06, "loss": 0.6602, "step": 194 }, { "epoch": 0.569620253164557, "grad_norm": 0.698590106062137, "learning_rate": 9.756857792595555e-06, "loss": 0.654, "step": 195 }, { "epoch": 0.572541382667965, "grad_norm": 0.6098458915248055, "learning_rate": 9.751587810217398e-06, "loss": 0.6571, "step": 196 }, { "epoch": 0.5754625121713729, "grad_norm": 0.6600018141821303, "learning_rate": 9.746262780708919e-06, "loss": 0.6572, "step": 197 }, { "epoch": 0.578383641674781, "grad_norm": 0.6601639946678165, "learning_rate": 9.740882765760567e-06, "loss": 0.6593, "step": 198 }, { "epoch": 0.5813047711781889, "grad_norm": 0.726967683938266, "learning_rate": 9.735447827699798e-06, "loss": 0.6573, "step": 199 }, { "epoch": 0.5842259006815969, "grad_norm": 0.6076134837821863, "learning_rate": 9.729958029490353e-06, "loss": 0.6495, "step": 200 }, { "epoch": 0.5871470301850049, "grad_norm": 0.693728259825805, "learning_rate": 9.72441343473153e-06, "loss": 0.6384, "step": 201 }, { "epoch": 0.5900681596884129, "grad_norm": 0.7278423168034551, "learning_rate": 9.718814107657441e-06, "loss": 0.6584, "step": 202 }, { "epoch": 0.5929892891918208, "grad_norm": 0.7100671054561837, "learning_rate": 9.713160113136272e-06, "loss": 0.6555, "step": 203 }, { "epoch": 0.5959104186952289, "grad_norm": 0.6955835624438068, "learning_rate": 9.707451516669533e-06, "loss": 0.6581, "step": 204 }, { "epoch": 0.5988315481986368, "grad_norm": 0.6862859891275203, "learning_rate": 9.701688384391296e-06, "loss": 0.6471, "step": 205 }, { "epoch": 0.6017526777020448, "grad_norm": 0.7918106833642026, "learning_rate": 9.695870783067434e-06, "loss": 0.6351, "step": 206 }, { "epoch": 0.6046738072054528, "grad_norm": 0.762255183423834, "learning_rate": 9.689998780094839e-06, "loss": 0.6464, "step": 207 }, { "epoch": 0.6075949367088608, "grad_norm": 0.5926349601655899, "learning_rate": 9.684072443500645e-06, "loss": 0.6342, "step": 208 }, { "epoch": 0.6105160662122687, "grad_norm": 0.863234457455766, "learning_rate": 9.678091841941446e-06, "loss": 0.653, "step": 209 }, { "epoch": 0.6134371957156767, "grad_norm": 0.7588656251837851, "learning_rate": 9.672057044702492e-06, "loss": 0.6379, "step": 210 }, { "epoch": 0.6163583252190847, "grad_norm": 0.6108635991637165, "learning_rate": 9.665968121696892e-06, "loss": 0.6605, "step": 211 }, { "epoch": 0.6192794547224927, "grad_norm": 0.8356230484629192, "learning_rate": 9.659825143464798e-06, "loss": 0.6458, "step": 212 }, { "epoch": 0.6222005842259006, "grad_norm": 0.6240730332192024, "learning_rate": 9.653628181172596e-06, "loss": 0.6506, "step": 213 }, { "epoch": 0.6251217137293087, "grad_norm": 0.6689297135107584, "learning_rate": 9.647377306612075e-06, "loss": 0.6299, "step": 214 }, { "epoch": 0.6280428432327166, "grad_norm": 0.7685374427252067, "learning_rate": 9.641072592199599e-06, "loss": 0.6634, "step": 215 }, { "epoch": 0.6309639727361246, "grad_norm": 0.6331465350705314, "learning_rate": 9.634714110975263e-06, "loss": 0.6705, "step": 216 }, { "epoch": 0.6338851022395326, "grad_norm": 0.7142109675799595, "learning_rate": 9.628301936602053e-06, "loss": 0.6539, "step": 217 }, { "epoch": 0.6368062317429406, "grad_norm": 0.8729837863809322, "learning_rate": 9.62183614336499e-06, "loss": 0.6596, "step": 218 }, { "epoch": 0.6397273612463485, "grad_norm": 0.5991176634004923, "learning_rate": 9.61531680617027e-06, "loss": 0.6656, "step": 219 }, { "epoch": 0.6426484907497566, "grad_norm": 0.7145894148388473, "learning_rate": 9.608744000544392e-06, "loss": 0.6643, "step": 220 }, { "epoch": 0.6455696202531646, "grad_norm": 0.7598912110336243, "learning_rate": 9.602117802633293e-06, "loss": 0.6291, "step": 221 }, { "epoch": 0.6484907497565725, "grad_norm": 0.6815208811474045, "learning_rate": 9.595438289201453e-06, "loss": 0.6472, "step": 222 }, { "epoch": 0.6514118792599806, "grad_norm": 0.5760568859368258, "learning_rate": 9.588705537631014e-06, "loss": 0.6563, "step": 223 }, { "epoch": 0.6543330087633885, "grad_norm": 0.7022056147869815, "learning_rate": 9.581919625920886e-06, "loss": 0.6524, "step": 224 }, { "epoch": 0.6572541382667965, "grad_norm": 0.67470404537858, "learning_rate": 9.575080632685832e-06, "loss": 0.6436, "step": 225 }, { "epoch": 0.6601752677702045, "grad_norm": 0.6156654623500659, "learning_rate": 9.568188637155569e-06, "loss": 0.6256, "step": 226 }, { "epoch": 0.6630963972736125, "grad_norm": 0.6870167927139845, "learning_rate": 9.561243719173844e-06, "loss": 0.628, "step": 227 }, { "epoch": 0.6660175267770204, "grad_norm": 0.6043043393160271, "learning_rate": 9.554245959197511e-06, "loss": 0.6631, "step": 228 }, { "epoch": 0.6689386562804285, "grad_norm": 0.6424377779531785, "learning_rate": 9.5471954382956e-06, "loss": 0.6455, "step": 229 }, { "epoch": 0.6718597857838364, "grad_norm": 0.6747246252989533, "learning_rate": 9.54009223814837e-06, "loss": 0.6482, "step": 230 }, { "epoch": 0.6747809152872444, "grad_norm": 0.6198749042634925, "learning_rate": 9.532936441046376e-06, "loss": 0.6679, "step": 231 }, { "epoch": 0.6777020447906524, "grad_norm": 0.6700871145900489, "learning_rate": 9.525728129889505e-06, "loss": 0.6704, "step": 232 }, { "epoch": 0.6806231742940604, "grad_norm": 0.6486221862584836, "learning_rate": 9.51846738818602e-06, "loss": 0.6533, "step": 233 }, { "epoch": 0.6835443037974683, "grad_norm": 0.6158430372621209, "learning_rate": 9.511154300051591e-06, "loss": 0.6391, "step": 234 }, { "epoch": 0.6864654333008764, "grad_norm": 0.6059458334544432, "learning_rate": 9.503788950208324e-06, "loss": 0.6326, "step": 235 }, { "epoch": 0.6893865628042843, "grad_norm": 0.6630441105155737, "learning_rate": 9.49637142398377e-06, "loss": 0.6419, "step": 236 }, { "epoch": 0.6923076923076923, "grad_norm": 0.7121010143311686, "learning_rate": 9.48890180730995e-06, "loss": 0.6366, "step": 237 }, { "epoch": 0.6952288218111002, "grad_norm": 0.6536311654933115, "learning_rate": 9.481380186722354e-06, "loss": 0.6475, "step": 238 }, { "epoch": 0.6981499513145083, "grad_norm": 0.5870586837084283, "learning_rate": 9.473806649358929e-06, "loss": 0.6664, "step": 239 }, { "epoch": 0.7010710808179162, "grad_norm": 0.7293444660664181, "learning_rate": 9.466181282959083e-06, "loss": 0.6294, "step": 240 }, { "epoch": 0.7039922103213242, "grad_norm": 0.6803091119725557, "learning_rate": 9.458504175862665e-06, "loss": 0.6543, "step": 241 }, { "epoch": 0.7069133398247323, "grad_norm": 0.513109460376802, "learning_rate": 9.450775417008936e-06, "loss": 0.6529, "step": 242 }, { "epoch": 0.7098344693281402, "grad_norm": 0.6591044352211995, "learning_rate": 9.442995095935542e-06, "loss": 0.6485, "step": 243 }, { "epoch": 0.7127555988315482, "grad_norm": 0.5639394652214005, "learning_rate": 9.43516330277748e-06, "loss": 0.6354, "step": 244 }, { "epoch": 0.7156767283349562, "grad_norm": 0.5382276491132706, "learning_rate": 9.427280128266049e-06, "loss": 0.6338, "step": 245 }, { "epoch": 0.7185978578383642, "grad_norm": 0.5783621915913141, "learning_rate": 9.419345663727805e-06, "loss": 0.6541, "step": 246 }, { "epoch": 0.7215189873417721, "grad_norm": 0.5457758477722148, "learning_rate": 9.411360001083496e-06, "loss": 0.6649, "step": 247 }, { "epoch": 0.7244401168451802, "grad_norm": 0.5701118395223765, "learning_rate": 9.403323232846994e-06, "loss": 0.6305, "step": 248 }, { "epoch": 0.7273612463485881, "grad_norm": 0.5986045250901076, "learning_rate": 9.395235452124239e-06, "loss": 0.6315, "step": 249 }, { "epoch": 0.7302823758519961, "grad_norm": 0.5915757822980239, "learning_rate": 9.387096752612144e-06, "loss": 0.6563, "step": 250 }, { "epoch": 0.7332035053554041, "grad_norm": 0.6447044009504002, "learning_rate": 9.378907228597518e-06, "loss": 0.6543, "step": 251 }, { "epoch": 0.7361246348588121, "grad_norm": 0.6146190085950654, "learning_rate": 9.370666974955973e-06, "loss": 0.6474, "step": 252 }, { "epoch": 0.73904576436222, "grad_norm": 0.6186340229955254, "learning_rate": 9.362376087150822e-06, "loss": 0.6498, "step": 253 }, { "epoch": 0.7419668938656281, "grad_norm": 0.6697188050730257, "learning_rate": 9.354034661231976e-06, "loss": 0.629, "step": 254 }, { "epoch": 0.744888023369036, "grad_norm": 0.7166126450253048, "learning_rate": 9.345642793834825e-06, "loss": 0.6476, "step": 255 }, { "epoch": 0.747809152872444, "grad_norm": 0.5909733136537622, "learning_rate": 9.337200582179134e-06, "loss": 0.6338, "step": 256 }, { "epoch": 0.750730282375852, "grad_norm": 0.6690338784895201, "learning_rate": 9.328708124067893e-06, "loss": 0.6425, "step": 257 }, { "epoch": 0.75365141187926, "grad_norm": 0.6170552061635086, "learning_rate": 9.320165517886207e-06, "loss": 0.649, "step": 258 }, { "epoch": 0.7565725413826679, "grad_norm": 0.6767600272762853, "learning_rate": 9.31157286260014e-06, "loss": 0.6496, "step": 259 }, { "epoch": 0.759493670886076, "grad_norm": 0.7264560653599718, "learning_rate": 9.302930257755579e-06, "loss": 0.6583, "step": 260 }, { "epoch": 0.762414800389484, "grad_norm": 0.5453999254984644, "learning_rate": 9.294237803477076e-06, "loss": 0.648, "step": 261 }, { "epoch": 0.7653359298928919, "grad_norm": 0.7523663818704205, "learning_rate": 9.285495600466683e-06, "loss": 0.6488, "step": 262 }, { "epoch": 0.7682570593963, "grad_norm": 0.7485014295621598, "learning_rate": 9.276703750002801e-06, "loss": 0.6255, "step": 263 }, { "epoch": 0.7711781888997079, "grad_norm": 0.686777506781002, "learning_rate": 9.267862353938988e-06, "loss": 0.6534, "step": 264 }, { "epoch": 0.7740993184031159, "grad_norm": 0.7655220545156425, "learning_rate": 9.258971514702789e-06, "loss": 0.6439, "step": 265 }, { "epoch": 0.7770204479065238, "grad_norm": 0.5995690372205543, "learning_rate": 9.250031335294551e-06, "loss": 0.6264, "step": 266 }, { "epoch": 0.7799415774099319, "grad_norm": 0.6306408667729854, "learning_rate": 9.241041919286227e-06, "loss": 0.633, "step": 267 }, { "epoch": 0.7828627069133398, "grad_norm": 0.7405970033463782, "learning_rate": 9.232003370820171e-06, "loss": 0.6355, "step": 268 }, { "epoch": 0.7857838364167478, "grad_norm": 0.5761877077710947, "learning_rate": 9.222915794607942e-06, "loss": 0.6431, "step": 269 }, { "epoch": 0.7887049659201558, "grad_norm": 0.6417671555725009, "learning_rate": 9.213779295929082e-06, "loss": 0.6302, "step": 270 }, { "epoch": 0.7916260954235638, "grad_norm": 0.5476966311891922, "learning_rate": 9.204593980629898e-06, "loss": 0.6307, "step": 271 }, { "epoch": 0.7945472249269717, "grad_norm": 0.6325575856222458, "learning_rate": 9.195359955122244e-06, "loss": 0.6316, "step": 272 }, { "epoch": 0.7974683544303798, "grad_norm": 0.6487642715668982, "learning_rate": 9.186077326382275e-06, "loss": 0.6324, "step": 273 }, { "epoch": 0.8003894839337877, "grad_norm": 0.5751549331819923, "learning_rate": 9.176746201949216e-06, "loss": 0.6585, "step": 274 }, { "epoch": 0.8033106134371957, "grad_norm": 0.6208214920916966, "learning_rate": 9.167366689924116e-06, "loss": 0.6517, "step": 275 }, { "epoch": 0.8062317429406037, "grad_norm": 0.6579852013112687, "learning_rate": 9.157938898968594e-06, "loss": 0.643, "step": 276 }, { "epoch": 0.8091528724440117, "grad_norm": 0.5985871969783593, "learning_rate": 9.14846293830358e-06, "loss": 0.6386, "step": 277 }, { "epoch": 0.8120740019474196, "grad_norm": 0.5776495482063276, "learning_rate": 9.138938917708047e-06, "loss": 0.6367, "step": 278 }, { "epoch": 0.8149951314508277, "grad_norm": 0.6461097088775256, "learning_rate": 9.129366947517746e-06, "loss": 0.6311, "step": 279 }, { "epoch": 0.8179162609542356, "grad_norm": 0.6054324701596054, "learning_rate": 9.119747138623925e-06, "loss": 0.6365, "step": 280 }, { "epoch": 0.8208373904576436, "grad_norm": 0.6046664159813615, "learning_rate": 9.110079602472035e-06, "loss": 0.6549, "step": 281 }, { "epoch": 0.8237585199610516, "grad_norm": 0.690747949343666, "learning_rate": 9.100364451060457e-06, "loss": 0.6477, "step": 282 }, { "epoch": 0.8266796494644596, "grad_norm": 0.6352884441285447, "learning_rate": 9.090601796939192e-06, "loss": 0.6315, "step": 283 }, { "epoch": 0.8296007789678675, "grad_norm": 0.7031617950372325, "learning_rate": 9.080791753208553e-06, "loss": 0.6304, "step": 284 }, { "epoch": 0.8325219084712756, "grad_norm": 0.6955969158961154, "learning_rate": 9.070934433517872e-06, "loss": 0.6371, "step": 285 }, { "epoch": 0.8354430379746836, "grad_norm": 0.8508450672946094, "learning_rate": 9.061029952064165e-06, "loss": 0.6392, "step": 286 }, { "epoch": 0.8383641674780915, "grad_norm": 0.6682143045272909, "learning_rate": 9.05107842359082e-06, "loss": 0.6354, "step": 287 }, { "epoch": 0.8412852969814996, "grad_norm": 0.730935179569577, "learning_rate": 9.041079963386263e-06, "loss": 0.6365, "step": 288 }, { "epoch": 0.8442064264849075, "grad_norm": 0.9284592753787645, "learning_rate": 9.031034687282627e-06, "loss": 0.6512, "step": 289 }, { "epoch": 0.8471275559883155, "grad_norm": 0.5711740420462373, "learning_rate": 9.020942711654404e-06, "loss": 0.6253, "step": 290 }, { "epoch": 0.8500486854917235, "grad_norm": 0.7521676041243963, "learning_rate": 9.0108041534171e-06, "loss": 0.6346, "step": 291 }, { "epoch": 0.8529698149951315, "grad_norm": 0.7268819094436544, "learning_rate": 9.000619130025885e-06, "loss": 0.6321, "step": 292 }, { "epoch": 0.8558909444985394, "grad_norm": 0.6255002973375909, "learning_rate": 8.99038775947422e-06, "loss": 0.6448, "step": 293 }, { "epoch": 0.8588120740019474, "grad_norm": 0.5654580978730834, "learning_rate": 8.980110160292503e-06, "loss": 0.6546, "step": 294 }, { "epoch": 0.8617332035053554, "grad_norm": 0.5744283394449378, "learning_rate": 8.969786451546691e-06, "loss": 0.6354, "step": 295 }, { "epoch": 0.8646543330087634, "grad_norm": 0.5607277367845812, "learning_rate": 8.959416752836915e-06, "loss": 0.6315, "step": 296 }, { "epoch": 0.8675754625121713, "grad_norm": 0.6883032069319117, "learning_rate": 8.949001184296107e-06, "loss": 0.6284, "step": 297 }, { "epoch": 0.8704965920155794, "grad_norm": 0.5793320184156081, "learning_rate": 8.938539866588593e-06, "loss": 0.6299, "step": 298 }, { "epoch": 0.8734177215189873, "grad_norm": 0.6122270930586687, "learning_rate": 8.928032920908709e-06, "loss": 0.6463, "step": 299 }, { "epoch": 0.8763388510223953, "grad_norm": 0.6718320925279334, "learning_rate": 8.917480468979387e-06, "loss": 0.6405, "step": 300 }, { "epoch": 0.8792599805258033, "grad_norm": 0.6132181340646334, "learning_rate": 8.906882633050753e-06, "loss": 0.6426, "step": 301 }, { "epoch": 0.8821811100292113, "grad_norm": 0.6206805066838101, "learning_rate": 8.896239535898702e-06, "loss": 0.65, "step": 302 }, { "epoch": 0.8851022395326192, "grad_norm": 0.7456156741229341, "learning_rate": 8.885551300823483e-06, "loss": 0.6395, "step": 303 }, { "epoch": 0.8880233690360273, "grad_norm": 0.6186279743363227, "learning_rate": 8.874818051648267e-06, "loss": 0.6236, "step": 304 }, { "epoch": 0.8909444985394352, "grad_norm": 0.7192434268898347, "learning_rate": 8.864039912717713e-06, "loss": 0.6427, "step": 305 }, { "epoch": 0.8938656280428432, "grad_norm": 0.6528244466221247, "learning_rate": 8.853217008896526e-06, "loss": 0.6478, "step": 306 }, { "epoch": 0.8967867575462513, "grad_norm": 0.6240468261028331, "learning_rate": 8.842349465568018e-06, "loss": 0.6354, "step": 307 }, { "epoch": 0.8997078870496592, "grad_norm": 0.6086442105860419, "learning_rate": 8.831437408632639e-06, "loss": 0.6175, "step": 308 }, { "epoch": 0.9026290165530672, "grad_norm": 0.6206978097636743, "learning_rate": 8.820480964506542e-06, "loss": 0.6329, "step": 309 }, { "epoch": 0.9055501460564752, "grad_norm": 0.67571865721595, "learning_rate": 8.809480260120096e-06, "loss": 0.6302, "step": 310 }, { "epoch": 0.9084712755598832, "grad_norm": 0.612775453081801, "learning_rate": 8.798435422916425e-06, "loss": 0.6248, "step": 311 }, { "epoch": 0.9113924050632911, "grad_norm": 0.5939423168299965, "learning_rate": 8.787346580849939e-06, "loss": 0.6274, "step": 312 }, { "epoch": 0.9143135345666992, "grad_norm": 0.6578489213437513, "learning_rate": 8.776213862384838e-06, "loss": 0.6367, "step": 313 }, { "epoch": 0.9172346640701071, "grad_norm": 0.7439730364901744, "learning_rate": 8.76503739649363e-06, "loss": 0.6584, "step": 314 }, { "epoch": 0.9201557935735151, "grad_norm": 0.5758626313580656, "learning_rate": 8.753817312655642e-06, "loss": 0.638, "step": 315 }, { "epoch": 0.9230769230769231, "grad_norm": 0.7372145224476075, "learning_rate": 8.742553740855507e-06, "loss": 0.6391, "step": 316 }, { "epoch": 0.9259980525803311, "grad_norm": 0.7004119821194315, "learning_rate": 8.73124681158167e-06, "loss": 0.6426, "step": 317 }, { "epoch": 0.928919182083739, "grad_norm": 0.5393578063337247, "learning_rate": 8.719896655824878e-06, "loss": 0.6326, "step": 318 }, { "epoch": 0.9318403115871471, "grad_norm": 0.5698772754745327, "learning_rate": 8.708503405076646e-06, "loss": 0.634, "step": 319 }, { "epoch": 0.934761441090555, "grad_norm": 0.588256486663727, "learning_rate": 8.697067191327748e-06, "loss": 0.6328, "step": 320 }, { "epoch": 0.937682570593963, "grad_norm": 0.6194561945270076, "learning_rate": 8.685588147066688e-06, "loss": 0.6303, "step": 321 }, { "epoch": 0.9406037000973709, "grad_norm": 0.5910869397544244, "learning_rate": 8.67406640527816e-06, "loss": 0.6494, "step": 322 }, { "epoch": 0.943524829600779, "grad_norm": 0.6055073823945398, "learning_rate": 8.662502099441505e-06, "loss": 0.6334, "step": 323 }, { "epoch": 0.9464459591041869, "grad_norm": 0.5417594595582104, "learning_rate": 8.650895363529172e-06, "loss": 0.6279, "step": 324 }, { "epoch": 0.9493670886075949, "grad_norm": 0.5878187442362004, "learning_rate": 8.639246332005163e-06, "loss": 0.639, "step": 325 }, { "epoch": 0.952288218111003, "grad_norm": 0.5233625621707794, "learning_rate": 8.627555139823468e-06, "loss": 0.614, "step": 326 }, { "epoch": 0.9552093476144109, "grad_norm": 0.5656952115933153, "learning_rate": 8.615821922426517e-06, "loss": 0.6214, "step": 327 }, { "epoch": 0.9581304771178188, "grad_norm": 0.5806200502649499, "learning_rate": 8.604046815743598e-06, "loss": 0.6424, "step": 328 }, { "epoch": 0.9610516066212269, "grad_norm": 0.5561534960958242, "learning_rate": 8.592229956189283e-06, "loss": 0.638, "step": 329 }, { "epoch": 0.9639727361246349, "grad_norm": 0.5340425065767049, "learning_rate": 8.580371480661857e-06, "loss": 0.6238, "step": 330 }, { "epoch": 0.9668938656280428, "grad_norm": 0.5501524304300768, "learning_rate": 8.568471526541721e-06, "loss": 0.6518, "step": 331 }, { "epoch": 0.9698149951314509, "grad_norm": 0.571968183721703, "learning_rate": 8.556530231689809e-06, "loss": 0.6588, "step": 332 }, { "epoch": 0.9727361246348588, "grad_norm": 0.5870934675467651, "learning_rate": 8.544547734445983e-06, "loss": 0.6441, "step": 333 }, { "epoch": 0.9756572541382668, "grad_norm": 0.5633142366316923, "learning_rate": 8.532524173627438e-06, "loss": 0.6558, "step": 334 }, { "epoch": 0.9785783836416748, "grad_norm": 0.6080503864783372, "learning_rate": 8.520459688527091e-06, "loss": 0.6239, "step": 335 }, { "epoch": 0.9814995131450828, "grad_norm": 0.6010534017830508, "learning_rate": 8.508354418911966e-06, "loss": 0.6392, "step": 336 }, { "epoch": 0.9844206426484907, "grad_norm": 0.6110941694269748, "learning_rate": 8.496208505021572e-06, "loss": 0.6334, "step": 337 }, { "epoch": 0.9873417721518988, "grad_norm": 0.5748899254330844, "learning_rate": 8.484022087566284e-06, "loss": 0.6213, "step": 338 }, { "epoch": 0.9902629016553067, "grad_norm": 0.5910067428214469, "learning_rate": 8.471795307725713e-06, "loss": 0.6313, "step": 339 }, { "epoch": 0.9931840311587147, "grad_norm": 0.6409845798522262, "learning_rate": 8.459528307147066e-06, "loss": 0.6223, "step": 340 }, { "epoch": 0.9961051606621227, "grad_norm": 0.5798933129205326, "learning_rate": 8.447221227943507e-06, "loss": 0.6375, "step": 341 }, { "epoch": 0.9990262901655307, "grad_norm": 0.5323356127721257, "learning_rate": 8.434874212692513e-06, "loss": 0.631, "step": 342 }, { "epoch": 1.0019474196689386, "grad_norm": 1.1893567012708939, "learning_rate": 8.422487404434214e-06, "loss": 1.0195, "step": 343 }, { "epoch": 1.0048685491723466, "grad_norm": 0.6455689118062985, "learning_rate": 8.41006094666975e-06, "loss": 0.5891, "step": 344 }, { "epoch": 1.0077896786757545, "grad_norm": 0.6583690984771934, "learning_rate": 8.397594983359591e-06, "loss": 0.5549, "step": 345 }, { "epoch": 1.0107108081791627, "grad_norm": 0.6011562108844719, "learning_rate": 8.385089658921892e-06, "loss": 0.4942, "step": 346 }, { "epoch": 1.0136319376825706, "grad_norm": 0.7194629629820957, "learning_rate": 8.372545118230793e-06, "loss": 0.5879, "step": 347 }, { "epoch": 1.0165530671859786, "grad_norm": 0.7106922347864785, "learning_rate": 8.35996150661476e-06, "loss": 0.6464, "step": 348 }, { "epoch": 1.0194741966893865, "grad_norm": 0.7004876649145284, "learning_rate": 8.347338969854898e-06, "loss": 0.5635, "step": 349 }, { "epoch": 1.0223953261927945, "grad_norm": 0.7176313477690597, "learning_rate": 8.334677654183254e-06, "loss": 0.6121, "step": 350 }, { "epoch": 1.0253164556962024, "grad_norm": 0.6767419736782746, "learning_rate": 8.321977706281135e-06, "loss": 0.5923, "step": 351 }, { "epoch": 1.0282375851996106, "grad_norm": 0.6606355795156919, "learning_rate": 8.309239273277394e-06, "loss": 0.5375, "step": 352 }, { "epoch": 1.0311587147030186, "grad_norm": 0.7777491012749531, "learning_rate": 8.296462502746743e-06, "loss": 0.5971, "step": 353 }, { "epoch": 1.0340798442064265, "grad_norm": 0.6239185790928177, "learning_rate": 8.283647542708026e-06, "loss": 0.6017, "step": 354 }, { "epoch": 1.0370009737098345, "grad_norm": 0.659621365452411, "learning_rate": 8.27079454162252e-06, "loss": 0.523, "step": 355 }, { "epoch": 1.0399221032132424, "grad_norm": 0.7314219237543246, "learning_rate": 8.2579036483922e-06, "loss": 0.5992, "step": 356 }, { "epoch": 1.0428432327166504, "grad_norm": 0.6317967581739343, "learning_rate": 8.244975012358028e-06, "loss": 0.58, "step": 357 }, { "epoch": 1.0457643622200585, "grad_norm": 0.709552964482946, "learning_rate": 8.232008783298211e-06, "loss": 0.5703, "step": 358 }, { "epoch": 1.0486854917234665, "grad_norm": 0.5856088171700015, "learning_rate": 8.219005111426472e-06, "loss": 0.5851, "step": 359 }, { "epoch": 1.0516066212268744, "grad_norm": 0.72335786152197, "learning_rate": 8.205964147390313e-06, "loss": 0.5762, "step": 360 }, { "epoch": 1.0545277507302824, "grad_norm": 0.6082315528170907, "learning_rate": 8.19288604226926e-06, "loss": 0.6045, "step": 361 }, { "epoch": 1.0574488802336903, "grad_norm": 0.6904542946382585, "learning_rate": 8.179770947573124e-06, "loss": 0.5649, "step": 362 }, { "epoch": 1.0603700097370983, "grad_norm": 0.5889683647522176, "learning_rate": 8.166619015240236e-06, "loss": 0.5681, "step": 363 }, { "epoch": 1.0632911392405062, "grad_norm": 0.5621253187560026, "learning_rate": 8.15343039763569e-06, "loss": 0.6013, "step": 364 }, { "epoch": 1.0662122687439144, "grad_norm": 0.5587583825073225, "learning_rate": 8.140205247549583e-06, "loss": 0.6026, "step": 365 }, { "epoch": 1.0691333982473223, "grad_norm": 0.5053935078299595, "learning_rate": 8.126943718195239e-06, "loss": 0.5693, "step": 366 }, { "epoch": 1.0720545277507303, "grad_norm": 0.5469581161571481, "learning_rate": 8.113645963207432e-06, "loss": 0.6007, "step": 367 }, { "epoch": 1.0749756572541382, "grad_norm": 0.5430360387939689, "learning_rate": 8.100312136640618e-06, "loss": 0.5754, "step": 368 }, { "epoch": 1.0778967867575462, "grad_norm": 0.529838189847217, "learning_rate": 8.086942392967131e-06, "loss": 0.5758, "step": 369 }, { "epoch": 1.0808179162609541, "grad_norm": 0.5537166527008645, "learning_rate": 8.073536887075417e-06, "loss": 0.5596, "step": 370 }, { "epoch": 1.0837390457643623, "grad_norm": 0.575683387457894, "learning_rate": 8.060095774268217e-06, "loss": 0.6149, "step": 371 }, { "epoch": 1.0866601752677703, "grad_norm": 0.5212823944156575, "learning_rate": 8.046619210260785e-06, "loss": 0.5376, "step": 372 }, { "epoch": 1.0895813047711782, "grad_norm": 0.6132577983715938, "learning_rate": 8.03310735117907e-06, "loss": 0.611, "step": 373 }, { "epoch": 1.0925024342745862, "grad_norm": 0.5741650450345563, "learning_rate": 8.019560353557923e-06, "loss": 0.5428, "step": 374 }, { "epoch": 1.095423563777994, "grad_norm": 0.6898994227426227, "learning_rate": 8.005978374339264e-06, "loss": 0.6122, "step": 375 }, { "epoch": 1.098344693281402, "grad_norm": 0.6195159909025971, "learning_rate": 7.992361570870289e-06, "loss": 0.6298, "step": 376 }, { "epoch": 1.1012658227848102, "grad_norm": 0.5592176863621418, "learning_rate": 7.978710100901617e-06, "loss": 0.527, "step": 377 }, { "epoch": 1.1041869522882182, "grad_norm": 0.6077274644298606, "learning_rate": 7.965024122585491e-06, "loss": 0.5733, "step": 378 }, { "epoch": 1.1071080817916261, "grad_norm": 0.5404386498094536, "learning_rate": 7.951303794473926e-06, "loss": 0.5786, "step": 379 }, { "epoch": 1.110029211295034, "grad_norm": 0.6285033579389189, "learning_rate": 7.937549275516882e-06, "loss": 0.5593, "step": 380 }, { "epoch": 1.112950340798442, "grad_norm": 0.6136636789874864, "learning_rate": 7.92376072506042e-06, "loss": 0.5887, "step": 381 }, { "epoch": 1.11587147030185, "grad_norm": 0.5563073653338333, "learning_rate": 7.909938302844856e-06, "loss": 0.637, "step": 382 }, { "epoch": 1.1187925998052581, "grad_norm": 0.4984935085580574, "learning_rate": 7.896082169002903e-06, "loss": 0.537, "step": 383 }, { "epoch": 1.121713729308666, "grad_norm": 0.5688731697204966, "learning_rate": 7.882192484057837e-06, "loss": 0.5977, "step": 384 }, { "epoch": 1.124634858812074, "grad_norm": 0.5409938929481367, "learning_rate": 7.868269408921614e-06, "loss": 0.5477, "step": 385 }, { "epoch": 1.127555988315482, "grad_norm": 0.6311630545441865, "learning_rate": 7.854313104893014e-06, "loss": 0.5595, "step": 386 }, { "epoch": 1.13047711781889, "grad_norm": 0.5517528259351719, "learning_rate": 7.84032373365578e-06, "loss": 0.5588, "step": 387 }, { "epoch": 1.1333982473222979, "grad_norm": 0.515690277716415, "learning_rate": 7.826301457276733e-06, "loss": 0.5767, "step": 388 }, { "epoch": 1.136319376825706, "grad_norm": 0.5872560884226455, "learning_rate": 7.812246438203905e-06, "loss": 0.5618, "step": 389 }, { "epoch": 1.139240506329114, "grad_norm": 0.5449550621943328, "learning_rate": 7.798158839264645e-06, "loss": 0.5373, "step": 390 }, { "epoch": 1.142161635832522, "grad_norm": 0.6192261787272578, "learning_rate": 7.784038823663746e-06, "loss": 0.6117, "step": 391 }, { "epoch": 1.14508276533593, "grad_norm": 0.5911797274948096, "learning_rate": 7.769886554981549e-06, "loss": 0.5832, "step": 392 }, { "epoch": 1.1480038948393378, "grad_norm": 0.5393860544929336, "learning_rate": 7.755702197172036e-06, "loss": 0.5511, "step": 393 }, { "epoch": 1.1509250243427458, "grad_norm": 0.5686675548722897, "learning_rate": 7.741485914560958e-06, "loss": 0.601, "step": 394 }, { "epoch": 1.1538461538461537, "grad_norm": 0.5965421717249094, "learning_rate": 7.7272378718439e-06, "loss": 0.6142, "step": 395 }, { "epoch": 1.156767283349562, "grad_norm": 0.5976378598496599, "learning_rate": 7.712958234084395e-06, "loss": 0.532, "step": 396 }, { "epoch": 1.1596884128529699, "grad_norm": 0.5866514447746561, "learning_rate": 7.698647166712003e-06, "loss": 0.6436, "step": 397 }, { "epoch": 1.1626095423563778, "grad_norm": 0.5577115650955226, "learning_rate": 7.684304835520395e-06, "loss": 0.5524, "step": 398 }, { "epoch": 1.1655306718597858, "grad_norm": 0.6320330036922427, "learning_rate": 7.669931406665437e-06, "loss": 0.631, "step": 399 }, { "epoch": 1.1684518013631937, "grad_norm": 0.5274101108563934, "learning_rate": 7.655527046663254e-06, "loss": 0.5369, "step": 400 }, { "epoch": 1.1713729308666017, "grad_norm": 0.5563468540404465, "learning_rate": 7.641091922388316e-06, "loss": 0.577, "step": 401 }, { "epoch": 1.1742940603700098, "grad_norm": 0.5043411412651241, "learning_rate": 7.626626201071494e-06, "loss": 0.5623, "step": 402 }, { "epoch": 1.1772151898734178, "grad_norm": 0.5658997847969963, "learning_rate": 7.612130050298126e-06, "loss": 0.5613, "step": 403 }, { "epoch": 1.1801363193768257, "grad_norm": 0.5617715925867288, "learning_rate": 7.597603638006071e-06, "loss": 0.5796, "step": 404 }, { "epoch": 1.1830574488802337, "grad_norm": 0.5730674124102592, "learning_rate": 7.5830471324837765e-06, "loss": 0.6102, "step": 405 }, { "epoch": 1.1859785783836416, "grad_norm": 0.5322280548113558, "learning_rate": 7.56846070236831e-06, "loss": 0.5392, "step": 406 }, { "epoch": 1.1888997078870496, "grad_norm": 0.5636446588596294, "learning_rate": 7.55384451664342e-06, "loss": 0.5805, "step": 407 }, { "epoch": 1.1918208373904577, "grad_norm": 0.5515192754192108, "learning_rate": 7.539198744637577e-06, "loss": 0.5647, "step": 408 }, { "epoch": 1.1947419668938657, "grad_norm": 0.649611846767762, "learning_rate": 7.524523556022003e-06, "loss": 0.5804, "step": 409 }, { "epoch": 1.1976630963972736, "grad_norm": 0.5400876783820088, "learning_rate": 7.5098191208087144e-06, "loss": 0.5295, "step": 410 }, { "epoch": 1.2005842259006816, "grad_norm": 0.727016701945529, "learning_rate": 7.495085609348549e-06, "loss": 0.6035, "step": 411 }, { "epoch": 1.2035053554040895, "grad_norm": 0.6474638029028308, "learning_rate": 7.4803231923291905e-06, "loss": 0.5905, "step": 412 }, { "epoch": 1.2064264849074975, "grad_norm": 0.5322418094330414, "learning_rate": 7.465532040773195e-06, "loss": 0.5696, "step": 413 }, { "epoch": 1.2093476144109054, "grad_norm": 0.6033615790880061, "learning_rate": 7.45071232603601e-06, "loss": 0.5495, "step": 414 }, { "epoch": 1.2122687439143136, "grad_norm": 0.5696185950396394, "learning_rate": 7.4358642198039835e-06, "loss": 0.5761, "step": 415 }, { "epoch": 1.2151898734177216, "grad_norm": 0.58236484440868, "learning_rate": 7.420987894092383e-06, "loss": 0.6225, "step": 416 }, { "epoch": 1.2181110029211295, "grad_norm": 0.6023725223738031, "learning_rate": 7.406083521243396e-06, "loss": 0.5539, "step": 417 }, { "epoch": 1.2210321324245375, "grad_norm": 0.5964833664567649, "learning_rate": 7.391151273924135e-06, "loss": 0.5766, "step": 418 }, { "epoch": 1.2239532619279454, "grad_norm": 0.5579327351939204, "learning_rate": 7.376191325124644e-06, "loss": 0.6037, "step": 419 }, { "epoch": 1.2268743914313536, "grad_norm": 0.662691054067833, "learning_rate": 7.36120384815588e-06, "loss": 0.5775, "step": 420 }, { "epoch": 1.2297955209347615, "grad_norm": 0.5457870053427097, "learning_rate": 7.34618901664772e-06, "loss": 0.6032, "step": 421 }, { "epoch": 1.2327166504381695, "grad_norm": 0.528437147629629, "learning_rate": 7.33114700454694e-06, "loss": 0.5551, "step": 422 }, { "epoch": 1.2356377799415774, "grad_norm": 0.6396983335282087, "learning_rate": 7.316077986115206e-06, "loss": 0.5546, "step": 423 }, { "epoch": 1.2385589094449854, "grad_norm": 0.6059608136412865, "learning_rate": 7.300982135927051e-06, "loss": 0.5889, "step": 424 }, { "epoch": 1.2414800389483933, "grad_norm": 0.5395833042938699, "learning_rate": 7.285859628867851e-06, "loss": 0.5765, "step": 425 }, { "epoch": 1.2444011684518013, "grad_norm": 0.5938150264550871, "learning_rate": 7.270710640131806e-06, "loss": 0.6058, "step": 426 }, { "epoch": 1.2473222979552094, "grad_norm": 0.5435408340953944, "learning_rate": 7.255535345219905e-06, "loss": 0.5933, "step": 427 }, { "epoch": 1.2502434274586174, "grad_norm": 0.5055763559518435, "learning_rate": 7.240333919937893e-06, "loss": 0.5989, "step": 428 }, { "epoch": 1.2531645569620253, "grad_norm": 0.5862621574043387, "learning_rate": 7.2251065403942355e-06, "loss": 0.5888, "step": 429 }, { "epoch": 1.2560856864654333, "grad_norm": 0.5234746081250162, "learning_rate": 7.209853382998077e-06, "loss": 0.5537, "step": 430 }, { "epoch": 1.2590068159688412, "grad_norm": 0.5736429396541277, "learning_rate": 7.1945746244572e-06, "loss": 0.547, "step": 431 }, { "epoch": 1.2619279454722494, "grad_norm": 0.5662950178895197, "learning_rate": 7.179270441775976e-06, "loss": 0.6502, "step": 432 }, { "epoch": 1.2648490749756571, "grad_norm": 0.6193942417064999, "learning_rate": 7.163941012253317e-06, "loss": 0.5182, "step": 433 }, { "epoch": 1.2677702044790653, "grad_norm": 0.5954433533450075, "learning_rate": 7.148586513480614e-06, "loss": 0.5961, "step": 434 }, { "epoch": 1.2706913339824732, "grad_norm": 0.5407062847357796, "learning_rate": 7.133207123339689e-06, "loss": 0.5475, "step": 435 }, { "epoch": 1.2736124634858812, "grad_norm": 0.6133121400132443, "learning_rate": 7.117803020000733e-06, "loss": 0.59, "step": 436 }, { "epoch": 1.2765335929892891, "grad_norm": 0.5272070801107334, "learning_rate": 7.102374381920233e-06, "loss": 0.542, "step": 437 }, { "epoch": 1.279454722492697, "grad_norm": 0.6014559091825684, "learning_rate": 7.086921387838916e-06, "loss": 0.6136, "step": 438 }, { "epoch": 1.2823758519961053, "grad_norm": 0.4925204082847554, "learning_rate": 7.071444216779669e-06, "loss": 0.5707, "step": 439 }, { "epoch": 1.2852969814995132, "grad_norm": 0.5297523991584766, "learning_rate": 7.055943048045476e-06, "loss": 0.5917, "step": 440 }, { "epoch": 1.2882181110029212, "grad_norm": 0.5742994575633306, "learning_rate": 7.040418061217325e-06, "loss": 0.6161, "step": 441 }, { "epoch": 1.2911392405063291, "grad_norm": 0.5144128093922836, "learning_rate": 7.024869436152144e-06, "loss": 0.5481, "step": 442 }, { "epoch": 1.294060370009737, "grad_norm": 0.5651127309438425, "learning_rate": 7.009297352980706e-06, "loss": 0.5789, "step": 443 }, { "epoch": 1.296981499513145, "grad_norm": 0.46327530559493646, "learning_rate": 6.99370199210555e-06, "loss": 0.5435, "step": 444 }, { "epoch": 1.299902629016553, "grad_norm": 0.5306445037454199, "learning_rate": 6.978083534198878e-06, "loss": 0.6135, "step": 445 }, { "epoch": 1.3028237585199611, "grad_norm": 0.6551321587944974, "learning_rate": 6.962442160200484e-06, "loss": 0.6111, "step": 446 }, { "epoch": 1.305744888023369, "grad_norm": 0.5887843505365383, "learning_rate": 6.9467780513156335e-06, "loss": 0.5428, "step": 447 }, { "epoch": 1.308666017526777, "grad_norm": 0.5871163724329924, "learning_rate": 6.931091389012983e-06, "loss": 0.5595, "step": 448 }, { "epoch": 1.311587147030185, "grad_norm": 0.6876137396335796, "learning_rate": 6.915382355022465e-06, "loss": 0.5776, "step": 449 }, { "epoch": 1.314508276533593, "grad_norm": 0.5978349360901288, "learning_rate": 6.899651131333194e-06, "loss": 0.559, "step": 450 }, { "epoch": 1.317429406037001, "grad_norm": 0.5600538501243326, "learning_rate": 6.8838979001913454e-06, "loss": 0.5782, "step": 451 }, { "epoch": 1.3203505355404088, "grad_norm": 0.6670604315541686, "learning_rate": 6.868122844098057e-06, "loss": 0.6669, "step": 452 }, { "epoch": 1.323271665043817, "grad_norm": 0.6019460261148103, "learning_rate": 6.852326145807302e-06, "loss": 0.6006, "step": 453 }, { "epoch": 1.326192794547225, "grad_norm": 0.4932938142869286, "learning_rate": 6.836507988323785e-06, "loss": 0.4971, "step": 454 }, { "epoch": 1.3291139240506329, "grad_norm": 0.6389612562939274, "learning_rate": 6.82066855490081e-06, "loss": 0.5994, "step": 455 }, { "epoch": 1.3320350535540408, "grad_norm": 0.569391901245458, "learning_rate": 6.804808029038168e-06, "loss": 0.5776, "step": 456 }, { "epoch": 1.3349561830574488, "grad_norm": 0.5627619450291675, "learning_rate": 6.788926594480001e-06, "loss": 0.5894, "step": 457 }, { "epoch": 1.337877312560857, "grad_norm": 0.5525138776130833, "learning_rate": 6.773024435212678e-06, "loss": 0.5507, "step": 458 }, { "epoch": 1.340798442064265, "grad_norm": 0.6243616187017386, "learning_rate": 6.75710173546267e-06, "loss": 0.6052, "step": 459 }, { "epoch": 1.3437195715676729, "grad_norm": 0.46296070627628405, "learning_rate": 6.741158679694403e-06, "loss": 0.5284, "step": 460 }, { "epoch": 1.3466407010710808, "grad_norm": 0.556033009188379, "learning_rate": 6.7251954526081294e-06, "loss": 0.6225, "step": 461 }, { "epoch": 1.3495618305744888, "grad_norm": 0.5425579804205766, "learning_rate": 6.709212239137785e-06, "loss": 0.5919, "step": 462 }, { "epoch": 1.352482960077897, "grad_norm": 0.5628452971401441, "learning_rate": 6.693209224448853e-06, "loss": 0.6162, "step": 463 }, { "epoch": 1.3554040895813046, "grad_norm": 0.5283138648819659, "learning_rate": 6.677186593936207e-06, "loss": 0.5022, "step": 464 }, { "epoch": 1.3583252190847128, "grad_norm": 0.5688480741205626, "learning_rate": 6.661144533221974e-06, "loss": 0.594, "step": 465 }, { "epoch": 1.3612463485881208, "grad_norm": 0.6112831896983953, "learning_rate": 6.645083228153377e-06, "loss": 0.5803, "step": 466 }, { "epoch": 1.3641674780915287, "grad_norm": 0.5385606514191121, "learning_rate": 6.629002864800589e-06, "loss": 0.5834, "step": 467 }, { "epoch": 1.3670886075949367, "grad_norm": 0.5366278285884805, "learning_rate": 6.612903629454568e-06, "loss": 0.5343, "step": 468 }, { "epoch": 1.3700097370983446, "grad_norm": 0.6713338764048196, "learning_rate": 6.5967857086249065e-06, "loss": 0.5573, "step": 469 }, { "epoch": 1.3729308666017528, "grad_norm": 0.5538883029206723, "learning_rate": 6.58064928903767e-06, "loss": 0.6164, "step": 470 }, { "epoch": 1.3758519961051607, "grad_norm": 0.5005412280561692, "learning_rate": 6.56449455763323e-06, "loss": 0.5576, "step": 471 }, { "epoch": 1.3787731256085687, "grad_norm": 0.5868063901075268, "learning_rate": 6.548321701564099e-06, "loss": 0.5933, "step": 472 }, { "epoch": 1.3816942551119766, "grad_norm": 0.5499777044713934, "learning_rate": 6.5321309081927665e-06, "loss": 0.562, "step": 473 }, { "epoch": 1.3846153846153846, "grad_norm": 0.5209457126764143, "learning_rate": 6.515922365089524e-06, "loss": 0.5793, "step": 474 }, { "epoch": 1.3875365141187925, "grad_norm": 0.5045261657801648, "learning_rate": 6.499696260030297e-06, "loss": 0.594, "step": 475 }, { "epoch": 1.3904576436222005, "grad_norm": 0.5243175331043878, "learning_rate": 6.483452780994459e-06, "loss": 0.534, "step": 476 }, { "epoch": 1.3933787731256086, "grad_norm": 0.5169962119701911, "learning_rate": 6.467192116162668e-06, "loss": 0.5403, "step": 477 }, { "epoch": 1.3962999026290166, "grad_norm": 0.5140189723701601, "learning_rate": 6.450914453914674e-06, "loss": 0.6058, "step": 478 }, { "epoch": 1.3992210321324245, "grad_norm": 0.5330866177837568, "learning_rate": 6.434619982827147e-06, "loss": 0.5882, "step": 479 }, { "epoch": 1.4021421616358325, "grad_norm": 0.5414814038608168, "learning_rate": 6.418308891671484e-06, "loss": 0.5954, "step": 480 }, { "epoch": 1.4050632911392404, "grad_norm": 0.5376659643418761, "learning_rate": 6.401981369411626e-06, "loss": 0.5633, "step": 481 }, { "epoch": 1.4079844206426486, "grad_norm": 0.5438938648229659, "learning_rate": 6.385637605201871e-06, "loss": 0.5677, "step": 482 }, { "epoch": 1.4109055501460563, "grad_norm": 0.533597042074437, "learning_rate": 6.3692777883846746e-06, "loss": 0.6217, "step": 483 }, { "epoch": 1.4138266796494645, "grad_norm": 0.5530416412659575, "learning_rate": 6.3529021084884655e-06, "loss": 0.5135, "step": 484 }, { "epoch": 1.4167478091528725, "grad_norm": 0.6426214488830264, "learning_rate": 6.336510755225447e-06, "loss": 0.6039, "step": 485 }, { "epoch": 1.4196689386562804, "grad_norm": 0.5321542466408685, "learning_rate": 6.320103918489395e-06, "loss": 0.5615, "step": 486 }, { "epoch": 1.4225900681596884, "grad_norm": 0.5807878117878842, "learning_rate": 6.303681788353465e-06, "loss": 0.6417, "step": 487 }, { "epoch": 1.4255111976630963, "grad_norm": 0.6006151574662654, "learning_rate": 6.287244555067984e-06, "loss": 0.5794, "step": 488 }, { "epoch": 1.4284323271665045, "grad_norm": 0.5126167984608686, "learning_rate": 6.270792409058247e-06, "loss": 0.5242, "step": 489 }, { "epoch": 1.4313534566699124, "grad_norm": 0.509995152360523, "learning_rate": 6.25432554092232e-06, "loss": 0.5828, "step": 490 }, { "epoch": 1.4342745861733204, "grad_norm": 0.5960222110178056, "learning_rate": 6.237844141428817e-06, "loss": 0.568, "step": 491 }, { "epoch": 1.4371957156767283, "grad_norm": 0.5173153132658744, "learning_rate": 6.221348401514703e-06, "loss": 0.5796, "step": 492 }, { "epoch": 1.4401168451801363, "grad_norm": 0.5671800858222442, "learning_rate": 6.204838512283073e-06, "loss": 0.5506, "step": 493 }, { "epoch": 1.4430379746835442, "grad_norm": 0.6040430910549448, "learning_rate": 6.188314665000944e-06, "loss": 0.5444, "step": 494 }, { "epoch": 1.4459591041869522, "grad_norm": 0.5391742690410167, "learning_rate": 6.171777051097037e-06, "loss": 0.5649, "step": 495 }, { "epoch": 1.4488802336903603, "grad_norm": 0.597749185130725, "learning_rate": 6.155225862159558e-06, "loss": 0.6408, "step": 496 }, { "epoch": 1.4518013631937683, "grad_norm": 0.5229846439398993, "learning_rate": 6.138661289933981e-06, "loss": 0.5494, "step": 497 }, { "epoch": 1.4547224926971762, "grad_norm": 0.5920796120572965, "learning_rate": 6.1220835263208256e-06, "loss": 0.5992, "step": 498 }, { "epoch": 1.4576436222005842, "grad_norm": 0.5775415962111137, "learning_rate": 6.105492763373431e-06, "loss": 0.5911, "step": 499 }, { "epoch": 1.4605647517039921, "grad_norm": 0.5917654314551983, "learning_rate": 6.088889193295738e-06, "loss": 0.5643, "step": 500 }, { "epoch": 1.4634858812074003, "grad_norm": 0.5199838430286094, "learning_rate": 6.072273008440052e-06, "loss": 0.5472, "step": 501 }, { "epoch": 1.4664070107108083, "grad_norm": 0.5522718232673819, "learning_rate": 6.0556444013048265e-06, "loss": 0.546, "step": 502 }, { "epoch": 1.4693281402142162, "grad_norm": 0.525908351198303, "learning_rate": 6.039003564532423e-06, "loss": 0.6015, "step": 503 }, { "epoch": 1.4722492697176242, "grad_norm": 0.5105926082337128, "learning_rate": 6.0223506909068875e-06, "loss": 0.5785, "step": 504 }, { "epoch": 1.475170399221032, "grad_norm": 0.5832929477157227, "learning_rate": 6.005685973351708e-06, "loss": 0.6027, "step": 505 }, { "epoch": 1.47809152872444, "grad_norm": 0.5447366515820563, "learning_rate": 5.989009604927587e-06, "loss": 0.5833, "step": 506 }, { "epoch": 1.481012658227848, "grad_norm": 0.5245433114840506, "learning_rate": 5.972321778830202e-06, "loss": 0.4944, "step": 507 }, { "epoch": 1.4839337877312562, "grad_norm": 0.5377622062276557, "learning_rate": 5.9556226883879685e-06, "loss": 0.5518, "step": 508 }, { "epoch": 1.4868549172346641, "grad_norm": 0.6351508971275112, "learning_rate": 5.938912527059798e-06, "loss": 0.5982, "step": 509 }, { "epoch": 1.489776046738072, "grad_norm": 0.5264532680548314, "learning_rate": 5.922191488432857e-06, "loss": 0.5472, "step": 510 }, { "epoch": 1.49269717624148, "grad_norm": 0.5618733075790842, "learning_rate": 5.90545976622033e-06, "loss": 0.6387, "step": 511 }, { "epoch": 1.495618305744888, "grad_norm": 0.5042404211846953, "learning_rate": 5.888717554259165e-06, "loss": 0.5188, "step": 512 }, { "epoch": 1.4985394352482961, "grad_norm": 0.5996827758246558, "learning_rate": 5.871965046507838e-06, "loss": 0.5845, "step": 513 }, { "epoch": 1.5014605647517039, "grad_norm": 0.513160385214818, "learning_rate": 5.855202437044102e-06, "loss": 0.5407, "step": 514 }, { "epoch": 1.504381694255112, "grad_norm": 0.49930093522201097, "learning_rate": 5.838429920062734e-06, "loss": 0.5888, "step": 515 }, { "epoch": 1.50730282375852, "grad_norm": 0.5820994183473087, "learning_rate": 5.8216476898732935e-06, "loss": 0.6026, "step": 516 }, { "epoch": 1.510223953261928, "grad_norm": 0.5965529197255538, "learning_rate": 5.804855940897866e-06, "loss": 0.5772, "step": 517 }, { "epoch": 1.5131450827653359, "grad_norm": 0.4891063083201174, "learning_rate": 5.788054867668811e-06, "loss": 0.5383, "step": 518 }, { "epoch": 1.5160662122687438, "grad_norm": 0.549598064227309, "learning_rate": 5.771244664826512e-06, "loss": 0.5701, "step": 519 }, { "epoch": 1.518987341772152, "grad_norm": 0.5201662235289578, "learning_rate": 5.754425527117118e-06, "loss": 0.5221, "step": 520 }, { "epoch": 1.5219084712755597, "grad_norm": 0.5727695969794172, "learning_rate": 5.737597649390288e-06, "loss": 0.5849, "step": 521 }, { "epoch": 1.524829600778968, "grad_norm": 0.5387267148344115, "learning_rate": 5.720761226596935e-06, "loss": 0.5899, "step": 522 }, { "epoch": 1.5277507302823758, "grad_norm": 0.5196417961644373, "learning_rate": 5.703916453786965e-06, "loss": 0.6075, "step": 523 }, { "epoch": 1.5306718597857838, "grad_norm": 0.5162526391071218, "learning_rate": 5.6870635261070186e-06, "loss": 0.5524, "step": 524 }, { "epoch": 1.533592989289192, "grad_norm": 0.5712502081281199, "learning_rate": 5.670202638798213e-06, "loss": 0.5639, "step": 525 }, { "epoch": 1.5365141187925997, "grad_norm": 0.569067876773126, "learning_rate": 5.653333987193876e-06, "loss": 0.5437, "step": 526 }, { "epoch": 1.5394352482960079, "grad_norm": 0.5925750045532109, "learning_rate": 5.636457766717286e-06, "loss": 0.591, "step": 527 }, { "epoch": 1.5423563777994158, "grad_norm": 0.5141531017914309, "learning_rate": 5.619574172879405e-06, "loss": 0.5471, "step": 528 }, { "epoch": 1.5452775073028238, "grad_norm": 0.6597290642553237, "learning_rate": 5.6026834012766155e-06, "loss": 0.5901, "step": 529 }, { "epoch": 1.5481986368062317, "grad_norm": 0.5884715259486929, "learning_rate": 5.585785647588458e-06, "loss": 0.5878, "step": 530 }, { "epoch": 1.5511197663096397, "grad_norm": 0.5255721650946908, "learning_rate": 5.568881107575353e-06, "loss": 0.5447, "step": 531 }, { "epoch": 1.5540408958130478, "grad_norm": 0.5091097959641512, "learning_rate": 5.55196997707635e-06, "loss": 0.6276, "step": 532 }, { "epoch": 1.5569620253164556, "grad_norm": 0.5119232421542075, "learning_rate": 5.53505245200684e-06, "loss": 0.5658, "step": 533 }, { "epoch": 1.5598831548198637, "grad_norm": 0.556141615427262, "learning_rate": 5.518128728356303e-06, "loss": 0.5951, "step": 534 }, { "epoch": 1.5628042843232717, "grad_norm": 0.4969794250430925, "learning_rate": 5.501199002186024e-06, "loss": 0.5457, "step": 535 }, { "epoch": 1.5657254138266796, "grad_norm": 0.5377430121732353, "learning_rate": 5.48426346962683e-06, "loss": 0.587, "step": 536 }, { "epoch": 1.5686465433300878, "grad_norm": 0.4763376786579273, "learning_rate": 5.467322326876813e-06, "loss": 0.5352, "step": 537 }, { "epoch": 1.5715676728334955, "grad_norm": 0.5630031793193719, "learning_rate": 5.450375770199063e-06, "loss": 0.652, "step": 538 }, { "epoch": 1.5744888023369037, "grad_norm": 0.505233569384271, "learning_rate": 5.433423995919383e-06, "loss": 0.5633, "step": 539 }, { "epoch": 1.5774099318403116, "grad_norm": 0.4918114442897718, "learning_rate": 5.416467200424032e-06, "loss": 0.5422, "step": 540 }, { "epoch": 1.5803310613437196, "grad_norm": 0.5991334895146722, "learning_rate": 5.399505580157428e-06, "loss": 0.558, "step": 541 }, { "epoch": 1.5832521908471275, "grad_norm": 0.5363474139996335, "learning_rate": 5.382539331619896e-06, "loss": 0.5699, "step": 542 }, { "epoch": 1.5861733203505355, "grad_norm": 0.5807115654118472, "learning_rate": 5.365568651365369e-06, "loss": 0.5688, "step": 543 }, { "epoch": 1.5890944498539437, "grad_norm": 0.5904366045058206, "learning_rate": 5.34859373599913e-06, "loss": 0.5969, "step": 544 }, { "epoch": 1.5920155793573514, "grad_norm": 0.491932507277215, "learning_rate": 5.33161478217552e-06, "loss": 0.5542, "step": 545 }, { "epoch": 1.5949367088607596, "grad_norm": 0.5011301426277018, "learning_rate": 5.314631986595669e-06, "loss": 0.569, "step": 546 }, { "epoch": 1.5978578383641675, "grad_norm": 0.6106059864005979, "learning_rate": 5.297645546005208e-06, "loss": 0.6051, "step": 547 }, { "epoch": 1.6007789678675755, "grad_norm": 0.5958177689979132, "learning_rate": 5.280655657192003e-06, "loss": 0.5266, "step": 548 }, { "epoch": 1.6037000973709834, "grad_norm": 0.5395470595824067, "learning_rate": 5.263662516983863e-06, "loss": 0.5657, "step": 549 }, { "epoch": 1.6066212268743914, "grad_norm": 0.6212970783782318, "learning_rate": 5.246666322246267e-06, "loss": 0.5619, "step": 550 }, { "epoch": 1.6095423563777995, "grad_norm": 0.6613037032143712, "learning_rate": 5.229667269880078e-06, "loss": 0.6092, "step": 551 }, { "epoch": 1.6124634858812072, "grad_norm": 0.48332526807644743, "learning_rate": 5.212665556819264e-06, "loss": 0.5401, "step": 552 }, { "epoch": 1.6153846153846154, "grad_norm": 0.5325857980417065, "learning_rate": 5.195661380028625e-06, "loss": 0.6035, "step": 553 }, { "epoch": 1.6183057448880234, "grad_norm": 0.5442785714221408, "learning_rate": 5.178654936501492e-06, "loss": 0.5494, "step": 554 }, { "epoch": 1.6212268743914313, "grad_norm": 0.6437807568653632, "learning_rate": 5.1616464232574635e-06, "loss": 0.6254, "step": 555 }, { "epoch": 1.6241480038948395, "grad_norm": 0.4492125406814553, "learning_rate": 5.1446360373401125e-06, "loss": 0.5301, "step": 556 }, { "epoch": 1.6270691333982472, "grad_norm": 0.5622471364004906, "learning_rate": 5.127623975814709e-06, "loss": 0.5715, "step": 557 }, { "epoch": 1.6299902629016554, "grad_norm": 0.6240851711756873, "learning_rate": 5.110610435765935e-06, "loss": 0.5876, "step": 558 }, { "epoch": 1.6329113924050633, "grad_norm": 0.4713533254066173, "learning_rate": 5.093595614295599e-06, "loss": 0.4901, "step": 559 }, { "epoch": 1.6358325219084713, "grad_norm": 0.5796358881475803, "learning_rate": 5.076579708520355e-06, "loss": 0.6325, "step": 560 }, { "epoch": 1.6387536514118792, "grad_norm": 0.4866837205814519, "learning_rate": 5.059562915569424e-06, "loss": 0.5202, "step": 561 }, { "epoch": 1.6416747809152872, "grad_norm": 0.5697352103252339, "learning_rate": 5.0425454325822946e-06, "loss": 0.6339, "step": 562 }, { "epoch": 1.6445959104186954, "grad_norm": 0.5721771538016244, "learning_rate": 5.0255274567064594e-06, "loss": 0.5364, "step": 563 }, { "epoch": 1.647517039922103, "grad_norm": 0.5605730886016151, "learning_rate": 5.008509185095114e-06, "loss": 0.5717, "step": 564 }, { "epoch": 1.6504381694255112, "grad_norm": 0.472681806190808, "learning_rate": 4.991490814904888e-06, "loss": 0.555, "step": 565 }, { "epoch": 1.6533592989289192, "grad_norm": 0.5791761573446532, "learning_rate": 4.974472543293544e-06, "loss": 0.6077, "step": 566 }, { "epoch": 1.6562804284323271, "grad_norm": 0.5167447517838185, "learning_rate": 4.957454567417708e-06, "loss": 0.5805, "step": 567 }, { "epoch": 1.6592015579357353, "grad_norm": 0.5561818381815233, "learning_rate": 4.940437084430579e-06, "loss": 0.594, "step": 568 }, { "epoch": 1.662122687439143, "grad_norm": 0.44095968784824047, "learning_rate": 4.923420291479646e-06, "loss": 0.5235, "step": 569 }, { "epoch": 1.6650438169425512, "grad_norm": 0.49918372558187907, "learning_rate": 4.906404385704402e-06, "loss": 0.6011, "step": 570 }, { "epoch": 1.667964946445959, "grad_norm": 0.5067863457643954, "learning_rate": 4.8893895642340665e-06, "loss": 0.5415, "step": 571 }, { "epoch": 1.6708860759493671, "grad_norm": 0.49737844087035643, "learning_rate": 4.872376024185291e-06, "loss": 0.583, "step": 572 }, { "epoch": 1.673807205452775, "grad_norm": 0.4659962080708305, "learning_rate": 4.855363962659889e-06, "loss": 0.5627, "step": 573 }, { "epoch": 1.676728334956183, "grad_norm": 0.4972582768404136, "learning_rate": 4.838353576742538e-06, "loss": 0.5911, "step": 574 }, { "epoch": 1.6796494644595912, "grad_norm": 0.44592731268572705, "learning_rate": 4.82134506349851e-06, "loss": 0.5459, "step": 575 }, { "epoch": 1.682570593962999, "grad_norm": 0.506185991053395, "learning_rate": 4.804338619971377e-06, "loss": 0.5306, "step": 576 }, { "epoch": 1.685491723466407, "grad_norm": 0.5177923049058465, "learning_rate": 4.787334443180735e-06, "loss": 0.6209, "step": 577 }, { "epoch": 1.688412852969815, "grad_norm": 0.4601309987230691, "learning_rate": 4.7703327301199244e-06, "loss": 0.5145, "step": 578 }, { "epoch": 1.691333982473223, "grad_norm": 0.4643978865329562, "learning_rate": 4.753333677753734e-06, "loss": 0.5788, "step": 579 }, { "epoch": 1.694255111976631, "grad_norm": 0.5283219640600681, "learning_rate": 4.736337483016138e-06, "loss": 0.6056, "step": 580 }, { "epoch": 1.6971762414800389, "grad_norm": 0.5571423915108132, "learning_rate": 4.719344342808e-06, "loss": 0.5791, "step": 581 }, { "epoch": 1.700097370983447, "grad_norm": 0.49563030247464135, "learning_rate": 4.702354453994794e-06, "loss": 0.5373, "step": 582 }, { "epoch": 1.7030185004868548, "grad_norm": 0.49902926183967655, "learning_rate": 4.6853680134043345e-06, "loss": 0.5823, "step": 583 }, { "epoch": 1.705939629990263, "grad_norm": 0.5116289187531037, "learning_rate": 4.668385217824482e-06, "loss": 0.5557, "step": 584 }, { "epoch": 1.7088607594936709, "grad_norm": 0.5412976872055105, "learning_rate": 4.651406264000871e-06, "loss": 0.5535, "step": 585 }, { "epoch": 1.7117818889970788, "grad_norm": 0.5297049362432723, "learning_rate": 4.634431348634632e-06, "loss": 0.5872, "step": 586 }, { "epoch": 1.714703018500487, "grad_norm": 0.46815419228452076, "learning_rate": 4.6174606683801055e-06, "loss": 0.5612, "step": 587 }, { "epoch": 1.7176241480038947, "grad_norm": 0.5047028920506614, "learning_rate": 4.600494419842573e-06, "loss": 0.5457, "step": 588 }, { "epoch": 1.720545277507303, "grad_norm": 0.5021848097551853, "learning_rate": 4.58353279957597e-06, "loss": 0.5512, "step": 589 }, { "epoch": 1.7234664070107109, "grad_norm": 0.5567736723329815, "learning_rate": 4.5665760040806174e-06, "loss": 0.577, "step": 590 }, { "epoch": 1.7263875365141188, "grad_norm": 0.4982671338092423, "learning_rate": 4.549624229800938e-06, "loss": 0.548, "step": 591 }, { "epoch": 1.7293086660175268, "grad_norm": 0.49329094074335034, "learning_rate": 4.532677673123188e-06, "loss": 0.5563, "step": 592 }, { "epoch": 1.7322297955209347, "grad_norm": 0.4890277332079673, "learning_rate": 4.5157365303731705e-06, "loss": 0.5747, "step": 593 }, { "epoch": 1.7351509250243429, "grad_norm": 0.5600537499346627, "learning_rate": 4.498800997813976e-06, "loss": 0.5816, "step": 594 }, { "epoch": 1.7380720545277506, "grad_norm": 0.5086959673884651, "learning_rate": 4.481871271643698e-06, "loss": 0.5671, "step": 595 }, { "epoch": 1.7409931840311588, "grad_norm": 0.5157431177024923, "learning_rate": 4.464947547993162e-06, "loss": 0.5731, "step": 596 }, { "epoch": 1.7439143135345667, "grad_norm": 0.5137961841228513, "learning_rate": 4.4480300229236525e-06, "loss": 0.544, "step": 597 }, { "epoch": 1.7468354430379747, "grad_norm": 0.5937636838238008, "learning_rate": 4.431118892424649e-06, "loss": 0.6082, "step": 598 }, { "epoch": 1.7497565725413826, "grad_norm": 0.5529433304939138, "learning_rate": 4.414214352411544e-06, "loss": 0.5569, "step": 599 }, { "epoch": 1.7526777020447906, "grad_norm": 0.5159596510951358, "learning_rate": 4.397316598723385e-06, "loss": 0.5455, "step": 600 }, { "epoch": 1.7555988315481987, "grad_norm": 0.5506758519748508, "learning_rate": 4.3804258271205965e-06, "loss": 0.5505, "step": 601 }, { "epoch": 1.7585199610516065, "grad_norm": 0.5523849090934733, "learning_rate": 4.363542233282715e-06, "loss": 0.5355, "step": 602 }, { "epoch": 1.7614410905550146, "grad_norm": 0.6508686455735765, "learning_rate": 4.346666012806126e-06, "loss": 0.602, "step": 603 }, { "epoch": 1.7643622200584226, "grad_norm": 0.46084404235460724, "learning_rate": 4.329797361201788e-06, "loss": 0.5028, "step": 604 }, { "epoch": 1.7672833495618305, "grad_norm": 0.5857372990733761, "learning_rate": 4.312936473892984e-06, "loss": 0.6468, "step": 605 }, { "epoch": 1.7702044790652387, "grad_norm": 0.4935847914878853, "learning_rate": 4.296083546213037e-06, "loss": 0.5578, "step": 606 }, { "epoch": 1.7731256085686464, "grad_norm": 0.5144278362398599, "learning_rate": 4.279238773403066e-06, "loss": 0.5337, "step": 607 }, { "epoch": 1.7760467380720546, "grad_norm": 0.46428380216043086, "learning_rate": 4.2624023506097116e-06, "loss": 0.6219, "step": 608 }, { "epoch": 1.7789678675754625, "grad_norm": 0.4492920362378749, "learning_rate": 4.245574472882882e-06, "loss": 0.5061, "step": 609 }, { "epoch": 1.7818889970788705, "grad_norm": 0.48814416218826545, "learning_rate": 4.228755335173488e-06, "loss": 0.5045, "step": 610 }, { "epoch": 1.7848101265822784, "grad_norm": 0.4935317746542341, "learning_rate": 4.21194513233119e-06, "loss": 0.6334, "step": 611 }, { "epoch": 1.7877312560856864, "grad_norm": 0.5119284354029839, "learning_rate": 4.1951440591021375e-06, "loss": 0.5599, "step": 612 }, { "epoch": 1.7906523855890946, "grad_norm": 0.4827371863106343, "learning_rate": 4.17835231012671e-06, "loss": 0.5692, "step": 613 }, { "epoch": 1.7935735150925023, "grad_norm": 0.5212508350905807, "learning_rate": 4.161570079937268e-06, "loss": 0.5947, "step": 614 }, { "epoch": 1.7964946445959105, "grad_norm": 0.5235515777898554, "learning_rate": 4.1447975629559e-06, "loss": 0.5808, "step": 615 }, { "epoch": 1.7994157740993184, "grad_norm": 0.5336005328533747, "learning_rate": 4.128034953492163e-06, "loss": 0.5358, "step": 616 }, { "epoch": 1.8023369036027264, "grad_norm": 0.5597776990519019, "learning_rate": 4.1112824457408375e-06, "loss": 0.5884, "step": 617 }, { "epoch": 1.8052580331061345, "grad_norm": 0.5036193557727575, "learning_rate": 4.094540233779672e-06, "loss": 0.5688, "step": 618 }, { "epoch": 1.8081791626095423, "grad_norm": 0.5338396120419266, "learning_rate": 4.077808511567145e-06, "loss": 0.5608, "step": 619 }, { "epoch": 1.8111002921129504, "grad_norm": 0.4958715717085618, "learning_rate": 4.061087472940204e-06, "loss": 0.5543, "step": 620 }, { "epoch": 1.8140214216163584, "grad_norm": 0.6029334087991165, "learning_rate": 4.044377311612033e-06, "loss": 0.5691, "step": 621 }, { "epoch": 1.8169425511197663, "grad_norm": 0.47915777984754265, "learning_rate": 4.0276782211698e-06, "loss": 0.5391, "step": 622 }, { "epoch": 1.8198636806231743, "grad_norm": 0.5235008297199194, "learning_rate": 4.010990395072414e-06, "loss": 0.5736, "step": 623 }, { "epoch": 1.8227848101265822, "grad_norm": 0.5590747505915074, "learning_rate": 3.9943140266482935e-06, "loss": 0.6075, "step": 624 }, { "epoch": 1.8257059396299904, "grad_norm": 0.47598042280541497, "learning_rate": 3.977649309093113e-06, "loss": 0.5344, "step": 625 }, { "epoch": 1.8286270691333981, "grad_norm": 0.48621841809715133, "learning_rate": 3.960996435467577e-06, "loss": 0.6179, "step": 626 }, { "epoch": 1.8315481986368063, "grad_norm": 0.5017298117872958, "learning_rate": 3.944355598695174e-06, "loss": 0.5443, "step": 627 }, { "epoch": 1.8344693281402142, "grad_norm": 0.5380565465468942, "learning_rate": 3.9277269915599505e-06, "loss": 0.57, "step": 628 }, { "epoch": 1.8373904576436222, "grad_norm": 0.5990209575217125, "learning_rate": 3.911110806704265e-06, "loss": 0.5784, "step": 629 }, { "epoch": 1.8403115871470301, "grad_norm": 0.49509817541993284, "learning_rate": 3.89450723662657e-06, "loss": 0.5628, "step": 630 }, { "epoch": 1.843232716650438, "grad_norm": 0.550605497861562, "learning_rate": 3.877916473679176e-06, "loss": 0.6147, "step": 631 }, { "epoch": 1.8461538461538463, "grad_norm": 0.5524680706044535, "learning_rate": 3.861338710066021e-06, "loss": 0.5125, "step": 632 }, { "epoch": 1.849074975657254, "grad_norm": 0.5494313577644265, "learning_rate": 3.8447741378404436e-06, "loss": 0.5562, "step": 633 }, { "epoch": 1.8519961051606622, "grad_norm": 0.5058622045652387, "learning_rate": 3.828222948902965e-06, "loss": 0.5967, "step": 634 }, { "epoch": 1.85491723466407, "grad_norm": 0.4488395706127018, "learning_rate": 3.8116853349990574e-06, "loss": 0.4735, "step": 635 }, { "epoch": 1.857838364167478, "grad_norm": 0.5073104983823334, "learning_rate": 3.7951614877169285e-06, "loss": 0.6214, "step": 636 }, { "epoch": 1.8607594936708862, "grad_norm": 0.5020623526859631, "learning_rate": 3.7786515984852977e-06, "loss": 0.5801, "step": 637 }, { "epoch": 1.863680623174294, "grad_norm": 0.5066558405528695, "learning_rate": 3.762155858571184e-06, "loss": 0.5571, "step": 638 }, { "epoch": 1.8666017526777021, "grad_norm": 0.5016476079074594, "learning_rate": 3.7456744590776807e-06, "loss": 0.554, "step": 639 }, { "epoch": 1.86952288218111, "grad_norm": 0.5522667996432958, "learning_rate": 3.729207590941753e-06, "loss": 0.5684, "step": 640 }, { "epoch": 1.872444011684518, "grad_norm": 0.5063071640192653, "learning_rate": 3.712755444932018e-06, "loss": 0.5536, "step": 641 }, { "epoch": 1.875365141187926, "grad_norm": 0.47979895935138983, "learning_rate": 3.6963182116465358e-06, "loss": 0.5607, "step": 642 }, { "epoch": 1.878286270691334, "grad_norm": 0.5466280214439997, "learning_rate": 3.6798960815106065e-06, "loss": 0.5912, "step": 643 }, { "epoch": 1.881207400194742, "grad_norm": 0.50137359619968, "learning_rate": 3.663489244774555e-06, "loss": 0.6032, "step": 644 }, { "epoch": 1.8841285296981498, "grad_norm": 0.5176702766961846, "learning_rate": 3.647097891511536e-06, "loss": 0.5441, "step": 645 }, { "epoch": 1.887049659201558, "grad_norm": 0.45011045640480307, "learning_rate": 3.630722211615328e-06, "loss": 0.5064, "step": 646 }, { "epoch": 1.889970788704966, "grad_norm": 0.5394614468332969, "learning_rate": 3.614362394798131e-06, "loss": 0.6368, "step": 647 }, { "epoch": 1.8928919182083739, "grad_norm": 0.4874425981153431, "learning_rate": 3.5980186305883746e-06, "loss": 0.5701, "step": 648 }, { "epoch": 1.895813047711782, "grad_norm": 0.507766689537816, "learning_rate": 3.5816911083285165e-06, "loss": 0.5798, "step": 649 }, { "epoch": 1.8987341772151898, "grad_norm": 0.4863274459783003, "learning_rate": 3.565380017172854e-06, "loss": 0.5244, "step": 650 }, { "epoch": 1.901655306718598, "grad_norm": 0.5805740988868203, "learning_rate": 3.5490855460853275e-06, "loss": 0.6026, "step": 651 }, { "epoch": 1.904576436222006, "grad_norm": 0.5151920826178348, "learning_rate": 3.5328078838373338e-06, "loss": 0.5578, "step": 652 }, { "epoch": 1.9074975657254138, "grad_norm": 0.4988300821200546, "learning_rate": 3.516547219005542e-06, "loss": 0.5639, "step": 653 }, { "epoch": 1.9104186952288218, "grad_norm": 0.5020574341639038, "learning_rate": 3.500303739969704e-06, "loss": 0.5553, "step": 654 }, { "epoch": 1.9133398247322297, "grad_norm": 0.47286299047516567, "learning_rate": 3.4840776349104755e-06, "loss": 0.5649, "step": 655 }, { "epoch": 1.916260954235638, "grad_norm": 0.4681307130388586, "learning_rate": 3.4678690918072335e-06, "loss": 0.5682, "step": 656 }, { "epoch": 1.9191820837390456, "grad_norm": 0.46768532251073486, "learning_rate": 3.451678298435902e-06, "loss": 0.5486, "step": 657 }, { "epoch": 1.9221032132424538, "grad_norm": 0.4674675744224569, "learning_rate": 3.4355054423667712e-06, "loss": 0.5421, "step": 658 }, { "epoch": 1.9250243427458618, "grad_norm": 0.5073113729998869, "learning_rate": 3.4193507109623323e-06, "loss": 0.5802, "step": 659 }, { "epoch": 1.9279454722492697, "grad_norm": 0.47796072007470125, "learning_rate": 3.4032142913750956e-06, "loss": 0.5454, "step": 660 }, { "epoch": 1.9308666017526777, "grad_norm": 0.44978612526614165, "learning_rate": 3.3870963705454353e-06, "loss": 0.6016, "step": 661 }, { "epoch": 1.9337877312560856, "grad_norm": 0.49209754086259494, "learning_rate": 3.370997135199413e-06, "loss": 0.5898, "step": 662 }, { "epoch": 1.9367088607594938, "grad_norm": 0.5261340174516683, "learning_rate": 3.3549167718466245e-06, "loss": 0.5438, "step": 663 }, { "epoch": 1.9396299902629015, "grad_norm": 0.5215555650313894, "learning_rate": 3.3388554667780272e-06, "loss": 0.59, "step": 664 }, { "epoch": 1.9425511197663097, "grad_norm": 0.4360310531256464, "learning_rate": 3.322813406063794e-06, "loss": 0.5361, "step": 665 }, { "epoch": 1.9454722492697176, "grad_norm": 0.4338971494593997, "learning_rate": 3.3067907755511473e-06, "loss": 0.5217, "step": 666 }, { "epoch": 1.9483933787731256, "grad_norm": 0.5993978224848855, "learning_rate": 3.290787760862215e-06, "loss": 0.6499, "step": 667 }, { "epoch": 1.9513145082765337, "grad_norm": 0.4277968919618892, "learning_rate": 3.274804547391872e-06, "loss": 0.5059, "step": 668 }, { "epoch": 1.9542356377799415, "grad_norm": 0.4833235638298786, "learning_rate": 3.258841320305598e-06, "loss": 0.6043, "step": 669 }, { "epoch": 1.9571567672833496, "grad_norm": 0.4876241950113585, "learning_rate": 3.242898264537331e-06, "loss": 0.5982, "step": 670 }, { "epoch": 1.9600778967867576, "grad_norm": 0.4287569825122473, "learning_rate": 3.226975564787322e-06, "loss": 0.5291, "step": 671 }, { "epoch": 1.9629990262901655, "grad_norm": 0.559198431896861, "learning_rate": 3.211073405520001e-06, "loss": 0.5722, "step": 672 }, { "epoch": 1.9659201557935735, "grad_norm": 0.5529747384120326, "learning_rate": 3.195191970961833e-06, "loss": 0.5766, "step": 673 }, { "epoch": 1.9688412852969814, "grad_norm": 0.4620747878088523, "learning_rate": 3.1793314450991895e-06, "loss": 0.5911, "step": 674 }, { "epoch": 1.9717624148003896, "grad_norm": 0.46006255679284874, "learning_rate": 3.1634920116762175e-06, "loss": 0.5622, "step": 675 }, { "epoch": 1.9746835443037973, "grad_norm": 0.4259271639046577, "learning_rate": 3.1476738541926993e-06, "loss": 0.5287, "step": 676 }, { "epoch": 1.9776046738072055, "grad_norm": 0.4729000058273843, "learning_rate": 3.1318771559019455e-06, "loss": 0.5863, "step": 677 }, { "epoch": 1.9805258033106135, "grad_norm": 0.5100371866983662, "learning_rate": 3.1161020998086566e-06, "loss": 0.5687, "step": 678 }, { "epoch": 1.9834469328140214, "grad_norm": 0.4379626484222088, "learning_rate": 3.1003488686668076e-06, "loss": 0.5155, "step": 679 }, { "epoch": 1.9863680623174296, "grad_norm": 0.4844416917506815, "learning_rate": 3.0846176449775363e-06, "loss": 0.5782, "step": 680 }, { "epoch": 1.9892891918208373, "grad_norm": 0.48815074844900985, "learning_rate": 3.0689086109870188e-06, "loss": 0.5709, "step": 681 }, { "epoch": 1.9922103213242455, "grad_norm": 0.5064903715857387, "learning_rate": 3.0532219486843686e-06, "loss": 0.596, "step": 682 }, { "epoch": 1.9951314508276532, "grad_norm": 0.492762116678178, "learning_rate": 3.0375578397995178e-06, "loss": 0.5766, "step": 683 }, { "epoch": 1.9980525803310614, "grad_norm": 0.44675106308795903, "learning_rate": 3.021916465801122e-06, "loss": 0.5206, "step": 684 }, { "epoch": 2.0009737098344695, "grad_norm": 0.9800800703433351, "learning_rate": 3.0062980078944515e-06, "loss": 0.8991, "step": 685 }, { "epoch": 2.0038948393378773, "grad_norm": 0.4864022689636578, "learning_rate": 2.990702647019294e-06, "loss": 0.4938, "step": 686 }, { "epoch": 2.0068159688412854, "grad_norm": 0.47889455472294706, "learning_rate": 2.9751305638478555e-06, "loss": 0.5913, "step": 687 }, { "epoch": 2.009737098344693, "grad_norm": 0.41456980849801855, "learning_rate": 2.9595819387826753e-06, "loss": 0.4387, "step": 688 }, { "epoch": 2.0126582278481013, "grad_norm": 0.5149845371997606, "learning_rate": 2.9440569519545258e-06, "loss": 0.5165, "step": 689 }, { "epoch": 2.015579357351509, "grad_norm": 0.5752113401087547, "learning_rate": 2.9285557832203328e-06, "loss": 0.5373, "step": 690 }, { "epoch": 2.0185004868549172, "grad_norm": 0.5062618329379694, "learning_rate": 2.9130786121610866e-06, "loss": 0.4947, "step": 691 }, { "epoch": 2.0214216163583254, "grad_norm": 0.49684882097652555, "learning_rate": 2.897625618079769e-06, "loss": 0.5208, "step": 692 }, { "epoch": 2.024342745861733, "grad_norm": 0.53299035788322, "learning_rate": 2.88219697999927e-06, "loss": 0.5438, "step": 693 }, { "epoch": 2.0272638753651413, "grad_norm": 0.5543675569999611, "learning_rate": 2.8667928766603115e-06, "loss": 0.5444, "step": 694 }, { "epoch": 2.030185004868549, "grad_norm": 0.4572356814121919, "learning_rate": 2.851413486519388e-06, "loss": 0.4634, "step": 695 }, { "epoch": 2.033106134371957, "grad_norm": 0.48340574114252344, "learning_rate": 2.8360589877466848e-06, "loss": 0.5391, "step": 696 }, { "epoch": 2.036027263875365, "grad_norm": 0.5141120553439197, "learning_rate": 2.8207295582240248e-06, "loss": 0.4871, "step": 697 }, { "epoch": 2.038948393378773, "grad_norm": 0.4913089013833701, "learning_rate": 2.8054253755428017e-06, "loss": 0.5574, "step": 698 }, { "epoch": 2.0418695228821813, "grad_norm": 0.4601265937011902, "learning_rate": 2.7901466170019242e-06, "loss": 0.4977, "step": 699 }, { "epoch": 2.044790652385589, "grad_norm": 0.5120604434660413, "learning_rate": 2.774893459605766e-06, "loss": 0.56, "step": 700 }, { "epoch": 2.047711781888997, "grad_norm": 0.47951033771571855, "learning_rate": 2.7596660800621076e-06, "loss": 0.5028, "step": 701 }, { "epoch": 2.050632911392405, "grad_norm": 0.48501372243432067, "learning_rate": 2.744464654780096e-06, "loss": 0.5174, "step": 702 }, { "epoch": 2.053554040895813, "grad_norm": 0.4741865276944199, "learning_rate": 2.7292893598681934e-06, "loss": 0.5104, "step": 703 }, { "epoch": 2.0564751703992212, "grad_norm": 0.48778437702309935, "learning_rate": 2.714140371132149e-06, "loss": 0.5375, "step": 704 }, { "epoch": 2.059396299902629, "grad_norm": 0.49018615472755583, "learning_rate": 2.69901786407295e-06, "loss": 0.5105, "step": 705 }, { "epoch": 2.062317429406037, "grad_norm": 0.48782945068406763, "learning_rate": 2.6839220138847966e-06, "loss": 0.4899, "step": 706 }, { "epoch": 2.065238558909445, "grad_norm": 0.4732567901538321, "learning_rate": 2.6688529954530612e-06, "loss": 0.5388, "step": 707 }, { "epoch": 2.068159688412853, "grad_norm": 0.4935373768584108, "learning_rate": 2.653810983352282e-06, "loss": 0.5822, "step": 708 }, { "epoch": 2.0710808179162608, "grad_norm": 0.44379266935434936, "learning_rate": 2.6387961518441223e-06, "loss": 0.4639, "step": 709 }, { "epoch": 2.074001947419669, "grad_norm": 0.47555161755659076, "learning_rate": 2.6238086748753587e-06, "loss": 0.5153, "step": 710 }, { "epoch": 2.076923076923077, "grad_norm": 0.4924498329014108, "learning_rate": 2.6088487260758643e-06, "loss": 0.5245, "step": 711 }, { "epoch": 2.079844206426485, "grad_norm": 0.4492288702216835, "learning_rate": 2.593916478756605e-06, "loss": 0.4816, "step": 712 }, { "epoch": 2.082765335929893, "grad_norm": 0.5008247473603998, "learning_rate": 2.579012105907619e-06, "loss": 0.5824, "step": 713 }, { "epoch": 2.0856864654333007, "grad_norm": 0.479453244836006, "learning_rate": 2.5641357801960186e-06, "loss": 0.4738, "step": 714 }, { "epoch": 2.088607594936709, "grad_norm": 0.49048895532662706, "learning_rate": 2.5492876739639912e-06, "loss": 0.5299, "step": 715 }, { "epoch": 2.091528724440117, "grad_norm": 0.5369724976128885, "learning_rate": 2.534467959226806e-06, "loss": 0.5162, "step": 716 }, { "epoch": 2.094449853943525, "grad_norm": 0.5137260447049424, "learning_rate": 2.519676807670811e-06, "loss": 0.5081, "step": 717 }, { "epoch": 2.097370983446933, "grad_norm": 0.4930859741487903, "learning_rate": 2.504914390651453e-06, "loss": 0.5576, "step": 718 }, { "epoch": 2.1002921129503407, "grad_norm": 0.4775125343363975, "learning_rate": 2.4901808791912864e-06, "loss": 0.4998, "step": 719 }, { "epoch": 2.103213242453749, "grad_norm": 0.5214779418043487, "learning_rate": 2.475476443977996e-06, "loss": 0.5696, "step": 720 }, { "epoch": 2.1061343719571566, "grad_norm": 0.47977761871037206, "learning_rate": 2.460801255362425e-06, "loss": 0.4764, "step": 721 }, { "epoch": 2.1090555014605648, "grad_norm": 0.48061705012654404, "learning_rate": 2.446155483356582e-06, "loss": 0.5357, "step": 722 }, { "epoch": 2.111976630963973, "grad_norm": 0.45629547557990546, "learning_rate": 2.4315392976316923e-06, "loss": 0.4915, "step": 723 }, { "epoch": 2.1148977604673806, "grad_norm": 0.4780522247938775, "learning_rate": 2.4169528675162256e-06, "loss": 0.4838, "step": 724 }, { "epoch": 2.117818889970789, "grad_norm": 0.49801842180761746, "learning_rate": 2.40239636199393e-06, "loss": 0.537, "step": 725 }, { "epoch": 2.1207400194741965, "grad_norm": 0.4846689678826097, "learning_rate": 2.3878699497018763e-06, "loss": 0.549, "step": 726 }, { "epoch": 2.1236611489776047, "grad_norm": 0.5018761582798917, "learning_rate": 2.373373798928507e-06, "loss": 0.5042, "step": 727 }, { "epoch": 2.1265822784810124, "grad_norm": 0.5072064832309269, "learning_rate": 2.358908077611684e-06, "loss": 0.5795, "step": 728 }, { "epoch": 2.1295034079844206, "grad_norm": 0.46610002849697485, "learning_rate": 2.344472953336747e-06, "loss": 0.468, "step": 729 }, { "epoch": 2.132424537487829, "grad_norm": 0.5002554866584938, "learning_rate": 2.3300685933345656e-06, "loss": 0.5022, "step": 730 }, { "epoch": 2.1353456669912365, "grad_norm": 0.5324744280306557, "learning_rate": 2.3156951644796065e-06, "loss": 0.5877, "step": 731 }, { "epoch": 2.1382667964946447, "grad_norm": 0.4804432916120911, "learning_rate": 2.3013528332879976e-06, "loss": 0.4814, "step": 732 }, { "epoch": 2.1411879259980524, "grad_norm": 0.47091232830887136, "learning_rate": 2.287041765915606e-06, "loss": 0.4953, "step": 733 }, { "epoch": 2.1441090555014606, "grad_norm": 0.4535727155626379, "learning_rate": 2.272762128156101e-06, "loss": 0.5269, "step": 734 }, { "epoch": 2.1470301850048688, "grad_norm": 0.47112551281102016, "learning_rate": 2.2585140854390432e-06, "loss": 0.5244, "step": 735 }, { "epoch": 2.1499513145082765, "grad_norm": 0.4545532988740129, "learning_rate": 2.2442978028279634e-06, "loss": 0.5341, "step": 736 }, { "epoch": 2.1528724440116846, "grad_norm": 0.5024829916260531, "learning_rate": 2.2301134450184535e-06, "loss": 0.5243, "step": 737 }, { "epoch": 2.1557935735150924, "grad_norm": 0.457679806572502, "learning_rate": 2.215961176336255e-06, "loss": 0.5157, "step": 738 }, { "epoch": 2.1587147030185005, "grad_norm": 0.41799952326682965, "learning_rate": 2.2018411607353572e-06, "loss": 0.4653, "step": 739 }, { "epoch": 2.1616358325219083, "grad_norm": 0.4787265660707973, "learning_rate": 2.187753561796097e-06, "loss": 0.5521, "step": 740 }, { "epoch": 2.1645569620253164, "grad_norm": 0.48532252970802797, "learning_rate": 2.1736985427232684e-06, "loss": 0.5736, "step": 741 }, { "epoch": 2.1674780915287246, "grad_norm": 0.46670045472968474, "learning_rate": 2.159676266344222e-06, "loss": 0.4856, "step": 742 }, { "epoch": 2.1703992210321323, "grad_norm": 0.456907097163635, "learning_rate": 2.1456868951069875e-06, "loss": 0.513, "step": 743 }, { "epoch": 2.1733203505355405, "grad_norm": 0.43480032818372993, "learning_rate": 2.1317305910783886e-06, "loss": 0.5183, "step": 744 }, { "epoch": 2.1762414800389482, "grad_norm": 0.4646674951011356, "learning_rate": 2.117807515942163e-06, "loss": 0.5169, "step": 745 }, { "epoch": 2.1791626095423564, "grad_norm": 0.5402952715425572, "learning_rate": 2.1039178309970975e-06, "loss": 0.5256, "step": 746 }, { "epoch": 2.1820837390457646, "grad_norm": 0.44604300646980855, "learning_rate": 2.090061697155147e-06, "loss": 0.5264, "step": 747 }, { "epoch": 2.1850048685491723, "grad_norm": 0.4855629840620991, "learning_rate": 2.076239274939582e-06, "loss": 0.497, "step": 748 }, { "epoch": 2.1879259980525805, "grad_norm": 0.450371819506377, "learning_rate": 2.062450724483118e-06, "loss": 0.492, "step": 749 }, { "epoch": 2.190847127555988, "grad_norm": 0.4675791607919664, "learning_rate": 2.0486962055260744e-06, "loss": 0.4833, "step": 750 }, { "epoch": 2.1937682570593964, "grad_norm": 0.4931924675592549, "learning_rate": 2.03497587741451e-06, "loss": 0.5325, "step": 751 }, { "epoch": 2.196689386562804, "grad_norm": 0.4728333199503108, "learning_rate": 2.021289899098384e-06, "loss": 0.5753, "step": 752 }, { "epoch": 2.1996105160662123, "grad_norm": 0.44477975654172386, "learning_rate": 2.0076384291297134e-06, "loss": 0.5095, "step": 753 }, { "epoch": 2.2025316455696204, "grad_norm": 0.42502487470067785, "learning_rate": 1.994021625660737e-06, "loss": 0.5004, "step": 754 }, { "epoch": 2.205452775073028, "grad_norm": 0.4682554593875959, "learning_rate": 1.9804396464420798e-06, "loss": 0.5377, "step": 755 }, { "epoch": 2.2083739045764363, "grad_norm": 0.478543398252641, "learning_rate": 1.966892648820932e-06, "loss": 0.5214, "step": 756 }, { "epoch": 2.211295034079844, "grad_norm": 0.4638445072960299, "learning_rate": 1.953380789739216e-06, "loss": 0.522, "step": 757 }, { "epoch": 2.2142161635832522, "grad_norm": 0.44455755591464263, "learning_rate": 1.939904225731783e-06, "loss": 0.5279, "step": 758 }, { "epoch": 2.21713729308666, "grad_norm": 0.43381712587340415, "learning_rate": 1.9264631129245836e-06, "loss": 0.5082, "step": 759 }, { "epoch": 2.220058422590068, "grad_norm": 0.46180168699161817, "learning_rate": 1.9130576070328695e-06, "loss": 0.5129, "step": 760 }, { "epoch": 2.2229795520934763, "grad_norm": 0.4554045010755852, "learning_rate": 1.8996878633593829e-06, "loss": 0.499, "step": 761 }, { "epoch": 2.225900681596884, "grad_norm": 0.4749182830100526, "learning_rate": 1.8863540367925676e-06, "loss": 0.5343, "step": 762 }, { "epoch": 2.228821811100292, "grad_norm": 0.45836107720097835, "learning_rate": 1.873056281804762e-06, "loss": 0.4869, "step": 763 }, { "epoch": 2.2317429406037, "grad_norm": 0.45195906937962993, "learning_rate": 1.8597947524504178e-06, "loss": 0.5201, "step": 764 }, { "epoch": 2.234664070107108, "grad_norm": 0.45112191767699233, "learning_rate": 1.8465696023643115e-06, "loss": 0.5354, "step": 765 }, { "epoch": 2.2375851996105163, "grad_norm": 0.49025509449815313, "learning_rate": 1.8333809847597644e-06, "loss": 0.558, "step": 766 }, { "epoch": 2.240506329113924, "grad_norm": 0.4486301410086504, "learning_rate": 1.8202290524268761e-06, "loss": 0.4795, "step": 767 }, { "epoch": 2.243427458617332, "grad_norm": 0.45776452388300076, "learning_rate": 1.8071139577307416e-06, "loss": 0.5228, "step": 768 }, { "epoch": 2.24634858812074, "grad_norm": 0.4414886979982915, "learning_rate": 1.7940358526096885e-06, "loss": 0.4903, "step": 769 }, { "epoch": 2.249269717624148, "grad_norm": 0.44672029042681155, "learning_rate": 1.7809948885735295e-06, "loss": 0.5906, "step": 770 }, { "epoch": 2.252190847127556, "grad_norm": 0.40774965094554166, "learning_rate": 1.7679912167017922e-06, "loss": 0.4726, "step": 771 }, { "epoch": 2.255111976630964, "grad_norm": 0.42904714451405873, "learning_rate": 1.7550249876419746e-06, "loss": 0.5263, "step": 772 }, { "epoch": 2.258033106134372, "grad_norm": 0.46232426979933267, "learning_rate": 1.7420963516078016e-06, "loss": 0.5207, "step": 773 }, { "epoch": 2.26095423563778, "grad_norm": 0.43964498189412426, "learning_rate": 1.7292054583774809e-06, "loss": 0.5634, "step": 774 }, { "epoch": 2.263875365141188, "grad_norm": 0.4614885797674806, "learning_rate": 1.7163524572919748e-06, "loss": 0.5167, "step": 775 }, { "epoch": 2.2667964946445958, "grad_norm": 0.4326167271312955, "learning_rate": 1.7035374972532593e-06, "loss": 0.4734, "step": 776 }, { "epoch": 2.269717624148004, "grad_norm": 0.477108582339082, "learning_rate": 1.6907607267226079e-06, "loss": 0.5436, "step": 777 }, { "epoch": 2.272638753651412, "grad_norm": 0.45201769752670545, "learning_rate": 1.6780222937188662e-06, "loss": 0.5207, "step": 778 }, { "epoch": 2.27555988315482, "grad_norm": 0.48262208131523315, "learning_rate": 1.665322345816746e-06, "loss": 0.5304, "step": 779 }, { "epoch": 2.278481012658228, "grad_norm": 0.45623936201495924, "learning_rate": 1.6526610301451028e-06, "loss": 0.4604, "step": 780 }, { "epoch": 2.2814021421616357, "grad_norm": 0.45941621411519823, "learning_rate": 1.6400384933852403e-06, "loss": 0.508, "step": 781 }, { "epoch": 2.284323271665044, "grad_norm": 0.4640162246141768, "learning_rate": 1.6274548817692088e-06, "loss": 0.5429, "step": 782 }, { "epoch": 2.2872444011684516, "grad_norm": 0.4517120340093356, "learning_rate": 1.6149103410781086e-06, "loss": 0.5126, "step": 783 }, { "epoch": 2.29016553067186, "grad_norm": 0.4528239429323758, "learning_rate": 1.6024050166404097e-06, "loss": 0.4995, "step": 784 }, { "epoch": 2.293086660175268, "grad_norm": 0.4440949852742247, "learning_rate": 1.5899390533302538e-06, "loss": 0.5333, "step": 785 }, { "epoch": 2.2960077896786757, "grad_norm": 0.3970036577732869, "learning_rate": 1.5775125955657877e-06, "loss": 0.5012, "step": 786 }, { "epoch": 2.298928919182084, "grad_norm": 0.4484643953022932, "learning_rate": 1.5651257873074898e-06, "loss": 0.5413, "step": 787 }, { "epoch": 2.3018500486854916, "grad_norm": 0.4679410901074801, "learning_rate": 1.5527787720564946e-06, "loss": 0.5229, "step": 788 }, { "epoch": 2.3047711781888998, "grad_norm": 0.450714999626221, "learning_rate": 1.5404716928529356e-06, "loss": 0.5079, "step": 789 }, { "epoch": 2.3076923076923075, "grad_norm": 0.46250776338412736, "learning_rate": 1.5282046922742876e-06, "loss": 0.5056, "step": 790 }, { "epoch": 2.3106134371957157, "grad_norm": 0.47191060686599, "learning_rate": 1.515977912433717e-06, "loss": 0.5279, "step": 791 }, { "epoch": 2.313534566699124, "grad_norm": 0.493016587785878, "learning_rate": 1.50379149497843e-06, "loss": 0.5483, "step": 792 }, { "epoch": 2.3164556962025316, "grad_norm": 0.47810281591953624, "learning_rate": 1.4916455810880358e-06, "loss": 0.4898, "step": 793 }, { "epoch": 2.3193768257059397, "grad_norm": 0.4478075131535074, "learning_rate": 1.4795403114729095e-06, "loss": 0.5101, "step": 794 }, { "epoch": 2.3222979552093475, "grad_norm": 0.4875641000416008, "learning_rate": 1.4674758263725614e-06, "loss": 0.5854, "step": 795 }, { "epoch": 2.3252190847127556, "grad_norm": 0.42525631573701944, "learning_rate": 1.4554522655540176e-06, "loss": 0.4989, "step": 796 }, { "epoch": 2.3281402142161634, "grad_norm": 0.41706543335088775, "learning_rate": 1.4434697683101928e-06, "loss": 0.516, "step": 797 }, { "epoch": 2.3310613437195715, "grad_norm": 0.4524630863860165, "learning_rate": 1.4315284734582802e-06, "loss": 0.4936, "step": 798 }, { "epoch": 2.3339824732229797, "grad_norm": 0.5066460944480237, "learning_rate": 1.4196285193381431e-06, "loss": 0.5665, "step": 799 }, { "epoch": 2.3369036027263874, "grad_norm": 0.4886439078673485, "learning_rate": 1.4077700438107183e-06, "loss": 0.5191, "step": 800 }, { "epoch": 2.3398247322297956, "grad_norm": 0.42502615563238366, "learning_rate": 1.3959531842564046e-06, "loss": 0.4785, "step": 801 }, { "epoch": 2.3427458617332033, "grad_norm": 0.45982217381683227, "learning_rate": 1.3841780775734847e-06, "loss": 0.5643, "step": 802 }, { "epoch": 2.3456669912366115, "grad_norm": 0.43225051785834356, "learning_rate": 1.3724448601765328e-06, "loss": 0.5629, "step": 803 }, { "epoch": 2.3485881207400197, "grad_norm": 0.44393943057319624, "learning_rate": 1.3607536679948397e-06, "loss": 0.456, "step": 804 }, { "epoch": 2.3515092502434274, "grad_norm": 0.44935770423521926, "learning_rate": 1.3491046364708294e-06, "loss": 0.52, "step": 805 }, { "epoch": 2.3544303797468356, "grad_norm": 0.4573183127076476, "learning_rate": 1.337497900558497e-06, "loss": 0.5231, "step": 806 }, { "epoch": 2.3573515092502433, "grad_norm": 0.4471873705684597, "learning_rate": 1.325933594721841e-06, "loss": 0.5191, "step": 807 }, { "epoch": 2.3602726387536515, "grad_norm": 0.46232878806411926, "learning_rate": 1.3144118529333126e-06, "loss": 0.4641, "step": 808 }, { "epoch": 2.3631937682570596, "grad_norm": 0.46569289026754496, "learning_rate": 1.3029328086722537e-06, "loss": 0.5363, "step": 809 }, { "epoch": 2.3661148977604674, "grad_norm": 0.4515090725602521, "learning_rate": 1.2914965949233572e-06, "loss": 0.5278, "step": 810 }, { "epoch": 2.3690360272638755, "grad_norm": 0.45127137221258873, "learning_rate": 1.2801033441751244e-06, "loss": 0.4919, "step": 811 }, { "epoch": 2.3719571567672832, "grad_norm": 0.45307816143888285, "learning_rate": 1.268753188418329e-06, "loss": 0.5063, "step": 812 }, { "epoch": 2.3748782862706914, "grad_norm": 0.4681309102712795, "learning_rate": 1.257446259144494e-06, "loss": 0.5185, "step": 813 }, { "epoch": 2.377799415774099, "grad_norm": 0.4665552455306928, "learning_rate": 1.24618268734436e-06, "loss": 0.5426, "step": 814 }, { "epoch": 2.3807205452775073, "grad_norm": 0.45997968978470405, "learning_rate": 1.2349626035063705e-06, "loss": 0.4996, "step": 815 }, { "epoch": 2.3836416747809155, "grad_norm": 0.470113793180687, "learning_rate": 1.2237861376151632e-06, "loss": 0.5392, "step": 816 }, { "epoch": 2.386562804284323, "grad_norm": 0.47230105708445114, "learning_rate": 1.2126534191500622e-06, "loss": 0.5361, "step": 817 }, { "epoch": 2.3894839337877314, "grad_norm": 0.4439367141535501, "learning_rate": 1.2015645770835765e-06, "loss": 0.4732, "step": 818 }, { "epoch": 2.392405063291139, "grad_norm": 0.4656016270422327, "learning_rate": 1.1905197398799074e-06, "loss": 0.5227, "step": 819 }, { "epoch": 2.3953261927945473, "grad_norm": 0.46417888892808024, "learning_rate": 1.1795190354934587e-06, "loss": 0.5606, "step": 820 }, { "epoch": 2.398247322297955, "grad_norm": 0.41153069172218604, "learning_rate": 1.168562591367361e-06, "loss": 0.5035, "step": 821 }, { "epoch": 2.401168451801363, "grad_norm": 0.4591933177826969, "learning_rate": 1.1576505344319843e-06, "loss": 0.5485, "step": 822 }, { "epoch": 2.4040895813047714, "grad_norm": 0.48641885058814227, "learning_rate": 1.146782991103475e-06, "loss": 0.5122, "step": 823 }, { "epoch": 2.407010710808179, "grad_norm": 0.452636414728882, "learning_rate": 1.1359600872822879e-06, "loss": 0.528, "step": 824 }, { "epoch": 2.4099318403115872, "grad_norm": 0.4637194713579063, "learning_rate": 1.1251819483517334e-06, "loss": 0.5335, "step": 825 }, { "epoch": 2.412852969814995, "grad_norm": 0.46499658066713967, "learning_rate": 1.1144486991765175e-06, "loss": 0.519, "step": 826 }, { "epoch": 2.415774099318403, "grad_norm": 0.4519765342482912, "learning_rate": 1.1037604641012995e-06, "loss": 0.5488, "step": 827 }, { "epoch": 2.418695228821811, "grad_norm": 0.4382247666420335, "learning_rate": 1.0931173669492472e-06, "loss": 0.4853, "step": 828 }, { "epoch": 2.421616358325219, "grad_norm": 0.45306336606053793, "learning_rate": 1.0825195310206132e-06, "loss": 0.5224, "step": 829 }, { "epoch": 2.424537487828627, "grad_norm": 0.4473053801916244, "learning_rate": 1.0719670790912928e-06, "loss": 0.4971, "step": 830 }, { "epoch": 2.427458617332035, "grad_norm": 0.44465614500766465, "learning_rate": 1.0614601334114099e-06, "loss": 0.5404, "step": 831 }, { "epoch": 2.430379746835443, "grad_norm": 0.4251639911624754, "learning_rate": 1.0509988157038952e-06, "loss": 0.5179, "step": 832 }, { "epoch": 2.433300876338851, "grad_norm": 0.40203118130358384, "learning_rate": 1.0405832471630862e-06, "loss": 0.4465, "step": 833 }, { "epoch": 2.436222005842259, "grad_norm": 0.49627881222358805, "learning_rate": 1.030213548453311e-06, "loss": 0.5782, "step": 834 }, { "epoch": 2.439143135345667, "grad_norm": 0.4352033965269781, "learning_rate": 1.019889839707498e-06, "loss": 0.471, "step": 835 }, { "epoch": 2.442064264849075, "grad_norm": 0.4112365791877552, "learning_rate": 1.0096122405257802e-06, "loss": 0.525, "step": 836 }, { "epoch": 2.444985394352483, "grad_norm": 0.46113597408698487, "learning_rate": 9.99380869974116e-07, "loss": 0.55, "step": 837 }, { "epoch": 2.447906523855891, "grad_norm": 0.42894572321057406, "learning_rate": 9.891958465828999e-07, "loss": 0.4883, "step": 838 }, { "epoch": 2.450827653359299, "grad_norm": 0.4306576340142323, "learning_rate": 9.790572883455974e-07, "loss": 0.5127, "step": 839 }, { "epoch": 2.453748782862707, "grad_norm": 0.47843377943065934, "learning_rate": 9.689653127173743e-07, "loss": 0.5376, "step": 840 }, { "epoch": 2.456669912366115, "grad_norm": 0.48286119801655825, "learning_rate": 9.589200366137375e-07, "loss": 0.5108, "step": 841 }, { "epoch": 2.459591041869523, "grad_norm": 0.4840439892081914, "learning_rate": 9.489215764091813e-07, "loss": 0.5436, "step": 842 }, { "epoch": 2.4625121713729308, "grad_norm": 0.4360507831964012, "learning_rate": 9.389700479358365e-07, "loss": 0.4733, "step": 843 }, { "epoch": 2.465433300876339, "grad_norm": 0.449450069287443, "learning_rate": 9.290655664821296e-07, "loss": 0.5599, "step": 844 }, { "epoch": 2.4683544303797467, "grad_norm": 0.41280059188581175, "learning_rate": 9.192082467914465e-07, "loss": 0.4792, "step": 845 }, { "epoch": 2.471275559883155, "grad_norm": 0.5093589634384059, "learning_rate": 9.093982030608095e-07, "loss": 0.5397, "step": 846 }, { "epoch": 2.474196689386563, "grad_norm": 0.45448988250512634, "learning_rate": 8.996355489395442e-07, "loss": 0.4975, "step": 847 }, { "epoch": 2.4771178188899707, "grad_norm": 0.44480146821328176, "learning_rate": 8.899203975279674e-07, "loss": 0.4914, "step": 848 }, { "epoch": 2.480038948393379, "grad_norm": 0.4896804183179627, "learning_rate": 8.802528613760775e-07, "loss": 0.5599, "step": 849 }, { "epoch": 2.4829600778967866, "grad_norm": 0.4529609749837371, "learning_rate": 8.706330524822548e-07, "loss": 0.4598, "step": 850 }, { "epoch": 2.485881207400195, "grad_norm": 0.4141270566207645, "learning_rate": 8.610610822919546e-07, "loss": 0.5195, "step": 851 }, { "epoch": 2.4888023369036025, "grad_norm": 0.4413472687231943, "learning_rate": 8.515370616964219e-07, "loss": 0.5181, "step": 852 }, { "epoch": 2.4917234664070107, "grad_norm": 0.4120555785416349, "learning_rate": 8.420611010314062e-07, "loss": 0.5267, "step": 853 }, { "epoch": 2.494644595910419, "grad_norm": 0.4661650695044543, "learning_rate": 8.326333100758843e-07, "loss": 0.5086, "step": 854 }, { "epoch": 2.4975657254138266, "grad_norm": 0.5073361612635612, "learning_rate": 8.232537980507848e-07, "loss": 0.5536, "step": 855 }, { "epoch": 2.5004868549172348, "grad_norm": 0.4961949890878462, "learning_rate": 8.13922673617727e-07, "loss": 0.5238, "step": 856 }, { "epoch": 2.5034079844206425, "grad_norm": 0.4720376448013444, "learning_rate": 8.046400448777575e-07, "loss": 0.4823, "step": 857 }, { "epoch": 2.5063291139240507, "grad_norm": 0.49809468171059407, "learning_rate": 7.954060193701019e-07, "loss": 0.5466, "step": 858 }, { "epoch": 2.5092502434274584, "grad_norm": 0.4639329084639315, "learning_rate": 7.862207040709191e-07, "loss": 0.5104, "step": 859 }, { "epoch": 2.5121713729308666, "grad_norm": 0.47993248882289974, "learning_rate": 7.770842053920585e-07, "loss": 0.5379, "step": 860 }, { "epoch": 2.5150925024342747, "grad_norm": 0.42897728377766037, "learning_rate": 7.679966291798297e-07, "loss": 0.482, "step": 861 }, { "epoch": 2.5180136319376825, "grad_norm": 0.45778561195288503, "learning_rate": 7.589580807137742e-07, "loss": 0.5032, "step": 862 }, { "epoch": 2.5209347614410906, "grad_norm": 0.46270673489951974, "learning_rate": 7.4996866470545e-07, "loss": 0.5706, "step": 863 }, { "epoch": 2.523855890944499, "grad_norm": 0.42072713048943156, "learning_rate": 7.410284852972127e-07, "loss": 0.5204, "step": 864 }, { "epoch": 2.5267770204479065, "grad_norm": 0.40846266187572267, "learning_rate": 7.321376460610136e-07, "loss": 0.4916, "step": 865 }, { "epoch": 2.5296981499513143, "grad_norm": 0.4443443498829621, "learning_rate": 7.232962499972002e-07, "loss": 0.5395, "step": 866 }, { "epoch": 2.5326192794547224, "grad_norm": 0.459368960082765, "learning_rate": 7.145043995333173e-07, "loss": 0.4924, "step": 867 }, { "epoch": 2.5355404089581306, "grad_norm": 0.47824210909228654, "learning_rate": 7.057621965229267e-07, "loss": 0.513, "step": 868 }, { "epoch": 2.5384615384615383, "grad_norm": 0.438214605329431, "learning_rate": 6.970697422444228e-07, "loss": 0.5695, "step": 869 }, { "epoch": 2.5413826679649465, "grad_norm": 0.45798584370168316, "learning_rate": 6.884271373998608e-07, "loss": 0.4925, "step": 870 }, { "epoch": 2.5443037974683547, "grad_norm": 0.4811217026678713, "learning_rate": 6.798344821137947e-07, "loss": 0.5083, "step": 871 }, { "epoch": 2.5472249269717624, "grad_norm": 0.4418809012550695, "learning_rate": 6.71291875932108e-07, "loss": 0.5352, "step": 872 }, { "epoch": 2.5501460564751706, "grad_norm": 0.42576085849329387, "learning_rate": 6.62799417820868e-07, "loss": 0.5262, "step": 873 }, { "epoch": 2.5530671859785783, "grad_norm": 0.4287609653415205, "learning_rate": 6.543572061651738e-07, "loss": 0.5399, "step": 874 }, { "epoch": 2.5559883154819865, "grad_norm": 0.45232434071279953, "learning_rate": 6.459653387680248e-07, "loss": 0.4725, "step": 875 }, { "epoch": 2.558909444985394, "grad_norm": 0.4554887075692202, "learning_rate": 6.376239128491784e-07, "loss": 0.5122, "step": 876 }, { "epoch": 2.5618305744888024, "grad_norm": 0.46277880760014206, "learning_rate": 6.293330250440277e-07, "loss": 0.4941, "step": 877 }, { "epoch": 2.5647517039922105, "grad_norm": 0.4962082268099505, "learning_rate": 6.210927714024834e-07, "loss": 0.5592, "step": 878 }, { "epoch": 2.5676728334956183, "grad_norm": 0.44324064458458867, "learning_rate": 6.129032473878582e-07, "loss": 0.5282, "step": 879 }, { "epoch": 2.5705939629990264, "grad_norm": 0.43320029335686094, "learning_rate": 6.047645478757635e-07, "loss": 0.4562, "step": 880 }, { "epoch": 2.573515092502434, "grad_norm": 0.45163259818241996, "learning_rate": 5.966767671530078e-07, "loss": 0.5501, "step": 881 }, { "epoch": 2.5764362220058423, "grad_norm": 0.4552021870162526, "learning_rate": 5.88639998916506e-07, "loss": 0.5359, "step": 882 }, { "epoch": 2.57935735150925, "grad_norm": 0.4604883255796744, "learning_rate": 5.806543362721945e-07, "loss": 0.5274, "step": 883 }, { "epoch": 2.5822784810126582, "grad_norm": 0.41454954591229876, "learning_rate": 5.727198717339511e-07, "loss": 0.4765, "step": 884 }, { "epoch": 2.5851996105160664, "grad_norm": 0.4325932298835654, "learning_rate": 5.648366972225222e-07, "loss": 0.5282, "step": 885 }, { "epoch": 2.588120740019474, "grad_norm": 0.433038404211758, "learning_rate": 5.570049040644609e-07, "loss": 0.5137, "step": 886 }, { "epoch": 2.5910418695228823, "grad_norm": 0.45496347053471886, "learning_rate": 5.492245829910664e-07, "loss": 0.4973, "step": 887 }, { "epoch": 2.59396299902629, "grad_norm": 0.44790019595001196, "learning_rate": 5.414958241373358e-07, "loss": 0.4881, "step": 888 }, { "epoch": 2.596884128529698, "grad_norm": 0.4506105888324243, "learning_rate": 5.33818717040917e-07, "loss": 0.5446, "step": 889 }, { "epoch": 2.599805258033106, "grad_norm": 0.45920237343152387, "learning_rate": 5.261933506410722e-07, "loss": 0.5544, "step": 890 }, { "epoch": 2.602726387536514, "grad_norm": 0.433968288589714, "learning_rate": 5.186198132776459e-07, "loss": 0.4792, "step": 891 }, { "epoch": 2.6056475170399223, "grad_norm": 0.45054587006233904, "learning_rate": 5.110981926900488e-07, "loss": 0.5271, "step": 892 }, { "epoch": 2.60856864654333, "grad_norm": 0.4402130423120147, "learning_rate": 5.036285760162307e-07, "loss": 0.4772, "step": 893 }, { "epoch": 2.611489776046738, "grad_norm": 0.4560401508104557, "learning_rate": 4.962110497916789e-07, "loss": 0.5078, "step": 894 }, { "epoch": 2.6144109055501463, "grad_norm": 0.4574934740247989, "learning_rate": 4.888456999484098e-07, "loss": 0.5177, "step": 895 }, { "epoch": 2.617332035053554, "grad_norm": 0.42797938213430475, "learning_rate": 4.815326118139813e-07, "loss": 0.5068, "step": 896 }, { "epoch": 2.620253164556962, "grad_norm": 0.4508198643702282, "learning_rate": 4.742718701104965e-07, "loss": 0.4836, "step": 897 }, { "epoch": 2.62317429406037, "grad_norm": 0.4936637557435792, "learning_rate": 4.670635589536254e-07, "loss": 0.5688, "step": 898 }, { "epoch": 2.626095423563778, "grad_norm": 0.4446810251135343, "learning_rate": 4.599077618516312e-07, "loss": 0.5213, "step": 899 }, { "epoch": 2.629016553067186, "grad_norm": 0.4393121918571539, "learning_rate": 4.528045617044019e-07, "loss": 0.515, "step": 900 }, { "epoch": 2.631937682570594, "grad_norm": 0.4546453714565344, "learning_rate": 4.457540408024896e-07, "loss": 0.5119, "step": 901 }, { "epoch": 2.634858812074002, "grad_norm": 0.43919005512991516, "learning_rate": 4.387562808261575e-07, "loss": 0.5273, "step": 902 }, { "epoch": 2.63777994157741, "grad_norm": 0.4491615831317555, "learning_rate": 4.3181136284443204e-07, "loss": 0.4994, "step": 903 }, { "epoch": 2.6407010710808176, "grad_norm": 0.48434931343594284, "learning_rate": 4.249193673141694e-07, "loss": 0.5186, "step": 904 }, { "epoch": 2.643622200584226, "grad_norm": 0.41259506287384623, "learning_rate": 4.180803740791156e-07, "loss": 0.5319, "step": 905 }, { "epoch": 2.646543330087634, "grad_norm": 0.4076433170920647, "learning_rate": 4.112944623689863e-07, "loss": 0.5178, "step": 906 }, { "epoch": 2.6494644595910417, "grad_norm": 0.421345630456112, "learning_rate": 4.0456171079854833e-07, "loss": 0.4951, "step": 907 }, { "epoch": 2.65238558909445, "grad_norm": 0.4435092242276153, "learning_rate": 3.978821973667074e-07, "loss": 0.4837, "step": 908 }, { "epoch": 2.655306718597858, "grad_norm": 0.44521873058867134, "learning_rate": 3.9125599945560866e-07, "loss": 0.5526, "step": 909 }, { "epoch": 2.6582278481012658, "grad_norm": 0.45505539256105115, "learning_rate": 3.846831938297324e-07, "loss": 0.5158, "step": 910 }, { "epoch": 2.661148977604674, "grad_norm": 0.433074017412774, "learning_rate": 3.7816385663501097e-07, "loss": 0.4809, "step": 911 }, { "epoch": 2.6640701071080817, "grad_norm": 0.4502224951078667, "learning_rate": 3.716980633979489e-07, "loss": 0.5363, "step": 912 }, { "epoch": 2.66699123661149, "grad_norm": 0.43979723124115566, "learning_rate": 3.6528588902473905e-07, "loss": 0.508, "step": 913 }, { "epoch": 2.6699123661148976, "grad_norm": 0.46901546979749387, "learning_rate": 3.589274078004029e-07, "loss": 0.5489, "step": 914 }, { "epoch": 2.6728334956183057, "grad_norm": 0.4432062648822634, "learning_rate": 3.5262269338792623e-07, "loss": 0.526, "step": 915 }, { "epoch": 2.675754625121714, "grad_norm": 0.43511333463780544, "learning_rate": 3.463718188274046e-07, "loss": 0.5332, "step": 916 }, { "epoch": 2.6786757546251216, "grad_norm": 0.38843912794513447, "learning_rate": 3.401748565352031e-07, "loss": 0.4997, "step": 917 }, { "epoch": 2.68159688412853, "grad_norm": 0.41986039406023057, "learning_rate": 3.340318783031099e-07, "loss": 0.4804, "step": 918 }, { "epoch": 2.6845180136319375, "grad_norm": 0.44648799987073357, "learning_rate": 3.279429552975094e-07, "loss": 0.57, "step": 919 }, { "epoch": 2.6874391431353457, "grad_norm": 0.38625226404630086, "learning_rate": 3.219081580585548e-07, "loss": 0.418, "step": 920 }, { "epoch": 2.6903602726387534, "grad_norm": 0.46495901890128216, "learning_rate": 3.159275564993558e-07, "loss": 0.562, "step": 921 }, { "epoch": 2.6932814021421616, "grad_norm": 0.4542095138457356, "learning_rate": 3.100012199051627e-07, "loss": 0.48, "step": 922 }, { "epoch": 2.6962025316455698, "grad_norm": 0.461034837465003, "learning_rate": 3.0412921693256657e-07, "loss": 0.5802, "step": 923 }, { "epoch": 2.6991236611489775, "grad_norm": 0.4124404226906978, "learning_rate": 2.9831161560870346e-07, "loss": 0.5124, "step": 924 }, { "epoch": 2.7020447906523857, "grad_norm": 0.420405602577591, "learning_rate": 2.9254848333046817e-07, "loss": 0.4959, "step": 925 }, { "epoch": 2.704965920155794, "grad_norm": 0.42637463336396936, "learning_rate": 2.8683988686372956e-07, "loss": 0.529, "step": 926 }, { "epoch": 2.7078870496592016, "grad_norm": 0.4156040595949887, "learning_rate": 2.811858923425609e-07, "loss": 0.5184, "step": 927 }, { "epoch": 2.7108081791626093, "grad_norm": 0.4312166608407746, "learning_rate": 2.755865652684703e-07, "loss": 0.5438, "step": 928 }, { "epoch": 2.7137293086660175, "grad_norm": 0.4369289965460609, "learning_rate": 2.7004197050964744e-07, "loss": 0.4978, "step": 929 }, { "epoch": 2.7166504381694256, "grad_norm": 0.426344227648129, "learning_rate": 2.645521723002037e-07, "loss": 0.5041, "step": 930 }, { "epoch": 2.7195715676728334, "grad_norm": 0.4417752376579622, "learning_rate": 2.591172342394349e-07, "loss": 0.5587, "step": 931 }, { "epoch": 2.7224926971762415, "grad_norm": 0.457814912149392, "learning_rate": 2.537372192910825e-07, "loss": 0.5175, "step": 932 }, { "epoch": 2.7254138266796497, "grad_norm": 0.45549699443779185, "learning_rate": 2.4841218978260194e-07, "loss": 0.5033, "step": 933 }, { "epoch": 2.7283349561830574, "grad_norm": 0.44509990954989737, "learning_rate": 2.4314220740444595e-07, "loss": 0.5192, "step": 934 }, { "epoch": 2.731256085686465, "grad_norm": 0.4542595414389855, "learning_rate": 2.3792733320934348e-07, "loss": 0.5193, "step": 935 }, { "epoch": 2.7341772151898733, "grad_norm": 0.46788823553431785, "learning_rate": 2.3276762761159588e-07, "loss": 0.5226, "step": 936 }, { "epoch": 2.7370983446932815, "grad_norm": 0.46548939441743264, "learning_rate": 2.276631503863763e-07, "loss": 0.5234, "step": 937 }, { "epoch": 2.7400194741966892, "grad_norm": 0.43071296003232673, "learning_rate": 2.2261396066903916e-07, "loss": 0.4968, "step": 938 }, { "epoch": 2.7429406037000974, "grad_norm": 0.4422590771192652, "learning_rate": 2.176201169544312e-07, "loss": 0.5135, "step": 939 }, { "epoch": 2.7458617332035056, "grad_norm": 0.43839294000442797, "learning_rate": 2.12681677096217e-07, "loss": 0.5459, "step": 940 }, { "epoch": 2.7487828627069133, "grad_norm": 0.4296191280132258, "learning_rate": 2.0779869830620692e-07, "loss": 0.4538, "step": 941 }, { "epoch": 2.7517039922103215, "grad_norm": 0.4119484440139385, "learning_rate": 2.029712371536957e-07, "loss": 0.4839, "step": 942 }, { "epoch": 2.754625121713729, "grad_norm": 0.4512158763524109, "learning_rate": 1.9819934956480603e-07, "loss": 0.5445, "step": 943 }, { "epoch": 2.7575462512171374, "grad_norm": 0.4212292286246555, "learning_rate": 1.9348309082184102e-07, "loss": 0.4934, "step": 944 }, { "epoch": 2.760467380720545, "grad_norm": 0.44496919909800775, "learning_rate": 1.888225155626433e-07, "loss": 0.5456, "step": 945 }, { "epoch": 2.7633885102239533, "grad_norm": 0.41713723572900646, "learning_rate": 1.8421767777996425e-07, "loss": 0.4462, "step": 946 }, { "epoch": 2.7663096397273614, "grad_norm": 0.4457624171080001, "learning_rate": 1.7966863082083507e-07, "loss": 0.5427, "step": 947 }, { "epoch": 2.769230769230769, "grad_norm": 0.4348007177346548, "learning_rate": 1.7517542738595071e-07, "loss": 0.5568, "step": 948 }, { "epoch": 2.7721518987341773, "grad_norm": 0.41399848976523823, "learning_rate": 1.7073811952905862e-07, "loss": 0.5083, "step": 949 }, { "epoch": 2.775073028237585, "grad_norm": 0.43232436981734923, "learning_rate": 1.6635675865635859e-07, "loss": 0.4505, "step": 950 }, { "epoch": 2.7779941577409932, "grad_norm": 0.4750164143516587, "learning_rate": 1.6203139552590175e-07, "loss": 0.5831, "step": 951 }, { "epoch": 2.780915287244401, "grad_norm": 0.39972505599387415, "learning_rate": 1.5776208024700702e-07, "loss": 0.4751, "step": 952 }, { "epoch": 2.783836416747809, "grad_norm": 0.42708956986618357, "learning_rate": 1.5354886227967934e-07, "loss": 0.532, "step": 953 }, { "epoch": 2.7867575462512173, "grad_norm": 0.4435213719711565, "learning_rate": 1.4939179043403474e-07, "loss": 0.5347, "step": 954 }, { "epoch": 2.789678675754625, "grad_norm": 0.4416374818886015, "learning_rate": 1.4529091286973994e-07, "loss": 0.5459, "step": 955 }, { "epoch": 2.792599805258033, "grad_norm": 0.45153869521194906, "learning_rate": 1.4124627709544814e-07, "loss": 0.4918, "step": 956 }, { "epoch": 2.7955209347614414, "grad_norm": 0.4440135260222283, "learning_rate": 1.3725792996825083e-07, "loss": 0.5016, "step": 957 }, { "epoch": 2.798442064264849, "grad_norm": 0.39484795107770926, "learning_rate": 1.3332591769314006e-07, "loss": 0.5043, "step": 958 }, { "epoch": 2.801363193768257, "grad_norm": 0.444235962409132, "learning_rate": 1.2945028582246445e-07, "loss": 0.5637, "step": 959 }, { "epoch": 2.804284323271665, "grad_norm": 0.4583481416669065, "learning_rate": 1.2563107925540774e-07, "loss": 0.5017, "step": 960 }, { "epoch": 2.807205452775073, "grad_norm": 0.4215520268166574, "learning_rate": 1.2186834223746612e-07, "loss": 0.5136, "step": 961 }, { "epoch": 2.810126582278481, "grad_norm": 0.43820018627048524, "learning_rate": 1.1816211835993684e-07, "loss": 0.4964, "step": 962 }, { "epoch": 2.813047711781889, "grad_norm": 0.4420212484091613, "learning_rate": 1.1451245055941251e-07, "loss": 0.5112, "step": 963 }, { "epoch": 2.8159688412852972, "grad_norm": 0.4365334218610012, "learning_rate": 1.1091938111728374e-07, "loss": 0.5273, "step": 964 }, { "epoch": 2.818889970788705, "grad_norm": 0.4377608647741322, "learning_rate": 1.0738295165924783e-07, "loss": 0.5398, "step": 965 }, { "epoch": 2.8218111002921127, "grad_norm": 0.45046127023142907, "learning_rate": 1.0390320315482982e-07, "loss": 0.5062, "step": 966 }, { "epoch": 2.824732229795521, "grad_norm": 0.4019879301034501, "learning_rate": 1.0048017591690607e-07, "loss": 0.5114, "step": 967 }, { "epoch": 2.827653359298929, "grad_norm": 0.4402917529148519, "learning_rate": 9.711390960123634e-08, "loss": 0.5359, "step": 968 }, { "epoch": 2.8305744888023368, "grad_norm": 0.41904135734811376, "learning_rate": 9.380444320600591e-08, "loss": 0.4684, "step": 969 }, { "epoch": 2.833495618305745, "grad_norm": 0.4659811321879393, "learning_rate": 9.055181507137245e-08, "loss": 0.5426, "step": 970 }, { "epoch": 2.836416747809153, "grad_norm": 0.440634928654449, "learning_rate": 8.73560628790232e-08, "loss": 0.5258, "step": 971 }, { "epoch": 2.839337877312561, "grad_norm": 0.44241019235537177, "learning_rate": 8.421722365173856e-08, "loss": 0.553, "step": 972 }, { "epoch": 2.842259006815969, "grad_norm": 0.4085933163681831, "learning_rate": 8.113533375295968e-08, "loss": 0.4804, "step": 973 }, { "epoch": 2.8451801363193767, "grad_norm": 0.42283609192144683, "learning_rate": 7.81104288863721e-08, "loss": 0.5268, "step": 974 }, { "epoch": 2.848101265822785, "grad_norm": 0.4414452447566111, "learning_rate": 7.514254409549005e-08, "loss": 0.5059, "step": 975 }, { "epoch": 2.8510223953261926, "grad_norm": 0.477876908869347, "learning_rate": 7.223171376325e-08, "loss": 0.5401, "step": 976 }, { "epoch": 2.853943524829601, "grad_norm": 0.45667617596842697, "learning_rate": 6.93779716116122e-08, "loss": 0.5179, "step": 977 }, { "epoch": 2.856864654333009, "grad_norm": 0.43661296400815486, "learning_rate": 6.658135070116978e-08, "loss": 0.479, "step": 978 }, { "epoch": 2.8597857838364167, "grad_norm": 0.43447801900111893, "learning_rate": 6.384188343076914e-08, "loss": 0.4933, "step": 979 }, { "epoch": 2.862706913339825, "grad_norm": 0.4567696696725672, "learning_rate": 6.115960153712963e-08, "loss": 0.5225, "step": 980 }, { "epoch": 2.8656280428432326, "grad_norm": 0.4423237005020546, "learning_rate": 5.8534536094478345e-08, "loss": 0.5649, "step": 981 }, { "epoch": 2.8685491723466408, "grad_norm": 0.41379887497949286, "learning_rate": 5.596671751419147e-08, "loss": 0.474, "step": 982 }, { "epoch": 2.8714703018500485, "grad_norm": 0.4382842717392665, "learning_rate": 5.3456175544437936e-08, "loss": 0.529, "step": 983 }, { "epoch": 2.8743914313534566, "grad_norm": 0.4463860151059245, "learning_rate": 5.100293926984023e-08, "loss": 0.4909, "step": 984 }, { "epoch": 2.877312560856865, "grad_norm": 0.4057084453405752, "learning_rate": 4.860703711113246e-08, "loss": 0.5029, "step": 985 }, { "epoch": 2.8802336903602725, "grad_norm": 0.42518495194251527, "learning_rate": 4.626849682483392e-08, "loss": 0.5376, "step": 986 }, { "epoch": 2.8831548198636807, "grad_norm": 0.45413911287155506, "learning_rate": 4.398734550292716e-08, "loss": 0.5467, "step": 987 }, { "epoch": 2.8860759493670884, "grad_norm": 0.4142515305933239, "learning_rate": 4.176360957254266e-08, "loss": 0.4568, "step": 988 }, { "epoch": 2.8889970788704966, "grad_norm": 0.435060369756503, "learning_rate": 3.959731479565465e-08, "loss": 0.5319, "step": 989 }, { "epoch": 2.8919182083739043, "grad_norm": 0.43929179209673974, "learning_rate": 3.748848626878132e-08, "loss": 0.5782, "step": 990 }, { "epoch": 2.8948393378773125, "grad_norm": 0.4061515904911967, "learning_rate": 3.543714842269508e-08, "loss": 0.462, "step": 991 }, { "epoch": 2.8977604673807207, "grad_norm": 0.46420928088916874, "learning_rate": 3.344332502213887e-08, "loss": 0.5218, "step": 992 }, { "epoch": 2.9006815968841284, "grad_norm": 0.4346617171265348, "learning_rate": 3.150703916554976e-08, "loss": 0.4832, "step": 993 }, { "epoch": 2.9036027263875366, "grad_norm": 0.45371971197536476, "learning_rate": 2.962831328479465e-08, "loss": 0.5087, "step": 994 }, { "epoch": 2.9065238558909448, "grad_norm": 0.47344580171350403, "learning_rate": 2.7807169144906108e-08, "loss": 0.5639, "step": 995 }, { "epoch": 2.9094449853943525, "grad_norm": 0.4343079692320858, "learning_rate": 2.6043627843834185e-08, "loss": 0.534, "step": 996 }, { "epoch": 2.91236611489776, "grad_norm": 0.41472366849725406, "learning_rate": 2.4337709812199405e-08, "loss": 0.471, "step": 997 }, { "epoch": 2.9152872444011684, "grad_norm": 0.41847164257771846, "learning_rate": 2.2689434813056856e-08, "loss": 0.5252, "step": 998 }, { "epoch": 2.9182083739045765, "grad_norm": 0.4540400278681399, "learning_rate": 2.1098821941667457e-08, "loss": 0.5683, "step": 999 }, { "epoch": 2.9211295034079843, "grad_norm": 0.43916840754064074, "learning_rate": 1.9565889625275945e-08, "loss": 0.4844, "step": 1000 }, { "epoch": 2.9240506329113924, "grad_norm": 0.4345074027775515, "learning_rate": 1.8090655622899356e-08, "loss": 0.4881, "step": 1001 }, { "epoch": 2.9269717624148006, "grad_norm": 0.4796907378778938, "learning_rate": 1.6673137025118303e-08, "loss": 0.5576, "step": 1002 }, { "epoch": 2.9298928919182083, "grad_norm": 0.4751897325081864, "learning_rate": 1.5313350253882143e-08, "loss": 0.5248, "step": 1003 }, { "epoch": 2.9328140214216165, "grad_norm": 0.4566958532328187, "learning_rate": 1.4011311062316347e-08, "loss": 0.5207, "step": 1004 }, { "epoch": 2.9357351509250242, "grad_norm": 0.4143592301627743, "learning_rate": 1.2767034534540978e-08, "loss": 0.483, "step": 1005 }, { "epoch": 2.9386562804284324, "grad_norm": 0.4151441658060871, "learning_rate": 1.158053508549528e-08, "loss": 0.4624, "step": 1006 }, { "epoch": 2.94157740993184, "grad_norm": 0.44443161265224573, "learning_rate": 1.0451826460772252e-08, "loss": 0.5295, "step": 1007 }, { "epoch": 2.9444985394352483, "grad_norm": 0.4244569182328679, "learning_rate": 9.38092173645766e-09, "loss": 0.5164, "step": 1008 }, { "epoch": 2.9474196689386565, "grad_norm": 0.42619195275616983, "learning_rate": 8.367833318980168e-09, "loss": 0.5474, "step": 1009 }, { "epoch": 2.950340798442064, "grad_norm": 0.44631991928366016, "learning_rate": 7.412572944965335e-09, "loss": 0.4825, "step": 1010 }, { "epoch": 2.9532619279454724, "grad_norm": 0.4028816094942567, "learning_rate": 6.515151681101284e-09, "loss": 0.507, "step": 1011 }, { "epoch": 2.95618305744888, "grad_norm": 0.4346618256973241, "learning_rate": 5.675579924011576e-09, "loss": 0.5173, "step": 1012 }, { "epoch": 2.9591041869522883, "grad_norm": 0.4213345827232603, "learning_rate": 4.89386740013198e-09, "loss": 0.4984, "step": 1013 }, { "epoch": 2.962025316455696, "grad_norm": 0.40444422520787165, "learning_rate": 4.170023165600001e-09, "loss": 0.5463, "step": 1014 }, { "epoch": 2.964946445959104, "grad_norm": 0.4158643063044851, "learning_rate": 3.5040556061483043e-09, "loss": 0.4929, "step": 1015 }, { "epoch": 2.9678675754625123, "grad_norm": 0.45968510533530904, "learning_rate": 2.8959724370092313e-09, "loss": 0.5618, "step": 1016 }, { "epoch": 2.97078870496592, "grad_norm": 0.45874234552575655, "learning_rate": 2.3457807028248734e-09, "loss": 0.5245, "step": 1017 }, { "epoch": 2.9737098344693282, "grad_norm": 0.4715342361612812, "learning_rate": 1.8534867775649169e-09, "loss": 0.5129, "step": 1018 }, { "epoch": 2.976630963972736, "grad_norm": 0.4184866954459166, "learning_rate": 1.4190963644522549e-09, "loss": 0.4844, "step": 1019 }, { "epoch": 2.979552093476144, "grad_norm": 0.41399082545193333, "learning_rate": 1.0426144958985974e-09, "loss": 0.5588, "step": 1020 }, { "epoch": 2.982473222979552, "grad_norm": 0.445097303921185, "learning_rate": 7.240455334456276e-10, "loss": 0.4614, "step": 1021 }, { "epoch": 2.98539435248296, "grad_norm": 0.4214572233508389, "learning_rate": 4.633931677139325e-10, "loss": 0.5183, "step": 1022 }, { "epoch": 2.988315481986368, "grad_norm": 0.4246075700283406, "learning_rate": 2.60660418360259e-10, "loss": 0.5797, "step": 1023 }, { "epoch": 2.991236611489776, "grad_norm": 0.4332145327952596, "learning_rate": 1.158496340419868e-10, "loss": 0.5132, "step": 1024 }, { "epoch": 2.994157740993184, "grad_norm": 0.4486911892623868, "learning_rate": 2.8962492393258546e-11, "loss": 0.5159, "step": 1025 }, { "epoch": 2.9970788704965923, "grad_norm": 0.4412334672273621, "learning_rate": 0.0, "loss": 0.5245, "step": 1026 }, { "epoch": 2.9970788704965923, "step": 1026, "total_flos": 883131557314560.0, "train_loss": 0.5971221869800523, "train_runtime": 15429.9437, "train_samples_per_second": 6.384, "train_steps_per_second": 0.066 } ], "logging_steps": 1, "max_steps": 1026, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 883131557314560.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }