diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,7224 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.9970788704965923, + "eval_steps": 500, + "global_step": 1026, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0029211295034079843, + "grad_norm": 6.789194746397651, + "learning_rate": 9.70873786407767e-08, + "loss": 1.0797, + "step": 1 + }, + { + "epoch": 0.005842259006815969, + "grad_norm": 6.623744256841628, + "learning_rate": 1.941747572815534e-07, + "loss": 1.0936, + "step": 2 + }, + { + "epoch": 0.008763388510223954, + "grad_norm": 6.814087231706784, + "learning_rate": 2.9126213592233014e-07, + "loss": 1.0915, + "step": 3 + }, + { + "epoch": 0.011684518013631937, + "grad_norm": 6.979864404527602, + "learning_rate": 3.883495145631068e-07, + "loss": 1.113, + "step": 4 + }, + { + "epoch": 0.014605647517039922, + "grad_norm": 7.130506329535252, + "learning_rate": 4.854368932038835e-07, + "loss": 1.1201, + "step": 5 + }, + { + "epoch": 0.017526777020447908, + "grad_norm": 6.457316225305797, + "learning_rate": 5.825242718446603e-07, + "loss": 1.0589, + "step": 6 + }, + { + "epoch": 0.02044790652385589, + "grad_norm": 6.634803833935598, + "learning_rate": 6.79611650485437e-07, + "loss": 1.0948, + "step": 7 + }, + { + "epoch": 0.023369036027263874, + "grad_norm": 6.51924054198785, + "learning_rate": 7.766990291262136e-07, + "loss": 1.1033, + "step": 8 + }, + { + "epoch": 0.02629016553067186, + "grad_norm": 6.172807303107381, + "learning_rate": 8.737864077669904e-07, + "loss": 1.1025, + "step": 9 + }, + { + "epoch": 0.029211295034079845, + "grad_norm": 6.123481829151969, + "learning_rate": 9.70873786407767e-07, + "loss": 1.0593, + "step": 10 + }, + { + "epoch": 0.03213242453748783, + "grad_norm": 5.067066736988861, + "learning_rate": 1.0679611650485437e-06, + "loss": 1.0514, + "step": 11 + }, + { + "epoch": 0.035053554040895815, + "grad_norm": 5.316963931526159, + "learning_rate": 1.1650485436893206e-06, + "loss": 1.0649, + "step": 12 + }, + { + "epoch": 0.0379746835443038, + "grad_norm": 4.8023708760386326, + "learning_rate": 1.2621359223300972e-06, + "loss": 1.0595, + "step": 13 + }, + { + "epoch": 0.04089581304771178, + "grad_norm": 4.307477266760391, + "learning_rate": 1.359223300970874e-06, + "loss": 1.0045, + "step": 14 + }, + { + "epoch": 0.043816942551119765, + "grad_norm": 2.9737946154856254, + "learning_rate": 1.4563106796116506e-06, + "loss": 0.9864, + "step": 15 + }, + { + "epoch": 0.04673807205452775, + "grad_norm": 2.969891758892917, + "learning_rate": 1.5533980582524272e-06, + "loss": 1.0038, + "step": 16 + }, + { + "epoch": 0.04965920155793573, + "grad_norm": 2.6410254985919264, + "learning_rate": 1.650485436893204e-06, + "loss": 0.9688, + "step": 17 + }, + { + "epoch": 0.05258033106134372, + "grad_norm": 2.6767839969522385, + "learning_rate": 1.7475728155339808e-06, + "loss": 0.9777, + "step": 18 + }, + { + "epoch": 0.055501460564751706, + "grad_norm": 2.3845837305117867, + "learning_rate": 1.8446601941747574e-06, + "loss": 0.9803, + "step": 19 + }, + { + "epoch": 0.05842259006815969, + "grad_norm": 2.485668257022799, + "learning_rate": 1.941747572815534e-06, + "loss": 0.9512, + "step": 20 + }, + { + "epoch": 0.06134371957156767, + "grad_norm": 2.9203150216318057, + "learning_rate": 2.0388349514563107e-06, + "loss": 0.93, + "step": 21 + }, + { + "epoch": 0.06426484907497566, + "grad_norm": 2.868722547204036, + "learning_rate": 2.1359223300970874e-06, + "loss": 0.9171, + "step": 22 + }, + { + "epoch": 0.06718597857838364, + "grad_norm": 2.679648431038279, + "learning_rate": 2.2330097087378645e-06, + "loss": 0.9199, + "step": 23 + }, + { + "epoch": 0.07010710808179163, + "grad_norm": 2.459073576199394, + "learning_rate": 2.330097087378641e-06, + "loss": 0.9157, + "step": 24 + }, + { + "epoch": 0.0730282375851996, + "grad_norm": 2.014055908464458, + "learning_rate": 2.427184466019418e-06, + "loss": 0.9164, + "step": 25 + }, + { + "epoch": 0.0759493670886076, + "grad_norm": 1.5955860458303692, + "learning_rate": 2.5242718446601945e-06, + "loss": 0.8909, + "step": 26 + }, + { + "epoch": 0.07887049659201557, + "grad_norm": 1.3790379745407235, + "learning_rate": 2.621359223300971e-06, + "loss": 0.8631, + "step": 27 + }, + { + "epoch": 0.08179162609542356, + "grad_norm": 1.6462939916147095, + "learning_rate": 2.718446601941748e-06, + "loss": 0.8384, + "step": 28 + }, + { + "epoch": 0.08471275559883155, + "grad_norm": 1.7416274692927092, + "learning_rate": 2.8155339805825245e-06, + "loss": 0.8643, + "step": 29 + }, + { + "epoch": 0.08763388510223953, + "grad_norm": 1.4689264445022938, + "learning_rate": 2.912621359223301e-06, + "loss": 0.8401, + "step": 30 + }, + { + "epoch": 0.09055501460564752, + "grad_norm": 1.2970139226424346, + "learning_rate": 3.0097087378640778e-06, + "loss": 0.8423, + "step": 31 + }, + { + "epoch": 0.0934761441090555, + "grad_norm": 1.1674121947058942, + "learning_rate": 3.1067961165048544e-06, + "loss": 0.837, + "step": 32 + }, + { + "epoch": 0.09639727361246349, + "grad_norm": 1.1908632755730892, + "learning_rate": 3.2038834951456315e-06, + "loss": 0.8203, + "step": 33 + }, + { + "epoch": 0.09931840311587146, + "grad_norm": 1.1042928737436872, + "learning_rate": 3.300970873786408e-06, + "loss": 0.799, + "step": 34 + }, + { + "epoch": 0.10223953261927946, + "grad_norm": 1.1603936762022113, + "learning_rate": 3.398058252427185e-06, + "loss": 0.8303, + "step": 35 + }, + { + "epoch": 0.10516066212268745, + "grad_norm": 1.0681312398436076, + "learning_rate": 3.4951456310679615e-06, + "loss": 0.8052, + "step": 36 + }, + { + "epoch": 0.10808179162609542, + "grad_norm": 0.8925175456575719, + "learning_rate": 3.592233009708738e-06, + "loss": 0.8069, + "step": 37 + }, + { + "epoch": 0.11100292112950341, + "grad_norm": 0.8822021824942768, + "learning_rate": 3.689320388349515e-06, + "loss": 0.7812, + "step": 38 + }, + { + "epoch": 0.11392405063291139, + "grad_norm": 0.8935693976115221, + "learning_rate": 3.7864077669902915e-06, + "loss": 0.7787, + "step": 39 + }, + { + "epoch": 0.11684518013631938, + "grad_norm": 0.9243397791705507, + "learning_rate": 3.883495145631068e-06, + "loss": 0.7936, + "step": 40 + }, + { + "epoch": 0.11976630963972736, + "grad_norm": 0.8181188534202998, + "learning_rate": 3.980582524271845e-06, + "loss": 0.7752, + "step": 41 + }, + { + "epoch": 0.12268743914313535, + "grad_norm": 0.723717052367605, + "learning_rate": 4.0776699029126215e-06, + "loss": 0.7544, + "step": 42 + }, + { + "epoch": 0.12560856864654332, + "grad_norm": 0.7268642480625847, + "learning_rate": 4.1747572815533986e-06, + "loss": 0.7871, + "step": 43 + }, + { + "epoch": 0.12852969814995133, + "grad_norm": 0.7154200711509474, + "learning_rate": 4.271844660194175e-06, + "loss": 0.7617, + "step": 44 + }, + { + "epoch": 0.1314508276533593, + "grad_norm": 0.7425189888635864, + "learning_rate": 4.368932038834952e-06, + "loss": 0.7661, + "step": 45 + }, + { + "epoch": 0.13437195715676728, + "grad_norm": 0.6964525874804132, + "learning_rate": 4.466019417475729e-06, + "loss": 0.7578, + "step": 46 + }, + { + "epoch": 0.13729308666017526, + "grad_norm": 0.7077568254698756, + "learning_rate": 4.563106796116505e-06, + "loss": 0.7466, + "step": 47 + }, + { + "epoch": 0.14021421616358326, + "grad_norm": 0.728109048064247, + "learning_rate": 4.660194174757282e-06, + "loss": 0.7536, + "step": 48 + }, + { + "epoch": 0.14313534566699124, + "grad_norm": 0.7073551382013681, + "learning_rate": 4.7572815533980585e-06, + "loss": 0.7386, + "step": 49 + }, + { + "epoch": 0.1460564751703992, + "grad_norm": 0.6810942703152736, + "learning_rate": 4.854368932038836e-06, + "loss": 0.7771, + "step": 50 + }, + { + "epoch": 0.14897760467380722, + "grad_norm": 0.7280450266974076, + "learning_rate": 4.951456310679612e-06, + "loss": 0.739, + "step": 51 + }, + { + "epoch": 0.1518987341772152, + "grad_norm": 0.7021483972343961, + "learning_rate": 5.048543689320389e-06, + "loss": 0.7437, + "step": 52 + }, + { + "epoch": 0.15481986368062317, + "grad_norm": 0.7167957676962433, + "learning_rate": 5.145631067961165e-06, + "loss": 0.764, + "step": 53 + }, + { + "epoch": 0.15774099318403115, + "grad_norm": 0.6140023964252928, + "learning_rate": 5.242718446601942e-06, + "loss": 0.7338, + "step": 54 + }, + { + "epoch": 0.16066212268743915, + "grad_norm": 0.7304557287827438, + "learning_rate": 5.3398058252427185e-06, + "loss": 0.7815, + "step": 55 + }, + { + "epoch": 0.16358325219084713, + "grad_norm": 0.6396185255120146, + "learning_rate": 5.436893203883496e-06, + "loss": 0.7349, + "step": 56 + }, + { + "epoch": 0.1665043816942551, + "grad_norm": 0.6558732255969539, + "learning_rate": 5.533980582524272e-06, + "loss": 0.7212, + "step": 57 + }, + { + "epoch": 0.1694255111976631, + "grad_norm": 0.5940933645304805, + "learning_rate": 5.631067961165049e-06, + "loss": 0.7577, + "step": 58 + }, + { + "epoch": 0.17234664070107109, + "grad_norm": 0.6218585913117457, + "learning_rate": 5.728155339805825e-06, + "loss": 0.7431, + "step": 59 + }, + { + "epoch": 0.17526777020447906, + "grad_norm": 0.6908136674947283, + "learning_rate": 5.825242718446602e-06, + "loss": 0.7368, + "step": 60 + }, + { + "epoch": 0.17818889970788704, + "grad_norm": 0.6522752947974526, + "learning_rate": 5.9223300970873785e-06, + "loss": 0.7126, + "step": 61 + }, + { + "epoch": 0.18111002921129504, + "grad_norm": 0.6780346336214896, + "learning_rate": 6.0194174757281556e-06, + "loss": 0.7367, + "step": 62 + }, + { + "epoch": 0.18403115871470302, + "grad_norm": 0.7722253573433183, + "learning_rate": 6.116504854368932e-06, + "loss": 0.7228, + "step": 63 + }, + { + "epoch": 0.186952288218111, + "grad_norm": 0.6399768773236866, + "learning_rate": 6.213592233009709e-06, + "loss": 0.7244, + "step": 64 + }, + { + "epoch": 0.189873417721519, + "grad_norm": 0.6674128228125541, + "learning_rate": 6.310679611650487e-06, + "loss": 0.7376, + "step": 65 + }, + { + "epoch": 0.19279454722492698, + "grad_norm": 0.6775138710637573, + "learning_rate": 6.407766990291263e-06, + "loss": 0.7286, + "step": 66 + }, + { + "epoch": 0.19571567672833495, + "grad_norm": 0.7443299390748634, + "learning_rate": 6.50485436893204e-06, + "loss": 0.7123, + "step": 67 + }, + { + "epoch": 0.19863680623174293, + "grad_norm": 0.6712217876186718, + "learning_rate": 6.601941747572816e-06, + "loss": 0.7118, + "step": 68 + }, + { + "epoch": 0.20155793573515093, + "grad_norm": 0.6173482204736721, + "learning_rate": 6.6990291262135935e-06, + "loss": 0.7141, + "step": 69 + }, + { + "epoch": 0.2044790652385589, + "grad_norm": 0.7586706797857404, + "learning_rate": 6.79611650485437e-06, + "loss": 0.7182, + "step": 70 + }, + { + "epoch": 0.2074001947419669, + "grad_norm": 0.6711328351671544, + "learning_rate": 6.893203883495147e-06, + "loss": 0.7182, + "step": 71 + }, + { + "epoch": 0.2103213242453749, + "grad_norm": 0.6423529972707454, + "learning_rate": 6.990291262135923e-06, + "loss": 0.7147, + "step": 72 + }, + { + "epoch": 0.21324245374878287, + "grad_norm": 0.7175157009799245, + "learning_rate": 7.0873786407767e-06, + "loss": 0.7078, + "step": 73 + }, + { + "epoch": 0.21616358325219084, + "grad_norm": 0.7156449836663106, + "learning_rate": 7.184466019417476e-06, + "loss": 0.7026, + "step": 74 + }, + { + "epoch": 0.21908471275559882, + "grad_norm": 0.591120785534527, + "learning_rate": 7.2815533980582534e-06, + "loss": 0.6996, + "step": 75 + }, + { + "epoch": 0.22200584225900682, + "grad_norm": 0.6568235675952798, + "learning_rate": 7.37864077669903e-06, + "loss": 0.7098, + "step": 76 + }, + { + "epoch": 0.2249269717624148, + "grad_norm": 0.6969907394816692, + "learning_rate": 7.475728155339807e-06, + "loss": 0.7112, + "step": 77 + }, + { + "epoch": 0.22784810126582278, + "grad_norm": 0.674624972312595, + "learning_rate": 7.572815533980583e-06, + "loss": 0.6905, + "step": 78 + }, + { + "epoch": 0.23076923076923078, + "grad_norm": 0.6004655479528318, + "learning_rate": 7.66990291262136e-06, + "loss": 0.6987, + "step": 79 + }, + { + "epoch": 0.23369036027263876, + "grad_norm": 0.672439140786889, + "learning_rate": 7.766990291262136e-06, + "loss": 0.7059, + "step": 80 + }, + { + "epoch": 0.23661148977604674, + "grad_norm": 0.6379167803971234, + "learning_rate": 7.864077669902913e-06, + "loss": 0.6993, + "step": 81 + }, + { + "epoch": 0.2395326192794547, + "grad_norm": 0.624979376741808, + "learning_rate": 7.96116504854369e-06, + "loss": 0.6961, + "step": 82 + }, + { + "epoch": 0.24245374878286272, + "grad_norm": 0.5967840653189634, + "learning_rate": 8.058252427184466e-06, + "loss": 0.7115, + "step": 83 + }, + { + "epoch": 0.2453748782862707, + "grad_norm": 0.6448351126797235, + "learning_rate": 8.155339805825243e-06, + "loss": 0.7058, + "step": 84 + }, + { + "epoch": 0.24829600778967867, + "grad_norm": 0.5456246974361787, + "learning_rate": 8.25242718446602e-06, + "loss": 0.6971, + "step": 85 + }, + { + "epoch": 0.25121713729308665, + "grad_norm": 0.5981523190285354, + "learning_rate": 8.349514563106797e-06, + "loss": 0.6644, + "step": 86 + }, + { + "epoch": 0.25413826679649465, + "grad_norm": 0.6094281430676193, + "learning_rate": 8.446601941747573e-06, + "loss": 0.6822, + "step": 87 + }, + { + "epoch": 0.25705939629990265, + "grad_norm": 0.645486530934357, + "learning_rate": 8.54368932038835e-06, + "loss": 0.6767, + "step": 88 + }, + { + "epoch": 0.2599805258033106, + "grad_norm": 0.6225273389721123, + "learning_rate": 8.640776699029127e-06, + "loss": 0.7152, + "step": 89 + }, + { + "epoch": 0.2629016553067186, + "grad_norm": 0.6432259026110182, + "learning_rate": 8.737864077669904e-06, + "loss": 0.7159, + "step": 90 + }, + { + "epoch": 0.26582278481012656, + "grad_norm": 0.6162731782600254, + "learning_rate": 8.834951456310681e-06, + "loss": 0.7132, + "step": 91 + }, + { + "epoch": 0.26874391431353456, + "grad_norm": 0.6542304058964258, + "learning_rate": 8.932038834951458e-06, + "loss": 0.6809, + "step": 92 + }, + { + "epoch": 0.27166504381694256, + "grad_norm": 0.6079172030969984, + "learning_rate": 9.029126213592233e-06, + "loss": 0.6824, + "step": 93 + }, + { + "epoch": 0.2745861733203505, + "grad_norm": 0.6519383743853913, + "learning_rate": 9.12621359223301e-06, + "loss": 0.6899, + "step": 94 + }, + { + "epoch": 0.2775073028237585, + "grad_norm": 0.7282902638094394, + "learning_rate": 9.223300970873788e-06, + "loss": 0.6922, + "step": 95 + }, + { + "epoch": 0.2804284323271665, + "grad_norm": 0.5989786182176935, + "learning_rate": 9.320388349514565e-06, + "loss": 0.6734, + "step": 96 + }, + { + "epoch": 0.28334956183057447, + "grad_norm": 0.7160709546948157, + "learning_rate": 9.41747572815534e-06, + "loss": 0.6815, + "step": 97 + }, + { + "epoch": 0.2862706913339825, + "grad_norm": 0.6813096412081009, + "learning_rate": 9.514563106796117e-06, + "loss": 0.6885, + "step": 98 + }, + { + "epoch": 0.2891918208373905, + "grad_norm": 0.6540975722149734, + "learning_rate": 9.611650485436894e-06, + "loss": 0.69, + "step": 99 + }, + { + "epoch": 0.2921129503407984, + "grad_norm": 0.8050571281257926, + "learning_rate": 9.708737864077671e-06, + "loss": 0.691, + "step": 100 + }, + { + "epoch": 0.29503407984420643, + "grad_norm": 0.6433559989032654, + "learning_rate": 9.805825242718447e-06, + "loss": 0.6986, + "step": 101 + }, + { + "epoch": 0.29795520934761444, + "grad_norm": 0.8412736504392738, + "learning_rate": 9.902912621359224e-06, + "loss": 0.6926, + "step": 102 + }, + { + "epoch": 0.3008763388510224, + "grad_norm": 0.7443726086923518, + "learning_rate": 1e-05, + "loss": 0.6906, + "step": 103 + }, + { + "epoch": 0.3037974683544304, + "grad_norm": 0.6597015006842325, + "learning_rate": 9.999971037507608e-06, + "loss": 0.675, + "step": 104 + }, + { + "epoch": 0.30671859785783834, + "grad_norm": 0.5931435595661035, + "learning_rate": 9.99988415036596e-06, + "loss": 0.6802, + "step": 105 + }, + { + "epoch": 0.30963972736124634, + "grad_norm": 0.7214217523040783, + "learning_rate": 9.99973933958164e-06, + "loss": 0.7041, + "step": 106 + }, + { + "epoch": 0.31256085686465435, + "grad_norm": 0.7234513096207073, + "learning_rate": 9.999536606832288e-06, + "loss": 0.6872, + "step": 107 + }, + { + "epoch": 0.3154819863680623, + "grad_norm": 0.7879752038918911, + "learning_rate": 9.999275954466555e-06, + "loss": 0.6873, + "step": 108 + }, + { + "epoch": 0.3184031158714703, + "grad_norm": 0.6655845938433153, + "learning_rate": 9.998957385504103e-06, + "loss": 0.6976, + "step": 109 + }, + { + "epoch": 0.3213242453748783, + "grad_norm": 0.8522730059744493, + "learning_rate": 9.99858090363555e-06, + "loss": 0.6719, + "step": 110 + }, + { + "epoch": 0.32424537487828625, + "grad_norm": 0.6292219731062122, + "learning_rate": 9.998146513222436e-06, + "loss": 0.6993, + "step": 111 + }, + { + "epoch": 0.32716650438169426, + "grad_norm": 0.7181176135878521, + "learning_rate": 9.997654219297176e-06, + "loss": 0.6901, + "step": 112 + }, + { + "epoch": 0.33008763388510226, + "grad_norm": 0.6962928810640735, + "learning_rate": 9.997104027562991e-06, + "loss": 0.6951, + "step": 113 + }, + { + "epoch": 0.3330087633885102, + "grad_norm": 0.6849230790401417, + "learning_rate": 9.996495944393853e-06, + "loss": 0.6828, + "step": 114 + }, + { + "epoch": 0.3359298928919182, + "grad_norm": 0.7135171270031142, + "learning_rate": 9.995829976834402e-06, + "loss": 0.6737, + "step": 115 + }, + { + "epoch": 0.3388510223953262, + "grad_norm": 0.6814570642632325, + "learning_rate": 9.995106132599869e-06, + "loss": 0.6875, + "step": 116 + }, + { + "epoch": 0.34177215189873417, + "grad_norm": 0.8248995841087691, + "learning_rate": 9.99432442007599e-06, + "loss": 0.6873, + "step": 117 + }, + { + "epoch": 0.34469328140214217, + "grad_norm": 0.6919031766051941, + "learning_rate": 9.993484848318899e-06, + "loss": 0.6835, + "step": 118 + }, + { + "epoch": 0.3476144109055501, + "grad_norm": 0.7748697122331325, + "learning_rate": 9.992587427055036e-06, + "loss": 0.6933, + "step": 119 + }, + { + "epoch": 0.3505355404089581, + "grad_norm": 0.6232356121853884, + "learning_rate": 9.99163216668102e-06, + "loss": 0.6634, + "step": 120 + }, + { + "epoch": 0.35345666991236613, + "grad_norm": 0.8152682417581196, + "learning_rate": 9.990619078263543e-06, + "loss": 0.6833, + "step": 121 + }, + { + "epoch": 0.3563777994157741, + "grad_norm": 0.6829212246748637, + "learning_rate": 9.989548173539229e-06, + "loss": 0.6904, + "step": 122 + }, + { + "epoch": 0.3592989289191821, + "grad_norm": 0.7764702428398512, + "learning_rate": 9.988419464914505e-06, + "loss": 0.6911, + "step": 123 + }, + { + "epoch": 0.3622200584225901, + "grad_norm": 0.792332480824063, + "learning_rate": 9.98723296546546e-06, + "loss": 0.6817, + "step": 124 + }, + { + "epoch": 0.36514118792599803, + "grad_norm": 0.6594623726005864, + "learning_rate": 9.985988688937684e-06, + "loss": 0.6873, + "step": 125 + }, + { + "epoch": 0.36806231742940604, + "grad_norm": 0.8269823477396988, + "learning_rate": 9.984686649746119e-06, + "loss": 0.693, + "step": 126 + }, + { + "epoch": 0.37098344693281404, + "grad_norm": 0.6422126697095933, + "learning_rate": 9.983326862974882e-06, + "loss": 0.6576, + "step": 127 + }, + { + "epoch": 0.373904576436222, + "grad_norm": 0.8200568586438982, + "learning_rate": 9.981909344377101e-06, + "loss": 0.6929, + "step": 128 + }, + { + "epoch": 0.37682570593963, + "grad_norm": 0.7400126500706381, + "learning_rate": 9.980434110374725e-06, + "loss": 0.6557, + "step": 129 + }, + { + "epoch": 0.379746835443038, + "grad_norm": 0.713854878991036, + "learning_rate": 9.978901178058333e-06, + "loss": 0.6942, + "step": 130 + }, + { + "epoch": 0.38266796494644595, + "grad_norm": 0.7483541113835968, + "learning_rate": 9.977310565186945e-06, + "loss": 0.6781, + "step": 131 + }, + { + "epoch": 0.38558909444985395, + "grad_norm": 0.7131870241688308, + "learning_rate": 9.975662290187802e-06, + "loss": 0.6564, + "step": 132 + }, + { + "epoch": 0.3885102239532619, + "grad_norm": 0.7880618680074065, + "learning_rate": 9.973956372156166e-06, + "loss": 0.6752, + "step": 133 + }, + { + "epoch": 0.3914313534566699, + "grad_norm": 0.5977156256261835, + "learning_rate": 9.972192830855095e-06, + "loss": 0.6763, + "step": 134 + }, + { + "epoch": 0.3943524829600779, + "grad_norm": 0.76186235107972, + "learning_rate": 9.970371686715205e-06, + "loss": 0.7014, + "step": 135 + }, + { + "epoch": 0.39727361246348586, + "grad_norm": 0.8438134698760479, + "learning_rate": 9.96849296083445e-06, + "loss": 0.6902, + "step": 136 + }, + { + "epoch": 0.40019474196689386, + "grad_norm": 0.6120514167187466, + "learning_rate": 9.966556674977864e-06, + "loss": 0.663, + "step": 137 + }, + { + "epoch": 0.40311587147030187, + "grad_norm": 0.7891889540366791, + "learning_rate": 9.964562851577307e-06, + "loss": 0.6458, + "step": 138 + }, + { + "epoch": 0.4060370009737098, + "grad_norm": 0.8381022052177086, + "learning_rate": 9.962511513731219e-06, + "loss": 0.6728, + "step": 139 + }, + { + "epoch": 0.4089581304771178, + "grad_norm": 0.7489332182994131, + "learning_rate": 9.960402685204347e-06, + "loss": 0.6585, + "step": 140 + }, + { + "epoch": 0.4118792599805258, + "grad_norm": 0.804869829411989, + "learning_rate": 9.958236390427458e-06, + "loss": 0.6784, + "step": 141 + }, + { + "epoch": 0.4148003894839338, + "grad_norm": 0.7474713341428297, + "learning_rate": 9.956012654497073e-06, + "loss": 0.6488, + "step": 142 + }, + { + "epoch": 0.4177215189873418, + "grad_norm": 0.8115239389486957, + "learning_rate": 9.953731503175166e-06, + "loss": 0.6894, + "step": 143 + }, + { + "epoch": 0.4206426484907498, + "grad_norm": 0.781168520393115, + "learning_rate": 9.951392962888868e-06, + "loss": 0.6534, + "step": 144 + }, + { + "epoch": 0.42356377799415773, + "grad_norm": 0.7722795439014734, + "learning_rate": 9.948997060730161e-06, + "loss": 0.6504, + "step": 145 + }, + { + "epoch": 0.42648490749756574, + "grad_norm": 0.8668435801796398, + "learning_rate": 9.946543824455563e-06, + "loss": 0.6507, + "step": 146 + }, + { + "epoch": 0.4294060370009737, + "grad_norm": 0.8391877814865175, + "learning_rate": 9.94403328248581e-06, + "loss": 0.6702, + "step": 147 + }, + { + "epoch": 0.4323271665043817, + "grad_norm": 0.6805575537389376, + "learning_rate": 9.941465463905522e-06, + "loss": 0.6744, + "step": 148 + }, + { + "epoch": 0.4352482960077897, + "grad_norm": 0.7033559759136386, + "learning_rate": 9.938840398462872e-06, + "loss": 0.6732, + "step": 149 + }, + { + "epoch": 0.43816942551119764, + "grad_norm": 0.8071774421297472, + "learning_rate": 9.936158116569231e-06, + "loss": 0.6704, + "step": 150 + }, + { + "epoch": 0.44109055501460565, + "grad_norm": 0.7788309365778397, + "learning_rate": 9.933418649298831e-06, + "loss": 0.6697, + "step": 151 + }, + { + "epoch": 0.44401168451801365, + "grad_norm": 0.633484330960411, + "learning_rate": 9.930622028388388e-06, + "loss": 0.6533, + "step": 152 + }, + { + "epoch": 0.4469328140214216, + "grad_norm": 0.6677671353777382, + "learning_rate": 9.92776828623675e-06, + "loss": 0.6323, + "step": 153 + }, + { + "epoch": 0.4498539435248296, + "grad_norm": 0.6580113405415786, + "learning_rate": 9.924857455904511e-06, + "loss": 0.6569, + "step": 154 + }, + { + "epoch": 0.4527750730282376, + "grad_norm": 0.7153967279296722, + "learning_rate": 9.921889571113629e-06, + "loss": 0.6651, + "step": 155 + }, + { + "epoch": 0.45569620253164556, + "grad_norm": 0.6507727168616313, + "learning_rate": 9.918864666247042e-06, + "loss": 0.6709, + "step": 156 + }, + { + "epoch": 0.45861733203505356, + "grad_norm": 0.6614595067208825, + "learning_rate": 9.915782776348263e-06, + "loss": 0.6558, + "step": 157 + }, + { + "epoch": 0.46153846153846156, + "grad_norm": 0.8137858393206525, + "learning_rate": 9.912643937120978e-06, + "loss": 0.6756, + "step": 158 + }, + { + "epoch": 0.4644595910418695, + "grad_norm": 0.6306212983973334, + "learning_rate": 9.909448184928629e-06, + "loss": 0.6416, + "step": 159 + }, + { + "epoch": 0.4673807205452775, + "grad_norm": 0.6836208681318199, + "learning_rate": 9.906195556793996e-06, + "loss": 0.6662, + "step": 160 + }, + { + "epoch": 0.47030185004868547, + "grad_norm": 0.6141476817991299, + "learning_rate": 9.902886090398764e-06, + "loss": 0.6774, + "step": 161 + }, + { + "epoch": 0.47322297955209347, + "grad_norm": 0.7052661650958192, + "learning_rate": 9.899519824083095e-06, + "loss": 0.6618, + "step": 162 + }, + { + "epoch": 0.4761441090555015, + "grad_norm": 0.5986401947237227, + "learning_rate": 9.896096796845172e-06, + "loss": 0.6738, + "step": 163 + }, + { + "epoch": 0.4790652385589094, + "grad_norm": 0.6916970139226505, + "learning_rate": 9.892617048340754e-06, + "loss": 0.6588, + "step": 164 + }, + { + "epoch": 0.4819863680623174, + "grad_norm": 0.6805676670949773, + "learning_rate": 9.889080618882719e-06, + "loss": 0.6826, + "step": 165 + }, + { + "epoch": 0.48490749756572543, + "grad_norm": 0.6658431348883423, + "learning_rate": 9.88548754944059e-06, + "loss": 0.6702, + "step": 166 + }, + { + "epoch": 0.4878286270691334, + "grad_norm": 0.682259295410329, + "learning_rate": 9.881837881640064e-06, + "loss": 0.6735, + "step": 167 + }, + { + "epoch": 0.4907497565725414, + "grad_norm": 0.5993509164653038, + "learning_rate": 9.878131657762535e-06, + "loss": 0.6545, + "step": 168 + }, + { + "epoch": 0.4936708860759494, + "grad_norm": 0.7534948952289838, + "learning_rate": 9.874368920744594e-06, + "loss": 0.6812, + "step": 169 + }, + { + "epoch": 0.49659201557935734, + "grad_norm": 0.707512873305316, + "learning_rate": 9.870549714177538e-06, + "loss": 0.6513, + "step": 170 + }, + { + "epoch": 0.49951314508276534, + "grad_norm": 0.7200355139629895, + "learning_rate": 9.866674082306861e-06, + "loss": 0.6438, + "step": 171 + }, + { + "epoch": 0.5024342745861733, + "grad_norm": 0.8156295486958224, + "learning_rate": 9.86274207003175e-06, + "loss": 0.6564, + "step": 172 + }, + { + "epoch": 0.5053554040895814, + "grad_norm": 0.6117511012510686, + "learning_rate": 9.858753722904552e-06, + "loss": 0.6827, + "step": 173 + }, + { + "epoch": 0.5082765335929893, + "grad_norm": 0.8219493047877929, + "learning_rate": 9.854709087130261e-06, + "loss": 0.6718, + "step": 174 + }, + { + "epoch": 0.5111976630963972, + "grad_norm": 0.6662487126424073, + "learning_rate": 9.850608209565967e-06, + "loss": 0.6388, + "step": 175 + }, + { + "epoch": 0.5141187925998053, + "grad_norm": 0.7125438710540766, + "learning_rate": 9.84645113772032e-06, + "loss": 0.6589, + "step": 176 + }, + { + "epoch": 0.5170399221032133, + "grad_norm": 0.7487609228616714, + "learning_rate": 9.842237919752994e-06, + "loss": 0.6544, + "step": 177 + }, + { + "epoch": 0.5199610516066212, + "grad_norm": 0.6468420309416394, + "learning_rate": 9.8379686044741e-06, + "loss": 0.6565, + "step": 178 + }, + { + "epoch": 0.5228821811100293, + "grad_norm": 0.7546649500098744, + "learning_rate": 9.833643241343642e-06, + "loss": 0.6647, + "step": 179 + }, + { + "epoch": 0.5258033106134372, + "grad_norm": 0.6258186572488958, + "learning_rate": 9.829261880470941e-06, + "loss": 0.6392, + "step": 180 + }, + { + "epoch": 0.5287244401168452, + "grad_norm": 0.5997447293335689, + "learning_rate": 9.82482457261405e-06, + "loss": 0.6398, + "step": 181 + }, + { + "epoch": 0.5316455696202531, + "grad_norm": 0.6452320212378018, + "learning_rate": 9.820331369179166e-06, + "loss": 0.6611, + "step": 182 + }, + { + "epoch": 0.5345666991236612, + "grad_norm": 0.5580735581912285, + "learning_rate": 9.815782322220036e-06, + "loss": 0.6548, + "step": 183 + }, + { + "epoch": 0.5374878286270691, + "grad_norm": 0.6058080142971995, + "learning_rate": 9.811177484437357e-06, + "loss": 0.6664, + "step": 184 + }, + { + "epoch": 0.5404089581304771, + "grad_norm": 0.7487147947448509, + "learning_rate": 9.806516909178161e-06, + "loss": 0.665, + "step": 185 + }, + { + "epoch": 0.5433300876338851, + "grad_norm": 0.5994532711002538, + "learning_rate": 9.801800650435194e-06, + "loss": 0.6345, + "step": 186 + }, + { + "epoch": 0.5462512171372931, + "grad_norm": 0.6402638834025774, + "learning_rate": 9.797028762846305e-06, + "loss": 0.6689, + "step": 187 + }, + { + "epoch": 0.549172346640701, + "grad_norm": 0.8110120277002857, + "learning_rate": 9.792201301693793e-06, + "loss": 0.6623, + "step": 188 + }, + { + "epoch": 0.5520934761441091, + "grad_norm": 0.6022982451173915, + "learning_rate": 9.787318322903784e-06, + "loss": 0.642, + "step": 189 + }, + { + "epoch": 0.555014605647517, + "grad_norm": 0.8359964918822578, + "learning_rate": 9.78237988304557e-06, + "loss": 0.6828, + "step": 190 + }, + { + "epoch": 0.557935735150925, + "grad_norm": 0.5971974835914099, + "learning_rate": 9.77738603933096e-06, + "loss": 0.6637, + "step": 191 + }, + { + "epoch": 0.560856864654333, + "grad_norm": 0.9007553762322157, + "learning_rate": 9.772336849613624e-06, + "loss": 0.6489, + "step": 192 + }, + { + "epoch": 0.563777994157741, + "grad_norm": 0.6097333173160772, + "learning_rate": 9.767232372388406e-06, + "loss": 0.6195, + "step": 193 + }, + { + "epoch": 0.5666991236611489, + "grad_norm": 0.7430323349181741, + "learning_rate": 9.762072666790658e-06, + "loss": 0.6602, + "step": 194 + }, + { + "epoch": 0.569620253164557, + "grad_norm": 0.698590106062137, + "learning_rate": 9.756857792595555e-06, + "loss": 0.654, + "step": 195 + }, + { + "epoch": 0.572541382667965, + "grad_norm": 0.6098458915248055, + "learning_rate": 9.751587810217398e-06, + "loss": 0.6571, + "step": 196 + }, + { + "epoch": 0.5754625121713729, + "grad_norm": 0.6600018141821303, + "learning_rate": 9.746262780708919e-06, + "loss": 0.6572, + "step": 197 + }, + { + "epoch": 0.578383641674781, + "grad_norm": 0.6601639946678165, + "learning_rate": 9.740882765760567e-06, + "loss": 0.6593, + "step": 198 + }, + { + "epoch": 0.5813047711781889, + "grad_norm": 0.726967683938266, + "learning_rate": 9.735447827699798e-06, + "loss": 0.6573, + "step": 199 + }, + { + "epoch": 0.5842259006815969, + "grad_norm": 0.6076134837821863, + "learning_rate": 9.729958029490353e-06, + "loss": 0.6495, + "step": 200 + }, + { + "epoch": 0.5871470301850049, + "grad_norm": 0.693728259825805, + "learning_rate": 9.72441343473153e-06, + "loss": 0.6384, + "step": 201 + }, + { + "epoch": 0.5900681596884129, + "grad_norm": 0.7278423168034551, + "learning_rate": 9.718814107657441e-06, + "loss": 0.6584, + "step": 202 + }, + { + "epoch": 0.5929892891918208, + "grad_norm": 0.7100671054561837, + "learning_rate": 9.713160113136272e-06, + "loss": 0.6555, + "step": 203 + }, + { + "epoch": 0.5959104186952289, + "grad_norm": 0.6955835624438068, + "learning_rate": 9.707451516669533e-06, + "loss": 0.6581, + "step": 204 + }, + { + "epoch": 0.5988315481986368, + "grad_norm": 0.6862859891275203, + "learning_rate": 9.701688384391296e-06, + "loss": 0.6471, + "step": 205 + }, + { + "epoch": 0.6017526777020448, + "grad_norm": 0.7918106833642026, + "learning_rate": 9.695870783067434e-06, + "loss": 0.6351, + "step": 206 + }, + { + "epoch": 0.6046738072054528, + "grad_norm": 0.762255183423834, + "learning_rate": 9.689998780094839e-06, + "loss": 0.6464, + "step": 207 + }, + { + "epoch": 0.6075949367088608, + "grad_norm": 0.5926349601655899, + "learning_rate": 9.684072443500645e-06, + "loss": 0.6342, + "step": 208 + }, + { + "epoch": 0.6105160662122687, + "grad_norm": 0.863234457455766, + "learning_rate": 9.678091841941446e-06, + "loss": 0.653, + "step": 209 + }, + { + "epoch": 0.6134371957156767, + "grad_norm": 0.7588656251837851, + "learning_rate": 9.672057044702492e-06, + "loss": 0.6379, + "step": 210 + }, + { + "epoch": 0.6163583252190847, + "grad_norm": 0.6108635991637165, + "learning_rate": 9.665968121696892e-06, + "loss": 0.6605, + "step": 211 + }, + { + "epoch": 0.6192794547224927, + "grad_norm": 0.8356230484629192, + "learning_rate": 9.659825143464798e-06, + "loss": 0.6458, + "step": 212 + }, + { + "epoch": 0.6222005842259006, + "grad_norm": 0.6240730332192024, + "learning_rate": 9.653628181172596e-06, + "loss": 0.6506, + "step": 213 + }, + { + "epoch": 0.6251217137293087, + "grad_norm": 0.6689297135107584, + "learning_rate": 9.647377306612075e-06, + "loss": 0.6299, + "step": 214 + }, + { + "epoch": 0.6280428432327166, + "grad_norm": 0.7685374427252067, + "learning_rate": 9.641072592199599e-06, + "loss": 0.6634, + "step": 215 + }, + { + "epoch": 0.6309639727361246, + "grad_norm": 0.6331465350705314, + "learning_rate": 9.634714110975263e-06, + "loss": 0.6705, + "step": 216 + }, + { + "epoch": 0.6338851022395326, + "grad_norm": 0.7142109675799595, + "learning_rate": 9.628301936602053e-06, + "loss": 0.6539, + "step": 217 + }, + { + "epoch": 0.6368062317429406, + "grad_norm": 0.8729837863809322, + "learning_rate": 9.62183614336499e-06, + "loss": 0.6596, + "step": 218 + }, + { + "epoch": 0.6397273612463485, + "grad_norm": 0.5991176634004923, + "learning_rate": 9.61531680617027e-06, + "loss": 0.6656, + "step": 219 + }, + { + "epoch": 0.6426484907497566, + "grad_norm": 0.7145894148388473, + "learning_rate": 9.608744000544392e-06, + "loss": 0.6643, + "step": 220 + }, + { + "epoch": 0.6455696202531646, + "grad_norm": 0.7598912110336243, + "learning_rate": 9.602117802633293e-06, + "loss": 0.6291, + "step": 221 + }, + { + "epoch": 0.6484907497565725, + "grad_norm": 0.6815208811474045, + "learning_rate": 9.595438289201453e-06, + "loss": 0.6472, + "step": 222 + }, + { + "epoch": 0.6514118792599806, + "grad_norm": 0.5760568859368258, + "learning_rate": 9.588705537631014e-06, + "loss": 0.6563, + "step": 223 + }, + { + "epoch": 0.6543330087633885, + "grad_norm": 0.7022056147869815, + "learning_rate": 9.581919625920886e-06, + "loss": 0.6524, + "step": 224 + }, + { + "epoch": 0.6572541382667965, + "grad_norm": 0.67470404537858, + "learning_rate": 9.575080632685832e-06, + "loss": 0.6436, + "step": 225 + }, + { + "epoch": 0.6601752677702045, + "grad_norm": 0.6156654623500659, + "learning_rate": 9.568188637155569e-06, + "loss": 0.6256, + "step": 226 + }, + { + "epoch": 0.6630963972736125, + "grad_norm": 0.6870167927139845, + "learning_rate": 9.561243719173844e-06, + "loss": 0.628, + "step": 227 + }, + { + "epoch": 0.6660175267770204, + "grad_norm": 0.6043043393160271, + "learning_rate": 9.554245959197511e-06, + "loss": 0.6631, + "step": 228 + }, + { + "epoch": 0.6689386562804285, + "grad_norm": 0.6424377779531785, + "learning_rate": 9.5471954382956e-06, + "loss": 0.6455, + "step": 229 + }, + { + "epoch": 0.6718597857838364, + "grad_norm": 0.6747246252989533, + "learning_rate": 9.54009223814837e-06, + "loss": 0.6482, + "step": 230 + }, + { + "epoch": 0.6747809152872444, + "grad_norm": 0.6198749042634925, + "learning_rate": 9.532936441046376e-06, + "loss": 0.6679, + "step": 231 + }, + { + "epoch": 0.6777020447906524, + "grad_norm": 0.6700871145900489, + "learning_rate": 9.525728129889505e-06, + "loss": 0.6704, + "step": 232 + }, + { + "epoch": 0.6806231742940604, + "grad_norm": 0.6486221862584836, + "learning_rate": 9.51846738818602e-06, + "loss": 0.6533, + "step": 233 + }, + { + "epoch": 0.6835443037974683, + "grad_norm": 0.6158430372621209, + "learning_rate": 9.511154300051591e-06, + "loss": 0.6391, + "step": 234 + }, + { + "epoch": 0.6864654333008764, + "grad_norm": 0.6059458334544432, + "learning_rate": 9.503788950208324e-06, + "loss": 0.6326, + "step": 235 + }, + { + "epoch": 0.6893865628042843, + "grad_norm": 0.6630441105155737, + "learning_rate": 9.49637142398377e-06, + "loss": 0.6419, + "step": 236 + }, + { + "epoch": 0.6923076923076923, + "grad_norm": 0.7121010143311686, + "learning_rate": 9.48890180730995e-06, + "loss": 0.6366, + "step": 237 + }, + { + "epoch": 0.6952288218111002, + "grad_norm": 0.6536311654933115, + "learning_rate": 9.481380186722354e-06, + "loss": 0.6475, + "step": 238 + }, + { + "epoch": 0.6981499513145083, + "grad_norm": 0.5870586837084283, + "learning_rate": 9.473806649358929e-06, + "loss": 0.6664, + "step": 239 + }, + { + "epoch": 0.7010710808179162, + "grad_norm": 0.7293444660664181, + "learning_rate": 9.466181282959083e-06, + "loss": 0.6294, + "step": 240 + }, + { + "epoch": 0.7039922103213242, + "grad_norm": 0.6803091119725557, + "learning_rate": 9.458504175862665e-06, + "loss": 0.6543, + "step": 241 + }, + { + "epoch": 0.7069133398247323, + "grad_norm": 0.513109460376802, + "learning_rate": 9.450775417008936e-06, + "loss": 0.6529, + "step": 242 + }, + { + "epoch": 0.7098344693281402, + "grad_norm": 0.6591044352211995, + "learning_rate": 9.442995095935542e-06, + "loss": 0.6485, + "step": 243 + }, + { + "epoch": 0.7127555988315482, + "grad_norm": 0.5639394652214005, + "learning_rate": 9.43516330277748e-06, + "loss": 0.6354, + "step": 244 + }, + { + "epoch": 0.7156767283349562, + "grad_norm": 0.5382276491132706, + "learning_rate": 9.427280128266049e-06, + "loss": 0.6338, + "step": 245 + }, + { + "epoch": 0.7185978578383642, + "grad_norm": 0.5783621915913141, + "learning_rate": 9.419345663727805e-06, + "loss": 0.6541, + "step": 246 + }, + { + "epoch": 0.7215189873417721, + "grad_norm": 0.5457758477722148, + "learning_rate": 9.411360001083496e-06, + "loss": 0.6649, + "step": 247 + }, + { + "epoch": 0.7244401168451802, + "grad_norm": 0.5701118395223765, + "learning_rate": 9.403323232846994e-06, + "loss": 0.6305, + "step": 248 + }, + { + "epoch": 0.7273612463485881, + "grad_norm": 0.5986045250901076, + "learning_rate": 9.395235452124239e-06, + "loss": 0.6315, + "step": 249 + }, + { + "epoch": 0.7302823758519961, + "grad_norm": 0.5915757822980239, + "learning_rate": 9.387096752612144e-06, + "loss": 0.6563, + "step": 250 + }, + { + "epoch": 0.7332035053554041, + "grad_norm": 0.6447044009504002, + "learning_rate": 9.378907228597518e-06, + "loss": 0.6543, + "step": 251 + }, + { + "epoch": 0.7361246348588121, + "grad_norm": 0.6146190085950654, + "learning_rate": 9.370666974955973e-06, + "loss": 0.6474, + "step": 252 + }, + { + "epoch": 0.73904576436222, + "grad_norm": 0.6186340229955254, + "learning_rate": 9.362376087150822e-06, + "loss": 0.6498, + "step": 253 + }, + { + "epoch": 0.7419668938656281, + "grad_norm": 0.6697188050730257, + "learning_rate": 9.354034661231976e-06, + "loss": 0.629, + "step": 254 + }, + { + "epoch": 0.744888023369036, + "grad_norm": 0.7166126450253048, + "learning_rate": 9.345642793834825e-06, + "loss": 0.6476, + "step": 255 + }, + { + "epoch": 0.747809152872444, + "grad_norm": 0.5909733136537622, + "learning_rate": 9.337200582179134e-06, + "loss": 0.6338, + "step": 256 + }, + { + "epoch": 0.750730282375852, + "grad_norm": 0.6690338784895201, + "learning_rate": 9.328708124067893e-06, + "loss": 0.6425, + "step": 257 + }, + { + "epoch": 0.75365141187926, + "grad_norm": 0.6170552061635086, + "learning_rate": 9.320165517886207e-06, + "loss": 0.649, + "step": 258 + }, + { + "epoch": 0.7565725413826679, + "grad_norm": 0.6767600272762853, + "learning_rate": 9.31157286260014e-06, + "loss": 0.6496, + "step": 259 + }, + { + "epoch": 0.759493670886076, + "grad_norm": 0.7264560653599718, + "learning_rate": 9.302930257755579e-06, + "loss": 0.6583, + "step": 260 + }, + { + "epoch": 0.762414800389484, + "grad_norm": 0.5453999254984644, + "learning_rate": 9.294237803477076e-06, + "loss": 0.648, + "step": 261 + }, + { + "epoch": 0.7653359298928919, + "grad_norm": 0.7523663818704205, + "learning_rate": 9.285495600466683e-06, + "loss": 0.6488, + "step": 262 + }, + { + "epoch": 0.7682570593963, + "grad_norm": 0.7485014295621598, + "learning_rate": 9.276703750002801e-06, + "loss": 0.6255, + "step": 263 + }, + { + "epoch": 0.7711781888997079, + "grad_norm": 0.686777506781002, + "learning_rate": 9.267862353938988e-06, + "loss": 0.6534, + "step": 264 + }, + { + "epoch": 0.7740993184031159, + "grad_norm": 0.7655220545156425, + "learning_rate": 9.258971514702789e-06, + "loss": 0.6439, + "step": 265 + }, + { + "epoch": 0.7770204479065238, + "grad_norm": 0.5995690372205543, + "learning_rate": 9.250031335294551e-06, + "loss": 0.6264, + "step": 266 + }, + { + "epoch": 0.7799415774099319, + "grad_norm": 0.6306408667729854, + "learning_rate": 9.241041919286227e-06, + "loss": 0.633, + "step": 267 + }, + { + "epoch": 0.7828627069133398, + "grad_norm": 0.7405970033463782, + "learning_rate": 9.232003370820171e-06, + "loss": 0.6355, + "step": 268 + }, + { + "epoch": 0.7857838364167478, + "grad_norm": 0.5761877077710947, + "learning_rate": 9.222915794607942e-06, + "loss": 0.6431, + "step": 269 + }, + { + "epoch": 0.7887049659201558, + "grad_norm": 0.6417671555725009, + "learning_rate": 9.213779295929082e-06, + "loss": 0.6302, + "step": 270 + }, + { + "epoch": 0.7916260954235638, + "grad_norm": 0.5476966311891922, + "learning_rate": 9.204593980629898e-06, + "loss": 0.6307, + "step": 271 + }, + { + "epoch": 0.7945472249269717, + "grad_norm": 0.6325575856222458, + "learning_rate": 9.195359955122244e-06, + "loss": 0.6316, + "step": 272 + }, + { + "epoch": 0.7974683544303798, + "grad_norm": 0.6487642715668982, + "learning_rate": 9.186077326382275e-06, + "loss": 0.6324, + "step": 273 + }, + { + "epoch": 0.8003894839337877, + "grad_norm": 0.5751549331819923, + "learning_rate": 9.176746201949216e-06, + "loss": 0.6585, + "step": 274 + }, + { + "epoch": 0.8033106134371957, + "grad_norm": 0.6208214920916966, + "learning_rate": 9.167366689924116e-06, + "loss": 0.6517, + "step": 275 + }, + { + "epoch": 0.8062317429406037, + "grad_norm": 0.6579852013112687, + "learning_rate": 9.157938898968594e-06, + "loss": 0.643, + "step": 276 + }, + { + "epoch": 0.8091528724440117, + "grad_norm": 0.5985871969783593, + "learning_rate": 9.14846293830358e-06, + "loss": 0.6386, + "step": 277 + }, + { + "epoch": 0.8120740019474196, + "grad_norm": 0.5776495482063276, + "learning_rate": 9.138938917708047e-06, + "loss": 0.6367, + "step": 278 + }, + { + "epoch": 0.8149951314508277, + "grad_norm": 0.6461097088775256, + "learning_rate": 9.129366947517746e-06, + "loss": 0.6311, + "step": 279 + }, + { + "epoch": 0.8179162609542356, + "grad_norm": 0.6054324701596054, + "learning_rate": 9.119747138623925e-06, + "loss": 0.6365, + "step": 280 + }, + { + "epoch": 0.8208373904576436, + "grad_norm": 0.6046664159813615, + "learning_rate": 9.110079602472035e-06, + "loss": 0.6549, + "step": 281 + }, + { + "epoch": 0.8237585199610516, + "grad_norm": 0.690747949343666, + "learning_rate": 9.100364451060457e-06, + "loss": 0.6477, + "step": 282 + }, + { + "epoch": 0.8266796494644596, + "grad_norm": 0.6352884441285447, + "learning_rate": 9.090601796939192e-06, + "loss": 0.6315, + "step": 283 + }, + { + "epoch": 0.8296007789678675, + "grad_norm": 0.7031617950372325, + "learning_rate": 9.080791753208553e-06, + "loss": 0.6304, + "step": 284 + }, + { + "epoch": 0.8325219084712756, + "grad_norm": 0.6955969158961154, + "learning_rate": 9.070934433517872e-06, + "loss": 0.6371, + "step": 285 + }, + { + "epoch": 0.8354430379746836, + "grad_norm": 0.8508450672946094, + "learning_rate": 9.061029952064165e-06, + "loss": 0.6392, + "step": 286 + }, + { + "epoch": 0.8383641674780915, + "grad_norm": 0.6682143045272909, + "learning_rate": 9.05107842359082e-06, + "loss": 0.6354, + "step": 287 + }, + { + "epoch": 0.8412852969814996, + "grad_norm": 0.730935179569577, + "learning_rate": 9.041079963386263e-06, + "loss": 0.6365, + "step": 288 + }, + { + "epoch": 0.8442064264849075, + "grad_norm": 0.9284592753787645, + "learning_rate": 9.031034687282627e-06, + "loss": 0.6512, + "step": 289 + }, + { + "epoch": 0.8471275559883155, + "grad_norm": 0.5711740420462373, + "learning_rate": 9.020942711654404e-06, + "loss": 0.6253, + "step": 290 + }, + { + "epoch": 0.8500486854917235, + "grad_norm": 0.7521676041243963, + "learning_rate": 9.0108041534171e-06, + "loss": 0.6346, + "step": 291 + }, + { + "epoch": 0.8529698149951315, + "grad_norm": 0.7268819094436544, + "learning_rate": 9.000619130025885e-06, + "loss": 0.6321, + "step": 292 + }, + { + "epoch": 0.8558909444985394, + "grad_norm": 0.6255002973375909, + "learning_rate": 8.99038775947422e-06, + "loss": 0.6448, + "step": 293 + }, + { + "epoch": 0.8588120740019474, + "grad_norm": 0.5654580978730834, + "learning_rate": 8.980110160292503e-06, + "loss": 0.6546, + "step": 294 + }, + { + "epoch": 0.8617332035053554, + "grad_norm": 0.5744283394449378, + "learning_rate": 8.969786451546691e-06, + "loss": 0.6354, + "step": 295 + }, + { + "epoch": 0.8646543330087634, + "grad_norm": 0.5607277367845812, + "learning_rate": 8.959416752836915e-06, + "loss": 0.6315, + "step": 296 + }, + { + "epoch": 0.8675754625121713, + "grad_norm": 0.6883032069319117, + "learning_rate": 8.949001184296107e-06, + "loss": 0.6284, + "step": 297 + }, + { + "epoch": 0.8704965920155794, + "grad_norm": 0.5793320184156081, + "learning_rate": 8.938539866588593e-06, + "loss": 0.6299, + "step": 298 + }, + { + "epoch": 0.8734177215189873, + "grad_norm": 0.6122270930586687, + "learning_rate": 8.928032920908709e-06, + "loss": 0.6463, + "step": 299 + }, + { + "epoch": 0.8763388510223953, + "grad_norm": 0.6718320925279334, + "learning_rate": 8.917480468979387e-06, + "loss": 0.6405, + "step": 300 + }, + { + "epoch": 0.8792599805258033, + "grad_norm": 0.6132181340646334, + "learning_rate": 8.906882633050753e-06, + "loss": 0.6426, + "step": 301 + }, + { + "epoch": 0.8821811100292113, + "grad_norm": 0.6206805066838101, + "learning_rate": 8.896239535898702e-06, + "loss": 0.65, + "step": 302 + }, + { + "epoch": 0.8851022395326192, + "grad_norm": 0.7456156741229341, + "learning_rate": 8.885551300823483e-06, + "loss": 0.6395, + "step": 303 + }, + { + "epoch": 0.8880233690360273, + "grad_norm": 0.6186279743363227, + "learning_rate": 8.874818051648267e-06, + "loss": 0.6236, + "step": 304 + }, + { + "epoch": 0.8909444985394352, + "grad_norm": 0.7192434268898347, + "learning_rate": 8.864039912717713e-06, + "loss": 0.6427, + "step": 305 + }, + { + "epoch": 0.8938656280428432, + "grad_norm": 0.6528244466221247, + "learning_rate": 8.853217008896526e-06, + "loss": 0.6478, + "step": 306 + }, + { + "epoch": 0.8967867575462513, + "grad_norm": 0.6240468261028331, + "learning_rate": 8.842349465568018e-06, + "loss": 0.6354, + "step": 307 + }, + { + "epoch": 0.8997078870496592, + "grad_norm": 0.6086442105860419, + "learning_rate": 8.831437408632639e-06, + "loss": 0.6175, + "step": 308 + }, + { + "epoch": 0.9026290165530672, + "grad_norm": 0.6206978097636743, + "learning_rate": 8.820480964506542e-06, + "loss": 0.6329, + "step": 309 + }, + { + "epoch": 0.9055501460564752, + "grad_norm": 0.67571865721595, + "learning_rate": 8.809480260120096e-06, + "loss": 0.6302, + "step": 310 + }, + { + "epoch": 0.9084712755598832, + "grad_norm": 0.612775453081801, + "learning_rate": 8.798435422916425e-06, + "loss": 0.6248, + "step": 311 + }, + { + "epoch": 0.9113924050632911, + "grad_norm": 0.5939423168299965, + "learning_rate": 8.787346580849939e-06, + "loss": 0.6274, + "step": 312 + }, + { + "epoch": 0.9143135345666992, + "grad_norm": 0.6578489213437513, + "learning_rate": 8.776213862384838e-06, + "loss": 0.6367, + "step": 313 + }, + { + "epoch": 0.9172346640701071, + "grad_norm": 0.7439730364901744, + "learning_rate": 8.76503739649363e-06, + "loss": 0.6584, + "step": 314 + }, + { + "epoch": 0.9201557935735151, + "grad_norm": 0.5758626313580656, + "learning_rate": 8.753817312655642e-06, + "loss": 0.638, + "step": 315 + }, + { + "epoch": 0.9230769230769231, + "grad_norm": 0.7372145224476075, + "learning_rate": 8.742553740855507e-06, + "loss": 0.6391, + "step": 316 + }, + { + "epoch": 0.9259980525803311, + "grad_norm": 0.7004119821194315, + "learning_rate": 8.73124681158167e-06, + "loss": 0.6426, + "step": 317 + }, + { + "epoch": 0.928919182083739, + "grad_norm": 0.5393578063337247, + "learning_rate": 8.719896655824878e-06, + "loss": 0.6326, + "step": 318 + }, + { + "epoch": 0.9318403115871471, + "grad_norm": 0.5698772754745327, + "learning_rate": 8.708503405076646e-06, + "loss": 0.634, + "step": 319 + }, + { + "epoch": 0.934761441090555, + "grad_norm": 0.588256486663727, + "learning_rate": 8.697067191327748e-06, + "loss": 0.6328, + "step": 320 + }, + { + "epoch": 0.937682570593963, + "grad_norm": 0.6194561945270076, + "learning_rate": 8.685588147066688e-06, + "loss": 0.6303, + "step": 321 + }, + { + "epoch": 0.9406037000973709, + "grad_norm": 0.5910869397544244, + "learning_rate": 8.67406640527816e-06, + "loss": 0.6494, + "step": 322 + }, + { + "epoch": 0.943524829600779, + "grad_norm": 0.6055073823945398, + "learning_rate": 8.662502099441505e-06, + "loss": 0.6334, + "step": 323 + }, + { + "epoch": 0.9464459591041869, + "grad_norm": 0.5417594595582104, + "learning_rate": 8.650895363529172e-06, + "loss": 0.6279, + "step": 324 + }, + { + "epoch": 0.9493670886075949, + "grad_norm": 0.5878187442362004, + "learning_rate": 8.639246332005163e-06, + "loss": 0.639, + "step": 325 + }, + { + "epoch": 0.952288218111003, + "grad_norm": 0.5233625621707794, + "learning_rate": 8.627555139823468e-06, + "loss": 0.614, + "step": 326 + }, + { + "epoch": 0.9552093476144109, + "grad_norm": 0.5656952115933153, + "learning_rate": 8.615821922426517e-06, + "loss": 0.6214, + "step": 327 + }, + { + "epoch": 0.9581304771178188, + "grad_norm": 0.5806200502649499, + "learning_rate": 8.604046815743598e-06, + "loss": 0.6424, + "step": 328 + }, + { + "epoch": 0.9610516066212269, + "grad_norm": 0.5561534960958242, + "learning_rate": 8.592229956189283e-06, + "loss": 0.638, + "step": 329 + }, + { + "epoch": 0.9639727361246349, + "grad_norm": 0.5340425065767049, + "learning_rate": 8.580371480661857e-06, + "loss": 0.6238, + "step": 330 + }, + { + "epoch": 0.9668938656280428, + "grad_norm": 0.5501524304300768, + "learning_rate": 8.568471526541721e-06, + "loss": 0.6518, + "step": 331 + }, + { + "epoch": 0.9698149951314509, + "grad_norm": 0.571968183721703, + "learning_rate": 8.556530231689809e-06, + "loss": 0.6588, + "step": 332 + }, + { + "epoch": 0.9727361246348588, + "grad_norm": 0.5870934675467651, + "learning_rate": 8.544547734445983e-06, + "loss": 0.6441, + "step": 333 + }, + { + "epoch": 0.9756572541382668, + "grad_norm": 0.5633142366316923, + "learning_rate": 8.532524173627438e-06, + "loss": 0.6558, + "step": 334 + }, + { + "epoch": 0.9785783836416748, + "grad_norm": 0.6080503864783372, + "learning_rate": 8.520459688527091e-06, + "loss": 0.6239, + "step": 335 + }, + { + "epoch": 0.9814995131450828, + "grad_norm": 0.6010534017830508, + "learning_rate": 8.508354418911966e-06, + "loss": 0.6392, + "step": 336 + }, + { + "epoch": 0.9844206426484907, + "grad_norm": 0.6110941694269748, + "learning_rate": 8.496208505021572e-06, + "loss": 0.6334, + "step": 337 + }, + { + "epoch": 0.9873417721518988, + "grad_norm": 0.5748899254330844, + "learning_rate": 8.484022087566284e-06, + "loss": 0.6213, + "step": 338 + }, + { + "epoch": 0.9902629016553067, + "grad_norm": 0.5910067428214469, + "learning_rate": 8.471795307725713e-06, + "loss": 0.6313, + "step": 339 + }, + { + "epoch": 0.9931840311587147, + "grad_norm": 0.6409845798522262, + "learning_rate": 8.459528307147066e-06, + "loss": 0.6223, + "step": 340 + }, + { + "epoch": 0.9961051606621227, + "grad_norm": 0.5798933129205326, + "learning_rate": 8.447221227943507e-06, + "loss": 0.6375, + "step": 341 + }, + { + "epoch": 0.9990262901655307, + "grad_norm": 0.5323356127721257, + "learning_rate": 8.434874212692513e-06, + "loss": 0.631, + "step": 342 + }, + { + "epoch": 1.0019474196689386, + "grad_norm": 1.1893567012708939, + "learning_rate": 8.422487404434214e-06, + "loss": 1.0195, + "step": 343 + }, + { + "epoch": 1.0048685491723466, + "grad_norm": 0.6455689118062985, + "learning_rate": 8.41006094666975e-06, + "loss": 0.5891, + "step": 344 + }, + { + "epoch": 1.0077896786757545, + "grad_norm": 0.6583690984771934, + "learning_rate": 8.397594983359591e-06, + "loss": 0.5549, + "step": 345 + }, + { + "epoch": 1.0107108081791627, + "grad_norm": 0.6011562108844719, + "learning_rate": 8.385089658921892e-06, + "loss": 0.4942, + "step": 346 + }, + { + "epoch": 1.0136319376825706, + "grad_norm": 0.7194629629820957, + "learning_rate": 8.372545118230793e-06, + "loss": 0.5879, + "step": 347 + }, + { + "epoch": 1.0165530671859786, + "grad_norm": 0.7106922347864785, + "learning_rate": 8.35996150661476e-06, + "loss": 0.6464, + "step": 348 + }, + { + "epoch": 1.0194741966893865, + "grad_norm": 0.7004876649145284, + "learning_rate": 8.347338969854898e-06, + "loss": 0.5635, + "step": 349 + }, + { + "epoch": 1.0223953261927945, + "grad_norm": 0.7176313477690597, + "learning_rate": 8.334677654183254e-06, + "loss": 0.6121, + "step": 350 + }, + { + "epoch": 1.0253164556962024, + "grad_norm": 0.6767419736782746, + "learning_rate": 8.321977706281135e-06, + "loss": 0.5923, + "step": 351 + }, + { + "epoch": 1.0282375851996106, + "grad_norm": 0.6606355795156919, + "learning_rate": 8.309239273277394e-06, + "loss": 0.5375, + "step": 352 + }, + { + "epoch": 1.0311587147030186, + "grad_norm": 0.7777491012749531, + "learning_rate": 8.296462502746743e-06, + "loss": 0.5971, + "step": 353 + }, + { + "epoch": 1.0340798442064265, + "grad_norm": 0.6239185790928177, + "learning_rate": 8.283647542708026e-06, + "loss": 0.6017, + "step": 354 + }, + { + "epoch": 1.0370009737098345, + "grad_norm": 0.659621365452411, + "learning_rate": 8.27079454162252e-06, + "loss": 0.523, + "step": 355 + }, + { + "epoch": 1.0399221032132424, + "grad_norm": 0.7314219237543246, + "learning_rate": 8.2579036483922e-06, + "loss": 0.5992, + "step": 356 + }, + { + "epoch": 1.0428432327166504, + "grad_norm": 0.6317967581739343, + "learning_rate": 8.244975012358028e-06, + "loss": 0.58, + "step": 357 + }, + { + "epoch": 1.0457643622200585, + "grad_norm": 0.709552964482946, + "learning_rate": 8.232008783298211e-06, + "loss": 0.5703, + "step": 358 + }, + { + "epoch": 1.0486854917234665, + "grad_norm": 0.5856088171700015, + "learning_rate": 8.219005111426472e-06, + "loss": 0.5851, + "step": 359 + }, + { + "epoch": 1.0516066212268744, + "grad_norm": 0.72335786152197, + "learning_rate": 8.205964147390313e-06, + "loss": 0.5762, + "step": 360 + }, + { + "epoch": 1.0545277507302824, + "grad_norm": 0.6082315528170907, + "learning_rate": 8.19288604226926e-06, + "loss": 0.6045, + "step": 361 + }, + { + "epoch": 1.0574488802336903, + "grad_norm": 0.6904542946382585, + "learning_rate": 8.179770947573124e-06, + "loss": 0.5649, + "step": 362 + }, + { + "epoch": 1.0603700097370983, + "grad_norm": 0.5889683647522176, + "learning_rate": 8.166619015240236e-06, + "loss": 0.5681, + "step": 363 + }, + { + "epoch": 1.0632911392405062, + "grad_norm": 0.5621253187560026, + "learning_rate": 8.15343039763569e-06, + "loss": 0.6013, + "step": 364 + }, + { + "epoch": 1.0662122687439144, + "grad_norm": 0.5587583825073225, + "learning_rate": 8.140205247549583e-06, + "loss": 0.6026, + "step": 365 + }, + { + "epoch": 1.0691333982473223, + "grad_norm": 0.5053935078299595, + "learning_rate": 8.126943718195239e-06, + "loss": 0.5693, + "step": 366 + }, + { + "epoch": 1.0720545277507303, + "grad_norm": 0.5469581161571481, + "learning_rate": 8.113645963207432e-06, + "loss": 0.6007, + "step": 367 + }, + { + "epoch": 1.0749756572541382, + "grad_norm": 0.5430360387939689, + "learning_rate": 8.100312136640618e-06, + "loss": 0.5754, + "step": 368 + }, + { + "epoch": 1.0778967867575462, + "grad_norm": 0.529838189847217, + "learning_rate": 8.086942392967131e-06, + "loss": 0.5758, + "step": 369 + }, + { + "epoch": 1.0808179162609541, + "grad_norm": 0.5537166527008645, + "learning_rate": 8.073536887075417e-06, + "loss": 0.5596, + "step": 370 + }, + { + "epoch": 1.0837390457643623, + "grad_norm": 0.575683387457894, + "learning_rate": 8.060095774268217e-06, + "loss": 0.6149, + "step": 371 + }, + { + "epoch": 1.0866601752677703, + "grad_norm": 0.5212823944156575, + "learning_rate": 8.046619210260785e-06, + "loss": 0.5376, + "step": 372 + }, + { + "epoch": 1.0895813047711782, + "grad_norm": 0.6132577983715938, + "learning_rate": 8.03310735117907e-06, + "loss": 0.611, + "step": 373 + }, + { + "epoch": 1.0925024342745862, + "grad_norm": 0.5741650450345563, + "learning_rate": 8.019560353557923e-06, + "loss": 0.5428, + "step": 374 + }, + { + "epoch": 1.095423563777994, + "grad_norm": 0.6898994227426227, + "learning_rate": 8.005978374339264e-06, + "loss": 0.6122, + "step": 375 + }, + { + "epoch": 1.098344693281402, + "grad_norm": 0.6195159909025971, + "learning_rate": 7.992361570870289e-06, + "loss": 0.6298, + "step": 376 + }, + { + "epoch": 1.1012658227848102, + "grad_norm": 0.5592176863621418, + "learning_rate": 7.978710100901617e-06, + "loss": 0.527, + "step": 377 + }, + { + "epoch": 1.1041869522882182, + "grad_norm": 0.6077274644298606, + "learning_rate": 7.965024122585491e-06, + "loss": 0.5733, + "step": 378 + }, + { + "epoch": 1.1071080817916261, + "grad_norm": 0.5404386498094536, + "learning_rate": 7.951303794473926e-06, + "loss": 0.5786, + "step": 379 + }, + { + "epoch": 1.110029211295034, + "grad_norm": 0.6285033579389189, + "learning_rate": 7.937549275516882e-06, + "loss": 0.5593, + "step": 380 + }, + { + "epoch": 1.112950340798442, + "grad_norm": 0.6136636789874864, + "learning_rate": 7.92376072506042e-06, + "loss": 0.5887, + "step": 381 + }, + { + "epoch": 1.11587147030185, + "grad_norm": 0.5563073653338333, + "learning_rate": 7.909938302844856e-06, + "loss": 0.637, + "step": 382 + }, + { + "epoch": 1.1187925998052581, + "grad_norm": 0.4984935085580574, + "learning_rate": 7.896082169002903e-06, + "loss": 0.537, + "step": 383 + }, + { + "epoch": 1.121713729308666, + "grad_norm": 0.5688731697204966, + "learning_rate": 7.882192484057837e-06, + "loss": 0.5977, + "step": 384 + }, + { + "epoch": 1.124634858812074, + "grad_norm": 0.5409938929481367, + "learning_rate": 7.868269408921614e-06, + "loss": 0.5477, + "step": 385 + }, + { + "epoch": 1.127555988315482, + "grad_norm": 0.6311630545441865, + "learning_rate": 7.854313104893014e-06, + "loss": 0.5595, + "step": 386 + }, + { + "epoch": 1.13047711781889, + "grad_norm": 0.5517528259351719, + "learning_rate": 7.84032373365578e-06, + "loss": 0.5588, + "step": 387 + }, + { + "epoch": 1.1333982473222979, + "grad_norm": 0.515690277716415, + "learning_rate": 7.826301457276733e-06, + "loss": 0.5767, + "step": 388 + }, + { + "epoch": 1.136319376825706, + "grad_norm": 0.5872560884226455, + "learning_rate": 7.812246438203905e-06, + "loss": 0.5618, + "step": 389 + }, + { + "epoch": 1.139240506329114, + "grad_norm": 0.5449550621943328, + "learning_rate": 7.798158839264645e-06, + "loss": 0.5373, + "step": 390 + }, + { + "epoch": 1.142161635832522, + "grad_norm": 0.6192261787272578, + "learning_rate": 7.784038823663746e-06, + "loss": 0.6117, + "step": 391 + }, + { + "epoch": 1.14508276533593, + "grad_norm": 0.5911797274948096, + "learning_rate": 7.769886554981549e-06, + "loss": 0.5832, + "step": 392 + }, + { + "epoch": 1.1480038948393378, + "grad_norm": 0.5393860544929336, + "learning_rate": 7.755702197172036e-06, + "loss": 0.5511, + "step": 393 + }, + { + "epoch": 1.1509250243427458, + "grad_norm": 0.5686675548722897, + "learning_rate": 7.741485914560958e-06, + "loss": 0.601, + "step": 394 + }, + { + "epoch": 1.1538461538461537, + "grad_norm": 0.5965421717249094, + "learning_rate": 7.7272378718439e-06, + "loss": 0.6142, + "step": 395 + }, + { + "epoch": 1.156767283349562, + "grad_norm": 0.5976378598496599, + "learning_rate": 7.712958234084395e-06, + "loss": 0.532, + "step": 396 + }, + { + "epoch": 1.1596884128529699, + "grad_norm": 0.5866514447746561, + "learning_rate": 7.698647166712003e-06, + "loss": 0.6436, + "step": 397 + }, + { + "epoch": 1.1626095423563778, + "grad_norm": 0.5577115650955226, + "learning_rate": 7.684304835520395e-06, + "loss": 0.5524, + "step": 398 + }, + { + "epoch": 1.1655306718597858, + "grad_norm": 0.6320330036922427, + "learning_rate": 7.669931406665437e-06, + "loss": 0.631, + "step": 399 + }, + { + "epoch": 1.1684518013631937, + "grad_norm": 0.5274101108563934, + "learning_rate": 7.655527046663254e-06, + "loss": 0.5369, + "step": 400 + }, + { + "epoch": 1.1713729308666017, + "grad_norm": 0.5563468540404465, + "learning_rate": 7.641091922388316e-06, + "loss": 0.577, + "step": 401 + }, + { + "epoch": 1.1742940603700098, + "grad_norm": 0.5043411412651241, + "learning_rate": 7.626626201071494e-06, + "loss": 0.5623, + "step": 402 + }, + { + "epoch": 1.1772151898734178, + "grad_norm": 0.5658997847969963, + "learning_rate": 7.612130050298126e-06, + "loss": 0.5613, + "step": 403 + }, + { + "epoch": 1.1801363193768257, + "grad_norm": 0.5617715925867288, + "learning_rate": 7.597603638006071e-06, + "loss": 0.5796, + "step": 404 + }, + { + "epoch": 1.1830574488802337, + "grad_norm": 0.5730674124102592, + "learning_rate": 7.5830471324837765e-06, + "loss": 0.6102, + "step": 405 + }, + { + "epoch": 1.1859785783836416, + "grad_norm": 0.5322280548113558, + "learning_rate": 7.56846070236831e-06, + "loss": 0.5392, + "step": 406 + }, + { + "epoch": 1.1888997078870496, + "grad_norm": 0.5636446588596294, + "learning_rate": 7.55384451664342e-06, + "loss": 0.5805, + "step": 407 + }, + { + "epoch": 1.1918208373904577, + "grad_norm": 0.5515192754192108, + "learning_rate": 7.539198744637577e-06, + "loss": 0.5647, + "step": 408 + }, + { + "epoch": 1.1947419668938657, + "grad_norm": 0.649611846767762, + "learning_rate": 7.524523556022003e-06, + "loss": 0.5804, + "step": 409 + }, + { + "epoch": 1.1976630963972736, + "grad_norm": 0.5400876783820088, + "learning_rate": 7.5098191208087144e-06, + "loss": 0.5295, + "step": 410 + }, + { + "epoch": 1.2005842259006816, + "grad_norm": 0.727016701945529, + "learning_rate": 7.495085609348549e-06, + "loss": 0.6035, + "step": 411 + }, + { + "epoch": 1.2035053554040895, + "grad_norm": 0.6474638029028308, + "learning_rate": 7.4803231923291905e-06, + "loss": 0.5905, + "step": 412 + }, + { + "epoch": 1.2064264849074975, + "grad_norm": 0.5322418094330414, + "learning_rate": 7.465532040773195e-06, + "loss": 0.5696, + "step": 413 + }, + { + "epoch": 1.2093476144109054, + "grad_norm": 0.6033615790880061, + "learning_rate": 7.45071232603601e-06, + "loss": 0.5495, + "step": 414 + }, + { + "epoch": 1.2122687439143136, + "grad_norm": 0.5696185950396394, + "learning_rate": 7.4358642198039835e-06, + "loss": 0.5761, + "step": 415 + }, + { + "epoch": 1.2151898734177216, + "grad_norm": 0.58236484440868, + "learning_rate": 7.420987894092383e-06, + "loss": 0.6225, + "step": 416 + }, + { + "epoch": 1.2181110029211295, + "grad_norm": 0.6023725223738031, + "learning_rate": 7.406083521243396e-06, + "loss": 0.5539, + "step": 417 + }, + { + "epoch": 1.2210321324245375, + "grad_norm": 0.5964833664567649, + "learning_rate": 7.391151273924135e-06, + "loss": 0.5766, + "step": 418 + }, + { + "epoch": 1.2239532619279454, + "grad_norm": 0.5579327351939204, + "learning_rate": 7.376191325124644e-06, + "loss": 0.6037, + "step": 419 + }, + { + "epoch": 1.2268743914313536, + "grad_norm": 0.662691054067833, + "learning_rate": 7.36120384815588e-06, + "loss": 0.5775, + "step": 420 + }, + { + "epoch": 1.2297955209347615, + "grad_norm": 0.5457870053427097, + "learning_rate": 7.34618901664772e-06, + "loss": 0.6032, + "step": 421 + }, + { + "epoch": 1.2327166504381695, + "grad_norm": 0.528437147629629, + "learning_rate": 7.33114700454694e-06, + "loss": 0.5551, + "step": 422 + }, + { + "epoch": 1.2356377799415774, + "grad_norm": 0.6396983335282087, + "learning_rate": 7.316077986115206e-06, + "loss": 0.5546, + "step": 423 + }, + { + "epoch": 1.2385589094449854, + "grad_norm": 0.6059608136412865, + "learning_rate": 7.300982135927051e-06, + "loss": 0.5889, + "step": 424 + }, + { + "epoch": 1.2414800389483933, + "grad_norm": 0.5395833042938699, + "learning_rate": 7.285859628867851e-06, + "loss": 0.5765, + "step": 425 + }, + { + "epoch": 1.2444011684518013, + "grad_norm": 0.5938150264550871, + "learning_rate": 7.270710640131806e-06, + "loss": 0.6058, + "step": 426 + }, + { + "epoch": 1.2473222979552094, + "grad_norm": 0.5435408340953944, + "learning_rate": 7.255535345219905e-06, + "loss": 0.5933, + "step": 427 + }, + { + "epoch": 1.2502434274586174, + "grad_norm": 0.5055763559518435, + "learning_rate": 7.240333919937893e-06, + "loss": 0.5989, + "step": 428 + }, + { + "epoch": 1.2531645569620253, + "grad_norm": 0.5862621574043387, + "learning_rate": 7.2251065403942355e-06, + "loss": 0.5888, + "step": 429 + }, + { + "epoch": 1.2560856864654333, + "grad_norm": 0.5234746081250162, + "learning_rate": 7.209853382998077e-06, + "loss": 0.5537, + "step": 430 + }, + { + "epoch": 1.2590068159688412, + "grad_norm": 0.5736429396541277, + "learning_rate": 7.1945746244572e-06, + "loss": 0.547, + "step": 431 + }, + { + "epoch": 1.2619279454722494, + "grad_norm": 0.5662950178895197, + "learning_rate": 7.179270441775976e-06, + "loss": 0.6502, + "step": 432 + }, + { + "epoch": 1.2648490749756571, + "grad_norm": 0.6193942417064999, + "learning_rate": 7.163941012253317e-06, + "loss": 0.5182, + "step": 433 + }, + { + "epoch": 1.2677702044790653, + "grad_norm": 0.5954433533450075, + "learning_rate": 7.148586513480614e-06, + "loss": 0.5961, + "step": 434 + }, + { + "epoch": 1.2706913339824732, + "grad_norm": 0.5407062847357796, + "learning_rate": 7.133207123339689e-06, + "loss": 0.5475, + "step": 435 + }, + { + "epoch": 1.2736124634858812, + "grad_norm": 0.6133121400132443, + "learning_rate": 7.117803020000733e-06, + "loss": 0.59, + "step": 436 + }, + { + "epoch": 1.2765335929892891, + "grad_norm": 0.5272070801107334, + "learning_rate": 7.102374381920233e-06, + "loss": 0.542, + "step": 437 + }, + { + "epoch": 1.279454722492697, + "grad_norm": 0.6014559091825684, + "learning_rate": 7.086921387838916e-06, + "loss": 0.6136, + "step": 438 + }, + { + "epoch": 1.2823758519961053, + "grad_norm": 0.4925204082847554, + "learning_rate": 7.071444216779669e-06, + "loss": 0.5707, + "step": 439 + }, + { + "epoch": 1.2852969814995132, + "grad_norm": 0.5297523991584766, + "learning_rate": 7.055943048045476e-06, + "loss": 0.5917, + "step": 440 + }, + { + "epoch": 1.2882181110029212, + "grad_norm": 0.5742994575633306, + "learning_rate": 7.040418061217325e-06, + "loss": 0.6161, + "step": 441 + }, + { + "epoch": 1.2911392405063291, + "grad_norm": 0.5144128093922836, + "learning_rate": 7.024869436152144e-06, + "loss": 0.5481, + "step": 442 + }, + { + "epoch": 1.294060370009737, + "grad_norm": 0.5651127309438425, + "learning_rate": 7.009297352980706e-06, + "loss": 0.5789, + "step": 443 + }, + { + "epoch": 1.296981499513145, + "grad_norm": 0.46327530559493646, + "learning_rate": 6.99370199210555e-06, + "loss": 0.5435, + "step": 444 + }, + { + "epoch": 1.299902629016553, + "grad_norm": 0.5306445037454199, + "learning_rate": 6.978083534198878e-06, + "loss": 0.6135, + "step": 445 + }, + { + "epoch": 1.3028237585199611, + "grad_norm": 0.6551321587944974, + "learning_rate": 6.962442160200484e-06, + "loss": 0.6111, + "step": 446 + }, + { + "epoch": 1.305744888023369, + "grad_norm": 0.5887843505365383, + "learning_rate": 6.9467780513156335e-06, + "loss": 0.5428, + "step": 447 + }, + { + "epoch": 1.308666017526777, + "grad_norm": 0.5871163724329924, + "learning_rate": 6.931091389012983e-06, + "loss": 0.5595, + "step": 448 + }, + { + "epoch": 1.311587147030185, + "grad_norm": 0.6876137396335796, + "learning_rate": 6.915382355022465e-06, + "loss": 0.5776, + "step": 449 + }, + { + "epoch": 1.314508276533593, + "grad_norm": 0.5978349360901288, + "learning_rate": 6.899651131333194e-06, + "loss": 0.559, + "step": 450 + }, + { + "epoch": 1.317429406037001, + "grad_norm": 0.5600538501243326, + "learning_rate": 6.8838979001913454e-06, + "loss": 0.5782, + "step": 451 + }, + { + "epoch": 1.3203505355404088, + "grad_norm": 0.6670604315541686, + "learning_rate": 6.868122844098057e-06, + "loss": 0.6669, + "step": 452 + }, + { + "epoch": 1.323271665043817, + "grad_norm": 0.6019460261148103, + "learning_rate": 6.852326145807302e-06, + "loss": 0.6006, + "step": 453 + }, + { + "epoch": 1.326192794547225, + "grad_norm": 0.4932938142869286, + "learning_rate": 6.836507988323785e-06, + "loss": 0.4971, + "step": 454 + }, + { + "epoch": 1.3291139240506329, + "grad_norm": 0.6389612562939274, + "learning_rate": 6.82066855490081e-06, + "loss": 0.5994, + "step": 455 + }, + { + "epoch": 1.3320350535540408, + "grad_norm": 0.569391901245458, + "learning_rate": 6.804808029038168e-06, + "loss": 0.5776, + "step": 456 + }, + { + "epoch": 1.3349561830574488, + "grad_norm": 0.5627619450291675, + "learning_rate": 6.788926594480001e-06, + "loss": 0.5894, + "step": 457 + }, + { + "epoch": 1.337877312560857, + "grad_norm": 0.5525138776130833, + "learning_rate": 6.773024435212678e-06, + "loss": 0.5507, + "step": 458 + }, + { + "epoch": 1.340798442064265, + "grad_norm": 0.6243616187017386, + "learning_rate": 6.75710173546267e-06, + "loss": 0.6052, + "step": 459 + }, + { + "epoch": 1.3437195715676729, + "grad_norm": 0.46296070627628405, + "learning_rate": 6.741158679694403e-06, + "loss": 0.5284, + "step": 460 + }, + { + "epoch": 1.3466407010710808, + "grad_norm": 0.556033009188379, + "learning_rate": 6.7251954526081294e-06, + "loss": 0.6225, + "step": 461 + }, + { + "epoch": 1.3495618305744888, + "grad_norm": 0.5425579804205766, + "learning_rate": 6.709212239137785e-06, + "loss": 0.5919, + "step": 462 + }, + { + "epoch": 1.352482960077897, + "grad_norm": 0.5628452971401441, + "learning_rate": 6.693209224448853e-06, + "loss": 0.6162, + "step": 463 + }, + { + "epoch": 1.3554040895813046, + "grad_norm": 0.5283138648819659, + "learning_rate": 6.677186593936207e-06, + "loss": 0.5022, + "step": 464 + }, + { + "epoch": 1.3583252190847128, + "grad_norm": 0.5688480741205626, + "learning_rate": 6.661144533221974e-06, + "loss": 0.594, + "step": 465 + }, + { + "epoch": 1.3612463485881208, + "grad_norm": 0.6112831896983953, + "learning_rate": 6.645083228153377e-06, + "loss": 0.5803, + "step": 466 + }, + { + "epoch": 1.3641674780915287, + "grad_norm": 0.5385606514191121, + "learning_rate": 6.629002864800589e-06, + "loss": 0.5834, + "step": 467 + }, + { + "epoch": 1.3670886075949367, + "grad_norm": 0.5366278285884805, + "learning_rate": 6.612903629454568e-06, + "loss": 0.5343, + "step": 468 + }, + { + "epoch": 1.3700097370983446, + "grad_norm": 0.6713338764048196, + "learning_rate": 6.5967857086249065e-06, + "loss": 0.5573, + "step": 469 + }, + { + "epoch": 1.3729308666017528, + "grad_norm": 0.5538883029206723, + "learning_rate": 6.58064928903767e-06, + "loss": 0.6164, + "step": 470 + }, + { + "epoch": 1.3758519961051607, + "grad_norm": 0.5005412280561692, + "learning_rate": 6.56449455763323e-06, + "loss": 0.5576, + "step": 471 + }, + { + "epoch": 1.3787731256085687, + "grad_norm": 0.5868063901075268, + "learning_rate": 6.548321701564099e-06, + "loss": 0.5933, + "step": 472 + }, + { + "epoch": 1.3816942551119766, + "grad_norm": 0.5499777044713934, + "learning_rate": 6.5321309081927665e-06, + "loss": 0.562, + "step": 473 + }, + { + "epoch": 1.3846153846153846, + "grad_norm": 0.5209457126764143, + "learning_rate": 6.515922365089524e-06, + "loss": 0.5793, + "step": 474 + }, + { + "epoch": 1.3875365141187925, + "grad_norm": 0.5045261657801648, + "learning_rate": 6.499696260030297e-06, + "loss": 0.594, + "step": 475 + }, + { + "epoch": 1.3904576436222005, + "grad_norm": 0.5243175331043878, + "learning_rate": 6.483452780994459e-06, + "loss": 0.534, + "step": 476 + }, + { + "epoch": 1.3933787731256086, + "grad_norm": 0.5169962119701911, + "learning_rate": 6.467192116162668e-06, + "loss": 0.5403, + "step": 477 + }, + { + "epoch": 1.3962999026290166, + "grad_norm": 0.5140189723701601, + "learning_rate": 6.450914453914674e-06, + "loss": 0.6058, + "step": 478 + }, + { + "epoch": 1.3992210321324245, + "grad_norm": 0.5330866177837568, + "learning_rate": 6.434619982827147e-06, + "loss": 0.5882, + "step": 479 + }, + { + "epoch": 1.4021421616358325, + "grad_norm": 0.5414814038608168, + "learning_rate": 6.418308891671484e-06, + "loss": 0.5954, + "step": 480 + }, + { + "epoch": 1.4050632911392404, + "grad_norm": 0.5376659643418761, + "learning_rate": 6.401981369411626e-06, + "loss": 0.5633, + "step": 481 + }, + { + "epoch": 1.4079844206426486, + "grad_norm": 0.5438938648229659, + "learning_rate": 6.385637605201871e-06, + "loss": 0.5677, + "step": 482 + }, + { + "epoch": 1.4109055501460563, + "grad_norm": 0.533597042074437, + "learning_rate": 6.3692777883846746e-06, + "loss": 0.6217, + "step": 483 + }, + { + "epoch": 1.4138266796494645, + "grad_norm": 0.5530416412659575, + "learning_rate": 6.3529021084884655e-06, + "loss": 0.5135, + "step": 484 + }, + { + "epoch": 1.4167478091528725, + "grad_norm": 0.6426214488830264, + "learning_rate": 6.336510755225447e-06, + "loss": 0.6039, + "step": 485 + }, + { + "epoch": 1.4196689386562804, + "grad_norm": 0.5321542466408685, + "learning_rate": 6.320103918489395e-06, + "loss": 0.5615, + "step": 486 + }, + { + "epoch": 1.4225900681596884, + "grad_norm": 0.5807878117878842, + "learning_rate": 6.303681788353465e-06, + "loss": 0.6417, + "step": 487 + }, + { + "epoch": 1.4255111976630963, + "grad_norm": 0.6006151574662654, + "learning_rate": 6.287244555067984e-06, + "loss": 0.5794, + "step": 488 + }, + { + "epoch": 1.4284323271665045, + "grad_norm": 0.5126167984608686, + "learning_rate": 6.270792409058247e-06, + "loss": 0.5242, + "step": 489 + }, + { + "epoch": 1.4313534566699124, + "grad_norm": 0.509995152360523, + "learning_rate": 6.25432554092232e-06, + "loss": 0.5828, + "step": 490 + }, + { + "epoch": 1.4342745861733204, + "grad_norm": 0.5960222110178056, + "learning_rate": 6.237844141428817e-06, + "loss": 0.568, + "step": 491 + }, + { + "epoch": 1.4371957156767283, + "grad_norm": 0.5173153132658744, + "learning_rate": 6.221348401514703e-06, + "loss": 0.5796, + "step": 492 + }, + { + "epoch": 1.4401168451801363, + "grad_norm": 0.5671800858222442, + "learning_rate": 6.204838512283073e-06, + "loss": 0.5506, + "step": 493 + }, + { + "epoch": 1.4430379746835442, + "grad_norm": 0.6040430910549448, + "learning_rate": 6.188314665000944e-06, + "loss": 0.5444, + "step": 494 + }, + { + "epoch": 1.4459591041869522, + "grad_norm": 0.5391742690410167, + "learning_rate": 6.171777051097037e-06, + "loss": 0.5649, + "step": 495 + }, + { + "epoch": 1.4488802336903603, + "grad_norm": 0.597749185130725, + "learning_rate": 6.155225862159558e-06, + "loss": 0.6408, + "step": 496 + }, + { + "epoch": 1.4518013631937683, + "grad_norm": 0.5229846439398993, + "learning_rate": 6.138661289933981e-06, + "loss": 0.5494, + "step": 497 + }, + { + "epoch": 1.4547224926971762, + "grad_norm": 0.5920796120572965, + "learning_rate": 6.1220835263208256e-06, + "loss": 0.5992, + "step": 498 + }, + { + "epoch": 1.4576436222005842, + "grad_norm": 0.5775415962111137, + "learning_rate": 6.105492763373431e-06, + "loss": 0.5911, + "step": 499 + }, + { + "epoch": 1.4605647517039921, + "grad_norm": 0.5917654314551983, + "learning_rate": 6.088889193295738e-06, + "loss": 0.5643, + "step": 500 + }, + { + "epoch": 1.4634858812074003, + "grad_norm": 0.5199838430286094, + "learning_rate": 6.072273008440052e-06, + "loss": 0.5472, + "step": 501 + }, + { + "epoch": 1.4664070107108083, + "grad_norm": 0.5522718232673819, + "learning_rate": 6.0556444013048265e-06, + "loss": 0.546, + "step": 502 + }, + { + "epoch": 1.4693281402142162, + "grad_norm": 0.525908351198303, + "learning_rate": 6.039003564532423e-06, + "loss": 0.6015, + "step": 503 + }, + { + "epoch": 1.4722492697176242, + "grad_norm": 0.5105926082337128, + "learning_rate": 6.0223506909068875e-06, + "loss": 0.5785, + "step": 504 + }, + { + "epoch": 1.475170399221032, + "grad_norm": 0.5832929477157227, + "learning_rate": 6.005685973351708e-06, + "loss": 0.6027, + "step": 505 + }, + { + "epoch": 1.47809152872444, + "grad_norm": 0.5447366515820563, + "learning_rate": 5.989009604927587e-06, + "loss": 0.5833, + "step": 506 + }, + { + "epoch": 1.481012658227848, + "grad_norm": 0.5245433114840506, + "learning_rate": 5.972321778830202e-06, + "loss": 0.4944, + "step": 507 + }, + { + "epoch": 1.4839337877312562, + "grad_norm": 0.5377622062276557, + "learning_rate": 5.9556226883879685e-06, + "loss": 0.5518, + "step": 508 + }, + { + "epoch": 1.4868549172346641, + "grad_norm": 0.6351508971275112, + "learning_rate": 5.938912527059798e-06, + "loss": 0.5982, + "step": 509 + }, + { + "epoch": 1.489776046738072, + "grad_norm": 0.5264532680548314, + "learning_rate": 5.922191488432857e-06, + "loss": 0.5472, + "step": 510 + }, + { + "epoch": 1.49269717624148, + "grad_norm": 0.5618733075790842, + "learning_rate": 5.90545976622033e-06, + "loss": 0.6387, + "step": 511 + }, + { + "epoch": 1.495618305744888, + "grad_norm": 0.5042404211846953, + "learning_rate": 5.888717554259165e-06, + "loss": 0.5188, + "step": 512 + }, + { + "epoch": 1.4985394352482961, + "grad_norm": 0.5996827758246558, + "learning_rate": 5.871965046507838e-06, + "loss": 0.5845, + "step": 513 + }, + { + "epoch": 1.5014605647517039, + "grad_norm": 0.513160385214818, + "learning_rate": 5.855202437044102e-06, + "loss": 0.5407, + "step": 514 + }, + { + "epoch": 1.504381694255112, + "grad_norm": 0.49930093522201097, + "learning_rate": 5.838429920062734e-06, + "loss": 0.5888, + "step": 515 + }, + { + "epoch": 1.50730282375852, + "grad_norm": 0.5820994183473087, + "learning_rate": 5.8216476898732935e-06, + "loss": 0.6026, + "step": 516 + }, + { + "epoch": 1.510223953261928, + "grad_norm": 0.5965529197255538, + "learning_rate": 5.804855940897866e-06, + "loss": 0.5772, + "step": 517 + }, + { + "epoch": 1.5131450827653359, + "grad_norm": 0.4891063083201174, + "learning_rate": 5.788054867668811e-06, + "loss": 0.5383, + "step": 518 + }, + { + "epoch": 1.5160662122687438, + "grad_norm": 0.549598064227309, + "learning_rate": 5.771244664826512e-06, + "loss": 0.5701, + "step": 519 + }, + { + "epoch": 1.518987341772152, + "grad_norm": 0.5201662235289578, + "learning_rate": 5.754425527117118e-06, + "loss": 0.5221, + "step": 520 + }, + { + "epoch": 1.5219084712755597, + "grad_norm": 0.5727695969794172, + "learning_rate": 5.737597649390288e-06, + "loss": 0.5849, + "step": 521 + }, + { + "epoch": 1.524829600778968, + "grad_norm": 0.5387267148344115, + "learning_rate": 5.720761226596935e-06, + "loss": 0.5899, + "step": 522 + }, + { + "epoch": 1.5277507302823758, + "grad_norm": 0.5196417961644373, + "learning_rate": 5.703916453786965e-06, + "loss": 0.6075, + "step": 523 + }, + { + "epoch": 1.5306718597857838, + "grad_norm": 0.5162526391071218, + "learning_rate": 5.6870635261070186e-06, + "loss": 0.5524, + "step": 524 + }, + { + "epoch": 1.533592989289192, + "grad_norm": 0.5712502081281199, + "learning_rate": 5.670202638798213e-06, + "loss": 0.5639, + "step": 525 + }, + { + "epoch": 1.5365141187925997, + "grad_norm": 0.569067876773126, + "learning_rate": 5.653333987193876e-06, + "loss": 0.5437, + "step": 526 + }, + { + "epoch": 1.5394352482960079, + "grad_norm": 0.5925750045532109, + "learning_rate": 5.636457766717286e-06, + "loss": 0.591, + "step": 527 + }, + { + "epoch": 1.5423563777994158, + "grad_norm": 0.5141531017914309, + "learning_rate": 5.619574172879405e-06, + "loss": 0.5471, + "step": 528 + }, + { + "epoch": 1.5452775073028238, + "grad_norm": 0.6597290642553237, + "learning_rate": 5.6026834012766155e-06, + "loss": 0.5901, + "step": 529 + }, + { + "epoch": 1.5481986368062317, + "grad_norm": 0.5884715259486929, + "learning_rate": 5.585785647588458e-06, + "loss": 0.5878, + "step": 530 + }, + { + "epoch": 1.5511197663096397, + "grad_norm": 0.5255721650946908, + "learning_rate": 5.568881107575353e-06, + "loss": 0.5447, + "step": 531 + }, + { + "epoch": 1.5540408958130478, + "grad_norm": 0.5091097959641512, + "learning_rate": 5.55196997707635e-06, + "loss": 0.6276, + "step": 532 + }, + { + "epoch": 1.5569620253164556, + "grad_norm": 0.5119232421542075, + "learning_rate": 5.53505245200684e-06, + "loss": 0.5658, + "step": 533 + }, + { + "epoch": 1.5598831548198637, + "grad_norm": 0.556141615427262, + "learning_rate": 5.518128728356303e-06, + "loss": 0.5951, + "step": 534 + }, + { + "epoch": 1.5628042843232717, + "grad_norm": 0.4969794250430925, + "learning_rate": 5.501199002186024e-06, + "loss": 0.5457, + "step": 535 + }, + { + "epoch": 1.5657254138266796, + "grad_norm": 0.5377430121732353, + "learning_rate": 5.48426346962683e-06, + "loss": 0.587, + "step": 536 + }, + { + "epoch": 1.5686465433300878, + "grad_norm": 0.4763376786579273, + "learning_rate": 5.467322326876813e-06, + "loss": 0.5352, + "step": 537 + }, + { + "epoch": 1.5715676728334955, + "grad_norm": 0.5630031793193719, + "learning_rate": 5.450375770199063e-06, + "loss": 0.652, + "step": 538 + }, + { + "epoch": 1.5744888023369037, + "grad_norm": 0.505233569384271, + "learning_rate": 5.433423995919383e-06, + "loss": 0.5633, + "step": 539 + }, + { + "epoch": 1.5774099318403116, + "grad_norm": 0.4918114442897718, + "learning_rate": 5.416467200424032e-06, + "loss": 0.5422, + "step": 540 + }, + { + "epoch": 1.5803310613437196, + "grad_norm": 0.5991334895146722, + "learning_rate": 5.399505580157428e-06, + "loss": 0.558, + "step": 541 + }, + { + "epoch": 1.5832521908471275, + "grad_norm": 0.5363474139996335, + "learning_rate": 5.382539331619896e-06, + "loss": 0.5699, + "step": 542 + }, + { + "epoch": 1.5861733203505355, + "grad_norm": 0.5807115654118472, + "learning_rate": 5.365568651365369e-06, + "loss": 0.5688, + "step": 543 + }, + { + "epoch": 1.5890944498539437, + "grad_norm": 0.5904366045058206, + "learning_rate": 5.34859373599913e-06, + "loss": 0.5969, + "step": 544 + }, + { + "epoch": 1.5920155793573514, + "grad_norm": 0.491932507277215, + "learning_rate": 5.33161478217552e-06, + "loss": 0.5542, + "step": 545 + }, + { + "epoch": 1.5949367088607596, + "grad_norm": 0.5011301426277018, + "learning_rate": 5.314631986595669e-06, + "loss": 0.569, + "step": 546 + }, + { + "epoch": 1.5978578383641675, + "grad_norm": 0.6106059864005979, + "learning_rate": 5.297645546005208e-06, + "loss": 0.6051, + "step": 547 + }, + { + "epoch": 1.6007789678675755, + "grad_norm": 0.5958177689979132, + "learning_rate": 5.280655657192003e-06, + "loss": 0.5266, + "step": 548 + }, + { + "epoch": 1.6037000973709834, + "grad_norm": 0.5395470595824067, + "learning_rate": 5.263662516983863e-06, + "loss": 0.5657, + "step": 549 + }, + { + "epoch": 1.6066212268743914, + "grad_norm": 0.6212970783782318, + "learning_rate": 5.246666322246267e-06, + "loss": 0.5619, + "step": 550 + }, + { + "epoch": 1.6095423563777995, + "grad_norm": 0.6613037032143712, + "learning_rate": 5.229667269880078e-06, + "loss": 0.6092, + "step": 551 + }, + { + "epoch": 1.6124634858812072, + "grad_norm": 0.48332526807644743, + "learning_rate": 5.212665556819264e-06, + "loss": 0.5401, + "step": 552 + }, + { + "epoch": 1.6153846153846154, + "grad_norm": 0.5325857980417065, + "learning_rate": 5.195661380028625e-06, + "loss": 0.6035, + "step": 553 + }, + { + "epoch": 1.6183057448880234, + "grad_norm": 0.5442785714221408, + "learning_rate": 5.178654936501492e-06, + "loss": 0.5494, + "step": 554 + }, + { + "epoch": 1.6212268743914313, + "grad_norm": 0.6437807568653632, + "learning_rate": 5.1616464232574635e-06, + "loss": 0.6254, + "step": 555 + }, + { + "epoch": 1.6241480038948395, + "grad_norm": 0.4492125406814553, + "learning_rate": 5.1446360373401125e-06, + "loss": 0.5301, + "step": 556 + }, + { + "epoch": 1.6270691333982472, + "grad_norm": 0.5622471364004906, + "learning_rate": 5.127623975814709e-06, + "loss": 0.5715, + "step": 557 + }, + { + "epoch": 1.6299902629016554, + "grad_norm": 0.6240851711756873, + "learning_rate": 5.110610435765935e-06, + "loss": 0.5876, + "step": 558 + }, + { + "epoch": 1.6329113924050633, + "grad_norm": 0.4713533254066173, + "learning_rate": 5.093595614295599e-06, + "loss": 0.4901, + "step": 559 + }, + { + "epoch": 1.6358325219084713, + "grad_norm": 0.5796358881475803, + "learning_rate": 5.076579708520355e-06, + "loss": 0.6325, + "step": 560 + }, + { + "epoch": 1.6387536514118792, + "grad_norm": 0.4866837205814519, + "learning_rate": 5.059562915569424e-06, + "loss": 0.5202, + "step": 561 + }, + { + "epoch": 1.6416747809152872, + "grad_norm": 0.5697352103252339, + "learning_rate": 5.0425454325822946e-06, + "loss": 0.6339, + "step": 562 + }, + { + "epoch": 1.6445959104186954, + "grad_norm": 0.5721771538016244, + "learning_rate": 5.0255274567064594e-06, + "loss": 0.5364, + "step": 563 + }, + { + "epoch": 1.647517039922103, + "grad_norm": 0.5605730886016151, + "learning_rate": 5.008509185095114e-06, + "loss": 0.5717, + "step": 564 + }, + { + "epoch": 1.6504381694255112, + "grad_norm": 0.472681806190808, + "learning_rate": 4.991490814904888e-06, + "loss": 0.555, + "step": 565 + }, + { + "epoch": 1.6533592989289192, + "grad_norm": 0.5791761573446532, + "learning_rate": 4.974472543293544e-06, + "loss": 0.6077, + "step": 566 + }, + { + "epoch": 1.6562804284323271, + "grad_norm": 0.5167447517838185, + "learning_rate": 4.957454567417708e-06, + "loss": 0.5805, + "step": 567 + }, + { + "epoch": 1.6592015579357353, + "grad_norm": 0.5561818381815233, + "learning_rate": 4.940437084430579e-06, + "loss": 0.594, + "step": 568 + }, + { + "epoch": 1.662122687439143, + "grad_norm": 0.44095968784824047, + "learning_rate": 4.923420291479646e-06, + "loss": 0.5235, + "step": 569 + }, + { + "epoch": 1.6650438169425512, + "grad_norm": 0.49918372558187907, + "learning_rate": 4.906404385704402e-06, + "loss": 0.6011, + "step": 570 + }, + { + "epoch": 1.667964946445959, + "grad_norm": 0.5067863457643954, + "learning_rate": 4.8893895642340665e-06, + "loss": 0.5415, + "step": 571 + }, + { + "epoch": 1.6708860759493671, + "grad_norm": 0.49737844087035643, + "learning_rate": 4.872376024185291e-06, + "loss": 0.583, + "step": 572 + }, + { + "epoch": 1.673807205452775, + "grad_norm": 0.4659962080708305, + "learning_rate": 4.855363962659889e-06, + "loss": 0.5627, + "step": 573 + }, + { + "epoch": 1.676728334956183, + "grad_norm": 0.4972582768404136, + "learning_rate": 4.838353576742538e-06, + "loss": 0.5911, + "step": 574 + }, + { + "epoch": 1.6796494644595912, + "grad_norm": 0.44592731268572705, + "learning_rate": 4.82134506349851e-06, + "loss": 0.5459, + "step": 575 + }, + { + "epoch": 1.682570593962999, + "grad_norm": 0.506185991053395, + "learning_rate": 4.804338619971377e-06, + "loss": 0.5306, + "step": 576 + }, + { + "epoch": 1.685491723466407, + "grad_norm": 0.5177923049058465, + "learning_rate": 4.787334443180735e-06, + "loss": 0.6209, + "step": 577 + }, + { + "epoch": 1.688412852969815, + "grad_norm": 0.4601309987230691, + "learning_rate": 4.7703327301199244e-06, + "loss": 0.5145, + "step": 578 + }, + { + "epoch": 1.691333982473223, + "grad_norm": 0.4643978865329562, + "learning_rate": 4.753333677753734e-06, + "loss": 0.5788, + "step": 579 + }, + { + "epoch": 1.694255111976631, + "grad_norm": 0.5283219640600681, + "learning_rate": 4.736337483016138e-06, + "loss": 0.6056, + "step": 580 + }, + { + "epoch": 1.6971762414800389, + "grad_norm": 0.5571423915108132, + "learning_rate": 4.719344342808e-06, + "loss": 0.5791, + "step": 581 + }, + { + "epoch": 1.700097370983447, + "grad_norm": 0.49563030247464135, + "learning_rate": 4.702354453994794e-06, + "loss": 0.5373, + "step": 582 + }, + { + "epoch": 1.7030185004868548, + "grad_norm": 0.49902926183967655, + "learning_rate": 4.6853680134043345e-06, + "loss": 0.5823, + "step": 583 + }, + { + "epoch": 1.705939629990263, + "grad_norm": 0.5116289187531037, + "learning_rate": 4.668385217824482e-06, + "loss": 0.5557, + "step": 584 + }, + { + "epoch": 1.7088607594936709, + "grad_norm": 0.5412976872055105, + "learning_rate": 4.651406264000871e-06, + "loss": 0.5535, + "step": 585 + }, + { + "epoch": 1.7117818889970788, + "grad_norm": 0.5297049362432723, + "learning_rate": 4.634431348634632e-06, + "loss": 0.5872, + "step": 586 + }, + { + "epoch": 1.714703018500487, + "grad_norm": 0.46815419228452076, + "learning_rate": 4.6174606683801055e-06, + "loss": 0.5612, + "step": 587 + }, + { + "epoch": 1.7176241480038947, + "grad_norm": 0.5047028920506614, + "learning_rate": 4.600494419842573e-06, + "loss": 0.5457, + "step": 588 + }, + { + "epoch": 1.720545277507303, + "grad_norm": 0.5021848097551853, + "learning_rate": 4.58353279957597e-06, + "loss": 0.5512, + "step": 589 + }, + { + "epoch": 1.7234664070107109, + "grad_norm": 0.5567736723329815, + "learning_rate": 4.5665760040806174e-06, + "loss": 0.577, + "step": 590 + }, + { + "epoch": 1.7263875365141188, + "grad_norm": 0.4982671338092423, + "learning_rate": 4.549624229800938e-06, + "loss": 0.548, + "step": 591 + }, + { + "epoch": 1.7293086660175268, + "grad_norm": 0.49329094074335034, + "learning_rate": 4.532677673123188e-06, + "loss": 0.5563, + "step": 592 + }, + { + "epoch": 1.7322297955209347, + "grad_norm": 0.4890277332079673, + "learning_rate": 4.5157365303731705e-06, + "loss": 0.5747, + "step": 593 + }, + { + "epoch": 1.7351509250243429, + "grad_norm": 0.5600537499346627, + "learning_rate": 4.498800997813976e-06, + "loss": 0.5816, + "step": 594 + }, + { + "epoch": 1.7380720545277506, + "grad_norm": 0.5086959673884651, + "learning_rate": 4.481871271643698e-06, + "loss": 0.5671, + "step": 595 + }, + { + "epoch": 1.7409931840311588, + "grad_norm": 0.5157431177024923, + "learning_rate": 4.464947547993162e-06, + "loss": 0.5731, + "step": 596 + }, + { + "epoch": 1.7439143135345667, + "grad_norm": 0.5137961841228513, + "learning_rate": 4.4480300229236525e-06, + "loss": 0.544, + "step": 597 + }, + { + "epoch": 1.7468354430379747, + "grad_norm": 0.5937636838238008, + "learning_rate": 4.431118892424649e-06, + "loss": 0.6082, + "step": 598 + }, + { + "epoch": 1.7497565725413826, + "grad_norm": 0.5529433304939138, + "learning_rate": 4.414214352411544e-06, + "loss": 0.5569, + "step": 599 + }, + { + "epoch": 1.7526777020447906, + "grad_norm": 0.5159596510951358, + "learning_rate": 4.397316598723385e-06, + "loss": 0.5455, + "step": 600 + }, + { + "epoch": 1.7555988315481987, + "grad_norm": 0.5506758519748508, + "learning_rate": 4.3804258271205965e-06, + "loss": 0.5505, + "step": 601 + }, + { + "epoch": 1.7585199610516065, + "grad_norm": 0.5523849090934733, + "learning_rate": 4.363542233282715e-06, + "loss": 0.5355, + "step": 602 + }, + { + "epoch": 1.7614410905550146, + "grad_norm": 0.6508686455735765, + "learning_rate": 4.346666012806126e-06, + "loss": 0.602, + "step": 603 + }, + { + "epoch": 1.7643622200584226, + "grad_norm": 0.46084404235460724, + "learning_rate": 4.329797361201788e-06, + "loss": 0.5028, + "step": 604 + }, + { + "epoch": 1.7672833495618305, + "grad_norm": 0.5857372990733761, + "learning_rate": 4.312936473892984e-06, + "loss": 0.6468, + "step": 605 + }, + { + "epoch": 1.7702044790652387, + "grad_norm": 0.4935847914878853, + "learning_rate": 4.296083546213037e-06, + "loss": 0.5578, + "step": 606 + }, + { + "epoch": 1.7731256085686464, + "grad_norm": 0.5144278362398599, + "learning_rate": 4.279238773403066e-06, + "loss": 0.5337, + "step": 607 + }, + { + "epoch": 1.7760467380720546, + "grad_norm": 0.46428380216043086, + "learning_rate": 4.2624023506097116e-06, + "loss": 0.6219, + "step": 608 + }, + { + "epoch": 1.7789678675754625, + "grad_norm": 0.4492920362378749, + "learning_rate": 4.245574472882882e-06, + "loss": 0.5061, + "step": 609 + }, + { + "epoch": 1.7818889970788705, + "grad_norm": 0.48814416218826545, + "learning_rate": 4.228755335173488e-06, + "loss": 0.5045, + "step": 610 + }, + { + "epoch": 1.7848101265822784, + "grad_norm": 0.4935317746542341, + "learning_rate": 4.21194513233119e-06, + "loss": 0.6334, + "step": 611 + }, + { + "epoch": 1.7877312560856864, + "grad_norm": 0.5119284354029839, + "learning_rate": 4.1951440591021375e-06, + "loss": 0.5599, + "step": 612 + }, + { + "epoch": 1.7906523855890946, + "grad_norm": 0.4827371863106343, + "learning_rate": 4.17835231012671e-06, + "loss": 0.5692, + "step": 613 + }, + { + "epoch": 1.7935735150925023, + "grad_norm": 0.5212508350905807, + "learning_rate": 4.161570079937268e-06, + "loss": 0.5947, + "step": 614 + }, + { + "epoch": 1.7964946445959105, + "grad_norm": 0.5235515777898554, + "learning_rate": 4.1447975629559e-06, + "loss": 0.5808, + "step": 615 + }, + { + "epoch": 1.7994157740993184, + "grad_norm": 0.5336005328533747, + "learning_rate": 4.128034953492163e-06, + "loss": 0.5358, + "step": 616 + }, + { + "epoch": 1.8023369036027264, + "grad_norm": 0.5597776990519019, + "learning_rate": 4.1112824457408375e-06, + "loss": 0.5884, + "step": 617 + }, + { + "epoch": 1.8052580331061345, + "grad_norm": 0.5036193557727575, + "learning_rate": 4.094540233779672e-06, + "loss": 0.5688, + "step": 618 + }, + { + "epoch": 1.8081791626095423, + "grad_norm": 0.5338396120419266, + "learning_rate": 4.077808511567145e-06, + "loss": 0.5608, + "step": 619 + }, + { + "epoch": 1.8111002921129504, + "grad_norm": 0.4958715717085618, + "learning_rate": 4.061087472940204e-06, + "loss": 0.5543, + "step": 620 + }, + { + "epoch": 1.8140214216163584, + "grad_norm": 0.6029334087991165, + "learning_rate": 4.044377311612033e-06, + "loss": 0.5691, + "step": 621 + }, + { + "epoch": 1.8169425511197663, + "grad_norm": 0.47915777984754265, + "learning_rate": 4.0276782211698e-06, + "loss": 0.5391, + "step": 622 + }, + { + "epoch": 1.8198636806231743, + "grad_norm": 0.5235008297199194, + "learning_rate": 4.010990395072414e-06, + "loss": 0.5736, + "step": 623 + }, + { + "epoch": 1.8227848101265822, + "grad_norm": 0.5590747505915074, + "learning_rate": 3.9943140266482935e-06, + "loss": 0.6075, + "step": 624 + }, + { + "epoch": 1.8257059396299904, + "grad_norm": 0.47598042280541497, + "learning_rate": 3.977649309093113e-06, + "loss": 0.5344, + "step": 625 + }, + { + "epoch": 1.8286270691333981, + "grad_norm": 0.48621841809715133, + "learning_rate": 3.960996435467577e-06, + "loss": 0.6179, + "step": 626 + }, + { + "epoch": 1.8315481986368063, + "grad_norm": 0.5017298117872958, + "learning_rate": 3.944355598695174e-06, + "loss": 0.5443, + "step": 627 + }, + { + "epoch": 1.8344693281402142, + "grad_norm": 0.5380565465468942, + "learning_rate": 3.9277269915599505e-06, + "loss": 0.57, + "step": 628 + }, + { + "epoch": 1.8373904576436222, + "grad_norm": 0.5990209575217125, + "learning_rate": 3.911110806704265e-06, + "loss": 0.5784, + "step": 629 + }, + { + "epoch": 1.8403115871470301, + "grad_norm": 0.49509817541993284, + "learning_rate": 3.89450723662657e-06, + "loss": 0.5628, + "step": 630 + }, + { + "epoch": 1.843232716650438, + "grad_norm": 0.550605497861562, + "learning_rate": 3.877916473679176e-06, + "loss": 0.6147, + "step": 631 + }, + { + "epoch": 1.8461538461538463, + "grad_norm": 0.5524680706044535, + "learning_rate": 3.861338710066021e-06, + "loss": 0.5125, + "step": 632 + }, + { + "epoch": 1.849074975657254, + "grad_norm": 0.5494313577644265, + "learning_rate": 3.8447741378404436e-06, + "loss": 0.5562, + "step": 633 + }, + { + "epoch": 1.8519961051606622, + "grad_norm": 0.5058622045652387, + "learning_rate": 3.828222948902965e-06, + "loss": 0.5967, + "step": 634 + }, + { + "epoch": 1.85491723466407, + "grad_norm": 0.4488395706127018, + "learning_rate": 3.8116853349990574e-06, + "loss": 0.4735, + "step": 635 + }, + { + "epoch": 1.857838364167478, + "grad_norm": 0.5073104983823334, + "learning_rate": 3.7951614877169285e-06, + "loss": 0.6214, + "step": 636 + }, + { + "epoch": 1.8607594936708862, + "grad_norm": 0.5020623526859631, + "learning_rate": 3.7786515984852977e-06, + "loss": 0.5801, + "step": 637 + }, + { + "epoch": 1.863680623174294, + "grad_norm": 0.5066558405528695, + "learning_rate": 3.762155858571184e-06, + "loss": 0.5571, + "step": 638 + }, + { + "epoch": 1.8666017526777021, + "grad_norm": 0.5016476079074594, + "learning_rate": 3.7456744590776807e-06, + "loss": 0.554, + "step": 639 + }, + { + "epoch": 1.86952288218111, + "grad_norm": 0.5522667996432958, + "learning_rate": 3.729207590941753e-06, + "loss": 0.5684, + "step": 640 + }, + { + "epoch": 1.872444011684518, + "grad_norm": 0.5063071640192653, + "learning_rate": 3.712755444932018e-06, + "loss": 0.5536, + "step": 641 + }, + { + "epoch": 1.875365141187926, + "grad_norm": 0.47979895935138983, + "learning_rate": 3.6963182116465358e-06, + "loss": 0.5607, + "step": 642 + }, + { + "epoch": 1.878286270691334, + "grad_norm": 0.5466280214439997, + "learning_rate": 3.6798960815106065e-06, + "loss": 0.5912, + "step": 643 + }, + { + "epoch": 1.881207400194742, + "grad_norm": 0.50137359619968, + "learning_rate": 3.663489244774555e-06, + "loss": 0.6032, + "step": 644 + }, + { + "epoch": 1.8841285296981498, + "grad_norm": 0.5176702766961846, + "learning_rate": 3.647097891511536e-06, + "loss": 0.5441, + "step": 645 + }, + { + "epoch": 1.887049659201558, + "grad_norm": 0.45011045640480307, + "learning_rate": 3.630722211615328e-06, + "loss": 0.5064, + "step": 646 + }, + { + "epoch": 1.889970788704966, + "grad_norm": 0.5394614468332969, + "learning_rate": 3.614362394798131e-06, + "loss": 0.6368, + "step": 647 + }, + { + "epoch": 1.8928919182083739, + "grad_norm": 0.4874425981153431, + "learning_rate": 3.5980186305883746e-06, + "loss": 0.5701, + "step": 648 + }, + { + "epoch": 1.895813047711782, + "grad_norm": 0.507766689537816, + "learning_rate": 3.5816911083285165e-06, + "loss": 0.5798, + "step": 649 + }, + { + "epoch": 1.8987341772151898, + "grad_norm": 0.4863274459783003, + "learning_rate": 3.565380017172854e-06, + "loss": 0.5244, + "step": 650 + }, + { + "epoch": 1.901655306718598, + "grad_norm": 0.5805740988868203, + "learning_rate": 3.5490855460853275e-06, + "loss": 0.6026, + "step": 651 + }, + { + "epoch": 1.904576436222006, + "grad_norm": 0.5151920826178348, + "learning_rate": 3.5328078838373338e-06, + "loss": 0.5578, + "step": 652 + }, + { + "epoch": 1.9074975657254138, + "grad_norm": 0.4988300821200546, + "learning_rate": 3.516547219005542e-06, + "loss": 0.5639, + "step": 653 + }, + { + "epoch": 1.9104186952288218, + "grad_norm": 0.5020574341639038, + "learning_rate": 3.500303739969704e-06, + "loss": 0.5553, + "step": 654 + }, + { + "epoch": 1.9133398247322297, + "grad_norm": 0.47286299047516567, + "learning_rate": 3.4840776349104755e-06, + "loss": 0.5649, + "step": 655 + }, + { + "epoch": 1.916260954235638, + "grad_norm": 0.4681307130388586, + "learning_rate": 3.4678690918072335e-06, + "loss": 0.5682, + "step": 656 + }, + { + "epoch": 1.9191820837390456, + "grad_norm": 0.46768532251073486, + "learning_rate": 3.451678298435902e-06, + "loss": 0.5486, + "step": 657 + }, + { + "epoch": 1.9221032132424538, + "grad_norm": 0.4674675744224569, + "learning_rate": 3.4355054423667712e-06, + "loss": 0.5421, + "step": 658 + }, + { + "epoch": 1.9250243427458618, + "grad_norm": 0.5073113729998869, + "learning_rate": 3.4193507109623323e-06, + "loss": 0.5802, + "step": 659 + }, + { + "epoch": 1.9279454722492697, + "grad_norm": 0.47796072007470125, + "learning_rate": 3.4032142913750956e-06, + "loss": 0.5454, + "step": 660 + }, + { + "epoch": 1.9308666017526777, + "grad_norm": 0.44978612526614165, + "learning_rate": 3.3870963705454353e-06, + "loss": 0.6016, + "step": 661 + }, + { + "epoch": 1.9337877312560856, + "grad_norm": 0.49209754086259494, + "learning_rate": 3.370997135199413e-06, + "loss": 0.5898, + "step": 662 + }, + { + "epoch": 1.9367088607594938, + "grad_norm": 0.5261340174516683, + "learning_rate": 3.3549167718466245e-06, + "loss": 0.5438, + "step": 663 + }, + { + "epoch": 1.9396299902629015, + "grad_norm": 0.5215555650313894, + "learning_rate": 3.3388554667780272e-06, + "loss": 0.59, + "step": 664 + }, + { + "epoch": 1.9425511197663097, + "grad_norm": 0.4360310531256464, + "learning_rate": 3.322813406063794e-06, + "loss": 0.5361, + "step": 665 + }, + { + "epoch": 1.9454722492697176, + "grad_norm": 0.4338971494593997, + "learning_rate": 3.3067907755511473e-06, + "loss": 0.5217, + "step": 666 + }, + { + "epoch": 1.9483933787731256, + "grad_norm": 0.5993978224848855, + "learning_rate": 3.290787760862215e-06, + "loss": 0.6499, + "step": 667 + }, + { + "epoch": 1.9513145082765337, + "grad_norm": 0.4277968919618892, + "learning_rate": 3.274804547391872e-06, + "loss": 0.5059, + "step": 668 + }, + { + "epoch": 1.9542356377799415, + "grad_norm": 0.4833235638298786, + "learning_rate": 3.258841320305598e-06, + "loss": 0.6043, + "step": 669 + }, + { + "epoch": 1.9571567672833496, + "grad_norm": 0.4876241950113585, + "learning_rate": 3.242898264537331e-06, + "loss": 0.5982, + "step": 670 + }, + { + "epoch": 1.9600778967867576, + "grad_norm": 0.4287569825122473, + "learning_rate": 3.226975564787322e-06, + "loss": 0.5291, + "step": 671 + }, + { + "epoch": 1.9629990262901655, + "grad_norm": 0.559198431896861, + "learning_rate": 3.211073405520001e-06, + "loss": 0.5722, + "step": 672 + }, + { + "epoch": 1.9659201557935735, + "grad_norm": 0.5529747384120326, + "learning_rate": 3.195191970961833e-06, + "loss": 0.5766, + "step": 673 + }, + { + "epoch": 1.9688412852969814, + "grad_norm": 0.4620747878088523, + "learning_rate": 3.1793314450991895e-06, + "loss": 0.5911, + "step": 674 + }, + { + "epoch": 1.9717624148003896, + "grad_norm": 0.46006255679284874, + "learning_rate": 3.1634920116762175e-06, + "loss": 0.5622, + "step": 675 + }, + { + "epoch": 1.9746835443037973, + "grad_norm": 0.4259271639046577, + "learning_rate": 3.1476738541926993e-06, + "loss": 0.5287, + "step": 676 + }, + { + "epoch": 1.9776046738072055, + "grad_norm": 0.4729000058273843, + "learning_rate": 3.1318771559019455e-06, + "loss": 0.5863, + "step": 677 + }, + { + "epoch": 1.9805258033106135, + "grad_norm": 0.5100371866983662, + "learning_rate": 3.1161020998086566e-06, + "loss": 0.5687, + "step": 678 + }, + { + "epoch": 1.9834469328140214, + "grad_norm": 0.4379626484222088, + "learning_rate": 3.1003488686668076e-06, + "loss": 0.5155, + "step": 679 + }, + { + "epoch": 1.9863680623174296, + "grad_norm": 0.4844416917506815, + "learning_rate": 3.0846176449775363e-06, + "loss": 0.5782, + "step": 680 + }, + { + "epoch": 1.9892891918208373, + "grad_norm": 0.48815074844900985, + "learning_rate": 3.0689086109870188e-06, + "loss": 0.5709, + "step": 681 + }, + { + "epoch": 1.9922103213242455, + "grad_norm": 0.5064903715857387, + "learning_rate": 3.0532219486843686e-06, + "loss": 0.596, + "step": 682 + }, + { + "epoch": 1.9951314508276532, + "grad_norm": 0.492762116678178, + "learning_rate": 3.0375578397995178e-06, + "loss": 0.5766, + "step": 683 + }, + { + "epoch": 1.9980525803310614, + "grad_norm": 0.44675106308795903, + "learning_rate": 3.021916465801122e-06, + "loss": 0.5206, + "step": 684 + }, + { + "epoch": 2.0009737098344695, + "grad_norm": 0.9800800703433351, + "learning_rate": 3.0062980078944515e-06, + "loss": 0.8991, + "step": 685 + }, + { + "epoch": 2.0038948393378773, + "grad_norm": 0.4864022689636578, + "learning_rate": 2.990702647019294e-06, + "loss": 0.4938, + "step": 686 + }, + { + "epoch": 2.0068159688412854, + "grad_norm": 0.47889455472294706, + "learning_rate": 2.9751305638478555e-06, + "loss": 0.5913, + "step": 687 + }, + { + "epoch": 2.009737098344693, + "grad_norm": 0.41456980849801855, + "learning_rate": 2.9595819387826753e-06, + "loss": 0.4387, + "step": 688 + }, + { + "epoch": 2.0126582278481013, + "grad_norm": 0.5149845371997606, + "learning_rate": 2.9440569519545258e-06, + "loss": 0.5165, + "step": 689 + }, + { + "epoch": 2.015579357351509, + "grad_norm": 0.5752113401087547, + "learning_rate": 2.9285557832203328e-06, + "loss": 0.5373, + "step": 690 + }, + { + "epoch": 2.0185004868549172, + "grad_norm": 0.5062618329379694, + "learning_rate": 2.9130786121610866e-06, + "loss": 0.4947, + "step": 691 + }, + { + "epoch": 2.0214216163583254, + "grad_norm": 0.49684882097652555, + "learning_rate": 2.897625618079769e-06, + "loss": 0.5208, + "step": 692 + }, + { + "epoch": 2.024342745861733, + "grad_norm": 0.53299035788322, + "learning_rate": 2.88219697999927e-06, + "loss": 0.5438, + "step": 693 + }, + { + "epoch": 2.0272638753651413, + "grad_norm": 0.5543675569999611, + "learning_rate": 2.8667928766603115e-06, + "loss": 0.5444, + "step": 694 + }, + { + "epoch": 2.030185004868549, + "grad_norm": 0.4572356814121919, + "learning_rate": 2.851413486519388e-06, + "loss": 0.4634, + "step": 695 + }, + { + "epoch": 2.033106134371957, + "grad_norm": 0.48340574114252344, + "learning_rate": 2.8360589877466848e-06, + "loss": 0.5391, + "step": 696 + }, + { + "epoch": 2.036027263875365, + "grad_norm": 0.5141120553439197, + "learning_rate": 2.8207295582240248e-06, + "loss": 0.4871, + "step": 697 + }, + { + "epoch": 2.038948393378773, + "grad_norm": 0.4913089013833701, + "learning_rate": 2.8054253755428017e-06, + "loss": 0.5574, + "step": 698 + }, + { + "epoch": 2.0418695228821813, + "grad_norm": 0.4601265937011902, + "learning_rate": 2.7901466170019242e-06, + "loss": 0.4977, + "step": 699 + }, + { + "epoch": 2.044790652385589, + "grad_norm": 0.5120604434660413, + "learning_rate": 2.774893459605766e-06, + "loss": 0.56, + "step": 700 + }, + { + "epoch": 2.047711781888997, + "grad_norm": 0.47951033771571855, + "learning_rate": 2.7596660800621076e-06, + "loss": 0.5028, + "step": 701 + }, + { + "epoch": 2.050632911392405, + "grad_norm": 0.48501372243432067, + "learning_rate": 2.744464654780096e-06, + "loss": 0.5174, + "step": 702 + }, + { + "epoch": 2.053554040895813, + "grad_norm": 0.4741865276944199, + "learning_rate": 2.7292893598681934e-06, + "loss": 0.5104, + "step": 703 + }, + { + "epoch": 2.0564751703992212, + "grad_norm": 0.48778437702309935, + "learning_rate": 2.714140371132149e-06, + "loss": 0.5375, + "step": 704 + }, + { + "epoch": 2.059396299902629, + "grad_norm": 0.49018615472755583, + "learning_rate": 2.69901786407295e-06, + "loss": 0.5105, + "step": 705 + }, + { + "epoch": 2.062317429406037, + "grad_norm": 0.48782945068406763, + "learning_rate": 2.6839220138847966e-06, + "loss": 0.4899, + "step": 706 + }, + { + "epoch": 2.065238558909445, + "grad_norm": 0.4732567901538321, + "learning_rate": 2.6688529954530612e-06, + "loss": 0.5388, + "step": 707 + }, + { + "epoch": 2.068159688412853, + "grad_norm": 0.4935373768584108, + "learning_rate": 2.653810983352282e-06, + "loss": 0.5822, + "step": 708 + }, + { + "epoch": 2.0710808179162608, + "grad_norm": 0.44379266935434936, + "learning_rate": 2.6387961518441223e-06, + "loss": 0.4639, + "step": 709 + }, + { + "epoch": 2.074001947419669, + "grad_norm": 0.47555161755659076, + "learning_rate": 2.6238086748753587e-06, + "loss": 0.5153, + "step": 710 + }, + { + "epoch": 2.076923076923077, + "grad_norm": 0.4924498329014108, + "learning_rate": 2.6088487260758643e-06, + "loss": 0.5245, + "step": 711 + }, + { + "epoch": 2.079844206426485, + "grad_norm": 0.4492288702216835, + "learning_rate": 2.593916478756605e-06, + "loss": 0.4816, + "step": 712 + }, + { + "epoch": 2.082765335929893, + "grad_norm": 0.5008247473603998, + "learning_rate": 2.579012105907619e-06, + "loss": 0.5824, + "step": 713 + }, + { + "epoch": 2.0856864654333007, + "grad_norm": 0.479453244836006, + "learning_rate": 2.5641357801960186e-06, + "loss": 0.4738, + "step": 714 + }, + { + "epoch": 2.088607594936709, + "grad_norm": 0.49048895532662706, + "learning_rate": 2.5492876739639912e-06, + "loss": 0.5299, + "step": 715 + }, + { + "epoch": 2.091528724440117, + "grad_norm": 0.5369724976128885, + "learning_rate": 2.534467959226806e-06, + "loss": 0.5162, + "step": 716 + }, + { + "epoch": 2.094449853943525, + "grad_norm": 0.5137260447049424, + "learning_rate": 2.519676807670811e-06, + "loss": 0.5081, + "step": 717 + }, + { + "epoch": 2.097370983446933, + "grad_norm": 0.4930859741487903, + "learning_rate": 2.504914390651453e-06, + "loss": 0.5576, + "step": 718 + }, + { + "epoch": 2.1002921129503407, + "grad_norm": 0.4775125343363975, + "learning_rate": 2.4901808791912864e-06, + "loss": 0.4998, + "step": 719 + }, + { + "epoch": 2.103213242453749, + "grad_norm": 0.5214779418043487, + "learning_rate": 2.475476443977996e-06, + "loss": 0.5696, + "step": 720 + }, + { + "epoch": 2.1061343719571566, + "grad_norm": 0.47977761871037206, + "learning_rate": 2.460801255362425e-06, + "loss": 0.4764, + "step": 721 + }, + { + "epoch": 2.1090555014605648, + "grad_norm": 0.48061705012654404, + "learning_rate": 2.446155483356582e-06, + "loss": 0.5357, + "step": 722 + }, + { + "epoch": 2.111976630963973, + "grad_norm": 0.45629547557990546, + "learning_rate": 2.4315392976316923e-06, + "loss": 0.4915, + "step": 723 + }, + { + "epoch": 2.1148977604673806, + "grad_norm": 0.4780522247938775, + "learning_rate": 2.4169528675162256e-06, + "loss": 0.4838, + "step": 724 + }, + { + "epoch": 2.117818889970789, + "grad_norm": 0.49801842180761746, + "learning_rate": 2.40239636199393e-06, + "loss": 0.537, + "step": 725 + }, + { + "epoch": 2.1207400194741965, + "grad_norm": 0.4846689678826097, + "learning_rate": 2.3878699497018763e-06, + "loss": 0.549, + "step": 726 + }, + { + "epoch": 2.1236611489776047, + "grad_norm": 0.5018761582798917, + "learning_rate": 2.373373798928507e-06, + "loss": 0.5042, + "step": 727 + }, + { + "epoch": 2.1265822784810124, + "grad_norm": 0.5072064832309269, + "learning_rate": 2.358908077611684e-06, + "loss": 0.5795, + "step": 728 + }, + { + "epoch": 2.1295034079844206, + "grad_norm": 0.46610002849697485, + "learning_rate": 2.344472953336747e-06, + "loss": 0.468, + "step": 729 + }, + { + "epoch": 2.132424537487829, + "grad_norm": 0.5002554866584938, + "learning_rate": 2.3300685933345656e-06, + "loss": 0.5022, + "step": 730 + }, + { + "epoch": 2.1353456669912365, + "grad_norm": 0.5324744280306557, + "learning_rate": 2.3156951644796065e-06, + "loss": 0.5877, + "step": 731 + }, + { + "epoch": 2.1382667964946447, + "grad_norm": 0.4804432916120911, + "learning_rate": 2.3013528332879976e-06, + "loss": 0.4814, + "step": 732 + }, + { + "epoch": 2.1411879259980524, + "grad_norm": 0.47091232830887136, + "learning_rate": 2.287041765915606e-06, + "loss": 0.4953, + "step": 733 + }, + { + "epoch": 2.1441090555014606, + "grad_norm": 0.4535727155626379, + "learning_rate": 2.272762128156101e-06, + "loss": 0.5269, + "step": 734 + }, + { + "epoch": 2.1470301850048688, + "grad_norm": 0.47112551281102016, + "learning_rate": 2.2585140854390432e-06, + "loss": 0.5244, + "step": 735 + }, + { + "epoch": 2.1499513145082765, + "grad_norm": 0.4545532988740129, + "learning_rate": 2.2442978028279634e-06, + "loss": 0.5341, + "step": 736 + }, + { + "epoch": 2.1528724440116846, + "grad_norm": 0.5024829916260531, + "learning_rate": 2.2301134450184535e-06, + "loss": 0.5243, + "step": 737 + }, + { + "epoch": 2.1557935735150924, + "grad_norm": 0.457679806572502, + "learning_rate": 2.215961176336255e-06, + "loss": 0.5157, + "step": 738 + }, + { + "epoch": 2.1587147030185005, + "grad_norm": 0.41799952326682965, + "learning_rate": 2.2018411607353572e-06, + "loss": 0.4653, + "step": 739 + }, + { + "epoch": 2.1616358325219083, + "grad_norm": 0.4787265660707973, + "learning_rate": 2.187753561796097e-06, + "loss": 0.5521, + "step": 740 + }, + { + "epoch": 2.1645569620253164, + "grad_norm": 0.48532252970802797, + "learning_rate": 2.1736985427232684e-06, + "loss": 0.5736, + "step": 741 + }, + { + "epoch": 2.1674780915287246, + "grad_norm": 0.46670045472968474, + "learning_rate": 2.159676266344222e-06, + "loss": 0.4856, + "step": 742 + }, + { + "epoch": 2.1703992210321323, + "grad_norm": 0.456907097163635, + "learning_rate": 2.1456868951069875e-06, + "loss": 0.513, + "step": 743 + }, + { + "epoch": 2.1733203505355405, + "grad_norm": 0.43480032818372993, + "learning_rate": 2.1317305910783886e-06, + "loss": 0.5183, + "step": 744 + }, + { + "epoch": 2.1762414800389482, + "grad_norm": 0.4646674951011356, + "learning_rate": 2.117807515942163e-06, + "loss": 0.5169, + "step": 745 + }, + { + "epoch": 2.1791626095423564, + "grad_norm": 0.5402952715425572, + "learning_rate": 2.1039178309970975e-06, + "loss": 0.5256, + "step": 746 + }, + { + "epoch": 2.1820837390457646, + "grad_norm": 0.44604300646980855, + "learning_rate": 2.090061697155147e-06, + "loss": 0.5264, + "step": 747 + }, + { + "epoch": 2.1850048685491723, + "grad_norm": 0.4855629840620991, + "learning_rate": 2.076239274939582e-06, + "loss": 0.497, + "step": 748 + }, + { + "epoch": 2.1879259980525805, + "grad_norm": 0.450371819506377, + "learning_rate": 2.062450724483118e-06, + "loss": 0.492, + "step": 749 + }, + { + "epoch": 2.190847127555988, + "grad_norm": 0.4675791607919664, + "learning_rate": 2.0486962055260744e-06, + "loss": 0.4833, + "step": 750 + }, + { + "epoch": 2.1937682570593964, + "grad_norm": 0.4931924675592549, + "learning_rate": 2.03497587741451e-06, + "loss": 0.5325, + "step": 751 + }, + { + "epoch": 2.196689386562804, + "grad_norm": 0.4728333199503108, + "learning_rate": 2.021289899098384e-06, + "loss": 0.5753, + "step": 752 + }, + { + "epoch": 2.1996105160662123, + "grad_norm": 0.44477975654172386, + "learning_rate": 2.0076384291297134e-06, + "loss": 0.5095, + "step": 753 + }, + { + "epoch": 2.2025316455696204, + "grad_norm": 0.42502487470067785, + "learning_rate": 1.994021625660737e-06, + "loss": 0.5004, + "step": 754 + }, + { + "epoch": 2.205452775073028, + "grad_norm": 0.4682554593875959, + "learning_rate": 1.9804396464420798e-06, + "loss": 0.5377, + "step": 755 + }, + { + "epoch": 2.2083739045764363, + "grad_norm": 0.478543398252641, + "learning_rate": 1.966892648820932e-06, + "loss": 0.5214, + "step": 756 + }, + { + "epoch": 2.211295034079844, + "grad_norm": 0.4638445072960299, + "learning_rate": 1.953380789739216e-06, + "loss": 0.522, + "step": 757 + }, + { + "epoch": 2.2142161635832522, + "grad_norm": 0.44455755591464263, + "learning_rate": 1.939904225731783e-06, + "loss": 0.5279, + "step": 758 + }, + { + "epoch": 2.21713729308666, + "grad_norm": 0.43381712587340415, + "learning_rate": 1.9264631129245836e-06, + "loss": 0.5082, + "step": 759 + }, + { + "epoch": 2.220058422590068, + "grad_norm": 0.46180168699161817, + "learning_rate": 1.9130576070328695e-06, + "loss": 0.5129, + "step": 760 + }, + { + "epoch": 2.2229795520934763, + "grad_norm": 0.4554045010755852, + "learning_rate": 1.8996878633593829e-06, + "loss": 0.499, + "step": 761 + }, + { + "epoch": 2.225900681596884, + "grad_norm": 0.4749182830100526, + "learning_rate": 1.8863540367925676e-06, + "loss": 0.5343, + "step": 762 + }, + { + "epoch": 2.228821811100292, + "grad_norm": 0.45836107720097835, + "learning_rate": 1.873056281804762e-06, + "loss": 0.4869, + "step": 763 + }, + { + "epoch": 2.2317429406037, + "grad_norm": 0.45195906937962993, + "learning_rate": 1.8597947524504178e-06, + "loss": 0.5201, + "step": 764 + }, + { + "epoch": 2.234664070107108, + "grad_norm": 0.45112191767699233, + "learning_rate": 1.8465696023643115e-06, + "loss": 0.5354, + "step": 765 + }, + { + "epoch": 2.2375851996105163, + "grad_norm": 0.49025509449815313, + "learning_rate": 1.8333809847597644e-06, + "loss": 0.558, + "step": 766 + }, + { + "epoch": 2.240506329113924, + "grad_norm": 0.4486301410086504, + "learning_rate": 1.8202290524268761e-06, + "loss": 0.4795, + "step": 767 + }, + { + "epoch": 2.243427458617332, + "grad_norm": 0.45776452388300076, + "learning_rate": 1.8071139577307416e-06, + "loss": 0.5228, + "step": 768 + }, + { + "epoch": 2.24634858812074, + "grad_norm": 0.4414886979982915, + "learning_rate": 1.7940358526096885e-06, + "loss": 0.4903, + "step": 769 + }, + { + "epoch": 2.249269717624148, + "grad_norm": 0.44672029042681155, + "learning_rate": 1.7809948885735295e-06, + "loss": 0.5906, + "step": 770 + }, + { + "epoch": 2.252190847127556, + "grad_norm": 0.40774965094554166, + "learning_rate": 1.7679912167017922e-06, + "loss": 0.4726, + "step": 771 + }, + { + "epoch": 2.255111976630964, + "grad_norm": 0.42904714451405873, + "learning_rate": 1.7550249876419746e-06, + "loss": 0.5263, + "step": 772 + }, + { + "epoch": 2.258033106134372, + "grad_norm": 0.46232426979933267, + "learning_rate": 1.7420963516078016e-06, + "loss": 0.5207, + "step": 773 + }, + { + "epoch": 2.26095423563778, + "grad_norm": 0.43964498189412426, + "learning_rate": 1.7292054583774809e-06, + "loss": 0.5634, + "step": 774 + }, + { + "epoch": 2.263875365141188, + "grad_norm": 0.4614885797674806, + "learning_rate": 1.7163524572919748e-06, + "loss": 0.5167, + "step": 775 + }, + { + "epoch": 2.2667964946445958, + "grad_norm": 0.4326167271312955, + "learning_rate": 1.7035374972532593e-06, + "loss": 0.4734, + "step": 776 + }, + { + "epoch": 2.269717624148004, + "grad_norm": 0.477108582339082, + "learning_rate": 1.6907607267226079e-06, + "loss": 0.5436, + "step": 777 + }, + { + "epoch": 2.272638753651412, + "grad_norm": 0.45201769752670545, + "learning_rate": 1.6780222937188662e-06, + "loss": 0.5207, + "step": 778 + }, + { + "epoch": 2.27555988315482, + "grad_norm": 0.48262208131523315, + "learning_rate": 1.665322345816746e-06, + "loss": 0.5304, + "step": 779 + }, + { + "epoch": 2.278481012658228, + "grad_norm": 0.45623936201495924, + "learning_rate": 1.6526610301451028e-06, + "loss": 0.4604, + "step": 780 + }, + { + "epoch": 2.2814021421616357, + "grad_norm": 0.45941621411519823, + "learning_rate": 1.6400384933852403e-06, + "loss": 0.508, + "step": 781 + }, + { + "epoch": 2.284323271665044, + "grad_norm": 0.4640162246141768, + "learning_rate": 1.6274548817692088e-06, + "loss": 0.5429, + "step": 782 + }, + { + "epoch": 2.2872444011684516, + "grad_norm": 0.4517120340093356, + "learning_rate": 1.6149103410781086e-06, + "loss": 0.5126, + "step": 783 + }, + { + "epoch": 2.29016553067186, + "grad_norm": 0.4528239429323758, + "learning_rate": 1.6024050166404097e-06, + "loss": 0.4995, + "step": 784 + }, + { + "epoch": 2.293086660175268, + "grad_norm": 0.4440949852742247, + "learning_rate": 1.5899390533302538e-06, + "loss": 0.5333, + "step": 785 + }, + { + "epoch": 2.2960077896786757, + "grad_norm": 0.3970036577732869, + "learning_rate": 1.5775125955657877e-06, + "loss": 0.5012, + "step": 786 + }, + { + "epoch": 2.298928919182084, + "grad_norm": 0.4484643953022932, + "learning_rate": 1.5651257873074898e-06, + "loss": 0.5413, + "step": 787 + }, + { + "epoch": 2.3018500486854916, + "grad_norm": 0.4679410901074801, + "learning_rate": 1.5527787720564946e-06, + "loss": 0.5229, + "step": 788 + }, + { + "epoch": 2.3047711781888998, + "grad_norm": 0.450714999626221, + "learning_rate": 1.5404716928529356e-06, + "loss": 0.5079, + "step": 789 + }, + { + "epoch": 2.3076923076923075, + "grad_norm": 0.46250776338412736, + "learning_rate": 1.5282046922742876e-06, + "loss": 0.5056, + "step": 790 + }, + { + "epoch": 2.3106134371957157, + "grad_norm": 0.47191060686599, + "learning_rate": 1.515977912433717e-06, + "loss": 0.5279, + "step": 791 + }, + { + "epoch": 2.313534566699124, + "grad_norm": 0.493016587785878, + "learning_rate": 1.50379149497843e-06, + "loss": 0.5483, + "step": 792 + }, + { + "epoch": 2.3164556962025316, + "grad_norm": 0.47810281591953624, + "learning_rate": 1.4916455810880358e-06, + "loss": 0.4898, + "step": 793 + }, + { + "epoch": 2.3193768257059397, + "grad_norm": 0.4478075131535074, + "learning_rate": 1.4795403114729095e-06, + "loss": 0.5101, + "step": 794 + }, + { + "epoch": 2.3222979552093475, + "grad_norm": 0.4875641000416008, + "learning_rate": 1.4674758263725614e-06, + "loss": 0.5854, + "step": 795 + }, + { + "epoch": 2.3252190847127556, + "grad_norm": 0.42525631573701944, + "learning_rate": 1.4554522655540176e-06, + "loss": 0.4989, + "step": 796 + }, + { + "epoch": 2.3281402142161634, + "grad_norm": 0.41706543335088775, + "learning_rate": 1.4434697683101928e-06, + "loss": 0.516, + "step": 797 + }, + { + "epoch": 2.3310613437195715, + "grad_norm": 0.4524630863860165, + "learning_rate": 1.4315284734582802e-06, + "loss": 0.4936, + "step": 798 + }, + { + "epoch": 2.3339824732229797, + "grad_norm": 0.5066460944480237, + "learning_rate": 1.4196285193381431e-06, + "loss": 0.5665, + "step": 799 + }, + { + "epoch": 2.3369036027263874, + "grad_norm": 0.4886439078673485, + "learning_rate": 1.4077700438107183e-06, + "loss": 0.5191, + "step": 800 + }, + { + "epoch": 2.3398247322297956, + "grad_norm": 0.42502615563238366, + "learning_rate": 1.3959531842564046e-06, + "loss": 0.4785, + "step": 801 + }, + { + "epoch": 2.3427458617332033, + "grad_norm": 0.45982217381683227, + "learning_rate": 1.3841780775734847e-06, + "loss": 0.5643, + "step": 802 + }, + { + "epoch": 2.3456669912366115, + "grad_norm": 0.43225051785834356, + "learning_rate": 1.3724448601765328e-06, + "loss": 0.5629, + "step": 803 + }, + { + "epoch": 2.3485881207400197, + "grad_norm": 0.44393943057319624, + "learning_rate": 1.3607536679948397e-06, + "loss": 0.456, + "step": 804 + }, + { + "epoch": 2.3515092502434274, + "grad_norm": 0.44935770423521926, + "learning_rate": 1.3491046364708294e-06, + "loss": 0.52, + "step": 805 + }, + { + "epoch": 2.3544303797468356, + "grad_norm": 0.4573183127076476, + "learning_rate": 1.337497900558497e-06, + "loss": 0.5231, + "step": 806 + }, + { + "epoch": 2.3573515092502433, + "grad_norm": 0.4471873705684597, + "learning_rate": 1.325933594721841e-06, + "loss": 0.5191, + "step": 807 + }, + { + "epoch": 2.3602726387536515, + "grad_norm": 0.46232878806411926, + "learning_rate": 1.3144118529333126e-06, + "loss": 0.4641, + "step": 808 + }, + { + "epoch": 2.3631937682570596, + "grad_norm": 0.46569289026754496, + "learning_rate": 1.3029328086722537e-06, + "loss": 0.5363, + "step": 809 + }, + { + "epoch": 2.3661148977604674, + "grad_norm": 0.4515090725602521, + "learning_rate": 1.2914965949233572e-06, + "loss": 0.5278, + "step": 810 + }, + { + "epoch": 2.3690360272638755, + "grad_norm": 0.45127137221258873, + "learning_rate": 1.2801033441751244e-06, + "loss": 0.4919, + "step": 811 + }, + { + "epoch": 2.3719571567672832, + "grad_norm": 0.45307816143888285, + "learning_rate": 1.268753188418329e-06, + "loss": 0.5063, + "step": 812 + }, + { + "epoch": 2.3748782862706914, + "grad_norm": 0.4681309102712795, + "learning_rate": 1.257446259144494e-06, + "loss": 0.5185, + "step": 813 + }, + { + "epoch": 2.377799415774099, + "grad_norm": 0.4665552455306928, + "learning_rate": 1.24618268734436e-06, + "loss": 0.5426, + "step": 814 + }, + { + "epoch": 2.3807205452775073, + "grad_norm": 0.45997968978470405, + "learning_rate": 1.2349626035063705e-06, + "loss": 0.4996, + "step": 815 + }, + { + "epoch": 2.3836416747809155, + "grad_norm": 0.470113793180687, + "learning_rate": 1.2237861376151632e-06, + "loss": 0.5392, + "step": 816 + }, + { + "epoch": 2.386562804284323, + "grad_norm": 0.47230105708445114, + "learning_rate": 1.2126534191500622e-06, + "loss": 0.5361, + "step": 817 + }, + { + "epoch": 2.3894839337877314, + "grad_norm": 0.4439367141535501, + "learning_rate": 1.2015645770835765e-06, + "loss": 0.4732, + "step": 818 + }, + { + "epoch": 2.392405063291139, + "grad_norm": 0.4656016270422327, + "learning_rate": 1.1905197398799074e-06, + "loss": 0.5227, + "step": 819 + }, + { + "epoch": 2.3953261927945473, + "grad_norm": 0.46417888892808024, + "learning_rate": 1.1795190354934587e-06, + "loss": 0.5606, + "step": 820 + }, + { + "epoch": 2.398247322297955, + "grad_norm": 0.41153069172218604, + "learning_rate": 1.168562591367361e-06, + "loss": 0.5035, + "step": 821 + }, + { + "epoch": 2.401168451801363, + "grad_norm": 0.4591933177826969, + "learning_rate": 1.1576505344319843e-06, + "loss": 0.5485, + "step": 822 + }, + { + "epoch": 2.4040895813047714, + "grad_norm": 0.48641885058814227, + "learning_rate": 1.146782991103475e-06, + "loss": 0.5122, + "step": 823 + }, + { + "epoch": 2.407010710808179, + "grad_norm": 0.452636414728882, + "learning_rate": 1.1359600872822879e-06, + "loss": 0.528, + "step": 824 + }, + { + "epoch": 2.4099318403115872, + "grad_norm": 0.4637194713579063, + "learning_rate": 1.1251819483517334e-06, + "loss": 0.5335, + "step": 825 + }, + { + "epoch": 2.412852969814995, + "grad_norm": 0.46499658066713967, + "learning_rate": 1.1144486991765175e-06, + "loss": 0.519, + "step": 826 + }, + { + "epoch": 2.415774099318403, + "grad_norm": 0.4519765342482912, + "learning_rate": 1.1037604641012995e-06, + "loss": 0.5488, + "step": 827 + }, + { + "epoch": 2.418695228821811, + "grad_norm": 0.4382247666420335, + "learning_rate": 1.0931173669492472e-06, + "loss": 0.4853, + "step": 828 + }, + { + "epoch": 2.421616358325219, + "grad_norm": 0.45306336606053793, + "learning_rate": 1.0825195310206132e-06, + "loss": 0.5224, + "step": 829 + }, + { + "epoch": 2.424537487828627, + "grad_norm": 0.4473053801916244, + "learning_rate": 1.0719670790912928e-06, + "loss": 0.4971, + "step": 830 + }, + { + "epoch": 2.427458617332035, + "grad_norm": 0.44465614500766465, + "learning_rate": 1.0614601334114099e-06, + "loss": 0.5404, + "step": 831 + }, + { + "epoch": 2.430379746835443, + "grad_norm": 0.4251639911624754, + "learning_rate": 1.0509988157038952e-06, + "loss": 0.5179, + "step": 832 + }, + { + "epoch": 2.433300876338851, + "grad_norm": 0.40203118130358384, + "learning_rate": 1.0405832471630862e-06, + "loss": 0.4465, + "step": 833 + }, + { + "epoch": 2.436222005842259, + "grad_norm": 0.49627881222358805, + "learning_rate": 1.030213548453311e-06, + "loss": 0.5782, + "step": 834 + }, + { + "epoch": 2.439143135345667, + "grad_norm": 0.4352033965269781, + "learning_rate": 1.019889839707498e-06, + "loss": 0.471, + "step": 835 + }, + { + "epoch": 2.442064264849075, + "grad_norm": 0.4112365791877552, + "learning_rate": 1.0096122405257802e-06, + "loss": 0.525, + "step": 836 + }, + { + "epoch": 2.444985394352483, + "grad_norm": 0.46113597408698487, + "learning_rate": 9.99380869974116e-07, + "loss": 0.55, + "step": 837 + }, + { + "epoch": 2.447906523855891, + "grad_norm": 0.42894572321057406, + "learning_rate": 9.891958465828999e-07, + "loss": 0.4883, + "step": 838 + }, + { + "epoch": 2.450827653359299, + "grad_norm": 0.4306576340142323, + "learning_rate": 9.790572883455974e-07, + "loss": 0.5127, + "step": 839 + }, + { + "epoch": 2.453748782862707, + "grad_norm": 0.47843377943065934, + "learning_rate": 9.689653127173743e-07, + "loss": 0.5376, + "step": 840 + }, + { + "epoch": 2.456669912366115, + "grad_norm": 0.48286119801655825, + "learning_rate": 9.589200366137375e-07, + "loss": 0.5108, + "step": 841 + }, + { + "epoch": 2.459591041869523, + "grad_norm": 0.4840439892081914, + "learning_rate": 9.489215764091813e-07, + "loss": 0.5436, + "step": 842 + }, + { + "epoch": 2.4625121713729308, + "grad_norm": 0.4360507831964012, + "learning_rate": 9.389700479358365e-07, + "loss": 0.4733, + "step": 843 + }, + { + "epoch": 2.465433300876339, + "grad_norm": 0.449450069287443, + "learning_rate": 9.290655664821296e-07, + "loss": 0.5599, + "step": 844 + }, + { + "epoch": 2.4683544303797467, + "grad_norm": 0.41280059188581175, + "learning_rate": 9.192082467914465e-07, + "loss": 0.4792, + "step": 845 + }, + { + "epoch": 2.471275559883155, + "grad_norm": 0.5093589634384059, + "learning_rate": 9.093982030608095e-07, + "loss": 0.5397, + "step": 846 + }, + { + "epoch": 2.474196689386563, + "grad_norm": 0.45448988250512634, + "learning_rate": 8.996355489395442e-07, + "loss": 0.4975, + "step": 847 + }, + { + "epoch": 2.4771178188899707, + "grad_norm": 0.44480146821328176, + "learning_rate": 8.899203975279674e-07, + "loss": 0.4914, + "step": 848 + }, + { + "epoch": 2.480038948393379, + "grad_norm": 0.4896804183179627, + "learning_rate": 8.802528613760775e-07, + "loss": 0.5599, + "step": 849 + }, + { + "epoch": 2.4829600778967866, + "grad_norm": 0.4529609749837371, + "learning_rate": 8.706330524822548e-07, + "loss": 0.4598, + "step": 850 + }, + { + "epoch": 2.485881207400195, + "grad_norm": 0.4141270566207645, + "learning_rate": 8.610610822919546e-07, + "loss": 0.5195, + "step": 851 + }, + { + "epoch": 2.4888023369036025, + "grad_norm": 0.4413472687231943, + "learning_rate": 8.515370616964219e-07, + "loss": 0.5181, + "step": 852 + }, + { + "epoch": 2.4917234664070107, + "grad_norm": 0.4120555785416349, + "learning_rate": 8.420611010314062e-07, + "loss": 0.5267, + "step": 853 + }, + { + "epoch": 2.494644595910419, + "grad_norm": 0.4661650695044543, + "learning_rate": 8.326333100758843e-07, + "loss": 0.5086, + "step": 854 + }, + { + "epoch": 2.4975657254138266, + "grad_norm": 0.5073361612635612, + "learning_rate": 8.232537980507848e-07, + "loss": 0.5536, + "step": 855 + }, + { + "epoch": 2.5004868549172348, + "grad_norm": 0.4961949890878462, + "learning_rate": 8.13922673617727e-07, + "loss": 0.5238, + "step": 856 + }, + { + "epoch": 2.5034079844206425, + "grad_norm": 0.4720376448013444, + "learning_rate": 8.046400448777575e-07, + "loss": 0.4823, + "step": 857 + }, + { + "epoch": 2.5063291139240507, + "grad_norm": 0.49809468171059407, + "learning_rate": 7.954060193701019e-07, + "loss": 0.5466, + "step": 858 + }, + { + "epoch": 2.5092502434274584, + "grad_norm": 0.4639329084639315, + "learning_rate": 7.862207040709191e-07, + "loss": 0.5104, + "step": 859 + }, + { + "epoch": 2.5121713729308666, + "grad_norm": 0.47993248882289974, + "learning_rate": 7.770842053920585e-07, + "loss": 0.5379, + "step": 860 + }, + { + "epoch": 2.5150925024342747, + "grad_norm": 0.42897728377766037, + "learning_rate": 7.679966291798297e-07, + "loss": 0.482, + "step": 861 + }, + { + "epoch": 2.5180136319376825, + "grad_norm": 0.45778561195288503, + "learning_rate": 7.589580807137742e-07, + "loss": 0.5032, + "step": 862 + }, + { + "epoch": 2.5209347614410906, + "grad_norm": 0.46270673489951974, + "learning_rate": 7.4996866470545e-07, + "loss": 0.5706, + "step": 863 + }, + { + "epoch": 2.523855890944499, + "grad_norm": 0.42072713048943156, + "learning_rate": 7.410284852972127e-07, + "loss": 0.5204, + "step": 864 + }, + { + "epoch": 2.5267770204479065, + "grad_norm": 0.40846266187572267, + "learning_rate": 7.321376460610136e-07, + "loss": 0.4916, + "step": 865 + }, + { + "epoch": 2.5296981499513143, + "grad_norm": 0.4443443498829621, + "learning_rate": 7.232962499972002e-07, + "loss": 0.5395, + "step": 866 + }, + { + "epoch": 2.5326192794547224, + "grad_norm": 0.459368960082765, + "learning_rate": 7.145043995333173e-07, + "loss": 0.4924, + "step": 867 + }, + { + "epoch": 2.5355404089581306, + "grad_norm": 0.47824210909228654, + "learning_rate": 7.057621965229267e-07, + "loss": 0.513, + "step": 868 + }, + { + "epoch": 2.5384615384615383, + "grad_norm": 0.438214605329431, + "learning_rate": 6.970697422444228e-07, + "loss": 0.5695, + "step": 869 + }, + { + "epoch": 2.5413826679649465, + "grad_norm": 0.45798584370168316, + "learning_rate": 6.884271373998608e-07, + "loss": 0.4925, + "step": 870 + }, + { + "epoch": 2.5443037974683547, + "grad_norm": 0.4811217026678713, + "learning_rate": 6.798344821137947e-07, + "loss": 0.5083, + "step": 871 + }, + { + "epoch": 2.5472249269717624, + "grad_norm": 0.4418809012550695, + "learning_rate": 6.71291875932108e-07, + "loss": 0.5352, + "step": 872 + }, + { + "epoch": 2.5501460564751706, + "grad_norm": 0.42576085849329387, + "learning_rate": 6.62799417820868e-07, + "loss": 0.5262, + "step": 873 + }, + { + "epoch": 2.5530671859785783, + "grad_norm": 0.4287609653415205, + "learning_rate": 6.543572061651738e-07, + "loss": 0.5399, + "step": 874 + }, + { + "epoch": 2.5559883154819865, + "grad_norm": 0.45232434071279953, + "learning_rate": 6.459653387680248e-07, + "loss": 0.4725, + "step": 875 + }, + { + "epoch": 2.558909444985394, + "grad_norm": 0.4554887075692202, + "learning_rate": 6.376239128491784e-07, + "loss": 0.5122, + "step": 876 + }, + { + "epoch": 2.5618305744888024, + "grad_norm": 0.46277880760014206, + "learning_rate": 6.293330250440277e-07, + "loss": 0.4941, + "step": 877 + }, + { + "epoch": 2.5647517039922105, + "grad_norm": 0.4962082268099505, + "learning_rate": 6.210927714024834e-07, + "loss": 0.5592, + "step": 878 + }, + { + "epoch": 2.5676728334956183, + "grad_norm": 0.44324064458458867, + "learning_rate": 6.129032473878582e-07, + "loss": 0.5282, + "step": 879 + }, + { + "epoch": 2.5705939629990264, + "grad_norm": 0.43320029335686094, + "learning_rate": 6.047645478757635e-07, + "loss": 0.4562, + "step": 880 + }, + { + "epoch": 2.573515092502434, + "grad_norm": 0.45163259818241996, + "learning_rate": 5.966767671530078e-07, + "loss": 0.5501, + "step": 881 + }, + { + "epoch": 2.5764362220058423, + "grad_norm": 0.4552021870162526, + "learning_rate": 5.88639998916506e-07, + "loss": 0.5359, + "step": 882 + }, + { + "epoch": 2.57935735150925, + "grad_norm": 0.4604883255796744, + "learning_rate": 5.806543362721945e-07, + "loss": 0.5274, + "step": 883 + }, + { + "epoch": 2.5822784810126582, + "grad_norm": 0.41454954591229876, + "learning_rate": 5.727198717339511e-07, + "loss": 0.4765, + "step": 884 + }, + { + "epoch": 2.5851996105160664, + "grad_norm": 0.4325932298835654, + "learning_rate": 5.648366972225222e-07, + "loss": 0.5282, + "step": 885 + }, + { + "epoch": 2.588120740019474, + "grad_norm": 0.433038404211758, + "learning_rate": 5.570049040644609e-07, + "loss": 0.5137, + "step": 886 + }, + { + "epoch": 2.5910418695228823, + "grad_norm": 0.45496347053471886, + "learning_rate": 5.492245829910664e-07, + "loss": 0.4973, + "step": 887 + }, + { + "epoch": 2.59396299902629, + "grad_norm": 0.44790019595001196, + "learning_rate": 5.414958241373358e-07, + "loss": 0.4881, + "step": 888 + }, + { + "epoch": 2.596884128529698, + "grad_norm": 0.4506105888324243, + "learning_rate": 5.33818717040917e-07, + "loss": 0.5446, + "step": 889 + }, + { + "epoch": 2.599805258033106, + "grad_norm": 0.45920237343152387, + "learning_rate": 5.261933506410722e-07, + "loss": 0.5544, + "step": 890 + }, + { + "epoch": 2.602726387536514, + "grad_norm": 0.433968288589714, + "learning_rate": 5.186198132776459e-07, + "loss": 0.4792, + "step": 891 + }, + { + "epoch": 2.6056475170399223, + "grad_norm": 0.45054587006233904, + "learning_rate": 5.110981926900488e-07, + "loss": 0.5271, + "step": 892 + }, + { + "epoch": 2.60856864654333, + "grad_norm": 0.4402130423120147, + "learning_rate": 5.036285760162307e-07, + "loss": 0.4772, + "step": 893 + }, + { + "epoch": 2.611489776046738, + "grad_norm": 0.4560401508104557, + "learning_rate": 4.962110497916789e-07, + "loss": 0.5078, + "step": 894 + }, + { + "epoch": 2.6144109055501463, + "grad_norm": 0.4574934740247989, + "learning_rate": 4.888456999484098e-07, + "loss": 0.5177, + "step": 895 + }, + { + "epoch": 2.617332035053554, + "grad_norm": 0.42797938213430475, + "learning_rate": 4.815326118139813e-07, + "loss": 0.5068, + "step": 896 + }, + { + "epoch": 2.620253164556962, + "grad_norm": 0.4508198643702282, + "learning_rate": 4.742718701104965e-07, + "loss": 0.4836, + "step": 897 + }, + { + "epoch": 2.62317429406037, + "grad_norm": 0.4936637557435792, + "learning_rate": 4.670635589536254e-07, + "loss": 0.5688, + "step": 898 + }, + { + "epoch": 2.626095423563778, + "grad_norm": 0.4446810251135343, + "learning_rate": 4.599077618516312e-07, + "loss": 0.5213, + "step": 899 + }, + { + "epoch": 2.629016553067186, + "grad_norm": 0.4393121918571539, + "learning_rate": 4.528045617044019e-07, + "loss": 0.515, + "step": 900 + }, + { + "epoch": 2.631937682570594, + "grad_norm": 0.4546453714565344, + "learning_rate": 4.457540408024896e-07, + "loss": 0.5119, + "step": 901 + }, + { + "epoch": 2.634858812074002, + "grad_norm": 0.43919005512991516, + "learning_rate": 4.387562808261575e-07, + "loss": 0.5273, + "step": 902 + }, + { + "epoch": 2.63777994157741, + "grad_norm": 0.4491615831317555, + "learning_rate": 4.3181136284443204e-07, + "loss": 0.4994, + "step": 903 + }, + { + "epoch": 2.6407010710808176, + "grad_norm": 0.48434931343594284, + "learning_rate": 4.249193673141694e-07, + "loss": 0.5186, + "step": 904 + }, + { + "epoch": 2.643622200584226, + "grad_norm": 0.41259506287384623, + "learning_rate": 4.180803740791156e-07, + "loss": 0.5319, + "step": 905 + }, + { + "epoch": 2.646543330087634, + "grad_norm": 0.4076433170920647, + "learning_rate": 4.112944623689863e-07, + "loss": 0.5178, + "step": 906 + }, + { + "epoch": 2.6494644595910417, + "grad_norm": 0.421345630456112, + "learning_rate": 4.0456171079854833e-07, + "loss": 0.4951, + "step": 907 + }, + { + "epoch": 2.65238558909445, + "grad_norm": 0.4435092242276153, + "learning_rate": 3.978821973667074e-07, + "loss": 0.4837, + "step": 908 + }, + { + "epoch": 2.655306718597858, + "grad_norm": 0.44521873058867134, + "learning_rate": 3.9125599945560866e-07, + "loss": 0.5526, + "step": 909 + }, + { + "epoch": 2.6582278481012658, + "grad_norm": 0.45505539256105115, + "learning_rate": 3.846831938297324e-07, + "loss": 0.5158, + "step": 910 + }, + { + "epoch": 2.661148977604674, + "grad_norm": 0.433074017412774, + "learning_rate": 3.7816385663501097e-07, + "loss": 0.4809, + "step": 911 + }, + { + "epoch": 2.6640701071080817, + "grad_norm": 0.4502224951078667, + "learning_rate": 3.716980633979489e-07, + "loss": 0.5363, + "step": 912 + }, + { + "epoch": 2.66699123661149, + "grad_norm": 0.43979723124115566, + "learning_rate": 3.6528588902473905e-07, + "loss": 0.508, + "step": 913 + }, + { + "epoch": 2.6699123661148976, + "grad_norm": 0.46901546979749387, + "learning_rate": 3.589274078004029e-07, + "loss": 0.5489, + "step": 914 + }, + { + "epoch": 2.6728334956183057, + "grad_norm": 0.4432062648822634, + "learning_rate": 3.5262269338792623e-07, + "loss": 0.526, + "step": 915 + }, + { + "epoch": 2.675754625121714, + "grad_norm": 0.43511333463780544, + "learning_rate": 3.463718188274046e-07, + "loss": 0.5332, + "step": 916 + }, + { + "epoch": 2.6786757546251216, + "grad_norm": 0.38843912794513447, + "learning_rate": 3.401748565352031e-07, + "loss": 0.4997, + "step": 917 + }, + { + "epoch": 2.68159688412853, + "grad_norm": 0.41986039406023057, + "learning_rate": 3.340318783031099e-07, + "loss": 0.4804, + "step": 918 + }, + { + "epoch": 2.6845180136319375, + "grad_norm": 0.44648799987073357, + "learning_rate": 3.279429552975094e-07, + "loss": 0.57, + "step": 919 + }, + { + "epoch": 2.6874391431353457, + "grad_norm": 0.38625226404630086, + "learning_rate": 3.219081580585548e-07, + "loss": 0.418, + "step": 920 + }, + { + "epoch": 2.6903602726387534, + "grad_norm": 0.46495901890128216, + "learning_rate": 3.159275564993558e-07, + "loss": 0.562, + "step": 921 + }, + { + "epoch": 2.6932814021421616, + "grad_norm": 0.4542095138457356, + "learning_rate": 3.100012199051627e-07, + "loss": 0.48, + "step": 922 + }, + { + "epoch": 2.6962025316455698, + "grad_norm": 0.461034837465003, + "learning_rate": 3.0412921693256657e-07, + "loss": 0.5802, + "step": 923 + }, + { + "epoch": 2.6991236611489775, + "grad_norm": 0.4124404226906978, + "learning_rate": 2.9831161560870346e-07, + "loss": 0.5124, + "step": 924 + }, + { + "epoch": 2.7020447906523857, + "grad_norm": 0.420405602577591, + "learning_rate": 2.9254848333046817e-07, + "loss": 0.4959, + "step": 925 + }, + { + "epoch": 2.704965920155794, + "grad_norm": 0.42637463336396936, + "learning_rate": 2.8683988686372956e-07, + "loss": 0.529, + "step": 926 + }, + { + "epoch": 2.7078870496592016, + "grad_norm": 0.4156040595949887, + "learning_rate": 2.811858923425609e-07, + "loss": 0.5184, + "step": 927 + }, + { + "epoch": 2.7108081791626093, + "grad_norm": 0.4312166608407746, + "learning_rate": 2.755865652684703e-07, + "loss": 0.5438, + "step": 928 + }, + { + "epoch": 2.7137293086660175, + "grad_norm": 0.4369289965460609, + "learning_rate": 2.7004197050964744e-07, + "loss": 0.4978, + "step": 929 + }, + { + "epoch": 2.7166504381694256, + "grad_norm": 0.426344227648129, + "learning_rate": 2.645521723002037e-07, + "loss": 0.5041, + "step": 930 + }, + { + "epoch": 2.7195715676728334, + "grad_norm": 0.4417752376579622, + "learning_rate": 2.591172342394349e-07, + "loss": 0.5587, + "step": 931 + }, + { + "epoch": 2.7224926971762415, + "grad_norm": 0.457814912149392, + "learning_rate": 2.537372192910825e-07, + "loss": 0.5175, + "step": 932 + }, + { + "epoch": 2.7254138266796497, + "grad_norm": 0.45549699443779185, + "learning_rate": 2.4841218978260194e-07, + "loss": 0.5033, + "step": 933 + }, + { + "epoch": 2.7283349561830574, + "grad_norm": 0.44509990954989737, + "learning_rate": 2.4314220740444595e-07, + "loss": 0.5192, + "step": 934 + }, + { + "epoch": 2.731256085686465, + "grad_norm": 0.4542595414389855, + "learning_rate": 2.3792733320934348e-07, + "loss": 0.5193, + "step": 935 + }, + { + "epoch": 2.7341772151898733, + "grad_norm": 0.46788823553431785, + "learning_rate": 2.3276762761159588e-07, + "loss": 0.5226, + "step": 936 + }, + { + "epoch": 2.7370983446932815, + "grad_norm": 0.46548939441743264, + "learning_rate": 2.276631503863763e-07, + "loss": 0.5234, + "step": 937 + }, + { + "epoch": 2.7400194741966892, + "grad_norm": 0.43071296003232673, + "learning_rate": 2.2261396066903916e-07, + "loss": 0.4968, + "step": 938 + }, + { + "epoch": 2.7429406037000974, + "grad_norm": 0.4422590771192652, + "learning_rate": 2.176201169544312e-07, + "loss": 0.5135, + "step": 939 + }, + { + "epoch": 2.7458617332035056, + "grad_norm": 0.43839294000442797, + "learning_rate": 2.12681677096217e-07, + "loss": 0.5459, + "step": 940 + }, + { + "epoch": 2.7487828627069133, + "grad_norm": 0.4296191280132258, + "learning_rate": 2.0779869830620692e-07, + "loss": 0.4538, + "step": 941 + }, + { + "epoch": 2.7517039922103215, + "grad_norm": 0.4119484440139385, + "learning_rate": 2.029712371536957e-07, + "loss": 0.4839, + "step": 942 + }, + { + "epoch": 2.754625121713729, + "grad_norm": 0.4512158763524109, + "learning_rate": 1.9819934956480603e-07, + "loss": 0.5445, + "step": 943 + }, + { + "epoch": 2.7575462512171374, + "grad_norm": 0.4212292286246555, + "learning_rate": 1.9348309082184102e-07, + "loss": 0.4934, + "step": 944 + }, + { + "epoch": 2.760467380720545, + "grad_norm": 0.44496919909800775, + "learning_rate": 1.888225155626433e-07, + "loss": 0.5456, + "step": 945 + }, + { + "epoch": 2.7633885102239533, + "grad_norm": 0.41713723572900646, + "learning_rate": 1.8421767777996425e-07, + "loss": 0.4462, + "step": 946 + }, + { + "epoch": 2.7663096397273614, + "grad_norm": 0.4457624171080001, + "learning_rate": 1.7966863082083507e-07, + "loss": 0.5427, + "step": 947 + }, + { + "epoch": 2.769230769230769, + "grad_norm": 0.4348007177346548, + "learning_rate": 1.7517542738595071e-07, + "loss": 0.5568, + "step": 948 + }, + { + "epoch": 2.7721518987341773, + "grad_norm": 0.41399848976523823, + "learning_rate": 1.7073811952905862e-07, + "loss": 0.5083, + "step": 949 + }, + { + "epoch": 2.775073028237585, + "grad_norm": 0.43232436981734923, + "learning_rate": 1.6635675865635859e-07, + "loss": 0.4505, + "step": 950 + }, + { + "epoch": 2.7779941577409932, + "grad_norm": 0.4750164143516587, + "learning_rate": 1.6203139552590175e-07, + "loss": 0.5831, + "step": 951 + }, + { + "epoch": 2.780915287244401, + "grad_norm": 0.39972505599387415, + "learning_rate": 1.5776208024700702e-07, + "loss": 0.4751, + "step": 952 + }, + { + "epoch": 2.783836416747809, + "grad_norm": 0.42708956986618357, + "learning_rate": 1.5354886227967934e-07, + "loss": 0.532, + "step": 953 + }, + { + "epoch": 2.7867575462512173, + "grad_norm": 0.4435213719711565, + "learning_rate": 1.4939179043403474e-07, + "loss": 0.5347, + "step": 954 + }, + { + "epoch": 2.789678675754625, + "grad_norm": 0.4416374818886015, + "learning_rate": 1.4529091286973994e-07, + "loss": 0.5459, + "step": 955 + }, + { + "epoch": 2.792599805258033, + "grad_norm": 0.45153869521194906, + "learning_rate": 1.4124627709544814e-07, + "loss": 0.4918, + "step": 956 + }, + { + "epoch": 2.7955209347614414, + "grad_norm": 0.4440135260222283, + "learning_rate": 1.3725792996825083e-07, + "loss": 0.5016, + "step": 957 + }, + { + "epoch": 2.798442064264849, + "grad_norm": 0.39484795107770926, + "learning_rate": 1.3332591769314006e-07, + "loss": 0.5043, + "step": 958 + }, + { + "epoch": 2.801363193768257, + "grad_norm": 0.444235962409132, + "learning_rate": 1.2945028582246445e-07, + "loss": 0.5637, + "step": 959 + }, + { + "epoch": 2.804284323271665, + "grad_norm": 0.4583481416669065, + "learning_rate": 1.2563107925540774e-07, + "loss": 0.5017, + "step": 960 + }, + { + "epoch": 2.807205452775073, + "grad_norm": 0.4215520268166574, + "learning_rate": 1.2186834223746612e-07, + "loss": 0.5136, + "step": 961 + }, + { + "epoch": 2.810126582278481, + "grad_norm": 0.43820018627048524, + "learning_rate": 1.1816211835993684e-07, + "loss": 0.4964, + "step": 962 + }, + { + "epoch": 2.813047711781889, + "grad_norm": 0.4420212484091613, + "learning_rate": 1.1451245055941251e-07, + "loss": 0.5112, + "step": 963 + }, + { + "epoch": 2.8159688412852972, + "grad_norm": 0.4365334218610012, + "learning_rate": 1.1091938111728374e-07, + "loss": 0.5273, + "step": 964 + }, + { + "epoch": 2.818889970788705, + "grad_norm": 0.4377608647741322, + "learning_rate": 1.0738295165924783e-07, + "loss": 0.5398, + "step": 965 + }, + { + "epoch": 2.8218111002921127, + "grad_norm": 0.45046127023142907, + "learning_rate": 1.0390320315482982e-07, + "loss": 0.5062, + "step": 966 + }, + { + "epoch": 2.824732229795521, + "grad_norm": 0.4019879301034501, + "learning_rate": 1.0048017591690607e-07, + "loss": 0.5114, + "step": 967 + }, + { + "epoch": 2.827653359298929, + "grad_norm": 0.4402917529148519, + "learning_rate": 9.711390960123634e-08, + "loss": 0.5359, + "step": 968 + }, + { + "epoch": 2.8305744888023368, + "grad_norm": 0.41904135734811376, + "learning_rate": 9.380444320600591e-08, + "loss": 0.4684, + "step": 969 + }, + { + "epoch": 2.833495618305745, + "grad_norm": 0.4659811321879393, + "learning_rate": 9.055181507137245e-08, + "loss": 0.5426, + "step": 970 + }, + { + "epoch": 2.836416747809153, + "grad_norm": 0.440634928654449, + "learning_rate": 8.73560628790232e-08, + "loss": 0.5258, + "step": 971 + }, + { + "epoch": 2.839337877312561, + "grad_norm": 0.44241019235537177, + "learning_rate": 8.421722365173856e-08, + "loss": 0.553, + "step": 972 + }, + { + "epoch": 2.842259006815969, + "grad_norm": 0.4085933163681831, + "learning_rate": 8.113533375295968e-08, + "loss": 0.4804, + "step": 973 + }, + { + "epoch": 2.8451801363193767, + "grad_norm": 0.42283609192144683, + "learning_rate": 7.81104288863721e-08, + "loss": 0.5268, + "step": 974 + }, + { + "epoch": 2.848101265822785, + "grad_norm": 0.4414452447566111, + "learning_rate": 7.514254409549005e-08, + "loss": 0.5059, + "step": 975 + }, + { + "epoch": 2.8510223953261926, + "grad_norm": 0.477876908869347, + "learning_rate": 7.223171376325e-08, + "loss": 0.5401, + "step": 976 + }, + { + "epoch": 2.853943524829601, + "grad_norm": 0.45667617596842697, + "learning_rate": 6.93779716116122e-08, + "loss": 0.5179, + "step": 977 + }, + { + "epoch": 2.856864654333009, + "grad_norm": 0.43661296400815486, + "learning_rate": 6.658135070116978e-08, + "loss": 0.479, + "step": 978 + }, + { + "epoch": 2.8597857838364167, + "grad_norm": 0.43447801900111893, + "learning_rate": 6.384188343076914e-08, + "loss": 0.4933, + "step": 979 + }, + { + "epoch": 2.862706913339825, + "grad_norm": 0.4567696696725672, + "learning_rate": 6.115960153712963e-08, + "loss": 0.5225, + "step": 980 + }, + { + "epoch": 2.8656280428432326, + "grad_norm": 0.4423237005020546, + "learning_rate": 5.8534536094478345e-08, + "loss": 0.5649, + "step": 981 + }, + { + "epoch": 2.8685491723466408, + "grad_norm": 0.41379887497949286, + "learning_rate": 5.596671751419147e-08, + "loss": 0.474, + "step": 982 + }, + { + "epoch": 2.8714703018500485, + "grad_norm": 0.4382842717392665, + "learning_rate": 5.3456175544437936e-08, + "loss": 0.529, + "step": 983 + }, + { + "epoch": 2.8743914313534566, + "grad_norm": 0.4463860151059245, + "learning_rate": 5.100293926984023e-08, + "loss": 0.4909, + "step": 984 + }, + { + "epoch": 2.877312560856865, + "grad_norm": 0.4057084453405752, + "learning_rate": 4.860703711113246e-08, + "loss": 0.5029, + "step": 985 + }, + { + "epoch": 2.8802336903602725, + "grad_norm": 0.42518495194251527, + "learning_rate": 4.626849682483392e-08, + "loss": 0.5376, + "step": 986 + }, + { + "epoch": 2.8831548198636807, + "grad_norm": 0.45413911287155506, + "learning_rate": 4.398734550292716e-08, + "loss": 0.5467, + "step": 987 + }, + { + "epoch": 2.8860759493670884, + "grad_norm": 0.4142515305933239, + "learning_rate": 4.176360957254266e-08, + "loss": 0.4568, + "step": 988 + }, + { + "epoch": 2.8889970788704966, + "grad_norm": 0.435060369756503, + "learning_rate": 3.959731479565465e-08, + "loss": 0.5319, + "step": 989 + }, + { + "epoch": 2.8919182083739043, + "grad_norm": 0.43929179209673974, + "learning_rate": 3.748848626878132e-08, + "loss": 0.5782, + "step": 990 + }, + { + "epoch": 2.8948393378773125, + "grad_norm": 0.4061515904911967, + "learning_rate": 3.543714842269508e-08, + "loss": 0.462, + "step": 991 + }, + { + "epoch": 2.8977604673807207, + "grad_norm": 0.46420928088916874, + "learning_rate": 3.344332502213887e-08, + "loss": 0.5218, + "step": 992 + }, + { + "epoch": 2.9006815968841284, + "grad_norm": 0.4346617171265348, + "learning_rate": 3.150703916554976e-08, + "loss": 0.4832, + "step": 993 + }, + { + "epoch": 2.9036027263875366, + "grad_norm": 0.45371971197536476, + "learning_rate": 2.962831328479465e-08, + "loss": 0.5087, + "step": 994 + }, + { + "epoch": 2.9065238558909448, + "grad_norm": 0.47344580171350403, + "learning_rate": 2.7807169144906108e-08, + "loss": 0.5639, + "step": 995 + }, + { + "epoch": 2.9094449853943525, + "grad_norm": 0.4343079692320858, + "learning_rate": 2.6043627843834185e-08, + "loss": 0.534, + "step": 996 + }, + { + "epoch": 2.91236611489776, + "grad_norm": 0.41472366849725406, + "learning_rate": 2.4337709812199405e-08, + "loss": 0.471, + "step": 997 + }, + { + "epoch": 2.9152872444011684, + "grad_norm": 0.41847164257771846, + "learning_rate": 2.2689434813056856e-08, + "loss": 0.5252, + "step": 998 + }, + { + "epoch": 2.9182083739045765, + "grad_norm": 0.4540400278681399, + "learning_rate": 2.1098821941667457e-08, + "loss": 0.5683, + "step": 999 + }, + { + "epoch": 2.9211295034079843, + "grad_norm": 0.43916840754064074, + "learning_rate": 1.9565889625275945e-08, + "loss": 0.4844, + "step": 1000 + }, + { + "epoch": 2.9240506329113924, + "grad_norm": 0.4345074027775515, + "learning_rate": 1.8090655622899356e-08, + "loss": 0.4881, + "step": 1001 + }, + { + "epoch": 2.9269717624148006, + "grad_norm": 0.4796907378778938, + "learning_rate": 1.6673137025118303e-08, + "loss": 0.5576, + "step": 1002 + }, + { + "epoch": 2.9298928919182083, + "grad_norm": 0.4751897325081864, + "learning_rate": 1.5313350253882143e-08, + "loss": 0.5248, + "step": 1003 + }, + { + "epoch": 2.9328140214216165, + "grad_norm": 0.4566958532328187, + "learning_rate": 1.4011311062316347e-08, + "loss": 0.5207, + "step": 1004 + }, + { + "epoch": 2.9357351509250242, + "grad_norm": 0.4143592301627743, + "learning_rate": 1.2767034534540978e-08, + "loss": 0.483, + "step": 1005 + }, + { + "epoch": 2.9386562804284324, + "grad_norm": 0.4151441658060871, + "learning_rate": 1.158053508549528e-08, + "loss": 0.4624, + "step": 1006 + }, + { + "epoch": 2.94157740993184, + "grad_norm": 0.44443161265224573, + "learning_rate": 1.0451826460772252e-08, + "loss": 0.5295, + "step": 1007 + }, + { + "epoch": 2.9444985394352483, + "grad_norm": 0.4244569182328679, + "learning_rate": 9.38092173645766e-09, + "loss": 0.5164, + "step": 1008 + }, + { + "epoch": 2.9474196689386565, + "grad_norm": 0.42619195275616983, + "learning_rate": 8.367833318980168e-09, + "loss": 0.5474, + "step": 1009 + }, + { + "epoch": 2.950340798442064, + "grad_norm": 0.44631991928366016, + "learning_rate": 7.412572944965335e-09, + "loss": 0.4825, + "step": 1010 + }, + { + "epoch": 2.9532619279454724, + "grad_norm": 0.4028816094942567, + "learning_rate": 6.515151681101284e-09, + "loss": 0.507, + "step": 1011 + }, + { + "epoch": 2.95618305744888, + "grad_norm": 0.4346618256973241, + "learning_rate": 5.675579924011576e-09, + "loss": 0.5173, + "step": 1012 + }, + { + "epoch": 2.9591041869522883, + "grad_norm": 0.4213345827232603, + "learning_rate": 4.89386740013198e-09, + "loss": 0.4984, + "step": 1013 + }, + { + "epoch": 2.962025316455696, + "grad_norm": 0.40444422520787165, + "learning_rate": 4.170023165600001e-09, + "loss": 0.5463, + "step": 1014 + }, + { + "epoch": 2.964946445959104, + "grad_norm": 0.4158643063044851, + "learning_rate": 3.5040556061483043e-09, + "loss": 0.4929, + "step": 1015 + }, + { + "epoch": 2.9678675754625123, + "grad_norm": 0.45968510533530904, + "learning_rate": 2.8959724370092313e-09, + "loss": 0.5618, + "step": 1016 + }, + { + "epoch": 2.97078870496592, + "grad_norm": 0.45874234552575655, + "learning_rate": 2.3457807028248734e-09, + "loss": 0.5245, + "step": 1017 + }, + { + "epoch": 2.9737098344693282, + "grad_norm": 0.4715342361612812, + "learning_rate": 1.8534867775649169e-09, + "loss": 0.5129, + "step": 1018 + }, + { + "epoch": 2.976630963972736, + "grad_norm": 0.4184866954459166, + "learning_rate": 1.4190963644522549e-09, + "loss": 0.4844, + "step": 1019 + }, + { + "epoch": 2.979552093476144, + "grad_norm": 0.41399082545193333, + "learning_rate": 1.0426144958985974e-09, + "loss": 0.5588, + "step": 1020 + }, + { + "epoch": 2.982473222979552, + "grad_norm": 0.445097303921185, + "learning_rate": 7.240455334456276e-10, + "loss": 0.4614, + "step": 1021 + }, + { + "epoch": 2.98539435248296, + "grad_norm": 0.4214572233508389, + "learning_rate": 4.633931677139325e-10, + "loss": 0.5183, + "step": 1022 + }, + { + "epoch": 2.988315481986368, + "grad_norm": 0.4246075700283406, + "learning_rate": 2.60660418360259e-10, + "loss": 0.5797, + "step": 1023 + }, + { + "epoch": 2.991236611489776, + "grad_norm": 0.4332145327952596, + "learning_rate": 1.158496340419868e-10, + "loss": 0.5132, + "step": 1024 + }, + { + "epoch": 2.994157740993184, + "grad_norm": 0.4486911892623868, + "learning_rate": 2.8962492393258546e-11, + "loss": 0.5159, + "step": 1025 + }, + { + "epoch": 2.9970788704965923, + "grad_norm": 0.4412334672273621, + "learning_rate": 0.0, + "loss": 0.5245, + "step": 1026 + }, + { + "epoch": 2.9970788704965923, + "step": 1026, + "total_flos": 883131557314560.0, + "train_loss": 0.5971221869800523, + "train_runtime": 15429.9437, + "train_samples_per_second": 6.384, + "train_steps_per_second": 0.066 + } + ], + "logging_steps": 1, + "max_steps": 1026, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 883131557314560.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}