| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9972122660294703, | |
| "eval_steps": 500, | |
| "global_step": 313, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "grad_norm": 10.092560768127441, | |
| "learning_rate": 1.0000000000000002e-06, | |
| "loss": 1.9497, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 8.963287353515625, | |
| "learning_rate": 2.0000000000000003e-06, | |
| "loss": 1.951, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 9.27600383758545, | |
| "learning_rate": 3e-06, | |
| "loss": 1.9345, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 5.632405757904053, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 1.9157, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 5.668580055236816, | |
| "learning_rate": 5e-06, | |
| "loss": 1.9026, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 3.8046553134918213, | |
| "learning_rate": 6e-06, | |
| "loss": 1.8923, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 4.357985973358154, | |
| "learning_rate": 7e-06, | |
| "loss": 1.8241, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 4.685062885284424, | |
| "learning_rate": 8.000000000000001e-06, | |
| "loss": 1.8467, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 4.768229961395264, | |
| "learning_rate": 9e-06, | |
| "loss": 1.8199, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 4.796407699584961, | |
| "learning_rate": 1e-05, | |
| "loss": 1.8374, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 5.7536139488220215, | |
| "learning_rate": 9.999731248679734e-06, | |
| "loss": 1.779, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 4.202663898468018, | |
| "learning_rate": 9.99892502360984e-06, | |
| "loss": 1.7579, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 2.9114131927490234, | |
| "learning_rate": 9.99758141145994e-06, | |
| "loss": 1.7433, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 2.3823723793029785, | |
| "learning_rate": 9.995700556669052e-06, | |
| "loss": 1.7212, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 2.3254876136779785, | |
| "learning_rate": 9.993282661430058e-06, | |
| "loss": 1.7218, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 2.053166151046753, | |
| "learning_rate": 9.990327985667972e-06, | |
| "loss": 1.7256, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 2.3782012462615967, | |
| "learning_rate": 9.986836847012001e-06, | |
| "loss": 1.713, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 2.1311683654785156, | |
| "learning_rate": 9.98280962076139e-06, | |
| "loss": 1.6785, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 2.325747489929199, | |
| "learning_rate": 9.978246739845095e-06, | |
| "loss": 1.7167, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 2.1330366134643555, | |
| "learning_rate": 9.973148694775217e-06, | |
| "loss": 1.676, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 2.2632806301116943, | |
| "learning_rate": 9.967516033594295e-06, | |
| "loss": 1.7033, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 2.6582744121551514, | |
| "learning_rate": 9.961349361816384e-06, | |
| "loss": 1.6957, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 1.9663337469100952, | |
| "learning_rate": 9.954649342361952e-06, | |
| "loss": 1.6729, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 2.2557435035705566, | |
| "learning_rate": 9.947416695486633e-06, | |
| "loss": 1.6399, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 2.090054512023926, | |
| "learning_rate": 9.939652198703785e-06, | |
| "loss": 1.6792, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 2.1648776531219482, | |
| "learning_rate": 9.93135668670091e-06, | |
| "loss": 1.6625, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 2.1209850311279297, | |
| "learning_rate": 9.92253105124993e-06, | |
| "loss": 1.7057, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 2.2777392864227295, | |
| "learning_rate": 9.91317624111132e-06, | |
| "loss": 1.6052, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 2.1739561557769775, | |
| "learning_rate": 9.903293261932106e-06, | |
| "loss": 1.6139, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 2.4518871307373047, | |
| "learning_rate": 9.89288317613777e-06, | |
| "loss": 1.6759, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 1.795519471168518, | |
| "learning_rate": 9.881947102818036e-06, | |
| "loss": 1.7036, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 2.150223731994629, | |
| "learning_rate": 9.870486217606557e-06, | |
| "loss": 1.6469, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 1.91805899143219, | |
| "learning_rate": 9.858501752554548e-06, | |
| "loss": 1.6353, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 1.8129810094833374, | |
| "learning_rate": 9.845994995998332e-06, | |
| "loss": 1.6551, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 2.1308679580688477, | |
| "learning_rate": 9.83296729242084e-06, | |
| "loss": 1.617, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 1.8321012258529663, | |
| "learning_rate": 9.819420042307091e-06, | |
| "loss": 1.6213, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 1.9112164974212646, | |
| "learning_rate": 9.805354701993624e-06, | |
| "loss": 1.6245, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 2.1160471439361572, | |
| "learning_rate": 9.79077278351195e-06, | |
| "loss": 1.6405, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 2.6318371295928955, | |
| "learning_rate": 9.77567585442601e-06, | |
| "loss": 1.6234, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 3.238373279571533, | |
| "learning_rate": 9.76006553766365e-06, | |
| "loss": 1.6452, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 2.056736469268799, | |
| "learning_rate": 9.743943511342168e-06, | |
| "loss": 1.679, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 2.000826358795166, | |
| "learning_rate": 9.727311508587907e-06, | |
| "loss": 1.6904, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 2.089482069015503, | |
| "learning_rate": 9.710171317349946e-06, | |
| "loss": 1.62, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 2.5748705863952637, | |
| "learning_rate": 9.692524780207897e-06, | |
| "loss": 1.6669, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 1.9039987325668335, | |
| "learning_rate": 9.674373794173818e-06, | |
| "loss": 1.6489, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 1.8047181367874146, | |
| "learning_rate": 9.655720310488298e-06, | |
| "loss": 1.6227, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 2.0552868843078613, | |
| "learning_rate": 9.636566334410682e-06, | |
| "loss": 1.5898, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 2.252218008041382, | |
| "learning_rate": 9.616913925003514e-06, | |
| "loss": 1.6667, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 2.015887498855591, | |
| "learning_rate": 9.596765194911182e-06, | |
| "loss": 1.6668, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 2.582007884979248, | |
| "learning_rate": 9.576122310132814e-06, | |
| "loss": 1.6542, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 1.9070011377334595, | |
| "learning_rate": 9.554987489789426e-06, | |
| "loss": 1.691, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 1.9671483039855957, | |
| "learning_rate": 9.533363005885362e-06, | |
| "loss": 1.6116, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 2.4808287620544434, | |
| "learning_rate": 9.511251183064068e-06, | |
| "loss": 1.6653, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 2.237518548965454, | |
| "learning_rate": 9.48865439835817e-06, | |
| "loss": 1.6571, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 2.142627477645874, | |
| "learning_rate": 9.465575080933959e-06, | |
| "loss": 1.6247, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 2.0365779399871826, | |
| "learning_rate": 9.442015711830246e-06, | |
| "loss": 1.6323, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 2.427443742752075, | |
| "learning_rate": 9.417978823691652e-06, | |
| "loss": 1.6646, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 2.014340877532959, | |
| "learning_rate": 9.393467000496345e-06, | |
| "loss": 1.5754, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 2.6110949516296387, | |
| "learning_rate": 9.368482877278264e-06, | |
| "loss": 1.657, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 2.048231363296509, | |
| "learning_rate": 9.34302913984385e-06, | |
| "loss": 1.61, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 2.672330379486084, | |
| "learning_rate": 9.317108524483319e-06, | |
| "loss": 1.5894, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 2.6745121479034424, | |
| "learning_rate": 9.29072381767651e-06, | |
| "loss": 1.5623, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 2.1490442752838135, | |
| "learning_rate": 9.26387785579334e-06, | |
| "loss": 1.6668, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 3.805147171020508, | |
| "learning_rate": 9.236573524788888e-06, | |
| "loss": 1.6274, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 2.243023157119751, | |
| "learning_rate": 9.208813759893158e-06, | |
| "loss": 1.6496, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 2.5260114669799805, | |
| "learning_rate": 9.180601545295535e-06, | |
| "loss": 1.623, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 2.0329606533050537, | |
| "learning_rate": 9.151939913823988e-06, | |
| "loss": 1.5873, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 3.3269598484039307, | |
| "learning_rate": 9.122831946619038e-06, | |
| "loss": 1.6327, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 2.1953978538513184, | |
| "learning_rate": 9.093280772802527e-06, | |
| "loss": 1.6321, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 3.2001328468322754, | |
| "learning_rate": 9.063289569141251e-06, | |
| "loss": 1.6338, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 2.076502799987793, | |
| "learning_rate": 9.032861559705442e-06, | |
| "loss": 1.6648, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 2.0800352096557617, | |
| "learning_rate": 9.002000015522182e-06, | |
| "loss": 1.5694, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 2.0901432037353516, | |
| "learning_rate": 8.970708254223768e-06, | |
| "loss": 1.6571, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 3.1389923095703125, | |
| "learning_rate": 8.938989639691068e-06, | |
| "loss": 1.6151, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 2.0514352321624756, | |
| "learning_rate": 8.90684758169189e-06, | |
| "loss": 1.618, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 1.9462100267410278, | |
| "learning_rate": 8.87428553551445e-06, | |
| "loss": 1.6367, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 2.2343838214874268, | |
| "learning_rate": 8.841307001595904e-06, | |
| "loss": 1.6321, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 2.6064207553863525, | |
| "learning_rate": 8.807915525146065e-06, | |
| "loss": 1.5913, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 2.589186191558838, | |
| "learning_rate": 8.774114695766286e-06, | |
| "loss": 1.6, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 2.0082924365997314, | |
| "learning_rate": 8.739908147063576e-06, | |
| "loss": 1.6266, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 1.9739817380905151, | |
| "learning_rate": 8.705299556259986e-06, | |
| "loss": 1.5926, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 2.9566433429718018, | |
| "learning_rate": 8.670292643797302e-06, | |
| "loss": 1.624, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 2.10595703125, | |
| "learning_rate": 8.634891172937102e-06, | |
| "loss": 1.6437, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 1.7080891132354736, | |
| "learning_rate": 8.599098949356201e-06, | |
| "loss": 1.6131, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 2.358856678009033, | |
| "learning_rate": 8.562919820737537e-06, | |
| "loss": 1.5748, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 2.030383825302124, | |
| "learning_rate": 8.526357676356538e-06, | |
| "loss": 1.6389, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 2.3567280769348145, | |
| "learning_rate": 8.489416446663037e-06, | |
| "loss": 1.6288, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 2.1183040142059326, | |
| "learning_rate": 8.452100102858734e-06, | |
| "loss": 1.5929, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 2.467944860458374, | |
| "learning_rate": 8.414412656470297e-06, | |
| "loss": 1.5579, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 2.200274705886841, | |
| "learning_rate": 8.376358158918114e-06, | |
| "loss": 1.587, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 2.3279902935028076, | |
| "learning_rate": 8.33794070108077e-06, | |
| "loss": 1.6496, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 2.7403910160064697, | |
| "learning_rate": 8.299164412855268e-06, | |
| "loss": 1.5665, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 2.5097343921661377, | |
| "learning_rate": 8.260033462713073e-06, | |
| "loss": 1.6245, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 2.2657604217529297, | |
| "learning_rate": 8.22055205725199e-06, | |
| "loss": 1.5785, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 2.4038476943969727, | |
| "learning_rate": 8.180724440743957e-06, | |
| "loss": 1.6912, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 2.811049461364746, | |
| "learning_rate": 8.14055489467878e-06, | |
| "loss": 1.6059, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 2.474153518676758, | |
| "learning_rate": 8.100047737303877e-06, | |
| "loss": 1.6423, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 2.351736307144165, | |
| "learning_rate": 8.059207323160057e-06, | |
| "loss": 1.6199, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 2.3526997566223145, | |
| "learning_rate": 8.018038042613407e-06, | |
| "loss": 1.5856, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 2.075406789779663, | |
| "learning_rate": 7.97654432138333e-06, | |
| "loss": 1.5901, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 2.6973438262939453, | |
| "learning_rate": 7.93473062006677e-06, | |
| "loss": 1.5696, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 2.6449167728424072, | |
| "learning_rate": 7.892601433658705e-06, | |
| "loss": 1.5939, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 2.360849380493164, | |
| "learning_rate": 7.850161291068915e-06, | |
| "loss": 1.5449, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 2.467226028442383, | |
| "learning_rate": 7.807414754635145e-06, | |
| "loss": 1.5926, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 2.449989080429077, | |
| "learning_rate": 7.764366419632636e-06, | |
| "loss": 1.5591, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 5.518803119659424, | |
| "learning_rate": 7.721020913780137e-06, | |
| "loss": 1.5406, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 2.6841485500335693, | |
| "learning_rate": 7.677382896742417e-06, | |
| "loss": 1.5797, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 2.475085496902466, | |
| "learning_rate": 7.63345705962935e-06, | |
| "loss": 1.5653, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 2.3905832767486572, | |
| "learning_rate": 7.589248124491627e-06, | |
| "loss": 1.5641, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 2.7013638019561768, | |
| "learning_rate": 7.544760843813122e-06, | |
| "loss": 1.5837, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 2.500148296356201, | |
| "learning_rate": 7.500000000000001e-06, | |
| "loss": 1.6255, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 2.106887102127075, | |
| "learning_rate": 7.454970404866612e-06, | |
| "loss": 1.6069, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 2.1840200424194336, | |
| "learning_rate": 7.409676899118213e-06, | |
| "loss": 1.5959, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 5.758022308349609, | |
| "learning_rate": 7.3641243518305915e-06, | |
| "loss": 1.616, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 2.24625301361084, | |
| "learning_rate": 7.318317659926637e-06, | |
| "loss": 1.639, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 3.8651440143585205, | |
| "learning_rate": 7.272261747649922e-06, | |
| "loss": 1.6287, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 1.9262839555740356, | |
| "learning_rate": 7.225961566035335e-06, | |
| "loss": 1.5851, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 2.4312078952789307, | |
| "learning_rate": 7.179422092376856e-06, | |
| "loss": 1.5934, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 2.391693592071533, | |
| "learning_rate": 7.132648329692478e-06, | |
| "loss": 1.5719, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 2.10438871383667, | |
| "learning_rate": 7.085645306186391e-06, | |
| "loss": 1.5876, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 1.984710931777954, | |
| "learning_rate": 7.038418074708444e-06, | |
| "loss": 1.6506, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 2.4563817977905273, | |
| "learning_rate": 6.990971712210966e-06, | |
| "loss": 1.5578, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 1.856941819190979, | |
| "learning_rate": 6.943311319202976e-06, | |
| "loss": 1.5806, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 2.228983163833618, | |
| "learning_rate": 6.895442019201898e-06, | |
| "loss": 1.5639, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 2.9688351154327393, | |
| "learning_rate": 6.8473689581827585e-06, | |
| "loss": 1.5939, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 2.028876304626465, | |
| "learning_rate": 6.7990973040250055e-06, | |
| "loss": 1.6096, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 2.385794162750244, | |
| "learning_rate": 6.750632245956954e-06, | |
| "loss": 1.5617, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 2.0037996768951416, | |
| "learning_rate": 6.701978993997942e-06, | |
| "loss": 1.6028, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 2.0149664878845215, | |
| "learning_rate": 6.653142778398247e-06, | |
| "loss": 1.583, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 2.466585874557495, | |
| "learning_rate": 6.6041288490768385e-06, | |
| "loss": 1.6368, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 2.181319236755371, | |
| "learning_rate": 6.554942475057003e-06, | |
| "loss": 1.5819, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 1.9432387351989746, | |
| "learning_rate": 6.505588943899923e-06, | |
| "loss": 1.5551, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 2.277068614959717, | |
| "learning_rate": 6.456073561136261e-06, | |
| "loss": 1.5788, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 1.9018107652664185, | |
| "learning_rate": 6.406401649695814e-06, | |
| "loss": 1.5639, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 2.755958318710327, | |
| "learning_rate": 6.356578549335295e-06, | |
| "loss": 1.6015, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 2.2761142253875732, | |
| "learning_rate": 6.306609616064304e-06, | |
| "loss": 1.6054, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 3.212486505508423, | |
| "learning_rate": 6.256500221569556e-06, | |
| "loss": 1.5953, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 2.2175703048706055, | |
| "learning_rate": 6.2062557526374226e-06, | |
| "loss": 1.5745, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 1.9958215951919556, | |
| "learning_rate": 6.15588161057485e-06, | |
| "loss": 1.5981, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 2.4180896282196045, | |
| "learning_rate": 6.10538321062871e-06, | |
| "loss": 1.5836, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 2.0187716484069824, | |
| "learning_rate": 6.0547659814036664e-06, | |
| "loss": 1.6062, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 2.3086986541748047, | |
| "learning_rate": 6.004035364278593e-06, | |
| "loss": 1.5764, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 1.9949414730072021, | |
| "learning_rate": 5.953196812821622e-06, | |
| "loss": 1.6037, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 1.8632932901382446, | |
| "learning_rate": 5.902255792203882e-06, | |
| "loss": 1.6051, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 2.034745931625366, | |
| "learning_rate": 5.851217778611994e-06, | |
| "loss": 1.6386, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 2.1993603706359863, | |
| "learning_rate": 5.800088258659371e-06, | |
| "loss": 1.6081, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 2.0585718154907227, | |
| "learning_rate": 5.748872728796409e-06, | |
| "loss": 1.5955, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 2.2753782272338867, | |
| "learning_rate": 5.697576694719616e-06, | |
| "loss": 1.5655, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 3.237483024597168, | |
| "learning_rate": 5.646205670779745e-06, | |
| "loss": 1.6004, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 3.019814968109131, | |
| "learning_rate": 5.594765179389003e-06, | |
| "loss": 1.5721, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 2.3698084354400635, | |
| "learning_rate": 5.543260750427373e-06, | |
| "loss": 1.5946, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 2.441636085510254, | |
| "learning_rate": 5.4916979206481745e-06, | |
| "loss": 1.5522, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 2.149003267288208, | |
| "learning_rate": 5.440082233082837e-06, | |
| "loss": 1.5935, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 2.03369402885437, | |
| "learning_rate": 5.388419236445033e-06, | |
| "loss": 1.5912, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 1.9619389772415161, | |
| "learning_rate": 5.336714484534183e-06, | |
| "loss": 1.5324, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 3.591277837753296, | |
| "learning_rate": 5.284973535638424e-06, | |
| "loss": 1.5662, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 2.212984323501587, | |
| "learning_rate": 5.233201951937088e-06, | |
| "loss": 1.6589, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 1.835900902748108, | |
| "learning_rate": 5.181405298902763e-06, | |
| "loss": 1.597, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 2.046597480773926, | |
| "learning_rate": 5.1295891447030056e-06, | |
| "loss": 1.5742, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 2.5161941051483154, | |
| "learning_rate": 5.077759059601756e-06, | |
| "loss": 1.5848, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 2.1808998584747314, | |
| "learning_rate": 5.025920615360532e-06, | |
| "loss": 1.5684, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 2.5023722648620605, | |
| "learning_rate": 4.974079384639469e-06, | |
| "loss": 1.5815, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 2.056931734085083, | |
| "learning_rate": 4.922240940398246e-06, | |
| "loss": 1.5334, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 2.1214613914489746, | |
| "learning_rate": 4.870410855296994e-06, | |
| "loss": 1.5726, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 3.080734968185425, | |
| "learning_rate": 4.818594701097239e-06, | |
| "loss": 1.5869, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 2.13873028755188, | |
| "learning_rate": 4.766798048062913e-06, | |
| "loss": 1.6214, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 2.4381258487701416, | |
| "learning_rate": 4.715026464361576e-06, | |
| "loss": 1.5921, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 2.3654768466949463, | |
| "learning_rate": 4.663285515465818e-06, | |
| "loss": 1.5598, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 2.3201935291290283, | |
| "learning_rate": 4.611580763554969e-06, | |
| "loss": 1.5339, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 2.3482742309570312, | |
| "learning_rate": 4.559917766917166e-06, | |
| "loss": 1.5501, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 2.296869993209839, | |
| "learning_rate": 4.508302079351827e-06, | |
| "loss": 1.542, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 2.0012331008911133, | |
| "learning_rate": 4.456739249572628e-06, | |
| "loss": 1.5223, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 2.4259610176086426, | |
| "learning_rate": 4.405234820611001e-06, | |
| "loss": 1.5929, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 2.1162846088409424, | |
| "learning_rate": 4.3537943292202555e-06, | |
| "loss": 1.5213, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 4.398637771606445, | |
| "learning_rate": 4.3024233052803855e-06, | |
| "loss": 1.5698, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 2.08495831489563, | |
| "learning_rate": 4.251127271203593e-06, | |
| "loss": 1.6086, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 5.053515911102295, | |
| "learning_rate": 4.199911741340631e-06, | |
| "loss": 1.6036, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 3.257359266281128, | |
| "learning_rate": 4.148782221388007e-06, | |
| "loss": 1.6246, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 2.1245572566986084, | |
| "learning_rate": 4.097744207796119e-06, | |
| "loss": 1.5544, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 2.556488275527954, | |
| "learning_rate": 4.04680318717838e-06, | |
| "loss": 1.5516, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 2.438620090484619, | |
| "learning_rate": 3.995964635721409e-06, | |
| "loss": 1.5524, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 1.9453068971633911, | |
| "learning_rate": 3.945234018596335e-06, | |
| "loss": 1.5657, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 2.2514045238494873, | |
| "learning_rate": 3.8946167893712916e-06, | |
| "loss": 1.5916, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 2.2144711017608643, | |
| "learning_rate": 3.844118389425154e-06, | |
| "loss": 1.6051, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 2.4618723392486572, | |
| "learning_rate": 3.7937442473625787e-06, | |
| "loss": 1.5841, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 2.2561230659484863, | |
| "learning_rate": 3.743499778430445e-06, | |
| "loss": 1.5469, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 2.2580950260162354, | |
| "learning_rate": 3.6933903839356983e-06, | |
| "loss": 1.5618, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 1.8674182891845703, | |
| "learning_rate": 3.6434214506647064e-06, | |
| "loss": 1.5266, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 2.33337664604187, | |
| "learning_rate": 3.5935983503041864e-06, | |
| "loss": 1.5223, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 1.912142038345337, | |
| "learning_rate": 3.5439264388637407e-06, | |
| "loss": 1.5331, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 2.8374595642089844, | |
| "learning_rate": 3.4944110561000785e-06, | |
| "loss": 1.6068, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 2.329653024673462, | |
| "learning_rate": 3.4450575249429975e-06, | |
| "loss": 1.5324, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 2.3587117195129395, | |
| "learning_rate": 3.3958711509231627e-06, | |
| "loss": 1.5542, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 2.657672643661499, | |
| "learning_rate": 3.3468572216017536e-06, | |
| "loss": 1.5755, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 2.17305064201355, | |
| "learning_rate": 3.29802100600206e-06, | |
| "loss": 1.5489, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 2.5755293369293213, | |
| "learning_rate": 3.249367754043047e-06, | |
| "loss": 1.5835, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 3.943556070327759, | |
| "learning_rate": 3.200902695974995e-06, | |
| "loss": 1.6003, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 2.425208568572998, | |
| "learning_rate": 3.152631041817244e-06, | |
| "loss": 1.5649, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 2.0712738037109375, | |
| "learning_rate": 3.104557980798104e-06, | |
| "loss": 1.5363, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 2.3942067623138428, | |
| "learning_rate": 3.056688680797024e-06, | |
| "loss": 1.5522, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 2.139435052871704, | |
| "learning_rate": 3.0090282877890376e-06, | |
| "loss": 1.5759, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 2.4167027473449707, | |
| "learning_rate": 2.961581925291557e-06, | |
| "loss": 1.5254, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 1.9939026832580566, | |
| "learning_rate": 2.9143546938136093e-06, | |
| "loss": 1.5688, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 1.8560909032821655, | |
| "learning_rate": 2.8673516703075247e-06, | |
| "loss": 1.5346, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 4.581928730010986, | |
| "learning_rate": 2.820577907623145e-06, | |
| "loss": 1.5447, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 2.128434419631958, | |
| "learning_rate": 2.7740384339646655e-06, | |
| "loss": 1.5557, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 2.027505397796631, | |
| "learning_rate": 2.7277382523500804e-06, | |
| "loss": 1.527, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 2.5379273891448975, | |
| "learning_rate": 2.6816823400733628e-06, | |
| "loss": 1.5794, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 6.262667655944824, | |
| "learning_rate": 2.6358756481694115e-06, | |
| "loss": 1.5452, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 2.3182244300842285, | |
| "learning_rate": 2.5903231008817888e-06, | |
| "loss": 1.5354, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 2.3638322353363037, | |
| "learning_rate": 2.5450295951333896e-06, | |
| "loss": 1.5699, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 2.2807657718658447, | |
| "learning_rate": 2.5000000000000015e-06, | |
| "loss": 1.5958, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 2.4435348510742188, | |
| "learning_rate": 2.4552391561868783e-06, | |
| "loss": 1.5601, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 2.2643187046051025, | |
| "learning_rate": 2.410751875508373e-06, | |
| "loss": 1.5474, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 2.2094953060150146, | |
| "learning_rate": 2.3665429403706506e-06, | |
| "loss": 1.5888, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 2.147650718688965, | |
| "learning_rate": 2.3226171032575856e-06, | |
| "loss": 1.5937, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 1.9609376192092896, | |
| "learning_rate": 2.278979086219863e-06, | |
| "loss": 1.5873, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 2.404101610183716, | |
| "learning_rate": 2.2356335803673655e-06, | |
| "loss": 1.5707, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 2.1139962673187256, | |
| "learning_rate": 2.192585245364856e-06, | |
| "loss": 1.5839, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 2.373765707015991, | |
| "learning_rate": 2.149838708931087e-06, | |
| "loss": 1.5289, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 2.1613967418670654, | |
| "learning_rate": 2.1073985663412984e-06, | |
| "loss": 1.5599, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 2.6613214015960693, | |
| "learning_rate": 2.0652693799332286e-06, | |
| "loss": 1.5657, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 1.8200939893722534, | |
| "learning_rate": 2.0234556786166715e-06, | |
| "loss": 1.4937, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 1.8578108549118042, | |
| "learning_rate": 1.9819619573865932e-06, | |
| "loss": 1.5268, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 2.355733633041382, | |
| "learning_rate": 1.9407926768399456e-06, | |
| "loss": 1.5658, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 2.7907326221466064, | |
| "learning_rate": 1.8999522626961254e-06, | |
| "loss": 1.5777, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 2.2081241607666016, | |
| "learning_rate": 1.859445105321221e-06, | |
| "loss": 1.6039, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 2.7307095527648926, | |
| "learning_rate": 1.8192755592560446e-06, | |
| "loss": 1.554, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 2.100717306137085, | |
| "learning_rate": 1.7794479427480115e-06, | |
| "loss": 1.5687, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 2.900230884552002, | |
| "learning_rate": 1.739966537286929e-06, | |
| "loss": 1.5894, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 2.7056214809417725, | |
| "learning_rate": 1.7008355871447345e-06, | |
| "loss": 1.5262, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 2.2790305614471436, | |
| "learning_rate": 1.6620592989192318e-06, | |
| "loss": 1.5334, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 2.336102247238159, | |
| "learning_rate": 1.6236418410818872e-06, | |
| "loss": 1.6231, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 2.4769365787506104, | |
| "learning_rate": 1.5855873435297042e-06, | |
| "loss": 1.5542, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 2.2314493656158447, | |
| "learning_rate": 1.5478998971412669e-06, | |
| "loss": 1.582, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 2.37534236907959, | |
| "learning_rate": 1.510583553336964e-06, | |
| "loss": 1.5618, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 2.2861831188201904, | |
| "learning_rate": 1.473642323643465e-06, | |
| "loss": 1.5557, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 2.0637292861938477, | |
| "learning_rate": 1.4370801792624656e-06, | |
| "loss": 1.5812, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 2.4701573848724365, | |
| "learning_rate": 1.4009010506437997e-06, | |
| "loss": 1.5535, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 2.5278735160827637, | |
| "learning_rate": 1.3651088270628992e-06, | |
| "loss": 1.5792, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 2.0601749420166016, | |
| "learning_rate": 1.3297073562026992e-06, | |
| "loss": 1.5739, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 2.125802755355835, | |
| "learning_rate": 1.2947004437400161e-06, | |
| "loss": 1.579, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 2.429272413253784, | |
| "learning_rate": 1.2600918529364253e-06, | |
| "loss": 1.5575, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 2.331495761871338, | |
| "learning_rate": 1.225885304233716e-06, | |
| "loss": 1.5796, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 2.5032925605773926, | |
| "learning_rate": 1.1920844748539373e-06, | |
| "loss": 1.5205, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 2.1888771057128906, | |
| "learning_rate": 1.1586929984040974e-06, | |
| "loss": 1.5877, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 2.270721673965454, | |
| "learning_rate": 1.125714464485551e-06, | |
| "loss": 1.5382, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 3.260922431945801, | |
| "learning_rate": 1.0931524183081105e-06, | |
| "loss": 1.585, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 2.52512526512146, | |
| "learning_rate": 1.0610103603089345e-06, | |
| "loss": 1.5353, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 2.214529275894165, | |
| "learning_rate": 1.0292917457762325e-06, | |
| "loss": 1.6081, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 3.466500997543335, | |
| "learning_rate": 9.979999844778203e-07, | |
| "loss": 1.5661, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 2.2533483505249023, | |
| "learning_rate": 9.671384402945588e-07, | |
| "loss": 1.5718, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 1.9847800731658936, | |
| "learning_rate": 9.367104308587493e-07, | |
| "loss": 1.561, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 2.6065824031829834, | |
| "learning_rate": 9.06719227197474e-07, | |
| "loss": 1.5482, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 2.5435142517089844, | |
| "learning_rate": 8.771680533809634e-07, | |
| "loss": 1.5918, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 2.8875584602355957, | |
| "learning_rate": 8.480600861760124e-07, | |
| "loss": 1.6082, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 2.1082193851470947, | |
| "learning_rate": 8.193984547044659e-07, | |
| "loss": 1.594, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 2.4410855770111084, | |
| "learning_rate": 7.911862401068431e-07, | |
| "loss": 1.5938, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 2.8396708965301514, | |
| "learning_rate": 7.634264752111131e-07, | |
| "loss": 1.6215, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 2.493769407272339, | |
| "learning_rate": 7.361221442066607e-07, | |
| "loss": 1.5734, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 1.8893070220947266, | |
| "learning_rate": 7.092761823234911e-07, | |
| "loss": 1.5264, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 2.5776591300964355, | |
| "learning_rate": 6.828914755166826e-07, | |
| "loss": 1.5733, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 2.7408130168914795, | |
| "learning_rate": 6.569708601561515e-07, | |
| "loss": 1.5617, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 2.9031057357788086, | |
| "learning_rate": 6.315171227217365e-07, | |
| "loss": 1.5979, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 2.713923931121826, | |
| "learning_rate": 6.065329995036573e-07, | |
| "loss": 1.5364, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 2.8765623569488525, | |
| "learning_rate": 5.820211763083494e-07, | |
| "loss": 1.5547, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 2.5959739685058594, | |
| "learning_rate": 5.579842881697556e-07, | |
| "loss": 1.5372, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 2.1984755992889404, | |
| "learning_rate": 5.344249190660427e-07, | |
| "loss": 1.6001, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 2.3963303565979004, | |
| "learning_rate": 5.113456016418305e-07, | |
| "loss": 1.6344, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 4.728706359863281, | |
| "learning_rate": 4.88748816935934e-07, | |
| "loss": 1.5381, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 2.736025333404541, | |
| "learning_rate": 4.666369941146376e-07, | |
| "loss": 1.5458, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 3.1046762466430664, | |
| "learning_rate": 4.4501251021057566e-07, | |
| "loss": 1.5871, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 3.1005470752716064, | |
| "learning_rate": 4.2387768986718644e-07, | |
| "loss": 1.545, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 5.466526985168457, | |
| "learning_rate": 4.03234805088818e-07, | |
| "loss": 1.5337, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 2.6507673263549805, | |
| "learning_rate": 3.8308607499648765e-07, | |
| "loss": 1.5436, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 2.5536324977874756, | |
| "learning_rate": 3.634336655893189e-07, | |
| "loss": 1.5853, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 3.4438483715057373, | |
| "learning_rate": 3.4427968951170287e-07, | |
| "loss": 1.5655, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 2.799037456512451, | |
| "learning_rate": 3.256262058261816e-07, | |
| "loss": 1.6025, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 2.272508144378662, | |
| "learning_rate": 3.0747521979210436e-07, | |
| "loss": 1.5339, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 2.6367502212524414, | |
| "learning_rate": 2.8982868265005457e-07, | |
| "loss": 1.487, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 2.486656665802002, | |
| "learning_rate": 2.726884914120936e-07, | |
| "loss": 1.5738, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 2.579956293106079, | |
| "learning_rate": 2.5605648865783315e-07, | |
| "loss": 1.5955, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 2.506580114364624, | |
| "learning_rate": 2.399344623363503e-07, | |
| "loss": 1.5778, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 2.4960708618164062, | |
| "learning_rate": 2.2432414557399197e-07, | |
| "loss": 1.5922, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 3.1219966411590576, | |
| "learning_rate": 2.0922721648805045e-07, | |
| "loss": 1.5548, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 2.500851631164551, | |
| "learning_rate": 1.9464529800637731e-07, | |
| "loss": 1.5491, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 2.394024133682251, | |
| "learning_rate": 1.805799576929107e-07, | |
| "loss": 1.5433, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 2.6824209690093994, | |
| "learning_rate": 1.6703270757916e-07, | |
| "loss": 1.5465, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 2.4885001182556152, | |
| "learning_rate": 1.540050040016694e-07, | |
| "loss": 1.5682, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 4.021599769592285, | |
| "learning_rate": 1.414982474454524e-07, | |
| "loss": 1.5516, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 2.3498406410217285, | |
| "learning_rate": 1.2951378239344337e-07, | |
| "loss": 1.5837, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 2.3095269203186035, | |
| "learning_rate": 1.1805289718196499e-07, | |
| "loss": 1.5519, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 2.113964080810547, | |
| "learning_rate": 1.0711682386222943e-07, | |
| "loss": 1.5897, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 2.6248152256011963, | |
| "learning_rate": 9.670673806789543e-08, | |
| "loss": 1.5596, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 2.7190983295440674, | |
| "learning_rate": 8.682375888868167e-08, | |
| "loss": 1.5829, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 2.4610188007354736, | |
| "learning_rate": 7.746894875007016e-08, | |
| "loss": 1.6058, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 3.32197904586792, | |
| "learning_rate": 6.864331329909102e-08, | |
| "loss": 1.6013, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 3.1023738384246826, | |
| "learning_rate": 6.034780129621664e-08, | |
| "loss": 1.5767, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 2.545488119125366, | |
| "learning_rate": 5.258330451336724e-08, | |
| "loss": 1.5267, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 2.106576442718506, | |
| "learning_rate": 4.535065763804802e-08, | |
| "loss": 1.5447, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 2.0030553340911865, | |
| "learning_rate": 3.8650638183617695e-08, | |
| "loss": 1.5714, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 2.5803191661834717, | |
| "learning_rate": 3.248396640570528e-08, | |
| "loss": 1.5273, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 2.743903160095215, | |
| "learning_rate": 2.685130522478485e-08, | |
| "loss": 1.5455, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 2.4540319442749023, | |
| "learning_rate": 2.1753260154906973e-08, | |
| "loss": 1.5372, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 2.002195358276367, | |
| "learning_rate": 1.7190379238609666e-08, | |
| "loss": 1.5892, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 2.3258426189422607, | |
| "learning_rate": 1.3163152988000527e-08, | |
| "loss": 1.5832, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 3.36910343170166, | |
| "learning_rate": 9.672014332028357e-09, | |
| "loss": 1.5319, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 2.2311582565307617, | |
| "learning_rate": 6.717338569942611e-09, | |
| "loss": 1.5475, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 3.657569646835327, | |
| "learning_rate": 4.299443330947895e-09, | |
| "loss": 1.5227, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 2.1514527797698975, | |
| "learning_rate": 2.4185885400596076e-09, | |
| "loss": 1.5543, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 2.7922263145446777, | |
| "learning_rate": 1.0749763901607425e-09, | |
| "loss": 1.575, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 2.211779832839966, | |
| "learning_rate": 2.6875132026760173e-10, | |
| "loss": 1.5689, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 2.225865125656128, | |
| "learning_rate": 0.0, | |
| "loss": 1.5786, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 313, | |
| "total_flos": 399880339259392.0, | |
| "train_loss": 1.6020259019284964, | |
| "train_runtime": 68378.6303, | |
| "train_samples_per_second": 0.588, | |
| "train_steps_per_second": 0.005 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 313, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 50, | |
| "total_flos": 399880339259392.0, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |