{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.989690721649485, "eval_steps": 500, "global_step": 1815, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0027491408934707906, "grad_norm": 6.006607135034925, "learning_rate": 4.395604395604396e-07, "loss": 1.1009, "step": 1 }, { "epoch": 0.005498281786941581, "grad_norm": 6.018597667128207, "learning_rate": 8.791208791208792e-07, "loss": 1.1005, "step": 2 }, { "epoch": 0.008247422680412371, "grad_norm": 6.073361462181881, "learning_rate": 1.3186813186813187e-06, "loss": 1.1055, "step": 3 }, { "epoch": 0.010996563573883162, "grad_norm": 5.961879860389406, "learning_rate": 1.7582417582417585e-06, "loss": 1.1099, "step": 4 }, { "epoch": 0.013745704467353952, "grad_norm": 5.617774190916827, "learning_rate": 2.197802197802198e-06, "loss": 1.081, "step": 5 }, { "epoch": 0.016494845360824743, "grad_norm": 4.531649230749474, "learning_rate": 2.6373626373626375e-06, "loss": 1.0487, "step": 6 }, { "epoch": 0.019243986254295534, "grad_norm": 4.217358609239622, "learning_rate": 3.0769230769230774e-06, "loss": 1.0556, "step": 7 }, { "epoch": 0.021993127147766325, "grad_norm": 2.3655606096533908, "learning_rate": 3.516483516483517e-06, "loss": 1.0049, "step": 8 }, { "epoch": 0.024742268041237112, "grad_norm": 2.0449739951596695, "learning_rate": 3.9560439560439565e-06, "loss": 0.9803, "step": 9 }, { "epoch": 0.027491408934707903, "grad_norm": 2.8045358100308406, "learning_rate": 4.395604395604396e-06, "loss": 0.9589, "step": 10 }, { "epoch": 0.030240549828178694, "grad_norm": 3.662777544001548, "learning_rate": 4.8351648351648355e-06, "loss": 0.9677, "step": 11 }, { "epoch": 0.032989690721649485, "grad_norm": 3.5905401528260374, "learning_rate": 5.274725274725275e-06, "loss": 0.96, "step": 12 }, { "epoch": 0.03573883161512027, "grad_norm": 3.3209223305008964, "learning_rate": 5.7142857142857145e-06, "loss": 0.9481, "step": 13 }, { "epoch": 0.03848797250859107, "grad_norm": 2.3880367341332343, "learning_rate": 6.153846153846155e-06, "loss": 0.9057, "step": 14 }, { "epoch": 0.041237113402061855, "grad_norm": 2.269345734711135, "learning_rate": 6.5934065934065935e-06, "loss": 0.9048, "step": 15 }, { "epoch": 0.04398625429553265, "grad_norm": 1.7351295951376182, "learning_rate": 7.032967032967034e-06, "loss": 0.8869, "step": 16 }, { "epoch": 0.04673539518900344, "grad_norm": 1.5080912194950085, "learning_rate": 7.472527472527473e-06, "loss": 0.8503, "step": 17 }, { "epoch": 0.049484536082474224, "grad_norm": 1.4431136736456789, "learning_rate": 7.912087912087913e-06, "loss": 0.8405, "step": 18 }, { "epoch": 0.05223367697594502, "grad_norm": 1.4712422848105347, "learning_rate": 8.351648351648353e-06, "loss": 0.8274, "step": 19 }, { "epoch": 0.054982817869415807, "grad_norm": 1.3178930536171367, "learning_rate": 8.791208791208792e-06, "loss": 0.8428, "step": 20 }, { "epoch": 0.0577319587628866, "grad_norm": 1.2703340647552794, "learning_rate": 9.230769230769232e-06, "loss": 0.8211, "step": 21 }, { "epoch": 0.06048109965635739, "grad_norm": 1.2871317510857054, "learning_rate": 9.670329670329671e-06, "loss": 0.8053, "step": 22 }, { "epoch": 0.06323024054982818, "grad_norm": 1.1237165728026322, "learning_rate": 1.010989010989011e-05, "loss": 0.8101, "step": 23 }, { "epoch": 0.06597938144329897, "grad_norm": 0.9168689840127058, "learning_rate": 1.054945054945055e-05, "loss": 0.7837, "step": 24 }, { "epoch": 0.06872852233676977, "grad_norm": 0.9355697137818322, "learning_rate": 1.098901098901099e-05, "loss": 0.79, "step": 25 }, { "epoch": 0.07147766323024055, "grad_norm": 0.8919194529855764, "learning_rate": 1.1428571428571429e-05, "loss": 0.7927, "step": 26 }, { "epoch": 0.07422680412371134, "grad_norm": 0.8013095068191387, "learning_rate": 1.186813186813187e-05, "loss": 0.7826, "step": 27 }, { "epoch": 0.07697594501718213, "grad_norm": 0.7074089625670721, "learning_rate": 1.230769230769231e-05, "loss": 0.77, "step": 28 }, { "epoch": 0.07972508591065292, "grad_norm": 0.5939851673949501, "learning_rate": 1.2747252747252747e-05, "loss": 0.764, "step": 29 }, { "epoch": 0.08247422680412371, "grad_norm": 0.7291769706418423, "learning_rate": 1.3186813186813187e-05, "loss": 0.7667, "step": 30 }, { "epoch": 0.0852233676975945, "grad_norm": 0.5476356706171729, "learning_rate": 1.3626373626373627e-05, "loss": 0.7669, "step": 31 }, { "epoch": 0.0879725085910653, "grad_norm": 0.496258207605204, "learning_rate": 1.4065934065934068e-05, "loss": 0.7617, "step": 32 }, { "epoch": 0.09072164948453608, "grad_norm": 0.5382769338196303, "learning_rate": 1.4505494505494506e-05, "loss": 0.7554, "step": 33 }, { "epoch": 0.09347079037800687, "grad_norm": 0.4580628373124881, "learning_rate": 1.4945054945054947e-05, "loss": 0.75, "step": 34 }, { "epoch": 0.09621993127147767, "grad_norm": 0.40847997114284523, "learning_rate": 1.5384615384615387e-05, "loss": 0.7467, "step": 35 }, { "epoch": 0.09896907216494845, "grad_norm": 0.5031600722774845, "learning_rate": 1.5824175824175826e-05, "loss": 0.7383, "step": 36 }, { "epoch": 0.10171821305841924, "grad_norm": 0.3529000118564782, "learning_rate": 1.6263736263736265e-05, "loss": 0.7524, "step": 37 }, { "epoch": 0.10446735395189004, "grad_norm": 0.40270165373486655, "learning_rate": 1.6703296703296707e-05, "loss": 0.7371, "step": 38 }, { "epoch": 0.10721649484536082, "grad_norm": 0.3743730526645748, "learning_rate": 1.7142857142857142e-05, "loss": 0.7332, "step": 39 }, { "epoch": 0.10996563573883161, "grad_norm": 0.3802324817797124, "learning_rate": 1.7582417582417584e-05, "loss": 0.7301, "step": 40 }, { "epoch": 0.11271477663230241, "grad_norm": 0.38122013544181327, "learning_rate": 1.8021978021978023e-05, "loss": 0.7481, "step": 41 }, { "epoch": 0.1154639175257732, "grad_norm": 0.3807629715206352, "learning_rate": 1.8461538461538465e-05, "loss": 0.7234, "step": 42 }, { "epoch": 0.11821305841924398, "grad_norm": 0.3690276896533298, "learning_rate": 1.8901098901098903e-05, "loss": 0.7382, "step": 43 }, { "epoch": 0.12096219931271478, "grad_norm": 0.39354181368093777, "learning_rate": 1.9340659340659342e-05, "loss": 0.7274, "step": 44 }, { "epoch": 0.12371134020618557, "grad_norm": 0.45930210506482494, "learning_rate": 1.9780219780219784e-05, "loss": 0.7222, "step": 45 }, { "epoch": 0.12646048109965635, "grad_norm": 0.3257242749282543, "learning_rate": 2.021978021978022e-05, "loss": 0.7062, "step": 46 }, { "epoch": 0.12920962199312716, "grad_norm": 0.4894322683246899, "learning_rate": 2.0659340659340665e-05, "loss": 0.7273, "step": 47 }, { "epoch": 0.13195876288659794, "grad_norm": 0.4217505425727167, "learning_rate": 2.10989010989011e-05, "loss": 0.7144, "step": 48 }, { "epoch": 0.13470790378006872, "grad_norm": 0.34926375879792276, "learning_rate": 2.153846153846154e-05, "loss": 0.7284, "step": 49 }, { "epoch": 0.13745704467353953, "grad_norm": 0.37706775972489026, "learning_rate": 2.197802197802198e-05, "loss": 0.7197, "step": 50 }, { "epoch": 0.1402061855670103, "grad_norm": 0.4461248020784725, "learning_rate": 2.241758241758242e-05, "loss": 0.7168, "step": 51 }, { "epoch": 0.1429553264604811, "grad_norm": 0.45683903751388155, "learning_rate": 2.2857142857142858e-05, "loss": 0.7047, "step": 52 }, { "epoch": 0.1457044673539519, "grad_norm": 0.4941785029111688, "learning_rate": 2.32967032967033e-05, "loss": 0.7203, "step": 53 }, { "epoch": 0.14845360824742268, "grad_norm": 0.7101247658482969, "learning_rate": 2.373626373626374e-05, "loss": 0.7132, "step": 54 }, { "epoch": 0.15120274914089346, "grad_norm": 1.3845226821923962, "learning_rate": 2.4175824175824177e-05, "loss": 0.7139, "step": 55 }, { "epoch": 0.15395189003436427, "grad_norm": 0.9145522126443197, "learning_rate": 2.461538461538462e-05, "loss": 0.7201, "step": 56 }, { "epoch": 0.15670103092783505, "grad_norm": 0.7316554124403013, "learning_rate": 2.5054945054945058e-05, "loss": 0.6997, "step": 57 }, { "epoch": 0.15945017182130583, "grad_norm": 1.525604886016204, "learning_rate": 2.5494505494505493e-05, "loss": 0.708, "step": 58 }, { "epoch": 0.16219931271477664, "grad_norm": 0.9569983399077112, "learning_rate": 2.593406593406594e-05, "loss": 0.7023, "step": 59 }, { "epoch": 0.16494845360824742, "grad_norm": 1.222421966279637, "learning_rate": 2.6373626373626374e-05, "loss": 0.699, "step": 60 }, { "epoch": 0.1676975945017182, "grad_norm": 1.1284003579139785, "learning_rate": 2.6813186813186813e-05, "loss": 0.6944, "step": 61 }, { "epoch": 0.170446735395189, "grad_norm": 1.323080634931126, "learning_rate": 2.7252747252747255e-05, "loss": 0.6897, "step": 62 }, { "epoch": 0.1731958762886598, "grad_norm": 0.9918797937768813, "learning_rate": 2.7692307692307694e-05, "loss": 0.6937, "step": 63 }, { "epoch": 0.1759450171821306, "grad_norm": 1.4485583433380003, "learning_rate": 2.8131868131868136e-05, "loss": 0.6924, "step": 64 }, { "epoch": 0.17869415807560138, "grad_norm": 0.9586990948210379, "learning_rate": 2.8571428571428574e-05, "loss": 0.7036, "step": 65 }, { "epoch": 0.18144329896907216, "grad_norm": 1.5401125351907554, "learning_rate": 2.9010989010989013e-05, "loss": 0.7067, "step": 66 }, { "epoch": 0.18419243986254297, "grad_norm": 1.149941673251316, "learning_rate": 2.9450549450549455e-05, "loss": 0.6846, "step": 67 }, { "epoch": 0.18694158075601375, "grad_norm": 1.398728105276103, "learning_rate": 2.9890109890109894e-05, "loss": 0.7001, "step": 68 }, { "epoch": 0.18969072164948453, "grad_norm": 1.3380685614311483, "learning_rate": 3.0329670329670332e-05, "loss": 0.7017, "step": 69 }, { "epoch": 0.19243986254295534, "grad_norm": 0.966967948159332, "learning_rate": 3.0769230769230774e-05, "loss": 0.6945, "step": 70 }, { "epoch": 0.19518900343642612, "grad_norm": 1.2949156130535089, "learning_rate": 3.120879120879121e-05, "loss": 0.6983, "step": 71 }, { "epoch": 0.1979381443298969, "grad_norm": 0.9121501279786434, "learning_rate": 3.164835164835165e-05, "loss": 0.6944, "step": 72 }, { "epoch": 0.2006872852233677, "grad_norm": 0.7744380611271706, "learning_rate": 3.2087912087912094e-05, "loss": 0.6797, "step": 73 }, { "epoch": 0.2034364261168385, "grad_norm": 1.171893254170078, "learning_rate": 3.252747252747253e-05, "loss": 0.6916, "step": 74 }, { "epoch": 0.20618556701030927, "grad_norm": 1.432067598703686, "learning_rate": 3.296703296703297e-05, "loss": 0.696, "step": 75 }, { "epoch": 0.20893470790378008, "grad_norm": 2.015800500571526, "learning_rate": 3.340659340659341e-05, "loss": 0.6946, "step": 76 }, { "epoch": 0.21168384879725086, "grad_norm": 1.1450267893677495, "learning_rate": 3.384615384615385e-05, "loss": 0.6938, "step": 77 }, { "epoch": 0.21443298969072164, "grad_norm": 3.012444154989314, "learning_rate": 3.4285714285714284e-05, "loss": 0.708, "step": 78 }, { "epoch": 0.21718213058419245, "grad_norm": 2.938842593069413, "learning_rate": 3.4725274725274726e-05, "loss": 0.7043, "step": 79 }, { "epoch": 0.21993127147766323, "grad_norm": 1.2620210326118846, "learning_rate": 3.516483516483517e-05, "loss": 0.7015, "step": 80 }, { "epoch": 0.22268041237113403, "grad_norm": 1.8136600496263213, "learning_rate": 3.56043956043956e-05, "loss": 0.6934, "step": 81 }, { "epoch": 0.22542955326460482, "grad_norm": 1.4037155277378568, "learning_rate": 3.6043956043956045e-05, "loss": 0.6923, "step": 82 }, { "epoch": 0.2281786941580756, "grad_norm": 1.3955889629454836, "learning_rate": 3.648351648351649e-05, "loss": 0.6932, "step": 83 }, { "epoch": 0.2309278350515464, "grad_norm": 1.4401203218805492, "learning_rate": 3.692307692307693e-05, "loss": 0.6837, "step": 84 }, { "epoch": 0.23367697594501718, "grad_norm": 1.1506286717804464, "learning_rate": 3.7362637362637365e-05, "loss": 0.6824, "step": 85 }, { "epoch": 0.23642611683848797, "grad_norm": 1.4015178509079977, "learning_rate": 3.7802197802197807e-05, "loss": 0.6818, "step": 86 }, { "epoch": 0.23917525773195877, "grad_norm": 0.8632967007668437, "learning_rate": 3.824175824175825e-05, "loss": 0.6881, "step": 87 }, { "epoch": 0.24192439862542955, "grad_norm": 1.7108832289817475, "learning_rate": 3.8681318681318684e-05, "loss": 0.694, "step": 88 }, { "epoch": 0.24467353951890033, "grad_norm": 1.2330487896649007, "learning_rate": 3.9120879120879126e-05, "loss": 0.682, "step": 89 }, { "epoch": 0.24742268041237114, "grad_norm": 1.2807204902880718, "learning_rate": 3.956043956043957e-05, "loss": 0.6851, "step": 90 }, { "epoch": 0.2501718213058419, "grad_norm": 1.327068957190879, "learning_rate": 4e-05, "loss": 0.6807, "step": 91 }, { "epoch": 0.2529209621993127, "grad_norm": 1.0022125576592875, "learning_rate": 4.043956043956044e-05, "loss": 0.6782, "step": 92 }, { "epoch": 0.2556701030927835, "grad_norm": 1.3221553663136452, "learning_rate": 4.087912087912088e-05, "loss": 0.6759, "step": 93 }, { "epoch": 0.2584192439862543, "grad_norm": 0.9568766990099716, "learning_rate": 4.131868131868133e-05, "loss": 0.6873, "step": 94 }, { "epoch": 0.2611683848797251, "grad_norm": 1.7172395359463215, "learning_rate": 4.1758241758241765e-05, "loss": 0.6814, "step": 95 }, { "epoch": 0.2639175257731959, "grad_norm": 1.5142722720152748, "learning_rate": 4.21978021978022e-05, "loss": 0.682, "step": 96 }, { "epoch": 0.26666666666666666, "grad_norm": 0.8979115368433407, "learning_rate": 4.263736263736264e-05, "loss": 0.6828, "step": 97 }, { "epoch": 0.26941580756013744, "grad_norm": 1.3321313501935168, "learning_rate": 4.307692307692308e-05, "loss": 0.687, "step": 98 }, { "epoch": 0.2721649484536082, "grad_norm": 1.6997841277714232, "learning_rate": 4.351648351648352e-05, "loss": 0.6759, "step": 99 }, { "epoch": 0.27491408934707906, "grad_norm": 1.0580779198597339, "learning_rate": 4.395604395604396e-05, "loss": 0.6743, "step": 100 }, { "epoch": 0.27766323024054984, "grad_norm": 2.3143610998320896, "learning_rate": 4.4395604395604403e-05, "loss": 0.6684, "step": 101 }, { "epoch": 0.2804123711340206, "grad_norm": 1.370198490883026, "learning_rate": 4.483516483516484e-05, "loss": 0.6643, "step": 102 }, { "epoch": 0.2831615120274914, "grad_norm": 2.7976008957802088, "learning_rate": 4.527472527472528e-05, "loss": 0.6878, "step": 103 }, { "epoch": 0.2859106529209622, "grad_norm": 2.539570133020213, "learning_rate": 4.5714285714285716e-05, "loss": 0.6889, "step": 104 }, { "epoch": 0.28865979381443296, "grad_norm": 1.7701837403812288, "learning_rate": 4.615384615384615e-05, "loss": 0.6794, "step": 105 }, { "epoch": 0.2914089347079038, "grad_norm": 1.6244664002638955, "learning_rate": 4.65934065934066e-05, "loss": 0.6798, "step": 106 }, { "epoch": 0.2941580756013746, "grad_norm": 1.658296656226508, "learning_rate": 4.7032967032967035e-05, "loss": 0.6852, "step": 107 }, { "epoch": 0.29690721649484536, "grad_norm": 1.3133889484198134, "learning_rate": 4.747252747252748e-05, "loss": 0.6695, "step": 108 }, { "epoch": 0.29965635738831614, "grad_norm": 1.5315701499343, "learning_rate": 4.791208791208791e-05, "loss": 0.6842, "step": 109 }, { "epoch": 0.3024054982817869, "grad_norm": 1.5106029998832964, "learning_rate": 4.8351648351648355e-05, "loss": 0.6728, "step": 110 }, { "epoch": 0.30515463917525776, "grad_norm": 1.2198017486153612, "learning_rate": 4.87912087912088e-05, "loss": 0.6804, "step": 111 }, { "epoch": 0.30790378006872854, "grad_norm": 0.9408294254788567, "learning_rate": 4.923076923076924e-05, "loss": 0.6673, "step": 112 }, { "epoch": 0.3106529209621993, "grad_norm": 1.4299436206183789, "learning_rate": 4.9670329670329674e-05, "loss": 0.6782, "step": 113 }, { "epoch": 0.3134020618556701, "grad_norm": 1.3010524254683744, "learning_rate": 5.0109890109890116e-05, "loss": 0.6719, "step": 114 }, { "epoch": 0.3161512027491409, "grad_norm": 1.0294736794964319, "learning_rate": 5.054945054945055e-05, "loss": 0.6734, "step": 115 }, { "epoch": 0.31890034364261166, "grad_norm": 1.9844841183716888, "learning_rate": 5.098901098901099e-05, "loss": 0.6689, "step": 116 }, { "epoch": 0.3216494845360825, "grad_norm": 1.0769609751269622, "learning_rate": 5.1428571428571436e-05, "loss": 0.6725, "step": 117 }, { "epoch": 0.3243986254295533, "grad_norm": 1.9412353253336612, "learning_rate": 5.186813186813188e-05, "loss": 0.6788, "step": 118 }, { "epoch": 0.32714776632302406, "grad_norm": 1.2983006823972247, "learning_rate": 5.230769230769231e-05, "loss": 0.6717, "step": 119 }, { "epoch": 0.32989690721649484, "grad_norm": 1.78349719868111, "learning_rate": 5.274725274725275e-05, "loss": 0.6754, "step": 120 }, { "epoch": 0.3326460481099656, "grad_norm": 1.2685409855209113, "learning_rate": 5.318681318681319e-05, "loss": 0.6689, "step": 121 }, { "epoch": 0.3353951890034364, "grad_norm": 1.601626046067794, "learning_rate": 5.3626373626373626e-05, "loss": 0.6822, "step": 122 }, { "epoch": 0.33814432989690724, "grad_norm": 1.6399160217312283, "learning_rate": 5.4065934065934074e-05, "loss": 0.6939, "step": 123 }, { "epoch": 0.340893470790378, "grad_norm": 1.0496938208520434, "learning_rate": 5.450549450549451e-05, "loss": 0.6685, "step": 124 }, { "epoch": 0.3436426116838488, "grad_norm": 1.8210641674833647, "learning_rate": 5.494505494505495e-05, "loss": 0.6726, "step": 125 }, { "epoch": 0.3463917525773196, "grad_norm": 1.0994677896279192, "learning_rate": 5.538461538461539e-05, "loss": 0.6609, "step": 126 }, { "epoch": 0.34914089347079036, "grad_norm": 1.7375889267050222, "learning_rate": 5.582417582417583e-05, "loss": 0.68, "step": 127 }, { "epoch": 0.3518900343642612, "grad_norm": 1.1687845090860516, "learning_rate": 5.626373626373627e-05, "loss": 0.6715, "step": 128 }, { "epoch": 0.354639175257732, "grad_norm": 1.395445256796641, "learning_rate": 5.670329670329671e-05, "loss": 0.6681, "step": 129 }, { "epoch": 0.35738831615120276, "grad_norm": 1.355862476688157, "learning_rate": 5.714285714285715e-05, "loss": 0.6818, "step": 130 }, { "epoch": 0.36013745704467354, "grad_norm": 1.0924489864173972, "learning_rate": 5.7582417582417584e-05, "loss": 0.6605, "step": 131 }, { "epoch": 0.3628865979381443, "grad_norm": 1.9384854103627827, "learning_rate": 5.8021978021978026e-05, "loss": 0.6731, "step": 132 }, { "epoch": 0.3656357388316151, "grad_norm": 1.5529834412680505, "learning_rate": 5.846153846153846e-05, "loss": 0.675, "step": 133 }, { "epoch": 0.36838487972508593, "grad_norm": 0.9671293876862966, "learning_rate": 5.890109890109891e-05, "loss": 0.661, "step": 134 }, { "epoch": 0.3711340206185567, "grad_norm": 1.5434547051402783, "learning_rate": 5.9340659340659345e-05, "loss": 0.673, "step": 135 }, { "epoch": 0.3738831615120275, "grad_norm": 1.466260971358487, "learning_rate": 5.978021978021979e-05, "loss": 0.6799, "step": 136 }, { "epoch": 0.3766323024054983, "grad_norm": 1.3025182971790092, "learning_rate": 6.021978021978022e-05, "loss": 0.656, "step": 137 }, { "epoch": 0.37938144329896906, "grad_norm": 1.5407519214603846, "learning_rate": 6.0659340659340665e-05, "loss": 0.6703, "step": 138 }, { "epoch": 0.38213058419243984, "grad_norm": 1.2435864649597408, "learning_rate": 6.10989010989011e-05, "loss": 0.6699, "step": 139 }, { "epoch": 0.3848797250859107, "grad_norm": 1.32415071563785, "learning_rate": 6.153846153846155e-05, "loss": 0.6557, "step": 140 }, { "epoch": 0.38762886597938145, "grad_norm": 1.5121087379761295, "learning_rate": 6.197802197802199e-05, "loss": 0.6659, "step": 141 }, { "epoch": 0.39037800687285223, "grad_norm": 1.3570253258903069, "learning_rate": 6.241758241758242e-05, "loss": 0.6568, "step": 142 }, { "epoch": 0.393127147766323, "grad_norm": 1.2775583571637077, "learning_rate": 6.285714285714286e-05, "loss": 0.6558, "step": 143 }, { "epoch": 0.3958762886597938, "grad_norm": 1.029079383256909, "learning_rate": 6.32967032967033e-05, "loss": 0.6569, "step": 144 }, { "epoch": 0.39862542955326463, "grad_norm": 1.379942185445747, "learning_rate": 6.373626373626373e-05, "loss": 0.6664, "step": 145 }, { "epoch": 0.4013745704467354, "grad_norm": 1.7317621636752105, "learning_rate": 6.417582417582419e-05, "loss": 0.6675, "step": 146 }, { "epoch": 0.4041237113402062, "grad_norm": 1.0553856530228192, "learning_rate": 6.461538461538462e-05, "loss": 0.6623, "step": 147 }, { "epoch": 0.406872852233677, "grad_norm": 1.310533202428338, "learning_rate": 6.505494505494506e-05, "loss": 0.6658, "step": 148 }, { "epoch": 0.40962199312714775, "grad_norm": 1.0365623836624396, "learning_rate": 6.54945054945055e-05, "loss": 0.6539, "step": 149 }, { "epoch": 0.41237113402061853, "grad_norm": 1.5824321098093788, "learning_rate": 6.593406593406594e-05, "loss": 0.672, "step": 150 }, { "epoch": 0.41512027491408937, "grad_norm": 1.2342253922087731, "learning_rate": 6.637362637362638e-05, "loss": 0.6609, "step": 151 }, { "epoch": 0.41786941580756015, "grad_norm": 2.2157482513280797, "learning_rate": 6.681318681318683e-05, "loss": 0.6614, "step": 152 }, { "epoch": 0.42061855670103093, "grad_norm": 0.8519496697821355, "learning_rate": 6.725274725274725e-05, "loss": 0.6558, "step": 153 }, { "epoch": 0.4233676975945017, "grad_norm": 2.585667143290517, "learning_rate": 6.76923076923077e-05, "loss": 0.6712, "step": 154 }, { "epoch": 0.4261168384879725, "grad_norm": 1.561949044082342, "learning_rate": 6.813186813186814e-05, "loss": 0.6606, "step": 155 }, { "epoch": 0.4288659793814433, "grad_norm": 1.6541712978836747, "learning_rate": 6.857142857142857e-05, "loss": 0.6662, "step": 156 }, { "epoch": 0.4316151202749141, "grad_norm": 1.471980620525164, "learning_rate": 6.901098901098902e-05, "loss": 0.6688, "step": 157 }, { "epoch": 0.4343642611683849, "grad_norm": 1.4565326176621909, "learning_rate": 6.945054945054945e-05, "loss": 0.668, "step": 158 }, { "epoch": 0.43711340206185567, "grad_norm": 1.3548150534245453, "learning_rate": 6.98901098901099e-05, "loss": 0.6628, "step": 159 }, { "epoch": 0.43986254295532645, "grad_norm": 1.2045551451888672, "learning_rate": 7.032967032967034e-05, "loss": 0.6616, "step": 160 }, { "epoch": 0.44261168384879723, "grad_norm": 1.7494121190993142, "learning_rate": 7.076923076923078e-05, "loss": 0.6581, "step": 161 }, { "epoch": 0.44536082474226807, "grad_norm": 1.0147994550817765, "learning_rate": 7.12087912087912e-05, "loss": 0.6608, "step": 162 }, { "epoch": 0.44810996563573885, "grad_norm": 1.0859192635780746, "learning_rate": 7.164835164835166e-05, "loss": 0.6451, "step": 163 }, { "epoch": 0.45085910652920963, "grad_norm": 1.0492942035079653, "learning_rate": 7.208791208791209e-05, "loss": 0.6591, "step": 164 }, { "epoch": 0.4536082474226804, "grad_norm": 2.379607115922538, "learning_rate": 7.252747252747253e-05, "loss": 0.6634, "step": 165 }, { "epoch": 0.4563573883161512, "grad_norm": 1.112381416151378, "learning_rate": 7.296703296703297e-05, "loss": 0.6611, "step": 166 }, { "epoch": 0.45910652920962197, "grad_norm": 2.0850617091598886, "learning_rate": 7.34065934065934e-05, "loss": 0.6569, "step": 167 }, { "epoch": 0.4618556701030928, "grad_norm": 1.5623269496352417, "learning_rate": 7.384615384615386e-05, "loss": 0.672, "step": 168 }, { "epoch": 0.4646048109965636, "grad_norm": 1.6783665454838335, "learning_rate": 7.42857142857143e-05, "loss": 0.6567, "step": 169 }, { "epoch": 0.46735395189003437, "grad_norm": 1.0376996339879248, "learning_rate": 7.472527472527473e-05, "loss": 0.6606, "step": 170 }, { "epoch": 0.47010309278350515, "grad_norm": 2.2216262071159356, "learning_rate": 7.516483516483517e-05, "loss": 0.6636, "step": 171 }, { "epoch": 0.47285223367697593, "grad_norm": 1.3858824784808275, "learning_rate": 7.560439560439561e-05, "loss": 0.6673, "step": 172 }, { "epoch": 0.4756013745704467, "grad_norm": 1.9605283100011852, "learning_rate": 7.604395604395604e-05, "loss": 0.6572, "step": 173 }, { "epoch": 0.47835051546391755, "grad_norm": 1.4278083885748436, "learning_rate": 7.64835164835165e-05, "loss": 0.6645, "step": 174 }, { "epoch": 0.48109965635738833, "grad_norm": 1.191754742216612, "learning_rate": 7.692307692307693e-05, "loss": 0.6631, "step": 175 }, { "epoch": 0.4838487972508591, "grad_norm": 0.9597081578943443, "learning_rate": 7.736263736263737e-05, "loss": 0.655, "step": 176 }, { "epoch": 0.4865979381443299, "grad_norm": 1.9487418698749246, "learning_rate": 7.780219780219781e-05, "loss": 0.6617, "step": 177 }, { "epoch": 0.48934707903780067, "grad_norm": 1.4578651378155227, "learning_rate": 7.824175824175825e-05, "loss": 0.665, "step": 178 }, { "epoch": 0.49209621993127145, "grad_norm": 0.9204633011470749, "learning_rate": 7.868131868131868e-05, "loss": 0.6559, "step": 179 }, { "epoch": 0.4948453608247423, "grad_norm": 2.3519272780997893, "learning_rate": 7.912087912087914e-05, "loss": 0.661, "step": 180 }, { "epoch": 0.49759450171821307, "grad_norm": 1.4425041211579737, "learning_rate": 7.956043956043956e-05, "loss": 0.6555, "step": 181 }, { "epoch": 0.5003436426116838, "grad_norm": 2.2406732058079917, "learning_rate": 8e-05, "loss": 0.6672, "step": 182 }, { "epoch": 0.5030927835051546, "grad_norm": 1.717445277482838, "learning_rate": 7.999992597860977e-05, "loss": 0.669, "step": 183 }, { "epoch": 0.5058419243986254, "grad_norm": 1.9554853415214135, "learning_rate": 7.999970391471297e-05, "loss": 0.6484, "step": 184 }, { "epoch": 0.5085910652920962, "grad_norm": 1.3337471142504986, "learning_rate": 7.99993338091315e-05, "loss": 0.6542, "step": 185 }, { "epoch": 0.511340206185567, "grad_norm": 2.3628021636514913, "learning_rate": 7.999881566323518e-05, "loss": 0.6666, "step": 186 }, { "epoch": 0.5140893470790378, "grad_norm": 1.6284990281451381, "learning_rate": 7.999814947894166e-05, "loss": 0.6713, "step": 187 }, { "epoch": 0.5168384879725086, "grad_norm": 2.208780281202899, "learning_rate": 7.999733525871655e-05, "loss": 0.6564, "step": 188 }, { "epoch": 0.5195876288659794, "grad_norm": 2.0324845611311653, "learning_rate": 7.999637300557334e-05, "loss": 0.6643, "step": 189 }, { "epoch": 0.5223367697594502, "grad_norm": 1.411207407439057, "learning_rate": 7.999526272307338e-05, "loss": 0.6466, "step": 190 }, { "epoch": 0.525085910652921, "grad_norm": 1.67795112571074, "learning_rate": 7.999400441532593e-05, "loss": 0.6567, "step": 191 }, { "epoch": 0.5278350515463918, "grad_norm": 1.0171507403303963, "learning_rate": 7.999259808698805e-05, "loss": 0.6535, "step": 192 }, { "epoch": 0.5305841924398625, "grad_norm": 1.8103499540778905, "learning_rate": 7.999104374326465e-05, "loss": 0.6546, "step": 193 }, { "epoch": 0.5333333333333333, "grad_norm": 1.3619232333701012, "learning_rate": 7.99893413899085e-05, "loss": 0.6495, "step": 194 }, { "epoch": 0.5360824742268041, "grad_norm": 1.7537074766579162, "learning_rate": 7.99874910332201e-05, "loss": 0.6513, "step": 195 }, { "epoch": 0.5388316151202749, "grad_norm": 1.5125359865682328, "learning_rate": 7.998549268004776e-05, "loss": 0.6627, "step": 196 }, { "epoch": 0.5415807560137457, "grad_norm": 1.8694904140870736, "learning_rate": 7.998334633778752e-05, "loss": 0.6552, "step": 197 }, { "epoch": 0.5443298969072164, "grad_norm": 1.0313289192419033, "learning_rate": 7.998105201438315e-05, "loss": 0.6572, "step": 198 }, { "epoch": 0.5470790378006873, "grad_norm": 1.7524234342414688, "learning_rate": 7.997860971832609e-05, "loss": 0.6485, "step": 199 }, { "epoch": 0.5498281786941581, "grad_norm": 1.674466921847352, "learning_rate": 7.997601945865545e-05, "loss": 0.6567, "step": 200 }, { "epoch": 0.5525773195876289, "grad_norm": 0.9044207656617541, "learning_rate": 7.997328124495797e-05, "loss": 0.6571, "step": 201 }, { "epoch": 0.5553264604810997, "grad_norm": 1.5698427976284253, "learning_rate": 7.997039508736794e-05, "loss": 0.6552, "step": 202 }, { "epoch": 0.5580756013745705, "grad_norm": 1.2149289312154958, "learning_rate": 7.996736099656728e-05, "loss": 0.6504, "step": 203 }, { "epoch": 0.5608247422680412, "grad_norm": 1.5016916008686747, "learning_rate": 7.996417898378532e-05, "loss": 0.6501, "step": 204 }, { "epoch": 0.563573883161512, "grad_norm": 1.3674415360691992, "learning_rate": 7.996084906079895e-05, "loss": 0.6521, "step": 205 }, { "epoch": 0.5663230240549828, "grad_norm": 1.012486835623592, "learning_rate": 7.995737123993242e-05, "loss": 0.6533, "step": 206 }, { "epoch": 0.5690721649484536, "grad_norm": 1.2333701554709846, "learning_rate": 7.99537455340574e-05, "loss": 0.6423, "step": 207 }, { "epoch": 0.5718213058419244, "grad_norm": 1.432895391786843, "learning_rate": 7.994997195659288e-05, "loss": 0.6588, "step": 208 }, { "epoch": 0.5745704467353951, "grad_norm": 1.3560618026372162, "learning_rate": 7.994605052150512e-05, "loss": 0.6419, "step": 209 }, { "epoch": 0.5773195876288659, "grad_norm": 1.1169005294274361, "learning_rate": 7.994198124330764e-05, "loss": 0.643, "step": 210 }, { "epoch": 0.5800687285223368, "grad_norm": 1.1551183527776216, "learning_rate": 7.99377641370611e-05, "loss": 0.6658, "step": 211 }, { "epoch": 0.5828178694158076, "grad_norm": 1.494094722393518, "learning_rate": 7.993339921837333e-05, "loss": 0.6566, "step": 212 }, { "epoch": 0.5855670103092784, "grad_norm": 1.1087501045845025, "learning_rate": 7.992888650339918e-05, "loss": 0.6389, "step": 213 }, { "epoch": 0.5883161512027492, "grad_norm": 1.9002883231044971, "learning_rate": 7.992422600884052e-05, "loss": 0.6539, "step": 214 }, { "epoch": 0.5910652920962199, "grad_norm": 0.7299685549898423, "learning_rate": 7.991941775194619e-05, "loss": 0.6465, "step": 215 }, { "epoch": 0.5938144329896907, "grad_norm": 1.712564742362983, "learning_rate": 7.991446175051184e-05, "loss": 0.6474, "step": 216 }, { "epoch": 0.5965635738831615, "grad_norm": 1.4830236426687247, "learning_rate": 7.990935802288002e-05, "loss": 0.6512, "step": 217 }, { "epoch": 0.5993127147766323, "grad_norm": 1.1265182423363587, "learning_rate": 7.990410658793994e-05, "loss": 0.6541, "step": 218 }, { "epoch": 0.6020618556701031, "grad_norm": 1.2971103607944785, "learning_rate": 7.989870746512756e-05, "loss": 0.6479, "step": 219 }, { "epoch": 0.6048109965635738, "grad_norm": 1.2538177052036752, "learning_rate": 7.989316067442539e-05, "loss": 0.6497, "step": 220 }, { "epoch": 0.6075601374570446, "grad_norm": 0.7957257777767685, "learning_rate": 7.98874662363625e-05, "loss": 0.6352, "step": 221 }, { "epoch": 0.6103092783505155, "grad_norm": 0.9768965818664068, "learning_rate": 7.988162417201437e-05, "loss": 0.6443, "step": 222 }, { "epoch": 0.6130584192439863, "grad_norm": 1.9580245398506602, "learning_rate": 7.987563450300293e-05, "loss": 0.6652, "step": 223 }, { "epoch": 0.6158075601374571, "grad_norm": 1.065435200900762, "learning_rate": 7.986949725149634e-05, "loss": 0.6456, "step": 224 }, { "epoch": 0.6185567010309279, "grad_norm": 1.9649018261160003, "learning_rate": 7.986321244020901e-05, "loss": 0.6479, "step": 225 }, { "epoch": 0.6213058419243986, "grad_norm": 1.6213807682895127, "learning_rate": 7.985678009240142e-05, "loss": 0.6595, "step": 226 }, { "epoch": 0.6240549828178694, "grad_norm": 1.4900550716535215, "learning_rate": 7.985020023188018e-05, "loss": 0.6359, "step": 227 }, { "epoch": 0.6268041237113402, "grad_norm": 1.3214648767288044, "learning_rate": 7.98434728829978e-05, "loss": 0.647, "step": 228 }, { "epoch": 0.629553264604811, "grad_norm": 1.4301295600290387, "learning_rate": 7.983659807065267e-05, "loss": 0.6475, "step": 229 }, { "epoch": 0.6323024054982818, "grad_norm": 1.0375560871972622, "learning_rate": 7.982957582028892e-05, "loss": 0.6364, "step": 230 }, { "epoch": 0.6350515463917525, "grad_norm": 1.3894087032049172, "learning_rate": 7.982240615789641e-05, "loss": 0.6426, "step": 231 }, { "epoch": 0.6378006872852233, "grad_norm": 0.9850693356026884, "learning_rate": 7.981508911001057e-05, "loss": 0.6459, "step": 232 }, { "epoch": 0.6405498281786941, "grad_norm": 1.4426682351463136, "learning_rate": 7.980762470371228e-05, "loss": 0.6474, "step": 233 }, { "epoch": 0.643298969072165, "grad_norm": 1.0547468177759594, "learning_rate": 7.980001296662784e-05, "loss": 0.6469, "step": 234 }, { "epoch": 0.6460481099656358, "grad_norm": 1.2056014455360582, "learning_rate": 7.979225392692882e-05, "loss": 0.6354, "step": 235 }, { "epoch": 0.6487972508591066, "grad_norm": 1.535046209332006, "learning_rate": 7.978434761333195e-05, "loss": 0.6432, "step": 236 }, { "epoch": 0.6515463917525773, "grad_norm": 0.7757101221034997, "learning_rate": 7.977629405509905e-05, "loss": 0.633, "step": 237 }, { "epoch": 0.6542955326460481, "grad_norm": 0.992595763686358, "learning_rate": 7.976809328203693e-05, "loss": 0.64, "step": 238 }, { "epoch": 0.6570446735395189, "grad_norm": 1.3919630066257997, "learning_rate": 7.975974532449718e-05, "loss": 0.6495, "step": 239 }, { "epoch": 0.6597938144329897, "grad_norm": 1.130072368133697, "learning_rate": 7.975125021337618e-05, "loss": 0.6337, "step": 240 }, { "epoch": 0.6625429553264605, "grad_norm": 1.0428416388066186, "learning_rate": 7.974260798011494e-05, "loss": 0.6515, "step": 241 }, { "epoch": 0.6652920962199312, "grad_norm": 1.4678385771525282, "learning_rate": 7.973381865669897e-05, "loss": 0.6446, "step": 242 }, { "epoch": 0.668041237113402, "grad_norm": 1.1731453018255897, "learning_rate": 7.972488227565814e-05, "loss": 0.6424, "step": 243 }, { "epoch": 0.6707903780068728, "grad_norm": 1.1037532875258624, "learning_rate": 7.971579887006663e-05, "loss": 0.6356, "step": 244 }, { "epoch": 0.6735395189003437, "grad_norm": 1.3601555121892395, "learning_rate": 7.970656847354277e-05, "loss": 0.6332, "step": 245 }, { "epoch": 0.6762886597938145, "grad_norm": 0.6674453682989576, "learning_rate": 7.969719112024889e-05, "loss": 0.6341, "step": 246 }, { "epoch": 0.6790378006872853, "grad_norm": 0.8185867421910552, "learning_rate": 7.968766684489122e-05, "loss": 0.6351, "step": 247 }, { "epoch": 0.681786941580756, "grad_norm": 1.2859620347833656, "learning_rate": 7.967799568271978e-05, "loss": 0.636, "step": 248 }, { "epoch": 0.6845360824742268, "grad_norm": 1.0894539357769442, "learning_rate": 7.96681776695282e-05, "loss": 0.6483, "step": 249 }, { "epoch": 0.6872852233676976, "grad_norm": 1.536917534466485, "learning_rate": 7.965821284165362e-05, "loss": 0.6358, "step": 250 }, { "epoch": 0.6900343642611684, "grad_norm": 0.8985662521364918, "learning_rate": 7.964810123597659e-05, "loss": 0.6345, "step": 251 }, { "epoch": 0.6927835051546392, "grad_norm": 1.274319886309132, "learning_rate": 7.963784288992085e-05, "loss": 0.6416, "step": 252 }, { "epoch": 0.6955326460481099, "grad_norm": 0.8642801778285559, "learning_rate": 7.962743784145323e-05, "loss": 0.6386, "step": 253 }, { "epoch": 0.6982817869415807, "grad_norm": 0.9360706797234668, "learning_rate": 7.961688612908358e-05, "loss": 0.6311, "step": 254 }, { "epoch": 0.7010309278350515, "grad_norm": 1.2347692000194708, "learning_rate": 7.96061877918645e-05, "loss": 0.6342, "step": 255 }, { "epoch": 0.7037800687285224, "grad_norm": 1.3067449184972966, "learning_rate": 7.959534286939126e-05, "loss": 0.6348, "step": 256 }, { "epoch": 0.7065292096219932, "grad_norm": 1.0853797825255465, "learning_rate": 7.95843514018017e-05, "loss": 0.6387, "step": 257 }, { "epoch": 0.709278350515464, "grad_norm": 1.3876365868084906, "learning_rate": 7.9573213429776e-05, "loss": 0.6408, "step": 258 }, { "epoch": 0.7120274914089347, "grad_norm": 1.2720820097889098, "learning_rate": 7.956192899453656e-05, "loss": 0.6403, "step": 259 }, { "epoch": 0.7147766323024055, "grad_norm": 1.083330147390025, "learning_rate": 7.955049813784787e-05, "loss": 0.6326, "step": 260 }, { "epoch": 0.7175257731958763, "grad_norm": 0.8536354360684208, "learning_rate": 7.953892090201633e-05, "loss": 0.6289, "step": 261 }, { "epoch": 0.7202749140893471, "grad_norm": 0.6669283801432064, "learning_rate": 7.952719732989007e-05, "loss": 0.6311, "step": 262 }, { "epoch": 0.7230240549828179, "grad_norm": 0.6028131310980143, "learning_rate": 7.951532746485886e-05, "loss": 0.6383, "step": 263 }, { "epoch": 0.7257731958762886, "grad_norm": 1.0830419136461447, "learning_rate": 7.950331135085389e-05, "loss": 0.6257, "step": 264 }, { "epoch": 0.7285223367697594, "grad_norm": 1.7725196403596395, "learning_rate": 7.949114903234766e-05, "loss": 0.6431, "step": 265 }, { "epoch": 0.7312714776632302, "grad_norm": 0.7205576501203266, "learning_rate": 7.947884055435371e-05, "loss": 0.636, "step": 266 }, { "epoch": 0.734020618556701, "grad_norm": 2.0339470836036386, "learning_rate": 7.946638596242661e-05, "loss": 0.6435, "step": 267 }, { "epoch": 0.7367697594501719, "grad_norm": 1.0158450970399455, "learning_rate": 7.945378530266166e-05, "loss": 0.6342, "step": 268 }, { "epoch": 0.7395189003436426, "grad_norm": 2.328741839753608, "learning_rate": 7.944103862169478e-05, "loss": 0.6496, "step": 269 }, { "epoch": 0.7422680412371134, "grad_norm": 1.5377779904132143, "learning_rate": 7.94281459667023e-05, "loss": 0.6536, "step": 270 }, { "epoch": 0.7450171821305842, "grad_norm": 2.1537049583171064, "learning_rate": 7.941510738540086e-05, "loss": 0.6411, "step": 271 }, { "epoch": 0.747766323024055, "grad_norm": 1.7517265079351534, "learning_rate": 7.940192292604714e-05, "loss": 0.6489, "step": 272 }, { "epoch": 0.7505154639175258, "grad_norm": 2.1942121182211527, "learning_rate": 7.938859263743776e-05, "loss": 0.6345, "step": 273 }, { "epoch": 0.7532646048109966, "grad_norm": 1.863240282026407, "learning_rate": 7.937511656890903e-05, "loss": 0.65, "step": 274 }, { "epoch": 0.7560137457044673, "grad_norm": 1.4298523411886686, "learning_rate": 7.936149477033682e-05, "loss": 0.6391, "step": 275 }, { "epoch": 0.7587628865979381, "grad_norm": 1.2516811961606267, "learning_rate": 7.934772729213634e-05, "loss": 0.638, "step": 276 }, { "epoch": 0.7615120274914089, "grad_norm": 1.5514968401316607, "learning_rate": 7.9333814185262e-05, "loss": 0.6464, "step": 277 }, { "epoch": 0.7642611683848797, "grad_norm": 1.1588148872170223, "learning_rate": 7.931975550120716e-05, "loss": 0.642, "step": 278 }, { "epoch": 0.7670103092783506, "grad_norm": 2.0239434861105403, "learning_rate": 7.930555129200402e-05, "loss": 0.6428, "step": 279 }, { "epoch": 0.7697594501718213, "grad_norm": 1.6949238476000754, "learning_rate": 7.929120161022329e-05, "loss": 0.6333, "step": 280 }, { "epoch": 0.7725085910652921, "grad_norm": 1.2740472233054914, "learning_rate": 7.927670650897421e-05, "loss": 0.6392, "step": 281 }, { "epoch": 0.7752577319587629, "grad_norm": 1.2888896268808947, "learning_rate": 7.92620660419041e-05, "loss": 0.6324, "step": 282 }, { "epoch": 0.7780068728522337, "grad_norm": 1.2148413669943177, "learning_rate": 7.924728026319837e-05, "loss": 0.646, "step": 283 }, { "epoch": 0.7807560137457045, "grad_norm": 0.7141756639924294, "learning_rate": 7.923234922758021e-05, "loss": 0.6403, "step": 284 }, { "epoch": 0.7835051546391752, "grad_norm": 1.0058322496482315, "learning_rate": 7.921727299031042e-05, "loss": 0.6418, "step": 285 }, { "epoch": 0.786254295532646, "grad_norm": 0.794851634713387, "learning_rate": 7.920205160718721e-05, "loss": 0.6383, "step": 286 }, { "epoch": 0.7890034364261168, "grad_norm": 0.7763515931120876, "learning_rate": 7.918668513454598e-05, "loss": 0.629, "step": 287 }, { "epoch": 0.7917525773195876, "grad_norm": 1.0206658407382474, "learning_rate": 7.917117362925907e-05, "loss": 0.6369, "step": 288 }, { "epoch": 0.7945017182130584, "grad_norm": 0.9044945170274734, "learning_rate": 7.915551714873571e-05, "loss": 0.6265, "step": 289 }, { "epoch": 0.7972508591065293, "grad_norm": 1.7861242632179295, "learning_rate": 7.913971575092157e-05, "loss": 0.6354, "step": 290 }, { "epoch": 0.8, "grad_norm": 0.8352377362963846, "learning_rate": 7.912376949429875e-05, "loss": 0.6258, "step": 291 }, { "epoch": 0.8027491408934708, "grad_norm": 1.1285143530097073, "learning_rate": 7.910767843788543e-05, "loss": 0.6411, "step": 292 }, { "epoch": 0.8054982817869416, "grad_norm": 2.0986758482850774, "learning_rate": 7.909144264123575e-05, "loss": 0.6432, "step": 293 }, { "epoch": 0.8082474226804124, "grad_norm": 1.0319530318383152, "learning_rate": 7.90750621644395e-05, "loss": 0.6489, "step": 294 }, { "epoch": 0.8109965635738832, "grad_norm": 2.628617525239813, "learning_rate": 7.905853706812199e-05, "loss": 0.6439, "step": 295 }, { "epoch": 0.813745704467354, "grad_norm": 1.977433582449252, "learning_rate": 7.904186741344373e-05, "loss": 0.6482, "step": 296 }, { "epoch": 0.8164948453608247, "grad_norm": 1.8360229102761563, "learning_rate": 7.902505326210028e-05, "loss": 0.65, "step": 297 }, { "epoch": 0.8192439862542955, "grad_norm": 1.5433944283254346, "learning_rate": 7.900809467632197e-05, "loss": 0.6434, "step": 298 }, { "epoch": 0.8219931271477663, "grad_norm": 1.6379518485922502, "learning_rate": 7.899099171887373e-05, "loss": 0.6553, "step": 299 }, { "epoch": 0.8247422680412371, "grad_norm": 1.477099597856914, "learning_rate": 7.897374445305478e-05, "loss": 0.6392, "step": 300 }, { "epoch": 0.8274914089347079, "grad_norm": 1.1857730742135408, "learning_rate": 7.895635294269843e-05, "loss": 0.6316, "step": 301 }, { "epoch": 0.8302405498281787, "grad_norm": 1.361216152623904, "learning_rate": 7.893881725217191e-05, "loss": 0.6352, "step": 302 }, { "epoch": 0.8329896907216495, "grad_norm": 0.8010015883602253, "learning_rate": 7.892113744637599e-05, "loss": 0.6382, "step": 303 }, { "epoch": 0.8357388316151203, "grad_norm": 1.102241513039724, "learning_rate": 7.890331359074488e-05, "loss": 0.6447, "step": 304 }, { "epoch": 0.8384879725085911, "grad_norm": 1.1821390144057153, "learning_rate": 7.888534575124591e-05, "loss": 0.6292, "step": 305 }, { "epoch": 0.8412371134020619, "grad_norm": 0.7269296274858387, "learning_rate": 7.886723399437931e-05, "loss": 0.6319, "step": 306 }, { "epoch": 0.8439862542955326, "grad_norm": 1.630523895781267, "learning_rate": 7.884897838717792e-05, "loss": 0.6169, "step": 307 }, { "epoch": 0.8467353951890034, "grad_norm": 0.740638554856235, "learning_rate": 7.883057899720703e-05, "loss": 0.6355, "step": 308 }, { "epoch": 0.8494845360824742, "grad_norm": 1.8497443172662162, "learning_rate": 7.881203589256408e-05, "loss": 0.6274, "step": 309 }, { "epoch": 0.852233676975945, "grad_norm": 1.1910232798072693, "learning_rate": 7.879334914187836e-05, "loss": 0.6392, "step": 310 }, { "epoch": 0.8549828178694158, "grad_norm": 2.1003839885669815, "learning_rate": 7.877451881431086e-05, "loss": 0.6296, "step": 311 }, { "epoch": 0.8577319587628865, "grad_norm": 2.041585074702372, "learning_rate": 7.87555449795539e-05, "loss": 0.6352, "step": 312 }, { "epoch": 0.8604810996563574, "grad_norm": 1.0287133005035243, "learning_rate": 7.873642770783098e-05, "loss": 0.6361, "step": 313 }, { "epoch": 0.8632302405498282, "grad_norm": 1.4803932962834592, "learning_rate": 7.871716706989645e-05, "loss": 0.6433, "step": 314 }, { "epoch": 0.865979381443299, "grad_norm": 0.9388113471557215, "learning_rate": 7.869776313703528e-05, "loss": 0.6346, "step": 315 }, { "epoch": 0.8687285223367698, "grad_norm": 0.8575055727089447, "learning_rate": 7.867821598106275e-05, "loss": 0.6287, "step": 316 }, { "epoch": 0.8714776632302406, "grad_norm": 0.9753760114729692, "learning_rate": 7.865852567432428e-05, "loss": 0.6386, "step": 317 }, { "epoch": 0.8742268041237113, "grad_norm": 0.9688066608476287, "learning_rate": 7.863869228969501e-05, "loss": 0.631, "step": 318 }, { "epoch": 0.8769759450171821, "grad_norm": 0.9593995798512949, "learning_rate": 7.861871590057971e-05, "loss": 0.6347, "step": 319 }, { "epoch": 0.8797250859106529, "grad_norm": 1.2389383575654551, "learning_rate": 7.85985965809124e-05, "loss": 0.6346, "step": 320 }, { "epoch": 0.8824742268041237, "grad_norm": 0.8827816805243792, "learning_rate": 7.857833440515605e-05, "loss": 0.6265, "step": 321 }, { "epoch": 0.8852233676975945, "grad_norm": 0.6061777594451044, "learning_rate": 7.85579294483024e-05, "loss": 0.6286, "step": 322 }, { "epoch": 0.8879725085910652, "grad_norm": 0.6113638623646376, "learning_rate": 7.85373817858716e-05, "loss": 0.6256, "step": 323 }, { "epoch": 0.8907216494845361, "grad_norm": 0.8948079812828788, "learning_rate": 7.851669149391198e-05, "loss": 0.6238, "step": 324 }, { "epoch": 0.8934707903780069, "grad_norm": 1.4121615992032686, "learning_rate": 7.849585864899976e-05, "loss": 0.633, "step": 325 }, { "epoch": 0.8962199312714777, "grad_norm": 0.7940592789567055, "learning_rate": 7.847488332823873e-05, "loss": 0.6359, "step": 326 }, { "epoch": 0.8989690721649485, "grad_norm": 0.5764914369379771, "learning_rate": 7.845376560926002e-05, "loss": 0.6375, "step": 327 }, { "epoch": 0.9017182130584193, "grad_norm": 1.071092163645109, "learning_rate": 7.843250557022177e-05, "loss": 0.6196, "step": 328 }, { "epoch": 0.90446735395189, "grad_norm": 1.6020470680476175, "learning_rate": 7.841110328980887e-05, "loss": 0.6293, "step": 329 }, { "epoch": 0.9072164948453608, "grad_norm": 0.4752698709356062, "learning_rate": 7.838955884723265e-05, "loss": 0.626, "step": 330 }, { "epoch": 0.9099656357388316, "grad_norm": 1.5726092706691321, "learning_rate": 7.836787232223058e-05, "loss": 0.6294, "step": 331 }, { "epoch": 0.9127147766323024, "grad_norm": 1.01011634702882, "learning_rate": 7.8346043795066e-05, "loss": 0.6431, "step": 332 }, { "epoch": 0.9154639175257732, "grad_norm": 0.8236758132081481, "learning_rate": 7.83240733465278e-05, "loss": 0.6237, "step": 333 }, { "epoch": 0.9182130584192439, "grad_norm": 0.816087984512738, "learning_rate": 7.830196105793017e-05, "loss": 0.6304, "step": 334 }, { "epoch": 0.9209621993127147, "grad_norm": 0.7514093388017495, "learning_rate": 7.827970701111219e-05, "loss": 0.6311, "step": 335 }, { "epoch": 0.9237113402061856, "grad_norm": 0.9502215222777328, "learning_rate": 7.825731128843762e-05, "loss": 0.6317, "step": 336 }, { "epoch": 0.9264604810996564, "grad_norm": 1.2376836455505118, "learning_rate": 7.823477397279464e-05, "loss": 0.6199, "step": 337 }, { "epoch": 0.9292096219931272, "grad_norm": 0.7535199306698559, "learning_rate": 7.821209514759539e-05, "loss": 0.6248, "step": 338 }, { "epoch": 0.931958762886598, "grad_norm": 0.8490031147278277, "learning_rate": 7.818927489677577e-05, "loss": 0.6267, "step": 339 }, { "epoch": 0.9347079037800687, "grad_norm": 1.3764847868272945, "learning_rate": 7.816631330479514e-05, "loss": 0.6307, "step": 340 }, { "epoch": 0.9374570446735395, "grad_norm": 0.6440823293174261, "learning_rate": 7.814321045663594e-05, "loss": 0.6366, "step": 341 }, { "epoch": 0.9402061855670103, "grad_norm": 1.117461654132214, "learning_rate": 7.811996643780339e-05, "loss": 0.6284, "step": 342 }, { "epoch": 0.9429553264604811, "grad_norm": 1.198712384928446, "learning_rate": 7.809658133432526e-05, "loss": 0.6179, "step": 343 }, { "epoch": 0.9457044673539519, "grad_norm": 0.7800718683120617, "learning_rate": 7.807305523275142e-05, "loss": 0.6246, "step": 344 }, { "epoch": 0.9484536082474226, "grad_norm": 1.106890600114741, "learning_rate": 7.804938822015361e-05, "loss": 0.6186, "step": 345 }, { "epoch": 0.9512027491408934, "grad_norm": 1.1783968723372122, "learning_rate": 7.802558038412509e-05, "loss": 0.6249, "step": 346 }, { "epoch": 0.9539518900343643, "grad_norm": 0.909115741821025, "learning_rate": 7.800163181278033e-05, "loss": 0.6244, "step": 347 }, { "epoch": 0.9567010309278351, "grad_norm": 1.0133199245495836, "learning_rate": 7.797754259475464e-05, "loss": 0.6264, "step": 348 }, { "epoch": 0.9594501718213059, "grad_norm": 1.195705135014382, "learning_rate": 7.795331281920387e-05, "loss": 0.6278, "step": 349 }, { "epoch": 0.9621993127147767, "grad_norm": 0.7727285070098746, "learning_rate": 7.792894257580415e-05, "loss": 0.6322, "step": 350 }, { "epoch": 0.9649484536082474, "grad_norm": 0.9523131612793756, "learning_rate": 7.790443195475142e-05, "loss": 0.6251, "step": 351 }, { "epoch": 0.9676975945017182, "grad_norm": 1.3036288679788564, "learning_rate": 7.78797810467612e-05, "loss": 0.6218, "step": 352 }, { "epoch": 0.970446735395189, "grad_norm": 0.5634490374165979, "learning_rate": 7.785498994306821e-05, "loss": 0.6221, "step": 353 }, { "epoch": 0.9731958762886598, "grad_norm": 1.5260597480048004, "learning_rate": 7.783005873542605e-05, "loss": 0.619, "step": 354 }, { "epoch": 0.9759450171821306, "grad_norm": 0.6409092943149695, "learning_rate": 7.780498751610684e-05, "loss": 0.6251, "step": 355 }, { "epoch": 0.9786941580756013, "grad_norm": 1.6664465662420271, "learning_rate": 7.777977637790092e-05, "loss": 0.6357, "step": 356 }, { "epoch": 0.9814432989690721, "grad_norm": 0.928708660543268, "learning_rate": 7.775442541411647e-05, "loss": 0.6218, "step": 357 }, { "epoch": 0.9841924398625429, "grad_norm": 1.8449391562901856, "learning_rate": 7.772893471857915e-05, "loss": 0.6396, "step": 358 }, { "epoch": 0.9869415807560138, "grad_norm": 1.6259312779533432, "learning_rate": 7.77033043856318e-05, "loss": 0.6368, "step": 359 }, { "epoch": 0.9896907216494846, "grad_norm": 0.9340296294033434, "learning_rate": 7.767753451013408e-05, "loss": 0.6204, "step": 360 }, { "epoch": 0.9924398625429554, "grad_norm": 1.2307139404659375, "learning_rate": 7.765162518746207e-05, "loss": 0.6177, "step": 361 }, { "epoch": 0.9951890034364261, "grad_norm": 0.7744777080187526, "learning_rate": 7.762557651350798e-05, "loss": 0.6288, "step": 362 }, { "epoch": 0.9979381443298969, "grad_norm": 1.0676481742466821, "learning_rate": 7.759938858467979e-05, "loss": 0.6272, "step": 363 }, { "epoch": 1.0006872852233677, "grad_norm": 0.9232629543637857, "learning_rate": 7.757306149790082e-05, "loss": 0.7685, "step": 364 }, { "epoch": 1.0034364261168385, "grad_norm": 1.0678574538408068, "learning_rate": 7.754659535060943e-05, "loss": 0.5916, "step": 365 }, { "epoch": 1.0061855670103093, "grad_norm": 1.2551799774905172, "learning_rate": 7.751999024075871e-05, "loss": 0.5988, "step": 366 }, { "epoch": 1.00893470790378, "grad_norm": 0.9263667600041875, "learning_rate": 7.749324626681599e-05, "loss": 0.5966, "step": 367 }, { "epoch": 1.0116838487972508, "grad_norm": 0.731842781771422, "learning_rate": 7.746636352776259e-05, "loss": 0.5928, "step": 368 }, { "epoch": 1.0144329896907216, "grad_norm": 0.8372632610382791, "learning_rate": 7.74393421230934e-05, "loss": 0.5919, "step": 369 }, { "epoch": 1.0171821305841924, "grad_norm": 1.0188957663159135, "learning_rate": 7.741218215281652e-05, "loss": 0.5932, "step": 370 }, { "epoch": 1.0199312714776632, "grad_norm": 1.3394999167844803, "learning_rate": 7.738488371745287e-05, "loss": 0.5879, "step": 371 }, { "epoch": 1.022680412371134, "grad_norm": 0.8980665393748412, "learning_rate": 7.735744691803588e-05, "loss": 0.5942, "step": 372 }, { "epoch": 1.0254295532646047, "grad_norm": 0.9856701598024391, "learning_rate": 7.732987185611102e-05, "loss": 0.5793, "step": 373 }, { "epoch": 1.0281786941580755, "grad_norm": 0.8156719105288647, "learning_rate": 7.730215863373554e-05, "loss": 0.5953, "step": 374 }, { "epoch": 1.0309278350515463, "grad_norm": 0.5095314934411239, "learning_rate": 7.727430735347799e-05, "loss": 0.5848, "step": 375 }, { "epoch": 1.0336769759450173, "grad_norm": 0.5914859323410439, "learning_rate": 7.724631811841789e-05, "loss": 0.5877, "step": 376 }, { "epoch": 1.036426116838488, "grad_norm": 0.6410263828936001, "learning_rate": 7.721819103214536e-05, "loss": 0.5852, "step": 377 }, { "epoch": 1.0391752577319588, "grad_norm": 0.9724802661391164, "learning_rate": 7.71899261987607e-05, "loss": 0.5949, "step": 378 }, { "epoch": 1.0419243986254296, "grad_norm": 1.3805105132332318, "learning_rate": 7.716152372287399e-05, "loss": 0.5912, "step": 379 }, { "epoch": 1.0446735395189004, "grad_norm": 0.5961514316349491, "learning_rate": 7.713298370960481e-05, "loss": 0.5989, "step": 380 }, { "epoch": 1.0474226804123712, "grad_norm": 1.0381265315374555, "learning_rate": 7.710430626458171e-05, "loss": 0.5853, "step": 381 }, { "epoch": 1.050171821305842, "grad_norm": 0.88611297798641, "learning_rate": 7.707549149394192e-05, "loss": 0.5971, "step": 382 }, { "epoch": 1.0529209621993127, "grad_norm": 0.9894621130663098, "learning_rate": 7.704653950433092e-05, "loss": 0.5839, "step": 383 }, { "epoch": 1.0556701030927835, "grad_norm": 1.4525179754862534, "learning_rate": 7.7017450402902e-05, "loss": 0.5841, "step": 384 }, { "epoch": 1.0584192439862543, "grad_norm": 0.6031744444046739, "learning_rate": 7.698822429731595e-05, "loss": 0.577, "step": 385 }, { "epoch": 1.061168384879725, "grad_norm": 1.0814979553674235, "learning_rate": 7.695886129574065e-05, "loss": 0.5963, "step": 386 }, { "epoch": 1.0639175257731959, "grad_norm": 1.2035521726093799, "learning_rate": 7.692936150685059e-05, "loss": 0.59, "step": 387 }, { "epoch": 1.0666666666666667, "grad_norm": 1.065535271974634, "learning_rate": 7.689972503982654e-05, "loss": 0.59, "step": 388 }, { "epoch": 1.0694158075601374, "grad_norm": 0.9885511702158032, "learning_rate": 7.686995200435513e-05, "loss": 0.5908, "step": 389 }, { "epoch": 1.0721649484536082, "grad_norm": 1.0056851158221545, "learning_rate": 7.684004251062844e-05, "loss": 0.592, "step": 390 }, { "epoch": 1.074914089347079, "grad_norm": 1.0202284999862656, "learning_rate": 7.680999666934358e-05, "loss": 0.5809, "step": 391 }, { "epoch": 1.0776632302405498, "grad_norm": 1.1513878865642915, "learning_rate": 7.677981459170228e-05, "loss": 0.5878, "step": 392 }, { "epoch": 1.0804123711340206, "grad_norm": 0.8699810682515629, "learning_rate": 7.674949638941053e-05, "loss": 0.5915, "step": 393 }, { "epoch": 1.0831615120274913, "grad_norm": 0.4496053235798414, "learning_rate": 7.671904217467808e-05, "loss": 0.5801, "step": 394 }, { "epoch": 1.0859106529209621, "grad_norm": 0.6745209087686392, "learning_rate": 7.668845206021812e-05, "loss": 0.5979, "step": 395 }, { "epoch": 1.088659793814433, "grad_norm": 1.2754326592180227, "learning_rate": 7.66577261592468e-05, "loss": 0.5807, "step": 396 }, { "epoch": 1.0914089347079037, "grad_norm": 0.7069250995821895, "learning_rate": 7.662686458548276e-05, "loss": 0.582, "step": 397 }, { "epoch": 1.0941580756013747, "grad_norm": 0.4973387506195001, "learning_rate": 7.659586745314689e-05, "loss": 0.5855, "step": 398 }, { "epoch": 1.0969072164948455, "grad_norm": 0.5388764587620251, "learning_rate": 7.65647348769617e-05, "loss": 0.5901, "step": 399 }, { "epoch": 1.0996563573883162, "grad_norm": 0.5432430222716064, "learning_rate": 7.653346697215103e-05, "loss": 0.5852, "step": 400 }, { "epoch": 1.102405498281787, "grad_norm": 0.7845045531047907, "learning_rate": 7.650206385443958e-05, "loss": 0.5815, "step": 401 }, { "epoch": 1.1051546391752578, "grad_norm": 1.1593889092948502, "learning_rate": 7.647052564005243e-05, "loss": 0.5884, "step": 402 }, { "epoch": 1.1079037800687286, "grad_norm": 0.9849246083500859, "learning_rate": 7.643885244571474e-05, "loss": 0.5884, "step": 403 }, { "epoch": 1.1106529209621994, "grad_norm": 1.2327221523818575, "learning_rate": 7.64070443886512e-05, "loss": 0.5808, "step": 404 }, { "epoch": 1.1134020618556701, "grad_norm": 0.8894760613915432, "learning_rate": 7.637510158658563e-05, "loss": 0.5808, "step": 405 }, { "epoch": 1.116151202749141, "grad_norm": 0.8458010101721215, "learning_rate": 7.634302415774057e-05, "loss": 0.5921, "step": 406 }, { "epoch": 1.1189003436426117, "grad_norm": 1.0269578881644645, "learning_rate": 7.63108122208368e-05, "loss": 0.5876, "step": 407 }, { "epoch": 1.1216494845360825, "grad_norm": 0.9883024415526525, "learning_rate": 7.627846589509295e-05, "loss": 0.5938, "step": 408 }, { "epoch": 1.1243986254295533, "grad_norm": 1.0374738987854506, "learning_rate": 7.624598530022502e-05, "loss": 0.5759, "step": 409 }, { "epoch": 1.127147766323024, "grad_norm": 0.8968993730271081, "learning_rate": 7.621337055644596e-05, "loss": 0.5971, "step": 410 }, { "epoch": 1.1298969072164948, "grad_norm": 0.645336647760953, "learning_rate": 7.618062178446518e-05, "loss": 0.5843, "step": 411 }, { "epoch": 1.1326460481099656, "grad_norm": 0.8156198060430901, "learning_rate": 7.614773910548816e-05, "loss": 0.5834, "step": 412 }, { "epoch": 1.1353951890034364, "grad_norm": 0.9969205762211508, "learning_rate": 7.6114722641216e-05, "loss": 0.5874, "step": 413 }, { "epoch": 1.1381443298969072, "grad_norm": 0.9512507581396507, "learning_rate": 7.608157251384493e-05, "loss": 0.5907, "step": 414 }, { "epoch": 1.140893470790378, "grad_norm": 0.46469337738492333, "learning_rate": 7.604828884606587e-05, "loss": 0.5854, "step": 415 }, { "epoch": 1.1436426116838487, "grad_norm": 0.6875741181821637, "learning_rate": 7.601487176106397e-05, "loss": 0.5907, "step": 416 }, { "epoch": 1.1463917525773195, "grad_norm": 0.8229065410063605, "learning_rate": 7.59813213825182e-05, "loss": 0.5835, "step": 417 }, { "epoch": 1.1491408934707903, "grad_norm": 0.5535331052196206, "learning_rate": 7.594763783460086e-05, "loss": 0.5913, "step": 418 }, { "epoch": 1.151890034364261, "grad_norm": 0.609930416365321, "learning_rate": 7.591382124197708e-05, "loss": 0.5861, "step": 419 }, { "epoch": 1.1546391752577319, "grad_norm": 0.549391385583707, "learning_rate": 7.587987172980443e-05, "loss": 0.5774, "step": 420 }, { "epoch": 1.1573883161512026, "grad_norm": 0.5299626474050247, "learning_rate": 7.58457894237324e-05, "loss": 0.5915, "step": 421 }, { "epoch": 1.1601374570446734, "grad_norm": 0.5702316031822813, "learning_rate": 7.581157444990199e-05, "loss": 0.5771, "step": 422 }, { "epoch": 1.1628865979381444, "grad_norm": 0.8908922680435231, "learning_rate": 7.577722693494519e-05, "loss": 0.5932, "step": 423 }, { "epoch": 1.1656357388316152, "grad_norm": 1.4431986503442844, "learning_rate": 7.574274700598453e-05, "loss": 0.5813, "step": 424 }, { "epoch": 1.168384879725086, "grad_norm": 0.6369720186684446, "learning_rate": 7.570813479063265e-05, "loss": 0.5845, "step": 425 }, { "epoch": 1.1711340206185568, "grad_norm": 0.5256518804042303, "learning_rate": 7.567339041699175e-05, "loss": 0.5924, "step": 426 }, { "epoch": 1.1738831615120275, "grad_norm": 0.8165966324851138, "learning_rate": 7.563851401365316e-05, "loss": 0.588, "step": 427 }, { "epoch": 1.1766323024054983, "grad_norm": 1.1011809927414538, "learning_rate": 7.56035057096969e-05, "loss": 0.5878, "step": 428 }, { "epoch": 1.179381443298969, "grad_norm": 1.2202082201319733, "learning_rate": 7.556836563469111e-05, "loss": 0.589, "step": 429 }, { "epoch": 1.1821305841924399, "grad_norm": 0.8269982439978933, "learning_rate": 7.553309391869167e-05, "loss": 0.5846, "step": 430 }, { "epoch": 1.1848797250859107, "grad_norm": 1.0285455684872933, "learning_rate": 7.549769069224164e-05, "loss": 0.5883, "step": 431 }, { "epoch": 1.1876288659793814, "grad_norm": 1.4122193045161404, "learning_rate": 7.546215608637083e-05, "loss": 0.5832, "step": 432 }, { "epoch": 1.1903780068728522, "grad_norm": 0.3041741502072138, "learning_rate": 7.542649023259527e-05, "loss": 0.5813, "step": 433 }, { "epoch": 1.193127147766323, "grad_norm": 1.5989692268879376, "learning_rate": 7.53906932629168e-05, "loss": 0.5896, "step": 434 }, { "epoch": 1.1958762886597938, "grad_norm": 0.5519101760067153, "learning_rate": 7.535476530982244e-05, "loss": 0.5719, "step": 435 }, { "epoch": 1.1986254295532646, "grad_norm": 0.770417484741293, "learning_rate": 7.53187065062841e-05, "loss": 0.5881, "step": 436 }, { "epoch": 1.2013745704467353, "grad_norm": 1.3484068230851232, "learning_rate": 7.528251698575788e-05, "loss": 0.5868, "step": 437 }, { "epoch": 1.2041237113402061, "grad_norm": 0.5995693974246964, "learning_rate": 7.524619688218372e-05, "loss": 0.5812, "step": 438 }, { "epoch": 1.206872852233677, "grad_norm": 0.9168760247232807, "learning_rate": 7.520974632998485e-05, "loss": 0.5877, "step": 439 }, { "epoch": 1.2096219931271477, "grad_norm": 1.1013744178468492, "learning_rate": 7.51731654640673e-05, "loss": 0.592, "step": 440 }, { "epoch": 1.2123711340206185, "grad_norm": 0.5790863138582544, "learning_rate": 7.51364544198194e-05, "loss": 0.5749, "step": 441 }, { "epoch": 1.2151202749140895, "grad_norm": 0.5723305467246104, "learning_rate": 7.509961333311126e-05, "loss": 0.5905, "step": 442 }, { "epoch": 1.2178694158075603, "grad_norm": 0.6893985276435765, "learning_rate": 7.506264234029432e-05, "loss": 0.5799, "step": 443 }, { "epoch": 1.220618556701031, "grad_norm": 0.6306086582019484, "learning_rate": 7.502554157820079e-05, "loss": 0.583, "step": 444 }, { "epoch": 1.2233676975945018, "grad_norm": 0.5828509036657884, "learning_rate": 7.498831118414316e-05, "loss": 0.5791, "step": 445 }, { "epoch": 1.2261168384879726, "grad_norm": 0.9766394221436852, "learning_rate": 7.495095129591373e-05, "loss": 0.5876, "step": 446 }, { "epoch": 1.2288659793814434, "grad_norm": 1.416296834720271, "learning_rate": 7.4913462051784e-05, "loss": 0.5876, "step": 447 }, { "epoch": 1.2316151202749142, "grad_norm": 0.6702851111393332, "learning_rate": 7.487584359050431e-05, "loss": 0.5881, "step": 448 }, { "epoch": 1.234364261168385, "grad_norm": 0.9384647234340839, "learning_rate": 7.483809605130319e-05, "loss": 0.5959, "step": 449 }, { "epoch": 1.2371134020618557, "grad_norm": 1.2907110973467573, "learning_rate": 7.480021957388691e-05, "loss": 0.5785, "step": 450 }, { "epoch": 1.2398625429553265, "grad_norm": 0.8298126531860124, "learning_rate": 7.476221429843894e-05, "loss": 0.5934, "step": 451 }, { "epoch": 1.2426116838487973, "grad_norm": 1.1127124873762135, "learning_rate": 7.472408036561942e-05, "loss": 0.5833, "step": 452 }, { "epoch": 1.245360824742268, "grad_norm": 0.7148549185439407, "learning_rate": 7.468581791656473e-05, "loss": 0.5927, "step": 453 }, { "epoch": 1.2481099656357388, "grad_norm": 0.863808291677187, "learning_rate": 7.464742709288683e-05, "loss": 0.5955, "step": 454 }, { "epoch": 1.2508591065292096, "grad_norm": 0.8446688366988994, "learning_rate": 7.460890803667283e-05, "loss": 0.5939, "step": 455 }, { "epoch": 1.2536082474226804, "grad_norm": 1.0639580002350018, "learning_rate": 7.457026089048445e-05, "loss": 0.5839, "step": 456 }, { "epoch": 1.2563573883161512, "grad_norm": 1.1091140834360853, "learning_rate": 7.453148579735743e-05, "loss": 0.5829, "step": 457 }, { "epoch": 1.259106529209622, "grad_norm": 0.80462944337638, "learning_rate": 7.449258290080112e-05, "loss": 0.5845, "step": 458 }, { "epoch": 1.2618556701030927, "grad_norm": 0.8412084155727457, "learning_rate": 7.445355234479783e-05, "loss": 0.5767, "step": 459 }, { "epoch": 1.2646048109965635, "grad_norm": 0.6330472076770527, "learning_rate": 7.441439427380235e-05, "loss": 0.5816, "step": 460 }, { "epoch": 1.2673539518900343, "grad_norm": 0.597903788175847, "learning_rate": 7.437510883274144e-05, "loss": 0.5729, "step": 461 }, { "epoch": 1.270103092783505, "grad_norm": 0.4287714351702764, "learning_rate": 7.433569616701324e-05, "loss": 0.5932, "step": 462 }, { "epoch": 1.2728522336769759, "grad_norm": 0.47550714789032805, "learning_rate": 7.429615642248677e-05, "loss": 0.5929, "step": 463 }, { "epoch": 1.2756013745704466, "grad_norm": 0.5414468519874974, "learning_rate": 7.425648974550138e-05, "loss": 0.5774, "step": 464 }, { "epoch": 1.2783505154639174, "grad_norm": 0.5640877600721679, "learning_rate": 7.421669628286617e-05, "loss": 0.586, "step": 465 }, { "epoch": 1.2810996563573882, "grad_norm": 0.6048691932188638, "learning_rate": 7.417677618185955e-05, "loss": 0.597, "step": 466 }, { "epoch": 1.283848797250859, "grad_norm": 0.7508265029471074, "learning_rate": 7.413672959022856e-05, "loss": 0.5961, "step": 467 }, { "epoch": 1.2865979381443298, "grad_norm": 1.066042006928496, "learning_rate": 7.409655665618843e-05, "loss": 0.5837, "step": 468 }, { "epoch": 1.2893470790378008, "grad_norm": 1.4460442111626333, "learning_rate": 7.405625752842198e-05, "loss": 0.5771, "step": 469 }, { "epoch": 1.2920962199312716, "grad_norm": 0.40527253568707877, "learning_rate": 7.401583235607908e-05, "loss": 0.5769, "step": 470 }, { "epoch": 1.2948453608247423, "grad_norm": 0.8934936601794983, "learning_rate": 7.397528128877611e-05, "loss": 0.585, "step": 471 }, { "epoch": 1.2975945017182131, "grad_norm": 1.6705979909682174, "learning_rate": 7.393460447659539e-05, "loss": 0.5932, "step": 472 }, { "epoch": 1.300343642611684, "grad_norm": 0.4422750143504963, "learning_rate": 7.389380207008462e-05, "loss": 0.5878, "step": 473 }, { "epoch": 1.3030927835051547, "grad_norm": 1.5622517777634235, "learning_rate": 7.385287422025635e-05, "loss": 0.5895, "step": 474 }, { "epoch": 1.3058419243986255, "grad_norm": 0.6900852059622544, "learning_rate": 7.381182107858738e-05, "loss": 0.5883, "step": 475 }, { "epoch": 1.3085910652920962, "grad_norm": 0.9882097928385363, "learning_rate": 7.377064279701827e-05, "loss": 0.5894, "step": 476 }, { "epoch": 1.311340206185567, "grad_norm": 0.9430491893537721, "learning_rate": 7.37293395279527e-05, "loss": 0.5921, "step": 477 }, { "epoch": 1.3140893470790378, "grad_norm": 0.7682283056356898, "learning_rate": 7.368791142425691e-05, "loss": 0.5977, "step": 478 }, { "epoch": 1.3168384879725086, "grad_norm": 1.0703629836228026, "learning_rate": 7.364635863925922e-05, "loss": 0.5829, "step": 479 }, { "epoch": 1.3195876288659794, "grad_norm": 0.8976152595403926, "learning_rate": 7.360468132674935e-05, "loss": 0.5902, "step": 480 }, { "epoch": 1.3223367697594501, "grad_norm": 0.7540663375222781, "learning_rate": 7.356287964097795e-05, "loss": 0.588, "step": 481 }, { "epoch": 1.325085910652921, "grad_norm": 1.1037972218662753, "learning_rate": 7.352095373665598e-05, "loss": 0.5814, "step": 482 }, { "epoch": 1.3278350515463917, "grad_norm": 0.6721571418198156, "learning_rate": 7.347890376895407e-05, "loss": 0.586, "step": 483 }, { "epoch": 1.3305841924398625, "grad_norm": 0.6580601865881139, "learning_rate": 7.343672989350214e-05, "loss": 0.5726, "step": 484 }, { "epoch": 1.3333333333333333, "grad_norm": 0.9232629737168565, "learning_rate": 7.33944322663886e-05, "loss": 0.5841, "step": 485 }, { "epoch": 1.3360824742268043, "grad_norm": 0.921611093133066, "learning_rate": 7.335201104415992e-05, "loss": 0.5961, "step": 486 }, { "epoch": 1.338831615120275, "grad_norm": 0.7555439012244609, "learning_rate": 7.330946638381998e-05, "loss": 0.5877, "step": 487 }, { "epoch": 1.3415807560137458, "grad_norm": 0.8027350119891178, "learning_rate": 7.326679844282953e-05, "loss": 0.5917, "step": 488 }, { "epoch": 1.3443298969072166, "grad_norm": 1.1401383602299868, "learning_rate": 7.322400737910558e-05, "loss": 0.5852, "step": 489 }, { "epoch": 1.3470790378006874, "grad_norm": 0.805628145096471, "learning_rate": 7.318109335102083e-05, "loss": 0.5916, "step": 490 }, { "epoch": 1.3498281786941582, "grad_norm": 0.5776321913613638, "learning_rate": 7.31380565174031e-05, "loss": 0.5846, "step": 491 }, { "epoch": 1.352577319587629, "grad_norm": 0.730623963485602, "learning_rate": 7.309489703753471e-05, "loss": 0.5859, "step": 492 }, { "epoch": 1.3553264604810997, "grad_norm": 0.6579197440168587, "learning_rate": 7.305161507115185e-05, "loss": 0.5796, "step": 493 }, { "epoch": 1.3580756013745705, "grad_norm": 0.5630346265554919, "learning_rate": 7.300821077844413e-05, "loss": 0.5822, "step": 494 }, { "epoch": 1.3608247422680413, "grad_norm": 0.7908819829643783, "learning_rate": 7.296468432005382e-05, "loss": 0.5709, "step": 495 }, { "epoch": 1.363573883161512, "grad_norm": 1.1004873982746464, "learning_rate": 7.292103585707541e-05, "loss": 0.5923, "step": 496 }, { "epoch": 1.3663230240549828, "grad_norm": 1.0680631626760493, "learning_rate": 7.287726555105485e-05, "loss": 0.5932, "step": 497 }, { "epoch": 1.3690721649484536, "grad_norm": 0.9112494358591099, "learning_rate": 7.283337356398911e-05, "loss": 0.5785, "step": 498 }, { "epoch": 1.3718213058419244, "grad_norm": 1.0337260282053968, "learning_rate": 7.278936005832549e-05, "loss": 0.5882, "step": 499 }, { "epoch": 1.3745704467353952, "grad_norm": 1.3007806381700902, "learning_rate": 7.274522519696102e-05, "loss": 0.5798, "step": 500 }, { "epoch": 1.377319587628866, "grad_norm": 0.43519904375561846, "learning_rate": 7.270096914324189e-05, "loss": 0.5791, "step": 501 }, { "epoch": 1.3800687285223368, "grad_norm": 0.8544940867252295, "learning_rate": 7.265659206096285e-05, "loss": 0.5901, "step": 502 }, { "epoch": 1.3828178694158075, "grad_norm": 1.2429689674231412, "learning_rate": 7.261209411436654e-05, "loss": 0.5809, "step": 503 }, { "epoch": 1.3855670103092783, "grad_norm": 0.6308748621084995, "learning_rate": 7.256747546814298e-05, "loss": 0.5791, "step": 504 }, { "epoch": 1.388316151202749, "grad_norm": 0.7148204479546915, "learning_rate": 7.252273628742885e-05, "loss": 0.5796, "step": 505 }, { "epoch": 1.3910652920962199, "grad_norm": 0.9925558904292439, "learning_rate": 7.2477876737807e-05, "loss": 0.5858, "step": 506 }, { "epoch": 1.3938144329896907, "grad_norm": 0.8578115038769086, "learning_rate": 7.243289698530572e-05, "loss": 0.5845, "step": 507 }, { "epoch": 1.3965635738831614, "grad_norm": 0.7229862425355326, "learning_rate": 7.23877971963982e-05, "loss": 0.5755, "step": 508 }, { "epoch": 1.3993127147766322, "grad_norm": 0.8939365007653768, "learning_rate": 7.234257753800191e-05, "loss": 0.5875, "step": 509 }, { "epoch": 1.402061855670103, "grad_norm": 0.9280862590900643, "learning_rate": 7.229723817747793e-05, "loss": 0.5916, "step": 510 }, { "epoch": 1.4048109965635738, "grad_norm": 0.7722106893797273, "learning_rate": 7.225177928263042e-05, "loss": 0.5817, "step": 511 }, { "epoch": 1.4075601374570446, "grad_norm": 0.5590378429645152, "learning_rate": 7.220620102170585e-05, "loss": 0.5804, "step": 512 }, { "epoch": 1.4103092783505153, "grad_norm": 0.5415497371424565, "learning_rate": 7.21605035633926e-05, "loss": 0.5829, "step": 513 }, { "epoch": 1.4130584192439861, "grad_norm": 0.6720235507203421, "learning_rate": 7.21146870768201e-05, "loss": 0.5818, "step": 514 }, { "epoch": 1.4158075601374571, "grad_norm": 0.5829950782167737, "learning_rate": 7.206875173155834e-05, "loss": 0.5922, "step": 515 }, { "epoch": 1.418556701030928, "grad_norm": 0.3621229457523989, "learning_rate": 7.202269769761726e-05, "loss": 0.5842, "step": 516 }, { "epoch": 1.4213058419243987, "grad_norm": 0.5023031289705603, "learning_rate": 7.197652514544601e-05, "loss": 0.576, "step": 517 }, { "epoch": 1.4240549828178695, "grad_norm": 0.6099312033435136, "learning_rate": 7.193023424593245e-05, "loss": 0.5795, "step": 518 }, { "epoch": 1.4268041237113402, "grad_norm": 0.5933868127441155, "learning_rate": 7.188382517040237e-05, "loss": 0.5736, "step": 519 }, { "epoch": 1.429553264604811, "grad_norm": 0.582367928707779, "learning_rate": 7.183729809061904e-05, "loss": 0.5794, "step": 520 }, { "epoch": 1.4323024054982818, "grad_norm": 0.7208839572689155, "learning_rate": 7.179065317878237e-05, "loss": 0.587, "step": 521 }, { "epoch": 1.4350515463917526, "grad_norm": 1.0049567135040567, "learning_rate": 7.174389060752845e-05, "loss": 0.5863, "step": 522 }, { "epoch": 1.4378006872852234, "grad_norm": 1.3116471278229, "learning_rate": 7.16970105499288e-05, "loss": 0.5873, "step": 523 }, { "epoch": 1.4405498281786941, "grad_norm": 0.7052150815292837, "learning_rate": 7.165001317948976e-05, "loss": 0.5854, "step": 524 }, { "epoch": 1.443298969072165, "grad_norm": 0.6674498164029434, "learning_rate": 7.160289867015187e-05, "loss": 0.58, "step": 525 }, { "epoch": 1.4460481099656357, "grad_norm": 0.6479049706905784, "learning_rate": 7.155566719628921e-05, "loss": 0.583, "step": 526 }, { "epoch": 1.4487972508591065, "grad_norm": 0.6003355447355602, "learning_rate": 7.150831893270874e-05, "loss": 0.5792, "step": 527 }, { "epoch": 1.4515463917525773, "grad_norm": 0.7215636928866146, "learning_rate": 7.146085405464968e-05, "loss": 0.5746, "step": 528 }, { "epoch": 1.454295532646048, "grad_norm": 0.6022527234482411, "learning_rate": 7.141327273778284e-05, "loss": 0.5808, "step": 529 }, { "epoch": 1.4570446735395188, "grad_norm": 0.44820302089966574, "learning_rate": 7.136557515820999e-05, "loss": 0.5745, "step": 530 }, { "epoch": 1.4597938144329896, "grad_norm": 0.42351586228242205, "learning_rate": 7.13177614924632e-05, "loss": 0.5858, "step": 531 }, { "epoch": 1.4625429553264606, "grad_norm": 0.4466996033955547, "learning_rate": 7.126983191750412e-05, "loss": 0.5746, "step": 532 }, { "epoch": 1.4652920962199314, "grad_norm": 0.40737004869624294, "learning_rate": 7.12217866107235e-05, "loss": 0.5818, "step": 533 }, { "epoch": 1.4680412371134022, "grad_norm": 0.47191169515150533, "learning_rate": 7.11736257499403e-05, "loss": 0.5809, "step": 534 }, { "epoch": 1.470790378006873, "grad_norm": 0.4813204578717719, "learning_rate": 7.112534951340126e-05, "loss": 0.5844, "step": 535 }, { "epoch": 1.4735395189003437, "grad_norm": 0.5549078938525966, "learning_rate": 7.107695807978007e-05, "loss": 0.5812, "step": 536 }, { "epoch": 1.4762886597938145, "grad_norm": 0.6761669063164707, "learning_rate": 7.10284516281768e-05, "loss": 0.5766, "step": 537 }, { "epoch": 1.4790378006872853, "grad_norm": 0.6480701339305898, "learning_rate": 7.097983033811718e-05, "loss": 0.5805, "step": 538 }, { "epoch": 1.481786941580756, "grad_norm": 0.5593193077179679, "learning_rate": 7.0931094389552e-05, "loss": 0.575, "step": 539 }, { "epoch": 1.4845360824742269, "grad_norm": 0.4942110243815134, "learning_rate": 7.088224396285638e-05, "loss": 0.5833, "step": 540 }, { "epoch": 1.4872852233676976, "grad_norm": 0.5797979347815979, "learning_rate": 7.083327923882917e-05, "loss": 0.5747, "step": 541 }, { "epoch": 1.4900343642611684, "grad_norm": 0.9315007373980465, "learning_rate": 7.078420039869218e-05, "loss": 0.592, "step": 542 }, { "epoch": 1.4927835051546392, "grad_norm": 1.616429951355777, "learning_rate": 7.073500762408964e-05, "loss": 0.5944, "step": 543 }, { "epoch": 1.49553264604811, "grad_norm": 0.5055431608140836, "learning_rate": 7.068570109708741e-05, "loss": 0.5901, "step": 544 }, { "epoch": 1.4982817869415808, "grad_norm": 1.0997422339745508, "learning_rate": 7.06362810001724e-05, "loss": 0.5902, "step": 545 }, { "epoch": 1.5010309278350515, "grad_norm": 1.7657567775705625, "learning_rate": 7.05867475162518e-05, "loss": 0.5885, "step": 546 }, { "epoch": 1.5037800687285223, "grad_norm": 0.7356302747726631, "learning_rate": 7.053710082865247e-05, "loss": 0.5861, "step": 547 }, { "epoch": 1.506529209621993, "grad_norm": 2.1759109547514943, "learning_rate": 7.048734112112026e-05, "loss": 0.5997, "step": 548 }, { "epoch": 1.5092783505154639, "grad_norm": 1.2978110027459113, "learning_rate": 7.043746857781933e-05, "loss": 0.6157, "step": 549 }, { "epoch": 1.5120274914089347, "grad_norm": 2.219306449832288, "learning_rate": 7.03874833833314e-05, "loss": 0.607, "step": 550 }, { "epoch": 1.5147766323024054, "grad_norm": 1.9591339401200079, "learning_rate": 7.033738572265517e-05, "loss": 0.6369, "step": 551 }, { "epoch": 1.5175257731958762, "grad_norm": 1.478246394702369, "learning_rate": 7.028717578120555e-05, "loss": 0.6063, "step": 552 }, { "epoch": 1.520274914089347, "grad_norm": 1.2579901505180804, "learning_rate": 7.023685374481301e-05, "loss": 0.6019, "step": 553 }, { "epoch": 1.5230240549828178, "grad_norm": 1.46105914035749, "learning_rate": 7.018641979972295e-05, "loss": 0.6094, "step": 554 }, { "epoch": 1.5257731958762886, "grad_norm": 0.8090152261985227, "learning_rate": 7.013587413259486e-05, "loss": 0.5913, "step": 555 }, { "epoch": 1.5285223367697593, "grad_norm": 1.2609062811324916, "learning_rate": 7.008521693050179e-05, "loss": 0.5856, "step": 556 }, { "epoch": 1.5312714776632301, "grad_norm": 1.0043747397077383, "learning_rate": 7.003444838092957e-05, "loss": 0.5989, "step": 557 }, { "epoch": 1.534020618556701, "grad_norm": 0.7445889690645772, "learning_rate": 6.998356867177613e-05, "loss": 0.5865, "step": 558 }, { "epoch": 1.5367697594501717, "grad_norm": 1.2566147555053582, "learning_rate": 6.993257799135078e-05, "loss": 0.5855, "step": 559 }, { "epoch": 1.5395189003436425, "grad_norm": 0.623650378420435, "learning_rate": 6.988147652837362e-05, "loss": 0.5931, "step": 560 }, { "epoch": 1.5422680412371133, "grad_norm": 1.0346995401552566, "learning_rate": 6.983026447197469e-05, "loss": 0.5914, "step": 561 }, { "epoch": 1.545017182130584, "grad_norm": 0.9620047133783202, "learning_rate": 6.977894201169336e-05, "loss": 0.5915, "step": 562 }, { "epoch": 1.547766323024055, "grad_norm": 0.6073721542597159, "learning_rate": 6.972750933747765e-05, "loss": 0.5834, "step": 563 }, { "epoch": 1.5505154639175258, "grad_norm": 0.9240637148715215, "learning_rate": 6.967596663968346e-05, "loss": 0.5891, "step": 564 }, { "epoch": 1.5532646048109966, "grad_norm": 0.7729700501754245, "learning_rate": 6.962431410907387e-05, "loss": 0.5864, "step": 565 }, { "epoch": 1.5560137457044674, "grad_norm": 0.7023186108460778, "learning_rate": 6.957255193681852e-05, "loss": 0.5792, "step": 566 }, { "epoch": 1.5587628865979382, "grad_norm": 0.8156415963129797, "learning_rate": 6.952068031449278e-05, "loss": 0.5804, "step": 567 }, { "epoch": 1.561512027491409, "grad_norm": 0.4418458963414597, "learning_rate": 6.946869943407714e-05, "loss": 0.5796, "step": 568 }, { "epoch": 1.5642611683848797, "grad_norm": 0.7517403935755494, "learning_rate": 6.941660948795646e-05, "loss": 0.5757, "step": 569 }, { "epoch": 1.5670103092783505, "grad_norm": 0.6880319452283528, "learning_rate": 6.936441066891924e-05, "loss": 0.5866, "step": 570 }, { "epoch": 1.5697594501718213, "grad_norm": 0.44431908466865444, "learning_rate": 6.931210317015694e-05, "loss": 0.5831, "step": 571 }, { "epoch": 1.572508591065292, "grad_norm": 0.5206213485516508, "learning_rate": 6.925968718526325e-05, "loss": 0.5835, "step": 572 }, { "epoch": 1.5752577319587628, "grad_norm": 0.4865149316480254, "learning_rate": 6.920716290823337e-05, "loss": 0.5872, "step": 573 }, { "epoch": 1.5780068728522336, "grad_norm": 0.45702506236242896, "learning_rate": 6.915453053346332e-05, "loss": 0.5918, "step": 574 }, { "epoch": 1.5807560137457046, "grad_norm": 0.44494079428514854, "learning_rate": 6.910179025574916e-05, "loss": 0.5889, "step": 575 }, { "epoch": 1.5835051546391754, "grad_norm": 0.40161179223147797, "learning_rate": 6.904894227028631e-05, "loss": 0.5793, "step": 576 }, { "epoch": 1.5862542955326462, "grad_norm": 0.40020351103635327, "learning_rate": 6.899598677266888e-05, "loss": 0.5791, "step": 577 }, { "epoch": 1.589003436426117, "grad_norm": 0.5533864087207034, "learning_rate": 6.89429239588888e-05, "loss": 0.5686, "step": 578 }, { "epoch": 1.5917525773195877, "grad_norm": 0.4609556358868059, "learning_rate": 6.888975402533528e-05, "loss": 0.5696, "step": 579 }, { "epoch": 1.5945017182130585, "grad_norm": 0.376485016123922, "learning_rate": 6.883647716879389e-05, "loss": 0.5787, "step": 580 }, { "epoch": 1.5972508591065293, "grad_norm": 0.3703464382281221, "learning_rate": 6.878309358644602e-05, "loss": 0.5786, "step": 581 }, { "epoch": 1.6, "grad_norm": 0.36242738510972944, "learning_rate": 6.8729603475868e-05, "loss": 0.5838, "step": 582 }, { "epoch": 1.6027491408934709, "grad_norm": 0.4005050809946878, "learning_rate": 6.867600703503044e-05, "loss": 0.5828, "step": 583 }, { "epoch": 1.6054982817869417, "grad_norm": 0.42329454686286444, "learning_rate": 6.86223044622975e-05, "loss": 0.5803, "step": 584 }, { "epoch": 1.6082474226804124, "grad_norm": 0.3329212652227201, "learning_rate": 6.856849595642617e-05, "loss": 0.5822, "step": 585 }, { "epoch": 1.6109965635738832, "grad_norm": 0.46886870029058014, "learning_rate": 6.851458171656541e-05, "loss": 0.58, "step": 586 }, { "epoch": 1.613745704467354, "grad_norm": 0.5556659975442035, "learning_rate": 6.846056194225562e-05, "loss": 0.5702, "step": 587 }, { "epoch": 1.6164948453608248, "grad_norm": 0.5260721900968394, "learning_rate": 6.84064368334277e-05, "loss": 0.5867, "step": 588 }, { "epoch": 1.6192439862542956, "grad_norm": 0.40786759602060935, "learning_rate": 6.835220659040246e-05, "loss": 0.5834, "step": 589 }, { "epoch": 1.6219931271477663, "grad_norm": 0.2996276604956069, "learning_rate": 6.82978714138898e-05, "loss": 0.5776, "step": 590 }, { "epoch": 1.6247422680412371, "grad_norm": 0.35270615435100505, "learning_rate": 6.824343150498798e-05, "loss": 0.5783, "step": 591 }, { "epoch": 1.627491408934708, "grad_norm": 0.28376863619687237, "learning_rate": 6.81888870651829e-05, "loss": 0.5767, "step": 592 }, { "epoch": 1.6302405498281787, "grad_norm": 0.36443318149486764, "learning_rate": 6.813423829634732e-05, "loss": 0.579, "step": 593 }, { "epoch": 1.6329896907216495, "grad_norm": 0.481183239710988, "learning_rate": 6.80794854007401e-05, "loss": 0.5815, "step": 594 }, { "epoch": 1.6357388316151202, "grad_norm": 0.4948913070938149, "learning_rate": 6.802462858100556e-05, "loss": 0.5847, "step": 595 }, { "epoch": 1.638487972508591, "grad_norm": 0.3815071178544923, "learning_rate": 6.796966804017258e-05, "loss": 0.5724, "step": 596 }, { "epoch": 1.6412371134020618, "grad_norm": 0.4407973877140981, "learning_rate": 6.791460398165393e-05, "loss": 0.5811, "step": 597 }, { "epoch": 1.6439862542955326, "grad_norm": 0.43391575615156724, "learning_rate": 6.785943660924553e-05, "loss": 0.5827, "step": 598 }, { "epoch": 1.6467353951890034, "grad_norm": 0.4035737413547384, "learning_rate": 6.780416612712568e-05, "loss": 0.5819, "step": 599 }, { "epoch": 1.6494845360824741, "grad_norm": 0.5281104884380362, "learning_rate": 6.774879273985423e-05, "loss": 0.5775, "step": 600 }, { "epoch": 1.652233676975945, "grad_norm": 0.7760361977252223, "learning_rate": 6.769331665237198e-05, "loss": 0.5899, "step": 601 }, { "epoch": 1.6549828178694157, "grad_norm": 1.09403271674242, "learning_rate": 6.763773806999977e-05, "loss": 0.5859, "step": 602 }, { "epoch": 1.6577319587628865, "grad_norm": 0.9348914419022682, "learning_rate": 6.75820571984378e-05, "loss": 0.5918, "step": 603 }, { "epoch": 1.6604810996563573, "grad_norm": 0.7435836588260047, "learning_rate": 6.752627424376484e-05, "loss": 0.582, "step": 604 }, { "epoch": 1.663230240549828, "grad_norm": 0.5416202424064285, "learning_rate": 6.747038941243748e-05, "loss": 0.5845, "step": 605 }, { "epoch": 1.6659793814432988, "grad_norm": 0.4289104135369005, "learning_rate": 6.741440291128938e-05, "loss": 0.5841, "step": 606 }, { "epoch": 1.6687285223367696, "grad_norm": 0.3842777168497921, "learning_rate": 6.735831494753046e-05, "loss": 0.5783, "step": 607 }, { "epoch": 1.6714776632302404, "grad_norm": 0.3676465027204242, "learning_rate": 6.730212572874618e-05, "loss": 0.5708, "step": 608 }, { "epoch": 1.6742268041237114, "grad_norm": 0.3960487648888393, "learning_rate": 6.724583546289672e-05, "loss": 0.5724, "step": 609 }, { "epoch": 1.6769759450171822, "grad_norm": 0.4083333422479102, "learning_rate": 6.71894443583163e-05, "loss": 0.5925, "step": 610 }, { "epoch": 1.679725085910653, "grad_norm": 0.4578552218287685, "learning_rate": 6.713295262371232e-05, "loss": 0.5715, "step": 611 }, { "epoch": 1.6824742268041237, "grad_norm": 0.5656117537132986, "learning_rate": 6.707636046816457e-05, "loss": 0.5766, "step": 612 }, { "epoch": 1.6852233676975945, "grad_norm": 0.6850771753028508, "learning_rate": 6.70196681011246e-05, "loss": 0.5928, "step": 613 }, { "epoch": 1.6879725085910653, "grad_norm": 0.698725362743093, "learning_rate": 6.696287573241478e-05, "loss": 0.5897, "step": 614 }, { "epoch": 1.690721649484536, "grad_norm": 0.7382677867943621, "learning_rate": 6.690598357222764e-05, "loss": 0.5668, "step": 615 }, { "epoch": 1.6934707903780069, "grad_norm": 0.7105270810835767, "learning_rate": 6.684899183112498e-05, "loss": 0.5736, "step": 616 }, { "epoch": 1.6962199312714776, "grad_norm": 0.6862323082435818, "learning_rate": 6.679190072003722e-05, "loss": 0.5765, "step": 617 }, { "epoch": 1.6989690721649484, "grad_norm": 0.6299104085981699, "learning_rate": 6.67347104502625e-05, "loss": 0.5642, "step": 618 }, { "epoch": 1.7017182130584192, "grad_norm": 0.6026002947844302, "learning_rate": 6.667742123346603e-05, "loss": 0.5804, "step": 619 }, { "epoch": 1.7044673539518902, "grad_norm": 0.4462813007489562, "learning_rate": 6.662003328167917e-05, "loss": 0.5807, "step": 620 }, { "epoch": 1.707216494845361, "grad_norm": 0.29472463703813456, "learning_rate": 6.65625468072987e-05, "loss": 0.5702, "step": 621 }, { "epoch": 1.7099656357388318, "grad_norm": 0.320893033856143, "learning_rate": 6.650496202308607e-05, "loss": 0.5835, "step": 622 }, { "epoch": 1.7127147766323025, "grad_norm": 0.39010773443389357, "learning_rate": 6.644727914216656e-05, "loss": 0.576, "step": 623 }, { "epoch": 1.7154639175257733, "grad_norm": 0.5319160826521161, "learning_rate": 6.638949837802855e-05, "loss": 0.5798, "step": 624 }, { "epoch": 1.718213058419244, "grad_norm": 0.6498383386047935, "learning_rate": 6.633161994452262e-05, "loss": 0.5795, "step": 625 }, { "epoch": 1.7209621993127149, "grad_norm": 0.7259267311718851, "learning_rate": 6.627364405586091e-05, "loss": 0.5919, "step": 626 }, { "epoch": 1.7237113402061857, "grad_norm": 0.8162977870752441, "learning_rate": 6.621557092661621e-05, "loss": 0.5753, "step": 627 }, { "epoch": 1.7264604810996564, "grad_norm": 0.7728017259091454, "learning_rate": 6.61574007717212e-05, "loss": 0.5843, "step": 628 }, { "epoch": 1.7292096219931272, "grad_norm": 0.7443225127869316, "learning_rate": 6.609913380646767e-05, "loss": 0.5802, "step": 629 }, { "epoch": 1.731958762886598, "grad_norm": 0.7541481509168282, "learning_rate": 6.604077024650571e-05, "loss": 0.5841, "step": 630 }, { "epoch": 1.7347079037800688, "grad_norm": 0.6874877805251565, "learning_rate": 6.598231030784289e-05, "loss": 0.5861, "step": 631 }, { "epoch": 1.7374570446735396, "grad_norm": 0.6302784745574853, "learning_rate": 6.592375420684354e-05, "loss": 0.579, "step": 632 }, { "epoch": 1.7402061855670103, "grad_norm": 0.3605005517279551, "learning_rate": 6.586510216022785e-05, "loss": 0.5761, "step": 633 }, { "epoch": 1.7429553264604811, "grad_norm": 0.40851626295602705, "learning_rate": 6.58063543850711e-05, "loss": 0.5809, "step": 634 }, { "epoch": 1.745704467353952, "grad_norm": 0.5889121925134353, "learning_rate": 6.574751109880295e-05, "loss": 0.589, "step": 635 }, { "epoch": 1.7484536082474227, "grad_norm": 0.6855740977784578, "learning_rate": 6.568857251920641e-05, "loss": 0.5791, "step": 636 }, { "epoch": 1.7512027491408935, "grad_norm": 0.7763253623405283, "learning_rate": 6.562953886441731e-05, "loss": 0.5698, "step": 637 }, { "epoch": 1.7539518900343642, "grad_norm": 0.8027574724353924, "learning_rate": 6.557041035292331e-05, "loss": 0.5695, "step": 638 }, { "epoch": 1.756701030927835, "grad_norm": 0.8124682042384597, "learning_rate": 6.551118720356313e-05, "loss": 0.5793, "step": 639 }, { "epoch": 1.7594501718213058, "grad_norm": 0.7285676570172558, "learning_rate": 6.545186963552578e-05, "loss": 0.5809, "step": 640 }, { "epoch": 1.7621993127147766, "grad_norm": 0.5493034420656492, "learning_rate": 6.539245786834965e-05, "loss": 0.5779, "step": 641 }, { "epoch": 1.7649484536082474, "grad_norm": 0.4100237013175215, "learning_rate": 6.533295212192189e-05, "loss": 0.5634, "step": 642 }, { "epoch": 1.7676975945017182, "grad_norm": 0.44386016701061204, "learning_rate": 6.527335261647735e-05, "loss": 0.5865, "step": 643 }, { "epoch": 1.770446735395189, "grad_norm": 0.7605693123034108, "learning_rate": 6.521365957259798e-05, "loss": 0.5758, "step": 644 }, { "epoch": 1.7731958762886597, "grad_norm": 0.9531489941794372, "learning_rate": 6.515387321121185e-05, "loss": 0.5845, "step": 645 }, { "epoch": 1.7759450171821305, "grad_norm": 0.9748291614843003, "learning_rate": 6.509399375359247e-05, "loss": 0.5719, "step": 646 }, { "epoch": 1.7786941580756013, "grad_norm": 0.7628578645914652, "learning_rate": 6.503402142135787e-05, "loss": 0.5768, "step": 647 }, { "epoch": 1.781443298969072, "grad_norm": 0.5799935111185708, "learning_rate": 6.497395643646981e-05, "loss": 0.5701, "step": 648 }, { "epoch": 1.7841924398625428, "grad_norm": 0.5277145245316653, "learning_rate": 6.491379902123297e-05, "loss": 0.5683, "step": 649 }, { "epoch": 1.7869415807560136, "grad_norm": 0.4736528081812741, "learning_rate": 6.485354939829414e-05, "loss": 0.5779, "step": 650 }, { "epoch": 1.7896907216494844, "grad_norm": 0.5669836094412534, "learning_rate": 6.479320779064134e-05, "loss": 0.5764, "step": 651 }, { "epoch": 1.7924398625429552, "grad_norm": 0.559780538593531, "learning_rate": 6.473277442160309e-05, "loss": 0.5832, "step": 652 }, { "epoch": 1.795189003436426, "grad_norm": 0.5065885370576576, "learning_rate": 6.467224951484746e-05, "loss": 0.5783, "step": 653 }, { "epoch": 1.797938144329897, "grad_norm": 0.4559459768927469, "learning_rate": 6.461163329438135e-05, "loss": 0.5715, "step": 654 }, { "epoch": 1.8006872852233677, "grad_norm": 0.42548224770937315, "learning_rate": 6.45509259845496e-05, "loss": 0.5751, "step": 655 }, { "epoch": 1.8034364261168385, "grad_norm": 0.3455007524085001, "learning_rate": 6.449012781003419e-05, "loss": 0.5754, "step": 656 }, { "epoch": 1.8061855670103093, "grad_norm": 0.4400225976066889, "learning_rate": 6.442923899585338e-05, "loss": 0.5941, "step": 657 }, { "epoch": 1.80893470790378, "grad_norm": 0.5084943205879741, "learning_rate": 6.436825976736092e-05, "loss": 0.579, "step": 658 }, { "epoch": 1.8116838487972509, "grad_norm": 0.5358918155905726, "learning_rate": 6.430719035024515e-05, "loss": 0.5748, "step": 659 }, { "epoch": 1.8144329896907216, "grad_norm": 0.503047756943303, "learning_rate": 6.424603097052826e-05, "loss": 0.5794, "step": 660 }, { "epoch": 1.8171821305841924, "grad_norm": 0.44321729048134617, "learning_rate": 6.418478185456534e-05, "loss": 0.5865, "step": 661 }, { "epoch": 1.8199312714776632, "grad_norm": 0.3737976145350619, "learning_rate": 6.412344322904362e-05, "loss": 0.5762, "step": 662 }, { "epoch": 1.822680412371134, "grad_norm": 0.2731612660975183, "learning_rate": 6.406201532098165e-05, "loss": 0.5691, "step": 663 }, { "epoch": 1.8254295532646048, "grad_norm": 0.3311008105208644, "learning_rate": 6.400049835772836e-05, "loss": 0.5729, "step": 664 }, { "epoch": 1.8281786941580758, "grad_norm": 0.3627533269491876, "learning_rate": 6.393889256696231e-05, "loss": 0.5729, "step": 665 }, { "epoch": 1.8309278350515465, "grad_norm": 0.3316310187075983, "learning_rate": 6.387719817669081e-05, "loss": 0.582, "step": 666 }, { "epoch": 1.8336769759450173, "grad_norm": 0.3431221379909935, "learning_rate": 6.381541541524911e-05, "loss": 0.5775, "step": 667 }, { "epoch": 1.836426116838488, "grad_norm": 0.308724522020036, "learning_rate": 6.375354451129949e-05, "loss": 0.5759, "step": 668 }, { "epoch": 1.839175257731959, "grad_norm": 0.2799060976007718, "learning_rate": 6.369158569383045e-05, "loss": 0.5826, "step": 669 }, { "epoch": 1.8419243986254297, "grad_norm": 0.3820386705421886, "learning_rate": 6.362953919215591e-05, "loss": 0.5778, "step": 670 }, { "epoch": 1.8446735395189005, "grad_norm": 0.38246444479750036, "learning_rate": 6.356740523591427e-05, "loss": 0.5749, "step": 671 }, { "epoch": 1.8474226804123712, "grad_norm": 0.3271098200229515, "learning_rate": 6.350518405506764e-05, "loss": 0.5817, "step": 672 }, { "epoch": 1.850171821305842, "grad_norm": 0.34228419827242806, "learning_rate": 6.344287587990091e-05, "loss": 0.5819, "step": 673 }, { "epoch": 1.8529209621993128, "grad_norm": 0.32655548951800334, "learning_rate": 6.338048094102096e-05, "loss": 0.5749, "step": 674 }, { "epoch": 1.8556701030927836, "grad_norm": 0.32871268032009765, "learning_rate": 6.331799946935583e-05, "loss": 0.5685, "step": 675 }, { "epoch": 1.8584192439862544, "grad_norm": 0.4199684721453658, "learning_rate": 6.325543169615378e-05, "loss": 0.5725, "step": 676 }, { "epoch": 1.8611683848797251, "grad_norm": 0.5756048364369786, "learning_rate": 6.319277785298247e-05, "loss": 0.5805, "step": 677 }, { "epoch": 1.863917525773196, "grad_norm": 0.6547196962591697, "learning_rate": 6.313003817172812e-05, "loss": 0.5662, "step": 678 }, { "epoch": 1.8666666666666667, "grad_norm": 0.7656825534443144, "learning_rate": 6.30672128845947e-05, "loss": 0.5864, "step": 679 }, { "epoch": 1.8694158075601375, "grad_norm": 0.9358535366007116, "learning_rate": 6.300430222410292e-05, "loss": 0.5798, "step": 680 }, { "epoch": 1.8721649484536083, "grad_norm": 0.9703236167568001, "learning_rate": 6.294130642308952e-05, "loss": 0.5752, "step": 681 }, { "epoch": 1.874914089347079, "grad_norm": 0.8892433252848201, "learning_rate": 6.287822571470636e-05, "loss": 0.5709, "step": 682 }, { "epoch": 1.8776632302405498, "grad_norm": 0.8810114036796293, "learning_rate": 6.281506033241948e-05, "loss": 0.5733, "step": 683 }, { "epoch": 1.8804123711340206, "grad_norm": 0.853242199524319, "learning_rate": 6.275181051000842e-05, "loss": 0.5758, "step": 684 }, { "epoch": 1.8831615120274914, "grad_norm": 0.6715868178268443, "learning_rate": 6.26884764815651e-05, "loss": 0.5681, "step": 685 }, { "epoch": 1.8859106529209622, "grad_norm": 0.41927983624654164, "learning_rate": 6.26250584814932e-05, "loss": 0.5706, "step": 686 }, { "epoch": 1.888659793814433, "grad_norm": 0.35123714928190247, "learning_rate": 6.256155674450712e-05, "loss": 0.5742, "step": 687 }, { "epoch": 1.8914089347079037, "grad_norm": 0.5653996664319825, "learning_rate": 6.249797150563124e-05, "loss": 0.5731, "step": 688 }, { "epoch": 1.8941580756013745, "grad_norm": 0.6649259671430193, "learning_rate": 6.243430300019891e-05, "loss": 0.5797, "step": 689 }, { "epoch": 1.8969072164948453, "grad_norm": 0.60181480524739, "learning_rate": 6.237055146385173e-05, "loss": 0.5758, "step": 690 }, { "epoch": 1.899656357388316, "grad_norm": 0.6235277209689051, "learning_rate": 6.230671713253854e-05, "loss": 0.5778, "step": 691 }, { "epoch": 1.9024054982817868, "grad_norm": 0.6196187849528243, "learning_rate": 6.224280024251466e-05, "loss": 0.5698, "step": 692 }, { "epoch": 1.9051546391752576, "grad_norm": 0.42709196066527666, "learning_rate": 6.217880103034091e-05, "loss": 0.5736, "step": 693 }, { "epoch": 1.9079037800687284, "grad_norm": 0.34692776304267553, "learning_rate": 6.211471973288285e-05, "loss": 0.5712, "step": 694 }, { "epoch": 1.9106529209621992, "grad_norm": 0.3110542487628964, "learning_rate": 6.205055658730983e-05, "loss": 0.5653, "step": 695 }, { "epoch": 1.91340206185567, "grad_norm": 0.34540933819310177, "learning_rate": 6.198631183109408e-05, "loss": 0.576, "step": 696 }, { "epoch": 1.9161512027491407, "grad_norm": 0.3126265820113025, "learning_rate": 6.192198570200992e-05, "loss": 0.5906, "step": 697 }, { "epoch": 1.9189003436426115, "grad_norm": 0.29077210119155705, "learning_rate": 6.185757843813282e-05, "loss": 0.5799, "step": 698 }, { "epoch": 1.9216494845360823, "grad_norm": 0.43001232514386034, "learning_rate": 6.179309027783856e-05, "loss": 0.5784, "step": 699 }, { "epoch": 1.9243986254295533, "grad_norm": 0.4962679375465972, "learning_rate": 6.172852145980228e-05, "loss": 0.5731, "step": 700 }, { "epoch": 1.927147766323024, "grad_norm": 0.5311714605519653, "learning_rate": 6.166387222299767e-05, "loss": 0.5812, "step": 701 }, { "epoch": 1.9298969072164949, "grad_norm": 0.586368926903193, "learning_rate": 6.159914280669607e-05, "loss": 0.5638, "step": 702 }, { "epoch": 1.9326460481099657, "grad_norm": 0.6549081294272183, "learning_rate": 6.153433345046552e-05, "loss": 0.5785, "step": 703 }, { "epoch": 1.9353951890034364, "grad_norm": 0.7252681437363886, "learning_rate": 6.146944439416997e-05, "loss": 0.5751, "step": 704 }, { "epoch": 1.9381443298969072, "grad_norm": 0.7845460102188683, "learning_rate": 6.140447587796832e-05, "loss": 0.5646, "step": 705 }, { "epoch": 1.940893470790378, "grad_norm": 0.8048678290944055, "learning_rate": 6.133942814231357e-05, "loss": 0.5742, "step": 706 }, { "epoch": 1.9436426116838488, "grad_norm": 0.7734169093084938, "learning_rate": 6.127430142795192e-05, "loss": 0.5876, "step": 707 }, { "epoch": 1.9463917525773196, "grad_norm": 0.5484007054312127, "learning_rate": 6.120909597592185e-05, "loss": 0.5843, "step": 708 }, { "epoch": 1.9491408934707903, "grad_norm": 0.3242608423209931, "learning_rate": 6.114381202755328e-05, "loss": 0.5654, "step": 709 }, { "epoch": 1.9518900343642611, "grad_norm": 0.3539205469330434, "learning_rate": 6.107844982446662e-05, "loss": 0.5833, "step": 710 }, { "epoch": 1.9546391752577321, "grad_norm": 0.5156015064072683, "learning_rate": 6.1013009608571954e-05, "loss": 0.583, "step": 711 }, { "epoch": 1.957388316151203, "grad_norm": 0.6131503596319059, "learning_rate": 6.094749162206807e-05, "loss": 0.5742, "step": 712 }, { "epoch": 1.9601374570446737, "grad_norm": 0.6333732757531064, "learning_rate": 6.0881896107441575e-05, "loss": 0.5751, "step": 713 }, { "epoch": 1.9628865979381445, "grad_norm": 0.456390717989884, "learning_rate": 6.081622330746603e-05, "loss": 0.5844, "step": 714 }, { "epoch": 1.9656357388316152, "grad_norm": 0.41591397010146025, "learning_rate": 6.075047346520105e-05, "loss": 0.5756, "step": 715 }, { "epoch": 1.968384879725086, "grad_norm": 0.5054704227601636, "learning_rate": 6.0684646823991345e-05, "loss": 0.5688, "step": 716 }, { "epoch": 1.9711340206185568, "grad_norm": 0.6526494286710874, "learning_rate": 6.061874362746592e-05, "loss": 0.568, "step": 717 }, { "epoch": 1.9738831615120276, "grad_norm": 0.7155431076783062, "learning_rate": 6.055276411953705e-05, "loss": 0.5746, "step": 718 }, { "epoch": 1.9766323024054984, "grad_norm": 0.5467249394073967, "learning_rate": 6.048670854439952e-05, "loss": 0.5793, "step": 719 }, { "epoch": 1.9793814432989691, "grad_norm": 0.44545561246216775, "learning_rate": 6.042057714652958e-05, "loss": 0.5793, "step": 720 }, { "epoch": 1.98213058419244, "grad_norm": 0.43668565117704844, "learning_rate": 6.035437017068413e-05, "loss": 0.5711, "step": 721 }, { "epoch": 1.9848797250859107, "grad_norm": 0.4794911786968242, "learning_rate": 6.02880878618998e-05, "loss": 0.5764, "step": 722 }, { "epoch": 1.9876288659793815, "grad_norm": 0.45936527802619287, "learning_rate": 6.022173046549201e-05, "loss": 0.5728, "step": 723 }, { "epoch": 1.9903780068728523, "grad_norm": 0.3710590792543285, "learning_rate": 6.0155298227054125e-05, "loss": 0.5764, "step": 724 }, { "epoch": 1.993127147766323, "grad_norm": 0.3684226778884953, "learning_rate": 6.008879139245645e-05, "loss": 0.5711, "step": 725 }, { "epoch": 1.9958762886597938, "grad_norm": 0.37506727135987533, "learning_rate": 6.00222102078454e-05, "loss": 0.5655, "step": 726 }, { "epoch": 1.9986254295532646, "grad_norm": 0.42203860157012285, "learning_rate": 5.995555491964259e-05, "loss": 0.6314, "step": 727 }, { "epoch": 2.0013745704467354, "grad_norm": 0.6121901786626351, "learning_rate": 5.988882577454386e-05, "loss": 0.6437, "step": 728 }, { "epoch": 2.004123711340206, "grad_norm": 0.838029390831891, "learning_rate": 5.982202301951841e-05, "loss": 0.5246, "step": 729 }, { "epoch": 2.006872852233677, "grad_norm": 1.0035111815398425, "learning_rate": 5.97551469018079e-05, "loss": 0.532, "step": 730 }, { "epoch": 2.0096219931271477, "grad_norm": 1.0550172252909944, "learning_rate": 5.968819766892546e-05, "loss": 0.5356, "step": 731 }, { "epoch": 2.0123711340206185, "grad_norm": 0.7212990022902317, "learning_rate": 5.962117556865489e-05, "loss": 0.5149, "step": 732 }, { "epoch": 2.0151202749140893, "grad_norm": 0.5084983736290086, "learning_rate": 5.955408084904962e-05, "loss": 0.5248, "step": 733 }, { "epoch": 2.01786941580756, "grad_norm": 0.508857242926003, "learning_rate": 5.948691375843187e-05, "loss": 0.5309, "step": 734 }, { "epoch": 2.020618556701031, "grad_norm": 0.5715174569632097, "learning_rate": 5.941967454539171e-05, "loss": 0.5247, "step": 735 }, { "epoch": 2.0233676975945016, "grad_norm": 0.5565322020849539, "learning_rate": 5.935236345878616e-05, "loss": 0.5175, "step": 736 }, { "epoch": 2.0261168384879724, "grad_norm": 0.5745604042771598, "learning_rate": 5.928498074773823e-05, "loss": 0.517, "step": 737 }, { "epoch": 2.028865979381443, "grad_norm": 0.500350720005458, "learning_rate": 5.921752666163598e-05, "loss": 0.5215, "step": 738 }, { "epoch": 2.031615120274914, "grad_norm": 0.4067063681679997, "learning_rate": 5.9150001450131716e-05, "loss": 0.5212, "step": 739 }, { "epoch": 2.0343642611683848, "grad_norm": 0.4770810449857095, "learning_rate": 5.9082405363140916e-05, "loss": 0.521, "step": 740 }, { "epoch": 2.0371134020618555, "grad_norm": 0.5487285643494075, "learning_rate": 5.901473865084141e-05, "loss": 0.5268, "step": 741 }, { "epoch": 2.0398625429553263, "grad_norm": 0.4395320063115494, "learning_rate": 5.894700156367238e-05, "loss": 0.5243, "step": 742 }, { "epoch": 2.042611683848797, "grad_norm": 0.29982074556969307, "learning_rate": 5.887919435233352e-05, "loss": 0.5298, "step": 743 }, { "epoch": 2.045360824742268, "grad_norm": 0.41119709116602937, "learning_rate": 5.881131726778402e-05, "loss": 0.5145, "step": 744 }, { "epoch": 2.0481099656357387, "grad_norm": 0.384320461752947, "learning_rate": 5.8743370561241715e-05, "loss": 0.5253, "step": 745 }, { "epoch": 2.0508591065292094, "grad_norm": 0.3420792544611106, "learning_rate": 5.867535448418203e-05, "loss": 0.5183, "step": 746 }, { "epoch": 2.05360824742268, "grad_norm": 0.36905823023095596, "learning_rate": 5.8607269288337244e-05, "loss": 0.5208, "step": 747 }, { "epoch": 2.056357388316151, "grad_norm": 0.32511574569133067, "learning_rate": 5.853911522569538e-05, "loss": 0.5252, "step": 748 }, { "epoch": 2.059106529209622, "grad_norm": 0.42000247500887966, "learning_rate": 5.8470892548499384e-05, "loss": 0.5182, "step": 749 }, { "epoch": 2.0618556701030926, "grad_norm": 0.5749785701633202, "learning_rate": 5.840260150924609e-05, "loss": 0.5172, "step": 750 }, { "epoch": 2.064604810996564, "grad_norm": 0.6483835292108968, "learning_rate": 5.833424236068541e-05, "loss": 0.5336, "step": 751 }, { "epoch": 2.0673539518900346, "grad_norm": 0.6231601301759575, "learning_rate": 5.8265815355819284e-05, "loss": 0.527, "step": 752 }, { "epoch": 2.0701030927835053, "grad_norm": 0.6400482537396563, "learning_rate": 5.819732074790084e-05, "loss": 0.5187, "step": 753 }, { "epoch": 2.072852233676976, "grad_norm": 0.6112696586750945, "learning_rate": 5.812875879043336e-05, "loss": 0.5198, "step": 754 }, { "epoch": 2.075601374570447, "grad_norm": 0.5115706970358354, "learning_rate": 5.8060129737169416e-05, "loss": 0.5279, "step": 755 }, { "epoch": 2.0783505154639177, "grad_norm": 0.4058486284297744, "learning_rate": 5.799143384210992e-05, "loss": 0.5268, "step": 756 }, { "epoch": 2.0810996563573885, "grad_norm": 0.32047749300200185, "learning_rate": 5.7922671359503145e-05, "loss": 0.5252, "step": 757 }, { "epoch": 2.0838487972508593, "grad_norm": 0.37004296113308727, "learning_rate": 5.785384254384382e-05, "loss": 0.5239, "step": 758 }, { "epoch": 2.08659793814433, "grad_norm": 0.4458786395281163, "learning_rate": 5.7784947649872176e-05, "loss": 0.5294, "step": 759 }, { "epoch": 2.089347079037801, "grad_norm": 0.4272640389491108, "learning_rate": 5.7715986932573e-05, "loss": 0.5249, "step": 760 }, { "epoch": 2.0920962199312716, "grad_norm": 0.3526739356483497, "learning_rate": 5.7646960647174706e-05, "loss": 0.5184, "step": 761 }, { "epoch": 2.0948453608247424, "grad_norm": 0.27188265621194263, "learning_rate": 5.757786904914838e-05, "loss": 0.5223, "step": 762 }, { "epoch": 2.097594501718213, "grad_norm": 0.3457734998315455, "learning_rate": 5.750871239420681e-05, "loss": 0.5227, "step": 763 }, { "epoch": 2.100343642611684, "grad_norm": 0.4222137704724379, "learning_rate": 5.7439490938303604e-05, "loss": 0.5273, "step": 764 }, { "epoch": 2.1030927835051547, "grad_norm": 0.45756362105195786, "learning_rate": 5.7370204937632167e-05, "loss": 0.5222, "step": 765 }, { "epoch": 2.1058419243986255, "grad_norm": 0.3435489204211452, "learning_rate": 5.7300854648624824e-05, "loss": 0.5102, "step": 766 }, { "epoch": 2.1085910652920963, "grad_norm": 0.24611635871938473, "learning_rate": 5.723144032795179e-05, "loss": 0.5221, "step": 767 }, { "epoch": 2.111340206185567, "grad_norm": 0.3013065227913591, "learning_rate": 5.7161962232520304e-05, "loss": 0.532, "step": 768 }, { "epoch": 2.114089347079038, "grad_norm": 0.3470862495086026, "learning_rate": 5.7092420619473626e-05, "loss": 0.5258, "step": 769 }, { "epoch": 2.1168384879725086, "grad_norm": 0.38943011157017127, "learning_rate": 5.70228157461901e-05, "loss": 0.5264, "step": 770 }, { "epoch": 2.1195876288659794, "grad_norm": 0.39287644862912424, "learning_rate": 5.69531478702822e-05, "loss": 0.5114, "step": 771 }, { "epoch": 2.12233676975945, "grad_norm": 0.2916898291723961, "learning_rate": 5.688341724959557e-05, "loss": 0.5275, "step": 772 }, { "epoch": 2.125085910652921, "grad_norm": 0.22962256743492102, "learning_rate": 5.681362414220811e-05, "loss": 0.5214, "step": 773 }, { "epoch": 2.1278350515463917, "grad_norm": 0.27944270631912393, "learning_rate": 5.674376880642893e-05, "loss": 0.5264, "step": 774 }, { "epoch": 2.1305841924398625, "grad_norm": 0.29044286341885533, "learning_rate": 5.66738515007975e-05, "loss": 0.5148, "step": 775 }, { "epoch": 2.1333333333333333, "grad_norm": 0.28701568805373207, "learning_rate": 5.6603872484082614e-05, "loss": 0.5218, "step": 776 }, { "epoch": 2.136082474226804, "grad_norm": 0.33569332451853506, "learning_rate": 5.653383201528151e-05, "loss": 0.5244, "step": 777 }, { "epoch": 2.138831615120275, "grad_norm": 0.2997007948052843, "learning_rate": 5.6463730353618795e-05, "loss": 0.5207, "step": 778 }, { "epoch": 2.1415807560137456, "grad_norm": 0.21237174047717156, "learning_rate": 5.6393567758545616e-05, "loss": 0.5103, "step": 779 }, { "epoch": 2.1443298969072164, "grad_norm": 0.2360600600158792, "learning_rate": 5.63233444897386e-05, "loss": 0.5293, "step": 780 }, { "epoch": 2.147079037800687, "grad_norm": 0.23546258348715238, "learning_rate": 5.625306080709895e-05, "loss": 0.526, "step": 781 }, { "epoch": 2.149828178694158, "grad_norm": 0.23742640059602604, "learning_rate": 5.618271697075147e-05, "loss": 0.5207, "step": 782 }, { "epoch": 2.1525773195876288, "grad_norm": 0.2697802862240549, "learning_rate": 5.611231324104358e-05, "loss": 0.5185, "step": 783 }, { "epoch": 2.1553264604810995, "grad_norm": 0.33962286687598364, "learning_rate": 5.604184987854437e-05, "loss": 0.5324, "step": 784 }, { "epoch": 2.1580756013745703, "grad_norm": 0.3496574863191211, "learning_rate": 5.597132714404366e-05, "loss": 0.5149, "step": 785 }, { "epoch": 2.160824742268041, "grad_norm": 0.3002103817256998, "learning_rate": 5.590074529855099e-05, "loss": 0.5271, "step": 786 }, { "epoch": 2.163573883161512, "grad_norm": 0.33431676858651316, "learning_rate": 5.583010460329465e-05, "loss": 0.5304, "step": 787 }, { "epoch": 2.1663230240549827, "grad_norm": 0.3298687279509311, "learning_rate": 5.5759405319720806e-05, "loss": 0.5167, "step": 788 }, { "epoch": 2.1690721649484535, "grad_norm": 0.21499388088920707, "learning_rate": 5.568864770949237e-05, "loss": 0.536, "step": 789 }, { "epoch": 2.1718213058419242, "grad_norm": 0.26397548444272334, "learning_rate": 5.5617832034488236e-05, "loss": 0.5251, "step": 790 }, { "epoch": 2.174570446735395, "grad_norm": 0.40340132985540694, "learning_rate": 5.554695855680209e-05, "loss": 0.5283, "step": 791 }, { "epoch": 2.177319587628866, "grad_norm": 0.4098335019644846, "learning_rate": 5.547602753874163e-05, "loss": 0.5238, "step": 792 }, { "epoch": 2.1800687285223366, "grad_norm": 0.3531497244525965, "learning_rate": 5.540503924282746e-05, "loss": 0.5231, "step": 793 }, { "epoch": 2.1828178694158074, "grad_norm": 0.38987792040546504, "learning_rate": 5.5333993931792224e-05, "loss": 0.5366, "step": 794 }, { "epoch": 2.1855670103092786, "grad_norm": 0.2647915747419579, "learning_rate": 5.526289186857953e-05, "loss": 0.5348, "step": 795 }, { "epoch": 2.1883161512027494, "grad_norm": 0.26788341550791417, "learning_rate": 5.5191733316343074e-05, "loss": 0.5253, "step": 796 }, { "epoch": 2.19106529209622, "grad_norm": 0.31976809906665954, "learning_rate": 5.5120518538445595e-05, "loss": 0.5225, "step": 797 }, { "epoch": 2.193814432989691, "grad_norm": 0.27229292264708094, "learning_rate": 5.504924779845794e-05, "loss": 0.5169, "step": 798 }, { "epoch": 2.1965635738831617, "grad_norm": 0.3204542340097227, "learning_rate": 5.4977921360158096e-05, "loss": 0.5256, "step": 799 }, { "epoch": 2.1993127147766325, "grad_norm": 0.33648692429827426, "learning_rate": 5.490653948753013e-05, "loss": 0.523, "step": 800 }, { "epoch": 2.2020618556701033, "grad_norm": 0.25849474311106246, "learning_rate": 5.483510244476332e-05, "loss": 0.5345, "step": 801 }, { "epoch": 2.204810996563574, "grad_norm": 0.2777163411097152, "learning_rate": 5.476361049625115e-05, "loss": 0.5217, "step": 802 }, { "epoch": 2.207560137457045, "grad_norm": 0.259120862015438, "learning_rate": 5.469206390659028e-05, "loss": 0.5293, "step": 803 }, { "epoch": 2.2103092783505156, "grad_norm": 0.25014178976390844, "learning_rate": 5.46204629405796e-05, "loss": 0.5269, "step": 804 }, { "epoch": 2.2130584192439864, "grad_norm": 0.223711086431831, "learning_rate": 5.454880786321928e-05, "loss": 0.5187, "step": 805 }, { "epoch": 2.215807560137457, "grad_norm": 0.2438903572639829, "learning_rate": 5.447709893970974e-05, "loss": 0.5209, "step": 806 }, { "epoch": 2.218556701030928, "grad_norm": 0.2545794798823825, "learning_rate": 5.4405336435450684e-05, "loss": 0.527, "step": 807 }, { "epoch": 2.2213058419243987, "grad_norm": 0.18607426816746073, "learning_rate": 5.4333520616040116e-05, "loss": 0.5184, "step": 808 }, { "epoch": 2.2240549828178695, "grad_norm": 0.29001196046138, "learning_rate": 5.42616517472734e-05, "loss": 0.5195, "step": 809 }, { "epoch": 2.2268041237113403, "grad_norm": 0.2664754443847999, "learning_rate": 5.418973009514221e-05, "loss": 0.5254, "step": 810 }, { "epoch": 2.229553264604811, "grad_norm": 0.2763244689974836, "learning_rate": 5.411775592583358e-05, "loss": 0.5225, "step": 811 }, { "epoch": 2.232302405498282, "grad_norm": 0.2952802637754747, "learning_rate": 5.4045729505728884e-05, "loss": 0.5227, "step": 812 }, { "epoch": 2.2350515463917526, "grad_norm": 0.23994298354796337, "learning_rate": 5.397365110140295e-05, "loss": 0.5305, "step": 813 }, { "epoch": 2.2378006872852234, "grad_norm": 0.22190622512563482, "learning_rate": 5.390152097962295e-05, "loss": 0.532, "step": 814 }, { "epoch": 2.240549828178694, "grad_norm": 0.2958910982821988, "learning_rate": 5.382933940734747e-05, "loss": 0.5215, "step": 815 }, { "epoch": 2.243298969072165, "grad_norm": 0.3582968358435383, "learning_rate": 5.375710665172554e-05, "loss": 0.5206, "step": 816 }, { "epoch": 2.2460481099656358, "grad_norm": 0.4048685164576461, "learning_rate": 5.368482298009559e-05, "loss": 0.5322, "step": 817 }, { "epoch": 2.2487972508591065, "grad_norm": 0.35928125952347756, "learning_rate": 5.3612488659984534e-05, "loss": 0.5256, "step": 818 }, { "epoch": 2.2515463917525773, "grad_norm": 0.27863313688997904, "learning_rate": 5.3540103959106696e-05, "loss": 0.5243, "step": 819 }, { "epoch": 2.254295532646048, "grad_norm": 0.3504201948483768, "learning_rate": 5.3467669145362914e-05, "loss": 0.5369, "step": 820 }, { "epoch": 2.257044673539519, "grad_norm": 0.4374644885930723, "learning_rate": 5.339518448683945e-05, "loss": 0.5206, "step": 821 }, { "epoch": 2.2597938144329897, "grad_norm": 0.2820104015984913, "learning_rate": 5.332265025180707e-05, "loss": 0.5201, "step": 822 }, { "epoch": 2.2625429553264604, "grad_norm": 0.3051225307345294, "learning_rate": 5.325006670872002e-05, "loss": 0.5248, "step": 823 }, { "epoch": 2.265292096219931, "grad_norm": 0.41761093729690574, "learning_rate": 5.3177434126215034e-05, "loss": 0.5378, "step": 824 }, { "epoch": 2.268041237113402, "grad_norm": 0.3487478544658144, "learning_rate": 5.3104752773110345e-05, "loss": 0.5245, "step": 825 }, { "epoch": 2.270790378006873, "grad_norm": 0.32194879135312154, "learning_rate": 5.303202291840471e-05, "loss": 0.526, "step": 826 }, { "epoch": 2.2735395189003436, "grad_norm": 0.27245901521747345, "learning_rate": 5.295924483127635e-05, "loss": 0.5184, "step": 827 }, { "epoch": 2.2762886597938143, "grad_norm": 0.26804164140224673, "learning_rate": 5.2886418781082066e-05, "loss": 0.5201, "step": 828 }, { "epoch": 2.279037800687285, "grad_norm": 0.23960902170409884, "learning_rate": 5.2813545037356086e-05, "loss": 0.5256, "step": 829 }, { "epoch": 2.281786941580756, "grad_norm": 0.23457140639322555, "learning_rate": 5.2740623869809234e-05, "loss": 0.5209, "step": 830 }, { "epoch": 2.2845360824742267, "grad_norm": 0.27297058783320943, "learning_rate": 5.2667655548327796e-05, "loss": 0.5315, "step": 831 }, { "epoch": 2.2872852233676975, "grad_norm": 0.28200029368869456, "learning_rate": 5.259464034297262e-05, "loss": 0.527, "step": 832 }, { "epoch": 2.2900343642611682, "grad_norm": 0.34366587435096785, "learning_rate": 5.2521578523978046e-05, "loss": 0.5166, "step": 833 }, { "epoch": 2.292783505154639, "grad_norm": 0.33716783962672386, "learning_rate": 5.2448470361750955e-05, "loss": 0.5271, "step": 834 }, { "epoch": 2.29553264604811, "grad_norm": 0.2883538098562466, "learning_rate": 5.237531612686973e-05, "loss": 0.5326, "step": 835 }, { "epoch": 2.2982817869415806, "grad_norm": 0.23816842715139308, "learning_rate": 5.2302116090083274e-05, "loss": 0.521, "step": 836 }, { "epoch": 2.3010309278350514, "grad_norm": 0.2676295491191559, "learning_rate": 5.222887052231003e-05, "loss": 0.5307, "step": 837 }, { "epoch": 2.303780068728522, "grad_norm": 0.21899952831754288, "learning_rate": 5.2155579694636916e-05, "loss": 0.5254, "step": 838 }, { "epoch": 2.306529209621993, "grad_norm": 0.2725930609511096, "learning_rate": 5.208224387831839e-05, "loss": 0.5243, "step": 839 }, { "epoch": 2.3092783505154637, "grad_norm": 0.2670269822849005, "learning_rate": 5.200886334477541e-05, "loss": 0.5263, "step": 840 }, { "epoch": 2.3120274914089345, "grad_norm": 0.2896259602041522, "learning_rate": 5.1935438365594424e-05, "loss": 0.5212, "step": 841 }, { "epoch": 2.3147766323024053, "grad_norm": 0.3040525797226736, "learning_rate": 5.186196921252638e-05, "loss": 0.5272, "step": 842 }, { "epoch": 2.317525773195876, "grad_norm": 0.402527650035371, "learning_rate": 5.178845615748573e-05, "loss": 0.5317, "step": 843 }, { "epoch": 2.320274914089347, "grad_norm": 0.28589840098643327, "learning_rate": 5.1714899472549394e-05, "loss": 0.5302, "step": 844 }, { "epoch": 2.323024054982818, "grad_norm": 0.2746598297288049, "learning_rate": 5.164129942995578e-05, "loss": 0.5237, "step": 845 }, { "epoch": 2.325773195876289, "grad_norm": 0.35895254123689857, "learning_rate": 5.156765630210375e-05, "loss": 0.5321, "step": 846 }, { "epoch": 2.3285223367697596, "grad_norm": 0.3075206472125044, "learning_rate": 5.1493970361551666e-05, "loss": 0.5261, "step": 847 }, { "epoch": 2.3312714776632304, "grad_norm": 0.2631726158525417, "learning_rate": 5.14202418810163e-05, "loss": 0.5295, "step": 848 }, { "epoch": 2.334020618556701, "grad_norm": 0.26253313412596246, "learning_rate": 5.134647113337187e-05, "loss": 0.5278, "step": 849 }, { "epoch": 2.336769759450172, "grad_norm": 0.27292981975275177, "learning_rate": 5.127265839164906e-05, "loss": 0.5174, "step": 850 }, { "epoch": 2.3395189003436427, "grad_norm": 0.29830820148659387, "learning_rate": 5.119880392903396e-05, "loss": 0.5235, "step": 851 }, { "epoch": 2.3422680412371135, "grad_norm": 0.319178636597951, "learning_rate": 5.112490801886706e-05, "loss": 0.5249, "step": 852 }, { "epoch": 2.3450171821305843, "grad_norm": 0.39509509521849606, "learning_rate": 5.105097093464225e-05, "loss": 0.53, "step": 853 }, { "epoch": 2.347766323024055, "grad_norm": 0.43481008500350815, "learning_rate": 5.0976992950005836e-05, "loss": 0.5222, "step": 854 }, { "epoch": 2.350515463917526, "grad_norm": 0.35580076547200273, "learning_rate": 5.090297433875549e-05, "loss": 0.5341, "step": 855 }, { "epoch": 2.3532646048109966, "grad_norm": 0.27387444348548706, "learning_rate": 5.082891537483921e-05, "loss": 0.527, "step": 856 }, { "epoch": 2.3560137457044674, "grad_norm": 0.28242295705533316, "learning_rate": 5.0754816332354384e-05, "loss": 0.5313, "step": 857 }, { "epoch": 2.358762886597938, "grad_norm": 0.23633783144067932, "learning_rate": 5.0680677485546724e-05, "loss": 0.5256, "step": 858 }, { "epoch": 2.361512027491409, "grad_norm": 0.20611827366274735, "learning_rate": 5.060649910880926e-05, "loss": 0.5246, "step": 859 }, { "epoch": 2.3642611683848798, "grad_norm": 0.24527029533975683, "learning_rate": 5.0532281476681295e-05, "loss": 0.526, "step": 860 }, { "epoch": 2.3670103092783505, "grad_norm": 0.2988866489058128, "learning_rate": 5.0458024863847455e-05, "loss": 0.5318, "step": 861 }, { "epoch": 2.3697594501718213, "grad_norm": 0.3278941979649635, "learning_rate": 5.038372954513664e-05, "loss": 0.5208, "step": 862 }, { "epoch": 2.372508591065292, "grad_norm": 0.3590976948020781, "learning_rate": 5.030939579552098e-05, "loss": 0.5288, "step": 863 }, { "epoch": 2.375257731958763, "grad_norm": 0.3257482610923462, "learning_rate": 5.0235023890114844e-05, "loss": 0.5342, "step": 864 }, { "epoch": 2.3780068728522337, "grad_norm": 0.26418250148170586, "learning_rate": 5.016061410417384e-05, "loss": 0.534, "step": 865 }, { "epoch": 2.3807560137457044, "grad_norm": 0.25433731862506237, "learning_rate": 5.008616671309374e-05, "loss": 0.5349, "step": 866 }, { "epoch": 2.3835051546391752, "grad_norm": 0.3098360496587967, "learning_rate": 5.0011681992409526e-05, "loss": 0.536, "step": 867 }, { "epoch": 2.386254295532646, "grad_norm": 0.48700691274762764, "learning_rate": 4.993716021779431e-05, "loss": 0.5281, "step": 868 }, { "epoch": 2.389003436426117, "grad_norm": 0.4325007530550783, "learning_rate": 4.986260166505838e-05, "loss": 0.5287, "step": 869 }, { "epoch": 2.3917525773195876, "grad_norm": 0.40355654494603005, "learning_rate": 4.978800661014811e-05, "loss": 0.5211, "step": 870 }, { "epoch": 2.3945017182130583, "grad_norm": 0.26544768260770013, "learning_rate": 4.9713375329145e-05, "loss": 0.5195, "step": 871 }, { "epoch": 2.397250859106529, "grad_norm": 0.2791642651045501, "learning_rate": 4.963870809826458e-05, "loss": 0.5265, "step": 872 }, { "epoch": 2.4, "grad_norm": 0.3408862865130159, "learning_rate": 4.9564005193855486e-05, "loss": 0.5391, "step": 873 }, { "epoch": 2.4027491408934707, "grad_norm": 0.3331718549433266, "learning_rate": 4.9489266892398346e-05, "loss": 0.5216, "step": 874 }, { "epoch": 2.4054982817869415, "grad_norm": 0.33450428391908577, "learning_rate": 4.9414493470504834e-05, "loss": 0.5274, "step": 875 }, { "epoch": 2.4082474226804123, "grad_norm": 0.324321171570019, "learning_rate": 4.933968520491654e-05, "loss": 0.5373, "step": 876 }, { "epoch": 2.410996563573883, "grad_norm": 0.28829591882553246, "learning_rate": 4.9264842372504095e-05, "loss": 0.5286, "step": 877 }, { "epoch": 2.413745704467354, "grad_norm": 0.28746004519536006, "learning_rate": 4.9189965250265994e-05, "loss": 0.5199, "step": 878 }, { "epoch": 2.4164948453608246, "grad_norm": 0.31445005880135923, "learning_rate": 4.911505411532769e-05, "loss": 0.5278, "step": 879 }, { "epoch": 2.4192439862542954, "grad_norm": 0.47892737901970533, "learning_rate": 4.9040109244940505e-05, "loss": 0.5234, "step": 880 }, { "epoch": 2.421993127147766, "grad_norm": 0.45215275967709107, "learning_rate": 4.896513091648058e-05, "loss": 0.535, "step": 881 }, { "epoch": 2.424742268041237, "grad_norm": 0.34138582396119665, "learning_rate": 4.889011940744796e-05, "loss": 0.5164, "step": 882 }, { "epoch": 2.4274914089347077, "grad_norm": 0.2682120490670279, "learning_rate": 4.881507499546545e-05, "loss": 0.5342, "step": 883 }, { "epoch": 2.430240549828179, "grad_norm": 0.2692048784370494, "learning_rate": 4.873999795827761e-05, "loss": 0.5344, "step": 884 }, { "epoch": 2.4329896907216497, "grad_norm": 0.3048032911190028, "learning_rate": 4.866488857374979e-05, "loss": 0.5131, "step": 885 }, { "epoch": 2.4357388316151205, "grad_norm": 0.43798907642210033, "learning_rate": 4.858974711986704e-05, "loss": 0.5277, "step": 886 }, { "epoch": 2.4384879725085913, "grad_norm": 0.4537544665998809, "learning_rate": 4.851457387473312e-05, "loss": 0.5272, "step": 887 }, { "epoch": 2.441237113402062, "grad_norm": 0.275619974234131, "learning_rate": 4.843936911656941e-05, "loss": 0.5229, "step": 888 }, { "epoch": 2.443986254295533, "grad_norm": 0.24697116802006067, "learning_rate": 4.836413312371394e-05, "loss": 0.5237, "step": 889 }, { "epoch": 2.4467353951890036, "grad_norm": 0.2889992360936922, "learning_rate": 4.828886617462039e-05, "loss": 0.5214, "step": 890 }, { "epoch": 2.4494845360824744, "grad_norm": 0.34236892659433343, "learning_rate": 4.821356854785695e-05, "loss": 0.5204, "step": 891 }, { "epoch": 2.452233676975945, "grad_norm": 0.26986931392665453, "learning_rate": 4.8138240522105365e-05, "loss": 0.5285, "step": 892 }, { "epoch": 2.454982817869416, "grad_norm": 0.2386727777106889, "learning_rate": 4.806288237615989e-05, "loss": 0.5258, "step": 893 }, { "epoch": 2.4577319587628867, "grad_norm": 0.2400107390785469, "learning_rate": 4.7987494388926275e-05, "loss": 0.523, "step": 894 }, { "epoch": 2.4604810996563575, "grad_norm": 0.30720815238105814, "learning_rate": 4.7912076839420695e-05, "loss": 0.5232, "step": 895 }, { "epoch": 2.4632302405498283, "grad_norm": 0.25563044453886097, "learning_rate": 4.7836630006768746e-05, "loss": 0.5289, "step": 896 }, { "epoch": 2.465979381443299, "grad_norm": 0.2618995944120738, "learning_rate": 4.77611541702044e-05, "loss": 0.5256, "step": 897 }, { "epoch": 2.46872852233677, "grad_norm": 0.2872261399026541, "learning_rate": 4.768564960906897e-05, "loss": 0.527, "step": 898 }, { "epoch": 2.4714776632302407, "grad_norm": 0.22134537204990035, "learning_rate": 4.7610116602810096e-05, "loss": 0.5343, "step": 899 }, { "epoch": 2.4742268041237114, "grad_norm": 0.2565389994931295, "learning_rate": 4.753455543098067e-05, "loss": 0.5375, "step": 900 }, { "epoch": 2.476975945017182, "grad_norm": 0.3740627121098048, "learning_rate": 4.745896637323785e-05, "loss": 0.5227, "step": 901 }, { "epoch": 2.479725085910653, "grad_norm": 0.3714016967888814, "learning_rate": 4.7383349709341994e-05, "loss": 0.5216, "step": 902 }, { "epoch": 2.4824742268041238, "grad_norm": 0.2863979507280106, "learning_rate": 4.730770571915562e-05, "loss": 0.5289, "step": 903 }, { "epoch": 2.4852233676975946, "grad_norm": 0.2639569172110602, "learning_rate": 4.7232034682642417e-05, "loss": 0.5207, "step": 904 }, { "epoch": 2.4879725085910653, "grad_norm": 0.26506986793397436, "learning_rate": 4.715633687986613e-05, "loss": 0.5276, "step": 905 }, { "epoch": 2.490721649484536, "grad_norm": 0.3034302978268054, "learning_rate": 4.7080612590989596e-05, "loss": 0.5185, "step": 906 }, { "epoch": 2.493470790378007, "grad_norm": 0.3242741496552742, "learning_rate": 4.7004862096273674e-05, "loss": 0.5292, "step": 907 }, { "epoch": 2.4962199312714777, "grad_norm": 0.25198934191589384, "learning_rate": 4.692908567607621e-05, "loss": 0.5209, "step": 908 }, { "epoch": 2.4989690721649485, "grad_norm": 0.2851823027369076, "learning_rate": 4.6853283610851004e-05, "loss": 0.5332, "step": 909 }, { "epoch": 2.5017182130584192, "grad_norm": 0.2250586051327087, "learning_rate": 4.677745618114674e-05, "loss": 0.5326, "step": 910 }, { "epoch": 2.50446735395189, "grad_norm": 0.2137916864449506, "learning_rate": 4.670160366760606e-05, "loss": 0.5214, "step": 911 }, { "epoch": 2.507216494845361, "grad_norm": 0.23692573516235416, "learning_rate": 4.6625726350964355e-05, "loss": 0.5209, "step": 912 }, { "epoch": 2.5099656357388316, "grad_norm": 0.2503089565651227, "learning_rate": 4.654982451204885e-05, "loss": 0.5228, "step": 913 }, { "epoch": 2.5127147766323024, "grad_norm": 0.187246659026728, "learning_rate": 4.6473898431777535e-05, "loss": 0.5279, "step": 914 }, { "epoch": 2.515463917525773, "grad_norm": 0.22968241737897607, "learning_rate": 4.6397948391158104e-05, "loss": 0.5229, "step": 915 }, { "epoch": 2.518213058419244, "grad_norm": 0.34560044485812286, "learning_rate": 4.632197467128695e-05, "loss": 0.521, "step": 916 }, { "epoch": 2.5209621993127147, "grad_norm": 0.36369842042382355, "learning_rate": 4.624597755334807e-05, "loss": 0.5222, "step": 917 }, { "epoch": 2.5237113402061855, "grad_norm": 0.3273459802424622, "learning_rate": 4.6169957318612096e-05, "loss": 0.5269, "step": 918 }, { "epoch": 2.5264604810996563, "grad_norm": 0.27046737413984473, "learning_rate": 4.609391424843519e-05, "loss": 0.5207, "step": 919 }, { "epoch": 2.529209621993127, "grad_norm": 0.23973509685168032, "learning_rate": 4.601784862425807e-05, "loss": 0.5305, "step": 920 }, { "epoch": 2.531958762886598, "grad_norm": 0.207584971564237, "learning_rate": 4.594176072760485e-05, "loss": 0.5391, "step": 921 }, { "epoch": 2.5347079037800686, "grad_norm": 0.26389513176009577, "learning_rate": 4.586565084008217e-05, "loss": 0.5257, "step": 922 }, { "epoch": 2.5374570446735394, "grad_norm": 0.27492964134545905, "learning_rate": 4.5789519243377975e-05, "loss": 0.5198, "step": 923 }, { "epoch": 2.54020618556701, "grad_norm": 0.2550669423258001, "learning_rate": 4.571336621926065e-05, "loss": 0.5161, "step": 924 }, { "epoch": 2.542955326460481, "grad_norm": 0.2434323842617188, "learning_rate": 4.563719204957776e-05, "loss": 0.5266, "step": 925 }, { "epoch": 2.5457044673539517, "grad_norm": 0.2381537603353079, "learning_rate": 4.5560997016255265e-05, "loss": 0.5306, "step": 926 }, { "epoch": 2.5484536082474225, "grad_norm": 0.25877756360414456, "learning_rate": 4.548478140129624e-05, "loss": 0.5279, "step": 927 }, { "epoch": 2.5512027491408933, "grad_norm": 0.23404565240810715, "learning_rate": 4.5408545486779996e-05, "loss": 0.5325, "step": 928 }, { "epoch": 2.553951890034364, "grad_norm": 0.20244068111276767, "learning_rate": 4.533228955486094e-05, "loss": 0.5245, "step": 929 }, { "epoch": 2.556701030927835, "grad_norm": 0.20414770829038065, "learning_rate": 4.525601388776758e-05, "loss": 0.5193, "step": 930 }, { "epoch": 2.5594501718213056, "grad_norm": 0.22610091916102834, "learning_rate": 4.517971876780147e-05, "loss": 0.5335, "step": 931 }, { "epoch": 2.5621993127147764, "grad_norm": 0.2030707662947604, "learning_rate": 4.5103404477336144e-05, "loss": 0.5247, "step": 932 }, { "epoch": 2.564948453608247, "grad_norm": 0.19565686931462759, "learning_rate": 4.502707129881609e-05, "loss": 0.5223, "step": 933 }, { "epoch": 2.567697594501718, "grad_norm": 0.22241975043349305, "learning_rate": 4.495071951475572e-05, "loss": 0.5289, "step": 934 }, { "epoch": 2.5704467353951888, "grad_norm": 0.25405146475942325, "learning_rate": 4.487434940773828e-05, "loss": 0.5216, "step": 935 }, { "epoch": 2.5731958762886595, "grad_norm": 0.24090685844792697, "learning_rate": 4.479796126041487e-05, "loss": 0.5347, "step": 936 }, { "epoch": 2.5759450171821303, "grad_norm": 0.20268315657201996, "learning_rate": 4.472155535550331e-05, "loss": 0.5174, "step": 937 }, { "epoch": 2.5786941580756015, "grad_norm": 0.2164104030727069, "learning_rate": 4.464513197578717e-05, "loss": 0.5306, "step": 938 }, { "epoch": 2.5814432989690723, "grad_norm": 0.20675687136906504, "learning_rate": 4.45686914041147e-05, "loss": 0.5217, "step": 939 }, { "epoch": 2.584192439862543, "grad_norm": 0.2495476073768104, "learning_rate": 4.449223392339776e-05, "loss": 0.5233, "step": 940 }, { "epoch": 2.586941580756014, "grad_norm": 0.2188055318879783, "learning_rate": 4.44157598166108e-05, "loss": 0.536, "step": 941 }, { "epoch": 2.5896907216494847, "grad_norm": 0.19661959776334886, "learning_rate": 4.43392693667898e-05, "loss": 0.5167, "step": 942 }, { "epoch": 2.5924398625429554, "grad_norm": 0.21281895098022574, "learning_rate": 4.426276285703125e-05, "loss": 0.5241, "step": 943 }, { "epoch": 2.5951890034364262, "grad_norm": 0.24612261354981668, "learning_rate": 4.418624057049106e-05, "loss": 0.5288, "step": 944 }, { "epoch": 2.597938144329897, "grad_norm": 0.26256604053792143, "learning_rate": 4.410970279038351e-05, "loss": 0.528, "step": 945 }, { "epoch": 2.600687285223368, "grad_norm": 0.2146570261022671, "learning_rate": 4.4033149799980265e-05, "loss": 0.5336, "step": 946 }, { "epoch": 2.6034364261168386, "grad_norm": 0.20864992098912383, "learning_rate": 4.395658188260924e-05, "loss": 0.5289, "step": 947 }, { "epoch": 2.6061855670103093, "grad_norm": 0.23475812934644366, "learning_rate": 4.3879999321653664e-05, "loss": 0.5294, "step": 948 }, { "epoch": 2.60893470790378, "grad_norm": 0.21552160156786482, "learning_rate": 4.380340240055087e-05, "loss": 0.5225, "step": 949 }, { "epoch": 2.611683848797251, "grad_norm": 0.21451249484405446, "learning_rate": 4.3726791402791405e-05, "loss": 0.5309, "step": 950 }, { "epoch": 2.6144329896907217, "grad_norm": 0.2228499608780446, "learning_rate": 4.3650166611917904e-05, "loss": 0.5218, "step": 951 }, { "epoch": 2.6171821305841925, "grad_norm": 0.19761673711898892, "learning_rate": 4.3573528311524056e-05, "loss": 0.5263, "step": 952 }, { "epoch": 2.6199312714776632, "grad_norm": 0.2717760446302159, "learning_rate": 4.34968767852535e-05, "loss": 0.5192, "step": 953 }, { "epoch": 2.622680412371134, "grad_norm": 0.2307774736780584, "learning_rate": 4.3420212316798895e-05, "loss": 0.5224, "step": 954 }, { "epoch": 2.625429553264605, "grad_norm": 0.26499404963083745, "learning_rate": 4.334353518990075e-05, "loss": 0.5284, "step": 955 }, { "epoch": 2.6281786941580756, "grad_norm": 0.33949301525073333, "learning_rate": 4.326684568834647e-05, "loss": 0.5215, "step": 956 }, { "epoch": 2.6309278350515464, "grad_norm": 0.33940964847094485, "learning_rate": 4.31901440959692e-05, "loss": 0.527, "step": 957 }, { "epoch": 2.633676975945017, "grad_norm": 0.3034703314779094, "learning_rate": 4.311343069664688e-05, "loss": 0.5315, "step": 958 }, { "epoch": 2.636426116838488, "grad_norm": 0.3285064736514596, "learning_rate": 4.303670577430111e-05, "loss": 0.5328, "step": 959 }, { "epoch": 2.6391752577319587, "grad_norm": 0.3453164976173602, "learning_rate": 4.295996961289619e-05, "loss": 0.5323, "step": 960 }, { "epoch": 2.6419243986254295, "grad_norm": 0.38444181745132805, "learning_rate": 4.288322249643799e-05, "loss": 0.5288, "step": 961 }, { "epoch": 2.6446735395189003, "grad_norm": 0.2686311128572819, "learning_rate": 4.2806464708972905e-05, "loss": 0.525, "step": 962 }, { "epoch": 2.647422680412371, "grad_norm": 0.3232991261053769, "learning_rate": 4.272969653458685e-05, "loss": 0.5278, "step": 963 }, { "epoch": 2.650171821305842, "grad_norm": 0.32877598170142874, "learning_rate": 4.265291825740416e-05, "loss": 0.5176, "step": 964 }, { "epoch": 2.6529209621993126, "grad_norm": 0.30844002727607134, "learning_rate": 4.257613016158661e-05, "loss": 0.5326, "step": 965 }, { "epoch": 2.6556701030927834, "grad_norm": 0.2255639863284666, "learning_rate": 4.249933253133224e-05, "loss": 0.5266, "step": 966 }, { "epoch": 2.658419243986254, "grad_norm": 0.25200678515207303, "learning_rate": 4.2422525650874446e-05, "loss": 0.5292, "step": 967 }, { "epoch": 2.661168384879725, "grad_norm": 0.22993658765349623, "learning_rate": 4.234570980448081e-05, "loss": 0.5251, "step": 968 }, { "epoch": 2.6639175257731957, "grad_norm": 0.2642010038722434, "learning_rate": 4.226888527645215e-05, "loss": 0.529, "step": 969 }, { "epoch": 2.6666666666666665, "grad_norm": 0.28025388603567647, "learning_rate": 4.219205235112135e-05, "loss": 0.523, "step": 970 }, { "epoch": 2.6694158075601373, "grad_norm": 0.24286835186859662, "learning_rate": 4.2115211312852435e-05, "loss": 0.5241, "step": 971 }, { "epoch": 2.6721649484536085, "grad_norm": 0.260911275304071, "learning_rate": 4.203836244603941e-05, "loss": 0.5251, "step": 972 }, { "epoch": 2.6749140893470793, "grad_norm": 0.23336974876081842, "learning_rate": 4.1961506035105285e-05, "loss": 0.5337, "step": 973 }, { "epoch": 2.67766323024055, "grad_norm": 0.19717415494962492, "learning_rate": 4.188464236450098e-05, "loss": 0.518, "step": 974 }, { "epoch": 2.680412371134021, "grad_norm": 0.2635705526316712, "learning_rate": 4.180777171870427e-05, "loss": 0.519, "step": 975 }, { "epoch": 2.6831615120274916, "grad_norm": 0.2420795496999561, "learning_rate": 4.173089438221876e-05, "loss": 0.5329, "step": 976 }, { "epoch": 2.6859106529209624, "grad_norm": 0.24604439574458017, "learning_rate": 4.165401063957283e-05, "loss": 0.5331, "step": 977 }, { "epoch": 2.688659793814433, "grad_norm": 0.1948605941664403, "learning_rate": 4.157712077531856e-05, "loss": 0.5221, "step": 978 }, { "epoch": 2.691408934707904, "grad_norm": 0.21270792355056445, "learning_rate": 4.1500225074030654e-05, "loss": 0.5355, "step": 979 }, { "epoch": 2.6941580756013748, "grad_norm": 0.251861479262255, "learning_rate": 4.142332382030547e-05, "loss": 0.5336, "step": 980 }, { "epoch": 2.6969072164948455, "grad_norm": 0.2112157139724383, "learning_rate": 4.13464172987599e-05, "loss": 0.5243, "step": 981 }, { "epoch": 2.6996563573883163, "grad_norm": 0.20387549860274823, "learning_rate": 4.1269505794030296e-05, "loss": 0.5237, "step": 982 }, { "epoch": 2.702405498281787, "grad_norm": 0.2515717159268574, "learning_rate": 4.119258959077151e-05, "loss": 0.5319, "step": 983 }, { "epoch": 2.705154639175258, "grad_norm": 0.20847504323318589, "learning_rate": 4.111566897365575e-05, "loss": 0.52, "step": 984 }, { "epoch": 2.7079037800687287, "grad_norm": 0.21911840052600298, "learning_rate": 4.103874422737157e-05, "loss": 0.5283, "step": 985 }, { "epoch": 2.7106529209621995, "grad_norm": 0.2049343201857368, "learning_rate": 4.096181563662279e-05, "loss": 0.5263, "step": 986 }, { "epoch": 2.7134020618556702, "grad_norm": 0.1947687795273586, "learning_rate": 4.088488348612748e-05, "loss": 0.5239, "step": 987 }, { "epoch": 2.716151202749141, "grad_norm": 0.23438600392965853, "learning_rate": 4.080794806061688e-05, "loss": 0.5235, "step": 988 }, { "epoch": 2.718900343642612, "grad_norm": 0.19851824128921414, "learning_rate": 4.073100964483435e-05, "loss": 0.5236, "step": 989 }, { "epoch": 2.7216494845360826, "grad_norm": 0.23677872375792366, "learning_rate": 4.06540685235343e-05, "loss": 0.5307, "step": 990 }, { "epoch": 2.7243986254295534, "grad_norm": 0.19484452355877876, "learning_rate": 4.057712498148119e-05, "loss": 0.533, "step": 991 }, { "epoch": 2.727147766323024, "grad_norm": 0.24691106342879912, "learning_rate": 4.05001793034484e-05, "loss": 0.5282, "step": 992 }, { "epoch": 2.729896907216495, "grad_norm": 0.23911573502762684, "learning_rate": 4.0423231774217246e-05, "loss": 0.5291, "step": 993 }, { "epoch": 2.7326460481099657, "grad_norm": 0.24784394785326042, "learning_rate": 4.034628267857587e-05, "loss": 0.5299, "step": 994 }, { "epoch": 2.7353951890034365, "grad_norm": 0.2530406886888911, "learning_rate": 4.026933230131823e-05, "loss": 0.5155, "step": 995 }, { "epoch": 2.7381443298969073, "grad_norm": 0.24815572290601853, "learning_rate": 4.0192380927243014e-05, "loss": 0.5331, "step": 996 }, { "epoch": 2.740893470790378, "grad_norm": 0.2756774245673258, "learning_rate": 4.0115428841152637e-05, "loss": 0.5334, "step": 997 }, { "epoch": 2.743642611683849, "grad_norm": 0.24077726898420207, "learning_rate": 4.0038476327852065e-05, "loss": 0.5337, "step": 998 }, { "epoch": 2.7463917525773196, "grad_norm": 0.20128889819397291, "learning_rate": 3.9961523672147955e-05, "loss": 0.5302, "step": 999 }, { "epoch": 2.7491408934707904, "grad_norm": 0.20443940819422327, "learning_rate": 3.988457115884739e-05, "loss": 0.5284, "step": 1000 }, { "epoch": 2.751890034364261, "grad_norm": 0.2208826727313727, "learning_rate": 3.9807619072757e-05, "loss": 0.5271, "step": 1001 }, { "epoch": 2.754639175257732, "grad_norm": 0.24447959308366699, "learning_rate": 3.973066769868178e-05, "loss": 0.5283, "step": 1002 }, { "epoch": 2.7573883161512027, "grad_norm": 0.1966957237573264, "learning_rate": 3.965371732142415e-05, "loss": 0.5292, "step": 1003 }, { "epoch": 2.7601374570446735, "grad_norm": 0.21603427189344282, "learning_rate": 3.957676822578276e-05, "loss": 0.5237, "step": 1004 }, { "epoch": 2.7628865979381443, "grad_norm": 0.18664057521742963, "learning_rate": 3.949982069655161e-05, "loss": 0.5233, "step": 1005 }, { "epoch": 2.765635738831615, "grad_norm": 0.2234326703320951, "learning_rate": 3.942287501851881e-05, "loss": 0.5256, "step": 1006 }, { "epoch": 2.768384879725086, "grad_norm": 0.2278011506232875, "learning_rate": 3.9345931476465706e-05, "loss": 0.5279, "step": 1007 }, { "epoch": 2.7711340206185566, "grad_norm": 0.19798832981596853, "learning_rate": 3.9268990355165664e-05, "loss": 0.5211, "step": 1008 }, { "epoch": 2.7738831615120274, "grad_norm": 0.2183804005479682, "learning_rate": 3.9192051939383126e-05, "loss": 0.5239, "step": 1009 }, { "epoch": 2.776632302405498, "grad_norm": 0.22246588980190424, "learning_rate": 3.911511651387253e-05, "loss": 0.527, "step": 1010 }, { "epoch": 2.779381443298969, "grad_norm": 0.1780086146278406, "learning_rate": 3.903818436337722e-05, "loss": 0.5284, "step": 1011 }, { "epoch": 2.7821305841924397, "grad_norm": 0.19604513054901485, "learning_rate": 3.896125577262845e-05, "loss": 0.5162, "step": 1012 }, { "epoch": 2.7848797250859105, "grad_norm": 0.18794858233130177, "learning_rate": 3.888433102634425e-05, "loss": 0.5216, "step": 1013 }, { "epoch": 2.7876288659793813, "grad_norm": 0.21435085046499355, "learning_rate": 3.8807410409228496e-05, "loss": 0.5305, "step": 1014 }, { "epoch": 2.790378006872852, "grad_norm": 0.17410246023922102, "learning_rate": 3.8730494205969724e-05, "loss": 0.5312, "step": 1015 }, { "epoch": 2.793127147766323, "grad_norm": 0.1822314934502394, "learning_rate": 3.865358270124013e-05, "loss": 0.5189, "step": 1016 }, { "epoch": 2.7958762886597937, "grad_norm": 0.206717747798629, "learning_rate": 3.857667617969454e-05, "loss": 0.5269, "step": 1017 }, { "epoch": 2.7986254295532644, "grad_norm": 0.16863448749904908, "learning_rate": 3.849977492596936e-05, "loss": 0.5303, "step": 1018 }, { "epoch": 2.801374570446735, "grad_norm": 0.23449573918173383, "learning_rate": 3.8422879224681456e-05, "loss": 0.5322, "step": 1019 }, { "epoch": 2.804123711340206, "grad_norm": 0.17234697630039422, "learning_rate": 3.8345989360427174e-05, "loss": 0.5229, "step": 1020 }, { "epoch": 2.8068728522336768, "grad_norm": 0.21339720138725765, "learning_rate": 3.826910561778124e-05, "loss": 0.5271, "step": 1021 }, { "epoch": 2.8096219931271476, "grad_norm": 0.22865589786306695, "learning_rate": 3.819222828129574e-05, "loss": 0.5261, "step": 1022 }, { "epoch": 2.8123711340206183, "grad_norm": 0.19395333537771012, "learning_rate": 3.8115357635499045e-05, "loss": 0.5307, "step": 1023 }, { "epoch": 2.815120274914089, "grad_norm": 0.26952248831929015, "learning_rate": 3.803849396489473e-05, "loss": 0.5304, "step": 1024 }, { "epoch": 2.81786941580756, "grad_norm": 0.2078676429133822, "learning_rate": 3.7961637553960605e-05, "loss": 0.521, "step": 1025 }, { "epoch": 2.8206185567010307, "grad_norm": 0.2437865253535147, "learning_rate": 3.788478868714758e-05, "loss": 0.5284, "step": 1026 }, { "epoch": 2.8233676975945015, "grad_norm": 0.23244204464146223, "learning_rate": 3.780794764887866e-05, "loss": 0.5347, "step": 1027 }, { "epoch": 2.8261168384879722, "grad_norm": 0.1883460034103066, "learning_rate": 3.7731114723547856e-05, "loss": 0.5332, "step": 1028 }, { "epoch": 2.8288659793814435, "grad_norm": 0.18889412109996112, "learning_rate": 3.7654290195519195e-05, "loss": 0.527, "step": 1029 }, { "epoch": 2.8316151202749142, "grad_norm": 0.23037525353226912, "learning_rate": 3.757747434912556e-05, "loss": 0.5293, "step": 1030 }, { "epoch": 2.834364261168385, "grad_norm": 0.21033380977426863, "learning_rate": 3.750066746866778e-05, "loss": 0.5226, "step": 1031 }, { "epoch": 2.837113402061856, "grad_norm": 0.17225724075961382, "learning_rate": 3.742386983841341e-05, "loss": 0.5244, "step": 1032 }, { "epoch": 2.8398625429553266, "grad_norm": 0.22879502842160185, "learning_rate": 3.734708174259585e-05, "loss": 0.5291, "step": 1033 }, { "epoch": 2.8426116838487974, "grad_norm": 0.1859425423671784, "learning_rate": 3.727030346541317e-05, "loss": 0.5207, "step": 1034 }, { "epoch": 2.845360824742268, "grad_norm": 0.21138789242685446, "learning_rate": 3.7193535291027115e-05, "loss": 0.5283, "step": 1035 }, { "epoch": 2.848109965635739, "grad_norm": 0.22384899016314974, "learning_rate": 3.7116777503562016e-05, "loss": 0.528, "step": 1036 }, { "epoch": 2.8508591065292097, "grad_norm": 0.2431535434052746, "learning_rate": 3.7040030387103815e-05, "loss": 0.5187, "step": 1037 }, { "epoch": 2.8536082474226805, "grad_norm": 0.24399503353207871, "learning_rate": 3.696329422569889e-05, "loss": 0.5179, "step": 1038 }, { "epoch": 2.8563573883161513, "grad_norm": 0.16921161494449857, "learning_rate": 3.6886569303353136e-05, "loss": 0.5292, "step": 1039 }, { "epoch": 2.859106529209622, "grad_norm": 0.26728522647583636, "learning_rate": 3.680985590403082e-05, "loss": 0.5262, "step": 1040 }, { "epoch": 2.861855670103093, "grad_norm": 0.27421229932307095, "learning_rate": 3.673315431165355e-05, "loss": 0.5252, "step": 1041 }, { "epoch": 2.8646048109965636, "grad_norm": 0.2601105298366, "learning_rate": 3.665646481009926e-05, "loss": 0.5203, "step": 1042 }, { "epoch": 2.8673539518900344, "grad_norm": 0.2997837262368804, "learning_rate": 3.657978768320111e-05, "loss": 0.5266, "step": 1043 }, { "epoch": 2.870103092783505, "grad_norm": 0.1739319697716107, "learning_rate": 3.650312321474651e-05, "loss": 0.5285, "step": 1044 }, { "epoch": 2.872852233676976, "grad_norm": 0.2868154357660894, "learning_rate": 3.642647168847596e-05, "loss": 0.5349, "step": 1045 }, { "epoch": 2.8756013745704467, "grad_norm": 0.3150816781576196, "learning_rate": 3.6349833388082096e-05, "loss": 0.5213, "step": 1046 }, { "epoch": 2.8783505154639175, "grad_norm": 0.21703226563550534, "learning_rate": 3.6273208597208595e-05, "loss": 0.5302, "step": 1047 }, { "epoch": 2.8810996563573883, "grad_norm": 0.24416409278757065, "learning_rate": 3.619659759944916e-05, "loss": 0.5249, "step": 1048 }, { "epoch": 2.883848797250859, "grad_norm": 0.2807527048682551, "learning_rate": 3.612000067834636e-05, "loss": 0.5249, "step": 1049 }, { "epoch": 2.88659793814433, "grad_norm": 0.2744380467215706, "learning_rate": 3.604341811739077e-05, "loss": 0.5305, "step": 1050 }, { "epoch": 2.8893470790378006, "grad_norm": 0.2271292993546667, "learning_rate": 3.596685020001975e-05, "loss": 0.5194, "step": 1051 }, { "epoch": 2.8920962199312714, "grad_norm": 0.21515311729133313, "learning_rate": 3.5890297209616507e-05, "loss": 0.5242, "step": 1052 }, { "epoch": 2.894845360824742, "grad_norm": 0.2904407669489142, "learning_rate": 3.581375942950895e-05, "loss": 0.5163, "step": 1053 }, { "epoch": 2.897594501718213, "grad_norm": 0.23987566400445187, "learning_rate": 3.5737237142968755e-05, "loss": 0.5321, "step": 1054 }, { "epoch": 2.9003436426116838, "grad_norm": 0.203928269390139, "learning_rate": 3.56607306332102e-05, "loss": 0.5203, "step": 1055 }, { "epoch": 2.9030927835051545, "grad_norm": 0.26314705579076864, "learning_rate": 3.558424018338922e-05, "loss": 0.5267, "step": 1056 }, { "epoch": 2.9058419243986253, "grad_norm": 0.22978092693646193, "learning_rate": 3.5507766076602264e-05, "loss": 0.5211, "step": 1057 }, { "epoch": 2.908591065292096, "grad_norm": 0.1831775606016915, "learning_rate": 3.5431308595885316e-05, "loss": 0.5186, "step": 1058 }, { "epoch": 2.911340206185567, "grad_norm": 0.2557072250330763, "learning_rate": 3.535486802421284e-05, "loss": 0.5204, "step": 1059 }, { "epoch": 2.9140893470790377, "grad_norm": 0.2285012181274199, "learning_rate": 3.5278444644496695e-05, "loss": 0.5273, "step": 1060 }, { "epoch": 2.9168384879725084, "grad_norm": 0.18355449593774995, "learning_rate": 3.520203873958514e-05, "loss": 0.531, "step": 1061 }, { "epoch": 2.9195876288659792, "grad_norm": 0.21328292549083275, "learning_rate": 3.512565059226172e-05, "loss": 0.5229, "step": 1062 }, { "epoch": 2.9223367697594504, "grad_norm": 0.21044538382380515, "learning_rate": 3.5049280485244286e-05, "loss": 0.5231, "step": 1063 }, { "epoch": 2.9250859106529212, "grad_norm": 0.17119562274819797, "learning_rate": 3.4972928701183925e-05, "loss": 0.522, "step": 1064 }, { "epoch": 2.927835051546392, "grad_norm": 0.22581900383360864, "learning_rate": 3.489659552266388e-05, "loss": 0.5246, "step": 1065 }, { "epoch": 2.930584192439863, "grad_norm": 0.301379881236332, "learning_rate": 3.482028123219855e-05, "loss": 0.529, "step": 1066 }, { "epoch": 2.9333333333333336, "grad_norm": 0.295319936733804, "learning_rate": 3.4743986112232434e-05, "loss": 0.5289, "step": 1067 }, { "epoch": 2.9360824742268044, "grad_norm": 0.2388909601572767, "learning_rate": 3.466771044513907e-05, "loss": 0.5252, "step": 1068 }, { "epoch": 2.938831615120275, "grad_norm": 0.2678249461923104, "learning_rate": 3.459145451322002e-05, "loss": 0.5262, "step": 1069 }, { "epoch": 2.941580756013746, "grad_norm": 0.22322825864211238, "learning_rate": 3.4515218598703765e-05, "loss": 0.5297, "step": 1070 }, { "epoch": 2.9443298969072167, "grad_norm": 0.179417097733591, "learning_rate": 3.443900298374475e-05, "loss": 0.5249, "step": 1071 }, { "epoch": 2.9470790378006875, "grad_norm": 0.2551012950816274, "learning_rate": 3.436280795042225e-05, "loss": 0.5212, "step": 1072 }, { "epoch": 2.9498281786941583, "grad_norm": 0.25848561223138594, "learning_rate": 3.428663378073937e-05, "loss": 0.528, "step": 1073 }, { "epoch": 2.952577319587629, "grad_norm": 0.17137059703558605, "learning_rate": 3.421048075662203e-05, "loss": 0.524, "step": 1074 }, { "epoch": 2.9553264604811, "grad_norm": 0.21809992041709833, "learning_rate": 3.413434915991784e-05, "loss": 0.5225, "step": 1075 }, { "epoch": 2.9580756013745706, "grad_norm": 0.19281232602243703, "learning_rate": 3.4058239272395156e-05, "loss": 0.5198, "step": 1076 }, { "epoch": 2.9608247422680414, "grad_norm": 0.18758042777019787, "learning_rate": 3.398215137574194e-05, "loss": 0.5145, "step": 1077 }, { "epoch": 2.963573883161512, "grad_norm": 0.24212160366698007, "learning_rate": 3.390608575156481e-05, "loss": 0.529, "step": 1078 }, { "epoch": 2.966323024054983, "grad_norm": 0.19965723896338422, "learning_rate": 3.3830042681387904e-05, "loss": 0.5345, "step": 1079 }, { "epoch": 2.9690721649484537, "grad_norm": 0.16936178177221572, "learning_rate": 3.375402244665194e-05, "loss": 0.5208, "step": 1080 }, { "epoch": 2.9718213058419245, "grad_norm": 0.20843282321759193, "learning_rate": 3.367802532871306e-05, "loss": 0.5254, "step": 1081 }, { "epoch": 2.9745704467353953, "grad_norm": 0.1837127741621114, "learning_rate": 3.360205160884191e-05, "loss": 0.5313, "step": 1082 }, { "epoch": 2.977319587628866, "grad_norm": 0.199460376426822, "learning_rate": 3.352610156822248e-05, "loss": 0.5292, "step": 1083 }, { "epoch": 2.980068728522337, "grad_norm": 0.24820207311148457, "learning_rate": 3.345017548795116e-05, "loss": 0.5257, "step": 1084 }, { "epoch": 2.9828178694158076, "grad_norm": 0.24493105512545987, "learning_rate": 3.337427364903565e-05, "loss": 0.5224, "step": 1085 }, { "epoch": 2.9855670103092784, "grad_norm": 0.23786424160796374, "learning_rate": 3.329839633239395e-05, "loss": 0.5304, "step": 1086 }, { "epoch": 2.988316151202749, "grad_norm": 0.27502699218694276, "learning_rate": 3.322254381885325e-05, "loss": 0.5281, "step": 1087 }, { "epoch": 2.99106529209622, "grad_norm": 0.27049323273784326, "learning_rate": 3.314671638914902e-05, "loss": 0.5301, "step": 1088 }, { "epoch": 2.9938144329896907, "grad_norm": 0.19737402579388286, "learning_rate": 3.307091432392382e-05, "loss": 0.5288, "step": 1089 }, { "epoch": 2.9965635738831615, "grad_norm": 0.26232022578272246, "learning_rate": 3.299513790372634e-05, "loss": 0.521, "step": 1090 }, { "epoch": 2.9993127147766323, "grad_norm": 0.23892201559966744, "learning_rate": 3.2919387409010424e-05, "loss": 0.6184, "step": 1091 }, { "epoch": 3.002061855670103, "grad_norm": 0.3594917888207816, "learning_rate": 3.284366312013388e-05, "loss": 0.5355, "step": 1092 }, { "epoch": 3.004810996563574, "grad_norm": 0.25599834111682307, "learning_rate": 3.27679653173576e-05, "loss": 0.4746, "step": 1093 }, { "epoch": 3.0075601374570446, "grad_norm": 0.2254148080474178, "learning_rate": 3.2692294280844374e-05, "loss": 0.4669, "step": 1094 }, { "epoch": 3.0103092783505154, "grad_norm": 0.2506660939835137, "learning_rate": 3.261665029065801e-05, "loss": 0.4682, "step": 1095 }, { "epoch": 3.013058419243986, "grad_norm": 0.267195831442756, "learning_rate": 3.254103362676217e-05, "loss": 0.4715, "step": 1096 }, { "epoch": 3.015807560137457, "grad_norm": 0.24179014490822495, "learning_rate": 3.246544456901934e-05, "loss": 0.4778, "step": 1097 }, { "epoch": 3.0185567010309278, "grad_norm": 0.2329847121267379, "learning_rate": 3.238988339718992e-05, "loss": 0.4747, "step": 1098 }, { "epoch": 3.0213058419243985, "grad_norm": 0.2043691678048163, "learning_rate": 3.231435039093104e-05, "loss": 0.4786, "step": 1099 }, { "epoch": 3.0240549828178693, "grad_norm": 0.22745344698265332, "learning_rate": 3.223884582979561e-05, "loss": 0.4657, "step": 1100 }, { "epoch": 3.02680412371134, "grad_norm": 0.22934476133037818, "learning_rate": 3.216336999323127e-05, "loss": 0.4889, "step": 1101 }, { "epoch": 3.029553264604811, "grad_norm": 0.21801654584396532, "learning_rate": 3.208792316057931e-05, "loss": 0.4784, "step": 1102 }, { "epoch": 3.0323024054982817, "grad_norm": 0.2431718990003574, "learning_rate": 3.201250561107374e-05, "loss": 0.4797, "step": 1103 }, { "epoch": 3.0350515463917525, "grad_norm": 0.26915847727829645, "learning_rate": 3.1937117623840114e-05, "loss": 0.4826, "step": 1104 }, { "epoch": 3.0378006872852232, "grad_norm": 0.2351108919147438, "learning_rate": 3.1861759477894656e-05, "loss": 0.4772, "step": 1105 }, { "epoch": 3.040549828178694, "grad_norm": 0.2904532318483722, "learning_rate": 3.178643145214307e-05, "loss": 0.4803, "step": 1106 }, { "epoch": 3.043298969072165, "grad_norm": 0.19635036055007402, "learning_rate": 3.171113382537962e-05, "loss": 0.4701, "step": 1107 }, { "epoch": 3.0460481099656356, "grad_norm": 0.22223894507419642, "learning_rate": 3.1635866876286064e-05, "loss": 0.4741, "step": 1108 }, { "epoch": 3.0487972508591064, "grad_norm": 0.2152706605758009, "learning_rate": 3.156063088343061e-05, "loss": 0.4746, "step": 1109 }, { "epoch": 3.051546391752577, "grad_norm": 0.17409361205810703, "learning_rate": 3.1485426125266896e-05, "loss": 0.478, "step": 1110 }, { "epoch": 3.054295532646048, "grad_norm": 0.21539232573809372, "learning_rate": 3.141025288013296e-05, "loss": 0.4706, "step": 1111 }, { "epoch": 3.0570446735395187, "grad_norm": 0.19308259974547204, "learning_rate": 3.1335111426250216e-05, "loss": 0.4777, "step": 1112 }, { "epoch": 3.0597938144329895, "grad_norm": 0.2276406440928922, "learning_rate": 3.1260002041722405e-05, "loss": 0.4823, "step": 1113 }, { "epoch": 3.0625429553264603, "grad_norm": 0.1977395826548686, "learning_rate": 3.118492500453456e-05, "loss": 0.4781, "step": 1114 }, { "epoch": 3.0652920962199315, "grad_norm": 0.17401988491729847, "learning_rate": 3.1109880592552046e-05, "loss": 0.474, "step": 1115 }, { "epoch": 3.0680412371134023, "grad_norm": 0.1912931420538338, "learning_rate": 3.103486908351943e-05, "loss": 0.4763, "step": 1116 }, { "epoch": 3.070790378006873, "grad_norm": 0.16713338636093636, "learning_rate": 3.0959890755059515e-05, "loss": 0.4796, "step": 1117 }, { "epoch": 3.073539518900344, "grad_norm": 0.20702313839329717, "learning_rate": 3.088494588467232e-05, "loss": 0.4778, "step": 1118 }, { "epoch": 3.0762886597938146, "grad_norm": 0.1616831225744788, "learning_rate": 3.081003474973401e-05, "loss": 0.4756, "step": 1119 }, { "epoch": 3.0790378006872854, "grad_norm": 0.24558159371403487, "learning_rate": 3.0735157627495925e-05, "loss": 0.4789, "step": 1120 }, { "epoch": 3.081786941580756, "grad_norm": 0.17163392192032517, "learning_rate": 3.0660314795083475e-05, "loss": 0.4768, "step": 1121 }, { "epoch": 3.084536082474227, "grad_norm": 0.21495669147653343, "learning_rate": 3.0585506529495186e-05, "loss": 0.4757, "step": 1122 }, { "epoch": 3.0872852233676977, "grad_norm": 0.15045704205058852, "learning_rate": 3.0510733107601664e-05, "loss": 0.473, "step": 1123 }, { "epoch": 3.0900343642611685, "grad_norm": 0.2020087192391845, "learning_rate": 3.0435994806144524e-05, "loss": 0.4735, "step": 1124 }, { "epoch": 3.0927835051546393, "grad_norm": 0.15940495290357568, "learning_rate": 3.0361291901735428e-05, "loss": 0.4709, "step": 1125 }, { "epoch": 3.09553264604811, "grad_norm": 0.18904118815363216, "learning_rate": 3.028662467085501e-05, "loss": 0.4838, "step": 1126 }, { "epoch": 3.098281786941581, "grad_norm": 0.17129808608598124, "learning_rate": 3.0211993389851893e-05, "loss": 0.4768, "step": 1127 }, { "epoch": 3.1010309278350516, "grad_norm": 0.18416866902814874, "learning_rate": 3.013739833494162e-05, "loss": 0.4718, "step": 1128 }, { "epoch": 3.1037800687285224, "grad_norm": 0.1705293761247234, "learning_rate": 3.00628397822057e-05, "loss": 0.4791, "step": 1129 }, { "epoch": 3.106529209621993, "grad_norm": 0.20209221200923272, "learning_rate": 2.9988318007590494e-05, "loss": 0.4823, "step": 1130 }, { "epoch": 3.109278350515464, "grad_norm": 0.2118645657673755, "learning_rate": 2.991383328690627e-05, "loss": 0.4782, "step": 1131 }, { "epoch": 3.1120274914089348, "grad_norm": 0.17556823451574127, "learning_rate": 2.9839385895826175e-05, "loss": 0.4783, "step": 1132 }, { "epoch": 3.1147766323024055, "grad_norm": 0.19795461740770434, "learning_rate": 2.9764976109885166e-05, "loss": 0.4818, "step": 1133 }, { "epoch": 3.1175257731958763, "grad_norm": 0.16701933368048089, "learning_rate": 2.9690604204479026e-05, "loss": 0.476, "step": 1134 }, { "epoch": 3.120274914089347, "grad_norm": 0.20618027859459478, "learning_rate": 2.9616270454863368e-05, "loss": 0.4798, "step": 1135 }, { "epoch": 3.123024054982818, "grad_norm": 0.1431586178012281, "learning_rate": 2.9541975136152548e-05, "loss": 0.478, "step": 1136 }, { "epoch": 3.1257731958762887, "grad_norm": 0.17793879100992058, "learning_rate": 2.946771852331873e-05, "loss": 0.4756, "step": 1137 }, { "epoch": 3.1285223367697594, "grad_norm": 0.16681103898774272, "learning_rate": 2.939350089119077e-05, "loss": 0.4715, "step": 1138 }, { "epoch": 3.13127147766323, "grad_norm": 0.17261193531928995, "learning_rate": 2.9319322514453286e-05, "loss": 0.4759, "step": 1139 }, { "epoch": 3.134020618556701, "grad_norm": 0.1582640372762462, "learning_rate": 2.9245183667645626e-05, "loss": 0.4749, "step": 1140 }, { "epoch": 3.136769759450172, "grad_norm": 0.17322436049835935, "learning_rate": 2.9171084625160797e-05, "loss": 0.4703, "step": 1141 }, { "epoch": 3.1395189003436426, "grad_norm": 0.17367955729692558, "learning_rate": 2.909702566124452e-05, "loss": 0.4792, "step": 1142 }, { "epoch": 3.1422680412371133, "grad_norm": 0.16722751357988883, "learning_rate": 2.9023007049994157e-05, "loss": 0.4717, "step": 1143 }, { "epoch": 3.145017182130584, "grad_norm": 0.1728042548273763, "learning_rate": 2.8949029065357753e-05, "loss": 0.4818, "step": 1144 }, { "epoch": 3.147766323024055, "grad_norm": 0.18069700472194267, "learning_rate": 2.8875091981132963e-05, "loss": 0.476, "step": 1145 }, { "epoch": 3.1505154639175257, "grad_norm": 0.15820067646687608, "learning_rate": 2.8801196070966054e-05, "loss": 0.4666, "step": 1146 }, { "epoch": 3.1532646048109965, "grad_norm": 0.15258574864023677, "learning_rate": 2.8727341608350952e-05, "loss": 0.4759, "step": 1147 }, { "epoch": 3.1560137457044672, "grad_norm": 0.1960418259666068, "learning_rate": 2.8653528866628132e-05, "loss": 0.4769, "step": 1148 }, { "epoch": 3.158762886597938, "grad_norm": 0.15371731621163243, "learning_rate": 2.8579758118983716e-05, "loss": 0.4842, "step": 1149 }, { "epoch": 3.161512027491409, "grad_norm": 0.19068409005542658, "learning_rate": 2.8506029638448347e-05, "loss": 0.4776, "step": 1150 }, { "epoch": 3.1642611683848796, "grad_norm": 0.17886606119360748, "learning_rate": 2.843234369789625e-05, "loss": 0.4744, "step": 1151 }, { "epoch": 3.1670103092783504, "grad_norm": 0.19710185432909114, "learning_rate": 2.8358700570044232e-05, "loss": 0.4813, "step": 1152 }, { "epoch": 3.169759450171821, "grad_norm": 0.18041901946396155, "learning_rate": 2.8285100527450623e-05, "loss": 0.4833, "step": 1153 }, { "epoch": 3.172508591065292, "grad_norm": 0.16581599590174512, "learning_rate": 2.8211543842514288e-05, "loss": 0.4739, "step": 1154 }, { "epoch": 3.1752577319587627, "grad_norm": 0.15872506391538035, "learning_rate": 2.8138030787473635e-05, "loss": 0.4854, "step": 1155 }, { "epoch": 3.1780068728522335, "grad_norm": 0.18595712745784665, "learning_rate": 2.806456163440559e-05, "loss": 0.4791, "step": 1156 }, { "epoch": 3.1807560137457043, "grad_norm": 0.1585604017012495, "learning_rate": 2.7991136655224602e-05, "loss": 0.4794, "step": 1157 }, { "epoch": 3.183505154639175, "grad_norm": 0.16724675223224328, "learning_rate": 2.791775612168161e-05, "loss": 0.484, "step": 1158 }, { "epoch": 3.1862542955326463, "grad_norm": 0.1598241225286718, "learning_rate": 2.784442030536309e-05, "loss": 0.4816, "step": 1159 }, { "epoch": 3.189003436426117, "grad_norm": 0.16901071883801533, "learning_rate": 2.7771129477689972e-05, "loss": 0.473, "step": 1160 }, { "epoch": 3.191752577319588, "grad_norm": 0.13853638340372265, "learning_rate": 2.7697883909916732e-05, "loss": 0.4736, "step": 1161 }, { "epoch": 3.1945017182130586, "grad_norm": 0.16091561527631862, "learning_rate": 2.7624683873130287e-05, "loss": 0.4805, "step": 1162 }, { "epoch": 3.1972508591065294, "grad_norm": 0.14598087437104398, "learning_rate": 2.755152963824905e-05, "loss": 0.4835, "step": 1163 }, { "epoch": 3.2, "grad_norm": 0.175532838361529, "learning_rate": 2.7478421476021968e-05, "loss": 0.4743, "step": 1164 }, { "epoch": 3.202749140893471, "grad_norm": 0.1337685529203635, "learning_rate": 2.740535965702739e-05, "loss": 0.4734, "step": 1165 }, { "epoch": 3.2054982817869417, "grad_norm": 0.1735634483137809, "learning_rate": 2.7332344451672214e-05, "loss": 0.4709, "step": 1166 }, { "epoch": 3.2082474226804125, "grad_norm": 0.15693421713978023, "learning_rate": 2.7259376130190783e-05, "loss": 0.4741, "step": 1167 }, { "epoch": 3.2109965635738833, "grad_norm": 0.1330436144853641, "learning_rate": 2.718645496264392e-05, "loss": 0.4739, "step": 1168 }, { "epoch": 3.213745704467354, "grad_norm": 0.17309959852487516, "learning_rate": 2.711358121891795e-05, "loss": 0.4705, "step": 1169 }, { "epoch": 3.216494845360825, "grad_norm": 0.1642984807751716, "learning_rate": 2.704075516872366e-05, "loss": 0.4726, "step": 1170 }, { "epoch": 3.2192439862542956, "grad_norm": 0.14722892838781462, "learning_rate": 2.6967977081595304e-05, "loss": 0.4775, "step": 1171 }, { "epoch": 3.2219931271477664, "grad_norm": 0.16887717979369643, "learning_rate": 2.6895247226889672e-05, "loss": 0.4738, "step": 1172 }, { "epoch": 3.224742268041237, "grad_norm": 0.17049853048281954, "learning_rate": 2.682256587378498e-05, "loss": 0.4719, "step": 1173 }, { "epoch": 3.227491408934708, "grad_norm": 0.15138446686573354, "learning_rate": 2.6749933291279994e-05, "loss": 0.4856, "step": 1174 }, { "epoch": 3.2302405498281788, "grad_norm": 0.16148723512701157, "learning_rate": 2.6677349748192934e-05, "loss": 0.4762, "step": 1175 }, { "epoch": 3.2329896907216495, "grad_norm": 0.14912677660800724, "learning_rate": 2.6604815513160556e-05, "loss": 0.4653, "step": 1176 }, { "epoch": 3.2357388316151203, "grad_norm": 0.13862596201937702, "learning_rate": 2.6532330854637086e-05, "loss": 0.4743, "step": 1177 }, { "epoch": 3.238487972508591, "grad_norm": 0.15927565474045197, "learning_rate": 2.645989604089331e-05, "loss": 0.4746, "step": 1178 }, { "epoch": 3.241237113402062, "grad_norm": 0.17434623830674337, "learning_rate": 2.638751134001549e-05, "loss": 0.48, "step": 1179 }, { "epoch": 3.2439862542955327, "grad_norm": 0.1447492759368682, "learning_rate": 2.6315177019904423e-05, "loss": 0.4781, "step": 1180 }, { "epoch": 3.2467353951890034, "grad_norm": 0.16927963210867344, "learning_rate": 2.624289334827448e-05, "loss": 0.4704, "step": 1181 }, { "epoch": 3.2494845360824742, "grad_norm": 0.1665258090219738, "learning_rate": 2.6170660592652545e-05, "loss": 0.4781, "step": 1182 }, { "epoch": 3.252233676975945, "grad_norm": 0.15855482153444905, "learning_rate": 2.609847902037706e-05, "loss": 0.4743, "step": 1183 }, { "epoch": 3.254982817869416, "grad_norm": 0.17971555626146168, "learning_rate": 2.6026348898597057e-05, "loss": 0.4803, "step": 1184 }, { "epoch": 3.2577319587628866, "grad_norm": 0.2816535496494584, "learning_rate": 2.5954270494271116e-05, "loss": 0.4761, "step": 1185 }, { "epoch": 3.2604810996563574, "grad_norm": 0.23781070038647936, "learning_rate": 2.588224407416645e-05, "loss": 0.4729, "step": 1186 }, { "epoch": 3.263230240549828, "grad_norm": 0.23075459880557803, "learning_rate": 2.581026990485781e-05, "loss": 0.4764, "step": 1187 }, { "epoch": 3.265979381443299, "grad_norm": 0.14086275865969622, "learning_rate": 2.5738348252726607e-05, "loss": 0.4809, "step": 1188 }, { "epoch": 3.2687285223367697, "grad_norm": 0.18978376603435007, "learning_rate": 2.566647938395989e-05, "loss": 0.4817, "step": 1189 }, { "epoch": 3.2714776632302405, "grad_norm": 0.16648197235336828, "learning_rate": 2.559466356454933e-05, "loss": 0.4703, "step": 1190 }, { "epoch": 3.2742268041237113, "grad_norm": 0.17731919695127468, "learning_rate": 2.5522901060290272e-05, "loss": 0.4757, "step": 1191 }, { "epoch": 3.276975945017182, "grad_norm": 0.22533634711616554, "learning_rate": 2.545119213678072e-05, "loss": 0.4779, "step": 1192 }, { "epoch": 3.279725085910653, "grad_norm": 0.29341047419363175, "learning_rate": 2.53795370594204e-05, "loss": 0.4809, "step": 1193 }, { "epoch": 3.2824742268041236, "grad_norm": 0.17396435349550476, "learning_rate": 2.530793609340974e-05, "loss": 0.4796, "step": 1194 }, { "epoch": 3.2852233676975944, "grad_norm": 0.20383849713641408, "learning_rate": 2.523638950374886e-05, "loss": 0.4721, "step": 1195 }, { "epoch": 3.287972508591065, "grad_norm": 0.16601271855097022, "learning_rate": 2.5164897555236686e-05, "loss": 0.4804, "step": 1196 }, { "epoch": 3.290721649484536, "grad_norm": 0.30470051683065796, "learning_rate": 2.509346051246988e-05, "loss": 0.4792, "step": 1197 }, { "epoch": 3.2934707903780067, "grad_norm": 0.2812298427453436, "learning_rate": 2.5022078639841918e-05, "loss": 0.4782, "step": 1198 }, { "epoch": 3.2962199312714775, "grad_norm": 0.1597994542210946, "learning_rate": 2.495075220154206e-05, "loss": 0.4823, "step": 1199 }, { "epoch": 3.2989690721649483, "grad_norm": 0.3095999568945906, "learning_rate": 2.4879481461554405e-05, "loss": 0.4762, "step": 1200 }, { "epoch": 3.301718213058419, "grad_norm": 0.24822013052998354, "learning_rate": 2.4808266683656932e-05, "loss": 0.4805, "step": 1201 }, { "epoch": 3.30446735395189, "grad_norm": 0.19304267454670382, "learning_rate": 2.473710813142049e-05, "loss": 0.4709, "step": 1202 }, { "epoch": 3.3072164948453606, "grad_norm": 0.22511180187157134, "learning_rate": 2.4666006068207793e-05, "loss": 0.4791, "step": 1203 }, { "epoch": 3.3099656357388314, "grad_norm": 0.17597702188477718, "learning_rate": 2.4594960757172555e-05, "loss": 0.4825, "step": 1204 }, { "epoch": 3.312714776632302, "grad_norm": 0.18231300214622423, "learning_rate": 2.4523972461258386e-05, "loss": 0.4794, "step": 1205 }, { "epoch": 3.315463917525773, "grad_norm": 0.25489199823717157, "learning_rate": 2.445304144319792e-05, "loss": 0.4795, "step": 1206 }, { "epoch": 3.3182130584192437, "grad_norm": 0.15105215480831594, "learning_rate": 2.4382167965511774e-05, "loss": 0.4902, "step": 1207 }, { "epoch": 3.320962199312715, "grad_norm": 0.2799009342897516, "learning_rate": 2.431135229050763e-05, "loss": 0.4818, "step": 1208 }, { "epoch": 3.3237113402061857, "grad_norm": 0.21951158609310864, "learning_rate": 2.4240594680279204e-05, "loss": 0.479, "step": 1209 }, { "epoch": 3.3264604810996565, "grad_norm": 0.19395691144843719, "learning_rate": 2.416989539670536e-05, "loss": 0.479, "step": 1210 }, { "epoch": 3.3292096219931273, "grad_norm": 0.17094551246350398, "learning_rate": 2.409925470144903e-05, "loss": 0.4808, "step": 1211 }, { "epoch": 3.331958762886598, "grad_norm": 0.23964152078624051, "learning_rate": 2.402867285595635e-05, "loss": 0.4764, "step": 1212 }, { "epoch": 3.334707903780069, "grad_norm": 0.2782973993431327, "learning_rate": 2.3958150121455638e-05, "loss": 0.4811, "step": 1213 }, { "epoch": 3.3374570446735397, "grad_norm": 0.1640795235166955, "learning_rate": 2.3887686758956425e-05, "loss": 0.4793, "step": 1214 }, { "epoch": 3.3402061855670104, "grad_norm": 0.3070072205373487, "learning_rate": 2.3817283029248536e-05, "loss": 0.475, "step": 1215 }, { "epoch": 3.342955326460481, "grad_norm": 0.23275102863915093, "learning_rate": 2.3746939192901052e-05, "loss": 0.479, "step": 1216 }, { "epoch": 3.345704467353952, "grad_norm": 0.23028817088895479, "learning_rate": 2.3676655510261402e-05, "loss": 0.4735, "step": 1217 }, { "epoch": 3.3484536082474228, "grad_norm": 0.320766094957754, "learning_rate": 2.3606432241454398e-05, "loss": 0.4877, "step": 1218 }, { "epoch": 3.3512027491408936, "grad_norm": 0.2518366256607952, "learning_rate": 2.353626964638122e-05, "loss": 0.4733, "step": 1219 }, { "epoch": 3.3539518900343643, "grad_norm": 0.1904395352124866, "learning_rate": 2.34661679847185e-05, "loss": 0.4895, "step": 1220 }, { "epoch": 3.356701030927835, "grad_norm": 0.3279385370633572, "learning_rate": 2.3396127515917392e-05, "loss": 0.4777, "step": 1221 }, { "epoch": 3.359450171821306, "grad_norm": 0.20836286167731985, "learning_rate": 2.3326148499202518e-05, "loss": 0.4765, "step": 1222 }, { "epoch": 3.3621993127147767, "grad_norm": 0.22895782643501986, "learning_rate": 2.325623119357107e-05, "loss": 0.4883, "step": 1223 }, { "epoch": 3.3649484536082475, "grad_norm": 0.29237657945287854, "learning_rate": 2.3186375857791896e-05, "loss": 0.483, "step": 1224 }, { "epoch": 3.3676975945017182, "grad_norm": 0.18465319808299466, "learning_rate": 2.3116582750404427e-05, "loss": 0.472, "step": 1225 }, { "epoch": 3.370446735395189, "grad_norm": 0.20550527495867124, "learning_rate": 2.304685212971781e-05, "loss": 0.477, "step": 1226 }, { "epoch": 3.37319587628866, "grad_norm": 0.2692822007314835, "learning_rate": 2.2977184253809913e-05, "loss": 0.4763, "step": 1227 }, { "epoch": 3.3759450171821306, "grad_norm": 0.1708056032464783, "learning_rate": 2.2907579380526387e-05, "loss": 0.48, "step": 1228 }, { "epoch": 3.3786941580756014, "grad_norm": 0.368849523723143, "learning_rate": 2.2838037767479713e-05, "loss": 0.4797, "step": 1229 }, { "epoch": 3.381443298969072, "grad_norm": 0.21646292002270615, "learning_rate": 2.2768559672048218e-05, "loss": 0.474, "step": 1230 }, { "epoch": 3.384192439862543, "grad_norm": 0.26373109002232037, "learning_rate": 2.2699145351375186e-05, "loss": 0.4803, "step": 1231 }, { "epoch": 3.3869415807560137, "grad_norm": 0.2843624854578763, "learning_rate": 2.262979506236784e-05, "loss": 0.4826, "step": 1232 }, { "epoch": 3.3896907216494845, "grad_norm": 0.19833041217898748, "learning_rate": 2.256050906169641e-05, "loss": 0.491, "step": 1233 }, { "epoch": 3.3924398625429553, "grad_norm": 0.2800244352558394, "learning_rate": 2.2491287605793204e-05, "loss": 0.4707, "step": 1234 }, { "epoch": 3.395189003436426, "grad_norm": 0.18325778358692565, "learning_rate": 2.2422130950851643e-05, "loss": 0.4744, "step": 1235 }, { "epoch": 3.397938144329897, "grad_norm": 0.3008410533335858, "learning_rate": 2.235303935282531e-05, "loss": 0.4846, "step": 1236 }, { "epoch": 3.4006872852233676, "grad_norm": 0.19194203147930453, "learning_rate": 2.2284013067427025e-05, "loss": 0.4718, "step": 1237 }, { "epoch": 3.4034364261168384, "grad_norm": 0.34898181416099955, "learning_rate": 2.2215052350127834e-05, "loss": 0.484, "step": 1238 }, { "epoch": 3.406185567010309, "grad_norm": 0.2356425803363679, "learning_rate": 2.214615745615619e-05, "loss": 0.4774, "step": 1239 }, { "epoch": 3.40893470790378, "grad_norm": 0.21466246650328097, "learning_rate": 2.207732864049686e-05, "loss": 0.4812, "step": 1240 }, { "epoch": 3.4116838487972507, "grad_norm": 0.26377375192350677, "learning_rate": 2.200856615789009e-05, "loss": 0.4912, "step": 1241 }, { "epoch": 3.4144329896907215, "grad_norm": 0.15695872082264978, "learning_rate": 2.1939870262830577e-05, "loss": 0.4715, "step": 1242 }, { "epoch": 3.4171821305841923, "grad_norm": 0.28736994518322967, "learning_rate": 2.187124120956666e-05, "loss": 0.4811, "step": 1243 }, { "epoch": 3.419931271477663, "grad_norm": 0.24847231178584678, "learning_rate": 2.1802679252099184e-05, "loss": 0.4784, "step": 1244 }, { "epoch": 3.422680412371134, "grad_norm": 0.17957114598770363, "learning_rate": 2.1734184644180715e-05, "loss": 0.4706, "step": 1245 }, { "epoch": 3.4254295532646046, "grad_norm": 0.3386000360336158, "learning_rate": 2.16657576393146e-05, "loss": 0.4764, "step": 1246 }, { "epoch": 3.4281786941580754, "grad_norm": 0.14262617741372224, "learning_rate": 2.1597398490753917e-05, "loss": 0.4645, "step": 1247 }, { "epoch": 3.4309278350515466, "grad_norm": 0.2371298714043362, "learning_rate": 2.152910745150063e-05, "loss": 0.4778, "step": 1248 }, { "epoch": 3.4336769759450174, "grad_norm": 0.15007549750939145, "learning_rate": 2.1460884774304614e-05, "loss": 0.481, "step": 1249 }, { "epoch": 3.436426116838488, "grad_norm": 0.2666157528473445, "learning_rate": 2.1392730711662755e-05, "loss": 0.473, "step": 1250 }, { "epoch": 3.439175257731959, "grad_norm": 0.19370621015791017, "learning_rate": 2.1324645515817988e-05, "loss": 0.4916, "step": 1251 }, { "epoch": 3.4419243986254298, "grad_norm": 0.1363626153767201, "learning_rate": 2.125662943875832e-05, "loss": 0.4783, "step": 1252 }, { "epoch": 3.4446735395189005, "grad_norm": 0.1558927651589561, "learning_rate": 2.1188682732215978e-05, "loss": 0.4863, "step": 1253 }, { "epoch": 3.4474226804123713, "grad_norm": 0.1878171719282548, "learning_rate": 2.1120805647666484e-05, "loss": 0.483, "step": 1254 }, { "epoch": 3.450171821305842, "grad_norm": 0.1706433093556562, "learning_rate": 2.1052998436327624e-05, "loss": 0.4766, "step": 1255 }, { "epoch": 3.452920962199313, "grad_norm": 0.14594078534672697, "learning_rate": 2.0985261349158606e-05, "loss": 0.4701, "step": 1256 }, { "epoch": 3.4556701030927837, "grad_norm": 0.2713964023439289, "learning_rate": 2.0917594636859084e-05, "loss": 0.4734, "step": 1257 }, { "epoch": 3.4584192439862544, "grad_norm": 0.20301631548067403, "learning_rate": 2.084999854986829e-05, "loss": 0.4664, "step": 1258 }, { "epoch": 3.4611683848797252, "grad_norm": 0.21017329606796895, "learning_rate": 2.0782473338364034e-05, "loss": 0.4801, "step": 1259 }, { "epoch": 3.463917525773196, "grad_norm": 0.2024174569645721, "learning_rate": 2.0715019252261786e-05, "loss": 0.473, "step": 1260 }, { "epoch": 3.466666666666667, "grad_norm": 0.13513087820592787, "learning_rate": 2.0647636541213843e-05, "loss": 0.479, "step": 1261 }, { "epoch": 3.4694158075601376, "grad_norm": 0.15323818419427804, "learning_rate": 2.0580325454608294e-05, "loss": 0.479, "step": 1262 }, { "epoch": 3.4721649484536083, "grad_norm": 0.1883180981275938, "learning_rate": 2.051308624156815e-05, "loss": 0.4797, "step": 1263 }, { "epoch": 3.474914089347079, "grad_norm": 0.18535688822536886, "learning_rate": 2.0445919150950387e-05, "loss": 0.4828, "step": 1264 }, { "epoch": 3.47766323024055, "grad_norm": 0.1735738790195007, "learning_rate": 2.0378824431345116e-05, "loss": 0.4752, "step": 1265 }, { "epoch": 3.4804123711340207, "grad_norm": 0.25380262888772, "learning_rate": 2.0311802331074543e-05, "loss": 0.479, "step": 1266 }, { "epoch": 3.4831615120274915, "grad_norm": 0.17442121730064505, "learning_rate": 2.024485309819213e-05, "loss": 0.4739, "step": 1267 }, { "epoch": 3.4859106529209622, "grad_norm": 0.284320944684907, "learning_rate": 2.01779769804816e-05, "loss": 0.4843, "step": 1268 }, { "epoch": 3.488659793814433, "grad_norm": 0.1625892440951168, "learning_rate": 2.011117422545616e-05, "loss": 0.4763, "step": 1269 }, { "epoch": 3.491408934707904, "grad_norm": 0.28275001646900066, "learning_rate": 2.004444508035743e-05, "loss": 0.4699, "step": 1270 }, { "epoch": 3.4941580756013746, "grad_norm": 0.17127917891669825, "learning_rate": 1.9977789792154615e-05, "loss": 0.4882, "step": 1271 }, { "epoch": 3.4969072164948454, "grad_norm": 0.155941741872769, "learning_rate": 1.991120860754356e-05, "loss": 0.4692, "step": 1272 }, { "epoch": 3.499656357388316, "grad_norm": 0.15459720154148457, "learning_rate": 1.984470177294588e-05, "loss": 0.4841, "step": 1273 }, { "epoch": 3.502405498281787, "grad_norm": 0.2196341305641022, "learning_rate": 1.9778269534507987e-05, "loss": 0.4713, "step": 1274 }, { "epoch": 3.5051546391752577, "grad_norm": 0.22478563840945578, "learning_rate": 1.971191213810021e-05, "loss": 0.4892, "step": 1275 }, { "epoch": 3.5079037800687285, "grad_norm": 0.1739449107806518, "learning_rate": 1.964562982931588e-05, "loss": 0.4765, "step": 1276 }, { "epoch": 3.5106529209621993, "grad_norm": 0.27214067372219086, "learning_rate": 1.9579422853470436e-05, "loss": 0.4815, "step": 1277 }, { "epoch": 3.51340206185567, "grad_norm": 0.17400788077272825, "learning_rate": 1.95132914556005e-05, "loss": 0.4727, "step": 1278 }, { "epoch": 3.516151202749141, "grad_norm": 0.19337030209527964, "learning_rate": 1.9447235880462947e-05, "loss": 0.4804, "step": 1279 }, { "epoch": 3.5189003436426116, "grad_norm": 0.1514665359437226, "learning_rate": 1.938125637253409e-05, "loss": 0.4729, "step": 1280 }, { "epoch": 3.5216494845360824, "grad_norm": 0.27810047698132523, "learning_rate": 1.9315353176008655e-05, "loss": 0.478, "step": 1281 }, { "epoch": 3.524398625429553, "grad_norm": 0.2052552676995858, "learning_rate": 1.9249526534798965e-05, "loss": 0.4825, "step": 1282 }, { "epoch": 3.527147766323024, "grad_norm": 0.1638869338895521, "learning_rate": 1.9183776692533977e-05, "loss": 0.4754, "step": 1283 }, { "epoch": 3.5298969072164947, "grad_norm": 0.1623552411012285, "learning_rate": 1.9118103892558438e-05, "loss": 0.4721, "step": 1284 }, { "epoch": 3.5326460481099655, "grad_norm": 0.2765823979801074, "learning_rate": 1.9052508377931945e-05, "loss": 0.4721, "step": 1285 }, { "epoch": 3.5353951890034363, "grad_norm": 0.20787425872060086, "learning_rate": 1.8986990391428056e-05, "loss": 0.4721, "step": 1286 }, { "epoch": 3.538144329896907, "grad_norm": 0.23103945726058686, "learning_rate": 1.8921550175533384e-05, "loss": 0.4804, "step": 1287 }, { "epoch": 3.540893470790378, "grad_norm": 0.14819667238986017, "learning_rate": 1.885618797244674e-05, "loss": 0.4781, "step": 1288 }, { "epoch": 3.5436426116838486, "grad_norm": 0.34832469601339244, "learning_rate": 1.8790904024078166e-05, "loss": 0.4853, "step": 1289 }, { "epoch": 3.5463917525773194, "grad_norm": 0.17762443199441277, "learning_rate": 1.8725698572048096e-05, "loss": 0.4687, "step": 1290 }, { "epoch": 3.54914089347079, "grad_norm": 0.220110309410904, "learning_rate": 1.866057185768644e-05, "loss": 0.4722, "step": 1291 }, { "epoch": 3.551890034364261, "grad_norm": 0.1568411202207363, "learning_rate": 1.8595524122031695e-05, "loss": 0.477, "step": 1292 }, { "epoch": 3.5546391752577318, "grad_norm": 0.1487231862533628, "learning_rate": 1.8530555605830048e-05, "loss": 0.4812, "step": 1293 }, { "epoch": 3.5573883161512025, "grad_norm": 0.27212850028975205, "learning_rate": 1.8465666549534487e-05, "loss": 0.4716, "step": 1294 }, { "epoch": 3.5601374570446733, "grad_norm": 0.26360242336986534, "learning_rate": 1.840085719330394e-05, "loss": 0.4794, "step": 1295 }, { "epoch": 3.562886597938144, "grad_norm": 0.20568964864290215, "learning_rate": 1.8336127777002336e-05, "loss": 0.4808, "step": 1296 }, { "epoch": 3.565635738831615, "grad_norm": 0.282051987566169, "learning_rate": 1.8271478540197735e-05, "loss": 0.4761, "step": 1297 }, { "epoch": 3.5683848797250857, "grad_norm": 0.16289176663769198, "learning_rate": 1.8206909722161444e-05, "loss": 0.4774, "step": 1298 }, { "epoch": 3.5711340206185564, "grad_norm": 0.2575145343236498, "learning_rate": 1.814242156186718e-05, "loss": 0.4795, "step": 1299 }, { "epoch": 3.5738831615120272, "grad_norm": 0.1486906599375562, "learning_rate": 1.8078014297990104e-05, "loss": 0.4824, "step": 1300 }, { "epoch": 3.576632302405498, "grad_norm": 0.1501909622248065, "learning_rate": 1.8013688168905946e-05, "loss": 0.4804, "step": 1301 }, { "epoch": 3.5793814432989692, "grad_norm": 0.15230793704069007, "learning_rate": 1.7949443412690182e-05, "loss": 0.4751, "step": 1302 }, { "epoch": 3.58213058419244, "grad_norm": 0.16635368754820404, "learning_rate": 1.788528026711715e-05, "loss": 0.4806, "step": 1303 }, { "epoch": 3.584879725085911, "grad_norm": 0.1347568536197528, "learning_rate": 1.78211989696591e-05, "loss": 0.4723, "step": 1304 }, { "epoch": 3.5876288659793816, "grad_norm": 0.19251735027744715, "learning_rate": 1.7757199757485363e-05, "loss": 0.4796, "step": 1305 }, { "epoch": 3.5903780068728524, "grad_norm": 0.14511887693354616, "learning_rate": 1.7693282867461464e-05, "loss": 0.4875, "step": 1306 }, { "epoch": 3.593127147766323, "grad_norm": 0.21823490371477205, "learning_rate": 1.762944853614828e-05, "loss": 0.4831, "step": 1307 }, { "epoch": 3.595876288659794, "grad_norm": 0.14789140619202265, "learning_rate": 1.756569699980111e-05, "loss": 0.4719, "step": 1308 }, { "epoch": 3.5986254295532647, "grad_norm": 0.1537689543052464, "learning_rate": 1.7502028494368777e-05, "loss": 0.4775, "step": 1309 }, { "epoch": 3.6013745704467355, "grad_norm": 0.16438166123180578, "learning_rate": 1.743844325549289e-05, "loss": 0.4778, "step": 1310 }, { "epoch": 3.6041237113402063, "grad_norm": 0.1309598142497098, "learning_rate": 1.737494151850682e-05, "loss": 0.4799, "step": 1311 }, { "epoch": 3.606872852233677, "grad_norm": 0.13501349366551338, "learning_rate": 1.731152351843492e-05, "loss": 0.4778, "step": 1312 }, { "epoch": 3.609621993127148, "grad_norm": 0.12271151015542832, "learning_rate": 1.7248189489991594e-05, "loss": 0.4804, "step": 1313 }, { "epoch": 3.6123711340206186, "grad_norm": 0.1579684841055174, "learning_rate": 1.7184939667580512e-05, "loss": 0.4756, "step": 1314 }, { "epoch": 3.6151202749140894, "grad_norm": 0.11550758392122117, "learning_rate": 1.712177428529365e-05, "loss": 0.4813, "step": 1315 }, { "epoch": 3.61786941580756, "grad_norm": 0.1446519336760662, "learning_rate": 1.7058693576910495e-05, "loss": 0.4761, "step": 1316 }, { "epoch": 3.620618556701031, "grad_norm": 0.1387283733233433, "learning_rate": 1.6995697775897097e-05, "loss": 0.4855, "step": 1317 }, { "epoch": 3.6233676975945017, "grad_norm": 0.12450900277305127, "learning_rate": 1.6932787115405318e-05, "loss": 0.4768, "step": 1318 }, { "epoch": 3.6261168384879725, "grad_norm": 0.15118692423046173, "learning_rate": 1.6869961828271892e-05, "loss": 0.4795, "step": 1319 }, { "epoch": 3.6288659793814433, "grad_norm": 0.1275953061475887, "learning_rate": 1.6807222147017558e-05, "loss": 0.4844, "step": 1320 }, { "epoch": 3.631615120274914, "grad_norm": 0.12357741430968726, "learning_rate": 1.6744568303846237e-05, "loss": 0.4718, "step": 1321 }, { "epoch": 3.634364261168385, "grad_norm": 0.13324292495631146, "learning_rate": 1.6682000530644177e-05, "loss": 0.4828, "step": 1322 }, { "epoch": 3.6371134020618556, "grad_norm": 0.12727908902135235, "learning_rate": 1.6619519058979044e-05, "loss": 0.4731, "step": 1323 }, { "epoch": 3.6398625429553264, "grad_norm": 0.14031695435202224, "learning_rate": 1.6557124120099108e-05, "loss": 0.4723, "step": 1324 }, { "epoch": 3.642611683848797, "grad_norm": 0.1238822517218841, "learning_rate": 1.6494815944932376e-05, "loss": 0.4767, "step": 1325 }, { "epoch": 3.645360824742268, "grad_norm": 0.14383655055562664, "learning_rate": 1.6432594764085738e-05, "loss": 0.4744, "step": 1326 }, { "epoch": 3.6481099656357387, "grad_norm": 0.13074212391556805, "learning_rate": 1.6370460807844096e-05, "loss": 0.4719, "step": 1327 }, { "epoch": 3.6508591065292095, "grad_norm": 0.1408486433749572, "learning_rate": 1.6308414306169545e-05, "loss": 0.4792, "step": 1328 }, { "epoch": 3.6536082474226803, "grad_norm": 0.1425866228634816, "learning_rate": 1.6246455488700522e-05, "loss": 0.4759, "step": 1329 }, { "epoch": 3.656357388316151, "grad_norm": 0.13452543614110554, "learning_rate": 1.61845845847509e-05, "loss": 0.4825, "step": 1330 }, { "epoch": 3.659106529209622, "grad_norm": 0.123447733182089, "learning_rate": 1.6122801823309195e-05, "loss": 0.4839, "step": 1331 }, { "epoch": 3.6618556701030927, "grad_norm": 0.141262084894595, "learning_rate": 1.6061107433037707e-05, "loss": 0.4803, "step": 1332 }, { "epoch": 3.6646048109965634, "grad_norm": 0.16335769694535568, "learning_rate": 1.5999501642271664e-05, "loss": 0.491, "step": 1333 }, { "epoch": 3.667353951890034, "grad_norm": 0.13731203188970648, "learning_rate": 1.593798467901837e-05, "loss": 0.4764, "step": 1334 }, { "epoch": 3.670103092783505, "grad_norm": 0.14344242124627954, "learning_rate": 1.5876556770956394e-05, "loss": 0.4776, "step": 1335 }, { "epoch": 3.672852233676976, "grad_norm": 0.14287518229033996, "learning_rate": 1.581521814543467e-05, "loss": 0.4729, "step": 1336 }, { "epoch": 3.675601374570447, "grad_norm": 0.13919137334446374, "learning_rate": 1.575396902947175e-05, "loss": 0.4803, "step": 1337 }, { "epoch": 3.678350515463918, "grad_norm": 0.1332949080422285, "learning_rate": 1.5692809649754855e-05, "loss": 0.4732, "step": 1338 }, { "epoch": 3.6810996563573886, "grad_norm": 0.15251079258406947, "learning_rate": 1.563174023263909e-05, "loss": 0.4721, "step": 1339 }, { "epoch": 3.6838487972508593, "grad_norm": 0.12884291771360323, "learning_rate": 1.557076100414663e-05, "loss": 0.4771, "step": 1340 }, { "epoch": 3.68659793814433, "grad_norm": 0.14077765554999186, "learning_rate": 1.5509872189965826e-05, "loss": 0.4765, "step": 1341 }, { "epoch": 3.689347079037801, "grad_norm": 0.12097139782888651, "learning_rate": 1.5449074015450416e-05, "loss": 0.4816, "step": 1342 }, { "epoch": 3.6920962199312717, "grad_norm": 0.12394128179144817, "learning_rate": 1.5388366705618656e-05, "loss": 0.4719, "step": 1343 }, { "epoch": 3.6948453608247425, "grad_norm": 0.12482467627547454, "learning_rate": 1.5327750485152546e-05, "loss": 0.4763, "step": 1344 }, { "epoch": 3.6975945017182132, "grad_norm": 0.12282644196525853, "learning_rate": 1.5267225578396923e-05, "loss": 0.4773, "step": 1345 }, { "epoch": 3.700343642611684, "grad_norm": 0.12030648328844293, "learning_rate": 1.5206792209358665e-05, "loss": 0.486, "step": 1346 }, { "epoch": 3.703092783505155, "grad_norm": 0.1242904271715769, "learning_rate": 1.5146450601705866e-05, "loss": 0.4788, "step": 1347 }, { "epoch": 3.7058419243986256, "grad_norm": 0.12938487720592814, "learning_rate": 1.5086200978767051e-05, "loss": 0.482, "step": 1348 }, { "epoch": 3.7085910652920964, "grad_norm": 0.1181010757249468, "learning_rate": 1.5026043563530213e-05, "loss": 0.4817, "step": 1349 }, { "epoch": 3.711340206185567, "grad_norm": 0.12542244186954696, "learning_rate": 1.4965978578642152e-05, "loss": 0.4829, "step": 1350 }, { "epoch": 3.714089347079038, "grad_norm": 0.1301001364291388, "learning_rate": 1.4906006246407531e-05, "loss": 0.4801, "step": 1351 }, { "epoch": 3.7168384879725087, "grad_norm": 0.11752586529006614, "learning_rate": 1.4846126788788153e-05, "loss": 0.4824, "step": 1352 }, { "epoch": 3.7195876288659795, "grad_norm": 0.1250323838865372, "learning_rate": 1.4786340427402034e-05, "loss": 0.4735, "step": 1353 }, { "epoch": 3.7223367697594503, "grad_norm": 0.11762297628012565, "learning_rate": 1.472664738352266e-05, "loss": 0.4738, "step": 1354 }, { "epoch": 3.725085910652921, "grad_norm": 0.1295998839343961, "learning_rate": 1.4667047878078115e-05, "loss": 0.4797, "step": 1355 }, { "epoch": 3.727835051546392, "grad_norm": 0.12318111875882015, "learning_rate": 1.460754213165036e-05, "loss": 0.4749, "step": 1356 }, { "epoch": 3.7305841924398626, "grad_norm": 0.1372696026783743, "learning_rate": 1.4548130364474253e-05, "loss": 0.4828, "step": 1357 }, { "epoch": 3.7333333333333334, "grad_norm": 0.11717512263035454, "learning_rate": 1.4488812796436871e-05, "loss": 0.4672, "step": 1358 }, { "epoch": 3.736082474226804, "grad_norm": 0.13561308791703292, "learning_rate": 1.4429589647076693e-05, "loss": 0.4786, "step": 1359 }, { "epoch": 3.738831615120275, "grad_norm": 0.12422009286706612, "learning_rate": 1.4370461135582691e-05, "loss": 0.4851, "step": 1360 }, { "epoch": 3.7415807560137457, "grad_norm": 0.11832155403134417, "learning_rate": 1.4311427480793597e-05, "loss": 0.4766, "step": 1361 }, { "epoch": 3.7443298969072165, "grad_norm": 0.12867243461981434, "learning_rate": 1.4252488901197064e-05, "loss": 0.4724, "step": 1362 }, { "epoch": 3.7470790378006873, "grad_norm": 0.1313834080728087, "learning_rate": 1.4193645614928886e-05, "loss": 0.4714, "step": 1363 }, { "epoch": 3.749828178694158, "grad_norm": 0.129183260330846, "learning_rate": 1.4134897839772155e-05, "loss": 0.4809, "step": 1364 }, { "epoch": 3.752577319587629, "grad_norm": 0.12556934848044315, "learning_rate": 1.4076245793156477e-05, "loss": 0.4769, "step": 1365 }, { "epoch": 3.7553264604810996, "grad_norm": 0.12147447097394133, "learning_rate": 1.4017689692157114e-05, "loss": 0.4821, "step": 1366 }, { "epoch": 3.7580756013745704, "grad_norm": 0.13051373867173985, "learning_rate": 1.3959229753494312e-05, "loss": 0.472, "step": 1367 }, { "epoch": 3.760824742268041, "grad_norm": 0.1294424563340561, "learning_rate": 1.3900866193532347e-05, "loss": 0.4822, "step": 1368 }, { "epoch": 3.763573883161512, "grad_norm": 0.1319274091460674, "learning_rate": 1.3842599228278819e-05, "loss": 0.4808, "step": 1369 }, { "epoch": 3.7663230240549828, "grad_norm": 0.14039131963053228, "learning_rate": 1.3784429073383794e-05, "loss": 0.4878, "step": 1370 }, { "epoch": 3.7690721649484535, "grad_norm": 0.15726224881254797, "learning_rate": 1.3726355944139091e-05, "loss": 0.4858, "step": 1371 }, { "epoch": 3.7718213058419243, "grad_norm": 0.1414080124007642, "learning_rate": 1.3668380055477383e-05, "loss": 0.4843, "step": 1372 }, { "epoch": 3.774570446735395, "grad_norm": 0.14834309789500114, "learning_rate": 1.3610501621971466e-05, "loss": 0.4817, "step": 1373 }, { "epoch": 3.777319587628866, "grad_norm": 0.17635910401341698, "learning_rate": 1.3552720857833449e-05, "loss": 0.4791, "step": 1374 }, { "epoch": 3.7800687285223367, "grad_norm": 0.12924512390652784, "learning_rate": 1.3495037976913947e-05, "loss": 0.4786, "step": 1375 }, { "epoch": 3.7828178694158074, "grad_norm": 0.15614756263557175, "learning_rate": 1.3437453192701315e-05, "loss": 0.4805, "step": 1376 }, { "epoch": 3.7855670103092782, "grad_norm": 0.13352957029862794, "learning_rate": 1.3379966718320839e-05, "loss": 0.476, "step": 1377 }, { "epoch": 3.788316151202749, "grad_norm": 0.14136665810120994, "learning_rate": 1.3322578766533968e-05, "loss": 0.473, "step": 1378 }, { "epoch": 3.79106529209622, "grad_norm": 0.1527290850046059, "learning_rate": 1.3265289549737501e-05, "loss": 0.4826, "step": 1379 }, { "epoch": 3.7938144329896906, "grad_norm": 0.13142189843945631, "learning_rate": 1.32080992799628e-05, "loss": 0.4807, "step": 1380 }, { "epoch": 3.7965635738831613, "grad_norm": 0.13227897526677368, "learning_rate": 1.3151008168875037e-05, "loss": 0.47, "step": 1381 }, { "epoch": 3.799312714776632, "grad_norm": 0.1292721449879669, "learning_rate": 1.309401642777238e-05, "loss": 0.4747, "step": 1382 }, { "epoch": 3.802061855670103, "grad_norm": 0.11592891333671732, "learning_rate": 1.3037124267585228e-05, "loss": 0.4795, "step": 1383 }, { "epoch": 3.8048109965635737, "grad_norm": 0.12251003727405267, "learning_rate": 1.298033189887541e-05, "loss": 0.4841, "step": 1384 }, { "epoch": 3.8075601374570445, "grad_norm": 0.11989246216482814, "learning_rate": 1.2923639531835433e-05, "loss": 0.4688, "step": 1385 }, { "epoch": 3.8103092783505152, "grad_norm": 0.12756974223833703, "learning_rate": 1.28670473762877e-05, "loss": 0.4746, "step": 1386 }, { "epoch": 3.813058419243986, "grad_norm": 0.1180466527779192, "learning_rate": 1.2810555641683706e-05, "loss": 0.4779, "step": 1387 }, { "epoch": 3.815807560137457, "grad_norm": 0.11529174071752626, "learning_rate": 1.2754164537103289e-05, "loss": 0.4721, "step": 1388 }, { "epoch": 3.8185567010309276, "grad_norm": 0.11876628327578556, "learning_rate": 1.2697874271253844e-05, "loss": 0.4719, "step": 1389 }, { "epoch": 3.8213058419243984, "grad_norm": 0.1212315987997364, "learning_rate": 1.2641685052469556e-05, "loss": 0.4901, "step": 1390 }, { "epoch": 3.824054982817869, "grad_norm": 0.11790905831570075, "learning_rate": 1.2585597088710637e-05, "loss": 0.4755, "step": 1391 }, { "epoch": 3.82680412371134, "grad_norm": 0.11385644404016117, "learning_rate": 1.252961058756252e-05, "loss": 0.4844, "step": 1392 }, { "epoch": 3.829553264604811, "grad_norm": 0.11530386738039233, "learning_rate": 1.2473725756235165e-05, "loss": 0.4756, "step": 1393 }, { "epoch": 3.832302405498282, "grad_norm": 0.12379904902641005, "learning_rate": 1.2417942801562201e-05, "loss": 0.4702, "step": 1394 }, { "epoch": 3.8350515463917527, "grad_norm": 0.12225777044844523, "learning_rate": 1.236226193000023e-05, "loss": 0.4734, "step": 1395 }, { "epoch": 3.8378006872852235, "grad_norm": 0.12161313208268175, "learning_rate": 1.2306683347628012e-05, "loss": 0.4809, "step": 1396 }, { "epoch": 3.8405498281786943, "grad_norm": 0.13058946315743358, "learning_rate": 1.2251207260145775e-05, "loss": 0.4787, "step": 1397 }, { "epoch": 3.843298969072165, "grad_norm": 0.11523497518057338, "learning_rate": 1.2195833872874343e-05, "loss": 0.4723, "step": 1398 }, { "epoch": 3.846048109965636, "grad_norm": 0.13253286053473112, "learning_rate": 1.2140563390754485e-05, "loss": 0.4765, "step": 1399 }, { "epoch": 3.8487972508591066, "grad_norm": 0.12147651056509255, "learning_rate": 1.208539601834608e-05, "loss": 0.4715, "step": 1400 }, { "epoch": 3.8515463917525774, "grad_norm": 0.12552379756828472, "learning_rate": 1.203033195982744e-05, "loss": 0.4785, "step": 1401 }, { "epoch": 3.854295532646048, "grad_norm": 0.12156559670865116, "learning_rate": 1.1975371418994457e-05, "loss": 0.4865, "step": 1402 }, { "epoch": 3.857044673539519, "grad_norm": 0.13699296116050447, "learning_rate": 1.1920514599259908e-05, "loss": 0.4754, "step": 1403 }, { "epoch": 3.8597938144329897, "grad_norm": 0.14619958749768325, "learning_rate": 1.1865761703652691e-05, "loss": 0.4896, "step": 1404 }, { "epoch": 3.8625429553264605, "grad_norm": 0.11322022044471265, "learning_rate": 1.1811112934817115e-05, "loss": 0.4809, "step": 1405 }, { "epoch": 3.8652920962199313, "grad_norm": 0.12792071165635296, "learning_rate": 1.1756568495012032e-05, "loss": 0.4777, "step": 1406 }, { "epoch": 3.868041237113402, "grad_norm": 0.12840707488542488, "learning_rate": 1.1702128586110204e-05, "loss": 0.4704, "step": 1407 }, { "epoch": 3.870790378006873, "grad_norm": 0.12236846928122713, "learning_rate": 1.164779340959755e-05, "loss": 0.4734, "step": 1408 }, { "epoch": 3.8735395189003436, "grad_norm": 0.12828972122388524, "learning_rate": 1.1593563166572315e-05, "loss": 0.4728, "step": 1409 }, { "epoch": 3.8762886597938144, "grad_norm": 0.11963041349113263, "learning_rate": 1.1539438057744405e-05, "loss": 0.4798, "step": 1410 }, { "epoch": 3.879037800687285, "grad_norm": 0.11889755394099005, "learning_rate": 1.1485418283434591e-05, "loss": 0.4765, "step": 1411 }, { "epoch": 3.881786941580756, "grad_norm": 0.1220924766559957, "learning_rate": 1.1431504043573845e-05, "loss": 0.4792, "step": 1412 }, { "epoch": 3.8845360824742268, "grad_norm": 0.10991608717200368, "learning_rate": 1.1377695537702506e-05, "loss": 0.4758, "step": 1413 }, { "epoch": 3.8872852233676976, "grad_norm": 0.1185758525402704, "learning_rate": 1.1323992964969568e-05, "loss": 0.4764, "step": 1414 }, { "epoch": 3.8900343642611683, "grad_norm": 0.1285483122063387, "learning_rate": 1.1270396524132016e-05, "loss": 0.4726, "step": 1415 }, { "epoch": 3.892783505154639, "grad_norm": 0.11257438797691342, "learning_rate": 1.1216906413553996e-05, "loss": 0.4853, "step": 1416 }, { "epoch": 3.89553264604811, "grad_norm": 0.1265940578758485, "learning_rate": 1.1163522831206119e-05, "loss": 0.4853, "step": 1417 }, { "epoch": 3.8982817869415807, "grad_norm": 0.12742498500479152, "learning_rate": 1.1110245974664741e-05, "loss": 0.4848, "step": 1418 }, { "epoch": 3.9010309278350515, "grad_norm": 0.11699707249742979, "learning_rate": 1.1057076041111197e-05, "loss": 0.4728, "step": 1419 }, { "epoch": 3.9037800687285222, "grad_norm": 0.12027686818075006, "learning_rate": 1.100401322733113e-05, "loss": 0.4697, "step": 1420 }, { "epoch": 3.906529209621993, "grad_norm": 0.12028837431785216, "learning_rate": 1.0951057729713698e-05, "loss": 0.4735, "step": 1421 }, { "epoch": 3.909278350515464, "grad_norm": 0.11366367375993695, "learning_rate": 1.0898209744250852e-05, "loss": 0.4798, "step": 1422 }, { "epoch": 3.9120274914089346, "grad_norm": 0.12271108404053097, "learning_rate": 1.0845469466536689e-05, "loss": 0.4767, "step": 1423 }, { "epoch": 3.9147766323024054, "grad_norm": 0.11169948199394023, "learning_rate": 1.0792837091766635e-05, "loss": 0.4773, "step": 1424 }, { "epoch": 3.917525773195876, "grad_norm": 0.12185308524300838, "learning_rate": 1.0740312814736766e-05, "loss": 0.481, "step": 1425 }, { "epoch": 3.920274914089347, "grad_norm": 0.11498767957292035, "learning_rate": 1.0687896829843068e-05, "loss": 0.4815, "step": 1426 }, { "epoch": 3.923024054982818, "grad_norm": 0.10760167455092437, "learning_rate": 1.0635589331080771e-05, "loss": 0.4716, "step": 1427 }, { "epoch": 3.925773195876289, "grad_norm": 0.11652282613458785, "learning_rate": 1.0583390512043557e-05, "loss": 0.4716, "step": 1428 }, { "epoch": 3.9285223367697597, "grad_norm": 0.11095078761860352, "learning_rate": 1.0531300565922873e-05, "loss": 0.4773, "step": 1429 }, { "epoch": 3.9312714776632305, "grad_norm": 0.10567263314079245, "learning_rate": 1.0479319685507234e-05, "loss": 0.4818, "step": 1430 }, { "epoch": 3.9340206185567013, "grad_norm": 0.11290338751098791, "learning_rate": 1.04274480631815e-05, "loss": 0.4736, "step": 1431 }, { "epoch": 3.936769759450172, "grad_norm": 0.11142975595031625, "learning_rate": 1.0375685890926142e-05, "loss": 0.4759, "step": 1432 }, { "epoch": 3.939518900343643, "grad_norm": 0.12580032061681007, "learning_rate": 1.0324033360316559e-05, "loss": 0.4787, "step": 1433 }, { "epoch": 3.9422680412371136, "grad_norm": 0.10935541303336552, "learning_rate": 1.0272490662522347e-05, "loss": 0.4771, "step": 1434 }, { "epoch": 3.9450171821305844, "grad_norm": 0.11261196339303021, "learning_rate": 1.0221057988306637e-05, "loss": 0.4711, "step": 1435 }, { "epoch": 3.947766323024055, "grad_norm": 0.11321244000910012, "learning_rate": 1.0169735528025319e-05, "loss": 0.4851, "step": 1436 }, { "epoch": 3.950515463917526, "grad_norm": 0.11089992172259956, "learning_rate": 1.011852347162639e-05, "loss": 0.4751, "step": 1437 }, { "epoch": 3.9532646048109967, "grad_norm": 0.11766021872483373, "learning_rate": 1.0067422008649221e-05, "loss": 0.4797, "step": 1438 }, { "epoch": 3.9560137457044675, "grad_norm": 0.10628228690278965, "learning_rate": 1.0016431328223887e-05, "loss": 0.4782, "step": 1439 }, { "epoch": 3.9587628865979383, "grad_norm": 0.11378648141949942, "learning_rate": 9.96555161907044e-06, "loss": 0.4766, "step": 1440 }, { "epoch": 3.961512027491409, "grad_norm": 0.11864702105913649, "learning_rate": 9.91478306949821e-06, "loss": 0.4786, "step": 1441 }, { "epoch": 3.96426116838488, "grad_norm": 0.12126170054240307, "learning_rate": 9.864125867405146e-06, "loss": 0.4797, "step": 1442 }, { "epoch": 3.9670103092783506, "grad_norm": 0.11758636840862939, "learning_rate": 9.813580200277063e-06, "loss": 0.4768, "step": 1443 }, { "epoch": 3.9697594501718214, "grad_norm": 0.11275441649847655, "learning_rate": 9.76314625518699e-06, "loss": 0.4668, "step": 1444 }, { "epoch": 3.972508591065292, "grad_norm": 0.1131262643697182, "learning_rate": 9.712824218794457e-06, "loss": 0.4733, "step": 1445 }, { "epoch": 3.975257731958763, "grad_norm": 0.10279589610315747, "learning_rate": 9.662614277344846e-06, "loss": 0.479, "step": 1446 }, { "epoch": 3.9780068728522338, "grad_norm": 0.11106926427912463, "learning_rate": 9.612516616668612e-06, "loss": 0.4822, "step": 1447 }, { "epoch": 3.9807560137457045, "grad_norm": 0.10564345176951702, "learning_rate": 9.562531422180674e-06, "loss": 0.4745, "step": 1448 }, { "epoch": 3.9835051546391753, "grad_norm": 0.10829788647274623, "learning_rate": 9.512658878879737e-06, "loss": 0.4846, "step": 1449 }, { "epoch": 3.986254295532646, "grad_norm": 0.11018495327771764, "learning_rate": 9.462899171347538e-06, "loss": 0.4802, "step": 1450 }, { "epoch": 3.989003436426117, "grad_norm": 0.12608551718647237, "learning_rate": 9.413252483748216e-06, "loss": 0.4766, "step": 1451 }, { "epoch": 3.9917525773195877, "grad_norm": 0.11337813442101644, "learning_rate": 9.363718999827612e-06, "loss": 0.479, "step": 1452 }, { "epoch": 3.9945017182130584, "grad_norm": 0.12372671414303754, "learning_rate": 9.314298902912583e-06, "loss": 0.479, "step": 1453 }, { "epoch": 3.997250859106529, "grad_norm": 0.11011001077642653, "learning_rate": 9.26499237591037e-06, "loss": 0.4695, "step": 1454 }, { "epoch": 4.0, "grad_norm": 0.14840181176358444, "learning_rate": 9.215799601307828e-06, "loss": 0.6096, "step": 1455 }, { "epoch": 4.002749140893471, "grad_norm": 0.19421057731547003, "learning_rate": 9.16672076117084e-06, "loss": 0.4433, "step": 1456 }, { "epoch": 4.005498281786942, "grad_norm": 0.15015912458886393, "learning_rate": 9.117756037143622e-06, "loss": 0.4425, "step": 1457 }, { "epoch": 4.008247422680412, "grad_norm": 0.1383285525044493, "learning_rate": 9.068905610448011e-06, "loss": 0.4562, "step": 1458 }, { "epoch": 4.010996563573883, "grad_norm": 0.15430721574490833, "learning_rate": 9.02016966188283e-06, "loss": 0.4484, "step": 1459 }, { "epoch": 4.013745704467354, "grad_norm": 0.14905855161830828, "learning_rate": 8.971548371823205e-06, "loss": 0.446, "step": 1460 }, { "epoch": 4.016494845360825, "grad_norm": 0.15470678416362293, "learning_rate": 8.923041920219927e-06, "loss": 0.4412, "step": 1461 }, { "epoch": 4.0192439862542955, "grad_norm": 0.14564788987737176, "learning_rate": 8.87465048659875e-06, "loss": 0.4393, "step": 1462 }, { "epoch": 4.021993127147766, "grad_norm": 0.13342656861373697, "learning_rate": 8.826374250059704e-06, "loss": 0.4498, "step": 1463 }, { "epoch": 4.024742268041237, "grad_norm": 0.1302414374832906, "learning_rate": 8.778213389276523e-06, "loss": 0.4383, "step": 1464 }, { "epoch": 4.027491408934708, "grad_norm": 0.14007280746183526, "learning_rate": 8.730168082495885e-06, "loss": 0.4474, "step": 1465 }, { "epoch": 4.030240549828179, "grad_norm": 0.13171480431395038, "learning_rate": 8.682238507536823e-06, "loss": 0.4388, "step": 1466 }, { "epoch": 4.032989690721649, "grad_norm": 0.12287866481381172, "learning_rate": 8.634424841790014e-06, "loss": 0.4384, "step": 1467 }, { "epoch": 4.03573883161512, "grad_norm": 0.12293190753633977, "learning_rate": 8.586727262217156e-06, "loss": 0.4444, "step": 1468 }, { "epoch": 4.038487972508591, "grad_norm": 0.12156568952497382, "learning_rate": 8.539145945350324e-06, "loss": 0.4393, "step": 1469 }, { "epoch": 4.041237113402062, "grad_norm": 0.13520454469985085, "learning_rate": 8.491681067291279e-06, "loss": 0.4438, "step": 1470 }, { "epoch": 4.0439862542955325, "grad_norm": 0.12818857279590956, "learning_rate": 8.444332803710806e-06, "loss": 0.4443, "step": 1471 }, { "epoch": 4.046735395189003, "grad_norm": 0.11732641042539134, "learning_rate": 8.397101329848146e-06, "loss": 0.4443, "step": 1472 }, { "epoch": 4.049484536082474, "grad_norm": 0.12736631288786282, "learning_rate": 8.349986820510257e-06, "loss": 0.4427, "step": 1473 }, { "epoch": 4.052233676975945, "grad_norm": 0.1304818919917951, "learning_rate": 8.302989450071219e-06, "loss": 0.4478, "step": 1474 }, { "epoch": 4.054982817869416, "grad_norm": 0.13641319860640636, "learning_rate": 8.256109392471549e-06, "loss": 0.4595, "step": 1475 }, { "epoch": 4.057731958762886, "grad_norm": 0.11825978977034925, "learning_rate": 8.20934682121763e-06, "loss": 0.4461, "step": 1476 }, { "epoch": 4.060481099656357, "grad_norm": 0.13030109067712098, "learning_rate": 8.162701909380967e-06, "loss": 0.4473, "step": 1477 }, { "epoch": 4.063230240549828, "grad_norm": 0.12977389788164925, "learning_rate": 8.116174829597625e-06, "loss": 0.4506, "step": 1478 }, { "epoch": 4.065979381443299, "grad_norm": 0.12148360180077224, "learning_rate": 8.069765754067562e-06, "loss": 0.4454, "step": 1479 }, { "epoch": 4.0687285223367695, "grad_norm": 0.13076375704534893, "learning_rate": 8.023474854553996e-06, "loss": 0.4459, "step": 1480 }, { "epoch": 4.07147766323024, "grad_norm": 0.13403343398060516, "learning_rate": 7.977302302382757e-06, "loss": 0.4392, "step": 1481 }, { "epoch": 4.074226804123711, "grad_norm": 0.11166777845488875, "learning_rate": 7.93124826844167e-06, "loss": 0.4411, "step": 1482 }, { "epoch": 4.076975945017182, "grad_norm": 0.13721899419089614, "learning_rate": 7.885312923179912e-06, "loss": 0.444, "step": 1483 }, { "epoch": 4.079725085910653, "grad_norm": 0.13263433098857536, "learning_rate": 7.839496436607406e-06, "loss": 0.4459, "step": 1484 }, { "epoch": 4.082474226804123, "grad_norm": 0.11590942502418243, "learning_rate": 7.793798978294144e-06, "loss": 0.4453, "step": 1485 }, { "epoch": 4.085223367697594, "grad_norm": 0.13293144505263688, "learning_rate": 7.748220717369594e-06, "loss": 0.4413, "step": 1486 }, { "epoch": 4.087972508591065, "grad_norm": 0.13483699801148805, "learning_rate": 7.702761822522072e-06, "loss": 0.4366, "step": 1487 }, { "epoch": 4.090721649484536, "grad_norm": 0.11307879927149218, "learning_rate": 7.6574224619981e-06, "loss": 0.4352, "step": 1488 }, { "epoch": 4.0934707903780065, "grad_norm": 0.11327258461306543, "learning_rate": 7.612202803601812e-06, "loss": 0.4513, "step": 1489 }, { "epoch": 4.096219931271477, "grad_norm": 0.11998728903473971, "learning_rate": 7.56710301469429e-06, "loss": 0.4462, "step": 1490 }, { "epoch": 4.098969072164948, "grad_norm": 0.11148518956657245, "learning_rate": 7.522123262193011e-06, "loss": 0.4414, "step": 1491 }, { "epoch": 4.101718213058419, "grad_norm": 0.10980782127466164, "learning_rate": 7.477263712571154e-06, "loss": 0.4505, "step": 1492 }, { "epoch": 4.10446735395189, "grad_norm": 0.13007209291051053, "learning_rate": 7.432524531857037e-06, "loss": 0.4404, "step": 1493 }, { "epoch": 4.10721649484536, "grad_norm": 0.10633901487852752, "learning_rate": 7.38790588563346e-06, "loss": 0.4432, "step": 1494 }, { "epoch": 4.109965635738831, "grad_norm": 0.10337495591575992, "learning_rate": 7.343407939037166e-06, "loss": 0.4442, "step": 1495 }, { "epoch": 4.112714776632302, "grad_norm": 0.11699522345473407, "learning_rate": 7.29903085675812e-06, "loss": 0.4404, "step": 1496 }, { "epoch": 4.115463917525773, "grad_norm": 0.1071569023006855, "learning_rate": 7.254774803038986e-06, "loss": 0.4448, "step": 1497 }, { "epoch": 4.118213058419244, "grad_norm": 0.10590236218912877, "learning_rate": 7.210639941674515e-06, "loss": 0.4429, "step": 1498 }, { "epoch": 4.120962199312714, "grad_norm": 0.10210210146806667, "learning_rate": 7.166626436010893e-06, "loss": 0.4415, "step": 1499 }, { "epoch": 4.123711340206185, "grad_norm": 0.11125955970384037, "learning_rate": 7.122734448945157e-06, "loss": 0.4526, "step": 1500 }, { "epoch": 4.126460481099656, "grad_norm": 0.12013542728499116, "learning_rate": 7.078964142924607e-06, "loss": 0.4419, "step": 1501 }, { "epoch": 4.129209621993128, "grad_norm": 0.10484380190061886, "learning_rate": 7.035315679946176e-06, "loss": 0.4489, "step": 1502 }, { "epoch": 4.131958762886598, "grad_norm": 0.09567503999966823, "learning_rate": 6.991789221555887e-06, "loss": 0.4372, "step": 1503 }, { "epoch": 4.134707903780069, "grad_norm": 0.10540468880005585, "learning_rate": 6.94838492884816e-06, "loss": 0.4464, "step": 1504 }, { "epoch": 4.13745704467354, "grad_norm": 0.11225166249471208, "learning_rate": 6.905102962465302e-06, "loss": 0.4439, "step": 1505 }, { "epoch": 4.140206185567011, "grad_norm": 0.10759681388030853, "learning_rate": 6.861943482596896e-06, "loss": 0.4436, "step": 1506 }, { "epoch": 4.1429553264604815, "grad_norm": 0.10130359403512029, "learning_rate": 6.818906648979169e-06, "loss": 0.4388, "step": 1507 }, { "epoch": 4.145704467353952, "grad_norm": 0.1121174956770293, "learning_rate": 6.775992620894434e-06, "loss": 0.4507, "step": 1508 }, { "epoch": 4.148453608247423, "grad_norm": 0.1014146829605852, "learning_rate": 6.733201557170481e-06, "loss": 0.4432, "step": 1509 }, { "epoch": 4.151202749140894, "grad_norm": 0.09837158537375995, "learning_rate": 6.690533616180031e-06, "loss": 0.4416, "step": 1510 }, { "epoch": 4.153951890034365, "grad_norm": 0.10857273055540849, "learning_rate": 6.647988955840099e-06, "loss": 0.4283, "step": 1511 }, { "epoch": 4.156701030927835, "grad_norm": 0.11827148688247635, "learning_rate": 6.6055677336114024e-06, "loss": 0.4435, "step": 1512 }, { "epoch": 4.159450171821306, "grad_norm": 0.10661184402237826, "learning_rate": 6.563270106497866e-06, "loss": 0.4493, "step": 1513 }, { "epoch": 4.162199312714777, "grad_norm": 0.10027884158801532, "learning_rate": 6.521096231045927e-06, "loss": 0.4424, "step": 1514 }, { "epoch": 4.164948453608248, "grad_norm": 0.10677714613702075, "learning_rate": 6.479046263344044e-06, "loss": 0.4407, "step": 1515 }, { "epoch": 4.1676975945017185, "grad_norm": 0.10120025403165747, "learning_rate": 6.43712035902206e-06, "loss": 0.4482, "step": 1516 }, { "epoch": 4.170446735395189, "grad_norm": 0.09576847158285014, "learning_rate": 6.395318673250655e-06, "loss": 0.4455, "step": 1517 }, { "epoch": 4.17319587628866, "grad_norm": 0.10456502041919041, "learning_rate": 6.353641360740796e-06, "loss": 0.4442, "step": 1518 }, { "epoch": 4.175945017182131, "grad_norm": 0.10975215070737157, "learning_rate": 6.312088575743112e-06, "loss": 0.4421, "step": 1519 }, { "epoch": 4.178694158075602, "grad_norm": 0.09670376981272215, "learning_rate": 6.270660472047318e-06, "loss": 0.4357, "step": 1520 }, { "epoch": 4.181443298969072, "grad_norm": 0.0989487056180283, "learning_rate": 6.229357202981736e-06, "loss": 0.4446, "step": 1521 }, { "epoch": 4.184192439862543, "grad_norm": 0.11575820428367671, "learning_rate": 6.188178921412622e-06, "loss": 0.4376, "step": 1522 }, { "epoch": 4.186941580756014, "grad_norm": 0.09506731515085455, "learning_rate": 6.147125779743666e-06, "loss": 0.4399, "step": 1523 }, { "epoch": 4.189690721649485, "grad_norm": 0.10319725228868595, "learning_rate": 6.106197929915385e-06, "loss": 0.4494, "step": 1524 }, { "epoch": 4.1924398625429555, "grad_norm": 0.10930094347383024, "learning_rate": 6.065395523404616e-06, "loss": 0.4416, "step": 1525 }, { "epoch": 4.195189003436426, "grad_norm": 0.10333832350838476, "learning_rate": 6.0247187112238936e-06, "loss": 0.4419, "step": 1526 }, { "epoch": 4.197938144329897, "grad_norm": 0.09863833187141094, "learning_rate": 5.984167643920926e-06, "loss": 0.45, "step": 1527 }, { "epoch": 4.200687285223368, "grad_norm": 0.10338899938814769, "learning_rate": 5.943742471578029e-06, "loss": 0.4432, "step": 1528 }, { "epoch": 4.203436426116839, "grad_norm": 0.10080408781457796, "learning_rate": 5.903443343811583e-06, "loss": 0.4498, "step": 1529 }, { "epoch": 4.206185567010309, "grad_norm": 0.0995238925477759, "learning_rate": 5.863270409771451e-06, "loss": 0.4441, "step": 1530 }, { "epoch": 4.20893470790378, "grad_norm": 0.09871347524798951, "learning_rate": 5.823223818140458e-06, "loss": 0.4465, "step": 1531 }, { "epoch": 4.211683848797251, "grad_norm": 0.09741808192782717, "learning_rate": 5.78330371713383e-06, "loss": 0.4521, "step": 1532 }, { "epoch": 4.214432989690722, "grad_norm": 0.10108809287384478, "learning_rate": 5.7435102544986325e-06, "loss": 0.4476, "step": 1533 }, { "epoch": 4.217182130584193, "grad_norm": 0.09586236341789252, "learning_rate": 5.70384357751323e-06, "loss": 0.4417, "step": 1534 }, { "epoch": 4.219931271477663, "grad_norm": 0.0982846297983466, "learning_rate": 5.664303832986764e-06, "loss": 0.4496, "step": 1535 }, { "epoch": 4.222680412371134, "grad_norm": 0.09404435584314635, "learning_rate": 5.624891167258569e-06, "loss": 0.4415, "step": 1536 }, { "epoch": 4.225429553264605, "grad_norm": 0.09792884175585087, "learning_rate": 5.585605726197663e-06, "loss": 0.4364, "step": 1537 }, { "epoch": 4.228178694158076, "grad_norm": 0.0966404647926693, "learning_rate": 5.54644765520219e-06, "loss": 0.4464, "step": 1538 }, { "epoch": 4.2309278350515465, "grad_norm": 0.09803372963199505, "learning_rate": 5.507417099198886e-06, "loss": 0.443, "step": 1539 }, { "epoch": 4.233676975945017, "grad_norm": 0.10220414170891072, "learning_rate": 5.468514202642574e-06, "loss": 0.4416, "step": 1540 }, { "epoch": 4.236426116838488, "grad_norm": 0.09940583554566296, "learning_rate": 5.429739109515564e-06, "loss": 0.4483, "step": 1541 }, { "epoch": 4.239175257731959, "grad_norm": 0.09639208839569098, "learning_rate": 5.3910919633271755e-06, "loss": 0.442, "step": 1542 }, { "epoch": 4.24192439862543, "grad_norm": 0.1117369320333374, "learning_rate": 5.352572907113178e-06, "loss": 0.4461, "step": 1543 }, { "epoch": 4.2446735395189, "grad_norm": 0.09611791874864287, "learning_rate": 5.314182083435282e-06, "loss": 0.4346, "step": 1544 }, { "epoch": 4.247422680412371, "grad_norm": 0.10794580998258632, "learning_rate": 5.2759196343805885e-06, "loss": 0.4415, "step": 1545 }, { "epoch": 4.250171821305842, "grad_norm": 0.0952864724999734, "learning_rate": 5.237785701561078e-06, "loss": 0.4443, "step": 1546 }, { "epoch": 4.252920962199313, "grad_norm": 0.10156924345931388, "learning_rate": 5.1997804261131015e-06, "loss": 0.449, "step": 1547 }, { "epoch": 4.2556701030927835, "grad_norm": 0.09725478784660825, "learning_rate": 5.161903948696813e-06, "loss": 0.439, "step": 1548 }, { "epoch": 4.258419243986254, "grad_norm": 0.10252085604277364, "learning_rate": 5.124156409495693e-06, "loss": 0.4385, "step": 1549 }, { "epoch": 4.261168384879725, "grad_norm": 0.10131443009760514, "learning_rate": 5.086537948216008e-06, "loss": 0.4407, "step": 1550 }, { "epoch": 4.263917525773196, "grad_norm": 0.09875784055342578, "learning_rate": 5.049048704086295e-06, "loss": 0.4469, "step": 1551 }, { "epoch": 4.266666666666667, "grad_norm": 0.10157130046545569, "learning_rate": 5.011688815856856e-06, "loss": 0.4502, "step": 1552 }, { "epoch": 4.269415807560137, "grad_norm": 0.10941012381717943, "learning_rate": 4.974458421799231e-06, "loss": 0.4411, "step": 1553 }, { "epoch": 4.272164948453608, "grad_norm": 0.11064889415177875, "learning_rate": 4.937357659705688e-06, "loss": 0.4419, "step": 1554 }, { "epoch": 4.274914089347079, "grad_norm": 0.10002641057536234, "learning_rate": 4.90038666688875e-06, "loss": 0.4558, "step": 1555 }, { "epoch": 4.27766323024055, "grad_norm": 0.10016807122767976, "learning_rate": 4.863545580180615e-06, "loss": 0.4435, "step": 1556 }, { "epoch": 4.2804123711340205, "grad_norm": 0.10851136506221658, "learning_rate": 4.8268345359327075e-06, "loss": 0.4398, "step": 1557 }, { "epoch": 4.283161512027491, "grad_norm": 0.09488321817571037, "learning_rate": 4.7902536700151504e-06, "loss": 0.4429, "step": 1558 }, { "epoch": 4.285910652920962, "grad_norm": 0.09962394867509372, "learning_rate": 4.75380311781628e-06, "loss": 0.4435, "step": 1559 }, { "epoch": 4.288659793814433, "grad_norm": 0.10525730743423926, "learning_rate": 4.717483014242134e-06, "loss": 0.4488, "step": 1560 }, { "epoch": 4.291408934707904, "grad_norm": 0.09444016052953329, "learning_rate": 4.681293493715906e-06, "loss": 0.4383, "step": 1561 }, { "epoch": 4.294158075601374, "grad_norm": 0.0935555125757194, "learning_rate": 4.645234690177556e-06, "loss": 0.441, "step": 1562 }, { "epoch": 4.296907216494845, "grad_norm": 0.09875362647621048, "learning_rate": 4.6093067370832145e-06, "loss": 0.4467, "step": 1563 }, { "epoch": 4.299656357388316, "grad_norm": 0.10106687338028049, "learning_rate": 4.573509767404733e-06, "loss": 0.4398, "step": 1564 }, { "epoch": 4.302405498281787, "grad_norm": 0.09430960821120914, "learning_rate": 4.537843913629178e-06, "loss": 0.4498, "step": 1565 }, { "epoch": 4.3051546391752575, "grad_norm": 0.09341919086027475, "learning_rate": 4.502309307758368e-06, "loss": 0.4427, "step": 1566 }, { "epoch": 4.307903780068728, "grad_norm": 0.09579646490730007, "learning_rate": 4.46690608130834e-06, "loss": 0.4504, "step": 1567 }, { "epoch": 4.310652920962199, "grad_norm": 0.10140215061324537, "learning_rate": 4.431634365308904e-06, "loss": 0.4445, "step": 1568 }, { "epoch": 4.31340206185567, "grad_norm": 0.09687704144452364, "learning_rate": 4.39649429030311e-06, "loss": 0.4478, "step": 1569 }, { "epoch": 4.316151202749141, "grad_norm": 0.09438778342768989, "learning_rate": 4.3614859863468425e-06, "loss": 0.4386, "step": 1570 }, { "epoch": 4.318900343642611, "grad_norm": 0.09156131739398227, "learning_rate": 4.326609583008261e-06, "loss": 0.4338, "step": 1571 }, { "epoch": 4.321649484536082, "grad_norm": 0.09078660520618338, "learning_rate": 4.2918652093673606e-06, "loss": 0.442, "step": 1572 }, { "epoch": 4.324398625429553, "grad_norm": 0.10260056112083875, "learning_rate": 4.257252994015466e-06, "loss": 0.4441, "step": 1573 }, { "epoch": 4.327147766323024, "grad_norm": 0.10384708279251556, "learning_rate": 4.222773065054817e-06, "loss": 0.4466, "step": 1574 }, { "epoch": 4.329896907216495, "grad_norm": 0.09444726168415894, "learning_rate": 4.188425550098019e-06, "loss": 0.444, "step": 1575 }, { "epoch": 4.332646048109965, "grad_norm": 0.1139460077036934, "learning_rate": 4.15421057626761e-06, "loss": 0.4446, "step": 1576 }, { "epoch": 4.335395189003436, "grad_norm": 0.09822740636361162, "learning_rate": 4.120128270195585e-06, "loss": 0.4424, "step": 1577 }, { "epoch": 4.338144329896907, "grad_norm": 0.09913730861031846, "learning_rate": 4.086178758022929e-06, "loss": 0.4477, "step": 1578 }, { "epoch": 4.340893470790378, "grad_norm": 0.09892692023825955, "learning_rate": 4.052362165399148e-06, "loss": 0.4462, "step": 1579 }, { "epoch": 4.3436426116838485, "grad_norm": 0.09559985988178789, "learning_rate": 4.018678617481797e-06, "loss": 0.4518, "step": 1580 }, { "epoch": 4.346391752577319, "grad_norm": 0.09645896728376689, "learning_rate": 3.9851282389360336e-06, "loss": 0.4411, "step": 1581 }, { "epoch": 4.34914089347079, "grad_norm": 0.09621093774200687, "learning_rate": 3.951711153934143e-06, "loss": 0.4455, "step": 1582 }, { "epoch": 4.351890034364261, "grad_norm": 0.09860601420290797, "learning_rate": 3.918427486155079e-06, "loss": 0.4452, "step": 1583 }, { "epoch": 4.354639175257732, "grad_norm": 0.09460049776906895, "learning_rate": 3.885277358784003e-06, "loss": 0.4476, "step": 1584 }, { "epoch": 4.357388316151202, "grad_norm": 0.0970914145620547, "learning_rate": 3.85226089451185e-06, "loss": 0.4486, "step": 1585 }, { "epoch": 4.360137457044673, "grad_norm": 0.10970056178061441, "learning_rate": 3.819378215534842e-06, "loss": 0.4404, "step": 1586 }, { "epoch": 4.362886597938144, "grad_norm": 0.09671374353988552, "learning_rate": 3.7866294435540574e-06, "loss": 0.4534, "step": 1587 }, { "epoch": 4.365635738831615, "grad_norm": 0.09276644771447388, "learning_rate": 3.7540146997749793e-06, "loss": 0.4489, "step": 1588 }, { "epoch": 4.368384879725086, "grad_norm": 0.09400993795205304, "learning_rate": 3.72153410490705e-06, "loss": 0.4467, "step": 1589 }, { "epoch": 4.371134020618557, "grad_norm": 0.10312570233843393, "learning_rate": 3.689187779163206e-06, "loss": 0.4491, "step": 1590 }, { "epoch": 4.373883161512028, "grad_norm": 0.10124141768595926, "learning_rate": 3.6569758422594446e-06, "loss": 0.4528, "step": 1591 }, { "epoch": 4.376632302405499, "grad_norm": 0.0947775184692594, "learning_rate": 3.6248984134143794e-06, "loss": 0.4493, "step": 1592 }, { "epoch": 4.3793814432989695, "grad_norm": 0.0949068368311883, "learning_rate": 3.5929556113488117e-06, "loss": 0.4502, "step": 1593 }, { "epoch": 4.38213058419244, "grad_norm": 0.09225617183677981, "learning_rate": 3.5611475542852714e-06, "loss": 0.4378, "step": 1594 }, { "epoch": 4.384879725085911, "grad_norm": 0.09712650449251893, "learning_rate": 3.5294743599475754e-06, "loss": 0.446, "step": 1595 }, { "epoch": 4.387628865979382, "grad_norm": 0.10510272619911021, "learning_rate": 3.4979361455604386e-06, "loss": 0.4438, "step": 1596 }, { "epoch": 4.390378006872853, "grad_norm": 0.10034299553088664, "learning_rate": 3.4665330278489708e-06, "loss": 0.4487, "step": 1597 }, { "epoch": 4.393127147766323, "grad_norm": 0.09360138161898929, "learning_rate": 3.4352651230383026e-06, "loss": 0.4409, "step": 1598 }, { "epoch": 4.395876288659794, "grad_norm": 0.09887187647716278, "learning_rate": 3.4041325468531094e-06, "loss": 0.4409, "step": 1599 }, { "epoch": 4.398625429553265, "grad_norm": 0.10473071373698492, "learning_rate": 3.3731354145172435e-06, "loss": 0.4516, "step": 1600 }, { "epoch": 4.401374570446736, "grad_norm": 0.1007375745415408, "learning_rate": 3.342273840753221e-06, "loss": 0.4383, "step": 1601 }, { "epoch": 4.4041237113402065, "grad_norm": 0.09735792228082213, "learning_rate": 3.311547939781887e-06, "loss": 0.4409, "step": 1602 }, { "epoch": 4.406872852233677, "grad_norm": 0.09459778267684918, "learning_rate": 3.280957825321922e-06, "loss": 0.4442, "step": 1603 }, { "epoch": 4.409621993127148, "grad_norm": 0.09281792628016262, "learning_rate": 3.250503610589482e-06, "loss": 0.4491, "step": 1604 }, { "epoch": 4.412371134020619, "grad_norm": 0.1007767987975053, "learning_rate": 3.2201854082977292e-06, "loss": 0.4517, "step": 1605 }, { "epoch": 4.41512027491409, "grad_norm": 0.09789821294262532, "learning_rate": 3.190003330656435e-06, "loss": 0.4451, "step": 1606 }, { "epoch": 4.41786941580756, "grad_norm": 0.10028034923814605, "learning_rate": 3.159957489371559e-06, "loss": 0.4367, "step": 1607 }, { "epoch": 4.420618556701031, "grad_norm": 0.09081132192790976, "learning_rate": 3.1300479956448693e-06, "loss": 0.4387, "step": 1608 }, { "epoch": 4.423367697594502, "grad_norm": 0.09711016308139389, "learning_rate": 3.1002749601734618e-06, "loss": 0.4428, "step": 1609 }, { "epoch": 4.426116838487973, "grad_norm": 0.09435493012986142, "learning_rate": 3.0706384931494137e-06, "loss": 0.4328, "step": 1610 }, { "epoch": 4.4288659793814436, "grad_norm": 0.09951785033015637, "learning_rate": 3.0411387042593545e-06, "loss": 0.4512, "step": 1611 }, { "epoch": 4.431615120274914, "grad_norm": 0.0960124562000595, "learning_rate": 3.0117757026840543e-06, "loss": 0.4485, "step": 1612 }, { "epoch": 4.434364261168385, "grad_norm": 0.09374569219824876, "learning_rate": 2.9825495970980234e-06, "loss": 0.4448, "step": 1613 }, { "epoch": 4.437113402061856, "grad_norm": 0.09287833110723459, "learning_rate": 2.953460495669096e-06, "loss": 0.4473, "step": 1614 }, { "epoch": 4.439862542955327, "grad_norm": 0.09330518761309428, "learning_rate": 2.924508506058077e-06, "loss": 0.4368, "step": 1615 }, { "epoch": 4.4426116838487975, "grad_norm": 0.09851414675120082, "learning_rate": 2.8956937354182923e-06, "loss": 0.4438, "step": 1616 }, { "epoch": 4.445360824742268, "grad_norm": 0.09503089697922397, "learning_rate": 2.867016290395199e-06, "loss": 0.4557, "step": 1617 }, { "epoch": 4.448109965635739, "grad_norm": 0.09322925477662775, "learning_rate": 2.838476277126012e-06, "loss": 0.4428, "step": 1618 }, { "epoch": 4.45085910652921, "grad_norm": 0.09242195549495436, "learning_rate": 2.810073801239317e-06, "loss": 0.446, "step": 1619 }, { "epoch": 4.453608247422681, "grad_norm": 0.08945993365343624, "learning_rate": 2.781808967854649e-06, "loss": 0.4412, "step": 1620 }, { "epoch": 4.456357388316151, "grad_norm": 0.09428643807864204, "learning_rate": 2.753681881582115e-06, "loss": 0.4416, "step": 1621 }, { "epoch": 4.459106529209622, "grad_norm": 0.0931736838114746, "learning_rate": 2.7256926465220177e-06, "loss": 0.4472, "step": 1622 }, { "epoch": 4.461855670103093, "grad_norm": 0.09181260953087753, "learning_rate": 2.697841366264471e-06, "loss": 0.4481, "step": 1623 }, { "epoch": 4.464604810996564, "grad_norm": 0.09223991013911755, "learning_rate": 2.670128143888988e-06, "loss": 0.4417, "step": 1624 }, { "epoch": 4.4673539518900345, "grad_norm": 0.09242485950713865, "learning_rate": 2.6425530819641364e-06, "loss": 0.4501, "step": 1625 }, { "epoch": 4.470103092783505, "grad_norm": 0.09113105402491876, "learning_rate": 2.6151162825471364e-06, "loss": 0.4442, "step": 1626 }, { "epoch": 4.472852233676976, "grad_norm": 0.09212808247265639, "learning_rate": 2.587817847183489e-06, "loss": 0.4429, "step": 1627 }, { "epoch": 4.475601374570447, "grad_norm": 0.09767261371119439, "learning_rate": 2.5606578769066026e-06, "loss": 0.4432, "step": 1628 }, { "epoch": 4.478350515463918, "grad_norm": 0.09743572389157769, "learning_rate": 2.5336364722374106e-06, "loss": 0.4487, "step": 1629 }, { "epoch": 4.481099656357388, "grad_norm": 0.09102922314529087, "learning_rate": 2.506753733184013e-06, "loss": 0.4503, "step": 1630 }, { "epoch": 4.483848797250859, "grad_norm": 0.09266255759834628, "learning_rate": 2.480009759241302e-06, "loss": 0.4405, "step": 1631 }, { "epoch": 4.48659793814433, "grad_norm": 0.0925039627946117, "learning_rate": 2.453404649390576e-06, "loss": 0.4398, "step": 1632 }, { "epoch": 4.489347079037801, "grad_norm": 0.09521064172876387, "learning_rate": 2.4269385020991987e-06, "loss": 0.4486, "step": 1633 }, { "epoch": 4.4920962199312715, "grad_norm": 0.09113818118580695, "learning_rate": 2.40061141532022e-06, "loss": 0.4431, "step": 1634 }, { "epoch": 4.494845360824742, "grad_norm": 0.10000887227672747, "learning_rate": 2.374423486492021e-06, "loss": 0.4388, "step": 1635 }, { "epoch": 4.497594501718213, "grad_norm": 0.09939158503866538, "learning_rate": 2.3483748125379434e-06, "loss": 0.4526, "step": 1636 }, { "epoch": 4.500343642611684, "grad_norm": 0.08865852194600819, "learning_rate": 2.3224654898659348e-06, "loss": 0.4458, "step": 1637 }, { "epoch": 4.503092783505155, "grad_norm": 0.09118943093226906, "learning_rate": 2.2966956143682094e-06, "loss": 0.444, "step": 1638 }, { "epoch": 4.505841924398625, "grad_norm": 0.09085329322674768, "learning_rate": 2.2710652814208656e-06, "loss": 0.4516, "step": 1639 }, { "epoch": 4.508591065292096, "grad_norm": 0.09474948162033754, "learning_rate": 2.2455745858835474e-06, "loss": 0.4439, "step": 1640 }, { "epoch": 4.511340206185567, "grad_norm": 0.09100873289075298, "learning_rate": 2.2202236220990913e-06, "loss": 0.4441, "step": 1641 }, { "epoch": 4.514089347079038, "grad_norm": 0.09556416343413056, "learning_rate": 2.1950124838931685e-06, "loss": 0.4426, "step": 1642 }, { "epoch": 4.5168384879725085, "grad_norm": 0.09124176627878805, "learning_rate": 2.169941264573967e-06, "loss": 0.4515, "step": 1643 }, { "epoch": 4.519587628865979, "grad_norm": 0.08967523905753207, "learning_rate": 2.1450100569317954e-06, "loss": 0.4427, "step": 1644 }, { "epoch": 4.52233676975945, "grad_norm": 0.08908910909605862, "learning_rate": 2.1202189532387994e-06, "loss": 0.4421, "step": 1645 }, { "epoch": 4.525085910652921, "grad_norm": 0.08930544495492708, "learning_rate": 2.0955680452485795e-06, "loss": 0.4386, "step": 1646 }, { "epoch": 4.527835051546392, "grad_norm": 0.09705957099996088, "learning_rate": 2.0710574241958527e-06, "loss": 0.4499, "step": 1647 }, { "epoch": 4.530584192439862, "grad_norm": 0.09629213253097813, "learning_rate": 2.0466871807961253e-06, "loss": 0.4483, "step": 1648 }, { "epoch": 4.533333333333333, "grad_norm": 0.08724520563465142, "learning_rate": 2.0224574052453817e-06, "loss": 0.4449, "step": 1649 }, { "epoch": 4.536082474226804, "grad_norm": 0.09135773175337733, "learning_rate": 1.9983681872196836e-06, "loss": 0.4399, "step": 1650 }, { "epoch": 4.538831615120275, "grad_norm": 0.09009008354611814, "learning_rate": 1.9744196158749183e-06, "loss": 0.4418, "step": 1651 }, { "epoch": 4.541580756013746, "grad_norm": 0.09186018004376852, "learning_rate": 1.9506117798463944e-06, "loss": 0.4478, "step": 1652 }, { "epoch": 4.544329896907216, "grad_norm": 0.08971975311980089, "learning_rate": 1.9269447672485864e-06, "loss": 0.4414, "step": 1653 }, { "epoch": 4.547079037800687, "grad_norm": 0.08979949876575676, "learning_rate": 1.9034186656747478e-06, "loss": 0.4452, "step": 1654 }, { "epoch": 4.549828178694158, "grad_norm": 0.0936250091293909, "learning_rate": 1.8800335621966148e-06, "loss": 0.4482, "step": 1655 }, { "epoch": 4.552577319587629, "grad_norm": 0.08981143391720423, "learning_rate": 1.8567895433640705e-06, "loss": 0.4448, "step": 1656 }, { "epoch": 4.5553264604810995, "grad_norm": 0.08587438163737723, "learning_rate": 1.8336866952048683e-06, "loss": 0.435, "step": 1657 }, { "epoch": 4.55807560137457, "grad_norm": 0.09326486613993211, "learning_rate": 1.8107251032242335e-06, "loss": 0.4508, "step": 1658 }, { "epoch": 4.560824742268041, "grad_norm": 0.09331518445515743, "learning_rate": 1.7879048524046182e-06, "loss": 0.4499, "step": 1659 }, { "epoch": 4.563573883161512, "grad_norm": 0.08976792064689015, "learning_rate": 1.765226027205369e-06, "loss": 0.4466, "step": 1660 }, { "epoch": 4.566323024054983, "grad_norm": 0.0875193893760576, "learning_rate": 1.7426887115623791e-06, "loss": 0.4452, "step": 1661 }, { "epoch": 4.569072164948453, "grad_norm": 0.08920117183529058, "learning_rate": 1.7202929888878329e-06, "loss": 0.4401, "step": 1662 }, { "epoch": 4.571821305841924, "grad_norm": 0.08832595498688149, "learning_rate": 1.6980389420698395e-06, "loss": 0.4451, "step": 1663 }, { "epoch": 4.574570446735395, "grad_norm": 0.08859855283203406, "learning_rate": 1.6759266534721952e-06, "loss": 0.4431, "step": 1664 }, { "epoch": 4.577319587628866, "grad_norm": 0.09152185332529074, "learning_rate": 1.6539562049340085e-06, "loss": 0.4402, "step": 1665 }, { "epoch": 4.5800687285223365, "grad_norm": 0.08871925909133102, "learning_rate": 1.6321276777694307e-06, "loss": 0.4487, "step": 1666 }, { "epoch": 4.582817869415807, "grad_norm": 0.09057802642441674, "learning_rate": 1.6104411527673613e-06, "loss": 0.4435, "step": 1667 }, { "epoch": 4.585567010309278, "grad_norm": 0.09402629602885212, "learning_rate": 1.588896710191139e-06, "loss": 0.4489, "step": 1668 }, { "epoch": 4.588316151202749, "grad_norm": 0.08990522146704653, "learning_rate": 1.567494429778238e-06, "loss": 0.4427, "step": 1669 }, { "epoch": 4.59106529209622, "grad_norm": 0.09102458376196323, "learning_rate": 1.546234390739998e-06, "loss": 0.4498, "step": 1670 }, { "epoch": 4.59381443298969, "grad_norm": 0.09383015002767303, "learning_rate": 1.525116671761282e-06, "loss": 0.4456, "step": 1671 }, { "epoch": 4.596563573883161, "grad_norm": 0.087955107242829, "learning_rate": 1.5041413510002544e-06, "loss": 0.4495, "step": 1672 }, { "epoch": 4.599312714776632, "grad_norm": 0.08874300385624094, "learning_rate": 1.4833085060880349e-06, "loss": 0.4412, "step": 1673 }, { "epoch": 4.602061855670103, "grad_norm": 0.08466557475239722, "learning_rate": 1.4626182141284085e-06, "loss": 0.4455, "step": 1674 }, { "epoch": 4.6048109965635735, "grad_norm": 0.08895580241312337, "learning_rate": 1.4420705516976097e-06, "loss": 0.4406, "step": 1675 }, { "epoch": 4.607560137457044, "grad_norm": 0.08951149546757349, "learning_rate": 1.421665594843953e-06, "loss": 0.4485, "step": 1676 }, { "epoch": 4.610309278350515, "grad_norm": 0.09239895115644584, "learning_rate": 1.4014034190876057e-06, "loss": 0.448, "step": 1677 }, { "epoch": 4.613058419243986, "grad_norm": 0.08778403074207193, "learning_rate": 1.3812840994202792e-06, "loss": 0.44, "step": 1678 }, { "epoch": 4.615807560137457, "grad_norm": 0.08817137101596591, "learning_rate": 1.361307710304991e-06, "loss": 0.4394, "step": 1679 }, { "epoch": 4.618556701030927, "grad_norm": 0.09227710988150696, "learning_rate": 1.3414743256757334e-06, "loss": 0.4501, "step": 1680 }, { "epoch": 4.621305841924398, "grad_norm": 0.08760382434561047, "learning_rate": 1.3217840189372555e-06, "loss": 0.4461, "step": 1681 }, { "epoch": 4.624054982817869, "grad_norm": 0.09031853519071162, "learning_rate": 1.3022368629647253e-06, "loss": 0.4496, "step": 1682 }, { "epoch": 4.62680412371134, "grad_norm": 0.09206239051131887, "learning_rate": 1.2828329301035481e-06, "loss": 0.4432, "step": 1683 }, { "epoch": 4.6295532646048105, "grad_norm": 0.08843350392995919, "learning_rate": 1.263572292169024e-06, "loss": 0.4501, "step": 1684 }, { "epoch": 4.632302405498281, "grad_norm": 0.09481473881768035, "learning_rate": 1.2444550204461092e-06, "loss": 0.4403, "step": 1685 }, { "epoch": 4.635051546391752, "grad_norm": 0.08689601784166971, "learning_rate": 1.2254811856891524e-06, "loss": 0.4435, "step": 1686 }, { "epoch": 4.637800687285223, "grad_norm": 0.08893870660419893, "learning_rate": 1.2066508581216429e-06, "loss": 0.4503, "step": 1687 }, { "epoch": 4.640549828178694, "grad_norm": 0.08937431163006333, "learning_rate": 1.187964107435926e-06, "loss": 0.4456, "step": 1688 }, { "epoch": 4.643298969072165, "grad_norm": 0.08865313646975571, "learning_rate": 1.169421002792972e-06, "loss": 0.4433, "step": 1689 }, { "epoch": 4.646048109965636, "grad_norm": 0.08865228962259511, "learning_rate": 1.151021612822092e-06, "loss": 0.4491, "step": 1690 }, { "epoch": 4.648797250859107, "grad_norm": 0.08885659723955662, "learning_rate": 1.1327660056207113e-06, "loss": 0.4368, "step": 1691 }, { "epoch": 4.651546391752578, "grad_norm": 0.08962612308335349, "learning_rate": 1.114654248754099e-06, "loss": 0.4495, "step": 1692 }, { "epoch": 4.6542955326460484, "grad_norm": 0.08766658442624285, "learning_rate": 1.0966864092551233e-06, "loss": 0.445, "step": 1693 }, { "epoch": 4.657044673539519, "grad_norm": 0.08709960621618447, "learning_rate": 1.0788625536240206e-06, "loss": 0.4471, "step": 1694 }, { "epoch": 4.65979381443299, "grad_norm": 0.09261985426164675, "learning_rate": 1.0611827478281067e-06, "loss": 0.4456, "step": 1695 }, { "epoch": 4.662542955326461, "grad_norm": 0.08650196395240495, "learning_rate": 1.0436470573015733e-06, "loss": 0.4403, "step": 1696 }, { "epoch": 4.665292096219932, "grad_norm": 0.08590676857219628, "learning_rate": 1.026255546945234e-06, "loss": 0.4449, "step": 1697 }, { "epoch": 4.668041237113402, "grad_norm": 0.08741332778858373, "learning_rate": 1.0090082811262802e-06, "loss": 0.4492, "step": 1698 }, { "epoch": 4.670790378006873, "grad_norm": 0.08738071184604224, "learning_rate": 9.919053236780328e-07, "loss": 0.444, "step": 1699 }, { "epoch": 4.673539518900344, "grad_norm": 0.08949580602300157, "learning_rate": 9.74946737899729e-07, "loss": 0.4376, "step": 1700 }, { "epoch": 4.676288659793815, "grad_norm": 0.08859670553407732, "learning_rate": 9.581325865562775e-07, "loss": 0.4486, "step": 1701 }, { "epoch": 4.6790378006872855, "grad_norm": 0.08694870133225886, "learning_rate": 9.414629318780189e-07, "loss": 0.4483, "step": 1702 }, { "epoch": 4.681786941580756, "grad_norm": 0.08702985035490321, "learning_rate": 9.249378355605043e-07, "loss": 0.4419, "step": 1703 }, { "epoch": 4.684536082474227, "grad_norm": 0.08584597159667766, "learning_rate": 9.085573587642637e-07, "loss": 0.4517, "step": 1704 }, { "epoch": 4.687285223367698, "grad_norm": 0.08578695182725778, "learning_rate": 8.923215621145753e-07, "loss": 0.4399, "step": 1705 }, { "epoch": 4.690034364261169, "grad_norm": 0.08592571403893629, "learning_rate": 8.762305057012654e-07, "loss": 0.4515, "step": 1706 }, { "epoch": 4.692783505154639, "grad_norm": 0.08693493065817048, "learning_rate": 8.60284249078438e-07, "loss": 0.4461, "step": 1707 }, { "epoch": 4.69553264604811, "grad_norm": 0.08872474686819293, "learning_rate": 8.444828512642966e-07, "loss": 0.4545, "step": 1708 }, { "epoch": 4.698281786941581, "grad_norm": 0.09336681238214921, "learning_rate": 8.288263707409272e-07, "loss": 0.4543, "step": 1709 }, { "epoch": 4.701030927835052, "grad_norm": 0.0862849071211661, "learning_rate": 8.133148654540402e-07, "loss": 0.4409, "step": 1710 }, { "epoch": 4.7037800687285225, "grad_norm": 0.08478357720964781, "learning_rate": 7.979483928127974e-07, "loss": 0.4422, "step": 1711 }, { "epoch": 4.706529209621993, "grad_norm": 0.08654065268244254, "learning_rate": 7.827270096895811e-07, "loss": 0.4479, "step": 1712 }, { "epoch": 4.709278350515464, "grad_norm": 0.08647792501003562, "learning_rate": 7.676507724197946e-07, "loss": 0.4484, "step": 1713 }, { "epoch": 4.712027491408935, "grad_norm": 0.08597489991064104, "learning_rate": 7.527197368016437e-07, "loss": 0.4417, "step": 1714 }, { "epoch": 4.714776632302406, "grad_norm": 0.0898277616665661, "learning_rate": 7.37933958095911e-07, "loss": 0.4514, "step": 1715 }, { "epoch": 4.717525773195876, "grad_norm": 0.08697107341130339, "learning_rate": 7.232934910258004e-07, "loss": 0.4456, "step": 1716 }, { "epoch": 4.720274914089347, "grad_norm": 0.08637312554758676, "learning_rate": 7.087983897767059e-07, "loss": 0.4477, "step": 1717 }, { "epoch": 4.723024054982818, "grad_norm": 0.0882085833054912, "learning_rate": 6.944487079959982e-07, "loss": 0.4395, "step": 1718 }, { "epoch": 4.725773195876289, "grad_norm": 0.085015391595313, "learning_rate": 6.802444987928436e-07, "loss": 0.4395, "step": 1719 }, { "epoch": 4.7285223367697595, "grad_norm": 0.08451996713089384, "learning_rate": 6.661858147380118e-07, "loss": 0.437, "step": 1720 }, { "epoch": 4.73127147766323, "grad_norm": 0.08738636017583253, "learning_rate": 6.522727078636681e-07, "loss": 0.4508, "step": 1721 }, { "epoch": 4.734020618556701, "grad_norm": 0.08558858353740155, "learning_rate": 6.385052296631955e-07, "loss": 0.4464, "step": 1722 }, { "epoch": 4.736769759450172, "grad_norm": 0.08661882774459452, "learning_rate": 6.248834310909768e-07, "loss": 0.4464, "step": 1723 }, { "epoch": 4.739518900343643, "grad_norm": 0.08559450787680475, "learning_rate": 6.114073625622396e-07, "loss": 0.4372, "step": 1724 }, { "epoch": 4.742268041237113, "grad_norm": 0.0894381151555916, "learning_rate": 5.980770739528563e-07, "loss": 0.4394, "step": 1725 }, { "epoch": 4.745017182130584, "grad_norm": 0.0897214937051112, "learning_rate": 5.84892614599144e-07, "loss": 0.4511, "step": 1726 }, { "epoch": 4.747766323024055, "grad_norm": 0.08890953858034158, "learning_rate": 5.718540332977007e-07, "loss": 0.4504, "step": 1727 }, { "epoch": 4.750515463917526, "grad_norm": 0.08567497991113034, "learning_rate": 5.589613783052317e-07, "loss": 0.4422, "step": 1728 }, { "epoch": 4.7532646048109966, "grad_norm": 0.08728822000081427, "learning_rate": 5.462146973383453e-07, "loss": 0.4589, "step": 1729 }, { "epoch": 4.756013745704467, "grad_norm": 0.08723350175277615, "learning_rate": 5.336140375733934e-07, "loss": 0.4444, "step": 1730 }, { "epoch": 4.758762886597938, "grad_norm": 0.0895784856053207, "learning_rate": 5.211594456462932e-07, "loss": 0.4502, "step": 1731 }, { "epoch": 4.761512027491409, "grad_norm": 0.08733169216974808, "learning_rate": 5.088509676523545e-07, "loss": 0.4482, "step": 1732 }, { "epoch": 4.76426116838488, "grad_norm": 0.08877952523971627, "learning_rate": 4.966886491461109e-07, "loss": 0.4478, "step": 1733 }, { "epoch": 4.7670103092783505, "grad_norm": 0.08664355725321998, "learning_rate": 4.846725351411507e-07, "loss": 0.4463, "step": 1734 }, { "epoch": 4.769759450171821, "grad_norm": 0.08722460501048601, "learning_rate": 4.7280267010993974e-07, "loss": 0.4484, "step": 1735 }, { "epoch": 4.772508591065292, "grad_norm": 0.0869689702035606, "learning_rate": 4.6107909798368324e-07, "loss": 0.4449, "step": 1736 }, { "epoch": 4.775257731958763, "grad_norm": 0.09408455047074887, "learning_rate": 4.495018621521352e-07, "loss": 0.4436, "step": 1737 }, { "epoch": 4.778006872852234, "grad_norm": 0.09000664206402534, "learning_rate": 4.3807100546344296e-07, "loss": 0.44, "step": 1738 }, { "epoch": 4.780756013745704, "grad_norm": 0.08893238330077334, "learning_rate": 4.267865702240048e-07, "loss": 0.4418, "step": 1739 }, { "epoch": 4.783505154639175, "grad_norm": 0.08753276685918031, "learning_rate": 4.1564859819830607e-07, "loss": 0.4407, "step": 1740 }, { "epoch": 4.786254295532646, "grad_norm": 0.08418852751539188, "learning_rate": 4.046571306087499e-07, "loss": 0.4429, "step": 1741 }, { "epoch": 4.789003436426117, "grad_norm": 0.0880689357278319, "learning_rate": 3.9381220813551555e-07, "loss": 0.4517, "step": 1742 }, { "epoch": 4.7917525773195875, "grad_norm": 0.08935991058559813, "learning_rate": 3.831138709164295e-07, "loss": 0.4464, "step": 1743 }, { "epoch": 4.794501718213058, "grad_norm": 0.086751085582121, "learning_rate": 3.725621585467698e-07, "loss": 0.4504, "step": 1744 }, { "epoch": 4.797250859106529, "grad_norm": 0.08660059471433466, "learning_rate": 3.6215711007916434e-07, "loss": 0.4457, "step": 1745 }, { "epoch": 4.8, "grad_norm": 0.08795409500802658, "learning_rate": 3.5189876402341727e-07, "loss": 0.4487, "step": 1746 }, { "epoch": 4.802749140893471, "grad_norm": 0.08517686102366333, "learning_rate": 3.417871583463805e-07, "loss": 0.442, "step": 1747 }, { "epoch": 4.805498281786941, "grad_norm": 0.08656026625325718, "learning_rate": 3.3182233047181154e-07, "loss": 0.4482, "step": 1748 }, { "epoch": 4.808247422680412, "grad_norm": 0.08742296055893738, "learning_rate": 3.2200431728022676e-07, "loss": 0.4497, "step": 1749 }, { "epoch": 4.810996563573883, "grad_norm": 0.08706269427544457, "learning_rate": 3.1233315510877714e-07, "loss": 0.4399, "step": 1750 }, { "epoch": 4.813745704467354, "grad_norm": 0.08541652658402427, "learning_rate": 3.0280887975111087e-07, "loss": 0.4463, "step": 1751 }, { "epoch": 4.8164948453608245, "grad_norm": 0.08561518369119724, "learning_rate": 2.9343152645723075e-07, "loss": 0.4496, "step": 1752 }, { "epoch": 4.819243986254295, "grad_norm": 0.08671209749993644, "learning_rate": 2.8420112993337026e-07, "loss": 0.4441, "step": 1753 }, { "epoch": 4.821993127147766, "grad_norm": 0.0868113057408611, "learning_rate": 2.75117724341869e-07, "loss": 0.4539, "step": 1754 }, { "epoch": 4.824742268041237, "grad_norm": 0.08787299513468588, "learning_rate": 2.661813433010485e-07, "loss": 0.4469, "step": 1755 }, { "epoch": 4.827491408934708, "grad_norm": 0.08426954798208526, "learning_rate": 2.5739201988506544e-07, "loss": 0.4452, "step": 1756 }, { "epoch": 4.830240549828178, "grad_norm": 0.08925999331930048, "learning_rate": 2.487497866238231e-07, "loss": 0.4431, "step": 1757 }, { "epoch": 4.832989690721649, "grad_norm": 0.09009930022107024, "learning_rate": 2.4025467550283345e-07, "loss": 0.4428, "step": 1758 }, { "epoch": 4.83573883161512, "grad_norm": 0.08903112575869974, "learning_rate": 2.3190671796307963e-07, "loss": 0.4376, "step": 1759 }, { "epoch": 4.838487972508591, "grad_norm": 0.087097271598177, "learning_rate": 2.237059449009449e-07, "loss": 0.4514, "step": 1760 }, { "epoch": 4.8412371134020615, "grad_norm": 0.08665721242897642, "learning_rate": 2.1565238666805266e-07, "loss": 0.4507, "step": 1761 }, { "epoch": 4.843986254295532, "grad_norm": 0.08570901423189047, "learning_rate": 2.0774607307118665e-07, "loss": 0.444, "step": 1762 }, { "epoch": 4.846735395189003, "grad_norm": 0.08855478539796897, "learning_rate": 1.9998703337216207e-07, "loss": 0.4459, "step": 1763 }, { "epoch": 4.849484536082474, "grad_norm": 0.08830013933034192, "learning_rate": 1.9237529628772345e-07, "loss": 0.4471, "step": 1764 }, { "epoch": 4.852233676975945, "grad_norm": 0.08632909822266122, "learning_rate": 1.8491088998943806e-07, "loss": 0.4443, "step": 1765 }, { "epoch": 4.854982817869415, "grad_norm": 0.08788988835658808, "learning_rate": 1.7759384210358943e-07, "loss": 0.4495, "step": 1766 }, { "epoch": 4.857731958762886, "grad_norm": 0.08641182871335162, "learning_rate": 1.7042417971108837e-07, "loss": 0.448, "step": 1767 }, { "epoch": 4.860481099656358, "grad_norm": 0.08684528157216327, "learning_rate": 1.6340192934734879e-07, "loss": 0.4503, "step": 1768 }, { "epoch": 4.863230240549829, "grad_norm": 0.08740681191652898, "learning_rate": 1.5652711700220756e-07, "loss": 0.4476, "step": 1769 }, { "epoch": 4.8659793814432994, "grad_norm": 0.08455281840311087, "learning_rate": 1.4979976811982267e-07, "loss": 0.4501, "step": 1770 }, { "epoch": 4.86872852233677, "grad_norm": 0.08621142637068548, "learning_rate": 1.4321990759858406e-07, "loss": 0.4427, "step": 1771 }, { "epoch": 4.871477663230241, "grad_norm": 0.09048418177504676, "learning_rate": 1.3678755979100734e-07, "loss": 0.4414, "step": 1772 }, { "epoch": 4.874226804123712, "grad_norm": 0.08552997152050767, "learning_rate": 1.3050274850366252e-07, "loss": 0.4452, "step": 1773 }, { "epoch": 4.876975945017183, "grad_norm": 0.09163336347075388, "learning_rate": 1.2436549699706756e-07, "loss": 0.4421, "step": 1774 }, { "epoch": 4.879725085910653, "grad_norm": 0.08581324177498632, "learning_rate": 1.183758279856262e-07, "loss": 0.445, "step": 1775 }, { "epoch": 4.882474226804124, "grad_norm": 0.08520078540460027, "learning_rate": 1.1253376363751234e-07, "loss": 0.4398, "step": 1776 }, { "epoch": 4.885223367697595, "grad_norm": 0.08389381645053694, "learning_rate": 1.0683932557461696e-07, "loss": 0.4509, "step": 1777 }, { "epoch": 4.887972508591066, "grad_norm": 0.08463720514626481, "learning_rate": 1.0129253487244584e-07, "loss": 0.444, "step": 1778 }, { "epoch": 4.8907216494845365, "grad_norm": 0.08376091905777829, "learning_rate": 9.589341206006186e-08, "loss": 0.4424, "step": 1779 }, { "epoch": 4.893470790378007, "grad_norm": 0.0864473265773848, "learning_rate": 9.064197711999178e-08, "loss": 0.436, "step": 1780 }, { "epoch": 4.896219931271478, "grad_norm": 0.08780409654999521, "learning_rate": 8.553824948816402e-08, "loss": 0.4439, "step": 1781 }, { "epoch": 4.898969072164949, "grad_norm": 0.08643060928003499, "learning_rate": 8.058224805382431e-08, "loss": 0.457, "step": 1782 }, { "epoch": 4.90171821305842, "grad_norm": 0.08999003837315114, "learning_rate": 7.577399115948236e-08, "loss": 0.4536, "step": 1783 }, { "epoch": 4.90446735395189, "grad_norm": 0.08500920989818493, "learning_rate": 7.111349660082756e-08, "loss": 0.4524, "step": 1784 }, { "epoch": 4.907216494845361, "grad_norm": 0.08713394706241256, "learning_rate": 6.660078162667561e-08, "loss": 0.4474, "step": 1785 }, { "epoch": 4.909965635738832, "grad_norm": 0.08415301103290775, "learning_rate": 6.223586293890193e-08, "loss": 0.4388, "step": 1786 }, { "epoch": 4.912714776632303, "grad_norm": 0.08641909751480724, "learning_rate": 5.8018756692370673e-08, "loss": 0.4452, "step": 1787 }, { "epoch": 4.9154639175257735, "grad_norm": 0.085407296137395, "learning_rate": 5.3949478494885745e-08, "loss": 0.4442, "step": 1788 }, { "epoch": 4.918213058419244, "grad_norm": 0.08620621228120745, "learning_rate": 5.002804340712875e-08, "loss": 0.4395, "step": 1789 }, { "epoch": 4.920962199312715, "grad_norm": 0.08440435770555102, "learning_rate": 4.625446594260563e-08, "loss": 0.4475, "step": 1790 }, { "epoch": 4.923711340206186, "grad_norm": 0.08498966775862096, "learning_rate": 4.262876006758454e-08, "loss": 0.4461, "step": 1791 }, { "epoch": 4.926460481099657, "grad_norm": 0.0852940765401518, "learning_rate": 3.915093920105584e-08, "loss": 0.4508, "step": 1792 }, { "epoch": 4.929209621993127, "grad_norm": 0.08708725826340913, "learning_rate": 3.582101621467882e-08, "loss": 0.4478, "step": 1793 }, { "epoch": 4.931958762886598, "grad_norm": 0.08563954373853248, "learning_rate": 3.263900343272841e-08, "loss": 0.4506, "step": 1794 }, { "epoch": 4.934707903780069, "grad_norm": 0.08351208793444787, "learning_rate": 2.960491263205523e-08, "loss": 0.4358, "step": 1795 }, { "epoch": 4.93745704467354, "grad_norm": 0.08968304296903096, "learning_rate": 2.6718755042041134e-08, "loss": 0.4408, "step": 1796 }, { "epoch": 4.9402061855670105, "grad_norm": 0.08687120751973657, "learning_rate": 2.3980541344554852e-08, "loss": 0.4472, "step": 1797 }, { "epoch": 4.942955326460481, "grad_norm": 0.0843041818695698, "learning_rate": 2.1390281673916435e-08, "loss": 0.4433, "step": 1798 }, { "epoch": 4.945704467353952, "grad_norm": 0.08678186311343745, "learning_rate": 1.894798561685729e-08, "loss": 0.4401, "step": 1799 }, { "epoch": 4.948453608247423, "grad_norm": 0.08552567460187925, "learning_rate": 1.6653662212484656e-08, "loss": 0.4463, "step": 1800 }, { "epoch": 4.951202749140894, "grad_norm": 0.08327662566119383, "learning_rate": 1.4507319952246079e-08, "loss": 0.4428, "step": 1801 }, { "epoch": 4.953951890034364, "grad_norm": 0.08558816066436331, "learning_rate": 1.2508966779907206e-08, "loss": 0.4439, "step": 1802 }, { "epoch": 4.956701030927835, "grad_norm": 0.08685318316335923, "learning_rate": 1.0658610091507371e-08, "loss": 0.4433, "step": 1803 }, { "epoch": 4.959450171821306, "grad_norm": 0.08552301347515416, "learning_rate": 8.956256735355162e-09, "loss": 0.4462, "step": 1804 }, { "epoch": 4.962199312714777, "grad_norm": 0.08918690199355128, "learning_rate": 7.401913011966244e-09, "loss": 0.4516, "step": 1805 }, { "epoch": 4.9649484536082475, "grad_norm": 0.08439607746487938, "learning_rate": 5.995584674085564e-09, "loss": 0.4443, "step": 1806 }, { "epoch": 4.967697594501718, "grad_norm": 0.0859086823143405, "learning_rate": 4.737276926620738e-09, "loss": 0.4477, "step": 1807 }, { "epoch": 4.970446735395189, "grad_norm": 0.08603802799919767, "learning_rate": 3.6269944266686953e-09, "loss": 0.4398, "step": 1808 }, { "epoch": 4.97319587628866, "grad_norm": 0.08797273612741212, "learning_rate": 2.664741283453509e-09, "loss": 0.4455, "step": 1809 }, { "epoch": 4.975945017182131, "grad_norm": 0.08498147628736628, "learning_rate": 1.8505210583441568e-09, "loss": 0.4417, "step": 1810 }, { "epoch": 4.9786941580756015, "grad_norm": 0.08573341738089828, "learning_rate": 1.184336764823435e-09, "loss": 0.4457, "step": 1811 }, { "epoch": 4.981443298969072, "grad_norm": 0.08322231115466387, "learning_rate": 6.661908684924002e-10, "loss": 0.454, "step": 1812 }, { "epoch": 4.984192439862543, "grad_norm": 0.08656315765889926, "learning_rate": 2.9608528703928274e-10, "loss": 0.4449, "step": 1813 }, { "epoch": 4.986941580756014, "grad_norm": 0.08364091527237393, "learning_rate": 7.402139024836886e-11, "loss": 0.455, "step": 1814 }, { "epoch": 4.989690721649485, "grad_norm": 0.08519652883753787, "learning_rate": 0.0, "loss": 0.4445, "step": 1815 }, { "epoch": 4.989690721649485, "step": 1815, "total_flos": 3.73703706563429e+19, "train_loss": 0.5429748183767986, "train_runtime": 94067.7478, "train_samples_per_second": 9.898, "train_steps_per_second": 0.019 } ], "logging_steps": 1.0, "max_steps": 1815, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.73703706563429e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }