{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.5434782608695652, "eval_steps": 500, "global_step": 2000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0002717391304347826, "grad_norm": Infinity, "learning_rate": 0.0, "loss": 3.8119, "step": 1 }, { "epoch": 0.0005434782608695652, "grad_norm": Infinity, "learning_rate": 0.0, "loss": 5.6312, "step": 2 }, { "epoch": 0.0008152173913043478, "grad_norm": 3.606865644454956, "learning_rate": 1.0000000000000001e-07, "loss": 3.6085, "step": 3 }, { "epoch": 0.0010869565217391304, "grad_norm": 5.723954677581787, "learning_rate": 2.0000000000000002e-07, "loss": 4.5497, "step": 4 }, { "epoch": 0.001358695652173913, "grad_norm": 8.396666526794434, "learning_rate": 3.0000000000000004e-07, "loss": 5.6655, "step": 5 }, { "epoch": 0.0016304347826086956, "grad_norm": 6.602017402648926, "learning_rate": 4.0000000000000003e-07, "loss": 4.9113, "step": 6 }, { "epoch": 0.0019021739130434783, "grad_norm": 2.8019049167633057, "learning_rate": 5e-07, "loss": 3.0843, "step": 7 }, { "epoch": 0.002173913043478261, "grad_norm": 5.959436893463135, "learning_rate": 6.000000000000001e-07, "loss": 4.4235, "step": 8 }, { "epoch": 0.0024456521739130437, "grad_norm": 5.623807430267334, "learning_rate": 7.000000000000001e-07, "loss": 4.5667, "step": 9 }, { "epoch": 0.002717391304347826, "grad_norm": 5.9944844245910645, "learning_rate": 8.000000000000001e-07, "loss": 4.6548, "step": 10 }, { "epoch": 0.002989130434782609, "grad_norm": 7.308701038360596, "learning_rate": 9e-07, "loss": 5.1422, "step": 11 }, { "epoch": 0.003260869565217391, "grad_norm": 5.805826187133789, "learning_rate": 1e-06, "loss": 4.3458, "step": 12 }, { "epoch": 0.003532608695652174, "grad_norm": 5.096138000488281, "learning_rate": 1.1e-06, "loss": 4.0761, "step": 13 }, { "epoch": 0.0038043478260869567, "grad_norm": 4.222947120666504, "learning_rate": 1.2000000000000002e-06, "loss": 3.7963, "step": 14 }, { "epoch": 0.004076086956521739, "grad_norm": 8.304177284240723, "learning_rate": 1.3e-06, "loss": 5.4697, "step": 15 }, { "epoch": 0.004347826086956522, "grad_norm": 6.131305694580078, "learning_rate": 1.4000000000000001e-06, "loss": 4.5078, "step": 16 }, { "epoch": 0.004619565217391305, "grad_norm": 5.1205315589904785, "learning_rate": 1.5e-06, "loss": 4.1935, "step": 17 }, { "epoch": 0.004891304347826087, "grad_norm": 3.348874568939209, "learning_rate": 1.6000000000000001e-06, "loss": 3.165, "step": 18 }, { "epoch": 0.005163043478260869, "grad_norm": 11.620071411132812, "learning_rate": 1.7e-06, "loss": 6.6692, "step": 19 }, { "epoch": 0.005434782608695652, "grad_norm": 5.148247241973877, "learning_rate": 1.8e-06, "loss": 3.7967, "step": 20 }, { "epoch": 0.005706521739130435, "grad_norm": 6.738089561462402, "learning_rate": 1.9000000000000002e-06, "loss": 4.6053, "step": 21 }, { "epoch": 0.005978260869565218, "grad_norm": 7.383264541625977, "learning_rate": 2e-06, "loss": 4.7862, "step": 22 }, { "epoch": 0.00625, "grad_norm": 9.380542755126953, "learning_rate": 2.1000000000000002e-06, "loss": 5.5961, "step": 23 }, { "epoch": 0.006521739130434782, "grad_norm": 5.481351852416992, "learning_rate": 2.2e-06, "loss": 3.9489, "step": 24 }, { "epoch": 0.006793478260869565, "grad_norm": 3.769026756286621, "learning_rate": 2.3e-06, "loss": 3.3576, "step": 25 }, { "epoch": 0.007065217391304348, "grad_norm": 6.088125705718994, "learning_rate": 2.4000000000000003e-06, "loss": 4.3102, "step": 26 }, { "epoch": 0.007336956521739131, "grad_norm": 3.7228612899780273, "learning_rate": 2.4999999999999998e-06, "loss": 3.2898, "step": 27 }, { "epoch": 0.007608695652173913, "grad_norm": 5.163967609405518, "learning_rate": 2.6e-06, "loss": 3.9394, "step": 28 }, { "epoch": 0.007880434782608695, "grad_norm": 3.577744960784912, "learning_rate": 2.7e-06, "loss": 3.4044, "step": 29 }, { "epoch": 0.008152173913043478, "grad_norm": 5.421440124511719, "learning_rate": 2.8000000000000003e-06, "loss": 4.2156, "step": 30 }, { "epoch": 0.00842391304347826, "grad_norm": 4.284877777099609, "learning_rate": 2.9e-06, "loss": 3.519, "step": 31 }, { "epoch": 0.008695652173913044, "grad_norm": 6.597541809082031, "learning_rate": 3e-06, "loss": 4.5028, "step": 32 }, { "epoch": 0.008967391304347826, "grad_norm": 7.851593494415283, "learning_rate": 3.1e-06, "loss": 5.1247, "step": 33 }, { "epoch": 0.00923913043478261, "grad_norm": 2.644531726837158, "learning_rate": 3.2000000000000003e-06, "loss": 2.9878, "step": 34 }, { "epoch": 0.009510869565217392, "grad_norm": 3.316633462905884, "learning_rate": 3.3e-06, "loss": 3.4112, "step": 35 }, { "epoch": 0.009782608695652175, "grad_norm": 3.0373375415802, "learning_rate": 3.4e-06, "loss": 3.301, "step": 36 }, { "epoch": 0.010054347826086956, "grad_norm": 5.997847080230713, "learning_rate": 3.5e-06, "loss": 4.9763, "step": 37 }, { "epoch": 0.010326086956521738, "grad_norm": 5.2729268074035645, "learning_rate": 3.6e-06, "loss": 4.7279, "step": 38 }, { "epoch": 0.010597826086956521, "grad_norm": 2.7055575847625732, "learning_rate": 3.7e-06, "loss": 3.4656, "step": 39 }, { "epoch": 0.010869565217391304, "grad_norm": 3.5352964401245117, "learning_rate": 3.8000000000000005e-06, "loss": 3.9642, "step": 40 }, { "epoch": 0.011141304347826087, "grad_norm": 2.8491668701171875, "learning_rate": 3.9e-06, "loss": 3.8092, "step": 41 }, { "epoch": 0.01141304347826087, "grad_norm": 1.8910856246948242, "learning_rate": 4e-06, "loss": 3.0861, "step": 42 }, { "epoch": 0.011684782608695652, "grad_norm": 2.873784303665161, "learning_rate": 4.1e-06, "loss": 3.788, "step": 43 }, { "epoch": 0.011956521739130435, "grad_norm": 3.121163845062256, "learning_rate": 4.2000000000000004e-06, "loss": 4.0837, "step": 44 }, { "epoch": 0.012228260869565218, "grad_norm": 2.977182388305664, "learning_rate": 4.3e-06, "loss": 4.0382, "step": 45 }, { "epoch": 0.0125, "grad_norm": 3.935685396194458, "learning_rate": 4.4e-06, "loss": 5.0603, "step": 46 }, { "epoch": 0.012771739130434783, "grad_norm": 1.5644341707229614, "learning_rate": 4.5e-06, "loss": 3.0085, "step": 47 }, { "epoch": 0.013043478260869565, "grad_norm": 2.2639412879943848, "learning_rate": 4.6e-06, "loss": 3.5606, "step": 48 }, { "epoch": 0.013315217391304347, "grad_norm": 3.8248438835144043, "learning_rate": 4.700000000000001e-06, "loss": 5.1055, "step": 49 }, { "epoch": 0.01358695652173913, "grad_norm": 2.0189621448516846, "learning_rate": 4.800000000000001e-06, "loss": 3.6357, "step": 50 }, { "epoch": 0.013858695652173913, "grad_norm": 1.7053027153015137, "learning_rate": 4.9e-06, "loss": 3.3706, "step": 51 }, { "epoch": 0.014130434782608696, "grad_norm": 2.996800184249878, "learning_rate": 4.9999999999999996e-06, "loss": 4.2643, "step": 52 }, { "epoch": 0.014402173913043478, "grad_norm": 2.2045717239379883, "learning_rate": 5.1e-06, "loss": 3.4676, "step": 53 }, { "epoch": 0.014673913043478261, "grad_norm": 1.9880719184875488, "learning_rate": 5.2e-06, "loss": 3.7901, "step": 54 }, { "epoch": 0.014945652173913044, "grad_norm": 1.4993855953216553, "learning_rate": 5.3e-06, "loss": 3.3175, "step": 55 }, { "epoch": 0.015217391304347827, "grad_norm": 1.3958892822265625, "learning_rate": 5.4e-06, "loss": 3.0166, "step": 56 }, { "epoch": 0.01548913043478261, "grad_norm": 1.5511572360992432, "learning_rate": 5.5e-06, "loss": 3.4503, "step": 57 }, { "epoch": 0.01576086956521739, "grad_norm": 2.274249792098999, "learning_rate": 5.600000000000001e-06, "loss": 4.1636, "step": 58 }, { "epoch": 0.016032608695652175, "grad_norm": 1.7447724342346191, "learning_rate": 5.7000000000000005e-06, "loss": 3.7056, "step": 59 }, { "epoch": 0.016304347826086956, "grad_norm": 2.4315755367279053, "learning_rate": 5.8e-06, "loss": 3.9477, "step": 60 }, { "epoch": 0.01657608695652174, "grad_norm": 1.5694116353988647, "learning_rate": 5.899999999999999e-06, "loss": 3.321, "step": 61 }, { "epoch": 0.01684782608695652, "grad_norm": 1.856467366218567, "learning_rate": 6e-06, "loss": 3.8384, "step": 62 }, { "epoch": 0.017119565217391303, "grad_norm": 1.3913215398788452, "learning_rate": 6.1e-06, "loss": 3.1717, "step": 63 }, { "epoch": 0.017391304347826087, "grad_norm": 1.1280888319015503, "learning_rate": 6.2e-06, "loss": 3.0487, "step": 64 }, { "epoch": 0.017663043478260868, "grad_norm": 1.3553597927093506, "learning_rate": 6.3e-06, "loss": 3.4274, "step": 65 }, { "epoch": 0.017934782608695653, "grad_norm": 1.5680698156356812, "learning_rate": 6.4000000000000006e-06, "loss": 3.2402, "step": 66 }, { "epoch": 0.018206521739130434, "grad_norm": 1.2775722742080688, "learning_rate": 6.5000000000000004e-06, "loss": 3.1252, "step": 67 }, { "epoch": 0.01847826086956522, "grad_norm": 1.547057867050171, "learning_rate": 6.6e-06, "loss": 3.6654, "step": 68 }, { "epoch": 0.01875, "grad_norm": 1.6398184299468994, "learning_rate": 6.7e-06, "loss": 3.4637, "step": 69 }, { "epoch": 0.019021739130434784, "grad_norm": 0.8923302292823792, "learning_rate": 6.8e-06, "loss": 2.6619, "step": 70 }, { "epoch": 0.019293478260869565, "grad_norm": 1.1862719058990479, "learning_rate": 6.900000000000001e-06, "loss": 3.1319, "step": 71 }, { "epoch": 0.01956521739130435, "grad_norm": 1.0350522994995117, "learning_rate": 7e-06, "loss": 2.8874, "step": 72 }, { "epoch": 0.01983695652173913, "grad_norm": 1.7756403684616089, "learning_rate": 7.1e-06, "loss": 3.7356, "step": 73 }, { "epoch": 0.02010869565217391, "grad_norm": 1.2788705825805664, "learning_rate": 7.2e-06, "loss": 3.2409, "step": 74 }, { "epoch": 0.020380434782608696, "grad_norm": 1.4206314086914062, "learning_rate": 7.3e-06, "loss": 3.5884, "step": 75 }, { "epoch": 0.020652173913043477, "grad_norm": 2.0677270889282227, "learning_rate": 7.4e-06, "loss": 4.3218, "step": 76 }, { "epoch": 0.02092391304347826, "grad_norm": 0.700586199760437, "learning_rate": 7.5e-06, "loss": 2.4596, "step": 77 }, { "epoch": 0.021195652173913043, "grad_norm": 2.1819796562194824, "learning_rate": 7.600000000000001e-06, "loss": 3.9885, "step": 78 }, { "epoch": 0.021467391304347827, "grad_norm": 1.0155264139175415, "learning_rate": 7.699999999999999e-06, "loss": 2.7548, "step": 79 }, { "epoch": 0.021739130434782608, "grad_norm": 1.2022877931594849, "learning_rate": 7.8e-06, "loss": 3.1796, "step": 80 }, { "epoch": 0.022010869565217393, "grad_norm": 1.1962783336639404, "learning_rate": 7.899999999999999e-06, "loss": 2.907, "step": 81 }, { "epoch": 0.022282608695652174, "grad_norm": 1.0273329019546509, "learning_rate": 8e-06, "loss": 2.7207, "step": 82 }, { "epoch": 0.022554347826086958, "grad_norm": 2.3930468559265137, "learning_rate": 8.1e-06, "loss": 3.9644, "step": 83 }, { "epoch": 0.02282608695652174, "grad_norm": 2.244040012359619, "learning_rate": 8.2e-06, "loss": 3.5003, "step": 84 }, { "epoch": 0.02309782608695652, "grad_norm": 1.685024380683899, "learning_rate": 8.3e-06, "loss": 3.0757, "step": 85 }, { "epoch": 0.023369565217391305, "grad_norm": 3.1130995750427246, "learning_rate": 8.400000000000001e-06, "loss": 3.9588, "step": 86 }, { "epoch": 0.023641304347826086, "grad_norm": 2.6688392162323, "learning_rate": 8.5e-06, "loss": 3.6427, "step": 87 }, { "epoch": 0.02391304347826087, "grad_norm": 1.3578976392745972, "learning_rate": 8.6e-06, "loss": 2.9824, "step": 88 }, { "epoch": 0.02418478260869565, "grad_norm": 1.2437134981155396, "learning_rate": 8.7e-06, "loss": 3.1183, "step": 89 }, { "epoch": 0.024456521739130436, "grad_norm": 2.143594741821289, "learning_rate": 8.8e-06, "loss": 3.6833, "step": 90 }, { "epoch": 0.024728260869565217, "grad_norm": 1.2158911228179932, "learning_rate": 8.900000000000001e-06, "loss": 2.7814, "step": 91 }, { "epoch": 0.025, "grad_norm": 2.4209673404693604, "learning_rate": 9e-06, "loss": 4.3185, "step": 92 }, { "epoch": 0.025271739130434782, "grad_norm": 3.090715169906616, "learning_rate": 9.100000000000001e-06, "loss": 4.0217, "step": 93 }, { "epoch": 0.025543478260869567, "grad_norm": 3.1287500858306885, "learning_rate": 9.2e-06, "loss": 3.8665, "step": 94 }, { "epoch": 0.025815217391304348, "grad_norm": 2.6370797157287598, "learning_rate": 9.3e-06, "loss": 3.7443, "step": 95 }, { "epoch": 0.02608695652173913, "grad_norm": 4.120901107788086, "learning_rate": 9.400000000000001e-06, "loss": 4.3482, "step": 96 }, { "epoch": 0.026358695652173914, "grad_norm": NaN, "learning_rate": 9.400000000000001e-06, "loss": 3.7633, "step": 97 }, { "epoch": 0.026630434782608695, "grad_norm": 2.1682493686676025, "learning_rate": 9.5e-06, "loss": 3.135, "step": 98 }, { "epoch": 0.02690217391304348, "grad_norm": 9.943031311035156, "learning_rate": 9.600000000000001e-06, "loss": 4.4992, "step": 99 }, { "epoch": 0.02717391304347826, "grad_norm": 9.851699829101562, "learning_rate": 9.699999999999999e-06, "loss": 3.3346, "step": 100 }, { "epoch": 0.027445652173913045, "grad_norm": 9.068572044372559, "learning_rate": 9.8e-06, "loss": 3.4352, "step": 101 }, { "epoch": 0.027717391304347826, "grad_norm": 16.46036148071289, "learning_rate": 9.9e-06, "loss": 4.2803, "step": 102 }, { "epoch": 0.02798913043478261, "grad_norm": 4.916048049926758, "learning_rate": 9.999999999999999e-06, "loss": 5.2283, "step": 103 }, { "epoch": 0.02826086956521739, "grad_norm": 3.9112675189971924, "learning_rate": 1.01e-05, "loss": 3.9085, "step": 104 }, { "epoch": 0.028532608695652172, "grad_norm": 2.9353702068328857, "learning_rate": 1.02e-05, "loss": 3.436, "step": 105 }, { "epoch": 0.028804347826086957, "grad_norm": 2.1756107807159424, "learning_rate": 1.03e-05, "loss": 4.2118, "step": 106 }, { "epoch": 0.029076086956521738, "grad_norm": 2.4335989952087402, "learning_rate": 1.04e-05, "loss": 4.3144, "step": 107 }, { "epoch": 0.029347826086956522, "grad_norm": 2.6821694374084473, "learning_rate": 1.05e-05, "loss": 4.461, "step": 108 }, { "epoch": 0.029619565217391303, "grad_norm": 1.9325740337371826, "learning_rate": 1.06e-05, "loss": 4.1079, "step": 109 }, { "epoch": 0.029891304347826088, "grad_norm": 3.598806381225586, "learning_rate": 1.0700000000000001e-05, "loss": 3.6908, "step": 110 }, { "epoch": 0.03016304347826087, "grad_norm": 1.7592302560806274, "learning_rate": 1.08e-05, "loss": 3.2482, "step": 111 }, { "epoch": 0.030434782608695653, "grad_norm": 2.4772744178771973, "learning_rate": 1.09e-05, "loss": 4.3941, "step": 112 }, { "epoch": 0.030706521739130434, "grad_norm": 1.8405948877334595, "learning_rate": 1.1e-05, "loss": 3.6201, "step": 113 }, { "epoch": 0.03097826086956522, "grad_norm": 4.303748607635498, "learning_rate": 1.11e-05, "loss": 4.6995, "step": 114 }, { "epoch": 0.03125, "grad_norm": 2.199484348297119, "learning_rate": 1.1200000000000001e-05, "loss": 3.7969, "step": 115 }, { "epoch": 0.03152173913043478, "grad_norm": 1.0349135398864746, "learning_rate": 1.13e-05, "loss": 2.8461, "step": 116 }, { "epoch": 0.03179347826086956, "grad_norm": 2.374260425567627, "learning_rate": 1.1400000000000001e-05, "loss": 4.2862, "step": 117 }, { "epoch": 0.03206521739130435, "grad_norm": 1.283584713935852, "learning_rate": 1.1500000000000002e-05, "loss": 3.1288, "step": 118 }, { "epoch": 0.03233695652173913, "grad_norm": 3.549565076828003, "learning_rate": 1.16e-05, "loss": 4.1137, "step": 119 }, { "epoch": 0.03260869565217391, "grad_norm": 1.3312267065048218, "learning_rate": 1.1700000000000001e-05, "loss": 3.6003, "step": 120 }, { "epoch": 0.03288043478260869, "grad_norm": 1.655480146408081, "learning_rate": 1.1799999999999999e-05, "loss": 3.7798, "step": 121 }, { "epoch": 0.03315217391304348, "grad_norm": 1.6580840349197388, "learning_rate": 1.19e-05, "loss": 3.5654, "step": 122 }, { "epoch": 0.03342391304347826, "grad_norm": 2.8110294342041016, "learning_rate": 1.2e-05, "loss": 3.9673, "step": 123 }, { "epoch": 0.03369565217391304, "grad_norm": 1.2965394258499146, "learning_rate": 1.21e-05, "loss": 3.26, "step": 124 }, { "epoch": 0.033967391304347824, "grad_norm": 1.914302945137024, "learning_rate": 1.22e-05, "loss": 3.95, "step": 125 }, { "epoch": 0.034239130434782605, "grad_norm": 1.3706483840942383, "learning_rate": 1.2299999999999999e-05, "loss": 3.5317, "step": 126 }, { "epoch": 0.03451086956521739, "grad_norm": 2.016113042831421, "learning_rate": 1.24e-05, "loss": 3.2582, "step": 127 }, { "epoch": 0.034782608695652174, "grad_norm": 2.5129711627960205, "learning_rate": 1.25e-05, "loss": 3.3608, "step": 128 }, { "epoch": 0.035054347826086955, "grad_norm": Infinity, "learning_rate": 1.25e-05, "loss": 4.3213, "step": 129 }, { "epoch": 0.035326086956521736, "grad_norm": 3.073050022125244, "learning_rate": 1.26e-05, "loss": 4.4668, "step": 130 }, { "epoch": 0.035597826086956524, "grad_norm": 1.8033742904663086, "learning_rate": 1.27e-05, "loss": 3.5871, "step": 131 }, { "epoch": 0.035869565217391305, "grad_norm": 3.148634195327759, "learning_rate": 1.2800000000000001e-05, "loss": 3.7611, "step": 132 }, { "epoch": 0.036141304347826086, "grad_norm": 2.898167610168457, "learning_rate": 1.29e-05, "loss": 3.6928, "step": 133 }, { "epoch": 0.03641304347826087, "grad_norm": 1.9879661798477173, "learning_rate": 1.3000000000000001e-05, "loss": 3.607, "step": 134 }, { "epoch": 0.036684782608695655, "grad_norm": 2.0479657649993896, "learning_rate": 1.31e-05, "loss": 3.2456, "step": 135 }, { "epoch": 0.03695652173913044, "grad_norm": 1.5060527324676514, "learning_rate": 1.32e-05, "loss": 3.4124, "step": 136 }, { "epoch": 0.03722826086956522, "grad_norm": 2.2012810707092285, "learning_rate": 1.3300000000000001e-05, "loss": 4.1008, "step": 137 }, { "epoch": 0.0375, "grad_norm": 2.2929439544677734, "learning_rate": 1.34e-05, "loss": 4.5483, "step": 138 }, { "epoch": 0.03777173913043478, "grad_norm": 1.777136206626892, "learning_rate": 1.3500000000000001e-05, "loss": 3.9796, "step": 139 }, { "epoch": 0.03804347826086957, "grad_norm": 3.1867823600769043, "learning_rate": 1.36e-05, "loss": 4.13, "step": 140 }, { "epoch": 0.03831521739130435, "grad_norm": 2.070611000061035, "learning_rate": 1.3700000000000001e-05, "loss": 3.3474, "step": 141 }, { "epoch": 0.03858695652173913, "grad_norm": 1.9961167573928833, "learning_rate": 1.3800000000000002e-05, "loss": 4.0889, "step": 142 }, { "epoch": 0.03885869565217391, "grad_norm": 1.709995985031128, "learning_rate": 1.39e-05, "loss": 3.9576, "step": 143 }, { "epoch": 0.0391304347826087, "grad_norm": 1.793494462966919, "learning_rate": 1.4e-05, "loss": 4.419, "step": 144 }, { "epoch": 0.03940217391304348, "grad_norm": 2.1228878498077393, "learning_rate": 1.4099999999999999e-05, "loss": 3.7677, "step": 145 }, { "epoch": 0.03967391304347826, "grad_norm": 1.5662310123443604, "learning_rate": 1.42e-05, "loss": 3.5038, "step": 146 }, { "epoch": 0.03994565217391304, "grad_norm": 1.1859204769134521, "learning_rate": 1.43e-05, "loss": 3.0831, "step": 147 }, { "epoch": 0.04021739130434782, "grad_norm": 1.998694896697998, "learning_rate": 1.44e-05, "loss": 4.06, "step": 148 }, { "epoch": 0.04048913043478261, "grad_norm": 2.2709462642669678, "learning_rate": 1.45e-05, "loss": 3.4234, "step": 149 }, { "epoch": 0.04076086956521739, "grad_norm": 3.572169065475464, "learning_rate": 1.46e-05, "loss": 4.1884, "step": 150 }, { "epoch": 0.04103260869565217, "grad_norm": 2.4503207206726074, "learning_rate": 1.47e-05, "loss": 4.0506, "step": 151 }, { "epoch": 0.041304347826086954, "grad_norm": 2.66367244720459, "learning_rate": 1.48e-05, "loss": 4.6858, "step": 152 }, { "epoch": 0.04157608695652174, "grad_norm": 2.1236956119537354, "learning_rate": 1.49e-05, "loss": 3.2571, "step": 153 }, { "epoch": 0.04184782608695652, "grad_norm": 4.9067206382751465, "learning_rate": 1.5e-05, "loss": 5.32, "step": 154 }, { "epoch": 0.042119565217391304, "grad_norm": 2.0733673572540283, "learning_rate": 1.51e-05, "loss": 3.7548, "step": 155 }, { "epoch": 0.042391304347826085, "grad_norm": 1.186569333076477, "learning_rate": 1.5200000000000002e-05, "loss": 3.3953, "step": 156 }, { "epoch": 0.042663043478260866, "grad_norm": 2.5303616523742676, "learning_rate": 1.53e-05, "loss": 4.0742, "step": 157 }, { "epoch": 0.042934782608695654, "grad_norm": 2.3837692737579346, "learning_rate": 1.5399999999999998e-05, "loss": 4.3684, "step": 158 }, { "epoch": 0.043206521739130435, "grad_norm": 2.3704850673675537, "learning_rate": 1.55e-05, "loss": 4.1658, "step": 159 }, { "epoch": 0.043478260869565216, "grad_norm": 3.8034274578094482, "learning_rate": 1.56e-05, "loss": 4.0919, "step": 160 }, { "epoch": 0.04375, "grad_norm": 1.2763932943344116, "learning_rate": 1.57e-05, "loss": 3.8752, "step": 161 }, { "epoch": 0.044021739130434785, "grad_norm": 1.7155462503433228, "learning_rate": 1.5799999999999998e-05, "loss": 4.317, "step": 162 }, { "epoch": 0.044293478260869566, "grad_norm": 2.481580972671509, "learning_rate": 1.59e-05, "loss": 4.1314, "step": 163 }, { "epoch": 0.04456521739130435, "grad_norm": 2.3853437900543213, "learning_rate": 1.6e-05, "loss": 3.6408, "step": 164 }, { "epoch": 0.04483695652173913, "grad_norm": 1.7092500925064087, "learning_rate": 1.61e-05, "loss": 3.833, "step": 165 }, { "epoch": 0.045108695652173916, "grad_norm": 2.1334292888641357, "learning_rate": 1.62e-05, "loss": 4.5951, "step": 166 }, { "epoch": 0.0453804347826087, "grad_norm": 1.352221965789795, "learning_rate": 1.63e-05, "loss": 3.804, "step": 167 }, { "epoch": 0.04565217391304348, "grad_norm": 2.178356409072876, "learning_rate": 1.64e-05, "loss": 3.7594, "step": 168 }, { "epoch": 0.04592391304347826, "grad_norm": 2.9925286769866943, "learning_rate": 1.65e-05, "loss": 3.8819, "step": 169 }, { "epoch": 0.04619565217391304, "grad_norm": 3.0277819633483887, "learning_rate": 1.66e-05, "loss": 4.4896, "step": 170 }, { "epoch": 0.04646739130434783, "grad_norm": 1.459825873374939, "learning_rate": 1.67e-05, "loss": 4.1701, "step": 171 }, { "epoch": 0.04673913043478261, "grad_norm": 1.6663670539855957, "learning_rate": 1.6800000000000002e-05, "loss": 3.6651, "step": 172 }, { "epoch": 0.04701086956521739, "grad_norm": 1.7755922079086304, "learning_rate": 1.69e-05, "loss": 3.3178, "step": 173 }, { "epoch": 0.04728260869565217, "grad_norm": 1.4884556531906128, "learning_rate": 1.7e-05, "loss": 3.6467, "step": 174 }, { "epoch": 0.04755434782608696, "grad_norm": 2.701298713684082, "learning_rate": 1.71e-05, "loss": 4.773, "step": 175 }, { "epoch": 0.04782608695652174, "grad_norm": 3.169405937194824, "learning_rate": 1.72e-05, "loss": 4.1844, "step": 176 }, { "epoch": 0.04809782608695652, "grad_norm": 2.4385275840759277, "learning_rate": 1.73e-05, "loss": 3.8691, "step": 177 }, { "epoch": 0.0483695652173913, "grad_norm": 3.446741819381714, "learning_rate": 1.74e-05, "loss": 3.6733, "step": 178 }, { "epoch": 0.048641304347826084, "grad_norm": 4.095767974853516, "learning_rate": 1.7500000000000002e-05, "loss": 3.927, "step": 179 }, { "epoch": 0.04891304347826087, "grad_norm": 3.42309308052063, "learning_rate": 1.76e-05, "loss": 4.152, "step": 180 }, { "epoch": 0.04918478260869565, "grad_norm": 1.8505162000656128, "learning_rate": 1.77e-05, "loss": 3.5489, "step": 181 }, { "epoch": 0.049456521739130434, "grad_norm": 3.3934953212738037, "learning_rate": 1.7800000000000002e-05, "loss": 3.9161, "step": 182 }, { "epoch": 0.049728260869565215, "grad_norm": 3.2479214668273926, "learning_rate": 1.79e-05, "loss": 3.3177, "step": 183 }, { "epoch": 0.05, "grad_norm": 4.14975643157959, "learning_rate": 1.8e-05, "loss": 4.0072, "step": 184 }, { "epoch": 0.050271739130434784, "grad_norm": 4.540154457092285, "learning_rate": 1.8100000000000003e-05, "loss": 3.878, "step": 185 }, { "epoch": 0.050543478260869565, "grad_norm": 2.0964109897613525, "learning_rate": 1.8200000000000002e-05, "loss": 3.6273, "step": 186 }, { "epoch": 0.050815217391304346, "grad_norm": 2.119013547897339, "learning_rate": 1.83e-05, "loss": 3.5527, "step": 187 }, { "epoch": 0.051086956521739134, "grad_norm": 4.674685478210449, "learning_rate": 1.84e-05, "loss": 4.1131, "step": 188 }, { "epoch": 0.051358695652173915, "grad_norm": 7.127579689025879, "learning_rate": 1.8500000000000002e-05, "loss": 4.0545, "step": 189 }, { "epoch": 0.051630434782608696, "grad_norm": 3.285233497619629, "learning_rate": 1.86e-05, "loss": 3.6065, "step": 190 }, { "epoch": 0.05190217391304348, "grad_norm": 1.405059576034546, "learning_rate": 1.87e-05, "loss": 3.4746, "step": 191 }, { "epoch": 0.05217391304347826, "grad_norm": 3.002393960952759, "learning_rate": 1.8800000000000003e-05, "loss": 3.9168, "step": 192 }, { "epoch": 0.052445652173913046, "grad_norm": 2.253774404525757, "learning_rate": 1.8900000000000002e-05, "loss": 3.0605, "step": 193 }, { "epoch": 0.05271739130434783, "grad_norm": 3.4486026763916016, "learning_rate": 1.9e-05, "loss": 3.6955, "step": 194 }, { "epoch": 0.05298913043478261, "grad_norm": 2.6076107025146484, "learning_rate": 1.9100000000000003e-05, "loss": 4.7168, "step": 195 }, { "epoch": 0.05326086956521739, "grad_norm": 2.173379421234131, "learning_rate": 1.9200000000000003e-05, "loss": 3.9563, "step": 196 }, { "epoch": 0.05353260869565218, "grad_norm": 1.8571799993515015, "learning_rate": 1.9299999999999998e-05, "loss": 3.531, "step": 197 }, { "epoch": 0.05380434782608696, "grad_norm": 2.940171957015991, "learning_rate": 1.9399999999999997e-05, "loss": 3.9534, "step": 198 }, { "epoch": 0.05407608695652174, "grad_norm": 1.6403425931930542, "learning_rate": 1.95e-05, "loss": 3.7979, "step": 199 }, { "epoch": 0.05434782608695652, "grad_norm": 1.9086182117462158, "learning_rate": 1.96e-05, "loss": 3.3714, "step": 200 }, { "epoch": 0.0546195652173913, "grad_norm": 2.42441463470459, "learning_rate": 1.9699999999999998e-05, "loss": 3.6953, "step": 201 }, { "epoch": 0.05489130434782609, "grad_norm": 1.7879725694656372, "learning_rate": 1.98e-05, "loss": 3.424, "step": 202 }, { "epoch": 0.05516304347826087, "grad_norm": 1.951125144958496, "learning_rate": 1.99e-05, "loss": 4.0087, "step": 203 }, { "epoch": 0.05543478260869565, "grad_norm": 1.8883895874023438, "learning_rate": 1.9999999999999998e-05, "loss": 3.4857, "step": 204 }, { "epoch": 0.05570652173913043, "grad_norm": 2.1979916095733643, "learning_rate": 2.01e-05, "loss": 3.8898, "step": 205 }, { "epoch": 0.05597826086956522, "grad_norm": 1.6748876571655273, "learning_rate": 2.02e-05, "loss": 3.5928, "step": 206 }, { "epoch": 0.05625, "grad_norm": 4.129452705383301, "learning_rate": 2.03e-05, "loss": 4.2804, "step": 207 }, { "epoch": 0.05652173913043478, "grad_norm": 2.1873629093170166, "learning_rate": 2.04e-05, "loss": 3.5242, "step": 208 }, { "epoch": 0.05679347826086956, "grad_norm": 3.064663887023926, "learning_rate": 2.05e-05, "loss": 3.379, "step": 209 }, { "epoch": 0.057065217391304345, "grad_norm": 2.8077127933502197, "learning_rate": 2.06e-05, "loss": 3.4032, "step": 210 }, { "epoch": 0.05733695652173913, "grad_norm": 2.902247905731201, "learning_rate": 2.07e-05, "loss": 3.9496, "step": 211 }, { "epoch": 0.057608695652173914, "grad_norm": 2.3877575397491455, "learning_rate": 2.08e-05, "loss": 3.5266, "step": 212 }, { "epoch": 0.057880434782608695, "grad_norm": 4.157260417938232, "learning_rate": 2.09e-05, "loss": 4.0795, "step": 213 }, { "epoch": 0.058152173913043476, "grad_norm": 4.117393970489502, "learning_rate": 2.1e-05, "loss": 3.3584, "step": 214 }, { "epoch": 0.058423913043478264, "grad_norm": 3.1118719577789307, "learning_rate": 2.11e-05, "loss": 3.5, "step": 215 }, { "epoch": 0.058695652173913045, "grad_norm": 3.3953042030334473, "learning_rate": 2.12e-05, "loss": 3.6948, "step": 216 }, { "epoch": 0.058967391304347826, "grad_norm": 4.036997318267822, "learning_rate": 2.13e-05, "loss": 3.6871, "step": 217 }, { "epoch": 0.05923913043478261, "grad_norm": 4.334664821624756, "learning_rate": 2.1400000000000002e-05, "loss": 4.0403, "step": 218 }, { "epoch": 0.059510869565217395, "grad_norm": 3.2473344802856445, "learning_rate": 2.15e-05, "loss": 2.8208, "step": 219 }, { "epoch": 0.059782608695652176, "grad_norm": 5.333037376403809, "learning_rate": 2.16e-05, "loss": 3.4915, "step": 220 }, { "epoch": 0.06005434782608696, "grad_norm": 4.594522953033447, "learning_rate": 2.1700000000000002e-05, "loss": 3.773, "step": 221 }, { "epoch": 0.06032608695652174, "grad_norm": 2.8913493156433105, "learning_rate": 2.18e-05, "loss": 3.5039, "step": 222 }, { "epoch": 0.06059782608695652, "grad_norm": 3.1354737281799316, "learning_rate": 2.19e-05, "loss": 3.5823, "step": 223 }, { "epoch": 0.06086956521739131, "grad_norm": 4.392479419708252, "learning_rate": 2.2e-05, "loss": 3.8093, "step": 224 }, { "epoch": 0.06114130434782609, "grad_norm": 3.4441139698028564, "learning_rate": 2.2100000000000002e-05, "loss": 3.9232, "step": 225 }, { "epoch": 0.06141304347826087, "grad_norm": 2.799862861633301, "learning_rate": 2.22e-05, "loss": 3.9281, "step": 226 }, { "epoch": 0.06168478260869565, "grad_norm": 2.910454750061035, "learning_rate": 2.23e-05, "loss": 3.635, "step": 227 }, { "epoch": 0.06195652173913044, "grad_norm": 3.9778056144714355, "learning_rate": 2.2400000000000002e-05, "loss": 3.5988, "step": 228 }, { "epoch": 0.06222826086956522, "grad_norm": 3.500877857208252, "learning_rate": 2.25e-05, "loss": 3.2595, "step": 229 }, { "epoch": 0.0625, "grad_norm": 3.3998258113861084, "learning_rate": 2.26e-05, "loss": 3.9116, "step": 230 }, { "epoch": 0.06277173913043478, "grad_norm": 2.585266351699829, "learning_rate": 2.2700000000000003e-05, "loss": 3.6087, "step": 231 }, { "epoch": 0.06304347826086956, "grad_norm": 2.313265085220337, "learning_rate": 2.2800000000000002e-05, "loss": 3.8507, "step": 232 }, { "epoch": 0.06331521739130434, "grad_norm": 6.233187198638916, "learning_rate": 2.29e-05, "loss": 4.6271, "step": 233 }, { "epoch": 0.06358695652173912, "grad_norm": 4.387447357177734, "learning_rate": 2.3000000000000003e-05, "loss": 3.7921, "step": 234 }, { "epoch": 0.06385869565217392, "grad_norm": 3.573700189590454, "learning_rate": 2.3100000000000002e-05, "loss": 3.6567, "step": 235 }, { "epoch": 0.0641304347826087, "grad_norm": 1.9834128618240356, "learning_rate": 2.32e-05, "loss": 3.715, "step": 236 }, { "epoch": 0.06440217391304348, "grad_norm": 3.0220143795013428, "learning_rate": 2.33e-05, "loss": 4.302, "step": 237 }, { "epoch": 0.06467391304347826, "grad_norm": 5.85452938079834, "learning_rate": 2.3400000000000003e-05, "loss": 4.3755, "step": 238 }, { "epoch": 0.06494565217391304, "grad_norm": 6.263114929199219, "learning_rate": 2.3500000000000002e-05, "loss": 4.0045, "step": 239 }, { "epoch": 0.06521739130434782, "grad_norm": 4.6512322425842285, "learning_rate": 2.3599999999999998e-05, "loss": 4.202, "step": 240 }, { "epoch": 0.0654891304347826, "grad_norm": 1.589857578277588, "learning_rate": 2.37e-05, "loss": 3.0368, "step": 241 }, { "epoch": 0.06576086956521739, "grad_norm": 3.0660102367401123, "learning_rate": 2.38e-05, "loss": 3.959, "step": 242 }, { "epoch": 0.06603260869565217, "grad_norm": 3.7164723873138428, "learning_rate": 2.3899999999999998e-05, "loss": 3.5374, "step": 243 }, { "epoch": 0.06630434782608696, "grad_norm": 3.772892713546753, "learning_rate": 2.4e-05, "loss": 3.3464, "step": 244 }, { "epoch": 0.06657608695652174, "grad_norm": 3.2034425735473633, "learning_rate": 2.41e-05, "loss": 3.3077, "step": 245 }, { "epoch": 0.06684782608695652, "grad_norm": 4.294682502746582, "learning_rate": 2.42e-05, "loss": 3.813, "step": 246 }, { "epoch": 0.0671195652173913, "grad_norm": 3.805955171585083, "learning_rate": 2.43e-05, "loss": 3.6854, "step": 247 }, { "epoch": 0.06739130434782609, "grad_norm": 4.587597370147705, "learning_rate": 2.44e-05, "loss": 3.3058, "step": 248 }, { "epoch": 0.06766304347826087, "grad_norm": 5.513325214385986, "learning_rate": 2.45e-05, "loss": 3.6714, "step": 249 }, { "epoch": 0.06793478260869565, "grad_norm": 5.820981025695801, "learning_rate": 2.4599999999999998e-05, "loss": 3.6253, "step": 250 }, { "epoch": 0.06820652173913043, "grad_norm": 4.236602306365967, "learning_rate": 2.47e-05, "loss": 3.605, "step": 251 }, { "epoch": 0.06847826086956521, "grad_norm": 3.524596929550171, "learning_rate": 2.48e-05, "loss": 3.6746, "step": 252 }, { "epoch": 0.06875, "grad_norm": 4.912044525146484, "learning_rate": 2.49e-05, "loss": 3.6485, "step": 253 }, { "epoch": 0.06902173913043479, "grad_norm": 5.730941295623779, "learning_rate": 2.5e-05, "loss": 3.7662, "step": 254 }, { "epoch": 0.06929347826086957, "grad_norm": 7.801654815673828, "learning_rate": 2.51e-05, "loss": 4.8335, "step": 255 }, { "epoch": 0.06956521739130435, "grad_norm": 4.116402626037598, "learning_rate": 2.52e-05, "loss": 3.9456, "step": 256 }, { "epoch": 0.06983695652173913, "grad_norm": 6.192750930786133, "learning_rate": 2.5300000000000002e-05, "loss": 4.0458, "step": 257 }, { "epoch": 0.07010869565217391, "grad_norm": 4.565896511077881, "learning_rate": 2.54e-05, "loss": 3.3915, "step": 258 }, { "epoch": 0.07038043478260869, "grad_norm": 4.502013206481934, "learning_rate": 2.55e-05, "loss": 3.4612, "step": 259 }, { "epoch": 0.07065217391304347, "grad_norm": 3.8033788204193115, "learning_rate": 2.5600000000000002e-05, "loss": 2.9393, "step": 260 }, { "epoch": 0.07092391304347827, "grad_norm": 4.34993839263916, "learning_rate": 2.57e-05, "loss": 3.6575, "step": 261 }, { "epoch": 0.07119565217391305, "grad_norm": 3.9064600467681885, "learning_rate": 2.58e-05, "loss": 3.6109, "step": 262 }, { "epoch": 0.07146739130434783, "grad_norm": 3.737865447998047, "learning_rate": 2.59e-05, "loss": 3.7551, "step": 263 }, { "epoch": 0.07173913043478261, "grad_norm": 4.92514181137085, "learning_rate": 2.6000000000000002e-05, "loss": 3.9133, "step": 264 }, { "epoch": 0.07201086956521739, "grad_norm": 4.838226318359375, "learning_rate": 2.61e-05, "loss": 3.4457, "step": 265 }, { "epoch": 0.07228260869565217, "grad_norm": 5.465858459472656, "learning_rate": 2.62e-05, "loss": 3.3492, "step": 266 }, { "epoch": 0.07255434782608695, "grad_norm": 4.452864646911621, "learning_rate": 2.6300000000000002e-05, "loss": 4.0503, "step": 267 }, { "epoch": 0.07282608695652174, "grad_norm": 2.657294750213623, "learning_rate": 2.64e-05, "loss": 3.5952, "step": 268 }, { "epoch": 0.07309782608695652, "grad_norm": 3.738781213760376, "learning_rate": 2.65e-05, "loss": 3.2922, "step": 269 }, { "epoch": 0.07336956521739131, "grad_norm": 4.5520524978637695, "learning_rate": 2.6600000000000003e-05, "loss": 3.4737, "step": 270 }, { "epoch": 0.07364130434782609, "grad_norm": 6.709564685821533, "learning_rate": 2.6700000000000002e-05, "loss": 3.8822, "step": 271 }, { "epoch": 0.07391304347826087, "grad_norm": 5.555806636810303, "learning_rate": 2.68e-05, "loss": 4.1245, "step": 272 }, { "epoch": 0.07418478260869565, "grad_norm": 3.99271821975708, "learning_rate": 2.69e-05, "loss": 3.535, "step": 273 }, { "epoch": 0.07445652173913044, "grad_norm": 3.0163462162017822, "learning_rate": 2.7000000000000002e-05, "loss": 3.4727, "step": 274 }, { "epoch": 0.07472826086956522, "grad_norm": 5.363269329071045, "learning_rate": 2.71e-05, "loss": 4.1682, "step": 275 }, { "epoch": 0.075, "grad_norm": 4.282686710357666, "learning_rate": 2.72e-05, "loss": 3.1355, "step": 276 }, { "epoch": 0.07527173913043478, "grad_norm": 5.983767509460449, "learning_rate": 2.7300000000000003e-05, "loss": 3.8695, "step": 277 }, { "epoch": 0.07554347826086956, "grad_norm": 5.345359802246094, "learning_rate": 2.7400000000000002e-05, "loss": 4.1882, "step": 278 }, { "epoch": 0.07581521739130435, "grad_norm": 2.5548171997070312, "learning_rate": 2.75e-05, "loss": 3.2401, "step": 279 }, { "epoch": 0.07608695652173914, "grad_norm": 3.790065050125122, "learning_rate": 2.7600000000000003e-05, "loss": 3.8057, "step": 280 }, { "epoch": 0.07635869565217392, "grad_norm": 4.037810325622559, "learning_rate": 2.7700000000000002e-05, "loss": 3.8397, "step": 281 }, { "epoch": 0.0766304347826087, "grad_norm": 4.571537017822266, "learning_rate": 2.78e-05, "loss": 3.1902, "step": 282 }, { "epoch": 0.07690217391304348, "grad_norm": 4.4513702392578125, "learning_rate": 2.79e-05, "loss": 3.6552, "step": 283 }, { "epoch": 0.07717391304347826, "grad_norm": 2.919421911239624, "learning_rate": 2.8e-05, "loss": 3.0449, "step": 284 }, { "epoch": 0.07744565217391304, "grad_norm": 3.9303693771362305, "learning_rate": 2.81e-05, "loss": 3.8186, "step": 285 }, { "epoch": 0.07771739130434782, "grad_norm": 3.8404746055603027, "learning_rate": 2.8199999999999998e-05, "loss": 3.2603, "step": 286 }, { "epoch": 0.0779891304347826, "grad_norm": 3.2469253540039062, "learning_rate": 2.83e-05, "loss": 2.917, "step": 287 }, { "epoch": 0.0782608695652174, "grad_norm": 2.8970375061035156, "learning_rate": 2.84e-05, "loss": 3.6446, "step": 288 }, { "epoch": 0.07853260869565218, "grad_norm": 4.029104232788086, "learning_rate": 2.8499999999999998e-05, "loss": 4.1754, "step": 289 }, { "epoch": 0.07880434782608696, "grad_norm": 3.967705726623535, "learning_rate": 2.86e-05, "loss": 3.3913, "step": 290 }, { "epoch": 0.07907608695652174, "grad_norm": 3.084944009780884, "learning_rate": 2.87e-05, "loss": 3.4598, "step": 291 }, { "epoch": 0.07934782608695652, "grad_norm": 3.2562334537506104, "learning_rate": 2.88e-05, "loss": 3.7326, "step": 292 }, { "epoch": 0.0796195652173913, "grad_norm": 3.9668397903442383, "learning_rate": 2.89e-05, "loss": 3.8312, "step": 293 }, { "epoch": 0.07989130434782608, "grad_norm": 3.7607946395874023, "learning_rate": 2.9e-05, "loss": 3.9309, "step": 294 }, { "epoch": 0.08016304347826086, "grad_norm": 4.704547882080078, "learning_rate": 2.91e-05, "loss": 3.9181, "step": 295 }, { "epoch": 0.08043478260869565, "grad_norm": 5.68735408782959, "learning_rate": 2.92e-05, "loss": 3.6405, "step": 296 }, { "epoch": 0.08070652173913044, "grad_norm": 4.839725971221924, "learning_rate": 2.93e-05, "loss": 3.6735, "step": 297 }, { "epoch": 0.08097826086956522, "grad_norm": 3.6157710552215576, "learning_rate": 2.94e-05, "loss": 4.1132, "step": 298 }, { "epoch": 0.08125, "grad_norm": 3.6725997924804688, "learning_rate": 2.95e-05, "loss": 3.9299, "step": 299 }, { "epoch": 0.08152173913043478, "grad_norm": 4.408240795135498, "learning_rate": 2.96e-05, "loss": 3.4141, "step": 300 }, { "epoch": 0.08179347826086956, "grad_norm": 3.8871631622314453, "learning_rate": 2.97e-05, "loss": 2.7057, "step": 301 }, { "epoch": 0.08206521739130435, "grad_norm": 5.148252010345459, "learning_rate": 2.98e-05, "loss": 3.0604, "step": 302 }, { "epoch": 0.08233695652173913, "grad_norm": 3.8631784915924072, "learning_rate": 2.9900000000000002e-05, "loss": 3.4581, "step": 303 }, { "epoch": 0.08260869565217391, "grad_norm": 2.67698073387146, "learning_rate": 3e-05, "loss": 3.1465, "step": 304 }, { "epoch": 0.08288043478260869, "grad_norm": 2.6791391372680664, "learning_rate": 3e-05, "loss": 2.8593, "step": 305 }, { "epoch": 0.08315217391304348, "grad_norm": 3.844003677368164, "learning_rate": 3e-05, "loss": 3.4371, "step": 306 }, { "epoch": 0.08342391304347826, "grad_norm": 5.624002933502197, "learning_rate": 3e-05, "loss": 3.8372, "step": 307 }, { "epoch": 0.08369565217391305, "grad_norm": 4.992491245269775, "learning_rate": 3e-05, "loss": 3.6231, "step": 308 }, { "epoch": 0.08396739130434783, "grad_norm": 4.717808246612549, "learning_rate": 3e-05, "loss": 3.9652, "step": 309 }, { "epoch": 0.08423913043478261, "grad_norm": 3.6492786407470703, "learning_rate": 3e-05, "loss": 3.7065, "step": 310 }, { "epoch": 0.08451086956521739, "grad_norm": 4.453199863433838, "learning_rate": 3e-05, "loss": 3.7317, "step": 311 }, { "epoch": 0.08478260869565217, "grad_norm": 5.009824275970459, "learning_rate": 3e-05, "loss": 3.7651, "step": 312 }, { "epoch": 0.08505434782608695, "grad_norm": 6.800715446472168, "learning_rate": 3e-05, "loss": 4.0778, "step": 313 }, { "epoch": 0.08532608695652173, "grad_norm": 6.225503444671631, "learning_rate": 3e-05, "loss": 4.3679, "step": 314 }, { "epoch": 0.08559782608695653, "grad_norm": 2.9776594638824463, "learning_rate": 3e-05, "loss": 3.1434, "step": 315 }, { "epoch": 0.08586956521739131, "grad_norm": 3.1179118156433105, "learning_rate": 3e-05, "loss": 3.5896, "step": 316 }, { "epoch": 0.08614130434782609, "grad_norm": 2.81669282913208, "learning_rate": 3e-05, "loss": 2.9054, "step": 317 }, { "epoch": 0.08641304347826087, "grad_norm": 3.894665002822876, "learning_rate": 3e-05, "loss": 3.2122, "step": 318 }, { "epoch": 0.08668478260869565, "grad_norm": 5.43796443939209, "learning_rate": 3e-05, "loss": 3.8039, "step": 319 }, { "epoch": 0.08695652173913043, "grad_norm": 3.9177141189575195, "learning_rate": 3e-05, "loss": 3.6393, "step": 320 }, { "epoch": 0.08722826086956521, "grad_norm": 3.0576508045196533, "learning_rate": 3e-05, "loss": 3.67, "step": 321 }, { "epoch": 0.0875, "grad_norm": 5.208736896514893, "learning_rate": 3e-05, "loss": 3.9719, "step": 322 }, { "epoch": 0.08777173913043479, "grad_norm": 4.839489936828613, "learning_rate": 3e-05, "loss": 3.7197, "step": 323 }, { "epoch": 0.08804347826086957, "grad_norm": 4.195175647735596, "learning_rate": 3e-05, "loss": 3.3963, "step": 324 }, { "epoch": 0.08831521739130435, "grad_norm": 3.565690279006958, "learning_rate": 3e-05, "loss": 3.177, "step": 325 }, { "epoch": 0.08858695652173913, "grad_norm": 2.6973624229431152, "learning_rate": 3e-05, "loss": 3.0416, "step": 326 }, { "epoch": 0.08885869565217391, "grad_norm": 2.620365619659424, "learning_rate": 3e-05, "loss": 3.3155, "step": 327 }, { "epoch": 0.0891304347826087, "grad_norm": 3.7461090087890625, "learning_rate": 3e-05, "loss": 3.6553, "step": 328 }, { "epoch": 0.08940217391304348, "grad_norm": 6.702042102813721, "learning_rate": 3e-05, "loss": 4.2574, "step": 329 }, { "epoch": 0.08967391304347826, "grad_norm": 5.888139247894287, "learning_rate": 3e-05, "loss": 4.148, "step": 330 }, { "epoch": 0.08994565217391304, "grad_norm": 4.737085342407227, "learning_rate": 3e-05, "loss": 3.8761, "step": 331 }, { "epoch": 0.09021739130434783, "grad_norm": 3.9981937408447266, "learning_rate": 3e-05, "loss": 4.4605, "step": 332 }, { "epoch": 0.09048913043478261, "grad_norm": 5.0993266105651855, "learning_rate": 3e-05, "loss": 4.095, "step": 333 }, { "epoch": 0.0907608695652174, "grad_norm": 4.450422763824463, "learning_rate": 3e-05, "loss": 2.8801, "step": 334 }, { "epoch": 0.09103260869565218, "grad_norm": 6.846070766448975, "learning_rate": 3e-05, "loss": 3.5946, "step": 335 }, { "epoch": 0.09130434782608696, "grad_norm": 5.887484550476074, "learning_rate": 3e-05, "loss": 3.4395, "step": 336 }, { "epoch": 0.09157608695652174, "grad_norm": 3.8128914833068848, "learning_rate": 3e-05, "loss": 3.0863, "step": 337 }, { "epoch": 0.09184782608695652, "grad_norm": 2.493584394454956, "learning_rate": 3e-05, "loss": 3.0949, "step": 338 }, { "epoch": 0.0921195652173913, "grad_norm": 3.304899215698242, "learning_rate": 3e-05, "loss": 3.3348, "step": 339 }, { "epoch": 0.09239130434782608, "grad_norm": 4.633394241333008, "learning_rate": 3e-05, "loss": 3.5233, "step": 340 }, { "epoch": 0.09266304347826088, "grad_norm": 5.405637264251709, "learning_rate": 3e-05, "loss": 3.776, "step": 341 }, { "epoch": 0.09293478260869566, "grad_norm": 5.505592346191406, "learning_rate": 3e-05, "loss": 3.9894, "step": 342 }, { "epoch": 0.09320652173913044, "grad_norm": 3.901547431945801, "learning_rate": 3e-05, "loss": 4.1865, "step": 343 }, { "epoch": 0.09347826086956522, "grad_norm": 3.3303725719451904, "learning_rate": 3e-05, "loss": 3.3656, "step": 344 }, { "epoch": 0.09375, "grad_norm": 5.3323493003845215, "learning_rate": 3e-05, "loss": 4.2369, "step": 345 }, { "epoch": 0.09402173913043478, "grad_norm": 5.352675914764404, "learning_rate": 3e-05, "loss": 3.7106, "step": 346 }, { "epoch": 0.09429347826086956, "grad_norm": 5.6517839431762695, "learning_rate": 3e-05, "loss": 3.4388, "step": 347 }, { "epoch": 0.09456521739130434, "grad_norm": 4.7645111083984375, "learning_rate": 3e-05, "loss": 3.4332, "step": 348 }, { "epoch": 0.09483695652173912, "grad_norm": 3.140958309173584, "learning_rate": 3e-05, "loss": 3.2139, "step": 349 }, { "epoch": 0.09510869565217392, "grad_norm": 2.692319393157959, "learning_rate": 3e-05, "loss": 3.3015, "step": 350 }, { "epoch": 0.0953804347826087, "grad_norm": 4.1542134284973145, "learning_rate": 3e-05, "loss": 3.29, "step": 351 }, { "epoch": 0.09565217391304348, "grad_norm": 4.946714401245117, "learning_rate": 3e-05, "loss": 3.2509, "step": 352 }, { "epoch": 0.09592391304347826, "grad_norm": 5.529204368591309, "learning_rate": 3e-05, "loss": 4.0578, "step": 353 }, { "epoch": 0.09619565217391304, "grad_norm": 3.837975025177002, "learning_rate": 3e-05, "loss": 3.7921, "step": 354 }, { "epoch": 0.09646739130434782, "grad_norm": 3.0146334171295166, "learning_rate": 3e-05, "loss": 4.0369, "step": 355 }, { "epoch": 0.0967391304347826, "grad_norm": 4.283068656921387, "learning_rate": 3e-05, "loss": 3.9506, "step": 356 }, { "epoch": 0.09701086956521739, "grad_norm": 4.30133581161499, "learning_rate": 3e-05, "loss": 3.4218, "step": 357 }, { "epoch": 0.09728260869565217, "grad_norm": 6.228667736053467, "learning_rate": 3e-05, "loss": 4.0054, "step": 358 }, { "epoch": 0.09755434782608696, "grad_norm": 3.927131414413452, "learning_rate": 3e-05, "loss": 2.9401, "step": 359 }, { "epoch": 0.09782608695652174, "grad_norm": 4.374425888061523, "learning_rate": 3e-05, "loss": 3.5279, "step": 360 }, { "epoch": 0.09809782608695652, "grad_norm": 2.559403419494629, "learning_rate": 3e-05, "loss": 2.8306, "step": 361 }, { "epoch": 0.0983695652173913, "grad_norm": 2.9556596279144287, "learning_rate": 3e-05, "loss": 3.4532, "step": 362 }, { "epoch": 0.09864130434782609, "grad_norm": 4.841480731964111, "learning_rate": 3e-05, "loss": 3.474, "step": 363 }, { "epoch": 0.09891304347826087, "grad_norm": 5.785233497619629, "learning_rate": 3e-05, "loss": 3.6698, "step": 364 }, { "epoch": 0.09918478260869565, "grad_norm": 5.180464267730713, "learning_rate": 3e-05, "loss": 3.4199, "step": 365 }, { "epoch": 0.09945652173913043, "grad_norm": 3.5438666343688965, "learning_rate": 3e-05, "loss": 3.5603, "step": 366 }, { "epoch": 0.09972826086956521, "grad_norm": 3.1015937328338623, "learning_rate": 3e-05, "loss": 2.9937, "step": 367 }, { "epoch": 0.1, "grad_norm": 4.174754619598389, "learning_rate": 3e-05, "loss": 3.5994, "step": 368 }, { "epoch": 0.10027173913043479, "grad_norm": 4.678737163543701, "learning_rate": 3e-05, "loss": 4.0816, "step": 369 }, { "epoch": 0.10054347826086957, "grad_norm": 3.884875774383545, "learning_rate": 3e-05, "loss": 2.8268, "step": 370 }, { "epoch": 0.10081521739130435, "grad_norm": 4.060309886932373, "learning_rate": 3e-05, "loss": 3.5038, "step": 371 }, { "epoch": 0.10108695652173913, "grad_norm": 4.1803812980651855, "learning_rate": 3e-05, "loss": 3.8122, "step": 372 }, { "epoch": 0.10135869565217391, "grad_norm": 3.1725456714630127, "learning_rate": 3e-05, "loss": 3.7945, "step": 373 }, { "epoch": 0.10163043478260869, "grad_norm": 3.1784756183624268, "learning_rate": 3e-05, "loss": 3.0676, "step": 374 }, { "epoch": 0.10190217391304347, "grad_norm": 4.812568187713623, "learning_rate": 3e-05, "loss": 3.593, "step": 375 }, { "epoch": 0.10217391304347827, "grad_norm": 5.944811820983887, "learning_rate": 3e-05, "loss": 3.843, "step": 376 }, { "epoch": 0.10244565217391305, "grad_norm": 4.13525915145874, "learning_rate": 3e-05, "loss": 3.9103, "step": 377 }, { "epoch": 0.10271739130434783, "grad_norm": 5.860477447509766, "learning_rate": 3e-05, "loss": 4.9384, "step": 378 }, { "epoch": 0.10298913043478261, "grad_norm": 3.2515783309936523, "learning_rate": 3e-05, "loss": 3.0586, "step": 379 }, { "epoch": 0.10326086956521739, "grad_norm": 4.5452494621276855, "learning_rate": 3e-05, "loss": 3.7046, "step": 380 }, { "epoch": 0.10353260869565217, "grad_norm": 3.8891849517822266, "learning_rate": 3e-05, "loss": 3.34, "step": 381 }, { "epoch": 0.10380434782608695, "grad_norm": 4.083732604980469, "learning_rate": 3e-05, "loss": 3.3112, "step": 382 }, { "epoch": 0.10407608695652174, "grad_norm": 3.438572406768799, "learning_rate": 3e-05, "loss": 2.9908, "step": 383 }, { "epoch": 0.10434782608695652, "grad_norm": 3.0750012397766113, "learning_rate": 3e-05, "loss": 3.3847, "step": 384 }, { "epoch": 0.10461956521739131, "grad_norm": 3.0833847522735596, "learning_rate": 3e-05, "loss": 3.2274, "step": 385 }, { "epoch": 0.10489130434782609, "grad_norm": 4.356601238250732, "learning_rate": 3e-05, "loss": 3.9247, "step": 386 }, { "epoch": 0.10516304347826087, "grad_norm": 4.894799709320068, "learning_rate": 3e-05, "loss": 3.6005, "step": 387 }, { "epoch": 0.10543478260869565, "grad_norm": 4.887790203094482, "learning_rate": 3e-05, "loss": 4.3258, "step": 388 }, { "epoch": 0.10570652173913044, "grad_norm": 4.861374855041504, "learning_rate": 3e-05, "loss": 3.6857, "step": 389 }, { "epoch": 0.10597826086956522, "grad_norm": 4.775892734527588, "learning_rate": 3e-05, "loss": 4.3695, "step": 390 }, { "epoch": 0.10625, "grad_norm": 4.2715582847595215, "learning_rate": 3e-05, "loss": 3.7657, "step": 391 }, { "epoch": 0.10652173913043478, "grad_norm": 3.8440310955047607, "learning_rate": 3e-05, "loss": 3.4856, "step": 392 }, { "epoch": 0.10679347826086956, "grad_norm": 4.285994529724121, "learning_rate": 3e-05, "loss": 3.4081, "step": 393 }, { "epoch": 0.10706521739130435, "grad_norm": 2.1318747997283936, "learning_rate": 3e-05, "loss": 2.9345, "step": 394 }, { "epoch": 0.10733695652173914, "grad_norm": 2.4292728900909424, "learning_rate": 3e-05, "loss": 3.091, "step": 395 }, { "epoch": 0.10760869565217392, "grad_norm": 2.3496248722076416, "learning_rate": 3e-05, "loss": 3.1994, "step": 396 }, { "epoch": 0.1078804347826087, "grad_norm": 5.604576110839844, "learning_rate": 3e-05, "loss": 4.6392, "step": 397 }, { "epoch": 0.10815217391304348, "grad_norm": 5.448110580444336, "learning_rate": 3e-05, "loss": 4.4596, "step": 398 }, { "epoch": 0.10842391304347826, "grad_norm": 3.2921388149261475, "learning_rate": 3e-05, "loss": 3.5248, "step": 399 }, { "epoch": 0.10869565217391304, "grad_norm": 3.1395716667175293, "learning_rate": 3e-05, "loss": 4.1658, "step": 400 }, { "epoch": 0.10896739130434782, "grad_norm": 2.3313381671905518, "learning_rate": 3e-05, "loss": 3.7592, "step": 401 }, { "epoch": 0.1092391304347826, "grad_norm": 2.9619717597961426, "learning_rate": 3e-05, "loss": 3.5212, "step": 402 }, { "epoch": 0.1095108695652174, "grad_norm": 3.2616071701049805, "learning_rate": 3e-05, "loss": 3.2052, "step": 403 }, { "epoch": 0.10978260869565218, "grad_norm": 4.842064380645752, "learning_rate": 3e-05, "loss": 3.907, "step": 404 }, { "epoch": 0.11005434782608696, "grad_norm": 3.6140100955963135, "learning_rate": 3e-05, "loss": 4.061, "step": 405 }, { "epoch": 0.11032608695652174, "grad_norm": 1.4534454345703125, "learning_rate": 3e-05, "loss": 3.0512, "step": 406 }, { "epoch": 0.11059782608695652, "grad_norm": 2.4747161865234375, "learning_rate": 3e-05, "loss": 3.7008, "step": 407 }, { "epoch": 0.1108695652173913, "grad_norm": 3.0130977630615234, "learning_rate": 3e-05, "loss": 3.9765, "step": 408 }, { "epoch": 0.11114130434782608, "grad_norm": 2.256558656692505, "learning_rate": 3e-05, "loss": 3.2158, "step": 409 }, { "epoch": 0.11141304347826086, "grad_norm": 4.283812522888184, "learning_rate": 3e-05, "loss": 4.4811, "step": 410 }, { "epoch": 0.11168478260869565, "grad_norm": 2.566425323486328, "learning_rate": 3e-05, "loss": 3.9403, "step": 411 }, { "epoch": 0.11195652173913044, "grad_norm": 2.4623820781707764, "learning_rate": 3e-05, "loss": 3.2361, "step": 412 }, { "epoch": 0.11222826086956522, "grad_norm": 2.934722661972046, "learning_rate": 3e-05, "loss": 3.0554, "step": 413 }, { "epoch": 0.1125, "grad_norm": 5.524335861206055, "learning_rate": 3e-05, "loss": 4.0103, "step": 414 }, { "epoch": 0.11277173913043478, "grad_norm": 3.285830020904541, "learning_rate": 3e-05, "loss": 3.6652, "step": 415 }, { "epoch": 0.11304347826086956, "grad_norm": 2.5553982257843018, "learning_rate": 3e-05, "loss": 3.8115, "step": 416 }, { "epoch": 0.11331521739130435, "grad_norm": 1.8953039646148682, "learning_rate": 3e-05, "loss": 3.6769, "step": 417 }, { "epoch": 0.11358695652173913, "grad_norm": 2.4014382362365723, "learning_rate": 3e-05, "loss": 3.6343, "step": 418 }, { "epoch": 0.11385869565217391, "grad_norm": 3.2362847328186035, "learning_rate": 3e-05, "loss": 3.722, "step": 419 }, { "epoch": 0.11413043478260869, "grad_norm": 2.525406837463379, "learning_rate": 3e-05, "loss": 3.2201, "step": 420 }, { "epoch": 0.11440217391304348, "grad_norm": 2.492753028869629, "learning_rate": 3e-05, "loss": 3.3864, "step": 421 }, { "epoch": 0.11467391304347826, "grad_norm": 1.6389284133911133, "learning_rate": 3e-05, "loss": 3.1505, "step": 422 }, { "epoch": 0.11494565217391305, "grad_norm": 2.721895933151245, "learning_rate": 3e-05, "loss": 4.02, "step": 423 }, { "epoch": 0.11521739130434783, "grad_norm": 2.1239542961120605, "learning_rate": 3e-05, "loss": 3.3395, "step": 424 }, { "epoch": 0.11548913043478261, "grad_norm": 2.3277628421783447, "learning_rate": 3e-05, "loss": 4.3061, "step": 425 }, { "epoch": 0.11576086956521739, "grad_norm": 1.7965201139450073, "learning_rate": 3e-05, "loss": 3.8778, "step": 426 }, { "epoch": 0.11603260869565217, "grad_norm": 2.9034996032714844, "learning_rate": 3e-05, "loss": 3.8348, "step": 427 }, { "epoch": 0.11630434782608695, "grad_norm": 2.0713295936584473, "learning_rate": 3e-05, "loss": 3.6901, "step": 428 }, { "epoch": 0.11657608695652173, "grad_norm": 2.612128257751465, "learning_rate": 3e-05, "loss": 3.1308, "step": 429 }, { "epoch": 0.11684782608695653, "grad_norm": 3.535771369934082, "learning_rate": 3e-05, "loss": 3.7629, "step": 430 }, { "epoch": 0.11711956521739131, "grad_norm": 2.6418893337249756, "learning_rate": 3e-05, "loss": 3.351, "step": 431 }, { "epoch": 0.11739130434782609, "grad_norm": 1.9122227430343628, "learning_rate": 3e-05, "loss": 2.9162, "step": 432 }, { "epoch": 0.11766304347826087, "grad_norm": 3.12231707572937, "learning_rate": 3e-05, "loss": 3.6215, "step": 433 }, { "epoch": 0.11793478260869565, "grad_norm": 2.5034008026123047, "learning_rate": 3e-05, "loss": 3.6675, "step": 434 }, { "epoch": 0.11820652173913043, "grad_norm": 2.69254732131958, "learning_rate": 3e-05, "loss": 3.8847, "step": 435 }, { "epoch": 0.11847826086956521, "grad_norm": 2.026611566543579, "learning_rate": 3e-05, "loss": 3.9927, "step": 436 }, { "epoch": 0.11875, "grad_norm": 3.05465030670166, "learning_rate": 3e-05, "loss": 3.4363, "step": 437 }, { "epoch": 0.11902173913043479, "grad_norm": 2.800114631652832, "learning_rate": 3e-05, "loss": 4.0161, "step": 438 }, { "epoch": 0.11929347826086957, "grad_norm": 2.452852725982666, "learning_rate": 3e-05, "loss": 3.8825, "step": 439 }, { "epoch": 0.11956521739130435, "grad_norm": 2.425046920776367, "learning_rate": 3e-05, "loss": 3.8342, "step": 440 }, { "epoch": 0.11983695652173913, "grad_norm": 1.6898943185806274, "learning_rate": 3e-05, "loss": 3.4458, "step": 441 }, { "epoch": 0.12010869565217391, "grad_norm": 2.265521764755249, "learning_rate": 3e-05, "loss": 3.498, "step": 442 }, { "epoch": 0.1203804347826087, "grad_norm": 3.3182766437530518, "learning_rate": 3e-05, "loss": 4.0802, "step": 443 }, { "epoch": 0.12065217391304348, "grad_norm": 3.441931962966919, "learning_rate": 3e-05, "loss": 3.7694, "step": 444 }, { "epoch": 0.12092391304347826, "grad_norm": 2.1982293128967285, "learning_rate": 3e-05, "loss": 3.3313, "step": 445 }, { "epoch": 0.12119565217391304, "grad_norm": 1.986126184463501, "learning_rate": 3e-05, "loss": 3.3758, "step": 446 }, { "epoch": 0.12146739130434783, "grad_norm": 1.6694939136505127, "learning_rate": 3e-05, "loss": 3.1291, "step": 447 }, { "epoch": 0.12173913043478261, "grad_norm": 3.418254852294922, "learning_rate": 3e-05, "loss": 3.9384, "step": 448 }, { "epoch": 0.1220108695652174, "grad_norm": 2.4987125396728516, "learning_rate": 3e-05, "loss": 4.0415, "step": 449 }, { "epoch": 0.12228260869565218, "grad_norm": 2.4183948040008545, "learning_rate": 3e-05, "loss": 3.4816, "step": 450 }, { "epoch": 0.12255434782608696, "grad_norm": 2.753169059753418, "learning_rate": 3e-05, "loss": 3.3146, "step": 451 }, { "epoch": 0.12282608695652174, "grad_norm": 2.2018933296203613, "learning_rate": 3e-05, "loss": 3.5041, "step": 452 }, { "epoch": 0.12309782608695652, "grad_norm": 1.910427451133728, "learning_rate": 3e-05, "loss": 3.1185, "step": 453 }, { "epoch": 0.1233695652173913, "grad_norm": 2.2100231647491455, "learning_rate": 3e-05, "loss": 3.1726, "step": 454 }, { "epoch": 0.12364130434782608, "grad_norm": 2.6990854740142822, "learning_rate": 3e-05, "loss": 3.7998, "step": 455 }, { "epoch": 0.12391304347826088, "grad_norm": 3.1403119564056396, "learning_rate": 3e-05, "loss": 3.3689, "step": 456 }, { "epoch": 0.12418478260869566, "grad_norm": 1.9951990842819214, "learning_rate": 3e-05, "loss": 3.0753, "step": 457 }, { "epoch": 0.12445652173913044, "grad_norm": 2.348395347595215, "learning_rate": 3e-05, "loss": 3.8548, "step": 458 }, { "epoch": 0.12472826086956522, "grad_norm": 2.324969530105591, "learning_rate": 3e-05, "loss": 3.3703, "step": 459 }, { "epoch": 0.125, "grad_norm": 2.897873640060425, "learning_rate": 3e-05, "loss": 3.4589, "step": 460 }, { "epoch": 0.1252717391304348, "grad_norm": 2.81925892829895, "learning_rate": 3e-05, "loss": 3.4818, "step": 461 }, { "epoch": 0.12554347826086956, "grad_norm": 4.320977687835693, "learning_rate": 3e-05, "loss": 4.3366, "step": 462 }, { "epoch": 0.12581521739130436, "grad_norm": 3.2348084449768066, "learning_rate": 3e-05, "loss": 4.1263, "step": 463 }, { "epoch": 0.12608695652173912, "grad_norm": 2.6263833045959473, "learning_rate": 3e-05, "loss": 2.9206, "step": 464 }, { "epoch": 0.12635869565217392, "grad_norm": 4.796882629394531, "learning_rate": 3e-05, "loss": 3.8013, "step": 465 }, { "epoch": 0.1266304347826087, "grad_norm": 5.0804619789123535, "learning_rate": 3e-05, "loss": 3.8332, "step": 466 }, { "epoch": 0.12690217391304348, "grad_norm": 3.0131075382232666, "learning_rate": 3e-05, "loss": 3.5667, "step": 467 }, { "epoch": 0.12717391304347825, "grad_norm": 1.6071690320968628, "learning_rate": 3e-05, "loss": 2.8863, "step": 468 }, { "epoch": 0.12744565217391304, "grad_norm": 2.211411476135254, "learning_rate": 3e-05, "loss": 3.2423, "step": 469 }, { "epoch": 0.12771739130434784, "grad_norm": 2.4000439643859863, "learning_rate": 3e-05, "loss": 3.3725, "step": 470 }, { "epoch": 0.1279891304347826, "grad_norm": 2.421668529510498, "learning_rate": 3e-05, "loss": 3.6009, "step": 471 }, { "epoch": 0.1282608695652174, "grad_norm": 1.906244158744812, "learning_rate": 3e-05, "loss": 3.4958, "step": 472 }, { "epoch": 0.12853260869565217, "grad_norm": 1.7820132970809937, "learning_rate": 3e-05, "loss": 3.0321, "step": 473 }, { "epoch": 0.12880434782608696, "grad_norm": 2.929561138153076, "learning_rate": 3e-05, "loss": 3.788, "step": 474 }, { "epoch": 0.12907608695652173, "grad_norm": 2.740144968032837, "learning_rate": 3e-05, "loss": 3.6582, "step": 475 }, { "epoch": 0.12934782608695652, "grad_norm": 2.153079032897949, "learning_rate": 3e-05, "loss": 3.6739, "step": 476 }, { "epoch": 0.1296195652173913, "grad_norm": 1.366892695426941, "learning_rate": 3e-05, "loss": 3.5163, "step": 477 }, { "epoch": 0.1298913043478261, "grad_norm": 2.3881099224090576, "learning_rate": 3e-05, "loss": 3.7265, "step": 478 }, { "epoch": 0.13016304347826088, "grad_norm": 1.72128164768219, "learning_rate": 3e-05, "loss": 2.9679, "step": 479 }, { "epoch": 0.13043478260869565, "grad_norm": 1.8024574518203735, "learning_rate": 3e-05, "loss": 3.6759, "step": 480 }, { "epoch": 0.13070652173913044, "grad_norm": 2.3349437713623047, "learning_rate": 3e-05, "loss": 3.9231, "step": 481 }, { "epoch": 0.1309782608695652, "grad_norm": 2.60579776763916, "learning_rate": 3e-05, "loss": 3.5484, "step": 482 }, { "epoch": 0.13125, "grad_norm": 1.823648452758789, "learning_rate": 3e-05, "loss": 3.409, "step": 483 }, { "epoch": 0.13152173913043477, "grad_norm": 1.939772367477417, "learning_rate": 3e-05, "loss": 3.7185, "step": 484 }, { "epoch": 0.13179347826086957, "grad_norm": 1.7059251070022583, "learning_rate": 3e-05, "loss": 3.2223, "step": 485 }, { "epoch": 0.13206521739130433, "grad_norm": 2.1309432983398438, "learning_rate": 3e-05, "loss": 3.5143, "step": 486 }, { "epoch": 0.13233695652173913, "grad_norm": 2.5169553756713867, "learning_rate": 3e-05, "loss": 3.9375, "step": 487 }, { "epoch": 0.13260869565217392, "grad_norm": 1.6783947944641113, "learning_rate": 3e-05, "loss": 3.0069, "step": 488 }, { "epoch": 0.1328804347826087, "grad_norm": 2.5426528453826904, "learning_rate": 3e-05, "loss": 3.4032, "step": 489 }, { "epoch": 0.1331521739130435, "grad_norm": 2.8521511554718018, "learning_rate": 3e-05, "loss": 3.1236, "step": 490 }, { "epoch": 0.13342391304347825, "grad_norm": 2.7477147579193115, "learning_rate": 3e-05, "loss": 3.0802, "step": 491 }, { "epoch": 0.13369565217391305, "grad_norm": 2.785457134246826, "learning_rate": 3e-05, "loss": 3.8152, "step": 492 }, { "epoch": 0.13396739130434782, "grad_norm": 3.6600492000579834, "learning_rate": 3e-05, "loss": 3.6564, "step": 493 }, { "epoch": 0.1342391304347826, "grad_norm": 4.777431488037109, "learning_rate": 3e-05, "loss": 3.7449, "step": 494 }, { "epoch": 0.13451086956521738, "grad_norm": 4.060612201690674, "learning_rate": 3e-05, "loss": 4.1282, "step": 495 }, { "epoch": 0.13478260869565217, "grad_norm": 4.2701592445373535, "learning_rate": 3e-05, "loss": 4.6185, "step": 496 }, { "epoch": 0.13505434782608697, "grad_norm": 3.9511125087738037, "learning_rate": 3e-05, "loss": 3.2463, "step": 497 }, { "epoch": 0.13532608695652174, "grad_norm": 3.8475091457366943, "learning_rate": 3e-05, "loss": 3.0689, "step": 498 }, { "epoch": 0.13559782608695653, "grad_norm": 5.655093669891357, "learning_rate": 3e-05, "loss": 4.024, "step": 499 }, { "epoch": 0.1358695652173913, "grad_norm": 3.6835744380950928, "learning_rate": 3e-05, "loss": 3.1939, "step": 500 }, { "epoch": 0.1361413043478261, "grad_norm": 3.2142796516418457, "learning_rate": 3e-05, "loss": 4.0483, "step": 501 }, { "epoch": 0.13641304347826086, "grad_norm": 3.1059064865112305, "learning_rate": 3e-05, "loss": 3.8777, "step": 502 }, { "epoch": 0.13668478260869565, "grad_norm": 2.3298041820526123, "learning_rate": 3e-05, "loss": 3.2728, "step": 503 }, { "epoch": 0.13695652173913042, "grad_norm": 3.318361759185791, "learning_rate": 3e-05, "loss": 4.0421, "step": 504 }, { "epoch": 0.13722826086956522, "grad_norm": 2.8279385566711426, "learning_rate": 3e-05, "loss": 3.49, "step": 505 }, { "epoch": 0.1375, "grad_norm": 2.233304023742676, "learning_rate": 3e-05, "loss": 3.3737, "step": 506 }, { "epoch": 0.13777173913043478, "grad_norm": 2.943725824356079, "learning_rate": 3e-05, "loss": 3.5774, "step": 507 }, { "epoch": 0.13804347826086957, "grad_norm": 2.764371871948242, "learning_rate": 3e-05, "loss": 3.7041, "step": 508 }, { "epoch": 0.13831521739130434, "grad_norm": 2.647728443145752, "learning_rate": 3e-05, "loss": 3.0545, "step": 509 }, { "epoch": 0.13858695652173914, "grad_norm": 3.1184072494506836, "learning_rate": 3e-05, "loss": 3.2394, "step": 510 }, { "epoch": 0.1388586956521739, "grad_norm": 1.9611910581588745, "learning_rate": 3e-05, "loss": 3.4171, "step": 511 }, { "epoch": 0.1391304347826087, "grad_norm": 1.5122188329696655, "learning_rate": 3e-05, "loss": 3.1593, "step": 512 }, { "epoch": 0.13940217391304346, "grad_norm": 1.9166432619094849, "learning_rate": 3e-05, "loss": 3.21, "step": 513 }, { "epoch": 0.13967391304347826, "grad_norm": 2.1466691493988037, "learning_rate": 3e-05, "loss": 3.4333, "step": 514 }, { "epoch": 0.13994565217391305, "grad_norm": 2.217186689376831, "learning_rate": 3e-05, "loss": 3.3885, "step": 515 }, { "epoch": 0.14021739130434782, "grad_norm": 2.03560733795166, "learning_rate": 3e-05, "loss": 3.9298, "step": 516 }, { "epoch": 0.14048913043478262, "grad_norm": 2.214279890060425, "learning_rate": 3e-05, "loss": 3.2791, "step": 517 }, { "epoch": 0.14076086956521738, "grad_norm": 3.428405523300171, "learning_rate": 3e-05, "loss": 3.668, "step": 518 }, { "epoch": 0.14103260869565218, "grad_norm": 3.077150344848633, "learning_rate": 3e-05, "loss": 3.7187, "step": 519 }, { "epoch": 0.14130434782608695, "grad_norm": 1.3983957767486572, "learning_rate": 3e-05, "loss": 3.3129, "step": 520 }, { "epoch": 0.14157608695652174, "grad_norm": 2.0241615772247314, "learning_rate": 3e-05, "loss": 3.6655, "step": 521 }, { "epoch": 0.14184782608695654, "grad_norm": 2.20249080657959, "learning_rate": 3e-05, "loss": 3.2375, "step": 522 }, { "epoch": 0.1421195652173913, "grad_norm": 2.2980117797851562, "learning_rate": 3e-05, "loss": 3.7075, "step": 523 }, { "epoch": 0.1423913043478261, "grad_norm": 1.2177653312683105, "learning_rate": 3e-05, "loss": 3.2919, "step": 524 }, { "epoch": 0.14266304347826086, "grad_norm": 1.0807304382324219, "learning_rate": 3e-05, "loss": 3.2761, "step": 525 }, { "epoch": 0.14293478260869566, "grad_norm": 1.4918198585510254, "learning_rate": 3e-05, "loss": 3.6412, "step": 526 }, { "epoch": 0.14320652173913043, "grad_norm": 1.3037655353546143, "learning_rate": 3e-05, "loss": 3.4596, "step": 527 }, { "epoch": 0.14347826086956522, "grad_norm": 1.1929515600204468, "learning_rate": 3e-05, "loss": 3.4554, "step": 528 }, { "epoch": 0.14375, "grad_norm": 1.7983393669128418, "learning_rate": 3e-05, "loss": 3.0941, "step": 529 }, { "epoch": 0.14402173913043478, "grad_norm": 1.4893770217895508, "learning_rate": 3e-05, "loss": 3.688, "step": 530 }, { "epoch": 0.14429347826086958, "grad_norm": 1.4558385610580444, "learning_rate": 3e-05, "loss": 3.6866, "step": 531 }, { "epoch": 0.14456521739130435, "grad_norm": 1.4489972591400146, "learning_rate": 3e-05, "loss": 3.1238, "step": 532 }, { "epoch": 0.14483695652173914, "grad_norm": 1.603597640991211, "learning_rate": 3e-05, "loss": 2.9605, "step": 533 }, { "epoch": 0.1451086956521739, "grad_norm": 1.1347497701644897, "learning_rate": 3e-05, "loss": 3.2066, "step": 534 }, { "epoch": 0.1453804347826087, "grad_norm": 1.9727925062179565, "learning_rate": 3e-05, "loss": 3.6686, "step": 535 }, { "epoch": 0.14565217391304347, "grad_norm": 2.5670254230499268, "learning_rate": 3e-05, "loss": 3.3345, "step": 536 }, { "epoch": 0.14592391304347826, "grad_norm": 2.017517328262329, "learning_rate": 3e-05, "loss": 3.7092, "step": 537 }, { "epoch": 0.14619565217391303, "grad_norm": 2.1940996646881104, "learning_rate": 3e-05, "loss": 3.807, "step": 538 }, { "epoch": 0.14646739130434783, "grad_norm": 1.7565933465957642, "learning_rate": 3e-05, "loss": 3.6645, "step": 539 }, { "epoch": 0.14673913043478262, "grad_norm": 2.248276948928833, "learning_rate": 3e-05, "loss": 3.979, "step": 540 }, { "epoch": 0.1470108695652174, "grad_norm": 2.3259811401367188, "learning_rate": 3e-05, "loss": 3.3435, "step": 541 }, { "epoch": 0.14728260869565218, "grad_norm": 2.7300784587860107, "learning_rate": 3e-05, "loss": 3.7696, "step": 542 }, { "epoch": 0.14755434782608695, "grad_norm": 1.406214952468872, "learning_rate": 3e-05, "loss": 2.6787, "step": 543 }, { "epoch": 0.14782608695652175, "grad_norm": 2.6304099559783936, "learning_rate": 3e-05, "loss": 3.7273, "step": 544 }, { "epoch": 0.1480978260869565, "grad_norm": 4.7364912033081055, "learning_rate": 3e-05, "loss": 4.4759, "step": 545 }, { "epoch": 0.1483695652173913, "grad_norm": 5.363766193389893, "learning_rate": 3e-05, "loss": 4.1787, "step": 546 }, { "epoch": 0.14864130434782608, "grad_norm": 2.660940408706665, "learning_rate": 3e-05, "loss": 3.2534, "step": 547 }, { "epoch": 0.14891304347826087, "grad_norm": 3.186602830886841, "learning_rate": 3e-05, "loss": 3.5762, "step": 548 }, { "epoch": 0.14918478260869567, "grad_norm": 3.962216377258301, "learning_rate": 3e-05, "loss": 3.5002, "step": 549 }, { "epoch": 0.14945652173913043, "grad_norm": 5.419135093688965, "learning_rate": 3e-05, "loss": 3.6202, "step": 550 }, { "epoch": 0.14972826086956523, "grad_norm": 4.870014667510986, "learning_rate": 3e-05, "loss": 3.5467, "step": 551 }, { "epoch": 0.15, "grad_norm": 3.175389289855957, "learning_rate": 3e-05, "loss": 3.9385, "step": 552 }, { "epoch": 0.1502717391304348, "grad_norm": 2.783637046813965, "learning_rate": 3e-05, "loss": 3.8554, "step": 553 }, { "epoch": 0.15054347826086956, "grad_norm": 2.6941978931427, "learning_rate": 3e-05, "loss": 3.211, "step": 554 }, { "epoch": 0.15081521739130435, "grad_norm": 3.185209274291992, "learning_rate": 3e-05, "loss": 3.3476, "step": 555 }, { "epoch": 0.15108695652173912, "grad_norm": 2.922362804412842, "learning_rate": 3e-05, "loss": 3.0941, "step": 556 }, { "epoch": 0.1513586956521739, "grad_norm": 2.7833430767059326, "learning_rate": 3e-05, "loss": 3.3851, "step": 557 }, { "epoch": 0.1516304347826087, "grad_norm": 1.4858049154281616, "learning_rate": 3e-05, "loss": 2.7326, "step": 558 }, { "epoch": 0.15190217391304348, "grad_norm": 2.9602174758911133, "learning_rate": 3e-05, "loss": 3.2037, "step": 559 }, { "epoch": 0.15217391304347827, "grad_norm": 3.0462303161621094, "learning_rate": 3e-05, "loss": 3.2449, "step": 560 }, { "epoch": 0.15244565217391304, "grad_norm": 3.038954019546509, "learning_rate": 3e-05, "loss": 3.3796, "step": 561 }, { "epoch": 0.15271739130434783, "grad_norm": 1.572611689567566, "learning_rate": 3e-05, "loss": 3.1521, "step": 562 }, { "epoch": 0.1529891304347826, "grad_norm": 2.4640631675720215, "learning_rate": 3e-05, "loss": 3.9771, "step": 563 }, { "epoch": 0.1532608695652174, "grad_norm": 1.757712483406067, "learning_rate": 3e-05, "loss": 2.9893, "step": 564 }, { "epoch": 0.15353260869565216, "grad_norm": 2.606873035430908, "learning_rate": 3e-05, "loss": 3.702, "step": 565 }, { "epoch": 0.15380434782608696, "grad_norm": 1.677304983139038, "learning_rate": 3e-05, "loss": 3.2607, "step": 566 }, { "epoch": 0.15407608695652175, "grad_norm": 2.4729526042938232, "learning_rate": 3e-05, "loss": 3.9865, "step": 567 }, { "epoch": 0.15434782608695652, "grad_norm": 2.162191867828369, "learning_rate": 3e-05, "loss": 3.6349, "step": 568 }, { "epoch": 0.1546195652173913, "grad_norm": 3.027984380722046, "learning_rate": 3e-05, "loss": 3.8996, "step": 569 }, { "epoch": 0.15489130434782608, "grad_norm": 1.9820539951324463, "learning_rate": 3e-05, "loss": 3.3207, "step": 570 }, { "epoch": 0.15516304347826088, "grad_norm": 1.7880254983901978, "learning_rate": 3e-05, "loss": 3.6082, "step": 571 }, { "epoch": 0.15543478260869564, "grad_norm": 1.9113103151321411, "learning_rate": 3e-05, "loss": 3.2639, "step": 572 }, { "epoch": 0.15570652173913044, "grad_norm": 1.7418012619018555, "learning_rate": 3e-05, "loss": 3.4052, "step": 573 }, { "epoch": 0.1559782608695652, "grad_norm": 1.6183768510818481, "learning_rate": 3e-05, "loss": 3.8061, "step": 574 }, { "epoch": 0.15625, "grad_norm": 1.33986496925354, "learning_rate": 3e-05, "loss": 3.3393, "step": 575 }, { "epoch": 0.1565217391304348, "grad_norm": 1.4486945867538452, "learning_rate": 3e-05, "loss": 3.7729, "step": 576 }, { "epoch": 0.15679347826086956, "grad_norm": 1.7886120080947876, "learning_rate": 3e-05, "loss": 3.8346, "step": 577 }, { "epoch": 0.15706521739130436, "grad_norm": 1.2352441549301147, "learning_rate": 3e-05, "loss": 3.2718, "step": 578 }, { "epoch": 0.15733695652173912, "grad_norm": 1.8679944276809692, "learning_rate": 3e-05, "loss": 3.4247, "step": 579 }, { "epoch": 0.15760869565217392, "grad_norm": 1.689501404762268, "learning_rate": 3e-05, "loss": 4.0971, "step": 580 }, { "epoch": 0.1578804347826087, "grad_norm": 1.6779184341430664, "learning_rate": 3e-05, "loss": 3.8803, "step": 581 }, { "epoch": 0.15815217391304348, "grad_norm": 2.0760598182678223, "learning_rate": 3e-05, "loss": 3.5429, "step": 582 }, { "epoch": 0.15842391304347825, "grad_norm": 1.8308002948760986, "learning_rate": 3e-05, "loss": 3.4797, "step": 583 }, { "epoch": 0.15869565217391304, "grad_norm": 2.4416396617889404, "learning_rate": 3e-05, "loss": 4.2491, "step": 584 }, { "epoch": 0.15896739130434784, "grad_norm": 2.365509271621704, "learning_rate": 3e-05, "loss": 3.4862, "step": 585 }, { "epoch": 0.1592391304347826, "grad_norm": 4.06919002532959, "learning_rate": 3e-05, "loss": 4.0302, "step": 586 }, { "epoch": 0.1595108695652174, "grad_norm": 2.430508613586426, "learning_rate": 3e-05, "loss": 3.8575, "step": 587 }, { "epoch": 0.15978260869565217, "grad_norm": 2.2944397926330566, "learning_rate": 3e-05, "loss": 4.3022, "step": 588 }, { "epoch": 0.16005434782608696, "grad_norm": 4.096346855163574, "learning_rate": 3e-05, "loss": 4.0663, "step": 589 }, { "epoch": 0.16032608695652173, "grad_norm": 3.8741347789764404, "learning_rate": 3e-05, "loss": 3.3398, "step": 590 }, { "epoch": 0.16059782608695652, "grad_norm": 3.7039456367492676, "learning_rate": 3e-05, "loss": 3.1813, "step": 591 }, { "epoch": 0.1608695652173913, "grad_norm": 2.129674196243286, "learning_rate": 3e-05, "loss": 3.0778, "step": 592 }, { "epoch": 0.1611413043478261, "grad_norm": 2.404423952102661, "learning_rate": 3e-05, "loss": 3.7275, "step": 593 }, { "epoch": 0.16141304347826088, "grad_norm": 2.341529607772827, "learning_rate": 3e-05, "loss": 2.9763, "step": 594 }, { "epoch": 0.16168478260869565, "grad_norm": 4.927234649658203, "learning_rate": 3e-05, "loss": 4.1287, "step": 595 }, { "epoch": 0.16195652173913044, "grad_norm": 4.529046535491943, "learning_rate": 3e-05, "loss": 3.7712, "step": 596 }, { "epoch": 0.1622282608695652, "grad_norm": 4.280128479003906, "learning_rate": 3e-05, "loss": 3.7557, "step": 597 }, { "epoch": 0.1625, "grad_norm": 2.845654249191284, "learning_rate": 3e-05, "loss": 3.4817, "step": 598 }, { "epoch": 0.16277173913043477, "grad_norm": 2.5575075149536133, "learning_rate": 3e-05, "loss": 3.0366, "step": 599 }, { "epoch": 0.16304347826086957, "grad_norm": 3.475111722946167, "learning_rate": 3e-05, "loss": 3.2081, "step": 600 }, { "epoch": 0.16331521739130433, "grad_norm": 3.926103353500366, "learning_rate": 3e-05, "loss": 2.8455, "step": 601 }, { "epoch": 0.16358695652173913, "grad_norm": 3.796124219894409, "learning_rate": 3e-05, "loss": 3.4857, "step": 602 }, { "epoch": 0.16385869565217392, "grad_norm": 2.974625825881958, "learning_rate": 3e-05, "loss": 3.3609, "step": 603 }, { "epoch": 0.1641304347826087, "grad_norm": 1.9188250303268433, "learning_rate": 3e-05, "loss": 3.0467, "step": 604 }, { "epoch": 0.1644021739130435, "grad_norm": 2.922881841659546, "learning_rate": 3e-05, "loss": 3.969, "step": 605 }, { "epoch": 0.16467391304347825, "grad_norm": 2.9243340492248535, "learning_rate": 3e-05, "loss": 3.418, "step": 606 }, { "epoch": 0.16494565217391305, "grad_norm": 2.3607518672943115, "learning_rate": 3e-05, "loss": 3.3568, "step": 607 }, { "epoch": 0.16521739130434782, "grad_norm": 1.4109339714050293, "learning_rate": 3e-05, "loss": 3.1942, "step": 608 }, { "epoch": 0.1654891304347826, "grad_norm": 2.339900016784668, "learning_rate": 3e-05, "loss": 3.6621, "step": 609 }, { "epoch": 0.16576086956521738, "grad_norm": 3.0392491817474365, "learning_rate": 3e-05, "loss": 3.6538, "step": 610 }, { "epoch": 0.16603260869565217, "grad_norm": 2.5297353267669678, "learning_rate": 3e-05, "loss": 3.72, "step": 611 }, { "epoch": 0.16630434782608697, "grad_norm": 2.1570205688476562, "learning_rate": 3e-05, "loss": 3.1765, "step": 612 }, { "epoch": 0.16657608695652174, "grad_norm": 1.5125519037246704, "learning_rate": 3e-05, "loss": 3.5076, "step": 613 }, { "epoch": 0.16684782608695653, "grad_norm": 1.5017513036727905, "learning_rate": 3e-05, "loss": 3.5962, "step": 614 }, { "epoch": 0.1671195652173913, "grad_norm": 2.1280760765075684, "learning_rate": 3e-05, "loss": 3.4612, "step": 615 }, { "epoch": 0.1673913043478261, "grad_norm": 1.6511037349700928, "learning_rate": 3e-05, "loss": 3.3281, "step": 616 }, { "epoch": 0.16766304347826086, "grad_norm": 1.842786431312561, "learning_rate": 3e-05, "loss": 2.876, "step": 617 }, { "epoch": 0.16793478260869565, "grad_norm": 2.008168935775757, "learning_rate": 3e-05, "loss": 3.7421, "step": 618 }, { "epoch": 0.16820652173913042, "grad_norm": 2.965365409851074, "learning_rate": 3e-05, "loss": 3.9864, "step": 619 }, { "epoch": 0.16847826086956522, "grad_norm": 3.619001626968384, "learning_rate": 3e-05, "loss": 4.1181, "step": 620 }, { "epoch": 0.16875, "grad_norm": 3.3966426849365234, "learning_rate": 3e-05, "loss": 3.2355, "step": 621 }, { "epoch": 0.16902173913043478, "grad_norm": 2.4561333656311035, "learning_rate": 3e-05, "loss": 3.0077, "step": 622 }, { "epoch": 0.16929347826086957, "grad_norm": 3.7762365341186523, "learning_rate": 3e-05, "loss": 3.8245, "step": 623 }, { "epoch": 0.16956521739130434, "grad_norm": 4.286952495574951, "learning_rate": 3e-05, "loss": 3.4895, "step": 624 }, { "epoch": 0.16983695652173914, "grad_norm": 3.7213828563690186, "learning_rate": 3e-05, "loss": 2.9799, "step": 625 }, { "epoch": 0.1701086956521739, "grad_norm": 3.109027862548828, "learning_rate": 3e-05, "loss": 3.8767, "step": 626 }, { "epoch": 0.1703804347826087, "grad_norm": 2.4418489933013916, "learning_rate": 3e-05, "loss": 3.7539, "step": 627 }, { "epoch": 0.17065217391304346, "grad_norm": 2.3365840911865234, "learning_rate": 3e-05, "loss": 3.0382, "step": 628 }, { "epoch": 0.17092391304347826, "grad_norm": 2.649116039276123, "learning_rate": 3e-05, "loss": 3.0011, "step": 629 }, { "epoch": 0.17119565217391305, "grad_norm": 4.0966105461120605, "learning_rate": 3e-05, "loss": 3.5889, "step": 630 }, { "epoch": 0.17146739130434782, "grad_norm": 4.70590877532959, "learning_rate": 3e-05, "loss": 3.4292, "step": 631 }, { "epoch": 0.17173913043478262, "grad_norm": 4.246743202209473, "learning_rate": 3e-05, "loss": 3.3927, "step": 632 }, { "epoch": 0.17201086956521738, "grad_norm": 3.4085919857025146, "learning_rate": 3e-05, "loss": 3.5786, "step": 633 }, { "epoch": 0.17228260869565218, "grad_norm": 3.164703607559204, "learning_rate": 3e-05, "loss": 3.4236, "step": 634 }, { "epoch": 0.17255434782608695, "grad_norm": 4.914327144622803, "learning_rate": 3e-05, "loss": 4.2673, "step": 635 }, { "epoch": 0.17282608695652174, "grad_norm": 4.016328811645508, "learning_rate": 3e-05, "loss": 3.5927, "step": 636 }, { "epoch": 0.17309782608695654, "grad_norm": 3.3188321590423584, "learning_rate": 3e-05, "loss": 2.9687, "step": 637 }, { "epoch": 0.1733695652173913, "grad_norm": 3.4671499729156494, "learning_rate": 3e-05, "loss": 3.3144, "step": 638 }, { "epoch": 0.1736413043478261, "grad_norm": 2.273728609085083, "learning_rate": 3e-05, "loss": 3.2453, "step": 639 }, { "epoch": 0.17391304347826086, "grad_norm": 2.1210081577301025, "learning_rate": 3e-05, "loss": 3.0778, "step": 640 }, { "epoch": 0.17418478260869566, "grad_norm": 3.6642115116119385, "learning_rate": 3e-05, "loss": 3.7503, "step": 641 }, { "epoch": 0.17445652173913043, "grad_norm": 3.753002166748047, "learning_rate": 3e-05, "loss": 3.9504, "step": 642 }, { "epoch": 0.17472826086956522, "grad_norm": 2.683023691177368, "learning_rate": 3e-05, "loss": 3.5103, "step": 643 }, { "epoch": 0.175, "grad_norm": 2.7764129638671875, "learning_rate": 3e-05, "loss": 3.8473, "step": 644 }, { "epoch": 0.17527173913043478, "grad_norm": 2.340895652770996, "learning_rate": 3e-05, "loss": 3.3381, "step": 645 }, { "epoch": 0.17554347826086958, "grad_norm": 3.224985122680664, "learning_rate": 3e-05, "loss": 3.7218, "step": 646 }, { "epoch": 0.17581521739130435, "grad_norm": 4.798543930053711, "learning_rate": 3e-05, "loss": 4.2621, "step": 647 }, { "epoch": 0.17608695652173914, "grad_norm": 3.2494759559631348, "learning_rate": 3e-05, "loss": 3.4175, "step": 648 }, { "epoch": 0.1763586956521739, "grad_norm": 1.9543368816375732, "learning_rate": 3e-05, "loss": 2.9133, "step": 649 }, { "epoch": 0.1766304347826087, "grad_norm": 1.7534743547439575, "learning_rate": 3e-05, "loss": 3.2838, "step": 650 }, { "epoch": 0.17690217391304347, "grad_norm": 1.8143278360366821, "learning_rate": 3e-05, "loss": 3.015, "step": 651 }, { "epoch": 0.17717391304347826, "grad_norm": 3.4449610710144043, "learning_rate": 3e-05, "loss": 3.7868, "step": 652 }, { "epoch": 0.17744565217391303, "grad_norm": 2.250290632247925, "learning_rate": 3e-05, "loss": 3.6572, "step": 653 }, { "epoch": 0.17771739130434783, "grad_norm": 2.3306338787078857, "learning_rate": 3e-05, "loss": 3.6463, "step": 654 }, { "epoch": 0.17798913043478262, "grad_norm": 1.6108341217041016, "learning_rate": 3e-05, "loss": 3.3977, "step": 655 }, { "epoch": 0.1782608695652174, "grad_norm": 1.7349133491516113, "learning_rate": 3e-05, "loss": 3.5826, "step": 656 }, { "epoch": 0.17853260869565218, "grad_norm": 2.5024263858795166, "learning_rate": 3e-05, "loss": 4.2271, "step": 657 }, { "epoch": 0.17880434782608695, "grad_norm": 2.2973721027374268, "learning_rate": 3e-05, "loss": 3.8238, "step": 658 }, { "epoch": 0.17907608695652175, "grad_norm": 2.167056083679199, "learning_rate": 3e-05, "loss": 3.5719, "step": 659 }, { "epoch": 0.1793478260869565, "grad_norm": 1.8604989051818848, "learning_rate": 3e-05, "loss": 3.9091, "step": 660 }, { "epoch": 0.1796195652173913, "grad_norm": 1.7601895332336426, "learning_rate": 3e-05, "loss": 3.8582, "step": 661 }, { "epoch": 0.17989130434782608, "grad_norm": 1.5462661981582642, "learning_rate": 3e-05, "loss": 2.9604, "step": 662 }, { "epoch": 0.18016304347826087, "grad_norm": 1.9906418323516846, "learning_rate": 3e-05, "loss": 3.8877, "step": 663 }, { "epoch": 0.18043478260869567, "grad_norm": 2.1721136569976807, "learning_rate": 3e-05, "loss": 4.241, "step": 664 }, { "epoch": 0.18070652173913043, "grad_norm": 1.7870231866836548, "learning_rate": 3e-05, "loss": 3.0484, "step": 665 }, { "epoch": 0.18097826086956523, "grad_norm": 1.543925166130066, "learning_rate": 3e-05, "loss": 3.1115, "step": 666 }, { "epoch": 0.18125, "grad_norm": 1.9159563779830933, "learning_rate": 3e-05, "loss": 3.3634, "step": 667 }, { "epoch": 0.1815217391304348, "grad_norm": 2.8006904125213623, "learning_rate": 3e-05, "loss": 3.7725, "step": 668 }, { "epoch": 0.18179347826086956, "grad_norm": 2.8094773292541504, "learning_rate": 3e-05, "loss": 3.5421, "step": 669 }, { "epoch": 0.18206521739130435, "grad_norm": 2.0162885189056396, "learning_rate": 3e-05, "loss": 3.4805, "step": 670 }, { "epoch": 0.18233695652173912, "grad_norm": 2.175633430480957, "learning_rate": 3e-05, "loss": 3.3937, "step": 671 }, { "epoch": 0.1826086956521739, "grad_norm": 2.4267466068267822, "learning_rate": 3e-05, "loss": 3.1429, "step": 672 }, { "epoch": 0.1828804347826087, "grad_norm": 4.546562194824219, "learning_rate": 3e-05, "loss": 4.1168, "step": 673 }, { "epoch": 0.18315217391304348, "grad_norm": 3.3911025524139404, "learning_rate": 3e-05, "loss": 3.7966, "step": 674 }, { "epoch": 0.18342391304347827, "grad_norm": 1.7894442081451416, "learning_rate": 3e-05, "loss": 3.8492, "step": 675 }, { "epoch": 0.18369565217391304, "grad_norm": 2.7028391361236572, "learning_rate": 3e-05, "loss": 3.5555, "step": 676 }, { "epoch": 0.18396739130434783, "grad_norm": 3.052639961242676, "learning_rate": 3e-05, "loss": 3.9254, "step": 677 }, { "epoch": 0.1842391304347826, "grad_norm": 2.4930381774902344, "learning_rate": 3e-05, "loss": 3.4053, "step": 678 }, { "epoch": 0.1845108695652174, "grad_norm": 1.9123059511184692, "learning_rate": 3e-05, "loss": 3.5101, "step": 679 }, { "epoch": 0.18478260869565216, "grad_norm": 1.5995222330093384, "learning_rate": 3e-05, "loss": 3.1762, "step": 680 }, { "epoch": 0.18505434782608696, "grad_norm": 4.272251129150391, "learning_rate": 3e-05, "loss": 4.397, "step": 681 }, { "epoch": 0.18532608695652175, "grad_norm": 3.0316965579986572, "learning_rate": 3e-05, "loss": 3.8721, "step": 682 }, { "epoch": 0.18559782608695652, "grad_norm": 2.0125246047973633, "learning_rate": 3e-05, "loss": 3.3782, "step": 683 }, { "epoch": 0.1858695652173913, "grad_norm": 1.8079962730407715, "learning_rate": 3e-05, "loss": 3.4851, "step": 684 }, { "epoch": 0.18614130434782608, "grad_norm": 1.8277610540390015, "learning_rate": 3e-05, "loss": 2.8756, "step": 685 }, { "epoch": 0.18641304347826088, "grad_norm": 2.4962170124053955, "learning_rate": 3e-05, "loss": 4.2237, "step": 686 }, { "epoch": 0.18668478260869564, "grad_norm": 1.3268901109695435, "learning_rate": 3e-05, "loss": 3.1177, "step": 687 }, { "epoch": 0.18695652173913044, "grad_norm": 2.3978965282440186, "learning_rate": 3e-05, "loss": 3.583, "step": 688 }, { "epoch": 0.1872282608695652, "grad_norm": 1.2693482637405396, "learning_rate": 3e-05, "loss": 3.3973, "step": 689 }, { "epoch": 0.1875, "grad_norm": 2.0299911499023438, "learning_rate": 3e-05, "loss": 3.6627, "step": 690 }, { "epoch": 0.1877717391304348, "grad_norm": 1.2860682010650635, "learning_rate": 3e-05, "loss": 3.2345, "step": 691 }, { "epoch": 0.18804347826086956, "grad_norm": 2.264256477355957, "learning_rate": 3e-05, "loss": 4.2411, "step": 692 }, { "epoch": 0.18831521739130436, "grad_norm": 1.5267597436904907, "learning_rate": 3e-05, "loss": 3.1942, "step": 693 }, { "epoch": 0.18858695652173912, "grad_norm": 1.2684136629104614, "learning_rate": 3e-05, "loss": 3.192, "step": 694 }, { "epoch": 0.18885869565217392, "grad_norm": 1.6705724000930786, "learning_rate": 3e-05, "loss": 3.6082, "step": 695 }, { "epoch": 0.1891304347826087, "grad_norm": 2.0569725036621094, "learning_rate": 3e-05, "loss": 3.9019, "step": 696 }, { "epoch": 0.18940217391304348, "grad_norm": 2.03778076171875, "learning_rate": 3e-05, "loss": 4.0118, "step": 697 }, { "epoch": 0.18967391304347825, "grad_norm": 1.7906421422958374, "learning_rate": 3e-05, "loss": 3.7659, "step": 698 }, { "epoch": 0.18994565217391304, "grad_norm": 1.6202226877212524, "learning_rate": 3e-05, "loss": 3.3228, "step": 699 }, { "epoch": 0.19021739130434784, "grad_norm": 1.568548560142517, "learning_rate": 3e-05, "loss": 3.3992, "step": 700 }, { "epoch": 0.1904891304347826, "grad_norm": 1.3486888408660889, "learning_rate": 3e-05, "loss": 2.7886, "step": 701 }, { "epoch": 0.1907608695652174, "grad_norm": 2.013223171234131, "learning_rate": 3e-05, "loss": 3.365, "step": 702 }, { "epoch": 0.19103260869565217, "grad_norm": 1.7130358219146729, "learning_rate": 3e-05, "loss": 3.7516, "step": 703 }, { "epoch": 0.19130434782608696, "grad_norm": 1.7324903011322021, "learning_rate": 3e-05, "loss": 3.1275, "step": 704 }, { "epoch": 0.19157608695652173, "grad_norm": 1.7227182388305664, "learning_rate": 3e-05, "loss": 3.1415, "step": 705 }, { "epoch": 0.19184782608695652, "grad_norm": 1.3472079038619995, "learning_rate": 3e-05, "loss": 3.1949, "step": 706 }, { "epoch": 0.1921195652173913, "grad_norm": 1.1573243141174316, "learning_rate": 3e-05, "loss": 2.8314, "step": 707 }, { "epoch": 0.1923913043478261, "grad_norm": 1.6137231588363647, "learning_rate": 3e-05, "loss": 3.6169, "step": 708 }, { "epoch": 0.19266304347826088, "grad_norm": 1.4710439443588257, "learning_rate": 3e-05, "loss": 3.4097, "step": 709 }, { "epoch": 0.19293478260869565, "grad_norm": 2.347607374191284, "learning_rate": 3e-05, "loss": 4.1638, "step": 710 }, { "epoch": 0.19320652173913044, "grad_norm": 1.4299064874649048, "learning_rate": 3e-05, "loss": 3.811, "step": 711 }, { "epoch": 0.1934782608695652, "grad_norm": 1.5702636241912842, "learning_rate": 3e-05, "loss": 3.5396, "step": 712 }, { "epoch": 0.19375, "grad_norm": 1.8417882919311523, "learning_rate": 3e-05, "loss": 3.5613, "step": 713 }, { "epoch": 0.19402173913043477, "grad_norm": 1.4125804901123047, "learning_rate": 3e-05, "loss": 3.1808, "step": 714 }, { "epoch": 0.19429347826086957, "grad_norm": 1.3733165264129639, "learning_rate": 3e-05, "loss": 3.3127, "step": 715 }, { "epoch": 0.19456521739130433, "grad_norm": 2.1651833057403564, "learning_rate": 3e-05, "loss": 4.1732, "step": 716 }, { "epoch": 0.19483695652173913, "grad_norm": 1.4096370935440063, "learning_rate": 3e-05, "loss": 3.5565, "step": 717 }, { "epoch": 0.19510869565217392, "grad_norm": 1.7251945734024048, "learning_rate": 3e-05, "loss": 3.7342, "step": 718 }, { "epoch": 0.1953804347826087, "grad_norm": 1.643470287322998, "learning_rate": 3e-05, "loss": 2.9678, "step": 719 }, { "epoch": 0.1956521739130435, "grad_norm": 1.5827432870864868, "learning_rate": 3e-05, "loss": 3.1044, "step": 720 }, { "epoch": 0.19592391304347825, "grad_norm": 1.4647548198699951, "learning_rate": 3e-05, "loss": 2.8413, "step": 721 }, { "epoch": 0.19619565217391305, "grad_norm": 2.1561598777770996, "learning_rate": 3e-05, "loss": 3.4492, "step": 722 }, { "epoch": 0.19646739130434782, "grad_norm": 1.6970568895339966, "learning_rate": 3e-05, "loss": 3.1848, "step": 723 }, { "epoch": 0.1967391304347826, "grad_norm": 2.248117685317993, "learning_rate": 3e-05, "loss": 3.6296, "step": 724 }, { "epoch": 0.19701086956521738, "grad_norm": 3.168339252471924, "learning_rate": 3e-05, "loss": 4.291, "step": 725 }, { "epoch": 0.19728260869565217, "grad_norm": 3.802591323852539, "learning_rate": 3e-05, "loss": 3.9495, "step": 726 }, { "epoch": 0.19755434782608697, "grad_norm": 3.848099708557129, "learning_rate": 3e-05, "loss": 3.6353, "step": 727 }, { "epoch": 0.19782608695652174, "grad_norm": 3.534794569015503, "learning_rate": 3e-05, "loss": 3.9035, "step": 728 }, { "epoch": 0.19809782608695653, "grad_norm": 1.7165888547897339, "learning_rate": 3e-05, "loss": 3.3266, "step": 729 }, { "epoch": 0.1983695652173913, "grad_norm": 2.5262253284454346, "learning_rate": 3e-05, "loss": 3.6359, "step": 730 }, { "epoch": 0.1986413043478261, "grad_norm": 2.5349607467651367, "learning_rate": 3e-05, "loss": 3.129, "step": 731 }, { "epoch": 0.19891304347826086, "grad_norm": 5.680181503295898, "learning_rate": 3e-05, "loss": 4.8488, "step": 732 }, { "epoch": 0.19918478260869565, "grad_norm": 3.0405654907226562, "learning_rate": 3e-05, "loss": 3.3313, "step": 733 }, { "epoch": 0.19945652173913042, "grad_norm": 1.6890621185302734, "learning_rate": 3e-05, "loss": 3.2893, "step": 734 }, { "epoch": 0.19972826086956522, "grad_norm": 2.204172372817993, "learning_rate": 3e-05, "loss": 3.5067, "step": 735 }, { "epoch": 0.2, "grad_norm": 2.858398914337158, "learning_rate": 3e-05, "loss": 3.4036, "step": 736 }, { "epoch": 0.20027173913043478, "grad_norm": 2.4763529300689697, "learning_rate": 3e-05, "loss": 3.5171, "step": 737 }, { "epoch": 0.20054347826086957, "grad_norm": 2.957118034362793, "learning_rate": 3e-05, "loss": 4.1728, "step": 738 }, { "epoch": 0.20081521739130434, "grad_norm": 1.8309071063995361, "learning_rate": 3e-05, "loss": 2.9663, "step": 739 }, { "epoch": 0.20108695652173914, "grad_norm": 2.270033121109009, "learning_rate": 3e-05, "loss": 3.6115, "step": 740 }, { "epoch": 0.2013586956521739, "grad_norm": 2.1604442596435547, "learning_rate": 3e-05, "loss": 3.2511, "step": 741 }, { "epoch": 0.2016304347826087, "grad_norm": 3.5561349391937256, "learning_rate": 3e-05, "loss": 3.8886, "step": 742 }, { "epoch": 0.20190217391304346, "grad_norm": 2.510963201522827, "learning_rate": 3e-05, "loss": 3.4812, "step": 743 }, { "epoch": 0.20217391304347826, "grad_norm": 1.5206657648086548, "learning_rate": 3e-05, "loss": 3.0231, "step": 744 }, { "epoch": 0.20244565217391305, "grad_norm": 1.645155906677246, "learning_rate": 3e-05, "loss": 3.0951, "step": 745 }, { "epoch": 0.20271739130434782, "grad_norm": 2.075399160385132, "learning_rate": 3e-05, "loss": 3.4057, "step": 746 }, { "epoch": 0.20298913043478262, "grad_norm": 2.36306095123291, "learning_rate": 3e-05, "loss": 3.6449, "step": 747 }, { "epoch": 0.20326086956521738, "grad_norm": 1.9054253101348877, "learning_rate": 3e-05, "loss": 3.3983, "step": 748 }, { "epoch": 0.20353260869565218, "grad_norm": 1.7470569610595703, "learning_rate": 3e-05, "loss": 3.7903, "step": 749 }, { "epoch": 0.20380434782608695, "grad_norm": 1.6297615766525269, "learning_rate": 3e-05, "loss": 3.0304, "step": 750 }, { "epoch": 0.20407608695652174, "grad_norm": 3.2830233573913574, "learning_rate": 3e-05, "loss": 4.2781, "step": 751 }, { "epoch": 0.20434782608695654, "grad_norm": 1.361449122428894, "learning_rate": 3e-05, "loss": 3.2224, "step": 752 }, { "epoch": 0.2046195652173913, "grad_norm": 1.4894417524337769, "learning_rate": 3e-05, "loss": 3.262, "step": 753 }, { "epoch": 0.2048913043478261, "grad_norm": 1.7271361351013184, "learning_rate": 3e-05, "loss": 3.3017, "step": 754 }, { "epoch": 0.20516304347826086, "grad_norm": 2.033215284347534, "learning_rate": 3e-05, "loss": 2.9264, "step": 755 }, { "epoch": 0.20543478260869566, "grad_norm": 2.808903455734253, "learning_rate": 3e-05, "loss": 3.7223, "step": 756 }, { "epoch": 0.20570652173913043, "grad_norm": 1.450661540031433, "learning_rate": 3e-05, "loss": 4.0346, "step": 757 }, { "epoch": 0.20597826086956522, "grad_norm": 1.9796857833862305, "learning_rate": 3e-05, "loss": 3.2978, "step": 758 }, { "epoch": 0.20625, "grad_norm": 2.9408633708953857, "learning_rate": 3e-05, "loss": 3.7247, "step": 759 }, { "epoch": 0.20652173913043478, "grad_norm": 1.4558497667312622, "learning_rate": 3e-05, "loss": 3.0258, "step": 760 }, { "epoch": 0.20679347826086958, "grad_norm": 1.1229232549667358, "learning_rate": 3e-05, "loss": 2.9328, "step": 761 }, { "epoch": 0.20706521739130435, "grad_norm": 1.663891077041626, "learning_rate": 3e-05, "loss": 3.2935, "step": 762 }, { "epoch": 0.20733695652173914, "grad_norm": 1.8489412069320679, "learning_rate": 3e-05, "loss": 2.8728, "step": 763 }, { "epoch": 0.2076086956521739, "grad_norm": 1.359864592552185, "learning_rate": 3e-05, "loss": 3.6552, "step": 764 }, { "epoch": 0.2078804347826087, "grad_norm": 1.5933263301849365, "learning_rate": 3e-05, "loss": 3.2384, "step": 765 }, { "epoch": 0.20815217391304347, "grad_norm": 3.0126466751098633, "learning_rate": 3e-05, "loss": 3.4475, "step": 766 }, { "epoch": 0.20842391304347826, "grad_norm": 2.608242988586426, "learning_rate": 3e-05, "loss": 3.2701, "step": 767 }, { "epoch": 0.20869565217391303, "grad_norm": 2.7247979640960693, "learning_rate": 3e-05, "loss": 3.96, "step": 768 }, { "epoch": 0.20896739130434783, "grad_norm": 1.8706130981445312, "learning_rate": 3e-05, "loss": 3.0889, "step": 769 }, { "epoch": 0.20923913043478262, "grad_norm": 2.668093681335449, "learning_rate": 3e-05, "loss": 3.4351, "step": 770 }, { "epoch": 0.2095108695652174, "grad_norm": 3.017181396484375, "learning_rate": 3e-05, "loss": 3.2463, "step": 771 }, { "epoch": 0.20978260869565218, "grad_norm": 3.0812008380889893, "learning_rate": 3e-05, "loss": 3.3998, "step": 772 }, { "epoch": 0.21005434782608695, "grad_norm": 2.7510030269622803, "learning_rate": 3e-05, "loss": 4.2353, "step": 773 }, { "epoch": 0.21032608695652175, "grad_norm": 2.4355452060699463, "learning_rate": 3e-05, "loss": 3.8853, "step": 774 }, { "epoch": 0.2105978260869565, "grad_norm": 4.041244983673096, "learning_rate": 3e-05, "loss": 4.0771, "step": 775 }, { "epoch": 0.2108695652173913, "grad_norm": 3.403510332107544, "learning_rate": 3e-05, "loss": 3.6963, "step": 776 }, { "epoch": 0.21114130434782608, "grad_norm": 3.216482400894165, "learning_rate": 3e-05, "loss": 3.6608, "step": 777 }, { "epoch": 0.21141304347826087, "grad_norm": 2.5989348888397217, "learning_rate": 3e-05, "loss": 3.689, "step": 778 }, { "epoch": 0.21168478260869567, "grad_norm": 2.0239574909210205, "learning_rate": 3e-05, "loss": 3.681, "step": 779 }, { "epoch": 0.21195652173913043, "grad_norm": 3.3878705501556396, "learning_rate": 3e-05, "loss": 3.9301, "step": 780 }, { "epoch": 0.21222826086956523, "grad_norm": 4.106480598449707, "learning_rate": 3e-05, "loss": 3.2929, "step": 781 }, { "epoch": 0.2125, "grad_norm": 3.4818196296691895, "learning_rate": 3e-05, "loss": 2.8562, "step": 782 }, { "epoch": 0.2127717391304348, "grad_norm": 3.4106242656707764, "learning_rate": 3e-05, "loss": 3.6078, "step": 783 }, { "epoch": 0.21304347826086956, "grad_norm": 2.401315450668335, "learning_rate": 3e-05, "loss": 3.693, "step": 784 }, { "epoch": 0.21331521739130435, "grad_norm": 3.092789649963379, "learning_rate": 3e-05, "loss": 3.6907, "step": 785 }, { "epoch": 0.21358695652173912, "grad_norm": 2.8016254901885986, "learning_rate": 3e-05, "loss": 3.7602, "step": 786 }, { "epoch": 0.2138586956521739, "grad_norm": 2.281668186187744, "learning_rate": 3e-05, "loss": 2.9451, "step": 787 }, { "epoch": 0.2141304347826087, "grad_norm": 2.3392529487609863, "learning_rate": 3e-05, "loss": 3.1969, "step": 788 }, { "epoch": 0.21440217391304348, "grad_norm": 1.7437257766723633, "learning_rate": 3e-05, "loss": 3.2468, "step": 789 }, { "epoch": 0.21467391304347827, "grad_norm": 2.128056049346924, "learning_rate": 3e-05, "loss": 3.5476, "step": 790 }, { "epoch": 0.21494565217391304, "grad_norm": 2.1912710666656494, "learning_rate": 3e-05, "loss": 3.7689, "step": 791 }, { "epoch": 0.21521739130434783, "grad_norm": 2.5414609909057617, "learning_rate": 3e-05, "loss": 3.4014, "step": 792 }, { "epoch": 0.2154891304347826, "grad_norm": 2.6355223655700684, "learning_rate": 3e-05, "loss": 3.3787, "step": 793 }, { "epoch": 0.2157608695652174, "grad_norm": 2.1927592754364014, "learning_rate": 3e-05, "loss": 3.2381, "step": 794 }, { "epoch": 0.21603260869565216, "grad_norm": 1.7231847047805786, "learning_rate": 3e-05, "loss": 3.4627, "step": 795 }, { "epoch": 0.21630434782608696, "grad_norm": 1.8754600286483765, "learning_rate": 3e-05, "loss": 3.7409, "step": 796 }, { "epoch": 0.21657608695652175, "grad_norm": 1.7413159608840942, "learning_rate": 3e-05, "loss": 3.6012, "step": 797 }, { "epoch": 0.21684782608695652, "grad_norm": 1.9197379350662231, "learning_rate": 3e-05, "loss": 3.15, "step": 798 }, { "epoch": 0.2171195652173913, "grad_norm": 2.100706100463867, "learning_rate": 3e-05, "loss": 2.8748, "step": 799 }, { "epoch": 0.21739130434782608, "grad_norm": 2.7185757160186768, "learning_rate": 3e-05, "loss": 3.7859, "step": 800 }, { "epoch": 0.21766304347826088, "grad_norm": 3.241926908493042, "learning_rate": 3e-05, "loss": 3.7198, "step": 801 }, { "epoch": 0.21793478260869564, "grad_norm": 2.365151882171631, "learning_rate": 3e-05, "loss": 3.7607, "step": 802 }, { "epoch": 0.21820652173913044, "grad_norm": 2.0234482288360596, "learning_rate": 3e-05, "loss": 3.1843, "step": 803 }, { "epoch": 0.2184782608695652, "grad_norm": 3.084533929824829, "learning_rate": 3e-05, "loss": 4.1945, "step": 804 }, { "epoch": 0.21875, "grad_norm": 2.927222490310669, "learning_rate": 3e-05, "loss": 3.3531, "step": 805 }, { "epoch": 0.2190217391304348, "grad_norm": 2.6623260974884033, "learning_rate": 3e-05, "loss": 3.5994, "step": 806 }, { "epoch": 0.21929347826086956, "grad_norm": 2.8344593048095703, "learning_rate": 3e-05, "loss": 3.6001, "step": 807 }, { "epoch": 0.21956521739130436, "grad_norm": 1.8181391954421997, "learning_rate": 3e-05, "loss": 2.9434, "step": 808 }, { "epoch": 0.21983695652173912, "grad_norm": 2.103512763977051, "learning_rate": 3e-05, "loss": 3.8034, "step": 809 }, { "epoch": 0.22010869565217392, "grad_norm": 1.961503505706787, "learning_rate": 3e-05, "loss": 3.0883, "step": 810 }, { "epoch": 0.2203804347826087, "grad_norm": 2.3394432067871094, "learning_rate": 3e-05, "loss": 2.8568, "step": 811 }, { "epoch": 0.22065217391304348, "grad_norm": 2.6501801013946533, "learning_rate": 3e-05, "loss": 3.2469, "step": 812 }, { "epoch": 0.22092391304347825, "grad_norm": 2.7663819789886475, "learning_rate": 3e-05, "loss": 3.3066, "step": 813 }, { "epoch": 0.22119565217391304, "grad_norm": 2.4567878246307373, "learning_rate": 3e-05, "loss": 3.0416, "step": 814 }, { "epoch": 0.22146739130434784, "grad_norm": 2.378188371658325, "learning_rate": 3e-05, "loss": 3.6954, "step": 815 }, { "epoch": 0.2217391304347826, "grad_norm": 2.611842632293701, "learning_rate": 3e-05, "loss": 3.3243, "step": 816 }, { "epoch": 0.2220108695652174, "grad_norm": 2.1804463863372803, "learning_rate": 3e-05, "loss": 2.9858, "step": 817 }, { "epoch": 0.22228260869565217, "grad_norm": 2.03049373626709, "learning_rate": 3e-05, "loss": 2.787, "step": 818 }, { "epoch": 0.22255434782608696, "grad_norm": 1.8013122081756592, "learning_rate": 3e-05, "loss": 2.9299, "step": 819 }, { "epoch": 0.22282608695652173, "grad_norm": 2.039745569229126, "learning_rate": 3e-05, "loss": 3.2247, "step": 820 }, { "epoch": 0.22309782608695652, "grad_norm": 1.6249157190322876, "learning_rate": 3e-05, "loss": 2.8942, "step": 821 }, { "epoch": 0.2233695652173913, "grad_norm": 1.8277668952941895, "learning_rate": 3e-05, "loss": 3.5634, "step": 822 }, { "epoch": 0.2236413043478261, "grad_norm": 2.2749369144439697, "learning_rate": 3e-05, "loss": 3.6253, "step": 823 }, { "epoch": 0.22391304347826088, "grad_norm": 2.3847649097442627, "learning_rate": 3e-05, "loss": 3.3725, "step": 824 }, { "epoch": 0.22418478260869565, "grad_norm": 1.9465465545654297, "learning_rate": 3e-05, "loss": 3.4556, "step": 825 }, { "epoch": 0.22445652173913044, "grad_norm": 2.793731927871704, "learning_rate": 3e-05, "loss": 4.0801, "step": 826 }, { "epoch": 0.2247282608695652, "grad_norm": 2.605842113494873, "learning_rate": 3e-05, "loss": 4.0396, "step": 827 }, { "epoch": 0.225, "grad_norm": 2.200025796890259, "learning_rate": 3e-05, "loss": 3.7514, "step": 828 }, { "epoch": 0.22527173913043477, "grad_norm": 1.8318731784820557, "learning_rate": 3e-05, "loss": 3.3619, "step": 829 }, { "epoch": 0.22554347826086957, "grad_norm": 1.7559170722961426, "learning_rate": 3e-05, "loss": 3.3115, "step": 830 }, { "epoch": 0.22581521739130433, "grad_norm": 1.7032020092010498, "learning_rate": 3e-05, "loss": 3.113, "step": 831 }, { "epoch": 0.22608695652173913, "grad_norm": 1.8972278833389282, "learning_rate": 3e-05, "loss": 3.1965, "step": 832 }, { "epoch": 0.22635869565217392, "grad_norm": 1.9449539184570312, "learning_rate": 3e-05, "loss": 3.2422, "step": 833 }, { "epoch": 0.2266304347826087, "grad_norm": 1.6739164590835571, "learning_rate": 3e-05, "loss": 3.2255, "step": 834 }, { "epoch": 0.2269021739130435, "grad_norm": 1.7071807384490967, "learning_rate": 3e-05, "loss": 3.4032, "step": 835 }, { "epoch": 0.22717391304347825, "grad_norm": 3.5912764072418213, "learning_rate": 3e-05, "loss": 4.1757, "step": 836 }, { "epoch": 0.22744565217391305, "grad_norm": 1.969997763633728, "learning_rate": 3e-05, "loss": 3.1138, "step": 837 }, { "epoch": 0.22771739130434782, "grad_norm": 1.9781277179718018, "learning_rate": 3e-05, "loss": 3.5066, "step": 838 }, { "epoch": 0.2279891304347826, "grad_norm": 1.6115202903747559, "learning_rate": 3e-05, "loss": 2.9759, "step": 839 }, { "epoch": 0.22826086956521738, "grad_norm": 1.3473846912384033, "learning_rate": 3e-05, "loss": 3.0617, "step": 840 }, { "epoch": 0.22853260869565217, "grad_norm": 1.6643739938735962, "learning_rate": 3e-05, "loss": 3.4268, "step": 841 }, { "epoch": 0.22880434782608697, "grad_norm": 1.9249311685562134, "learning_rate": 3e-05, "loss": 3.6207, "step": 842 }, { "epoch": 0.22907608695652174, "grad_norm": 2.079634666442871, "learning_rate": 3e-05, "loss": 3.534, "step": 843 }, { "epoch": 0.22934782608695653, "grad_norm": 1.8367491960525513, "learning_rate": 3e-05, "loss": 3.9664, "step": 844 }, { "epoch": 0.2296195652173913, "grad_norm": 1.7043336629867554, "learning_rate": 3e-05, "loss": 3.4275, "step": 845 }, { "epoch": 0.2298913043478261, "grad_norm": 1.6460905075073242, "learning_rate": 3e-05, "loss": 3.1771, "step": 846 }, { "epoch": 0.23016304347826086, "grad_norm": 2.0371439456939697, "learning_rate": 3e-05, "loss": 3.3044, "step": 847 }, { "epoch": 0.23043478260869565, "grad_norm": 1.5710325241088867, "learning_rate": 3e-05, "loss": 3.0044, "step": 848 }, { "epoch": 0.23070652173913042, "grad_norm": 1.6866766214370728, "learning_rate": 3e-05, "loss": 2.9916, "step": 849 }, { "epoch": 0.23097826086956522, "grad_norm": 2.127194404602051, "learning_rate": 3e-05, "loss": 3.5753, "step": 850 }, { "epoch": 0.23125, "grad_norm": 2.0354788303375244, "learning_rate": 3e-05, "loss": 2.8992, "step": 851 }, { "epoch": 0.23152173913043478, "grad_norm": 1.5567448139190674, "learning_rate": 3e-05, "loss": 3.4486, "step": 852 }, { "epoch": 0.23179347826086957, "grad_norm": 1.3115332126617432, "learning_rate": 3e-05, "loss": 3.1354, "step": 853 }, { "epoch": 0.23206521739130434, "grad_norm": 1.3727189302444458, "learning_rate": 3e-05, "loss": 3.0521, "step": 854 }, { "epoch": 0.23233695652173914, "grad_norm": 1.3423751592636108, "learning_rate": 3e-05, "loss": 2.908, "step": 855 }, { "epoch": 0.2326086956521739, "grad_norm": 1.61993408203125, "learning_rate": 3e-05, "loss": 3.2873, "step": 856 }, { "epoch": 0.2328804347826087, "grad_norm": 1.8327138423919678, "learning_rate": 3e-05, "loss": 3.6789, "step": 857 }, { "epoch": 0.23315217391304346, "grad_norm": 1.626006007194519, "learning_rate": 3e-05, "loss": 3.0027, "step": 858 }, { "epoch": 0.23342391304347826, "grad_norm": 2.236844539642334, "learning_rate": 3e-05, "loss": 4.083, "step": 859 }, { "epoch": 0.23369565217391305, "grad_norm": 2.1321239471435547, "learning_rate": 3e-05, "loss": 3.455, "step": 860 }, { "epoch": 0.23396739130434782, "grad_norm": 2.1679024696350098, "learning_rate": 3e-05, "loss": 3.8644, "step": 861 }, { "epoch": 0.23423913043478262, "grad_norm": 1.9821666479110718, "learning_rate": 3e-05, "loss": 3.5024, "step": 862 }, { "epoch": 0.23451086956521738, "grad_norm": 2.8352460861206055, "learning_rate": 3e-05, "loss": 4.0082, "step": 863 }, { "epoch": 0.23478260869565218, "grad_norm": 2.2241246700286865, "learning_rate": 3e-05, "loss": 3.8031, "step": 864 }, { "epoch": 0.23505434782608695, "grad_norm": 2.1706037521362305, "learning_rate": 3e-05, "loss": 4.0421, "step": 865 }, { "epoch": 0.23532608695652174, "grad_norm": 1.695703148841858, "learning_rate": 3e-05, "loss": 2.9672, "step": 866 }, { "epoch": 0.23559782608695654, "grad_norm": 1.994674563407898, "learning_rate": 3e-05, "loss": 3.3449, "step": 867 }, { "epoch": 0.2358695652173913, "grad_norm": 2.692945957183838, "learning_rate": 3e-05, "loss": 4.1761, "step": 868 }, { "epoch": 0.2361413043478261, "grad_norm": 2.5395121574401855, "learning_rate": 3e-05, "loss": 3.921, "step": 869 }, { "epoch": 0.23641304347826086, "grad_norm": 1.991248607635498, "learning_rate": 3e-05, "loss": 2.9, "step": 870 }, { "epoch": 0.23668478260869566, "grad_norm": 2.2419328689575195, "learning_rate": 3e-05, "loss": 3.501, "step": 871 }, { "epoch": 0.23695652173913043, "grad_norm": 1.9551500082015991, "learning_rate": 3e-05, "loss": 3.8966, "step": 872 }, { "epoch": 0.23722826086956522, "grad_norm": 1.9755557775497437, "learning_rate": 3e-05, "loss": 3.7852, "step": 873 }, { "epoch": 0.2375, "grad_norm": 1.964316487312317, "learning_rate": 3e-05, "loss": 3.4841, "step": 874 }, { "epoch": 0.23777173913043478, "grad_norm": 1.667006015777588, "learning_rate": 3e-05, "loss": 3.3479, "step": 875 }, { "epoch": 0.23804347826086958, "grad_norm": 1.3252580165863037, "learning_rate": 3e-05, "loss": 3.1133, "step": 876 }, { "epoch": 0.23831521739130435, "grad_norm": 1.904585361480713, "learning_rate": 3e-05, "loss": 3.7604, "step": 877 }, { "epoch": 0.23858695652173914, "grad_norm": 1.581666350364685, "learning_rate": 3e-05, "loss": 3.3015, "step": 878 }, { "epoch": 0.2388586956521739, "grad_norm": 1.2517045736312866, "learning_rate": 3e-05, "loss": 3.2687, "step": 879 }, { "epoch": 0.2391304347826087, "grad_norm": 1.5604926347732544, "learning_rate": 3e-05, "loss": 3.0638, "step": 880 }, { "epoch": 0.23940217391304347, "grad_norm": 1.9475339651107788, "learning_rate": 3e-05, "loss": 3.8968, "step": 881 }, { "epoch": 0.23967391304347826, "grad_norm": 1.528891921043396, "learning_rate": 3e-05, "loss": 3.0379, "step": 882 }, { "epoch": 0.23994565217391303, "grad_norm": 1.838294506072998, "learning_rate": 3e-05, "loss": 3.7645, "step": 883 }, { "epoch": 0.24021739130434783, "grad_norm": 1.5220407247543335, "learning_rate": 3e-05, "loss": 3.9417, "step": 884 }, { "epoch": 0.24048913043478262, "grad_norm": 1.4615991115570068, "learning_rate": 3e-05, "loss": 3.4286, "step": 885 }, { "epoch": 0.2407608695652174, "grad_norm": 2.569244623184204, "learning_rate": 3e-05, "loss": 3.9001, "step": 886 }, { "epoch": 0.24103260869565218, "grad_norm": 2.3717269897460938, "learning_rate": 3e-05, "loss": 4.2076, "step": 887 }, { "epoch": 0.24130434782608695, "grad_norm": 1.8509140014648438, "learning_rate": 3e-05, "loss": 3.2101, "step": 888 }, { "epoch": 0.24157608695652175, "grad_norm": 2.0046916007995605, "learning_rate": 3e-05, "loss": 2.9609, "step": 889 }, { "epoch": 0.2418478260869565, "grad_norm": 2.2359912395477295, "learning_rate": 3e-05, "loss": 3.6416, "step": 890 }, { "epoch": 0.2421195652173913, "grad_norm": 1.8625085353851318, "learning_rate": 3e-05, "loss": 3.3007, "step": 891 }, { "epoch": 0.24239130434782608, "grad_norm": 2.0958662033081055, "learning_rate": 3e-05, "loss": 2.7472, "step": 892 }, { "epoch": 0.24266304347826087, "grad_norm": 2.5501413345336914, "learning_rate": 3e-05, "loss": 3.0367, "step": 893 }, { "epoch": 0.24293478260869567, "grad_norm": 3.020587921142578, "learning_rate": 3e-05, "loss": 3.7357, "step": 894 }, { "epoch": 0.24320652173913043, "grad_norm": 2.0165812969207764, "learning_rate": 3e-05, "loss": 3.6235, "step": 895 }, { "epoch": 0.24347826086956523, "grad_norm": 3.1172547340393066, "learning_rate": 3e-05, "loss": 4.1981, "step": 896 }, { "epoch": 0.24375, "grad_norm": 2.381237268447876, "learning_rate": 3e-05, "loss": 3.104, "step": 897 }, { "epoch": 0.2440217391304348, "grad_norm": 3.345118522644043, "learning_rate": 3e-05, "loss": 4.1412, "step": 898 }, { "epoch": 0.24429347826086956, "grad_norm": 2.8453972339630127, "learning_rate": 3e-05, "loss": 3.3134, "step": 899 }, { "epoch": 0.24456521739130435, "grad_norm": 1.9727962017059326, "learning_rate": 3e-05, "loss": 3.1409, "step": 900 }, { "epoch": 0.24483695652173912, "grad_norm": 1.3924411535263062, "learning_rate": 3e-05, "loss": 3.2174, "step": 901 }, { "epoch": 0.2451086956521739, "grad_norm": 1.8885339498519897, "learning_rate": 3e-05, "loss": 3.4122, "step": 902 }, { "epoch": 0.2453804347826087, "grad_norm": 2.3274450302124023, "learning_rate": 3e-05, "loss": 3.2905, "step": 903 }, { "epoch": 0.24565217391304348, "grad_norm": 1.877142310142517, "learning_rate": 3e-05, "loss": 3.4118, "step": 904 }, { "epoch": 0.24592391304347827, "grad_norm": 1.8108350038528442, "learning_rate": 3e-05, "loss": 3.206, "step": 905 }, { "epoch": 0.24619565217391304, "grad_norm": 1.6787278652191162, "learning_rate": 3e-05, "loss": 3.2236, "step": 906 }, { "epoch": 0.24646739130434783, "grad_norm": 2.007154703140259, "learning_rate": 3e-05, "loss": 4.1134, "step": 907 }, { "epoch": 0.2467391304347826, "grad_norm": 1.9029121398925781, "learning_rate": 3e-05, "loss": 3.2066, "step": 908 }, { "epoch": 0.2470108695652174, "grad_norm": 1.5358937978744507, "learning_rate": 3e-05, "loss": 3.3406, "step": 909 }, { "epoch": 0.24728260869565216, "grad_norm": 1.706233263015747, "learning_rate": 3e-05, "loss": 3.368, "step": 910 }, { "epoch": 0.24755434782608696, "grad_norm": 2.283039093017578, "learning_rate": 3e-05, "loss": 4.1815, "step": 911 }, { "epoch": 0.24782608695652175, "grad_norm": 1.6811976432800293, "learning_rate": 3e-05, "loss": 3.8339, "step": 912 }, { "epoch": 0.24809782608695652, "grad_norm": 1.6838414669036865, "learning_rate": 3e-05, "loss": 3.6558, "step": 913 }, { "epoch": 0.2483695652173913, "grad_norm": 1.451900839805603, "learning_rate": 3e-05, "loss": 3.2279, "step": 914 }, { "epoch": 0.24864130434782608, "grad_norm": 2.3503222465515137, "learning_rate": 3e-05, "loss": 3.9484, "step": 915 }, { "epoch": 0.24891304347826088, "grad_norm": 2.001974582672119, "learning_rate": 3e-05, "loss": 3.3028, "step": 916 }, { "epoch": 0.24918478260869564, "grad_norm": 1.5036247968673706, "learning_rate": 3e-05, "loss": 3.2459, "step": 917 }, { "epoch": 0.24945652173913044, "grad_norm": 2.033236026763916, "learning_rate": 3e-05, "loss": 3.4323, "step": 918 }, { "epoch": 0.2497282608695652, "grad_norm": 1.4043093919754028, "learning_rate": 3e-05, "loss": 3.0965, "step": 919 }, { "epoch": 0.25, "grad_norm": 1.203482985496521, "learning_rate": 3e-05, "loss": 3.1118, "step": 920 }, { "epoch": 0.25027173913043477, "grad_norm": 1.4571930170059204, "learning_rate": 3e-05, "loss": 3.351, "step": 921 }, { "epoch": 0.2505434782608696, "grad_norm": 0.9827572703361511, "learning_rate": 3e-05, "loss": 2.8122, "step": 922 }, { "epoch": 0.25081521739130436, "grad_norm": 1.520359754562378, "learning_rate": 3e-05, "loss": 3.6678, "step": 923 }, { "epoch": 0.2510869565217391, "grad_norm": 1.277106761932373, "learning_rate": 3e-05, "loss": 2.9916, "step": 924 }, { "epoch": 0.2513586956521739, "grad_norm": 1.7534111738204956, "learning_rate": 3e-05, "loss": 3.9959, "step": 925 }, { "epoch": 0.2516304347826087, "grad_norm": 1.031008005142212, "learning_rate": 3e-05, "loss": 2.9768, "step": 926 }, { "epoch": 0.2519021739130435, "grad_norm": 1.1608198881149292, "learning_rate": 3e-05, "loss": 3.456, "step": 927 }, { "epoch": 0.25217391304347825, "grad_norm": 1.4016661643981934, "learning_rate": 3e-05, "loss": 3.4423, "step": 928 }, { "epoch": 0.25244565217391307, "grad_norm": 1.1200820207595825, "learning_rate": 3e-05, "loss": 2.897, "step": 929 }, { "epoch": 0.25271739130434784, "grad_norm": 1.4193557500839233, "learning_rate": 3e-05, "loss": 3.7882, "step": 930 }, { "epoch": 0.2529891304347826, "grad_norm": 1.2631326913833618, "learning_rate": 3e-05, "loss": 3.8642, "step": 931 }, { "epoch": 0.2532608695652174, "grad_norm": 1.6645845174789429, "learning_rate": 3e-05, "loss": 3.7794, "step": 932 }, { "epoch": 0.2535326086956522, "grad_norm": 1.7016172409057617, "learning_rate": 3e-05, "loss": 3.3343, "step": 933 }, { "epoch": 0.25380434782608696, "grad_norm": 1.1917381286621094, "learning_rate": 3e-05, "loss": 3.3563, "step": 934 }, { "epoch": 0.25407608695652173, "grad_norm": 1.4421312808990479, "learning_rate": 3e-05, "loss": 3.6694, "step": 935 }, { "epoch": 0.2543478260869565, "grad_norm": 1.2393507957458496, "learning_rate": 3e-05, "loss": 2.8572, "step": 936 }, { "epoch": 0.2546195652173913, "grad_norm": 1.4665427207946777, "learning_rate": 3e-05, "loss": 3.5944, "step": 937 }, { "epoch": 0.2548913043478261, "grad_norm": 1.3151808977127075, "learning_rate": 3e-05, "loss": 3.1136, "step": 938 }, { "epoch": 0.25516304347826085, "grad_norm": 1.3733018636703491, "learning_rate": 3e-05, "loss": 3.1559, "step": 939 }, { "epoch": 0.2554347826086957, "grad_norm": 1.638494610786438, "learning_rate": 3e-05, "loss": 3.5616, "step": 940 }, { "epoch": 0.25570652173913044, "grad_norm": 1.5186861753463745, "learning_rate": 3e-05, "loss": 3.7529, "step": 941 }, { "epoch": 0.2559782608695652, "grad_norm": 1.7179397344589233, "learning_rate": 3e-05, "loss": 3.8634, "step": 942 }, { "epoch": 0.25625, "grad_norm": 1.9366356134414673, "learning_rate": 3e-05, "loss": 3.785, "step": 943 }, { "epoch": 0.2565217391304348, "grad_norm": 2.2738037109375, "learning_rate": 3e-05, "loss": 3.7442, "step": 944 }, { "epoch": 0.25679347826086957, "grad_norm": 2.67051100730896, "learning_rate": 3e-05, "loss": 3.8186, "step": 945 }, { "epoch": 0.25706521739130433, "grad_norm": 3.3741190433502197, "learning_rate": 3e-05, "loss": 4.2118, "step": 946 }, { "epoch": 0.25733695652173916, "grad_norm": 1.6262043714523315, "learning_rate": 3e-05, "loss": 3.0208, "step": 947 }, { "epoch": 0.2576086956521739, "grad_norm": 1.7342244386672974, "learning_rate": 3e-05, "loss": 3.1897, "step": 948 }, { "epoch": 0.2578804347826087, "grad_norm": 2.0924177169799805, "learning_rate": 3e-05, "loss": 2.7324, "step": 949 }, { "epoch": 0.25815217391304346, "grad_norm": 2.21530818939209, "learning_rate": 3e-05, "loss": 3.1098, "step": 950 }, { "epoch": 0.2584239130434783, "grad_norm": 1.642945647239685, "learning_rate": 3e-05, "loss": 3.58, "step": 951 }, { "epoch": 0.25869565217391305, "grad_norm": 1.4641486406326294, "learning_rate": 3e-05, "loss": 3.4771, "step": 952 }, { "epoch": 0.2589673913043478, "grad_norm": 1.4831671714782715, "learning_rate": 3e-05, "loss": 3.0967, "step": 953 }, { "epoch": 0.2592391304347826, "grad_norm": 1.605242133140564, "learning_rate": 3e-05, "loss": 3.3319, "step": 954 }, { "epoch": 0.2595108695652174, "grad_norm": 2.312819480895996, "learning_rate": 3e-05, "loss": 4.214, "step": 955 }, { "epoch": 0.2597826086956522, "grad_norm": 1.751943588256836, "learning_rate": 3e-05, "loss": 3.3116, "step": 956 }, { "epoch": 0.26005434782608694, "grad_norm": 1.951372504234314, "learning_rate": 3e-05, "loss": 3.4478, "step": 957 }, { "epoch": 0.26032608695652176, "grad_norm": 2.3981149196624756, "learning_rate": 3e-05, "loss": 3.4107, "step": 958 }, { "epoch": 0.26059782608695653, "grad_norm": 2.448025703430176, "learning_rate": 3e-05, "loss": 4.0686, "step": 959 }, { "epoch": 0.2608695652173913, "grad_norm": 2.1298911571502686, "learning_rate": 3e-05, "loss": 3.4676, "step": 960 }, { "epoch": 0.26114130434782606, "grad_norm": 1.9788066148757935, "learning_rate": 3e-05, "loss": 3.4239, "step": 961 }, { "epoch": 0.2614130434782609, "grad_norm": 1.6203289031982422, "learning_rate": 3e-05, "loss": 3.2623, "step": 962 }, { "epoch": 0.26168478260869565, "grad_norm": 1.7709441184997559, "learning_rate": 3e-05, "loss": 3.5855, "step": 963 }, { "epoch": 0.2619565217391304, "grad_norm": 1.4671415090560913, "learning_rate": 3e-05, "loss": 3.2081, "step": 964 }, { "epoch": 0.26222826086956524, "grad_norm": 1.2123912572860718, "learning_rate": 3e-05, "loss": 3.0287, "step": 965 }, { "epoch": 0.2625, "grad_norm": 1.9638895988464355, "learning_rate": 3e-05, "loss": 3.4101, "step": 966 }, { "epoch": 0.2627717391304348, "grad_norm": 1.6470272541046143, "learning_rate": 3e-05, "loss": 3.1761, "step": 967 }, { "epoch": 0.26304347826086955, "grad_norm": 1.7481547594070435, "learning_rate": 3e-05, "loss": 3.7102, "step": 968 }, { "epoch": 0.26331521739130437, "grad_norm": 1.589220404624939, "learning_rate": 3e-05, "loss": 2.9376, "step": 969 }, { "epoch": 0.26358695652173914, "grad_norm": 2.317216634750366, "learning_rate": 3e-05, "loss": 3.888, "step": 970 }, { "epoch": 0.2638586956521739, "grad_norm": 1.9831353425979614, "learning_rate": 3e-05, "loss": 3.831, "step": 971 }, { "epoch": 0.26413043478260867, "grad_norm": 2.201885223388672, "learning_rate": 3e-05, "loss": 3.5633, "step": 972 }, { "epoch": 0.2644021739130435, "grad_norm": 1.755304217338562, "learning_rate": 3e-05, "loss": 3.9027, "step": 973 }, { "epoch": 0.26467391304347826, "grad_norm": 1.7514090538024902, "learning_rate": 3e-05, "loss": 3.3256, "step": 974 }, { "epoch": 0.264945652173913, "grad_norm": 2.1359901428222656, "learning_rate": 3e-05, "loss": 3.9904, "step": 975 }, { "epoch": 0.26521739130434785, "grad_norm": 1.5768872499465942, "learning_rate": 3e-05, "loss": 3.3242, "step": 976 }, { "epoch": 0.2654891304347826, "grad_norm": 1.8486355543136597, "learning_rate": 3e-05, "loss": 3.7133, "step": 977 }, { "epoch": 0.2657608695652174, "grad_norm": 1.5333871841430664, "learning_rate": 3e-05, "loss": 3.4191, "step": 978 }, { "epoch": 0.26603260869565215, "grad_norm": 1.6924866437911987, "learning_rate": 3e-05, "loss": 3.3763, "step": 979 }, { "epoch": 0.266304347826087, "grad_norm": 2.1424450874328613, "learning_rate": 3e-05, "loss": 3.8962, "step": 980 }, { "epoch": 0.26657608695652174, "grad_norm": 1.5438109636306763, "learning_rate": 3e-05, "loss": 3.2176, "step": 981 }, { "epoch": 0.2668478260869565, "grad_norm": 1.5958982706069946, "learning_rate": 3e-05, "loss": 3.8708, "step": 982 }, { "epoch": 0.26711956521739133, "grad_norm": 1.129310965538025, "learning_rate": 3e-05, "loss": 2.8674, "step": 983 }, { "epoch": 0.2673913043478261, "grad_norm": 1.6280345916748047, "learning_rate": 3e-05, "loss": 4.165, "step": 984 }, { "epoch": 0.26766304347826086, "grad_norm": 1.499475121498108, "learning_rate": 3e-05, "loss": 2.9953, "step": 985 }, { "epoch": 0.26793478260869563, "grad_norm": 1.253732442855835, "learning_rate": 3e-05, "loss": 3.4718, "step": 986 }, { "epoch": 0.26820652173913045, "grad_norm": 1.113315463066101, "learning_rate": 3e-05, "loss": 3.3101, "step": 987 }, { "epoch": 0.2684782608695652, "grad_norm": 1.4947681427001953, "learning_rate": 3e-05, "loss": 3.217, "step": 988 }, { "epoch": 0.26875, "grad_norm": 1.9959630966186523, "learning_rate": 3e-05, "loss": 4.0953, "step": 989 }, { "epoch": 0.26902173913043476, "grad_norm": 1.6297112703323364, "learning_rate": 3e-05, "loss": 3.4108, "step": 990 }, { "epoch": 0.2692934782608696, "grad_norm": 1.9028682708740234, "learning_rate": 3e-05, "loss": 4.1301, "step": 991 }, { "epoch": 0.26956521739130435, "grad_norm": 1.6655460596084595, "learning_rate": 3e-05, "loss": 3.5511, "step": 992 }, { "epoch": 0.2698369565217391, "grad_norm": 1.8980700969696045, "learning_rate": 3e-05, "loss": 3.5143, "step": 993 }, { "epoch": 0.27010869565217394, "grad_norm": 1.5208625793457031, "learning_rate": 3e-05, "loss": 3.5514, "step": 994 }, { "epoch": 0.2703804347826087, "grad_norm": 1.1632534265518188, "learning_rate": 3e-05, "loss": 3.2167, "step": 995 }, { "epoch": 0.27065217391304347, "grad_norm": 2.094667673110962, "learning_rate": 3e-05, "loss": 3.0362, "step": 996 }, { "epoch": 0.27092391304347824, "grad_norm": 1.4889837503433228, "learning_rate": 3e-05, "loss": 3.4279, "step": 997 }, { "epoch": 0.27119565217391306, "grad_norm": 1.9965155124664307, "learning_rate": 3e-05, "loss": 3.774, "step": 998 }, { "epoch": 0.2714673913043478, "grad_norm": 1.5410970449447632, "learning_rate": 3e-05, "loss": 3.2647, "step": 999 }, { "epoch": 0.2717391304347826, "grad_norm": 1.9614113569259644, "learning_rate": 3e-05, "loss": 3.5648, "step": 1000 }, { "epoch": 0.2720108695652174, "grad_norm": 1.3656377792358398, "learning_rate": 3e-05, "loss": 3.3263, "step": 1001 }, { "epoch": 0.2722826086956522, "grad_norm": 2.057823419570923, "learning_rate": 3e-05, "loss": 3.1952, "step": 1002 }, { "epoch": 0.27255434782608695, "grad_norm": 1.9812930822372437, "learning_rate": 3e-05, "loss": 3.5128, "step": 1003 }, { "epoch": 0.2728260869565217, "grad_norm": 1.443973183631897, "learning_rate": 3e-05, "loss": 3.5108, "step": 1004 }, { "epoch": 0.27309782608695654, "grad_norm": 1.1789878606796265, "learning_rate": 3e-05, "loss": 3.142, "step": 1005 }, { "epoch": 0.2733695652173913, "grad_norm": 1.5677335262298584, "learning_rate": 3e-05, "loss": 3.0946, "step": 1006 }, { "epoch": 0.2736413043478261, "grad_norm": 1.6985656023025513, "learning_rate": 3e-05, "loss": 3.8017, "step": 1007 }, { "epoch": 0.27391304347826084, "grad_norm": 0.9420762658119202, "learning_rate": 3e-05, "loss": 2.8754, "step": 1008 }, { "epoch": 0.27418478260869567, "grad_norm": 1.1210711002349854, "learning_rate": 3e-05, "loss": 3.1734, "step": 1009 }, { "epoch": 0.27445652173913043, "grad_norm": 0.9131266474723816, "learning_rate": 3e-05, "loss": 2.5223, "step": 1010 }, { "epoch": 0.2747282608695652, "grad_norm": 1.4077900648117065, "learning_rate": 3e-05, "loss": 3.2054, "step": 1011 }, { "epoch": 0.275, "grad_norm": 1.4042190313339233, "learning_rate": 3e-05, "loss": 3.3916, "step": 1012 }, { "epoch": 0.2752717391304348, "grad_norm": 1.1815489530563354, "learning_rate": 3e-05, "loss": 3.2108, "step": 1013 }, { "epoch": 0.27554347826086956, "grad_norm": 1.007114052772522, "learning_rate": 3e-05, "loss": 2.8035, "step": 1014 }, { "epoch": 0.2758152173913043, "grad_norm": 1.6630133390426636, "learning_rate": 3e-05, "loss": 3.5364, "step": 1015 }, { "epoch": 0.27608695652173915, "grad_norm": 1.3151155710220337, "learning_rate": 3e-05, "loss": 3.8982, "step": 1016 }, { "epoch": 0.2763586956521739, "grad_norm": 1.3191099166870117, "learning_rate": 3e-05, "loss": 3.3812, "step": 1017 }, { "epoch": 0.2766304347826087, "grad_norm": 0.9553658366203308, "learning_rate": 3e-05, "loss": 3.2285, "step": 1018 }, { "epoch": 0.2769021739130435, "grad_norm": 1.2714715003967285, "learning_rate": 3e-05, "loss": 4.1331, "step": 1019 }, { "epoch": 0.27717391304347827, "grad_norm": 1.4878464937210083, "learning_rate": 3e-05, "loss": 3.5607, "step": 1020 }, { "epoch": 0.27744565217391304, "grad_norm": 1.3467168807983398, "learning_rate": 3e-05, "loss": 3.4943, "step": 1021 }, { "epoch": 0.2777173913043478, "grad_norm": 1.052966833114624, "learning_rate": 3e-05, "loss": 2.9399, "step": 1022 }, { "epoch": 0.2779891304347826, "grad_norm": 1.1948199272155762, "learning_rate": 3e-05, "loss": 3.0152, "step": 1023 }, { "epoch": 0.2782608695652174, "grad_norm": 1.1960163116455078, "learning_rate": 3e-05, "loss": 2.9338, "step": 1024 }, { "epoch": 0.27853260869565216, "grad_norm": 1.2721116542816162, "learning_rate": 3e-05, "loss": 3.2096, "step": 1025 }, { "epoch": 0.27880434782608693, "grad_norm": 1.620215654373169, "learning_rate": 3e-05, "loss": 3.3828, "step": 1026 }, { "epoch": 0.27907608695652175, "grad_norm": 1.4918383359909058, "learning_rate": 3e-05, "loss": 3.5746, "step": 1027 }, { "epoch": 0.2793478260869565, "grad_norm": 1.9811660051345825, "learning_rate": 3e-05, "loss": 4.0559, "step": 1028 }, { "epoch": 0.2796195652173913, "grad_norm": 1.3812586069107056, "learning_rate": 3e-05, "loss": 3.1452, "step": 1029 }, { "epoch": 0.2798913043478261, "grad_norm": 1.53925359249115, "learning_rate": 3e-05, "loss": 3.9887, "step": 1030 }, { "epoch": 0.2801630434782609, "grad_norm": 2.2425036430358887, "learning_rate": 3e-05, "loss": 4.2516, "step": 1031 }, { "epoch": 0.28043478260869564, "grad_norm": 1.2638776302337646, "learning_rate": 3e-05, "loss": 3.5207, "step": 1032 }, { "epoch": 0.2807065217391304, "grad_norm": 1.1238954067230225, "learning_rate": 3e-05, "loss": 3.4072, "step": 1033 }, { "epoch": 0.28097826086956523, "grad_norm": 1.2834945917129517, "learning_rate": 3e-05, "loss": 3.0585, "step": 1034 }, { "epoch": 0.28125, "grad_norm": 1.5284252166748047, "learning_rate": 3e-05, "loss": 3.1547, "step": 1035 }, { "epoch": 0.28152173913043477, "grad_norm": 1.8825403451919556, "learning_rate": 3e-05, "loss": 3.892, "step": 1036 }, { "epoch": 0.2817934782608696, "grad_norm": 1.620376467704773, "learning_rate": 3e-05, "loss": 2.9044, "step": 1037 }, { "epoch": 0.28206521739130436, "grad_norm": 2.1728601455688477, "learning_rate": 3e-05, "loss": 3.1266, "step": 1038 }, { "epoch": 0.2823369565217391, "grad_norm": 2.090522050857544, "learning_rate": 3e-05, "loss": 4.4403, "step": 1039 }, { "epoch": 0.2826086956521739, "grad_norm": 2.0362656116485596, "learning_rate": 3e-05, "loss": 3.6416, "step": 1040 }, { "epoch": 0.2828804347826087, "grad_norm": 1.493579387664795, "learning_rate": 3e-05, "loss": 3.2111, "step": 1041 }, { "epoch": 0.2831521739130435, "grad_norm": 1.2144675254821777, "learning_rate": 3e-05, "loss": 2.911, "step": 1042 }, { "epoch": 0.28342391304347825, "grad_norm": 1.833510398864746, "learning_rate": 3e-05, "loss": 3.0022, "step": 1043 }, { "epoch": 0.28369565217391307, "grad_norm": 2.0041656494140625, "learning_rate": 3e-05, "loss": 3.1412, "step": 1044 }, { "epoch": 0.28396739130434784, "grad_norm": 1.2218122482299805, "learning_rate": 3e-05, "loss": 3.4581, "step": 1045 }, { "epoch": 0.2842391304347826, "grad_norm": 1.6633751392364502, "learning_rate": 3e-05, "loss": 3.6047, "step": 1046 }, { "epoch": 0.2845108695652174, "grad_norm": 2.4305849075317383, "learning_rate": 3e-05, "loss": 3.3797, "step": 1047 }, { "epoch": 0.2847826086956522, "grad_norm": 2.309483289718628, "learning_rate": 3e-05, "loss": 3.6304, "step": 1048 }, { "epoch": 0.28505434782608696, "grad_norm": 1.678916335105896, "learning_rate": 3e-05, "loss": 3.4073, "step": 1049 }, { "epoch": 0.28532608695652173, "grad_norm": 1.9873173236846924, "learning_rate": 3e-05, "loss": 3.9298, "step": 1050 }, { "epoch": 0.2855978260869565, "grad_norm": 2.085996389389038, "learning_rate": 3e-05, "loss": 4.4263, "step": 1051 }, { "epoch": 0.2858695652173913, "grad_norm": 3.049558639526367, "learning_rate": 3e-05, "loss": 3.6507, "step": 1052 }, { "epoch": 0.2861413043478261, "grad_norm": 1.4257125854492188, "learning_rate": 3e-05, "loss": 3.3861, "step": 1053 }, { "epoch": 0.28641304347826085, "grad_norm": 1.5436605215072632, "learning_rate": 3e-05, "loss": 3.1414, "step": 1054 }, { "epoch": 0.2866847826086957, "grad_norm": 3.0967252254486084, "learning_rate": 3e-05, "loss": 3.9067, "step": 1055 }, { "epoch": 0.28695652173913044, "grad_norm": 1.6564278602600098, "learning_rate": 3e-05, "loss": 3.5354, "step": 1056 }, { "epoch": 0.2872282608695652, "grad_norm": 2.498901605606079, "learning_rate": 3e-05, "loss": 3.3153, "step": 1057 }, { "epoch": 0.2875, "grad_norm": 2.2996490001678467, "learning_rate": 3e-05, "loss": 3.4023, "step": 1058 }, { "epoch": 0.2877717391304348, "grad_norm": 2.0972297191619873, "learning_rate": 3e-05, "loss": 3.038, "step": 1059 }, { "epoch": 0.28804347826086957, "grad_norm": 1.221664547920227, "learning_rate": 3e-05, "loss": 3.0761, "step": 1060 }, { "epoch": 0.28831521739130433, "grad_norm": 2.2344908714294434, "learning_rate": 3e-05, "loss": 2.8314, "step": 1061 }, { "epoch": 0.28858695652173916, "grad_norm": 1.9931285381317139, "learning_rate": 3e-05, "loss": 3.3899, "step": 1062 }, { "epoch": 0.2888586956521739, "grad_norm": 1.924612045288086, "learning_rate": 3e-05, "loss": 3.4548, "step": 1063 }, { "epoch": 0.2891304347826087, "grad_norm": 1.9507229328155518, "learning_rate": 3e-05, "loss": 3.9302, "step": 1064 }, { "epoch": 0.28940217391304346, "grad_norm": 1.8761188983917236, "learning_rate": 3e-05, "loss": 3.195, "step": 1065 }, { "epoch": 0.2896739130434783, "grad_norm": 2.233293294906616, "learning_rate": 3e-05, "loss": 3.1672, "step": 1066 }, { "epoch": 0.28994565217391305, "grad_norm": 1.7154324054718018, "learning_rate": 3e-05, "loss": 3.1413, "step": 1067 }, { "epoch": 0.2902173913043478, "grad_norm": 1.8050798177719116, "learning_rate": 3e-05, "loss": 3.4341, "step": 1068 }, { "epoch": 0.2904891304347826, "grad_norm": 1.838165044784546, "learning_rate": 3e-05, "loss": 3.3827, "step": 1069 }, { "epoch": 0.2907608695652174, "grad_norm": 2.7229554653167725, "learning_rate": 3e-05, "loss": 3.3938, "step": 1070 }, { "epoch": 0.2910326086956522, "grad_norm": 3.150635242462158, "learning_rate": 3e-05, "loss": 3.8396, "step": 1071 }, { "epoch": 0.29130434782608694, "grad_norm": 2.2429847717285156, "learning_rate": 3e-05, "loss": 3.9437, "step": 1072 }, { "epoch": 0.29157608695652176, "grad_norm": 1.305802345275879, "learning_rate": 3e-05, "loss": 3.5962, "step": 1073 }, { "epoch": 0.29184782608695653, "grad_norm": 1.4509013891220093, "learning_rate": 3e-05, "loss": 3.2975, "step": 1074 }, { "epoch": 0.2921195652173913, "grad_norm": 2.059382200241089, "learning_rate": 3e-05, "loss": 4.21, "step": 1075 }, { "epoch": 0.29239130434782606, "grad_norm": 0.9824801087379456, "learning_rate": 3e-05, "loss": 2.6315, "step": 1076 }, { "epoch": 0.2926630434782609, "grad_norm": 1.0867946147918701, "learning_rate": 3e-05, "loss": 2.6838, "step": 1077 }, { "epoch": 0.29293478260869565, "grad_norm": 1.992459774017334, "learning_rate": 3e-05, "loss": 4.2482, "step": 1078 }, { "epoch": 0.2932065217391304, "grad_norm": 1.6963497400283813, "learning_rate": 3e-05, "loss": 3.5684, "step": 1079 }, { "epoch": 0.29347826086956524, "grad_norm": 1.2850168943405151, "learning_rate": 3e-05, "loss": 3.399, "step": 1080 }, { "epoch": 0.29375, "grad_norm": 1.652712106704712, "learning_rate": 3e-05, "loss": 3.3213, "step": 1081 }, { "epoch": 0.2940217391304348, "grad_norm": 1.733959436416626, "learning_rate": 3e-05, "loss": 3.1942, "step": 1082 }, { "epoch": 0.29429347826086955, "grad_norm": 1.8719747066497803, "learning_rate": 3e-05, "loss": 3.8789, "step": 1083 }, { "epoch": 0.29456521739130437, "grad_norm": 1.5426567792892456, "learning_rate": 3e-05, "loss": 3.1319, "step": 1084 }, { "epoch": 0.29483695652173914, "grad_norm": 2.0286598205566406, "learning_rate": 3e-05, "loss": 3.9002, "step": 1085 }, { "epoch": 0.2951086956521739, "grad_norm": 1.5031648874282837, "learning_rate": 3e-05, "loss": 3.3979, "step": 1086 }, { "epoch": 0.29538043478260867, "grad_norm": 1.556738257408142, "learning_rate": 3e-05, "loss": 3.9843, "step": 1087 }, { "epoch": 0.2956521739130435, "grad_norm": 1.8954527378082275, "learning_rate": 3e-05, "loss": 4.6858, "step": 1088 }, { "epoch": 0.29592391304347826, "grad_norm": 1.7330039739608765, "learning_rate": 3e-05, "loss": 2.8431, "step": 1089 }, { "epoch": 0.296195652173913, "grad_norm": 2.0785205364227295, "learning_rate": 3e-05, "loss": 3.6168, "step": 1090 }, { "epoch": 0.29646739130434785, "grad_norm": 1.708628535270691, "learning_rate": 3e-05, "loss": 3.6387, "step": 1091 }, { "epoch": 0.2967391304347826, "grad_norm": 2.5519232749938965, "learning_rate": 3e-05, "loss": 4.0809, "step": 1092 }, { "epoch": 0.2970108695652174, "grad_norm": 1.5967369079589844, "learning_rate": 3e-05, "loss": 3.5908, "step": 1093 }, { "epoch": 0.29728260869565215, "grad_norm": 1.345123052597046, "learning_rate": 3e-05, "loss": 3.297, "step": 1094 }, { "epoch": 0.297554347826087, "grad_norm": 2.123605251312256, "learning_rate": 3e-05, "loss": 4.1267, "step": 1095 }, { "epoch": 0.29782608695652174, "grad_norm": 1.6863003969192505, "learning_rate": 3e-05, "loss": 3.4321, "step": 1096 }, { "epoch": 0.2980978260869565, "grad_norm": 2.0726163387298584, "learning_rate": 3e-05, "loss": 3.386, "step": 1097 }, { "epoch": 0.29836956521739133, "grad_norm": 1.5322779417037964, "learning_rate": 3e-05, "loss": 2.9161, "step": 1098 }, { "epoch": 0.2986413043478261, "grad_norm": 1.1131296157836914, "learning_rate": 3e-05, "loss": 2.7584, "step": 1099 }, { "epoch": 0.29891304347826086, "grad_norm": 1.9430760145187378, "learning_rate": 3e-05, "loss": 3.4661, "step": 1100 }, { "epoch": 0.29918478260869563, "grad_norm": 2.6745107173919678, "learning_rate": 3e-05, "loss": 3.0976, "step": 1101 }, { "epoch": 0.29945652173913045, "grad_norm": 3.127723217010498, "learning_rate": 3e-05, "loss": 3.7013, "step": 1102 }, { "epoch": 0.2997282608695652, "grad_norm": 1.894990086555481, "learning_rate": 3e-05, "loss": 3.4841, "step": 1103 }, { "epoch": 0.3, "grad_norm": 1.9465398788452148, "learning_rate": 3e-05, "loss": 3.48, "step": 1104 }, { "epoch": 0.30027173913043476, "grad_norm": 2.5459132194519043, "learning_rate": 3e-05, "loss": 3.239, "step": 1105 }, { "epoch": 0.3005434782608696, "grad_norm": 2.382801055908203, "learning_rate": 3e-05, "loss": 3.2947, "step": 1106 }, { "epoch": 0.30081521739130435, "grad_norm": 1.913734793663025, "learning_rate": 3e-05, "loss": 3.0162, "step": 1107 }, { "epoch": 0.3010869565217391, "grad_norm": 1.1874632835388184, "learning_rate": 3e-05, "loss": 2.6723, "step": 1108 }, { "epoch": 0.30135869565217394, "grad_norm": 1.6929433345794678, "learning_rate": 3e-05, "loss": 3.1864, "step": 1109 }, { "epoch": 0.3016304347826087, "grad_norm": 2.1046361923217773, "learning_rate": 3e-05, "loss": 3.442, "step": 1110 }, { "epoch": 0.30190217391304347, "grad_norm": 2.196091413497925, "learning_rate": 3e-05, "loss": 3.2791, "step": 1111 }, { "epoch": 0.30217391304347824, "grad_norm": 2.6548757553100586, "learning_rate": 3e-05, "loss": 4.0965, "step": 1112 }, { "epoch": 0.30244565217391306, "grad_norm": 1.7033164501190186, "learning_rate": 3e-05, "loss": 3.5031, "step": 1113 }, { "epoch": 0.3027173913043478, "grad_norm": 1.5866395235061646, "learning_rate": 3e-05, "loss": 2.9015, "step": 1114 }, { "epoch": 0.3029891304347826, "grad_norm": 2.0246646404266357, "learning_rate": 3e-05, "loss": 3.1981, "step": 1115 }, { "epoch": 0.3032608695652174, "grad_norm": 2.0733206272125244, "learning_rate": 3e-05, "loss": 3.3524, "step": 1116 }, { "epoch": 0.3035326086956522, "grad_norm": 1.6320792436599731, "learning_rate": 3e-05, "loss": 3.6667, "step": 1117 }, { "epoch": 0.30380434782608695, "grad_norm": 1.4059102535247803, "learning_rate": 3e-05, "loss": 3.4399, "step": 1118 }, { "epoch": 0.3040760869565217, "grad_norm": 1.6570987701416016, "learning_rate": 3e-05, "loss": 3.2745, "step": 1119 }, { "epoch": 0.30434782608695654, "grad_norm": 1.9712703227996826, "learning_rate": 3e-05, "loss": 3.751, "step": 1120 }, { "epoch": 0.3046195652173913, "grad_norm": 1.4697049856185913, "learning_rate": 3e-05, "loss": 3.4402, "step": 1121 }, { "epoch": 0.3048913043478261, "grad_norm": 1.8021398782730103, "learning_rate": 3e-05, "loss": 3.8907, "step": 1122 }, { "epoch": 0.30516304347826084, "grad_norm": 1.5529664754867554, "learning_rate": 3e-05, "loss": 3.7741, "step": 1123 }, { "epoch": 0.30543478260869567, "grad_norm": 1.5865482091903687, "learning_rate": 3e-05, "loss": 3.7243, "step": 1124 }, { "epoch": 0.30570652173913043, "grad_norm": 2.221069097518921, "learning_rate": 3e-05, "loss": 3.65, "step": 1125 }, { "epoch": 0.3059782608695652, "grad_norm": 2.1728017330169678, "learning_rate": 3e-05, "loss": 3.0357, "step": 1126 }, { "epoch": 0.30625, "grad_norm": 1.6755625009536743, "learning_rate": 3e-05, "loss": 3.2292, "step": 1127 }, { "epoch": 0.3065217391304348, "grad_norm": 2.135910987854004, "learning_rate": 3e-05, "loss": 4.0064, "step": 1128 }, { "epoch": 0.30679347826086956, "grad_norm": 1.7224137783050537, "learning_rate": 3e-05, "loss": 3.6076, "step": 1129 }, { "epoch": 0.3070652173913043, "grad_norm": 1.5196242332458496, "learning_rate": 3e-05, "loss": 3.2132, "step": 1130 }, { "epoch": 0.30733695652173915, "grad_norm": 2.016763210296631, "learning_rate": 3e-05, "loss": 4.1373, "step": 1131 }, { "epoch": 0.3076086956521739, "grad_norm": 1.3778748512268066, "learning_rate": 3e-05, "loss": 3.4815, "step": 1132 }, { "epoch": 0.3078804347826087, "grad_norm": 1.5161099433898926, "learning_rate": 3e-05, "loss": 3.2094, "step": 1133 }, { "epoch": 0.3081521739130435, "grad_norm": 1.2191746234893799, "learning_rate": 3e-05, "loss": 3.0117, "step": 1134 }, { "epoch": 0.30842391304347827, "grad_norm": 1.6270561218261719, "learning_rate": 3e-05, "loss": 3.8738, "step": 1135 }, { "epoch": 0.30869565217391304, "grad_norm": 1.3209682703018188, "learning_rate": 3e-05, "loss": 3.1873, "step": 1136 }, { "epoch": 0.3089673913043478, "grad_norm": 1.7960654497146606, "learning_rate": 3e-05, "loss": 3.9211, "step": 1137 }, { "epoch": 0.3092391304347826, "grad_norm": 1.5645345449447632, "learning_rate": 3e-05, "loss": 3.8176, "step": 1138 }, { "epoch": 0.3095108695652174, "grad_norm": 1.2274796962738037, "learning_rate": 3e-05, "loss": 3.0552, "step": 1139 }, { "epoch": 0.30978260869565216, "grad_norm": 3.1637630462646484, "learning_rate": 3e-05, "loss": 4.0908, "step": 1140 }, { "epoch": 0.31005434782608693, "grad_norm": 1.7452456951141357, "learning_rate": 3e-05, "loss": 3.2379, "step": 1141 }, { "epoch": 0.31032608695652175, "grad_norm": 1.3035019636154175, "learning_rate": 3e-05, "loss": 3.1619, "step": 1142 }, { "epoch": 0.3105978260869565, "grad_norm": 1.5721486806869507, "learning_rate": 3e-05, "loss": 3.3092, "step": 1143 }, { "epoch": 0.3108695652173913, "grad_norm": 1.5714532136917114, "learning_rate": 3e-05, "loss": 3.7343, "step": 1144 }, { "epoch": 0.3111413043478261, "grad_norm": 1.4483362436294556, "learning_rate": 3e-05, "loss": 3.5843, "step": 1145 }, { "epoch": 0.3114130434782609, "grad_norm": 1.2886680364608765, "learning_rate": 3e-05, "loss": 3.4743, "step": 1146 }, { "epoch": 0.31168478260869564, "grad_norm": 1.5609358549118042, "learning_rate": 3e-05, "loss": 3.7947, "step": 1147 }, { "epoch": 0.3119565217391304, "grad_norm": 1.7903861999511719, "learning_rate": 3e-05, "loss": 3.3165, "step": 1148 }, { "epoch": 0.31222826086956523, "grad_norm": 1.917932152748108, "learning_rate": 3e-05, "loss": 3.3985, "step": 1149 }, { "epoch": 0.3125, "grad_norm": 1.5185085535049438, "learning_rate": 3e-05, "loss": 3.571, "step": 1150 }, { "epoch": 0.31277173913043477, "grad_norm": 1.3498207330703735, "learning_rate": 3e-05, "loss": 3.3921, "step": 1151 }, { "epoch": 0.3130434782608696, "grad_norm": 1.6622629165649414, "learning_rate": 3e-05, "loss": 3.4101, "step": 1152 }, { "epoch": 0.31331521739130436, "grad_norm": 1.516265630722046, "learning_rate": 3e-05, "loss": 3.1228, "step": 1153 }, { "epoch": 0.3135869565217391, "grad_norm": 1.5660970211029053, "learning_rate": 3e-05, "loss": 3.3751, "step": 1154 }, { "epoch": 0.3138586956521739, "grad_norm": 1.8406776189804077, "learning_rate": 3e-05, "loss": 3.3181, "step": 1155 }, { "epoch": 0.3141304347826087, "grad_norm": 1.8162765502929688, "learning_rate": 3e-05, "loss": 4.0117, "step": 1156 }, { "epoch": 0.3144021739130435, "grad_norm": 1.2983214855194092, "learning_rate": 3e-05, "loss": 3.2412, "step": 1157 }, { "epoch": 0.31467391304347825, "grad_norm": 1.1409294605255127, "learning_rate": 3e-05, "loss": 3.0329, "step": 1158 }, { "epoch": 0.31494565217391307, "grad_norm": 1.754345178604126, "learning_rate": 3e-05, "loss": 3.6711, "step": 1159 }, { "epoch": 0.31521739130434784, "grad_norm": 1.868580937385559, "learning_rate": 3e-05, "loss": 3.6684, "step": 1160 }, { "epoch": 0.3154891304347826, "grad_norm": 1.8012367486953735, "learning_rate": 3e-05, "loss": 3.2884, "step": 1161 }, { "epoch": 0.3157608695652174, "grad_norm": 1.7688064575195312, "learning_rate": 3e-05, "loss": 3.3512, "step": 1162 }, { "epoch": 0.3160326086956522, "grad_norm": 1.5674185752868652, "learning_rate": 3e-05, "loss": 3.443, "step": 1163 }, { "epoch": 0.31630434782608696, "grad_norm": 1.5843913555145264, "learning_rate": 3e-05, "loss": 3.8519, "step": 1164 }, { "epoch": 0.31657608695652173, "grad_norm": 1.4522455930709839, "learning_rate": 3e-05, "loss": 3.123, "step": 1165 }, { "epoch": 0.3168478260869565, "grad_norm": 1.5610785484313965, "learning_rate": 3e-05, "loss": 3.1963, "step": 1166 }, { "epoch": 0.3171195652173913, "grad_norm": 1.4327051639556885, "learning_rate": 3e-05, "loss": 2.9382, "step": 1167 }, { "epoch": 0.3173913043478261, "grad_norm": 1.6065129041671753, "learning_rate": 3e-05, "loss": 3.4928, "step": 1168 }, { "epoch": 0.31766304347826085, "grad_norm": 1.9382498264312744, "learning_rate": 3e-05, "loss": 4.3174, "step": 1169 }, { "epoch": 0.3179347826086957, "grad_norm": 1.164689064025879, "learning_rate": 3e-05, "loss": 3.0678, "step": 1170 }, { "epoch": 0.31820652173913044, "grad_norm": 1.084055781364441, "learning_rate": 3e-05, "loss": 3.0878, "step": 1171 }, { "epoch": 0.3184782608695652, "grad_norm": 1.4881168603897095, "learning_rate": 3e-05, "loss": 3.5084, "step": 1172 }, { "epoch": 0.31875, "grad_norm": 1.5508403778076172, "learning_rate": 3e-05, "loss": 3.6359, "step": 1173 }, { "epoch": 0.3190217391304348, "grad_norm": 1.7135038375854492, "learning_rate": 3e-05, "loss": 2.819, "step": 1174 }, { "epoch": 0.31929347826086957, "grad_norm": 1.9360867738723755, "learning_rate": 3e-05, "loss": 3.711, "step": 1175 }, { "epoch": 0.31956521739130433, "grad_norm": 1.3898394107818604, "learning_rate": 3e-05, "loss": 3.0356, "step": 1176 }, { "epoch": 0.31983695652173916, "grad_norm": 1.37100088596344, "learning_rate": 3e-05, "loss": 2.9441, "step": 1177 }, { "epoch": 0.3201086956521739, "grad_norm": 1.7066446542739868, "learning_rate": 3e-05, "loss": 3.8696, "step": 1178 }, { "epoch": 0.3203804347826087, "grad_norm": 1.4085134267807007, "learning_rate": 3e-05, "loss": 3.4616, "step": 1179 }, { "epoch": 0.32065217391304346, "grad_norm": 1.5345913171768188, "learning_rate": 3e-05, "loss": 3.6033, "step": 1180 }, { "epoch": 0.3209239130434783, "grad_norm": 1.5240819454193115, "learning_rate": 3e-05, "loss": 3.1863, "step": 1181 }, { "epoch": 0.32119565217391305, "grad_norm": 1.147559404373169, "learning_rate": 3e-05, "loss": 3.2533, "step": 1182 }, { "epoch": 0.3214673913043478, "grad_norm": 1.3267796039581299, "learning_rate": 3e-05, "loss": 3.3342, "step": 1183 }, { "epoch": 0.3217391304347826, "grad_norm": 1.567126989364624, "learning_rate": 3e-05, "loss": 3.5055, "step": 1184 }, { "epoch": 0.3220108695652174, "grad_norm": 2.085658073425293, "learning_rate": 3e-05, "loss": 3.1383, "step": 1185 }, { "epoch": 0.3222826086956522, "grad_norm": 1.7624791860580444, "learning_rate": 3e-05, "loss": 3.2409, "step": 1186 }, { "epoch": 0.32255434782608694, "grad_norm": 1.4368826150894165, "learning_rate": 3e-05, "loss": 3.7424, "step": 1187 }, { "epoch": 0.32282608695652176, "grad_norm": 1.5510128736495972, "learning_rate": 3e-05, "loss": 3.3895, "step": 1188 }, { "epoch": 0.32309782608695653, "grad_norm": 1.471490740776062, "learning_rate": 3e-05, "loss": 3.6772, "step": 1189 }, { "epoch": 0.3233695652173913, "grad_norm": 1.2040995359420776, "learning_rate": 3e-05, "loss": 3.3302, "step": 1190 }, { "epoch": 0.32364130434782606, "grad_norm": 1.4731234312057495, "learning_rate": 3e-05, "loss": 3.3842, "step": 1191 }, { "epoch": 0.3239130434782609, "grad_norm": 1.3904873132705688, "learning_rate": 3e-05, "loss": 3.8537, "step": 1192 }, { "epoch": 0.32418478260869565, "grad_norm": 1.187960147857666, "learning_rate": 3e-05, "loss": 3.0656, "step": 1193 }, { "epoch": 0.3244565217391304, "grad_norm": 1.1819937229156494, "learning_rate": 3e-05, "loss": 2.7234, "step": 1194 }, { "epoch": 0.32472826086956524, "grad_norm": 1.6428884267807007, "learning_rate": 3e-05, "loss": 3.4446, "step": 1195 }, { "epoch": 0.325, "grad_norm": 1.6877864599227905, "learning_rate": 3e-05, "loss": 3.2674, "step": 1196 }, { "epoch": 0.3252717391304348, "grad_norm": 1.596166729927063, "learning_rate": 3e-05, "loss": 3.371, "step": 1197 }, { "epoch": 0.32554347826086955, "grad_norm": 1.3886607885360718, "learning_rate": 3e-05, "loss": 3.3149, "step": 1198 }, { "epoch": 0.32581521739130437, "grad_norm": 1.4295148849487305, "learning_rate": 3e-05, "loss": 3.0519, "step": 1199 }, { "epoch": 0.32608695652173914, "grad_norm": 1.5317387580871582, "learning_rate": 3e-05, "loss": 3.3816, "step": 1200 }, { "epoch": 0.3263586956521739, "grad_norm": 1.1524282693862915, "learning_rate": 3e-05, "loss": 3.2564, "step": 1201 }, { "epoch": 0.32663043478260867, "grad_norm": 1.3997209072113037, "learning_rate": 3e-05, "loss": 3.3173, "step": 1202 }, { "epoch": 0.3269021739130435, "grad_norm": 1.357279896736145, "learning_rate": 3e-05, "loss": 3.7316, "step": 1203 }, { "epoch": 0.32717391304347826, "grad_norm": 1.1962755918502808, "learning_rate": 3e-05, "loss": 3.5203, "step": 1204 }, { "epoch": 0.327445652173913, "grad_norm": 1.4180536270141602, "learning_rate": 3e-05, "loss": 3.7679, "step": 1205 }, { "epoch": 0.32771739130434785, "grad_norm": 1.4997798204421997, "learning_rate": 3e-05, "loss": 3.9842, "step": 1206 }, { "epoch": 0.3279891304347826, "grad_norm": 1.5224746465682983, "learning_rate": 3e-05, "loss": 3.723, "step": 1207 }, { "epoch": 0.3282608695652174, "grad_norm": 1.4814677238464355, "learning_rate": 3e-05, "loss": 3.3444, "step": 1208 }, { "epoch": 0.32853260869565215, "grad_norm": 1.8869359493255615, "learning_rate": 3e-05, "loss": 3.8663, "step": 1209 }, { "epoch": 0.328804347826087, "grad_norm": 1.2305731773376465, "learning_rate": 3e-05, "loss": 3.1855, "step": 1210 }, { "epoch": 0.32907608695652174, "grad_norm": 1.3822002410888672, "learning_rate": 3e-05, "loss": 3.9625, "step": 1211 }, { "epoch": 0.3293478260869565, "grad_norm": 1.2895426750183105, "learning_rate": 3e-05, "loss": 2.9574, "step": 1212 }, { "epoch": 0.32961956521739133, "grad_norm": 1.2724965810775757, "learning_rate": 3e-05, "loss": 3.2984, "step": 1213 }, { "epoch": 0.3298913043478261, "grad_norm": 1.569366455078125, "learning_rate": 3e-05, "loss": 3.4928, "step": 1214 }, { "epoch": 0.33016304347826086, "grad_norm": 1.538090467453003, "learning_rate": 3e-05, "loss": 3.8567, "step": 1215 }, { "epoch": 0.33043478260869563, "grad_norm": 1.3484816551208496, "learning_rate": 3e-05, "loss": 3.1096, "step": 1216 }, { "epoch": 0.33070652173913045, "grad_norm": 1.1450049877166748, "learning_rate": 3e-05, "loss": 2.773, "step": 1217 }, { "epoch": 0.3309782608695652, "grad_norm": 1.5292000770568848, "learning_rate": 3e-05, "loss": 3.3972, "step": 1218 }, { "epoch": 0.33125, "grad_norm": 1.3615137338638306, "learning_rate": 3e-05, "loss": 3.2462, "step": 1219 }, { "epoch": 0.33152173913043476, "grad_norm": 1.7036170959472656, "learning_rate": 3e-05, "loss": 3.7338, "step": 1220 }, { "epoch": 0.3317934782608696, "grad_norm": 1.7132785320281982, "learning_rate": 3e-05, "loss": 3.3378, "step": 1221 }, { "epoch": 0.33206521739130435, "grad_norm": 1.9641152620315552, "learning_rate": 3e-05, "loss": 3.8121, "step": 1222 }, { "epoch": 0.3323369565217391, "grad_norm": 1.3222547769546509, "learning_rate": 3e-05, "loss": 3.2911, "step": 1223 }, { "epoch": 0.33260869565217394, "grad_norm": 1.3695377111434937, "learning_rate": 3e-05, "loss": 3.3004, "step": 1224 }, { "epoch": 0.3328804347826087, "grad_norm": 1.8578277826309204, "learning_rate": 3e-05, "loss": 4.1831, "step": 1225 }, { "epoch": 0.33315217391304347, "grad_norm": 1.7893558740615845, "learning_rate": 3e-05, "loss": 3.359, "step": 1226 }, { "epoch": 0.33342391304347824, "grad_norm": 1.474753499031067, "learning_rate": 3e-05, "loss": 2.9094, "step": 1227 }, { "epoch": 0.33369565217391306, "grad_norm": 1.973402738571167, "learning_rate": 3e-05, "loss": 4.0505, "step": 1228 }, { "epoch": 0.3339673913043478, "grad_norm": 1.7613177299499512, "learning_rate": 3e-05, "loss": 3.3559, "step": 1229 }, { "epoch": 0.3342391304347826, "grad_norm": 2.031470775604248, "learning_rate": 3e-05, "loss": 3.3066, "step": 1230 }, { "epoch": 0.3345108695652174, "grad_norm": 2.1547582149505615, "learning_rate": 3e-05, "loss": 3.6043, "step": 1231 }, { "epoch": 0.3347826086956522, "grad_norm": 2.2331910133361816, "learning_rate": 3e-05, "loss": 2.9871, "step": 1232 }, { "epoch": 0.33505434782608695, "grad_norm": 2.156770944595337, "learning_rate": 3e-05, "loss": 3.6777, "step": 1233 }, { "epoch": 0.3353260869565217, "grad_norm": 1.717960000038147, "learning_rate": 3e-05, "loss": 3.6268, "step": 1234 }, { "epoch": 0.33559782608695654, "grad_norm": 1.649511694908142, "learning_rate": 3e-05, "loss": 3.3226, "step": 1235 }, { "epoch": 0.3358695652173913, "grad_norm": 1.8623225688934326, "learning_rate": 3e-05, "loss": 3.9329, "step": 1236 }, { "epoch": 0.3361413043478261, "grad_norm": 2.167483329772949, "learning_rate": 3e-05, "loss": 3.445, "step": 1237 }, { "epoch": 0.33641304347826084, "grad_norm": 1.7240395545959473, "learning_rate": 3e-05, "loss": 3.493, "step": 1238 }, { "epoch": 0.33668478260869567, "grad_norm": 1.3711011409759521, "learning_rate": 3e-05, "loss": 2.7254, "step": 1239 }, { "epoch": 0.33695652173913043, "grad_norm": 2.091745615005493, "learning_rate": 3e-05, "loss": 4.0388, "step": 1240 }, { "epoch": 0.3372282608695652, "grad_norm": 1.5960465669631958, "learning_rate": 3e-05, "loss": 3.5036, "step": 1241 }, { "epoch": 0.3375, "grad_norm": 1.1491891145706177, "learning_rate": 3e-05, "loss": 3.2282, "step": 1242 }, { "epoch": 0.3377717391304348, "grad_norm": 1.1655139923095703, "learning_rate": 3e-05, "loss": 3.3578, "step": 1243 }, { "epoch": 0.33804347826086956, "grad_norm": 1.4522464275360107, "learning_rate": 3e-05, "loss": 3.4827, "step": 1244 }, { "epoch": 0.3383152173913043, "grad_norm": 1.494310736656189, "learning_rate": 3e-05, "loss": 3.7874, "step": 1245 }, { "epoch": 0.33858695652173915, "grad_norm": 1.98121976852417, "learning_rate": 3e-05, "loss": 3.2323, "step": 1246 }, { "epoch": 0.3388586956521739, "grad_norm": 1.772510290145874, "learning_rate": 3e-05, "loss": 2.9783, "step": 1247 }, { "epoch": 0.3391304347826087, "grad_norm": 1.3050928115844727, "learning_rate": 3e-05, "loss": 3.3902, "step": 1248 }, { "epoch": 0.3394021739130435, "grad_norm": 1.590279221534729, "learning_rate": 3e-05, "loss": 3.0477, "step": 1249 }, { "epoch": 0.33967391304347827, "grad_norm": 2.1541528701782227, "learning_rate": 3e-05, "loss": 3.5042, "step": 1250 }, { "epoch": 0.33994565217391304, "grad_norm": 2.138265609741211, "learning_rate": 3e-05, "loss": 3.704, "step": 1251 }, { "epoch": 0.3402173913043478, "grad_norm": 1.5365831851959229, "learning_rate": 3e-05, "loss": 3.3562, "step": 1252 }, { "epoch": 0.3404891304347826, "grad_norm": 1.3827742338180542, "learning_rate": 3e-05, "loss": 3.2202, "step": 1253 }, { "epoch": 0.3407608695652174, "grad_norm": 1.4450342655181885, "learning_rate": 3e-05, "loss": 3.4887, "step": 1254 }, { "epoch": 0.34103260869565216, "grad_norm": 1.2322391271591187, "learning_rate": 3e-05, "loss": 2.7866, "step": 1255 }, { "epoch": 0.34130434782608693, "grad_norm": 1.3767555952072144, "learning_rate": 3e-05, "loss": 2.8691, "step": 1256 }, { "epoch": 0.34157608695652175, "grad_norm": 1.9902418851852417, "learning_rate": 3e-05, "loss": 3.5811, "step": 1257 }, { "epoch": 0.3418478260869565, "grad_norm": 1.7038296461105347, "learning_rate": 3e-05, "loss": 3.363, "step": 1258 }, { "epoch": 0.3421195652173913, "grad_norm": 1.82673978805542, "learning_rate": 3e-05, "loss": 3.0324, "step": 1259 }, { "epoch": 0.3423913043478261, "grad_norm": 1.4924451112747192, "learning_rate": 3e-05, "loss": 3.5862, "step": 1260 }, { "epoch": 0.3426630434782609, "grad_norm": 1.1290408372879028, "learning_rate": 3e-05, "loss": 3.0983, "step": 1261 }, { "epoch": 0.34293478260869564, "grad_norm": 1.3652050495147705, "learning_rate": 3e-05, "loss": 3.2806, "step": 1262 }, { "epoch": 0.3432065217391304, "grad_norm": 1.3618534803390503, "learning_rate": 3e-05, "loss": 3.0708, "step": 1263 }, { "epoch": 0.34347826086956523, "grad_norm": 1.3616082668304443, "learning_rate": 3e-05, "loss": 2.8319, "step": 1264 }, { "epoch": 0.34375, "grad_norm": 1.3398898839950562, "learning_rate": 3e-05, "loss": 3.1065, "step": 1265 }, { "epoch": 0.34402173913043477, "grad_norm": 1.6687146425247192, "learning_rate": 3e-05, "loss": 3.3584, "step": 1266 }, { "epoch": 0.3442934782608696, "grad_norm": 1.2948143482208252, "learning_rate": 3e-05, "loss": 3.2489, "step": 1267 }, { "epoch": 0.34456521739130436, "grad_norm": 1.7158524990081787, "learning_rate": 3e-05, "loss": 3.3558, "step": 1268 }, { "epoch": 0.3448369565217391, "grad_norm": 1.6835163831710815, "learning_rate": 3e-05, "loss": 2.9967, "step": 1269 }, { "epoch": 0.3451086956521739, "grad_norm": 1.971177101135254, "learning_rate": 3e-05, "loss": 3.829, "step": 1270 }, { "epoch": 0.3453804347826087, "grad_norm": 1.660911202430725, "learning_rate": 3e-05, "loss": 2.9975, "step": 1271 }, { "epoch": 0.3456521739130435, "grad_norm": 1.8289529085159302, "learning_rate": 3e-05, "loss": 3.4466, "step": 1272 }, { "epoch": 0.34592391304347825, "grad_norm": 1.9580225944519043, "learning_rate": 3e-05, "loss": 3.4431, "step": 1273 }, { "epoch": 0.34619565217391307, "grad_norm": 1.9081965684890747, "learning_rate": 3e-05, "loss": 3.1295, "step": 1274 }, { "epoch": 0.34646739130434784, "grad_norm": 2.8098397254943848, "learning_rate": 3e-05, "loss": 3.9265, "step": 1275 }, { "epoch": 0.3467391304347826, "grad_norm": 1.8996559381484985, "learning_rate": 3e-05, "loss": 2.9118, "step": 1276 }, { "epoch": 0.3470108695652174, "grad_norm": 1.8447978496551514, "learning_rate": 3e-05, "loss": 3.8908, "step": 1277 }, { "epoch": 0.3472826086956522, "grad_norm": 1.450376033782959, "learning_rate": 3e-05, "loss": 3.1262, "step": 1278 }, { "epoch": 0.34755434782608696, "grad_norm": 2.3735806941986084, "learning_rate": 3e-05, "loss": 3.4959, "step": 1279 }, { "epoch": 0.34782608695652173, "grad_norm": 2.2202954292297363, "learning_rate": 3e-05, "loss": 3.2127, "step": 1280 }, { "epoch": 0.3480978260869565, "grad_norm": 1.6313713788986206, "learning_rate": 3e-05, "loss": 2.7593, "step": 1281 }, { "epoch": 0.3483695652173913, "grad_norm": 1.8261613845825195, "learning_rate": 3e-05, "loss": 3.5207, "step": 1282 }, { "epoch": 0.3486413043478261, "grad_norm": 1.6101502180099487, "learning_rate": 3e-05, "loss": 3.1187, "step": 1283 }, { "epoch": 0.34891304347826085, "grad_norm": 1.4293962717056274, "learning_rate": 3e-05, "loss": 3.3783, "step": 1284 }, { "epoch": 0.3491847826086957, "grad_norm": 1.335721731185913, "learning_rate": 3e-05, "loss": 3.8365, "step": 1285 }, { "epoch": 0.34945652173913044, "grad_norm": 1.5614105463027954, "learning_rate": 3e-05, "loss": 3.7661, "step": 1286 }, { "epoch": 0.3497282608695652, "grad_norm": 1.38046133518219, "learning_rate": 3e-05, "loss": 3.1717, "step": 1287 }, { "epoch": 0.35, "grad_norm": 1.3673638105392456, "learning_rate": 3e-05, "loss": 3.7002, "step": 1288 }, { "epoch": 0.3502717391304348, "grad_norm": 1.2321789264678955, "learning_rate": 3e-05, "loss": 2.8849, "step": 1289 }, { "epoch": 0.35054347826086957, "grad_norm": 1.3166979551315308, "learning_rate": 3e-05, "loss": 3.1094, "step": 1290 }, { "epoch": 0.35081521739130433, "grad_norm": 1.5346661806106567, "learning_rate": 3e-05, "loss": 3.2696, "step": 1291 }, { "epoch": 0.35108695652173916, "grad_norm": 1.550195336341858, "learning_rate": 3e-05, "loss": 3.3044, "step": 1292 }, { "epoch": 0.3513586956521739, "grad_norm": 1.7930152416229248, "learning_rate": 3e-05, "loss": 3.7782, "step": 1293 }, { "epoch": 0.3516304347826087, "grad_norm": 1.9929698705673218, "learning_rate": 3e-05, "loss": 3.12, "step": 1294 }, { "epoch": 0.35190217391304346, "grad_norm": 1.244729995727539, "learning_rate": 3e-05, "loss": 2.9368, "step": 1295 }, { "epoch": 0.3521739130434783, "grad_norm": 1.3026173114776611, "learning_rate": 3e-05, "loss": 3.1947, "step": 1296 }, { "epoch": 0.35244565217391305, "grad_norm": 1.297813892364502, "learning_rate": 3e-05, "loss": 2.5978, "step": 1297 }, { "epoch": 0.3527173913043478, "grad_norm": 1.4547146558761597, "learning_rate": 3e-05, "loss": 3.4649, "step": 1298 }, { "epoch": 0.3529891304347826, "grad_norm": 1.2365776300430298, "learning_rate": 3e-05, "loss": 3.1037, "step": 1299 }, { "epoch": 0.3532608695652174, "grad_norm": 1.2477210760116577, "learning_rate": 3e-05, "loss": 3.1045, "step": 1300 }, { "epoch": 0.3535326086956522, "grad_norm": 1.4197901487350464, "learning_rate": 3e-05, "loss": 3.2411, "step": 1301 }, { "epoch": 0.35380434782608694, "grad_norm": 1.1791913509368896, "learning_rate": 3e-05, "loss": 3.4775, "step": 1302 }, { "epoch": 0.35407608695652176, "grad_norm": 1.1935291290283203, "learning_rate": 3e-05, "loss": 2.9038, "step": 1303 }, { "epoch": 0.35434782608695653, "grad_norm": 1.515292763710022, "learning_rate": 3e-05, "loss": 3.6569, "step": 1304 }, { "epoch": 0.3546195652173913, "grad_norm": 1.6459676027297974, "learning_rate": 3e-05, "loss": 3.638, "step": 1305 }, { "epoch": 0.35489130434782606, "grad_norm": 1.2799147367477417, "learning_rate": 3e-05, "loss": 3.3, "step": 1306 }, { "epoch": 0.3551630434782609, "grad_norm": 1.1415566205978394, "learning_rate": 3e-05, "loss": 2.7568, "step": 1307 }, { "epoch": 0.35543478260869565, "grad_norm": 1.4160751104354858, "learning_rate": 3e-05, "loss": 3.5389, "step": 1308 }, { "epoch": 0.3557065217391304, "grad_norm": 1.5589027404785156, "learning_rate": 3e-05, "loss": 3.5721, "step": 1309 }, { "epoch": 0.35597826086956524, "grad_norm": 1.4486863613128662, "learning_rate": 3e-05, "loss": 3.5145, "step": 1310 }, { "epoch": 0.35625, "grad_norm": 1.1204112768173218, "learning_rate": 3e-05, "loss": 3.2109, "step": 1311 }, { "epoch": 0.3565217391304348, "grad_norm": 1.4160523414611816, "learning_rate": 3e-05, "loss": 3.9734, "step": 1312 }, { "epoch": 0.35679347826086955, "grad_norm": 1.359877347946167, "learning_rate": 3e-05, "loss": 3.1197, "step": 1313 }, { "epoch": 0.35706521739130437, "grad_norm": 1.3469575643539429, "learning_rate": 3e-05, "loss": 2.7031, "step": 1314 }, { "epoch": 0.35733695652173914, "grad_norm": 1.877429485321045, "learning_rate": 3e-05, "loss": 3.6433, "step": 1315 }, { "epoch": 0.3576086956521739, "grad_norm": 1.6881115436553955, "learning_rate": 3e-05, "loss": 3.8429, "step": 1316 }, { "epoch": 0.35788043478260867, "grad_norm": 0.9885056018829346, "learning_rate": 3e-05, "loss": 2.9943, "step": 1317 }, { "epoch": 0.3581521739130435, "grad_norm": 1.1146260499954224, "learning_rate": 3e-05, "loss": 3.0583, "step": 1318 }, { "epoch": 0.35842391304347826, "grad_norm": 1.0930460691452026, "learning_rate": 3e-05, "loss": 3.1491, "step": 1319 }, { "epoch": 0.358695652173913, "grad_norm": 1.7721790075302124, "learning_rate": 3e-05, "loss": 3.682, "step": 1320 }, { "epoch": 0.35896739130434785, "grad_norm": 1.8233963251113892, "learning_rate": 3e-05, "loss": 4.0548, "step": 1321 }, { "epoch": 0.3592391304347826, "grad_norm": 2.3196828365325928, "learning_rate": 3e-05, "loss": 3.8065, "step": 1322 }, { "epoch": 0.3595108695652174, "grad_norm": 2.0975351333618164, "learning_rate": 3e-05, "loss": 3.3098, "step": 1323 }, { "epoch": 0.35978260869565215, "grad_norm": 1.4240121841430664, "learning_rate": 3e-05, "loss": 3.327, "step": 1324 }, { "epoch": 0.360054347826087, "grad_norm": 1.4401570558547974, "learning_rate": 3e-05, "loss": 2.9099, "step": 1325 }, { "epoch": 0.36032608695652174, "grad_norm": 1.7879765033721924, "learning_rate": 3e-05, "loss": 2.9776, "step": 1326 }, { "epoch": 0.3605978260869565, "grad_norm": 1.8819643259048462, "learning_rate": 3e-05, "loss": 3.3526, "step": 1327 }, { "epoch": 0.36086956521739133, "grad_norm": 1.8750115633010864, "learning_rate": 3e-05, "loss": 3.7411, "step": 1328 }, { "epoch": 0.3611413043478261, "grad_norm": 1.802832007408142, "learning_rate": 3e-05, "loss": 2.8997, "step": 1329 }, { "epoch": 0.36141304347826086, "grad_norm": 1.9432928562164307, "learning_rate": 3e-05, "loss": 2.9711, "step": 1330 }, { "epoch": 0.36168478260869563, "grad_norm": 2.222259998321533, "learning_rate": 3e-05, "loss": 3.5956, "step": 1331 }, { "epoch": 0.36195652173913045, "grad_norm": 1.0173529386520386, "learning_rate": 3e-05, "loss": 3.1368, "step": 1332 }, { "epoch": 0.3622282608695652, "grad_norm": 1.3840457201004028, "learning_rate": 3e-05, "loss": 3.3336, "step": 1333 }, { "epoch": 0.3625, "grad_norm": 2.199570417404175, "learning_rate": 3e-05, "loss": 3.1846, "step": 1334 }, { "epoch": 0.36277173913043476, "grad_norm": 2.3063576221466064, "learning_rate": 3e-05, "loss": 3.1996, "step": 1335 }, { "epoch": 0.3630434782608696, "grad_norm": 2.0329015254974365, "learning_rate": 3e-05, "loss": 3.6631, "step": 1336 }, { "epoch": 0.36331521739130435, "grad_norm": 0.9080404043197632, "learning_rate": 3e-05, "loss": 2.8746, "step": 1337 }, { "epoch": 0.3635869565217391, "grad_norm": 2.173773765563965, "learning_rate": 3e-05, "loss": 3.8486, "step": 1338 }, { "epoch": 0.36385869565217394, "grad_norm": 2.493887186050415, "learning_rate": 3e-05, "loss": 3.5707, "step": 1339 }, { "epoch": 0.3641304347826087, "grad_norm": 1.8862491846084595, "learning_rate": 3e-05, "loss": 3.8585, "step": 1340 }, { "epoch": 0.36440217391304347, "grad_norm": 1.23379647731781, "learning_rate": 3e-05, "loss": 3.5806, "step": 1341 }, { "epoch": 0.36467391304347824, "grad_norm": 1.3684219121932983, "learning_rate": 3e-05, "loss": 3.1318, "step": 1342 }, { "epoch": 0.36494565217391306, "grad_norm": 1.2184761762619019, "learning_rate": 3e-05, "loss": 3.4761, "step": 1343 }, { "epoch": 0.3652173913043478, "grad_norm": 1.230040431022644, "learning_rate": 3e-05, "loss": 4.1053, "step": 1344 }, { "epoch": 0.3654891304347826, "grad_norm": 1.4862629175186157, "learning_rate": 3e-05, "loss": 3.4211, "step": 1345 }, { "epoch": 0.3657608695652174, "grad_norm": 1.2832411527633667, "learning_rate": 3e-05, "loss": 2.7821, "step": 1346 }, { "epoch": 0.3660326086956522, "grad_norm": 1.1333979368209839, "learning_rate": 3e-05, "loss": 3.4129, "step": 1347 }, { "epoch": 0.36630434782608695, "grad_norm": 1.9462121725082397, "learning_rate": 3e-05, "loss": 3.8102, "step": 1348 }, { "epoch": 0.3665760869565217, "grad_norm": 1.314512848854065, "learning_rate": 3e-05, "loss": 3.5802, "step": 1349 }, { "epoch": 0.36684782608695654, "grad_norm": 0.9899564981460571, "learning_rate": 3e-05, "loss": 3.2397, "step": 1350 }, { "epoch": 0.3671195652173913, "grad_norm": 0.9129654765129089, "learning_rate": 3e-05, "loss": 3.0704, "step": 1351 }, { "epoch": 0.3673913043478261, "grad_norm": 1.2658374309539795, "learning_rate": 3e-05, "loss": 2.9416, "step": 1352 }, { "epoch": 0.36766304347826084, "grad_norm": 1.156530499458313, "learning_rate": 3e-05, "loss": 3.2254, "step": 1353 }, { "epoch": 0.36793478260869567, "grad_norm": 1.4215689897537231, "learning_rate": 3e-05, "loss": 3.5928, "step": 1354 }, { "epoch": 0.36820652173913043, "grad_norm": 1.4695452451705933, "learning_rate": 3e-05, "loss": 3.5298, "step": 1355 }, { "epoch": 0.3684782608695652, "grad_norm": 1.185398817062378, "learning_rate": 3e-05, "loss": 2.9786, "step": 1356 }, { "epoch": 0.36875, "grad_norm": 1.6181942224502563, "learning_rate": 3e-05, "loss": 3.8328, "step": 1357 }, { "epoch": 0.3690217391304348, "grad_norm": 1.7642178535461426, "learning_rate": 3e-05, "loss": 3.6084, "step": 1358 }, { "epoch": 0.36929347826086956, "grad_norm": 1.3343443870544434, "learning_rate": 3e-05, "loss": 4.0491, "step": 1359 }, { "epoch": 0.3695652173913043, "grad_norm": 2.167877435684204, "learning_rate": 3e-05, "loss": 3.6741, "step": 1360 }, { "epoch": 0.36983695652173915, "grad_norm": 2.0322484970092773, "learning_rate": 3e-05, "loss": 4.1298, "step": 1361 }, { "epoch": 0.3701086956521739, "grad_norm": 1.1205507516860962, "learning_rate": 3e-05, "loss": 3.2478, "step": 1362 }, { "epoch": 0.3703804347826087, "grad_norm": 1.4173895120620728, "learning_rate": 3e-05, "loss": 3.2024, "step": 1363 }, { "epoch": 0.3706521739130435, "grad_norm": 1.4186296463012695, "learning_rate": 3e-05, "loss": 2.9463, "step": 1364 }, { "epoch": 0.37092391304347827, "grad_norm": 1.4897836446762085, "learning_rate": 3e-05, "loss": 3.4138, "step": 1365 }, { "epoch": 0.37119565217391304, "grad_norm": 0.9213505983352661, "learning_rate": 3e-05, "loss": 2.7726, "step": 1366 }, { "epoch": 0.3714673913043478, "grad_norm": 1.4468334913253784, "learning_rate": 3e-05, "loss": 3.3552, "step": 1367 }, { "epoch": 0.3717391304347826, "grad_norm": 1.6020596027374268, "learning_rate": 3e-05, "loss": 3.3098, "step": 1368 }, { "epoch": 0.3720108695652174, "grad_norm": 1.2568944692611694, "learning_rate": 3e-05, "loss": 2.9896, "step": 1369 }, { "epoch": 0.37228260869565216, "grad_norm": 1.2092691659927368, "learning_rate": 3e-05, "loss": 3.2886, "step": 1370 }, { "epoch": 0.37255434782608693, "grad_norm": 1.0227844715118408, "learning_rate": 3e-05, "loss": 2.9555, "step": 1371 }, { "epoch": 0.37282608695652175, "grad_norm": 1.400106430053711, "learning_rate": 3e-05, "loss": 3.2697, "step": 1372 }, { "epoch": 0.3730978260869565, "grad_norm": 1.3000365495681763, "learning_rate": 3e-05, "loss": 3.1748, "step": 1373 }, { "epoch": 0.3733695652173913, "grad_norm": 1.197072148323059, "learning_rate": 3e-05, "loss": 3.3346, "step": 1374 }, { "epoch": 0.3736413043478261, "grad_norm": 1.1778289079666138, "learning_rate": 3e-05, "loss": 3.089, "step": 1375 }, { "epoch": 0.3739130434782609, "grad_norm": 1.23442542552948, "learning_rate": 3e-05, "loss": 3.0972, "step": 1376 }, { "epoch": 0.37418478260869564, "grad_norm": 1.135533094406128, "learning_rate": 3e-05, "loss": 2.9638, "step": 1377 }, { "epoch": 0.3744565217391304, "grad_norm": 1.4869073629379272, "learning_rate": 3e-05, "loss": 3.7439, "step": 1378 }, { "epoch": 0.37472826086956523, "grad_norm": 1.5528923273086548, "learning_rate": 3e-05, "loss": 2.8187, "step": 1379 }, { "epoch": 0.375, "grad_norm": 1.0053822994232178, "learning_rate": 3e-05, "loss": 3.2647, "step": 1380 }, { "epoch": 0.37527173913043477, "grad_norm": 1.2217366695404053, "learning_rate": 3e-05, "loss": 3.5844, "step": 1381 }, { "epoch": 0.3755434782608696, "grad_norm": 1.245322823524475, "learning_rate": 3e-05, "loss": 2.9917, "step": 1382 }, { "epoch": 0.37581521739130436, "grad_norm": 1.3003109693527222, "learning_rate": 3e-05, "loss": 3.7163, "step": 1383 }, { "epoch": 0.3760869565217391, "grad_norm": 2.398679733276367, "learning_rate": 3e-05, "loss": 4.2985, "step": 1384 }, { "epoch": 0.3763586956521739, "grad_norm": 1.2174817323684692, "learning_rate": 3e-05, "loss": 3.3311, "step": 1385 }, { "epoch": 0.3766304347826087, "grad_norm": 1.2073396444320679, "learning_rate": 3e-05, "loss": 3.4313, "step": 1386 }, { "epoch": 0.3769021739130435, "grad_norm": 1.2685997486114502, "learning_rate": 3e-05, "loss": 3.1358, "step": 1387 }, { "epoch": 0.37717391304347825, "grad_norm": 1.3566850423812866, "learning_rate": 3e-05, "loss": 3.0956, "step": 1388 }, { "epoch": 0.37744565217391307, "grad_norm": 1.6564099788665771, "learning_rate": 3e-05, "loss": 3.7791, "step": 1389 }, { "epoch": 0.37771739130434784, "grad_norm": 1.3698561191558838, "learning_rate": 3e-05, "loss": 3.3446, "step": 1390 }, { "epoch": 0.3779891304347826, "grad_norm": 1.2702010869979858, "learning_rate": 3e-05, "loss": 4.0951, "step": 1391 }, { "epoch": 0.3782608695652174, "grad_norm": 1.4307619333267212, "learning_rate": 3e-05, "loss": 3.3821, "step": 1392 }, { "epoch": 0.3785326086956522, "grad_norm": 1.207396149635315, "learning_rate": 3e-05, "loss": 3.6105, "step": 1393 }, { "epoch": 0.37880434782608696, "grad_norm": 1.1929562091827393, "learning_rate": 3e-05, "loss": 2.8538, "step": 1394 }, { "epoch": 0.37907608695652173, "grad_norm": 1.2328208684921265, "learning_rate": 3e-05, "loss": 3.0622, "step": 1395 }, { "epoch": 0.3793478260869565, "grad_norm": 1.4275541305541992, "learning_rate": 3e-05, "loss": 3.4482, "step": 1396 }, { "epoch": 0.3796195652173913, "grad_norm": 0.9579852223396301, "learning_rate": 3e-05, "loss": 2.8646, "step": 1397 }, { "epoch": 0.3798913043478261, "grad_norm": 1.6573379039764404, "learning_rate": 3e-05, "loss": 3.4417, "step": 1398 }, { "epoch": 0.38016304347826085, "grad_norm": 1.5388156175613403, "learning_rate": 3e-05, "loss": 3.2041, "step": 1399 }, { "epoch": 0.3804347826086957, "grad_norm": 1.4631450176239014, "learning_rate": 3e-05, "loss": 3.8626, "step": 1400 }, { "epoch": 0.38070652173913044, "grad_norm": 1.3570584058761597, "learning_rate": 3e-05, "loss": 3.441, "step": 1401 }, { "epoch": 0.3809782608695652, "grad_norm": 1.8016875982284546, "learning_rate": 3e-05, "loss": 3.1616, "step": 1402 }, { "epoch": 0.38125, "grad_norm": 1.4279495477676392, "learning_rate": 3e-05, "loss": 3.6129, "step": 1403 }, { "epoch": 0.3815217391304348, "grad_norm": 1.3305083513259888, "learning_rate": 3e-05, "loss": 3.3956, "step": 1404 }, { "epoch": 0.38179347826086957, "grad_norm": 1.7289437055587769, "learning_rate": 3e-05, "loss": 3.6499, "step": 1405 }, { "epoch": 0.38206521739130433, "grad_norm": 1.75969398021698, "learning_rate": 3e-05, "loss": 3.104, "step": 1406 }, { "epoch": 0.38233695652173916, "grad_norm": 1.545464277267456, "learning_rate": 3e-05, "loss": 4.1157, "step": 1407 }, { "epoch": 0.3826086956521739, "grad_norm": 1.5299899578094482, "learning_rate": 3e-05, "loss": 2.9975, "step": 1408 }, { "epoch": 0.3828804347826087, "grad_norm": 1.7116879224777222, "learning_rate": 3e-05, "loss": 3.0605, "step": 1409 }, { "epoch": 0.38315217391304346, "grad_norm": 1.4728469848632812, "learning_rate": 3e-05, "loss": 3.2079, "step": 1410 }, { "epoch": 0.3834239130434783, "grad_norm": 2.216775894165039, "learning_rate": 3e-05, "loss": 3.8937, "step": 1411 }, { "epoch": 0.38369565217391305, "grad_norm": 1.6279774904251099, "learning_rate": 3e-05, "loss": 3.7957, "step": 1412 }, { "epoch": 0.3839673913043478, "grad_norm": 0.9390795230865479, "learning_rate": 3e-05, "loss": 2.8752, "step": 1413 }, { "epoch": 0.3842391304347826, "grad_norm": 1.3048573732376099, "learning_rate": 3e-05, "loss": 3.0132, "step": 1414 }, { "epoch": 0.3845108695652174, "grad_norm": 1.132607340812683, "learning_rate": 3e-05, "loss": 2.965, "step": 1415 }, { "epoch": 0.3847826086956522, "grad_norm": 1.6570241451263428, "learning_rate": 3e-05, "loss": 3.1474, "step": 1416 }, { "epoch": 0.38505434782608694, "grad_norm": 1.5501550436019897, "learning_rate": 3e-05, "loss": 3.4049, "step": 1417 }, { "epoch": 0.38532608695652176, "grad_norm": 1.1150649785995483, "learning_rate": 3e-05, "loss": 3.4114, "step": 1418 }, { "epoch": 0.38559782608695653, "grad_norm": 1.3886046409606934, "learning_rate": 3e-05, "loss": 3.7872, "step": 1419 }, { "epoch": 0.3858695652173913, "grad_norm": 1.178165078163147, "learning_rate": 3e-05, "loss": 3.1258, "step": 1420 }, { "epoch": 0.38614130434782606, "grad_norm": 1.2557275295257568, "learning_rate": 3e-05, "loss": 3.4292, "step": 1421 }, { "epoch": 0.3864130434782609, "grad_norm": 0.9806421995162964, "learning_rate": 3e-05, "loss": 2.8931, "step": 1422 }, { "epoch": 0.38668478260869565, "grad_norm": 1.6238960027694702, "learning_rate": 3e-05, "loss": 3.5711, "step": 1423 }, { "epoch": 0.3869565217391304, "grad_norm": 1.3578873872756958, "learning_rate": 3e-05, "loss": 2.9015, "step": 1424 }, { "epoch": 0.38722826086956524, "grad_norm": 1.2328130006790161, "learning_rate": 3e-05, "loss": 3.2193, "step": 1425 }, { "epoch": 0.3875, "grad_norm": 2.146702289581299, "learning_rate": 3e-05, "loss": 4.0931, "step": 1426 }, { "epoch": 0.3877717391304348, "grad_norm": 1.9415775537490845, "learning_rate": 3e-05, "loss": 3.0855, "step": 1427 }, { "epoch": 0.38804347826086955, "grad_norm": 1.7980471849441528, "learning_rate": 3e-05, "loss": 3.3307, "step": 1428 }, { "epoch": 0.38831521739130437, "grad_norm": 1.429101824760437, "learning_rate": 3e-05, "loss": 3.1628, "step": 1429 }, { "epoch": 0.38858695652173914, "grad_norm": 2.225107431411743, "learning_rate": 3e-05, "loss": 3.5771, "step": 1430 }, { "epoch": 0.3888586956521739, "grad_norm": 2.46109938621521, "learning_rate": 3e-05, "loss": 3.7408, "step": 1431 }, { "epoch": 0.38913043478260867, "grad_norm": 2.695655345916748, "learning_rate": 3e-05, "loss": 3.4776, "step": 1432 }, { "epoch": 0.3894021739130435, "grad_norm": 2.3022725582122803, "learning_rate": 3e-05, "loss": 3.4633, "step": 1433 }, { "epoch": 0.38967391304347826, "grad_norm": 1.5871728658676147, "learning_rate": 3e-05, "loss": 2.8911, "step": 1434 }, { "epoch": 0.389945652173913, "grad_norm": 1.3922131061553955, "learning_rate": 3e-05, "loss": 2.7701, "step": 1435 }, { "epoch": 0.39021739130434785, "grad_norm": 2.131774663925171, "learning_rate": 3e-05, "loss": 3.6811, "step": 1436 }, { "epoch": 0.3904891304347826, "grad_norm": 1.8943536281585693, "learning_rate": 3e-05, "loss": 2.6957, "step": 1437 }, { "epoch": 0.3907608695652174, "grad_norm": 1.7963645458221436, "learning_rate": 3e-05, "loss": 3.4574, "step": 1438 }, { "epoch": 0.39103260869565215, "grad_norm": 1.613253116607666, "learning_rate": 3e-05, "loss": 3.5328, "step": 1439 }, { "epoch": 0.391304347826087, "grad_norm": 1.8886051177978516, "learning_rate": 3e-05, "loss": 2.7514, "step": 1440 }, { "epoch": 0.39157608695652174, "grad_norm": 2.6512999534606934, "learning_rate": 3e-05, "loss": 3.9306, "step": 1441 }, { "epoch": 0.3918478260869565, "grad_norm": 1.6623197793960571, "learning_rate": 3e-05, "loss": 3.4983, "step": 1442 }, { "epoch": 0.39211956521739133, "grad_norm": 1.2289677858352661, "learning_rate": 3e-05, "loss": 3.1967, "step": 1443 }, { "epoch": 0.3923913043478261, "grad_norm": 0.9220126867294312, "learning_rate": 3e-05, "loss": 2.8288, "step": 1444 }, { "epoch": 0.39266304347826086, "grad_norm": 1.853948950767517, "learning_rate": 3e-05, "loss": 3.2876, "step": 1445 }, { "epoch": 0.39293478260869563, "grad_norm": 2.040001392364502, "learning_rate": 3e-05, "loss": 3.6805, "step": 1446 }, { "epoch": 0.39320652173913045, "grad_norm": 1.4386035203933716, "learning_rate": 3e-05, "loss": 3.1173, "step": 1447 }, { "epoch": 0.3934782608695652, "grad_norm": 1.48452889919281, "learning_rate": 3e-05, "loss": 3.3276, "step": 1448 }, { "epoch": 0.39375, "grad_norm": 1.058858871459961, "learning_rate": 3e-05, "loss": 2.5619, "step": 1449 }, { "epoch": 0.39402173913043476, "grad_norm": 2.131944179534912, "learning_rate": 3e-05, "loss": 3.7079, "step": 1450 }, { "epoch": 0.3942934782608696, "grad_norm": 1.5783636569976807, "learning_rate": 3e-05, "loss": 3.5856, "step": 1451 }, { "epoch": 0.39456521739130435, "grad_norm": 1.249036192893982, "learning_rate": 3e-05, "loss": 3.0789, "step": 1452 }, { "epoch": 0.3948369565217391, "grad_norm": 1.6562423706054688, "learning_rate": 3e-05, "loss": 3.9025, "step": 1453 }, { "epoch": 0.39510869565217394, "grad_norm": 1.6861488819122314, "learning_rate": 3e-05, "loss": 3.5286, "step": 1454 }, { "epoch": 0.3953804347826087, "grad_norm": 1.503812313079834, "learning_rate": 3e-05, "loss": 3.0909, "step": 1455 }, { "epoch": 0.39565217391304347, "grad_norm": 1.8524895906448364, "learning_rate": 3e-05, "loss": 3.4152, "step": 1456 }, { "epoch": 0.39592391304347824, "grad_norm": 1.3550132513046265, "learning_rate": 3e-05, "loss": 3.5444, "step": 1457 }, { "epoch": 0.39619565217391306, "grad_norm": 1.454416036605835, "learning_rate": 3e-05, "loss": 3.6969, "step": 1458 }, { "epoch": 0.3964673913043478, "grad_norm": 1.5491242408752441, "learning_rate": 3e-05, "loss": 3.4632, "step": 1459 }, { "epoch": 0.3967391304347826, "grad_norm": 1.177980661392212, "learning_rate": 3e-05, "loss": 3.2844, "step": 1460 }, { "epoch": 0.3970108695652174, "grad_norm": 1.2985824346542358, "learning_rate": 3e-05, "loss": 3.6743, "step": 1461 }, { "epoch": 0.3972826086956522, "grad_norm": 1.545882225036621, "learning_rate": 3e-05, "loss": 3.4353, "step": 1462 }, { "epoch": 0.39755434782608695, "grad_norm": 1.3580163717269897, "learning_rate": 3e-05, "loss": 3.7926, "step": 1463 }, { "epoch": 0.3978260869565217, "grad_norm": 1.4254475831985474, "learning_rate": 3e-05, "loss": 3.7358, "step": 1464 }, { "epoch": 0.39809782608695654, "grad_norm": 1.2237005233764648, "learning_rate": 3e-05, "loss": 3.3169, "step": 1465 }, { "epoch": 0.3983695652173913, "grad_norm": 1.1622488498687744, "learning_rate": 3e-05, "loss": 3.1261, "step": 1466 }, { "epoch": 0.3986413043478261, "grad_norm": 0.9447959065437317, "learning_rate": 3e-05, "loss": 2.8757, "step": 1467 }, { "epoch": 0.39891304347826084, "grad_norm": 1.5178827047348022, "learning_rate": 3e-05, "loss": 3.7686, "step": 1468 }, { "epoch": 0.39918478260869567, "grad_norm": 1.2081400156021118, "learning_rate": 3e-05, "loss": 2.7978, "step": 1469 }, { "epoch": 0.39945652173913043, "grad_norm": 1.3333414793014526, "learning_rate": 3e-05, "loss": 3.8005, "step": 1470 }, { "epoch": 0.3997282608695652, "grad_norm": 1.2284449338912964, "learning_rate": 3e-05, "loss": 3.2384, "step": 1471 }, { "epoch": 0.4, "grad_norm": 1.6261060237884521, "learning_rate": 3e-05, "loss": 3.6663, "step": 1472 }, { "epoch": 0.4002717391304348, "grad_norm": 1.2063846588134766, "learning_rate": 3e-05, "loss": 3.2808, "step": 1473 }, { "epoch": 0.40054347826086956, "grad_norm": 1.1668663024902344, "learning_rate": 3e-05, "loss": 3.6527, "step": 1474 }, { "epoch": 0.4008152173913043, "grad_norm": 1.1682161092758179, "learning_rate": 3e-05, "loss": 3.0487, "step": 1475 }, { "epoch": 0.40108695652173915, "grad_norm": 1.2479826211929321, "learning_rate": 3e-05, "loss": 3.8231, "step": 1476 }, { "epoch": 0.4013586956521739, "grad_norm": 1.3903083801269531, "learning_rate": 3e-05, "loss": 3.3589, "step": 1477 }, { "epoch": 0.4016304347826087, "grad_norm": 2.756995677947998, "learning_rate": 3e-05, "loss": 3.5956, "step": 1478 }, { "epoch": 0.4019021739130435, "grad_norm": 1.1125861406326294, "learning_rate": 3e-05, "loss": 2.9482, "step": 1479 }, { "epoch": 0.40217391304347827, "grad_norm": 1.7494776248931885, "learning_rate": 3e-05, "loss": 3.6373, "step": 1480 }, { "epoch": 0.40244565217391304, "grad_norm": 1.7383618354797363, "learning_rate": 3e-05, "loss": 3.5196, "step": 1481 }, { "epoch": 0.4027173913043478, "grad_norm": 2.0332143306732178, "learning_rate": 3e-05, "loss": 3.4659, "step": 1482 }, { "epoch": 0.4029891304347826, "grad_norm": 1.7549830675125122, "learning_rate": 3e-05, "loss": 3.6027, "step": 1483 }, { "epoch": 0.4032608695652174, "grad_norm": 1.4020575284957886, "learning_rate": 3e-05, "loss": 3.1451, "step": 1484 }, { "epoch": 0.40353260869565216, "grad_norm": 1.134787917137146, "learning_rate": 3e-05, "loss": 2.6124, "step": 1485 }, { "epoch": 0.40380434782608693, "grad_norm": 1.6569262742996216, "learning_rate": 3e-05, "loss": 3.8979, "step": 1486 }, { "epoch": 0.40407608695652175, "grad_norm": 1.7637388706207275, "learning_rate": 3e-05, "loss": 3.8109, "step": 1487 }, { "epoch": 0.4043478260869565, "grad_norm": 1.2442704439163208, "learning_rate": 3e-05, "loss": 3.6988, "step": 1488 }, { "epoch": 0.4046195652173913, "grad_norm": 1.0942063331604004, "learning_rate": 3e-05, "loss": 3.0464, "step": 1489 }, { "epoch": 0.4048913043478261, "grad_norm": 1.337386965751648, "learning_rate": 3e-05, "loss": 3.8127, "step": 1490 }, { "epoch": 0.4051630434782609, "grad_norm": 1.7307548522949219, "learning_rate": 3e-05, "loss": 3.3854, "step": 1491 }, { "epoch": 0.40543478260869564, "grad_norm": 3.1193931102752686, "learning_rate": 3e-05, "loss": 4.9053, "step": 1492 }, { "epoch": 0.4057065217391304, "grad_norm": 1.5950829982757568, "learning_rate": 3e-05, "loss": 3.6687, "step": 1493 }, { "epoch": 0.40597826086956523, "grad_norm": 1.6609991788864136, "learning_rate": 3e-05, "loss": 3.4985, "step": 1494 }, { "epoch": 0.40625, "grad_norm": 1.824634075164795, "learning_rate": 3e-05, "loss": 2.7576, "step": 1495 }, { "epoch": 0.40652173913043477, "grad_norm": 1.8175928592681885, "learning_rate": 3e-05, "loss": 3.7615, "step": 1496 }, { "epoch": 0.4067934782608696, "grad_norm": 1.3160499334335327, "learning_rate": 3e-05, "loss": 3.1647, "step": 1497 }, { "epoch": 0.40706521739130436, "grad_norm": 1.7744853496551514, "learning_rate": 3e-05, "loss": 3.0396, "step": 1498 }, { "epoch": 0.4073369565217391, "grad_norm": 2.0241026878356934, "learning_rate": 3e-05, "loss": 3.8311, "step": 1499 }, { "epoch": 0.4076086956521739, "grad_norm": 1.7789679765701294, "learning_rate": 3e-05, "loss": 2.9538, "step": 1500 }, { "epoch": 0.4078804347826087, "grad_norm": 1.4469610452651978, "learning_rate": 3e-05, "loss": 3.3083, "step": 1501 }, { "epoch": 0.4081521739130435, "grad_norm": 1.570556879043579, "learning_rate": 3e-05, "loss": 3.0793, "step": 1502 }, { "epoch": 0.40842391304347825, "grad_norm": 1.3964000940322876, "learning_rate": 3e-05, "loss": 2.9603, "step": 1503 }, { "epoch": 0.40869565217391307, "grad_norm": 1.1800280809402466, "learning_rate": 3e-05, "loss": 3.2305, "step": 1504 }, { "epoch": 0.40896739130434784, "grad_norm": 1.762847661972046, "learning_rate": 3e-05, "loss": 3.7564, "step": 1505 }, { "epoch": 0.4092391304347826, "grad_norm": 1.5233442783355713, "learning_rate": 3e-05, "loss": 3.2744, "step": 1506 }, { "epoch": 0.4095108695652174, "grad_norm": 1.9167475700378418, "learning_rate": 3e-05, "loss": 3.6962, "step": 1507 }, { "epoch": 0.4097826086956522, "grad_norm": 1.949634075164795, "learning_rate": 3e-05, "loss": 3.1084, "step": 1508 }, { "epoch": 0.41005434782608696, "grad_norm": 1.9483072757720947, "learning_rate": 3e-05, "loss": 3.1895, "step": 1509 }, { "epoch": 0.41032608695652173, "grad_norm": 1.368509292602539, "learning_rate": 3e-05, "loss": 2.6804, "step": 1510 }, { "epoch": 0.4105978260869565, "grad_norm": 1.6476247310638428, "learning_rate": 3e-05, "loss": 3.1565, "step": 1511 }, { "epoch": 0.4108695652173913, "grad_norm": 2.14841890335083, "learning_rate": 3e-05, "loss": 3.5723, "step": 1512 }, { "epoch": 0.4111413043478261, "grad_norm": 2.1899356842041016, "learning_rate": 3e-05, "loss": 3.9324, "step": 1513 }, { "epoch": 0.41141304347826085, "grad_norm": 1.971421480178833, "learning_rate": 3e-05, "loss": 3.3302, "step": 1514 }, { "epoch": 0.4116847826086957, "grad_norm": 1.555216908454895, "learning_rate": 3e-05, "loss": 2.8559, "step": 1515 }, { "epoch": 0.41195652173913044, "grad_norm": 2.025975227355957, "learning_rate": 3e-05, "loss": 3.8605, "step": 1516 }, { "epoch": 0.4122282608695652, "grad_norm": 1.609168291091919, "learning_rate": 3e-05, "loss": 3.521, "step": 1517 }, { "epoch": 0.4125, "grad_norm": 1.4011493921279907, "learning_rate": 3e-05, "loss": 2.905, "step": 1518 }, { "epoch": 0.4127717391304348, "grad_norm": 1.813499927520752, "learning_rate": 3e-05, "loss": 3.9861, "step": 1519 }, { "epoch": 0.41304347826086957, "grad_norm": 2.023961067199707, "learning_rate": 3e-05, "loss": 3.3814, "step": 1520 }, { "epoch": 0.41331521739130433, "grad_norm": 1.9356021881103516, "learning_rate": 3e-05, "loss": 2.9743, "step": 1521 }, { "epoch": 0.41358695652173916, "grad_norm": 1.6765419244766235, "learning_rate": 3e-05, "loss": 2.9504, "step": 1522 }, { "epoch": 0.4138586956521739, "grad_norm": 1.954119324684143, "learning_rate": 3e-05, "loss": 3.7254, "step": 1523 }, { "epoch": 0.4141304347826087, "grad_norm": 1.5637805461883545, "learning_rate": 3e-05, "loss": 3.8543, "step": 1524 }, { "epoch": 0.41440217391304346, "grad_norm": 1.175187587738037, "learning_rate": 3e-05, "loss": 3.2871, "step": 1525 }, { "epoch": 0.4146739130434783, "grad_norm": 1.3228040933609009, "learning_rate": 3e-05, "loss": 2.9662, "step": 1526 }, { "epoch": 0.41494565217391305, "grad_norm": 1.4776461124420166, "learning_rate": 3e-05, "loss": 2.9617, "step": 1527 }, { "epoch": 0.4152173913043478, "grad_norm": 1.5169886350631714, "learning_rate": 3e-05, "loss": 3.6288, "step": 1528 }, { "epoch": 0.4154891304347826, "grad_norm": 1.4785473346710205, "learning_rate": 3e-05, "loss": 3.0807, "step": 1529 }, { "epoch": 0.4157608695652174, "grad_norm": 1.4565258026123047, "learning_rate": 3e-05, "loss": 3.3099, "step": 1530 }, { "epoch": 0.4160326086956522, "grad_norm": 1.2732715606689453, "learning_rate": 3e-05, "loss": 3.2948, "step": 1531 }, { "epoch": 0.41630434782608694, "grad_norm": 1.0217174291610718, "learning_rate": 3e-05, "loss": 3.0084, "step": 1532 }, { "epoch": 0.41657608695652176, "grad_norm": 1.2822792530059814, "learning_rate": 3e-05, "loss": 2.9284, "step": 1533 }, { "epoch": 0.41684782608695653, "grad_norm": 1.641827940940857, "learning_rate": 3e-05, "loss": 3.4684, "step": 1534 }, { "epoch": 0.4171195652173913, "grad_norm": 1.5296562910079956, "learning_rate": 3e-05, "loss": 2.5836, "step": 1535 }, { "epoch": 0.41739130434782606, "grad_norm": 1.3570102453231812, "learning_rate": 3e-05, "loss": 3.1323, "step": 1536 }, { "epoch": 0.4176630434782609, "grad_norm": 0.9901162981987, "learning_rate": 3e-05, "loss": 2.7749, "step": 1537 }, { "epoch": 0.41793478260869565, "grad_norm": 1.2664462327957153, "learning_rate": 3e-05, "loss": 3.1213, "step": 1538 }, { "epoch": 0.4182065217391304, "grad_norm": 1.682426929473877, "learning_rate": 3e-05, "loss": 3.8292, "step": 1539 }, { "epoch": 0.41847826086956524, "grad_norm": 1.3441531658172607, "learning_rate": 3e-05, "loss": 2.8989, "step": 1540 }, { "epoch": 0.41875, "grad_norm": 1.6413555145263672, "learning_rate": 3e-05, "loss": 3.4333, "step": 1541 }, { "epoch": 0.4190217391304348, "grad_norm": 1.2601666450500488, "learning_rate": 3e-05, "loss": 3.2815, "step": 1542 }, { "epoch": 0.41929347826086955, "grad_norm": 1.5582549571990967, "learning_rate": 3e-05, "loss": 3.5814, "step": 1543 }, { "epoch": 0.41956521739130437, "grad_norm": 1.0220012664794922, "learning_rate": 3e-05, "loss": 2.7242, "step": 1544 }, { "epoch": 0.41983695652173914, "grad_norm": 1.710802674293518, "learning_rate": 3e-05, "loss": 3.5271, "step": 1545 }, { "epoch": 0.4201086956521739, "grad_norm": 2.3148810863494873, "learning_rate": 3e-05, "loss": 4.0931, "step": 1546 }, { "epoch": 0.42038043478260867, "grad_norm": 1.4409066438674927, "learning_rate": 3e-05, "loss": 3.4806, "step": 1547 }, { "epoch": 0.4206521739130435, "grad_norm": 1.4869251251220703, "learning_rate": 3e-05, "loss": 3.8039, "step": 1548 }, { "epoch": 0.42092391304347826, "grad_norm": 2.2359511852264404, "learning_rate": 3e-05, "loss": 4.1842, "step": 1549 }, { "epoch": 0.421195652173913, "grad_norm": 2.2962846755981445, "learning_rate": 3e-05, "loss": 3.8559, "step": 1550 }, { "epoch": 0.42146739130434785, "grad_norm": 1.5050580501556396, "learning_rate": 3e-05, "loss": 3.6678, "step": 1551 }, { "epoch": 0.4217391304347826, "grad_norm": 1.2851321697235107, "learning_rate": 3e-05, "loss": 3.4175, "step": 1552 }, { "epoch": 0.4220108695652174, "grad_norm": 1.2671363353729248, "learning_rate": 3e-05, "loss": 3.0412, "step": 1553 }, { "epoch": 0.42228260869565215, "grad_norm": 1.2936915159225464, "learning_rate": 3e-05, "loss": 3.0918, "step": 1554 }, { "epoch": 0.422554347826087, "grad_norm": 1.4296749830245972, "learning_rate": 3e-05, "loss": 3.3257, "step": 1555 }, { "epoch": 0.42282608695652174, "grad_norm": 1.2652101516723633, "learning_rate": 3e-05, "loss": 3.5451, "step": 1556 }, { "epoch": 0.4230978260869565, "grad_norm": 1.4265552759170532, "learning_rate": 3e-05, "loss": 3.4145, "step": 1557 }, { "epoch": 0.42336956521739133, "grad_norm": 1.3775297403335571, "learning_rate": 3e-05, "loss": 3.3248, "step": 1558 }, { "epoch": 0.4236413043478261, "grad_norm": 1.8461849689483643, "learning_rate": 3e-05, "loss": 3.9212, "step": 1559 }, { "epoch": 0.42391304347826086, "grad_norm": 1.412374496459961, "learning_rate": 3e-05, "loss": 3.4737, "step": 1560 }, { "epoch": 0.42418478260869563, "grad_norm": 1.6272127628326416, "learning_rate": 3e-05, "loss": 3.4379, "step": 1561 }, { "epoch": 0.42445652173913045, "grad_norm": 1.754745364189148, "learning_rate": 3e-05, "loss": 3.4164, "step": 1562 }, { "epoch": 0.4247282608695652, "grad_norm": 1.2676855325698853, "learning_rate": 3e-05, "loss": 3.3106, "step": 1563 }, { "epoch": 0.425, "grad_norm": 1.7627284526824951, "learning_rate": 3e-05, "loss": 3.841, "step": 1564 }, { "epoch": 0.42527173913043476, "grad_norm": 1.143531084060669, "learning_rate": 3e-05, "loss": 2.82, "step": 1565 }, { "epoch": 0.4255434782608696, "grad_norm": 1.0254485607147217, "learning_rate": 3e-05, "loss": 2.9556, "step": 1566 }, { "epoch": 0.42581521739130435, "grad_norm": 1.3668675422668457, "learning_rate": 3e-05, "loss": 3.7105, "step": 1567 }, { "epoch": 0.4260869565217391, "grad_norm": 2.208592176437378, "learning_rate": 3e-05, "loss": 3.4152, "step": 1568 }, { "epoch": 0.42635869565217394, "grad_norm": 1.1941901445388794, "learning_rate": 3e-05, "loss": 3.3036, "step": 1569 }, { "epoch": 0.4266304347826087, "grad_norm": 1.536787748336792, "learning_rate": 3e-05, "loss": 3.5194, "step": 1570 }, { "epoch": 0.42690217391304347, "grad_norm": 1.4054354429244995, "learning_rate": 3e-05, "loss": 3.263, "step": 1571 }, { "epoch": 0.42717391304347824, "grad_norm": 1.6471023559570312, "learning_rate": 3e-05, "loss": 3.4855, "step": 1572 }, { "epoch": 0.42744565217391306, "grad_norm": 0.9407289028167725, "learning_rate": 3e-05, "loss": 3.1382, "step": 1573 }, { "epoch": 0.4277173913043478, "grad_norm": 1.244971513748169, "learning_rate": 3e-05, "loss": 3.4904, "step": 1574 }, { "epoch": 0.4279891304347826, "grad_norm": 1.305245041847229, "learning_rate": 3e-05, "loss": 3.3049, "step": 1575 }, { "epoch": 0.4282608695652174, "grad_norm": 1.2542614936828613, "learning_rate": 3e-05, "loss": 3.4931, "step": 1576 }, { "epoch": 0.4285326086956522, "grad_norm": 0.9759632349014282, "learning_rate": 3e-05, "loss": 3.4818, "step": 1577 }, { "epoch": 0.42880434782608695, "grad_norm": 1.4484260082244873, "learning_rate": 3e-05, "loss": 3.2934, "step": 1578 }, { "epoch": 0.4290760869565217, "grad_norm": 1.1898603439331055, "learning_rate": 3e-05, "loss": 3.4601, "step": 1579 }, { "epoch": 0.42934782608695654, "grad_norm": 1.486471176147461, "learning_rate": 3e-05, "loss": 3.762, "step": 1580 }, { "epoch": 0.4296195652173913, "grad_norm": 1.2455788850784302, "learning_rate": 3e-05, "loss": 3.721, "step": 1581 }, { "epoch": 0.4298913043478261, "grad_norm": 1.1194945573806763, "learning_rate": 3e-05, "loss": 3.3573, "step": 1582 }, { "epoch": 0.43016304347826084, "grad_norm": 1.666195273399353, "learning_rate": 3e-05, "loss": 3.531, "step": 1583 }, { "epoch": 0.43043478260869567, "grad_norm": 1.192292332649231, "learning_rate": 3e-05, "loss": 3.1313, "step": 1584 }, { "epoch": 0.43070652173913043, "grad_norm": 1.1080567836761475, "learning_rate": 3e-05, "loss": 3.2448, "step": 1585 }, { "epoch": 0.4309782608695652, "grad_norm": 1.7732088565826416, "learning_rate": 3e-05, "loss": 3.4527, "step": 1586 }, { "epoch": 0.43125, "grad_norm": 1.2465471029281616, "learning_rate": 3e-05, "loss": 2.8915, "step": 1587 }, { "epoch": 0.4315217391304348, "grad_norm": 1.43854820728302, "learning_rate": 3e-05, "loss": 3.2366, "step": 1588 }, { "epoch": 0.43179347826086956, "grad_norm": 1.3088682889938354, "learning_rate": 3e-05, "loss": 3.3654, "step": 1589 }, { "epoch": 0.4320652173913043, "grad_norm": 1.0148249864578247, "learning_rate": 3e-05, "loss": 3.083, "step": 1590 }, { "epoch": 0.43233695652173915, "grad_norm": 1.2219303846359253, "learning_rate": 3e-05, "loss": 3.0298, "step": 1591 }, { "epoch": 0.4326086956521739, "grad_norm": 1.081207275390625, "learning_rate": 3e-05, "loss": 3.2116, "step": 1592 }, { "epoch": 0.4328804347826087, "grad_norm": 1.515478491783142, "learning_rate": 3e-05, "loss": 3.4368, "step": 1593 }, { "epoch": 0.4331521739130435, "grad_norm": 1.073887586593628, "learning_rate": 3e-05, "loss": 3.1832, "step": 1594 }, { "epoch": 0.43342391304347827, "grad_norm": 1.1452287435531616, "learning_rate": 3e-05, "loss": 3.5617, "step": 1595 }, { "epoch": 0.43369565217391304, "grad_norm": 1.512242078781128, "learning_rate": 3e-05, "loss": 3.4468, "step": 1596 }, { "epoch": 0.4339673913043478, "grad_norm": 1.1862598657608032, "learning_rate": 3e-05, "loss": 3.3795, "step": 1597 }, { "epoch": 0.4342391304347826, "grad_norm": 1.4696996212005615, "learning_rate": 3e-05, "loss": 3.5239, "step": 1598 }, { "epoch": 0.4345108695652174, "grad_norm": 1.5985097885131836, "learning_rate": 3e-05, "loss": 3.8709, "step": 1599 }, { "epoch": 0.43478260869565216, "grad_norm": 1.3134711980819702, "learning_rate": 3e-05, "loss": 3.4582, "step": 1600 }, { "epoch": 0.43505434782608693, "grad_norm": 1.2656311988830566, "learning_rate": 3e-05, "loss": 3.7025, "step": 1601 }, { "epoch": 0.43532608695652175, "grad_norm": 1.1780740022659302, "learning_rate": 3e-05, "loss": 3.3486, "step": 1602 }, { "epoch": 0.4355978260869565, "grad_norm": 1.2048178911209106, "learning_rate": 3e-05, "loss": 3.3115, "step": 1603 }, { "epoch": 0.4358695652173913, "grad_norm": 1.3589476346969604, "learning_rate": 3e-05, "loss": 3.2494, "step": 1604 }, { "epoch": 0.4361413043478261, "grad_norm": 1.3171757459640503, "learning_rate": 3e-05, "loss": 2.8635, "step": 1605 }, { "epoch": 0.4364130434782609, "grad_norm": 1.1358094215393066, "learning_rate": 3e-05, "loss": 2.9426, "step": 1606 }, { "epoch": 0.43668478260869564, "grad_norm": 1.376583456993103, "learning_rate": 3e-05, "loss": 3.2611, "step": 1607 }, { "epoch": 0.4369565217391304, "grad_norm": 1.7452890872955322, "learning_rate": 3e-05, "loss": 3.7445, "step": 1608 }, { "epoch": 0.43722826086956523, "grad_norm": 1.715166687965393, "learning_rate": 3e-05, "loss": 3.4322, "step": 1609 }, { "epoch": 0.4375, "grad_norm": 1.3614475727081299, "learning_rate": 3e-05, "loss": 3.5131, "step": 1610 }, { "epoch": 0.43777173913043477, "grad_norm": 1.4811841249465942, "learning_rate": 3e-05, "loss": 3.714, "step": 1611 }, { "epoch": 0.4380434782608696, "grad_norm": 1.071926236152649, "learning_rate": 3e-05, "loss": 3.0734, "step": 1612 }, { "epoch": 0.43831521739130436, "grad_norm": 1.5618953704833984, "learning_rate": 3e-05, "loss": 3.876, "step": 1613 }, { "epoch": 0.4385869565217391, "grad_norm": 1.1414976119995117, "learning_rate": 3e-05, "loss": 3.5318, "step": 1614 }, { "epoch": 0.4388586956521739, "grad_norm": 1.0475927591323853, "learning_rate": 3e-05, "loss": 3.293, "step": 1615 }, { "epoch": 0.4391304347826087, "grad_norm": 1.392609715461731, "learning_rate": 3e-05, "loss": 2.9974, "step": 1616 }, { "epoch": 0.4394021739130435, "grad_norm": 0.906802237033844, "learning_rate": 3e-05, "loss": 2.9825, "step": 1617 }, { "epoch": 0.43967391304347825, "grad_norm": 1.4019746780395508, "learning_rate": 3e-05, "loss": 3.8698, "step": 1618 }, { "epoch": 0.43994565217391307, "grad_norm": 1.321829915046692, "learning_rate": 3e-05, "loss": 3.7109, "step": 1619 }, { "epoch": 0.44021739130434784, "grad_norm": 1.0838751792907715, "learning_rate": 3e-05, "loss": 2.6205, "step": 1620 }, { "epoch": 0.4404891304347826, "grad_norm": 2.383323907852173, "learning_rate": 3e-05, "loss": 4.2267, "step": 1621 }, { "epoch": 0.4407608695652174, "grad_norm": 1.0130791664123535, "learning_rate": 3e-05, "loss": 2.788, "step": 1622 }, { "epoch": 0.4410326086956522, "grad_norm": 1.3985093832015991, "learning_rate": 3e-05, "loss": 3.1002, "step": 1623 }, { "epoch": 0.44130434782608696, "grad_norm": 1.3819383382797241, "learning_rate": 3e-05, "loss": 3.2219, "step": 1624 }, { "epoch": 0.44157608695652173, "grad_norm": 1.4096386432647705, "learning_rate": 3e-05, "loss": 3.6886, "step": 1625 }, { "epoch": 0.4418478260869565, "grad_norm": 1.4819231033325195, "learning_rate": 3e-05, "loss": 3.2191, "step": 1626 }, { "epoch": 0.4421195652173913, "grad_norm": 1.3344846963882446, "learning_rate": 3e-05, "loss": 3.3927, "step": 1627 }, { "epoch": 0.4423913043478261, "grad_norm": 1.469277024269104, "learning_rate": 3e-05, "loss": 3.4069, "step": 1628 }, { "epoch": 0.44266304347826085, "grad_norm": 1.0877147912979126, "learning_rate": 3e-05, "loss": 2.9891, "step": 1629 }, { "epoch": 0.4429347826086957, "grad_norm": 1.1775012016296387, "learning_rate": 3e-05, "loss": 3.3463, "step": 1630 }, { "epoch": 0.44320652173913044, "grad_norm": 1.240936517715454, "learning_rate": 3e-05, "loss": 3.5751, "step": 1631 }, { "epoch": 0.4434782608695652, "grad_norm": 0.877865195274353, "learning_rate": 3e-05, "loss": 2.7317, "step": 1632 }, { "epoch": 0.44375, "grad_norm": 1.1812602281570435, "learning_rate": 3e-05, "loss": 3.0307, "step": 1633 }, { "epoch": 0.4440217391304348, "grad_norm": 1.7113842964172363, "learning_rate": 3e-05, "loss": 4.1471, "step": 1634 }, { "epoch": 0.44429347826086957, "grad_norm": 1.1820313930511475, "learning_rate": 3e-05, "loss": 3.362, "step": 1635 }, { "epoch": 0.44456521739130433, "grad_norm": 1.3378487825393677, "learning_rate": 3e-05, "loss": 3.5815, "step": 1636 }, { "epoch": 0.44483695652173916, "grad_norm": 0.8756458163261414, "learning_rate": 3e-05, "loss": 2.6882, "step": 1637 }, { "epoch": 0.4451086956521739, "grad_norm": 1.020820140838623, "learning_rate": 3e-05, "loss": 3.1006, "step": 1638 }, { "epoch": 0.4453804347826087, "grad_norm": 1.5075011253356934, "learning_rate": 3e-05, "loss": 4.0147, "step": 1639 }, { "epoch": 0.44565217391304346, "grad_norm": 1.1874289512634277, "learning_rate": 3e-05, "loss": 3.0811, "step": 1640 }, { "epoch": 0.4459239130434783, "grad_norm": 1.5765689611434937, "learning_rate": 3e-05, "loss": 3.4697, "step": 1641 }, { "epoch": 0.44619565217391305, "grad_norm": 1.3812084197998047, "learning_rate": 3e-05, "loss": 3.6468, "step": 1642 }, { "epoch": 0.4464673913043478, "grad_norm": 1.1477484703063965, "learning_rate": 3e-05, "loss": 3.2347, "step": 1643 }, { "epoch": 0.4467391304347826, "grad_norm": 1.431220531463623, "learning_rate": 3e-05, "loss": 3.8073, "step": 1644 }, { "epoch": 0.4470108695652174, "grad_norm": 1.0271748304367065, "learning_rate": 3e-05, "loss": 3.2165, "step": 1645 }, { "epoch": 0.4472826086956522, "grad_norm": 1.2913216352462769, "learning_rate": 3e-05, "loss": 3.2766, "step": 1646 }, { "epoch": 0.44755434782608694, "grad_norm": 1.341670274734497, "learning_rate": 3e-05, "loss": 3.8426, "step": 1647 }, { "epoch": 0.44782608695652176, "grad_norm": 1.1871728897094727, "learning_rate": 3e-05, "loss": 3.1251, "step": 1648 }, { "epoch": 0.44809782608695653, "grad_norm": 1.2151528596878052, "learning_rate": 3e-05, "loss": 3.3811, "step": 1649 }, { "epoch": 0.4483695652173913, "grad_norm": 1.7120990753173828, "learning_rate": 3e-05, "loss": 3.7093, "step": 1650 }, { "epoch": 0.44864130434782606, "grad_norm": 1.1526025533676147, "learning_rate": 3e-05, "loss": 3.1806, "step": 1651 }, { "epoch": 0.4489130434782609, "grad_norm": 1.5061256885528564, "learning_rate": 3e-05, "loss": 3.7246, "step": 1652 }, { "epoch": 0.44918478260869565, "grad_norm": 1.442798376083374, "learning_rate": 3e-05, "loss": 2.8975, "step": 1653 }, { "epoch": 0.4494565217391304, "grad_norm": 1.5733429193496704, "learning_rate": 3e-05, "loss": 3.7047, "step": 1654 }, { "epoch": 0.44972826086956524, "grad_norm": 1.5891181230545044, "learning_rate": 3e-05, "loss": 3.5587, "step": 1655 }, { "epoch": 0.45, "grad_norm": 2.2529733180999756, "learning_rate": 3e-05, "loss": 3.079, "step": 1656 }, { "epoch": 0.4502717391304348, "grad_norm": 2.397771120071411, "learning_rate": 3e-05, "loss": 3.2517, "step": 1657 }, { "epoch": 0.45054347826086955, "grad_norm": 1.5895618200302124, "learning_rate": 3e-05, "loss": 3.289, "step": 1658 }, { "epoch": 0.45081521739130437, "grad_norm": 1.3885931968688965, "learning_rate": 3e-05, "loss": 3.8401, "step": 1659 }, { "epoch": 0.45108695652173914, "grad_norm": 1.848649263381958, "learning_rate": 3e-05, "loss": 3.2375, "step": 1660 }, { "epoch": 0.4513586956521739, "grad_norm": 2.288536787033081, "learning_rate": 3e-05, "loss": 3.0737, "step": 1661 }, { "epoch": 0.45163043478260867, "grad_norm": 1.5626355409622192, "learning_rate": 3e-05, "loss": 3.6127, "step": 1662 }, { "epoch": 0.4519021739130435, "grad_norm": 1.2159314155578613, "learning_rate": 3e-05, "loss": 3.1681, "step": 1663 }, { "epoch": 0.45217391304347826, "grad_norm": 2.7480335235595703, "learning_rate": 3e-05, "loss": 3.3753, "step": 1664 }, { "epoch": 0.452445652173913, "grad_norm": 1.5772864818572998, "learning_rate": 3e-05, "loss": 3.4648, "step": 1665 }, { "epoch": 0.45271739130434785, "grad_norm": 2.1393795013427734, "learning_rate": 3e-05, "loss": 4.067, "step": 1666 }, { "epoch": 0.4529891304347826, "grad_norm": 1.48007071018219, "learning_rate": 3e-05, "loss": 3.0416, "step": 1667 }, { "epoch": 0.4532608695652174, "grad_norm": 1.3958046436309814, "learning_rate": 3e-05, "loss": 2.9871, "step": 1668 }, { "epoch": 0.45353260869565215, "grad_norm": 1.6530804634094238, "learning_rate": 3e-05, "loss": 3.1665, "step": 1669 }, { "epoch": 0.453804347826087, "grad_norm": 1.5180555582046509, "learning_rate": 3e-05, "loss": 3.1545, "step": 1670 }, { "epoch": 0.45407608695652174, "grad_norm": 1.3946945667266846, "learning_rate": 3e-05, "loss": 3.0606, "step": 1671 }, { "epoch": 0.4543478260869565, "grad_norm": 1.655504584312439, "learning_rate": 3e-05, "loss": 3.5141, "step": 1672 }, { "epoch": 0.45461956521739133, "grad_norm": 1.8746256828308105, "learning_rate": 3e-05, "loss": 3.2537, "step": 1673 }, { "epoch": 0.4548913043478261, "grad_norm": 1.6152852773666382, "learning_rate": 3e-05, "loss": 3.1725, "step": 1674 }, { "epoch": 0.45516304347826086, "grad_norm": 2.030758857727051, "learning_rate": 3e-05, "loss": 4.1911, "step": 1675 }, { "epoch": 0.45543478260869563, "grad_norm": 1.573011040687561, "learning_rate": 3e-05, "loss": 2.7004, "step": 1676 }, { "epoch": 0.45570652173913045, "grad_norm": 1.7782936096191406, "learning_rate": 3e-05, "loss": 3.4129, "step": 1677 }, { "epoch": 0.4559782608695652, "grad_norm": 1.6364409923553467, "learning_rate": 3e-05, "loss": 3.7828, "step": 1678 }, { "epoch": 0.45625, "grad_norm": 1.095005750656128, "learning_rate": 3e-05, "loss": 2.9575, "step": 1679 }, { "epoch": 0.45652173913043476, "grad_norm": 1.4323863983154297, "learning_rate": 3e-05, "loss": 2.9855, "step": 1680 }, { "epoch": 0.4567934782608696, "grad_norm": 1.8281795978546143, "learning_rate": 3e-05, "loss": 3.2274, "step": 1681 }, { "epoch": 0.45706521739130435, "grad_norm": 1.553568959236145, "learning_rate": 3e-05, "loss": 3.4245, "step": 1682 }, { "epoch": 0.4573369565217391, "grad_norm": 1.229791522026062, "learning_rate": 3e-05, "loss": 2.8749, "step": 1683 }, { "epoch": 0.45760869565217394, "grad_norm": 1.4264166355133057, "learning_rate": 3e-05, "loss": 3.0987, "step": 1684 }, { "epoch": 0.4578804347826087, "grad_norm": 1.4009737968444824, "learning_rate": 3e-05, "loss": 3.2024, "step": 1685 }, { "epoch": 0.45815217391304347, "grad_norm": 1.2461766004562378, "learning_rate": 3e-05, "loss": 3.1414, "step": 1686 }, { "epoch": 0.45842391304347824, "grad_norm": 1.1765496730804443, "learning_rate": 3e-05, "loss": 3.7237, "step": 1687 }, { "epoch": 0.45869565217391306, "grad_norm": 1.2187762260437012, "learning_rate": 3e-05, "loss": 3.5997, "step": 1688 }, { "epoch": 0.4589673913043478, "grad_norm": 2.117086410522461, "learning_rate": 3e-05, "loss": 3.5497, "step": 1689 }, { "epoch": 0.4592391304347826, "grad_norm": 1.2980027198791504, "learning_rate": 3e-05, "loss": 3.6001, "step": 1690 }, { "epoch": 0.4595108695652174, "grad_norm": 1.1031181812286377, "learning_rate": 3e-05, "loss": 2.814, "step": 1691 }, { "epoch": 0.4597826086956522, "grad_norm": 1.201127052307129, "learning_rate": 3e-05, "loss": 3.4627, "step": 1692 }, { "epoch": 0.46005434782608695, "grad_norm": 1.9975850582122803, "learning_rate": 3e-05, "loss": 3.8157, "step": 1693 }, { "epoch": 0.4603260869565217, "grad_norm": 1.4874250888824463, "learning_rate": 3e-05, "loss": 3.5901, "step": 1694 }, { "epoch": 0.46059782608695654, "grad_norm": 1.9747381210327148, "learning_rate": 3e-05, "loss": 3.8642, "step": 1695 }, { "epoch": 0.4608695652173913, "grad_norm": 1.405049443244934, "learning_rate": 3e-05, "loss": 3.3367, "step": 1696 }, { "epoch": 0.4611413043478261, "grad_norm": 1.4599591493606567, "learning_rate": 3e-05, "loss": 3.6499, "step": 1697 }, { "epoch": 0.46141304347826084, "grad_norm": 1.6565916538238525, "learning_rate": 3e-05, "loss": 3.5368, "step": 1698 }, { "epoch": 0.46168478260869567, "grad_norm": 1.327285647392273, "learning_rate": 3e-05, "loss": 2.7457, "step": 1699 }, { "epoch": 0.46195652173913043, "grad_norm": 1.317681074142456, "learning_rate": 3e-05, "loss": 3.3931, "step": 1700 }, { "epoch": 0.4622282608695652, "grad_norm": 1.7210628986358643, "learning_rate": 3e-05, "loss": 4.0996, "step": 1701 }, { "epoch": 0.4625, "grad_norm": 1.4575893878936768, "learning_rate": 3e-05, "loss": 3.5176, "step": 1702 }, { "epoch": 0.4627717391304348, "grad_norm": 0.8817341327667236, "learning_rate": 3e-05, "loss": 2.9138, "step": 1703 }, { "epoch": 0.46304347826086956, "grad_norm": 1.6268318891525269, "learning_rate": 3e-05, "loss": 3.3346, "step": 1704 }, { "epoch": 0.4633152173913043, "grad_norm": 1.281497836112976, "learning_rate": 3e-05, "loss": 2.776, "step": 1705 }, { "epoch": 0.46358695652173915, "grad_norm": 1.6672241687774658, "learning_rate": 3e-05, "loss": 3.5205, "step": 1706 }, { "epoch": 0.4638586956521739, "grad_norm": 1.4224839210510254, "learning_rate": 3e-05, "loss": 3.7986, "step": 1707 }, { "epoch": 0.4641304347826087, "grad_norm": 1.1987775564193726, "learning_rate": 3e-05, "loss": 3.2364, "step": 1708 }, { "epoch": 0.4644021739130435, "grad_norm": 1.4384889602661133, "learning_rate": 3e-05, "loss": 3.6339, "step": 1709 }, { "epoch": 0.46467391304347827, "grad_norm": 1.1359198093414307, "learning_rate": 3e-05, "loss": 3.2827, "step": 1710 }, { "epoch": 0.46494565217391304, "grad_norm": 1.3200854063034058, "learning_rate": 3e-05, "loss": 3.2209, "step": 1711 }, { "epoch": 0.4652173913043478, "grad_norm": 1.613688588142395, "learning_rate": 3e-05, "loss": 3.7578, "step": 1712 }, { "epoch": 0.4654891304347826, "grad_norm": 1.7866336107254028, "learning_rate": 3e-05, "loss": 3.9267, "step": 1713 }, { "epoch": 0.4657608695652174, "grad_norm": 2.0430612564086914, "learning_rate": 3e-05, "loss": 3.6663, "step": 1714 }, { "epoch": 0.46603260869565216, "grad_norm": 1.848181128501892, "learning_rate": 3e-05, "loss": 3.3196, "step": 1715 }, { "epoch": 0.46630434782608693, "grad_norm": 1.4686863422393799, "learning_rate": 3e-05, "loss": 3.352, "step": 1716 }, { "epoch": 0.46657608695652175, "grad_norm": 1.1621712446212769, "learning_rate": 3e-05, "loss": 3.1954, "step": 1717 }, { "epoch": 0.4668478260869565, "grad_norm": 1.2042642831802368, "learning_rate": 3e-05, "loss": 2.7059, "step": 1718 }, { "epoch": 0.4671195652173913, "grad_norm": 2.4052507877349854, "learning_rate": 3e-05, "loss": 3.6323, "step": 1719 }, { "epoch": 0.4673913043478261, "grad_norm": 0.8568060994148254, "learning_rate": 3e-05, "loss": 2.5241, "step": 1720 }, { "epoch": 0.4676630434782609, "grad_norm": 0.9677852988243103, "learning_rate": 3e-05, "loss": 2.5949, "step": 1721 }, { "epoch": 0.46793478260869564, "grad_norm": 1.2894710302352905, "learning_rate": 3e-05, "loss": 2.989, "step": 1722 }, { "epoch": 0.4682065217391304, "grad_norm": 1.5006341934204102, "learning_rate": 3e-05, "loss": 3.6919, "step": 1723 }, { "epoch": 0.46847826086956523, "grad_norm": 1.2537362575531006, "learning_rate": 3e-05, "loss": 3.3816, "step": 1724 }, { "epoch": 0.46875, "grad_norm": 1.333892822265625, "learning_rate": 3e-05, "loss": 3.3992, "step": 1725 }, { "epoch": 0.46902173913043477, "grad_norm": 1.1863936185836792, "learning_rate": 3e-05, "loss": 3.1168, "step": 1726 }, { "epoch": 0.4692934782608696, "grad_norm": 1.2577567100524902, "learning_rate": 3e-05, "loss": 3.4444, "step": 1727 }, { "epoch": 0.46956521739130436, "grad_norm": 1.2792398929595947, "learning_rate": 3e-05, "loss": 2.8464, "step": 1728 }, { "epoch": 0.4698369565217391, "grad_norm": 2.50295090675354, "learning_rate": 3e-05, "loss": 3.2523, "step": 1729 }, { "epoch": 0.4701086956521739, "grad_norm": 1.842974066734314, "learning_rate": 3e-05, "loss": 3.3193, "step": 1730 }, { "epoch": 0.4703804347826087, "grad_norm": 1.1899735927581787, "learning_rate": 3e-05, "loss": 3.4586, "step": 1731 }, { "epoch": 0.4706521739130435, "grad_norm": 1.7522422075271606, "learning_rate": 3e-05, "loss": 3.44, "step": 1732 }, { "epoch": 0.47092391304347825, "grad_norm": 2.0558695793151855, "learning_rate": 3e-05, "loss": 3.3303, "step": 1733 }, { "epoch": 0.47119565217391307, "grad_norm": 1.5954774618148804, "learning_rate": 3e-05, "loss": 3.3366, "step": 1734 }, { "epoch": 0.47146739130434784, "grad_norm": 1.4976911544799805, "learning_rate": 3e-05, "loss": 3.7435, "step": 1735 }, { "epoch": 0.4717391304347826, "grad_norm": 1.4843627214431763, "learning_rate": 3e-05, "loss": 3.0997, "step": 1736 }, { "epoch": 0.4720108695652174, "grad_norm": 1.7294570207595825, "learning_rate": 3e-05, "loss": 3.4683, "step": 1737 }, { "epoch": 0.4722826086956522, "grad_norm": 1.7918407917022705, "learning_rate": 3e-05, "loss": 3.5634, "step": 1738 }, { "epoch": 0.47255434782608696, "grad_norm": 1.458388090133667, "learning_rate": 3e-05, "loss": 3.724, "step": 1739 }, { "epoch": 0.47282608695652173, "grad_norm": 1.7183854579925537, "learning_rate": 3e-05, "loss": 3.7768, "step": 1740 }, { "epoch": 0.4730978260869565, "grad_norm": 1.6272084712982178, "learning_rate": 3e-05, "loss": 3.1361, "step": 1741 }, { "epoch": 0.4733695652173913, "grad_norm": 1.2989143133163452, "learning_rate": 3e-05, "loss": 3.1645, "step": 1742 }, { "epoch": 0.4736413043478261, "grad_norm": 1.1352357864379883, "learning_rate": 3e-05, "loss": 2.9178, "step": 1743 }, { "epoch": 0.47391304347826085, "grad_norm": 1.6457839012145996, "learning_rate": 3e-05, "loss": 3.8608, "step": 1744 }, { "epoch": 0.4741847826086957, "grad_norm": 1.3702744245529175, "learning_rate": 3e-05, "loss": 2.8434, "step": 1745 }, { "epoch": 0.47445652173913044, "grad_norm": 1.2821577787399292, "learning_rate": 3e-05, "loss": 3.0827, "step": 1746 }, { "epoch": 0.4747282608695652, "grad_norm": 1.4484645128250122, "learning_rate": 3e-05, "loss": 3.222, "step": 1747 }, { "epoch": 0.475, "grad_norm": 1.7668582201004028, "learning_rate": 3e-05, "loss": 3.2441, "step": 1748 }, { "epoch": 0.4752717391304348, "grad_norm": 1.764919400215149, "learning_rate": 3e-05, "loss": 3.7584, "step": 1749 }, { "epoch": 0.47554347826086957, "grad_norm": 1.640993356704712, "learning_rate": 3e-05, "loss": 3.3807, "step": 1750 }, { "epoch": 0.47581521739130433, "grad_norm": 1.6005651950836182, "learning_rate": 3e-05, "loss": 4.2452, "step": 1751 }, { "epoch": 0.47608695652173916, "grad_norm": 1.421935796737671, "learning_rate": 3e-05, "loss": 3.0033, "step": 1752 }, { "epoch": 0.4763586956521739, "grad_norm": 1.7603181600570679, "learning_rate": 3e-05, "loss": 3.2585, "step": 1753 }, { "epoch": 0.4766304347826087, "grad_norm": 1.378020167350769, "learning_rate": 3e-05, "loss": 3.0198, "step": 1754 }, { "epoch": 0.47690217391304346, "grad_norm": 1.5630027055740356, "learning_rate": 3e-05, "loss": 3.0409, "step": 1755 }, { "epoch": 0.4771739130434783, "grad_norm": 1.1110471487045288, "learning_rate": 3e-05, "loss": 3.2521, "step": 1756 }, { "epoch": 0.47744565217391305, "grad_norm": 1.403874397277832, "learning_rate": 3e-05, "loss": 2.9834, "step": 1757 }, { "epoch": 0.4777173913043478, "grad_norm": 1.2942219972610474, "learning_rate": 3e-05, "loss": 3.1724, "step": 1758 }, { "epoch": 0.4779891304347826, "grad_norm": 1.152891993522644, "learning_rate": 3e-05, "loss": 3.2869, "step": 1759 }, { "epoch": 0.4782608695652174, "grad_norm": 1.1676504611968994, "learning_rate": 3e-05, "loss": 3.1582, "step": 1760 }, { "epoch": 0.4785326086956522, "grad_norm": 1.151185154914856, "learning_rate": 3e-05, "loss": 3.3322, "step": 1761 }, { "epoch": 0.47880434782608694, "grad_norm": 1.4192887544631958, "learning_rate": 3e-05, "loss": 3.3659, "step": 1762 }, { "epoch": 0.47907608695652176, "grad_norm": 1.5391749143600464, "learning_rate": 3e-05, "loss": 3.2556, "step": 1763 }, { "epoch": 0.47934782608695653, "grad_norm": 1.4667121171951294, "learning_rate": 3e-05, "loss": 3.8051, "step": 1764 }, { "epoch": 0.4796195652173913, "grad_norm": 1.3416023254394531, "learning_rate": 3e-05, "loss": 3.4739, "step": 1765 }, { "epoch": 0.47989130434782606, "grad_norm": 1.7379714250564575, "learning_rate": 3e-05, "loss": 3.2991, "step": 1766 }, { "epoch": 0.4801630434782609, "grad_norm": 1.6660994291305542, "learning_rate": 3e-05, "loss": 3.1885, "step": 1767 }, { "epoch": 0.48043478260869565, "grad_norm": 1.9886304140090942, "learning_rate": 3e-05, "loss": 3.3295, "step": 1768 }, { "epoch": 0.4807065217391304, "grad_norm": 2.345217227935791, "learning_rate": 3e-05, "loss": 3.6796, "step": 1769 }, { "epoch": 0.48097826086956524, "grad_norm": 1.6064239740371704, "learning_rate": 3e-05, "loss": 2.977, "step": 1770 }, { "epoch": 0.48125, "grad_norm": 1.7928065061569214, "learning_rate": 3e-05, "loss": 3.5848, "step": 1771 }, { "epoch": 0.4815217391304348, "grad_norm": 1.5981751680374146, "learning_rate": 3e-05, "loss": 3.1049, "step": 1772 }, { "epoch": 0.48179347826086955, "grad_norm": 1.6446112394332886, "learning_rate": 3e-05, "loss": 2.9694, "step": 1773 }, { "epoch": 0.48206521739130437, "grad_norm": 1.8991427421569824, "learning_rate": 3e-05, "loss": 3.3139, "step": 1774 }, { "epoch": 0.48233695652173914, "grad_norm": 2.3716299533843994, "learning_rate": 3e-05, "loss": 3.9361, "step": 1775 }, { "epoch": 0.4826086956521739, "grad_norm": 1.8573355674743652, "learning_rate": 3e-05, "loss": 3.1394, "step": 1776 }, { "epoch": 0.48288043478260867, "grad_norm": 1.8625982999801636, "learning_rate": 3e-05, "loss": 3.8017, "step": 1777 }, { "epoch": 0.4831521739130435, "grad_norm": 2.2275497913360596, "learning_rate": 3e-05, "loss": 3.4565, "step": 1778 }, { "epoch": 0.48342391304347826, "grad_norm": 1.9856013059616089, "learning_rate": 3e-05, "loss": 3.5517, "step": 1779 }, { "epoch": 0.483695652173913, "grad_norm": 1.3510781526565552, "learning_rate": 3e-05, "loss": 3.1976, "step": 1780 }, { "epoch": 0.48396739130434785, "grad_norm": 1.0482627153396606, "learning_rate": 3e-05, "loss": 3.0025, "step": 1781 }, { "epoch": 0.4842391304347826, "grad_norm": 1.3749173879623413, "learning_rate": 3e-05, "loss": 3.2774, "step": 1782 }, { "epoch": 0.4845108695652174, "grad_norm": 1.405247449874878, "learning_rate": 3e-05, "loss": 3.4099, "step": 1783 }, { "epoch": 0.48478260869565215, "grad_norm": 1.2630794048309326, "learning_rate": 3e-05, "loss": 3.7092, "step": 1784 }, { "epoch": 0.485054347826087, "grad_norm": 1.1290115118026733, "learning_rate": 3e-05, "loss": 3.432, "step": 1785 }, { "epoch": 0.48532608695652174, "grad_norm": 1.2523903846740723, "learning_rate": 3e-05, "loss": 3.1431, "step": 1786 }, { "epoch": 0.4855978260869565, "grad_norm": 1.218867540359497, "learning_rate": 3e-05, "loss": 3.4881, "step": 1787 }, { "epoch": 0.48586956521739133, "grad_norm": 1.6144737005233765, "learning_rate": 3e-05, "loss": 4.0674, "step": 1788 }, { "epoch": 0.4861413043478261, "grad_norm": 1.4506441354751587, "learning_rate": 3e-05, "loss": 3.5498, "step": 1789 }, { "epoch": 0.48641304347826086, "grad_norm": 1.5907642841339111, "learning_rate": 3e-05, "loss": 3.1341, "step": 1790 }, { "epoch": 0.48668478260869563, "grad_norm": 1.18953275680542, "learning_rate": 3e-05, "loss": 3.0503, "step": 1791 }, { "epoch": 0.48695652173913045, "grad_norm": 1.2186987400054932, "learning_rate": 3e-05, "loss": 2.7508, "step": 1792 }, { "epoch": 0.4872282608695652, "grad_norm": 1.111574649810791, "learning_rate": 3e-05, "loss": 2.8275, "step": 1793 }, { "epoch": 0.4875, "grad_norm": 1.275896430015564, "learning_rate": 3e-05, "loss": 3.4576, "step": 1794 }, { "epoch": 0.48777173913043476, "grad_norm": 1.182626485824585, "learning_rate": 3e-05, "loss": 2.9678, "step": 1795 }, { "epoch": 0.4880434782608696, "grad_norm": 1.098120927810669, "learning_rate": 3e-05, "loss": 3.1544, "step": 1796 }, { "epoch": 0.48831521739130435, "grad_norm": 0.9524686932563782, "learning_rate": 3e-05, "loss": 3.0739, "step": 1797 }, { "epoch": 0.4885869565217391, "grad_norm": 1.302868366241455, "learning_rate": 3e-05, "loss": 3.3036, "step": 1798 }, { "epoch": 0.48885869565217394, "grad_norm": 1.124603271484375, "learning_rate": 3e-05, "loss": 3.2132, "step": 1799 }, { "epoch": 0.4891304347826087, "grad_norm": 1.3126977682113647, "learning_rate": 3e-05, "loss": 3.6832, "step": 1800 }, { "epoch": 0.48940217391304347, "grad_norm": 1.436482310295105, "learning_rate": 3e-05, "loss": 3.9607, "step": 1801 }, { "epoch": 0.48967391304347824, "grad_norm": 1.6329675912857056, "learning_rate": 3e-05, "loss": 2.6751, "step": 1802 }, { "epoch": 0.48994565217391306, "grad_norm": 1.2873584032058716, "learning_rate": 3e-05, "loss": 3.3793, "step": 1803 }, { "epoch": 0.4902173913043478, "grad_norm": 1.0539244413375854, "learning_rate": 3e-05, "loss": 3.0761, "step": 1804 }, { "epoch": 0.4904891304347826, "grad_norm": 1.7557803392410278, "learning_rate": 3e-05, "loss": 3.0488, "step": 1805 }, { "epoch": 0.4907608695652174, "grad_norm": 1.8407940864562988, "learning_rate": 3e-05, "loss": 3.1567, "step": 1806 }, { "epoch": 0.4910326086956522, "grad_norm": 1.3472580909729004, "learning_rate": 3e-05, "loss": 3.2331, "step": 1807 }, { "epoch": 0.49130434782608695, "grad_norm": 1.093360424041748, "learning_rate": 3e-05, "loss": 3.1883, "step": 1808 }, { "epoch": 0.4915760869565217, "grad_norm": 1.650109887123108, "learning_rate": 3e-05, "loss": 3.7661, "step": 1809 }, { "epoch": 0.49184782608695654, "grad_norm": 1.454372525215149, "learning_rate": 3e-05, "loss": 3.1205, "step": 1810 }, { "epoch": 0.4921195652173913, "grad_norm": 1.4370765686035156, "learning_rate": 3e-05, "loss": 3.8767, "step": 1811 }, { "epoch": 0.4923913043478261, "grad_norm": 2.6512532234191895, "learning_rate": 3e-05, "loss": 4.0366, "step": 1812 }, { "epoch": 0.49266304347826084, "grad_norm": 2.1347718238830566, "learning_rate": 3e-05, "loss": 3.5459, "step": 1813 }, { "epoch": 0.49293478260869567, "grad_norm": 1.5831135511398315, "learning_rate": 3e-05, "loss": 3.4828, "step": 1814 }, { "epoch": 0.49320652173913043, "grad_norm": 1.0648609399795532, "learning_rate": 3e-05, "loss": 3.0868, "step": 1815 }, { "epoch": 0.4934782608695652, "grad_norm": 1.467032551765442, "learning_rate": 3e-05, "loss": 3.6537, "step": 1816 }, { "epoch": 0.49375, "grad_norm": 1.2471240758895874, "learning_rate": 3e-05, "loss": 3.2671, "step": 1817 }, { "epoch": 0.4940217391304348, "grad_norm": 0.9745358228683472, "learning_rate": 3e-05, "loss": 3.4916, "step": 1818 }, { "epoch": 0.49429347826086956, "grad_norm": 1.3291053771972656, "learning_rate": 3e-05, "loss": 3.7441, "step": 1819 }, { "epoch": 0.4945652173913043, "grad_norm": 1.3420668840408325, "learning_rate": 3e-05, "loss": 3.3106, "step": 1820 }, { "epoch": 0.49483695652173915, "grad_norm": 1.0097370147705078, "learning_rate": 3e-05, "loss": 3.1215, "step": 1821 }, { "epoch": 0.4951086956521739, "grad_norm": 1.1169931888580322, "learning_rate": 3e-05, "loss": 2.9595, "step": 1822 }, { "epoch": 0.4953804347826087, "grad_norm": 1.1370187997817993, "learning_rate": 3e-05, "loss": 3.5224, "step": 1823 }, { "epoch": 0.4956521739130435, "grad_norm": 1.5470448732376099, "learning_rate": 3e-05, "loss": 3.9269, "step": 1824 }, { "epoch": 0.49592391304347827, "grad_norm": 1.000895380973816, "learning_rate": 3e-05, "loss": 2.9649, "step": 1825 }, { "epoch": 0.49619565217391304, "grad_norm": 1.2407017946243286, "learning_rate": 3e-05, "loss": 3.5568, "step": 1826 }, { "epoch": 0.4964673913043478, "grad_norm": 1.226485013961792, "learning_rate": 3e-05, "loss": 2.5796, "step": 1827 }, { "epoch": 0.4967391304347826, "grad_norm": 1.1970787048339844, "learning_rate": 3e-05, "loss": 2.8772, "step": 1828 }, { "epoch": 0.4970108695652174, "grad_norm": 1.2727575302124023, "learning_rate": 3e-05, "loss": 3.6501, "step": 1829 }, { "epoch": 0.49728260869565216, "grad_norm": 1.2288799285888672, "learning_rate": 3e-05, "loss": 3.0078, "step": 1830 }, { "epoch": 0.49755434782608693, "grad_norm": 1.688232660293579, "learning_rate": 3e-05, "loss": 3.2542, "step": 1831 }, { "epoch": 0.49782608695652175, "grad_norm": 1.6907511949539185, "learning_rate": 3e-05, "loss": 3.6193, "step": 1832 }, { "epoch": 0.4980978260869565, "grad_norm": 1.1391464471817017, "learning_rate": 3e-05, "loss": 3.3688, "step": 1833 }, { "epoch": 0.4983695652173913, "grad_norm": 1.2691822052001953, "learning_rate": 3e-05, "loss": 3.2763, "step": 1834 }, { "epoch": 0.4986413043478261, "grad_norm": 1.1176376342773438, "learning_rate": 3e-05, "loss": 3.0038, "step": 1835 }, { "epoch": 0.4989130434782609, "grad_norm": 0.9173082709312439, "learning_rate": 3e-05, "loss": 2.7976, "step": 1836 }, { "epoch": 0.49918478260869564, "grad_norm": 1.0989121198654175, "learning_rate": 3e-05, "loss": 3.0254, "step": 1837 }, { "epoch": 0.4994565217391304, "grad_norm": 1.2885278463363647, "learning_rate": 3e-05, "loss": 3.2596, "step": 1838 }, { "epoch": 0.49972826086956523, "grad_norm": 1.1294320821762085, "learning_rate": 3e-05, "loss": 3.0833, "step": 1839 }, { "epoch": 0.5, "grad_norm": 0.9139266014099121, "learning_rate": 3e-05, "loss": 3.043, "step": 1840 }, { "epoch": 0.5002717391304348, "grad_norm": 0.9999741315841675, "learning_rate": 3e-05, "loss": 2.6746, "step": 1841 }, { "epoch": 0.5005434782608695, "grad_norm": 2.3439218997955322, "learning_rate": 3e-05, "loss": 4.0552, "step": 1842 }, { "epoch": 0.5008152173913043, "grad_norm": 1.3185632228851318, "learning_rate": 3e-05, "loss": 3.4046, "step": 1843 }, { "epoch": 0.5010869565217392, "grad_norm": 1.4254107475280762, "learning_rate": 3e-05, "loss": 3.5323, "step": 1844 }, { "epoch": 0.501358695652174, "grad_norm": 1.0706696510314941, "learning_rate": 3e-05, "loss": 3.0672, "step": 1845 }, { "epoch": 0.5016304347826087, "grad_norm": 1.316158413887024, "learning_rate": 3e-05, "loss": 3.3709, "step": 1846 }, { "epoch": 0.5019021739130435, "grad_norm": 1.2480392456054688, "learning_rate": 3e-05, "loss": 3.673, "step": 1847 }, { "epoch": 0.5021739130434782, "grad_norm": 1.596869707107544, "learning_rate": 3e-05, "loss": 3.3651, "step": 1848 }, { "epoch": 0.502445652173913, "grad_norm": 1.6036261320114136, "learning_rate": 3e-05, "loss": 3.68, "step": 1849 }, { "epoch": 0.5027173913043478, "grad_norm": 1.2420178651809692, "learning_rate": 3e-05, "loss": 2.9173, "step": 1850 }, { "epoch": 0.5029891304347827, "grad_norm": 1.588339924812317, "learning_rate": 3e-05, "loss": 3.7931, "step": 1851 }, { "epoch": 0.5032608695652174, "grad_norm": 1.2808369398117065, "learning_rate": 3e-05, "loss": 3.05, "step": 1852 }, { "epoch": 0.5035326086956522, "grad_norm": 1.234026551246643, "learning_rate": 3e-05, "loss": 3.3401, "step": 1853 }, { "epoch": 0.503804347826087, "grad_norm": 1.6384221315383911, "learning_rate": 3e-05, "loss": 3.5597, "step": 1854 }, { "epoch": 0.5040760869565217, "grad_norm": 1.8604109287261963, "learning_rate": 3e-05, "loss": 3.6333, "step": 1855 }, { "epoch": 0.5043478260869565, "grad_norm": 1.4895154237747192, "learning_rate": 3e-05, "loss": 3.4255, "step": 1856 }, { "epoch": 0.5046195652173913, "grad_norm": 1.3410334587097168, "learning_rate": 3e-05, "loss": 2.7753, "step": 1857 }, { "epoch": 0.5048913043478261, "grad_norm": 1.446807861328125, "learning_rate": 3e-05, "loss": 4.1229, "step": 1858 }, { "epoch": 0.5051630434782609, "grad_norm": 1.188097596168518, "learning_rate": 3e-05, "loss": 3.2846, "step": 1859 }, { "epoch": 0.5054347826086957, "grad_norm": 1.2120987176895142, "learning_rate": 3e-05, "loss": 3.147, "step": 1860 }, { "epoch": 0.5057065217391304, "grad_norm": 1.5114277601242065, "learning_rate": 3e-05, "loss": 3.0952, "step": 1861 }, { "epoch": 0.5059782608695652, "grad_norm": 1.360628366470337, "learning_rate": 3e-05, "loss": 2.907, "step": 1862 }, { "epoch": 0.50625, "grad_norm": 1.4083906412124634, "learning_rate": 3e-05, "loss": 3.4222, "step": 1863 }, { "epoch": 0.5065217391304347, "grad_norm": 1.245968222618103, "learning_rate": 3e-05, "loss": 3.5063, "step": 1864 }, { "epoch": 0.5067934782608695, "grad_norm": 1.3542323112487793, "learning_rate": 3e-05, "loss": 3.981, "step": 1865 }, { "epoch": 0.5070652173913044, "grad_norm": 2.4021403789520264, "learning_rate": 3e-05, "loss": 3.9367, "step": 1866 }, { "epoch": 0.5073369565217392, "grad_norm": 1.4897315502166748, "learning_rate": 3e-05, "loss": 2.7825, "step": 1867 }, { "epoch": 0.5076086956521739, "grad_norm": 1.7353469133377075, "learning_rate": 3e-05, "loss": 3.5709, "step": 1868 }, { "epoch": 0.5078804347826087, "grad_norm": 1.216436505317688, "learning_rate": 3e-05, "loss": 3.1314, "step": 1869 }, { "epoch": 0.5081521739130435, "grad_norm": 1.2940754890441895, "learning_rate": 3e-05, "loss": 3.3858, "step": 1870 }, { "epoch": 0.5084239130434782, "grad_norm": 0.9794735312461853, "learning_rate": 3e-05, "loss": 2.823, "step": 1871 }, { "epoch": 0.508695652173913, "grad_norm": 1.4511076211929321, "learning_rate": 3e-05, "loss": 4.0647, "step": 1872 }, { "epoch": 0.5089673913043479, "grad_norm": 1.559597373008728, "learning_rate": 3e-05, "loss": 3.093, "step": 1873 }, { "epoch": 0.5092391304347826, "grad_norm": 1.7877241373062134, "learning_rate": 3e-05, "loss": 3.742, "step": 1874 }, { "epoch": 0.5095108695652174, "grad_norm": 1.585822582244873, "learning_rate": 3e-05, "loss": 3.631, "step": 1875 }, { "epoch": 0.5097826086956522, "grad_norm": 1.284661054611206, "learning_rate": 3e-05, "loss": 3.228, "step": 1876 }, { "epoch": 0.5100543478260869, "grad_norm": 1.295775294303894, "learning_rate": 3e-05, "loss": 3.7922, "step": 1877 }, { "epoch": 0.5103260869565217, "grad_norm": 1.2768347263336182, "learning_rate": 3e-05, "loss": 3.0369, "step": 1878 }, { "epoch": 0.5105978260869565, "grad_norm": 1.4773716926574707, "learning_rate": 3e-05, "loss": 3.3407, "step": 1879 }, { "epoch": 0.5108695652173914, "grad_norm": 1.2089797258377075, "learning_rate": 3e-05, "loss": 3.0556, "step": 1880 }, { "epoch": 0.5111413043478261, "grad_norm": 1.4668608903884888, "learning_rate": 3e-05, "loss": 3.4665, "step": 1881 }, { "epoch": 0.5114130434782609, "grad_norm": 1.998399257659912, "learning_rate": 3e-05, "loss": 3.8042, "step": 1882 }, { "epoch": 0.5116847826086957, "grad_norm": 1.448439598083496, "learning_rate": 3e-05, "loss": 3.7303, "step": 1883 }, { "epoch": 0.5119565217391304, "grad_norm": 0.9362940192222595, "learning_rate": 3e-05, "loss": 2.665, "step": 1884 }, { "epoch": 0.5122282608695652, "grad_norm": 1.6077910661697388, "learning_rate": 3e-05, "loss": 3.4826, "step": 1885 }, { "epoch": 0.5125, "grad_norm": 1.5588127374649048, "learning_rate": 3e-05, "loss": 2.837, "step": 1886 }, { "epoch": 0.5127717391304348, "grad_norm": 1.473461389541626, "learning_rate": 3e-05, "loss": 3.2511, "step": 1887 }, { "epoch": 0.5130434782608696, "grad_norm": 1.6276698112487793, "learning_rate": 3e-05, "loss": 3.5821, "step": 1888 }, { "epoch": 0.5133152173913044, "grad_norm": 1.6201423406600952, "learning_rate": 3e-05, "loss": 3.195, "step": 1889 }, { "epoch": 0.5135869565217391, "grad_norm": 1.6411088705062866, "learning_rate": 3e-05, "loss": 3.1111, "step": 1890 }, { "epoch": 0.5138586956521739, "grad_norm": 1.2973185777664185, "learning_rate": 3e-05, "loss": 2.6324, "step": 1891 }, { "epoch": 0.5141304347826087, "grad_norm": 2.3508846759796143, "learning_rate": 3e-05, "loss": 3.7621, "step": 1892 }, { "epoch": 0.5144021739130434, "grad_norm": 1.718246340751648, "learning_rate": 3e-05, "loss": 3.759, "step": 1893 }, { "epoch": 0.5146739130434783, "grad_norm": 1.5278515815734863, "learning_rate": 3e-05, "loss": 3.2879, "step": 1894 }, { "epoch": 0.5149456521739131, "grad_norm": 1.308929681777954, "learning_rate": 3e-05, "loss": 3.1075, "step": 1895 }, { "epoch": 0.5152173913043478, "grad_norm": 1.40326988697052, "learning_rate": 3e-05, "loss": 3.5211, "step": 1896 }, { "epoch": 0.5154891304347826, "grad_norm": 1.5129557847976685, "learning_rate": 3e-05, "loss": 3.2549, "step": 1897 }, { "epoch": 0.5157608695652174, "grad_norm": 1.1121063232421875, "learning_rate": 3e-05, "loss": 2.9461, "step": 1898 }, { "epoch": 0.5160326086956522, "grad_norm": 1.4302382469177246, "learning_rate": 3e-05, "loss": 3.627, "step": 1899 }, { "epoch": 0.5163043478260869, "grad_norm": 1.6184173822402954, "learning_rate": 3e-05, "loss": 2.9157, "step": 1900 }, { "epoch": 0.5165760869565217, "grad_norm": 1.1715192794799805, "learning_rate": 3e-05, "loss": 2.9425, "step": 1901 }, { "epoch": 0.5168478260869566, "grad_norm": 2.065067768096924, "learning_rate": 3e-05, "loss": 3.3198, "step": 1902 }, { "epoch": 0.5171195652173913, "grad_norm": 1.6942062377929688, "learning_rate": 3e-05, "loss": 3.4998, "step": 1903 }, { "epoch": 0.5173913043478261, "grad_norm": 1.4067577123641968, "learning_rate": 3e-05, "loss": 3.501, "step": 1904 }, { "epoch": 0.5176630434782609, "grad_norm": 1.5794352293014526, "learning_rate": 3e-05, "loss": 3.1307, "step": 1905 }, { "epoch": 0.5179347826086956, "grad_norm": 1.5043370723724365, "learning_rate": 3e-05, "loss": 3.8714, "step": 1906 }, { "epoch": 0.5182065217391304, "grad_norm": 1.2884905338287354, "learning_rate": 3e-05, "loss": 3.4459, "step": 1907 }, { "epoch": 0.5184782608695652, "grad_norm": 1.3515671491622925, "learning_rate": 3e-05, "loss": 3.5685, "step": 1908 }, { "epoch": 0.51875, "grad_norm": 1.2115263938903809, "learning_rate": 3e-05, "loss": 3.2234, "step": 1909 }, { "epoch": 0.5190217391304348, "grad_norm": 1.3323924541473389, "learning_rate": 3e-05, "loss": 3.42, "step": 1910 }, { "epoch": 0.5192934782608696, "grad_norm": 1.2837159633636475, "learning_rate": 3e-05, "loss": 3.0186, "step": 1911 }, { "epoch": 0.5195652173913043, "grad_norm": 1.1012318134307861, "learning_rate": 3e-05, "loss": 2.9993, "step": 1912 }, { "epoch": 0.5198369565217391, "grad_norm": 1.253250002861023, "learning_rate": 3e-05, "loss": 3.371, "step": 1913 }, { "epoch": 0.5201086956521739, "grad_norm": 0.9867514967918396, "learning_rate": 3e-05, "loss": 2.7761, "step": 1914 }, { "epoch": 0.5203804347826086, "grad_norm": 1.0671703815460205, "learning_rate": 3e-05, "loss": 3.5288, "step": 1915 }, { "epoch": 0.5206521739130435, "grad_norm": 0.9514213800430298, "learning_rate": 3e-05, "loss": 2.7831, "step": 1916 }, { "epoch": 0.5209239130434783, "grad_norm": 1.3238492012023926, "learning_rate": 3e-05, "loss": 3.1874, "step": 1917 }, { "epoch": 0.5211956521739131, "grad_norm": 1.5408774614334106, "learning_rate": 3e-05, "loss": 3.1997, "step": 1918 }, { "epoch": 0.5214673913043478, "grad_norm": 1.1520451307296753, "learning_rate": 3e-05, "loss": 3.1287, "step": 1919 }, { "epoch": 0.5217391304347826, "grad_norm": 1.5356744527816772, "learning_rate": 3e-05, "loss": 3.5281, "step": 1920 }, { "epoch": 0.5220108695652174, "grad_norm": 1.0727752447128296, "learning_rate": 3e-05, "loss": 3.0281, "step": 1921 }, { "epoch": 0.5222826086956521, "grad_norm": 1.4737282991409302, "learning_rate": 3e-05, "loss": 4.0763, "step": 1922 }, { "epoch": 0.522554347826087, "grad_norm": 1.2212142944335938, "learning_rate": 3e-05, "loss": 3.6583, "step": 1923 }, { "epoch": 0.5228260869565218, "grad_norm": 1.42307710647583, "learning_rate": 3e-05, "loss": 3.9908, "step": 1924 }, { "epoch": 0.5230978260869565, "grad_norm": 1.2455291748046875, "learning_rate": 3e-05, "loss": 2.7686, "step": 1925 }, { "epoch": 0.5233695652173913, "grad_norm": 1.1880292892456055, "learning_rate": 3e-05, "loss": 2.8165, "step": 1926 }, { "epoch": 0.5236413043478261, "grad_norm": 1.329270362854004, "learning_rate": 3e-05, "loss": 3.421, "step": 1927 }, { "epoch": 0.5239130434782608, "grad_norm": 1.8960901498794556, "learning_rate": 3e-05, "loss": 4.0437, "step": 1928 }, { "epoch": 0.5241847826086956, "grad_norm": 1.223769187927246, "learning_rate": 3e-05, "loss": 3.3202, "step": 1929 }, { "epoch": 0.5244565217391305, "grad_norm": 1.5670809745788574, "learning_rate": 3e-05, "loss": 3.6805, "step": 1930 }, { "epoch": 0.5247282608695653, "grad_norm": 1.9276920557022095, "learning_rate": 3e-05, "loss": 3.4752, "step": 1931 }, { "epoch": 0.525, "grad_norm": 1.7917790412902832, "learning_rate": 3e-05, "loss": 3.4745, "step": 1932 }, { "epoch": 0.5252717391304348, "grad_norm": 1.1239231824874878, "learning_rate": 3e-05, "loss": 2.9622, "step": 1933 }, { "epoch": 0.5255434782608696, "grad_norm": 1.2177050113677979, "learning_rate": 3e-05, "loss": 2.691, "step": 1934 }, { "epoch": 0.5258152173913043, "grad_norm": 1.1526474952697754, "learning_rate": 3e-05, "loss": 2.7313, "step": 1935 }, { "epoch": 0.5260869565217391, "grad_norm": 1.3403104543685913, "learning_rate": 3e-05, "loss": 3.215, "step": 1936 }, { "epoch": 0.5263586956521739, "grad_norm": 1.1823958158493042, "learning_rate": 3e-05, "loss": 3.0382, "step": 1937 }, { "epoch": 0.5266304347826087, "grad_norm": 1.2578941583633423, "learning_rate": 3e-05, "loss": 2.9962, "step": 1938 }, { "epoch": 0.5269021739130435, "grad_norm": 1.2342561483383179, "learning_rate": 3e-05, "loss": 3.6362, "step": 1939 }, { "epoch": 0.5271739130434783, "grad_norm": 1.0608114004135132, "learning_rate": 3e-05, "loss": 3.2318, "step": 1940 }, { "epoch": 0.527445652173913, "grad_norm": 1.2502753734588623, "learning_rate": 3e-05, "loss": 3.1968, "step": 1941 }, { "epoch": 0.5277173913043478, "grad_norm": 0.9138715267181396, "learning_rate": 3e-05, "loss": 3.0353, "step": 1942 }, { "epoch": 0.5279891304347826, "grad_norm": 1.257987380027771, "learning_rate": 3e-05, "loss": 2.5984, "step": 1943 }, { "epoch": 0.5282608695652173, "grad_norm": 1.5247408151626587, "learning_rate": 3e-05, "loss": 3.4343, "step": 1944 }, { "epoch": 0.5285326086956522, "grad_norm": 1.1517291069030762, "learning_rate": 3e-05, "loss": 2.7655, "step": 1945 }, { "epoch": 0.528804347826087, "grad_norm": 1.5516606569290161, "learning_rate": 3e-05, "loss": 3.9976, "step": 1946 }, { "epoch": 0.5290760869565218, "grad_norm": 1.6808713674545288, "learning_rate": 3e-05, "loss": 3.6858, "step": 1947 }, { "epoch": 0.5293478260869565, "grad_norm": 1.0476493835449219, "learning_rate": 3e-05, "loss": 2.8933, "step": 1948 }, { "epoch": 0.5296195652173913, "grad_norm": 1.7297437191009521, "learning_rate": 3e-05, "loss": 3.1754, "step": 1949 }, { "epoch": 0.529891304347826, "grad_norm": 1.4908910989761353, "learning_rate": 3e-05, "loss": 3.9229, "step": 1950 }, { "epoch": 0.5301630434782608, "grad_norm": 1.3128184080123901, "learning_rate": 3e-05, "loss": 3.5215, "step": 1951 }, { "epoch": 0.5304347826086957, "grad_norm": 1.7025582790374756, "learning_rate": 3e-05, "loss": 3.5113, "step": 1952 }, { "epoch": 0.5307065217391305, "grad_norm": 1.4770921468734741, "learning_rate": 3e-05, "loss": 3.0171, "step": 1953 }, { "epoch": 0.5309782608695652, "grad_norm": 1.6477012634277344, "learning_rate": 3e-05, "loss": 3.5914, "step": 1954 }, { "epoch": 0.53125, "grad_norm": 1.5680129528045654, "learning_rate": 3e-05, "loss": 3.6162, "step": 1955 }, { "epoch": 0.5315217391304348, "grad_norm": 2.2198116779327393, "learning_rate": 3e-05, "loss": 3.2198, "step": 1956 }, { "epoch": 0.5317934782608695, "grad_norm": 1.9690414667129517, "learning_rate": 3e-05, "loss": 3.0914, "step": 1957 }, { "epoch": 0.5320652173913043, "grad_norm": 1.4856075048446655, "learning_rate": 3e-05, "loss": 3.206, "step": 1958 }, { "epoch": 0.5323369565217392, "grad_norm": 1.1297136545181274, "learning_rate": 3e-05, "loss": 2.8479, "step": 1959 }, { "epoch": 0.532608695652174, "grad_norm": 1.859328269958496, "learning_rate": 3e-05, "loss": 3.3152, "step": 1960 }, { "epoch": 0.5328804347826087, "grad_norm": 1.939894199371338, "learning_rate": 3e-05, "loss": 3.2557, "step": 1961 }, { "epoch": 0.5331521739130435, "grad_norm": 2.148735523223877, "learning_rate": 3e-05, "loss": 3.6686, "step": 1962 }, { "epoch": 0.5334239130434782, "grad_norm": 1.4653475284576416, "learning_rate": 3e-05, "loss": 3.8618, "step": 1963 }, { "epoch": 0.533695652173913, "grad_norm": 2.1480512619018555, "learning_rate": 3e-05, "loss": 3.4809, "step": 1964 }, { "epoch": 0.5339673913043478, "grad_norm": 2.2585768699645996, "learning_rate": 3e-05, "loss": 3.1685, "step": 1965 }, { "epoch": 0.5342391304347827, "grad_norm": 2.143669605255127, "learning_rate": 3e-05, "loss": 3.8206, "step": 1966 }, { "epoch": 0.5345108695652174, "grad_norm": 1.5735530853271484, "learning_rate": 3e-05, "loss": 3.5483, "step": 1967 }, { "epoch": 0.5347826086956522, "grad_norm": 1.5427213907241821, "learning_rate": 3e-05, "loss": 3.469, "step": 1968 }, { "epoch": 0.535054347826087, "grad_norm": 1.7961771488189697, "learning_rate": 3e-05, "loss": 3.2229, "step": 1969 }, { "epoch": 0.5353260869565217, "grad_norm": 1.5997658967971802, "learning_rate": 3e-05, "loss": 3.6575, "step": 1970 }, { "epoch": 0.5355978260869565, "grad_norm": 1.2979601621627808, "learning_rate": 3e-05, "loss": 3.7012, "step": 1971 }, { "epoch": 0.5358695652173913, "grad_norm": 1.5162733793258667, "learning_rate": 3e-05, "loss": 3.1471, "step": 1972 }, { "epoch": 0.5361413043478261, "grad_norm": 1.5043525695800781, "learning_rate": 3e-05, "loss": 2.8777, "step": 1973 }, { "epoch": 0.5364130434782609, "grad_norm": 1.1250759363174438, "learning_rate": 3e-05, "loss": 2.9937, "step": 1974 }, { "epoch": 0.5366847826086957, "grad_norm": 1.2006231546401978, "learning_rate": 3e-05, "loss": 3.5737, "step": 1975 }, { "epoch": 0.5369565217391304, "grad_norm": 0.9950990080833435, "learning_rate": 3e-05, "loss": 2.9763, "step": 1976 }, { "epoch": 0.5372282608695652, "grad_norm": 1.2111644744873047, "learning_rate": 3e-05, "loss": 3.5871, "step": 1977 }, { "epoch": 0.5375, "grad_norm": 1.4334638118743896, "learning_rate": 3e-05, "loss": 3.645, "step": 1978 }, { "epoch": 0.5377717391304347, "grad_norm": 1.376451849937439, "learning_rate": 3e-05, "loss": 3.5722, "step": 1979 }, { "epoch": 0.5380434782608695, "grad_norm": 1.1852439641952515, "learning_rate": 3e-05, "loss": 3.2733, "step": 1980 }, { "epoch": 0.5383152173913044, "grad_norm": 1.208321452140808, "learning_rate": 3e-05, "loss": 3.4223, "step": 1981 }, { "epoch": 0.5385869565217392, "grad_norm": 1.2566808462142944, "learning_rate": 3e-05, "loss": 3.227, "step": 1982 }, { "epoch": 0.5388586956521739, "grad_norm": 1.1662538051605225, "learning_rate": 3e-05, "loss": 3.9731, "step": 1983 }, { "epoch": 0.5391304347826087, "grad_norm": 0.9991174340248108, "learning_rate": 3e-05, "loss": 3.2856, "step": 1984 }, { "epoch": 0.5394021739130435, "grad_norm": 1.6792123317718506, "learning_rate": 3e-05, "loss": 4.0315, "step": 1985 }, { "epoch": 0.5396739130434782, "grad_norm": 1.3554564714431763, "learning_rate": 3e-05, "loss": 2.7884, "step": 1986 }, { "epoch": 0.539945652173913, "grad_norm": 2.0495409965515137, "learning_rate": 3e-05, "loss": 4.0834, "step": 1987 }, { "epoch": 0.5402173913043479, "grad_norm": 1.3076777458190918, "learning_rate": 3e-05, "loss": 2.9288, "step": 1988 }, { "epoch": 0.5404891304347826, "grad_norm": 1.2214155197143555, "learning_rate": 3e-05, "loss": 2.9785, "step": 1989 }, { "epoch": 0.5407608695652174, "grad_norm": 1.6654658317565918, "learning_rate": 3e-05, "loss": 3.6795, "step": 1990 }, { "epoch": 0.5410326086956522, "grad_norm": 1.224678874015808, "learning_rate": 3e-05, "loss": 2.9736, "step": 1991 }, { "epoch": 0.5413043478260869, "grad_norm": 1.128450870513916, "learning_rate": 3e-05, "loss": 3.0371, "step": 1992 }, { "epoch": 0.5415760869565217, "grad_norm": 1.1452875137329102, "learning_rate": 3e-05, "loss": 2.989, "step": 1993 }, { "epoch": 0.5418478260869565, "grad_norm": 1.2154463529586792, "learning_rate": 3e-05, "loss": 2.9404, "step": 1994 }, { "epoch": 0.5421195652173914, "grad_norm": 1.1099878549575806, "learning_rate": 3e-05, "loss": 3.1758, "step": 1995 }, { "epoch": 0.5423913043478261, "grad_norm": 0.9395751953125, "learning_rate": 3e-05, "loss": 2.9722, "step": 1996 }, { "epoch": 0.5426630434782609, "grad_norm": 1.1049753427505493, "learning_rate": 3e-05, "loss": 3.0526, "step": 1997 }, { "epoch": 0.5429347826086957, "grad_norm": 1.45954430103302, "learning_rate": 3e-05, "loss": 3.7409, "step": 1998 }, { "epoch": 0.5432065217391304, "grad_norm": 1.2864134311676025, "learning_rate": 3e-05, "loss": 3.4088, "step": 1999 }, { "epoch": 0.5434782608695652, "grad_norm": 1.2165127992630005, "learning_rate": 3e-05, "loss": 3.2604, "step": 2000 } ], "logging_steps": 1, "max_steps": 3680, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.508513578745856e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }