{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 3680, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0002717391304347826, "grad_norm": Infinity, "learning_rate": 0.0, "loss": 3.8119, "step": 1 }, { "epoch": 0.0005434782608695652, "grad_norm": Infinity, "learning_rate": 0.0, "loss": 5.6312, "step": 2 }, { "epoch": 0.0008152173913043478, "grad_norm": 3.606865644454956, "learning_rate": 1.0000000000000001e-07, "loss": 3.6085, "step": 3 }, { "epoch": 0.0010869565217391304, "grad_norm": 5.723954677581787, "learning_rate": 2.0000000000000002e-07, "loss": 4.5497, "step": 4 }, { "epoch": 0.001358695652173913, "grad_norm": 8.396666526794434, "learning_rate": 3.0000000000000004e-07, "loss": 5.6655, "step": 5 }, { "epoch": 0.0016304347826086956, "grad_norm": 6.602017402648926, "learning_rate": 4.0000000000000003e-07, "loss": 4.9113, "step": 6 }, { "epoch": 0.0019021739130434783, "grad_norm": 2.8019049167633057, "learning_rate": 5e-07, "loss": 3.0843, "step": 7 }, { "epoch": 0.002173913043478261, "grad_norm": 5.959436893463135, "learning_rate": 6.000000000000001e-07, "loss": 4.4235, "step": 8 }, { "epoch": 0.0024456521739130437, "grad_norm": 5.623807430267334, "learning_rate": 7.000000000000001e-07, "loss": 4.5667, "step": 9 }, { "epoch": 0.002717391304347826, "grad_norm": 5.9944844245910645, "learning_rate": 8.000000000000001e-07, "loss": 4.6548, "step": 10 }, { "epoch": 0.002989130434782609, "grad_norm": 7.308701038360596, "learning_rate": 9e-07, "loss": 5.1422, "step": 11 }, { "epoch": 0.003260869565217391, "grad_norm": 5.805826187133789, "learning_rate": 1e-06, "loss": 4.3458, "step": 12 }, { "epoch": 0.003532608695652174, "grad_norm": 5.096138000488281, "learning_rate": 1.1e-06, "loss": 4.0761, "step": 13 }, { "epoch": 0.0038043478260869567, "grad_norm": 4.222947120666504, "learning_rate": 1.2000000000000002e-06, "loss": 3.7963, "step": 14 }, { "epoch": 0.004076086956521739, "grad_norm": 8.304177284240723, "learning_rate": 1.3e-06, "loss": 5.4697, "step": 15 }, { "epoch": 0.004347826086956522, "grad_norm": 6.131305694580078, "learning_rate": 1.4000000000000001e-06, "loss": 4.5078, "step": 16 }, { "epoch": 0.004619565217391305, "grad_norm": 5.1205315589904785, "learning_rate": 1.5e-06, "loss": 4.1935, "step": 17 }, { "epoch": 0.004891304347826087, "grad_norm": 3.348874568939209, "learning_rate": 1.6000000000000001e-06, "loss": 3.165, "step": 18 }, { "epoch": 0.005163043478260869, "grad_norm": 11.620071411132812, "learning_rate": 1.7e-06, "loss": 6.6692, "step": 19 }, { "epoch": 0.005434782608695652, "grad_norm": 5.148247241973877, "learning_rate": 1.8e-06, "loss": 3.7967, "step": 20 }, { "epoch": 0.005706521739130435, "grad_norm": 6.738089561462402, "learning_rate": 1.9000000000000002e-06, "loss": 4.6053, "step": 21 }, { "epoch": 0.005978260869565218, "grad_norm": 7.383264541625977, "learning_rate": 2e-06, "loss": 4.7862, "step": 22 }, { "epoch": 0.00625, "grad_norm": 9.380542755126953, "learning_rate": 2.1000000000000002e-06, "loss": 5.5961, "step": 23 }, { "epoch": 0.006521739130434782, "grad_norm": 5.481351852416992, "learning_rate": 2.2e-06, "loss": 3.9489, "step": 24 }, { "epoch": 0.006793478260869565, "grad_norm": 3.769026756286621, "learning_rate": 2.3e-06, "loss": 3.3576, "step": 25 }, { "epoch": 0.007065217391304348, "grad_norm": 6.088125705718994, "learning_rate": 2.4000000000000003e-06, "loss": 4.3102, "step": 26 }, { "epoch": 0.007336956521739131, "grad_norm": 3.7228612899780273, "learning_rate": 2.4999999999999998e-06, "loss": 3.2898, "step": 27 }, { "epoch": 0.007608695652173913, "grad_norm": 5.163967609405518, "learning_rate": 2.6e-06, "loss": 3.9394, "step": 28 }, { "epoch": 0.007880434782608695, "grad_norm": 3.577744960784912, "learning_rate": 2.7e-06, "loss": 3.4044, "step": 29 }, { "epoch": 0.008152173913043478, "grad_norm": 5.421440124511719, "learning_rate": 2.8000000000000003e-06, "loss": 4.2156, "step": 30 }, { "epoch": 0.00842391304347826, "grad_norm": 4.284877777099609, "learning_rate": 2.9e-06, "loss": 3.519, "step": 31 }, { "epoch": 0.008695652173913044, "grad_norm": 6.597541809082031, "learning_rate": 3e-06, "loss": 4.5028, "step": 32 }, { "epoch": 0.008967391304347826, "grad_norm": 7.851593494415283, "learning_rate": 3.1e-06, "loss": 5.1247, "step": 33 }, { "epoch": 0.00923913043478261, "grad_norm": 2.644531726837158, "learning_rate": 3.2000000000000003e-06, "loss": 2.9878, "step": 34 }, { "epoch": 0.009510869565217392, "grad_norm": 3.316633462905884, "learning_rate": 3.3e-06, "loss": 3.4112, "step": 35 }, { "epoch": 0.009782608695652175, "grad_norm": 3.0373375415802, "learning_rate": 3.4e-06, "loss": 3.301, "step": 36 }, { "epoch": 0.010054347826086956, "grad_norm": 5.997847080230713, "learning_rate": 3.5e-06, "loss": 4.9763, "step": 37 }, { "epoch": 0.010326086956521738, "grad_norm": 5.2729268074035645, "learning_rate": 3.6e-06, "loss": 4.7279, "step": 38 }, { "epoch": 0.010597826086956521, "grad_norm": 2.7055575847625732, "learning_rate": 3.7e-06, "loss": 3.4656, "step": 39 }, { "epoch": 0.010869565217391304, "grad_norm": 3.5352964401245117, "learning_rate": 3.8000000000000005e-06, "loss": 3.9642, "step": 40 }, { "epoch": 0.011141304347826087, "grad_norm": 2.8491668701171875, "learning_rate": 3.9e-06, "loss": 3.8092, "step": 41 }, { "epoch": 0.01141304347826087, "grad_norm": 1.8910856246948242, "learning_rate": 4e-06, "loss": 3.0861, "step": 42 }, { "epoch": 0.011684782608695652, "grad_norm": 2.873784303665161, "learning_rate": 4.1e-06, "loss": 3.788, "step": 43 }, { "epoch": 0.011956521739130435, "grad_norm": 3.121163845062256, "learning_rate": 4.2000000000000004e-06, "loss": 4.0837, "step": 44 }, { "epoch": 0.012228260869565218, "grad_norm": 2.977182388305664, "learning_rate": 4.3e-06, "loss": 4.0382, "step": 45 }, { "epoch": 0.0125, "grad_norm": 3.935685396194458, "learning_rate": 4.4e-06, "loss": 5.0603, "step": 46 }, { "epoch": 0.012771739130434783, "grad_norm": 1.5644341707229614, "learning_rate": 4.5e-06, "loss": 3.0085, "step": 47 }, { "epoch": 0.013043478260869565, "grad_norm": 2.2639412879943848, "learning_rate": 4.6e-06, "loss": 3.5606, "step": 48 }, { "epoch": 0.013315217391304347, "grad_norm": 3.8248438835144043, "learning_rate": 4.700000000000001e-06, "loss": 5.1055, "step": 49 }, { "epoch": 0.01358695652173913, "grad_norm": 2.0189621448516846, "learning_rate": 4.800000000000001e-06, "loss": 3.6357, "step": 50 }, { "epoch": 0.013858695652173913, "grad_norm": 1.7053027153015137, "learning_rate": 4.9e-06, "loss": 3.3706, "step": 51 }, { "epoch": 0.014130434782608696, "grad_norm": 2.996800184249878, "learning_rate": 4.9999999999999996e-06, "loss": 4.2643, "step": 52 }, { "epoch": 0.014402173913043478, "grad_norm": 2.2045717239379883, "learning_rate": 5.1e-06, "loss": 3.4676, "step": 53 }, { "epoch": 0.014673913043478261, "grad_norm": 1.9880719184875488, "learning_rate": 5.2e-06, "loss": 3.7901, "step": 54 }, { "epoch": 0.014945652173913044, "grad_norm": 1.4993855953216553, "learning_rate": 5.3e-06, "loss": 3.3175, "step": 55 }, { "epoch": 0.015217391304347827, "grad_norm": 1.3958892822265625, "learning_rate": 5.4e-06, "loss": 3.0166, "step": 56 }, { "epoch": 0.01548913043478261, "grad_norm": 1.5511572360992432, "learning_rate": 5.5e-06, "loss": 3.4503, "step": 57 }, { "epoch": 0.01576086956521739, "grad_norm": 2.274249792098999, "learning_rate": 5.600000000000001e-06, "loss": 4.1636, "step": 58 }, { "epoch": 0.016032608695652175, "grad_norm": 1.7447724342346191, "learning_rate": 5.7000000000000005e-06, "loss": 3.7056, "step": 59 }, { "epoch": 0.016304347826086956, "grad_norm": 2.4315755367279053, "learning_rate": 5.8e-06, "loss": 3.9477, "step": 60 }, { "epoch": 0.01657608695652174, "grad_norm": 1.5694116353988647, "learning_rate": 5.899999999999999e-06, "loss": 3.321, "step": 61 }, { "epoch": 0.01684782608695652, "grad_norm": 1.856467366218567, "learning_rate": 6e-06, "loss": 3.8384, "step": 62 }, { "epoch": 0.017119565217391303, "grad_norm": 1.3913215398788452, "learning_rate": 6.1e-06, "loss": 3.1717, "step": 63 }, { "epoch": 0.017391304347826087, "grad_norm": 1.1280888319015503, "learning_rate": 6.2e-06, "loss": 3.0487, "step": 64 }, { "epoch": 0.017663043478260868, "grad_norm": 1.3553597927093506, "learning_rate": 6.3e-06, "loss": 3.4274, "step": 65 }, { "epoch": 0.017934782608695653, "grad_norm": 1.5680698156356812, "learning_rate": 6.4000000000000006e-06, "loss": 3.2402, "step": 66 }, { "epoch": 0.018206521739130434, "grad_norm": 1.2775722742080688, "learning_rate": 6.5000000000000004e-06, "loss": 3.1252, "step": 67 }, { "epoch": 0.01847826086956522, "grad_norm": 1.547057867050171, "learning_rate": 6.6e-06, "loss": 3.6654, "step": 68 }, { "epoch": 0.01875, "grad_norm": 1.6398184299468994, "learning_rate": 6.7e-06, "loss": 3.4637, "step": 69 }, { "epoch": 0.019021739130434784, "grad_norm": 0.8923302292823792, "learning_rate": 6.8e-06, "loss": 2.6619, "step": 70 }, { "epoch": 0.019293478260869565, "grad_norm": 1.1862719058990479, "learning_rate": 6.900000000000001e-06, "loss": 3.1319, "step": 71 }, { "epoch": 0.01956521739130435, "grad_norm": 1.0350522994995117, "learning_rate": 7e-06, "loss": 2.8874, "step": 72 }, { "epoch": 0.01983695652173913, "grad_norm": 1.7756403684616089, "learning_rate": 7.1e-06, "loss": 3.7356, "step": 73 }, { "epoch": 0.02010869565217391, "grad_norm": 1.2788705825805664, "learning_rate": 7.2e-06, "loss": 3.2409, "step": 74 }, { "epoch": 0.020380434782608696, "grad_norm": 1.4206314086914062, "learning_rate": 7.3e-06, "loss": 3.5884, "step": 75 }, { "epoch": 0.020652173913043477, "grad_norm": 2.0677270889282227, "learning_rate": 7.4e-06, "loss": 4.3218, "step": 76 }, { "epoch": 0.02092391304347826, "grad_norm": 0.700586199760437, "learning_rate": 7.5e-06, "loss": 2.4596, "step": 77 }, { "epoch": 0.021195652173913043, "grad_norm": 2.1819796562194824, "learning_rate": 7.600000000000001e-06, "loss": 3.9885, "step": 78 }, { "epoch": 0.021467391304347827, "grad_norm": 1.0155264139175415, "learning_rate": 7.699999999999999e-06, "loss": 2.7548, "step": 79 }, { "epoch": 0.021739130434782608, "grad_norm": 1.2022877931594849, "learning_rate": 7.8e-06, "loss": 3.1796, "step": 80 }, { "epoch": 0.022010869565217393, "grad_norm": 1.1962783336639404, "learning_rate": 7.899999999999999e-06, "loss": 2.907, "step": 81 }, { "epoch": 0.022282608695652174, "grad_norm": 1.0273329019546509, "learning_rate": 8e-06, "loss": 2.7207, "step": 82 }, { "epoch": 0.022554347826086958, "grad_norm": 2.3930468559265137, "learning_rate": 8.1e-06, "loss": 3.9644, "step": 83 }, { "epoch": 0.02282608695652174, "grad_norm": 2.244040012359619, "learning_rate": 8.2e-06, "loss": 3.5003, "step": 84 }, { "epoch": 0.02309782608695652, "grad_norm": 1.685024380683899, "learning_rate": 8.3e-06, "loss": 3.0757, "step": 85 }, { "epoch": 0.023369565217391305, "grad_norm": 3.1130995750427246, "learning_rate": 8.400000000000001e-06, "loss": 3.9588, "step": 86 }, { "epoch": 0.023641304347826086, "grad_norm": 2.6688392162323, "learning_rate": 8.5e-06, "loss": 3.6427, "step": 87 }, { "epoch": 0.02391304347826087, "grad_norm": 1.3578976392745972, "learning_rate": 8.6e-06, "loss": 2.9824, "step": 88 }, { "epoch": 0.02418478260869565, "grad_norm": 1.2437134981155396, "learning_rate": 8.7e-06, "loss": 3.1183, "step": 89 }, { "epoch": 0.024456521739130436, "grad_norm": 2.143594741821289, "learning_rate": 8.8e-06, "loss": 3.6833, "step": 90 }, { "epoch": 0.024728260869565217, "grad_norm": 1.2158911228179932, "learning_rate": 8.900000000000001e-06, "loss": 2.7814, "step": 91 }, { "epoch": 0.025, "grad_norm": 2.4209673404693604, "learning_rate": 9e-06, "loss": 4.3185, "step": 92 }, { "epoch": 0.025271739130434782, "grad_norm": 3.090715169906616, "learning_rate": 9.100000000000001e-06, "loss": 4.0217, "step": 93 }, { "epoch": 0.025543478260869567, "grad_norm": 3.1287500858306885, "learning_rate": 9.2e-06, "loss": 3.8665, "step": 94 }, { "epoch": 0.025815217391304348, "grad_norm": 2.6370797157287598, "learning_rate": 9.3e-06, "loss": 3.7443, "step": 95 }, { "epoch": 0.02608695652173913, "grad_norm": 4.120901107788086, "learning_rate": 9.400000000000001e-06, "loss": 4.3482, "step": 96 }, { "epoch": 0.026358695652173914, "grad_norm": NaN, "learning_rate": 9.400000000000001e-06, "loss": 3.7633, "step": 97 }, { "epoch": 0.026630434782608695, "grad_norm": 2.1682493686676025, "learning_rate": 9.5e-06, "loss": 3.135, "step": 98 }, { "epoch": 0.02690217391304348, "grad_norm": 9.943031311035156, "learning_rate": 9.600000000000001e-06, "loss": 4.4992, "step": 99 }, { "epoch": 0.02717391304347826, "grad_norm": 9.851699829101562, "learning_rate": 9.699999999999999e-06, "loss": 3.3346, "step": 100 }, { "epoch": 0.027445652173913045, "grad_norm": 9.068572044372559, "learning_rate": 9.8e-06, "loss": 3.4352, "step": 101 }, { "epoch": 0.027717391304347826, "grad_norm": 16.46036148071289, "learning_rate": 9.9e-06, "loss": 4.2803, "step": 102 }, { "epoch": 0.02798913043478261, "grad_norm": 4.916048049926758, "learning_rate": 9.999999999999999e-06, "loss": 5.2283, "step": 103 }, { "epoch": 0.02826086956521739, "grad_norm": 3.9112675189971924, "learning_rate": 1.01e-05, "loss": 3.9085, "step": 104 }, { "epoch": 0.028532608695652172, "grad_norm": 2.9353702068328857, "learning_rate": 1.02e-05, "loss": 3.436, "step": 105 }, { "epoch": 0.028804347826086957, "grad_norm": 2.1756107807159424, "learning_rate": 1.03e-05, "loss": 4.2118, "step": 106 }, { "epoch": 0.029076086956521738, "grad_norm": 2.4335989952087402, "learning_rate": 1.04e-05, "loss": 4.3144, "step": 107 }, { "epoch": 0.029347826086956522, "grad_norm": 2.6821694374084473, "learning_rate": 1.05e-05, "loss": 4.461, "step": 108 }, { "epoch": 0.029619565217391303, "grad_norm": 1.9325740337371826, "learning_rate": 1.06e-05, "loss": 4.1079, "step": 109 }, { "epoch": 0.029891304347826088, "grad_norm": 3.598806381225586, "learning_rate": 1.0700000000000001e-05, "loss": 3.6908, "step": 110 }, { "epoch": 0.03016304347826087, "grad_norm": 1.7592302560806274, "learning_rate": 1.08e-05, "loss": 3.2482, "step": 111 }, { "epoch": 0.030434782608695653, "grad_norm": 2.4772744178771973, "learning_rate": 1.09e-05, "loss": 4.3941, "step": 112 }, { "epoch": 0.030706521739130434, "grad_norm": 1.8405948877334595, "learning_rate": 1.1e-05, "loss": 3.6201, "step": 113 }, { "epoch": 0.03097826086956522, "grad_norm": 4.303748607635498, "learning_rate": 1.11e-05, "loss": 4.6995, "step": 114 }, { "epoch": 0.03125, "grad_norm": 2.199484348297119, "learning_rate": 1.1200000000000001e-05, "loss": 3.7969, "step": 115 }, { "epoch": 0.03152173913043478, "grad_norm": 1.0349135398864746, "learning_rate": 1.13e-05, "loss": 2.8461, "step": 116 }, { "epoch": 0.03179347826086956, "grad_norm": 2.374260425567627, "learning_rate": 1.1400000000000001e-05, "loss": 4.2862, "step": 117 }, { "epoch": 0.03206521739130435, "grad_norm": 1.283584713935852, "learning_rate": 1.1500000000000002e-05, "loss": 3.1288, "step": 118 }, { "epoch": 0.03233695652173913, "grad_norm": 3.549565076828003, "learning_rate": 1.16e-05, "loss": 4.1137, "step": 119 }, { "epoch": 0.03260869565217391, "grad_norm": 1.3312267065048218, "learning_rate": 1.1700000000000001e-05, "loss": 3.6003, "step": 120 }, { "epoch": 0.03288043478260869, "grad_norm": 1.655480146408081, "learning_rate": 1.1799999999999999e-05, "loss": 3.7798, "step": 121 }, { "epoch": 0.03315217391304348, "grad_norm": 1.6580840349197388, "learning_rate": 1.19e-05, "loss": 3.5654, "step": 122 }, { "epoch": 0.03342391304347826, "grad_norm": 2.8110294342041016, "learning_rate": 1.2e-05, "loss": 3.9673, "step": 123 }, { "epoch": 0.03369565217391304, "grad_norm": 1.2965394258499146, "learning_rate": 1.21e-05, "loss": 3.26, "step": 124 }, { "epoch": 0.033967391304347824, "grad_norm": 1.914302945137024, "learning_rate": 1.22e-05, "loss": 3.95, "step": 125 }, { "epoch": 0.034239130434782605, "grad_norm": 1.3706483840942383, "learning_rate": 1.2299999999999999e-05, "loss": 3.5317, "step": 126 }, { "epoch": 0.03451086956521739, "grad_norm": 2.016113042831421, "learning_rate": 1.24e-05, "loss": 3.2582, "step": 127 }, { "epoch": 0.034782608695652174, "grad_norm": 2.5129711627960205, "learning_rate": 1.25e-05, "loss": 3.3608, "step": 128 }, { "epoch": 0.035054347826086955, "grad_norm": Infinity, "learning_rate": 1.25e-05, "loss": 4.3213, "step": 129 }, { "epoch": 0.035326086956521736, "grad_norm": 3.073050022125244, "learning_rate": 1.26e-05, "loss": 4.4668, "step": 130 }, { "epoch": 0.035597826086956524, "grad_norm": 1.8033742904663086, "learning_rate": 1.27e-05, "loss": 3.5871, "step": 131 }, { "epoch": 0.035869565217391305, "grad_norm": 3.148634195327759, "learning_rate": 1.2800000000000001e-05, "loss": 3.7611, "step": 132 }, { "epoch": 0.036141304347826086, "grad_norm": 2.898167610168457, "learning_rate": 1.29e-05, "loss": 3.6928, "step": 133 }, { "epoch": 0.03641304347826087, "grad_norm": 1.9879661798477173, "learning_rate": 1.3000000000000001e-05, "loss": 3.607, "step": 134 }, { "epoch": 0.036684782608695655, "grad_norm": 2.0479657649993896, "learning_rate": 1.31e-05, "loss": 3.2456, "step": 135 }, { "epoch": 0.03695652173913044, "grad_norm": 1.5060527324676514, "learning_rate": 1.32e-05, "loss": 3.4124, "step": 136 }, { "epoch": 0.03722826086956522, "grad_norm": 2.2012810707092285, "learning_rate": 1.3300000000000001e-05, "loss": 4.1008, "step": 137 }, { "epoch": 0.0375, "grad_norm": 2.2929439544677734, "learning_rate": 1.34e-05, "loss": 4.5483, "step": 138 }, { "epoch": 0.03777173913043478, "grad_norm": 1.777136206626892, "learning_rate": 1.3500000000000001e-05, "loss": 3.9796, "step": 139 }, { "epoch": 0.03804347826086957, "grad_norm": 3.1867823600769043, "learning_rate": 1.36e-05, "loss": 4.13, "step": 140 }, { "epoch": 0.03831521739130435, "grad_norm": 2.070611000061035, "learning_rate": 1.3700000000000001e-05, "loss": 3.3474, "step": 141 }, { "epoch": 0.03858695652173913, "grad_norm": 1.9961167573928833, "learning_rate": 1.3800000000000002e-05, "loss": 4.0889, "step": 142 }, { "epoch": 0.03885869565217391, "grad_norm": 1.709995985031128, "learning_rate": 1.39e-05, "loss": 3.9576, "step": 143 }, { "epoch": 0.0391304347826087, "grad_norm": 1.793494462966919, "learning_rate": 1.4e-05, "loss": 4.419, "step": 144 }, { "epoch": 0.03940217391304348, "grad_norm": 2.1228878498077393, "learning_rate": 1.4099999999999999e-05, "loss": 3.7677, "step": 145 }, { "epoch": 0.03967391304347826, "grad_norm": 1.5662310123443604, "learning_rate": 1.42e-05, "loss": 3.5038, "step": 146 }, { "epoch": 0.03994565217391304, "grad_norm": 1.1859204769134521, "learning_rate": 1.43e-05, "loss": 3.0831, "step": 147 }, { "epoch": 0.04021739130434782, "grad_norm": 1.998694896697998, "learning_rate": 1.44e-05, "loss": 4.06, "step": 148 }, { "epoch": 0.04048913043478261, "grad_norm": 2.2709462642669678, "learning_rate": 1.45e-05, "loss": 3.4234, "step": 149 }, { "epoch": 0.04076086956521739, "grad_norm": 3.572169065475464, "learning_rate": 1.46e-05, "loss": 4.1884, "step": 150 }, { "epoch": 0.04103260869565217, "grad_norm": 2.4503207206726074, "learning_rate": 1.47e-05, "loss": 4.0506, "step": 151 }, { "epoch": 0.041304347826086954, "grad_norm": 2.66367244720459, "learning_rate": 1.48e-05, "loss": 4.6858, "step": 152 }, { "epoch": 0.04157608695652174, "grad_norm": 2.1236956119537354, "learning_rate": 1.49e-05, "loss": 3.2571, "step": 153 }, { "epoch": 0.04184782608695652, "grad_norm": 4.9067206382751465, "learning_rate": 1.5e-05, "loss": 5.32, "step": 154 }, { "epoch": 0.042119565217391304, "grad_norm": 2.0733673572540283, "learning_rate": 1.51e-05, "loss": 3.7548, "step": 155 }, { "epoch": 0.042391304347826085, "grad_norm": 1.186569333076477, "learning_rate": 1.5200000000000002e-05, "loss": 3.3953, "step": 156 }, { "epoch": 0.042663043478260866, "grad_norm": 2.5303616523742676, "learning_rate": 1.53e-05, "loss": 4.0742, "step": 157 }, { "epoch": 0.042934782608695654, "grad_norm": 2.3837692737579346, "learning_rate": 1.5399999999999998e-05, "loss": 4.3684, "step": 158 }, { "epoch": 0.043206521739130435, "grad_norm": 2.3704850673675537, "learning_rate": 1.55e-05, "loss": 4.1658, "step": 159 }, { "epoch": 0.043478260869565216, "grad_norm": 3.8034274578094482, "learning_rate": 1.56e-05, "loss": 4.0919, "step": 160 }, { "epoch": 0.04375, "grad_norm": 1.2763932943344116, "learning_rate": 1.57e-05, "loss": 3.8752, "step": 161 }, { "epoch": 0.044021739130434785, "grad_norm": 1.7155462503433228, "learning_rate": 1.5799999999999998e-05, "loss": 4.317, "step": 162 }, { "epoch": 0.044293478260869566, "grad_norm": 2.481580972671509, "learning_rate": 1.59e-05, "loss": 4.1314, "step": 163 }, { "epoch": 0.04456521739130435, "grad_norm": 2.3853437900543213, "learning_rate": 1.6e-05, "loss": 3.6408, "step": 164 }, { "epoch": 0.04483695652173913, "grad_norm": 1.7092500925064087, "learning_rate": 1.61e-05, "loss": 3.833, "step": 165 }, { "epoch": 0.045108695652173916, "grad_norm": 2.1334292888641357, "learning_rate": 1.62e-05, "loss": 4.5951, "step": 166 }, { "epoch": 0.0453804347826087, "grad_norm": 1.352221965789795, "learning_rate": 1.63e-05, "loss": 3.804, "step": 167 }, { "epoch": 0.04565217391304348, "grad_norm": 2.178356409072876, "learning_rate": 1.64e-05, "loss": 3.7594, "step": 168 }, { "epoch": 0.04592391304347826, "grad_norm": 2.9925286769866943, "learning_rate": 1.65e-05, "loss": 3.8819, "step": 169 }, { "epoch": 0.04619565217391304, "grad_norm": 3.0277819633483887, "learning_rate": 1.66e-05, "loss": 4.4896, "step": 170 }, { "epoch": 0.04646739130434783, "grad_norm": 1.459825873374939, "learning_rate": 1.67e-05, "loss": 4.1701, "step": 171 }, { "epoch": 0.04673913043478261, "grad_norm": 1.6663670539855957, "learning_rate": 1.6800000000000002e-05, "loss": 3.6651, "step": 172 }, { "epoch": 0.04701086956521739, "grad_norm": 1.7755922079086304, "learning_rate": 1.69e-05, "loss": 3.3178, "step": 173 }, { "epoch": 0.04728260869565217, "grad_norm": 1.4884556531906128, "learning_rate": 1.7e-05, "loss": 3.6467, "step": 174 }, { "epoch": 0.04755434782608696, "grad_norm": 2.701298713684082, "learning_rate": 1.71e-05, "loss": 4.773, "step": 175 }, { "epoch": 0.04782608695652174, "grad_norm": 3.169405937194824, "learning_rate": 1.72e-05, "loss": 4.1844, "step": 176 }, { "epoch": 0.04809782608695652, "grad_norm": 2.4385275840759277, "learning_rate": 1.73e-05, "loss": 3.8691, "step": 177 }, { "epoch": 0.0483695652173913, "grad_norm": 3.446741819381714, "learning_rate": 1.74e-05, "loss": 3.6733, "step": 178 }, { "epoch": 0.048641304347826084, "grad_norm": 4.095767974853516, "learning_rate": 1.7500000000000002e-05, "loss": 3.927, "step": 179 }, { "epoch": 0.04891304347826087, "grad_norm": 3.42309308052063, "learning_rate": 1.76e-05, "loss": 4.152, "step": 180 }, { "epoch": 0.04918478260869565, "grad_norm": 1.8505162000656128, "learning_rate": 1.77e-05, "loss": 3.5489, "step": 181 }, { "epoch": 0.049456521739130434, "grad_norm": 3.3934953212738037, "learning_rate": 1.7800000000000002e-05, "loss": 3.9161, "step": 182 }, { "epoch": 0.049728260869565215, "grad_norm": 3.2479214668273926, "learning_rate": 1.79e-05, "loss": 3.3177, "step": 183 }, { "epoch": 0.05, "grad_norm": 4.14975643157959, "learning_rate": 1.8e-05, "loss": 4.0072, "step": 184 }, { "epoch": 0.050271739130434784, "grad_norm": 4.540154457092285, "learning_rate": 1.8100000000000003e-05, "loss": 3.878, "step": 185 }, { "epoch": 0.050543478260869565, "grad_norm": 2.0964109897613525, "learning_rate": 1.8200000000000002e-05, "loss": 3.6273, "step": 186 }, { "epoch": 0.050815217391304346, "grad_norm": 2.119013547897339, "learning_rate": 1.83e-05, "loss": 3.5527, "step": 187 }, { "epoch": 0.051086956521739134, "grad_norm": 4.674685478210449, "learning_rate": 1.84e-05, "loss": 4.1131, "step": 188 }, { "epoch": 0.051358695652173915, "grad_norm": 7.127579689025879, "learning_rate": 1.8500000000000002e-05, "loss": 4.0545, "step": 189 }, { "epoch": 0.051630434782608696, "grad_norm": 3.285233497619629, "learning_rate": 1.86e-05, "loss": 3.6065, "step": 190 }, { "epoch": 0.05190217391304348, "grad_norm": 1.405059576034546, "learning_rate": 1.87e-05, "loss": 3.4746, "step": 191 }, { "epoch": 0.05217391304347826, "grad_norm": 3.002393960952759, "learning_rate": 1.8800000000000003e-05, "loss": 3.9168, "step": 192 }, { "epoch": 0.052445652173913046, "grad_norm": 2.253774404525757, "learning_rate": 1.8900000000000002e-05, "loss": 3.0605, "step": 193 }, { "epoch": 0.05271739130434783, "grad_norm": 3.4486026763916016, "learning_rate": 1.9e-05, "loss": 3.6955, "step": 194 }, { "epoch": 0.05298913043478261, "grad_norm": 2.6076107025146484, "learning_rate": 1.9100000000000003e-05, "loss": 4.7168, "step": 195 }, { "epoch": 0.05326086956521739, "grad_norm": 2.173379421234131, "learning_rate": 1.9200000000000003e-05, "loss": 3.9563, "step": 196 }, { "epoch": 0.05353260869565218, "grad_norm": 1.8571799993515015, "learning_rate": 1.9299999999999998e-05, "loss": 3.531, "step": 197 }, { "epoch": 0.05380434782608696, "grad_norm": 2.940171957015991, "learning_rate": 1.9399999999999997e-05, "loss": 3.9534, "step": 198 }, { "epoch": 0.05407608695652174, "grad_norm": 1.6403425931930542, "learning_rate": 1.95e-05, "loss": 3.7979, "step": 199 }, { "epoch": 0.05434782608695652, "grad_norm": 1.9086182117462158, "learning_rate": 1.96e-05, "loss": 3.3714, "step": 200 }, { "epoch": 0.0546195652173913, "grad_norm": 2.42441463470459, "learning_rate": 1.9699999999999998e-05, "loss": 3.6953, "step": 201 }, { "epoch": 0.05489130434782609, "grad_norm": 1.7879725694656372, "learning_rate": 1.98e-05, "loss": 3.424, "step": 202 }, { "epoch": 0.05516304347826087, "grad_norm": 1.951125144958496, "learning_rate": 1.99e-05, "loss": 4.0087, "step": 203 }, { "epoch": 0.05543478260869565, "grad_norm": 1.8883895874023438, "learning_rate": 1.9999999999999998e-05, "loss": 3.4857, "step": 204 }, { "epoch": 0.05570652173913043, "grad_norm": 2.1979916095733643, "learning_rate": 2.01e-05, "loss": 3.8898, "step": 205 }, { "epoch": 0.05597826086956522, "grad_norm": 1.6748876571655273, "learning_rate": 2.02e-05, "loss": 3.5928, "step": 206 }, { "epoch": 0.05625, "grad_norm": 4.129452705383301, "learning_rate": 2.03e-05, "loss": 4.2804, "step": 207 }, { "epoch": 0.05652173913043478, "grad_norm": 2.1873629093170166, "learning_rate": 2.04e-05, "loss": 3.5242, "step": 208 }, { "epoch": 0.05679347826086956, "grad_norm": 3.064663887023926, "learning_rate": 2.05e-05, "loss": 3.379, "step": 209 }, { "epoch": 0.057065217391304345, "grad_norm": 2.8077127933502197, "learning_rate": 2.06e-05, "loss": 3.4032, "step": 210 }, { "epoch": 0.05733695652173913, "grad_norm": 2.902247905731201, "learning_rate": 2.07e-05, "loss": 3.9496, "step": 211 }, { "epoch": 0.057608695652173914, "grad_norm": 2.3877575397491455, "learning_rate": 2.08e-05, "loss": 3.5266, "step": 212 }, { "epoch": 0.057880434782608695, "grad_norm": 4.157260417938232, "learning_rate": 2.09e-05, "loss": 4.0795, "step": 213 }, { "epoch": 0.058152173913043476, "grad_norm": 4.117393970489502, "learning_rate": 2.1e-05, "loss": 3.3584, "step": 214 }, { "epoch": 0.058423913043478264, "grad_norm": 3.1118719577789307, "learning_rate": 2.11e-05, "loss": 3.5, "step": 215 }, { "epoch": 0.058695652173913045, "grad_norm": 3.3953042030334473, "learning_rate": 2.12e-05, "loss": 3.6948, "step": 216 }, { "epoch": 0.058967391304347826, "grad_norm": 4.036997318267822, "learning_rate": 2.13e-05, "loss": 3.6871, "step": 217 }, { "epoch": 0.05923913043478261, "grad_norm": 4.334664821624756, "learning_rate": 2.1400000000000002e-05, "loss": 4.0403, "step": 218 }, { "epoch": 0.059510869565217395, "grad_norm": 3.2473344802856445, "learning_rate": 2.15e-05, "loss": 2.8208, "step": 219 }, { "epoch": 0.059782608695652176, "grad_norm": 5.333037376403809, "learning_rate": 2.16e-05, "loss": 3.4915, "step": 220 }, { "epoch": 0.06005434782608696, "grad_norm": 4.594522953033447, "learning_rate": 2.1700000000000002e-05, "loss": 3.773, "step": 221 }, { "epoch": 0.06032608695652174, "grad_norm": 2.8913493156433105, "learning_rate": 2.18e-05, "loss": 3.5039, "step": 222 }, { "epoch": 0.06059782608695652, "grad_norm": 3.1354737281799316, "learning_rate": 2.19e-05, "loss": 3.5823, "step": 223 }, { "epoch": 0.06086956521739131, "grad_norm": 4.392479419708252, "learning_rate": 2.2e-05, "loss": 3.8093, "step": 224 }, { "epoch": 0.06114130434782609, "grad_norm": 3.4441139698028564, "learning_rate": 2.2100000000000002e-05, "loss": 3.9232, "step": 225 }, { "epoch": 0.06141304347826087, "grad_norm": 2.799862861633301, "learning_rate": 2.22e-05, "loss": 3.9281, "step": 226 }, { "epoch": 0.06168478260869565, "grad_norm": 2.910454750061035, "learning_rate": 2.23e-05, "loss": 3.635, "step": 227 }, { "epoch": 0.06195652173913044, "grad_norm": 3.9778056144714355, "learning_rate": 2.2400000000000002e-05, "loss": 3.5988, "step": 228 }, { "epoch": 0.06222826086956522, "grad_norm": 3.500877857208252, "learning_rate": 2.25e-05, "loss": 3.2595, "step": 229 }, { "epoch": 0.0625, "grad_norm": 3.3998258113861084, "learning_rate": 2.26e-05, "loss": 3.9116, "step": 230 }, { "epoch": 0.06277173913043478, "grad_norm": 2.585266351699829, "learning_rate": 2.2700000000000003e-05, "loss": 3.6087, "step": 231 }, { "epoch": 0.06304347826086956, "grad_norm": 2.313265085220337, "learning_rate": 2.2800000000000002e-05, "loss": 3.8507, "step": 232 }, { "epoch": 0.06331521739130434, "grad_norm": 6.233187198638916, "learning_rate": 2.29e-05, "loss": 4.6271, "step": 233 }, { "epoch": 0.06358695652173912, "grad_norm": 4.387447357177734, "learning_rate": 2.3000000000000003e-05, "loss": 3.7921, "step": 234 }, { "epoch": 0.06385869565217392, "grad_norm": 3.573700189590454, "learning_rate": 2.3100000000000002e-05, "loss": 3.6567, "step": 235 }, { "epoch": 0.0641304347826087, "grad_norm": 1.9834128618240356, "learning_rate": 2.32e-05, "loss": 3.715, "step": 236 }, { "epoch": 0.06440217391304348, "grad_norm": 3.0220143795013428, "learning_rate": 2.33e-05, "loss": 4.302, "step": 237 }, { "epoch": 0.06467391304347826, "grad_norm": 5.85452938079834, "learning_rate": 2.3400000000000003e-05, "loss": 4.3755, "step": 238 }, { "epoch": 0.06494565217391304, "grad_norm": 6.263114929199219, "learning_rate": 2.3500000000000002e-05, "loss": 4.0045, "step": 239 }, { "epoch": 0.06521739130434782, "grad_norm": 4.6512322425842285, "learning_rate": 2.3599999999999998e-05, "loss": 4.202, "step": 240 }, { "epoch": 0.0654891304347826, "grad_norm": 1.589857578277588, "learning_rate": 2.37e-05, "loss": 3.0368, "step": 241 }, { "epoch": 0.06576086956521739, "grad_norm": 3.0660102367401123, "learning_rate": 2.38e-05, "loss": 3.959, "step": 242 }, { "epoch": 0.06603260869565217, "grad_norm": 3.7164723873138428, "learning_rate": 2.3899999999999998e-05, "loss": 3.5374, "step": 243 }, { "epoch": 0.06630434782608696, "grad_norm": 3.772892713546753, "learning_rate": 2.4e-05, "loss": 3.3464, "step": 244 }, { "epoch": 0.06657608695652174, "grad_norm": 3.2034425735473633, "learning_rate": 2.41e-05, "loss": 3.3077, "step": 245 }, { "epoch": 0.06684782608695652, "grad_norm": 4.294682502746582, "learning_rate": 2.42e-05, "loss": 3.813, "step": 246 }, { "epoch": 0.0671195652173913, "grad_norm": 3.805955171585083, "learning_rate": 2.43e-05, "loss": 3.6854, "step": 247 }, { "epoch": 0.06739130434782609, "grad_norm": 4.587597370147705, "learning_rate": 2.44e-05, "loss": 3.3058, "step": 248 }, { "epoch": 0.06766304347826087, "grad_norm": 5.513325214385986, "learning_rate": 2.45e-05, "loss": 3.6714, "step": 249 }, { "epoch": 0.06793478260869565, "grad_norm": 5.820981025695801, "learning_rate": 2.4599999999999998e-05, "loss": 3.6253, "step": 250 }, { "epoch": 0.06820652173913043, "grad_norm": 4.236602306365967, "learning_rate": 2.47e-05, "loss": 3.605, "step": 251 }, { "epoch": 0.06847826086956521, "grad_norm": 3.524596929550171, "learning_rate": 2.48e-05, "loss": 3.6746, "step": 252 }, { "epoch": 0.06875, "grad_norm": 4.912044525146484, "learning_rate": 2.49e-05, "loss": 3.6485, "step": 253 }, { "epoch": 0.06902173913043479, "grad_norm": 5.730941295623779, "learning_rate": 2.5e-05, "loss": 3.7662, "step": 254 }, { "epoch": 0.06929347826086957, "grad_norm": 7.801654815673828, "learning_rate": 2.51e-05, "loss": 4.8335, "step": 255 }, { "epoch": 0.06956521739130435, "grad_norm": 4.116402626037598, "learning_rate": 2.52e-05, "loss": 3.9456, "step": 256 }, { "epoch": 0.06983695652173913, "grad_norm": 6.192750930786133, "learning_rate": 2.5300000000000002e-05, "loss": 4.0458, "step": 257 }, { "epoch": 0.07010869565217391, "grad_norm": 4.565896511077881, "learning_rate": 2.54e-05, "loss": 3.3915, "step": 258 }, { "epoch": 0.07038043478260869, "grad_norm": 4.502013206481934, "learning_rate": 2.55e-05, "loss": 3.4612, "step": 259 }, { "epoch": 0.07065217391304347, "grad_norm": 3.8033788204193115, "learning_rate": 2.5600000000000002e-05, "loss": 2.9393, "step": 260 }, { "epoch": 0.07092391304347827, "grad_norm": 4.34993839263916, "learning_rate": 2.57e-05, "loss": 3.6575, "step": 261 }, { "epoch": 0.07119565217391305, "grad_norm": 3.9064600467681885, "learning_rate": 2.58e-05, "loss": 3.6109, "step": 262 }, { "epoch": 0.07146739130434783, "grad_norm": 3.737865447998047, "learning_rate": 2.59e-05, "loss": 3.7551, "step": 263 }, { "epoch": 0.07173913043478261, "grad_norm": 4.92514181137085, "learning_rate": 2.6000000000000002e-05, "loss": 3.9133, "step": 264 }, { "epoch": 0.07201086956521739, "grad_norm": 4.838226318359375, "learning_rate": 2.61e-05, "loss": 3.4457, "step": 265 }, { "epoch": 0.07228260869565217, "grad_norm": 5.465858459472656, "learning_rate": 2.62e-05, "loss": 3.3492, "step": 266 }, { "epoch": 0.07255434782608695, "grad_norm": 4.452864646911621, "learning_rate": 2.6300000000000002e-05, "loss": 4.0503, "step": 267 }, { "epoch": 0.07282608695652174, "grad_norm": 2.657294750213623, "learning_rate": 2.64e-05, "loss": 3.5952, "step": 268 }, { "epoch": 0.07309782608695652, "grad_norm": 3.738781213760376, "learning_rate": 2.65e-05, "loss": 3.2922, "step": 269 }, { "epoch": 0.07336956521739131, "grad_norm": 4.5520524978637695, "learning_rate": 2.6600000000000003e-05, "loss": 3.4737, "step": 270 }, { "epoch": 0.07364130434782609, "grad_norm": 6.709564685821533, "learning_rate": 2.6700000000000002e-05, "loss": 3.8822, "step": 271 }, { "epoch": 0.07391304347826087, "grad_norm": 5.555806636810303, "learning_rate": 2.68e-05, "loss": 4.1245, "step": 272 }, { "epoch": 0.07418478260869565, "grad_norm": 3.99271821975708, "learning_rate": 2.69e-05, "loss": 3.535, "step": 273 }, { "epoch": 0.07445652173913044, "grad_norm": 3.0163462162017822, "learning_rate": 2.7000000000000002e-05, "loss": 3.4727, "step": 274 }, { "epoch": 0.07472826086956522, "grad_norm": 5.363269329071045, "learning_rate": 2.71e-05, "loss": 4.1682, "step": 275 }, { "epoch": 0.075, "grad_norm": 4.282686710357666, "learning_rate": 2.72e-05, "loss": 3.1355, "step": 276 }, { "epoch": 0.07527173913043478, "grad_norm": 5.983767509460449, "learning_rate": 2.7300000000000003e-05, "loss": 3.8695, "step": 277 }, { "epoch": 0.07554347826086956, "grad_norm": 5.345359802246094, "learning_rate": 2.7400000000000002e-05, "loss": 4.1882, "step": 278 }, { "epoch": 0.07581521739130435, "grad_norm": 2.5548171997070312, "learning_rate": 2.75e-05, "loss": 3.2401, "step": 279 }, { "epoch": 0.07608695652173914, "grad_norm": 3.790065050125122, "learning_rate": 2.7600000000000003e-05, "loss": 3.8057, "step": 280 }, { "epoch": 0.07635869565217392, "grad_norm": 4.037810325622559, "learning_rate": 2.7700000000000002e-05, "loss": 3.8397, "step": 281 }, { "epoch": 0.0766304347826087, "grad_norm": 4.571537017822266, "learning_rate": 2.78e-05, "loss": 3.1902, "step": 282 }, { "epoch": 0.07690217391304348, "grad_norm": 4.4513702392578125, "learning_rate": 2.79e-05, "loss": 3.6552, "step": 283 }, { "epoch": 0.07717391304347826, "grad_norm": 2.919421911239624, "learning_rate": 2.8e-05, "loss": 3.0449, "step": 284 }, { "epoch": 0.07744565217391304, "grad_norm": 3.9303693771362305, "learning_rate": 2.81e-05, "loss": 3.8186, "step": 285 }, { "epoch": 0.07771739130434782, "grad_norm": 3.8404746055603027, "learning_rate": 2.8199999999999998e-05, "loss": 3.2603, "step": 286 }, { "epoch": 0.0779891304347826, "grad_norm": 3.2469253540039062, "learning_rate": 2.83e-05, "loss": 2.917, "step": 287 }, { "epoch": 0.0782608695652174, "grad_norm": 2.8970375061035156, "learning_rate": 2.84e-05, "loss": 3.6446, "step": 288 }, { "epoch": 0.07853260869565218, "grad_norm": 4.029104232788086, "learning_rate": 2.8499999999999998e-05, "loss": 4.1754, "step": 289 }, { "epoch": 0.07880434782608696, "grad_norm": 3.967705726623535, "learning_rate": 2.86e-05, "loss": 3.3913, "step": 290 }, { "epoch": 0.07907608695652174, "grad_norm": 3.084944009780884, "learning_rate": 2.87e-05, "loss": 3.4598, "step": 291 }, { "epoch": 0.07934782608695652, "grad_norm": 3.2562334537506104, "learning_rate": 2.88e-05, "loss": 3.7326, "step": 292 }, { "epoch": 0.0796195652173913, "grad_norm": 3.9668397903442383, "learning_rate": 2.89e-05, "loss": 3.8312, "step": 293 }, { "epoch": 0.07989130434782608, "grad_norm": 3.7607946395874023, "learning_rate": 2.9e-05, "loss": 3.9309, "step": 294 }, { "epoch": 0.08016304347826086, "grad_norm": 4.704547882080078, "learning_rate": 2.91e-05, "loss": 3.9181, "step": 295 }, { "epoch": 0.08043478260869565, "grad_norm": 5.68735408782959, "learning_rate": 2.92e-05, "loss": 3.6405, "step": 296 }, { "epoch": 0.08070652173913044, "grad_norm": 4.839725971221924, "learning_rate": 2.93e-05, "loss": 3.6735, "step": 297 }, { "epoch": 0.08097826086956522, "grad_norm": 3.6157710552215576, "learning_rate": 2.94e-05, "loss": 4.1132, "step": 298 }, { "epoch": 0.08125, "grad_norm": 3.6725997924804688, "learning_rate": 2.95e-05, "loss": 3.9299, "step": 299 }, { "epoch": 0.08152173913043478, "grad_norm": 4.408240795135498, "learning_rate": 2.96e-05, "loss": 3.4141, "step": 300 }, { "epoch": 0.08179347826086956, "grad_norm": 3.8871631622314453, "learning_rate": 2.97e-05, "loss": 2.7057, "step": 301 }, { "epoch": 0.08206521739130435, "grad_norm": 5.148252010345459, "learning_rate": 2.98e-05, "loss": 3.0604, "step": 302 }, { "epoch": 0.08233695652173913, "grad_norm": 3.8631784915924072, "learning_rate": 2.9900000000000002e-05, "loss": 3.4581, "step": 303 }, { "epoch": 0.08260869565217391, "grad_norm": 2.67698073387146, "learning_rate": 3e-05, "loss": 3.1465, "step": 304 }, { "epoch": 0.08288043478260869, "grad_norm": 2.6791391372680664, "learning_rate": 3e-05, "loss": 2.8593, "step": 305 }, { "epoch": 0.08315217391304348, "grad_norm": 3.844003677368164, "learning_rate": 3e-05, "loss": 3.4371, "step": 306 }, { "epoch": 0.08342391304347826, "grad_norm": 5.624002933502197, "learning_rate": 3e-05, "loss": 3.8372, "step": 307 }, { "epoch": 0.08369565217391305, "grad_norm": 4.992491245269775, "learning_rate": 3e-05, "loss": 3.6231, "step": 308 }, { "epoch": 0.08396739130434783, "grad_norm": 4.717808246612549, "learning_rate": 3e-05, "loss": 3.9652, "step": 309 }, { "epoch": 0.08423913043478261, "grad_norm": 3.6492786407470703, "learning_rate": 3e-05, "loss": 3.7065, "step": 310 }, { "epoch": 0.08451086956521739, "grad_norm": 4.453199863433838, "learning_rate": 3e-05, "loss": 3.7317, "step": 311 }, { "epoch": 0.08478260869565217, "grad_norm": 5.009824275970459, "learning_rate": 3e-05, "loss": 3.7651, "step": 312 }, { "epoch": 0.08505434782608695, "grad_norm": 6.800715446472168, "learning_rate": 3e-05, "loss": 4.0778, "step": 313 }, { "epoch": 0.08532608695652173, "grad_norm": 6.225503444671631, "learning_rate": 3e-05, "loss": 4.3679, "step": 314 }, { "epoch": 0.08559782608695653, "grad_norm": 2.9776594638824463, "learning_rate": 3e-05, "loss": 3.1434, "step": 315 }, { "epoch": 0.08586956521739131, "grad_norm": 3.1179118156433105, "learning_rate": 3e-05, "loss": 3.5896, "step": 316 }, { "epoch": 0.08614130434782609, "grad_norm": 2.81669282913208, "learning_rate": 3e-05, "loss": 2.9054, "step": 317 }, { "epoch": 0.08641304347826087, "grad_norm": 3.894665002822876, "learning_rate": 3e-05, "loss": 3.2122, "step": 318 }, { "epoch": 0.08668478260869565, "grad_norm": 5.43796443939209, "learning_rate": 3e-05, "loss": 3.8039, "step": 319 }, { "epoch": 0.08695652173913043, "grad_norm": 3.9177141189575195, "learning_rate": 3e-05, "loss": 3.6393, "step": 320 }, { "epoch": 0.08722826086956521, "grad_norm": 3.0576508045196533, "learning_rate": 3e-05, "loss": 3.67, "step": 321 }, { "epoch": 0.0875, "grad_norm": 5.208736896514893, "learning_rate": 3e-05, "loss": 3.9719, "step": 322 }, { "epoch": 0.08777173913043479, "grad_norm": 4.839489936828613, "learning_rate": 3e-05, "loss": 3.7197, "step": 323 }, { "epoch": 0.08804347826086957, "grad_norm": 4.195175647735596, "learning_rate": 3e-05, "loss": 3.3963, "step": 324 }, { "epoch": 0.08831521739130435, "grad_norm": 3.565690279006958, "learning_rate": 3e-05, "loss": 3.177, "step": 325 }, { "epoch": 0.08858695652173913, "grad_norm": 2.6973624229431152, "learning_rate": 3e-05, "loss": 3.0416, "step": 326 }, { "epoch": 0.08885869565217391, "grad_norm": 2.620365619659424, "learning_rate": 3e-05, "loss": 3.3155, "step": 327 }, { "epoch": 0.0891304347826087, "grad_norm": 3.7461090087890625, "learning_rate": 3e-05, "loss": 3.6553, "step": 328 }, { "epoch": 0.08940217391304348, "grad_norm": 6.702042102813721, "learning_rate": 3e-05, "loss": 4.2574, "step": 329 }, { "epoch": 0.08967391304347826, "grad_norm": 5.888139247894287, "learning_rate": 3e-05, "loss": 4.148, "step": 330 }, { "epoch": 0.08994565217391304, "grad_norm": 4.737085342407227, "learning_rate": 3e-05, "loss": 3.8761, "step": 331 }, { "epoch": 0.09021739130434783, "grad_norm": 3.9981937408447266, "learning_rate": 3e-05, "loss": 4.4605, "step": 332 }, { "epoch": 0.09048913043478261, "grad_norm": 5.0993266105651855, "learning_rate": 3e-05, "loss": 4.095, "step": 333 }, { "epoch": 0.0907608695652174, "grad_norm": 4.450422763824463, "learning_rate": 3e-05, "loss": 2.8801, "step": 334 }, { "epoch": 0.09103260869565218, "grad_norm": 6.846070766448975, "learning_rate": 3e-05, "loss": 3.5946, "step": 335 }, { "epoch": 0.09130434782608696, "grad_norm": 5.887484550476074, "learning_rate": 3e-05, "loss": 3.4395, "step": 336 }, { "epoch": 0.09157608695652174, "grad_norm": 3.8128914833068848, "learning_rate": 3e-05, "loss": 3.0863, "step": 337 }, { "epoch": 0.09184782608695652, "grad_norm": 2.493584394454956, "learning_rate": 3e-05, "loss": 3.0949, "step": 338 }, { "epoch": 0.0921195652173913, "grad_norm": 3.304899215698242, "learning_rate": 3e-05, "loss": 3.3348, "step": 339 }, { "epoch": 0.09239130434782608, "grad_norm": 4.633394241333008, "learning_rate": 3e-05, "loss": 3.5233, "step": 340 }, { "epoch": 0.09266304347826088, "grad_norm": 5.405637264251709, "learning_rate": 3e-05, "loss": 3.776, "step": 341 }, { "epoch": 0.09293478260869566, "grad_norm": 5.505592346191406, "learning_rate": 3e-05, "loss": 3.9894, "step": 342 }, { "epoch": 0.09320652173913044, "grad_norm": 3.901547431945801, "learning_rate": 3e-05, "loss": 4.1865, "step": 343 }, { "epoch": 0.09347826086956522, "grad_norm": 3.3303725719451904, "learning_rate": 3e-05, "loss": 3.3656, "step": 344 }, { "epoch": 0.09375, "grad_norm": 5.3323493003845215, "learning_rate": 3e-05, "loss": 4.2369, "step": 345 }, { "epoch": 0.09402173913043478, "grad_norm": 5.352675914764404, "learning_rate": 3e-05, "loss": 3.7106, "step": 346 }, { "epoch": 0.09429347826086956, "grad_norm": 5.6517839431762695, "learning_rate": 3e-05, "loss": 3.4388, "step": 347 }, { "epoch": 0.09456521739130434, "grad_norm": 4.7645111083984375, "learning_rate": 3e-05, "loss": 3.4332, "step": 348 }, { "epoch": 0.09483695652173912, "grad_norm": 3.140958309173584, "learning_rate": 3e-05, "loss": 3.2139, "step": 349 }, { "epoch": 0.09510869565217392, "grad_norm": 2.692319393157959, "learning_rate": 3e-05, "loss": 3.3015, "step": 350 }, { "epoch": 0.0953804347826087, "grad_norm": 4.1542134284973145, "learning_rate": 3e-05, "loss": 3.29, "step": 351 }, { "epoch": 0.09565217391304348, "grad_norm": 4.946714401245117, "learning_rate": 3e-05, "loss": 3.2509, "step": 352 }, { "epoch": 0.09592391304347826, "grad_norm": 5.529204368591309, "learning_rate": 3e-05, "loss": 4.0578, "step": 353 }, { "epoch": 0.09619565217391304, "grad_norm": 3.837975025177002, "learning_rate": 3e-05, "loss": 3.7921, "step": 354 }, { "epoch": 0.09646739130434782, "grad_norm": 3.0146334171295166, "learning_rate": 3e-05, "loss": 4.0369, "step": 355 }, { "epoch": 0.0967391304347826, "grad_norm": 4.283068656921387, "learning_rate": 3e-05, "loss": 3.9506, "step": 356 }, { "epoch": 0.09701086956521739, "grad_norm": 4.30133581161499, "learning_rate": 3e-05, "loss": 3.4218, "step": 357 }, { "epoch": 0.09728260869565217, "grad_norm": 6.228667736053467, "learning_rate": 3e-05, "loss": 4.0054, "step": 358 }, { "epoch": 0.09755434782608696, "grad_norm": 3.927131414413452, "learning_rate": 3e-05, "loss": 2.9401, "step": 359 }, { "epoch": 0.09782608695652174, "grad_norm": 4.374425888061523, "learning_rate": 3e-05, "loss": 3.5279, "step": 360 }, { "epoch": 0.09809782608695652, "grad_norm": 2.559403419494629, "learning_rate": 3e-05, "loss": 2.8306, "step": 361 }, { "epoch": 0.0983695652173913, "grad_norm": 2.9556596279144287, "learning_rate": 3e-05, "loss": 3.4532, "step": 362 }, { "epoch": 0.09864130434782609, "grad_norm": 4.841480731964111, "learning_rate": 3e-05, "loss": 3.474, "step": 363 }, { "epoch": 0.09891304347826087, "grad_norm": 5.785233497619629, "learning_rate": 3e-05, "loss": 3.6698, "step": 364 }, { "epoch": 0.09918478260869565, "grad_norm": 5.180464267730713, "learning_rate": 3e-05, "loss": 3.4199, "step": 365 }, { "epoch": 0.09945652173913043, "grad_norm": 3.5438666343688965, "learning_rate": 3e-05, "loss": 3.5603, "step": 366 }, { "epoch": 0.09972826086956521, "grad_norm": 3.1015937328338623, "learning_rate": 3e-05, "loss": 2.9937, "step": 367 }, { "epoch": 0.1, "grad_norm": 4.174754619598389, "learning_rate": 3e-05, "loss": 3.5994, "step": 368 }, { "epoch": 0.10027173913043479, "grad_norm": 4.678737163543701, "learning_rate": 3e-05, "loss": 4.0816, "step": 369 }, { "epoch": 0.10054347826086957, "grad_norm": 3.884875774383545, "learning_rate": 3e-05, "loss": 2.8268, "step": 370 }, { "epoch": 0.10081521739130435, "grad_norm": 4.060309886932373, "learning_rate": 3e-05, "loss": 3.5038, "step": 371 }, { "epoch": 0.10108695652173913, "grad_norm": 4.1803812980651855, "learning_rate": 3e-05, "loss": 3.8122, "step": 372 }, { "epoch": 0.10135869565217391, "grad_norm": 3.1725456714630127, "learning_rate": 3e-05, "loss": 3.7945, "step": 373 }, { "epoch": 0.10163043478260869, "grad_norm": 3.1784756183624268, "learning_rate": 3e-05, "loss": 3.0676, "step": 374 }, { "epoch": 0.10190217391304347, "grad_norm": 4.812568187713623, "learning_rate": 3e-05, "loss": 3.593, "step": 375 }, { "epoch": 0.10217391304347827, "grad_norm": 5.944811820983887, "learning_rate": 3e-05, "loss": 3.843, "step": 376 }, { "epoch": 0.10244565217391305, "grad_norm": 4.13525915145874, "learning_rate": 3e-05, "loss": 3.9103, "step": 377 }, { "epoch": 0.10271739130434783, "grad_norm": 5.860477447509766, "learning_rate": 3e-05, "loss": 4.9384, "step": 378 }, { "epoch": 0.10298913043478261, "grad_norm": 3.2515783309936523, "learning_rate": 3e-05, "loss": 3.0586, "step": 379 }, { "epoch": 0.10326086956521739, "grad_norm": 4.5452494621276855, "learning_rate": 3e-05, "loss": 3.7046, "step": 380 }, { "epoch": 0.10353260869565217, "grad_norm": 3.8891849517822266, "learning_rate": 3e-05, "loss": 3.34, "step": 381 }, { "epoch": 0.10380434782608695, "grad_norm": 4.083732604980469, "learning_rate": 3e-05, "loss": 3.3112, "step": 382 }, { "epoch": 0.10407608695652174, "grad_norm": 3.438572406768799, "learning_rate": 3e-05, "loss": 2.9908, "step": 383 }, { "epoch": 0.10434782608695652, "grad_norm": 3.0750012397766113, "learning_rate": 3e-05, "loss": 3.3847, "step": 384 }, { "epoch": 0.10461956521739131, "grad_norm": 3.0833847522735596, "learning_rate": 3e-05, "loss": 3.2274, "step": 385 }, { "epoch": 0.10489130434782609, "grad_norm": 4.356601238250732, "learning_rate": 3e-05, "loss": 3.9247, "step": 386 }, { "epoch": 0.10516304347826087, "grad_norm": 4.894799709320068, "learning_rate": 3e-05, "loss": 3.6005, "step": 387 }, { "epoch": 0.10543478260869565, "grad_norm": 4.887790203094482, "learning_rate": 3e-05, "loss": 4.3258, "step": 388 }, { "epoch": 0.10570652173913044, "grad_norm": 4.861374855041504, "learning_rate": 3e-05, "loss": 3.6857, "step": 389 }, { "epoch": 0.10597826086956522, "grad_norm": 4.775892734527588, "learning_rate": 3e-05, "loss": 4.3695, "step": 390 }, { "epoch": 0.10625, "grad_norm": 4.2715582847595215, "learning_rate": 3e-05, "loss": 3.7657, "step": 391 }, { "epoch": 0.10652173913043478, "grad_norm": 3.8440310955047607, "learning_rate": 3e-05, "loss": 3.4856, "step": 392 }, { "epoch": 0.10679347826086956, "grad_norm": 4.285994529724121, "learning_rate": 3e-05, "loss": 3.4081, "step": 393 }, { "epoch": 0.10706521739130435, "grad_norm": 2.1318747997283936, "learning_rate": 3e-05, "loss": 2.9345, "step": 394 }, { "epoch": 0.10733695652173914, "grad_norm": 2.4292728900909424, "learning_rate": 3e-05, "loss": 3.091, "step": 395 }, { "epoch": 0.10760869565217392, "grad_norm": 2.3496248722076416, "learning_rate": 3e-05, "loss": 3.1994, "step": 396 }, { "epoch": 0.1078804347826087, "grad_norm": 5.604576110839844, "learning_rate": 3e-05, "loss": 4.6392, "step": 397 }, { "epoch": 0.10815217391304348, "grad_norm": 5.448110580444336, "learning_rate": 3e-05, "loss": 4.4596, "step": 398 }, { "epoch": 0.10842391304347826, "grad_norm": 3.2921388149261475, "learning_rate": 3e-05, "loss": 3.5248, "step": 399 }, { "epoch": 0.10869565217391304, "grad_norm": 3.1395716667175293, "learning_rate": 3e-05, "loss": 4.1658, "step": 400 }, { "epoch": 0.10896739130434782, "grad_norm": 2.3313381671905518, "learning_rate": 3e-05, "loss": 3.7592, "step": 401 }, { "epoch": 0.1092391304347826, "grad_norm": 2.9619717597961426, "learning_rate": 3e-05, "loss": 3.5212, "step": 402 }, { "epoch": 0.1095108695652174, "grad_norm": 3.2616071701049805, "learning_rate": 3e-05, "loss": 3.2052, "step": 403 }, { "epoch": 0.10978260869565218, "grad_norm": 4.842064380645752, "learning_rate": 3e-05, "loss": 3.907, "step": 404 }, { "epoch": 0.11005434782608696, "grad_norm": 3.6140100955963135, "learning_rate": 3e-05, "loss": 4.061, "step": 405 }, { "epoch": 0.11032608695652174, "grad_norm": 1.4534454345703125, "learning_rate": 3e-05, "loss": 3.0512, "step": 406 }, { "epoch": 0.11059782608695652, "grad_norm": 2.4747161865234375, "learning_rate": 3e-05, "loss": 3.7008, "step": 407 }, { "epoch": 0.1108695652173913, "grad_norm": 3.0130977630615234, "learning_rate": 3e-05, "loss": 3.9765, "step": 408 }, { "epoch": 0.11114130434782608, "grad_norm": 2.256558656692505, "learning_rate": 3e-05, "loss": 3.2158, "step": 409 }, { "epoch": 0.11141304347826086, "grad_norm": 4.283812522888184, "learning_rate": 3e-05, "loss": 4.4811, "step": 410 }, { "epoch": 0.11168478260869565, "grad_norm": 2.566425323486328, "learning_rate": 3e-05, "loss": 3.9403, "step": 411 }, { "epoch": 0.11195652173913044, "grad_norm": 2.4623820781707764, "learning_rate": 3e-05, "loss": 3.2361, "step": 412 }, { "epoch": 0.11222826086956522, "grad_norm": 2.934722661972046, "learning_rate": 3e-05, "loss": 3.0554, "step": 413 }, { "epoch": 0.1125, "grad_norm": 5.524335861206055, "learning_rate": 3e-05, "loss": 4.0103, "step": 414 }, { "epoch": 0.11277173913043478, "grad_norm": 3.285830020904541, "learning_rate": 3e-05, "loss": 3.6652, "step": 415 }, { "epoch": 0.11304347826086956, "grad_norm": 2.5553982257843018, "learning_rate": 3e-05, "loss": 3.8115, "step": 416 }, { "epoch": 0.11331521739130435, "grad_norm": 1.8953039646148682, "learning_rate": 3e-05, "loss": 3.6769, "step": 417 }, { "epoch": 0.11358695652173913, "grad_norm": 2.4014382362365723, "learning_rate": 3e-05, "loss": 3.6343, "step": 418 }, { "epoch": 0.11385869565217391, "grad_norm": 3.2362847328186035, "learning_rate": 3e-05, "loss": 3.722, "step": 419 }, { "epoch": 0.11413043478260869, "grad_norm": 2.525406837463379, "learning_rate": 3e-05, "loss": 3.2201, "step": 420 }, { "epoch": 0.11440217391304348, "grad_norm": 2.492753028869629, "learning_rate": 3e-05, "loss": 3.3864, "step": 421 }, { "epoch": 0.11467391304347826, "grad_norm": 1.6389284133911133, "learning_rate": 3e-05, "loss": 3.1505, "step": 422 }, { "epoch": 0.11494565217391305, "grad_norm": 2.721895933151245, "learning_rate": 3e-05, "loss": 4.02, "step": 423 }, { "epoch": 0.11521739130434783, "grad_norm": 2.1239542961120605, "learning_rate": 3e-05, "loss": 3.3395, "step": 424 }, { "epoch": 0.11548913043478261, "grad_norm": 2.3277628421783447, "learning_rate": 3e-05, "loss": 4.3061, "step": 425 }, { "epoch": 0.11576086956521739, "grad_norm": 1.7965201139450073, "learning_rate": 3e-05, "loss": 3.8778, "step": 426 }, { "epoch": 0.11603260869565217, "grad_norm": 2.9034996032714844, "learning_rate": 3e-05, "loss": 3.8348, "step": 427 }, { "epoch": 0.11630434782608695, "grad_norm": 2.0713295936584473, "learning_rate": 3e-05, "loss": 3.6901, "step": 428 }, { "epoch": 0.11657608695652173, "grad_norm": 2.612128257751465, "learning_rate": 3e-05, "loss": 3.1308, "step": 429 }, { "epoch": 0.11684782608695653, "grad_norm": 3.535771369934082, "learning_rate": 3e-05, "loss": 3.7629, "step": 430 }, { "epoch": 0.11711956521739131, "grad_norm": 2.6418893337249756, "learning_rate": 3e-05, "loss": 3.351, "step": 431 }, { "epoch": 0.11739130434782609, "grad_norm": 1.9122227430343628, "learning_rate": 3e-05, "loss": 2.9162, "step": 432 }, { "epoch": 0.11766304347826087, "grad_norm": 3.12231707572937, "learning_rate": 3e-05, "loss": 3.6215, "step": 433 }, { "epoch": 0.11793478260869565, "grad_norm": 2.5034008026123047, "learning_rate": 3e-05, "loss": 3.6675, "step": 434 }, { "epoch": 0.11820652173913043, "grad_norm": 2.69254732131958, "learning_rate": 3e-05, "loss": 3.8847, "step": 435 }, { "epoch": 0.11847826086956521, "grad_norm": 2.026611566543579, "learning_rate": 3e-05, "loss": 3.9927, "step": 436 }, { "epoch": 0.11875, "grad_norm": 3.05465030670166, "learning_rate": 3e-05, "loss": 3.4363, "step": 437 }, { "epoch": 0.11902173913043479, "grad_norm": 2.800114631652832, "learning_rate": 3e-05, "loss": 4.0161, "step": 438 }, { "epoch": 0.11929347826086957, "grad_norm": 2.452852725982666, "learning_rate": 3e-05, "loss": 3.8825, "step": 439 }, { "epoch": 0.11956521739130435, "grad_norm": 2.425046920776367, "learning_rate": 3e-05, "loss": 3.8342, "step": 440 }, { "epoch": 0.11983695652173913, "grad_norm": 1.6898943185806274, "learning_rate": 3e-05, "loss": 3.4458, "step": 441 }, { "epoch": 0.12010869565217391, "grad_norm": 2.265521764755249, "learning_rate": 3e-05, "loss": 3.498, "step": 442 }, { "epoch": 0.1203804347826087, "grad_norm": 3.3182766437530518, "learning_rate": 3e-05, "loss": 4.0802, "step": 443 }, { "epoch": 0.12065217391304348, "grad_norm": 3.441931962966919, "learning_rate": 3e-05, "loss": 3.7694, "step": 444 }, { "epoch": 0.12092391304347826, "grad_norm": 2.1982293128967285, "learning_rate": 3e-05, "loss": 3.3313, "step": 445 }, { "epoch": 0.12119565217391304, "grad_norm": 1.986126184463501, "learning_rate": 3e-05, "loss": 3.3758, "step": 446 }, { "epoch": 0.12146739130434783, "grad_norm": 1.6694939136505127, "learning_rate": 3e-05, "loss": 3.1291, "step": 447 }, { "epoch": 0.12173913043478261, "grad_norm": 3.418254852294922, "learning_rate": 3e-05, "loss": 3.9384, "step": 448 }, { "epoch": 0.1220108695652174, "grad_norm": 2.4987125396728516, "learning_rate": 3e-05, "loss": 4.0415, "step": 449 }, { "epoch": 0.12228260869565218, "grad_norm": 2.4183948040008545, "learning_rate": 3e-05, "loss": 3.4816, "step": 450 }, { "epoch": 0.12255434782608696, "grad_norm": 2.753169059753418, "learning_rate": 3e-05, "loss": 3.3146, "step": 451 }, { "epoch": 0.12282608695652174, "grad_norm": 2.2018933296203613, "learning_rate": 3e-05, "loss": 3.5041, "step": 452 }, { "epoch": 0.12309782608695652, "grad_norm": 1.910427451133728, "learning_rate": 3e-05, "loss": 3.1185, "step": 453 }, { "epoch": 0.1233695652173913, "grad_norm": 2.2100231647491455, "learning_rate": 3e-05, "loss": 3.1726, "step": 454 }, { "epoch": 0.12364130434782608, "grad_norm": 2.6990854740142822, "learning_rate": 3e-05, "loss": 3.7998, "step": 455 }, { "epoch": 0.12391304347826088, "grad_norm": 3.1403119564056396, "learning_rate": 3e-05, "loss": 3.3689, "step": 456 }, { "epoch": 0.12418478260869566, "grad_norm": 1.9951990842819214, "learning_rate": 3e-05, "loss": 3.0753, "step": 457 }, { "epoch": 0.12445652173913044, "grad_norm": 2.348395347595215, "learning_rate": 3e-05, "loss": 3.8548, "step": 458 }, { "epoch": 0.12472826086956522, "grad_norm": 2.324969530105591, "learning_rate": 3e-05, "loss": 3.3703, "step": 459 }, { "epoch": 0.125, "grad_norm": 2.897873640060425, "learning_rate": 3e-05, "loss": 3.4589, "step": 460 }, { "epoch": 0.1252717391304348, "grad_norm": 2.81925892829895, "learning_rate": 3e-05, "loss": 3.4818, "step": 461 }, { "epoch": 0.12554347826086956, "grad_norm": 4.320977687835693, "learning_rate": 3e-05, "loss": 4.3366, "step": 462 }, { "epoch": 0.12581521739130436, "grad_norm": 3.2348084449768066, "learning_rate": 3e-05, "loss": 4.1263, "step": 463 }, { "epoch": 0.12608695652173912, "grad_norm": 2.6263833045959473, "learning_rate": 3e-05, "loss": 2.9206, "step": 464 }, { "epoch": 0.12635869565217392, "grad_norm": 4.796882629394531, "learning_rate": 3e-05, "loss": 3.8013, "step": 465 }, { "epoch": 0.1266304347826087, "grad_norm": 5.0804619789123535, "learning_rate": 3e-05, "loss": 3.8332, "step": 466 }, { "epoch": 0.12690217391304348, "grad_norm": 3.0131075382232666, "learning_rate": 3e-05, "loss": 3.5667, "step": 467 }, { "epoch": 0.12717391304347825, "grad_norm": 1.6071690320968628, "learning_rate": 3e-05, "loss": 2.8863, "step": 468 }, { "epoch": 0.12744565217391304, "grad_norm": 2.211411476135254, "learning_rate": 3e-05, "loss": 3.2423, "step": 469 }, { "epoch": 0.12771739130434784, "grad_norm": 2.4000439643859863, "learning_rate": 3e-05, "loss": 3.3725, "step": 470 }, { "epoch": 0.1279891304347826, "grad_norm": 2.421668529510498, "learning_rate": 3e-05, "loss": 3.6009, "step": 471 }, { "epoch": 0.1282608695652174, "grad_norm": 1.906244158744812, "learning_rate": 3e-05, "loss": 3.4958, "step": 472 }, { "epoch": 0.12853260869565217, "grad_norm": 1.7820132970809937, "learning_rate": 3e-05, "loss": 3.0321, "step": 473 }, { "epoch": 0.12880434782608696, "grad_norm": 2.929561138153076, "learning_rate": 3e-05, "loss": 3.788, "step": 474 }, { "epoch": 0.12907608695652173, "grad_norm": 2.740144968032837, "learning_rate": 3e-05, "loss": 3.6582, "step": 475 }, { "epoch": 0.12934782608695652, "grad_norm": 2.153079032897949, "learning_rate": 3e-05, "loss": 3.6739, "step": 476 }, { "epoch": 0.1296195652173913, "grad_norm": 1.366892695426941, "learning_rate": 3e-05, "loss": 3.5163, "step": 477 }, { "epoch": 0.1298913043478261, "grad_norm": 2.3881099224090576, "learning_rate": 3e-05, "loss": 3.7265, "step": 478 }, { "epoch": 0.13016304347826088, "grad_norm": 1.72128164768219, "learning_rate": 3e-05, "loss": 2.9679, "step": 479 }, { "epoch": 0.13043478260869565, "grad_norm": 1.8024574518203735, "learning_rate": 3e-05, "loss": 3.6759, "step": 480 }, { "epoch": 0.13070652173913044, "grad_norm": 2.3349437713623047, "learning_rate": 3e-05, "loss": 3.9231, "step": 481 }, { "epoch": 0.1309782608695652, "grad_norm": 2.60579776763916, "learning_rate": 3e-05, "loss": 3.5484, "step": 482 }, { "epoch": 0.13125, "grad_norm": 1.823648452758789, "learning_rate": 3e-05, "loss": 3.409, "step": 483 }, { "epoch": 0.13152173913043477, "grad_norm": 1.939772367477417, "learning_rate": 3e-05, "loss": 3.7185, "step": 484 }, { "epoch": 0.13179347826086957, "grad_norm": 1.7059251070022583, "learning_rate": 3e-05, "loss": 3.2223, "step": 485 }, { "epoch": 0.13206521739130433, "grad_norm": 2.1309432983398438, "learning_rate": 3e-05, "loss": 3.5143, "step": 486 }, { "epoch": 0.13233695652173913, "grad_norm": 2.5169553756713867, "learning_rate": 3e-05, "loss": 3.9375, "step": 487 }, { "epoch": 0.13260869565217392, "grad_norm": 1.6783947944641113, "learning_rate": 3e-05, "loss": 3.0069, "step": 488 }, { "epoch": 0.1328804347826087, "grad_norm": 2.5426528453826904, "learning_rate": 3e-05, "loss": 3.4032, "step": 489 }, { "epoch": 0.1331521739130435, "grad_norm": 2.8521511554718018, "learning_rate": 3e-05, "loss": 3.1236, "step": 490 }, { "epoch": 0.13342391304347825, "grad_norm": 2.7477147579193115, "learning_rate": 3e-05, "loss": 3.0802, "step": 491 }, { "epoch": 0.13369565217391305, "grad_norm": 2.785457134246826, "learning_rate": 3e-05, "loss": 3.8152, "step": 492 }, { "epoch": 0.13396739130434782, "grad_norm": 3.6600492000579834, "learning_rate": 3e-05, "loss": 3.6564, "step": 493 }, { "epoch": 0.1342391304347826, "grad_norm": 4.777431488037109, "learning_rate": 3e-05, "loss": 3.7449, "step": 494 }, { "epoch": 0.13451086956521738, "grad_norm": 4.060612201690674, "learning_rate": 3e-05, "loss": 4.1282, "step": 495 }, { "epoch": 0.13478260869565217, "grad_norm": 4.2701592445373535, "learning_rate": 3e-05, "loss": 4.6185, "step": 496 }, { "epoch": 0.13505434782608697, "grad_norm": 3.9511125087738037, "learning_rate": 3e-05, "loss": 3.2463, "step": 497 }, { "epoch": 0.13532608695652174, "grad_norm": 3.8475091457366943, "learning_rate": 3e-05, "loss": 3.0689, "step": 498 }, { "epoch": 0.13559782608695653, "grad_norm": 5.655093669891357, "learning_rate": 3e-05, "loss": 4.024, "step": 499 }, { "epoch": 0.1358695652173913, "grad_norm": 3.6835744380950928, "learning_rate": 3e-05, "loss": 3.1939, "step": 500 }, { "epoch": 0.1361413043478261, "grad_norm": 3.2142796516418457, "learning_rate": 3e-05, "loss": 4.0483, "step": 501 }, { "epoch": 0.13641304347826086, "grad_norm": 3.1059064865112305, "learning_rate": 3e-05, "loss": 3.8777, "step": 502 }, { "epoch": 0.13668478260869565, "grad_norm": 2.3298041820526123, "learning_rate": 3e-05, "loss": 3.2728, "step": 503 }, { "epoch": 0.13695652173913042, "grad_norm": 3.318361759185791, "learning_rate": 3e-05, "loss": 4.0421, "step": 504 }, { "epoch": 0.13722826086956522, "grad_norm": 2.8279385566711426, "learning_rate": 3e-05, "loss": 3.49, "step": 505 }, { "epoch": 0.1375, "grad_norm": 2.233304023742676, "learning_rate": 3e-05, "loss": 3.3737, "step": 506 }, { "epoch": 0.13777173913043478, "grad_norm": 2.943725824356079, "learning_rate": 3e-05, "loss": 3.5774, "step": 507 }, { "epoch": 0.13804347826086957, "grad_norm": 2.764371871948242, "learning_rate": 3e-05, "loss": 3.7041, "step": 508 }, { "epoch": 0.13831521739130434, "grad_norm": 2.647728443145752, "learning_rate": 3e-05, "loss": 3.0545, "step": 509 }, { "epoch": 0.13858695652173914, "grad_norm": 3.1184072494506836, "learning_rate": 3e-05, "loss": 3.2394, "step": 510 }, { "epoch": 0.1388586956521739, "grad_norm": 1.9611910581588745, "learning_rate": 3e-05, "loss": 3.4171, "step": 511 }, { "epoch": 0.1391304347826087, "grad_norm": 1.5122188329696655, "learning_rate": 3e-05, "loss": 3.1593, "step": 512 }, { "epoch": 0.13940217391304346, "grad_norm": 1.9166432619094849, "learning_rate": 3e-05, "loss": 3.21, "step": 513 }, { "epoch": 0.13967391304347826, "grad_norm": 2.1466691493988037, "learning_rate": 3e-05, "loss": 3.4333, "step": 514 }, { "epoch": 0.13994565217391305, "grad_norm": 2.217186689376831, "learning_rate": 3e-05, "loss": 3.3885, "step": 515 }, { "epoch": 0.14021739130434782, "grad_norm": 2.03560733795166, "learning_rate": 3e-05, "loss": 3.9298, "step": 516 }, { "epoch": 0.14048913043478262, "grad_norm": 2.214279890060425, "learning_rate": 3e-05, "loss": 3.2791, "step": 517 }, { "epoch": 0.14076086956521738, "grad_norm": 3.428405523300171, "learning_rate": 3e-05, "loss": 3.668, "step": 518 }, { "epoch": 0.14103260869565218, "grad_norm": 3.077150344848633, "learning_rate": 3e-05, "loss": 3.7187, "step": 519 }, { "epoch": 0.14130434782608695, "grad_norm": 1.3983957767486572, "learning_rate": 3e-05, "loss": 3.3129, "step": 520 }, { "epoch": 0.14157608695652174, "grad_norm": 2.0241615772247314, "learning_rate": 3e-05, "loss": 3.6655, "step": 521 }, { "epoch": 0.14184782608695654, "grad_norm": 2.20249080657959, "learning_rate": 3e-05, "loss": 3.2375, "step": 522 }, { "epoch": 0.1421195652173913, "grad_norm": 2.2980117797851562, "learning_rate": 3e-05, "loss": 3.7075, "step": 523 }, { "epoch": 0.1423913043478261, "grad_norm": 1.2177653312683105, "learning_rate": 3e-05, "loss": 3.2919, "step": 524 }, { "epoch": 0.14266304347826086, "grad_norm": 1.0807304382324219, "learning_rate": 3e-05, "loss": 3.2761, "step": 525 }, { "epoch": 0.14293478260869566, "grad_norm": 1.4918198585510254, "learning_rate": 3e-05, "loss": 3.6412, "step": 526 }, { "epoch": 0.14320652173913043, "grad_norm": 1.3037655353546143, "learning_rate": 3e-05, "loss": 3.4596, "step": 527 }, { "epoch": 0.14347826086956522, "grad_norm": 1.1929515600204468, "learning_rate": 3e-05, "loss": 3.4554, "step": 528 }, { "epoch": 0.14375, "grad_norm": 1.7983393669128418, "learning_rate": 3e-05, "loss": 3.0941, "step": 529 }, { "epoch": 0.14402173913043478, "grad_norm": 1.4893770217895508, "learning_rate": 3e-05, "loss": 3.688, "step": 530 }, { "epoch": 0.14429347826086958, "grad_norm": 1.4558385610580444, "learning_rate": 3e-05, "loss": 3.6866, "step": 531 }, { "epoch": 0.14456521739130435, "grad_norm": 1.4489972591400146, "learning_rate": 3e-05, "loss": 3.1238, "step": 532 }, { "epoch": 0.14483695652173914, "grad_norm": 1.603597640991211, "learning_rate": 3e-05, "loss": 2.9605, "step": 533 }, { "epoch": 0.1451086956521739, "grad_norm": 1.1347497701644897, "learning_rate": 3e-05, "loss": 3.2066, "step": 534 }, { "epoch": 0.1453804347826087, "grad_norm": 1.9727925062179565, "learning_rate": 3e-05, "loss": 3.6686, "step": 535 }, { "epoch": 0.14565217391304347, "grad_norm": 2.5670254230499268, "learning_rate": 3e-05, "loss": 3.3345, "step": 536 }, { "epoch": 0.14592391304347826, "grad_norm": 2.017517328262329, "learning_rate": 3e-05, "loss": 3.7092, "step": 537 }, { "epoch": 0.14619565217391303, "grad_norm": 2.1940996646881104, "learning_rate": 3e-05, "loss": 3.807, "step": 538 }, { "epoch": 0.14646739130434783, "grad_norm": 1.7565933465957642, "learning_rate": 3e-05, "loss": 3.6645, "step": 539 }, { "epoch": 0.14673913043478262, "grad_norm": 2.248276948928833, "learning_rate": 3e-05, "loss": 3.979, "step": 540 }, { "epoch": 0.1470108695652174, "grad_norm": 2.3259811401367188, "learning_rate": 3e-05, "loss": 3.3435, "step": 541 }, { "epoch": 0.14728260869565218, "grad_norm": 2.7300784587860107, "learning_rate": 3e-05, "loss": 3.7696, "step": 542 }, { "epoch": 0.14755434782608695, "grad_norm": 1.406214952468872, "learning_rate": 3e-05, "loss": 2.6787, "step": 543 }, { "epoch": 0.14782608695652175, "grad_norm": 2.6304099559783936, "learning_rate": 3e-05, "loss": 3.7273, "step": 544 }, { "epoch": 0.1480978260869565, "grad_norm": 4.7364912033081055, "learning_rate": 3e-05, "loss": 4.4759, "step": 545 }, { "epoch": 0.1483695652173913, "grad_norm": 5.363766193389893, "learning_rate": 3e-05, "loss": 4.1787, "step": 546 }, { "epoch": 0.14864130434782608, "grad_norm": 2.660940408706665, "learning_rate": 3e-05, "loss": 3.2534, "step": 547 }, { "epoch": 0.14891304347826087, "grad_norm": 3.186602830886841, "learning_rate": 3e-05, "loss": 3.5762, "step": 548 }, { "epoch": 0.14918478260869567, "grad_norm": 3.962216377258301, "learning_rate": 3e-05, "loss": 3.5002, "step": 549 }, { "epoch": 0.14945652173913043, "grad_norm": 5.419135093688965, "learning_rate": 3e-05, "loss": 3.6202, "step": 550 }, { "epoch": 0.14972826086956523, "grad_norm": 4.870014667510986, "learning_rate": 3e-05, "loss": 3.5467, "step": 551 }, { "epoch": 0.15, "grad_norm": 3.175389289855957, "learning_rate": 3e-05, "loss": 3.9385, "step": 552 }, { "epoch": 0.1502717391304348, "grad_norm": 2.783637046813965, "learning_rate": 3e-05, "loss": 3.8554, "step": 553 }, { "epoch": 0.15054347826086956, "grad_norm": 2.6941978931427, "learning_rate": 3e-05, "loss": 3.211, "step": 554 }, { "epoch": 0.15081521739130435, "grad_norm": 3.185209274291992, "learning_rate": 3e-05, "loss": 3.3476, "step": 555 }, { "epoch": 0.15108695652173912, "grad_norm": 2.922362804412842, "learning_rate": 3e-05, "loss": 3.0941, "step": 556 }, { "epoch": 0.1513586956521739, "grad_norm": 2.7833430767059326, "learning_rate": 3e-05, "loss": 3.3851, "step": 557 }, { "epoch": 0.1516304347826087, "grad_norm": 1.4858049154281616, "learning_rate": 3e-05, "loss": 2.7326, "step": 558 }, { "epoch": 0.15190217391304348, "grad_norm": 2.9602174758911133, "learning_rate": 3e-05, "loss": 3.2037, "step": 559 }, { "epoch": 0.15217391304347827, "grad_norm": 3.0462303161621094, "learning_rate": 3e-05, "loss": 3.2449, "step": 560 }, { "epoch": 0.15244565217391304, "grad_norm": 3.038954019546509, "learning_rate": 3e-05, "loss": 3.3796, "step": 561 }, { "epoch": 0.15271739130434783, "grad_norm": 1.572611689567566, "learning_rate": 3e-05, "loss": 3.1521, "step": 562 }, { "epoch": 0.1529891304347826, "grad_norm": 2.4640631675720215, "learning_rate": 3e-05, "loss": 3.9771, "step": 563 }, { "epoch": 0.1532608695652174, "grad_norm": 1.757712483406067, "learning_rate": 3e-05, "loss": 2.9893, "step": 564 }, { "epoch": 0.15353260869565216, "grad_norm": 2.606873035430908, "learning_rate": 3e-05, "loss": 3.702, "step": 565 }, { "epoch": 0.15380434782608696, "grad_norm": 1.677304983139038, "learning_rate": 3e-05, "loss": 3.2607, "step": 566 }, { "epoch": 0.15407608695652175, "grad_norm": 2.4729526042938232, "learning_rate": 3e-05, "loss": 3.9865, "step": 567 }, { "epoch": 0.15434782608695652, "grad_norm": 2.162191867828369, "learning_rate": 3e-05, "loss": 3.6349, "step": 568 }, { "epoch": 0.1546195652173913, "grad_norm": 3.027984380722046, "learning_rate": 3e-05, "loss": 3.8996, "step": 569 }, { "epoch": 0.15489130434782608, "grad_norm": 1.9820539951324463, "learning_rate": 3e-05, "loss": 3.3207, "step": 570 }, { "epoch": 0.15516304347826088, "grad_norm": 1.7880254983901978, "learning_rate": 3e-05, "loss": 3.6082, "step": 571 }, { "epoch": 0.15543478260869564, "grad_norm": 1.9113103151321411, "learning_rate": 3e-05, "loss": 3.2639, "step": 572 }, { "epoch": 0.15570652173913044, "grad_norm": 1.7418012619018555, "learning_rate": 3e-05, "loss": 3.4052, "step": 573 }, { "epoch": 0.1559782608695652, "grad_norm": 1.6183768510818481, "learning_rate": 3e-05, "loss": 3.8061, "step": 574 }, { "epoch": 0.15625, "grad_norm": 1.33986496925354, "learning_rate": 3e-05, "loss": 3.3393, "step": 575 }, { "epoch": 0.1565217391304348, "grad_norm": 1.4486945867538452, "learning_rate": 3e-05, "loss": 3.7729, "step": 576 }, { "epoch": 0.15679347826086956, "grad_norm": 1.7886120080947876, "learning_rate": 3e-05, "loss": 3.8346, "step": 577 }, { "epoch": 0.15706521739130436, "grad_norm": 1.2352441549301147, "learning_rate": 3e-05, "loss": 3.2718, "step": 578 }, { "epoch": 0.15733695652173912, "grad_norm": 1.8679944276809692, "learning_rate": 3e-05, "loss": 3.4247, "step": 579 }, { "epoch": 0.15760869565217392, "grad_norm": 1.689501404762268, "learning_rate": 3e-05, "loss": 4.0971, "step": 580 }, { "epoch": 0.1578804347826087, "grad_norm": 1.6779184341430664, "learning_rate": 3e-05, "loss": 3.8803, "step": 581 }, { "epoch": 0.15815217391304348, "grad_norm": 2.0760598182678223, "learning_rate": 3e-05, "loss": 3.5429, "step": 582 }, { "epoch": 0.15842391304347825, "grad_norm": 1.8308002948760986, "learning_rate": 3e-05, "loss": 3.4797, "step": 583 }, { "epoch": 0.15869565217391304, "grad_norm": 2.4416396617889404, "learning_rate": 3e-05, "loss": 4.2491, "step": 584 }, { "epoch": 0.15896739130434784, "grad_norm": 2.365509271621704, "learning_rate": 3e-05, "loss": 3.4862, "step": 585 }, { "epoch": 0.1592391304347826, "grad_norm": 4.06919002532959, "learning_rate": 3e-05, "loss": 4.0302, "step": 586 }, { "epoch": 0.1595108695652174, "grad_norm": 2.430508613586426, "learning_rate": 3e-05, "loss": 3.8575, "step": 587 }, { "epoch": 0.15978260869565217, "grad_norm": 2.2944397926330566, "learning_rate": 3e-05, "loss": 4.3022, "step": 588 }, { "epoch": 0.16005434782608696, "grad_norm": 4.096346855163574, "learning_rate": 3e-05, "loss": 4.0663, "step": 589 }, { "epoch": 0.16032608695652173, "grad_norm": 3.8741347789764404, "learning_rate": 3e-05, "loss": 3.3398, "step": 590 }, { "epoch": 0.16059782608695652, "grad_norm": 3.7039456367492676, "learning_rate": 3e-05, "loss": 3.1813, "step": 591 }, { "epoch": 0.1608695652173913, "grad_norm": 2.129674196243286, "learning_rate": 3e-05, "loss": 3.0778, "step": 592 }, { "epoch": 0.1611413043478261, "grad_norm": 2.404423952102661, "learning_rate": 3e-05, "loss": 3.7275, "step": 593 }, { "epoch": 0.16141304347826088, "grad_norm": 2.341529607772827, "learning_rate": 3e-05, "loss": 2.9763, "step": 594 }, { "epoch": 0.16168478260869565, "grad_norm": 4.927234649658203, "learning_rate": 3e-05, "loss": 4.1287, "step": 595 }, { "epoch": 0.16195652173913044, "grad_norm": 4.529046535491943, "learning_rate": 3e-05, "loss": 3.7712, "step": 596 }, { "epoch": 0.1622282608695652, "grad_norm": 4.280128479003906, "learning_rate": 3e-05, "loss": 3.7557, "step": 597 }, { "epoch": 0.1625, "grad_norm": 2.845654249191284, "learning_rate": 3e-05, "loss": 3.4817, "step": 598 }, { "epoch": 0.16277173913043477, "grad_norm": 2.5575075149536133, "learning_rate": 3e-05, "loss": 3.0366, "step": 599 }, { "epoch": 0.16304347826086957, "grad_norm": 3.475111722946167, "learning_rate": 3e-05, "loss": 3.2081, "step": 600 }, { "epoch": 0.16331521739130433, "grad_norm": 3.926103353500366, "learning_rate": 3e-05, "loss": 2.8455, "step": 601 }, { "epoch": 0.16358695652173913, "grad_norm": 3.796124219894409, "learning_rate": 3e-05, "loss": 3.4857, "step": 602 }, { "epoch": 0.16385869565217392, "grad_norm": 2.974625825881958, "learning_rate": 3e-05, "loss": 3.3609, "step": 603 }, { "epoch": 0.1641304347826087, "grad_norm": 1.9188250303268433, "learning_rate": 3e-05, "loss": 3.0467, "step": 604 }, { "epoch": 0.1644021739130435, "grad_norm": 2.922881841659546, "learning_rate": 3e-05, "loss": 3.969, "step": 605 }, { "epoch": 0.16467391304347825, "grad_norm": 2.9243340492248535, "learning_rate": 3e-05, "loss": 3.418, "step": 606 }, { "epoch": 0.16494565217391305, "grad_norm": 2.3607518672943115, "learning_rate": 3e-05, "loss": 3.3568, "step": 607 }, { "epoch": 0.16521739130434782, "grad_norm": 1.4109339714050293, "learning_rate": 3e-05, "loss": 3.1942, "step": 608 }, { "epoch": 0.1654891304347826, "grad_norm": 2.339900016784668, "learning_rate": 3e-05, "loss": 3.6621, "step": 609 }, { "epoch": 0.16576086956521738, "grad_norm": 3.0392491817474365, "learning_rate": 3e-05, "loss": 3.6538, "step": 610 }, { "epoch": 0.16603260869565217, "grad_norm": 2.5297353267669678, "learning_rate": 3e-05, "loss": 3.72, "step": 611 }, { "epoch": 0.16630434782608697, "grad_norm": 2.1570205688476562, "learning_rate": 3e-05, "loss": 3.1765, "step": 612 }, { "epoch": 0.16657608695652174, "grad_norm": 1.5125519037246704, "learning_rate": 3e-05, "loss": 3.5076, "step": 613 }, { "epoch": 0.16684782608695653, "grad_norm": 1.5017513036727905, "learning_rate": 3e-05, "loss": 3.5962, "step": 614 }, { "epoch": 0.1671195652173913, "grad_norm": 2.1280760765075684, "learning_rate": 3e-05, "loss": 3.4612, "step": 615 }, { "epoch": 0.1673913043478261, "grad_norm": 1.6511037349700928, "learning_rate": 3e-05, "loss": 3.3281, "step": 616 }, { "epoch": 0.16766304347826086, "grad_norm": 1.842786431312561, "learning_rate": 3e-05, "loss": 2.876, "step": 617 }, { "epoch": 0.16793478260869565, "grad_norm": 2.008168935775757, "learning_rate": 3e-05, "loss": 3.7421, "step": 618 }, { "epoch": 0.16820652173913042, "grad_norm": 2.965365409851074, "learning_rate": 3e-05, "loss": 3.9864, "step": 619 }, { "epoch": 0.16847826086956522, "grad_norm": 3.619001626968384, "learning_rate": 3e-05, "loss": 4.1181, "step": 620 }, { "epoch": 0.16875, "grad_norm": 3.3966426849365234, "learning_rate": 3e-05, "loss": 3.2355, "step": 621 }, { "epoch": 0.16902173913043478, "grad_norm": 2.4561333656311035, "learning_rate": 3e-05, "loss": 3.0077, "step": 622 }, { "epoch": 0.16929347826086957, "grad_norm": 3.7762365341186523, "learning_rate": 3e-05, "loss": 3.8245, "step": 623 }, { "epoch": 0.16956521739130434, "grad_norm": 4.286952495574951, "learning_rate": 3e-05, "loss": 3.4895, "step": 624 }, { "epoch": 0.16983695652173914, "grad_norm": 3.7213828563690186, "learning_rate": 3e-05, "loss": 2.9799, "step": 625 }, { "epoch": 0.1701086956521739, "grad_norm": 3.109027862548828, "learning_rate": 3e-05, "loss": 3.8767, "step": 626 }, { "epoch": 0.1703804347826087, "grad_norm": 2.4418489933013916, "learning_rate": 3e-05, "loss": 3.7539, "step": 627 }, { "epoch": 0.17065217391304346, "grad_norm": 2.3365840911865234, "learning_rate": 3e-05, "loss": 3.0382, "step": 628 }, { "epoch": 0.17092391304347826, "grad_norm": 2.649116039276123, "learning_rate": 3e-05, "loss": 3.0011, "step": 629 }, { "epoch": 0.17119565217391305, "grad_norm": 4.0966105461120605, "learning_rate": 3e-05, "loss": 3.5889, "step": 630 }, { "epoch": 0.17146739130434782, "grad_norm": 4.70590877532959, "learning_rate": 3e-05, "loss": 3.4292, "step": 631 }, { "epoch": 0.17173913043478262, "grad_norm": 4.246743202209473, "learning_rate": 3e-05, "loss": 3.3927, "step": 632 }, { "epoch": 0.17201086956521738, "grad_norm": 3.4085919857025146, "learning_rate": 3e-05, "loss": 3.5786, "step": 633 }, { "epoch": 0.17228260869565218, "grad_norm": 3.164703607559204, "learning_rate": 3e-05, "loss": 3.4236, "step": 634 }, { "epoch": 0.17255434782608695, "grad_norm": 4.914327144622803, "learning_rate": 3e-05, "loss": 4.2673, "step": 635 }, { "epoch": 0.17282608695652174, "grad_norm": 4.016328811645508, "learning_rate": 3e-05, "loss": 3.5927, "step": 636 }, { "epoch": 0.17309782608695654, "grad_norm": 3.3188321590423584, "learning_rate": 3e-05, "loss": 2.9687, "step": 637 }, { "epoch": 0.1733695652173913, "grad_norm": 3.4671499729156494, "learning_rate": 3e-05, "loss": 3.3144, "step": 638 }, { "epoch": 0.1736413043478261, "grad_norm": 2.273728609085083, "learning_rate": 3e-05, "loss": 3.2453, "step": 639 }, { "epoch": 0.17391304347826086, "grad_norm": 2.1210081577301025, "learning_rate": 3e-05, "loss": 3.0778, "step": 640 }, { "epoch": 0.17418478260869566, "grad_norm": 3.6642115116119385, "learning_rate": 3e-05, "loss": 3.7503, "step": 641 }, { "epoch": 0.17445652173913043, "grad_norm": 3.753002166748047, "learning_rate": 3e-05, "loss": 3.9504, "step": 642 }, { "epoch": 0.17472826086956522, "grad_norm": 2.683023691177368, "learning_rate": 3e-05, "loss": 3.5103, "step": 643 }, { "epoch": 0.175, "grad_norm": 2.7764129638671875, "learning_rate": 3e-05, "loss": 3.8473, "step": 644 }, { "epoch": 0.17527173913043478, "grad_norm": 2.340895652770996, "learning_rate": 3e-05, "loss": 3.3381, "step": 645 }, { "epoch": 0.17554347826086958, "grad_norm": 3.224985122680664, "learning_rate": 3e-05, "loss": 3.7218, "step": 646 }, { "epoch": 0.17581521739130435, "grad_norm": 4.798543930053711, "learning_rate": 3e-05, "loss": 4.2621, "step": 647 }, { "epoch": 0.17608695652173914, "grad_norm": 3.2494759559631348, "learning_rate": 3e-05, "loss": 3.4175, "step": 648 }, { "epoch": 0.1763586956521739, "grad_norm": 1.9543368816375732, "learning_rate": 3e-05, "loss": 2.9133, "step": 649 }, { "epoch": 0.1766304347826087, "grad_norm": 1.7534743547439575, "learning_rate": 3e-05, "loss": 3.2838, "step": 650 }, { "epoch": 0.17690217391304347, "grad_norm": 1.8143278360366821, "learning_rate": 3e-05, "loss": 3.015, "step": 651 }, { "epoch": 0.17717391304347826, "grad_norm": 3.4449610710144043, "learning_rate": 3e-05, "loss": 3.7868, "step": 652 }, { "epoch": 0.17744565217391303, "grad_norm": 2.250290632247925, "learning_rate": 3e-05, "loss": 3.6572, "step": 653 }, { "epoch": 0.17771739130434783, "grad_norm": 2.3306338787078857, "learning_rate": 3e-05, "loss": 3.6463, "step": 654 }, { "epoch": 0.17798913043478262, "grad_norm": 1.6108341217041016, "learning_rate": 3e-05, "loss": 3.3977, "step": 655 }, { "epoch": 0.1782608695652174, "grad_norm": 1.7349133491516113, "learning_rate": 3e-05, "loss": 3.5826, "step": 656 }, { "epoch": 0.17853260869565218, "grad_norm": 2.5024263858795166, "learning_rate": 3e-05, "loss": 4.2271, "step": 657 }, { "epoch": 0.17880434782608695, "grad_norm": 2.2973721027374268, "learning_rate": 3e-05, "loss": 3.8238, "step": 658 }, { "epoch": 0.17907608695652175, "grad_norm": 2.167056083679199, "learning_rate": 3e-05, "loss": 3.5719, "step": 659 }, { "epoch": 0.1793478260869565, "grad_norm": 1.8604989051818848, "learning_rate": 3e-05, "loss": 3.9091, "step": 660 }, { "epoch": 0.1796195652173913, "grad_norm": 1.7601895332336426, "learning_rate": 3e-05, "loss": 3.8582, "step": 661 }, { "epoch": 0.17989130434782608, "grad_norm": 1.5462661981582642, "learning_rate": 3e-05, "loss": 2.9604, "step": 662 }, { "epoch": 0.18016304347826087, "grad_norm": 1.9906418323516846, "learning_rate": 3e-05, "loss": 3.8877, "step": 663 }, { "epoch": 0.18043478260869567, "grad_norm": 2.1721136569976807, "learning_rate": 3e-05, "loss": 4.241, "step": 664 }, { "epoch": 0.18070652173913043, "grad_norm": 1.7870231866836548, "learning_rate": 3e-05, "loss": 3.0484, "step": 665 }, { "epoch": 0.18097826086956523, "grad_norm": 1.543925166130066, "learning_rate": 3e-05, "loss": 3.1115, "step": 666 }, { "epoch": 0.18125, "grad_norm": 1.9159563779830933, "learning_rate": 3e-05, "loss": 3.3634, "step": 667 }, { "epoch": 0.1815217391304348, "grad_norm": 2.8006904125213623, "learning_rate": 3e-05, "loss": 3.7725, "step": 668 }, { "epoch": 0.18179347826086956, "grad_norm": 2.8094773292541504, "learning_rate": 3e-05, "loss": 3.5421, "step": 669 }, { "epoch": 0.18206521739130435, "grad_norm": 2.0162885189056396, "learning_rate": 3e-05, "loss": 3.4805, "step": 670 }, { "epoch": 0.18233695652173912, "grad_norm": 2.175633430480957, "learning_rate": 3e-05, "loss": 3.3937, "step": 671 }, { "epoch": 0.1826086956521739, "grad_norm": 2.4267466068267822, "learning_rate": 3e-05, "loss": 3.1429, "step": 672 }, { "epoch": 0.1828804347826087, "grad_norm": 4.546562194824219, "learning_rate": 3e-05, "loss": 4.1168, "step": 673 }, { "epoch": 0.18315217391304348, "grad_norm": 3.3911025524139404, "learning_rate": 3e-05, "loss": 3.7966, "step": 674 }, { "epoch": 0.18342391304347827, "grad_norm": 1.7894442081451416, "learning_rate": 3e-05, "loss": 3.8492, "step": 675 }, { "epoch": 0.18369565217391304, "grad_norm": 2.7028391361236572, "learning_rate": 3e-05, "loss": 3.5555, "step": 676 }, { "epoch": 0.18396739130434783, "grad_norm": 3.052639961242676, "learning_rate": 3e-05, "loss": 3.9254, "step": 677 }, { "epoch": 0.1842391304347826, "grad_norm": 2.4930381774902344, "learning_rate": 3e-05, "loss": 3.4053, "step": 678 }, { "epoch": 0.1845108695652174, "grad_norm": 1.9123059511184692, "learning_rate": 3e-05, "loss": 3.5101, "step": 679 }, { "epoch": 0.18478260869565216, "grad_norm": 1.5995222330093384, "learning_rate": 3e-05, "loss": 3.1762, "step": 680 }, { "epoch": 0.18505434782608696, "grad_norm": 4.272251129150391, "learning_rate": 3e-05, "loss": 4.397, "step": 681 }, { "epoch": 0.18532608695652175, "grad_norm": 3.0316965579986572, "learning_rate": 3e-05, "loss": 3.8721, "step": 682 }, { "epoch": 0.18559782608695652, "grad_norm": 2.0125246047973633, "learning_rate": 3e-05, "loss": 3.3782, "step": 683 }, { "epoch": 0.1858695652173913, "grad_norm": 1.8079962730407715, "learning_rate": 3e-05, "loss": 3.4851, "step": 684 }, { "epoch": 0.18614130434782608, "grad_norm": 1.8277610540390015, "learning_rate": 3e-05, "loss": 2.8756, "step": 685 }, { "epoch": 0.18641304347826088, "grad_norm": 2.4962170124053955, "learning_rate": 3e-05, "loss": 4.2237, "step": 686 }, { "epoch": 0.18668478260869564, "grad_norm": 1.3268901109695435, "learning_rate": 3e-05, "loss": 3.1177, "step": 687 }, { "epoch": 0.18695652173913044, "grad_norm": 2.3978965282440186, "learning_rate": 3e-05, "loss": 3.583, "step": 688 }, { "epoch": 0.1872282608695652, "grad_norm": 1.2693482637405396, "learning_rate": 3e-05, "loss": 3.3973, "step": 689 }, { "epoch": 0.1875, "grad_norm": 2.0299911499023438, "learning_rate": 3e-05, "loss": 3.6627, "step": 690 }, { "epoch": 0.1877717391304348, "grad_norm": 1.2860682010650635, "learning_rate": 3e-05, "loss": 3.2345, "step": 691 }, { "epoch": 0.18804347826086956, "grad_norm": 2.264256477355957, "learning_rate": 3e-05, "loss": 4.2411, "step": 692 }, { "epoch": 0.18831521739130436, "grad_norm": 1.5267597436904907, "learning_rate": 3e-05, "loss": 3.1942, "step": 693 }, { "epoch": 0.18858695652173912, "grad_norm": 1.2684136629104614, "learning_rate": 3e-05, "loss": 3.192, "step": 694 }, { "epoch": 0.18885869565217392, "grad_norm": 1.6705724000930786, "learning_rate": 3e-05, "loss": 3.6082, "step": 695 }, { "epoch": 0.1891304347826087, "grad_norm": 2.0569725036621094, "learning_rate": 3e-05, "loss": 3.9019, "step": 696 }, { "epoch": 0.18940217391304348, "grad_norm": 2.03778076171875, "learning_rate": 3e-05, "loss": 4.0118, "step": 697 }, { "epoch": 0.18967391304347825, "grad_norm": 1.7906421422958374, "learning_rate": 3e-05, "loss": 3.7659, "step": 698 }, { "epoch": 0.18994565217391304, "grad_norm": 1.6202226877212524, "learning_rate": 3e-05, "loss": 3.3228, "step": 699 }, { "epoch": 0.19021739130434784, "grad_norm": 1.568548560142517, "learning_rate": 3e-05, "loss": 3.3992, "step": 700 }, { "epoch": 0.1904891304347826, "grad_norm": 1.3486888408660889, "learning_rate": 3e-05, "loss": 2.7886, "step": 701 }, { "epoch": 0.1907608695652174, "grad_norm": 2.013223171234131, "learning_rate": 3e-05, "loss": 3.365, "step": 702 }, { "epoch": 0.19103260869565217, "grad_norm": 1.7130358219146729, "learning_rate": 3e-05, "loss": 3.7516, "step": 703 }, { "epoch": 0.19130434782608696, "grad_norm": 1.7324903011322021, "learning_rate": 3e-05, "loss": 3.1275, "step": 704 }, { "epoch": 0.19157608695652173, "grad_norm": 1.7227182388305664, "learning_rate": 3e-05, "loss": 3.1415, "step": 705 }, { "epoch": 0.19184782608695652, "grad_norm": 1.3472079038619995, "learning_rate": 3e-05, "loss": 3.1949, "step": 706 }, { "epoch": 0.1921195652173913, "grad_norm": 1.1573243141174316, "learning_rate": 3e-05, "loss": 2.8314, "step": 707 }, { "epoch": 0.1923913043478261, "grad_norm": 1.6137231588363647, "learning_rate": 3e-05, "loss": 3.6169, "step": 708 }, { "epoch": 0.19266304347826088, "grad_norm": 1.4710439443588257, "learning_rate": 3e-05, "loss": 3.4097, "step": 709 }, { "epoch": 0.19293478260869565, "grad_norm": 2.347607374191284, "learning_rate": 3e-05, "loss": 4.1638, "step": 710 }, { "epoch": 0.19320652173913044, "grad_norm": 1.4299064874649048, "learning_rate": 3e-05, "loss": 3.811, "step": 711 }, { "epoch": 0.1934782608695652, "grad_norm": 1.5702636241912842, "learning_rate": 3e-05, "loss": 3.5396, "step": 712 }, { "epoch": 0.19375, "grad_norm": 1.8417882919311523, "learning_rate": 3e-05, "loss": 3.5613, "step": 713 }, { "epoch": 0.19402173913043477, "grad_norm": 1.4125804901123047, "learning_rate": 3e-05, "loss": 3.1808, "step": 714 }, { "epoch": 0.19429347826086957, "grad_norm": 1.3733165264129639, "learning_rate": 3e-05, "loss": 3.3127, "step": 715 }, { "epoch": 0.19456521739130433, "grad_norm": 2.1651833057403564, "learning_rate": 3e-05, "loss": 4.1732, "step": 716 }, { "epoch": 0.19483695652173913, "grad_norm": 1.4096370935440063, "learning_rate": 3e-05, "loss": 3.5565, "step": 717 }, { "epoch": 0.19510869565217392, "grad_norm": 1.7251945734024048, "learning_rate": 3e-05, "loss": 3.7342, "step": 718 }, { "epoch": 0.1953804347826087, "grad_norm": 1.643470287322998, "learning_rate": 3e-05, "loss": 2.9678, "step": 719 }, { "epoch": 0.1956521739130435, "grad_norm": 1.5827432870864868, "learning_rate": 3e-05, "loss": 3.1044, "step": 720 }, { "epoch": 0.19592391304347825, "grad_norm": 1.4647548198699951, "learning_rate": 3e-05, "loss": 2.8413, "step": 721 }, { "epoch": 0.19619565217391305, "grad_norm": 2.1561598777770996, "learning_rate": 3e-05, "loss": 3.4492, "step": 722 }, { "epoch": 0.19646739130434782, "grad_norm": 1.6970568895339966, "learning_rate": 3e-05, "loss": 3.1848, "step": 723 }, { "epoch": 0.1967391304347826, "grad_norm": 2.248117685317993, "learning_rate": 3e-05, "loss": 3.6296, "step": 724 }, { "epoch": 0.19701086956521738, "grad_norm": 3.168339252471924, "learning_rate": 3e-05, "loss": 4.291, "step": 725 }, { "epoch": 0.19728260869565217, "grad_norm": 3.802591323852539, "learning_rate": 3e-05, "loss": 3.9495, "step": 726 }, { "epoch": 0.19755434782608697, "grad_norm": 3.848099708557129, "learning_rate": 3e-05, "loss": 3.6353, "step": 727 }, { "epoch": 0.19782608695652174, "grad_norm": 3.534794569015503, "learning_rate": 3e-05, "loss": 3.9035, "step": 728 }, { "epoch": 0.19809782608695653, "grad_norm": 1.7165888547897339, "learning_rate": 3e-05, "loss": 3.3266, "step": 729 }, { "epoch": 0.1983695652173913, "grad_norm": 2.5262253284454346, "learning_rate": 3e-05, "loss": 3.6359, "step": 730 }, { "epoch": 0.1986413043478261, "grad_norm": 2.5349607467651367, "learning_rate": 3e-05, "loss": 3.129, "step": 731 }, { "epoch": 0.19891304347826086, "grad_norm": 5.680181503295898, "learning_rate": 3e-05, "loss": 4.8488, "step": 732 }, { "epoch": 0.19918478260869565, "grad_norm": 3.0405654907226562, "learning_rate": 3e-05, "loss": 3.3313, "step": 733 }, { "epoch": 0.19945652173913042, "grad_norm": 1.6890621185302734, "learning_rate": 3e-05, "loss": 3.2893, "step": 734 }, { "epoch": 0.19972826086956522, "grad_norm": 2.204172372817993, "learning_rate": 3e-05, "loss": 3.5067, "step": 735 }, { "epoch": 0.2, "grad_norm": 2.858398914337158, "learning_rate": 3e-05, "loss": 3.4036, "step": 736 }, { "epoch": 0.20027173913043478, "grad_norm": 2.4763529300689697, "learning_rate": 3e-05, "loss": 3.5171, "step": 737 }, { "epoch": 0.20054347826086957, "grad_norm": 2.957118034362793, "learning_rate": 3e-05, "loss": 4.1728, "step": 738 }, { "epoch": 0.20081521739130434, "grad_norm": 1.8309071063995361, "learning_rate": 3e-05, "loss": 2.9663, "step": 739 }, { "epoch": 0.20108695652173914, "grad_norm": 2.270033121109009, "learning_rate": 3e-05, "loss": 3.6115, "step": 740 }, { "epoch": 0.2013586956521739, "grad_norm": 2.1604442596435547, "learning_rate": 3e-05, "loss": 3.2511, "step": 741 }, { "epoch": 0.2016304347826087, "grad_norm": 3.5561349391937256, "learning_rate": 3e-05, "loss": 3.8886, "step": 742 }, { "epoch": 0.20190217391304346, "grad_norm": 2.510963201522827, "learning_rate": 3e-05, "loss": 3.4812, "step": 743 }, { "epoch": 0.20217391304347826, "grad_norm": 1.5206657648086548, "learning_rate": 3e-05, "loss": 3.0231, "step": 744 }, { "epoch": 0.20244565217391305, "grad_norm": 1.645155906677246, "learning_rate": 3e-05, "loss": 3.0951, "step": 745 }, { "epoch": 0.20271739130434782, "grad_norm": 2.075399160385132, "learning_rate": 3e-05, "loss": 3.4057, "step": 746 }, { "epoch": 0.20298913043478262, "grad_norm": 2.36306095123291, "learning_rate": 3e-05, "loss": 3.6449, "step": 747 }, { "epoch": 0.20326086956521738, "grad_norm": 1.9054253101348877, "learning_rate": 3e-05, "loss": 3.3983, "step": 748 }, { "epoch": 0.20353260869565218, "grad_norm": 1.7470569610595703, "learning_rate": 3e-05, "loss": 3.7903, "step": 749 }, { "epoch": 0.20380434782608695, "grad_norm": 1.6297615766525269, "learning_rate": 3e-05, "loss": 3.0304, "step": 750 }, { "epoch": 0.20407608695652174, "grad_norm": 3.2830233573913574, "learning_rate": 3e-05, "loss": 4.2781, "step": 751 }, { "epoch": 0.20434782608695654, "grad_norm": 1.361449122428894, "learning_rate": 3e-05, "loss": 3.2224, "step": 752 }, { "epoch": 0.2046195652173913, "grad_norm": 1.4894417524337769, "learning_rate": 3e-05, "loss": 3.262, "step": 753 }, { "epoch": 0.2048913043478261, "grad_norm": 1.7271361351013184, "learning_rate": 3e-05, "loss": 3.3017, "step": 754 }, { "epoch": 0.20516304347826086, "grad_norm": 2.033215284347534, "learning_rate": 3e-05, "loss": 2.9264, "step": 755 }, { "epoch": 0.20543478260869566, "grad_norm": 2.808903455734253, "learning_rate": 3e-05, "loss": 3.7223, "step": 756 }, { "epoch": 0.20570652173913043, "grad_norm": 1.450661540031433, "learning_rate": 3e-05, "loss": 4.0346, "step": 757 }, { "epoch": 0.20597826086956522, "grad_norm": 1.9796857833862305, "learning_rate": 3e-05, "loss": 3.2978, "step": 758 }, { "epoch": 0.20625, "grad_norm": 2.9408633708953857, "learning_rate": 3e-05, "loss": 3.7247, "step": 759 }, { "epoch": 0.20652173913043478, "grad_norm": 1.4558497667312622, "learning_rate": 3e-05, "loss": 3.0258, "step": 760 }, { "epoch": 0.20679347826086958, "grad_norm": 1.1229232549667358, "learning_rate": 3e-05, "loss": 2.9328, "step": 761 }, { "epoch": 0.20706521739130435, "grad_norm": 1.663891077041626, "learning_rate": 3e-05, "loss": 3.2935, "step": 762 }, { "epoch": 0.20733695652173914, "grad_norm": 1.8489412069320679, "learning_rate": 3e-05, "loss": 2.8728, "step": 763 }, { "epoch": 0.2076086956521739, "grad_norm": 1.359864592552185, "learning_rate": 3e-05, "loss": 3.6552, "step": 764 }, { "epoch": 0.2078804347826087, "grad_norm": 1.5933263301849365, "learning_rate": 3e-05, "loss": 3.2384, "step": 765 }, { "epoch": 0.20815217391304347, "grad_norm": 3.0126466751098633, "learning_rate": 3e-05, "loss": 3.4475, "step": 766 }, { "epoch": 0.20842391304347826, "grad_norm": 2.608242988586426, "learning_rate": 3e-05, "loss": 3.2701, "step": 767 }, { "epoch": 0.20869565217391303, "grad_norm": 2.7247979640960693, "learning_rate": 3e-05, "loss": 3.96, "step": 768 }, { "epoch": 0.20896739130434783, "grad_norm": 1.8706130981445312, "learning_rate": 3e-05, "loss": 3.0889, "step": 769 }, { "epoch": 0.20923913043478262, "grad_norm": 2.668093681335449, "learning_rate": 3e-05, "loss": 3.4351, "step": 770 }, { "epoch": 0.2095108695652174, "grad_norm": 3.017181396484375, "learning_rate": 3e-05, "loss": 3.2463, "step": 771 }, { "epoch": 0.20978260869565218, "grad_norm": 3.0812008380889893, "learning_rate": 3e-05, "loss": 3.3998, "step": 772 }, { "epoch": 0.21005434782608695, "grad_norm": 2.7510030269622803, "learning_rate": 3e-05, "loss": 4.2353, "step": 773 }, { "epoch": 0.21032608695652175, "grad_norm": 2.4355452060699463, "learning_rate": 3e-05, "loss": 3.8853, "step": 774 }, { "epoch": 0.2105978260869565, "grad_norm": 4.041244983673096, "learning_rate": 3e-05, "loss": 4.0771, "step": 775 }, { "epoch": 0.2108695652173913, "grad_norm": 3.403510332107544, "learning_rate": 3e-05, "loss": 3.6963, "step": 776 }, { "epoch": 0.21114130434782608, "grad_norm": 3.216482400894165, "learning_rate": 3e-05, "loss": 3.6608, "step": 777 }, { "epoch": 0.21141304347826087, "grad_norm": 2.5989348888397217, "learning_rate": 3e-05, "loss": 3.689, "step": 778 }, { "epoch": 0.21168478260869567, "grad_norm": 2.0239574909210205, "learning_rate": 3e-05, "loss": 3.681, "step": 779 }, { "epoch": 0.21195652173913043, "grad_norm": 3.3878705501556396, "learning_rate": 3e-05, "loss": 3.9301, "step": 780 }, { "epoch": 0.21222826086956523, "grad_norm": 4.106480598449707, "learning_rate": 3e-05, "loss": 3.2929, "step": 781 }, { "epoch": 0.2125, "grad_norm": 3.4818196296691895, "learning_rate": 3e-05, "loss": 2.8562, "step": 782 }, { "epoch": 0.2127717391304348, "grad_norm": 3.4106242656707764, "learning_rate": 3e-05, "loss": 3.6078, "step": 783 }, { "epoch": 0.21304347826086956, "grad_norm": 2.401315450668335, "learning_rate": 3e-05, "loss": 3.693, "step": 784 }, { "epoch": 0.21331521739130435, "grad_norm": 3.092789649963379, "learning_rate": 3e-05, "loss": 3.6907, "step": 785 }, { "epoch": 0.21358695652173912, "grad_norm": 2.8016254901885986, "learning_rate": 3e-05, "loss": 3.7602, "step": 786 }, { "epoch": 0.2138586956521739, "grad_norm": 2.281668186187744, "learning_rate": 3e-05, "loss": 2.9451, "step": 787 }, { "epoch": 0.2141304347826087, "grad_norm": 2.3392529487609863, "learning_rate": 3e-05, "loss": 3.1969, "step": 788 }, { "epoch": 0.21440217391304348, "grad_norm": 1.7437257766723633, "learning_rate": 3e-05, "loss": 3.2468, "step": 789 }, { "epoch": 0.21467391304347827, "grad_norm": 2.128056049346924, "learning_rate": 3e-05, "loss": 3.5476, "step": 790 }, { "epoch": 0.21494565217391304, "grad_norm": 2.1912710666656494, "learning_rate": 3e-05, "loss": 3.7689, "step": 791 }, { "epoch": 0.21521739130434783, "grad_norm": 2.5414609909057617, "learning_rate": 3e-05, "loss": 3.4014, "step": 792 }, { "epoch": 0.2154891304347826, "grad_norm": 2.6355223655700684, "learning_rate": 3e-05, "loss": 3.3787, "step": 793 }, { "epoch": 0.2157608695652174, "grad_norm": 2.1927592754364014, "learning_rate": 3e-05, "loss": 3.2381, "step": 794 }, { "epoch": 0.21603260869565216, "grad_norm": 1.7231847047805786, "learning_rate": 3e-05, "loss": 3.4627, "step": 795 }, { "epoch": 0.21630434782608696, "grad_norm": 1.8754600286483765, "learning_rate": 3e-05, "loss": 3.7409, "step": 796 }, { "epoch": 0.21657608695652175, "grad_norm": 1.7413159608840942, "learning_rate": 3e-05, "loss": 3.6012, "step": 797 }, { "epoch": 0.21684782608695652, "grad_norm": 1.9197379350662231, "learning_rate": 3e-05, "loss": 3.15, "step": 798 }, { "epoch": 0.2171195652173913, "grad_norm": 2.100706100463867, "learning_rate": 3e-05, "loss": 2.8748, "step": 799 }, { "epoch": 0.21739130434782608, "grad_norm": 2.7185757160186768, "learning_rate": 3e-05, "loss": 3.7859, "step": 800 }, { "epoch": 0.21766304347826088, "grad_norm": 3.241926908493042, "learning_rate": 3e-05, "loss": 3.7198, "step": 801 }, { "epoch": 0.21793478260869564, "grad_norm": 2.365151882171631, "learning_rate": 3e-05, "loss": 3.7607, "step": 802 }, { "epoch": 0.21820652173913044, "grad_norm": 2.0234482288360596, "learning_rate": 3e-05, "loss": 3.1843, "step": 803 }, { "epoch": 0.2184782608695652, "grad_norm": 3.084533929824829, "learning_rate": 3e-05, "loss": 4.1945, "step": 804 }, { "epoch": 0.21875, "grad_norm": 2.927222490310669, "learning_rate": 3e-05, "loss": 3.3531, "step": 805 }, { "epoch": 0.2190217391304348, "grad_norm": 2.6623260974884033, "learning_rate": 3e-05, "loss": 3.5994, "step": 806 }, { "epoch": 0.21929347826086956, "grad_norm": 2.8344593048095703, "learning_rate": 3e-05, "loss": 3.6001, "step": 807 }, { "epoch": 0.21956521739130436, "grad_norm": 1.8181391954421997, "learning_rate": 3e-05, "loss": 2.9434, "step": 808 }, { "epoch": 0.21983695652173912, "grad_norm": 2.103512763977051, "learning_rate": 3e-05, "loss": 3.8034, "step": 809 }, { "epoch": 0.22010869565217392, "grad_norm": 1.961503505706787, "learning_rate": 3e-05, "loss": 3.0883, "step": 810 }, { "epoch": 0.2203804347826087, "grad_norm": 2.3394432067871094, "learning_rate": 3e-05, "loss": 2.8568, "step": 811 }, { "epoch": 0.22065217391304348, "grad_norm": 2.6501801013946533, "learning_rate": 3e-05, "loss": 3.2469, "step": 812 }, { "epoch": 0.22092391304347825, "grad_norm": 2.7663819789886475, "learning_rate": 3e-05, "loss": 3.3066, "step": 813 }, { "epoch": 0.22119565217391304, "grad_norm": 2.4567878246307373, "learning_rate": 3e-05, "loss": 3.0416, "step": 814 }, { "epoch": 0.22146739130434784, "grad_norm": 2.378188371658325, "learning_rate": 3e-05, "loss": 3.6954, "step": 815 }, { "epoch": 0.2217391304347826, "grad_norm": 2.611842632293701, "learning_rate": 3e-05, "loss": 3.3243, "step": 816 }, { "epoch": 0.2220108695652174, "grad_norm": 2.1804463863372803, "learning_rate": 3e-05, "loss": 2.9858, "step": 817 }, { "epoch": 0.22228260869565217, "grad_norm": 2.03049373626709, "learning_rate": 3e-05, "loss": 2.787, "step": 818 }, { "epoch": 0.22255434782608696, "grad_norm": 1.8013122081756592, "learning_rate": 3e-05, "loss": 2.9299, "step": 819 }, { "epoch": 0.22282608695652173, "grad_norm": 2.039745569229126, "learning_rate": 3e-05, "loss": 3.2247, "step": 820 }, { "epoch": 0.22309782608695652, "grad_norm": 1.6249157190322876, "learning_rate": 3e-05, "loss": 2.8942, "step": 821 }, { "epoch": 0.2233695652173913, "grad_norm": 1.8277668952941895, "learning_rate": 3e-05, "loss": 3.5634, "step": 822 }, { "epoch": 0.2236413043478261, "grad_norm": 2.2749369144439697, "learning_rate": 3e-05, "loss": 3.6253, "step": 823 }, { "epoch": 0.22391304347826088, "grad_norm": 2.3847649097442627, "learning_rate": 3e-05, "loss": 3.3725, "step": 824 }, { "epoch": 0.22418478260869565, "grad_norm": 1.9465465545654297, "learning_rate": 3e-05, "loss": 3.4556, "step": 825 }, { "epoch": 0.22445652173913044, "grad_norm": 2.793731927871704, "learning_rate": 3e-05, "loss": 4.0801, "step": 826 }, { "epoch": 0.2247282608695652, "grad_norm": 2.605842113494873, "learning_rate": 3e-05, "loss": 4.0396, "step": 827 }, { "epoch": 0.225, "grad_norm": 2.200025796890259, "learning_rate": 3e-05, "loss": 3.7514, "step": 828 }, { "epoch": 0.22527173913043477, "grad_norm": 1.8318731784820557, "learning_rate": 3e-05, "loss": 3.3619, "step": 829 }, { "epoch": 0.22554347826086957, "grad_norm": 1.7559170722961426, "learning_rate": 3e-05, "loss": 3.3115, "step": 830 }, { "epoch": 0.22581521739130433, "grad_norm": 1.7032020092010498, "learning_rate": 3e-05, "loss": 3.113, "step": 831 }, { "epoch": 0.22608695652173913, "grad_norm": 1.8972278833389282, "learning_rate": 3e-05, "loss": 3.1965, "step": 832 }, { "epoch": 0.22635869565217392, "grad_norm": 1.9449539184570312, "learning_rate": 3e-05, "loss": 3.2422, "step": 833 }, { "epoch": 0.2266304347826087, "grad_norm": 1.6739164590835571, "learning_rate": 3e-05, "loss": 3.2255, "step": 834 }, { "epoch": 0.2269021739130435, "grad_norm": 1.7071807384490967, "learning_rate": 3e-05, "loss": 3.4032, "step": 835 }, { "epoch": 0.22717391304347825, "grad_norm": 3.5912764072418213, "learning_rate": 3e-05, "loss": 4.1757, "step": 836 }, { "epoch": 0.22744565217391305, "grad_norm": 1.969997763633728, "learning_rate": 3e-05, "loss": 3.1138, "step": 837 }, { "epoch": 0.22771739130434782, "grad_norm": 1.9781277179718018, "learning_rate": 3e-05, "loss": 3.5066, "step": 838 }, { "epoch": 0.2279891304347826, "grad_norm": 1.6115202903747559, "learning_rate": 3e-05, "loss": 2.9759, "step": 839 }, { "epoch": 0.22826086956521738, "grad_norm": 1.3473846912384033, "learning_rate": 3e-05, "loss": 3.0617, "step": 840 }, { "epoch": 0.22853260869565217, "grad_norm": 1.6643739938735962, "learning_rate": 3e-05, "loss": 3.4268, "step": 841 }, { "epoch": 0.22880434782608697, "grad_norm": 1.9249311685562134, "learning_rate": 3e-05, "loss": 3.6207, "step": 842 }, { "epoch": 0.22907608695652174, "grad_norm": 2.079634666442871, "learning_rate": 3e-05, "loss": 3.534, "step": 843 }, { "epoch": 0.22934782608695653, "grad_norm": 1.8367491960525513, "learning_rate": 3e-05, "loss": 3.9664, "step": 844 }, { "epoch": 0.2296195652173913, "grad_norm": 1.7043336629867554, "learning_rate": 3e-05, "loss": 3.4275, "step": 845 }, { "epoch": 0.2298913043478261, "grad_norm": 1.6460905075073242, "learning_rate": 3e-05, "loss": 3.1771, "step": 846 }, { "epoch": 0.23016304347826086, "grad_norm": 2.0371439456939697, "learning_rate": 3e-05, "loss": 3.3044, "step": 847 }, { "epoch": 0.23043478260869565, "grad_norm": 1.5710325241088867, "learning_rate": 3e-05, "loss": 3.0044, "step": 848 }, { "epoch": 0.23070652173913042, "grad_norm": 1.6866766214370728, "learning_rate": 3e-05, "loss": 2.9916, "step": 849 }, { "epoch": 0.23097826086956522, "grad_norm": 2.127194404602051, "learning_rate": 3e-05, "loss": 3.5753, "step": 850 }, { "epoch": 0.23125, "grad_norm": 2.0354788303375244, "learning_rate": 3e-05, "loss": 2.8992, "step": 851 }, { "epoch": 0.23152173913043478, "grad_norm": 1.5567448139190674, "learning_rate": 3e-05, "loss": 3.4486, "step": 852 }, { "epoch": 0.23179347826086957, "grad_norm": 1.3115332126617432, "learning_rate": 3e-05, "loss": 3.1354, "step": 853 }, { "epoch": 0.23206521739130434, "grad_norm": 1.3727189302444458, "learning_rate": 3e-05, "loss": 3.0521, "step": 854 }, { "epoch": 0.23233695652173914, "grad_norm": 1.3423751592636108, "learning_rate": 3e-05, "loss": 2.908, "step": 855 }, { "epoch": 0.2326086956521739, "grad_norm": 1.61993408203125, "learning_rate": 3e-05, "loss": 3.2873, "step": 856 }, { "epoch": 0.2328804347826087, "grad_norm": 1.8327138423919678, "learning_rate": 3e-05, "loss": 3.6789, "step": 857 }, { "epoch": 0.23315217391304346, "grad_norm": 1.626006007194519, "learning_rate": 3e-05, "loss": 3.0027, "step": 858 }, { "epoch": 0.23342391304347826, "grad_norm": 2.236844539642334, "learning_rate": 3e-05, "loss": 4.083, "step": 859 }, { "epoch": 0.23369565217391305, "grad_norm": 2.1321239471435547, "learning_rate": 3e-05, "loss": 3.455, "step": 860 }, { "epoch": 0.23396739130434782, "grad_norm": 2.1679024696350098, "learning_rate": 3e-05, "loss": 3.8644, "step": 861 }, { "epoch": 0.23423913043478262, "grad_norm": 1.9821666479110718, "learning_rate": 3e-05, "loss": 3.5024, "step": 862 }, { "epoch": 0.23451086956521738, "grad_norm": 2.8352460861206055, "learning_rate": 3e-05, "loss": 4.0082, "step": 863 }, { "epoch": 0.23478260869565218, "grad_norm": 2.2241246700286865, "learning_rate": 3e-05, "loss": 3.8031, "step": 864 }, { "epoch": 0.23505434782608695, "grad_norm": 2.1706037521362305, "learning_rate": 3e-05, "loss": 4.0421, "step": 865 }, { "epoch": 0.23532608695652174, "grad_norm": 1.695703148841858, "learning_rate": 3e-05, "loss": 2.9672, "step": 866 }, { "epoch": 0.23559782608695654, "grad_norm": 1.994674563407898, "learning_rate": 3e-05, "loss": 3.3449, "step": 867 }, { "epoch": 0.2358695652173913, "grad_norm": 2.692945957183838, "learning_rate": 3e-05, "loss": 4.1761, "step": 868 }, { "epoch": 0.2361413043478261, "grad_norm": 2.5395121574401855, "learning_rate": 3e-05, "loss": 3.921, "step": 869 }, { "epoch": 0.23641304347826086, "grad_norm": 1.991248607635498, "learning_rate": 3e-05, "loss": 2.9, "step": 870 }, { "epoch": 0.23668478260869566, "grad_norm": 2.2419328689575195, "learning_rate": 3e-05, "loss": 3.501, "step": 871 }, { "epoch": 0.23695652173913043, "grad_norm": 1.9551500082015991, "learning_rate": 3e-05, "loss": 3.8966, "step": 872 }, { "epoch": 0.23722826086956522, "grad_norm": 1.9755557775497437, "learning_rate": 3e-05, "loss": 3.7852, "step": 873 }, { "epoch": 0.2375, "grad_norm": 1.964316487312317, "learning_rate": 3e-05, "loss": 3.4841, "step": 874 }, { "epoch": 0.23777173913043478, "grad_norm": 1.667006015777588, "learning_rate": 3e-05, "loss": 3.3479, "step": 875 }, { "epoch": 0.23804347826086958, "grad_norm": 1.3252580165863037, "learning_rate": 3e-05, "loss": 3.1133, "step": 876 }, { "epoch": 0.23831521739130435, "grad_norm": 1.904585361480713, "learning_rate": 3e-05, "loss": 3.7604, "step": 877 }, { "epoch": 0.23858695652173914, "grad_norm": 1.581666350364685, "learning_rate": 3e-05, "loss": 3.3015, "step": 878 }, { "epoch": 0.2388586956521739, "grad_norm": 1.2517045736312866, "learning_rate": 3e-05, "loss": 3.2687, "step": 879 }, { "epoch": 0.2391304347826087, "grad_norm": 1.5604926347732544, "learning_rate": 3e-05, "loss": 3.0638, "step": 880 }, { "epoch": 0.23940217391304347, "grad_norm": 1.9475339651107788, "learning_rate": 3e-05, "loss": 3.8968, "step": 881 }, { "epoch": 0.23967391304347826, "grad_norm": 1.528891921043396, "learning_rate": 3e-05, "loss": 3.0379, "step": 882 }, { "epoch": 0.23994565217391303, "grad_norm": 1.838294506072998, "learning_rate": 3e-05, "loss": 3.7645, "step": 883 }, { "epoch": 0.24021739130434783, "grad_norm": 1.5220407247543335, "learning_rate": 3e-05, "loss": 3.9417, "step": 884 }, { "epoch": 0.24048913043478262, "grad_norm": 1.4615991115570068, "learning_rate": 3e-05, "loss": 3.4286, "step": 885 }, { "epoch": 0.2407608695652174, "grad_norm": 2.569244623184204, "learning_rate": 3e-05, "loss": 3.9001, "step": 886 }, { "epoch": 0.24103260869565218, "grad_norm": 2.3717269897460938, "learning_rate": 3e-05, "loss": 4.2076, "step": 887 }, { "epoch": 0.24130434782608695, "grad_norm": 1.8509140014648438, "learning_rate": 3e-05, "loss": 3.2101, "step": 888 }, { "epoch": 0.24157608695652175, "grad_norm": 2.0046916007995605, "learning_rate": 3e-05, "loss": 2.9609, "step": 889 }, { "epoch": 0.2418478260869565, "grad_norm": 2.2359912395477295, "learning_rate": 3e-05, "loss": 3.6416, "step": 890 }, { "epoch": 0.2421195652173913, "grad_norm": 1.8625085353851318, "learning_rate": 3e-05, "loss": 3.3007, "step": 891 }, { "epoch": 0.24239130434782608, "grad_norm": 2.0958662033081055, "learning_rate": 3e-05, "loss": 2.7472, "step": 892 }, { "epoch": 0.24266304347826087, "grad_norm": 2.5501413345336914, "learning_rate": 3e-05, "loss": 3.0367, "step": 893 }, { "epoch": 0.24293478260869567, "grad_norm": 3.020587921142578, "learning_rate": 3e-05, "loss": 3.7357, "step": 894 }, { "epoch": 0.24320652173913043, "grad_norm": 2.0165812969207764, "learning_rate": 3e-05, "loss": 3.6235, "step": 895 }, { "epoch": 0.24347826086956523, "grad_norm": 3.1172547340393066, "learning_rate": 3e-05, "loss": 4.1981, "step": 896 }, { "epoch": 0.24375, "grad_norm": 2.381237268447876, "learning_rate": 3e-05, "loss": 3.104, "step": 897 }, { "epoch": 0.2440217391304348, "grad_norm": 3.345118522644043, "learning_rate": 3e-05, "loss": 4.1412, "step": 898 }, { "epoch": 0.24429347826086956, "grad_norm": 2.8453972339630127, "learning_rate": 3e-05, "loss": 3.3134, "step": 899 }, { "epoch": 0.24456521739130435, "grad_norm": 1.9727962017059326, "learning_rate": 3e-05, "loss": 3.1409, "step": 900 }, { "epoch": 0.24483695652173912, "grad_norm": 1.3924411535263062, "learning_rate": 3e-05, "loss": 3.2174, "step": 901 }, { "epoch": 0.2451086956521739, "grad_norm": 1.8885339498519897, "learning_rate": 3e-05, "loss": 3.4122, "step": 902 }, { "epoch": 0.2453804347826087, "grad_norm": 2.3274450302124023, "learning_rate": 3e-05, "loss": 3.2905, "step": 903 }, { "epoch": 0.24565217391304348, "grad_norm": 1.877142310142517, "learning_rate": 3e-05, "loss": 3.4118, "step": 904 }, { "epoch": 0.24592391304347827, "grad_norm": 1.8108350038528442, "learning_rate": 3e-05, "loss": 3.206, "step": 905 }, { "epoch": 0.24619565217391304, "grad_norm": 1.6787278652191162, "learning_rate": 3e-05, "loss": 3.2236, "step": 906 }, { "epoch": 0.24646739130434783, "grad_norm": 2.007154703140259, "learning_rate": 3e-05, "loss": 4.1134, "step": 907 }, { "epoch": 0.2467391304347826, "grad_norm": 1.9029121398925781, "learning_rate": 3e-05, "loss": 3.2066, "step": 908 }, { "epoch": 0.2470108695652174, "grad_norm": 1.5358937978744507, "learning_rate": 3e-05, "loss": 3.3406, "step": 909 }, { "epoch": 0.24728260869565216, "grad_norm": 1.706233263015747, "learning_rate": 3e-05, "loss": 3.368, "step": 910 }, { "epoch": 0.24755434782608696, "grad_norm": 2.283039093017578, "learning_rate": 3e-05, "loss": 4.1815, "step": 911 }, { "epoch": 0.24782608695652175, "grad_norm": 1.6811976432800293, "learning_rate": 3e-05, "loss": 3.8339, "step": 912 }, { "epoch": 0.24809782608695652, "grad_norm": 1.6838414669036865, "learning_rate": 3e-05, "loss": 3.6558, "step": 913 }, { "epoch": 0.2483695652173913, "grad_norm": 1.451900839805603, "learning_rate": 3e-05, "loss": 3.2279, "step": 914 }, { "epoch": 0.24864130434782608, "grad_norm": 2.3503222465515137, "learning_rate": 3e-05, "loss": 3.9484, "step": 915 }, { "epoch": 0.24891304347826088, "grad_norm": 2.001974582672119, "learning_rate": 3e-05, "loss": 3.3028, "step": 916 }, { "epoch": 0.24918478260869564, "grad_norm": 1.5036247968673706, "learning_rate": 3e-05, "loss": 3.2459, "step": 917 }, { "epoch": 0.24945652173913044, "grad_norm": 2.033236026763916, "learning_rate": 3e-05, "loss": 3.4323, "step": 918 }, { "epoch": 0.2497282608695652, "grad_norm": 1.4043093919754028, "learning_rate": 3e-05, "loss": 3.0965, "step": 919 }, { "epoch": 0.25, "grad_norm": 1.203482985496521, "learning_rate": 3e-05, "loss": 3.1118, "step": 920 }, { "epoch": 0.25027173913043477, "grad_norm": 1.4571930170059204, "learning_rate": 3e-05, "loss": 3.351, "step": 921 }, { "epoch": 0.2505434782608696, "grad_norm": 0.9827572703361511, "learning_rate": 3e-05, "loss": 2.8122, "step": 922 }, { "epoch": 0.25081521739130436, "grad_norm": 1.520359754562378, "learning_rate": 3e-05, "loss": 3.6678, "step": 923 }, { "epoch": 0.2510869565217391, "grad_norm": 1.277106761932373, "learning_rate": 3e-05, "loss": 2.9916, "step": 924 }, { "epoch": 0.2513586956521739, "grad_norm": 1.7534111738204956, "learning_rate": 3e-05, "loss": 3.9959, "step": 925 }, { "epoch": 0.2516304347826087, "grad_norm": 1.031008005142212, "learning_rate": 3e-05, "loss": 2.9768, "step": 926 }, { "epoch": 0.2519021739130435, "grad_norm": 1.1608198881149292, "learning_rate": 3e-05, "loss": 3.456, "step": 927 }, { "epoch": 0.25217391304347825, "grad_norm": 1.4016661643981934, "learning_rate": 3e-05, "loss": 3.4423, "step": 928 }, { "epoch": 0.25244565217391307, "grad_norm": 1.1200820207595825, "learning_rate": 3e-05, "loss": 2.897, "step": 929 }, { "epoch": 0.25271739130434784, "grad_norm": 1.4193557500839233, "learning_rate": 3e-05, "loss": 3.7882, "step": 930 }, { "epoch": 0.2529891304347826, "grad_norm": 1.2631326913833618, "learning_rate": 3e-05, "loss": 3.8642, "step": 931 }, { "epoch": 0.2532608695652174, "grad_norm": 1.6645845174789429, "learning_rate": 3e-05, "loss": 3.7794, "step": 932 }, { "epoch": 0.2535326086956522, "grad_norm": 1.7016172409057617, "learning_rate": 3e-05, "loss": 3.3343, "step": 933 }, { "epoch": 0.25380434782608696, "grad_norm": 1.1917381286621094, "learning_rate": 3e-05, "loss": 3.3563, "step": 934 }, { "epoch": 0.25407608695652173, "grad_norm": 1.4421312808990479, "learning_rate": 3e-05, "loss": 3.6694, "step": 935 }, { "epoch": 0.2543478260869565, "grad_norm": 1.2393507957458496, "learning_rate": 3e-05, "loss": 2.8572, "step": 936 }, { "epoch": 0.2546195652173913, "grad_norm": 1.4665427207946777, "learning_rate": 3e-05, "loss": 3.5944, "step": 937 }, { "epoch": 0.2548913043478261, "grad_norm": 1.3151808977127075, "learning_rate": 3e-05, "loss": 3.1136, "step": 938 }, { "epoch": 0.25516304347826085, "grad_norm": 1.3733018636703491, "learning_rate": 3e-05, "loss": 3.1559, "step": 939 }, { "epoch": 0.2554347826086957, "grad_norm": 1.638494610786438, "learning_rate": 3e-05, "loss": 3.5616, "step": 940 }, { "epoch": 0.25570652173913044, "grad_norm": 1.5186861753463745, "learning_rate": 3e-05, "loss": 3.7529, "step": 941 }, { "epoch": 0.2559782608695652, "grad_norm": 1.7179397344589233, "learning_rate": 3e-05, "loss": 3.8634, "step": 942 }, { "epoch": 0.25625, "grad_norm": 1.9366356134414673, "learning_rate": 3e-05, "loss": 3.785, "step": 943 }, { "epoch": 0.2565217391304348, "grad_norm": 2.2738037109375, "learning_rate": 3e-05, "loss": 3.7442, "step": 944 }, { "epoch": 0.25679347826086957, "grad_norm": 2.67051100730896, "learning_rate": 3e-05, "loss": 3.8186, "step": 945 }, { "epoch": 0.25706521739130433, "grad_norm": 3.3741190433502197, "learning_rate": 3e-05, "loss": 4.2118, "step": 946 }, { "epoch": 0.25733695652173916, "grad_norm": 1.6262043714523315, "learning_rate": 3e-05, "loss": 3.0208, "step": 947 }, { "epoch": 0.2576086956521739, "grad_norm": 1.7342244386672974, "learning_rate": 3e-05, "loss": 3.1897, "step": 948 }, { "epoch": 0.2578804347826087, "grad_norm": 2.0924177169799805, "learning_rate": 3e-05, "loss": 2.7324, "step": 949 }, { "epoch": 0.25815217391304346, "grad_norm": 2.21530818939209, "learning_rate": 3e-05, "loss": 3.1098, "step": 950 }, { "epoch": 0.2584239130434783, "grad_norm": 1.642945647239685, "learning_rate": 3e-05, "loss": 3.58, "step": 951 }, { "epoch": 0.25869565217391305, "grad_norm": 1.4641486406326294, "learning_rate": 3e-05, "loss": 3.4771, "step": 952 }, { "epoch": 0.2589673913043478, "grad_norm": 1.4831671714782715, "learning_rate": 3e-05, "loss": 3.0967, "step": 953 }, { "epoch": 0.2592391304347826, "grad_norm": 1.605242133140564, "learning_rate": 3e-05, "loss": 3.3319, "step": 954 }, { "epoch": 0.2595108695652174, "grad_norm": 2.312819480895996, "learning_rate": 3e-05, "loss": 4.214, "step": 955 }, { "epoch": 0.2597826086956522, "grad_norm": 1.751943588256836, "learning_rate": 3e-05, "loss": 3.3116, "step": 956 }, { "epoch": 0.26005434782608694, "grad_norm": 1.951372504234314, "learning_rate": 3e-05, "loss": 3.4478, "step": 957 }, { "epoch": 0.26032608695652176, "grad_norm": 2.3981149196624756, "learning_rate": 3e-05, "loss": 3.4107, "step": 958 }, { "epoch": 0.26059782608695653, "grad_norm": 2.448025703430176, "learning_rate": 3e-05, "loss": 4.0686, "step": 959 }, { "epoch": 0.2608695652173913, "grad_norm": 2.1298911571502686, "learning_rate": 3e-05, "loss": 3.4676, "step": 960 }, { "epoch": 0.26114130434782606, "grad_norm": 1.9788066148757935, "learning_rate": 3e-05, "loss": 3.4239, "step": 961 }, { "epoch": 0.2614130434782609, "grad_norm": 1.6203289031982422, "learning_rate": 3e-05, "loss": 3.2623, "step": 962 }, { "epoch": 0.26168478260869565, "grad_norm": 1.7709441184997559, "learning_rate": 3e-05, "loss": 3.5855, "step": 963 }, { "epoch": 0.2619565217391304, "grad_norm": 1.4671415090560913, "learning_rate": 3e-05, "loss": 3.2081, "step": 964 }, { "epoch": 0.26222826086956524, "grad_norm": 1.2123912572860718, "learning_rate": 3e-05, "loss": 3.0287, "step": 965 }, { "epoch": 0.2625, "grad_norm": 1.9638895988464355, "learning_rate": 3e-05, "loss": 3.4101, "step": 966 }, { "epoch": 0.2627717391304348, "grad_norm": 1.6470272541046143, "learning_rate": 3e-05, "loss": 3.1761, "step": 967 }, { "epoch": 0.26304347826086955, "grad_norm": 1.7481547594070435, "learning_rate": 3e-05, "loss": 3.7102, "step": 968 }, { "epoch": 0.26331521739130437, "grad_norm": 1.589220404624939, "learning_rate": 3e-05, "loss": 2.9376, "step": 969 }, { "epoch": 0.26358695652173914, "grad_norm": 2.317216634750366, "learning_rate": 3e-05, "loss": 3.888, "step": 970 }, { "epoch": 0.2638586956521739, "grad_norm": 1.9831353425979614, "learning_rate": 3e-05, "loss": 3.831, "step": 971 }, { "epoch": 0.26413043478260867, "grad_norm": 2.201885223388672, "learning_rate": 3e-05, "loss": 3.5633, "step": 972 }, { "epoch": 0.2644021739130435, "grad_norm": 1.755304217338562, "learning_rate": 3e-05, "loss": 3.9027, "step": 973 }, { "epoch": 0.26467391304347826, "grad_norm": 1.7514090538024902, "learning_rate": 3e-05, "loss": 3.3256, "step": 974 }, { "epoch": 0.264945652173913, "grad_norm": 2.1359901428222656, "learning_rate": 3e-05, "loss": 3.9904, "step": 975 }, { "epoch": 0.26521739130434785, "grad_norm": 1.5768872499465942, "learning_rate": 3e-05, "loss": 3.3242, "step": 976 }, { "epoch": 0.2654891304347826, "grad_norm": 1.8486355543136597, "learning_rate": 3e-05, "loss": 3.7133, "step": 977 }, { "epoch": 0.2657608695652174, "grad_norm": 1.5333871841430664, "learning_rate": 3e-05, "loss": 3.4191, "step": 978 }, { "epoch": 0.26603260869565215, "grad_norm": 1.6924866437911987, "learning_rate": 3e-05, "loss": 3.3763, "step": 979 }, { "epoch": 0.266304347826087, "grad_norm": 2.1424450874328613, "learning_rate": 3e-05, "loss": 3.8962, "step": 980 }, { "epoch": 0.26657608695652174, "grad_norm": 1.5438109636306763, "learning_rate": 3e-05, "loss": 3.2176, "step": 981 }, { "epoch": 0.2668478260869565, "grad_norm": 1.5958982706069946, "learning_rate": 3e-05, "loss": 3.8708, "step": 982 }, { "epoch": 0.26711956521739133, "grad_norm": 1.129310965538025, "learning_rate": 3e-05, "loss": 2.8674, "step": 983 }, { "epoch": 0.2673913043478261, "grad_norm": 1.6280345916748047, "learning_rate": 3e-05, "loss": 4.165, "step": 984 }, { "epoch": 0.26766304347826086, "grad_norm": 1.499475121498108, "learning_rate": 3e-05, "loss": 2.9953, "step": 985 }, { "epoch": 0.26793478260869563, "grad_norm": 1.253732442855835, "learning_rate": 3e-05, "loss": 3.4718, "step": 986 }, { "epoch": 0.26820652173913045, "grad_norm": 1.113315463066101, "learning_rate": 3e-05, "loss": 3.3101, "step": 987 }, { "epoch": 0.2684782608695652, "grad_norm": 1.4947681427001953, "learning_rate": 3e-05, "loss": 3.217, "step": 988 }, { "epoch": 0.26875, "grad_norm": 1.9959630966186523, "learning_rate": 3e-05, "loss": 4.0953, "step": 989 }, { "epoch": 0.26902173913043476, "grad_norm": 1.6297112703323364, "learning_rate": 3e-05, "loss": 3.4108, "step": 990 }, { "epoch": 0.2692934782608696, "grad_norm": 1.9028682708740234, "learning_rate": 3e-05, "loss": 4.1301, "step": 991 }, { "epoch": 0.26956521739130435, "grad_norm": 1.6655460596084595, "learning_rate": 3e-05, "loss": 3.5511, "step": 992 }, { "epoch": 0.2698369565217391, "grad_norm": 1.8980700969696045, "learning_rate": 3e-05, "loss": 3.5143, "step": 993 }, { "epoch": 0.27010869565217394, "grad_norm": 1.5208625793457031, "learning_rate": 3e-05, "loss": 3.5514, "step": 994 }, { "epoch": 0.2703804347826087, "grad_norm": 1.1632534265518188, "learning_rate": 3e-05, "loss": 3.2167, "step": 995 }, { "epoch": 0.27065217391304347, "grad_norm": 2.094667673110962, "learning_rate": 3e-05, "loss": 3.0362, "step": 996 }, { "epoch": 0.27092391304347824, "grad_norm": 1.4889837503433228, "learning_rate": 3e-05, "loss": 3.4279, "step": 997 }, { "epoch": 0.27119565217391306, "grad_norm": 1.9965155124664307, "learning_rate": 3e-05, "loss": 3.774, "step": 998 }, { "epoch": 0.2714673913043478, "grad_norm": 1.5410970449447632, "learning_rate": 3e-05, "loss": 3.2647, "step": 999 }, { "epoch": 0.2717391304347826, "grad_norm": 1.9614113569259644, "learning_rate": 3e-05, "loss": 3.5648, "step": 1000 }, { "epoch": 0.2720108695652174, "grad_norm": 1.3656377792358398, "learning_rate": 3e-05, "loss": 3.3263, "step": 1001 }, { "epoch": 0.2722826086956522, "grad_norm": 2.057823419570923, "learning_rate": 3e-05, "loss": 3.1952, "step": 1002 }, { "epoch": 0.27255434782608695, "grad_norm": 1.9812930822372437, "learning_rate": 3e-05, "loss": 3.5128, "step": 1003 }, { "epoch": 0.2728260869565217, "grad_norm": 1.443973183631897, "learning_rate": 3e-05, "loss": 3.5108, "step": 1004 }, { "epoch": 0.27309782608695654, "grad_norm": 1.1789878606796265, "learning_rate": 3e-05, "loss": 3.142, "step": 1005 }, { "epoch": 0.2733695652173913, "grad_norm": 1.5677335262298584, "learning_rate": 3e-05, "loss": 3.0946, "step": 1006 }, { "epoch": 0.2736413043478261, "grad_norm": 1.6985656023025513, "learning_rate": 3e-05, "loss": 3.8017, "step": 1007 }, { "epoch": 0.27391304347826084, "grad_norm": 0.9420762658119202, "learning_rate": 3e-05, "loss": 2.8754, "step": 1008 }, { "epoch": 0.27418478260869567, "grad_norm": 1.1210711002349854, "learning_rate": 3e-05, "loss": 3.1734, "step": 1009 }, { "epoch": 0.27445652173913043, "grad_norm": 0.9131266474723816, "learning_rate": 3e-05, "loss": 2.5223, "step": 1010 }, { "epoch": 0.2747282608695652, "grad_norm": 1.4077900648117065, "learning_rate": 3e-05, "loss": 3.2054, "step": 1011 }, { "epoch": 0.275, "grad_norm": 1.4042190313339233, "learning_rate": 3e-05, "loss": 3.3916, "step": 1012 }, { "epoch": 0.2752717391304348, "grad_norm": 1.1815489530563354, "learning_rate": 3e-05, "loss": 3.2108, "step": 1013 }, { "epoch": 0.27554347826086956, "grad_norm": 1.007114052772522, "learning_rate": 3e-05, "loss": 2.8035, "step": 1014 }, { "epoch": 0.2758152173913043, "grad_norm": 1.6630133390426636, "learning_rate": 3e-05, "loss": 3.5364, "step": 1015 }, { "epoch": 0.27608695652173915, "grad_norm": 1.3151155710220337, "learning_rate": 3e-05, "loss": 3.8982, "step": 1016 }, { "epoch": 0.2763586956521739, "grad_norm": 1.3191099166870117, "learning_rate": 3e-05, "loss": 3.3812, "step": 1017 }, { "epoch": 0.2766304347826087, "grad_norm": 0.9553658366203308, "learning_rate": 3e-05, "loss": 3.2285, "step": 1018 }, { "epoch": 0.2769021739130435, "grad_norm": 1.2714715003967285, "learning_rate": 3e-05, "loss": 4.1331, "step": 1019 }, { "epoch": 0.27717391304347827, "grad_norm": 1.4878464937210083, "learning_rate": 3e-05, "loss": 3.5607, "step": 1020 }, { "epoch": 0.27744565217391304, "grad_norm": 1.3467168807983398, "learning_rate": 3e-05, "loss": 3.4943, "step": 1021 }, { "epoch": 0.2777173913043478, "grad_norm": 1.052966833114624, "learning_rate": 3e-05, "loss": 2.9399, "step": 1022 }, { "epoch": 0.2779891304347826, "grad_norm": 1.1948199272155762, "learning_rate": 3e-05, "loss": 3.0152, "step": 1023 }, { "epoch": 0.2782608695652174, "grad_norm": 1.1960163116455078, "learning_rate": 3e-05, "loss": 2.9338, "step": 1024 }, { "epoch": 0.27853260869565216, "grad_norm": 1.2721116542816162, "learning_rate": 3e-05, "loss": 3.2096, "step": 1025 }, { "epoch": 0.27880434782608693, "grad_norm": 1.620215654373169, "learning_rate": 3e-05, "loss": 3.3828, "step": 1026 }, { "epoch": 0.27907608695652175, "grad_norm": 1.4918383359909058, "learning_rate": 3e-05, "loss": 3.5746, "step": 1027 }, { "epoch": 0.2793478260869565, "grad_norm": 1.9811660051345825, "learning_rate": 3e-05, "loss": 4.0559, "step": 1028 }, { "epoch": 0.2796195652173913, "grad_norm": 1.3812586069107056, "learning_rate": 3e-05, "loss": 3.1452, "step": 1029 }, { "epoch": 0.2798913043478261, "grad_norm": 1.53925359249115, "learning_rate": 3e-05, "loss": 3.9887, "step": 1030 }, { "epoch": 0.2801630434782609, "grad_norm": 2.2425036430358887, "learning_rate": 3e-05, "loss": 4.2516, "step": 1031 }, { "epoch": 0.28043478260869564, "grad_norm": 1.2638776302337646, "learning_rate": 3e-05, "loss": 3.5207, "step": 1032 }, { "epoch": 0.2807065217391304, "grad_norm": 1.1238954067230225, "learning_rate": 3e-05, "loss": 3.4072, "step": 1033 }, { "epoch": 0.28097826086956523, "grad_norm": 1.2834945917129517, "learning_rate": 3e-05, "loss": 3.0585, "step": 1034 }, { "epoch": 0.28125, "grad_norm": 1.5284252166748047, "learning_rate": 3e-05, "loss": 3.1547, "step": 1035 }, { "epoch": 0.28152173913043477, "grad_norm": 1.8825403451919556, "learning_rate": 3e-05, "loss": 3.892, "step": 1036 }, { "epoch": 0.2817934782608696, "grad_norm": 1.620376467704773, "learning_rate": 3e-05, "loss": 2.9044, "step": 1037 }, { "epoch": 0.28206521739130436, "grad_norm": 2.1728601455688477, "learning_rate": 3e-05, "loss": 3.1266, "step": 1038 }, { "epoch": 0.2823369565217391, "grad_norm": 2.090522050857544, "learning_rate": 3e-05, "loss": 4.4403, "step": 1039 }, { "epoch": 0.2826086956521739, "grad_norm": 2.0362656116485596, "learning_rate": 3e-05, "loss": 3.6416, "step": 1040 }, { "epoch": 0.2828804347826087, "grad_norm": 1.493579387664795, "learning_rate": 3e-05, "loss": 3.2111, "step": 1041 }, { "epoch": 0.2831521739130435, "grad_norm": 1.2144675254821777, "learning_rate": 3e-05, "loss": 2.911, "step": 1042 }, { "epoch": 0.28342391304347825, "grad_norm": 1.833510398864746, "learning_rate": 3e-05, "loss": 3.0022, "step": 1043 }, { "epoch": 0.28369565217391307, "grad_norm": 2.0041656494140625, "learning_rate": 3e-05, "loss": 3.1412, "step": 1044 }, { "epoch": 0.28396739130434784, "grad_norm": 1.2218122482299805, "learning_rate": 3e-05, "loss": 3.4581, "step": 1045 }, { "epoch": 0.2842391304347826, "grad_norm": 1.6633751392364502, "learning_rate": 3e-05, "loss": 3.6047, "step": 1046 }, { "epoch": 0.2845108695652174, "grad_norm": 2.4305849075317383, "learning_rate": 3e-05, "loss": 3.3797, "step": 1047 }, { "epoch": 0.2847826086956522, "grad_norm": 2.309483289718628, "learning_rate": 3e-05, "loss": 3.6304, "step": 1048 }, { "epoch": 0.28505434782608696, "grad_norm": 1.678916335105896, "learning_rate": 3e-05, "loss": 3.4073, "step": 1049 }, { "epoch": 0.28532608695652173, "grad_norm": 1.9873173236846924, "learning_rate": 3e-05, "loss": 3.9298, "step": 1050 }, { "epoch": 0.2855978260869565, "grad_norm": 2.085996389389038, "learning_rate": 3e-05, "loss": 4.4263, "step": 1051 }, { "epoch": 0.2858695652173913, "grad_norm": 3.049558639526367, "learning_rate": 3e-05, "loss": 3.6507, "step": 1052 }, { "epoch": 0.2861413043478261, "grad_norm": 1.4257125854492188, "learning_rate": 3e-05, "loss": 3.3861, "step": 1053 }, { "epoch": 0.28641304347826085, "grad_norm": 1.5436605215072632, "learning_rate": 3e-05, "loss": 3.1414, "step": 1054 }, { "epoch": 0.2866847826086957, "grad_norm": 3.0967252254486084, "learning_rate": 3e-05, "loss": 3.9067, "step": 1055 }, { "epoch": 0.28695652173913044, "grad_norm": 1.6564278602600098, "learning_rate": 3e-05, "loss": 3.5354, "step": 1056 }, { "epoch": 0.2872282608695652, "grad_norm": 2.498901605606079, "learning_rate": 3e-05, "loss": 3.3153, "step": 1057 }, { "epoch": 0.2875, "grad_norm": 2.2996490001678467, "learning_rate": 3e-05, "loss": 3.4023, "step": 1058 }, { "epoch": 0.2877717391304348, "grad_norm": 2.0972297191619873, "learning_rate": 3e-05, "loss": 3.038, "step": 1059 }, { "epoch": 0.28804347826086957, "grad_norm": 1.221664547920227, "learning_rate": 3e-05, "loss": 3.0761, "step": 1060 }, { "epoch": 0.28831521739130433, "grad_norm": 2.2344908714294434, "learning_rate": 3e-05, "loss": 2.8314, "step": 1061 }, { "epoch": 0.28858695652173916, "grad_norm": 1.9931285381317139, "learning_rate": 3e-05, "loss": 3.3899, "step": 1062 }, { "epoch": 0.2888586956521739, "grad_norm": 1.924612045288086, "learning_rate": 3e-05, "loss": 3.4548, "step": 1063 }, { "epoch": 0.2891304347826087, "grad_norm": 1.9507229328155518, "learning_rate": 3e-05, "loss": 3.9302, "step": 1064 }, { "epoch": 0.28940217391304346, "grad_norm": 1.8761188983917236, "learning_rate": 3e-05, "loss": 3.195, "step": 1065 }, { "epoch": 0.2896739130434783, "grad_norm": 2.233293294906616, "learning_rate": 3e-05, "loss": 3.1672, "step": 1066 }, { "epoch": 0.28994565217391305, "grad_norm": 1.7154324054718018, "learning_rate": 3e-05, "loss": 3.1413, "step": 1067 }, { "epoch": 0.2902173913043478, "grad_norm": 1.8050798177719116, "learning_rate": 3e-05, "loss": 3.4341, "step": 1068 }, { "epoch": 0.2904891304347826, "grad_norm": 1.838165044784546, "learning_rate": 3e-05, "loss": 3.3827, "step": 1069 }, { "epoch": 0.2907608695652174, "grad_norm": 2.7229554653167725, "learning_rate": 3e-05, "loss": 3.3938, "step": 1070 }, { "epoch": 0.2910326086956522, "grad_norm": 3.150635242462158, "learning_rate": 3e-05, "loss": 3.8396, "step": 1071 }, { "epoch": 0.29130434782608694, "grad_norm": 2.2429847717285156, "learning_rate": 3e-05, "loss": 3.9437, "step": 1072 }, { "epoch": 0.29157608695652176, "grad_norm": 1.305802345275879, "learning_rate": 3e-05, "loss": 3.5962, "step": 1073 }, { "epoch": 0.29184782608695653, "grad_norm": 1.4509013891220093, "learning_rate": 3e-05, "loss": 3.2975, "step": 1074 }, { "epoch": 0.2921195652173913, "grad_norm": 2.059382200241089, "learning_rate": 3e-05, "loss": 4.21, "step": 1075 }, { "epoch": 0.29239130434782606, "grad_norm": 0.9824801087379456, "learning_rate": 3e-05, "loss": 2.6315, "step": 1076 }, { "epoch": 0.2926630434782609, "grad_norm": 1.0867946147918701, "learning_rate": 3e-05, "loss": 2.6838, "step": 1077 }, { "epoch": 0.29293478260869565, "grad_norm": 1.992459774017334, "learning_rate": 3e-05, "loss": 4.2482, "step": 1078 }, { "epoch": 0.2932065217391304, "grad_norm": 1.6963497400283813, "learning_rate": 3e-05, "loss": 3.5684, "step": 1079 }, { "epoch": 0.29347826086956524, "grad_norm": 1.2850168943405151, "learning_rate": 3e-05, "loss": 3.399, "step": 1080 }, { "epoch": 0.29375, "grad_norm": 1.652712106704712, "learning_rate": 3e-05, "loss": 3.3213, "step": 1081 }, { "epoch": 0.2940217391304348, "grad_norm": 1.733959436416626, "learning_rate": 3e-05, "loss": 3.1942, "step": 1082 }, { "epoch": 0.29429347826086955, "grad_norm": 1.8719747066497803, "learning_rate": 3e-05, "loss": 3.8789, "step": 1083 }, { "epoch": 0.29456521739130437, "grad_norm": 1.5426567792892456, "learning_rate": 3e-05, "loss": 3.1319, "step": 1084 }, { "epoch": 0.29483695652173914, "grad_norm": 2.0286598205566406, "learning_rate": 3e-05, "loss": 3.9002, "step": 1085 }, { "epoch": 0.2951086956521739, "grad_norm": 1.5031648874282837, "learning_rate": 3e-05, "loss": 3.3979, "step": 1086 }, { "epoch": 0.29538043478260867, "grad_norm": 1.556738257408142, "learning_rate": 3e-05, "loss": 3.9843, "step": 1087 }, { "epoch": 0.2956521739130435, "grad_norm": 1.8954527378082275, "learning_rate": 3e-05, "loss": 4.6858, "step": 1088 }, { "epoch": 0.29592391304347826, "grad_norm": 1.7330039739608765, "learning_rate": 3e-05, "loss": 2.8431, "step": 1089 }, { "epoch": 0.296195652173913, "grad_norm": 2.0785205364227295, "learning_rate": 3e-05, "loss": 3.6168, "step": 1090 }, { "epoch": 0.29646739130434785, "grad_norm": 1.708628535270691, "learning_rate": 3e-05, "loss": 3.6387, "step": 1091 }, { "epoch": 0.2967391304347826, "grad_norm": 2.5519232749938965, "learning_rate": 3e-05, "loss": 4.0809, "step": 1092 }, { "epoch": 0.2970108695652174, "grad_norm": 1.5967369079589844, "learning_rate": 3e-05, "loss": 3.5908, "step": 1093 }, { "epoch": 0.29728260869565215, "grad_norm": 1.345123052597046, "learning_rate": 3e-05, "loss": 3.297, "step": 1094 }, { "epoch": 0.297554347826087, "grad_norm": 2.123605251312256, "learning_rate": 3e-05, "loss": 4.1267, "step": 1095 }, { "epoch": 0.29782608695652174, "grad_norm": 1.6863003969192505, "learning_rate": 3e-05, "loss": 3.4321, "step": 1096 }, { "epoch": 0.2980978260869565, "grad_norm": 2.0726163387298584, "learning_rate": 3e-05, "loss": 3.386, "step": 1097 }, { "epoch": 0.29836956521739133, "grad_norm": 1.5322779417037964, "learning_rate": 3e-05, "loss": 2.9161, "step": 1098 }, { "epoch": 0.2986413043478261, "grad_norm": 1.1131296157836914, "learning_rate": 3e-05, "loss": 2.7584, "step": 1099 }, { "epoch": 0.29891304347826086, "grad_norm": 1.9430760145187378, "learning_rate": 3e-05, "loss": 3.4661, "step": 1100 }, { "epoch": 0.29918478260869563, "grad_norm": 2.6745107173919678, "learning_rate": 3e-05, "loss": 3.0976, "step": 1101 }, { "epoch": 0.29945652173913045, "grad_norm": 3.127723217010498, "learning_rate": 3e-05, "loss": 3.7013, "step": 1102 }, { "epoch": 0.2997282608695652, "grad_norm": 1.894990086555481, "learning_rate": 3e-05, "loss": 3.4841, "step": 1103 }, { "epoch": 0.3, "grad_norm": 1.9465398788452148, "learning_rate": 3e-05, "loss": 3.48, "step": 1104 }, { "epoch": 0.30027173913043476, "grad_norm": 2.5459132194519043, "learning_rate": 3e-05, "loss": 3.239, "step": 1105 }, { "epoch": 0.3005434782608696, "grad_norm": 2.382801055908203, "learning_rate": 3e-05, "loss": 3.2947, "step": 1106 }, { "epoch": 0.30081521739130435, "grad_norm": 1.913734793663025, "learning_rate": 3e-05, "loss": 3.0162, "step": 1107 }, { "epoch": 0.3010869565217391, "grad_norm": 1.1874632835388184, "learning_rate": 3e-05, "loss": 2.6723, "step": 1108 }, { "epoch": 0.30135869565217394, "grad_norm": 1.6929433345794678, "learning_rate": 3e-05, "loss": 3.1864, "step": 1109 }, { "epoch": 0.3016304347826087, "grad_norm": 2.1046361923217773, "learning_rate": 3e-05, "loss": 3.442, "step": 1110 }, { "epoch": 0.30190217391304347, "grad_norm": 2.196091413497925, "learning_rate": 3e-05, "loss": 3.2791, "step": 1111 }, { "epoch": 0.30217391304347824, "grad_norm": 2.6548757553100586, "learning_rate": 3e-05, "loss": 4.0965, "step": 1112 }, { "epoch": 0.30244565217391306, "grad_norm": 1.7033164501190186, "learning_rate": 3e-05, "loss": 3.5031, "step": 1113 }, { "epoch": 0.3027173913043478, "grad_norm": 1.5866395235061646, "learning_rate": 3e-05, "loss": 2.9015, "step": 1114 }, { "epoch": 0.3029891304347826, "grad_norm": 2.0246646404266357, "learning_rate": 3e-05, "loss": 3.1981, "step": 1115 }, { "epoch": 0.3032608695652174, "grad_norm": 2.0733206272125244, "learning_rate": 3e-05, "loss": 3.3524, "step": 1116 }, { "epoch": 0.3035326086956522, "grad_norm": 1.6320792436599731, "learning_rate": 3e-05, "loss": 3.6667, "step": 1117 }, { "epoch": 0.30380434782608695, "grad_norm": 1.4059102535247803, "learning_rate": 3e-05, "loss": 3.4399, "step": 1118 }, { "epoch": 0.3040760869565217, "grad_norm": 1.6570987701416016, "learning_rate": 3e-05, "loss": 3.2745, "step": 1119 }, { "epoch": 0.30434782608695654, "grad_norm": 1.9712703227996826, "learning_rate": 3e-05, "loss": 3.751, "step": 1120 }, { "epoch": 0.3046195652173913, "grad_norm": 1.4697049856185913, "learning_rate": 3e-05, "loss": 3.4402, "step": 1121 }, { "epoch": 0.3048913043478261, "grad_norm": 1.8021398782730103, "learning_rate": 3e-05, "loss": 3.8907, "step": 1122 }, { "epoch": 0.30516304347826084, "grad_norm": 1.5529664754867554, "learning_rate": 3e-05, "loss": 3.7741, "step": 1123 }, { "epoch": 0.30543478260869567, "grad_norm": 1.5865482091903687, "learning_rate": 3e-05, "loss": 3.7243, "step": 1124 }, { "epoch": 0.30570652173913043, "grad_norm": 2.221069097518921, "learning_rate": 3e-05, "loss": 3.65, "step": 1125 }, { "epoch": 0.3059782608695652, "grad_norm": 2.1728017330169678, "learning_rate": 3e-05, "loss": 3.0357, "step": 1126 }, { "epoch": 0.30625, "grad_norm": 1.6755625009536743, "learning_rate": 3e-05, "loss": 3.2292, "step": 1127 }, { "epoch": 0.3065217391304348, "grad_norm": 2.135910987854004, "learning_rate": 3e-05, "loss": 4.0064, "step": 1128 }, { "epoch": 0.30679347826086956, "grad_norm": 1.7224137783050537, "learning_rate": 3e-05, "loss": 3.6076, "step": 1129 }, { "epoch": 0.3070652173913043, "grad_norm": 1.5196242332458496, "learning_rate": 3e-05, "loss": 3.2132, "step": 1130 }, { "epoch": 0.30733695652173915, "grad_norm": 2.016763210296631, "learning_rate": 3e-05, "loss": 4.1373, "step": 1131 }, { "epoch": 0.3076086956521739, "grad_norm": 1.3778748512268066, "learning_rate": 3e-05, "loss": 3.4815, "step": 1132 }, { "epoch": 0.3078804347826087, "grad_norm": 1.5161099433898926, "learning_rate": 3e-05, "loss": 3.2094, "step": 1133 }, { "epoch": 0.3081521739130435, "grad_norm": 1.2191746234893799, "learning_rate": 3e-05, "loss": 3.0117, "step": 1134 }, { "epoch": 0.30842391304347827, "grad_norm": 1.6270561218261719, "learning_rate": 3e-05, "loss": 3.8738, "step": 1135 }, { "epoch": 0.30869565217391304, "grad_norm": 1.3209682703018188, "learning_rate": 3e-05, "loss": 3.1873, "step": 1136 }, { "epoch": 0.3089673913043478, "grad_norm": 1.7960654497146606, "learning_rate": 3e-05, "loss": 3.9211, "step": 1137 }, { "epoch": 0.3092391304347826, "grad_norm": 1.5645345449447632, "learning_rate": 3e-05, "loss": 3.8176, "step": 1138 }, { "epoch": 0.3095108695652174, "grad_norm": 1.2274796962738037, "learning_rate": 3e-05, "loss": 3.0552, "step": 1139 }, { "epoch": 0.30978260869565216, "grad_norm": 3.1637630462646484, "learning_rate": 3e-05, "loss": 4.0908, "step": 1140 }, { "epoch": 0.31005434782608693, "grad_norm": 1.7452456951141357, "learning_rate": 3e-05, "loss": 3.2379, "step": 1141 }, { "epoch": 0.31032608695652175, "grad_norm": 1.3035019636154175, "learning_rate": 3e-05, "loss": 3.1619, "step": 1142 }, { "epoch": 0.3105978260869565, "grad_norm": 1.5721486806869507, "learning_rate": 3e-05, "loss": 3.3092, "step": 1143 }, { "epoch": 0.3108695652173913, "grad_norm": 1.5714532136917114, "learning_rate": 3e-05, "loss": 3.7343, "step": 1144 }, { "epoch": 0.3111413043478261, "grad_norm": 1.4483362436294556, "learning_rate": 3e-05, "loss": 3.5843, "step": 1145 }, { "epoch": 0.3114130434782609, "grad_norm": 1.2886680364608765, "learning_rate": 3e-05, "loss": 3.4743, "step": 1146 }, { "epoch": 0.31168478260869564, "grad_norm": 1.5609358549118042, "learning_rate": 3e-05, "loss": 3.7947, "step": 1147 }, { "epoch": 0.3119565217391304, "grad_norm": 1.7903861999511719, "learning_rate": 3e-05, "loss": 3.3165, "step": 1148 }, { "epoch": 0.31222826086956523, "grad_norm": 1.917932152748108, "learning_rate": 3e-05, "loss": 3.3985, "step": 1149 }, { "epoch": 0.3125, "grad_norm": 1.5185085535049438, "learning_rate": 3e-05, "loss": 3.571, "step": 1150 }, { "epoch": 0.31277173913043477, "grad_norm": 1.3498207330703735, "learning_rate": 3e-05, "loss": 3.3921, "step": 1151 }, { "epoch": 0.3130434782608696, "grad_norm": 1.6622629165649414, "learning_rate": 3e-05, "loss": 3.4101, "step": 1152 }, { "epoch": 0.31331521739130436, "grad_norm": 1.516265630722046, "learning_rate": 3e-05, "loss": 3.1228, "step": 1153 }, { "epoch": 0.3135869565217391, "grad_norm": 1.5660970211029053, "learning_rate": 3e-05, "loss": 3.3751, "step": 1154 }, { "epoch": 0.3138586956521739, "grad_norm": 1.8406776189804077, "learning_rate": 3e-05, "loss": 3.3181, "step": 1155 }, { "epoch": 0.3141304347826087, "grad_norm": 1.8162765502929688, "learning_rate": 3e-05, "loss": 4.0117, "step": 1156 }, { "epoch": 0.3144021739130435, "grad_norm": 1.2983214855194092, "learning_rate": 3e-05, "loss": 3.2412, "step": 1157 }, { "epoch": 0.31467391304347825, "grad_norm": 1.1409294605255127, "learning_rate": 3e-05, "loss": 3.0329, "step": 1158 }, { "epoch": 0.31494565217391307, "grad_norm": 1.754345178604126, "learning_rate": 3e-05, "loss": 3.6711, "step": 1159 }, { "epoch": 0.31521739130434784, "grad_norm": 1.868580937385559, "learning_rate": 3e-05, "loss": 3.6684, "step": 1160 }, { "epoch": 0.3154891304347826, "grad_norm": 1.8012367486953735, "learning_rate": 3e-05, "loss": 3.2884, "step": 1161 }, { "epoch": 0.3157608695652174, "grad_norm": 1.7688064575195312, "learning_rate": 3e-05, "loss": 3.3512, "step": 1162 }, { "epoch": 0.3160326086956522, "grad_norm": 1.5674185752868652, "learning_rate": 3e-05, "loss": 3.443, "step": 1163 }, { "epoch": 0.31630434782608696, "grad_norm": 1.5843913555145264, "learning_rate": 3e-05, "loss": 3.8519, "step": 1164 }, { "epoch": 0.31657608695652173, "grad_norm": 1.4522455930709839, "learning_rate": 3e-05, "loss": 3.123, "step": 1165 }, { "epoch": 0.3168478260869565, "grad_norm": 1.5610785484313965, "learning_rate": 3e-05, "loss": 3.1963, "step": 1166 }, { "epoch": 0.3171195652173913, "grad_norm": 1.4327051639556885, "learning_rate": 3e-05, "loss": 2.9382, "step": 1167 }, { "epoch": 0.3173913043478261, "grad_norm": 1.6065129041671753, "learning_rate": 3e-05, "loss": 3.4928, "step": 1168 }, { "epoch": 0.31766304347826085, "grad_norm": 1.9382498264312744, "learning_rate": 3e-05, "loss": 4.3174, "step": 1169 }, { "epoch": 0.3179347826086957, "grad_norm": 1.164689064025879, "learning_rate": 3e-05, "loss": 3.0678, "step": 1170 }, { "epoch": 0.31820652173913044, "grad_norm": 1.084055781364441, "learning_rate": 3e-05, "loss": 3.0878, "step": 1171 }, { "epoch": 0.3184782608695652, "grad_norm": 1.4881168603897095, "learning_rate": 3e-05, "loss": 3.5084, "step": 1172 }, { "epoch": 0.31875, "grad_norm": 1.5508403778076172, "learning_rate": 3e-05, "loss": 3.6359, "step": 1173 }, { "epoch": 0.3190217391304348, "grad_norm": 1.7135038375854492, "learning_rate": 3e-05, "loss": 2.819, "step": 1174 }, { "epoch": 0.31929347826086957, "grad_norm": 1.9360867738723755, "learning_rate": 3e-05, "loss": 3.711, "step": 1175 }, { "epoch": 0.31956521739130433, "grad_norm": 1.3898394107818604, "learning_rate": 3e-05, "loss": 3.0356, "step": 1176 }, { "epoch": 0.31983695652173916, "grad_norm": 1.37100088596344, "learning_rate": 3e-05, "loss": 2.9441, "step": 1177 }, { "epoch": 0.3201086956521739, "grad_norm": 1.7066446542739868, "learning_rate": 3e-05, "loss": 3.8696, "step": 1178 }, { "epoch": 0.3203804347826087, "grad_norm": 1.4085134267807007, "learning_rate": 3e-05, "loss": 3.4616, "step": 1179 }, { "epoch": 0.32065217391304346, "grad_norm": 1.5345913171768188, "learning_rate": 3e-05, "loss": 3.6033, "step": 1180 }, { "epoch": 0.3209239130434783, "grad_norm": 1.5240819454193115, "learning_rate": 3e-05, "loss": 3.1863, "step": 1181 }, { "epoch": 0.32119565217391305, "grad_norm": 1.147559404373169, "learning_rate": 3e-05, "loss": 3.2533, "step": 1182 }, { "epoch": 0.3214673913043478, "grad_norm": 1.3267796039581299, "learning_rate": 3e-05, "loss": 3.3342, "step": 1183 }, { "epoch": 0.3217391304347826, "grad_norm": 1.567126989364624, "learning_rate": 3e-05, "loss": 3.5055, "step": 1184 }, { "epoch": 0.3220108695652174, "grad_norm": 2.085658073425293, "learning_rate": 3e-05, "loss": 3.1383, "step": 1185 }, { "epoch": 0.3222826086956522, "grad_norm": 1.7624791860580444, "learning_rate": 3e-05, "loss": 3.2409, "step": 1186 }, { "epoch": 0.32255434782608694, "grad_norm": 1.4368826150894165, "learning_rate": 3e-05, "loss": 3.7424, "step": 1187 }, { "epoch": 0.32282608695652176, "grad_norm": 1.5510128736495972, "learning_rate": 3e-05, "loss": 3.3895, "step": 1188 }, { "epoch": 0.32309782608695653, "grad_norm": 1.471490740776062, "learning_rate": 3e-05, "loss": 3.6772, "step": 1189 }, { "epoch": 0.3233695652173913, "grad_norm": 1.2040995359420776, "learning_rate": 3e-05, "loss": 3.3302, "step": 1190 }, { "epoch": 0.32364130434782606, "grad_norm": 1.4731234312057495, "learning_rate": 3e-05, "loss": 3.3842, "step": 1191 }, { "epoch": 0.3239130434782609, "grad_norm": 1.3904873132705688, "learning_rate": 3e-05, "loss": 3.8537, "step": 1192 }, { "epoch": 0.32418478260869565, "grad_norm": 1.187960147857666, "learning_rate": 3e-05, "loss": 3.0656, "step": 1193 }, { "epoch": 0.3244565217391304, "grad_norm": 1.1819937229156494, "learning_rate": 3e-05, "loss": 2.7234, "step": 1194 }, { "epoch": 0.32472826086956524, "grad_norm": 1.6428884267807007, "learning_rate": 3e-05, "loss": 3.4446, "step": 1195 }, { "epoch": 0.325, "grad_norm": 1.6877864599227905, "learning_rate": 3e-05, "loss": 3.2674, "step": 1196 }, { "epoch": 0.3252717391304348, "grad_norm": 1.596166729927063, "learning_rate": 3e-05, "loss": 3.371, "step": 1197 }, { "epoch": 0.32554347826086955, "grad_norm": 1.3886607885360718, "learning_rate": 3e-05, "loss": 3.3149, "step": 1198 }, { "epoch": 0.32581521739130437, "grad_norm": 1.4295148849487305, "learning_rate": 3e-05, "loss": 3.0519, "step": 1199 }, { "epoch": 0.32608695652173914, "grad_norm": 1.5317387580871582, "learning_rate": 3e-05, "loss": 3.3816, "step": 1200 }, { "epoch": 0.3263586956521739, "grad_norm": 1.1524282693862915, "learning_rate": 3e-05, "loss": 3.2564, "step": 1201 }, { "epoch": 0.32663043478260867, "grad_norm": 1.3997209072113037, "learning_rate": 3e-05, "loss": 3.3173, "step": 1202 }, { "epoch": 0.3269021739130435, "grad_norm": 1.357279896736145, "learning_rate": 3e-05, "loss": 3.7316, "step": 1203 }, { "epoch": 0.32717391304347826, "grad_norm": 1.1962755918502808, "learning_rate": 3e-05, "loss": 3.5203, "step": 1204 }, { "epoch": 0.327445652173913, "grad_norm": 1.4180536270141602, "learning_rate": 3e-05, "loss": 3.7679, "step": 1205 }, { "epoch": 0.32771739130434785, "grad_norm": 1.4997798204421997, "learning_rate": 3e-05, "loss": 3.9842, "step": 1206 }, { "epoch": 0.3279891304347826, "grad_norm": 1.5224746465682983, "learning_rate": 3e-05, "loss": 3.723, "step": 1207 }, { "epoch": 0.3282608695652174, "grad_norm": 1.4814677238464355, "learning_rate": 3e-05, "loss": 3.3444, "step": 1208 }, { "epoch": 0.32853260869565215, "grad_norm": 1.8869359493255615, "learning_rate": 3e-05, "loss": 3.8663, "step": 1209 }, { "epoch": 0.328804347826087, "grad_norm": 1.2305731773376465, "learning_rate": 3e-05, "loss": 3.1855, "step": 1210 }, { "epoch": 0.32907608695652174, "grad_norm": 1.3822002410888672, "learning_rate": 3e-05, "loss": 3.9625, "step": 1211 }, { "epoch": 0.3293478260869565, "grad_norm": 1.2895426750183105, "learning_rate": 3e-05, "loss": 2.9574, "step": 1212 }, { "epoch": 0.32961956521739133, "grad_norm": 1.2724965810775757, "learning_rate": 3e-05, "loss": 3.2984, "step": 1213 }, { "epoch": 0.3298913043478261, "grad_norm": 1.569366455078125, "learning_rate": 3e-05, "loss": 3.4928, "step": 1214 }, { "epoch": 0.33016304347826086, "grad_norm": 1.538090467453003, "learning_rate": 3e-05, "loss": 3.8567, "step": 1215 }, { "epoch": 0.33043478260869563, "grad_norm": 1.3484816551208496, "learning_rate": 3e-05, "loss": 3.1096, "step": 1216 }, { "epoch": 0.33070652173913045, "grad_norm": 1.1450049877166748, "learning_rate": 3e-05, "loss": 2.773, "step": 1217 }, { "epoch": 0.3309782608695652, "grad_norm": 1.5292000770568848, "learning_rate": 3e-05, "loss": 3.3972, "step": 1218 }, { "epoch": 0.33125, "grad_norm": 1.3615137338638306, "learning_rate": 3e-05, "loss": 3.2462, "step": 1219 }, { "epoch": 0.33152173913043476, "grad_norm": 1.7036170959472656, "learning_rate": 3e-05, "loss": 3.7338, "step": 1220 }, { "epoch": 0.3317934782608696, "grad_norm": 1.7132785320281982, "learning_rate": 3e-05, "loss": 3.3378, "step": 1221 }, { "epoch": 0.33206521739130435, "grad_norm": 1.9641152620315552, "learning_rate": 3e-05, "loss": 3.8121, "step": 1222 }, { "epoch": 0.3323369565217391, "grad_norm": 1.3222547769546509, "learning_rate": 3e-05, "loss": 3.2911, "step": 1223 }, { "epoch": 0.33260869565217394, "grad_norm": 1.3695377111434937, "learning_rate": 3e-05, "loss": 3.3004, "step": 1224 }, { "epoch": 0.3328804347826087, "grad_norm": 1.8578277826309204, "learning_rate": 3e-05, "loss": 4.1831, "step": 1225 }, { "epoch": 0.33315217391304347, "grad_norm": 1.7893558740615845, "learning_rate": 3e-05, "loss": 3.359, "step": 1226 }, { "epoch": 0.33342391304347824, "grad_norm": 1.474753499031067, "learning_rate": 3e-05, "loss": 2.9094, "step": 1227 }, { "epoch": 0.33369565217391306, "grad_norm": 1.973402738571167, "learning_rate": 3e-05, "loss": 4.0505, "step": 1228 }, { "epoch": 0.3339673913043478, "grad_norm": 1.7613177299499512, "learning_rate": 3e-05, "loss": 3.3559, "step": 1229 }, { "epoch": 0.3342391304347826, "grad_norm": 2.031470775604248, "learning_rate": 3e-05, "loss": 3.3066, "step": 1230 }, { "epoch": 0.3345108695652174, "grad_norm": 2.1547582149505615, "learning_rate": 3e-05, "loss": 3.6043, "step": 1231 }, { "epoch": 0.3347826086956522, "grad_norm": 2.2331910133361816, "learning_rate": 3e-05, "loss": 2.9871, "step": 1232 }, { "epoch": 0.33505434782608695, "grad_norm": 2.156770944595337, "learning_rate": 3e-05, "loss": 3.6777, "step": 1233 }, { "epoch": 0.3353260869565217, "grad_norm": 1.717960000038147, "learning_rate": 3e-05, "loss": 3.6268, "step": 1234 }, { "epoch": 0.33559782608695654, "grad_norm": 1.649511694908142, "learning_rate": 3e-05, "loss": 3.3226, "step": 1235 }, { "epoch": 0.3358695652173913, "grad_norm": 1.8623225688934326, "learning_rate": 3e-05, "loss": 3.9329, "step": 1236 }, { "epoch": 0.3361413043478261, "grad_norm": 2.167483329772949, "learning_rate": 3e-05, "loss": 3.445, "step": 1237 }, { "epoch": 0.33641304347826084, "grad_norm": 1.7240395545959473, "learning_rate": 3e-05, "loss": 3.493, "step": 1238 }, { "epoch": 0.33668478260869567, "grad_norm": 1.3711011409759521, "learning_rate": 3e-05, "loss": 2.7254, "step": 1239 }, { "epoch": 0.33695652173913043, "grad_norm": 2.091745615005493, "learning_rate": 3e-05, "loss": 4.0388, "step": 1240 }, { "epoch": 0.3372282608695652, "grad_norm": 1.5960465669631958, "learning_rate": 3e-05, "loss": 3.5036, "step": 1241 }, { "epoch": 0.3375, "grad_norm": 1.1491891145706177, "learning_rate": 3e-05, "loss": 3.2282, "step": 1242 }, { "epoch": 0.3377717391304348, "grad_norm": 1.1655139923095703, "learning_rate": 3e-05, "loss": 3.3578, "step": 1243 }, { "epoch": 0.33804347826086956, "grad_norm": 1.4522464275360107, "learning_rate": 3e-05, "loss": 3.4827, "step": 1244 }, { "epoch": 0.3383152173913043, "grad_norm": 1.494310736656189, "learning_rate": 3e-05, "loss": 3.7874, "step": 1245 }, { "epoch": 0.33858695652173915, "grad_norm": 1.98121976852417, "learning_rate": 3e-05, "loss": 3.2323, "step": 1246 }, { "epoch": 0.3388586956521739, "grad_norm": 1.772510290145874, "learning_rate": 3e-05, "loss": 2.9783, "step": 1247 }, { "epoch": 0.3391304347826087, "grad_norm": 1.3050928115844727, "learning_rate": 3e-05, "loss": 3.3902, "step": 1248 }, { "epoch": 0.3394021739130435, "grad_norm": 1.590279221534729, "learning_rate": 3e-05, "loss": 3.0477, "step": 1249 }, { "epoch": 0.33967391304347827, "grad_norm": 2.1541528701782227, "learning_rate": 3e-05, "loss": 3.5042, "step": 1250 }, { "epoch": 0.33994565217391304, "grad_norm": 2.138265609741211, "learning_rate": 3e-05, "loss": 3.704, "step": 1251 }, { "epoch": 0.3402173913043478, "grad_norm": 1.5365831851959229, "learning_rate": 3e-05, "loss": 3.3562, "step": 1252 }, { "epoch": 0.3404891304347826, "grad_norm": 1.3827742338180542, "learning_rate": 3e-05, "loss": 3.2202, "step": 1253 }, { "epoch": 0.3407608695652174, "grad_norm": 1.4450342655181885, "learning_rate": 3e-05, "loss": 3.4887, "step": 1254 }, { "epoch": 0.34103260869565216, "grad_norm": 1.2322391271591187, "learning_rate": 3e-05, "loss": 2.7866, "step": 1255 }, { "epoch": 0.34130434782608693, "grad_norm": 1.3767555952072144, "learning_rate": 3e-05, "loss": 2.8691, "step": 1256 }, { "epoch": 0.34157608695652175, "grad_norm": 1.9902418851852417, "learning_rate": 3e-05, "loss": 3.5811, "step": 1257 }, { "epoch": 0.3418478260869565, "grad_norm": 1.7038296461105347, "learning_rate": 3e-05, "loss": 3.363, "step": 1258 }, { "epoch": 0.3421195652173913, "grad_norm": 1.82673978805542, "learning_rate": 3e-05, "loss": 3.0324, "step": 1259 }, { "epoch": 0.3423913043478261, "grad_norm": 1.4924451112747192, "learning_rate": 3e-05, "loss": 3.5862, "step": 1260 }, { "epoch": 0.3426630434782609, "grad_norm": 1.1290408372879028, "learning_rate": 3e-05, "loss": 3.0983, "step": 1261 }, { "epoch": 0.34293478260869564, "grad_norm": 1.3652050495147705, "learning_rate": 3e-05, "loss": 3.2806, "step": 1262 }, { "epoch": 0.3432065217391304, "grad_norm": 1.3618534803390503, "learning_rate": 3e-05, "loss": 3.0708, "step": 1263 }, { "epoch": 0.34347826086956523, "grad_norm": 1.3616082668304443, "learning_rate": 3e-05, "loss": 2.8319, "step": 1264 }, { "epoch": 0.34375, "grad_norm": 1.3398898839950562, "learning_rate": 3e-05, "loss": 3.1065, "step": 1265 }, { "epoch": 0.34402173913043477, "grad_norm": 1.6687146425247192, "learning_rate": 3e-05, "loss": 3.3584, "step": 1266 }, { "epoch": 0.3442934782608696, "grad_norm": 1.2948143482208252, "learning_rate": 3e-05, "loss": 3.2489, "step": 1267 }, { "epoch": 0.34456521739130436, "grad_norm": 1.7158524990081787, "learning_rate": 3e-05, "loss": 3.3558, "step": 1268 }, { "epoch": 0.3448369565217391, "grad_norm": 1.6835163831710815, "learning_rate": 3e-05, "loss": 2.9967, "step": 1269 }, { "epoch": 0.3451086956521739, "grad_norm": 1.971177101135254, "learning_rate": 3e-05, "loss": 3.829, "step": 1270 }, { "epoch": 0.3453804347826087, "grad_norm": 1.660911202430725, "learning_rate": 3e-05, "loss": 2.9975, "step": 1271 }, { "epoch": 0.3456521739130435, "grad_norm": 1.8289529085159302, "learning_rate": 3e-05, "loss": 3.4466, "step": 1272 }, { "epoch": 0.34592391304347825, "grad_norm": 1.9580225944519043, "learning_rate": 3e-05, "loss": 3.4431, "step": 1273 }, { "epoch": 0.34619565217391307, "grad_norm": 1.9081965684890747, "learning_rate": 3e-05, "loss": 3.1295, "step": 1274 }, { "epoch": 0.34646739130434784, "grad_norm": 2.8098397254943848, "learning_rate": 3e-05, "loss": 3.9265, "step": 1275 }, { "epoch": 0.3467391304347826, "grad_norm": 1.8996559381484985, "learning_rate": 3e-05, "loss": 2.9118, "step": 1276 }, { "epoch": 0.3470108695652174, "grad_norm": 1.8447978496551514, "learning_rate": 3e-05, "loss": 3.8908, "step": 1277 }, { "epoch": 0.3472826086956522, "grad_norm": 1.450376033782959, "learning_rate": 3e-05, "loss": 3.1262, "step": 1278 }, { "epoch": 0.34755434782608696, "grad_norm": 2.3735806941986084, "learning_rate": 3e-05, "loss": 3.4959, "step": 1279 }, { "epoch": 0.34782608695652173, "grad_norm": 2.2202954292297363, "learning_rate": 3e-05, "loss": 3.2127, "step": 1280 }, { "epoch": 0.3480978260869565, "grad_norm": 1.6313713788986206, "learning_rate": 3e-05, "loss": 2.7593, "step": 1281 }, { "epoch": 0.3483695652173913, "grad_norm": 1.8261613845825195, "learning_rate": 3e-05, "loss": 3.5207, "step": 1282 }, { "epoch": 0.3486413043478261, "grad_norm": 1.6101502180099487, "learning_rate": 3e-05, "loss": 3.1187, "step": 1283 }, { "epoch": 0.34891304347826085, "grad_norm": 1.4293962717056274, "learning_rate": 3e-05, "loss": 3.3783, "step": 1284 }, { "epoch": 0.3491847826086957, "grad_norm": 1.335721731185913, "learning_rate": 3e-05, "loss": 3.8365, "step": 1285 }, { "epoch": 0.34945652173913044, "grad_norm": 1.5614105463027954, "learning_rate": 3e-05, "loss": 3.7661, "step": 1286 }, { "epoch": 0.3497282608695652, "grad_norm": 1.38046133518219, "learning_rate": 3e-05, "loss": 3.1717, "step": 1287 }, { "epoch": 0.35, "grad_norm": 1.3673638105392456, "learning_rate": 3e-05, "loss": 3.7002, "step": 1288 }, { "epoch": 0.3502717391304348, "grad_norm": 1.2321789264678955, "learning_rate": 3e-05, "loss": 2.8849, "step": 1289 }, { "epoch": 0.35054347826086957, "grad_norm": 1.3166979551315308, "learning_rate": 3e-05, "loss": 3.1094, "step": 1290 }, { "epoch": 0.35081521739130433, "grad_norm": 1.5346661806106567, "learning_rate": 3e-05, "loss": 3.2696, "step": 1291 }, { "epoch": 0.35108695652173916, "grad_norm": 1.550195336341858, "learning_rate": 3e-05, "loss": 3.3044, "step": 1292 }, { "epoch": 0.3513586956521739, "grad_norm": 1.7930152416229248, "learning_rate": 3e-05, "loss": 3.7782, "step": 1293 }, { "epoch": 0.3516304347826087, "grad_norm": 1.9929698705673218, "learning_rate": 3e-05, "loss": 3.12, "step": 1294 }, { "epoch": 0.35190217391304346, "grad_norm": 1.244729995727539, "learning_rate": 3e-05, "loss": 2.9368, "step": 1295 }, { "epoch": 0.3521739130434783, "grad_norm": 1.3026173114776611, "learning_rate": 3e-05, "loss": 3.1947, "step": 1296 }, { "epoch": 0.35244565217391305, "grad_norm": 1.297813892364502, "learning_rate": 3e-05, "loss": 2.5978, "step": 1297 }, { "epoch": 0.3527173913043478, "grad_norm": 1.4547146558761597, "learning_rate": 3e-05, "loss": 3.4649, "step": 1298 }, { "epoch": 0.3529891304347826, "grad_norm": 1.2365776300430298, "learning_rate": 3e-05, "loss": 3.1037, "step": 1299 }, { "epoch": 0.3532608695652174, "grad_norm": 1.2477210760116577, "learning_rate": 3e-05, "loss": 3.1045, "step": 1300 }, { "epoch": 0.3535326086956522, "grad_norm": 1.4197901487350464, "learning_rate": 3e-05, "loss": 3.2411, "step": 1301 }, { "epoch": 0.35380434782608694, "grad_norm": 1.1791913509368896, "learning_rate": 3e-05, "loss": 3.4775, "step": 1302 }, { "epoch": 0.35407608695652176, "grad_norm": 1.1935291290283203, "learning_rate": 3e-05, "loss": 2.9038, "step": 1303 }, { "epoch": 0.35434782608695653, "grad_norm": 1.515292763710022, "learning_rate": 3e-05, "loss": 3.6569, "step": 1304 }, { "epoch": 0.3546195652173913, "grad_norm": 1.6459676027297974, "learning_rate": 3e-05, "loss": 3.638, "step": 1305 }, { "epoch": 0.35489130434782606, "grad_norm": 1.2799147367477417, "learning_rate": 3e-05, "loss": 3.3, "step": 1306 }, { "epoch": 0.3551630434782609, "grad_norm": 1.1415566205978394, "learning_rate": 3e-05, "loss": 2.7568, "step": 1307 }, { "epoch": 0.35543478260869565, "grad_norm": 1.4160751104354858, "learning_rate": 3e-05, "loss": 3.5389, "step": 1308 }, { "epoch": 0.3557065217391304, "grad_norm": 1.5589027404785156, "learning_rate": 3e-05, "loss": 3.5721, "step": 1309 }, { "epoch": 0.35597826086956524, "grad_norm": 1.4486863613128662, "learning_rate": 3e-05, "loss": 3.5145, "step": 1310 }, { "epoch": 0.35625, "grad_norm": 1.1204112768173218, "learning_rate": 3e-05, "loss": 3.2109, "step": 1311 }, { "epoch": 0.3565217391304348, "grad_norm": 1.4160523414611816, "learning_rate": 3e-05, "loss": 3.9734, "step": 1312 }, { "epoch": 0.35679347826086955, "grad_norm": 1.359877347946167, "learning_rate": 3e-05, "loss": 3.1197, "step": 1313 }, { "epoch": 0.35706521739130437, "grad_norm": 1.3469575643539429, "learning_rate": 3e-05, "loss": 2.7031, "step": 1314 }, { "epoch": 0.35733695652173914, "grad_norm": 1.877429485321045, "learning_rate": 3e-05, "loss": 3.6433, "step": 1315 }, { "epoch": 0.3576086956521739, "grad_norm": 1.6881115436553955, "learning_rate": 3e-05, "loss": 3.8429, "step": 1316 }, { "epoch": 0.35788043478260867, "grad_norm": 0.9885056018829346, "learning_rate": 3e-05, "loss": 2.9943, "step": 1317 }, { "epoch": 0.3581521739130435, "grad_norm": 1.1146260499954224, "learning_rate": 3e-05, "loss": 3.0583, "step": 1318 }, { "epoch": 0.35842391304347826, "grad_norm": 1.0930460691452026, "learning_rate": 3e-05, "loss": 3.1491, "step": 1319 }, { "epoch": 0.358695652173913, "grad_norm": 1.7721790075302124, "learning_rate": 3e-05, "loss": 3.682, "step": 1320 }, { "epoch": 0.35896739130434785, "grad_norm": 1.8233963251113892, "learning_rate": 3e-05, "loss": 4.0548, "step": 1321 }, { "epoch": 0.3592391304347826, "grad_norm": 2.3196828365325928, "learning_rate": 3e-05, "loss": 3.8065, "step": 1322 }, { "epoch": 0.3595108695652174, "grad_norm": 2.0975351333618164, "learning_rate": 3e-05, "loss": 3.3098, "step": 1323 }, { "epoch": 0.35978260869565215, "grad_norm": 1.4240121841430664, "learning_rate": 3e-05, "loss": 3.327, "step": 1324 }, { "epoch": 0.360054347826087, "grad_norm": 1.4401570558547974, "learning_rate": 3e-05, "loss": 2.9099, "step": 1325 }, { "epoch": 0.36032608695652174, "grad_norm": 1.7879765033721924, "learning_rate": 3e-05, "loss": 2.9776, "step": 1326 }, { "epoch": 0.3605978260869565, "grad_norm": 1.8819643259048462, "learning_rate": 3e-05, "loss": 3.3526, "step": 1327 }, { "epoch": 0.36086956521739133, "grad_norm": 1.8750115633010864, "learning_rate": 3e-05, "loss": 3.7411, "step": 1328 }, { "epoch": 0.3611413043478261, "grad_norm": 1.802832007408142, "learning_rate": 3e-05, "loss": 2.8997, "step": 1329 }, { "epoch": 0.36141304347826086, "grad_norm": 1.9432928562164307, "learning_rate": 3e-05, "loss": 2.9711, "step": 1330 }, { "epoch": 0.36168478260869563, "grad_norm": 2.222259998321533, "learning_rate": 3e-05, "loss": 3.5956, "step": 1331 }, { "epoch": 0.36195652173913045, "grad_norm": 1.0173529386520386, "learning_rate": 3e-05, "loss": 3.1368, "step": 1332 }, { "epoch": 0.3622282608695652, "grad_norm": 1.3840457201004028, "learning_rate": 3e-05, "loss": 3.3336, "step": 1333 }, { "epoch": 0.3625, "grad_norm": 2.199570417404175, "learning_rate": 3e-05, "loss": 3.1846, "step": 1334 }, { "epoch": 0.36277173913043476, "grad_norm": 2.3063576221466064, "learning_rate": 3e-05, "loss": 3.1996, "step": 1335 }, { "epoch": 0.3630434782608696, "grad_norm": 2.0329015254974365, "learning_rate": 3e-05, "loss": 3.6631, "step": 1336 }, { "epoch": 0.36331521739130435, "grad_norm": 0.9080404043197632, "learning_rate": 3e-05, "loss": 2.8746, "step": 1337 }, { "epoch": 0.3635869565217391, "grad_norm": 2.173773765563965, "learning_rate": 3e-05, "loss": 3.8486, "step": 1338 }, { "epoch": 0.36385869565217394, "grad_norm": 2.493887186050415, "learning_rate": 3e-05, "loss": 3.5707, "step": 1339 }, { "epoch": 0.3641304347826087, "grad_norm": 1.8862491846084595, "learning_rate": 3e-05, "loss": 3.8585, "step": 1340 }, { "epoch": 0.36440217391304347, "grad_norm": 1.23379647731781, "learning_rate": 3e-05, "loss": 3.5806, "step": 1341 }, { "epoch": 0.36467391304347824, "grad_norm": 1.3684219121932983, "learning_rate": 3e-05, "loss": 3.1318, "step": 1342 }, { "epoch": 0.36494565217391306, "grad_norm": 1.2184761762619019, "learning_rate": 3e-05, "loss": 3.4761, "step": 1343 }, { "epoch": 0.3652173913043478, "grad_norm": 1.230040431022644, "learning_rate": 3e-05, "loss": 4.1053, "step": 1344 }, { "epoch": 0.3654891304347826, "grad_norm": 1.4862629175186157, "learning_rate": 3e-05, "loss": 3.4211, "step": 1345 }, { "epoch": 0.3657608695652174, "grad_norm": 1.2832411527633667, "learning_rate": 3e-05, "loss": 2.7821, "step": 1346 }, { "epoch": 0.3660326086956522, "grad_norm": 1.1333979368209839, "learning_rate": 3e-05, "loss": 3.4129, "step": 1347 }, { "epoch": 0.36630434782608695, "grad_norm": 1.9462121725082397, "learning_rate": 3e-05, "loss": 3.8102, "step": 1348 }, { "epoch": 0.3665760869565217, "grad_norm": 1.314512848854065, "learning_rate": 3e-05, "loss": 3.5802, "step": 1349 }, { "epoch": 0.36684782608695654, "grad_norm": 0.9899564981460571, "learning_rate": 3e-05, "loss": 3.2397, "step": 1350 }, { "epoch": 0.3671195652173913, "grad_norm": 0.9129654765129089, "learning_rate": 3e-05, "loss": 3.0704, "step": 1351 }, { "epoch": 0.3673913043478261, "grad_norm": 1.2658374309539795, "learning_rate": 3e-05, "loss": 2.9416, "step": 1352 }, { "epoch": 0.36766304347826084, "grad_norm": 1.156530499458313, "learning_rate": 3e-05, "loss": 3.2254, "step": 1353 }, { "epoch": 0.36793478260869567, "grad_norm": 1.4215689897537231, "learning_rate": 3e-05, "loss": 3.5928, "step": 1354 }, { "epoch": 0.36820652173913043, "grad_norm": 1.4695452451705933, "learning_rate": 3e-05, "loss": 3.5298, "step": 1355 }, { "epoch": 0.3684782608695652, "grad_norm": 1.185398817062378, "learning_rate": 3e-05, "loss": 2.9786, "step": 1356 }, { "epoch": 0.36875, "grad_norm": 1.6181942224502563, "learning_rate": 3e-05, "loss": 3.8328, "step": 1357 }, { "epoch": 0.3690217391304348, "grad_norm": 1.7642178535461426, "learning_rate": 3e-05, "loss": 3.6084, "step": 1358 }, { "epoch": 0.36929347826086956, "grad_norm": 1.3343443870544434, "learning_rate": 3e-05, "loss": 4.0491, "step": 1359 }, { "epoch": 0.3695652173913043, "grad_norm": 2.167877435684204, "learning_rate": 3e-05, "loss": 3.6741, "step": 1360 }, { "epoch": 0.36983695652173915, "grad_norm": 2.0322484970092773, "learning_rate": 3e-05, "loss": 4.1298, "step": 1361 }, { "epoch": 0.3701086956521739, "grad_norm": 1.1205507516860962, "learning_rate": 3e-05, "loss": 3.2478, "step": 1362 }, { "epoch": 0.3703804347826087, "grad_norm": 1.4173895120620728, "learning_rate": 3e-05, "loss": 3.2024, "step": 1363 }, { "epoch": 0.3706521739130435, "grad_norm": 1.4186296463012695, "learning_rate": 3e-05, "loss": 2.9463, "step": 1364 }, { "epoch": 0.37092391304347827, "grad_norm": 1.4897836446762085, "learning_rate": 3e-05, "loss": 3.4138, "step": 1365 }, { "epoch": 0.37119565217391304, "grad_norm": 0.9213505983352661, "learning_rate": 3e-05, "loss": 2.7726, "step": 1366 }, { "epoch": 0.3714673913043478, "grad_norm": 1.4468334913253784, "learning_rate": 3e-05, "loss": 3.3552, "step": 1367 }, { "epoch": 0.3717391304347826, "grad_norm": 1.6020596027374268, "learning_rate": 3e-05, "loss": 3.3098, "step": 1368 }, { "epoch": 0.3720108695652174, "grad_norm": 1.2568944692611694, "learning_rate": 3e-05, "loss": 2.9896, "step": 1369 }, { "epoch": 0.37228260869565216, "grad_norm": 1.2092691659927368, "learning_rate": 3e-05, "loss": 3.2886, "step": 1370 }, { "epoch": 0.37255434782608693, "grad_norm": 1.0227844715118408, "learning_rate": 3e-05, "loss": 2.9555, "step": 1371 }, { "epoch": 0.37282608695652175, "grad_norm": 1.400106430053711, "learning_rate": 3e-05, "loss": 3.2697, "step": 1372 }, { "epoch": 0.3730978260869565, "grad_norm": 1.3000365495681763, "learning_rate": 3e-05, "loss": 3.1748, "step": 1373 }, { "epoch": 0.3733695652173913, "grad_norm": 1.197072148323059, "learning_rate": 3e-05, "loss": 3.3346, "step": 1374 }, { "epoch": 0.3736413043478261, "grad_norm": 1.1778289079666138, "learning_rate": 3e-05, "loss": 3.089, "step": 1375 }, { "epoch": 0.3739130434782609, "grad_norm": 1.23442542552948, "learning_rate": 3e-05, "loss": 3.0972, "step": 1376 }, { "epoch": 0.37418478260869564, "grad_norm": 1.135533094406128, "learning_rate": 3e-05, "loss": 2.9638, "step": 1377 }, { "epoch": 0.3744565217391304, "grad_norm": 1.4869073629379272, "learning_rate": 3e-05, "loss": 3.7439, "step": 1378 }, { "epoch": 0.37472826086956523, "grad_norm": 1.5528923273086548, "learning_rate": 3e-05, "loss": 2.8187, "step": 1379 }, { "epoch": 0.375, "grad_norm": 1.0053822994232178, "learning_rate": 3e-05, "loss": 3.2647, "step": 1380 }, { "epoch": 0.37527173913043477, "grad_norm": 1.2217366695404053, "learning_rate": 3e-05, "loss": 3.5844, "step": 1381 }, { "epoch": 0.3755434782608696, "grad_norm": 1.245322823524475, "learning_rate": 3e-05, "loss": 2.9917, "step": 1382 }, { "epoch": 0.37581521739130436, "grad_norm": 1.3003109693527222, "learning_rate": 3e-05, "loss": 3.7163, "step": 1383 }, { "epoch": 0.3760869565217391, "grad_norm": 2.398679733276367, "learning_rate": 3e-05, "loss": 4.2985, "step": 1384 }, { "epoch": 0.3763586956521739, "grad_norm": 1.2174817323684692, "learning_rate": 3e-05, "loss": 3.3311, "step": 1385 }, { "epoch": 0.3766304347826087, "grad_norm": 1.2073396444320679, "learning_rate": 3e-05, "loss": 3.4313, "step": 1386 }, { "epoch": 0.3769021739130435, "grad_norm": 1.2685997486114502, "learning_rate": 3e-05, "loss": 3.1358, "step": 1387 }, { "epoch": 0.37717391304347825, "grad_norm": 1.3566850423812866, "learning_rate": 3e-05, "loss": 3.0956, "step": 1388 }, { "epoch": 0.37744565217391307, "grad_norm": 1.6564099788665771, "learning_rate": 3e-05, "loss": 3.7791, "step": 1389 }, { "epoch": 0.37771739130434784, "grad_norm": 1.3698561191558838, "learning_rate": 3e-05, "loss": 3.3446, "step": 1390 }, { "epoch": 0.3779891304347826, "grad_norm": 1.2702010869979858, "learning_rate": 3e-05, "loss": 4.0951, "step": 1391 }, { "epoch": 0.3782608695652174, "grad_norm": 1.4307619333267212, "learning_rate": 3e-05, "loss": 3.3821, "step": 1392 }, { "epoch": 0.3785326086956522, "grad_norm": 1.207396149635315, "learning_rate": 3e-05, "loss": 3.6105, "step": 1393 }, { "epoch": 0.37880434782608696, "grad_norm": 1.1929562091827393, "learning_rate": 3e-05, "loss": 2.8538, "step": 1394 }, { "epoch": 0.37907608695652173, "grad_norm": 1.2328208684921265, "learning_rate": 3e-05, "loss": 3.0622, "step": 1395 }, { "epoch": 0.3793478260869565, "grad_norm": 1.4275541305541992, "learning_rate": 3e-05, "loss": 3.4482, "step": 1396 }, { "epoch": 0.3796195652173913, "grad_norm": 0.9579852223396301, "learning_rate": 3e-05, "loss": 2.8646, "step": 1397 }, { "epoch": 0.3798913043478261, "grad_norm": 1.6573379039764404, "learning_rate": 3e-05, "loss": 3.4417, "step": 1398 }, { "epoch": 0.38016304347826085, "grad_norm": 1.5388156175613403, "learning_rate": 3e-05, "loss": 3.2041, "step": 1399 }, { "epoch": 0.3804347826086957, "grad_norm": 1.4631450176239014, "learning_rate": 3e-05, "loss": 3.8626, "step": 1400 }, { "epoch": 0.38070652173913044, "grad_norm": 1.3570584058761597, "learning_rate": 3e-05, "loss": 3.441, "step": 1401 }, { "epoch": 0.3809782608695652, "grad_norm": 1.8016875982284546, "learning_rate": 3e-05, "loss": 3.1616, "step": 1402 }, { "epoch": 0.38125, "grad_norm": 1.4279495477676392, "learning_rate": 3e-05, "loss": 3.6129, "step": 1403 }, { "epoch": 0.3815217391304348, "grad_norm": 1.3305083513259888, "learning_rate": 3e-05, "loss": 3.3956, "step": 1404 }, { "epoch": 0.38179347826086957, "grad_norm": 1.7289437055587769, "learning_rate": 3e-05, "loss": 3.6499, "step": 1405 }, { "epoch": 0.38206521739130433, "grad_norm": 1.75969398021698, "learning_rate": 3e-05, "loss": 3.104, "step": 1406 }, { "epoch": 0.38233695652173916, "grad_norm": 1.545464277267456, "learning_rate": 3e-05, "loss": 4.1157, "step": 1407 }, { "epoch": 0.3826086956521739, "grad_norm": 1.5299899578094482, "learning_rate": 3e-05, "loss": 2.9975, "step": 1408 }, { "epoch": 0.3828804347826087, "grad_norm": 1.7116879224777222, "learning_rate": 3e-05, "loss": 3.0605, "step": 1409 }, { "epoch": 0.38315217391304346, "grad_norm": 1.4728469848632812, "learning_rate": 3e-05, "loss": 3.2079, "step": 1410 }, { "epoch": 0.3834239130434783, "grad_norm": 2.216775894165039, "learning_rate": 3e-05, "loss": 3.8937, "step": 1411 }, { "epoch": 0.38369565217391305, "grad_norm": 1.6279774904251099, "learning_rate": 3e-05, "loss": 3.7957, "step": 1412 }, { "epoch": 0.3839673913043478, "grad_norm": 0.9390795230865479, "learning_rate": 3e-05, "loss": 2.8752, "step": 1413 }, { "epoch": 0.3842391304347826, "grad_norm": 1.3048573732376099, "learning_rate": 3e-05, "loss": 3.0132, "step": 1414 }, { "epoch": 0.3845108695652174, "grad_norm": 1.132607340812683, "learning_rate": 3e-05, "loss": 2.965, "step": 1415 }, { "epoch": 0.3847826086956522, "grad_norm": 1.6570241451263428, "learning_rate": 3e-05, "loss": 3.1474, "step": 1416 }, { "epoch": 0.38505434782608694, "grad_norm": 1.5501550436019897, "learning_rate": 3e-05, "loss": 3.4049, "step": 1417 }, { "epoch": 0.38532608695652176, "grad_norm": 1.1150649785995483, "learning_rate": 3e-05, "loss": 3.4114, "step": 1418 }, { "epoch": 0.38559782608695653, "grad_norm": 1.3886046409606934, "learning_rate": 3e-05, "loss": 3.7872, "step": 1419 }, { "epoch": 0.3858695652173913, "grad_norm": 1.178165078163147, "learning_rate": 3e-05, "loss": 3.1258, "step": 1420 }, { "epoch": 0.38614130434782606, "grad_norm": 1.2557275295257568, "learning_rate": 3e-05, "loss": 3.4292, "step": 1421 }, { "epoch": 0.3864130434782609, "grad_norm": 0.9806421995162964, "learning_rate": 3e-05, "loss": 2.8931, "step": 1422 }, { "epoch": 0.38668478260869565, "grad_norm": 1.6238960027694702, "learning_rate": 3e-05, "loss": 3.5711, "step": 1423 }, { "epoch": 0.3869565217391304, "grad_norm": 1.3578873872756958, "learning_rate": 3e-05, "loss": 2.9015, "step": 1424 }, { "epoch": 0.38722826086956524, "grad_norm": 1.2328130006790161, "learning_rate": 3e-05, "loss": 3.2193, "step": 1425 }, { "epoch": 0.3875, "grad_norm": 2.146702289581299, "learning_rate": 3e-05, "loss": 4.0931, "step": 1426 }, { "epoch": 0.3877717391304348, "grad_norm": 1.9415775537490845, "learning_rate": 3e-05, "loss": 3.0855, "step": 1427 }, { "epoch": 0.38804347826086955, "grad_norm": 1.7980471849441528, "learning_rate": 3e-05, "loss": 3.3307, "step": 1428 }, { "epoch": 0.38831521739130437, "grad_norm": 1.429101824760437, "learning_rate": 3e-05, "loss": 3.1628, "step": 1429 }, { "epoch": 0.38858695652173914, "grad_norm": 2.225107431411743, "learning_rate": 3e-05, "loss": 3.5771, "step": 1430 }, { "epoch": 0.3888586956521739, "grad_norm": 2.46109938621521, "learning_rate": 3e-05, "loss": 3.7408, "step": 1431 }, { "epoch": 0.38913043478260867, "grad_norm": 2.695655345916748, "learning_rate": 3e-05, "loss": 3.4776, "step": 1432 }, { "epoch": 0.3894021739130435, "grad_norm": 2.3022725582122803, "learning_rate": 3e-05, "loss": 3.4633, "step": 1433 }, { "epoch": 0.38967391304347826, "grad_norm": 1.5871728658676147, "learning_rate": 3e-05, "loss": 2.8911, "step": 1434 }, { "epoch": 0.389945652173913, "grad_norm": 1.3922131061553955, "learning_rate": 3e-05, "loss": 2.7701, "step": 1435 }, { "epoch": 0.39021739130434785, "grad_norm": 2.131774663925171, "learning_rate": 3e-05, "loss": 3.6811, "step": 1436 }, { "epoch": 0.3904891304347826, "grad_norm": 1.8943536281585693, "learning_rate": 3e-05, "loss": 2.6957, "step": 1437 }, { "epoch": 0.3907608695652174, "grad_norm": 1.7963645458221436, "learning_rate": 3e-05, "loss": 3.4574, "step": 1438 }, { "epoch": 0.39103260869565215, "grad_norm": 1.613253116607666, "learning_rate": 3e-05, "loss": 3.5328, "step": 1439 }, { "epoch": 0.391304347826087, "grad_norm": 1.8886051177978516, "learning_rate": 3e-05, "loss": 2.7514, "step": 1440 }, { "epoch": 0.39157608695652174, "grad_norm": 2.6512999534606934, "learning_rate": 3e-05, "loss": 3.9306, "step": 1441 }, { "epoch": 0.3918478260869565, "grad_norm": 1.6623197793960571, "learning_rate": 3e-05, "loss": 3.4983, "step": 1442 }, { "epoch": 0.39211956521739133, "grad_norm": 1.2289677858352661, "learning_rate": 3e-05, "loss": 3.1967, "step": 1443 }, { "epoch": 0.3923913043478261, "grad_norm": 0.9220126867294312, "learning_rate": 3e-05, "loss": 2.8288, "step": 1444 }, { "epoch": 0.39266304347826086, "grad_norm": 1.853948950767517, "learning_rate": 3e-05, "loss": 3.2876, "step": 1445 }, { "epoch": 0.39293478260869563, "grad_norm": 2.040001392364502, "learning_rate": 3e-05, "loss": 3.6805, "step": 1446 }, { "epoch": 0.39320652173913045, "grad_norm": 1.4386035203933716, "learning_rate": 3e-05, "loss": 3.1173, "step": 1447 }, { "epoch": 0.3934782608695652, "grad_norm": 1.48452889919281, "learning_rate": 3e-05, "loss": 3.3276, "step": 1448 }, { "epoch": 0.39375, "grad_norm": 1.058858871459961, "learning_rate": 3e-05, "loss": 2.5619, "step": 1449 }, { "epoch": 0.39402173913043476, "grad_norm": 2.131944179534912, "learning_rate": 3e-05, "loss": 3.7079, "step": 1450 }, { "epoch": 0.3942934782608696, "grad_norm": 1.5783636569976807, "learning_rate": 3e-05, "loss": 3.5856, "step": 1451 }, { "epoch": 0.39456521739130435, "grad_norm": 1.249036192893982, "learning_rate": 3e-05, "loss": 3.0789, "step": 1452 }, { "epoch": 0.3948369565217391, "grad_norm": 1.6562423706054688, "learning_rate": 3e-05, "loss": 3.9025, "step": 1453 }, { "epoch": 0.39510869565217394, "grad_norm": 1.6861488819122314, "learning_rate": 3e-05, "loss": 3.5286, "step": 1454 }, { "epoch": 0.3953804347826087, "grad_norm": 1.503812313079834, "learning_rate": 3e-05, "loss": 3.0909, "step": 1455 }, { "epoch": 0.39565217391304347, "grad_norm": 1.8524895906448364, "learning_rate": 3e-05, "loss": 3.4152, "step": 1456 }, { "epoch": 0.39592391304347824, "grad_norm": 1.3550132513046265, "learning_rate": 3e-05, "loss": 3.5444, "step": 1457 }, { "epoch": 0.39619565217391306, "grad_norm": 1.454416036605835, "learning_rate": 3e-05, "loss": 3.6969, "step": 1458 }, { "epoch": 0.3964673913043478, "grad_norm": 1.5491242408752441, "learning_rate": 3e-05, "loss": 3.4632, "step": 1459 }, { "epoch": 0.3967391304347826, "grad_norm": 1.177980661392212, "learning_rate": 3e-05, "loss": 3.2844, "step": 1460 }, { "epoch": 0.3970108695652174, "grad_norm": 1.2985824346542358, "learning_rate": 3e-05, "loss": 3.6743, "step": 1461 }, { "epoch": 0.3972826086956522, "grad_norm": 1.545882225036621, "learning_rate": 3e-05, "loss": 3.4353, "step": 1462 }, { "epoch": 0.39755434782608695, "grad_norm": 1.3580163717269897, "learning_rate": 3e-05, "loss": 3.7926, "step": 1463 }, { "epoch": 0.3978260869565217, "grad_norm": 1.4254475831985474, "learning_rate": 3e-05, "loss": 3.7358, "step": 1464 }, { "epoch": 0.39809782608695654, "grad_norm": 1.2237005233764648, "learning_rate": 3e-05, "loss": 3.3169, "step": 1465 }, { "epoch": 0.3983695652173913, "grad_norm": 1.1622488498687744, "learning_rate": 3e-05, "loss": 3.1261, "step": 1466 }, { "epoch": 0.3986413043478261, "grad_norm": 0.9447959065437317, "learning_rate": 3e-05, "loss": 2.8757, "step": 1467 }, { "epoch": 0.39891304347826084, "grad_norm": 1.5178827047348022, "learning_rate": 3e-05, "loss": 3.7686, "step": 1468 }, { "epoch": 0.39918478260869567, "grad_norm": 1.2081400156021118, "learning_rate": 3e-05, "loss": 2.7978, "step": 1469 }, { "epoch": 0.39945652173913043, "grad_norm": 1.3333414793014526, "learning_rate": 3e-05, "loss": 3.8005, "step": 1470 }, { "epoch": 0.3997282608695652, "grad_norm": 1.2284449338912964, "learning_rate": 3e-05, "loss": 3.2384, "step": 1471 }, { "epoch": 0.4, "grad_norm": 1.6261060237884521, "learning_rate": 3e-05, "loss": 3.6663, "step": 1472 }, { "epoch": 0.4002717391304348, "grad_norm": 1.2063846588134766, "learning_rate": 3e-05, "loss": 3.2808, "step": 1473 }, { "epoch": 0.40054347826086956, "grad_norm": 1.1668663024902344, "learning_rate": 3e-05, "loss": 3.6527, "step": 1474 }, { "epoch": 0.4008152173913043, "grad_norm": 1.1682161092758179, "learning_rate": 3e-05, "loss": 3.0487, "step": 1475 }, { "epoch": 0.40108695652173915, "grad_norm": 1.2479826211929321, "learning_rate": 3e-05, "loss": 3.8231, "step": 1476 }, { "epoch": 0.4013586956521739, "grad_norm": 1.3903083801269531, "learning_rate": 3e-05, "loss": 3.3589, "step": 1477 }, { "epoch": 0.4016304347826087, "grad_norm": 2.756995677947998, "learning_rate": 3e-05, "loss": 3.5956, "step": 1478 }, { "epoch": 0.4019021739130435, "grad_norm": 1.1125861406326294, "learning_rate": 3e-05, "loss": 2.9482, "step": 1479 }, { "epoch": 0.40217391304347827, "grad_norm": 1.7494776248931885, "learning_rate": 3e-05, "loss": 3.6373, "step": 1480 }, { "epoch": 0.40244565217391304, "grad_norm": 1.7383618354797363, "learning_rate": 3e-05, "loss": 3.5196, "step": 1481 }, { "epoch": 0.4027173913043478, "grad_norm": 2.0332143306732178, "learning_rate": 3e-05, "loss": 3.4659, "step": 1482 }, { "epoch": 0.4029891304347826, "grad_norm": 1.7549830675125122, "learning_rate": 3e-05, "loss": 3.6027, "step": 1483 }, { "epoch": 0.4032608695652174, "grad_norm": 1.4020575284957886, "learning_rate": 3e-05, "loss": 3.1451, "step": 1484 }, { "epoch": 0.40353260869565216, "grad_norm": 1.134787917137146, "learning_rate": 3e-05, "loss": 2.6124, "step": 1485 }, { "epoch": 0.40380434782608693, "grad_norm": 1.6569262742996216, "learning_rate": 3e-05, "loss": 3.8979, "step": 1486 }, { "epoch": 0.40407608695652175, "grad_norm": 1.7637388706207275, "learning_rate": 3e-05, "loss": 3.8109, "step": 1487 }, { "epoch": 0.4043478260869565, "grad_norm": 1.2442704439163208, "learning_rate": 3e-05, "loss": 3.6988, "step": 1488 }, { "epoch": 0.4046195652173913, "grad_norm": 1.0942063331604004, "learning_rate": 3e-05, "loss": 3.0464, "step": 1489 }, { "epoch": 0.4048913043478261, "grad_norm": 1.337386965751648, "learning_rate": 3e-05, "loss": 3.8127, "step": 1490 }, { "epoch": 0.4051630434782609, "grad_norm": 1.7307548522949219, "learning_rate": 3e-05, "loss": 3.3854, "step": 1491 }, { "epoch": 0.40543478260869564, "grad_norm": 3.1193931102752686, "learning_rate": 3e-05, "loss": 4.9053, "step": 1492 }, { "epoch": 0.4057065217391304, "grad_norm": 1.5950829982757568, "learning_rate": 3e-05, "loss": 3.6687, "step": 1493 }, { "epoch": 0.40597826086956523, "grad_norm": 1.6609991788864136, "learning_rate": 3e-05, "loss": 3.4985, "step": 1494 }, { "epoch": 0.40625, "grad_norm": 1.824634075164795, "learning_rate": 3e-05, "loss": 2.7576, "step": 1495 }, { "epoch": 0.40652173913043477, "grad_norm": 1.8175928592681885, "learning_rate": 3e-05, "loss": 3.7615, "step": 1496 }, { "epoch": 0.4067934782608696, "grad_norm": 1.3160499334335327, "learning_rate": 3e-05, "loss": 3.1647, "step": 1497 }, { "epoch": 0.40706521739130436, "grad_norm": 1.7744853496551514, "learning_rate": 3e-05, "loss": 3.0396, "step": 1498 }, { "epoch": 0.4073369565217391, "grad_norm": 2.0241026878356934, "learning_rate": 3e-05, "loss": 3.8311, "step": 1499 }, { "epoch": 0.4076086956521739, "grad_norm": 1.7789679765701294, "learning_rate": 3e-05, "loss": 2.9538, "step": 1500 }, { "epoch": 0.4078804347826087, "grad_norm": 1.4469610452651978, "learning_rate": 3e-05, "loss": 3.3083, "step": 1501 }, { "epoch": 0.4081521739130435, "grad_norm": 1.570556879043579, "learning_rate": 3e-05, "loss": 3.0793, "step": 1502 }, { "epoch": 0.40842391304347825, "grad_norm": 1.3964000940322876, "learning_rate": 3e-05, "loss": 2.9603, "step": 1503 }, { "epoch": 0.40869565217391307, "grad_norm": 1.1800280809402466, "learning_rate": 3e-05, "loss": 3.2305, "step": 1504 }, { "epoch": 0.40896739130434784, "grad_norm": 1.762847661972046, "learning_rate": 3e-05, "loss": 3.7564, "step": 1505 }, { "epoch": 0.4092391304347826, "grad_norm": 1.5233442783355713, "learning_rate": 3e-05, "loss": 3.2744, "step": 1506 }, { "epoch": 0.4095108695652174, "grad_norm": 1.9167475700378418, "learning_rate": 3e-05, "loss": 3.6962, "step": 1507 }, { "epoch": 0.4097826086956522, "grad_norm": 1.949634075164795, "learning_rate": 3e-05, "loss": 3.1084, "step": 1508 }, { "epoch": 0.41005434782608696, "grad_norm": 1.9483072757720947, "learning_rate": 3e-05, "loss": 3.1895, "step": 1509 }, { "epoch": 0.41032608695652173, "grad_norm": 1.368509292602539, "learning_rate": 3e-05, "loss": 2.6804, "step": 1510 }, { "epoch": 0.4105978260869565, "grad_norm": 1.6476247310638428, "learning_rate": 3e-05, "loss": 3.1565, "step": 1511 }, { "epoch": 0.4108695652173913, "grad_norm": 2.14841890335083, "learning_rate": 3e-05, "loss": 3.5723, "step": 1512 }, { "epoch": 0.4111413043478261, "grad_norm": 2.1899356842041016, "learning_rate": 3e-05, "loss": 3.9324, "step": 1513 }, { "epoch": 0.41141304347826085, "grad_norm": 1.971421480178833, "learning_rate": 3e-05, "loss": 3.3302, "step": 1514 }, { "epoch": 0.4116847826086957, "grad_norm": 1.555216908454895, "learning_rate": 3e-05, "loss": 2.8559, "step": 1515 }, { "epoch": 0.41195652173913044, "grad_norm": 2.025975227355957, "learning_rate": 3e-05, "loss": 3.8605, "step": 1516 }, { "epoch": 0.4122282608695652, "grad_norm": 1.609168291091919, "learning_rate": 3e-05, "loss": 3.521, "step": 1517 }, { "epoch": 0.4125, "grad_norm": 1.4011493921279907, "learning_rate": 3e-05, "loss": 2.905, "step": 1518 }, { "epoch": 0.4127717391304348, "grad_norm": 1.813499927520752, "learning_rate": 3e-05, "loss": 3.9861, "step": 1519 }, { "epoch": 0.41304347826086957, "grad_norm": 2.023961067199707, "learning_rate": 3e-05, "loss": 3.3814, "step": 1520 }, { "epoch": 0.41331521739130433, "grad_norm": 1.9356021881103516, "learning_rate": 3e-05, "loss": 2.9743, "step": 1521 }, { "epoch": 0.41358695652173916, "grad_norm": 1.6765419244766235, "learning_rate": 3e-05, "loss": 2.9504, "step": 1522 }, { "epoch": 0.4138586956521739, "grad_norm": 1.954119324684143, "learning_rate": 3e-05, "loss": 3.7254, "step": 1523 }, { "epoch": 0.4141304347826087, "grad_norm": 1.5637805461883545, "learning_rate": 3e-05, "loss": 3.8543, "step": 1524 }, { "epoch": 0.41440217391304346, "grad_norm": 1.175187587738037, "learning_rate": 3e-05, "loss": 3.2871, "step": 1525 }, { "epoch": 0.4146739130434783, "grad_norm": 1.3228040933609009, "learning_rate": 3e-05, "loss": 2.9662, "step": 1526 }, { "epoch": 0.41494565217391305, "grad_norm": 1.4776461124420166, "learning_rate": 3e-05, "loss": 2.9617, "step": 1527 }, { "epoch": 0.4152173913043478, "grad_norm": 1.5169886350631714, "learning_rate": 3e-05, "loss": 3.6288, "step": 1528 }, { "epoch": 0.4154891304347826, "grad_norm": 1.4785473346710205, "learning_rate": 3e-05, "loss": 3.0807, "step": 1529 }, { "epoch": 0.4157608695652174, "grad_norm": 1.4565258026123047, "learning_rate": 3e-05, "loss": 3.3099, "step": 1530 }, { "epoch": 0.4160326086956522, "grad_norm": 1.2732715606689453, "learning_rate": 3e-05, "loss": 3.2948, "step": 1531 }, { "epoch": 0.41630434782608694, "grad_norm": 1.0217174291610718, "learning_rate": 3e-05, "loss": 3.0084, "step": 1532 }, { "epoch": 0.41657608695652176, "grad_norm": 1.2822792530059814, "learning_rate": 3e-05, "loss": 2.9284, "step": 1533 }, { "epoch": 0.41684782608695653, "grad_norm": 1.641827940940857, "learning_rate": 3e-05, "loss": 3.4684, "step": 1534 }, { "epoch": 0.4171195652173913, "grad_norm": 1.5296562910079956, "learning_rate": 3e-05, "loss": 2.5836, "step": 1535 }, { "epoch": 0.41739130434782606, "grad_norm": 1.3570102453231812, "learning_rate": 3e-05, "loss": 3.1323, "step": 1536 }, { "epoch": 0.4176630434782609, "grad_norm": 0.9901162981987, "learning_rate": 3e-05, "loss": 2.7749, "step": 1537 }, { "epoch": 0.41793478260869565, "grad_norm": 1.2664462327957153, "learning_rate": 3e-05, "loss": 3.1213, "step": 1538 }, { "epoch": 0.4182065217391304, "grad_norm": 1.682426929473877, "learning_rate": 3e-05, "loss": 3.8292, "step": 1539 }, { "epoch": 0.41847826086956524, "grad_norm": 1.3441531658172607, "learning_rate": 3e-05, "loss": 2.8989, "step": 1540 }, { "epoch": 0.41875, "grad_norm": 1.6413555145263672, "learning_rate": 3e-05, "loss": 3.4333, "step": 1541 }, { "epoch": 0.4190217391304348, "grad_norm": 1.2601666450500488, "learning_rate": 3e-05, "loss": 3.2815, "step": 1542 }, { "epoch": 0.41929347826086955, "grad_norm": 1.5582549571990967, "learning_rate": 3e-05, "loss": 3.5814, "step": 1543 }, { "epoch": 0.41956521739130437, "grad_norm": 1.0220012664794922, "learning_rate": 3e-05, "loss": 2.7242, "step": 1544 }, { "epoch": 0.41983695652173914, "grad_norm": 1.710802674293518, "learning_rate": 3e-05, "loss": 3.5271, "step": 1545 }, { "epoch": 0.4201086956521739, "grad_norm": 2.3148810863494873, "learning_rate": 3e-05, "loss": 4.0931, "step": 1546 }, { "epoch": 0.42038043478260867, "grad_norm": 1.4409066438674927, "learning_rate": 3e-05, "loss": 3.4806, "step": 1547 }, { "epoch": 0.4206521739130435, "grad_norm": 1.4869251251220703, "learning_rate": 3e-05, "loss": 3.8039, "step": 1548 }, { "epoch": 0.42092391304347826, "grad_norm": 2.2359511852264404, "learning_rate": 3e-05, "loss": 4.1842, "step": 1549 }, { "epoch": 0.421195652173913, "grad_norm": 2.2962846755981445, "learning_rate": 3e-05, "loss": 3.8559, "step": 1550 }, { "epoch": 0.42146739130434785, "grad_norm": 1.5050580501556396, "learning_rate": 3e-05, "loss": 3.6678, "step": 1551 }, { "epoch": 0.4217391304347826, "grad_norm": 1.2851321697235107, "learning_rate": 3e-05, "loss": 3.4175, "step": 1552 }, { "epoch": 0.4220108695652174, "grad_norm": 1.2671363353729248, "learning_rate": 3e-05, "loss": 3.0412, "step": 1553 }, { "epoch": 0.42228260869565215, "grad_norm": 1.2936915159225464, "learning_rate": 3e-05, "loss": 3.0918, "step": 1554 }, { "epoch": 0.422554347826087, "grad_norm": 1.4296749830245972, "learning_rate": 3e-05, "loss": 3.3257, "step": 1555 }, { "epoch": 0.42282608695652174, "grad_norm": 1.2652101516723633, "learning_rate": 3e-05, "loss": 3.5451, "step": 1556 }, { "epoch": 0.4230978260869565, "grad_norm": 1.4265552759170532, "learning_rate": 3e-05, "loss": 3.4145, "step": 1557 }, { "epoch": 0.42336956521739133, "grad_norm": 1.3775297403335571, "learning_rate": 3e-05, "loss": 3.3248, "step": 1558 }, { "epoch": 0.4236413043478261, "grad_norm": 1.8461849689483643, "learning_rate": 3e-05, "loss": 3.9212, "step": 1559 }, { "epoch": 0.42391304347826086, "grad_norm": 1.412374496459961, "learning_rate": 3e-05, "loss": 3.4737, "step": 1560 }, { "epoch": 0.42418478260869563, "grad_norm": 1.6272127628326416, "learning_rate": 3e-05, "loss": 3.4379, "step": 1561 }, { "epoch": 0.42445652173913045, "grad_norm": 1.754745364189148, "learning_rate": 3e-05, "loss": 3.4164, "step": 1562 }, { "epoch": 0.4247282608695652, "grad_norm": 1.2676855325698853, "learning_rate": 3e-05, "loss": 3.3106, "step": 1563 }, { "epoch": 0.425, "grad_norm": 1.7627284526824951, "learning_rate": 3e-05, "loss": 3.841, "step": 1564 }, { "epoch": 0.42527173913043476, "grad_norm": 1.143531084060669, "learning_rate": 3e-05, "loss": 2.82, "step": 1565 }, { "epoch": 0.4255434782608696, "grad_norm": 1.0254485607147217, "learning_rate": 3e-05, "loss": 2.9556, "step": 1566 }, { "epoch": 0.42581521739130435, "grad_norm": 1.3668675422668457, "learning_rate": 3e-05, "loss": 3.7105, "step": 1567 }, { "epoch": 0.4260869565217391, "grad_norm": 2.208592176437378, "learning_rate": 3e-05, "loss": 3.4152, "step": 1568 }, { "epoch": 0.42635869565217394, "grad_norm": 1.1941901445388794, "learning_rate": 3e-05, "loss": 3.3036, "step": 1569 }, { "epoch": 0.4266304347826087, "grad_norm": 1.536787748336792, "learning_rate": 3e-05, "loss": 3.5194, "step": 1570 }, { "epoch": 0.42690217391304347, "grad_norm": 1.4054354429244995, "learning_rate": 3e-05, "loss": 3.263, "step": 1571 }, { "epoch": 0.42717391304347824, "grad_norm": 1.6471023559570312, "learning_rate": 3e-05, "loss": 3.4855, "step": 1572 }, { "epoch": 0.42744565217391306, "grad_norm": 0.9407289028167725, "learning_rate": 3e-05, "loss": 3.1382, "step": 1573 }, { "epoch": 0.4277173913043478, "grad_norm": 1.244971513748169, "learning_rate": 3e-05, "loss": 3.4904, "step": 1574 }, { "epoch": 0.4279891304347826, "grad_norm": 1.305245041847229, "learning_rate": 3e-05, "loss": 3.3049, "step": 1575 }, { "epoch": 0.4282608695652174, "grad_norm": 1.2542614936828613, "learning_rate": 3e-05, "loss": 3.4931, "step": 1576 }, { "epoch": 0.4285326086956522, "grad_norm": 0.9759632349014282, "learning_rate": 3e-05, "loss": 3.4818, "step": 1577 }, { "epoch": 0.42880434782608695, "grad_norm": 1.4484260082244873, "learning_rate": 3e-05, "loss": 3.2934, "step": 1578 }, { "epoch": 0.4290760869565217, "grad_norm": 1.1898603439331055, "learning_rate": 3e-05, "loss": 3.4601, "step": 1579 }, { "epoch": 0.42934782608695654, "grad_norm": 1.486471176147461, "learning_rate": 3e-05, "loss": 3.762, "step": 1580 }, { "epoch": 0.4296195652173913, "grad_norm": 1.2455788850784302, "learning_rate": 3e-05, "loss": 3.721, "step": 1581 }, { "epoch": 0.4298913043478261, "grad_norm": 1.1194945573806763, "learning_rate": 3e-05, "loss": 3.3573, "step": 1582 }, { "epoch": 0.43016304347826084, "grad_norm": 1.666195273399353, "learning_rate": 3e-05, "loss": 3.531, "step": 1583 }, { "epoch": 0.43043478260869567, "grad_norm": 1.192292332649231, "learning_rate": 3e-05, "loss": 3.1313, "step": 1584 }, { "epoch": 0.43070652173913043, "grad_norm": 1.1080567836761475, "learning_rate": 3e-05, "loss": 3.2448, "step": 1585 }, { "epoch": 0.4309782608695652, "grad_norm": 1.7732088565826416, "learning_rate": 3e-05, "loss": 3.4527, "step": 1586 }, { "epoch": 0.43125, "grad_norm": 1.2465471029281616, "learning_rate": 3e-05, "loss": 2.8915, "step": 1587 }, { "epoch": 0.4315217391304348, "grad_norm": 1.43854820728302, "learning_rate": 3e-05, "loss": 3.2366, "step": 1588 }, { "epoch": 0.43179347826086956, "grad_norm": 1.3088682889938354, "learning_rate": 3e-05, "loss": 3.3654, "step": 1589 }, { "epoch": 0.4320652173913043, "grad_norm": 1.0148249864578247, "learning_rate": 3e-05, "loss": 3.083, "step": 1590 }, { "epoch": 0.43233695652173915, "grad_norm": 1.2219303846359253, "learning_rate": 3e-05, "loss": 3.0298, "step": 1591 }, { "epoch": 0.4326086956521739, "grad_norm": 1.081207275390625, "learning_rate": 3e-05, "loss": 3.2116, "step": 1592 }, { "epoch": 0.4328804347826087, "grad_norm": 1.515478491783142, "learning_rate": 3e-05, "loss": 3.4368, "step": 1593 }, { "epoch": 0.4331521739130435, "grad_norm": 1.073887586593628, "learning_rate": 3e-05, "loss": 3.1832, "step": 1594 }, { "epoch": 0.43342391304347827, "grad_norm": 1.1452287435531616, "learning_rate": 3e-05, "loss": 3.5617, "step": 1595 }, { "epoch": 0.43369565217391304, "grad_norm": 1.512242078781128, "learning_rate": 3e-05, "loss": 3.4468, "step": 1596 }, { "epoch": 0.4339673913043478, "grad_norm": 1.1862598657608032, "learning_rate": 3e-05, "loss": 3.3795, "step": 1597 }, { "epoch": 0.4342391304347826, "grad_norm": 1.4696996212005615, "learning_rate": 3e-05, "loss": 3.5239, "step": 1598 }, { "epoch": 0.4345108695652174, "grad_norm": 1.5985097885131836, "learning_rate": 3e-05, "loss": 3.8709, "step": 1599 }, { "epoch": 0.43478260869565216, "grad_norm": 1.3134711980819702, "learning_rate": 3e-05, "loss": 3.4582, "step": 1600 }, { "epoch": 0.43505434782608693, "grad_norm": 1.2656311988830566, "learning_rate": 3e-05, "loss": 3.7025, "step": 1601 }, { "epoch": 0.43532608695652175, "grad_norm": 1.1780740022659302, "learning_rate": 3e-05, "loss": 3.3486, "step": 1602 }, { "epoch": 0.4355978260869565, "grad_norm": 1.2048178911209106, "learning_rate": 3e-05, "loss": 3.3115, "step": 1603 }, { "epoch": 0.4358695652173913, "grad_norm": 1.3589476346969604, "learning_rate": 3e-05, "loss": 3.2494, "step": 1604 }, { "epoch": 0.4361413043478261, "grad_norm": 1.3171757459640503, "learning_rate": 3e-05, "loss": 2.8635, "step": 1605 }, { "epoch": 0.4364130434782609, "grad_norm": 1.1358094215393066, "learning_rate": 3e-05, "loss": 2.9426, "step": 1606 }, { "epoch": 0.43668478260869564, "grad_norm": 1.376583456993103, "learning_rate": 3e-05, "loss": 3.2611, "step": 1607 }, { "epoch": 0.4369565217391304, "grad_norm": 1.7452890872955322, "learning_rate": 3e-05, "loss": 3.7445, "step": 1608 }, { "epoch": 0.43722826086956523, "grad_norm": 1.715166687965393, "learning_rate": 3e-05, "loss": 3.4322, "step": 1609 }, { "epoch": 0.4375, "grad_norm": 1.3614475727081299, "learning_rate": 3e-05, "loss": 3.5131, "step": 1610 }, { "epoch": 0.43777173913043477, "grad_norm": 1.4811841249465942, "learning_rate": 3e-05, "loss": 3.714, "step": 1611 }, { "epoch": 0.4380434782608696, "grad_norm": 1.071926236152649, "learning_rate": 3e-05, "loss": 3.0734, "step": 1612 }, { "epoch": 0.43831521739130436, "grad_norm": 1.5618953704833984, "learning_rate": 3e-05, "loss": 3.876, "step": 1613 }, { "epoch": 0.4385869565217391, "grad_norm": 1.1414976119995117, "learning_rate": 3e-05, "loss": 3.5318, "step": 1614 }, { "epoch": 0.4388586956521739, "grad_norm": 1.0475927591323853, "learning_rate": 3e-05, "loss": 3.293, "step": 1615 }, { "epoch": 0.4391304347826087, "grad_norm": 1.392609715461731, "learning_rate": 3e-05, "loss": 2.9974, "step": 1616 }, { "epoch": 0.4394021739130435, "grad_norm": 0.906802237033844, "learning_rate": 3e-05, "loss": 2.9825, "step": 1617 }, { "epoch": 0.43967391304347825, "grad_norm": 1.4019746780395508, "learning_rate": 3e-05, "loss": 3.8698, "step": 1618 }, { "epoch": 0.43994565217391307, "grad_norm": 1.321829915046692, "learning_rate": 3e-05, "loss": 3.7109, "step": 1619 }, { "epoch": 0.44021739130434784, "grad_norm": 1.0838751792907715, "learning_rate": 3e-05, "loss": 2.6205, "step": 1620 }, { "epoch": 0.4404891304347826, "grad_norm": 2.383323907852173, "learning_rate": 3e-05, "loss": 4.2267, "step": 1621 }, { "epoch": 0.4407608695652174, "grad_norm": 1.0130791664123535, "learning_rate": 3e-05, "loss": 2.788, "step": 1622 }, { "epoch": 0.4410326086956522, "grad_norm": 1.3985093832015991, "learning_rate": 3e-05, "loss": 3.1002, "step": 1623 }, { "epoch": 0.44130434782608696, "grad_norm": 1.3819383382797241, "learning_rate": 3e-05, "loss": 3.2219, "step": 1624 }, { "epoch": 0.44157608695652173, "grad_norm": 1.4096386432647705, "learning_rate": 3e-05, "loss": 3.6886, "step": 1625 }, { "epoch": 0.4418478260869565, "grad_norm": 1.4819231033325195, "learning_rate": 3e-05, "loss": 3.2191, "step": 1626 }, { "epoch": 0.4421195652173913, "grad_norm": 1.3344846963882446, "learning_rate": 3e-05, "loss": 3.3927, "step": 1627 }, { "epoch": 0.4423913043478261, "grad_norm": 1.469277024269104, "learning_rate": 3e-05, "loss": 3.4069, "step": 1628 }, { "epoch": 0.44266304347826085, "grad_norm": 1.0877147912979126, "learning_rate": 3e-05, "loss": 2.9891, "step": 1629 }, { "epoch": 0.4429347826086957, "grad_norm": 1.1775012016296387, "learning_rate": 3e-05, "loss": 3.3463, "step": 1630 }, { "epoch": 0.44320652173913044, "grad_norm": 1.240936517715454, "learning_rate": 3e-05, "loss": 3.5751, "step": 1631 }, { "epoch": 0.4434782608695652, "grad_norm": 0.877865195274353, "learning_rate": 3e-05, "loss": 2.7317, "step": 1632 }, { "epoch": 0.44375, "grad_norm": 1.1812602281570435, "learning_rate": 3e-05, "loss": 3.0307, "step": 1633 }, { "epoch": 0.4440217391304348, "grad_norm": 1.7113842964172363, "learning_rate": 3e-05, "loss": 4.1471, "step": 1634 }, { "epoch": 0.44429347826086957, "grad_norm": 1.1820313930511475, "learning_rate": 3e-05, "loss": 3.362, "step": 1635 }, { "epoch": 0.44456521739130433, "grad_norm": 1.3378487825393677, "learning_rate": 3e-05, "loss": 3.5815, "step": 1636 }, { "epoch": 0.44483695652173916, "grad_norm": 0.8756458163261414, "learning_rate": 3e-05, "loss": 2.6882, "step": 1637 }, { "epoch": 0.4451086956521739, "grad_norm": 1.020820140838623, "learning_rate": 3e-05, "loss": 3.1006, "step": 1638 }, { "epoch": 0.4453804347826087, "grad_norm": 1.5075011253356934, "learning_rate": 3e-05, "loss": 4.0147, "step": 1639 }, { "epoch": 0.44565217391304346, "grad_norm": 1.1874289512634277, "learning_rate": 3e-05, "loss": 3.0811, "step": 1640 }, { "epoch": 0.4459239130434783, "grad_norm": 1.5765689611434937, "learning_rate": 3e-05, "loss": 3.4697, "step": 1641 }, { "epoch": 0.44619565217391305, "grad_norm": 1.3812084197998047, "learning_rate": 3e-05, "loss": 3.6468, "step": 1642 }, { "epoch": 0.4464673913043478, "grad_norm": 1.1477484703063965, "learning_rate": 3e-05, "loss": 3.2347, "step": 1643 }, { "epoch": 0.4467391304347826, "grad_norm": 1.431220531463623, "learning_rate": 3e-05, "loss": 3.8073, "step": 1644 }, { "epoch": 0.4470108695652174, "grad_norm": 1.0271748304367065, "learning_rate": 3e-05, "loss": 3.2165, "step": 1645 }, { "epoch": 0.4472826086956522, "grad_norm": 1.2913216352462769, "learning_rate": 3e-05, "loss": 3.2766, "step": 1646 }, { "epoch": 0.44755434782608694, "grad_norm": 1.341670274734497, "learning_rate": 3e-05, "loss": 3.8426, "step": 1647 }, { "epoch": 0.44782608695652176, "grad_norm": 1.1871728897094727, "learning_rate": 3e-05, "loss": 3.1251, "step": 1648 }, { "epoch": 0.44809782608695653, "grad_norm": 1.2151528596878052, "learning_rate": 3e-05, "loss": 3.3811, "step": 1649 }, { "epoch": 0.4483695652173913, "grad_norm": 1.7120990753173828, "learning_rate": 3e-05, "loss": 3.7093, "step": 1650 }, { "epoch": 0.44864130434782606, "grad_norm": 1.1526025533676147, "learning_rate": 3e-05, "loss": 3.1806, "step": 1651 }, { "epoch": 0.4489130434782609, "grad_norm": 1.5061256885528564, "learning_rate": 3e-05, "loss": 3.7246, "step": 1652 }, { "epoch": 0.44918478260869565, "grad_norm": 1.442798376083374, "learning_rate": 3e-05, "loss": 2.8975, "step": 1653 }, { "epoch": 0.4494565217391304, "grad_norm": 1.5733429193496704, "learning_rate": 3e-05, "loss": 3.7047, "step": 1654 }, { "epoch": 0.44972826086956524, "grad_norm": 1.5891181230545044, "learning_rate": 3e-05, "loss": 3.5587, "step": 1655 }, { "epoch": 0.45, "grad_norm": 2.2529733180999756, "learning_rate": 3e-05, "loss": 3.079, "step": 1656 }, { "epoch": 0.4502717391304348, "grad_norm": 2.397771120071411, "learning_rate": 3e-05, "loss": 3.2517, "step": 1657 }, { "epoch": 0.45054347826086955, "grad_norm": 1.5895618200302124, "learning_rate": 3e-05, "loss": 3.289, "step": 1658 }, { "epoch": 0.45081521739130437, "grad_norm": 1.3885931968688965, "learning_rate": 3e-05, "loss": 3.8401, "step": 1659 }, { "epoch": 0.45108695652173914, "grad_norm": 1.848649263381958, "learning_rate": 3e-05, "loss": 3.2375, "step": 1660 }, { "epoch": 0.4513586956521739, "grad_norm": 2.288536787033081, "learning_rate": 3e-05, "loss": 3.0737, "step": 1661 }, { "epoch": 0.45163043478260867, "grad_norm": 1.5626355409622192, "learning_rate": 3e-05, "loss": 3.6127, "step": 1662 }, { "epoch": 0.4519021739130435, "grad_norm": 1.2159314155578613, "learning_rate": 3e-05, "loss": 3.1681, "step": 1663 }, { "epoch": 0.45217391304347826, "grad_norm": 2.7480335235595703, "learning_rate": 3e-05, "loss": 3.3753, "step": 1664 }, { "epoch": 0.452445652173913, "grad_norm": 1.5772864818572998, "learning_rate": 3e-05, "loss": 3.4648, "step": 1665 }, { "epoch": 0.45271739130434785, "grad_norm": 2.1393795013427734, "learning_rate": 3e-05, "loss": 4.067, "step": 1666 }, { "epoch": 0.4529891304347826, "grad_norm": 1.48007071018219, "learning_rate": 3e-05, "loss": 3.0416, "step": 1667 }, { "epoch": 0.4532608695652174, "grad_norm": 1.3958046436309814, "learning_rate": 3e-05, "loss": 2.9871, "step": 1668 }, { "epoch": 0.45353260869565215, "grad_norm": 1.6530804634094238, "learning_rate": 3e-05, "loss": 3.1665, "step": 1669 }, { "epoch": 0.453804347826087, "grad_norm": 1.5180555582046509, "learning_rate": 3e-05, "loss": 3.1545, "step": 1670 }, { "epoch": 0.45407608695652174, "grad_norm": 1.3946945667266846, "learning_rate": 3e-05, "loss": 3.0606, "step": 1671 }, { "epoch": 0.4543478260869565, "grad_norm": 1.655504584312439, "learning_rate": 3e-05, "loss": 3.5141, "step": 1672 }, { "epoch": 0.45461956521739133, "grad_norm": 1.8746256828308105, "learning_rate": 3e-05, "loss": 3.2537, "step": 1673 }, { "epoch": 0.4548913043478261, "grad_norm": 1.6152852773666382, "learning_rate": 3e-05, "loss": 3.1725, "step": 1674 }, { "epoch": 0.45516304347826086, "grad_norm": 2.030758857727051, "learning_rate": 3e-05, "loss": 4.1911, "step": 1675 }, { "epoch": 0.45543478260869563, "grad_norm": 1.573011040687561, "learning_rate": 3e-05, "loss": 2.7004, "step": 1676 }, { "epoch": 0.45570652173913045, "grad_norm": 1.7782936096191406, "learning_rate": 3e-05, "loss": 3.4129, "step": 1677 }, { "epoch": 0.4559782608695652, "grad_norm": 1.6364409923553467, "learning_rate": 3e-05, "loss": 3.7828, "step": 1678 }, { "epoch": 0.45625, "grad_norm": 1.095005750656128, "learning_rate": 3e-05, "loss": 2.9575, "step": 1679 }, { "epoch": 0.45652173913043476, "grad_norm": 1.4323863983154297, "learning_rate": 3e-05, "loss": 2.9855, "step": 1680 }, { "epoch": 0.4567934782608696, "grad_norm": 1.8281795978546143, "learning_rate": 3e-05, "loss": 3.2274, "step": 1681 }, { "epoch": 0.45706521739130435, "grad_norm": 1.553568959236145, "learning_rate": 3e-05, "loss": 3.4245, "step": 1682 }, { "epoch": 0.4573369565217391, "grad_norm": 1.229791522026062, "learning_rate": 3e-05, "loss": 2.8749, "step": 1683 }, { "epoch": 0.45760869565217394, "grad_norm": 1.4264166355133057, "learning_rate": 3e-05, "loss": 3.0987, "step": 1684 }, { "epoch": 0.4578804347826087, "grad_norm": 1.4009737968444824, "learning_rate": 3e-05, "loss": 3.2024, "step": 1685 }, { "epoch": 0.45815217391304347, "grad_norm": 1.2461766004562378, "learning_rate": 3e-05, "loss": 3.1414, "step": 1686 }, { "epoch": 0.45842391304347824, "grad_norm": 1.1765496730804443, "learning_rate": 3e-05, "loss": 3.7237, "step": 1687 }, { "epoch": 0.45869565217391306, "grad_norm": 1.2187762260437012, "learning_rate": 3e-05, "loss": 3.5997, "step": 1688 }, { "epoch": 0.4589673913043478, "grad_norm": 2.117086410522461, "learning_rate": 3e-05, "loss": 3.5497, "step": 1689 }, { "epoch": 0.4592391304347826, "grad_norm": 1.2980027198791504, "learning_rate": 3e-05, "loss": 3.6001, "step": 1690 }, { "epoch": 0.4595108695652174, "grad_norm": 1.1031181812286377, "learning_rate": 3e-05, "loss": 2.814, "step": 1691 }, { "epoch": 0.4597826086956522, "grad_norm": 1.201127052307129, "learning_rate": 3e-05, "loss": 3.4627, "step": 1692 }, { "epoch": 0.46005434782608695, "grad_norm": 1.9975850582122803, "learning_rate": 3e-05, "loss": 3.8157, "step": 1693 }, { "epoch": 0.4603260869565217, "grad_norm": 1.4874250888824463, "learning_rate": 3e-05, "loss": 3.5901, "step": 1694 }, { "epoch": 0.46059782608695654, "grad_norm": 1.9747381210327148, "learning_rate": 3e-05, "loss": 3.8642, "step": 1695 }, { "epoch": 0.4608695652173913, "grad_norm": 1.405049443244934, "learning_rate": 3e-05, "loss": 3.3367, "step": 1696 }, { "epoch": 0.4611413043478261, "grad_norm": 1.4599591493606567, "learning_rate": 3e-05, "loss": 3.6499, "step": 1697 }, { "epoch": 0.46141304347826084, "grad_norm": 1.6565916538238525, "learning_rate": 3e-05, "loss": 3.5368, "step": 1698 }, { "epoch": 0.46168478260869567, "grad_norm": 1.327285647392273, "learning_rate": 3e-05, "loss": 2.7457, "step": 1699 }, { "epoch": 0.46195652173913043, "grad_norm": 1.317681074142456, "learning_rate": 3e-05, "loss": 3.3931, "step": 1700 }, { "epoch": 0.4622282608695652, "grad_norm": 1.7210628986358643, "learning_rate": 3e-05, "loss": 4.0996, "step": 1701 }, { "epoch": 0.4625, "grad_norm": 1.4575893878936768, "learning_rate": 3e-05, "loss": 3.5176, "step": 1702 }, { "epoch": 0.4627717391304348, "grad_norm": 0.8817341327667236, "learning_rate": 3e-05, "loss": 2.9138, "step": 1703 }, { "epoch": 0.46304347826086956, "grad_norm": 1.6268318891525269, "learning_rate": 3e-05, "loss": 3.3346, "step": 1704 }, { "epoch": 0.4633152173913043, "grad_norm": 1.281497836112976, "learning_rate": 3e-05, "loss": 2.776, "step": 1705 }, { "epoch": 0.46358695652173915, "grad_norm": 1.6672241687774658, "learning_rate": 3e-05, "loss": 3.5205, "step": 1706 }, { "epoch": 0.4638586956521739, "grad_norm": 1.4224839210510254, "learning_rate": 3e-05, "loss": 3.7986, "step": 1707 }, { "epoch": 0.4641304347826087, "grad_norm": 1.1987775564193726, "learning_rate": 3e-05, "loss": 3.2364, "step": 1708 }, { "epoch": 0.4644021739130435, "grad_norm": 1.4384889602661133, "learning_rate": 3e-05, "loss": 3.6339, "step": 1709 }, { "epoch": 0.46467391304347827, "grad_norm": 1.1359198093414307, "learning_rate": 3e-05, "loss": 3.2827, "step": 1710 }, { "epoch": 0.46494565217391304, "grad_norm": 1.3200854063034058, "learning_rate": 3e-05, "loss": 3.2209, "step": 1711 }, { "epoch": 0.4652173913043478, "grad_norm": 1.613688588142395, "learning_rate": 3e-05, "loss": 3.7578, "step": 1712 }, { "epoch": 0.4654891304347826, "grad_norm": 1.7866336107254028, "learning_rate": 3e-05, "loss": 3.9267, "step": 1713 }, { "epoch": 0.4657608695652174, "grad_norm": 2.0430612564086914, "learning_rate": 3e-05, "loss": 3.6663, "step": 1714 }, { "epoch": 0.46603260869565216, "grad_norm": 1.848181128501892, "learning_rate": 3e-05, "loss": 3.3196, "step": 1715 }, { "epoch": 0.46630434782608693, "grad_norm": 1.4686863422393799, "learning_rate": 3e-05, "loss": 3.352, "step": 1716 }, { "epoch": 0.46657608695652175, "grad_norm": 1.1621712446212769, "learning_rate": 3e-05, "loss": 3.1954, "step": 1717 }, { "epoch": 0.4668478260869565, "grad_norm": 1.2042642831802368, "learning_rate": 3e-05, "loss": 2.7059, "step": 1718 }, { "epoch": 0.4671195652173913, "grad_norm": 2.4052507877349854, "learning_rate": 3e-05, "loss": 3.6323, "step": 1719 }, { "epoch": 0.4673913043478261, "grad_norm": 0.8568060994148254, "learning_rate": 3e-05, "loss": 2.5241, "step": 1720 }, { "epoch": 0.4676630434782609, "grad_norm": 0.9677852988243103, "learning_rate": 3e-05, "loss": 2.5949, "step": 1721 }, { "epoch": 0.46793478260869564, "grad_norm": 1.2894710302352905, "learning_rate": 3e-05, "loss": 2.989, "step": 1722 }, { "epoch": 0.4682065217391304, "grad_norm": 1.5006341934204102, "learning_rate": 3e-05, "loss": 3.6919, "step": 1723 }, { "epoch": 0.46847826086956523, "grad_norm": 1.2537362575531006, "learning_rate": 3e-05, "loss": 3.3816, "step": 1724 }, { "epoch": 0.46875, "grad_norm": 1.333892822265625, "learning_rate": 3e-05, "loss": 3.3992, "step": 1725 }, { "epoch": 0.46902173913043477, "grad_norm": 1.1863936185836792, "learning_rate": 3e-05, "loss": 3.1168, "step": 1726 }, { "epoch": 0.4692934782608696, "grad_norm": 1.2577567100524902, "learning_rate": 3e-05, "loss": 3.4444, "step": 1727 }, { "epoch": 0.46956521739130436, "grad_norm": 1.2792398929595947, "learning_rate": 3e-05, "loss": 2.8464, "step": 1728 }, { "epoch": 0.4698369565217391, "grad_norm": 2.50295090675354, "learning_rate": 3e-05, "loss": 3.2523, "step": 1729 }, { "epoch": 0.4701086956521739, "grad_norm": 1.842974066734314, "learning_rate": 3e-05, "loss": 3.3193, "step": 1730 }, { "epoch": 0.4703804347826087, "grad_norm": 1.1899735927581787, "learning_rate": 3e-05, "loss": 3.4586, "step": 1731 }, { "epoch": 0.4706521739130435, "grad_norm": 1.7522422075271606, "learning_rate": 3e-05, "loss": 3.44, "step": 1732 }, { "epoch": 0.47092391304347825, "grad_norm": 2.0558695793151855, "learning_rate": 3e-05, "loss": 3.3303, "step": 1733 }, { "epoch": 0.47119565217391307, "grad_norm": 1.5954774618148804, "learning_rate": 3e-05, "loss": 3.3366, "step": 1734 }, { "epoch": 0.47146739130434784, "grad_norm": 1.4976911544799805, "learning_rate": 3e-05, "loss": 3.7435, "step": 1735 }, { "epoch": 0.4717391304347826, "grad_norm": 1.4843627214431763, "learning_rate": 3e-05, "loss": 3.0997, "step": 1736 }, { "epoch": 0.4720108695652174, "grad_norm": 1.7294570207595825, "learning_rate": 3e-05, "loss": 3.4683, "step": 1737 }, { "epoch": 0.4722826086956522, "grad_norm": 1.7918407917022705, "learning_rate": 3e-05, "loss": 3.5634, "step": 1738 }, { "epoch": 0.47255434782608696, "grad_norm": 1.458388090133667, "learning_rate": 3e-05, "loss": 3.724, "step": 1739 }, { "epoch": 0.47282608695652173, "grad_norm": 1.7183854579925537, "learning_rate": 3e-05, "loss": 3.7768, "step": 1740 }, { "epoch": 0.4730978260869565, "grad_norm": 1.6272084712982178, "learning_rate": 3e-05, "loss": 3.1361, "step": 1741 }, { "epoch": 0.4733695652173913, "grad_norm": 1.2989143133163452, "learning_rate": 3e-05, "loss": 3.1645, "step": 1742 }, { "epoch": 0.4736413043478261, "grad_norm": 1.1352357864379883, "learning_rate": 3e-05, "loss": 2.9178, "step": 1743 }, { "epoch": 0.47391304347826085, "grad_norm": 1.6457839012145996, "learning_rate": 3e-05, "loss": 3.8608, "step": 1744 }, { "epoch": 0.4741847826086957, "grad_norm": 1.3702744245529175, "learning_rate": 3e-05, "loss": 2.8434, "step": 1745 }, { "epoch": 0.47445652173913044, "grad_norm": 1.2821577787399292, "learning_rate": 3e-05, "loss": 3.0827, "step": 1746 }, { "epoch": 0.4747282608695652, "grad_norm": 1.4484645128250122, "learning_rate": 3e-05, "loss": 3.222, "step": 1747 }, { "epoch": 0.475, "grad_norm": 1.7668582201004028, "learning_rate": 3e-05, "loss": 3.2441, "step": 1748 }, { "epoch": 0.4752717391304348, "grad_norm": 1.764919400215149, "learning_rate": 3e-05, "loss": 3.7584, "step": 1749 }, { "epoch": 0.47554347826086957, "grad_norm": 1.640993356704712, "learning_rate": 3e-05, "loss": 3.3807, "step": 1750 }, { "epoch": 0.47581521739130433, "grad_norm": 1.6005651950836182, "learning_rate": 3e-05, "loss": 4.2452, "step": 1751 }, { "epoch": 0.47608695652173916, "grad_norm": 1.421935796737671, "learning_rate": 3e-05, "loss": 3.0033, "step": 1752 }, { "epoch": 0.4763586956521739, "grad_norm": 1.7603181600570679, "learning_rate": 3e-05, "loss": 3.2585, "step": 1753 }, { "epoch": 0.4766304347826087, "grad_norm": 1.378020167350769, "learning_rate": 3e-05, "loss": 3.0198, "step": 1754 }, { "epoch": 0.47690217391304346, "grad_norm": 1.5630027055740356, "learning_rate": 3e-05, "loss": 3.0409, "step": 1755 }, { "epoch": 0.4771739130434783, "grad_norm": 1.1110471487045288, "learning_rate": 3e-05, "loss": 3.2521, "step": 1756 }, { "epoch": 0.47744565217391305, "grad_norm": 1.403874397277832, "learning_rate": 3e-05, "loss": 2.9834, "step": 1757 }, { "epoch": 0.4777173913043478, "grad_norm": 1.2942219972610474, "learning_rate": 3e-05, "loss": 3.1724, "step": 1758 }, { "epoch": 0.4779891304347826, "grad_norm": 1.152891993522644, "learning_rate": 3e-05, "loss": 3.2869, "step": 1759 }, { "epoch": 0.4782608695652174, "grad_norm": 1.1676504611968994, "learning_rate": 3e-05, "loss": 3.1582, "step": 1760 }, { "epoch": 0.4785326086956522, "grad_norm": 1.151185154914856, "learning_rate": 3e-05, "loss": 3.3322, "step": 1761 }, { "epoch": 0.47880434782608694, "grad_norm": 1.4192887544631958, "learning_rate": 3e-05, "loss": 3.3659, "step": 1762 }, { "epoch": 0.47907608695652176, "grad_norm": 1.5391749143600464, "learning_rate": 3e-05, "loss": 3.2556, "step": 1763 }, { "epoch": 0.47934782608695653, "grad_norm": 1.4667121171951294, "learning_rate": 3e-05, "loss": 3.8051, "step": 1764 }, { "epoch": 0.4796195652173913, "grad_norm": 1.3416023254394531, "learning_rate": 3e-05, "loss": 3.4739, "step": 1765 }, { "epoch": 0.47989130434782606, "grad_norm": 1.7379714250564575, "learning_rate": 3e-05, "loss": 3.2991, "step": 1766 }, { "epoch": 0.4801630434782609, "grad_norm": 1.6660994291305542, "learning_rate": 3e-05, "loss": 3.1885, "step": 1767 }, { "epoch": 0.48043478260869565, "grad_norm": 1.9886304140090942, "learning_rate": 3e-05, "loss": 3.3295, "step": 1768 }, { "epoch": 0.4807065217391304, "grad_norm": 2.345217227935791, "learning_rate": 3e-05, "loss": 3.6796, "step": 1769 }, { "epoch": 0.48097826086956524, "grad_norm": 1.6064239740371704, "learning_rate": 3e-05, "loss": 2.977, "step": 1770 }, { "epoch": 0.48125, "grad_norm": 1.7928065061569214, "learning_rate": 3e-05, "loss": 3.5848, "step": 1771 }, { "epoch": 0.4815217391304348, "grad_norm": 1.5981751680374146, "learning_rate": 3e-05, "loss": 3.1049, "step": 1772 }, { "epoch": 0.48179347826086955, "grad_norm": 1.6446112394332886, "learning_rate": 3e-05, "loss": 2.9694, "step": 1773 }, { "epoch": 0.48206521739130437, "grad_norm": 1.8991427421569824, "learning_rate": 3e-05, "loss": 3.3139, "step": 1774 }, { "epoch": 0.48233695652173914, "grad_norm": 2.3716299533843994, "learning_rate": 3e-05, "loss": 3.9361, "step": 1775 }, { "epoch": 0.4826086956521739, "grad_norm": 1.8573355674743652, "learning_rate": 3e-05, "loss": 3.1394, "step": 1776 }, { "epoch": 0.48288043478260867, "grad_norm": 1.8625982999801636, "learning_rate": 3e-05, "loss": 3.8017, "step": 1777 }, { "epoch": 0.4831521739130435, "grad_norm": 2.2275497913360596, "learning_rate": 3e-05, "loss": 3.4565, "step": 1778 }, { "epoch": 0.48342391304347826, "grad_norm": 1.9856013059616089, "learning_rate": 3e-05, "loss": 3.5517, "step": 1779 }, { "epoch": 0.483695652173913, "grad_norm": 1.3510781526565552, "learning_rate": 3e-05, "loss": 3.1976, "step": 1780 }, { "epoch": 0.48396739130434785, "grad_norm": 1.0482627153396606, "learning_rate": 3e-05, "loss": 3.0025, "step": 1781 }, { "epoch": 0.4842391304347826, "grad_norm": 1.3749173879623413, "learning_rate": 3e-05, "loss": 3.2774, "step": 1782 }, { "epoch": 0.4845108695652174, "grad_norm": 1.405247449874878, "learning_rate": 3e-05, "loss": 3.4099, "step": 1783 }, { "epoch": 0.48478260869565215, "grad_norm": 1.2630794048309326, "learning_rate": 3e-05, "loss": 3.7092, "step": 1784 }, { "epoch": 0.485054347826087, "grad_norm": 1.1290115118026733, "learning_rate": 3e-05, "loss": 3.432, "step": 1785 }, { "epoch": 0.48532608695652174, "grad_norm": 1.2523903846740723, "learning_rate": 3e-05, "loss": 3.1431, "step": 1786 }, { "epoch": 0.4855978260869565, "grad_norm": 1.218867540359497, "learning_rate": 3e-05, "loss": 3.4881, "step": 1787 }, { "epoch": 0.48586956521739133, "grad_norm": 1.6144737005233765, "learning_rate": 3e-05, "loss": 4.0674, "step": 1788 }, { "epoch": 0.4861413043478261, "grad_norm": 1.4506441354751587, "learning_rate": 3e-05, "loss": 3.5498, "step": 1789 }, { "epoch": 0.48641304347826086, "grad_norm": 1.5907642841339111, "learning_rate": 3e-05, "loss": 3.1341, "step": 1790 }, { "epoch": 0.48668478260869563, "grad_norm": 1.18953275680542, "learning_rate": 3e-05, "loss": 3.0503, "step": 1791 }, { "epoch": 0.48695652173913045, "grad_norm": 1.2186987400054932, "learning_rate": 3e-05, "loss": 2.7508, "step": 1792 }, { "epoch": 0.4872282608695652, "grad_norm": 1.111574649810791, "learning_rate": 3e-05, "loss": 2.8275, "step": 1793 }, { "epoch": 0.4875, "grad_norm": 1.275896430015564, "learning_rate": 3e-05, "loss": 3.4576, "step": 1794 }, { "epoch": 0.48777173913043476, "grad_norm": 1.182626485824585, "learning_rate": 3e-05, "loss": 2.9678, "step": 1795 }, { "epoch": 0.4880434782608696, "grad_norm": 1.098120927810669, "learning_rate": 3e-05, "loss": 3.1544, "step": 1796 }, { "epoch": 0.48831521739130435, "grad_norm": 0.9524686932563782, "learning_rate": 3e-05, "loss": 3.0739, "step": 1797 }, { "epoch": 0.4885869565217391, "grad_norm": 1.302868366241455, "learning_rate": 3e-05, "loss": 3.3036, "step": 1798 }, { "epoch": 0.48885869565217394, "grad_norm": 1.124603271484375, "learning_rate": 3e-05, "loss": 3.2132, "step": 1799 }, { "epoch": 0.4891304347826087, "grad_norm": 1.3126977682113647, "learning_rate": 3e-05, "loss": 3.6832, "step": 1800 }, { "epoch": 0.48940217391304347, "grad_norm": 1.436482310295105, "learning_rate": 3e-05, "loss": 3.9607, "step": 1801 }, { "epoch": 0.48967391304347824, "grad_norm": 1.6329675912857056, "learning_rate": 3e-05, "loss": 2.6751, "step": 1802 }, { "epoch": 0.48994565217391306, "grad_norm": 1.2873584032058716, "learning_rate": 3e-05, "loss": 3.3793, "step": 1803 }, { "epoch": 0.4902173913043478, "grad_norm": 1.0539244413375854, "learning_rate": 3e-05, "loss": 3.0761, "step": 1804 }, { "epoch": 0.4904891304347826, "grad_norm": 1.7557803392410278, "learning_rate": 3e-05, "loss": 3.0488, "step": 1805 }, { "epoch": 0.4907608695652174, "grad_norm": 1.8407940864562988, "learning_rate": 3e-05, "loss": 3.1567, "step": 1806 }, { "epoch": 0.4910326086956522, "grad_norm": 1.3472580909729004, "learning_rate": 3e-05, "loss": 3.2331, "step": 1807 }, { "epoch": 0.49130434782608695, "grad_norm": 1.093360424041748, "learning_rate": 3e-05, "loss": 3.1883, "step": 1808 }, { "epoch": 0.4915760869565217, "grad_norm": 1.650109887123108, "learning_rate": 3e-05, "loss": 3.7661, "step": 1809 }, { "epoch": 0.49184782608695654, "grad_norm": 1.454372525215149, "learning_rate": 3e-05, "loss": 3.1205, "step": 1810 }, { "epoch": 0.4921195652173913, "grad_norm": 1.4370765686035156, "learning_rate": 3e-05, "loss": 3.8767, "step": 1811 }, { "epoch": 0.4923913043478261, "grad_norm": 2.6512532234191895, "learning_rate": 3e-05, "loss": 4.0366, "step": 1812 }, { "epoch": 0.49266304347826084, "grad_norm": 2.1347718238830566, "learning_rate": 3e-05, "loss": 3.5459, "step": 1813 }, { "epoch": 0.49293478260869567, "grad_norm": 1.5831135511398315, "learning_rate": 3e-05, "loss": 3.4828, "step": 1814 }, { "epoch": 0.49320652173913043, "grad_norm": 1.0648609399795532, "learning_rate": 3e-05, "loss": 3.0868, "step": 1815 }, { "epoch": 0.4934782608695652, "grad_norm": 1.467032551765442, "learning_rate": 3e-05, "loss": 3.6537, "step": 1816 }, { "epoch": 0.49375, "grad_norm": 1.2471240758895874, "learning_rate": 3e-05, "loss": 3.2671, "step": 1817 }, { "epoch": 0.4940217391304348, "grad_norm": 0.9745358228683472, "learning_rate": 3e-05, "loss": 3.4916, "step": 1818 }, { "epoch": 0.49429347826086956, "grad_norm": 1.3291053771972656, "learning_rate": 3e-05, "loss": 3.7441, "step": 1819 }, { "epoch": 0.4945652173913043, "grad_norm": 1.3420668840408325, "learning_rate": 3e-05, "loss": 3.3106, "step": 1820 }, { "epoch": 0.49483695652173915, "grad_norm": 1.0097370147705078, "learning_rate": 3e-05, "loss": 3.1215, "step": 1821 }, { "epoch": 0.4951086956521739, "grad_norm": 1.1169931888580322, "learning_rate": 3e-05, "loss": 2.9595, "step": 1822 }, { "epoch": 0.4953804347826087, "grad_norm": 1.1370187997817993, "learning_rate": 3e-05, "loss": 3.5224, "step": 1823 }, { "epoch": 0.4956521739130435, "grad_norm": 1.5470448732376099, "learning_rate": 3e-05, "loss": 3.9269, "step": 1824 }, { "epoch": 0.49592391304347827, "grad_norm": 1.000895380973816, "learning_rate": 3e-05, "loss": 2.9649, "step": 1825 }, { "epoch": 0.49619565217391304, "grad_norm": 1.2407017946243286, "learning_rate": 3e-05, "loss": 3.5568, "step": 1826 }, { "epoch": 0.4964673913043478, "grad_norm": 1.226485013961792, "learning_rate": 3e-05, "loss": 2.5796, "step": 1827 }, { "epoch": 0.4967391304347826, "grad_norm": 1.1970787048339844, "learning_rate": 3e-05, "loss": 2.8772, "step": 1828 }, { "epoch": 0.4970108695652174, "grad_norm": 1.2727575302124023, "learning_rate": 3e-05, "loss": 3.6501, "step": 1829 }, { "epoch": 0.49728260869565216, "grad_norm": 1.2288799285888672, "learning_rate": 3e-05, "loss": 3.0078, "step": 1830 }, { "epoch": 0.49755434782608693, "grad_norm": 1.688232660293579, "learning_rate": 3e-05, "loss": 3.2542, "step": 1831 }, { "epoch": 0.49782608695652175, "grad_norm": 1.6907511949539185, "learning_rate": 3e-05, "loss": 3.6193, "step": 1832 }, { "epoch": 0.4980978260869565, "grad_norm": 1.1391464471817017, "learning_rate": 3e-05, "loss": 3.3688, "step": 1833 }, { "epoch": 0.4983695652173913, "grad_norm": 1.2691822052001953, "learning_rate": 3e-05, "loss": 3.2763, "step": 1834 }, { "epoch": 0.4986413043478261, "grad_norm": 1.1176376342773438, "learning_rate": 3e-05, "loss": 3.0038, "step": 1835 }, { "epoch": 0.4989130434782609, "grad_norm": 0.9173082709312439, "learning_rate": 3e-05, "loss": 2.7976, "step": 1836 }, { "epoch": 0.49918478260869564, "grad_norm": 1.0989121198654175, "learning_rate": 3e-05, "loss": 3.0254, "step": 1837 }, { "epoch": 0.4994565217391304, "grad_norm": 1.2885278463363647, "learning_rate": 3e-05, "loss": 3.2596, "step": 1838 }, { "epoch": 0.49972826086956523, "grad_norm": 1.1294320821762085, "learning_rate": 3e-05, "loss": 3.0833, "step": 1839 }, { "epoch": 0.5, "grad_norm": 0.9139266014099121, "learning_rate": 3e-05, "loss": 3.043, "step": 1840 }, { "epoch": 0.5002717391304348, "grad_norm": 0.9999741315841675, "learning_rate": 3e-05, "loss": 2.6746, "step": 1841 }, { "epoch": 0.5005434782608695, "grad_norm": 2.3439218997955322, "learning_rate": 3e-05, "loss": 4.0552, "step": 1842 }, { "epoch": 0.5008152173913043, "grad_norm": 1.3185632228851318, "learning_rate": 3e-05, "loss": 3.4046, "step": 1843 }, { "epoch": 0.5010869565217392, "grad_norm": 1.4254107475280762, "learning_rate": 3e-05, "loss": 3.5323, "step": 1844 }, { "epoch": 0.501358695652174, "grad_norm": 1.0706696510314941, "learning_rate": 3e-05, "loss": 3.0672, "step": 1845 }, { "epoch": 0.5016304347826087, "grad_norm": 1.316158413887024, "learning_rate": 3e-05, "loss": 3.3709, "step": 1846 }, { "epoch": 0.5019021739130435, "grad_norm": 1.2480392456054688, "learning_rate": 3e-05, "loss": 3.673, "step": 1847 }, { "epoch": 0.5021739130434782, "grad_norm": 1.596869707107544, "learning_rate": 3e-05, "loss": 3.3651, "step": 1848 }, { "epoch": 0.502445652173913, "grad_norm": 1.6036261320114136, "learning_rate": 3e-05, "loss": 3.68, "step": 1849 }, { "epoch": 0.5027173913043478, "grad_norm": 1.2420178651809692, "learning_rate": 3e-05, "loss": 2.9173, "step": 1850 }, { "epoch": 0.5029891304347827, "grad_norm": 1.588339924812317, "learning_rate": 3e-05, "loss": 3.7931, "step": 1851 }, { "epoch": 0.5032608695652174, "grad_norm": 1.2808369398117065, "learning_rate": 3e-05, "loss": 3.05, "step": 1852 }, { "epoch": 0.5035326086956522, "grad_norm": 1.234026551246643, "learning_rate": 3e-05, "loss": 3.3401, "step": 1853 }, { "epoch": 0.503804347826087, "grad_norm": 1.6384221315383911, "learning_rate": 3e-05, "loss": 3.5597, "step": 1854 }, { "epoch": 0.5040760869565217, "grad_norm": 1.8604109287261963, "learning_rate": 3e-05, "loss": 3.6333, "step": 1855 }, { "epoch": 0.5043478260869565, "grad_norm": 1.4895154237747192, "learning_rate": 3e-05, "loss": 3.4255, "step": 1856 }, { "epoch": 0.5046195652173913, "grad_norm": 1.3410334587097168, "learning_rate": 3e-05, "loss": 2.7753, "step": 1857 }, { "epoch": 0.5048913043478261, "grad_norm": 1.446807861328125, "learning_rate": 3e-05, "loss": 4.1229, "step": 1858 }, { "epoch": 0.5051630434782609, "grad_norm": 1.188097596168518, "learning_rate": 3e-05, "loss": 3.2846, "step": 1859 }, { "epoch": 0.5054347826086957, "grad_norm": 1.2120987176895142, "learning_rate": 3e-05, "loss": 3.147, "step": 1860 }, { "epoch": 0.5057065217391304, "grad_norm": 1.5114277601242065, "learning_rate": 3e-05, "loss": 3.0952, "step": 1861 }, { "epoch": 0.5059782608695652, "grad_norm": 1.360628366470337, "learning_rate": 3e-05, "loss": 2.907, "step": 1862 }, { "epoch": 0.50625, "grad_norm": 1.4083906412124634, "learning_rate": 3e-05, "loss": 3.4222, "step": 1863 }, { "epoch": 0.5065217391304347, "grad_norm": 1.245968222618103, "learning_rate": 3e-05, "loss": 3.5063, "step": 1864 }, { "epoch": 0.5067934782608695, "grad_norm": 1.3542323112487793, "learning_rate": 3e-05, "loss": 3.981, "step": 1865 }, { "epoch": 0.5070652173913044, "grad_norm": 2.4021403789520264, "learning_rate": 3e-05, "loss": 3.9367, "step": 1866 }, { "epoch": 0.5073369565217392, "grad_norm": 1.4897315502166748, "learning_rate": 3e-05, "loss": 2.7825, "step": 1867 }, { "epoch": 0.5076086956521739, "grad_norm": 1.7353469133377075, "learning_rate": 3e-05, "loss": 3.5709, "step": 1868 }, { "epoch": 0.5078804347826087, "grad_norm": 1.216436505317688, "learning_rate": 3e-05, "loss": 3.1314, "step": 1869 }, { "epoch": 0.5081521739130435, "grad_norm": 1.2940754890441895, "learning_rate": 3e-05, "loss": 3.3858, "step": 1870 }, { "epoch": 0.5084239130434782, "grad_norm": 0.9794735312461853, "learning_rate": 3e-05, "loss": 2.823, "step": 1871 }, { "epoch": 0.508695652173913, "grad_norm": 1.4511076211929321, "learning_rate": 3e-05, "loss": 4.0647, "step": 1872 }, { "epoch": 0.5089673913043479, "grad_norm": 1.559597373008728, "learning_rate": 3e-05, "loss": 3.093, "step": 1873 }, { "epoch": 0.5092391304347826, "grad_norm": 1.7877241373062134, "learning_rate": 3e-05, "loss": 3.742, "step": 1874 }, { "epoch": 0.5095108695652174, "grad_norm": 1.585822582244873, "learning_rate": 3e-05, "loss": 3.631, "step": 1875 }, { "epoch": 0.5097826086956522, "grad_norm": 1.284661054611206, "learning_rate": 3e-05, "loss": 3.228, "step": 1876 }, { "epoch": 0.5100543478260869, "grad_norm": 1.295775294303894, "learning_rate": 3e-05, "loss": 3.7922, "step": 1877 }, { "epoch": 0.5103260869565217, "grad_norm": 1.2768347263336182, "learning_rate": 3e-05, "loss": 3.0369, "step": 1878 }, { "epoch": 0.5105978260869565, "grad_norm": 1.4773716926574707, "learning_rate": 3e-05, "loss": 3.3407, "step": 1879 }, { "epoch": 0.5108695652173914, "grad_norm": 1.2089797258377075, "learning_rate": 3e-05, "loss": 3.0556, "step": 1880 }, { "epoch": 0.5111413043478261, "grad_norm": 1.4668608903884888, "learning_rate": 3e-05, "loss": 3.4665, "step": 1881 }, { "epoch": 0.5114130434782609, "grad_norm": 1.998399257659912, "learning_rate": 3e-05, "loss": 3.8042, "step": 1882 }, { "epoch": 0.5116847826086957, "grad_norm": 1.448439598083496, "learning_rate": 3e-05, "loss": 3.7303, "step": 1883 }, { "epoch": 0.5119565217391304, "grad_norm": 0.9362940192222595, "learning_rate": 3e-05, "loss": 2.665, "step": 1884 }, { "epoch": 0.5122282608695652, "grad_norm": 1.6077910661697388, "learning_rate": 3e-05, "loss": 3.4826, "step": 1885 }, { "epoch": 0.5125, "grad_norm": 1.5588127374649048, "learning_rate": 3e-05, "loss": 2.837, "step": 1886 }, { "epoch": 0.5127717391304348, "grad_norm": 1.473461389541626, "learning_rate": 3e-05, "loss": 3.2511, "step": 1887 }, { "epoch": 0.5130434782608696, "grad_norm": 1.6276698112487793, "learning_rate": 3e-05, "loss": 3.5821, "step": 1888 }, { "epoch": 0.5133152173913044, "grad_norm": 1.6201423406600952, "learning_rate": 3e-05, "loss": 3.195, "step": 1889 }, { "epoch": 0.5135869565217391, "grad_norm": 1.6411088705062866, "learning_rate": 3e-05, "loss": 3.1111, "step": 1890 }, { "epoch": 0.5138586956521739, "grad_norm": 1.2973185777664185, "learning_rate": 3e-05, "loss": 2.6324, "step": 1891 }, { "epoch": 0.5141304347826087, "grad_norm": 2.3508846759796143, "learning_rate": 3e-05, "loss": 3.7621, "step": 1892 }, { "epoch": 0.5144021739130434, "grad_norm": 1.718246340751648, "learning_rate": 3e-05, "loss": 3.759, "step": 1893 }, { "epoch": 0.5146739130434783, "grad_norm": 1.5278515815734863, "learning_rate": 3e-05, "loss": 3.2879, "step": 1894 }, { "epoch": 0.5149456521739131, "grad_norm": 1.308929681777954, "learning_rate": 3e-05, "loss": 3.1075, "step": 1895 }, { "epoch": 0.5152173913043478, "grad_norm": 1.40326988697052, "learning_rate": 3e-05, "loss": 3.5211, "step": 1896 }, { "epoch": 0.5154891304347826, "grad_norm": 1.5129557847976685, "learning_rate": 3e-05, "loss": 3.2549, "step": 1897 }, { "epoch": 0.5157608695652174, "grad_norm": 1.1121063232421875, "learning_rate": 3e-05, "loss": 2.9461, "step": 1898 }, { "epoch": 0.5160326086956522, "grad_norm": 1.4302382469177246, "learning_rate": 3e-05, "loss": 3.627, "step": 1899 }, { "epoch": 0.5163043478260869, "grad_norm": 1.6184173822402954, "learning_rate": 3e-05, "loss": 2.9157, "step": 1900 }, { "epoch": 0.5165760869565217, "grad_norm": 1.1715192794799805, "learning_rate": 3e-05, "loss": 2.9425, "step": 1901 }, { "epoch": 0.5168478260869566, "grad_norm": 2.065067768096924, "learning_rate": 3e-05, "loss": 3.3198, "step": 1902 }, { "epoch": 0.5171195652173913, "grad_norm": 1.6942062377929688, "learning_rate": 3e-05, "loss": 3.4998, "step": 1903 }, { "epoch": 0.5173913043478261, "grad_norm": 1.4067577123641968, "learning_rate": 3e-05, "loss": 3.501, "step": 1904 }, { "epoch": 0.5176630434782609, "grad_norm": 1.5794352293014526, "learning_rate": 3e-05, "loss": 3.1307, "step": 1905 }, { "epoch": 0.5179347826086956, "grad_norm": 1.5043370723724365, "learning_rate": 3e-05, "loss": 3.8714, "step": 1906 }, { "epoch": 0.5182065217391304, "grad_norm": 1.2884905338287354, "learning_rate": 3e-05, "loss": 3.4459, "step": 1907 }, { "epoch": 0.5184782608695652, "grad_norm": 1.3515671491622925, "learning_rate": 3e-05, "loss": 3.5685, "step": 1908 }, { "epoch": 0.51875, "grad_norm": 1.2115263938903809, "learning_rate": 3e-05, "loss": 3.2234, "step": 1909 }, { "epoch": 0.5190217391304348, "grad_norm": 1.3323924541473389, "learning_rate": 3e-05, "loss": 3.42, "step": 1910 }, { "epoch": 0.5192934782608696, "grad_norm": 1.2837159633636475, "learning_rate": 3e-05, "loss": 3.0186, "step": 1911 }, { "epoch": 0.5195652173913043, "grad_norm": 1.1012318134307861, "learning_rate": 3e-05, "loss": 2.9993, "step": 1912 }, { "epoch": 0.5198369565217391, "grad_norm": 1.253250002861023, "learning_rate": 3e-05, "loss": 3.371, "step": 1913 }, { "epoch": 0.5201086956521739, "grad_norm": 0.9867514967918396, "learning_rate": 3e-05, "loss": 2.7761, "step": 1914 }, { "epoch": 0.5203804347826086, "grad_norm": 1.0671703815460205, "learning_rate": 3e-05, "loss": 3.5288, "step": 1915 }, { "epoch": 0.5206521739130435, "grad_norm": 0.9514213800430298, "learning_rate": 3e-05, "loss": 2.7831, "step": 1916 }, { "epoch": 0.5209239130434783, "grad_norm": 1.3238492012023926, "learning_rate": 3e-05, "loss": 3.1874, "step": 1917 }, { "epoch": 0.5211956521739131, "grad_norm": 1.5408774614334106, "learning_rate": 3e-05, "loss": 3.1997, "step": 1918 }, { "epoch": 0.5214673913043478, "grad_norm": 1.1520451307296753, "learning_rate": 3e-05, "loss": 3.1287, "step": 1919 }, { "epoch": 0.5217391304347826, "grad_norm": 1.5356744527816772, "learning_rate": 3e-05, "loss": 3.5281, "step": 1920 }, { "epoch": 0.5220108695652174, "grad_norm": 1.0727752447128296, "learning_rate": 3e-05, "loss": 3.0281, "step": 1921 }, { "epoch": 0.5222826086956521, "grad_norm": 1.4737282991409302, "learning_rate": 3e-05, "loss": 4.0763, "step": 1922 }, { "epoch": 0.522554347826087, "grad_norm": 1.2212142944335938, "learning_rate": 3e-05, "loss": 3.6583, "step": 1923 }, { "epoch": 0.5228260869565218, "grad_norm": 1.42307710647583, "learning_rate": 3e-05, "loss": 3.9908, "step": 1924 }, { "epoch": 0.5230978260869565, "grad_norm": 1.2455291748046875, "learning_rate": 3e-05, "loss": 2.7686, "step": 1925 }, { "epoch": 0.5233695652173913, "grad_norm": 1.1880292892456055, "learning_rate": 3e-05, "loss": 2.8165, "step": 1926 }, { "epoch": 0.5236413043478261, "grad_norm": 1.329270362854004, "learning_rate": 3e-05, "loss": 3.421, "step": 1927 }, { "epoch": 0.5239130434782608, "grad_norm": 1.8960901498794556, "learning_rate": 3e-05, "loss": 4.0437, "step": 1928 }, { "epoch": 0.5241847826086956, "grad_norm": 1.223769187927246, "learning_rate": 3e-05, "loss": 3.3202, "step": 1929 }, { "epoch": 0.5244565217391305, "grad_norm": 1.5670809745788574, "learning_rate": 3e-05, "loss": 3.6805, "step": 1930 }, { "epoch": 0.5247282608695653, "grad_norm": 1.9276920557022095, "learning_rate": 3e-05, "loss": 3.4752, "step": 1931 }, { "epoch": 0.525, "grad_norm": 1.7917790412902832, "learning_rate": 3e-05, "loss": 3.4745, "step": 1932 }, { "epoch": 0.5252717391304348, "grad_norm": 1.1239231824874878, "learning_rate": 3e-05, "loss": 2.9622, "step": 1933 }, { "epoch": 0.5255434782608696, "grad_norm": 1.2177050113677979, "learning_rate": 3e-05, "loss": 2.691, "step": 1934 }, { "epoch": 0.5258152173913043, "grad_norm": 1.1526474952697754, "learning_rate": 3e-05, "loss": 2.7313, "step": 1935 }, { "epoch": 0.5260869565217391, "grad_norm": 1.3403104543685913, "learning_rate": 3e-05, "loss": 3.215, "step": 1936 }, { "epoch": 0.5263586956521739, "grad_norm": 1.1823958158493042, "learning_rate": 3e-05, "loss": 3.0382, "step": 1937 }, { "epoch": 0.5266304347826087, "grad_norm": 1.2578941583633423, "learning_rate": 3e-05, "loss": 2.9962, "step": 1938 }, { "epoch": 0.5269021739130435, "grad_norm": 1.2342561483383179, "learning_rate": 3e-05, "loss": 3.6362, "step": 1939 }, { "epoch": 0.5271739130434783, "grad_norm": 1.0608114004135132, "learning_rate": 3e-05, "loss": 3.2318, "step": 1940 }, { "epoch": 0.527445652173913, "grad_norm": 1.2502753734588623, "learning_rate": 3e-05, "loss": 3.1968, "step": 1941 }, { "epoch": 0.5277173913043478, "grad_norm": 0.9138715267181396, "learning_rate": 3e-05, "loss": 3.0353, "step": 1942 }, { "epoch": 0.5279891304347826, "grad_norm": 1.257987380027771, "learning_rate": 3e-05, "loss": 2.5984, "step": 1943 }, { "epoch": 0.5282608695652173, "grad_norm": 1.5247408151626587, "learning_rate": 3e-05, "loss": 3.4343, "step": 1944 }, { "epoch": 0.5285326086956522, "grad_norm": 1.1517291069030762, "learning_rate": 3e-05, "loss": 2.7655, "step": 1945 }, { "epoch": 0.528804347826087, "grad_norm": 1.5516606569290161, "learning_rate": 3e-05, "loss": 3.9976, "step": 1946 }, { "epoch": 0.5290760869565218, "grad_norm": 1.6808713674545288, "learning_rate": 3e-05, "loss": 3.6858, "step": 1947 }, { "epoch": 0.5293478260869565, "grad_norm": 1.0476493835449219, "learning_rate": 3e-05, "loss": 2.8933, "step": 1948 }, { "epoch": 0.5296195652173913, "grad_norm": 1.7297437191009521, "learning_rate": 3e-05, "loss": 3.1754, "step": 1949 }, { "epoch": 0.529891304347826, "grad_norm": 1.4908910989761353, "learning_rate": 3e-05, "loss": 3.9229, "step": 1950 }, { "epoch": 0.5301630434782608, "grad_norm": 1.3128184080123901, "learning_rate": 3e-05, "loss": 3.5215, "step": 1951 }, { "epoch": 0.5304347826086957, "grad_norm": 1.7025582790374756, "learning_rate": 3e-05, "loss": 3.5113, "step": 1952 }, { "epoch": 0.5307065217391305, "grad_norm": 1.4770921468734741, "learning_rate": 3e-05, "loss": 3.0171, "step": 1953 }, { "epoch": 0.5309782608695652, "grad_norm": 1.6477012634277344, "learning_rate": 3e-05, "loss": 3.5914, "step": 1954 }, { "epoch": 0.53125, "grad_norm": 1.5680129528045654, "learning_rate": 3e-05, "loss": 3.6162, "step": 1955 }, { "epoch": 0.5315217391304348, "grad_norm": 2.2198116779327393, "learning_rate": 3e-05, "loss": 3.2198, "step": 1956 }, { "epoch": 0.5317934782608695, "grad_norm": 1.9690414667129517, "learning_rate": 3e-05, "loss": 3.0914, "step": 1957 }, { "epoch": 0.5320652173913043, "grad_norm": 1.4856075048446655, "learning_rate": 3e-05, "loss": 3.206, "step": 1958 }, { "epoch": 0.5323369565217392, "grad_norm": 1.1297136545181274, "learning_rate": 3e-05, "loss": 2.8479, "step": 1959 }, { "epoch": 0.532608695652174, "grad_norm": 1.859328269958496, "learning_rate": 3e-05, "loss": 3.3152, "step": 1960 }, { "epoch": 0.5328804347826087, "grad_norm": 1.939894199371338, "learning_rate": 3e-05, "loss": 3.2557, "step": 1961 }, { "epoch": 0.5331521739130435, "grad_norm": 2.148735523223877, "learning_rate": 3e-05, "loss": 3.6686, "step": 1962 }, { "epoch": 0.5334239130434782, "grad_norm": 1.4653475284576416, "learning_rate": 3e-05, "loss": 3.8618, "step": 1963 }, { "epoch": 0.533695652173913, "grad_norm": 2.1480512619018555, "learning_rate": 3e-05, "loss": 3.4809, "step": 1964 }, { "epoch": 0.5339673913043478, "grad_norm": 2.2585768699645996, "learning_rate": 3e-05, "loss": 3.1685, "step": 1965 }, { "epoch": 0.5342391304347827, "grad_norm": 2.143669605255127, "learning_rate": 3e-05, "loss": 3.8206, "step": 1966 }, { "epoch": 0.5345108695652174, "grad_norm": 1.5735530853271484, "learning_rate": 3e-05, "loss": 3.5483, "step": 1967 }, { "epoch": 0.5347826086956522, "grad_norm": 1.5427213907241821, "learning_rate": 3e-05, "loss": 3.469, "step": 1968 }, { "epoch": 0.535054347826087, "grad_norm": 1.7961771488189697, "learning_rate": 3e-05, "loss": 3.2229, "step": 1969 }, { "epoch": 0.5353260869565217, "grad_norm": 1.5997658967971802, "learning_rate": 3e-05, "loss": 3.6575, "step": 1970 }, { "epoch": 0.5355978260869565, "grad_norm": 1.2979601621627808, "learning_rate": 3e-05, "loss": 3.7012, "step": 1971 }, { "epoch": 0.5358695652173913, "grad_norm": 1.5162733793258667, "learning_rate": 3e-05, "loss": 3.1471, "step": 1972 }, { "epoch": 0.5361413043478261, "grad_norm": 1.5043525695800781, "learning_rate": 3e-05, "loss": 2.8777, "step": 1973 }, { "epoch": 0.5364130434782609, "grad_norm": 1.1250759363174438, "learning_rate": 3e-05, "loss": 2.9937, "step": 1974 }, { "epoch": 0.5366847826086957, "grad_norm": 1.2006231546401978, "learning_rate": 3e-05, "loss": 3.5737, "step": 1975 }, { "epoch": 0.5369565217391304, "grad_norm": 0.9950990080833435, "learning_rate": 3e-05, "loss": 2.9763, "step": 1976 }, { "epoch": 0.5372282608695652, "grad_norm": 1.2111644744873047, "learning_rate": 3e-05, "loss": 3.5871, "step": 1977 }, { "epoch": 0.5375, "grad_norm": 1.4334638118743896, "learning_rate": 3e-05, "loss": 3.645, "step": 1978 }, { "epoch": 0.5377717391304347, "grad_norm": 1.376451849937439, "learning_rate": 3e-05, "loss": 3.5722, "step": 1979 }, { "epoch": 0.5380434782608695, "grad_norm": 1.1852439641952515, "learning_rate": 3e-05, "loss": 3.2733, "step": 1980 }, { "epoch": 0.5383152173913044, "grad_norm": 1.208321452140808, "learning_rate": 3e-05, "loss": 3.4223, "step": 1981 }, { "epoch": 0.5385869565217392, "grad_norm": 1.2566808462142944, "learning_rate": 3e-05, "loss": 3.227, "step": 1982 }, { "epoch": 0.5388586956521739, "grad_norm": 1.1662538051605225, "learning_rate": 3e-05, "loss": 3.9731, "step": 1983 }, { "epoch": 0.5391304347826087, "grad_norm": 0.9991174340248108, "learning_rate": 3e-05, "loss": 3.2856, "step": 1984 }, { "epoch": 0.5394021739130435, "grad_norm": 1.6792123317718506, "learning_rate": 3e-05, "loss": 4.0315, "step": 1985 }, { "epoch": 0.5396739130434782, "grad_norm": 1.3554564714431763, "learning_rate": 3e-05, "loss": 2.7884, "step": 1986 }, { "epoch": 0.539945652173913, "grad_norm": 2.0495409965515137, "learning_rate": 3e-05, "loss": 4.0834, "step": 1987 }, { "epoch": 0.5402173913043479, "grad_norm": 1.3076777458190918, "learning_rate": 3e-05, "loss": 2.9288, "step": 1988 }, { "epoch": 0.5404891304347826, "grad_norm": 1.2214155197143555, "learning_rate": 3e-05, "loss": 2.9785, "step": 1989 }, { "epoch": 0.5407608695652174, "grad_norm": 1.6654658317565918, "learning_rate": 3e-05, "loss": 3.6795, "step": 1990 }, { "epoch": 0.5410326086956522, "grad_norm": 1.224678874015808, "learning_rate": 3e-05, "loss": 2.9736, "step": 1991 }, { "epoch": 0.5413043478260869, "grad_norm": 1.128450870513916, "learning_rate": 3e-05, "loss": 3.0371, "step": 1992 }, { "epoch": 0.5415760869565217, "grad_norm": 1.1452875137329102, "learning_rate": 3e-05, "loss": 2.989, "step": 1993 }, { "epoch": 0.5418478260869565, "grad_norm": 1.2154463529586792, "learning_rate": 3e-05, "loss": 2.9404, "step": 1994 }, { "epoch": 0.5421195652173914, "grad_norm": 1.1099878549575806, "learning_rate": 3e-05, "loss": 3.1758, "step": 1995 }, { "epoch": 0.5423913043478261, "grad_norm": 0.9395751953125, "learning_rate": 3e-05, "loss": 2.9722, "step": 1996 }, { "epoch": 0.5426630434782609, "grad_norm": 1.1049753427505493, "learning_rate": 3e-05, "loss": 3.0526, "step": 1997 }, { "epoch": 0.5429347826086957, "grad_norm": 1.45954430103302, "learning_rate": 3e-05, "loss": 3.7409, "step": 1998 }, { "epoch": 0.5432065217391304, "grad_norm": 1.2864134311676025, "learning_rate": 3e-05, "loss": 3.4088, "step": 1999 }, { "epoch": 0.5434782608695652, "grad_norm": 1.2165127992630005, "learning_rate": 3e-05, "loss": 3.2604, "step": 2000 }, { "epoch": 0.54375, "grad_norm": 1.272796869277954, "learning_rate": 3e-05, "loss": 3.6574, "step": 2001 }, { "epoch": 0.5440217391304348, "grad_norm": 1.3245208263397217, "learning_rate": 3e-05, "loss": 3.1406, "step": 2002 }, { "epoch": 0.5442934782608696, "grad_norm": 1.4110616445541382, "learning_rate": 3e-05, "loss": 2.8426, "step": 2003 }, { "epoch": 0.5445652173913044, "grad_norm": 1.7510126829147339, "learning_rate": 3e-05, "loss": 3.926, "step": 2004 }, { "epoch": 0.5448369565217391, "grad_norm": 1.226434588432312, "learning_rate": 3e-05, "loss": 3.5155, "step": 2005 }, { "epoch": 0.5451086956521739, "grad_norm": 1.5343637466430664, "learning_rate": 3e-05, "loss": 2.7504, "step": 2006 }, { "epoch": 0.5453804347826087, "grad_norm": 1.3316307067871094, "learning_rate": 3e-05, "loss": 3.0848, "step": 2007 }, { "epoch": 0.5456521739130434, "grad_norm": 1.2327113151550293, "learning_rate": 3e-05, "loss": 3.0306, "step": 2008 }, { "epoch": 0.5459239130434783, "grad_norm": 1.180682897567749, "learning_rate": 3e-05, "loss": 3.0591, "step": 2009 }, { "epoch": 0.5461956521739131, "grad_norm": 1.3480994701385498, "learning_rate": 3e-05, "loss": 3.3818, "step": 2010 }, { "epoch": 0.5464673913043478, "grad_norm": 1.246237874031067, "learning_rate": 3e-05, "loss": 3.3416, "step": 2011 }, { "epoch": 0.5467391304347826, "grad_norm": 0.8937397003173828, "learning_rate": 3e-05, "loss": 3.0593, "step": 2012 }, { "epoch": 0.5470108695652174, "grad_norm": 1.3019452095031738, "learning_rate": 3e-05, "loss": 3.2755, "step": 2013 }, { "epoch": 0.5472826086956522, "grad_norm": 1.408234715461731, "learning_rate": 3e-05, "loss": 3.4283, "step": 2014 }, { "epoch": 0.5475543478260869, "grad_norm": 1.5709846019744873, "learning_rate": 3e-05, "loss": 3.2933, "step": 2015 }, { "epoch": 0.5478260869565217, "grad_norm": 1.6085715293884277, "learning_rate": 3e-05, "loss": 3.2472, "step": 2016 }, { "epoch": 0.5480978260869566, "grad_norm": 1.2559678554534912, "learning_rate": 3e-05, "loss": 2.8599, "step": 2017 }, { "epoch": 0.5483695652173913, "grad_norm": 1.272606372833252, "learning_rate": 3e-05, "loss": 3.7284, "step": 2018 }, { "epoch": 0.5486413043478261, "grad_norm": 1.098386287689209, "learning_rate": 3e-05, "loss": 3.0488, "step": 2019 }, { "epoch": 0.5489130434782609, "grad_norm": 1.707297444343567, "learning_rate": 3e-05, "loss": 3.122, "step": 2020 }, { "epoch": 0.5491847826086956, "grad_norm": 1.7398796081542969, "learning_rate": 3e-05, "loss": 3.9982, "step": 2021 }, { "epoch": 0.5494565217391304, "grad_norm": 1.1505324840545654, "learning_rate": 3e-05, "loss": 3.6328, "step": 2022 }, { "epoch": 0.5497282608695652, "grad_norm": 0.978933572769165, "learning_rate": 3e-05, "loss": 2.8985, "step": 2023 }, { "epoch": 0.55, "grad_norm": 1.2565099000930786, "learning_rate": 3e-05, "loss": 3.1667, "step": 2024 }, { "epoch": 0.5502717391304348, "grad_norm": 1.2417691946029663, "learning_rate": 3e-05, "loss": 3.5036, "step": 2025 }, { "epoch": 0.5505434782608696, "grad_norm": 1.2185314893722534, "learning_rate": 3e-05, "loss": 3.1685, "step": 2026 }, { "epoch": 0.5508152173913043, "grad_norm": 1.4619181156158447, "learning_rate": 3e-05, "loss": 4.2336, "step": 2027 }, { "epoch": 0.5510869565217391, "grad_norm": 1.2455370426177979, "learning_rate": 3e-05, "loss": 3.6007, "step": 2028 }, { "epoch": 0.5513586956521739, "grad_norm": 1.4226831197738647, "learning_rate": 3e-05, "loss": 3.7129, "step": 2029 }, { "epoch": 0.5516304347826086, "grad_norm": 1.2961459159851074, "learning_rate": 3e-05, "loss": 3.2041, "step": 2030 }, { "epoch": 0.5519021739130435, "grad_norm": 1.259477138519287, "learning_rate": 3e-05, "loss": 3.1745, "step": 2031 }, { "epoch": 0.5521739130434783, "grad_norm": 1.691286325454712, "learning_rate": 3e-05, "loss": 3.5502, "step": 2032 }, { "epoch": 0.5524456521739131, "grad_norm": 1.3523718118667603, "learning_rate": 3e-05, "loss": 3.4055, "step": 2033 }, { "epoch": 0.5527173913043478, "grad_norm": 1.2605832815170288, "learning_rate": 3e-05, "loss": 2.7262, "step": 2034 }, { "epoch": 0.5529891304347826, "grad_norm": 1.440855860710144, "learning_rate": 3e-05, "loss": 2.9385, "step": 2035 }, { "epoch": 0.5532608695652174, "grad_norm": 2.026885986328125, "learning_rate": 3e-05, "loss": 4.3039, "step": 2036 }, { "epoch": 0.5535326086956521, "grad_norm": 1.1887251138687134, "learning_rate": 3e-05, "loss": 2.8882, "step": 2037 }, { "epoch": 0.553804347826087, "grad_norm": 1.322061538696289, "learning_rate": 3e-05, "loss": 3.2703, "step": 2038 }, { "epoch": 0.5540760869565218, "grad_norm": 1.3264415264129639, "learning_rate": 3e-05, "loss": 3.2199, "step": 2039 }, { "epoch": 0.5543478260869565, "grad_norm": 1.6366621255874634, "learning_rate": 3e-05, "loss": 3.6402, "step": 2040 }, { "epoch": 0.5546195652173913, "grad_norm": 1.592647910118103, "learning_rate": 3e-05, "loss": 3.2358, "step": 2041 }, { "epoch": 0.5548913043478261, "grad_norm": 1.4537502527236938, "learning_rate": 3e-05, "loss": 3.7269, "step": 2042 }, { "epoch": 0.5551630434782608, "grad_norm": 1.3483185768127441, "learning_rate": 3e-05, "loss": 3.2456, "step": 2043 }, { "epoch": 0.5554347826086956, "grad_norm": 1.271835207939148, "learning_rate": 3e-05, "loss": 3.5684, "step": 2044 }, { "epoch": 0.5557065217391305, "grad_norm": 1.2505134344100952, "learning_rate": 3e-05, "loss": 3.2272, "step": 2045 }, { "epoch": 0.5559782608695653, "grad_norm": 1.3669381141662598, "learning_rate": 3e-05, "loss": 2.8729, "step": 2046 }, { "epoch": 0.55625, "grad_norm": 1.1156485080718994, "learning_rate": 3e-05, "loss": 3.003, "step": 2047 }, { "epoch": 0.5565217391304348, "grad_norm": 1.458754301071167, "learning_rate": 3e-05, "loss": 3.0448, "step": 2048 }, { "epoch": 0.5567934782608696, "grad_norm": 1.1323773860931396, "learning_rate": 3e-05, "loss": 3.1986, "step": 2049 }, { "epoch": 0.5570652173913043, "grad_norm": 1.1786192655563354, "learning_rate": 3e-05, "loss": 3.6913, "step": 2050 }, { "epoch": 0.5573369565217391, "grad_norm": 1.2700916528701782, "learning_rate": 3e-05, "loss": 2.9963, "step": 2051 }, { "epoch": 0.5576086956521739, "grad_norm": 1.21451735496521, "learning_rate": 3e-05, "loss": 3.0128, "step": 2052 }, { "epoch": 0.5578804347826087, "grad_norm": 1.5701782703399658, "learning_rate": 3e-05, "loss": 2.8858, "step": 2053 }, { "epoch": 0.5581521739130435, "grad_norm": 1.7219531536102295, "learning_rate": 3e-05, "loss": 2.8876, "step": 2054 }, { "epoch": 0.5584239130434783, "grad_norm": 1.694474220275879, "learning_rate": 3e-05, "loss": 3.1169, "step": 2055 }, { "epoch": 0.558695652173913, "grad_norm": 1.684638500213623, "learning_rate": 3e-05, "loss": 3.9718, "step": 2056 }, { "epoch": 0.5589673913043478, "grad_norm": 1.8228230476379395, "learning_rate": 3e-05, "loss": 3.0664, "step": 2057 }, { "epoch": 0.5592391304347826, "grad_norm": 1.8484151363372803, "learning_rate": 3e-05, "loss": 3.148, "step": 2058 }, { "epoch": 0.5595108695652173, "grad_norm": 1.6520676612854004, "learning_rate": 3e-05, "loss": 3.6549, "step": 2059 }, { "epoch": 0.5597826086956522, "grad_norm": 1.4072345495224, "learning_rate": 3e-05, "loss": 2.7737, "step": 2060 }, { "epoch": 0.560054347826087, "grad_norm": 1.459381341934204, "learning_rate": 3e-05, "loss": 3.2269, "step": 2061 }, { "epoch": 0.5603260869565218, "grad_norm": 1.6240458488464355, "learning_rate": 3e-05, "loss": 3.2216, "step": 2062 }, { "epoch": 0.5605978260869565, "grad_norm": 1.3474336862564087, "learning_rate": 3e-05, "loss": 3.191, "step": 2063 }, { "epoch": 0.5608695652173913, "grad_norm": 1.2936657667160034, "learning_rate": 3e-05, "loss": 3.1124, "step": 2064 }, { "epoch": 0.561141304347826, "grad_norm": 1.4614254236221313, "learning_rate": 3e-05, "loss": 3.5676, "step": 2065 }, { "epoch": 0.5614130434782608, "grad_norm": 1.4121429920196533, "learning_rate": 3e-05, "loss": 2.9918, "step": 2066 }, { "epoch": 0.5616847826086957, "grad_norm": 1.8185652494430542, "learning_rate": 3e-05, "loss": 3.2873, "step": 2067 }, { "epoch": 0.5619565217391305, "grad_norm": 1.7184209823608398, "learning_rate": 3e-05, "loss": 3.3629, "step": 2068 }, { "epoch": 0.5622282608695652, "grad_norm": 1.2696248292922974, "learning_rate": 3e-05, "loss": 2.8962, "step": 2069 }, { "epoch": 0.5625, "grad_norm": 1.2504596710205078, "learning_rate": 3e-05, "loss": 3.111, "step": 2070 }, { "epoch": 0.5627717391304348, "grad_norm": 1.2336483001708984, "learning_rate": 3e-05, "loss": 3.1268, "step": 2071 }, { "epoch": 0.5630434782608695, "grad_norm": 1.420301914215088, "learning_rate": 3e-05, "loss": 3.0844, "step": 2072 }, { "epoch": 0.5633152173913043, "grad_norm": 1.4836974143981934, "learning_rate": 3e-05, "loss": 3.305, "step": 2073 }, { "epoch": 0.5635869565217392, "grad_norm": 1.4846380949020386, "learning_rate": 3e-05, "loss": 2.9409, "step": 2074 }, { "epoch": 0.563858695652174, "grad_norm": 1.5037497282028198, "learning_rate": 3e-05, "loss": 3.691, "step": 2075 }, { "epoch": 0.5641304347826087, "grad_norm": 1.1565651893615723, "learning_rate": 3e-05, "loss": 3.678, "step": 2076 }, { "epoch": 0.5644021739130435, "grad_norm": 1.3678765296936035, "learning_rate": 3e-05, "loss": 3.7943, "step": 2077 }, { "epoch": 0.5646739130434782, "grad_norm": 1.6229230165481567, "learning_rate": 3e-05, "loss": 3.6795, "step": 2078 }, { "epoch": 0.564945652173913, "grad_norm": 1.3019323348999023, "learning_rate": 3e-05, "loss": 3.2213, "step": 2079 }, { "epoch": 0.5652173913043478, "grad_norm": 1.5816229581832886, "learning_rate": 3e-05, "loss": 3.9553, "step": 2080 }, { "epoch": 0.5654891304347827, "grad_norm": 1.5422357320785522, "learning_rate": 3e-05, "loss": 4.4725, "step": 2081 }, { "epoch": 0.5657608695652174, "grad_norm": 1.4155722856521606, "learning_rate": 3e-05, "loss": 3.6369, "step": 2082 }, { "epoch": 0.5660326086956522, "grad_norm": 1.52280592918396, "learning_rate": 3e-05, "loss": 3.205, "step": 2083 }, { "epoch": 0.566304347826087, "grad_norm": 1.4019720554351807, "learning_rate": 3e-05, "loss": 3.5101, "step": 2084 }, { "epoch": 0.5665760869565217, "grad_norm": 1.2434489727020264, "learning_rate": 3e-05, "loss": 3.5674, "step": 2085 }, { "epoch": 0.5668478260869565, "grad_norm": 1.4792324304580688, "learning_rate": 3e-05, "loss": 2.8302, "step": 2086 }, { "epoch": 0.5671195652173913, "grad_norm": 2.289886236190796, "learning_rate": 3e-05, "loss": 3.8648, "step": 2087 }, { "epoch": 0.5673913043478261, "grad_norm": 1.3174787759780884, "learning_rate": 3e-05, "loss": 3.2811, "step": 2088 }, { "epoch": 0.5676630434782609, "grad_norm": 1.2132010459899902, "learning_rate": 3e-05, "loss": 3.6173, "step": 2089 }, { "epoch": 0.5679347826086957, "grad_norm": 1.528756856918335, "learning_rate": 3e-05, "loss": 3.5117, "step": 2090 }, { "epoch": 0.5682065217391304, "grad_norm": 1.3807978630065918, "learning_rate": 3e-05, "loss": 3.0206, "step": 2091 }, { "epoch": 0.5684782608695652, "grad_norm": 1.0497171878814697, "learning_rate": 3e-05, "loss": 3.2278, "step": 2092 }, { "epoch": 0.56875, "grad_norm": 1.6013227701187134, "learning_rate": 3e-05, "loss": 3.4125, "step": 2093 }, { "epoch": 0.5690217391304347, "grad_norm": 0.9287932515144348, "learning_rate": 3e-05, "loss": 3.1837, "step": 2094 }, { "epoch": 0.5692934782608695, "grad_norm": 1.1170742511749268, "learning_rate": 3e-05, "loss": 3.5987, "step": 2095 }, { "epoch": 0.5695652173913044, "grad_norm": 0.9773249626159668, "learning_rate": 3e-05, "loss": 3.4219, "step": 2096 }, { "epoch": 0.5698369565217392, "grad_norm": 0.9633996486663818, "learning_rate": 3e-05, "loss": 3.3764, "step": 2097 }, { "epoch": 0.5701086956521739, "grad_norm": 1.2055789232254028, "learning_rate": 3e-05, "loss": 3.3378, "step": 2098 }, { "epoch": 0.5703804347826087, "grad_norm": 0.9966376423835754, "learning_rate": 3e-05, "loss": 2.8607, "step": 2099 }, { "epoch": 0.5706521739130435, "grad_norm": 1.2070221900939941, "learning_rate": 3e-05, "loss": 2.7183, "step": 2100 }, { "epoch": 0.5709239130434782, "grad_norm": 1.4068926572799683, "learning_rate": 3e-05, "loss": 3.8481, "step": 2101 }, { "epoch": 0.571195652173913, "grad_norm": 1.5947731733322144, "learning_rate": 3e-05, "loss": 2.9881, "step": 2102 }, { "epoch": 0.5714673913043479, "grad_norm": 1.2691866159439087, "learning_rate": 3e-05, "loss": 3.4337, "step": 2103 }, { "epoch": 0.5717391304347826, "grad_norm": 1.850817322731018, "learning_rate": 3e-05, "loss": 3.4459, "step": 2104 }, { "epoch": 0.5720108695652174, "grad_norm": 1.5024229288101196, "learning_rate": 3e-05, "loss": 3.1217, "step": 2105 }, { "epoch": 0.5722826086956522, "grad_norm": 1.167541742324829, "learning_rate": 3e-05, "loss": 3.584, "step": 2106 }, { "epoch": 0.5725543478260869, "grad_norm": 1.89936363697052, "learning_rate": 3e-05, "loss": 3.7818, "step": 2107 }, { "epoch": 0.5728260869565217, "grad_norm": 1.5929173231124878, "learning_rate": 3e-05, "loss": 3.4213, "step": 2108 }, { "epoch": 0.5730978260869565, "grad_norm": 1.122673749923706, "learning_rate": 3e-05, "loss": 3.6769, "step": 2109 }, { "epoch": 0.5733695652173914, "grad_norm": 1.0892047882080078, "learning_rate": 3e-05, "loss": 3.2629, "step": 2110 }, { "epoch": 0.5736413043478261, "grad_norm": 0.9661053419113159, "learning_rate": 3e-05, "loss": 3.0303, "step": 2111 }, { "epoch": 0.5739130434782609, "grad_norm": 1.2852338552474976, "learning_rate": 3e-05, "loss": 3.6586, "step": 2112 }, { "epoch": 0.5741847826086957, "grad_norm": 1.211471676826477, "learning_rate": 3e-05, "loss": 3.1448, "step": 2113 }, { "epoch": 0.5744565217391304, "grad_norm": 1.1296757459640503, "learning_rate": 3e-05, "loss": 2.9875, "step": 2114 }, { "epoch": 0.5747282608695652, "grad_norm": 1.3020541667938232, "learning_rate": 3e-05, "loss": 3.111, "step": 2115 }, { "epoch": 0.575, "grad_norm": 1.0208520889282227, "learning_rate": 3e-05, "loss": 2.7882, "step": 2116 }, { "epoch": 0.5752717391304348, "grad_norm": 1.4555299282073975, "learning_rate": 3e-05, "loss": 3.3721, "step": 2117 }, { "epoch": 0.5755434782608696, "grad_norm": 2.195035934448242, "learning_rate": 3e-05, "loss": 4.1556, "step": 2118 }, { "epoch": 0.5758152173913044, "grad_norm": 1.4949889183044434, "learning_rate": 3e-05, "loss": 2.6564, "step": 2119 }, { "epoch": 0.5760869565217391, "grad_norm": 1.1324836015701294, "learning_rate": 3e-05, "loss": 3.1073, "step": 2120 }, { "epoch": 0.5763586956521739, "grad_norm": 1.3076951503753662, "learning_rate": 3e-05, "loss": 3.6178, "step": 2121 }, { "epoch": 0.5766304347826087, "grad_norm": 1.322385311126709, "learning_rate": 3e-05, "loss": 3.9284, "step": 2122 }, { "epoch": 0.5769021739130434, "grad_norm": 1.567959189414978, "learning_rate": 3e-05, "loss": 4.0229, "step": 2123 }, { "epoch": 0.5771739130434783, "grad_norm": 1.4464517831802368, "learning_rate": 3e-05, "loss": 3.3669, "step": 2124 }, { "epoch": 0.5774456521739131, "grad_norm": 1.5971894264221191, "learning_rate": 3e-05, "loss": 4.1792, "step": 2125 }, { "epoch": 0.5777173913043478, "grad_norm": 1.4091331958770752, "learning_rate": 3e-05, "loss": 3.3196, "step": 2126 }, { "epoch": 0.5779891304347826, "grad_norm": 1.4950571060180664, "learning_rate": 3e-05, "loss": 3.6238, "step": 2127 }, { "epoch": 0.5782608695652174, "grad_norm": 1.0344115495681763, "learning_rate": 3e-05, "loss": 2.8622, "step": 2128 }, { "epoch": 0.5785326086956522, "grad_norm": 1.1362147331237793, "learning_rate": 3e-05, "loss": 2.7986, "step": 2129 }, { "epoch": 0.5788043478260869, "grad_norm": 1.1396723985671997, "learning_rate": 3e-05, "loss": 2.7951, "step": 2130 }, { "epoch": 0.5790760869565217, "grad_norm": 1.355010747909546, "learning_rate": 3e-05, "loss": 3.1708, "step": 2131 }, { "epoch": 0.5793478260869566, "grad_norm": 1.5592600107192993, "learning_rate": 3e-05, "loss": 3.5758, "step": 2132 }, { "epoch": 0.5796195652173913, "grad_norm": 1.56844961643219, "learning_rate": 3e-05, "loss": 3.4122, "step": 2133 }, { "epoch": 0.5798913043478261, "grad_norm": 1.3906667232513428, "learning_rate": 3e-05, "loss": 3.7639, "step": 2134 }, { "epoch": 0.5801630434782609, "grad_norm": 1.131856918334961, "learning_rate": 3e-05, "loss": 3.3024, "step": 2135 }, { "epoch": 0.5804347826086956, "grad_norm": 1.2349239587783813, "learning_rate": 3e-05, "loss": 3.2338, "step": 2136 }, { "epoch": 0.5807065217391304, "grad_norm": 1.320646047592163, "learning_rate": 3e-05, "loss": 3.2089, "step": 2137 }, { "epoch": 0.5809782608695652, "grad_norm": 1.0962297916412354, "learning_rate": 3e-05, "loss": 2.9312, "step": 2138 }, { "epoch": 0.58125, "grad_norm": 1.2109733819961548, "learning_rate": 3e-05, "loss": 2.6762, "step": 2139 }, { "epoch": 0.5815217391304348, "grad_norm": 1.3556448221206665, "learning_rate": 3e-05, "loss": 3.6022, "step": 2140 }, { "epoch": 0.5817934782608696, "grad_norm": 1.216602087020874, "learning_rate": 3e-05, "loss": 3.9107, "step": 2141 }, { "epoch": 0.5820652173913043, "grad_norm": 1.2305996417999268, "learning_rate": 3e-05, "loss": 3.2447, "step": 2142 }, { "epoch": 0.5823369565217391, "grad_norm": 1.2739956378936768, "learning_rate": 3e-05, "loss": 3.4232, "step": 2143 }, { "epoch": 0.5826086956521739, "grad_norm": 1.2404019832611084, "learning_rate": 3e-05, "loss": 3.467, "step": 2144 }, { "epoch": 0.5828804347826086, "grad_norm": 1.3793765306472778, "learning_rate": 3e-05, "loss": 3.4627, "step": 2145 }, { "epoch": 0.5831521739130435, "grad_norm": 1.6223127841949463, "learning_rate": 3e-05, "loss": 3.2725, "step": 2146 }, { "epoch": 0.5834239130434783, "grad_norm": 1.055362343788147, "learning_rate": 3e-05, "loss": 3.2517, "step": 2147 }, { "epoch": 0.5836956521739131, "grad_norm": 1.2135682106018066, "learning_rate": 3e-05, "loss": 3.0423, "step": 2148 }, { "epoch": 0.5839673913043478, "grad_norm": 1.7276960611343384, "learning_rate": 3e-05, "loss": 4.2823, "step": 2149 }, { "epoch": 0.5842391304347826, "grad_norm": 1.1459511518478394, "learning_rate": 3e-05, "loss": 3.129, "step": 2150 }, { "epoch": 0.5845108695652174, "grad_norm": 1.7933846712112427, "learning_rate": 3e-05, "loss": 3.6957, "step": 2151 }, { "epoch": 0.5847826086956521, "grad_norm": 1.3888658285140991, "learning_rate": 3e-05, "loss": 4.2824, "step": 2152 }, { "epoch": 0.585054347826087, "grad_norm": 1.18975031375885, "learning_rate": 3e-05, "loss": 2.9272, "step": 2153 }, { "epoch": 0.5853260869565218, "grad_norm": 1.0910744667053223, "learning_rate": 3e-05, "loss": 3.4617, "step": 2154 }, { "epoch": 0.5855978260869565, "grad_norm": 1.3888293504714966, "learning_rate": 3e-05, "loss": 2.9697, "step": 2155 }, { "epoch": 0.5858695652173913, "grad_norm": 1.1816877126693726, "learning_rate": 3e-05, "loss": 3.5301, "step": 2156 }, { "epoch": 0.5861413043478261, "grad_norm": 1.2659231424331665, "learning_rate": 3e-05, "loss": 3.5646, "step": 2157 }, { "epoch": 0.5864130434782608, "grad_norm": 1.5958353281021118, "learning_rate": 3e-05, "loss": 3.378, "step": 2158 }, { "epoch": 0.5866847826086956, "grad_norm": 1.2585597038269043, "learning_rate": 3e-05, "loss": 3.0523, "step": 2159 }, { "epoch": 0.5869565217391305, "grad_norm": 1.0975054502487183, "learning_rate": 3e-05, "loss": 3.0933, "step": 2160 }, { "epoch": 0.5872282608695653, "grad_norm": 1.5233417749404907, "learning_rate": 3e-05, "loss": 3.2665, "step": 2161 }, { "epoch": 0.5875, "grad_norm": 2.02248215675354, "learning_rate": 3e-05, "loss": 3.754, "step": 2162 }, { "epoch": 0.5877717391304348, "grad_norm": 0.9959807395935059, "learning_rate": 3e-05, "loss": 2.9276, "step": 2163 }, { "epoch": 0.5880434782608696, "grad_norm": 1.1323211193084717, "learning_rate": 3e-05, "loss": 3.0789, "step": 2164 }, { "epoch": 0.5883152173913043, "grad_norm": 1.0242100954055786, "learning_rate": 3e-05, "loss": 3.3039, "step": 2165 }, { "epoch": 0.5885869565217391, "grad_norm": 1.0082273483276367, "learning_rate": 3e-05, "loss": 3.3631, "step": 2166 }, { "epoch": 0.5888586956521739, "grad_norm": 1.1713404655456543, "learning_rate": 3e-05, "loss": 2.7448, "step": 2167 }, { "epoch": 0.5891304347826087, "grad_norm": 1.1318339109420776, "learning_rate": 3e-05, "loss": 3.3552, "step": 2168 }, { "epoch": 0.5894021739130435, "grad_norm": 1.3056354522705078, "learning_rate": 3e-05, "loss": 3.8349, "step": 2169 }, { "epoch": 0.5896739130434783, "grad_norm": 1.1058902740478516, "learning_rate": 3e-05, "loss": 2.9931, "step": 2170 }, { "epoch": 0.589945652173913, "grad_norm": 1.3379881381988525, "learning_rate": 3e-05, "loss": 3.0709, "step": 2171 }, { "epoch": 0.5902173913043478, "grad_norm": 1.3458565473556519, "learning_rate": 3e-05, "loss": 3.5521, "step": 2172 }, { "epoch": 0.5904891304347826, "grad_norm": 1.2749676704406738, "learning_rate": 3e-05, "loss": 3.3535, "step": 2173 }, { "epoch": 0.5907608695652173, "grad_norm": 1.3268214464187622, "learning_rate": 3e-05, "loss": 3.6918, "step": 2174 }, { "epoch": 0.5910326086956522, "grad_norm": 1.4857879877090454, "learning_rate": 3e-05, "loss": 3.6046, "step": 2175 }, { "epoch": 0.591304347826087, "grad_norm": 1.009945034980774, "learning_rate": 3e-05, "loss": 3.435, "step": 2176 }, { "epoch": 0.5915760869565218, "grad_norm": 1.1983013153076172, "learning_rate": 3e-05, "loss": 3.5871, "step": 2177 }, { "epoch": 0.5918478260869565, "grad_norm": 0.9718300700187683, "learning_rate": 3e-05, "loss": 3.3072, "step": 2178 }, { "epoch": 0.5921195652173913, "grad_norm": 1.101165771484375, "learning_rate": 3e-05, "loss": 3.2276, "step": 2179 }, { "epoch": 0.592391304347826, "grad_norm": 1.012749433517456, "learning_rate": 3e-05, "loss": 3.1112, "step": 2180 }, { "epoch": 0.5926630434782608, "grad_norm": 0.9688409566879272, "learning_rate": 3e-05, "loss": 3.0918, "step": 2181 }, { "epoch": 0.5929347826086957, "grad_norm": 0.9854469895362854, "learning_rate": 3e-05, "loss": 3.1777, "step": 2182 }, { "epoch": 0.5932065217391305, "grad_norm": 0.9278433918952942, "learning_rate": 3e-05, "loss": 3.1989, "step": 2183 }, { "epoch": 0.5934782608695652, "grad_norm": 0.983065664768219, "learning_rate": 3e-05, "loss": 3.3761, "step": 2184 }, { "epoch": 0.59375, "grad_norm": 1.473644733428955, "learning_rate": 3e-05, "loss": 3.6095, "step": 2185 }, { "epoch": 0.5940217391304348, "grad_norm": 1.2839797735214233, "learning_rate": 3e-05, "loss": 2.9857, "step": 2186 }, { "epoch": 0.5942934782608695, "grad_norm": 1.7509822845458984, "learning_rate": 3e-05, "loss": 3.7711, "step": 2187 }, { "epoch": 0.5945652173913043, "grad_norm": 1.171937346458435, "learning_rate": 3e-05, "loss": 3.6335, "step": 2188 }, { "epoch": 0.5948369565217392, "grad_norm": 1.4512766599655151, "learning_rate": 3e-05, "loss": 2.9629, "step": 2189 }, { "epoch": 0.595108695652174, "grad_norm": 1.0719918012619019, "learning_rate": 3e-05, "loss": 2.8943, "step": 2190 }, { "epoch": 0.5953804347826087, "grad_norm": 1.11574387550354, "learning_rate": 3e-05, "loss": 3.7779, "step": 2191 }, { "epoch": 0.5956521739130435, "grad_norm": 1.6633427143096924, "learning_rate": 3e-05, "loss": 3.7736, "step": 2192 }, { "epoch": 0.5959239130434782, "grad_norm": 1.1903231143951416, "learning_rate": 3e-05, "loss": 2.9944, "step": 2193 }, { "epoch": 0.596195652173913, "grad_norm": 1.2331459522247314, "learning_rate": 3e-05, "loss": 3.639, "step": 2194 }, { "epoch": 0.5964673913043478, "grad_norm": 1.8560791015625, "learning_rate": 3e-05, "loss": 4.0684, "step": 2195 }, { "epoch": 0.5967391304347827, "grad_norm": 1.2733112573623657, "learning_rate": 3e-05, "loss": 3.1905, "step": 2196 }, { "epoch": 0.5970108695652174, "grad_norm": 1.156823992729187, "learning_rate": 3e-05, "loss": 3.3605, "step": 2197 }, { "epoch": 0.5972826086956522, "grad_norm": 1.0290706157684326, "learning_rate": 3e-05, "loss": 3.2214, "step": 2198 }, { "epoch": 0.597554347826087, "grad_norm": 1.1795909404754639, "learning_rate": 3e-05, "loss": 2.9137, "step": 2199 }, { "epoch": 0.5978260869565217, "grad_norm": 1.3558768033981323, "learning_rate": 3e-05, "loss": 3.6129, "step": 2200 }, { "epoch": 0.5980978260869565, "grad_norm": 0.9279618859291077, "learning_rate": 3e-05, "loss": 2.5655, "step": 2201 }, { "epoch": 0.5983695652173913, "grad_norm": 1.2650532722473145, "learning_rate": 3e-05, "loss": 2.8838, "step": 2202 }, { "epoch": 0.5986413043478261, "grad_norm": 1.2165402173995972, "learning_rate": 3e-05, "loss": 3.5675, "step": 2203 }, { "epoch": 0.5989130434782609, "grad_norm": 1.1845487356185913, "learning_rate": 3e-05, "loss": 2.9909, "step": 2204 }, { "epoch": 0.5991847826086957, "grad_norm": 1.1865789890289307, "learning_rate": 3e-05, "loss": 3.6638, "step": 2205 }, { "epoch": 0.5994565217391304, "grad_norm": 0.9880366921424866, "learning_rate": 3e-05, "loss": 3.023, "step": 2206 }, { "epoch": 0.5997282608695652, "grad_norm": 1.1242929697036743, "learning_rate": 3e-05, "loss": 3.1062, "step": 2207 }, { "epoch": 0.6, "grad_norm": 1.1712896823883057, "learning_rate": 3e-05, "loss": 3.029, "step": 2208 }, { "epoch": 0.6002717391304347, "grad_norm": 1.5733115673065186, "learning_rate": 3e-05, "loss": 3.7851, "step": 2209 }, { "epoch": 0.6005434782608695, "grad_norm": 1.2813576459884644, "learning_rate": 3e-05, "loss": 3.4965, "step": 2210 }, { "epoch": 0.6008152173913044, "grad_norm": 1.3263084888458252, "learning_rate": 3e-05, "loss": 4.1356, "step": 2211 }, { "epoch": 0.6010869565217392, "grad_norm": 1.1176859140396118, "learning_rate": 3e-05, "loss": 3.4671, "step": 2212 }, { "epoch": 0.6013586956521739, "grad_norm": 1.2956467866897583, "learning_rate": 3e-05, "loss": 3.8482, "step": 2213 }, { "epoch": 0.6016304347826087, "grad_norm": 1.8947595357894897, "learning_rate": 3e-05, "loss": 3.2963, "step": 2214 }, { "epoch": 0.6019021739130435, "grad_norm": 1.3561947345733643, "learning_rate": 3e-05, "loss": 3.1028, "step": 2215 }, { "epoch": 0.6021739130434782, "grad_norm": 1.1945207118988037, "learning_rate": 3e-05, "loss": 3.2949, "step": 2216 }, { "epoch": 0.602445652173913, "grad_norm": 1.3198587894439697, "learning_rate": 3e-05, "loss": 3.5389, "step": 2217 }, { "epoch": 0.6027173913043479, "grad_norm": 1.6522905826568604, "learning_rate": 3e-05, "loss": 3.2054, "step": 2218 }, { "epoch": 0.6029891304347826, "grad_norm": 1.2735859155654907, "learning_rate": 3e-05, "loss": 3.4955, "step": 2219 }, { "epoch": 0.6032608695652174, "grad_norm": 1.2329673767089844, "learning_rate": 3e-05, "loss": 3.7713, "step": 2220 }, { "epoch": 0.6035326086956522, "grad_norm": 1.1366288661956787, "learning_rate": 3e-05, "loss": 3.333, "step": 2221 }, { "epoch": 0.6038043478260869, "grad_norm": 1.0295037031173706, "learning_rate": 3e-05, "loss": 3.2991, "step": 2222 }, { "epoch": 0.6040760869565217, "grad_norm": 1.0760974884033203, "learning_rate": 3e-05, "loss": 3.6266, "step": 2223 }, { "epoch": 0.6043478260869565, "grad_norm": 1.0044188499450684, "learning_rate": 3e-05, "loss": 3.1854, "step": 2224 }, { "epoch": 0.6046195652173914, "grad_norm": 0.79954993724823, "learning_rate": 3e-05, "loss": 2.5833, "step": 2225 }, { "epoch": 0.6048913043478261, "grad_norm": 1.3376729488372803, "learning_rate": 3e-05, "loss": 3.8299, "step": 2226 }, { "epoch": 0.6051630434782609, "grad_norm": 1.3323661088943481, "learning_rate": 3e-05, "loss": 3.5493, "step": 2227 }, { "epoch": 0.6054347826086957, "grad_norm": 0.9946414828300476, "learning_rate": 3e-05, "loss": 3.0806, "step": 2228 }, { "epoch": 0.6057065217391304, "grad_norm": 1.486099123954773, "learning_rate": 3e-05, "loss": 3.5046, "step": 2229 }, { "epoch": 0.6059782608695652, "grad_norm": 1.4763749837875366, "learning_rate": 3e-05, "loss": 3.2249, "step": 2230 }, { "epoch": 0.60625, "grad_norm": 1.1273298263549805, "learning_rate": 3e-05, "loss": 3.097, "step": 2231 }, { "epoch": 0.6065217391304348, "grad_norm": 1.273410677909851, "learning_rate": 3e-05, "loss": 3.3963, "step": 2232 }, { "epoch": 0.6067934782608696, "grad_norm": 1.6567277908325195, "learning_rate": 3e-05, "loss": 3.4074, "step": 2233 }, { "epoch": 0.6070652173913044, "grad_norm": 1.272204875946045, "learning_rate": 3e-05, "loss": 3.5062, "step": 2234 }, { "epoch": 0.6073369565217391, "grad_norm": 0.9623717665672302, "learning_rate": 3e-05, "loss": 3.348, "step": 2235 }, { "epoch": 0.6076086956521739, "grad_norm": 1.7306931018829346, "learning_rate": 3e-05, "loss": 3.5976, "step": 2236 }, { "epoch": 0.6078804347826087, "grad_norm": 1.2716137170791626, "learning_rate": 3e-05, "loss": 3.8244, "step": 2237 }, { "epoch": 0.6081521739130434, "grad_norm": 1.3230034112930298, "learning_rate": 3e-05, "loss": 3.4558, "step": 2238 }, { "epoch": 0.6084239130434783, "grad_norm": 1.733973503112793, "learning_rate": 3e-05, "loss": 3.0608, "step": 2239 }, { "epoch": 0.6086956521739131, "grad_norm": 1.2597112655639648, "learning_rate": 3e-05, "loss": 2.8695, "step": 2240 }, { "epoch": 0.6089673913043478, "grad_norm": 1.033078670501709, "learning_rate": 3e-05, "loss": 2.528, "step": 2241 }, { "epoch": 0.6092391304347826, "grad_norm": 1.4419554471969604, "learning_rate": 3e-05, "loss": 3.5569, "step": 2242 }, { "epoch": 0.6095108695652174, "grad_norm": 1.376265287399292, "learning_rate": 3e-05, "loss": 3.4915, "step": 2243 }, { "epoch": 0.6097826086956522, "grad_norm": 1.1557855606079102, "learning_rate": 3e-05, "loss": 3.4314, "step": 2244 }, { "epoch": 0.6100543478260869, "grad_norm": 1.3031344413757324, "learning_rate": 3e-05, "loss": 3.5682, "step": 2245 }, { "epoch": 0.6103260869565217, "grad_norm": 1.2761789560317993, "learning_rate": 3e-05, "loss": 3.4629, "step": 2246 }, { "epoch": 0.6105978260869566, "grad_norm": 1.1963956356048584, "learning_rate": 3e-05, "loss": 3.2122, "step": 2247 }, { "epoch": 0.6108695652173913, "grad_norm": 1.357904314994812, "learning_rate": 3e-05, "loss": 3.0868, "step": 2248 }, { "epoch": 0.6111413043478261, "grad_norm": 1.3014525175094604, "learning_rate": 3e-05, "loss": 2.9693, "step": 2249 }, { "epoch": 0.6114130434782609, "grad_norm": 1.3314582109451294, "learning_rate": 3e-05, "loss": 3.3714, "step": 2250 }, { "epoch": 0.6116847826086956, "grad_norm": 1.3849120140075684, "learning_rate": 3e-05, "loss": 2.7755, "step": 2251 }, { "epoch": 0.6119565217391304, "grad_norm": 0.9548225402832031, "learning_rate": 3e-05, "loss": 3.0885, "step": 2252 }, { "epoch": 0.6122282608695652, "grad_norm": 1.4415459632873535, "learning_rate": 3e-05, "loss": 3.8783, "step": 2253 }, { "epoch": 0.6125, "grad_norm": 1.7835477590560913, "learning_rate": 3e-05, "loss": 3.6217, "step": 2254 }, { "epoch": 0.6127717391304348, "grad_norm": 1.4011732339859009, "learning_rate": 3e-05, "loss": 3.9328, "step": 2255 }, { "epoch": 0.6130434782608696, "grad_norm": 1.0685231685638428, "learning_rate": 3e-05, "loss": 3.2949, "step": 2256 }, { "epoch": 0.6133152173913043, "grad_norm": 1.2185416221618652, "learning_rate": 3e-05, "loss": 3.5132, "step": 2257 }, { "epoch": 0.6135869565217391, "grad_norm": 1.0910768508911133, "learning_rate": 3e-05, "loss": 3.0972, "step": 2258 }, { "epoch": 0.6138586956521739, "grad_norm": 1.0861088037490845, "learning_rate": 3e-05, "loss": 3.3816, "step": 2259 }, { "epoch": 0.6141304347826086, "grad_norm": 1.0227428674697876, "learning_rate": 3e-05, "loss": 3.1733, "step": 2260 }, { "epoch": 0.6144021739130435, "grad_norm": 1.3343944549560547, "learning_rate": 3e-05, "loss": 3.2618, "step": 2261 }, { "epoch": 0.6146739130434783, "grad_norm": 1.3029199838638306, "learning_rate": 3e-05, "loss": 3.4307, "step": 2262 }, { "epoch": 0.6149456521739131, "grad_norm": 1.152552843093872, "learning_rate": 3e-05, "loss": 3.4878, "step": 2263 }, { "epoch": 0.6152173913043478, "grad_norm": 1.5189865827560425, "learning_rate": 3e-05, "loss": 3.4252, "step": 2264 }, { "epoch": 0.6154891304347826, "grad_norm": 1.1164250373840332, "learning_rate": 3e-05, "loss": 3.5027, "step": 2265 }, { "epoch": 0.6157608695652174, "grad_norm": 0.9906913638114929, "learning_rate": 3e-05, "loss": 2.9876, "step": 2266 }, { "epoch": 0.6160326086956521, "grad_norm": 1.1681506633758545, "learning_rate": 3e-05, "loss": 3.2722, "step": 2267 }, { "epoch": 0.616304347826087, "grad_norm": 1.183396816253662, "learning_rate": 3e-05, "loss": 3.2224, "step": 2268 }, { "epoch": 0.6165760869565218, "grad_norm": 1.0938737392425537, "learning_rate": 3e-05, "loss": 3.2871, "step": 2269 }, { "epoch": 0.6168478260869565, "grad_norm": 1.5161495208740234, "learning_rate": 3e-05, "loss": 3.7379, "step": 2270 }, { "epoch": 0.6171195652173913, "grad_norm": 1.2433841228485107, "learning_rate": 3e-05, "loss": 2.9899, "step": 2271 }, { "epoch": 0.6173913043478261, "grad_norm": 1.092914342880249, "learning_rate": 3e-05, "loss": 2.7804, "step": 2272 }, { "epoch": 0.6176630434782608, "grad_norm": 1.2224652767181396, "learning_rate": 3e-05, "loss": 3.628, "step": 2273 }, { "epoch": 0.6179347826086956, "grad_norm": 1.3725236654281616, "learning_rate": 3e-05, "loss": 3.6826, "step": 2274 }, { "epoch": 0.6182065217391305, "grad_norm": 1.0092295408248901, "learning_rate": 3e-05, "loss": 2.9988, "step": 2275 }, { "epoch": 0.6184782608695653, "grad_norm": 1.259446144104004, "learning_rate": 3e-05, "loss": 3.4234, "step": 2276 }, { "epoch": 0.61875, "grad_norm": 1.220897912979126, "learning_rate": 3e-05, "loss": 3.2565, "step": 2277 }, { "epoch": 0.6190217391304348, "grad_norm": 0.9857214093208313, "learning_rate": 3e-05, "loss": 3.2311, "step": 2278 }, { "epoch": 0.6192934782608696, "grad_norm": 1.256536841392517, "learning_rate": 3e-05, "loss": 3.4595, "step": 2279 }, { "epoch": 0.6195652173913043, "grad_norm": 1.0857690572738647, "learning_rate": 3e-05, "loss": 2.8696, "step": 2280 }, { "epoch": 0.6198369565217391, "grad_norm": 0.9030542969703674, "learning_rate": 3e-05, "loss": 2.931, "step": 2281 }, { "epoch": 0.6201086956521739, "grad_norm": 1.245595097541809, "learning_rate": 3e-05, "loss": 3.0565, "step": 2282 }, { "epoch": 0.6203804347826087, "grad_norm": 1.130018949508667, "learning_rate": 3e-05, "loss": 3.6569, "step": 2283 }, { "epoch": 0.6206521739130435, "grad_norm": 1.1791630983352661, "learning_rate": 3e-05, "loss": 2.9299, "step": 2284 }, { "epoch": 0.6209239130434783, "grad_norm": 1.6365814208984375, "learning_rate": 3e-05, "loss": 3.7201, "step": 2285 }, { "epoch": 0.621195652173913, "grad_norm": 1.023141622543335, "learning_rate": 3e-05, "loss": 3.4364, "step": 2286 }, { "epoch": 0.6214673913043478, "grad_norm": 1.3121358156204224, "learning_rate": 3e-05, "loss": 3.4811, "step": 2287 }, { "epoch": 0.6217391304347826, "grad_norm": 1.009194254875183, "learning_rate": 3e-05, "loss": 3.0576, "step": 2288 }, { "epoch": 0.6220108695652173, "grad_norm": 1.0715727806091309, "learning_rate": 3e-05, "loss": 3.193, "step": 2289 }, { "epoch": 0.6222826086956522, "grad_norm": 1.0925318002700806, "learning_rate": 3e-05, "loss": 3.1963, "step": 2290 }, { "epoch": 0.622554347826087, "grad_norm": 1.6540546417236328, "learning_rate": 3e-05, "loss": 3.7859, "step": 2291 }, { "epoch": 0.6228260869565218, "grad_norm": 1.0642199516296387, "learning_rate": 3e-05, "loss": 3.8112, "step": 2292 }, { "epoch": 0.6230978260869565, "grad_norm": 1.2560967206954956, "learning_rate": 3e-05, "loss": 3.528, "step": 2293 }, { "epoch": 0.6233695652173913, "grad_norm": 1.2816760540008545, "learning_rate": 3e-05, "loss": 3.1706, "step": 2294 }, { "epoch": 0.623641304347826, "grad_norm": 0.9735274314880371, "learning_rate": 3e-05, "loss": 3.2275, "step": 2295 }, { "epoch": 0.6239130434782608, "grad_norm": 1.1143537759780884, "learning_rate": 3e-05, "loss": 3.2471, "step": 2296 }, { "epoch": 0.6241847826086957, "grad_norm": 1.2723908424377441, "learning_rate": 3e-05, "loss": 2.8132, "step": 2297 }, { "epoch": 0.6244565217391305, "grad_norm": 1.269571304321289, "learning_rate": 3e-05, "loss": 3.2307, "step": 2298 }, { "epoch": 0.6247282608695652, "grad_norm": 1.0772123336791992, "learning_rate": 3e-05, "loss": 3.2448, "step": 2299 }, { "epoch": 0.625, "grad_norm": 1.5265069007873535, "learning_rate": 3e-05, "loss": 3.2314, "step": 2300 }, { "epoch": 0.6252717391304348, "grad_norm": 1.133500576019287, "learning_rate": 3e-05, "loss": 3.1259, "step": 2301 }, { "epoch": 0.6255434782608695, "grad_norm": 1.0310444831848145, "learning_rate": 3e-05, "loss": 3.1374, "step": 2302 }, { "epoch": 0.6258152173913043, "grad_norm": 1.789737582206726, "learning_rate": 3e-05, "loss": 3.657, "step": 2303 }, { "epoch": 0.6260869565217392, "grad_norm": 1.028426170349121, "learning_rate": 3e-05, "loss": 3.0235, "step": 2304 }, { "epoch": 0.626358695652174, "grad_norm": 1.0289982557296753, "learning_rate": 3e-05, "loss": 3.0216, "step": 2305 }, { "epoch": 0.6266304347826087, "grad_norm": 1.2136385440826416, "learning_rate": 3e-05, "loss": 3.2615, "step": 2306 }, { "epoch": 0.6269021739130435, "grad_norm": 1.1237839460372925, "learning_rate": 3e-05, "loss": 3.1519, "step": 2307 }, { "epoch": 0.6271739130434782, "grad_norm": 1.0916386842727661, "learning_rate": 3e-05, "loss": 3.404, "step": 2308 }, { "epoch": 0.627445652173913, "grad_norm": 1.1526830196380615, "learning_rate": 3e-05, "loss": 3.671, "step": 2309 }, { "epoch": 0.6277173913043478, "grad_norm": 1.0757122039794922, "learning_rate": 3e-05, "loss": 3.069, "step": 2310 }, { "epoch": 0.6279891304347827, "grad_norm": 1.0339561700820923, "learning_rate": 3e-05, "loss": 2.9702, "step": 2311 }, { "epoch": 0.6282608695652174, "grad_norm": 1.2000614404678345, "learning_rate": 3e-05, "loss": 2.9536, "step": 2312 }, { "epoch": 0.6285326086956522, "grad_norm": 1.3104127645492554, "learning_rate": 3e-05, "loss": 3.1653, "step": 2313 }, { "epoch": 0.628804347826087, "grad_norm": 1.2230424880981445, "learning_rate": 3e-05, "loss": 3.464, "step": 2314 }, { "epoch": 0.6290760869565217, "grad_norm": 1.258482813835144, "learning_rate": 3e-05, "loss": 3.4273, "step": 2315 }, { "epoch": 0.6293478260869565, "grad_norm": 1.1824918985366821, "learning_rate": 3e-05, "loss": 3.5311, "step": 2316 }, { "epoch": 0.6296195652173913, "grad_norm": 0.9511439800262451, "learning_rate": 3e-05, "loss": 2.979, "step": 2317 }, { "epoch": 0.6298913043478261, "grad_norm": 1.0357087850570679, "learning_rate": 3e-05, "loss": 2.8735, "step": 2318 }, { "epoch": 0.6301630434782609, "grad_norm": 1.392822504043579, "learning_rate": 3e-05, "loss": 3.3224, "step": 2319 }, { "epoch": 0.6304347826086957, "grad_norm": 1.4284133911132812, "learning_rate": 3e-05, "loss": 3.8771, "step": 2320 }, { "epoch": 0.6307065217391304, "grad_norm": 1.2343987226486206, "learning_rate": 3e-05, "loss": 3.4731, "step": 2321 }, { "epoch": 0.6309782608695652, "grad_norm": 1.3236418962478638, "learning_rate": 3e-05, "loss": 3.0388, "step": 2322 }, { "epoch": 0.63125, "grad_norm": 1.332114577293396, "learning_rate": 3e-05, "loss": 3.1262, "step": 2323 }, { "epoch": 0.6315217391304347, "grad_norm": 1.0655372142791748, "learning_rate": 3e-05, "loss": 3.2109, "step": 2324 }, { "epoch": 0.6317934782608695, "grad_norm": 1.0099713802337646, "learning_rate": 3e-05, "loss": 3.2809, "step": 2325 }, { "epoch": 0.6320652173913044, "grad_norm": 1.0744123458862305, "learning_rate": 3e-05, "loss": 2.862, "step": 2326 }, { "epoch": 0.6323369565217392, "grad_norm": 1.051435112953186, "learning_rate": 3e-05, "loss": 2.8705, "step": 2327 }, { "epoch": 0.6326086956521739, "grad_norm": 1.400758147239685, "learning_rate": 3e-05, "loss": 3.7773, "step": 2328 }, { "epoch": 0.6328804347826087, "grad_norm": 1.1519436836242676, "learning_rate": 3e-05, "loss": 2.9399, "step": 2329 }, { "epoch": 0.6331521739130435, "grad_norm": 1.264049768447876, "learning_rate": 3e-05, "loss": 3.7764, "step": 2330 }, { "epoch": 0.6334239130434782, "grad_norm": 1.0223599672317505, "learning_rate": 3e-05, "loss": 2.9021, "step": 2331 }, { "epoch": 0.633695652173913, "grad_norm": 1.255913257598877, "learning_rate": 3e-05, "loss": 3.6456, "step": 2332 }, { "epoch": 0.6339673913043479, "grad_norm": 1.7114654779434204, "learning_rate": 3e-05, "loss": 4.0404, "step": 2333 }, { "epoch": 0.6342391304347826, "grad_norm": 2.052917718887329, "learning_rate": 3e-05, "loss": 4.1781, "step": 2334 }, { "epoch": 0.6345108695652174, "grad_norm": 1.376478672027588, "learning_rate": 3e-05, "loss": 3.4097, "step": 2335 }, { "epoch": 0.6347826086956522, "grad_norm": 1.3323421478271484, "learning_rate": 3e-05, "loss": 3.0153, "step": 2336 }, { "epoch": 0.6350543478260869, "grad_norm": 1.7771837711334229, "learning_rate": 3e-05, "loss": 3.7671, "step": 2337 }, { "epoch": 0.6353260869565217, "grad_norm": 2.060122489929199, "learning_rate": 3e-05, "loss": 3.5854, "step": 2338 }, { "epoch": 0.6355978260869565, "grad_norm": 1.3856672048568726, "learning_rate": 3e-05, "loss": 3.0497, "step": 2339 }, { "epoch": 0.6358695652173914, "grad_norm": 1.3146904706954956, "learning_rate": 3e-05, "loss": 3.1533, "step": 2340 }, { "epoch": 0.6361413043478261, "grad_norm": 1.4848661422729492, "learning_rate": 3e-05, "loss": 3.5613, "step": 2341 }, { "epoch": 0.6364130434782609, "grad_norm": 1.4037156105041504, "learning_rate": 3e-05, "loss": 3.8576, "step": 2342 }, { "epoch": 0.6366847826086957, "grad_norm": 1.288294792175293, "learning_rate": 3e-05, "loss": 3.5626, "step": 2343 }, { "epoch": 0.6369565217391304, "grad_norm": 1.080474853515625, "learning_rate": 3e-05, "loss": 3.2761, "step": 2344 }, { "epoch": 0.6372282608695652, "grad_norm": 1.1792863607406616, "learning_rate": 3e-05, "loss": 2.9453, "step": 2345 }, { "epoch": 0.6375, "grad_norm": 1.4226300716400146, "learning_rate": 3e-05, "loss": 2.8206, "step": 2346 }, { "epoch": 0.6377717391304348, "grad_norm": 1.5433789491653442, "learning_rate": 3e-05, "loss": 3.4793, "step": 2347 }, { "epoch": 0.6380434782608696, "grad_norm": 1.0570566654205322, "learning_rate": 3e-05, "loss": 3.0796, "step": 2348 }, { "epoch": 0.6383152173913044, "grad_norm": 1.4246288537979126, "learning_rate": 3e-05, "loss": 3.259, "step": 2349 }, { "epoch": 0.6385869565217391, "grad_norm": 2.0091404914855957, "learning_rate": 3e-05, "loss": 3.2464, "step": 2350 }, { "epoch": 0.6388586956521739, "grad_norm": 1.7661538124084473, "learning_rate": 3e-05, "loss": 3.0348, "step": 2351 }, { "epoch": 0.6391304347826087, "grad_norm": 1.6644985675811768, "learning_rate": 3e-05, "loss": 3.2656, "step": 2352 }, { "epoch": 0.6394021739130434, "grad_norm": 1.192597508430481, "learning_rate": 3e-05, "loss": 3.2846, "step": 2353 }, { "epoch": 0.6396739130434783, "grad_norm": 1.234485149383545, "learning_rate": 3e-05, "loss": 3.2553, "step": 2354 }, { "epoch": 0.6399456521739131, "grad_norm": 1.2841020822525024, "learning_rate": 3e-05, "loss": 2.8951, "step": 2355 }, { "epoch": 0.6402173913043478, "grad_norm": 1.119943618774414, "learning_rate": 3e-05, "loss": 2.7422, "step": 2356 }, { "epoch": 0.6404891304347826, "grad_norm": 1.5436328649520874, "learning_rate": 3e-05, "loss": 3.9712, "step": 2357 }, { "epoch": 0.6407608695652174, "grad_norm": 1.1272965669631958, "learning_rate": 3e-05, "loss": 3.1857, "step": 2358 }, { "epoch": 0.6410326086956522, "grad_norm": 1.340259075164795, "learning_rate": 3e-05, "loss": 3.7488, "step": 2359 }, { "epoch": 0.6413043478260869, "grad_norm": 1.486802577972412, "learning_rate": 3e-05, "loss": 3.8155, "step": 2360 }, { "epoch": 0.6415760869565217, "grad_norm": 0.9581712484359741, "learning_rate": 3e-05, "loss": 3.117, "step": 2361 }, { "epoch": 0.6418478260869566, "grad_norm": 1.012144923210144, "learning_rate": 3e-05, "loss": 3.0744, "step": 2362 }, { "epoch": 0.6421195652173913, "grad_norm": 1.1179242134094238, "learning_rate": 3e-05, "loss": 3.2825, "step": 2363 }, { "epoch": 0.6423913043478261, "grad_norm": 1.3681845664978027, "learning_rate": 3e-05, "loss": 3.1833, "step": 2364 }, { "epoch": 0.6426630434782609, "grad_norm": 1.192631721496582, "learning_rate": 3e-05, "loss": 3.3842, "step": 2365 }, { "epoch": 0.6429347826086956, "grad_norm": 1.2317907810211182, "learning_rate": 3e-05, "loss": 3.1083, "step": 2366 }, { "epoch": 0.6432065217391304, "grad_norm": 1.0752184391021729, "learning_rate": 3e-05, "loss": 3.2455, "step": 2367 }, { "epoch": 0.6434782608695652, "grad_norm": 1.035239338874817, "learning_rate": 3e-05, "loss": 2.9436, "step": 2368 }, { "epoch": 0.64375, "grad_norm": 1.3296703100204468, "learning_rate": 3e-05, "loss": 3.2708, "step": 2369 }, { "epoch": 0.6440217391304348, "grad_norm": 1.2996970415115356, "learning_rate": 3e-05, "loss": 3.8312, "step": 2370 }, { "epoch": 0.6442934782608696, "grad_norm": 0.9831019639968872, "learning_rate": 3e-05, "loss": 3.2478, "step": 2371 }, { "epoch": 0.6445652173913043, "grad_norm": 0.9067811965942383, "learning_rate": 3e-05, "loss": 3.1556, "step": 2372 }, { "epoch": 0.6448369565217391, "grad_norm": 1.2001372575759888, "learning_rate": 3e-05, "loss": 3.7225, "step": 2373 }, { "epoch": 0.6451086956521739, "grad_norm": 1.1935441493988037, "learning_rate": 3e-05, "loss": 3.3109, "step": 2374 }, { "epoch": 0.6453804347826086, "grad_norm": 0.9141666889190674, "learning_rate": 3e-05, "loss": 3.0955, "step": 2375 }, { "epoch": 0.6456521739130435, "grad_norm": 1.0943185091018677, "learning_rate": 3e-05, "loss": 3.3094, "step": 2376 }, { "epoch": 0.6459239130434783, "grad_norm": 1.1921093463897705, "learning_rate": 3e-05, "loss": 3.3715, "step": 2377 }, { "epoch": 0.6461956521739131, "grad_norm": 1.0566505193710327, "learning_rate": 3e-05, "loss": 3.1073, "step": 2378 }, { "epoch": 0.6464673913043478, "grad_norm": 1.5691936016082764, "learning_rate": 3e-05, "loss": 3.4321, "step": 2379 }, { "epoch": 0.6467391304347826, "grad_norm": 1.041152000427246, "learning_rate": 3e-05, "loss": 2.6529, "step": 2380 }, { "epoch": 0.6470108695652174, "grad_norm": 1.2687933444976807, "learning_rate": 3e-05, "loss": 2.9547, "step": 2381 }, { "epoch": 0.6472826086956521, "grad_norm": 1.074779748916626, "learning_rate": 3e-05, "loss": 3.0635, "step": 2382 }, { "epoch": 0.647554347826087, "grad_norm": 1.0904431343078613, "learning_rate": 3e-05, "loss": 2.8928, "step": 2383 }, { "epoch": 0.6478260869565218, "grad_norm": 1.3329259157180786, "learning_rate": 3e-05, "loss": 3.6509, "step": 2384 }, { "epoch": 0.6480978260869565, "grad_norm": 1.2522311210632324, "learning_rate": 3e-05, "loss": 3.6357, "step": 2385 }, { "epoch": 0.6483695652173913, "grad_norm": 1.2754337787628174, "learning_rate": 3e-05, "loss": 3.7422, "step": 2386 }, { "epoch": 0.6486413043478261, "grad_norm": 1.338080883026123, "learning_rate": 3e-05, "loss": 3.3859, "step": 2387 }, { "epoch": 0.6489130434782608, "grad_norm": 1.1422319412231445, "learning_rate": 3e-05, "loss": 3.8218, "step": 2388 }, { "epoch": 0.6491847826086956, "grad_norm": 1.647072196006775, "learning_rate": 3e-05, "loss": 3.6864, "step": 2389 }, { "epoch": 0.6494565217391305, "grad_norm": 1.1428362131118774, "learning_rate": 3e-05, "loss": 2.8359, "step": 2390 }, { "epoch": 0.6497282608695653, "grad_norm": 1.1382925510406494, "learning_rate": 3e-05, "loss": 3.1163, "step": 2391 }, { "epoch": 0.65, "grad_norm": 1.0694186687469482, "learning_rate": 3e-05, "loss": 3.1842, "step": 2392 }, { "epoch": 0.6502717391304348, "grad_norm": 1.413272738456726, "learning_rate": 3e-05, "loss": 3.7837, "step": 2393 }, { "epoch": 0.6505434782608696, "grad_norm": 1.353119134902954, "learning_rate": 3e-05, "loss": 3.2845, "step": 2394 }, { "epoch": 0.6508152173913043, "grad_norm": 1.542140245437622, "learning_rate": 3e-05, "loss": 3.4894, "step": 2395 }, { "epoch": 0.6510869565217391, "grad_norm": 1.6058558225631714, "learning_rate": 3e-05, "loss": 3.4778, "step": 2396 }, { "epoch": 0.6513586956521739, "grad_norm": 1.719277262687683, "learning_rate": 3e-05, "loss": 3.5328, "step": 2397 }, { "epoch": 0.6516304347826087, "grad_norm": 1.7350586652755737, "learning_rate": 3e-05, "loss": 3.5851, "step": 2398 }, { "epoch": 0.6519021739130435, "grad_norm": 1.4615671634674072, "learning_rate": 3e-05, "loss": 4.099, "step": 2399 }, { "epoch": 0.6521739130434783, "grad_norm": 1.194122314453125, "learning_rate": 3e-05, "loss": 3.353, "step": 2400 }, { "epoch": 0.652445652173913, "grad_norm": 1.4127830266952515, "learning_rate": 3e-05, "loss": 3.8697, "step": 2401 }, { "epoch": 0.6527173913043478, "grad_norm": 1.6707136631011963, "learning_rate": 3e-05, "loss": 3.9621, "step": 2402 }, { "epoch": 0.6529891304347826, "grad_norm": 1.4606815576553345, "learning_rate": 3e-05, "loss": 3.7677, "step": 2403 }, { "epoch": 0.6532608695652173, "grad_norm": 1.2474243640899658, "learning_rate": 3e-05, "loss": 3.5581, "step": 2404 }, { "epoch": 0.6535326086956522, "grad_norm": 1.1824653148651123, "learning_rate": 3e-05, "loss": 2.9751, "step": 2405 }, { "epoch": 0.653804347826087, "grad_norm": 1.22133469581604, "learning_rate": 3e-05, "loss": 3.3986, "step": 2406 }, { "epoch": 0.6540760869565218, "grad_norm": 1.160557746887207, "learning_rate": 3e-05, "loss": 3.2008, "step": 2407 }, { "epoch": 0.6543478260869565, "grad_norm": 1.2002159357070923, "learning_rate": 3e-05, "loss": 3.6029, "step": 2408 }, { "epoch": 0.6546195652173913, "grad_norm": 1.1110812425613403, "learning_rate": 3e-05, "loss": 3.2051, "step": 2409 }, { "epoch": 0.654891304347826, "grad_norm": 0.9706496596336365, "learning_rate": 3e-05, "loss": 2.9527, "step": 2410 }, { "epoch": 0.6551630434782608, "grad_norm": 1.0990405082702637, "learning_rate": 3e-05, "loss": 2.6553, "step": 2411 }, { "epoch": 0.6554347826086957, "grad_norm": 1.065306544303894, "learning_rate": 3e-05, "loss": 3.3728, "step": 2412 }, { "epoch": 0.6557065217391305, "grad_norm": 0.956051766872406, "learning_rate": 3e-05, "loss": 2.9055, "step": 2413 }, { "epoch": 0.6559782608695652, "grad_norm": 1.1310175657272339, "learning_rate": 3e-05, "loss": 3.8269, "step": 2414 }, { "epoch": 0.65625, "grad_norm": 0.9513333439826965, "learning_rate": 3e-05, "loss": 3.2811, "step": 2415 }, { "epoch": 0.6565217391304348, "grad_norm": 1.301697015762329, "learning_rate": 3e-05, "loss": 3.5579, "step": 2416 }, { "epoch": 0.6567934782608695, "grad_norm": 0.9810994267463684, "learning_rate": 3e-05, "loss": 2.9914, "step": 2417 }, { "epoch": 0.6570652173913043, "grad_norm": 1.3848271369934082, "learning_rate": 3e-05, "loss": 3.8109, "step": 2418 }, { "epoch": 0.6573369565217392, "grad_norm": 1.169350504875183, "learning_rate": 3e-05, "loss": 2.8291, "step": 2419 }, { "epoch": 0.657608695652174, "grad_norm": 1.3478142023086548, "learning_rate": 3e-05, "loss": 3.5609, "step": 2420 }, { "epoch": 0.6578804347826087, "grad_norm": 1.1918085813522339, "learning_rate": 3e-05, "loss": 3.5339, "step": 2421 }, { "epoch": 0.6581521739130435, "grad_norm": 1.4257116317749023, "learning_rate": 3e-05, "loss": 3.4636, "step": 2422 }, { "epoch": 0.6584239130434782, "grad_norm": 1.2269481420516968, "learning_rate": 3e-05, "loss": 3.2741, "step": 2423 }, { "epoch": 0.658695652173913, "grad_norm": 1.2066056728363037, "learning_rate": 3e-05, "loss": 3.2177, "step": 2424 }, { "epoch": 0.6589673913043478, "grad_norm": 1.3328955173492432, "learning_rate": 3e-05, "loss": 3.527, "step": 2425 }, { "epoch": 0.6592391304347827, "grad_norm": 1.1746604442596436, "learning_rate": 3e-05, "loss": 3.1194, "step": 2426 }, { "epoch": 0.6595108695652174, "grad_norm": 1.028620958328247, "learning_rate": 3e-05, "loss": 3.1119, "step": 2427 }, { "epoch": 0.6597826086956522, "grad_norm": 1.0080275535583496, "learning_rate": 3e-05, "loss": 2.7994, "step": 2428 }, { "epoch": 0.660054347826087, "grad_norm": 0.8145027756690979, "learning_rate": 3e-05, "loss": 2.9428, "step": 2429 }, { "epoch": 0.6603260869565217, "grad_norm": 1.0437175035476685, "learning_rate": 3e-05, "loss": 2.9878, "step": 2430 }, { "epoch": 0.6605978260869565, "grad_norm": 1.130467414855957, "learning_rate": 3e-05, "loss": 3.7007, "step": 2431 }, { "epoch": 0.6608695652173913, "grad_norm": 1.0865567922592163, "learning_rate": 3e-05, "loss": 3.1175, "step": 2432 }, { "epoch": 0.6611413043478261, "grad_norm": 1.1606049537658691, "learning_rate": 3e-05, "loss": 3.3582, "step": 2433 }, { "epoch": 0.6614130434782609, "grad_norm": 0.9023473262786865, "learning_rate": 3e-05, "loss": 2.9861, "step": 2434 }, { "epoch": 0.6616847826086957, "grad_norm": 0.856779932975769, "learning_rate": 3e-05, "loss": 3.0402, "step": 2435 }, { "epoch": 0.6619565217391304, "grad_norm": 1.0253337621688843, "learning_rate": 3e-05, "loss": 3.0909, "step": 2436 }, { "epoch": 0.6622282608695652, "grad_norm": 1.0018707513809204, "learning_rate": 3e-05, "loss": 2.9996, "step": 2437 }, { "epoch": 0.6625, "grad_norm": 1.0722147226333618, "learning_rate": 3e-05, "loss": 3.3102, "step": 2438 }, { "epoch": 0.6627717391304347, "grad_norm": 1.0798076391220093, "learning_rate": 3e-05, "loss": 2.9907, "step": 2439 }, { "epoch": 0.6630434782608695, "grad_norm": 1.1593507528305054, "learning_rate": 3e-05, "loss": 3.3741, "step": 2440 }, { "epoch": 0.6633152173913044, "grad_norm": 1.0947400331497192, "learning_rate": 3e-05, "loss": 3.0087, "step": 2441 }, { "epoch": 0.6635869565217392, "grad_norm": 1.2882115840911865, "learning_rate": 3e-05, "loss": 3.1978, "step": 2442 }, { "epoch": 0.6638586956521739, "grad_norm": 1.2732001543045044, "learning_rate": 3e-05, "loss": 3.8371, "step": 2443 }, { "epoch": 0.6641304347826087, "grad_norm": 0.9186998605728149, "learning_rate": 3e-05, "loss": 2.643, "step": 2444 }, { "epoch": 0.6644021739130435, "grad_norm": 1.0889346599578857, "learning_rate": 3e-05, "loss": 2.9247, "step": 2445 }, { "epoch": 0.6646739130434782, "grad_norm": 1.094017505645752, "learning_rate": 3e-05, "loss": 3.1954, "step": 2446 }, { "epoch": 0.664945652173913, "grad_norm": 0.9859854578971863, "learning_rate": 3e-05, "loss": 2.7335, "step": 2447 }, { "epoch": 0.6652173913043479, "grad_norm": 1.8823328018188477, "learning_rate": 3e-05, "loss": 3.4934, "step": 2448 }, { "epoch": 0.6654891304347826, "grad_norm": 0.9667043685913086, "learning_rate": 3e-05, "loss": 3.0711, "step": 2449 }, { "epoch": 0.6657608695652174, "grad_norm": 1.0037490129470825, "learning_rate": 3e-05, "loss": 3.1836, "step": 2450 }, { "epoch": 0.6660326086956522, "grad_norm": 1.0636874437332153, "learning_rate": 3e-05, "loss": 3.3269, "step": 2451 }, { "epoch": 0.6663043478260869, "grad_norm": 1.02604079246521, "learning_rate": 3e-05, "loss": 3.4928, "step": 2452 }, { "epoch": 0.6665760869565217, "grad_norm": 1.050881028175354, "learning_rate": 3e-05, "loss": 3.3983, "step": 2453 }, { "epoch": 0.6668478260869565, "grad_norm": 1.1651184558868408, "learning_rate": 3e-05, "loss": 3.573, "step": 2454 }, { "epoch": 0.6671195652173914, "grad_norm": 0.9464408755302429, "learning_rate": 3e-05, "loss": 3.4415, "step": 2455 }, { "epoch": 0.6673913043478261, "grad_norm": 0.9894149899482727, "learning_rate": 3e-05, "loss": 3.4124, "step": 2456 }, { "epoch": 0.6676630434782609, "grad_norm": 1.1918611526489258, "learning_rate": 3e-05, "loss": 3.1563, "step": 2457 }, { "epoch": 0.6679347826086957, "grad_norm": 1.1251368522644043, "learning_rate": 3e-05, "loss": 3.0224, "step": 2458 }, { "epoch": 0.6682065217391304, "grad_norm": 0.9575350880622864, "learning_rate": 3e-05, "loss": 3.1342, "step": 2459 }, { "epoch": 0.6684782608695652, "grad_norm": 1.1674573421478271, "learning_rate": 3e-05, "loss": 3.5287, "step": 2460 }, { "epoch": 0.66875, "grad_norm": 1.3104429244995117, "learning_rate": 3e-05, "loss": 3.3732, "step": 2461 }, { "epoch": 0.6690217391304348, "grad_norm": 1.0205787420272827, "learning_rate": 3e-05, "loss": 3.3107, "step": 2462 }, { "epoch": 0.6692934782608696, "grad_norm": 1.1491562128067017, "learning_rate": 3e-05, "loss": 3.1578, "step": 2463 }, { "epoch": 0.6695652173913044, "grad_norm": 1.2550066709518433, "learning_rate": 3e-05, "loss": 2.9625, "step": 2464 }, { "epoch": 0.6698369565217391, "grad_norm": 1.396528720855713, "learning_rate": 3e-05, "loss": 3.4526, "step": 2465 }, { "epoch": 0.6701086956521739, "grad_norm": 1.3484258651733398, "learning_rate": 3e-05, "loss": 3.0718, "step": 2466 }, { "epoch": 0.6703804347826087, "grad_norm": 1.2837543487548828, "learning_rate": 3e-05, "loss": 3.9352, "step": 2467 }, { "epoch": 0.6706521739130434, "grad_norm": 1.126395583152771, "learning_rate": 3e-05, "loss": 3.2975, "step": 2468 }, { "epoch": 0.6709239130434783, "grad_norm": 1.0975403785705566, "learning_rate": 3e-05, "loss": 2.9911, "step": 2469 }, { "epoch": 0.6711956521739131, "grad_norm": 1.260852336883545, "learning_rate": 3e-05, "loss": 3.4067, "step": 2470 }, { "epoch": 0.6714673913043478, "grad_norm": 1.3457390069961548, "learning_rate": 3e-05, "loss": 3.3277, "step": 2471 }, { "epoch": 0.6717391304347826, "grad_norm": 1.6091570854187012, "learning_rate": 3e-05, "loss": 3.8467, "step": 2472 }, { "epoch": 0.6720108695652174, "grad_norm": 1.3068939447402954, "learning_rate": 3e-05, "loss": 3.5756, "step": 2473 }, { "epoch": 0.6722826086956522, "grad_norm": 1.532850742340088, "learning_rate": 3e-05, "loss": 3.6188, "step": 2474 }, { "epoch": 0.6725543478260869, "grad_norm": 1.179315447807312, "learning_rate": 3e-05, "loss": 3.3408, "step": 2475 }, { "epoch": 0.6728260869565217, "grad_norm": 1.278639793395996, "learning_rate": 3e-05, "loss": 3.9328, "step": 2476 }, { "epoch": 0.6730978260869566, "grad_norm": 0.9712943434715271, "learning_rate": 3e-05, "loss": 2.7786, "step": 2477 }, { "epoch": 0.6733695652173913, "grad_norm": 1.1749228239059448, "learning_rate": 3e-05, "loss": 2.7182, "step": 2478 }, { "epoch": 0.6736413043478261, "grad_norm": 1.3901921510696411, "learning_rate": 3e-05, "loss": 3.2455, "step": 2479 }, { "epoch": 0.6739130434782609, "grad_norm": 1.3088897466659546, "learning_rate": 3e-05, "loss": 3.4134, "step": 2480 }, { "epoch": 0.6741847826086956, "grad_norm": 1.2072702646255493, "learning_rate": 3e-05, "loss": 3.7264, "step": 2481 }, { "epoch": 0.6744565217391304, "grad_norm": 1.3006356954574585, "learning_rate": 3e-05, "loss": 3.1591, "step": 2482 }, { "epoch": 0.6747282608695652, "grad_norm": 2.3080663681030273, "learning_rate": 3e-05, "loss": 3.6425, "step": 2483 }, { "epoch": 0.675, "grad_norm": 1.3426471948623657, "learning_rate": 3e-05, "loss": 2.9936, "step": 2484 }, { "epoch": 0.6752717391304348, "grad_norm": 1.1594306230545044, "learning_rate": 3e-05, "loss": 3.3481, "step": 2485 }, { "epoch": 0.6755434782608696, "grad_norm": 1.0936838388442993, "learning_rate": 3e-05, "loss": 3.0749, "step": 2486 }, { "epoch": 0.6758152173913043, "grad_norm": 1.3417373895645142, "learning_rate": 3e-05, "loss": 3.2727, "step": 2487 }, { "epoch": 0.6760869565217391, "grad_norm": 1.6536362171173096, "learning_rate": 3e-05, "loss": 3.3898, "step": 2488 }, { "epoch": 0.6763586956521739, "grad_norm": 1.2044962644577026, "learning_rate": 3e-05, "loss": 3.06, "step": 2489 }, { "epoch": 0.6766304347826086, "grad_norm": 1.6772176027297974, "learning_rate": 3e-05, "loss": 3.6963, "step": 2490 }, { "epoch": 0.6769021739130435, "grad_norm": 1.2547991275787354, "learning_rate": 3e-05, "loss": 2.9258, "step": 2491 }, { "epoch": 0.6771739130434783, "grad_norm": 1.2318508625030518, "learning_rate": 3e-05, "loss": 3.4305, "step": 2492 }, { "epoch": 0.6774456521739131, "grad_norm": 1.4965567588806152, "learning_rate": 3e-05, "loss": 3.4216, "step": 2493 }, { "epoch": 0.6777173913043478, "grad_norm": 1.6711788177490234, "learning_rate": 3e-05, "loss": 2.8026, "step": 2494 }, { "epoch": 0.6779891304347826, "grad_norm": 1.285457968711853, "learning_rate": 3e-05, "loss": 2.9369, "step": 2495 }, { "epoch": 0.6782608695652174, "grad_norm": 1.1391198635101318, "learning_rate": 3e-05, "loss": 2.6831, "step": 2496 }, { "epoch": 0.6785326086956521, "grad_norm": 1.4873690605163574, "learning_rate": 3e-05, "loss": 3.5054, "step": 2497 }, { "epoch": 0.678804347826087, "grad_norm": 1.5623383522033691, "learning_rate": 3e-05, "loss": 3.5831, "step": 2498 }, { "epoch": 0.6790760869565218, "grad_norm": 1.0868808031082153, "learning_rate": 3e-05, "loss": 3.5378, "step": 2499 }, { "epoch": 0.6793478260869565, "grad_norm": 1.5260769128799438, "learning_rate": 3e-05, "loss": 3.6442, "step": 2500 }, { "epoch": 0.6796195652173913, "grad_norm": 1.5094234943389893, "learning_rate": 3e-05, "loss": 2.8721, "step": 2501 }, { "epoch": 0.6798913043478261, "grad_norm": 1.03408944606781, "learning_rate": 3e-05, "loss": 3.2489, "step": 2502 }, { "epoch": 0.6801630434782608, "grad_norm": 1.5139715671539307, "learning_rate": 3e-05, "loss": 3.1632, "step": 2503 }, { "epoch": 0.6804347826086956, "grad_norm": 0.8952838778495789, "learning_rate": 3e-05, "loss": 2.678, "step": 2504 }, { "epoch": 0.6807065217391305, "grad_norm": 0.886978268623352, "learning_rate": 3e-05, "loss": 2.9663, "step": 2505 }, { "epoch": 0.6809782608695653, "grad_norm": 1.4015369415283203, "learning_rate": 3e-05, "loss": 3.4536, "step": 2506 }, { "epoch": 0.68125, "grad_norm": 1.1252248287200928, "learning_rate": 3e-05, "loss": 3.1302, "step": 2507 }, { "epoch": 0.6815217391304348, "grad_norm": 1.2304083108901978, "learning_rate": 3e-05, "loss": 3.5204, "step": 2508 }, { "epoch": 0.6817934782608696, "grad_norm": 1.2945035696029663, "learning_rate": 3e-05, "loss": 3.7562, "step": 2509 }, { "epoch": 0.6820652173913043, "grad_norm": 1.1755324602127075, "learning_rate": 3e-05, "loss": 2.787, "step": 2510 }, { "epoch": 0.6823369565217391, "grad_norm": 1.413987159729004, "learning_rate": 3e-05, "loss": 3.5305, "step": 2511 }, { "epoch": 0.6826086956521739, "grad_norm": 1.0807994604110718, "learning_rate": 3e-05, "loss": 3.2698, "step": 2512 }, { "epoch": 0.6828804347826087, "grad_norm": 1.5065345764160156, "learning_rate": 3e-05, "loss": 3.3808, "step": 2513 }, { "epoch": 0.6831521739130435, "grad_norm": 1.2206026315689087, "learning_rate": 3e-05, "loss": 2.9492, "step": 2514 }, { "epoch": 0.6834239130434783, "grad_norm": 1.1851751804351807, "learning_rate": 3e-05, "loss": 3.425, "step": 2515 }, { "epoch": 0.683695652173913, "grad_norm": 1.1727910041809082, "learning_rate": 3e-05, "loss": 3.6993, "step": 2516 }, { "epoch": 0.6839673913043478, "grad_norm": 1.5769599676132202, "learning_rate": 3e-05, "loss": 3.3359, "step": 2517 }, { "epoch": 0.6842391304347826, "grad_norm": 1.5909152030944824, "learning_rate": 3e-05, "loss": 3.1606, "step": 2518 }, { "epoch": 0.6845108695652173, "grad_norm": 1.094890832901001, "learning_rate": 3e-05, "loss": 3.2958, "step": 2519 }, { "epoch": 0.6847826086956522, "grad_norm": 1.5951069593429565, "learning_rate": 3e-05, "loss": 3.6897, "step": 2520 }, { "epoch": 0.685054347826087, "grad_norm": 1.3260122537612915, "learning_rate": 3e-05, "loss": 3.5941, "step": 2521 }, { "epoch": 0.6853260869565218, "grad_norm": 1.431256651878357, "learning_rate": 3e-05, "loss": 2.8581, "step": 2522 }, { "epoch": 0.6855978260869565, "grad_norm": 1.3032617568969727, "learning_rate": 3e-05, "loss": 3.3358, "step": 2523 }, { "epoch": 0.6858695652173913, "grad_norm": 1.0501641035079956, "learning_rate": 3e-05, "loss": 2.9872, "step": 2524 }, { "epoch": 0.686141304347826, "grad_norm": 1.1217620372772217, "learning_rate": 3e-05, "loss": 3.1064, "step": 2525 }, { "epoch": 0.6864130434782608, "grad_norm": 1.0204346179962158, "learning_rate": 3e-05, "loss": 3.13, "step": 2526 }, { "epoch": 0.6866847826086957, "grad_norm": 1.479447603225708, "learning_rate": 3e-05, "loss": 3.3739, "step": 2527 }, { "epoch": 0.6869565217391305, "grad_norm": 1.3281017541885376, "learning_rate": 3e-05, "loss": 3.1517, "step": 2528 }, { "epoch": 0.6872282608695652, "grad_norm": 1.358286738395691, "learning_rate": 3e-05, "loss": 3.215, "step": 2529 }, { "epoch": 0.6875, "grad_norm": 1.169724464416504, "learning_rate": 3e-05, "loss": 2.9665, "step": 2530 }, { "epoch": 0.6877717391304348, "grad_norm": 1.373205542564392, "learning_rate": 3e-05, "loss": 3.3161, "step": 2531 }, { "epoch": 0.6880434782608695, "grad_norm": 1.0364346504211426, "learning_rate": 3e-05, "loss": 3.135, "step": 2532 }, { "epoch": 0.6883152173913043, "grad_norm": 1.2910298109054565, "learning_rate": 3e-05, "loss": 3.2629, "step": 2533 }, { "epoch": 0.6885869565217392, "grad_norm": 1.7945650815963745, "learning_rate": 3e-05, "loss": 3.5541, "step": 2534 }, { "epoch": 0.688858695652174, "grad_norm": 1.3837659358978271, "learning_rate": 3e-05, "loss": 3.5515, "step": 2535 }, { "epoch": 0.6891304347826087, "grad_norm": 1.3154077529907227, "learning_rate": 3e-05, "loss": 3.6571, "step": 2536 }, { "epoch": 0.6894021739130435, "grad_norm": 1.3466911315917969, "learning_rate": 3e-05, "loss": 3.0786, "step": 2537 }, { "epoch": 0.6896739130434782, "grad_norm": 1.3170411586761475, "learning_rate": 3e-05, "loss": 2.9075, "step": 2538 }, { "epoch": 0.689945652173913, "grad_norm": 1.1219418048858643, "learning_rate": 3e-05, "loss": 3.1026, "step": 2539 }, { "epoch": 0.6902173913043478, "grad_norm": 1.15764319896698, "learning_rate": 3e-05, "loss": 3.45, "step": 2540 }, { "epoch": 0.6904891304347827, "grad_norm": 1.196807861328125, "learning_rate": 3e-05, "loss": 3.185, "step": 2541 }, { "epoch": 0.6907608695652174, "grad_norm": 1.187324047088623, "learning_rate": 3e-05, "loss": 3.4707, "step": 2542 }, { "epoch": 0.6910326086956522, "grad_norm": 1.0559121370315552, "learning_rate": 3e-05, "loss": 2.9871, "step": 2543 }, { "epoch": 0.691304347826087, "grad_norm": 1.1243149042129517, "learning_rate": 3e-05, "loss": 3.5086, "step": 2544 }, { "epoch": 0.6915760869565217, "grad_norm": 1.1120831966400146, "learning_rate": 3e-05, "loss": 3.1197, "step": 2545 }, { "epoch": 0.6918478260869565, "grad_norm": 1.294692039489746, "learning_rate": 3e-05, "loss": 3.9801, "step": 2546 }, { "epoch": 0.6921195652173913, "grad_norm": 1.3371599912643433, "learning_rate": 3e-05, "loss": 3.8999, "step": 2547 }, { "epoch": 0.6923913043478261, "grad_norm": 1.219671368598938, "learning_rate": 3e-05, "loss": 3.3713, "step": 2548 }, { "epoch": 0.6926630434782609, "grad_norm": 0.9593479037284851, "learning_rate": 3e-05, "loss": 2.8705, "step": 2549 }, { "epoch": 0.6929347826086957, "grad_norm": 1.2361105680465698, "learning_rate": 3e-05, "loss": 3.6656, "step": 2550 }, { "epoch": 0.6932065217391304, "grad_norm": 0.9579165577888489, "learning_rate": 3e-05, "loss": 2.9437, "step": 2551 }, { "epoch": 0.6934782608695652, "grad_norm": 1.0255228281021118, "learning_rate": 3e-05, "loss": 3.0341, "step": 2552 }, { "epoch": 0.69375, "grad_norm": 1.2201510667800903, "learning_rate": 3e-05, "loss": 3.6407, "step": 2553 }, { "epoch": 0.6940217391304347, "grad_norm": 1.1329454183578491, "learning_rate": 3e-05, "loss": 3.4387, "step": 2554 }, { "epoch": 0.6942934782608695, "grad_norm": 1.0753947496414185, "learning_rate": 3e-05, "loss": 3.1663, "step": 2555 }, { "epoch": 0.6945652173913044, "grad_norm": 1.8046010732650757, "learning_rate": 3e-05, "loss": 3.624, "step": 2556 }, { "epoch": 0.6948369565217392, "grad_norm": 1.0631359815597534, "learning_rate": 3e-05, "loss": 3.2537, "step": 2557 }, { "epoch": 0.6951086956521739, "grad_norm": 1.0694079399108887, "learning_rate": 3e-05, "loss": 2.8613, "step": 2558 }, { "epoch": 0.6953804347826087, "grad_norm": 1.2790333032608032, "learning_rate": 3e-05, "loss": 3.0732, "step": 2559 }, { "epoch": 0.6956521739130435, "grad_norm": 1.253812551498413, "learning_rate": 3e-05, "loss": 3.0818, "step": 2560 }, { "epoch": 0.6959239130434782, "grad_norm": 1.2076342105865479, "learning_rate": 3e-05, "loss": 2.7011, "step": 2561 }, { "epoch": 0.696195652173913, "grad_norm": 1.2318416833877563, "learning_rate": 3e-05, "loss": 2.7658, "step": 2562 }, { "epoch": 0.6964673913043479, "grad_norm": 1.442963719367981, "learning_rate": 3e-05, "loss": 3.7564, "step": 2563 }, { "epoch": 0.6967391304347826, "grad_norm": 1.3947499990463257, "learning_rate": 3e-05, "loss": 3.4476, "step": 2564 }, { "epoch": 0.6970108695652174, "grad_norm": 1.8033784627914429, "learning_rate": 3e-05, "loss": 3.7798, "step": 2565 }, { "epoch": 0.6972826086956522, "grad_norm": 1.3678721189498901, "learning_rate": 3e-05, "loss": 3.3788, "step": 2566 }, { "epoch": 0.6975543478260869, "grad_norm": 1.0561424493789673, "learning_rate": 3e-05, "loss": 3.1749, "step": 2567 }, { "epoch": 0.6978260869565217, "grad_norm": 1.2744587659835815, "learning_rate": 3e-05, "loss": 3.3584, "step": 2568 }, { "epoch": 0.6980978260869565, "grad_norm": 1.2167631387710571, "learning_rate": 3e-05, "loss": 3.4043, "step": 2569 }, { "epoch": 0.6983695652173914, "grad_norm": 1.1461526155471802, "learning_rate": 3e-05, "loss": 3.381, "step": 2570 }, { "epoch": 0.6986413043478261, "grad_norm": 1.173651933670044, "learning_rate": 3e-05, "loss": 2.8328, "step": 2571 }, { "epoch": 0.6989130434782609, "grad_norm": 1.3456881046295166, "learning_rate": 3e-05, "loss": 3.9301, "step": 2572 }, { "epoch": 0.6991847826086957, "grad_norm": 1.6022355556488037, "learning_rate": 3e-05, "loss": 4.0052, "step": 2573 }, { "epoch": 0.6994565217391304, "grad_norm": 1.4133738279342651, "learning_rate": 3e-05, "loss": 3.6391, "step": 2574 }, { "epoch": 0.6997282608695652, "grad_norm": 1.2311711311340332, "learning_rate": 3e-05, "loss": 3.6895, "step": 2575 }, { "epoch": 0.7, "grad_norm": 1.0733444690704346, "learning_rate": 3e-05, "loss": 2.685, "step": 2576 }, { "epoch": 0.7002717391304348, "grad_norm": 1.596375823020935, "learning_rate": 3e-05, "loss": 3.2422, "step": 2577 }, { "epoch": 0.7005434782608696, "grad_norm": 1.098339319229126, "learning_rate": 3e-05, "loss": 3.0258, "step": 2578 }, { "epoch": 0.7008152173913044, "grad_norm": 1.215551495552063, "learning_rate": 3e-05, "loss": 3.6583, "step": 2579 }, { "epoch": 0.7010869565217391, "grad_norm": 1.175678014755249, "learning_rate": 3e-05, "loss": 2.9082, "step": 2580 }, { "epoch": 0.7013586956521739, "grad_norm": 1.285050868988037, "learning_rate": 3e-05, "loss": 3.0373, "step": 2581 }, { "epoch": 0.7016304347826087, "grad_norm": 1.0280994176864624, "learning_rate": 3e-05, "loss": 3.027, "step": 2582 }, { "epoch": 0.7019021739130434, "grad_norm": 1.1611984968185425, "learning_rate": 3e-05, "loss": 2.9426, "step": 2583 }, { "epoch": 0.7021739130434783, "grad_norm": 1.475341558456421, "learning_rate": 3e-05, "loss": 3.366, "step": 2584 }, { "epoch": 0.7024456521739131, "grad_norm": 1.6377277374267578, "learning_rate": 3e-05, "loss": 3.4074, "step": 2585 }, { "epoch": 0.7027173913043478, "grad_norm": 1.4452388286590576, "learning_rate": 3e-05, "loss": 2.9369, "step": 2586 }, { "epoch": 0.7029891304347826, "grad_norm": 1.533266305923462, "learning_rate": 3e-05, "loss": 3.9352, "step": 2587 }, { "epoch": 0.7032608695652174, "grad_norm": 1.6853458881378174, "learning_rate": 3e-05, "loss": 3.4523, "step": 2588 }, { "epoch": 0.7035326086956522, "grad_norm": 1.3552666902542114, "learning_rate": 3e-05, "loss": 2.7867, "step": 2589 }, { "epoch": 0.7038043478260869, "grad_norm": 1.5494905710220337, "learning_rate": 3e-05, "loss": 3.4699, "step": 2590 }, { "epoch": 0.7040760869565217, "grad_norm": 1.74055016040802, "learning_rate": 3e-05, "loss": 3.3874, "step": 2591 }, { "epoch": 0.7043478260869566, "grad_norm": 1.250354290008545, "learning_rate": 3e-05, "loss": 3.1795, "step": 2592 }, { "epoch": 0.7046195652173913, "grad_norm": 1.302441954612732, "learning_rate": 3e-05, "loss": 3.9465, "step": 2593 }, { "epoch": 0.7048913043478261, "grad_norm": 1.7971372604370117, "learning_rate": 3e-05, "loss": 3.6084, "step": 2594 }, { "epoch": 0.7051630434782609, "grad_norm": 1.188605546951294, "learning_rate": 3e-05, "loss": 3.6716, "step": 2595 }, { "epoch": 0.7054347826086956, "grad_norm": 0.9954262971878052, "learning_rate": 3e-05, "loss": 3.1982, "step": 2596 }, { "epoch": 0.7057065217391304, "grad_norm": 1.174482822418213, "learning_rate": 3e-05, "loss": 3.1708, "step": 2597 }, { "epoch": 0.7059782608695652, "grad_norm": 1.039506435394287, "learning_rate": 3e-05, "loss": 3.1857, "step": 2598 }, { "epoch": 0.70625, "grad_norm": 0.9666784405708313, "learning_rate": 3e-05, "loss": 3.6005, "step": 2599 }, { "epoch": 0.7065217391304348, "grad_norm": 0.921465277671814, "learning_rate": 3e-05, "loss": 3.0565, "step": 2600 }, { "epoch": 0.7067934782608696, "grad_norm": 1.018864393234253, "learning_rate": 3e-05, "loss": 2.7895, "step": 2601 }, { "epoch": 0.7070652173913043, "grad_norm": 1.0360243320465088, "learning_rate": 3e-05, "loss": 3.125, "step": 2602 }, { "epoch": 0.7073369565217391, "grad_norm": 0.9393570423126221, "learning_rate": 3e-05, "loss": 3.1148, "step": 2603 }, { "epoch": 0.7076086956521739, "grad_norm": 1.1420108079910278, "learning_rate": 3e-05, "loss": 3.1415, "step": 2604 }, { "epoch": 0.7078804347826086, "grad_norm": 1.1641017198562622, "learning_rate": 3e-05, "loss": 3.0936, "step": 2605 }, { "epoch": 0.7081521739130435, "grad_norm": 1.6107722520828247, "learning_rate": 3e-05, "loss": 4.061, "step": 2606 }, { "epoch": 0.7084239130434783, "grad_norm": 1.2036430835723877, "learning_rate": 3e-05, "loss": 3.1156, "step": 2607 }, { "epoch": 0.7086956521739131, "grad_norm": 1.2490811347961426, "learning_rate": 3e-05, "loss": 3.0078, "step": 2608 }, { "epoch": 0.7089673913043478, "grad_norm": 1.0128928422927856, "learning_rate": 3e-05, "loss": 2.8499, "step": 2609 }, { "epoch": 0.7092391304347826, "grad_norm": 1.363193154335022, "learning_rate": 3e-05, "loss": 3.9375, "step": 2610 }, { "epoch": 0.7095108695652174, "grad_norm": 0.8343818187713623, "learning_rate": 3e-05, "loss": 3.1519, "step": 2611 }, { "epoch": 0.7097826086956521, "grad_norm": 1.0047922134399414, "learning_rate": 3e-05, "loss": 2.8608, "step": 2612 }, { "epoch": 0.710054347826087, "grad_norm": 1.18521249294281, "learning_rate": 3e-05, "loss": 3.7399, "step": 2613 }, { "epoch": 0.7103260869565218, "grad_norm": 1.3591742515563965, "learning_rate": 3e-05, "loss": 3.5885, "step": 2614 }, { "epoch": 0.7105978260869565, "grad_norm": 1.1936426162719727, "learning_rate": 3e-05, "loss": 3.7192, "step": 2615 }, { "epoch": 0.7108695652173913, "grad_norm": 1.114531397819519, "learning_rate": 3e-05, "loss": 3.22, "step": 2616 }, { "epoch": 0.7111413043478261, "grad_norm": 1.1929512023925781, "learning_rate": 3e-05, "loss": 2.751, "step": 2617 }, { "epoch": 0.7114130434782608, "grad_norm": 1.3775954246520996, "learning_rate": 3e-05, "loss": 3.6234, "step": 2618 }, { "epoch": 0.7116847826086956, "grad_norm": 1.313889980316162, "learning_rate": 3e-05, "loss": 3.2312, "step": 2619 }, { "epoch": 0.7119565217391305, "grad_norm": 1.4679901599884033, "learning_rate": 3e-05, "loss": 3.2006, "step": 2620 }, { "epoch": 0.7122282608695653, "grad_norm": 0.9835330843925476, "learning_rate": 3e-05, "loss": 3.2365, "step": 2621 }, { "epoch": 0.7125, "grad_norm": 1.9934345483779907, "learning_rate": 3e-05, "loss": 4.4249, "step": 2622 }, { "epoch": 0.7127717391304348, "grad_norm": 1.2943999767303467, "learning_rate": 3e-05, "loss": 3.517, "step": 2623 }, { "epoch": 0.7130434782608696, "grad_norm": 1.1526103019714355, "learning_rate": 3e-05, "loss": 3.6404, "step": 2624 }, { "epoch": 0.7133152173913043, "grad_norm": 1.3137702941894531, "learning_rate": 3e-05, "loss": 3.305, "step": 2625 }, { "epoch": 0.7135869565217391, "grad_norm": 1.2102550268173218, "learning_rate": 3e-05, "loss": 3.0545, "step": 2626 }, { "epoch": 0.7138586956521739, "grad_norm": 1.0959724187850952, "learning_rate": 3e-05, "loss": 3.0383, "step": 2627 }, { "epoch": 0.7141304347826087, "grad_norm": 1.7450772523880005, "learning_rate": 3e-05, "loss": 3.5891, "step": 2628 }, { "epoch": 0.7144021739130435, "grad_norm": 1.3901242017745972, "learning_rate": 3e-05, "loss": 3.5405, "step": 2629 }, { "epoch": 0.7146739130434783, "grad_norm": 1.2243647575378418, "learning_rate": 3e-05, "loss": 3.4307, "step": 2630 }, { "epoch": 0.714945652173913, "grad_norm": 2.3953914642333984, "learning_rate": 3e-05, "loss": 4.3341, "step": 2631 }, { "epoch": 0.7152173913043478, "grad_norm": 1.8682715892791748, "learning_rate": 3e-05, "loss": 3.0704, "step": 2632 }, { "epoch": 0.7154891304347826, "grad_norm": 1.4622515439987183, "learning_rate": 3e-05, "loss": 3.3376, "step": 2633 }, { "epoch": 0.7157608695652173, "grad_norm": 1.2892646789550781, "learning_rate": 3e-05, "loss": 3.312, "step": 2634 }, { "epoch": 0.7160326086956522, "grad_norm": 1.4950429201126099, "learning_rate": 3e-05, "loss": 3.3923, "step": 2635 }, { "epoch": 0.716304347826087, "grad_norm": 1.380149245262146, "learning_rate": 3e-05, "loss": 3.5931, "step": 2636 }, { "epoch": 0.7165760869565218, "grad_norm": 1.433959722518921, "learning_rate": 3e-05, "loss": 2.9042, "step": 2637 }, { "epoch": 0.7168478260869565, "grad_norm": 1.2683796882629395, "learning_rate": 3e-05, "loss": 3.3012, "step": 2638 }, { "epoch": 0.7171195652173913, "grad_norm": 1.0387176275253296, "learning_rate": 3e-05, "loss": 2.9117, "step": 2639 }, { "epoch": 0.717391304347826, "grad_norm": 1.5096675157546997, "learning_rate": 3e-05, "loss": 3.125, "step": 2640 }, { "epoch": 0.7176630434782608, "grad_norm": 1.1815725564956665, "learning_rate": 3e-05, "loss": 2.9721, "step": 2641 }, { "epoch": 0.7179347826086957, "grad_norm": 1.1765364408493042, "learning_rate": 3e-05, "loss": 3.3088, "step": 2642 }, { "epoch": 0.7182065217391305, "grad_norm": 1.087518572807312, "learning_rate": 3e-05, "loss": 3.1549, "step": 2643 }, { "epoch": 0.7184782608695652, "grad_norm": 1.1884835958480835, "learning_rate": 3e-05, "loss": 3.2079, "step": 2644 }, { "epoch": 0.71875, "grad_norm": 1.1198290586471558, "learning_rate": 3e-05, "loss": 2.9726, "step": 2645 }, { "epoch": 0.7190217391304348, "grad_norm": 1.2286672592163086, "learning_rate": 3e-05, "loss": 3.2783, "step": 2646 }, { "epoch": 0.7192934782608695, "grad_norm": 1.3148912191390991, "learning_rate": 3e-05, "loss": 3.914, "step": 2647 }, { "epoch": 0.7195652173913043, "grad_norm": 1.1227190494537354, "learning_rate": 3e-05, "loss": 3.1269, "step": 2648 }, { "epoch": 0.7198369565217392, "grad_norm": 1.3971885442733765, "learning_rate": 3e-05, "loss": 3.3385, "step": 2649 }, { "epoch": 0.720108695652174, "grad_norm": 1.0572166442871094, "learning_rate": 3e-05, "loss": 2.8119, "step": 2650 }, { "epoch": 0.7203804347826087, "grad_norm": 1.3299329280853271, "learning_rate": 3e-05, "loss": 3.5842, "step": 2651 }, { "epoch": 0.7206521739130435, "grad_norm": 1.2632654905319214, "learning_rate": 3e-05, "loss": 2.8276, "step": 2652 }, { "epoch": 0.7209239130434782, "grad_norm": 1.2086924314498901, "learning_rate": 3e-05, "loss": 3.3622, "step": 2653 }, { "epoch": 0.721195652173913, "grad_norm": 1.1867645978927612, "learning_rate": 3e-05, "loss": 3.1258, "step": 2654 }, { "epoch": 0.7214673913043478, "grad_norm": 1.123406171798706, "learning_rate": 3e-05, "loss": 3.3561, "step": 2655 }, { "epoch": 0.7217391304347827, "grad_norm": 1.3678390979766846, "learning_rate": 3e-05, "loss": 3.4183, "step": 2656 }, { "epoch": 0.7220108695652174, "grad_norm": 1.21256685256958, "learning_rate": 3e-05, "loss": 3.5207, "step": 2657 }, { "epoch": 0.7222826086956522, "grad_norm": 1.1133040189743042, "learning_rate": 3e-05, "loss": 2.8218, "step": 2658 }, { "epoch": 0.722554347826087, "grad_norm": 1.0846513509750366, "learning_rate": 3e-05, "loss": 2.9519, "step": 2659 }, { "epoch": 0.7228260869565217, "grad_norm": 1.0870620012283325, "learning_rate": 3e-05, "loss": 2.8142, "step": 2660 }, { "epoch": 0.7230978260869565, "grad_norm": 1.2407505512237549, "learning_rate": 3e-05, "loss": 3.0965, "step": 2661 }, { "epoch": 0.7233695652173913, "grad_norm": 1.0677931308746338, "learning_rate": 3e-05, "loss": 2.841, "step": 2662 }, { "epoch": 0.7236413043478261, "grad_norm": 1.1866375207901, "learning_rate": 3e-05, "loss": 3.2199, "step": 2663 }, { "epoch": 0.7239130434782609, "grad_norm": 1.4401320219039917, "learning_rate": 3e-05, "loss": 4.1391, "step": 2664 }, { "epoch": 0.7241847826086957, "grad_norm": 1.8525235652923584, "learning_rate": 3e-05, "loss": 2.954, "step": 2665 }, { "epoch": 0.7244565217391304, "grad_norm": 1.4307048320770264, "learning_rate": 3e-05, "loss": 3.6121, "step": 2666 }, { "epoch": 0.7247282608695652, "grad_norm": 1.4132686853408813, "learning_rate": 3e-05, "loss": 3.2657, "step": 2667 }, { "epoch": 0.725, "grad_norm": 1.1710587739944458, "learning_rate": 3e-05, "loss": 3.5075, "step": 2668 }, { "epoch": 0.7252717391304347, "grad_norm": 1.4402532577514648, "learning_rate": 3e-05, "loss": 3.5403, "step": 2669 }, { "epoch": 0.7255434782608695, "grad_norm": 1.0295389890670776, "learning_rate": 3e-05, "loss": 3.2313, "step": 2670 }, { "epoch": 0.7258152173913044, "grad_norm": 1.9836853742599487, "learning_rate": 3e-05, "loss": 3.7605, "step": 2671 }, { "epoch": 0.7260869565217392, "grad_norm": 1.313320279121399, "learning_rate": 3e-05, "loss": 2.892, "step": 2672 }, { "epoch": 0.7263586956521739, "grad_norm": 1.227516770362854, "learning_rate": 3e-05, "loss": 3.4072, "step": 2673 }, { "epoch": 0.7266304347826087, "grad_norm": 1.343525767326355, "learning_rate": 3e-05, "loss": 3.1093, "step": 2674 }, { "epoch": 0.7269021739130435, "grad_norm": 1.6472978591918945, "learning_rate": 3e-05, "loss": 3.2588, "step": 2675 }, { "epoch": 0.7271739130434782, "grad_norm": 1.313114047050476, "learning_rate": 3e-05, "loss": 3.0819, "step": 2676 }, { "epoch": 0.727445652173913, "grad_norm": 1.3546169996261597, "learning_rate": 3e-05, "loss": 3.3884, "step": 2677 }, { "epoch": 0.7277173913043479, "grad_norm": 1.4308792352676392, "learning_rate": 3e-05, "loss": 3.2228, "step": 2678 }, { "epoch": 0.7279891304347826, "grad_norm": 1.2303006649017334, "learning_rate": 3e-05, "loss": 2.9464, "step": 2679 }, { "epoch": 0.7282608695652174, "grad_norm": 2.12964129447937, "learning_rate": 3e-05, "loss": 4.1515, "step": 2680 }, { "epoch": 0.7285326086956522, "grad_norm": 1.1187695264816284, "learning_rate": 3e-05, "loss": 3.2245, "step": 2681 }, { "epoch": 0.7288043478260869, "grad_norm": 1.440588355064392, "learning_rate": 3e-05, "loss": 4.0875, "step": 2682 }, { "epoch": 0.7290760869565217, "grad_norm": 1.2928602695465088, "learning_rate": 3e-05, "loss": 3.4652, "step": 2683 }, { "epoch": 0.7293478260869565, "grad_norm": 1.1873844861984253, "learning_rate": 3e-05, "loss": 2.8352, "step": 2684 }, { "epoch": 0.7296195652173914, "grad_norm": 1.1879850625991821, "learning_rate": 3e-05, "loss": 2.8956, "step": 2685 }, { "epoch": 0.7298913043478261, "grad_norm": 1.206374168395996, "learning_rate": 3e-05, "loss": 3.3039, "step": 2686 }, { "epoch": 0.7301630434782609, "grad_norm": 1.1491773128509521, "learning_rate": 3e-05, "loss": 3.0314, "step": 2687 }, { "epoch": 0.7304347826086957, "grad_norm": 1.3866368532180786, "learning_rate": 3e-05, "loss": 3.055, "step": 2688 }, { "epoch": 0.7307065217391304, "grad_norm": 1.5674668550491333, "learning_rate": 3e-05, "loss": 3.1808, "step": 2689 }, { "epoch": 0.7309782608695652, "grad_norm": 1.562813639640808, "learning_rate": 3e-05, "loss": 3.2706, "step": 2690 }, { "epoch": 0.73125, "grad_norm": 1.6883511543273926, "learning_rate": 3e-05, "loss": 3.2192, "step": 2691 }, { "epoch": 0.7315217391304348, "grad_norm": 1.1254132986068726, "learning_rate": 3e-05, "loss": 3.1529, "step": 2692 }, { "epoch": 0.7317934782608696, "grad_norm": 1.0081984996795654, "learning_rate": 3e-05, "loss": 3.0864, "step": 2693 }, { "epoch": 0.7320652173913044, "grad_norm": 1.453481674194336, "learning_rate": 3e-05, "loss": 3.2169, "step": 2694 }, { "epoch": 0.7323369565217391, "grad_norm": 1.3893227577209473, "learning_rate": 3e-05, "loss": 3.3761, "step": 2695 }, { "epoch": 0.7326086956521739, "grad_norm": 1.7900173664093018, "learning_rate": 3e-05, "loss": 3.4638, "step": 2696 }, { "epoch": 0.7328804347826087, "grad_norm": 1.1031737327575684, "learning_rate": 3e-05, "loss": 2.6962, "step": 2697 }, { "epoch": 0.7331521739130434, "grad_norm": 1.3984167575836182, "learning_rate": 3e-05, "loss": 3.3052, "step": 2698 }, { "epoch": 0.7334239130434783, "grad_norm": 1.6182535886764526, "learning_rate": 3e-05, "loss": 3.4213, "step": 2699 }, { "epoch": 0.7336956521739131, "grad_norm": 1.7064732313156128, "learning_rate": 3e-05, "loss": 4.0381, "step": 2700 }, { "epoch": 0.7339673913043478, "grad_norm": 1.357704520225525, "learning_rate": 3e-05, "loss": 3.6925, "step": 2701 }, { "epoch": 0.7342391304347826, "grad_norm": 1.6365474462509155, "learning_rate": 3e-05, "loss": 3.2891, "step": 2702 }, { "epoch": 0.7345108695652174, "grad_norm": 1.8574589490890503, "learning_rate": 3e-05, "loss": 3.3637, "step": 2703 }, { "epoch": 0.7347826086956522, "grad_norm": 1.3467528820037842, "learning_rate": 3e-05, "loss": 3.1417, "step": 2704 }, { "epoch": 0.7350543478260869, "grad_norm": 1.4658093452453613, "learning_rate": 3e-05, "loss": 3.3401, "step": 2705 }, { "epoch": 0.7353260869565217, "grad_norm": 1.2771729230880737, "learning_rate": 3e-05, "loss": 3.3731, "step": 2706 }, { "epoch": 0.7355978260869566, "grad_norm": 1.0597422122955322, "learning_rate": 3e-05, "loss": 3.1315, "step": 2707 }, { "epoch": 0.7358695652173913, "grad_norm": 1.1597671508789062, "learning_rate": 3e-05, "loss": 2.6791, "step": 2708 }, { "epoch": 0.7361413043478261, "grad_norm": 1.7157272100448608, "learning_rate": 3e-05, "loss": 3.8816, "step": 2709 }, { "epoch": 0.7364130434782609, "grad_norm": 1.3123220205307007, "learning_rate": 3e-05, "loss": 3.6255, "step": 2710 }, { "epoch": 0.7366847826086956, "grad_norm": 1.2993775606155396, "learning_rate": 3e-05, "loss": 3.6012, "step": 2711 }, { "epoch": 0.7369565217391304, "grad_norm": 1.0929709672927856, "learning_rate": 3e-05, "loss": 3.3147, "step": 2712 }, { "epoch": 0.7372282608695652, "grad_norm": 1.2103468179702759, "learning_rate": 3e-05, "loss": 3.3148, "step": 2713 }, { "epoch": 0.7375, "grad_norm": 0.9965810179710388, "learning_rate": 3e-05, "loss": 3.2745, "step": 2714 }, { "epoch": 0.7377717391304348, "grad_norm": 1.7804696559906006, "learning_rate": 3e-05, "loss": 3.6753, "step": 2715 }, { "epoch": 0.7380434782608696, "grad_norm": 1.2880659103393555, "learning_rate": 3e-05, "loss": 3.4252, "step": 2716 }, { "epoch": 0.7383152173913043, "grad_norm": 0.9781152606010437, "learning_rate": 3e-05, "loss": 2.9325, "step": 2717 }, { "epoch": 0.7385869565217391, "grad_norm": 1.724921464920044, "learning_rate": 3e-05, "loss": 3.6383, "step": 2718 }, { "epoch": 0.7388586956521739, "grad_norm": 1.3509818315505981, "learning_rate": 3e-05, "loss": 3.2334, "step": 2719 }, { "epoch": 0.7391304347826086, "grad_norm": 1.2037723064422607, "learning_rate": 3e-05, "loss": 3.242, "step": 2720 }, { "epoch": 0.7394021739130435, "grad_norm": 1.337058424949646, "learning_rate": 3e-05, "loss": 3.5805, "step": 2721 }, { "epoch": 0.7396739130434783, "grad_norm": 1.3748770952224731, "learning_rate": 3e-05, "loss": 2.9626, "step": 2722 }, { "epoch": 0.7399456521739131, "grad_norm": 1.3539892435073853, "learning_rate": 3e-05, "loss": 3.0298, "step": 2723 }, { "epoch": 0.7402173913043478, "grad_norm": 0.979083776473999, "learning_rate": 3e-05, "loss": 2.7295, "step": 2724 }, { "epoch": 0.7404891304347826, "grad_norm": 1.7237200736999512, "learning_rate": 3e-05, "loss": 3.5329, "step": 2725 }, { "epoch": 0.7407608695652174, "grad_norm": 1.4173531532287598, "learning_rate": 3e-05, "loss": 3.5202, "step": 2726 }, { "epoch": 0.7410326086956521, "grad_norm": 1.3420840501785278, "learning_rate": 3e-05, "loss": 3.3239, "step": 2727 }, { "epoch": 0.741304347826087, "grad_norm": 1.3799097537994385, "learning_rate": 3e-05, "loss": 3.4807, "step": 2728 }, { "epoch": 0.7415760869565218, "grad_norm": 1.2215850353240967, "learning_rate": 3e-05, "loss": 3.5144, "step": 2729 }, { "epoch": 0.7418478260869565, "grad_norm": 1.2649718523025513, "learning_rate": 3e-05, "loss": 3.4898, "step": 2730 }, { "epoch": 0.7421195652173913, "grad_norm": 1.1176096200942993, "learning_rate": 3e-05, "loss": 3.4977, "step": 2731 }, { "epoch": 0.7423913043478261, "grad_norm": 1.0005884170532227, "learning_rate": 3e-05, "loss": 3.1848, "step": 2732 }, { "epoch": 0.7426630434782608, "grad_norm": 1.6306564807891846, "learning_rate": 3e-05, "loss": 3.8107, "step": 2733 }, { "epoch": 0.7429347826086956, "grad_norm": 1.6788973808288574, "learning_rate": 3e-05, "loss": 3.985, "step": 2734 }, { "epoch": 0.7432065217391305, "grad_norm": 1.16837739944458, "learning_rate": 3e-05, "loss": 2.8648, "step": 2735 }, { "epoch": 0.7434782608695653, "grad_norm": 1.324316382408142, "learning_rate": 3e-05, "loss": 3.5445, "step": 2736 }, { "epoch": 0.74375, "grad_norm": 1.2631101608276367, "learning_rate": 3e-05, "loss": 3.2311, "step": 2737 }, { "epoch": 0.7440217391304348, "grad_norm": 1.030339241027832, "learning_rate": 3e-05, "loss": 2.8981, "step": 2738 }, { "epoch": 0.7442934782608696, "grad_norm": 1.5260175466537476, "learning_rate": 3e-05, "loss": 3.2335, "step": 2739 }, { "epoch": 0.7445652173913043, "grad_norm": 1.610879898071289, "learning_rate": 3e-05, "loss": 3.3666, "step": 2740 }, { "epoch": 0.7448369565217391, "grad_norm": 1.290809988975525, "learning_rate": 3e-05, "loss": 2.9172, "step": 2741 }, { "epoch": 0.7451086956521739, "grad_norm": 1.2643457651138306, "learning_rate": 3e-05, "loss": 3.1168, "step": 2742 }, { "epoch": 0.7453804347826087, "grad_norm": 1.2268829345703125, "learning_rate": 3e-05, "loss": 3.2188, "step": 2743 }, { "epoch": 0.7456521739130435, "grad_norm": 1.022884726524353, "learning_rate": 3e-05, "loss": 2.884, "step": 2744 }, { "epoch": 0.7459239130434783, "grad_norm": 1.9909472465515137, "learning_rate": 3e-05, "loss": 4.1557, "step": 2745 }, { "epoch": 0.746195652173913, "grad_norm": 2.2413318157196045, "learning_rate": 3e-05, "loss": 4.0246, "step": 2746 }, { "epoch": 0.7464673913043478, "grad_norm": 1.0311791896820068, "learning_rate": 3e-05, "loss": 2.8403, "step": 2747 }, { "epoch": 0.7467391304347826, "grad_norm": 1.0483710765838623, "learning_rate": 3e-05, "loss": 2.7621, "step": 2748 }, { "epoch": 0.7470108695652173, "grad_norm": 0.9712722897529602, "learning_rate": 3e-05, "loss": 2.7713, "step": 2749 }, { "epoch": 0.7472826086956522, "grad_norm": 1.5642272233963013, "learning_rate": 3e-05, "loss": 3.2916, "step": 2750 }, { "epoch": 0.747554347826087, "grad_norm": 1.0362199544906616, "learning_rate": 3e-05, "loss": 3.0884, "step": 2751 }, { "epoch": 0.7478260869565218, "grad_norm": 1.2575013637542725, "learning_rate": 3e-05, "loss": 3.3603, "step": 2752 }, { "epoch": 0.7480978260869565, "grad_norm": 1.2161589860916138, "learning_rate": 3e-05, "loss": 3.3827, "step": 2753 }, { "epoch": 0.7483695652173913, "grad_norm": 1.0151880979537964, "learning_rate": 3e-05, "loss": 2.9711, "step": 2754 }, { "epoch": 0.748641304347826, "grad_norm": 1.3773422241210938, "learning_rate": 3e-05, "loss": 3.9046, "step": 2755 }, { "epoch": 0.7489130434782608, "grad_norm": 0.8108665943145752, "learning_rate": 3e-05, "loss": 2.9052, "step": 2756 }, { "epoch": 0.7491847826086957, "grad_norm": 0.8243381381034851, "learning_rate": 3e-05, "loss": 2.8174, "step": 2757 }, { "epoch": 0.7494565217391305, "grad_norm": 1.3421036005020142, "learning_rate": 3e-05, "loss": 3.3869, "step": 2758 }, { "epoch": 0.7497282608695652, "grad_norm": 1.1268045902252197, "learning_rate": 3e-05, "loss": 3.4369, "step": 2759 }, { "epoch": 0.75, "grad_norm": 0.982552707195282, "learning_rate": 3e-05, "loss": 3.0077, "step": 2760 }, { "epoch": 0.7502717391304348, "grad_norm": 1.548680305480957, "learning_rate": 3e-05, "loss": 2.9879, "step": 2761 }, { "epoch": 0.7505434782608695, "grad_norm": 1.1556179523468018, "learning_rate": 3e-05, "loss": 2.9654, "step": 2762 }, { "epoch": 0.7508152173913043, "grad_norm": 1.310678482055664, "learning_rate": 3e-05, "loss": 3.6287, "step": 2763 }, { "epoch": 0.7510869565217392, "grad_norm": 1.4762641191482544, "learning_rate": 3e-05, "loss": 3.6961, "step": 2764 }, { "epoch": 0.751358695652174, "grad_norm": 1.3265812397003174, "learning_rate": 3e-05, "loss": 3.2219, "step": 2765 }, { "epoch": 0.7516304347826087, "grad_norm": 1.2346985340118408, "learning_rate": 3e-05, "loss": 3.259, "step": 2766 }, { "epoch": 0.7519021739130435, "grad_norm": 0.9740216135978699, "learning_rate": 3e-05, "loss": 2.854, "step": 2767 }, { "epoch": 0.7521739130434782, "grad_norm": 1.2616466283798218, "learning_rate": 3e-05, "loss": 2.9113, "step": 2768 }, { "epoch": 0.752445652173913, "grad_norm": 1.4515007734298706, "learning_rate": 3e-05, "loss": 4.4309, "step": 2769 }, { "epoch": 0.7527173913043478, "grad_norm": 1.2112640142440796, "learning_rate": 3e-05, "loss": 3.1847, "step": 2770 }, { "epoch": 0.7529891304347827, "grad_norm": 1.123841643333435, "learning_rate": 3e-05, "loss": 3.3345, "step": 2771 }, { "epoch": 0.7532608695652174, "grad_norm": 1.1541368961334229, "learning_rate": 3e-05, "loss": 2.8234, "step": 2772 }, { "epoch": 0.7535326086956522, "grad_norm": 1.0209699869155884, "learning_rate": 3e-05, "loss": 3.1726, "step": 2773 }, { "epoch": 0.753804347826087, "grad_norm": 1.2185944318771362, "learning_rate": 3e-05, "loss": 3.8311, "step": 2774 }, { "epoch": 0.7540760869565217, "grad_norm": 1.553025722503662, "learning_rate": 3e-05, "loss": 3.2896, "step": 2775 }, { "epoch": 0.7543478260869565, "grad_norm": 1.5103728771209717, "learning_rate": 3e-05, "loss": 3.0133, "step": 2776 }, { "epoch": 0.7546195652173913, "grad_norm": 1.0733397006988525, "learning_rate": 3e-05, "loss": 3.0122, "step": 2777 }, { "epoch": 0.7548913043478261, "grad_norm": 1.4341579675674438, "learning_rate": 3e-05, "loss": 3.5646, "step": 2778 }, { "epoch": 0.7551630434782609, "grad_norm": 1.6829897165298462, "learning_rate": 3e-05, "loss": 3.6338, "step": 2779 }, { "epoch": 0.7554347826086957, "grad_norm": 1.0816172361373901, "learning_rate": 3e-05, "loss": 3.1679, "step": 2780 }, { "epoch": 0.7557065217391304, "grad_norm": 1.3089653253555298, "learning_rate": 3e-05, "loss": 3.6569, "step": 2781 }, { "epoch": 0.7559782608695652, "grad_norm": 1.329774022102356, "learning_rate": 3e-05, "loss": 3.0589, "step": 2782 }, { "epoch": 0.75625, "grad_norm": 1.5232934951782227, "learning_rate": 3e-05, "loss": 3.5649, "step": 2783 }, { "epoch": 0.7565217391304347, "grad_norm": 1.501591444015503, "learning_rate": 3e-05, "loss": 3.7119, "step": 2784 }, { "epoch": 0.7567934782608695, "grad_norm": 0.89922696352005, "learning_rate": 3e-05, "loss": 2.9755, "step": 2785 }, { "epoch": 0.7570652173913044, "grad_norm": 1.0538735389709473, "learning_rate": 3e-05, "loss": 3.4823, "step": 2786 }, { "epoch": 0.7573369565217392, "grad_norm": 0.8761924505233765, "learning_rate": 3e-05, "loss": 3.1571, "step": 2787 }, { "epoch": 0.7576086956521739, "grad_norm": 1.2133067846298218, "learning_rate": 3e-05, "loss": 3.3668, "step": 2788 }, { "epoch": 0.7578804347826087, "grad_norm": 1.0436729192733765, "learning_rate": 3e-05, "loss": 3.338, "step": 2789 }, { "epoch": 0.7581521739130435, "grad_norm": 1.0040472745895386, "learning_rate": 3e-05, "loss": 3.314, "step": 2790 }, { "epoch": 0.7584239130434782, "grad_norm": 1.019578456878662, "learning_rate": 3e-05, "loss": 3.3716, "step": 2791 }, { "epoch": 0.758695652173913, "grad_norm": 1.209367036819458, "learning_rate": 3e-05, "loss": 3.7626, "step": 2792 }, { "epoch": 0.7589673913043479, "grad_norm": 1.0667492151260376, "learning_rate": 3e-05, "loss": 3.4984, "step": 2793 }, { "epoch": 0.7592391304347826, "grad_norm": 1.200366497039795, "learning_rate": 3e-05, "loss": 3.228, "step": 2794 }, { "epoch": 0.7595108695652174, "grad_norm": 1.2726914882659912, "learning_rate": 3e-05, "loss": 2.9393, "step": 2795 }, { "epoch": 0.7597826086956522, "grad_norm": 1.244170904159546, "learning_rate": 3e-05, "loss": 2.9552, "step": 2796 }, { "epoch": 0.7600543478260869, "grad_norm": 1.113420844078064, "learning_rate": 3e-05, "loss": 3.1033, "step": 2797 }, { "epoch": 0.7603260869565217, "grad_norm": 1.9228954315185547, "learning_rate": 3e-05, "loss": 3.7877, "step": 2798 }, { "epoch": 0.7605978260869565, "grad_norm": 1.1893031597137451, "learning_rate": 3e-05, "loss": 3.0852, "step": 2799 }, { "epoch": 0.7608695652173914, "grad_norm": 1.524493932723999, "learning_rate": 3e-05, "loss": 4.1925, "step": 2800 }, { "epoch": 0.7611413043478261, "grad_norm": 1.5600111484527588, "learning_rate": 3e-05, "loss": 3.4646, "step": 2801 }, { "epoch": 0.7614130434782609, "grad_norm": 0.9608951210975647, "learning_rate": 3e-05, "loss": 2.8357, "step": 2802 }, { "epoch": 0.7616847826086957, "grad_norm": 1.2598545551300049, "learning_rate": 3e-05, "loss": 3.2298, "step": 2803 }, { "epoch": 0.7619565217391304, "grad_norm": 1.4729243516921997, "learning_rate": 3e-05, "loss": 3.2004, "step": 2804 }, { "epoch": 0.7622282608695652, "grad_norm": 1.04935622215271, "learning_rate": 3e-05, "loss": 2.9192, "step": 2805 }, { "epoch": 0.7625, "grad_norm": 1.1317492723464966, "learning_rate": 3e-05, "loss": 3.5182, "step": 2806 }, { "epoch": 0.7627717391304348, "grad_norm": 0.8651010394096375, "learning_rate": 3e-05, "loss": 2.7038, "step": 2807 }, { "epoch": 0.7630434782608696, "grad_norm": 1.4115127325057983, "learning_rate": 3e-05, "loss": 3.225, "step": 2808 }, { "epoch": 0.7633152173913044, "grad_norm": 1.0244847536087036, "learning_rate": 3e-05, "loss": 3.1254, "step": 2809 }, { "epoch": 0.7635869565217391, "grad_norm": 1.4117242097854614, "learning_rate": 3e-05, "loss": 3.4427, "step": 2810 }, { "epoch": 0.7638586956521739, "grad_norm": 1.3557825088500977, "learning_rate": 3e-05, "loss": 3.2669, "step": 2811 }, { "epoch": 0.7641304347826087, "grad_norm": 1.2034603357315063, "learning_rate": 3e-05, "loss": 3.77, "step": 2812 }, { "epoch": 0.7644021739130434, "grad_norm": 1.2670207023620605, "learning_rate": 3e-05, "loss": 3.1149, "step": 2813 }, { "epoch": 0.7646739130434783, "grad_norm": 0.9181430339813232, "learning_rate": 3e-05, "loss": 3.0427, "step": 2814 }, { "epoch": 0.7649456521739131, "grad_norm": 1.2132922410964966, "learning_rate": 3e-05, "loss": 3.4703, "step": 2815 }, { "epoch": 0.7652173913043478, "grad_norm": 0.9925717115402222, "learning_rate": 3e-05, "loss": 3.1396, "step": 2816 }, { "epoch": 0.7654891304347826, "grad_norm": 1.0748741626739502, "learning_rate": 3e-05, "loss": 2.8827, "step": 2817 }, { "epoch": 0.7657608695652174, "grad_norm": 1.0484018325805664, "learning_rate": 3e-05, "loss": 3.1837, "step": 2818 }, { "epoch": 0.7660326086956522, "grad_norm": 1.1004911661148071, "learning_rate": 3e-05, "loss": 3.135, "step": 2819 }, { "epoch": 0.7663043478260869, "grad_norm": 1.3505610227584839, "learning_rate": 3e-05, "loss": 3.3415, "step": 2820 }, { "epoch": 0.7665760869565217, "grad_norm": 1.1511911153793335, "learning_rate": 3e-05, "loss": 3.3407, "step": 2821 }, { "epoch": 0.7668478260869566, "grad_norm": 1.2488864660263062, "learning_rate": 3e-05, "loss": 3.0697, "step": 2822 }, { "epoch": 0.7671195652173913, "grad_norm": 0.970241129398346, "learning_rate": 3e-05, "loss": 2.6801, "step": 2823 }, { "epoch": 0.7673913043478261, "grad_norm": 1.3036609888076782, "learning_rate": 3e-05, "loss": 3.7613, "step": 2824 }, { "epoch": 0.7676630434782609, "grad_norm": 1.1077839136123657, "learning_rate": 3e-05, "loss": 3.829, "step": 2825 }, { "epoch": 0.7679347826086956, "grad_norm": 1.1794723272323608, "learning_rate": 3e-05, "loss": 3.1147, "step": 2826 }, { "epoch": 0.7682065217391304, "grad_norm": 1.460880994796753, "learning_rate": 3e-05, "loss": 3.7514, "step": 2827 }, { "epoch": 0.7684782608695652, "grad_norm": 1.2111482620239258, "learning_rate": 3e-05, "loss": 3.5524, "step": 2828 }, { "epoch": 0.76875, "grad_norm": 1.426535725593567, "learning_rate": 3e-05, "loss": 4.0174, "step": 2829 }, { "epoch": 0.7690217391304348, "grad_norm": 1.4052951335906982, "learning_rate": 3e-05, "loss": 3.3475, "step": 2830 }, { "epoch": 0.7692934782608696, "grad_norm": 0.9852263331413269, "learning_rate": 3e-05, "loss": 2.9218, "step": 2831 }, { "epoch": 0.7695652173913043, "grad_norm": 1.2342568635940552, "learning_rate": 3e-05, "loss": 3.1937, "step": 2832 }, { "epoch": 0.7698369565217391, "grad_norm": 2.569143772125244, "learning_rate": 3e-05, "loss": 3.8627, "step": 2833 }, { "epoch": 0.7701086956521739, "grad_norm": 1.07598876953125, "learning_rate": 3e-05, "loss": 2.8528, "step": 2834 }, { "epoch": 0.7703804347826086, "grad_norm": 1.0930018424987793, "learning_rate": 3e-05, "loss": 3.6242, "step": 2835 }, { "epoch": 0.7706521739130435, "grad_norm": 1.571661353111267, "learning_rate": 3e-05, "loss": 3.2027, "step": 2836 }, { "epoch": 0.7709239130434783, "grad_norm": 1.623907446861267, "learning_rate": 3e-05, "loss": 3.9722, "step": 2837 }, { "epoch": 0.7711956521739131, "grad_norm": 1.4214915037155151, "learning_rate": 3e-05, "loss": 3.483, "step": 2838 }, { "epoch": 0.7714673913043478, "grad_norm": 1.1177326440811157, "learning_rate": 3e-05, "loss": 3.2388, "step": 2839 }, { "epoch": 0.7717391304347826, "grad_norm": 1.2823110818862915, "learning_rate": 3e-05, "loss": 3.2644, "step": 2840 }, { "epoch": 0.7720108695652174, "grad_norm": 1.169673204421997, "learning_rate": 3e-05, "loss": 3.8629, "step": 2841 }, { "epoch": 0.7722826086956521, "grad_norm": 1.5543997287750244, "learning_rate": 3e-05, "loss": 3.6887, "step": 2842 }, { "epoch": 0.772554347826087, "grad_norm": 0.8735353946685791, "learning_rate": 3e-05, "loss": 2.7277, "step": 2843 }, { "epoch": 0.7728260869565218, "grad_norm": 0.8694882988929749, "learning_rate": 3e-05, "loss": 2.9491, "step": 2844 }, { "epoch": 0.7730978260869565, "grad_norm": 1.2447638511657715, "learning_rate": 3e-05, "loss": 3.473, "step": 2845 }, { "epoch": 0.7733695652173913, "grad_norm": 0.869049072265625, "learning_rate": 3e-05, "loss": 3.0047, "step": 2846 }, { "epoch": 0.7736413043478261, "grad_norm": 1.1336791515350342, "learning_rate": 3e-05, "loss": 3.5196, "step": 2847 }, { "epoch": 0.7739130434782608, "grad_norm": 1.0776828527450562, "learning_rate": 3e-05, "loss": 3.2469, "step": 2848 }, { "epoch": 0.7741847826086956, "grad_norm": 1.1838096380233765, "learning_rate": 3e-05, "loss": 3.5876, "step": 2849 }, { "epoch": 0.7744565217391305, "grad_norm": 0.9228270649909973, "learning_rate": 3e-05, "loss": 3.0038, "step": 2850 }, { "epoch": 0.7747282608695653, "grad_norm": 0.9622864127159119, "learning_rate": 3e-05, "loss": 2.7338, "step": 2851 }, { "epoch": 0.775, "grad_norm": 1.3739656209945679, "learning_rate": 3e-05, "loss": 3.3167, "step": 2852 }, { "epoch": 0.7752717391304348, "grad_norm": 1.2962409257888794, "learning_rate": 3e-05, "loss": 2.9095, "step": 2853 }, { "epoch": 0.7755434782608696, "grad_norm": 1.3018954992294312, "learning_rate": 3e-05, "loss": 3.4756, "step": 2854 }, { "epoch": 0.7758152173913043, "grad_norm": 1.2301543951034546, "learning_rate": 3e-05, "loss": 3.658, "step": 2855 }, { "epoch": 0.7760869565217391, "grad_norm": 0.9971374869346619, "learning_rate": 3e-05, "loss": 2.8921, "step": 2856 }, { "epoch": 0.7763586956521739, "grad_norm": 1.0504611730575562, "learning_rate": 3e-05, "loss": 2.907, "step": 2857 }, { "epoch": 0.7766304347826087, "grad_norm": 0.9376839399337769, "learning_rate": 3e-05, "loss": 2.9993, "step": 2858 }, { "epoch": 0.7769021739130435, "grad_norm": 1.1995012760162354, "learning_rate": 3e-05, "loss": 3.505, "step": 2859 }, { "epoch": 0.7771739130434783, "grad_norm": 1.2391998767852783, "learning_rate": 3e-05, "loss": 3.5183, "step": 2860 }, { "epoch": 0.777445652173913, "grad_norm": 1.0271795988082886, "learning_rate": 3e-05, "loss": 2.764, "step": 2861 }, { "epoch": 0.7777173913043478, "grad_norm": 1.1367555856704712, "learning_rate": 3e-05, "loss": 3.4225, "step": 2862 }, { "epoch": 0.7779891304347826, "grad_norm": 1.258692979812622, "learning_rate": 3e-05, "loss": 3.9128, "step": 2863 }, { "epoch": 0.7782608695652173, "grad_norm": 1.050280213356018, "learning_rate": 3e-05, "loss": 3.1644, "step": 2864 }, { "epoch": 0.7785326086956522, "grad_norm": 1.4861702919006348, "learning_rate": 3e-05, "loss": 3.651, "step": 2865 }, { "epoch": 0.778804347826087, "grad_norm": 1.9443000555038452, "learning_rate": 3e-05, "loss": 3.1064, "step": 2866 }, { "epoch": 0.7790760869565218, "grad_norm": 1.2355011701583862, "learning_rate": 3e-05, "loss": 3.2686, "step": 2867 }, { "epoch": 0.7793478260869565, "grad_norm": 1.06037175655365, "learning_rate": 3e-05, "loss": 3.4073, "step": 2868 }, { "epoch": 0.7796195652173913, "grad_norm": 1.3563976287841797, "learning_rate": 3e-05, "loss": 3.9419, "step": 2869 }, { "epoch": 0.779891304347826, "grad_norm": 1.411892294883728, "learning_rate": 3e-05, "loss": 3.4305, "step": 2870 }, { "epoch": 0.7801630434782608, "grad_norm": 2.1324245929718018, "learning_rate": 3e-05, "loss": 3.5014, "step": 2871 }, { "epoch": 0.7804347826086957, "grad_norm": 1.3413655757904053, "learning_rate": 3e-05, "loss": 3.4645, "step": 2872 }, { "epoch": 0.7807065217391305, "grad_norm": 1.2617512941360474, "learning_rate": 3e-05, "loss": 3.1992, "step": 2873 }, { "epoch": 0.7809782608695652, "grad_norm": 1.1358786821365356, "learning_rate": 3e-05, "loss": 3.223, "step": 2874 }, { "epoch": 0.78125, "grad_norm": 1.284077525138855, "learning_rate": 3e-05, "loss": 3.2483, "step": 2875 }, { "epoch": 0.7815217391304348, "grad_norm": 2.2102537155151367, "learning_rate": 3e-05, "loss": 3.3587, "step": 2876 }, { "epoch": 0.7817934782608695, "grad_norm": 1.2199740409851074, "learning_rate": 3e-05, "loss": 3.0096, "step": 2877 }, { "epoch": 0.7820652173913043, "grad_norm": 1.0134555101394653, "learning_rate": 3e-05, "loss": 3.0385, "step": 2878 }, { "epoch": 0.7823369565217392, "grad_norm": 0.9114274978637695, "learning_rate": 3e-05, "loss": 2.7133, "step": 2879 }, { "epoch": 0.782608695652174, "grad_norm": 1.0371575355529785, "learning_rate": 3e-05, "loss": 3.1674, "step": 2880 }, { "epoch": 0.7828804347826087, "grad_norm": 1.2628506422042847, "learning_rate": 3e-05, "loss": 3.5175, "step": 2881 }, { "epoch": 0.7831521739130435, "grad_norm": 1.2910369634628296, "learning_rate": 3e-05, "loss": 3.1921, "step": 2882 }, { "epoch": 0.7834239130434782, "grad_norm": 1.460605263710022, "learning_rate": 3e-05, "loss": 3.2101, "step": 2883 }, { "epoch": 0.783695652173913, "grad_norm": 1.4789795875549316, "learning_rate": 3e-05, "loss": 3.5643, "step": 2884 }, { "epoch": 0.7839673913043478, "grad_norm": 1.3687986135482788, "learning_rate": 3e-05, "loss": 3.5187, "step": 2885 }, { "epoch": 0.7842391304347827, "grad_norm": 1.4191471338272095, "learning_rate": 3e-05, "loss": 3.8682, "step": 2886 }, { "epoch": 0.7845108695652174, "grad_norm": 1.3148168325424194, "learning_rate": 3e-05, "loss": 3.0195, "step": 2887 }, { "epoch": 0.7847826086956522, "grad_norm": 1.15743088722229, "learning_rate": 3e-05, "loss": 2.9205, "step": 2888 }, { "epoch": 0.785054347826087, "grad_norm": 1.167377233505249, "learning_rate": 3e-05, "loss": 3.3717, "step": 2889 }, { "epoch": 0.7853260869565217, "grad_norm": 1.369253158569336, "learning_rate": 3e-05, "loss": 3.3943, "step": 2890 }, { "epoch": 0.7855978260869565, "grad_norm": 1.3081823587417603, "learning_rate": 3e-05, "loss": 3.1458, "step": 2891 }, { "epoch": 0.7858695652173913, "grad_norm": 1.0637850761413574, "learning_rate": 3e-05, "loss": 3.5853, "step": 2892 }, { "epoch": 0.7861413043478261, "grad_norm": 1.325531005859375, "learning_rate": 3e-05, "loss": 3.6246, "step": 2893 }, { "epoch": 0.7864130434782609, "grad_norm": 1.2909876108169556, "learning_rate": 3e-05, "loss": 2.7209, "step": 2894 }, { "epoch": 0.7866847826086957, "grad_norm": 1.072770357131958, "learning_rate": 3e-05, "loss": 2.7364, "step": 2895 }, { "epoch": 0.7869565217391304, "grad_norm": 1.2623767852783203, "learning_rate": 3e-05, "loss": 3.193, "step": 2896 }, { "epoch": 0.7872282608695652, "grad_norm": 1.262104868888855, "learning_rate": 3e-05, "loss": 3.4054, "step": 2897 }, { "epoch": 0.7875, "grad_norm": 0.9568426012992859, "learning_rate": 3e-05, "loss": 3.0418, "step": 2898 }, { "epoch": 0.7877717391304347, "grad_norm": 1.0304343700408936, "learning_rate": 3e-05, "loss": 3.1849, "step": 2899 }, { "epoch": 0.7880434782608695, "grad_norm": 1.1511303186416626, "learning_rate": 3e-05, "loss": 3.3258, "step": 2900 }, { "epoch": 0.7883152173913044, "grad_norm": 0.9511053562164307, "learning_rate": 3e-05, "loss": 2.9477, "step": 2901 }, { "epoch": 0.7885869565217392, "grad_norm": 1.1826517581939697, "learning_rate": 3e-05, "loss": 2.9749, "step": 2902 }, { "epoch": 0.7888586956521739, "grad_norm": 0.9371961355209351, "learning_rate": 3e-05, "loss": 3.0142, "step": 2903 }, { "epoch": 0.7891304347826087, "grad_norm": 1.3846261501312256, "learning_rate": 3e-05, "loss": 3.3297, "step": 2904 }, { "epoch": 0.7894021739130435, "grad_norm": 1.09294593334198, "learning_rate": 3e-05, "loss": 3.1767, "step": 2905 }, { "epoch": 0.7896739130434782, "grad_norm": 1.326731562614441, "learning_rate": 3e-05, "loss": 3.8467, "step": 2906 }, { "epoch": 0.789945652173913, "grad_norm": 1.1063183546066284, "learning_rate": 3e-05, "loss": 3.0861, "step": 2907 }, { "epoch": 0.7902173913043479, "grad_norm": 1.174807071685791, "learning_rate": 3e-05, "loss": 3.0249, "step": 2908 }, { "epoch": 0.7904891304347826, "grad_norm": 0.9046750068664551, "learning_rate": 3e-05, "loss": 3.0051, "step": 2909 }, { "epoch": 0.7907608695652174, "grad_norm": 2.0610415935516357, "learning_rate": 3e-05, "loss": 3.9009, "step": 2910 }, { "epoch": 0.7910326086956522, "grad_norm": 1.4573287963867188, "learning_rate": 3e-05, "loss": 3.1832, "step": 2911 }, { "epoch": 0.7913043478260869, "grad_norm": 1.1538214683532715, "learning_rate": 3e-05, "loss": 2.7802, "step": 2912 }, { "epoch": 0.7915760869565217, "grad_norm": 1.0094918012619019, "learning_rate": 3e-05, "loss": 3.0695, "step": 2913 }, { "epoch": 0.7918478260869565, "grad_norm": 1.1341625452041626, "learning_rate": 3e-05, "loss": 3.4901, "step": 2914 }, { "epoch": 0.7921195652173914, "grad_norm": 1.0618107318878174, "learning_rate": 3e-05, "loss": 3.1211, "step": 2915 }, { "epoch": 0.7923913043478261, "grad_norm": 1.4313887357711792, "learning_rate": 3e-05, "loss": 3.8207, "step": 2916 }, { "epoch": 0.7926630434782609, "grad_norm": 1.4587217569351196, "learning_rate": 3e-05, "loss": 3.5706, "step": 2917 }, { "epoch": 0.7929347826086957, "grad_norm": 1.4848532676696777, "learning_rate": 3e-05, "loss": 3.6406, "step": 2918 }, { "epoch": 0.7932065217391304, "grad_norm": 1.006595253944397, "learning_rate": 3e-05, "loss": 3.0846, "step": 2919 }, { "epoch": 0.7934782608695652, "grad_norm": 1.184501051902771, "learning_rate": 3e-05, "loss": 3.3719, "step": 2920 }, { "epoch": 0.79375, "grad_norm": 1.3015398979187012, "learning_rate": 3e-05, "loss": 2.9707, "step": 2921 }, { "epoch": 0.7940217391304348, "grad_norm": 1.001300573348999, "learning_rate": 3e-05, "loss": 2.9032, "step": 2922 }, { "epoch": 0.7942934782608696, "grad_norm": 1.4572199583053589, "learning_rate": 3e-05, "loss": 3.7695, "step": 2923 }, { "epoch": 0.7945652173913044, "grad_norm": 1.3405942916870117, "learning_rate": 3e-05, "loss": 2.8765, "step": 2924 }, { "epoch": 0.7948369565217391, "grad_norm": 1.6628684997558594, "learning_rate": 3e-05, "loss": 3.6848, "step": 2925 }, { "epoch": 0.7951086956521739, "grad_norm": 1.3848426342010498, "learning_rate": 3e-05, "loss": 3.3296, "step": 2926 }, { "epoch": 0.7953804347826087, "grad_norm": 1.5421236753463745, "learning_rate": 3e-05, "loss": 3.363, "step": 2927 }, { "epoch": 0.7956521739130434, "grad_norm": 1.3111261129379272, "learning_rate": 3e-05, "loss": 2.8346, "step": 2928 }, { "epoch": 0.7959239130434783, "grad_norm": 1.0846003293991089, "learning_rate": 3e-05, "loss": 3.3071, "step": 2929 }, { "epoch": 0.7961956521739131, "grad_norm": 0.9776524305343628, "learning_rate": 3e-05, "loss": 2.9533, "step": 2930 }, { "epoch": 0.7964673913043478, "grad_norm": 1.3840489387512207, "learning_rate": 3e-05, "loss": 3.4739, "step": 2931 }, { "epoch": 0.7967391304347826, "grad_norm": 1.2059979438781738, "learning_rate": 3e-05, "loss": 2.9474, "step": 2932 }, { "epoch": 0.7970108695652174, "grad_norm": 1.2110824584960938, "learning_rate": 3e-05, "loss": 3.2581, "step": 2933 }, { "epoch": 0.7972826086956522, "grad_norm": 1.0525647401809692, "learning_rate": 3e-05, "loss": 2.9519, "step": 2934 }, { "epoch": 0.7975543478260869, "grad_norm": 1.2633483409881592, "learning_rate": 3e-05, "loss": 3.1686, "step": 2935 }, { "epoch": 0.7978260869565217, "grad_norm": 0.8925679922103882, "learning_rate": 3e-05, "loss": 2.8761, "step": 2936 }, { "epoch": 0.7980978260869566, "grad_norm": 0.9802154898643494, "learning_rate": 3e-05, "loss": 3.2334, "step": 2937 }, { "epoch": 0.7983695652173913, "grad_norm": 1.3679046630859375, "learning_rate": 3e-05, "loss": 3.2836, "step": 2938 }, { "epoch": 0.7986413043478261, "grad_norm": 1.0334969758987427, "learning_rate": 3e-05, "loss": 3.1355, "step": 2939 }, { "epoch": 0.7989130434782609, "grad_norm": 1.0484052896499634, "learning_rate": 3e-05, "loss": 3.1652, "step": 2940 }, { "epoch": 0.7991847826086956, "grad_norm": 1.0501008033752441, "learning_rate": 3e-05, "loss": 2.9963, "step": 2941 }, { "epoch": 0.7994565217391304, "grad_norm": 1.2090955972671509, "learning_rate": 3e-05, "loss": 3.3553, "step": 2942 }, { "epoch": 0.7997282608695652, "grad_norm": 1.1819127798080444, "learning_rate": 3e-05, "loss": 3.1212, "step": 2943 }, { "epoch": 0.8, "grad_norm": 1.1051278114318848, "learning_rate": 3e-05, "loss": 3.2308, "step": 2944 }, { "epoch": 0.8002717391304348, "grad_norm": 0.9563131928443909, "learning_rate": 3e-05, "loss": 2.7561, "step": 2945 }, { "epoch": 0.8005434782608696, "grad_norm": 1.1029945611953735, "learning_rate": 3e-05, "loss": 3.6245, "step": 2946 }, { "epoch": 0.8008152173913043, "grad_norm": 1.3487423658370972, "learning_rate": 3e-05, "loss": 3.4707, "step": 2947 }, { "epoch": 0.8010869565217391, "grad_norm": 1.7481412887573242, "learning_rate": 3e-05, "loss": 3.6706, "step": 2948 }, { "epoch": 0.8013586956521739, "grad_norm": 1.0554581880569458, "learning_rate": 3e-05, "loss": 3.0406, "step": 2949 }, { "epoch": 0.8016304347826086, "grad_norm": 1.0953891277313232, "learning_rate": 3e-05, "loss": 2.7938, "step": 2950 }, { "epoch": 0.8019021739130435, "grad_norm": 1.3923141956329346, "learning_rate": 3e-05, "loss": 3.0691, "step": 2951 }, { "epoch": 0.8021739130434783, "grad_norm": 1.1058765649795532, "learning_rate": 3e-05, "loss": 2.9356, "step": 2952 }, { "epoch": 0.8024456521739131, "grad_norm": 1.1275757551193237, "learning_rate": 3e-05, "loss": 2.8867, "step": 2953 }, { "epoch": 0.8027173913043478, "grad_norm": 1.2130517959594727, "learning_rate": 3e-05, "loss": 3.1133, "step": 2954 }, { "epoch": 0.8029891304347826, "grad_norm": 1.6094809770584106, "learning_rate": 3e-05, "loss": 3.8152, "step": 2955 }, { "epoch": 0.8032608695652174, "grad_norm": 0.9870564341545105, "learning_rate": 3e-05, "loss": 2.721, "step": 2956 }, { "epoch": 0.8035326086956521, "grad_norm": 1.2996047735214233, "learning_rate": 3e-05, "loss": 3.6848, "step": 2957 }, { "epoch": 0.803804347826087, "grad_norm": 1.2500205039978027, "learning_rate": 3e-05, "loss": 3.3352, "step": 2958 }, { "epoch": 0.8040760869565218, "grad_norm": 1.3211798667907715, "learning_rate": 3e-05, "loss": 3.7284, "step": 2959 }, { "epoch": 0.8043478260869565, "grad_norm": 1.514905571937561, "learning_rate": 3e-05, "loss": 3.2425, "step": 2960 }, { "epoch": 0.8046195652173913, "grad_norm": 1.2107200622558594, "learning_rate": 3e-05, "loss": 3.4771, "step": 2961 }, { "epoch": 0.8048913043478261, "grad_norm": 1.707463264465332, "learning_rate": 3e-05, "loss": 3.6518, "step": 2962 }, { "epoch": 0.8051630434782608, "grad_norm": 0.9443992376327515, "learning_rate": 3e-05, "loss": 2.8471, "step": 2963 }, { "epoch": 0.8054347826086956, "grad_norm": 1.3588606119155884, "learning_rate": 3e-05, "loss": 3.3993, "step": 2964 }, { "epoch": 0.8057065217391305, "grad_norm": 1.2722697257995605, "learning_rate": 3e-05, "loss": 3.2316, "step": 2965 }, { "epoch": 0.8059782608695653, "grad_norm": 1.6636009216308594, "learning_rate": 3e-05, "loss": 3.5648, "step": 2966 }, { "epoch": 0.80625, "grad_norm": 1.3926972150802612, "learning_rate": 3e-05, "loss": 3.6403, "step": 2967 }, { "epoch": 0.8065217391304348, "grad_norm": 1.2489240169525146, "learning_rate": 3e-05, "loss": 3.6428, "step": 2968 }, { "epoch": 0.8067934782608696, "grad_norm": 1.3040508031845093, "learning_rate": 3e-05, "loss": 3.3698, "step": 2969 }, { "epoch": 0.8070652173913043, "grad_norm": 1.024529218673706, "learning_rate": 3e-05, "loss": 3.1765, "step": 2970 }, { "epoch": 0.8073369565217391, "grad_norm": 1.127133846282959, "learning_rate": 3e-05, "loss": 3.4167, "step": 2971 }, { "epoch": 0.8076086956521739, "grad_norm": 1.5928646326065063, "learning_rate": 3e-05, "loss": 3.2939, "step": 2972 }, { "epoch": 0.8078804347826087, "grad_norm": 1.099817156791687, "learning_rate": 3e-05, "loss": 3.3166, "step": 2973 }, { "epoch": 0.8081521739130435, "grad_norm": 0.9949479699134827, "learning_rate": 3e-05, "loss": 3.4637, "step": 2974 }, { "epoch": 0.8084239130434783, "grad_norm": 1.3012653589248657, "learning_rate": 3e-05, "loss": 3.3314, "step": 2975 }, { "epoch": 0.808695652173913, "grad_norm": 1.1788009405136108, "learning_rate": 3e-05, "loss": 2.9917, "step": 2976 }, { "epoch": 0.8089673913043478, "grad_norm": 1.5235824584960938, "learning_rate": 3e-05, "loss": 3.0784, "step": 2977 }, { "epoch": 0.8092391304347826, "grad_norm": 1.4393073320388794, "learning_rate": 3e-05, "loss": 3.8917, "step": 2978 }, { "epoch": 0.8095108695652173, "grad_norm": 1.432785987854004, "learning_rate": 3e-05, "loss": 3.8753, "step": 2979 }, { "epoch": 0.8097826086956522, "grad_norm": 1.4102272987365723, "learning_rate": 3e-05, "loss": 3.0096, "step": 2980 }, { "epoch": 0.810054347826087, "grad_norm": 1.5349915027618408, "learning_rate": 3e-05, "loss": 3.2429, "step": 2981 }, { "epoch": 0.8103260869565218, "grad_norm": 0.9741056561470032, "learning_rate": 3e-05, "loss": 3.2216, "step": 2982 }, { "epoch": 0.8105978260869565, "grad_norm": 1.3699191808700562, "learning_rate": 3e-05, "loss": 2.8594, "step": 2983 }, { "epoch": 0.8108695652173913, "grad_norm": 1.0808976888656616, "learning_rate": 3e-05, "loss": 3.1742, "step": 2984 }, { "epoch": 0.811141304347826, "grad_norm": 0.9032256007194519, "learning_rate": 3e-05, "loss": 3.0112, "step": 2985 }, { "epoch": 0.8114130434782608, "grad_norm": 1.119470477104187, "learning_rate": 3e-05, "loss": 3.1507, "step": 2986 }, { "epoch": 0.8116847826086957, "grad_norm": 1.3185955286026, "learning_rate": 3e-05, "loss": 3.7403, "step": 2987 }, { "epoch": 0.8119565217391305, "grad_norm": 1.283615231513977, "learning_rate": 3e-05, "loss": 3.2814, "step": 2988 }, { "epoch": 0.8122282608695652, "grad_norm": 1.2075817584991455, "learning_rate": 3e-05, "loss": 3.5796, "step": 2989 }, { "epoch": 0.8125, "grad_norm": 1.1944336891174316, "learning_rate": 3e-05, "loss": 3.3505, "step": 2990 }, { "epoch": 0.8127717391304348, "grad_norm": 1.1111395359039307, "learning_rate": 3e-05, "loss": 3.2605, "step": 2991 }, { "epoch": 0.8130434782608695, "grad_norm": 1.2699017524719238, "learning_rate": 3e-05, "loss": 3.4753, "step": 2992 }, { "epoch": 0.8133152173913043, "grad_norm": 1.1877753734588623, "learning_rate": 3e-05, "loss": 3.0174, "step": 2993 }, { "epoch": 0.8135869565217392, "grad_norm": 1.198492407798767, "learning_rate": 3e-05, "loss": 3.4792, "step": 2994 }, { "epoch": 0.813858695652174, "grad_norm": 1.2535799741744995, "learning_rate": 3e-05, "loss": 3.4246, "step": 2995 }, { "epoch": 0.8141304347826087, "grad_norm": 1.0552626848220825, "learning_rate": 3e-05, "loss": 2.9214, "step": 2996 }, { "epoch": 0.8144021739130435, "grad_norm": 1.1369494199752808, "learning_rate": 3e-05, "loss": 2.9704, "step": 2997 }, { "epoch": 0.8146739130434782, "grad_norm": 1.04477858543396, "learning_rate": 3e-05, "loss": 2.916, "step": 2998 }, { "epoch": 0.814945652173913, "grad_norm": 0.9620361328125, "learning_rate": 3e-05, "loss": 2.6156, "step": 2999 }, { "epoch": 0.8152173913043478, "grad_norm": 1.0974937677383423, "learning_rate": 3e-05, "loss": 3.362, "step": 3000 }, { "epoch": 0.8154891304347827, "grad_norm": 1.0228878259658813, "learning_rate": 3e-05, "loss": 3.3142, "step": 3001 }, { "epoch": 0.8157608695652174, "grad_norm": 1.0699377059936523, "learning_rate": 3e-05, "loss": 3.2705, "step": 3002 }, { "epoch": 0.8160326086956522, "grad_norm": 1.5110363960266113, "learning_rate": 3e-05, "loss": 3.9863, "step": 3003 }, { "epoch": 0.816304347826087, "grad_norm": 1.0896297693252563, "learning_rate": 3e-05, "loss": 3.2418, "step": 3004 }, { "epoch": 0.8165760869565217, "grad_norm": 1.0841624736785889, "learning_rate": 3e-05, "loss": 3.3033, "step": 3005 }, { "epoch": 0.8168478260869565, "grad_norm": 1.3732579946517944, "learning_rate": 3e-05, "loss": 3.3752, "step": 3006 }, { "epoch": 0.8171195652173913, "grad_norm": 1.2641088962554932, "learning_rate": 3e-05, "loss": 3.4179, "step": 3007 }, { "epoch": 0.8173913043478261, "grad_norm": 1.0250513553619385, "learning_rate": 3e-05, "loss": 2.8453, "step": 3008 }, { "epoch": 0.8176630434782609, "grad_norm": 1.1399973630905151, "learning_rate": 3e-05, "loss": 3.2173, "step": 3009 }, { "epoch": 0.8179347826086957, "grad_norm": 1.1633583307266235, "learning_rate": 3e-05, "loss": 2.7839, "step": 3010 }, { "epoch": 0.8182065217391304, "grad_norm": 1.114752173423767, "learning_rate": 3e-05, "loss": 3.0048, "step": 3011 }, { "epoch": 0.8184782608695652, "grad_norm": 1.6521841287612915, "learning_rate": 3e-05, "loss": 4.2389, "step": 3012 }, { "epoch": 0.81875, "grad_norm": 1.2092944383621216, "learning_rate": 3e-05, "loss": 3.2571, "step": 3013 }, { "epoch": 0.8190217391304347, "grad_norm": 1.091602087020874, "learning_rate": 3e-05, "loss": 3.1996, "step": 3014 }, { "epoch": 0.8192934782608695, "grad_norm": 0.8646924495697021, "learning_rate": 3e-05, "loss": 3.0836, "step": 3015 }, { "epoch": 0.8195652173913044, "grad_norm": 1.0290727615356445, "learning_rate": 3e-05, "loss": 2.9714, "step": 3016 }, { "epoch": 0.8198369565217392, "grad_norm": 0.9315720200538635, "learning_rate": 3e-05, "loss": 2.8955, "step": 3017 }, { "epoch": 0.8201086956521739, "grad_norm": 1.333827018737793, "learning_rate": 3e-05, "loss": 4.1577, "step": 3018 }, { "epoch": 0.8203804347826087, "grad_norm": 1.0600897073745728, "learning_rate": 3e-05, "loss": 3.4434, "step": 3019 }, { "epoch": 0.8206521739130435, "grad_norm": 1.7272323369979858, "learning_rate": 3e-05, "loss": 3.9848, "step": 3020 }, { "epoch": 0.8209239130434782, "grad_norm": 1.0585877895355225, "learning_rate": 3e-05, "loss": 2.8034, "step": 3021 }, { "epoch": 0.821195652173913, "grad_norm": 0.9289888143539429, "learning_rate": 3e-05, "loss": 2.938, "step": 3022 }, { "epoch": 0.8214673913043479, "grad_norm": 1.376469373703003, "learning_rate": 3e-05, "loss": 3.636, "step": 3023 }, { "epoch": 0.8217391304347826, "grad_norm": 1.3433892726898193, "learning_rate": 3e-05, "loss": 3.118, "step": 3024 }, { "epoch": 0.8220108695652174, "grad_norm": 1.2670068740844727, "learning_rate": 3e-05, "loss": 3.4596, "step": 3025 }, { "epoch": 0.8222826086956522, "grad_norm": 1.0194498300552368, "learning_rate": 3e-05, "loss": 3.0972, "step": 3026 }, { "epoch": 0.8225543478260869, "grad_norm": 1.7343618869781494, "learning_rate": 3e-05, "loss": 3.3819, "step": 3027 }, { "epoch": 0.8228260869565217, "grad_norm": 1.1823663711547852, "learning_rate": 3e-05, "loss": 3.2435, "step": 3028 }, { "epoch": 0.8230978260869565, "grad_norm": 1.1160571575164795, "learning_rate": 3e-05, "loss": 3.2883, "step": 3029 }, { "epoch": 0.8233695652173914, "grad_norm": 1.6302410364151, "learning_rate": 3e-05, "loss": 2.6802, "step": 3030 }, { "epoch": 0.8236413043478261, "grad_norm": 2.147940158843994, "learning_rate": 3e-05, "loss": 3.4556, "step": 3031 }, { "epoch": 0.8239130434782609, "grad_norm": 1.2989320755004883, "learning_rate": 3e-05, "loss": 3.1819, "step": 3032 }, { "epoch": 0.8241847826086957, "grad_norm": 1.2049676179885864, "learning_rate": 3e-05, "loss": 3.4831, "step": 3033 }, { "epoch": 0.8244565217391304, "grad_norm": 1.4658135175704956, "learning_rate": 3e-05, "loss": 3.6825, "step": 3034 }, { "epoch": 0.8247282608695652, "grad_norm": 1.271639108657837, "learning_rate": 3e-05, "loss": 3.4924, "step": 3035 }, { "epoch": 0.825, "grad_norm": 1.0066713094711304, "learning_rate": 3e-05, "loss": 3.376, "step": 3036 }, { "epoch": 0.8252717391304348, "grad_norm": 1.2980883121490479, "learning_rate": 3e-05, "loss": 3.328, "step": 3037 }, { "epoch": 0.8255434782608696, "grad_norm": 1.4236774444580078, "learning_rate": 3e-05, "loss": 2.9312, "step": 3038 }, { "epoch": 0.8258152173913044, "grad_norm": 1.3184988498687744, "learning_rate": 3e-05, "loss": 3.0132, "step": 3039 }, { "epoch": 0.8260869565217391, "grad_norm": 1.0756793022155762, "learning_rate": 3e-05, "loss": 3.4814, "step": 3040 }, { "epoch": 0.8263586956521739, "grad_norm": 1.240256428718567, "learning_rate": 3e-05, "loss": 3.3656, "step": 3041 }, { "epoch": 0.8266304347826087, "grad_norm": 1.0330684185028076, "learning_rate": 3e-05, "loss": 3.0126, "step": 3042 }, { "epoch": 0.8269021739130434, "grad_norm": 0.9498866200447083, "learning_rate": 3e-05, "loss": 2.7548, "step": 3043 }, { "epoch": 0.8271739130434783, "grad_norm": 0.9580128788948059, "learning_rate": 3e-05, "loss": 2.9633, "step": 3044 }, { "epoch": 0.8274456521739131, "grad_norm": 0.9247738718986511, "learning_rate": 3e-05, "loss": 3.1245, "step": 3045 }, { "epoch": 0.8277173913043478, "grad_norm": 0.9420803785324097, "learning_rate": 3e-05, "loss": 3.2215, "step": 3046 }, { "epoch": 0.8279891304347826, "grad_norm": 0.9758520722389221, "learning_rate": 3e-05, "loss": 2.6412, "step": 3047 }, { "epoch": 0.8282608695652174, "grad_norm": 1.035325288772583, "learning_rate": 3e-05, "loss": 3.3213, "step": 3048 }, { "epoch": 0.8285326086956522, "grad_norm": 1.6161421537399292, "learning_rate": 3e-05, "loss": 4.0357, "step": 3049 }, { "epoch": 0.8288043478260869, "grad_norm": 1.3096668720245361, "learning_rate": 3e-05, "loss": 3.7121, "step": 3050 }, { "epoch": 0.8290760869565217, "grad_norm": 0.9739846587181091, "learning_rate": 3e-05, "loss": 3.2274, "step": 3051 }, { "epoch": 0.8293478260869566, "grad_norm": 1.4004825353622437, "learning_rate": 3e-05, "loss": 3.2123, "step": 3052 }, { "epoch": 0.8296195652173913, "grad_norm": 1.3245350122451782, "learning_rate": 3e-05, "loss": 3.3188, "step": 3053 }, { "epoch": 0.8298913043478261, "grad_norm": 1.277363896369934, "learning_rate": 3e-05, "loss": 3.4315, "step": 3054 }, { "epoch": 0.8301630434782609, "grad_norm": 1.1860307455062866, "learning_rate": 3e-05, "loss": 3.2444, "step": 3055 }, { "epoch": 0.8304347826086956, "grad_norm": 1.4342072010040283, "learning_rate": 3e-05, "loss": 4.2132, "step": 3056 }, { "epoch": 0.8307065217391304, "grad_norm": 1.2908415794372559, "learning_rate": 3e-05, "loss": 3.4113, "step": 3057 }, { "epoch": 0.8309782608695652, "grad_norm": 0.9878079891204834, "learning_rate": 3e-05, "loss": 2.9514, "step": 3058 }, { "epoch": 0.83125, "grad_norm": 1.8127835988998413, "learning_rate": 3e-05, "loss": 3.8518, "step": 3059 }, { "epoch": 0.8315217391304348, "grad_norm": 1.3008536100387573, "learning_rate": 3e-05, "loss": 2.9649, "step": 3060 }, { "epoch": 0.8317934782608696, "grad_norm": 1.5720804929733276, "learning_rate": 3e-05, "loss": 3.9692, "step": 3061 }, { "epoch": 0.8320652173913043, "grad_norm": 1.2732257843017578, "learning_rate": 3e-05, "loss": 3.598, "step": 3062 }, { "epoch": 0.8323369565217391, "grad_norm": 1.817098617553711, "learning_rate": 3e-05, "loss": 3.7726, "step": 3063 }, { "epoch": 0.8326086956521739, "grad_norm": 1.3054240942001343, "learning_rate": 3e-05, "loss": 3.6484, "step": 3064 }, { "epoch": 0.8328804347826086, "grad_norm": 1.387664556503296, "learning_rate": 3e-05, "loss": 3.9251, "step": 3065 }, { "epoch": 0.8331521739130435, "grad_norm": 1.0745038986206055, "learning_rate": 3e-05, "loss": 3.7806, "step": 3066 }, { "epoch": 0.8334239130434783, "grad_norm": 0.9727368950843811, "learning_rate": 3e-05, "loss": 2.7217, "step": 3067 }, { "epoch": 0.8336956521739131, "grad_norm": 1.0093727111816406, "learning_rate": 3e-05, "loss": 2.9546, "step": 3068 }, { "epoch": 0.8339673913043478, "grad_norm": 1.3285881280899048, "learning_rate": 3e-05, "loss": 3.5858, "step": 3069 }, { "epoch": 0.8342391304347826, "grad_norm": 1.1906943321228027, "learning_rate": 3e-05, "loss": 2.6882, "step": 3070 }, { "epoch": 0.8345108695652174, "grad_norm": 1.162499189376831, "learning_rate": 3e-05, "loss": 3.3941, "step": 3071 }, { "epoch": 0.8347826086956521, "grad_norm": 1.5060298442840576, "learning_rate": 3e-05, "loss": 3.802, "step": 3072 }, { "epoch": 0.835054347826087, "grad_norm": 1.0174952745437622, "learning_rate": 3e-05, "loss": 2.7806, "step": 3073 }, { "epoch": 0.8353260869565218, "grad_norm": 0.9338005781173706, "learning_rate": 3e-05, "loss": 2.5784, "step": 3074 }, { "epoch": 0.8355978260869565, "grad_norm": 1.202101707458496, "learning_rate": 3e-05, "loss": 3.7435, "step": 3075 }, { "epoch": 0.8358695652173913, "grad_norm": 0.9823382496833801, "learning_rate": 3e-05, "loss": 2.77, "step": 3076 }, { "epoch": 0.8361413043478261, "grad_norm": 1.2876083850860596, "learning_rate": 3e-05, "loss": 3.1384, "step": 3077 }, { "epoch": 0.8364130434782608, "grad_norm": 0.9423876404762268, "learning_rate": 3e-05, "loss": 2.867, "step": 3078 }, { "epoch": 0.8366847826086956, "grad_norm": 1.1517866849899292, "learning_rate": 3e-05, "loss": 3.5569, "step": 3079 }, { "epoch": 0.8369565217391305, "grad_norm": 0.8807352185249329, "learning_rate": 3e-05, "loss": 2.9173, "step": 3080 }, { "epoch": 0.8372282608695653, "grad_norm": 1.3675475120544434, "learning_rate": 3e-05, "loss": 3.233, "step": 3081 }, { "epoch": 0.8375, "grad_norm": 0.976692259311676, "learning_rate": 3e-05, "loss": 2.6329, "step": 3082 }, { "epoch": 0.8377717391304348, "grad_norm": 0.9960673451423645, "learning_rate": 3e-05, "loss": 3.3855, "step": 3083 }, { "epoch": 0.8380434782608696, "grad_norm": 0.8318918347358704, "learning_rate": 3e-05, "loss": 3.1319, "step": 3084 }, { "epoch": 0.8383152173913043, "grad_norm": 0.9730811715126038, "learning_rate": 3e-05, "loss": 3.6126, "step": 3085 }, { "epoch": 0.8385869565217391, "grad_norm": 1.1164414882659912, "learning_rate": 3e-05, "loss": 3.2457, "step": 3086 }, { "epoch": 0.8388586956521739, "grad_norm": 1.1134709119796753, "learning_rate": 3e-05, "loss": 3.1498, "step": 3087 }, { "epoch": 0.8391304347826087, "grad_norm": 1.0676651000976562, "learning_rate": 3e-05, "loss": 3.0339, "step": 3088 }, { "epoch": 0.8394021739130435, "grad_norm": 1.4226154088974, "learning_rate": 3e-05, "loss": 3.6055, "step": 3089 }, { "epoch": 0.8396739130434783, "grad_norm": 1.1832042932510376, "learning_rate": 3e-05, "loss": 3.272, "step": 3090 }, { "epoch": 0.839945652173913, "grad_norm": 1.240876317024231, "learning_rate": 3e-05, "loss": 2.8104, "step": 3091 }, { "epoch": 0.8402173913043478, "grad_norm": 1.0383797883987427, "learning_rate": 3e-05, "loss": 2.8623, "step": 3092 }, { "epoch": 0.8404891304347826, "grad_norm": 1.1206902265548706, "learning_rate": 3e-05, "loss": 3.0109, "step": 3093 }, { "epoch": 0.8407608695652173, "grad_norm": 1.1229643821716309, "learning_rate": 3e-05, "loss": 3.0196, "step": 3094 }, { "epoch": 0.8410326086956522, "grad_norm": 0.9007440805435181, "learning_rate": 3e-05, "loss": 2.8602, "step": 3095 }, { "epoch": 0.841304347826087, "grad_norm": 1.1762633323669434, "learning_rate": 3e-05, "loss": 3.5676, "step": 3096 }, { "epoch": 0.8415760869565218, "grad_norm": 0.9211594462394714, "learning_rate": 3e-05, "loss": 3.2209, "step": 3097 }, { "epoch": 0.8418478260869565, "grad_norm": 1.324461817741394, "learning_rate": 3e-05, "loss": 3.0463, "step": 3098 }, { "epoch": 0.8421195652173913, "grad_norm": 1.0562752485275269, "learning_rate": 3e-05, "loss": 2.4726, "step": 3099 }, { "epoch": 0.842391304347826, "grad_norm": 0.9731100797653198, "learning_rate": 3e-05, "loss": 3.2275, "step": 3100 }, { "epoch": 0.8426630434782608, "grad_norm": 1.2750035524368286, "learning_rate": 3e-05, "loss": 3.2301, "step": 3101 }, { "epoch": 0.8429347826086957, "grad_norm": 1.4293423891067505, "learning_rate": 3e-05, "loss": 3.5761, "step": 3102 }, { "epoch": 0.8432065217391305, "grad_norm": 1.1072107553482056, "learning_rate": 3e-05, "loss": 3.1385, "step": 3103 }, { "epoch": 0.8434782608695652, "grad_norm": 0.9941693544387817, "learning_rate": 3e-05, "loss": 2.7235, "step": 3104 }, { "epoch": 0.84375, "grad_norm": 1.4560940265655518, "learning_rate": 3e-05, "loss": 3.2744, "step": 3105 }, { "epoch": 0.8440217391304348, "grad_norm": 1.4062750339508057, "learning_rate": 3e-05, "loss": 3.1, "step": 3106 }, { "epoch": 0.8442934782608695, "grad_norm": 1.0290594100952148, "learning_rate": 3e-05, "loss": 3.2149, "step": 3107 }, { "epoch": 0.8445652173913043, "grad_norm": 1.1329820156097412, "learning_rate": 3e-05, "loss": 3.5868, "step": 3108 }, { "epoch": 0.8448369565217392, "grad_norm": 1.0938060283660889, "learning_rate": 3e-05, "loss": 3.552, "step": 3109 }, { "epoch": 0.845108695652174, "grad_norm": 1.1490503549575806, "learning_rate": 3e-05, "loss": 3.2931, "step": 3110 }, { "epoch": 0.8453804347826087, "grad_norm": 1.1616952419281006, "learning_rate": 3e-05, "loss": 2.7943, "step": 3111 }, { "epoch": 0.8456521739130435, "grad_norm": 1.220522403717041, "learning_rate": 3e-05, "loss": 3.2521, "step": 3112 }, { "epoch": 0.8459239130434782, "grad_norm": 1.1222445964813232, "learning_rate": 3e-05, "loss": 3.1286, "step": 3113 }, { "epoch": 0.846195652173913, "grad_norm": 1.2962206602096558, "learning_rate": 3e-05, "loss": 3.1257, "step": 3114 }, { "epoch": 0.8464673913043478, "grad_norm": 1.2435656785964966, "learning_rate": 3e-05, "loss": 2.8689, "step": 3115 }, { "epoch": 0.8467391304347827, "grad_norm": 1.1281633377075195, "learning_rate": 3e-05, "loss": 3.2344, "step": 3116 }, { "epoch": 0.8470108695652174, "grad_norm": 1.426363468170166, "learning_rate": 3e-05, "loss": 3.4592, "step": 3117 }, { "epoch": 0.8472826086956522, "grad_norm": 1.5670675039291382, "learning_rate": 3e-05, "loss": 3.672, "step": 3118 }, { "epoch": 0.847554347826087, "grad_norm": 1.7653952836990356, "learning_rate": 3e-05, "loss": 3.3268, "step": 3119 }, { "epoch": 0.8478260869565217, "grad_norm": 1.2641386985778809, "learning_rate": 3e-05, "loss": 3.2548, "step": 3120 }, { "epoch": 0.8480978260869565, "grad_norm": 1.7537630796432495, "learning_rate": 3e-05, "loss": 3.6132, "step": 3121 }, { "epoch": 0.8483695652173913, "grad_norm": 1.1872942447662354, "learning_rate": 3e-05, "loss": 3.4434, "step": 3122 }, { "epoch": 0.8486413043478261, "grad_norm": 1.4559836387634277, "learning_rate": 3e-05, "loss": 3.1914, "step": 3123 }, { "epoch": 0.8489130434782609, "grad_norm": 1.3567631244659424, "learning_rate": 3e-05, "loss": 3.8778, "step": 3124 }, { "epoch": 0.8491847826086957, "grad_norm": 0.893852710723877, "learning_rate": 3e-05, "loss": 2.7665, "step": 3125 }, { "epoch": 0.8494565217391304, "grad_norm": 1.2900928258895874, "learning_rate": 3e-05, "loss": 3.5335, "step": 3126 }, { "epoch": 0.8497282608695652, "grad_norm": 1.327531099319458, "learning_rate": 3e-05, "loss": 2.8448, "step": 3127 }, { "epoch": 0.85, "grad_norm": 1.0268340110778809, "learning_rate": 3e-05, "loss": 3.0363, "step": 3128 }, { "epoch": 0.8502717391304347, "grad_norm": 1.4568902254104614, "learning_rate": 3e-05, "loss": 3.4417, "step": 3129 }, { "epoch": 0.8505434782608695, "grad_norm": 1.4307156801223755, "learning_rate": 3e-05, "loss": 3.5491, "step": 3130 }, { "epoch": 0.8508152173913044, "grad_norm": 1.5473573207855225, "learning_rate": 3e-05, "loss": 3.8162, "step": 3131 }, { "epoch": 0.8510869565217392, "grad_norm": 1.2859346866607666, "learning_rate": 3e-05, "loss": 3.4063, "step": 3132 }, { "epoch": 0.8513586956521739, "grad_norm": 1.3043359518051147, "learning_rate": 3e-05, "loss": 3.6142, "step": 3133 }, { "epoch": 0.8516304347826087, "grad_norm": 1.0424467325210571, "learning_rate": 3e-05, "loss": 3.2006, "step": 3134 }, { "epoch": 0.8519021739130435, "grad_norm": 1.096401572227478, "learning_rate": 3e-05, "loss": 3.2939, "step": 3135 }, { "epoch": 0.8521739130434782, "grad_norm": 1.5654884576797485, "learning_rate": 3e-05, "loss": 3.5884, "step": 3136 }, { "epoch": 0.852445652173913, "grad_norm": 1.015584945678711, "learning_rate": 3e-05, "loss": 2.8241, "step": 3137 }, { "epoch": 0.8527173913043479, "grad_norm": 1.2449654340744019, "learning_rate": 3e-05, "loss": 3.2508, "step": 3138 }, { "epoch": 0.8529891304347826, "grad_norm": 1.3450238704681396, "learning_rate": 3e-05, "loss": 3.8399, "step": 3139 }, { "epoch": 0.8532608695652174, "grad_norm": 1.1816442012786865, "learning_rate": 3e-05, "loss": 3.3473, "step": 3140 }, { "epoch": 0.8535326086956522, "grad_norm": 1.1475837230682373, "learning_rate": 3e-05, "loss": 3.2443, "step": 3141 }, { "epoch": 0.8538043478260869, "grad_norm": 0.9881013035774231, "learning_rate": 3e-05, "loss": 3.2166, "step": 3142 }, { "epoch": 0.8540760869565217, "grad_norm": 1.2011451721191406, "learning_rate": 3e-05, "loss": 3.3848, "step": 3143 }, { "epoch": 0.8543478260869565, "grad_norm": 1.2280961275100708, "learning_rate": 3e-05, "loss": 3.31, "step": 3144 }, { "epoch": 0.8546195652173914, "grad_norm": 1.537950038909912, "learning_rate": 3e-05, "loss": 3.4143, "step": 3145 }, { "epoch": 0.8548913043478261, "grad_norm": 1.2004420757293701, "learning_rate": 3e-05, "loss": 3.5707, "step": 3146 }, { "epoch": 0.8551630434782609, "grad_norm": 1.6398743391036987, "learning_rate": 3e-05, "loss": 3.1702, "step": 3147 }, { "epoch": 0.8554347826086957, "grad_norm": 1.4232219457626343, "learning_rate": 3e-05, "loss": 3.3429, "step": 3148 }, { "epoch": 0.8557065217391304, "grad_norm": 1.1291358470916748, "learning_rate": 3e-05, "loss": 3.1175, "step": 3149 }, { "epoch": 0.8559782608695652, "grad_norm": 1.0319854021072388, "learning_rate": 3e-05, "loss": 2.6689, "step": 3150 }, { "epoch": 0.85625, "grad_norm": 1.158753752708435, "learning_rate": 3e-05, "loss": 3.2033, "step": 3151 }, { "epoch": 0.8565217391304348, "grad_norm": 1.1246051788330078, "learning_rate": 3e-05, "loss": 2.934, "step": 3152 }, { "epoch": 0.8567934782608696, "grad_norm": 0.9106318354606628, "learning_rate": 3e-05, "loss": 2.6092, "step": 3153 }, { "epoch": 0.8570652173913044, "grad_norm": 1.3964134454727173, "learning_rate": 3e-05, "loss": 3.3104, "step": 3154 }, { "epoch": 0.8573369565217391, "grad_norm": 1.1211168766021729, "learning_rate": 3e-05, "loss": 3.0519, "step": 3155 }, { "epoch": 0.8576086956521739, "grad_norm": 1.3285335302352905, "learning_rate": 3e-05, "loss": 3.6055, "step": 3156 }, { "epoch": 0.8578804347826087, "grad_norm": 1.3057726621627808, "learning_rate": 3e-05, "loss": 3.7215, "step": 3157 }, { "epoch": 0.8581521739130434, "grad_norm": 1.4764635562896729, "learning_rate": 3e-05, "loss": 3.567, "step": 3158 }, { "epoch": 0.8584239130434783, "grad_norm": 1.202000379562378, "learning_rate": 3e-05, "loss": 3.4822, "step": 3159 }, { "epoch": 0.8586956521739131, "grad_norm": 1.0822246074676514, "learning_rate": 3e-05, "loss": 3.2965, "step": 3160 }, { "epoch": 0.8589673913043478, "grad_norm": 0.9603115320205688, "learning_rate": 3e-05, "loss": 3.0937, "step": 3161 }, { "epoch": 0.8592391304347826, "grad_norm": 1.190800666809082, "learning_rate": 3e-05, "loss": 3.2091, "step": 3162 }, { "epoch": 0.8595108695652174, "grad_norm": 1.1524766683578491, "learning_rate": 3e-05, "loss": 3.1559, "step": 3163 }, { "epoch": 0.8597826086956522, "grad_norm": 1.0062344074249268, "learning_rate": 3e-05, "loss": 3.2455, "step": 3164 }, { "epoch": 0.8600543478260869, "grad_norm": 1.0846439599990845, "learning_rate": 3e-05, "loss": 3.3582, "step": 3165 }, { "epoch": 0.8603260869565217, "grad_norm": 1.1138943433761597, "learning_rate": 3e-05, "loss": 2.9957, "step": 3166 }, { "epoch": 0.8605978260869566, "grad_norm": 1.2220076322555542, "learning_rate": 3e-05, "loss": 3.6103, "step": 3167 }, { "epoch": 0.8608695652173913, "grad_norm": 1.534000277519226, "learning_rate": 3e-05, "loss": 3.798, "step": 3168 }, { "epoch": 0.8611413043478261, "grad_norm": 1.349120855331421, "learning_rate": 3e-05, "loss": 3.8167, "step": 3169 }, { "epoch": 0.8614130434782609, "grad_norm": 1.3121823072433472, "learning_rate": 3e-05, "loss": 3.9782, "step": 3170 }, { "epoch": 0.8616847826086956, "grad_norm": 1.3369883298873901, "learning_rate": 3e-05, "loss": 3.0335, "step": 3171 }, { "epoch": 0.8619565217391304, "grad_norm": 1.9359731674194336, "learning_rate": 3e-05, "loss": 3.4322, "step": 3172 }, { "epoch": 0.8622282608695652, "grad_norm": 1.61549973487854, "learning_rate": 3e-05, "loss": 3.6562, "step": 3173 }, { "epoch": 0.8625, "grad_norm": 1.3874539136886597, "learning_rate": 3e-05, "loss": 3.449, "step": 3174 }, { "epoch": 0.8627717391304348, "grad_norm": 1.870835542678833, "learning_rate": 3e-05, "loss": 2.8775, "step": 3175 }, { "epoch": 0.8630434782608696, "grad_norm": 2.2991247177124023, "learning_rate": 3e-05, "loss": 4.1491, "step": 3176 }, { "epoch": 0.8633152173913043, "grad_norm": 1.2391786575317383, "learning_rate": 3e-05, "loss": 3.4684, "step": 3177 }, { "epoch": 0.8635869565217391, "grad_norm": 1.2693606615066528, "learning_rate": 3e-05, "loss": 3.2055, "step": 3178 }, { "epoch": 0.8638586956521739, "grad_norm": 1.7415162324905396, "learning_rate": 3e-05, "loss": 2.692, "step": 3179 }, { "epoch": 0.8641304347826086, "grad_norm": 1.57305908203125, "learning_rate": 3e-05, "loss": 3.6357, "step": 3180 }, { "epoch": 0.8644021739130435, "grad_norm": 0.970230758190155, "learning_rate": 3e-05, "loss": 3.2492, "step": 3181 }, { "epoch": 0.8646739130434783, "grad_norm": 1.3662259578704834, "learning_rate": 3e-05, "loss": 3.4164, "step": 3182 }, { "epoch": 0.8649456521739131, "grad_norm": 1.7679014205932617, "learning_rate": 3e-05, "loss": 3.7308, "step": 3183 }, { "epoch": 0.8652173913043478, "grad_norm": 1.2088621854782104, "learning_rate": 3e-05, "loss": 2.9937, "step": 3184 }, { "epoch": 0.8654891304347826, "grad_norm": 0.8111527562141418, "learning_rate": 3e-05, "loss": 2.6007, "step": 3185 }, { "epoch": 0.8657608695652174, "grad_norm": 1.3473327159881592, "learning_rate": 3e-05, "loss": 3.2987, "step": 3186 }, { "epoch": 0.8660326086956521, "grad_norm": 1.5230708122253418, "learning_rate": 3e-05, "loss": 2.923, "step": 3187 }, { "epoch": 0.866304347826087, "grad_norm": 1.0928703546524048, "learning_rate": 3e-05, "loss": 2.79, "step": 3188 }, { "epoch": 0.8665760869565218, "grad_norm": 0.952970027923584, "learning_rate": 3e-05, "loss": 2.9718, "step": 3189 }, { "epoch": 0.8668478260869565, "grad_norm": 1.2661546468734741, "learning_rate": 3e-05, "loss": 3.3439, "step": 3190 }, { "epoch": 0.8671195652173913, "grad_norm": 1.1697795391082764, "learning_rate": 3e-05, "loss": 2.6279, "step": 3191 }, { "epoch": 0.8673913043478261, "grad_norm": 1.0569891929626465, "learning_rate": 3e-05, "loss": 2.7563, "step": 3192 }, { "epoch": 0.8676630434782608, "grad_norm": 1.3665391206741333, "learning_rate": 3e-05, "loss": 3.9482, "step": 3193 }, { "epoch": 0.8679347826086956, "grad_norm": 1.4404640197753906, "learning_rate": 3e-05, "loss": 3.1148, "step": 3194 }, { "epoch": 0.8682065217391305, "grad_norm": 1.6767247915267944, "learning_rate": 3e-05, "loss": 3.7592, "step": 3195 }, { "epoch": 0.8684782608695653, "grad_norm": 1.1897008419036865, "learning_rate": 3e-05, "loss": 3.3022, "step": 3196 }, { "epoch": 0.86875, "grad_norm": 1.2555018663406372, "learning_rate": 3e-05, "loss": 3.5042, "step": 3197 }, { "epoch": 0.8690217391304348, "grad_norm": 1.5334973335266113, "learning_rate": 3e-05, "loss": 3.2271, "step": 3198 }, { "epoch": 0.8692934782608696, "grad_norm": 1.4210233688354492, "learning_rate": 3e-05, "loss": 3.4401, "step": 3199 }, { "epoch": 0.8695652173913043, "grad_norm": 1.2457329034805298, "learning_rate": 3e-05, "loss": 3.0819, "step": 3200 }, { "epoch": 0.8698369565217391, "grad_norm": 1.4019778966903687, "learning_rate": 3e-05, "loss": 3.3093, "step": 3201 }, { "epoch": 0.8701086956521739, "grad_norm": 1.6472575664520264, "learning_rate": 3e-05, "loss": 2.9264, "step": 3202 }, { "epoch": 0.8703804347826087, "grad_norm": 1.7156784534454346, "learning_rate": 3e-05, "loss": 3.2732, "step": 3203 }, { "epoch": 0.8706521739130435, "grad_norm": 0.9824452996253967, "learning_rate": 3e-05, "loss": 3.5759, "step": 3204 }, { "epoch": 0.8709239130434783, "grad_norm": 1.0671724081039429, "learning_rate": 3e-05, "loss": 3.2141, "step": 3205 }, { "epoch": 0.871195652173913, "grad_norm": 1.466063141822815, "learning_rate": 3e-05, "loss": 3.3825, "step": 3206 }, { "epoch": 0.8714673913043478, "grad_norm": 1.1988921165466309, "learning_rate": 3e-05, "loss": 2.7659, "step": 3207 }, { "epoch": 0.8717391304347826, "grad_norm": 0.9468457698822021, "learning_rate": 3e-05, "loss": 3.3256, "step": 3208 }, { "epoch": 0.8720108695652173, "grad_norm": 1.1663044691085815, "learning_rate": 3e-05, "loss": 3.5735, "step": 3209 }, { "epoch": 0.8722826086956522, "grad_norm": 1.3148590326309204, "learning_rate": 3e-05, "loss": 3.2213, "step": 3210 }, { "epoch": 0.872554347826087, "grad_norm": 1.931660771369934, "learning_rate": 3e-05, "loss": 4.0029, "step": 3211 }, { "epoch": 0.8728260869565218, "grad_norm": 1.226717472076416, "learning_rate": 3e-05, "loss": 3.7347, "step": 3212 }, { "epoch": 0.8730978260869565, "grad_norm": 1.2581689357757568, "learning_rate": 3e-05, "loss": 3.4094, "step": 3213 }, { "epoch": 0.8733695652173913, "grad_norm": 1.5382301807403564, "learning_rate": 3e-05, "loss": 2.9884, "step": 3214 }, { "epoch": 0.873641304347826, "grad_norm": 1.7043620347976685, "learning_rate": 3e-05, "loss": 3.6118, "step": 3215 }, { "epoch": 0.8739130434782608, "grad_norm": 1.0008423328399658, "learning_rate": 3e-05, "loss": 2.739, "step": 3216 }, { "epoch": 0.8741847826086957, "grad_norm": 1.4270468950271606, "learning_rate": 3e-05, "loss": 3.9526, "step": 3217 }, { "epoch": 0.8744565217391305, "grad_norm": 1.465566635131836, "learning_rate": 3e-05, "loss": 3.2112, "step": 3218 }, { "epoch": 0.8747282608695652, "grad_norm": 1.1289091110229492, "learning_rate": 3e-05, "loss": 3.4004, "step": 3219 }, { "epoch": 0.875, "grad_norm": 1.1125469207763672, "learning_rate": 3e-05, "loss": 3.3945, "step": 3220 }, { "epoch": 0.8752717391304348, "grad_norm": 1.0403220653533936, "learning_rate": 3e-05, "loss": 3.1545, "step": 3221 }, { "epoch": 0.8755434782608695, "grad_norm": 1.0171555280685425, "learning_rate": 3e-05, "loss": 3.2336, "step": 3222 }, { "epoch": 0.8758152173913043, "grad_norm": 1.17891526222229, "learning_rate": 3e-05, "loss": 2.939, "step": 3223 }, { "epoch": 0.8760869565217392, "grad_norm": 1.0108678340911865, "learning_rate": 3e-05, "loss": 2.9652, "step": 3224 }, { "epoch": 0.876358695652174, "grad_norm": 0.8684659600257874, "learning_rate": 3e-05, "loss": 2.6627, "step": 3225 }, { "epoch": 0.8766304347826087, "grad_norm": 0.9675855040550232, "learning_rate": 3e-05, "loss": 3.2626, "step": 3226 }, { "epoch": 0.8769021739130435, "grad_norm": 1.1135610342025757, "learning_rate": 3e-05, "loss": 2.8306, "step": 3227 }, { "epoch": 0.8771739130434782, "grad_norm": 0.9596325755119324, "learning_rate": 3e-05, "loss": 2.8982, "step": 3228 }, { "epoch": 0.877445652173913, "grad_norm": 1.1345157623291016, "learning_rate": 3e-05, "loss": 3.3746, "step": 3229 }, { "epoch": 0.8777173913043478, "grad_norm": 1.0093885660171509, "learning_rate": 3e-05, "loss": 3.3811, "step": 3230 }, { "epoch": 0.8779891304347827, "grad_norm": 1.4701224565505981, "learning_rate": 3e-05, "loss": 3.8729, "step": 3231 }, { "epoch": 0.8782608695652174, "grad_norm": 0.8455809354782104, "learning_rate": 3e-05, "loss": 2.889, "step": 3232 }, { "epoch": 0.8785326086956522, "grad_norm": 1.0636589527130127, "learning_rate": 3e-05, "loss": 3.2561, "step": 3233 }, { "epoch": 0.878804347826087, "grad_norm": 0.8937243223190308, "learning_rate": 3e-05, "loss": 2.7556, "step": 3234 }, { "epoch": 0.8790760869565217, "grad_norm": 1.276914119720459, "learning_rate": 3e-05, "loss": 3.3748, "step": 3235 }, { "epoch": 0.8793478260869565, "grad_norm": 1.4135632514953613, "learning_rate": 3e-05, "loss": 3.3929, "step": 3236 }, { "epoch": 0.8796195652173913, "grad_norm": 0.9119330644607544, "learning_rate": 3e-05, "loss": 3.229, "step": 3237 }, { "epoch": 0.8798913043478261, "grad_norm": 1.120521903038025, "learning_rate": 3e-05, "loss": 3.6578, "step": 3238 }, { "epoch": 0.8801630434782609, "grad_norm": 1.4400333166122437, "learning_rate": 3e-05, "loss": 3.6631, "step": 3239 }, { "epoch": 0.8804347826086957, "grad_norm": 1.6965535879135132, "learning_rate": 3e-05, "loss": 3.3765, "step": 3240 }, { "epoch": 0.8807065217391304, "grad_norm": 1.210610270500183, "learning_rate": 3e-05, "loss": 3.6085, "step": 3241 }, { "epoch": 0.8809782608695652, "grad_norm": 1.1407628059387207, "learning_rate": 3e-05, "loss": 2.9603, "step": 3242 }, { "epoch": 0.88125, "grad_norm": 1.430307388305664, "learning_rate": 3e-05, "loss": 3.2644, "step": 3243 }, { "epoch": 0.8815217391304347, "grad_norm": 1.3139044046401978, "learning_rate": 3e-05, "loss": 3.3523, "step": 3244 }, { "epoch": 0.8817934782608695, "grad_norm": 1.0071911811828613, "learning_rate": 3e-05, "loss": 3.3524, "step": 3245 }, { "epoch": 0.8820652173913044, "grad_norm": 1.5679545402526855, "learning_rate": 3e-05, "loss": 3.7991, "step": 3246 }, { "epoch": 0.8823369565217392, "grad_norm": 1.3887618780136108, "learning_rate": 3e-05, "loss": 3.0084, "step": 3247 }, { "epoch": 0.8826086956521739, "grad_norm": 1.1782461404800415, "learning_rate": 3e-05, "loss": 3.1136, "step": 3248 }, { "epoch": 0.8828804347826087, "grad_norm": 0.9846748113632202, "learning_rate": 3e-05, "loss": 2.5478, "step": 3249 }, { "epoch": 0.8831521739130435, "grad_norm": 1.3239892721176147, "learning_rate": 3e-05, "loss": 3.3812, "step": 3250 }, { "epoch": 0.8834239130434782, "grad_norm": 1.1534532308578491, "learning_rate": 3e-05, "loss": 3.4699, "step": 3251 }, { "epoch": 0.883695652173913, "grad_norm": 1.0228947401046753, "learning_rate": 3e-05, "loss": 3.2995, "step": 3252 }, { "epoch": 0.8839673913043479, "grad_norm": 1.1715861558914185, "learning_rate": 3e-05, "loss": 2.9041, "step": 3253 }, { "epoch": 0.8842391304347826, "grad_norm": 1.44595205783844, "learning_rate": 3e-05, "loss": 3.2008, "step": 3254 }, { "epoch": 0.8845108695652174, "grad_norm": 0.9128062129020691, "learning_rate": 3e-05, "loss": 3.0872, "step": 3255 }, { "epoch": 0.8847826086956522, "grad_norm": 0.9995028972625732, "learning_rate": 3e-05, "loss": 3.1744, "step": 3256 }, { "epoch": 0.8850543478260869, "grad_norm": 0.8776714205741882, "learning_rate": 3e-05, "loss": 2.9815, "step": 3257 }, { "epoch": 0.8853260869565217, "grad_norm": 0.9242172241210938, "learning_rate": 3e-05, "loss": 2.9738, "step": 3258 }, { "epoch": 0.8855978260869565, "grad_norm": 1.305087924003601, "learning_rate": 3e-05, "loss": 3.6002, "step": 3259 }, { "epoch": 0.8858695652173914, "grad_norm": 1.218959927558899, "learning_rate": 3e-05, "loss": 2.9891, "step": 3260 }, { "epoch": 0.8861413043478261, "grad_norm": 1.0651310682296753, "learning_rate": 3e-05, "loss": 3.2723, "step": 3261 }, { "epoch": 0.8864130434782609, "grad_norm": 1.420615792274475, "learning_rate": 3e-05, "loss": 3.9783, "step": 3262 }, { "epoch": 0.8866847826086957, "grad_norm": 1.7433851957321167, "learning_rate": 3e-05, "loss": 4.3755, "step": 3263 }, { "epoch": 0.8869565217391304, "grad_norm": 1.3065041303634644, "learning_rate": 3e-05, "loss": 3.132, "step": 3264 }, { "epoch": 0.8872282608695652, "grad_norm": 1.3422194719314575, "learning_rate": 3e-05, "loss": 3.1821, "step": 3265 }, { "epoch": 0.8875, "grad_norm": 1.0343096256256104, "learning_rate": 3e-05, "loss": 3.2091, "step": 3266 }, { "epoch": 0.8877717391304348, "grad_norm": 1.1553071737289429, "learning_rate": 3e-05, "loss": 3.2962, "step": 3267 }, { "epoch": 0.8880434782608696, "grad_norm": 1.3175245523452759, "learning_rate": 3e-05, "loss": 2.9399, "step": 3268 }, { "epoch": 0.8883152173913044, "grad_norm": 1.2289122343063354, "learning_rate": 3e-05, "loss": 3.3945, "step": 3269 }, { "epoch": 0.8885869565217391, "grad_norm": 0.8441329598426819, "learning_rate": 3e-05, "loss": 2.9865, "step": 3270 }, { "epoch": 0.8888586956521739, "grad_norm": 0.955934464931488, "learning_rate": 3e-05, "loss": 2.71, "step": 3271 }, { "epoch": 0.8891304347826087, "grad_norm": 1.2322384119033813, "learning_rate": 3e-05, "loss": 2.6869, "step": 3272 }, { "epoch": 0.8894021739130434, "grad_norm": 1.4021873474121094, "learning_rate": 3e-05, "loss": 3.5457, "step": 3273 }, { "epoch": 0.8896739130434783, "grad_norm": 1.5176516771316528, "learning_rate": 3e-05, "loss": 3.4049, "step": 3274 }, { "epoch": 0.8899456521739131, "grad_norm": 1.0812374353408813, "learning_rate": 3e-05, "loss": 2.7631, "step": 3275 }, { "epoch": 0.8902173913043478, "grad_norm": 1.0888370275497437, "learning_rate": 3e-05, "loss": 2.796, "step": 3276 }, { "epoch": 0.8904891304347826, "grad_norm": 1.0360549688339233, "learning_rate": 3e-05, "loss": 2.6445, "step": 3277 }, { "epoch": 0.8907608695652174, "grad_norm": 1.3518834114074707, "learning_rate": 3e-05, "loss": 3.8285, "step": 3278 }, { "epoch": 0.8910326086956522, "grad_norm": 1.34071946144104, "learning_rate": 3e-05, "loss": 4.0549, "step": 3279 }, { "epoch": 0.8913043478260869, "grad_norm": 1.6384416818618774, "learning_rate": 3e-05, "loss": 3.2126, "step": 3280 }, { "epoch": 0.8915760869565217, "grad_norm": 1.2500330209732056, "learning_rate": 3e-05, "loss": 3.79, "step": 3281 }, { "epoch": 0.8918478260869566, "grad_norm": 1.0918939113616943, "learning_rate": 3e-05, "loss": 3.4657, "step": 3282 }, { "epoch": 0.8921195652173913, "grad_norm": 1.119855284690857, "learning_rate": 3e-05, "loss": 2.6402, "step": 3283 }, { "epoch": 0.8923913043478261, "grad_norm": 1.1706061363220215, "learning_rate": 3e-05, "loss": 2.8103, "step": 3284 }, { "epoch": 0.8926630434782609, "grad_norm": 1.0363508462905884, "learning_rate": 3e-05, "loss": 2.6887, "step": 3285 }, { "epoch": 0.8929347826086956, "grad_norm": 1.1271647214889526, "learning_rate": 3e-05, "loss": 3.2676, "step": 3286 }, { "epoch": 0.8932065217391304, "grad_norm": 1.208321213722229, "learning_rate": 3e-05, "loss": 2.8407, "step": 3287 }, { "epoch": 0.8934782608695652, "grad_norm": 1.2740390300750732, "learning_rate": 3e-05, "loss": 2.9297, "step": 3288 }, { "epoch": 0.89375, "grad_norm": 1.1269636154174805, "learning_rate": 3e-05, "loss": 2.9789, "step": 3289 }, { "epoch": 0.8940217391304348, "grad_norm": 1.10430109500885, "learning_rate": 3e-05, "loss": 3.1109, "step": 3290 }, { "epoch": 0.8942934782608696, "grad_norm": 1.1625525951385498, "learning_rate": 3e-05, "loss": 3.445, "step": 3291 }, { "epoch": 0.8945652173913043, "grad_norm": 1.0440852642059326, "learning_rate": 3e-05, "loss": 3.0465, "step": 3292 }, { "epoch": 0.8948369565217391, "grad_norm": 0.9030226469039917, "learning_rate": 3e-05, "loss": 2.7886, "step": 3293 }, { "epoch": 0.8951086956521739, "grad_norm": 1.0526118278503418, "learning_rate": 3e-05, "loss": 3.1399, "step": 3294 }, { "epoch": 0.8953804347826086, "grad_norm": 1.5798497200012207, "learning_rate": 3e-05, "loss": 3.6915, "step": 3295 }, { "epoch": 0.8956521739130435, "grad_norm": 1.4318645000457764, "learning_rate": 3e-05, "loss": 3.298, "step": 3296 }, { "epoch": 0.8959239130434783, "grad_norm": 1.045587182044983, "learning_rate": 3e-05, "loss": 3.2022, "step": 3297 }, { "epoch": 0.8961956521739131, "grad_norm": 1.0729094743728638, "learning_rate": 3e-05, "loss": 3.2743, "step": 3298 }, { "epoch": 0.8964673913043478, "grad_norm": 1.0698198080062866, "learning_rate": 3e-05, "loss": 3.2366, "step": 3299 }, { "epoch": 0.8967391304347826, "grad_norm": 1.113480806350708, "learning_rate": 3e-05, "loss": 3.0156, "step": 3300 }, { "epoch": 0.8970108695652174, "grad_norm": 1.5028330087661743, "learning_rate": 3e-05, "loss": 3.1359, "step": 3301 }, { "epoch": 0.8972826086956521, "grad_norm": 1.1735856533050537, "learning_rate": 3e-05, "loss": 3.1369, "step": 3302 }, { "epoch": 0.897554347826087, "grad_norm": 1.3237791061401367, "learning_rate": 3e-05, "loss": 3.1095, "step": 3303 }, { "epoch": 0.8978260869565218, "grad_norm": 1.0287765264511108, "learning_rate": 3e-05, "loss": 2.7054, "step": 3304 }, { "epoch": 0.8980978260869565, "grad_norm": 1.2575358152389526, "learning_rate": 3e-05, "loss": 3.3384, "step": 3305 }, { "epoch": 0.8983695652173913, "grad_norm": 1.3265289068222046, "learning_rate": 3e-05, "loss": 3.2811, "step": 3306 }, { "epoch": 0.8986413043478261, "grad_norm": 1.2578668594360352, "learning_rate": 3e-05, "loss": 3.4697, "step": 3307 }, { "epoch": 0.8989130434782608, "grad_norm": 1.0790678262710571, "learning_rate": 3e-05, "loss": 3.2552, "step": 3308 }, { "epoch": 0.8991847826086956, "grad_norm": 0.9312275648117065, "learning_rate": 3e-05, "loss": 2.8223, "step": 3309 }, { "epoch": 0.8994565217391305, "grad_norm": 1.0071370601654053, "learning_rate": 3e-05, "loss": 3.0593, "step": 3310 }, { "epoch": 0.8997282608695653, "grad_norm": 1.2697527408599854, "learning_rate": 3e-05, "loss": 3.4125, "step": 3311 }, { "epoch": 0.9, "grad_norm": 1.4389897584915161, "learning_rate": 3e-05, "loss": 3.2935, "step": 3312 }, { "epoch": 0.9002717391304348, "grad_norm": 1.0937294960021973, "learning_rate": 3e-05, "loss": 3.1291, "step": 3313 }, { "epoch": 0.9005434782608696, "grad_norm": 1.262343168258667, "learning_rate": 3e-05, "loss": 3.8242, "step": 3314 }, { "epoch": 0.9008152173913043, "grad_norm": 1.0597748756408691, "learning_rate": 3e-05, "loss": 2.9029, "step": 3315 }, { "epoch": 0.9010869565217391, "grad_norm": 1.157682180404663, "learning_rate": 3e-05, "loss": 2.8939, "step": 3316 }, { "epoch": 0.9013586956521739, "grad_norm": 1.3134533166885376, "learning_rate": 3e-05, "loss": 3.4735, "step": 3317 }, { "epoch": 0.9016304347826087, "grad_norm": 1.289238691329956, "learning_rate": 3e-05, "loss": 3.7402, "step": 3318 }, { "epoch": 0.9019021739130435, "grad_norm": 1.2546768188476562, "learning_rate": 3e-05, "loss": 3.3855, "step": 3319 }, { "epoch": 0.9021739130434783, "grad_norm": 2.1356825828552246, "learning_rate": 3e-05, "loss": 3.3276, "step": 3320 }, { "epoch": 0.902445652173913, "grad_norm": 1.6080052852630615, "learning_rate": 3e-05, "loss": 3.5265, "step": 3321 }, { "epoch": 0.9027173913043478, "grad_norm": 1.3165130615234375, "learning_rate": 3e-05, "loss": 3.7407, "step": 3322 }, { "epoch": 0.9029891304347826, "grad_norm": 1.4776076078414917, "learning_rate": 3e-05, "loss": 3.6802, "step": 3323 }, { "epoch": 0.9032608695652173, "grad_norm": 1.4746235609054565, "learning_rate": 3e-05, "loss": 3.825, "step": 3324 }, { "epoch": 0.9035326086956522, "grad_norm": 1.0087367296218872, "learning_rate": 3e-05, "loss": 2.9628, "step": 3325 }, { "epoch": 0.903804347826087, "grad_norm": 1.4356976747512817, "learning_rate": 3e-05, "loss": 3.4988, "step": 3326 }, { "epoch": 0.9040760869565218, "grad_norm": 1.1169854402542114, "learning_rate": 3e-05, "loss": 3.2374, "step": 3327 }, { "epoch": 0.9043478260869565, "grad_norm": 1.1063263416290283, "learning_rate": 3e-05, "loss": 3.2963, "step": 3328 }, { "epoch": 0.9046195652173913, "grad_norm": 0.9434551000595093, "learning_rate": 3e-05, "loss": 3.4459, "step": 3329 }, { "epoch": 0.904891304347826, "grad_norm": 0.9521903395652771, "learning_rate": 3e-05, "loss": 2.7153, "step": 3330 }, { "epoch": 0.9051630434782608, "grad_norm": 1.2482932806015015, "learning_rate": 3e-05, "loss": 3.5719, "step": 3331 }, { "epoch": 0.9054347826086957, "grad_norm": 1.378918170928955, "learning_rate": 3e-05, "loss": 3.9698, "step": 3332 }, { "epoch": 0.9057065217391305, "grad_norm": 1.339396357536316, "learning_rate": 3e-05, "loss": 3.8912, "step": 3333 }, { "epoch": 0.9059782608695652, "grad_norm": 1.024444818496704, "learning_rate": 3e-05, "loss": 2.7524, "step": 3334 }, { "epoch": 0.90625, "grad_norm": 1.0689425468444824, "learning_rate": 3e-05, "loss": 3.344, "step": 3335 }, { "epoch": 0.9065217391304348, "grad_norm": 1.5281565189361572, "learning_rate": 3e-05, "loss": 4.0596, "step": 3336 }, { "epoch": 0.9067934782608695, "grad_norm": 1.1774648427963257, "learning_rate": 3e-05, "loss": 3.0477, "step": 3337 }, { "epoch": 0.9070652173913043, "grad_norm": 1.1559933423995972, "learning_rate": 3e-05, "loss": 3.2315, "step": 3338 }, { "epoch": 0.9073369565217392, "grad_norm": 1.3525007963180542, "learning_rate": 3e-05, "loss": 3.2819, "step": 3339 }, { "epoch": 0.907608695652174, "grad_norm": 1.2264772653579712, "learning_rate": 3e-05, "loss": 3.9097, "step": 3340 }, { "epoch": 0.9078804347826087, "grad_norm": 0.9528146386146545, "learning_rate": 3e-05, "loss": 3.1695, "step": 3341 }, { "epoch": 0.9081521739130435, "grad_norm": 1.4491996765136719, "learning_rate": 3e-05, "loss": 3.3542, "step": 3342 }, { "epoch": 0.9084239130434782, "grad_norm": 1.010594367980957, "learning_rate": 3e-05, "loss": 3.0192, "step": 3343 }, { "epoch": 0.908695652173913, "grad_norm": 1.2385438680648804, "learning_rate": 3e-05, "loss": 3.2475, "step": 3344 }, { "epoch": 0.9089673913043478, "grad_norm": 1.3691420555114746, "learning_rate": 3e-05, "loss": 3.6146, "step": 3345 }, { "epoch": 0.9092391304347827, "grad_norm": 1.3039157390594482, "learning_rate": 3e-05, "loss": 3.0749, "step": 3346 }, { "epoch": 0.9095108695652174, "grad_norm": 1.388312816619873, "learning_rate": 3e-05, "loss": 3.2633, "step": 3347 }, { "epoch": 0.9097826086956522, "grad_norm": 1.9077258110046387, "learning_rate": 3e-05, "loss": 4.2682, "step": 3348 }, { "epoch": 0.910054347826087, "grad_norm": 0.8595343828201294, "learning_rate": 3e-05, "loss": 2.654, "step": 3349 }, { "epoch": 0.9103260869565217, "grad_norm": 1.0038890838623047, "learning_rate": 3e-05, "loss": 3.3885, "step": 3350 }, { "epoch": 0.9105978260869565, "grad_norm": 1.0451005697250366, "learning_rate": 3e-05, "loss": 3.4294, "step": 3351 }, { "epoch": 0.9108695652173913, "grad_norm": 1.2645078897476196, "learning_rate": 3e-05, "loss": 3.1675, "step": 3352 }, { "epoch": 0.9111413043478261, "grad_norm": 1.1192885637283325, "learning_rate": 3e-05, "loss": 3.3621, "step": 3353 }, { "epoch": 0.9114130434782609, "grad_norm": 1.009670376777649, "learning_rate": 3e-05, "loss": 3.1018, "step": 3354 }, { "epoch": 0.9116847826086957, "grad_norm": 1.1403791904449463, "learning_rate": 3e-05, "loss": 3.5981, "step": 3355 }, { "epoch": 0.9119565217391304, "grad_norm": 1.2948347330093384, "learning_rate": 3e-05, "loss": 3.269, "step": 3356 }, { "epoch": 0.9122282608695652, "grad_norm": 1.2009344100952148, "learning_rate": 3e-05, "loss": 3.1199, "step": 3357 }, { "epoch": 0.9125, "grad_norm": 1.1034295558929443, "learning_rate": 3e-05, "loss": 3.2019, "step": 3358 }, { "epoch": 0.9127717391304347, "grad_norm": 0.9627982974052429, "learning_rate": 3e-05, "loss": 2.9578, "step": 3359 }, { "epoch": 0.9130434782608695, "grad_norm": 1.1057910919189453, "learning_rate": 3e-05, "loss": 3.35, "step": 3360 }, { "epoch": 0.9133152173913044, "grad_norm": 1.0927932262420654, "learning_rate": 3e-05, "loss": 3.2735, "step": 3361 }, { "epoch": 0.9135869565217392, "grad_norm": 0.9952607750892639, "learning_rate": 3e-05, "loss": 3.0225, "step": 3362 }, { "epoch": 0.9138586956521739, "grad_norm": 1.2180891036987305, "learning_rate": 3e-05, "loss": 3.372, "step": 3363 }, { "epoch": 0.9141304347826087, "grad_norm": 1.0906060934066772, "learning_rate": 3e-05, "loss": 2.8433, "step": 3364 }, { "epoch": 0.9144021739130435, "grad_norm": 0.8863520622253418, "learning_rate": 3e-05, "loss": 2.9024, "step": 3365 }, { "epoch": 0.9146739130434782, "grad_norm": 1.4132126569747925, "learning_rate": 3e-05, "loss": 3.278, "step": 3366 }, { "epoch": 0.914945652173913, "grad_norm": 1.2365199327468872, "learning_rate": 3e-05, "loss": 2.8892, "step": 3367 }, { "epoch": 0.9152173913043479, "grad_norm": 1.2023848295211792, "learning_rate": 3e-05, "loss": 3.4696, "step": 3368 }, { "epoch": 0.9154891304347826, "grad_norm": 1.077582597732544, "learning_rate": 3e-05, "loss": 2.8075, "step": 3369 }, { "epoch": 0.9157608695652174, "grad_norm": 1.1916066408157349, "learning_rate": 3e-05, "loss": 3.1826, "step": 3370 }, { "epoch": 0.9160326086956522, "grad_norm": 0.9405144453048706, "learning_rate": 3e-05, "loss": 3.2263, "step": 3371 }, { "epoch": 0.9163043478260869, "grad_norm": 1.0515216588974, "learning_rate": 3e-05, "loss": 3.3207, "step": 3372 }, { "epoch": 0.9165760869565217, "grad_norm": 1.1539192199707031, "learning_rate": 3e-05, "loss": 2.9277, "step": 3373 }, { "epoch": 0.9168478260869565, "grad_norm": 1.1389083862304688, "learning_rate": 3e-05, "loss": 3.0375, "step": 3374 }, { "epoch": 0.9171195652173914, "grad_norm": 1.1162118911743164, "learning_rate": 3e-05, "loss": 3.3001, "step": 3375 }, { "epoch": 0.9173913043478261, "grad_norm": 1.0966061353683472, "learning_rate": 3e-05, "loss": 3.2992, "step": 3376 }, { "epoch": 0.9176630434782609, "grad_norm": 0.9830874800682068, "learning_rate": 3e-05, "loss": 3.5812, "step": 3377 }, { "epoch": 0.9179347826086957, "grad_norm": 0.982811689376831, "learning_rate": 3e-05, "loss": 3.3548, "step": 3378 }, { "epoch": 0.9182065217391304, "grad_norm": 1.2300105094909668, "learning_rate": 3e-05, "loss": 3.163, "step": 3379 }, { "epoch": 0.9184782608695652, "grad_norm": 0.9209604859352112, "learning_rate": 3e-05, "loss": 3.0993, "step": 3380 }, { "epoch": 0.91875, "grad_norm": 1.1216840744018555, "learning_rate": 3e-05, "loss": 3.3593, "step": 3381 }, { "epoch": 0.9190217391304348, "grad_norm": 1.3138092756271362, "learning_rate": 3e-05, "loss": 3.6874, "step": 3382 }, { "epoch": 0.9192934782608696, "grad_norm": 1.2389674186706543, "learning_rate": 3e-05, "loss": 3.1537, "step": 3383 }, { "epoch": 0.9195652173913044, "grad_norm": 0.9438672065734863, "learning_rate": 3e-05, "loss": 3.0144, "step": 3384 }, { "epoch": 0.9198369565217391, "grad_norm": 1.1143367290496826, "learning_rate": 3e-05, "loss": 3.02, "step": 3385 }, { "epoch": 0.9201086956521739, "grad_norm": 1.4048357009887695, "learning_rate": 3e-05, "loss": 3.609, "step": 3386 }, { "epoch": 0.9203804347826087, "grad_norm": 1.0668329000473022, "learning_rate": 3e-05, "loss": 3.2136, "step": 3387 }, { "epoch": 0.9206521739130434, "grad_norm": 1.2218222618103027, "learning_rate": 3e-05, "loss": 2.8058, "step": 3388 }, { "epoch": 0.9209239130434783, "grad_norm": 1.2170056104660034, "learning_rate": 3e-05, "loss": 3.977, "step": 3389 }, { "epoch": 0.9211956521739131, "grad_norm": 1.1389089822769165, "learning_rate": 3e-05, "loss": 3.1123, "step": 3390 }, { "epoch": 0.9214673913043478, "grad_norm": 1.1248456239700317, "learning_rate": 3e-05, "loss": 3.4194, "step": 3391 }, { "epoch": 0.9217391304347826, "grad_norm": 0.8805920481681824, "learning_rate": 3e-05, "loss": 3.3877, "step": 3392 }, { "epoch": 0.9220108695652174, "grad_norm": 1.1866713762283325, "learning_rate": 3e-05, "loss": 3.3229, "step": 3393 }, { "epoch": 0.9222826086956522, "grad_norm": 1.2017425298690796, "learning_rate": 3e-05, "loss": 3.1533, "step": 3394 }, { "epoch": 0.9225543478260869, "grad_norm": 1.175695776939392, "learning_rate": 3e-05, "loss": 2.9847, "step": 3395 }, { "epoch": 0.9228260869565217, "grad_norm": 0.8511581420898438, "learning_rate": 3e-05, "loss": 2.5601, "step": 3396 }, { "epoch": 0.9230978260869566, "grad_norm": 1.8484715223312378, "learning_rate": 3e-05, "loss": 3.9292, "step": 3397 }, { "epoch": 0.9233695652173913, "grad_norm": 1.6159744262695312, "learning_rate": 3e-05, "loss": 3.137, "step": 3398 }, { "epoch": 0.9236413043478261, "grad_norm": 1.3892807960510254, "learning_rate": 3e-05, "loss": 3.0093, "step": 3399 }, { "epoch": 0.9239130434782609, "grad_norm": 1.2842003107070923, "learning_rate": 3e-05, "loss": 3.9279, "step": 3400 }, { "epoch": 0.9241847826086956, "grad_norm": 1.041723370552063, "learning_rate": 3e-05, "loss": 2.8251, "step": 3401 }, { "epoch": 0.9244565217391304, "grad_norm": 1.9582113027572632, "learning_rate": 3e-05, "loss": 3.4511, "step": 3402 }, { "epoch": 0.9247282608695652, "grad_norm": 1.8087749481201172, "learning_rate": 3e-05, "loss": 4.1364, "step": 3403 }, { "epoch": 0.925, "grad_norm": 1.6327838897705078, "learning_rate": 3e-05, "loss": 3.3022, "step": 3404 }, { "epoch": 0.9252717391304348, "grad_norm": 1.4796149730682373, "learning_rate": 3e-05, "loss": 3.1832, "step": 3405 }, { "epoch": 0.9255434782608696, "grad_norm": 1.868077039718628, "learning_rate": 3e-05, "loss": 3.656, "step": 3406 }, { "epoch": 0.9258152173913043, "grad_norm": 1.5386629104614258, "learning_rate": 3e-05, "loss": 3.3966, "step": 3407 }, { "epoch": 0.9260869565217391, "grad_norm": 1.507689356803894, "learning_rate": 3e-05, "loss": 3.2339, "step": 3408 }, { "epoch": 0.9263586956521739, "grad_norm": 1.3086180686950684, "learning_rate": 3e-05, "loss": 3.5104, "step": 3409 }, { "epoch": 0.9266304347826086, "grad_norm": 1.3591047525405884, "learning_rate": 3e-05, "loss": 3.142, "step": 3410 }, { "epoch": 0.9269021739130435, "grad_norm": 1.41386079788208, "learning_rate": 3e-05, "loss": 3.2576, "step": 3411 }, { "epoch": 0.9271739130434783, "grad_norm": 0.9256390333175659, "learning_rate": 3e-05, "loss": 3.0655, "step": 3412 }, { "epoch": 0.9274456521739131, "grad_norm": 1.1578412055969238, "learning_rate": 3e-05, "loss": 3.0678, "step": 3413 }, { "epoch": 0.9277173913043478, "grad_norm": 1.11700439453125, "learning_rate": 3e-05, "loss": 2.8778, "step": 3414 }, { "epoch": 0.9279891304347826, "grad_norm": 1.1300835609436035, "learning_rate": 3e-05, "loss": 3.4262, "step": 3415 }, { "epoch": 0.9282608695652174, "grad_norm": 1.252642035484314, "learning_rate": 3e-05, "loss": 3.0171, "step": 3416 }, { "epoch": 0.9285326086956521, "grad_norm": 1.4064898490905762, "learning_rate": 3e-05, "loss": 3.2901, "step": 3417 }, { "epoch": 0.928804347826087, "grad_norm": 1.2212474346160889, "learning_rate": 3e-05, "loss": 3.6426, "step": 3418 }, { "epoch": 0.9290760869565218, "grad_norm": 1.091092586517334, "learning_rate": 3e-05, "loss": 2.976, "step": 3419 }, { "epoch": 0.9293478260869565, "grad_norm": 1.142661452293396, "learning_rate": 3e-05, "loss": 3.25, "step": 3420 }, { "epoch": 0.9296195652173913, "grad_norm": 1.2335182428359985, "learning_rate": 3e-05, "loss": 2.5737, "step": 3421 }, { "epoch": 0.9298913043478261, "grad_norm": 1.7686213254928589, "learning_rate": 3e-05, "loss": 3.2724, "step": 3422 }, { "epoch": 0.9301630434782608, "grad_norm": 1.1885805130004883, "learning_rate": 3e-05, "loss": 3.1966, "step": 3423 }, { "epoch": 0.9304347826086956, "grad_norm": 1.5023466348648071, "learning_rate": 3e-05, "loss": 3.3998, "step": 3424 }, { "epoch": 0.9307065217391305, "grad_norm": 1.540440320968628, "learning_rate": 3e-05, "loss": 3.3097, "step": 3425 }, { "epoch": 0.9309782608695653, "grad_norm": 1.225467562675476, "learning_rate": 3e-05, "loss": 3.0512, "step": 3426 }, { "epoch": 0.93125, "grad_norm": 1.2939883470535278, "learning_rate": 3e-05, "loss": 2.934, "step": 3427 }, { "epoch": 0.9315217391304348, "grad_norm": 1.0971630811691284, "learning_rate": 3e-05, "loss": 3.3035, "step": 3428 }, { "epoch": 0.9317934782608696, "grad_norm": 1.4477622509002686, "learning_rate": 3e-05, "loss": 3.9893, "step": 3429 }, { "epoch": 0.9320652173913043, "grad_norm": 1.1087270975112915, "learning_rate": 3e-05, "loss": 3.3953, "step": 3430 }, { "epoch": 0.9323369565217391, "grad_norm": 1.2528973817825317, "learning_rate": 3e-05, "loss": 3.3526, "step": 3431 }, { "epoch": 0.9326086956521739, "grad_norm": 1.2172976732254028, "learning_rate": 3e-05, "loss": 3.3441, "step": 3432 }, { "epoch": 0.9328804347826087, "grad_norm": 1.5181169509887695, "learning_rate": 3e-05, "loss": 3.5499, "step": 3433 }, { "epoch": 0.9331521739130435, "grad_norm": 1.2571719884872437, "learning_rate": 3e-05, "loss": 3.3293, "step": 3434 }, { "epoch": 0.9334239130434783, "grad_norm": 1.218339204788208, "learning_rate": 3e-05, "loss": 3.4331, "step": 3435 }, { "epoch": 0.933695652173913, "grad_norm": 1.5465924739837646, "learning_rate": 3e-05, "loss": 3.4983, "step": 3436 }, { "epoch": 0.9339673913043478, "grad_norm": 1.1110074520111084, "learning_rate": 3e-05, "loss": 2.9831, "step": 3437 }, { "epoch": 0.9342391304347826, "grad_norm": 0.9091246128082275, "learning_rate": 3e-05, "loss": 2.7443, "step": 3438 }, { "epoch": 0.9345108695652173, "grad_norm": 1.1912291049957275, "learning_rate": 3e-05, "loss": 3.4729, "step": 3439 }, { "epoch": 0.9347826086956522, "grad_norm": 1.7168614864349365, "learning_rate": 3e-05, "loss": 3.8396, "step": 3440 }, { "epoch": 0.935054347826087, "grad_norm": 1.0588736534118652, "learning_rate": 3e-05, "loss": 3.1406, "step": 3441 }, { "epoch": 0.9353260869565218, "grad_norm": 1.0852594375610352, "learning_rate": 3e-05, "loss": 3.5512, "step": 3442 }, { "epoch": 0.9355978260869565, "grad_norm": 1.1496193408966064, "learning_rate": 3e-05, "loss": 3.0309, "step": 3443 }, { "epoch": 0.9358695652173913, "grad_norm": 1.212507724761963, "learning_rate": 3e-05, "loss": 3.4744, "step": 3444 }, { "epoch": 0.936141304347826, "grad_norm": 1.3551394939422607, "learning_rate": 3e-05, "loss": 3.5327, "step": 3445 }, { "epoch": 0.9364130434782608, "grad_norm": 1.1063857078552246, "learning_rate": 3e-05, "loss": 2.7333, "step": 3446 }, { "epoch": 0.9366847826086957, "grad_norm": 2.019047260284424, "learning_rate": 3e-05, "loss": 3.8059, "step": 3447 }, { "epoch": 0.9369565217391305, "grad_norm": 1.3297158479690552, "learning_rate": 3e-05, "loss": 3.0954, "step": 3448 }, { "epoch": 0.9372282608695652, "grad_norm": 0.9672344326972961, "learning_rate": 3e-05, "loss": 3.2951, "step": 3449 }, { "epoch": 0.9375, "grad_norm": 1.2703860998153687, "learning_rate": 3e-05, "loss": 3.0017, "step": 3450 }, { "epoch": 0.9377717391304348, "grad_norm": 1.1668411493301392, "learning_rate": 3e-05, "loss": 2.9217, "step": 3451 }, { "epoch": 0.9380434782608695, "grad_norm": 0.9579916000366211, "learning_rate": 3e-05, "loss": 3.1859, "step": 3452 }, { "epoch": 0.9383152173913043, "grad_norm": 1.984029769897461, "learning_rate": 3e-05, "loss": 3.5545, "step": 3453 }, { "epoch": 0.9385869565217392, "grad_norm": 1.7892874479293823, "learning_rate": 3e-05, "loss": 3.8938, "step": 3454 }, { "epoch": 0.938858695652174, "grad_norm": 1.3318513631820679, "learning_rate": 3e-05, "loss": 3.1979, "step": 3455 }, { "epoch": 0.9391304347826087, "grad_norm": 1.3399767875671387, "learning_rate": 3e-05, "loss": 3.5799, "step": 3456 }, { "epoch": 0.9394021739130435, "grad_norm": 1.5759323835372925, "learning_rate": 3e-05, "loss": 3.3052, "step": 3457 }, { "epoch": 0.9396739130434782, "grad_norm": 1.0409594774246216, "learning_rate": 3e-05, "loss": 3.2472, "step": 3458 }, { "epoch": 0.939945652173913, "grad_norm": 1.0662540197372437, "learning_rate": 3e-05, "loss": 2.9652, "step": 3459 }, { "epoch": 0.9402173913043478, "grad_norm": 1.5648605823516846, "learning_rate": 3e-05, "loss": 3.6519, "step": 3460 }, { "epoch": 0.9404891304347827, "grad_norm": 1.1745872497558594, "learning_rate": 3e-05, "loss": 3.1436, "step": 3461 }, { "epoch": 0.9407608695652174, "grad_norm": 1.0777145624160767, "learning_rate": 3e-05, "loss": 2.7773, "step": 3462 }, { "epoch": 0.9410326086956522, "grad_norm": 0.9803733229637146, "learning_rate": 3e-05, "loss": 2.7498, "step": 3463 }, { "epoch": 0.941304347826087, "grad_norm": 1.0982425212860107, "learning_rate": 3e-05, "loss": 3.413, "step": 3464 }, { "epoch": 0.9415760869565217, "grad_norm": 1.3649747371673584, "learning_rate": 3e-05, "loss": 3.0967, "step": 3465 }, { "epoch": 0.9418478260869565, "grad_norm": 1.4934804439544678, "learning_rate": 3e-05, "loss": 3.1274, "step": 3466 }, { "epoch": 0.9421195652173913, "grad_norm": 1.1337597370147705, "learning_rate": 3e-05, "loss": 2.6965, "step": 3467 }, { "epoch": 0.9423913043478261, "grad_norm": 1.2504925727844238, "learning_rate": 3e-05, "loss": 3.2437, "step": 3468 }, { "epoch": 0.9426630434782609, "grad_norm": 1.4370914697647095, "learning_rate": 3e-05, "loss": 3.5155, "step": 3469 }, { "epoch": 0.9429347826086957, "grad_norm": 1.259074330329895, "learning_rate": 3e-05, "loss": 3.9323, "step": 3470 }, { "epoch": 0.9432065217391304, "grad_norm": 1.247605323791504, "learning_rate": 3e-05, "loss": 3.465, "step": 3471 }, { "epoch": 0.9434782608695652, "grad_norm": 1.1649402379989624, "learning_rate": 3e-05, "loss": 3.2201, "step": 3472 }, { "epoch": 0.94375, "grad_norm": 1.123704433441162, "learning_rate": 3e-05, "loss": 2.6608, "step": 3473 }, { "epoch": 0.9440217391304347, "grad_norm": 0.972827672958374, "learning_rate": 3e-05, "loss": 2.7123, "step": 3474 }, { "epoch": 0.9442934782608695, "grad_norm": 1.2300611734390259, "learning_rate": 3e-05, "loss": 3.8955, "step": 3475 }, { "epoch": 0.9445652173913044, "grad_norm": 0.8878415822982788, "learning_rate": 3e-05, "loss": 2.8022, "step": 3476 }, { "epoch": 0.9448369565217392, "grad_norm": 1.3141851425170898, "learning_rate": 3e-05, "loss": 3.4375, "step": 3477 }, { "epoch": 0.9451086956521739, "grad_norm": 1.4346436262130737, "learning_rate": 3e-05, "loss": 3.9118, "step": 3478 }, { "epoch": 0.9453804347826087, "grad_norm": 1.22647225856781, "learning_rate": 3e-05, "loss": 2.9002, "step": 3479 }, { "epoch": 0.9456521739130435, "grad_norm": 1.0952470302581787, "learning_rate": 3e-05, "loss": 3.0309, "step": 3480 }, { "epoch": 0.9459239130434782, "grad_norm": 1.0914591550827026, "learning_rate": 3e-05, "loss": 3.6288, "step": 3481 }, { "epoch": 0.946195652173913, "grad_norm": 0.9968863725662231, "learning_rate": 3e-05, "loss": 3.2322, "step": 3482 }, { "epoch": 0.9464673913043479, "grad_norm": 1.3081971406936646, "learning_rate": 3e-05, "loss": 3.4543, "step": 3483 }, { "epoch": 0.9467391304347826, "grad_norm": 1.3062400817871094, "learning_rate": 3e-05, "loss": 3.32, "step": 3484 }, { "epoch": 0.9470108695652174, "grad_norm": 1.3035545349121094, "learning_rate": 3e-05, "loss": 3.484, "step": 3485 }, { "epoch": 0.9472826086956522, "grad_norm": 1.464484691619873, "learning_rate": 3e-05, "loss": 3.6629, "step": 3486 }, { "epoch": 0.9475543478260869, "grad_norm": 1.2064253091812134, "learning_rate": 3e-05, "loss": 2.8672, "step": 3487 }, { "epoch": 0.9478260869565217, "grad_norm": 1.4320194721221924, "learning_rate": 3e-05, "loss": 3.6001, "step": 3488 }, { "epoch": 0.9480978260869565, "grad_norm": 1.1308506727218628, "learning_rate": 3e-05, "loss": 3.5162, "step": 3489 }, { "epoch": 0.9483695652173914, "grad_norm": 1.2574654817581177, "learning_rate": 3e-05, "loss": 2.9263, "step": 3490 }, { "epoch": 0.9486413043478261, "grad_norm": 1.5057936906814575, "learning_rate": 3e-05, "loss": 3.5313, "step": 3491 }, { "epoch": 0.9489130434782609, "grad_norm": 1.3879592418670654, "learning_rate": 3e-05, "loss": 3.2695, "step": 3492 }, { "epoch": 0.9491847826086957, "grad_norm": 1.2129595279693604, "learning_rate": 3e-05, "loss": 3.426, "step": 3493 }, { "epoch": 0.9494565217391304, "grad_norm": 1.1940505504608154, "learning_rate": 3e-05, "loss": 3.8041, "step": 3494 }, { "epoch": 0.9497282608695652, "grad_norm": 1.167638897895813, "learning_rate": 3e-05, "loss": 2.9507, "step": 3495 }, { "epoch": 0.95, "grad_norm": 1.067775011062622, "learning_rate": 3e-05, "loss": 2.9318, "step": 3496 }, { "epoch": 0.9502717391304348, "grad_norm": 0.9276910424232483, "learning_rate": 3e-05, "loss": 2.9022, "step": 3497 }, { "epoch": 0.9505434782608696, "grad_norm": 1.5623066425323486, "learning_rate": 3e-05, "loss": 3.2125, "step": 3498 }, { "epoch": 0.9508152173913044, "grad_norm": 1.0339078903198242, "learning_rate": 3e-05, "loss": 2.8643, "step": 3499 }, { "epoch": 0.9510869565217391, "grad_norm": 1.3569916486740112, "learning_rate": 3e-05, "loss": 3.5325, "step": 3500 }, { "epoch": 0.9513586956521739, "grad_norm": 1.05780029296875, "learning_rate": 3e-05, "loss": 3.2485, "step": 3501 }, { "epoch": 0.9516304347826087, "grad_norm": 1.0890077352523804, "learning_rate": 3e-05, "loss": 2.8429, "step": 3502 }, { "epoch": 0.9519021739130434, "grad_norm": 0.9849225282669067, "learning_rate": 3e-05, "loss": 3.1688, "step": 3503 }, { "epoch": 0.9521739130434783, "grad_norm": 1.0427829027175903, "learning_rate": 3e-05, "loss": 3.372, "step": 3504 }, { "epoch": 0.9524456521739131, "grad_norm": 1.0842779874801636, "learning_rate": 3e-05, "loss": 2.6946, "step": 3505 }, { "epoch": 0.9527173913043478, "grad_norm": 1.0191032886505127, "learning_rate": 3e-05, "loss": 2.9598, "step": 3506 }, { "epoch": 0.9529891304347826, "grad_norm": 1.155341625213623, "learning_rate": 3e-05, "loss": 3.4415, "step": 3507 }, { "epoch": 0.9532608695652174, "grad_norm": 0.9909238219261169, "learning_rate": 3e-05, "loss": 3.4154, "step": 3508 }, { "epoch": 0.9535326086956522, "grad_norm": 1.2157588005065918, "learning_rate": 3e-05, "loss": 3.034, "step": 3509 }, { "epoch": 0.9538043478260869, "grad_norm": 1.0498162508010864, "learning_rate": 3e-05, "loss": 3.4161, "step": 3510 }, { "epoch": 0.9540760869565217, "grad_norm": 1.4054638147354126, "learning_rate": 3e-05, "loss": 2.6351, "step": 3511 }, { "epoch": 0.9543478260869566, "grad_norm": 1.115883231163025, "learning_rate": 3e-05, "loss": 3.1396, "step": 3512 }, { "epoch": 0.9546195652173913, "grad_norm": 1.088685154914856, "learning_rate": 3e-05, "loss": 3.6186, "step": 3513 }, { "epoch": 0.9548913043478261, "grad_norm": 1.3723245859146118, "learning_rate": 3e-05, "loss": 3.8247, "step": 3514 }, { "epoch": 0.9551630434782609, "grad_norm": 1.1690044403076172, "learning_rate": 3e-05, "loss": 3.3618, "step": 3515 }, { "epoch": 0.9554347826086956, "grad_norm": 1.0477054119110107, "learning_rate": 3e-05, "loss": 2.7343, "step": 3516 }, { "epoch": 0.9557065217391304, "grad_norm": 0.9991785287857056, "learning_rate": 3e-05, "loss": 3.0105, "step": 3517 }, { "epoch": 0.9559782608695652, "grad_norm": 0.9993420839309692, "learning_rate": 3e-05, "loss": 2.6458, "step": 3518 }, { "epoch": 0.95625, "grad_norm": 1.4663909673690796, "learning_rate": 3e-05, "loss": 3.5408, "step": 3519 }, { "epoch": 0.9565217391304348, "grad_norm": 1.4742854833602905, "learning_rate": 3e-05, "loss": 3.8588, "step": 3520 }, { "epoch": 0.9567934782608696, "grad_norm": 1.3884458541870117, "learning_rate": 3e-05, "loss": 2.9437, "step": 3521 }, { "epoch": 0.9570652173913043, "grad_norm": 1.3806692361831665, "learning_rate": 3e-05, "loss": 3.117, "step": 3522 }, { "epoch": 0.9573369565217391, "grad_norm": 1.3221436738967896, "learning_rate": 3e-05, "loss": 3.3902, "step": 3523 }, { "epoch": 0.9576086956521739, "grad_norm": 1.224565863609314, "learning_rate": 3e-05, "loss": 3.5285, "step": 3524 }, { "epoch": 0.9578804347826086, "grad_norm": 1.1438764333724976, "learning_rate": 3e-05, "loss": 3.4393, "step": 3525 }, { "epoch": 0.9581521739130435, "grad_norm": 1.342108964920044, "learning_rate": 3e-05, "loss": 3.509, "step": 3526 }, { "epoch": 0.9584239130434783, "grad_norm": 1.5581278800964355, "learning_rate": 3e-05, "loss": 3.1904, "step": 3527 }, { "epoch": 0.9586956521739131, "grad_norm": 1.614974021911621, "learning_rate": 3e-05, "loss": 4.2878, "step": 3528 }, { "epoch": 0.9589673913043478, "grad_norm": 1.3613301515579224, "learning_rate": 3e-05, "loss": 2.7037, "step": 3529 }, { "epoch": 0.9592391304347826, "grad_norm": 1.5549826622009277, "learning_rate": 3e-05, "loss": 3.0614, "step": 3530 }, { "epoch": 0.9595108695652174, "grad_norm": 1.1398217678070068, "learning_rate": 3e-05, "loss": 2.9882, "step": 3531 }, { "epoch": 0.9597826086956521, "grad_norm": 0.9870666861534119, "learning_rate": 3e-05, "loss": 3.2269, "step": 3532 }, { "epoch": 0.960054347826087, "grad_norm": 1.0095468759536743, "learning_rate": 3e-05, "loss": 2.7401, "step": 3533 }, { "epoch": 0.9603260869565218, "grad_norm": 1.436185359954834, "learning_rate": 3e-05, "loss": 3.7449, "step": 3534 }, { "epoch": 0.9605978260869565, "grad_norm": 1.9407174587249756, "learning_rate": 3e-05, "loss": 3.9864, "step": 3535 }, { "epoch": 0.9608695652173913, "grad_norm": 1.2536238431930542, "learning_rate": 3e-05, "loss": 2.9101, "step": 3536 }, { "epoch": 0.9611413043478261, "grad_norm": 1.211982250213623, "learning_rate": 3e-05, "loss": 3.7581, "step": 3537 }, { "epoch": 0.9614130434782608, "grad_norm": 1.236450433731079, "learning_rate": 3e-05, "loss": 3.124, "step": 3538 }, { "epoch": 0.9616847826086956, "grad_norm": 1.2030739784240723, "learning_rate": 3e-05, "loss": 3.3469, "step": 3539 }, { "epoch": 0.9619565217391305, "grad_norm": 0.9902325868606567, "learning_rate": 3e-05, "loss": 3.0698, "step": 3540 }, { "epoch": 0.9622282608695653, "grad_norm": 1.1244782209396362, "learning_rate": 3e-05, "loss": 2.9155, "step": 3541 }, { "epoch": 0.9625, "grad_norm": 1.2215620279312134, "learning_rate": 3e-05, "loss": 3.6743, "step": 3542 }, { "epoch": 0.9627717391304348, "grad_norm": 1.3267731666564941, "learning_rate": 3e-05, "loss": 3.5695, "step": 3543 }, { "epoch": 0.9630434782608696, "grad_norm": 0.8988152742385864, "learning_rate": 3e-05, "loss": 3.1857, "step": 3544 }, { "epoch": 0.9633152173913043, "grad_norm": 1.0987584590911865, "learning_rate": 3e-05, "loss": 2.9891, "step": 3545 }, { "epoch": 0.9635869565217391, "grad_norm": 0.982984185218811, "learning_rate": 3e-05, "loss": 3.3133, "step": 3546 }, { "epoch": 0.9638586956521739, "grad_norm": 1.2083067893981934, "learning_rate": 3e-05, "loss": 3.4338, "step": 3547 }, { "epoch": 0.9641304347826087, "grad_norm": 1.291344165802002, "learning_rate": 3e-05, "loss": 3.0172, "step": 3548 }, { "epoch": 0.9644021739130435, "grad_norm": 1.1700007915496826, "learning_rate": 3e-05, "loss": 3.3979, "step": 3549 }, { "epoch": 0.9646739130434783, "grad_norm": 1.1201428174972534, "learning_rate": 3e-05, "loss": 3.0888, "step": 3550 }, { "epoch": 0.964945652173913, "grad_norm": 0.9528276324272156, "learning_rate": 3e-05, "loss": 3.0396, "step": 3551 }, { "epoch": 0.9652173913043478, "grad_norm": 1.6611183881759644, "learning_rate": 3e-05, "loss": 3.2075, "step": 3552 }, { "epoch": 0.9654891304347826, "grad_norm": 1.4275506734848022, "learning_rate": 3e-05, "loss": 3.0498, "step": 3553 }, { "epoch": 0.9657608695652173, "grad_norm": 1.22553551197052, "learning_rate": 3e-05, "loss": 3.1952, "step": 3554 }, { "epoch": 0.9660326086956522, "grad_norm": 1.3704451322555542, "learning_rate": 3e-05, "loss": 3.5987, "step": 3555 }, { "epoch": 0.966304347826087, "grad_norm": 1.6007192134857178, "learning_rate": 3e-05, "loss": 3.0668, "step": 3556 }, { "epoch": 0.9665760869565218, "grad_norm": 1.2752271890640259, "learning_rate": 3e-05, "loss": 2.7751, "step": 3557 }, { "epoch": 0.9668478260869565, "grad_norm": 0.951418936252594, "learning_rate": 3e-05, "loss": 3.4574, "step": 3558 }, { "epoch": 0.9671195652173913, "grad_norm": 1.452092170715332, "learning_rate": 3e-05, "loss": 3.4391, "step": 3559 }, { "epoch": 0.967391304347826, "grad_norm": 1.2884318828582764, "learning_rate": 3e-05, "loss": 3.2205, "step": 3560 }, { "epoch": 0.9676630434782608, "grad_norm": 1.5570851564407349, "learning_rate": 3e-05, "loss": 3.7137, "step": 3561 }, { "epoch": 0.9679347826086957, "grad_norm": 0.9922688007354736, "learning_rate": 3e-05, "loss": 3.4295, "step": 3562 }, { "epoch": 0.9682065217391305, "grad_norm": 1.7942497730255127, "learning_rate": 3e-05, "loss": 3.3189, "step": 3563 }, { "epoch": 0.9684782608695652, "grad_norm": 1.421077847480774, "learning_rate": 3e-05, "loss": 3.8193, "step": 3564 }, { "epoch": 0.96875, "grad_norm": 1.0820494890213013, "learning_rate": 3e-05, "loss": 3.1531, "step": 3565 }, { "epoch": 0.9690217391304348, "grad_norm": 0.8867875933647156, "learning_rate": 3e-05, "loss": 3.1468, "step": 3566 }, { "epoch": 0.9692934782608695, "grad_norm": 1.167862892150879, "learning_rate": 3e-05, "loss": 3.3449, "step": 3567 }, { "epoch": 0.9695652173913043, "grad_norm": 1.1946779489517212, "learning_rate": 3e-05, "loss": 3.2855, "step": 3568 }, { "epoch": 0.9698369565217392, "grad_norm": 1.0145072937011719, "learning_rate": 3e-05, "loss": 3.0731, "step": 3569 }, { "epoch": 0.970108695652174, "grad_norm": 0.9989908337593079, "learning_rate": 3e-05, "loss": 2.9282, "step": 3570 }, { "epoch": 0.9703804347826087, "grad_norm": 1.4625117778778076, "learning_rate": 3e-05, "loss": 3.5186, "step": 3571 }, { "epoch": 0.9706521739130435, "grad_norm": 1.1812211275100708, "learning_rate": 3e-05, "loss": 3.1334, "step": 3572 }, { "epoch": 0.9709239130434782, "grad_norm": 0.9377380609512329, "learning_rate": 3e-05, "loss": 3.2545, "step": 3573 }, { "epoch": 0.971195652173913, "grad_norm": 0.8793758749961853, "learning_rate": 3e-05, "loss": 2.8575, "step": 3574 }, { "epoch": 0.9714673913043478, "grad_norm": 1.0345982313156128, "learning_rate": 3e-05, "loss": 2.9211, "step": 3575 }, { "epoch": 0.9717391304347827, "grad_norm": 1.0395225286483765, "learning_rate": 3e-05, "loss": 3.1647, "step": 3576 }, { "epoch": 0.9720108695652174, "grad_norm": 1.144873023033142, "learning_rate": 3e-05, "loss": 3.2006, "step": 3577 }, { "epoch": 0.9722826086956522, "grad_norm": 0.9547160267829895, "learning_rate": 3e-05, "loss": 3.448, "step": 3578 }, { "epoch": 0.972554347826087, "grad_norm": 1.033669352531433, "learning_rate": 3e-05, "loss": 3.0595, "step": 3579 }, { "epoch": 0.9728260869565217, "grad_norm": 1.1905789375305176, "learning_rate": 3e-05, "loss": 3.2033, "step": 3580 }, { "epoch": 0.9730978260869565, "grad_norm": 1.0761208534240723, "learning_rate": 3e-05, "loss": 3.047, "step": 3581 }, { "epoch": 0.9733695652173913, "grad_norm": 1.1482232809066772, "learning_rate": 3e-05, "loss": 3.423, "step": 3582 }, { "epoch": 0.9736413043478261, "grad_norm": 1.2702175378799438, "learning_rate": 3e-05, "loss": 3.4102, "step": 3583 }, { "epoch": 0.9739130434782609, "grad_norm": 1.342377781867981, "learning_rate": 3e-05, "loss": 2.9249, "step": 3584 }, { "epoch": 0.9741847826086957, "grad_norm": 1.0270349979400635, "learning_rate": 3e-05, "loss": 2.7235, "step": 3585 }, { "epoch": 0.9744565217391304, "grad_norm": 1.1443332433700562, "learning_rate": 3e-05, "loss": 3.5823, "step": 3586 }, { "epoch": 0.9747282608695652, "grad_norm": 1.4275180101394653, "learning_rate": 3e-05, "loss": 2.9542, "step": 3587 }, { "epoch": 0.975, "grad_norm": 1.380171537399292, "learning_rate": 3e-05, "loss": 3.9873, "step": 3588 }, { "epoch": 0.9752717391304347, "grad_norm": 1.120719075202942, "learning_rate": 3e-05, "loss": 3.3419, "step": 3589 }, { "epoch": 0.9755434782608695, "grad_norm": 1.069496512413025, "learning_rate": 3e-05, "loss": 2.7635, "step": 3590 }, { "epoch": 0.9758152173913044, "grad_norm": 1.449858546257019, "learning_rate": 3e-05, "loss": 3.5481, "step": 3591 }, { "epoch": 0.9760869565217392, "grad_norm": 1.3732768297195435, "learning_rate": 3e-05, "loss": 3.1, "step": 3592 }, { "epoch": 0.9763586956521739, "grad_norm": 1.129608392715454, "learning_rate": 3e-05, "loss": 3.1518, "step": 3593 }, { "epoch": 0.9766304347826087, "grad_norm": 1.32887864112854, "learning_rate": 3e-05, "loss": 3.6887, "step": 3594 }, { "epoch": 0.9769021739130435, "grad_norm": 0.886527419090271, "learning_rate": 3e-05, "loss": 2.6675, "step": 3595 }, { "epoch": 0.9771739130434782, "grad_norm": 0.9199744462966919, "learning_rate": 3e-05, "loss": 3.0289, "step": 3596 }, { "epoch": 0.977445652173913, "grad_norm": 1.402631163597107, "learning_rate": 3e-05, "loss": 3.9755, "step": 3597 }, { "epoch": 0.9777173913043479, "grad_norm": 1.6861292123794556, "learning_rate": 3e-05, "loss": 3.5657, "step": 3598 }, { "epoch": 0.9779891304347826, "grad_norm": 1.02433443069458, "learning_rate": 3e-05, "loss": 3.5024, "step": 3599 }, { "epoch": 0.9782608695652174, "grad_norm": 1.1592869758605957, "learning_rate": 3e-05, "loss": 3.166, "step": 3600 }, { "epoch": 0.9785326086956522, "grad_norm": 0.9945769906044006, "learning_rate": 3e-05, "loss": 2.9187, "step": 3601 }, { "epoch": 0.9788043478260869, "grad_norm": 1.2773929834365845, "learning_rate": 3e-05, "loss": 3.097, "step": 3602 }, { "epoch": 0.9790760869565217, "grad_norm": 0.8966383337974548, "learning_rate": 3e-05, "loss": 2.853, "step": 3603 }, { "epoch": 0.9793478260869565, "grad_norm": 1.030605673789978, "learning_rate": 3e-05, "loss": 3.2744, "step": 3604 }, { "epoch": 0.9796195652173914, "grad_norm": 1.0262856483459473, "learning_rate": 3e-05, "loss": 3.2838, "step": 3605 }, { "epoch": 0.9798913043478261, "grad_norm": 1.6333345174789429, "learning_rate": 3e-05, "loss": 3.61, "step": 3606 }, { "epoch": 0.9801630434782609, "grad_norm": 1.315659761428833, "learning_rate": 3e-05, "loss": 3.8466, "step": 3607 }, { "epoch": 0.9804347826086957, "grad_norm": 1.2973676919937134, "learning_rate": 3e-05, "loss": 3.2553, "step": 3608 }, { "epoch": 0.9807065217391304, "grad_norm": 1.081075668334961, "learning_rate": 3e-05, "loss": 3.3063, "step": 3609 }, { "epoch": 0.9809782608695652, "grad_norm": 1.2473063468933105, "learning_rate": 3e-05, "loss": 2.8419, "step": 3610 }, { "epoch": 0.98125, "grad_norm": 1.31900954246521, "learning_rate": 3e-05, "loss": 3.265, "step": 3611 }, { "epoch": 0.9815217391304348, "grad_norm": 1.3697569370269775, "learning_rate": 3e-05, "loss": 4.0451, "step": 3612 }, { "epoch": 0.9817934782608696, "grad_norm": 1.3317036628723145, "learning_rate": 3e-05, "loss": 3.3117, "step": 3613 }, { "epoch": 0.9820652173913044, "grad_norm": 1.6604176759719849, "learning_rate": 3e-05, "loss": 2.6856, "step": 3614 }, { "epoch": 0.9823369565217391, "grad_norm": 0.9710596799850464, "learning_rate": 3e-05, "loss": 2.7038, "step": 3615 }, { "epoch": 0.9826086956521739, "grad_norm": 0.9689478874206543, "learning_rate": 3e-05, "loss": 3.4549, "step": 3616 }, { "epoch": 0.9828804347826087, "grad_norm": 1.5066816806793213, "learning_rate": 3e-05, "loss": 3.7657, "step": 3617 }, { "epoch": 0.9831521739130434, "grad_norm": 1.4727463722229004, "learning_rate": 3e-05, "loss": 3.3064, "step": 3618 }, { "epoch": 0.9834239130434783, "grad_norm": 1.3264317512512207, "learning_rate": 3e-05, "loss": 3.3373, "step": 3619 }, { "epoch": 0.9836956521739131, "grad_norm": 1.066543459892273, "learning_rate": 3e-05, "loss": 2.768, "step": 3620 }, { "epoch": 0.9839673913043478, "grad_norm": 1.572569727897644, "learning_rate": 3e-05, "loss": 3.5364, "step": 3621 }, { "epoch": 0.9842391304347826, "grad_norm": 1.2815150022506714, "learning_rate": 3e-05, "loss": 3.0281, "step": 3622 }, { "epoch": 0.9845108695652174, "grad_norm": 0.946145236492157, "learning_rate": 3e-05, "loss": 2.971, "step": 3623 }, { "epoch": 0.9847826086956522, "grad_norm": 1.1449916362762451, "learning_rate": 3e-05, "loss": 3.3291, "step": 3624 }, { "epoch": 0.9850543478260869, "grad_norm": 1.2760465145111084, "learning_rate": 3e-05, "loss": 3.0303, "step": 3625 }, { "epoch": 0.9853260869565217, "grad_norm": 1.2135612964630127, "learning_rate": 3e-05, "loss": 3.2659, "step": 3626 }, { "epoch": 0.9855978260869566, "grad_norm": 1.0900663137435913, "learning_rate": 3e-05, "loss": 3.4042, "step": 3627 }, { "epoch": 0.9858695652173913, "grad_norm": 0.8186370730400085, "learning_rate": 3e-05, "loss": 2.5236, "step": 3628 }, { "epoch": 0.9861413043478261, "grad_norm": 0.8277904391288757, "learning_rate": 3e-05, "loss": 2.3149, "step": 3629 }, { "epoch": 0.9864130434782609, "grad_norm": 1.5521610975265503, "learning_rate": 3e-05, "loss": 3.3599, "step": 3630 }, { "epoch": 0.9866847826086956, "grad_norm": 1.1803765296936035, "learning_rate": 3e-05, "loss": 3.1166, "step": 3631 }, { "epoch": 0.9869565217391304, "grad_norm": 0.8749210238456726, "learning_rate": 3e-05, "loss": 2.9637, "step": 3632 }, { "epoch": 0.9872282608695652, "grad_norm": 1.1592975854873657, "learning_rate": 3e-05, "loss": 3.8733, "step": 3633 }, { "epoch": 0.9875, "grad_norm": 1.4104913473129272, "learning_rate": 3e-05, "loss": 3.6689, "step": 3634 }, { "epoch": 0.9877717391304348, "grad_norm": 1.266935110092163, "learning_rate": 3e-05, "loss": 3.3954, "step": 3635 }, { "epoch": 0.9880434782608696, "grad_norm": 1.2432810068130493, "learning_rate": 3e-05, "loss": 2.9512, "step": 3636 }, { "epoch": 0.9883152173913043, "grad_norm": 1.2164114713668823, "learning_rate": 3e-05, "loss": 3.3445, "step": 3637 }, { "epoch": 0.9885869565217391, "grad_norm": 1.2408303022384644, "learning_rate": 3e-05, "loss": 3.3845, "step": 3638 }, { "epoch": 0.9888586956521739, "grad_norm": 0.9395049810409546, "learning_rate": 3e-05, "loss": 2.9335, "step": 3639 }, { "epoch": 0.9891304347826086, "grad_norm": 1.044998049736023, "learning_rate": 3e-05, "loss": 3.0976, "step": 3640 }, { "epoch": 0.9894021739130435, "grad_norm": 1.1814433336257935, "learning_rate": 3e-05, "loss": 3.7362, "step": 3641 }, { "epoch": 0.9896739130434783, "grad_norm": 1.259177565574646, "learning_rate": 3e-05, "loss": 3.6791, "step": 3642 }, { "epoch": 0.9899456521739131, "grad_norm": 0.9600855112075806, "learning_rate": 3e-05, "loss": 3.3029, "step": 3643 }, { "epoch": 0.9902173913043478, "grad_norm": 0.9591723680496216, "learning_rate": 3e-05, "loss": 3.3596, "step": 3644 }, { "epoch": 0.9904891304347826, "grad_norm": 1.02616286277771, "learning_rate": 3e-05, "loss": 3.4194, "step": 3645 }, { "epoch": 0.9907608695652174, "grad_norm": 1.1400789022445679, "learning_rate": 3e-05, "loss": 3.7649, "step": 3646 }, { "epoch": 0.9910326086956521, "grad_norm": 1.3779942989349365, "learning_rate": 3e-05, "loss": 4.034, "step": 3647 }, { "epoch": 0.991304347826087, "grad_norm": 1.3959063291549683, "learning_rate": 3e-05, "loss": 3.4148, "step": 3648 }, { "epoch": 0.9915760869565218, "grad_norm": 1.1560637950897217, "learning_rate": 3e-05, "loss": 3.8032, "step": 3649 }, { "epoch": 0.9918478260869565, "grad_norm": 1.1359471082687378, "learning_rate": 3e-05, "loss": 3.1336, "step": 3650 }, { "epoch": 0.9921195652173913, "grad_norm": 0.963304340839386, "learning_rate": 3e-05, "loss": 2.8243, "step": 3651 }, { "epoch": 0.9923913043478261, "grad_norm": 1.6775833368301392, "learning_rate": 3e-05, "loss": 3.4228, "step": 3652 }, { "epoch": 0.9926630434782608, "grad_norm": 1.4887362718582153, "learning_rate": 3e-05, "loss": 3.0177, "step": 3653 }, { "epoch": 0.9929347826086956, "grad_norm": 1.307173728942871, "learning_rate": 3e-05, "loss": 3.0837, "step": 3654 }, { "epoch": 0.9932065217391305, "grad_norm": 1.2672486305236816, "learning_rate": 3e-05, "loss": 3.3747, "step": 3655 }, { "epoch": 0.9934782608695653, "grad_norm": 1.458817958831787, "learning_rate": 3e-05, "loss": 2.8296, "step": 3656 }, { "epoch": 0.99375, "grad_norm": 1.5649539232254028, "learning_rate": 3e-05, "loss": 3.0636, "step": 3657 }, { "epoch": 0.9940217391304348, "grad_norm": 1.1796952486038208, "learning_rate": 3e-05, "loss": 3.5806, "step": 3658 }, { "epoch": 0.9942934782608696, "grad_norm": 1.1408332586288452, "learning_rate": 3e-05, "loss": 3.22, "step": 3659 }, { "epoch": 0.9945652173913043, "grad_norm": 1.0065451860427856, "learning_rate": 3e-05, "loss": 2.9525, "step": 3660 }, { "epoch": 0.9948369565217391, "grad_norm": 1.1519659757614136, "learning_rate": 3e-05, "loss": 3.2977, "step": 3661 }, { "epoch": 0.9951086956521739, "grad_norm": 1.0569127798080444, "learning_rate": 3e-05, "loss": 3.5838, "step": 3662 }, { "epoch": 0.9953804347826087, "grad_norm": 0.8638179898262024, "learning_rate": 3e-05, "loss": 2.8316, "step": 3663 }, { "epoch": 0.9956521739130435, "grad_norm": 1.159942626953125, "learning_rate": 3e-05, "loss": 2.9437, "step": 3664 }, { "epoch": 0.9959239130434783, "grad_norm": 0.9920229315757751, "learning_rate": 3e-05, "loss": 2.8094, "step": 3665 }, { "epoch": 0.996195652173913, "grad_norm": 1.4593944549560547, "learning_rate": 3e-05, "loss": 3.7312, "step": 3666 }, { "epoch": 0.9964673913043478, "grad_norm": 1.1406296491622925, "learning_rate": 3e-05, "loss": 3.8133, "step": 3667 }, { "epoch": 0.9967391304347826, "grad_norm": 1.103284239768982, "learning_rate": 3e-05, "loss": 3.6508, "step": 3668 }, { "epoch": 0.9970108695652173, "grad_norm": 1.1054434776306152, "learning_rate": 3e-05, "loss": 3.812, "step": 3669 }, { "epoch": 0.9972826086956522, "grad_norm": 1.160575032234192, "learning_rate": 3e-05, "loss": 3.5056, "step": 3670 }, { "epoch": 0.997554347826087, "grad_norm": 1.1030219793319702, "learning_rate": 3e-05, "loss": 3.4479, "step": 3671 }, { "epoch": 0.9978260869565218, "grad_norm": 1.3701746463775635, "learning_rate": 3e-05, "loss": 3.3705, "step": 3672 }, { "epoch": 0.9980978260869565, "grad_norm": 1.4558731317520142, "learning_rate": 3e-05, "loss": 3.4222, "step": 3673 }, { "epoch": 0.9983695652173913, "grad_norm": 1.5558485984802246, "learning_rate": 3e-05, "loss": 3.9325, "step": 3674 }, { "epoch": 0.998641304347826, "grad_norm": 1.1101593971252441, "learning_rate": 3e-05, "loss": 3.4766, "step": 3675 }, { "epoch": 0.9989130434782608, "grad_norm": 1.319823145866394, "learning_rate": 3e-05, "loss": 3.454, "step": 3676 }, { "epoch": 0.9991847826086957, "grad_norm": 1.5775163173675537, "learning_rate": 3e-05, "loss": 3.7463, "step": 3677 }, { "epoch": 0.9994565217391305, "grad_norm": 0.9642970561981201, "learning_rate": 3e-05, "loss": 3.1748, "step": 3678 }, { "epoch": 0.9997282608695652, "grad_norm": 1.125762701034546, "learning_rate": 3e-05, "loss": 2.7312, "step": 3679 }, { "epoch": 1.0, "grad_norm": 1.213797688484192, "learning_rate": 3e-05, "loss": 2.8341, "step": 3680 }, { "epoch": 1.0, "step": 3680, "total_flos": 2.775664984892375e+18, "train_loss": 3.3962435043376424, "train_runtime": 29589.5294, "train_samples_per_second": 0.995, "train_steps_per_second": 0.124 } ], "logging_steps": 1, "max_steps": 3680, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.775664984892375e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }