diff --git "a/checkpoint-3000/trainer_state.json" "b/checkpoint-3000/trainer_state.json" new file mode 100644--- /dev/null +++ "b/checkpoint-3000/trainer_state.json" @@ -0,0 +1,21033 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.8152173913043478, + "eval_steps": 500, + "global_step": 3000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0002717391304347826, + "grad_norm": Infinity, + "learning_rate": 0.0, + "loss": 3.8119, + "step": 1 + }, + { + "epoch": 0.0005434782608695652, + "grad_norm": Infinity, + "learning_rate": 0.0, + "loss": 5.6312, + "step": 2 + }, + { + "epoch": 0.0008152173913043478, + "grad_norm": 3.606865644454956, + "learning_rate": 1.0000000000000001e-07, + "loss": 3.6085, + "step": 3 + }, + { + "epoch": 0.0010869565217391304, + "grad_norm": 5.723954677581787, + "learning_rate": 2.0000000000000002e-07, + "loss": 4.5497, + "step": 4 + }, + { + "epoch": 0.001358695652173913, + "grad_norm": 8.396666526794434, + "learning_rate": 3.0000000000000004e-07, + "loss": 5.6655, + "step": 5 + }, + { + "epoch": 0.0016304347826086956, + "grad_norm": 6.602017402648926, + "learning_rate": 4.0000000000000003e-07, + "loss": 4.9113, + "step": 6 + }, + { + "epoch": 0.0019021739130434783, + "grad_norm": 2.8019049167633057, + "learning_rate": 5e-07, + "loss": 3.0843, + "step": 7 + }, + { + "epoch": 0.002173913043478261, + "grad_norm": 5.959436893463135, + "learning_rate": 6.000000000000001e-07, + "loss": 4.4235, + "step": 8 + }, + { + "epoch": 0.0024456521739130437, + "grad_norm": 5.623807430267334, + "learning_rate": 7.000000000000001e-07, + "loss": 4.5667, + "step": 9 + }, + { + "epoch": 0.002717391304347826, + "grad_norm": 5.9944844245910645, + "learning_rate": 8.000000000000001e-07, + "loss": 4.6548, + "step": 10 + }, + { + "epoch": 0.002989130434782609, + "grad_norm": 7.308701038360596, + "learning_rate": 9e-07, + "loss": 5.1422, + "step": 11 + }, + { + "epoch": 0.003260869565217391, + "grad_norm": 5.805826187133789, + "learning_rate": 1e-06, + "loss": 4.3458, + "step": 12 + }, + { + "epoch": 0.003532608695652174, + "grad_norm": 5.096138000488281, + "learning_rate": 1.1e-06, + "loss": 4.0761, + "step": 13 + }, + { + "epoch": 0.0038043478260869567, + "grad_norm": 4.222947120666504, + "learning_rate": 1.2000000000000002e-06, + "loss": 3.7963, + "step": 14 + }, + { + "epoch": 0.004076086956521739, + "grad_norm": 8.304177284240723, + "learning_rate": 1.3e-06, + "loss": 5.4697, + "step": 15 + }, + { + "epoch": 0.004347826086956522, + "grad_norm": 6.131305694580078, + "learning_rate": 1.4000000000000001e-06, + "loss": 4.5078, + "step": 16 + }, + { + "epoch": 0.004619565217391305, + "grad_norm": 5.1205315589904785, + "learning_rate": 1.5e-06, + "loss": 4.1935, + "step": 17 + }, + { + "epoch": 0.004891304347826087, + "grad_norm": 3.348874568939209, + "learning_rate": 1.6000000000000001e-06, + "loss": 3.165, + "step": 18 + }, + { + "epoch": 0.005163043478260869, + "grad_norm": 11.620071411132812, + "learning_rate": 1.7e-06, + "loss": 6.6692, + "step": 19 + }, + { + "epoch": 0.005434782608695652, + "grad_norm": 5.148247241973877, + "learning_rate": 1.8e-06, + "loss": 3.7967, + "step": 20 + }, + { + "epoch": 0.005706521739130435, + "grad_norm": 6.738089561462402, + "learning_rate": 1.9000000000000002e-06, + "loss": 4.6053, + "step": 21 + }, + { + "epoch": 0.005978260869565218, + "grad_norm": 7.383264541625977, + "learning_rate": 2e-06, + "loss": 4.7862, + "step": 22 + }, + { + "epoch": 0.00625, + "grad_norm": 9.380542755126953, + "learning_rate": 2.1000000000000002e-06, + "loss": 5.5961, + "step": 23 + }, + { + "epoch": 0.006521739130434782, + "grad_norm": 5.481351852416992, + "learning_rate": 2.2e-06, + "loss": 3.9489, + "step": 24 + }, + { + "epoch": 0.006793478260869565, + "grad_norm": 3.769026756286621, + "learning_rate": 2.3e-06, + "loss": 3.3576, + "step": 25 + }, + { + "epoch": 0.007065217391304348, + "grad_norm": 6.088125705718994, + "learning_rate": 2.4000000000000003e-06, + "loss": 4.3102, + "step": 26 + }, + { + "epoch": 0.007336956521739131, + "grad_norm": 3.7228612899780273, + "learning_rate": 2.4999999999999998e-06, + "loss": 3.2898, + "step": 27 + }, + { + "epoch": 0.007608695652173913, + "grad_norm": 5.163967609405518, + "learning_rate": 2.6e-06, + "loss": 3.9394, + "step": 28 + }, + { + "epoch": 0.007880434782608695, + "grad_norm": 3.577744960784912, + "learning_rate": 2.7e-06, + "loss": 3.4044, + "step": 29 + }, + { + "epoch": 0.008152173913043478, + "grad_norm": 5.421440124511719, + "learning_rate": 2.8000000000000003e-06, + "loss": 4.2156, + "step": 30 + }, + { + "epoch": 0.00842391304347826, + "grad_norm": 4.284877777099609, + "learning_rate": 2.9e-06, + "loss": 3.519, + "step": 31 + }, + { + "epoch": 0.008695652173913044, + "grad_norm": 6.597541809082031, + "learning_rate": 3e-06, + "loss": 4.5028, + "step": 32 + }, + { + "epoch": 0.008967391304347826, + "grad_norm": 7.851593494415283, + "learning_rate": 3.1e-06, + "loss": 5.1247, + "step": 33 + }, + { + "epoch": 0.00923913043478261, + "grad_norm": 2.644531726837158, + "learning_rate": 3.2000000000000003e-06, + "loss": 2.9878, + "step": 34 + }, + { + "epoch": 0.009510869565217392, + "grad_norm": 3.316633462905884, + "learning_rate": 3.3e-06, + "loss": 3.4112, + "step": 35 + }, + { + "epoch": 0.009782608695652175, + "grad_norm": 3.0373375415802, + "learning_rate": 3.4e-06, + "loss": 3.301, + "step": 36 + }, + { + "epoch": 0.010054347826086956, + "grad_norm": 5.997847080230713, + "learning_rate": 3.5e-06, + "loss": 4.9763, + "step": 37 + }, + { + "epoch": 0.010326086956521738, + "grad_norm": 5.2729268074035645, + "learning_rate": 3.6e-06, + "loss": 4.7279, + "step": 38 + }, + { + "epoch": 0.010597826086956521, + "grad_norm": 2.7055575847625732, + "learning_rate": 3.7e-06, + "loss": 3.4656, + "step": 39 + }, + { + "epoch": 0.010869565217391304, + "grad_norm": 3.5352964401245117, + "learning_rate": 3.8000000000000005e-06, + "loss": 3.9642, + "step": 40 + }, + { + "epoch": 0.011141304347826087, + "grad_norm": 2.8491668701171875, + "learning_rate": 3.9e-06, + "loss": 3.8092, + "step": 41 + }, + { + "epoch": 0.01141304347826087, + "grad_norm": 1.8910856246948242, + "learning_rate": 4e-06, + "loss": 3.0861, + "step": 42 + }, + { + "epoch": 0.011684782608695652, + "grad_norm": 2.873784303665161, + "learning_rate": 4.1e-06, + "loss": 3.788, + "step": 43 + }, + { + "epoch": 0.011956521739130435, + "grad_norm": 3.121163845062256, + "learning_rate": 4.2000000000000004e-06, + "loss": 4.0837, + "step": 44 + }, + { + "epoch": 0.012228260869565218, + "grad_norm": 2.977182388305664, + "learning_rate": 4.3e-06, + "loss": 4.0382, + "step": 45 + }, + { + "epoch": 0.0125, + "grad_norm": 3.935685396194458, + "learning_rate": 4.4e-06, + "loss": 5.0603, + "step": 46 + }, + { + "epoch": 0.012771739130434783, + "grad_norm": 1.5644341707229614, + "learning_rate": 4.5e-06, + "loss": 3.0085, + "step": 47 + }, + { + "epoch": 0.013043478260869565, + "grad_norm": 2.2639412879943848, + "learning_rate": 4.6e-06, + "loss": 3.5606, + "step": 48 + }, + { + "epoch": 0.013315217391304347, + "grad_norm": 3.8248438835144043, + "learning_rate": 4.700000000000001e-06, + "loss": 5.1055, + "step": 49 + }, + { + "epoch": 0.01358695652173913, + "grad_norm": 2.0189621448516846, + "learning_rate": 4.800000000000001e-06, + "loss": 3.6357, + "step": 50 + }, + { + "epoch": 0.013858695652173913, + "grad_norm": 1.7053027153015137, + "learning_rate": 4.9e-06, + "loss": 3.3706, + "step": 51 + }, + { + "epoch": 0.014130434782608696, + "grad_norm": 2.996800184249878, + "learning_rate": 4.9999999999999996e-06, + "loss": 4.2643, + "step": 52 + }, + { + "epoch": 0.014402173913043478, + "grad_norm": 2.2045717239379883, + "learning_rate": 5.1e-06, + "loss": 3.4676, + "step": 53 + }, + { + "epoch": 0.014673913043478261, + "grad_norm": 1.9880719184875488, + "learning_rate": 5.2e-06, + "loss": 3.7901, + "step": 54 + }, + { + "epoch": 0.014945652173913044, + "grad_norm": 1.4993855953216553, + "learning_rate": 5.3e-06, + "loss": 3.3175, + "step": 55 + }, + { + "epoch": 0.015217391304347827, + "grad_norm": 1.3958892822265625, + "learning_rate": 5.4e-06, + "loss": 3.0166, + "step": 56 + }, + { + "epoch": 0.01548913043478261, + "grad_norm": 1.5511572360992432, + "learning_rate": 5.5e-06, + "loss": 3.4503, + "step": 57 + }, + { + "epoch": 0.01576086956521739, + "grad_norm": 2.274249792098999, + "learning_rate": 5.600000000000001e-06, + "loss": 4.1636, + "step": 58 + }, + { + "epoch": 0.016032608695652175, + "grad_norm": 1.7447724342346191, + "learning_rate": 5.7000000000000005e-06, + "loss": 3.7056, + "step": 59 + }, + { + "epoch": 0.016304347826086956, + "grad_norm": 2.4315755367279053, + "learning_rate": 5.8e-06, + "loss": 3.9477, + "step": 60 + }, + { + "epoch": 0.01657608695652174, + "grad_norm": 1.5694116353988647, + "learning_rate": 5.899999999999999e-06, + "loss": 3.321, + "step": 61 + }, + { + "epoch": 0.01684782608695652, + "grad_norm": 1.856467366218567, + "learning_rate": 6e-06, + "loss": 3.8384, + "step": 62 + }, + { + "epoch": 0.017119565217391303, + "grad_norm": 1.3913215398788452, + "learning_rate": 6.1e-06, + "loss": 3.1717, + "step": 63 + }, + { + "epoch": 0.017391304347826087, + "grad_norm": 1.1280888319015503, + "learning_rate": 6.2e-06, + "loss": 3.0487, + "step": 64 + }, + { + "epoch": 0.017663043478260868, + "grad_norm": 1.3553597927093506, + "learning_rate": 6.3e-06, + "loss": 3.4274, + "step": 65 + }, + { + "epoch": 0.017934782608695653, + "grad_norm": 1.5680698156356812, + "learning_rate": 6.4000000000000006e-06, + "loss": 3.2402, + "step": 66 + }, + { + "epoch": 0.018206521739130434, + "grad_norm": 1.2775722742080688, + "learning_rate": 6.5000000000000004e-06, + "loss": 3.1252, + "step": 67 + }, + { + "epoch": 0.01847826086956522, + "grad_norm": 1.547057867050171, + "learning_rate": 6.6e-06, + "loss": 3.6654, + "step": 68 + }, + { + "epoch": 0.01875, + "grad_norm": 1.6398184299468994, + "learning_rate": 6.7e-06, + "loss": 3.4637, + "step": 69 + }, + { + "epoch": 0.019021739130434784, + "grad_norm": 0.8923302292823792, + "learning_rate": 6.8e-06, + "loss": 2.6619, + "step": 70 + }, + { + "epoch": 0.019293478260869565, + "grad_norm": 1.1862719058990479, + "learning_rate": 6.900000000000001e-06, + "loss": 3.1319, + "step": 71 + }, + { + "epoch": 0.01956521739130435, + "grad_norm": 1.0350522994995117, + "learning_rate": 7e-06, + "loss": 2.8874, + "step": 72 + }, + { + "epoch": 0.01983695652173913, + "grad_norm": 1.7756403684616089, + "learning_rate": 7.1e-06, + "loss": 3.7356, + "step": 73 + }, + { + "epoch": 0.02010869565217391, + "grad_norm": 1.2788705825805664, + "learning_rate": 7.2e-06, + "loss": 3.2409, + "step": 74 + }, + { + "epoch": 0.020380434782608696, + "grad_norm": 1.4206314086914062, + "learning_rate": 7.3e-06, + "loss": 3.5884, + "step": 75 + }, + { + "epoch": 0.020652173913043477, + "grad_norm": 2.0677270889282227, + "learning_rate": 7.4e-06, + "loss": 4.3218, + "step": 76 + }, + { + "epoch": 0.02092391304347826, + "grad_norm": 0.700586199760437, + "learning_rate": 7.5e-06, + "loss": 2.4596, + "step": 77 + }, + { + "epoch": 0.021195652173913043, + "grad_norm": 2.1819796562194824, + "learning_rate": 7.600000000000001e-06, + "loss": 3.9885, + "step": 78 + }, + { + "epoch": 0.021467391304347827, + "grad_norm": 1.0155264139175415, + "learning_rate": 7.699999999999999e-06, + "loss": 2.7548, + "step": 79 + }, + { + "epoch": 0.021739130434782608, + "grad_norm": 1.2022877931594849, + "learning_rate": 7.8e-06, + "loss": 3.1796, + "step": 80 + }, + { + "epoch": 0.022010869565217393, + "grad_norm": 1.1962783336639404, + "learning_rate": 7.899999999999999e-06, + "loss": 2.907, + "step": 81 + }, + { + "epoch": 0.022282608695652174, + "grad_norm": 1.0273329019546509, + "learning_rate": 8e-06, + "loss": 2.7207, + "step": 82 + }, + { + "epoch": 0.022554347826086958, + "grad_norm": 2.3930468559265137, + "learning_rate": 8.1e-06, + "loss": 3.9644, + "step": 83 + }, + { + "epoch": 0.02282608695652174, + "grad_norm": 2.244040012359619, + "learning_rate": 8.2e-06, + "loss": 3.5003, + "step": 84 + }, + { + "epoch": 0.02309782608695652, + "grad_norm": 1.685024380683899, + "learning_rate": 8.3e-06, + "loss": 3.0757, + "step": 85 + }, + { + "epoch": 0.023369565217391305, + "grad_norm": 3.1130995750427246, + "learning_rate": 8.400000000000001e-06, + "loss": 3.9588, + "step": 86 + }, + { + "epoch": 0.023641304347826086, + "grad_norm": 2.6688392162323, + "learning_rate": 8.5e-06, + "loss": 3.6427, + "step": 87 + }, + { + "epoch": 0.02391304347826087, + "grad_norm": 1.3578976392745972, + "learning_rate": 8.6e-06, + "loss": 2.9824, + "step": 88 + }, + { + "epoch": 0.02418478260869565, + "grad_norm": 1.2437134981155396, + "learning_rate": 8.7e-06, + "loss": 3.1183, + "step": 89 + }, + { + "epoch": 0.024456521739130436, + "grad_norm": 2.143594741821289, + "learning_rate": 8.8e-06, + "loss": 3.6833, + "step": 90 + }, + { + "epoch": 0.024728260869565217, + "grad_norm": 1.2158911228179932, + "learning_rate": 8.900000000000001e-06, + "loss": 2.7814, + "step": 91 + }, + { + "epoch": 0.025, + "grad_norm": 2.4209673404693604, + "learning_rate": 9e-06, + "loss": 4.3185, + "step": 92 + }, + { + "epoch": 0.025271739130434782, + "grad_norm": 3.090715169906616, + "learning_rate": 9.100000000000001e-06, + "loss": 4.0217, + "step": 93 + }, + { + "epoch": 0.025543478260869567, + "grad_norm": 3.1287500858306885, + "learning_rate": 9.2e-06, + "loss": 3.8665, + "step": 94 + }, + { + "epoch": 0.025815217391304348, + "grad_norm": 2.6370797157287598, + "learning_rate": 9.3e-06, + "loss": 3.7443, + "step": 95 + }, + { + "epoch": 0.02608695652173913, + "grad_norm": 4.120901107788086, + "learning_rate": 9.400000000000001e-06, + "loss": 4.3482, + "step": 96 + }, + { + "epoch": 0.026358695652173914, + "grad_norm": NaN, + "learning_rate": 9.400000000000001e-06, + "loss": 3.7633, + "step": 97 + }, + { + "epoch": 0.026630434782608695, + "grad_norm": 2.1682493686676025, + "learning_rate": 9.5e-06, + "loss": 3.135, + "step": 98 + }, + { + "epoch": 0.02690217391304348, + "grad_norm": 9.943031311035156, + "learning_rate": 9.600000000000001e-06, + "loss": 4.4992, + "step": 99 + }, + { + "epoch": 0.02717391304347826, + "grad_norm": 9.851699829101562, + "learning_rate": 9.699999999999999e-06, + "loss": 3.3346, + "step": 100 + }, + { + "epoch": 0.027445652173913045, + "grad_norm": 9.068572044372559, + "learning_rate": 9.8e-06, + "loss": 3.4352, + "step": 101 + }, + { + "epoch": 0.027717391304347826, + "grad_norm": 16.46036148071289, + "learning_rate": 9.9e-06, + "loss": 4.2803, + "step": 102 + }, + { + "epoch": 0.02798913043478261, + "grad_norm": 4.916048049926758, + "learning_rate": 9.999999999999999e-06, + "loss": 5.2283, + "step": 103 + }, + { + "epoch": 0.02826086956521739, + "grad_norm": 3.9112675189971924, + "learning_rate": 1.01e-05, + "loss": 3.9085, + "step": 104 + }, + { + "epoch": 0.028532608695652172, + "grad_norm": 2.9353702068328857, + "learning_rate": 1.02e-05, + "loss": 3.436, + "step": 105 + }, + { + "epoch": 0.028804347826086957, + "grad_norm": 2.1756107807159424, + "learning_rate": 1.03e-05, + "loss": 4.2118, + "step": 106 + }, + { + "epoch": 0.029076086956521738, + "grad_norm": 2.4335989952087402, + "learning_rate": 1.04e-05, + "loss": 4.3144, + "step": 107 + }, + { + "epoch": 0.029347826086956522, + "grad_norm": 2.6821694374084473, + "learning_rate": 1.05e-05, + "loss": 4.461, + "step": 108 + }, + { + "epoch": 0.029619565217391303, + "grad_norm": 1.9325740337371826, + "learning_rate": 1.06e-05, + "loss": 4.1079, + "step": 109 + }, + { + "epoch": 0.029891304347826088, + "grad_norm": 3.598806381225586, + "learning_rate": 1.0700000000000001e-05, + "loss": 3.6908, + "step": 110 + }, + { + "epoch": 0.03016304347826087, + "grad_norm": 1.7592302560806274, + "learning_rate": 1.08e-05, + "loss": 3.2482, + "step": 111 + }, + { + "epoch": 0.030434782608695653, + "grad_norm": 2.4772744178771973, + "learning_rate": 1.09e-05, + "loss": 4.3941, + "step": 112 + }, + { + "epoch": 0.030706521739130434, + "grad_norm": 1.8405948877334595, + "learning_rate": 1.1e-05, + "loss": 3.6201, + "step": 113 + }, + { + "epoch": 0.03097826086956522, + "grad_norm": 4.303748607635498, + "learning_rate": 1.11e-05, + "loss": 4.6995, + "step": 114 + }, + { + "epoch": 0.03125, + "grad_norm": 2.199484348297119, + "learning_rate": 1.1200000000000001e-05, + "loss": 3.7969, + "step": 115 + }, + { + "epoch": 0.03152173913043478, + "grad_norm": 1.0349135398864746, + "learning_rate": 1.13e-05, + "loss": 2.8461, + "step": 116 + }, + { + "epoch": 0.03179347826086956, + "grad_norm": 2.374260425567627, + "learning_rate": 1.1400000000000001e-05, + "loss": 4.2862, + "step": 117 + }, + { + "epoch": 0.03206521739130435, + "grad_norm": 1.283584713935852, + "learning_rate": 1.1500000000000002e-05, + "loss": 3.1288, + "step": 118 + }, + { + "epoch": 0.03233695652173913, + "grad_norm": 3.549565076828003, + "learning_rate": 1.16e-05, + "loss": 4.1137, + "step": 119 + }, + { + "epoch": 0.03260869565217391, + "grad_norm": 1.3312267065048218, + "learning_rate": 1.1700000000000001e-05, + "loss": 3.6003, + "step": 120 + }, + { + "epoch": 0.03288043478260869, + "grad_norm": 1.655480146408081, + "learning_rate": 1.1799999999999999e-05, + "loss": 3.7798, + "step": 121 + }, + { + "epoch": 0.03315217391304348, + "grad_norm": 1.6580840349197388, + "learning_rate": 1.19e-05, + "loss": 3.5654, + "step": 122 + }, + { + "epoch": 0.03342391304347826, + "grad_norm": 2.8110294342041016, + "learning_rate": 1.2e-05, + "loss": 3.9673, + "step": 123 + }, + { + "epoch": 0.03369565217391304, + "grad_norm": 1.2965394258499146, + "learning_rate": 1.21e-05, + "loss": 3.26, + "step": 124 + }, + { + "epoch": 0.033967391304347824, + "grad_norm": 1.914302945137024, + "learning_rate": 1.22e-05, + "loss": 3.95, + "step": 125 + }, + { + "epoch": 0.034239130434782605, + "grad_norm": 1.3706483840942383, + "learning_rate": 1.2299999999999999e-05, + "loss": 3.5317, + "step": 126 + }, + { + "epoch": 0.03451086956521739, + "grad_norm": 2.016113042831421, + "learning_rate": 1.24e-05, + "loss": 3.2582, + "step": 127 + }, + { + "epoch": 0.034782608695652174, + "grad_norm": 2.5129711627960205, + "learning_rate": 1.25e-05, + "loss": 3.3608, + "step": 128 + }, + { + "epoch": 0.035054347826086955, + "grad_norm": Infinity, + "learning_rate": 1.25e-05, + "loss": 4.3213, + "step": 129 + }, + { + "epoch": 0.035326086956521736, + "grad_norm": 3.073050022125244, + "learning_rate": 1.26e-05, + "loss": 4.4668, + "step": 130 + }, + { + "epoch": 0.035597826086956524, + "grad_norm": 1.8033742904663086, + "learning_rate": 1.27e-05, + "loss": 3.5871, + "step": 131 + }, + { + "epoch": 0.035869565217391305, + "grad_norm": 3.148634195327759, + "learning_rate": 1.2800000000000001e-05, + "loss": 3.7611, + "step": 132 + }, + { + "epoch": 0.036141304347826086, + "grad_norm": 2.898167610168457, + "learning_rate": 1.29e-05, + "loss": 3.6928, + "step": 133 + }, + { + "epoch": 0.03641304347826087, + "grad_norm": 1.9879661798477173, + "learning_rate": 1.3000000000000001e-05, + "loss": 3.607, + "step": 134 + }, + { + "epoch": 0.036684782608695655, + "grad_norm": 2.0479657649993896, + "learning_rate": 1.31e-05, + "loss": 3.2456, + "step": 135 + }, + { + "epoch": 0.03695652173913044, + "grad_norm": 1.5060527324676514, + "learning_rate": 1.32e-05, + "loss": 3.4124, + "step": 136 + }, + { + "epoch": 0.03722826086956522, + "grad_norm": 2.2012810707092285, + "learning_rate": 1.3300000000000001e-05, + "loss": 4.1008, + "step": 137 + }, + { + "epoch": 0.0375, + "grad_norm": 2.2929439544677734, + "learning_rate": 1.34e-05, + "loss": 4.5483, + "step": 138 + }, + { + "epoch": 0.03777173913043478, + "grad_norm": 1.777136206626892, + "learning_rate": 1.3500000000000001e-05, + "loss": 3.9796, + "step": 139 + }, + { + "epoch": 0.03804347826086957, + "grad_norm": 3.1867823600769043, + "learning_rate": 1.36e-05, + "loss": 4.13, + "step": 140 + }, + { + "epoch": 0.03831521739130435, + "grad_norm": 2.070611000061035, + "learning_rate": 1.3700000000000001e-05, + "loss": 3.3474, + "step": 141 + }, + { + "epoch": 0.03858695652173913, + "grad_norm": 1.9961167573928833, + "learning_rate": 1.3800000000000002e-05, + "loss": 4.0889, + "step": 142 + }, + { + "epoch": 0.03885869565217391, + "grad_norm": 1.709995985031128, + "learning_rate": 1.39e-05, + "loss": 3.9576, + "step": 143 + }, + { + "epoch": 0.0391304347826087, + "grad_norm": 1.793494462966919, + "learning_rate": 1.4e-05, + "loss": 4.419, + "step": 144 + }, + { + "epoch": 0.03940217391304348, + "grad_norm": 2.1228878498077393, + "learning_rate": 1.4099999999999999e-05, + "loss": 3.7677, + "step": 145 + }, + { + "epoch": 0.03967391304347826, + "grad_norm": 1.5662310123443604, + "learning_rate": 1.42e-05, + "loss": 3.5038, + "step": 146 + }, + { + "epoch": 0.03994565217391304, + "grad_norm": 1.1859204769134521, + "learning_rate": 1.43e-05, + "loss": 3.0831, + "step": 147 + }, + { + "epoch": 0.04021739130434782, + "grad_norm": 1.998694896697998, + "learning_rate": 1.44e-05, + "loss": 4.06, + "step": 148 + }, + { + "epoch": 0.04048913043478261, + "grad_norm": 2.2709462642669678, + "learning_rate": 1.45e-05, + "loss": 3.4234, + "step": 149 + }, + { + "epoch": 0.04076086956521739, + "grad_norm": 3.572169065475464, + "learning_rate": 1.46e-05, + "loss": 4.1884, + "step": 150 + }, + { + "epoch": 0.04103260869565217, + "grad_norm": 2.4503207206726074, + "learning_rate": 1.47e-05, + "loss": 4.0506, + "step": 151 + }, + { + "epoch": 0.041304347826086954, + "grad_norm": 2.66367244720459, + "learning_rate": 1.48e-05, + "loss": 4.6858, + "step": 152 + }, + { + "epoch": 0.04157608695652174, + "grad_norm": 2.1236956119537354, + "learning_rate": 1.49e-05, + "loss": 3.2571, + "step": 153 + }, + { + "epoch": 0.04184782608695652, + "grad_norm": 4.9067206382751465, + "learning_rate": 1.5e-05, + "loss": 5.32, + "step": 154 + }, + { + "epoch": 0.042119565217391304, + "grad_norm": 2.0733673572540283, + "learning_rate": 1.51e-05, + "loss": 3.7548, + "step": 155 + }, + { + "epoch": 0.042391304347826085, + "grad_norm": 1.186569333076477, + "learning_rate": 1.5200000000000002e-05, + "loss": 3.3953, + "step": 156 + }, + { + "epoch": 0.042663043478260866, + "grad_norm": 2.5303616523742676, + "learning_rate": 1.53e-05, + "loss": 4.0742, + "step": 157 + }, + { + "epoch": 0.042934782608695654, + "grad_norm": 2.3837692737579346, + "learning_rate": 1.5399999999999998e-05, + "loss": 4.3684, + "step": 158 + }, + { + "epoch": 0.043206521739130435, + "grad_norm": 2.3704850673675537, + "learning_rate": 1.55e-05, + "loss": 4.1658, + "step": 159 + }, + { + "epoch": 0.043478260869565216, + "grad_norm": 3.8034274578094482, + "learning_rate": 1.56e-05, + "loss": 4.0919, + "step": 160 + }, + { + "epoch": 0.04375, + "grad_norm": 1.2763932943344116, + "learning_rate": 1.57e-05, + "loss": 3.8752, + "step": 161 + }, + { + "epoch": 0.044021739130434785, + "grad_norm": 1.7155462503433228, + "learning_rate": 1.5799999999999998e-05, + "loss": 4.317, + "step": 162 + }, + { + "epoch": 0.044293478260869566, + "grad_norm": 2.481580972671509, + "learning_rate": 1.59e-05, + "loss": 4.1314, + "step": 163 + }, + { + "epoch": 0.04456521739130435, + "grad_norm": 2.3853437900543213, + "learning_rate": 1.6e-05, + "loss": 3.6408, + "step": 164 + }, + { + "epoch": 0.04483695652173913, + "grad_norm": 1.7092500925064087, + "learning_rate": 1.61e-05, + "loss": 3.833, + "step": 165 + }, + { + "epoch": 0.045108695652173916, + "grad_norm": 2.1334292888641357, + "learning_rate": 1.62e-05, + "loss": 4.5951, + "step": 166 + }, + { + "epoch": 0.0453804347826087, + "grad_norm": 1.352221965789795, + "learning_rate": 1.63e-05, + "loss": 3.804, + "step": 167 + }, + { + "epoch": 0.04565217391304348, + "grad_norm": 2.178356409072876, + "learning_rate": 1.64e-05, + "loss": 3.7594, + "step": 168 + }, + { + "epoch": 0.04592391304347826, + "grad_norm": 2.9925286769866943, + "learning_rate": 1.65e-05, + "loss": 3.8819, + "step": 169 + }, + { + "epoch": 0.04619565217391304, + "grad_norm": 3.0277819633483887, + "learning_rate": 1.66e-05, + "loss": 4.4896, + "step": 170 + }, + { + "epoch": 0.04646739130434783, + "grad_norm": 1.459825873374939, + "learning_rate": 1.67e-05, + "loss": 4.1701, + "step": 171 + }, + { + "epoch": 0.04673913043478261, + "grad_norm": 1.6663670539855957, + "learning_rate": 1.6800000000000002e-05, + "loss": 3.6651, + "step": 172 + }, + { + "epoch": 0.04701086956521739, + "grad_norm": 1.7755922079086304, + "learning_rate": 1.69e-05, + "loss": 3.3178, + "step": 173 + }, + { + "epoch": 0.04728260869565217, + "grad_norm": 1.4884556531906128, + "learning_rate": 1.7e-05, + "loss": 3.6467, + "step": 174 + }, + { + "epoch": 0.04755434782608696, + "grad_norm": 2.701298713684082, + "learning_rate": 1.71e-05, + "loss": 4.773, + "step": 175 + }, + { + "epoch": 0.04782608695652174, + "grad_norm": 3.169405937194824, + "learning_rate": 1.72e-05, + "loss": 4.1844, + "step": 176 + }, + { + "epoch": 0.04809782608695652, + "grad_norm": 2.4385275840759277, + "learning_rate": 1.73e-05, + "loss": 3.8691, + "step": 177 + }, + { + "epoch": 0.0483695652173913, + "grad_norm": 3.446741819381714, + "learning_rate": 1.74e-05, + "loss": 3.6733, + "step": 178 + }, + { + "epoch": 0.048641304347826084, + "grad_norm": 4.095767974853516, + "learning_rate": 1.7500000000000002e-05, + "loss": 3.927, + "step": 179 + }, + { + "epoch": 0.04891304347826087, + "grad_norm": 3.42309308052063, + "learning_rate": 1.76e-05, + "loss": 4.152, + "step": 180 + }, + { + "epoch": 0.04918478260869565, + "grad_norm": 1.8505162000656128, + "learning_rate": 1.77e-05, + "loss": 3.5489, + "step": 181 + }, + { + "epoch": 0.049456521739130434, + "grad_norm": 3.3934953212738037, + "learning_rate": 1.7800000000000002e-05, + "loss": 3.9161, + "step": 182 + }, + { + "epoch": 0.049728260869565215, + "grad_norm": 3.2479214668273926, + "learning_rate": 1.79e-05, + "loss": 3.3177, + "step": 183 + }, + { + "epoch": 0.05, + "grad_norm": 4.14975643157959, + "learning_rate": 1.8e-05, + "loss": 4.0072, + "step": 184 + }, + { + "epoch": 0.050271739130434784, + "grad_norm": 4.540154457092285, + "learning_rate": 1.8100000000000003e-05, + "loss": 3.878, + "step": 185 + }, + { + "epoch": 0.050543478260869565, + "grad_norm": 2.0964109897613525, + "learning_rate": 1.8200000000000002e-05, + "loss": 3.6273, + "step": 186 + }, + { + "epoch": 0.050815217391304346, + "grad_norm": 2.119013547897339, + "learning_rate": 1.83e-05, + "loss": 3.5527, + "step": 187 + }, + { + "epoch": 0.051086956521739134, + "grad_norm": 4.674685478210449, + "learning_rate": 1.84e-05, + "loss": 4.1131, + "step": 188 + }, + { + "epoch": 0.051358695652173915, + "grad_norm": 7.127579689025879, + "learning_rate": 1.8500000000000002e-05, + "loss": 4.0545, + "step": 189 + }, + { + "epoch": 0.051630434782608696, + "grad_norm": 3.285233497619629, + "learning_rate": 1.86e-05, + "loss": 3.6065, + "step": 190 + }, + { + "epoch": 0.05190217391304348, + "grad_norm": 1.405059576034546, + "learning_rate": 1.87e-05, + "loss": 3.4746, + "step": 191 + }, + { + "epoch": 0.05217391304347826, + "grad_norm": 3.002393960952759, + "learning_rate": 1.8800000000000003e-05, + "loss": 3.9168, + "step": 192 + }, + { + "epoch": 0.052445652173913046, + "grad_norm": 2.253774404525757, + "learning_rate": 1.8900000000000002e-05, + "loss": 3.0605, + "step": 193 + }, + { + "epoch": 0.05271739130434783, + "grad_norm": 3.4486026763916016, + "learning_rate": 1.9e-05, + "loss": 3.6955, + "step": 194 + }, + { + "epoch": 0.05298913043478261, + "grad_norm": 2.6076107025146484, + "learning_rate": 1.9100000000000003e-05, + "loss": 4.7168, + "step": 195 + }, + { + "epoch": 0.05326086956521739, + "grad_norm": 2.173379421234131, + "learning_rate": 1.9200000000000003e-05, + "loss": 3.9563, + "step": 196 + }, + { + "epoch": 0.05353260869565218, + "grad_norm": 1.8571799993515015, + "learning_rate": 1.9299999999999998e-05, + "loss": 3.531, + "step": 197 + }, + { + "epoch": 0.05380434782608696, + "grad_norm": 2.940171957015991, + "learning_rate": 1.9399999999999997e-05, + "loss": 3.9534, + "step": 198 + }, + { + "epoch": 0.05407608695652174, + "grad_norm": 1.6403425931930542, + "learning_rate": 1.95e-05, + "loss": 3.7979, + "step": 199 + }, + { + "epoch": 0.05434782608695652, + "grad_norm": 1.9086182117462158, + "learning_rate": 1.96e-05, + "loss": 3.3714, + "step": 200 + }, + { + "epoch": 0.0546195652173913, + "grad_norm": 2.42441463470459, + "learning_rate": 1.9699999999999998e-05, + "loss": 3.6953, + "step": 201 + }, + { + "epoch": 0.05489130434782609, + "grad_norm": 1.7879725694656372, + "learning_rate": 1.98e-05, + "loss": 3.424, + "step": 202 + }, + { + "epoch": 0.05516304347826087, + "grad_norm": 1.951125144958496, + "learning_rate": 1.99e-05, + "loss": 4.0087, + "step": 203 + }, + { + "epoch": 0.05543478260869565, + "grad_norm": 1.8883895874023438, + "learning_rate": 1.9999999999999998e-05, + "loss": 3.4857, + "step": 204 + }, + { + "epoch": 0.05570652173913043, + "grad_norm": 2.1979916095733643, + "learning_rate": 2.01e-05, + "loss": 3.8898, + "step": 205 + }, + { + "epoch": 0.05597826086956522, + "grad_norm": 1.6748876571655273, + "learning_rate": 2.02e-05, + "loss": 3.5928, + "step": 206 + }, + { + "epoch": 0.05625, + "grad_norm": 4.129452705383301, + "learning_rate": 2.03e-05, + "loss": 4.2804, + "step": 207 + }, + { + "epoch": 0.05652173913043478, + "grad_norm": 2.1873629093170166, + "learning_rate": 2.04e-05, + "loss": 3.5242, + "step": 208 + }, + { + "epoch": 0.05679347826086956, + "grad_norm": 3.064663887023926, + "learning_rate": 2.05e-05, + "loss": 3.379, + "step": 209 + }, + { + "epoch": 0.057065217391304345, + "grad_norm": 2.8077127933502197, + "learning_rate": 2.06e-05, + "loss": 3.4032, + "step": 210 + }, + { + "epoch": 0.05733695652173913, + "grad_norm": 2.902247905731201, + "learning_rate": 2.07e-05, + "loss": 3.9496, + "step": 211 + }, + { + "epoch": 0.057608695652173914, + "grad_norm": 2.3877575397491455, + "learning_rate": 2.08e-05, + "loss": 3.5266, + "step": 212 + }, + { + "epoch": 0.057880434782608695, + "grad_norm": 4.157260417938232, + "learning_rate": 2.09e-05, + "loss": 4.0795, + "step": 213 + }, + { + "epoch": 0.058152173913043476, + "grad_norm": 4.117393970489502, + "learning_rate": 2.1e-05, + "loss": 3.3584, + "step": 214 + }, + { + "epoch": 0.058423913043478264, + "grad_norm": 3.1118719577789307, + "learning_rate": 2.11e-05, + "loss": 3.5, + "step": 215 + }, + { + "epoch": 0.058695652173913045, + "grad_norm": 3.3953042030334473, + "learning_rate": 2.12e-05, + "loss": 3.6948, + "step": 216 + }, + { + "epoch": 0.058967391304347826, + "grad_norm": 4.036997318267822, + "learning_rate": 2.13e-05, + "loss": 3.6871, + "step": 217 + }, + { + "epoch": 0.05923913043478261, + "grad_norm": 4.334664821624756, + "learning_rate": 2.1400000000000002e-05, + "loss": 4.0403, + "step": 218 + }, + { + "epoch": 0.059510869565217395, + "grad_norm": 3.2473344802856445, + "learning_rate": 2.15e-05, + "loss": 2.8208, + "step": 219 + }, + { + "epoch": 0.059782608695652176, + "grad_norm": 5.333037376403809, + "learning_rate": 2.16e-05, + "loss": 3.4915, + "step": 220 + }, + { + "epoch": 0.06005434782608696, + "grad_norm": 4.594522953033447, + "learning_rate": 2.1700000000000002e-05, + "loss": 3.773, + "step": 221 + }, + { + "epoch": 0.06032608695652174, + "grad_norm": 2.8913493156433105, + "learning_rate": 2.18e-05, + "loss": 3.5039, + "step": 222 + }, + { + "epoch": 0.06059782608695652, + "grad_norm": 3.1354737281799316, + "learning_rate": 2.19e-05, + "loss": 3.5823, + "step": 223 + }, + { + "epoch": 0.06086956521739131, + "grad_norm": 4.392479419708252, + "learning_rate": 2.2e-05, + "loss": 3.8093, + "step": 224 + }, + { + "epoch": 0.06114130434782609, + "grad_norm": 3.4441139698028564, + "learning_rate": 2.2100000000000002e-05, + "loss": 3.9232, + "step": 225 + }, + { + "epoch": 0.06141304347826087, + "grad_norm": 2.799862861633301, + "learning_rate": 2.22e-05, + "loss": 3.9281, + "step": 226 + }, + { + "epoch": 0.06168478260869565, + "grad_norm": 2.910454750061035, + "learning_rate": 2.23e-05, + "loss": 3.635, + "step": 227 + }, + { + "epoch": 0.06195652173913044, + "grad_norm": 3.9778056144714355, + "learning_rate": 2.2400000000000002e-05, + "loss": 3.5988, + "step": 228 + }, + { + "epoch": 0.06222826086956522, + "grad_norm": 3.500877857208252, + "learning_rate": 2.25e-05, + "loss": 3.2595, + "step": 229 + }, + { + "epoch": 0.0625, + "grad_norm": 3.3998258113861084, + "learning_rate": 2.26e-05, + "loss": 3.9116, + "step": 230 + }, + { + "epoch": 0.06277173913043478, + "grad_norm": 2.585266351699829, + "learning_rate": 2.2700000000000003e-05, + "loss": 3.6087, + "step": 231 + }, + { + "epoch": 0.06304347826086956, + "grad_norm": 2.313265085220337, + "learning_rate": 2.2800000000000002e-05, + "loss": 3.8507, + "step": 232 + }, + { + "epoch": 0.06331521739130434, + "grad_norm": 6.233187198638916, + "learning_rate": 2.29e-05, + "loss": 4.6271, + "step": 233 + }, + { + "epoch": 0.06358695652173912, + "grad_norm": 4.387447357177734, + "learning_rate": 2.3000000000000003e-05, + "loss": 3.7921, + "step": 234 + }, + { + "epoch": 0.06385869565217392, + "grad_norm": 3.573700189590454, + "learning_rate": 2.3100000000000002e-05, + "loss": 3.6567, + "step": 235 + }, + { + "epoch": 0.0641304347826087, + "grad_norm": 1.9834128618240356, + "learning_rate": 2.32e-05, + "loss": 3.715, + "step": 236 + }, + { + "epoch": 0.06440217391304348, + "grad_norm": 3.0220143795013428, + "learning_rate": 2.33e-05, + "loss": 4.302, + "step": 237 + }, + { + "epoch": 0.06467391304347826, + "grad_norm": 5.85452938079834, + "learning_rate": 2.3400000000000003e-05, + "loss": 4.3755, + "step": 238 + }, + { + "epoch": 0.06494565217391304, + "grad_norm": 6.263114929199219, + "learning_rate": 2.3500000000000002e-05, + "loss": 4.0045, + "step": 239 + }, + { + "epoch": 0.06521739130434782, + "grad_norm": 4.6512322425842285, + "learning_rate": 2.3599999999999998e-05, + "loss": 4.202, + "step": 240 + }, + { + "epoch": 0.0654891304347826, + "grad_norm": 1.589857578277588, + "learning_rate": 2.37e-05, + "loss": 3.0368, + "step": 241 + }, + { + "epoch": 0.06576086956521739, + "grad_norm": 3.0660102367401123, + "learning_rate": 2.38e-05, + "loss": 3.959, + "step": 242 + }, + { + "epoch": 0.06603260869565217, + "grad_norm": 3.7164723873138428, + "learning_rate": 2.3899999999999998e-05, + "loss": 3.5374, + "step": 243 + }, + { + "epoch": 0.06630434782608696, + "grad_norm": 3.772892713546753, + "learning_rate": 2.4e-05, + "loss": 3.3464, + "step": 244 + }, + { + "epoch": 0.06657608695652174, + "grad_norm": 3.2034425735473633, + "learning_rate": 2.41e-05, + "loss": 3.3077, + "step": 245 + }, + { + "epoch": 0.06684782608695652, + "grad_norm": 4.294682502746582, + "learning_rate": 2.42e-05, + "loss": 3.813, + "step": 246 + }, + { + "epoch": 0.0671195652173913, + "grad_norm": 3.805955171585083, + "learning_rate": 2.43e-05, + "loss": 3.6854, + "step": 247 + }, + { + "epoch": 0.06739130434782609, + "grad_norm": 4.587597370147705, + "learning_rate": 2.44e-05, + "loss": 3.3058, + "step": 248 + }, + { + "epoch": 0.06766304347826087, + "grad_norm": 5.513325214385986, + "learning_rate": 2.45e-05, + "loss": 3.6714, + "step": 249 + }, + { + "epoch": 0.06793478260869565, + "grad_norm": 5.820981025695801, + "learning_rate": 2.4599999999999998e-05, + "loss": 3.6253, + "step": 250 + }, + { + "epoch": 0.06820652173913043, + "grad_norm": 4.236602306365967, + "learning_rate": 2.47e-05, + "loss": 3.605, + "step": 251 + }, + { + "epoch": 0.06847826086956521, + "grad_norm": 3.524596929550171, + "learning_rate": 2.48e-05, + "loss": 3.6746, + "step": 252 + }, + { + "epoch": 0.06875, + "grad_norm": 4.912044525146484, + "learning_rate": 2.49e-05, + "loss": 3.6485, + "step": 253 + }, + { + "epoch": 0.06902173913043479, + "grad_norm": 5.730941295623779, + "learning_rate": 2.5e-05, + "loss": 3.7662, + "step": 254 + }, + { + "epoch": 0.06929347826086957, + "grad_norm": 7.801654815673828, + "learning_rate": 2.51e-05, + "loss": 4.8335, + "step": 255 + }, + { + "epoch": 0.06956521739130435, + "grad_norm": 4.116402626037598, + "learning_rate": 2.52e-05, + "loss": 3.9456, + "step": 256 + }, + { + "epoch": 0.06983695652173913, + "grad_norm": 6.192750930786133, + "learning_rate": 2.5300000000000002e-05, + "loss": 4.0458, + "step": 257 + }, + { + "epoch": 0.07010869565217391, + "grad_norm": 4.565896511077881, + "learning_rate": 2.54e-05, + "loss": 3.3915, + "step": 258 + }, + { + "epoch": 0.07038043478260869, + "grad_norm": 4.502013206481934, + "learning_rate": 2.55e-05, + "loss": 3.4612, + "step": 259 + }, + { + "epoch": 0.07065217391304347, + "grad_norm": 3.8033788204193115, + "learning_rate": 2.5600000000000002e-05, + "loss": 2.9393, + "step": 260 + }, + { + "epoch": 0.07092391304347827, + "grad_norm": 4.34993839263916, + "learning_rate": 2.57e-05, + "loss": 3.6575, + "step": 261 + }, + { + "epoch": 0.07119565217391305, + "grad_norm": 3.9064600467681885, + "learning_rate": 2.58e-05, + "loss": 3.6109, + "step": 262 + }, + { + "epoch": 0.07146739130434783, + "grad_norm": 3.737865447998047, + "learning_rate": 2.59e-05, + "loss": 3.7551, + "step": 263 + }, + { + "epoch": 0.07173913043478261, + "grad_norm": 4.92514181137085, + "learning_rate": 2.6000000000000002e-05, + "loss": 3.9133, + "step": 264 + }, + { + "epoch": 0.07201086956521739, + "grad_norm": 4.838226318359375, + "learning_rate": 2.61e-05, + "loss": 3.4457, + "step": 265 + }, + { + "epoch": 0.07228260869565217, + "grad_norm": 5.465858459472656, + "learning_rate": 2.62e-05, + "loss": 3.3492, + "step": 266 + }, + { + "epoch": 0.07255434782608695, + "grad_norm": 4.452864646911621, + "learning_rate": 2.6300000000000002e-05, + "loss": 4.0503, + "step": 267 + }, + { + "epoch": 0.07282608695652174, + "grad_norm": 2.657294750213623, + "learning_rate": 2.64e-05, + "loss": 3.5952, + "step": 268 + }, + { + "epoch": 0.07309782608695652, + "grad_norm": 3.738781213760376, + "learning_rate": 2.65e-05, + "loss": 3.2922, + "step": 269 + }, + { + "epoch": 0.07336956521739131, + "grad_norm": 4.5520524978637695, + "learning_rate": 2.6600000000000003e-05, + "loss": 3.4737, + "step": 270 + }, + { + "epoch": 0.07364130434782609, + "grad_norm": 6.709564685821533, + "learning_rate": 2.6700000000000002e-05, + "loss": 3.8822, + "step": 271 + }, + { + "epoch": 0.07391304347826087, + "grad_norm": 5.555806636810303, + "learning_rate": 2.68e-05, + "loss": 4.1245, + "step": 272 + }, + { + "epoch": 0.07418478260869565, + "grad_norm": 3.99271821975708, + "learning_rate": 2.69e-05, + "loss": 3.535, + "step": 273 + }, + { + "epoch": 0.07445652173913044, + "grad_norm": 3.0163462162017822, + "learning_rate": 2.7000000000000002e-05, + "loss": 3.4727, + "step": 274 + }, + { + "epoch": 0.07472826086956522, + "grad_norm": 5.363269329071045, + "learning_rate": 2.71e-05, + "loss": 4.1682, + "step": 275 + }, + { + "epoch": 0.075, + "grad_norm": 4.282686710357666, + "learning_rate": 2.72e-05, + "loss": 3.1355, + "step": 276 + }, + { + "epoch": 0.07527173913043478, + "grad_norm": 5.983767509460449, + "learning_rate": 2.7300000000000003e-05, + "loss": 3.8695, + "step": 277 + }, + { + "epoch": 0.07554347826086956, + "grad_norm": 5.345359802246094, + "learning_rate": 2.7400000000000002e-05, + "loss": 4.1882, + "step": 278 + }, + { + "epoch": 0.07581521739130435, + "grad_norm": 2.5548171997070312, + "learning_rate": 2.75e-05, + "loss": 3.2401, + "step": 279 + }, + { + "epoch": 0.07608695652173914, + "grad_norm": 3.790065050125122, + "learning_rate": 2.7600000000000003e-05, + "loss": 3.8057, + "step": 280 + }, + { + "epoch": 0.07635869565217392, + "grad_norm": 4.037810325622559, + "learning_rate": 2.7700000000000002e-05, + "loss": 3.8397, + "step": 281 + }, + { + "epoch": 0.0766304347826087, + "grad_norm": 4.571537017822266, + "learning_rate": 2.78e-05, + "loss": 3.1902, + "step": 282 + }, + { + "epoch": 0.07690217391304348, + "grad_norm": 4.4513702392578125, + "learning_rate": 2.79e-05, + "loss": 3.6552, + "step": 283 + }, + { + "epoch": 0.07717391304347826, + "grad_norm": 2.919421911239624, + "learning_rate": 2.8e-05, + "loss": 3.0449, + "step": 284 + }, + { + "epoch": 0.07744565217391304, + "grad_norm": 3.9303693771362305, + "learning_rate": 2.81e-05, + "loss": 3.8186, + "step": 285 + }, + { + "epoch": 0.07771739130434782, + "grad_norm": 3.8404746055603027, + "learning_rate": 2.8199999999999998e-05, + "loss": 3.2603, + "step": 286 + }, + { + "epoch": 0.0779891304347826, + "grad_norm": 3.2469253540039062, + "learning_rate": 2.83e-05, + "loss": 2.917, + "step": 287 + }, + { + "epoch": 0.0782608695652174, + "grad_norm": 2.8970375061035156, + "learning_rate": 2.84e-05, + "loss": 3.6446, + "step": 288 + }, + { + "epoch": 0.07853260869565218, + "grad_norm": 4.029104232788086, + "learning_rate": 2.8499999999999998e-05, + "loss": 4.1754, + "step": 289 + }, + { + "epoch": 0.07880434782608696, + "grad_norm": 3.967705726623535, + "learning_rate": 2.86e-05, + "loss": 3.3913, + "step": 290 + }, + { + "epoch": 0.07907608695652174, + "grad_norm": 3.084944009780884, + "learning_rate": 2.87e-05, + "loss": 3.4598, + "step": 291 + }, + { + "epoch": 0.07934782608695652, + "grad_norm": 3.2562334537506104, + "learning_rate": 2.88e-05, + "loss": 3.7326, + "step": 292 + }, + { + "epoch": 0.0796195652173913, + "grad_norm": 3.9668397903442383, + "learning_rate": 2.89e-05, + "loss": 3.8312, + "step": 293 + }, + { + "epoch": 0.07989130434782608, + "grad_norm": 3.7607946395874023, + "learning_rate": 2.9e-05, + "loss": 3.9309, + "step": 294 + }, + { + "epoch": 0.08016304347826086, + "grad_norm": 4.704547882080078, + "learning_rate": 2.91e-05, + "loss": 3.9181, + "step": 295 + }, + { + "epoch": 0.08043478260869565, + "grad_norm": 5.68735408782959, + "learning_rate": 2.92e-05, + "loss": 3.6405, + "step": 296 + }, + { + "epoch": 0.08070652173913044, + "grad_norm": 4.839725971221924, + "learning_rate": 2.93e-05, + "loss": 3.6735, + "step": 297 + }, + { + "epoch": 0.08097826086956522, + "grad_norm": 3.6157710552215576, + "learning_rate": 2.94e-05, + "loss": 4.1132, + "step": 298 + }, + { + "epoch": 0.08125, + "grad_norm": 3.6725997924804688, + "learning_rate": 2.95e-05, + "loss": 3.9299, + "step": 299 + }, + { + "epoch": 0.08152173913043478, + "grad_norm": 4.408240795135498, + "learning_rate": 2.96e-05, + "loss": 3.4141, + "step": 300 + }, + { + "epoch": 0.08179347826086956, + "grad_norm": 3.8871631622314453, + "learning_rate": 2.97e-05, + "loss": 2.7057, + "step": 301 + }, + { + "epoch": 0.08206521739130435, + "grad_norm": 5.148252010345459, + "learning_rate": 2.98e-05, + "loss": 3.0604, + "step": 302 + }, + { + "epoch": 0.08233695652173913, + "grad_norm": 3.8631784915924072, + "learning_rate": 2.9900000000000002e-05, + "loss": 3.4581, + "step": 303 + }, + { + "epoch": 0.08260869565217391, + "grad_norm": 2.67698073387146, + "learning_rate": 3e-05, + "loss": 3.1465, + "step": 304 + }, + { + "epoch": 0.08288043478260869, + "grad_norm": 2.6791391372680664, + "learning_rate": 3e-05, + "loss": 2.8593, + "step": 305 + }, + { + "epoch": 0.08315217391304348, + "grad_norm": 3.844003677368164, + "learning_rate": 3e-05, + "loss": 3.4371, + "step": 306 + }, + { + "epoch": 0.08342391304347826, + "grad_norm": 5.624002933502197, + "learning_rate": 3e-05, + "loss": 3.8372, + "step": 307 + }, + { + "epoch": 0.08369565217391305, + "grad_norm": 4.992491245269775, + "learning_rate": 3e-05, + "loss": 3.6231, + "step": 308 + }, + { + "epoch": 0.08396739130434783, + "grad_norm": 4.717808246612549, + "learning_rate": 3e-05, + "loss": 3.9652, + "step": 309 + }, + { + "epoch": 0.08423913043478261, + "grad_norm": 3.6492786407470703, + "learning_rate": 3e-05, + "loss": 3.7065, + "step": 310 + }, + { + "epoch": 0.08451086956521739, + "grad_norm": 4.453199863433838, + "learning_rate": 3e-05, + "loss": 3.7317, + "step": 311 + }, + { + "epoch": 0.08478260869565217, + "grad_norm": 5.009824275970459, + "learning_rate": 3e-05, + "loss": 3.7651, + "step": 312 + }, + { + "epoch": 0.08505434782608695, + "grad_norm": 6.800715446472168, + "learning_rate": 3e-05, + "loss": 4.0778, + "step": 313 + }, + { + "epoch": 0.08532608695652173, + "grad_norm": 6.225503444671631, + "learning_rate": 3e-05, + "loss": 4.3679, + "step": 314 + }, + { + "epoch": 0.08559782608695653, + "grad_norm": 2.9776594638824463, + "learning_rate": 3e-05, + "loss": 3.1434, + "step": 315 + }, + { + "epoch": 0.08586956521739131, + "grad_norm": 3.1179118156433105, + "learning_rate": 3e-05, + "loss": 3.5896, + "step": 316 + }, + { + "epoch": 0.08614130434782609, + "grad_norm": 2.81669282913208, + "learning_rate": 3e-05, + "loss": 2.9054, + "step": 317 + }, + { + "epoch": 0.08641304347826087, + "grad_norm": 3.894665002822876, + "learning_rate": 3e-05, + "loss": 3.2122, + "step": 318 + }, + { + "epoch": 0.08668478260869565, + "grad_norm": 5.43796443939209, + "learning_rate": 3e-05, + "loss": 3.8039, + "step": 319 + }, + { + "epoch": 0.08695652173913043, + "grad_norm": 3.9177141189575195, + "learning_rate": 3e-05, + "loss": 3.6393, + "step": 320 + }, + { + "epoch": 0.08722826086956521, + "grad_norm": 3.0576508045196533, + "learning_rate": 3e-05, + "loss": 3.67, + "step": 321 + }, + { + "epoch": 0.0875, + "grad_norm": 5.208736896514893, + "learning_rate": 3e-05, + "loss": 3.9719, + "step": 322 + }, + { + "epoch": 0.08777173913043479, + "grad_norm": 4.839489936828613, + "learning_rate": 3e-05, + "loss": 3.7197, + "step": 323 + }, + { + "epoch": 0.08804347826086957, + "grad_norm": 4.195175647735596, + "learning_rate": 3e-05, + "loss": 3.3963, + "step": 324 + }, + { + "epoch": 0.08831521739130435, + "grad_norm": 3.565690279006958, + "learning_rate": 3e-05, + "loss": 3.177, + "step": 325 + }, + { + "epoch": 0.08858695652173913, + "grad_norm": 2.6973624229431152, + "learning_rate": 3e-05, + "loss": 3.0416, + "step": 326 + }, + { + "epoch": 0.08885869565217391, + "grad_norm": 2.620365619659424, + "learning_rate": 3e-05, + "loss": 3.3155, + "step": 327 + }, + { + "epoch": 0.0891304347826087, + "grad_norm": 3.7461090087890625, + "learning_rate": 3e-05, + "loss": 3.6553, + "step": 328 + }, + { + "epoch": 0.08940217391304348, + "grad_norm": 6.702042102813721, + "learning_rate": 3e-05, + "loss": 4.2574, + "step": 329 + }, + { + "epoch": 0.08967391304347826, + "grad_norm": 5.888139247894287, + "learning_rate": 3e-05, + "loss": 4.148, + "step": 330 + }, + { + "epoch": 0.08994565217391304, + "grad_norm": 4.737085342407227, + "learning_rate": 3e-05, + "loss": 3.8761, + "step": 331 + }, + { + "epoch": 0.09021739130434783, + "grad_norm": 3.9981937408447266, + "learning_rate": 3e-05, + "loss": 4.4605, + "step": 332 + }, + { + "epoch": 0.09048913043478261, + "grad_norm": 5.0993266105651855, + "learning_rate": 3e-05, + "loss": 4.095, + "step": 333 + }, + { + "epoch": 0.0907608695652174, + "grad_norm": 4.450422763824463, + "learning_rate": 3e-05, + "loss": 2.8801, + "step": 334 + }, + { + "epoch": 0.09103260869565218, + "grad_norm": 6.846070766448975, + "learning_rate": 3e-05, + "loss": 3.5946, + "step": 335 + }, + { + "epoch": 0.09130434782608696, + "grad_norm": 5.887484550476074, + "learning_rate": 3e-05, + "loss": 3.4395, + "step": 336 + }, + { + "epoch": 0.09157608695652174, + "grad_norm": 3.8128914833068848, + "learning_rate": 3e-05, + "loss": 3.0863, + "step": 337 + }, + { + "epoch": 0.09184782608695652, + "grad_norm": 2.493584394454956, + "learning_rate": 3e-05, + "loss": 3.0949, + "step": 338 + }, + { + "epoch": 0.0921195652173913, + "grad_norm": 3.304899215698242, + "learning_rate": 3e-05, + "loss": 3.3348, + "step": 339 + }, + { + "epoch": 0.09239130434782608, + "grad_norm": 4.633394241333008, + "learning_rate": 3e-05, + "loss": 3.5233, + "step": 340 + }, + { + "epoch": 0.09266304347826088, + "grad_norm": 5.405637264251709, + "learning_rate": 3e-05, + "loss": 3.776, + "step": 341 + }, + { + "epoch": 0.09293478260869566, + "grad_norm": 5.505592346191406, + "learning_rate": 3e-05, + "loss": 3.9894, + "step": 342 + }, + { + "epoch": 0.09320652173913044, + "grad_norm": 3.901547431945801, + "learning_rate": 3e-05, + "loss": 4.1865, + "step": 343 + }, + { + "epoch": 0.09347826086956522, + "grad_norm": 3.3303725719451904, + "learning_rate": 3e-05, + "loss": 3.3656, + "step": 344 + }, + { + "epoch": 0.09375, + "grad_norm": 5.3323493003845215, + "learning_rate": 3e-05, + "loss": 4.2369, + "step": 345 + }, + { + "epoch": 0.09402173913043478, + "grad_norm": 5.352675914764404, + "learning_rate": 3e-05, + "loss": 3.7106, + "step": 346 + }, + { + "epoch": 0.09429347826086956, + "grad_norm": 5.6517839431762695, + "learning_rate": 3e-05, + "loss": 3.4388, + "step": 347 + }, + { + "epoch": 0.09456521739130434, + "grad_norm": 4.7645111083984375, + "learning_rate": 3e-05, + "loss": 3.4332, + "step": 348 + }, + { + "epoch": 0.09483695652173912, + "grad_norm": 3.140958309173584, + "learning_rate": 3e-05, + "loss": 3.2139, + "step": 349 + }, + { + "epoch": 0.09510869565217392, + "grad_norm": 2.692319393157959, + "learning_rate": 3e-05, + "loss": 3.3015, + "step": 350 + }, + { + "epoch": 0.0953804347826087, + "grad_norm": 4.1542134284973145, + "learning_rate": 3e-05, + "loss": 3.29, + "step": 351 + }, + { + "epoch": 0.09565217391304348, + "grad_norm": 4.946714401245117, + "learning_rate": 3e-05, + "loss": 3.2509, + "step": 352 + }, + { + "epoch": 0.09592391304347826, + "grad_norm": 5.529204368591309, + "learning_rate": 3e-05, + "loss": 4.0578, + "step": 353 + }, + { + "epoch": 0.09619565217391304, + "grad_norm": 3.837975025177002, + "learning_rate": 3e-05, + "loss": 3.7921, + "step": 354 + }, + { + "epoch": 0.09646739130434782, + "grad_norm": 3.0146334171295166, + "learning_rate": 3e-05, + "loss": 4.0369, + "step": 355 + }, + { + "epoch": 0.0967391304347826, + "grad_norm": 4.283068656921387, + "learning_rate": 3e-05, + "loss": 3.9506, + "step": 356 + }, + { + "epoch": 0.09701086956521739, + "grad_norm": 4.30133581161499, + "learning_rate": 3e-05, + "loss": 3.4218, + "step": 357 + }, + { + "epoch": 0.09728260869565217, + "grad_norm": 6.228667736053467, + "learning_rate": 3e-05, + "loss": 4.0054, + "step": 358 + }, + { + "epoch": 0.09755434782608696, + "grad_norm": 3.927131414413452, + "learning_rate": 3e-05, + "loss": 2.9401, + "step": 359 + }, + { + "epoch": 0.09782608695652174, + "grad_norm": 4.374425888061523, + "learning_rate": 3e-05, + "loss": 3.5279, + "step": 360 + }, + { + "epoch": 0.09809782608695652, + "grad_norm": 2.559403419494629, + "learning_rate": 3e-05, + "loss": 2.8306, + "step": 361 + }, + { + "epoch": 0.0983695652173913, + "grad_norm": 2.9556596279144287, + "learning_rate": 3e-05, + "loss": 3.4532, + "step": 362 + }, + { + "epoch": 0.09864130434782609, + "grad_norm": 4.841480731964111, + "learning_rate": 3e-05, + "loss": 3.474, + "step": 363 + }, + { + "epoch": 0.09891304347826087, + "grad_norm": 5.785233497619629, + "learning_rate": 3e-05, + "loss": 3.6698, + "step": 364 + }, + { + "epoch": 0.09918478260869565, + "grad_norm": 5.180464267730713, + "learning_rate": 3e-05, + "loss": 3.4199, + "step": 365 + }, + { + "epoch": 0.09945652173913043, + "grad_norm": 3.5438666343688965, + "learning_rate": 3e-05, + "loss": 3.5603, + "step": 366 + }, + { + "epoch": 0.09972826086956521, + "grad_norm": 3.1015937328338623, + "learning_rate": 3e-05, + "loss": 2.9937, + "step": 367 + }, + { + "epoch": 0.1, + "grad_norm": 4.174754619598389, + "learning_rate": 3e-05, + "loss": 3.5994, + "step": 368 + }, + { + "epoch": 0.10027173913043479, + "grad_norm": 4.678737163543701, + "learning_rate": 3e-05, + "loss": 4.0816, + "step": 369 + }, + { + "epoch": 0.10054347826086957, + "grad_norm": 3.884875774383545, + "learning_rate": 3e-05, + "loss": 2.8268, + "step": 370 + }, + { + "epoch": 0.10081521739130435, + "grad_norm": 4.060309886932373, + "learning_rate": 3e-05, + "loss": 3.5038, + "step": 371 + }, + { + "epoch": 0.10108695652173913, + "grad_norm": 4.1803812980651855, + "learning_rate": 3e-05, + "loss": 3.8122, + "step": 372 + }, + { + "epoch": 0.10135869565217391, + "grad_norm": 3.1725456714630127, + "learning_rate": 3e-05, + "loss": 3.7945, + "step": 373 + }, + { + "epoch": 0.10163043478260869, + "grad_norm": 3.1784756183624268, + "learning_rate": 3e-05, + "loss": 3.0676, + "step": 374 + }, + { + "epoch": 0.10190217391304347, + "grad_norm": 4.812568187713623, + "learning_rate": 3e-05, + "loss": 3.593, + "step": 375 + }, + { + "epoch": 0.10217391304347827, + "grad_norm": 5.944811820983887, + "learning_rate": 3e-05, + "loss": 3.843, + "step": 376 + }, + { + "epoch": 0.10244565217391305, + "grad_norm": 4.13525915145874, + "learning_rate": 3e-05, + "loss": 3.9103, + "step": 377 + }, + { + "epoch": 0.10271739130434783, + "grad_norm": 5.860477447509766, + "learning_rate": 3e-05, + "loss": 4.9384, + "step": 378 + }, + { + "epoch": 0.10298913043478261, + "grad_norm": 3.2515783309936523, + "learning_rate": 3e-05, + "loss": 3.0586, + "step": 379 + }, + { + "epoch": 0.10326086956521739, + "grad_norm": 4.5452494621276855, + "learning_rate": 3e-05, + "loss": 3.7046, + "step": 380 + }, + { + "epoch": 0.10353260869565217, + "grad_norm": 3.8891849517822266, + "learning_rate": 3e-05, + "loss": 3.34, + "step": 381 + }, + { + "epoch": 0.10380434782608695, + "grad_norm": 4.083732604980469, + "learning_rate": 3e-05, + "loss": 3.3112, + "step": 382 + }, + { + "epoch": 0.10407608695652174, + "grad_norm": 3.438572406768799, + "learning_rate": 3e-05, + "loss": 2.9908, + "step": 383 + }, + { + "epoch": 0.10434782608695652, + "grad_norm": 3.0750012397766113, + "learning_rate": 3e-05, + "loss": 3.3847, + "step": 384 + }, + { + "epoch": 0.10461956521739131, + "grad_norm": 3.0833847522735596, + "learning_rate": 3e-05, + "loss": 3.2274, + "step": 385 + }, + { + "epoch": 0.10489130434782609, + "grad_norm": 4.356601238250732, + "learning_rate": 3e-05, + "loss": 3.9247, + "step": 386 + }, + { + "epoch": 0.10516304347826087, + "grad_norm": 4.894799709320068, + "learning_rate": 3e-05, + "loss": 3.6005, + "step": 387 + }, + { + "epoch": 0.10543478260869565, + "grad_norm": 4.887790203094482, + "learning_rate": 3e-05, + "loss": 4.3258, + "step": 388 + }, + { + "epoch": 0.10570652173913044, + "grad_norm": 4.861374855041504, + "learning_rate": 3e-05, + "loss": 3.6857, + "step": 389 + }, + { + "epoch": 0.10597826086956522, + "grad_norm": 4.775892734527588, + "learning_rate": 3e-05, + "loss": 4.3695, + "step": 390 + }, + { + "epoch": 0.10625, + "grad_norm": 4.2715582847595215, + "learning_rate": 3e-05, + "loss": 3.7657, + "step": 391 + }, + { + "epoch": 0.10652173913043478, + "grad_norm": 3.8440310955047607, + "learning_rate": 3e-05, + "loss": 3.4856, + "step": 392 + }, + { + "epoch": 0.10679347826086956, + "grad_norm": 4.285994529724121, + "learning_rate": 3e-05, + "loss": 3.4081, + "step": 393 + }, + { + "epoch": 0.10706521739130435, + "grad_norm": 2.1318747997283936, + "learning_rate": 3e-05, + "loss": 2.9345, + "step": 394 + }, + { + "epoch": 0.10733695652173914, + "grad_norm": 2.4292728900909424, + "learning_rate": 3e-05, + "loss": 3.091, + "step": 395 + }, + { + "epoch": 0.10760869565217392, + "grad_norm": 2.3496248722076416, + "learning_rate": 3e-05, + "loss": 3.1994, + "step": 396 + }, + { + "epoch": 0.1078804347826087, + "grad_norm": 5.604576110839844, + "learning_rate": 3e-05, + "loss": 4.6392, + "step": 397 + }, + { + "epoch": 0.10815217391304348, + "grad_norm": 5.448110580444336, + "learning_rate": 3e-05, + "loss": 4.4596, + "step": 398 + }, + { + "epoch": 0.10842391304347826, + "grad_norm": 3.2921388149261475, + "learning_rate": 3e-05, + "loss": 3.5248, + "step": 399 + }, + { + "epoch": 0.10869565217391304, + "grad_norm": 3.1395716667175293, + "learning_rate": 3e-05, + "loss": 4.1658, + "step": 400 + }, + { + "epoch": 0.10896739130434782, + "grad_norm": 2.3313381671905518, + "learning_rate": 3e-05, + "loss": 3.7592, + "step": 401 + }, + { + "epoch": 0.1092391304347826, + "grad_norm": 2.9619717597961426, + "learning_rate": 3e-05, + "loss": 3.5212, + "step": 402 + }, + { + "epoch": 0.1095108695652174, + "grad_norm": 3.2616071701049805, + "learning_rate": 3e-05, + "loss": 3.2052, + "step": 403 + }, + { + "epoch": 0.10978260869565218, + "grad_norm": 4.842064380645752, + "learning_rate": 3e-05, + "loss": 3.907, + "step": 404 + }, + { + "epoch": 0.11005434782608696, + "grad_norm": 3.6140100955963135, + "learning_rate": 3e-05, + "loss": 4.061, + "step": 405 + }, + { + "epoch": 0.11032608695652174, + "grad_norm": 1.4534454345703125, + "learning_rate": 3e-05, + "loss": 3.0512, + "step": 406 + }, + { + "epoch": 0.11059782608695652, + "grad_norm": 2.4747161865234375, + "learning_rate": 3e-05, + "loss": 3.7008, + "step": 407 + }, + { + "epoch": 0.1108695652173913, + "grad_norm": 3.0130977630615234, + "learning_rate": 3e-05, + "loss": 3.9765, + "step": 408 + }, + { + "epoch": 0.11114130434782608, + "grad_norm": 2.256558656692505, + "learning_rate": 3e-05, + "loss": 3.2158, + "step": 409 + }, + { + "epoch": 0.11141304347826086, + "grad_norm": 4.283812522888184, + "learning_rate": 3e-05, + "loss": 4.4811, + "step": 410 + }, + { + "epoch": 0.11168478260869565, + "grad_norm": 2.566425323486328, + "learning_rate": 3e-05, + "loss": 3.9403, + "step": 411 + }, + { + "epoch": 0.11195652173913044, + "grad_norm": 2.4623820781707764, + "learning_rate": 3e-05, + "loss": 3.2361, + "step": 412 + }, + { + "epoch": 0.11222826086956522, + "grad_norm": 2.934722661972046, + "learning_rate": 3e-05, + "loss": 3.0554, + "step": 413 + }, + { + "epoch": 0.1125, + "grad_norm": 5.524335861206055, + "learning_rate": 3e-05, + "loss": 4.0103, + "step": 414 + }, + { + "epoch": 0.11277173913043478, + "grad_norm": 3.285830020904541, + "learning_rate": 3e-05, + "loss": 3.6652, + "step": 415 + }, + { + "epoch": 0.11304347826086956, + "grad_norm": 2.5553982257843018, + "learning_rate": 3e-05, + "loss": 3.8115, + "step": 416 + }, + { + "epoch": 0.11331521739130435, + "grad_norm": 1.8953039646148682, + "learning_rate": 3e-05, + "loss": 3.6769, + "step": 417 + }, + { + "epoch": 0.11358695652173913, + "grad_norm": 2.4014382362365723, + "learning_rate": 3e-05, + "loss": 3.6343, + "step": 418 + }, + { + "epoch": 0.11385869565217391, + "grad_norm": 3.2362847328186035, + "learning_rate": 3e-05, + "loss": 3.722, + "step": 419 + }, + { + "epoch": 0.11413043478260869, + "grad_norm": 2.525406837463379, + "learning_rate": 3e-05, + "loss": 3.2201, + "step": 420 + }, + { + "epoch": 0.11440217391304348, + "grad_norm": 2.492753028869629, + "learning_rate": 3e-05, + "loss": 3.3864, + "step": 421 + }, + { + "epoch": 0.11467391304347826, + "grad_norm": 1.6389284133911133, + "learning_rate": 3e-05, + "loss": 3.1505, + "step": 422 + }, + { + "epoch": 0.11494565217391305, + "grad_norm": 2.721895933151245, + "learning_rate": 3e-05, + "loss": 4.02, + "step": 423 + }, + { + "epoch": 0.11521739130434783, + "grad_norm": 2.1239542961120605, + "learning_rate": 3e-05, + "loss": 3.3395, + "step": 424 + }, + { + "epoch": 0.11548913043478261, + "grad_norm": 2.3277628421783447, + "learning_rate": 3e-05, + "loss": 4.3061, + "step": 425 + }, + { + "epoch": 0.11576086956521739, + "grad_norm": 1.7965201139450073, + "learning_rate": 3e-05, + "loss": 3.8778, + "step": 426 + }, + { + "epoch": 0.11603260869565217, + "grad_norm": 2.9034996032714844, + "learning_rate": 3e-05, + "loss": 3.8348, + "step": 427 + }, + { + "epoch": 0.11630434782608695, + "grad_norm": 2.0713295936584473, + "learning_rate": 3e-05, + "loss": 3.6901, + "step": 428 + }, + { + "epoch": 0.11657608695652173, + "grad_norm": 2.612128257751465, + "learning_rate": 3e-05, + "loss": 3.1308, + "step": 429 + }, + { + "epoch": 0.11684782608695653, + "grad_norm": 3.535771369934082, + "learning_rate": 3e-05, + "loss": 3.7629, + "step": 430 + }, + { + "epoch": 0.11711956521739131, + "grad_norm": 2.6418893337249756, + "learning_rate": 3e-05, + "loss": 3.351, + "step": 431 + }, + { + "epoch": 0.11739130434782609, + "grad_norm": 1.9122227430343628, + "learning_rate": 3e-05, + "loss": 2.9162, + "step": 432 + }, + { + "epoch": 0.11766304347826087, + "grad_norm": 3.12231707572937, + "learning_rate": 3e-05, + "loss": 3.6215, + "step": 433 + }, + { + "epoch": 0.11793478260869565, + "grad_norm": 2.5034008026123047, + "learning_rate": 3e-05, + "loss": 3.6675, + "step": 434 + }, + { + "epoch": 0.11820652173913043, + "grad_norm": 2.69254732131958, + "learning_rate": 3e-05, + "loss": 3.8847, + "step": 435 + }, + { + "epoch": 0.11847826086956521, + "grad_norm": 2.026611566543579, + "learning_rate": 3e-05, + "loss": 3.9927, + "step": 436 + }, + { + "epoch": 0.11875, + "grad_norm": 3.05465030670166, + "learning_rate": 3e-05, + "loss": 3.4363, + "step": 437 + }, + { + "epoch": 0.11902173913043479, + "grad_norm": 2.800114631652832, + "learning_rate": 3e-05, + "loss": 4.0161, + "step": 438 + }, + { + "epoch": 0.11929347826086957, + "grad_norm": 2.452852725982666, + "learning_rate": 3e-05, + "loss": 3.8825, + "step": 439 + }, + { + "epoch": 0.11956521739130435, + "grad_norm": 2.425046920776367, + "learning_rate": 3e-05, + "loss": 3.8342, + "step": 440 + }, + { + "epoch": 0.11983695652173913, + "grad_norm": 1.6898943185806274, + "learning_rate": 3e-05, + "loss": 3.4458, + "step": 441 + }, + { + "epoch": 0.12010869565217391, + "grad_norm": 2.265521764755249, + "learning_rate": 3e-05, + "loss": 3.498, + "step": 442 + }, + { + "epoch": 0.1203804347826087, + "grad_norm": 3.3182766437530518, + "learning_rate": 3e-05, + "loss": 4.0802, + "step": 443 + }, + { + "epoch": 0.12065217391304348, + "grad_norm": 3.441931962966919, + "learning_rate": 3e-05, + "loss": 3.7694, + "step": 444 + }, + { + "epoch": 0.12092391304347826, + "grad_norm": 2.1982293128967285, + "learning_rate": 3e-05, + "loss": 3.3313, + "step": 445 + }, + { + "epoch": 0.12119565217391304, + "grad_norm": 1.986126184463501, + "learning_rate": 3e-05, + "loss": 3.3758, + "step": 446 + }, + { + "epoch": 0.12146739130434783, + "grad_norm": 1.6694939136505127, + "learning_rate": 3e-05, + "loss": 3.1291, + "step": 447 + }, + { + "epoch": 0.12173913043478261, + "grad_norm": 3.418254852294922, + "learning_rate": 3e-05, + "loss": 3.9384, + "step": 448 + }, + { + "epoch": 0.1220108695652174, + "grad_norm": 2.4987125396728516, + "learning_rate": 3e-05, + "loss": 4.0415, + "step": 449 + }, + { + "epoch": 0.12228260869565218, + "grad_norm": 2.4183948040008545, + "learning_rate": 3e-05, + "loss": 3.4816, + "step": 450 + }, + { + "epoch": 0.12255434782608696, + "grad_norm": 2.753169059753418, + "learning_rate": 3e-05, + "loss": 3.3146, + "step": 451 + }, + { + "epoch": 0.12282608695652174, + "grad_norm": 2.2018933296203613, + "learning_rate": 3e-05, + "loss": 3.5041, + "step": 452 + }, + { + "epoch": 0.12309782608695652, + "grad_norm": 1.910427451133728, + "learning_rate": 3e-05, + "loss": 3.1185, + "step": 453 + }, + { + "epoch": 0.1233695652173913, + "grad_norm": 2.2100231647491455, + "learning_rate": 3e-05, + "loss": 3.1726, + "step": 454 + }, + { + "epoch": 0.12364130434782608, + "grad_norm": 2.6990854740142822, + "learning_rate": 3e-05, + "loss": 3.7998, + "step": 455 + }, + { + "epoch": 0.12391304347826088, + "grad_norm": 3.1403119564056396, + "learning_rate": 3e-05, + "loss": 3.3689, + "step": 456 + }, + { + "epoch": 0.12418478260869566, + "grad_norm": 1.9951990842819214, + "learning_rate": 3e-05, + "loss": 3.0753, + "step": 457 + }, + { + "epoch": 0.12445652173913044, + "grad_norm": 2.348395347595215, + "learning_rate": 3e-05, + "loss": 3.8548, + "step": 458 + }, + { + "epoch": 0.12472826086956522, + "grad_norm": 2.324969530105591, + "learning_rate": 3e-05, + "loss": 3.3703, + "step": 459 + }, + { + "epoch": 0.125, + "grad_norm": 2.897873640060425, + "learning_rate": 3e-05, + "loss": 3.4589, + "step": 460 + }, + { + "epoch": 0.1252717391304348, + "grad_norm": 2.81925892829895, + "learning_rate": 3e-05, + "loss": 3.4818, + "step": 461 + }, + { + "epoch": 0.12554347826086956, + "grad_norm": 4.320977687835693, + "learning_rate": 3e-05, + "loss": 4.3366, + "step": 462 + }, + { + "epoch": 0.12581521739130436, + "grad_norm": 3.2348084449768066, + "learning_rate": 3e-05, + "loss": 4.1263, + "step": 463 + }, + { + "epoch": 0.12608695652173912, + "grad_norm": 2.6263833045959473, + "learning_rate": 3e-05, + "loss": 2.9206, + "step": 464 + }, + { + "epoch": 0.12635869565217392, + "grad_norm": 4.796882629394531, + "learning_rate": 3e-05, + "loss": 3.8013, + "step": 465 + }, + { + "epoch": 0.1266304347826087, + "grad_norm": 5.0804619789123535, + "learning_rate": 3e-05, + "loss": 3.8332, + "step": 466 + }, + { + "epoch": 0.12690217391304348, + "grad_norm": 3.0131075382232666, + "learning_rate": 3e-05, + "loss": 3.5667, + "step": 467 + }, + { + "epoch": 0.12717391304347825, + "grad_norm": 1.6071690320968628, + "learning_rate": 3e-05, + "loss": 2.8863, + "step": 468 + }, + { + "epoch": 0.12744565217391304, + "grad_norm": 2.211411476135254, + "learning_rate": 3e-05, + "loss": 3.2423, + "step": 469 + }, + { + "epoch": 0.12771739130434784, + "grad_norm": 2.4000439643859863, + "learning_rate": 3e-05, + "loss": 3.3725, + "step": 470 + }, + { + "epoch": 0.1279891304347826, + "grad_norm": 2.421668529510498, + "learning_rate": 3e-05, + "loss": 3.6009, + "step": 471 + }, + { + "epoch": 0.1282608695652174, + "grad_norm": 1.906244158744812, + "learning_rate": 3e-05, + "loss": 3.4958, + "step": 472 + }, + { + "epoch": 0.12853260869565217, + "grad_norm": 1.7820132970809937, + "learning_rate": 3e-05, + "loss": 3.0321, + "step": 473 + }, + { + "epoch": 0.12880434782608696, + "grad_norm": 2.929561138153076, + "learning_rate": 3e-05, + "loss": 3.788, + "step": 474 + }, + { + "epoch": 0.12907608695652173, + "grad_norm": 2.740144968032837, + "learning_rate": 3e-05, + "loss": 3.6582, + "step": 475 + }, + { + "epoch": 0.12934782608695652, + "grad_norm": 2.153079032897949, + "learning_rate": 3e-05, + "loss": 3.6739, + "step": 476 + }, + { + "epoch": 0.1296195652173913, + "grad_norm": 1.366892695426941, + "learning_rate": 3e-05, + "loss": 3.5163, + "step": 477 + }, + { + "epoch": 0.1298913043478261, + "grad_norm": 2.3881099224090576, + "learning_rate": 3e-05, + "loss": 3.7265, + "step": 478 + }, + { + "epoch": 0.13016304347826088, + "grad_norm": 1.72128164768219, + "learning_rate": 3e-05, + "loss": 2.9679, + "step": 479 + }, + { + "epoch": 0.13043478260869565, + "grad_norm": 1.8024574518203735, + "learning_rate": 3e-05, + "loss": 3.6759, + "step": 480 + }, + { + "epoch": 0.13070652173913044, + "grad_norm": 2.3349437713623047, + "learning_rate": 3e-05, + "loss": 3.9231, + "step": 481 + }, + { + "epoch": 0.1309782608695652, + "grad_norm": 2.60579776763916, + "learning_rate": 3e-05, + "loss": 3.5484, + "step": 482 + }, + { + "epoch": 0.13125, + "grad_norm": 1.823648452758789, + "learning_rate": 3e-05, + "loss": 3.409, + "step": 483 + }, + { + "epoch": 0.13152173913043477, + "grad_norm": 1.939772367477417, + "learning_rate": 3e-05, + "loss": 3.7185, + "step": 484 + }, + { + "epoch": 0.13179347826086957, + "grad_norm": 1.7059251070022583, + "learning_rate": 3e-05, + "loss": 3.2223, + "step": 485 + }, + { + "epoch": 0.13206521739130433, + "grad_norm": 2.1309432983398438, + "learning_rate": 3e-05, + "loss": 3.5143, + "step": 486 + }, + { + "epoch": 0.13233695652173913, + "grad_norm": 2.5169553756713867, + "learning_rate": 3e-05, + "loss": 3.9375, + "step": 487 + }, + { + "epoch": 0.13260869565217392, + "grad_norm": 1.6783947944641113, + "learning_rate": 3e-05, + "loss": 3.0069, + "step": 488 + }, + { + "epoch": 0.1328804347826087, + "grad_norm": 2.5426528453826904, + "learning_rate": 3e-05, + "loss": 3.4032, + "step": 489 + }, + { + "epoch": 0.1331521739130435, + "grad_norm": 2.8521511554718018, + "learning_rate": 3e-05, + "loss": 3.1236, + "step": 490 + }, + { + "epoch": 0.13342391304347825, + "grad_norm": 2.7477147579193115, + "learning_rate": 3e-05, + "loss": 3.0802, + "step": 491 + }, + { + "epoch": 0.13369565217391305, + "grad_norm": 2.785457134246826, + "learning_rate": 3e-05, + "loss": 3.8152, + "step": 492 + }, + { + "epoch": 0.13396739130434782, + "grad_norm": 3.6600492000579834, + "learning_rate": 3e-05, + "loss": 3.6564, + "step": 493 + }, + { + "epoch": 0.1342391304347826, + "grad_norm": 4.777431488037109, + "learning_rate": 3e-05, + "loss": 3.7449, + "step": 494 + }, + { + "epoch": 0.13451086956521738, + "grad_norm": 4.060612201690674, + "learning_rate": 3e-05, + "loss": 4.1282, + "step": 495 + }, + { + "epoch": 0.13478260869565217, + "grad_norm": 4.2701592445373535, + "learning_rate": 3e-05, + "loss": 4.6185, + "step": 496 + }, + { + "epoch": 0.13505434782608697, + "grad_norm": 3.9511125087738037, + "learning_rate": 3e-05, + "loss": 3.2463, + "step": 497 + }, + { + "epoch": 0.13532608695652174, + "grad_norm": 3.8475091457366943, + "learning_rate": 3e-05, + "loss": 3.0689, + "step": 498 + }, + { + "epoch": 0.13559782608695653, + "grad_norm": 5.655093669891357, + "learning_rate": 3e-05, + "loss": 4.024, + "step": 499 + }, + { + "epoch": 0.1358695652173913, + "grad_norm": 3.6835744380950928, + "learning_rate": 3e-05, + "loss": 3.1939, + "step": 500 + }, + { + "epoch": 0.1361413043478261, + "grad_norm": 3.2142796516418457, + "learning_rate": 3e-05, + "loss": 4.0483, + "step": 501 + }, + { + "epoch": 0.13641304347826086, + "grad_norm": 3.1059064865112305, + "learning_rate": 3e-05, + "loss": 3.8777, + "step": 502 + }, + { + "epoch": 0.13668478260869565, + "grad_norm": 2.3298041820526123, + "learning_rate": 3e-05, + "loss": 3.2728, + "step": 503 + }, + { + "epoch": 0.13695652173913042, + "grad_norm": 3.318361759185791, + "learning_rate": 3e-05, + "loss": 4.0421, + "step": 504 + }, + { + "epoch": 0.13722826086956522, + "grad_norm": 2.8279385566711426, + "learning_rate": 3e-05, + "loss": 3.49, + "step": 505 + }, + { + "epoch": 0.1375, + "grad_norm": 2.233304023742676, + "learning_rate": 3e-05, + "loss": 3.3737, + "step": 506 + }, + { + "epoch": 0.13777173913043478, + "grad_norm": 2.943725824356079, + "learning_rate": 3e-05, + "loss": 3.5774, + "step": 507 + }, + { + "epoch": 0.13804347826086957, + "grad_norm": 2.764371871948242, + "learning_rate": 3e-05, + "loss": 3.7041, + "step": 508 + }, + { + "epoch": 0.13831521739130434, + "grad_norm": 2.647728443145752, + "learning_rate": 3e-05, + "loss": 3.0545, + "step": 509 + }, + { + "epoch": 0.13858695652173914, + "grad_norm": 3.1184072494506836, + "learning_rate": 3e-05, + "loss": 3.2394, + "step": 510 + }, + { + "epoch": 0.1388586956521739, + "grad_norm": 1.9611910581588745, + "learning_rate": 3e-05, + "loss": 3.4171, + "step": 511 + }, + { + "epoch": 0.1391304347826087, + "grad_norm": 1.5122188329696655, + "learning_rate": 3e-05, + "loss": 3.1593, + "step": 512 + }, + { + "epoch": 0.13940217391304346, + "grad_norm": 1.9166432619094849, + "learning_rate": 3e-05, + "loss": 3.21, + "step": 513 + }, + { + "epoch": 0.13967391304347826, + "grad_norm": 2.1466691493988037, + "learning_rate": 3e-05, + "loss": 3.4333, + "step": 514 + }, + { + "epoch": 0.13994565217391305, + "grad_norm": 2.217186689376831, + "learning_rate": 3e-05, + "loss": 3.3885, + "step": 515 + }, + { + "epoch": 0.14021739130434782, + "grad_norm": 2.03560733795166, + "learning_rate": 3e-05, + "loss": 3.9298, + "step": 516 + }, + { + "epoch": 0.14048913043478262, + "grad_norm": 2.214279890060425, + "learning_rate": 3e-05, + "loss": 3.2791, + "step": 517 + }, + { + "epoch": 0.14076086956521738, + "grad_norm": 3.428405523300171, + "learning_rate": 3e-05, + "loss": 3.668, + "step": 518 + }, + { + "epoch": 0.14103260869565218, + "grad_norm": 3.077150344848633, + "learning_rate": 3e-05, + "loss": 3.7187, + "step": 519 + }, + { + "epoch": 0.14130434782608695, + "grad_norm": 1.3983957767486572, + "learning_rate": 3e-05, + "loss": 3.3129, + "step": 520 + }, + { + "epoch": 0.14157608695652174, + "grad_norm": 2.0241615772247314, + "learning_rate": 3e-05, + "loss": 3.6655, + "step": 521 + }, + { + "epoch": 0.14184782608695654, + "grad_norm": 2.20249080657959, + "learning_rate": 3e-05, + "loss": 3.2375, + "step": 522 + }, + { + "epoch": 0.1421195652173913, + "grad_norm": 2.2980117797851562, + "learning_rate": 3e-05, + "loss": 3.7075, + "step": 523 + }, + { + "epoch": 0.1423913043478261, + "grad_norm": 1.2177653312683105, + "learning_rate": 3e-05, + "loss": 3.2919, + "step": 524 + }, + { + "epoch": 0.14266304347826086, + "grad_norm": 1.0807304382324219, + "learning_rate": 3e-05, + "loss": 3.2761, + "step": 525 + }, + { + "epoch": 0.14293478260869566, + "grad_norm": 1.4918198585510254, + "learning_rate": 3e-05, + "loss": 3.6412, + "step": 526 + }, + { + "epoch": 0.14320652173913043, + "grad_norm": 1.3037655353546143, + "learning_rate": 3e-05, + "loss": 3.4596, + "step": 527 + }, + { + "epoch": 0.14347826086956522, + "grad_norm": 1.1929515600204468, + "learning_rate": 3e-05, + "loss": 3.4554, + "step": 528 + }, + { + "epoch": 0.14375, + "grad_norm": 1.7983393669128418, + "learning_rate": 3e-05, + "loss": 3.0941, + "step": 529 + }, + { + "epoch": 0.14402173913043478, + "grad_norm": 1.4893770217895508, + "learning_rate": 3e-05, + "loss": 3.688, + "step": 530 + }, + { + "epoch": 0.14429347826086958, + "grad_norm": 1.4558385610580444, + "learning_rate": 3e-05, + "loss": 3.6866, + "step": 531 + }, + { + "epoch": 0.14456521739130435, + "grad_norm": 1.4489972591400146, + "learning_rate": 3e-05, + "loss": 3.1238, + "step": 532 + }, + { + "epoch": 0.14483695652173914, + "grad_norm": 1.603597640991211, + "learning_rate": 3e-05, + "loss": 2.9605, + "step": 533 + }, + { + "epoch": 0.1451086956521739, + "grad_norm": 1.1347497701644897, + "learning_rate": 3e-05, + "loss": 3.2066, + "step": 534 + }, + { + "epoch": 0.1453804347826087, + "grad_norm": 1.9727925062179565, + "learning_rate": 3e-05, + "loss": 3.6686, + "step": 535 + }, + { + "epoch": 0.14565217391304347, + "grad_norm": 2.5670254230499268, + "learning_rate": 3e-05, + "loss": 3.3345, + "step": 536 + }, + { + "epoch": 0.14592391304347826, + "grad_norm": 2.017517328262329, + "learning_rate": 3e-05, + "loss": 3.7092, + "step": 537 + }, + { + "epoch": 0.14619565217391303, + "grad_norm": 2.1940996646881104, + "learning_rate": 3e-05, + "loss": 3.807, + "step": 538 + }, + { + "epoch": 0.14646739130434783, + "grad_norm": 1.7565933465957642, + "learning_rate": 3e-05, + "loss": 3.6645, + "step": 539 + }, + { + "epoch": 0.14673913043478262, + "grad_norm": 2.248276948928833, + "learning_rate": 3e-05, + "loss": 3.979, + "step": 540 + }, + { + "epoch": 0.1470108695652174, + "grad_norm": 2.3259811401367188, + "learning_rate": 3e-05, + "loss": 3.3435, + "step": 541 + }, + { + "epoch": 0.14728260869565218, + "grad_norm": 2.7300784587860107, + "learning_rate": 3e-05, + "loss": 3.7696, + "step": 542 + }, + { + "epoch": 0.14755434782608695, + "grad_norm": 1.406214952468872, + "learning_rate": 3e-05, + "loss": 2.6787, + "step": 543 + }, + { + "epoch": 0.14782608695652175, + "grad_norm": 2.6304099559783936, + "learning_rate": 3e-05, + "loss": 3.7273, + "step": 544 + }, + { + "epoch": 0.1480978260869565, + "grad_norm": 4.7364912033081055, + "learning_rate": 3e-05, + "loss": 4.4759, + "step": 545 + }, + { + "epoch": 0.1483695652173913, + "grad_norm": 5.363766193389893, + "learning_rate": 3e-05, + "loss": 4.1787, + "step": 546 + }, + { + "epoch": 0.14864130434782608, + "grad_norm": 2.660940408706665, + "learning_rate": 3e-05, + "loss": 3.2534, + "step": 547 + }, + { + "epoch": 0.14891304347826087, + "grad_norm": 3.186602830886841, + "learning_rate": 3e-05, + "loss": 3.5762, + "step": 548 + }, + { + "epoch": 0.14918478260869567, + "grad_norm": 3.962216377258301, + "learning_rate": 3e-05, + "loss": 3.5002, + "step": 549 + }, + { + "epoch": 0.14945652173913043, + "grad_norm": 5.419135093688965, + "learning_rate": 3e-05, + "loss": 3.6202, + "step": 550 + }, + { + "epoch": 0.14972826086956523, + "grad_norm": 4.870014667510986, + "learning_rate": 3e-05, + "loss": 3.5467, + "step": 551 + }, + { + "epoch": 0.15, + "grad_norm": 3.175389289855957, + "learning_rate": 3e-05, + "loss": 3.9385, + "step": 552 + }, + { + "epoch": 0.1502717391304348, + "grad_norm": 2.783637046813965, + "learning_rate": 3e-05, + "loss": 3.8554, + "step": 553 + }, + { + "epoch": 0.15054347826086956, + "grad_norm": 2.6941978931427, + "learning_rate": 3e-05, + "loss": 3.211, + "step": 554 + }, + { + "epoch": 0.15081521739130435, + "grad_norm": 3.185209274291992, + "learning_rate": 3e-05, + "loss": 3.3476, + "step": 555 + }, + { + "epoch": 0.15108695652173912, + "grad_norm": 2.922362804412842, + "learning_rate": 3e-05, + "loss": 3.0941, + "step": 556 + }, + { + "epoch": 0.1513586956521739, + "grad_norm": 2.7833430767059326, + "learning_rate": 3e-05, + "loss": 3.3851, + "step": 557 + }, + { + "epoch": 0.1516304347826087, + "grad_norm": 1.4858049154281616, + "learning_rate": 3e-05, + "loss": 2.7326, + "step": 558 + }, + { + "epoch": 0.15190217391304348, + "grad_norm": 2.9602174758911133, + "learning_rate": 3e-05, + "loss": 3.2037, + "step": 559 + }, + { + "epoch": 0.15217391304347827, + "grad_norm": 3.0462303161621094, + "learning_rate": 3e-05, + "loss": 3.2449, + "step": 560 + }, + { + "epoch": 0.15244565217391304, + "grad_norm": 3.038954019546509, + "learning_rate": 3e-05, + "loss": 3.3796, + "step": 561 + }, + { + "epoch": 0.15271739130434783, + "grad_norm": 1.572611689567566, + "learning_rate": 3e-05, + "loss": 3.1521, + "step": 562 + }, + { + "epoch": 0.1529891304347826, + "grad_norm": 2.4640631675720215, + "learning_rate": 3e-05, + "loss": 3.9771, + "step": 563 + }, + { + "epoch": 0.1532608695652174, + "grad_norm": 1.757712483406067, + "learning_rate": 3e-05, + "loss": 2.9893, + "step": 564 + }, + { + "epoch": 0.15353260869565216, + "grad_norm": 2.606873035430908, + "learning_rate": 3e-05, + "loss": 3.702, + "step": 565 + }, + { + "epoch": 0.15380434782608696, + "grad_norm": 1.677304983139038, + "learning_rate": 3e-05, + "loss": 3.2607, + "step": 566 + }, + { + "epoch": 0.15407608695652175, + "grad_norm": 2.4729526042938232, + "learning_rate": 3e-05, + "loss": 3.9865, + "step": 567 + }, + { + "epoch": 0.15434782608695652, + "grad_norm": 2.162191867828369, + "learning_rate": 3e-05, + "loss": 3.6349, + "step": 568 + }, + { + "epoch": 0.1546195652173913, + "grad_norm": 3.027984380722046, + "learning_rate": 3e-05, + "loss": 3.8996, + "step": 569 + }, + { + "epoch": 0.15489130434782608, + "grad_norm": 1.9820539951324463, + "learning_rate": 3e-05, + "loss": 3.3207, + "step": 570 + }, + { + "epoch": 0.15516304347826088, + "grad_norm": 1.7880254983901978, + "learning_rate": 3e-05, + "loss": 3.6082, + "step": 571 + }, + { + "epoch": 0.15543478260869564, + "grad_norm": 1.9113103151321411, + "learning_rate": 3e-05, + "loss": 3.2639, + "step": 572 + }, + { + "epoch": 0.15570652173913044, + "grad_norm": 1.7418012619018555, + "learning_rate": 3e-05, + "loss": 3.4052, + "step": 573 + }, + { + "epoch": 0.1559782608695652, + "grad_norm": 1.6183768510818481, + "learning_rate": 3e-05, + "loss": 3.8061, + "step": 574 + }, + { + "epoch": 0.15625, + "grad_norm": 1.33986496925354, + "learning_rate": 3e-05, + "loss": 3.3393, + "step": 575 + }, + { + "epoch": 0.1565217391304348, + "grad_norm": 1.4486945867538452, + "learning_rate": 3e-05, + "loss": 3.7729, + "step": 576 + }, + { + "epoch": 0.15679347826086956, + "grad_norm": 1.7886120080947876, + "learning_rate": 3e-05, + "loss": 3.8346, + "step": 577 + }, + { + "epoch": 0.15706521739130436, + "grad_norm": 1.2352441549301147, + "learning_rate": 3e-05, + "loss": 3.2718, + "step": 578 + }, + { + "epoch": 0.15733695652173912, + "grad_norm": 1.8679944276809692, + "learning_rate": 3e-05, + "loss": 3.4247, + "step": 579 + }, + { + "epoch": 0.15760869565217392, + "grad_norm": 1.689501404762268, + "learning_rate": 3e-05, + "loss": 4.0971, + "step": 580 + }, + { + "epoch": 0.1578804347826087, + "grad_norm": 1.6779184341430664, + "learning_rate": 3e-05, + "loss": 3.8803, + "step": 581 + }, + { + "epoch": 0.15815217391304348, + "grad_norm": 2.0760598182678223, + "learning_rate": 3e-05, + "loss": 3.5429, + "step": 582 + }, + { + "epoch": 0.15842391304347825, + "grad_norm": 1.8308002948760986, + "learning_rate": 3e-05, + "loss": 3.4797, + "step": 583 + }, + { + "epoch": 0.15869565217391304, + "grad_norm": 2.4416396617889404, + "learning_rate": 3e-05, + "loss": 4.2491, + "step": 584 + }, + { + "epoch": 0.15896739130434784, + "grad_norm": 2.365509271621704, + "learning_rate": 3e-05, + "loss": 3.4862, + "step": 585 + }, + { + "epoch": 0.1592391304347826, + "grad_norm": 4.06919002532959, + "learning_rate": 3e-05, + "loss": 4.0302, + "step": 586 + }, + { + "epoch": 0.1595108695652174, + "grad_norm": 2.430508613586426, + "learning_rate": 3e-05, + "loss": 3.8575, + "step": 587 + }, + { + "epoch": 0.15978260869565217, + "grad_norm": 2.2944397926330566, + "learning_rate": 3e-05, + "loss": 4.3022, + "step": 588 + }, + { + "epoch": 0.16005434782608696, + "grad_norm": 4.096346855163574, + "learning_rate": 3e-05, + "loss": 4.0663, + "step": 589 + }, + { + "epoch": 0.16032608695652173, + "grad_norm": 3.8741347789764404, + "learning_rate": 3e-05, + "loss": 3.3398, + "step": 590 + }, + { + "epoch": 0.16059782608695652, + "grad_norm": 3.7039456367492676, + "learning_rate": 3e-05, + "loss": 3.1813, + "step": 591 + }, + { + "epoch": 0.1608695652173913, + "grad_norm": 2.129674196243286, + "learning_rate": 3e-05, + "loss": 3.0778, + "step": 592 + }, + { + "epoch": 0.1611413043478261, + "grad_norm": 2.404423952102661, + "learning_rate": 3e-05, + "loss": 3.7275, + "step": 593 + }, + { + "epoch": 0.16141304347826088, + "grad_norm": 2.341529607772827, + "learning_rate": 3e-05, + "loss": 2.9763, + "step": 594 + }, + { + "epoch": 0.16168478260869565, + "grad_norm": 4.927234649658203, + "learning_rate": 3e-05, + "loss": 4.1287, + "step": 595 + }, + { + "epoch": 0.16195652173913044, + "grad_norm": 4.529046535491943, + "learning_rate": 3e-05, + "loss": 3.7712, + "step": 596 + }, + { + "epoch": 0.1622282608695652, + "grad_norm": 4.280128479003906, + "learning_rate": 3e-05, + "loss": 3.7557, + "step": 597 + }, + { + "epoch": 0.1625, + "grad_norm": 2.845654249191284, + "learning_rate": 3e-05, + "loss": 3.4817, + "step": 598 + }, + { + "epoch": 0.16277173913043477, + "grad_norm": 2.5575075149536133, + "learning_rate": 3e-05, + "loss": 3.0366, + "step": 599 + }, + { + "epoch": 0.16304347826086957, + "grad_norm": 3.475111722946167, + "learning_rate": 3e-05, + "loss": 3.2081, + "step": 600 + }, + { + "epoch": 0.16331521739130433, + "grad_norm": 3.926103353500366, + "learning_rate": 3e-05, + "loss": 2.8455, + "step": 601 + }, + { + "epoch": 0.16358695652173913, + "grad_norm": 3.796124219894409, + "learning_rate": 3e-05, + "loss": 3.4857, + "step": 602 + }, + { + "epoch": 0.16385869565217392, + "grad_norm": 2.974625825881958, + "learning_rate": 3e-05, + "loss": 3.3609, + "step": 603 + }, + { + "epoch": 0.1641304347826087, + "grad_norm": 1.9188250303268433, + "learning_rate": 3e-05, + "loss": 3.0467, + "step": 604 + }, + { + "epoch": 0.1644021739130435, + "grad_norm": 2.922881841659546, + "learning_rate": 3e-05, + "loss": 3.969, + "step": 605 + }, + { + "epoch": 0.16467391304347825, + "grad_norm": 2.9243340492248535, + "learning_rate": 3e-05, + "loss": 3.418, + "step": 606 + }, + { + "epoch": 0.16494565217391305, + "grad_norm": 2.3607518672943115, + "learning_rate": 3e-05, + "loss": 3.3568, + "step": 607 + }, + { + "epoch": 0.16521739130434782, + "grad_norm": 1.4109339714050293, + "learning_rate": 3e-05, + "loss": 3.1942, + "step": 608 + }, + { + "epoch": 0.1654891304347826, + "grad_norm": 2.339900016784668, + "learning_rate": 3e-05, + "loss": 3.6621, + "step": 609 + }, + { + "epoch": 0.16576086956521738, + "grad_norm": 3.0392491817474365, + "learning_rate": 3e-05, + "loss": 3.6538, + "step": 610 + }, + { + "epoch": 0.16603260869565217, + "grad_norm": 2.5297353267669678, + "learning_rate": 3e-05, + "loss": 3.72, + "step": 611 + }, + { + "epoch": 0.16630434782608697, + "grad_norm": 2.1570205688476562, + "learning_rate": 3e-05, + "loss": 3.1765, + "step": 612 + }, + { + "epoch": 0.16657608695652174, + "grad_norm": 1.5125519037246704, + "learning_rate": 3e-05, + "loss": 3.5076, + "step": 613 + }, + { + "epoch": 0.16684782608695653, + "grad_norm": 1.5017513036727905, + "learning_rate": 3e-05, + "loss": 3.5962, + "step": 614 + }, + { + "epoch": 0.1671195652173913, + "grad_norm": 2.1280760765075684, + "learning_rate": 3e-05, + "loss": 3.4612, + "step": 615 + }, + { + "epoch": 0.1673913043478261, + "grad_norm": 1.6511037349700928, + "learning_rate": 3e-05, + "loss": 3.3281, + "step": 616 + }, + { + "epoch": 0.16766304347826086, + "grad_norm": 1.842786431312561, + "learning_rate": 3e-05, + "loss": 2.876, + "step": 617 + }, + { + "epoch": 0.16793478260869565, + "grad_norm": 2.008168935775757, + "learning_rate": 3e-05, + "loss": 3.7421, + "step": 618 + }, + { + "epoch": 0.16820652173913042, + "grad_norm": 2.965365409851074, + "learning_rate": 3e-05, + "loss": 3.9864, + "step": 619 + }, + { + "epoch": 0.16847826086956522, + "grad_norm": 3.619001626968384, + "learning_rate": 3e-05, + "loss": 4.1181, + "step": 620 + }, + { + "epoch": 0.16875, + "grad_norm": 3.3966426849365234, + "learning_rate": 3e-05, + "loss": 3.2355, + "step": 621 + }, + { + "epoch": 0.16902173913043478, + "grad_norm": 2.4561333656311035, + "learning_rate": 3e-05, + "loss": 3.0077, + "step": 622 + }, + { + "epoch": 0.16929347826086957, + "grad_norm": 3.7762365341186523, + "learning_rate": 3e-05, + "loss": 3.8245, + "step": 623 + }, + { + "epoch": 0.16956521739130434, + "grad_norm": 4.286952495574951, + "learning_rate": 3e-05, + "loss": 3.4895, + "step": 624 + }, + { + "epoch": 0.16983695652173914, + "grad_norm": 3.7213828563690186, + "learning_rate": 3e-05, + "loss": 2.9799, + "step": 625 + }, + { + "epoch": 0.1701086956521739, + "grad_norm": 3.109027862548828, + "learning_rate": 3e-05, + "loss": 3.8767, + "step": 626 + }, + { + "epoch": 0.1703804347826087, + "grad_norm": 2.4418489933013916, + "learning_rate": 3e-05, + "loss": 3.7539, + "step": 627 + }, + { + "epoch": 0.17065217391304346, + "grad_norm": 2.3365840911865234, + "learning_rate": 3e-05, + "loss": 3.0382, + "step": 628 + }, + { + "epoch": 0.17092391304347826, + "grad_norm": 2.649116039276123, + "learning_rate": 3e-05, + "loss": 3.0011, + "step": 629 + }, + { + "epoch": 0.17119565217391305, + "grad_norm": 4.0966105461120605, + "learning_rate": 3e-05, + "loss": 3.5889, + "step": 630 + }, + { + "epoch": 0.17146739130434782, + "grad_norm": 4.70590877532959, + "learning_rate": 3e-05, + "loss": 3.4292, + "step": 631 + }, + { + "epoch": 0.17173913043478262, + "grad_norm": 4.246743202209473, + "learning_rate": 3e-05, + "loss": 3.3927, + "step": 632 + }, + { + "epoch": 0.17201086956521738, + "grad_norm": 3.4085919857025146, + "learning_rate": 3e-05, + "loss": 3.5786, + "step": 633 + }, + { + "epoch": 0.17228260869565218, + "grad_norm": 3.164703607559204, + "learning_rate": 3e-05, + "loss": 3.4236, + "step": 634 + }, + { + "epoch": 0.17255434782608695, + "grad_norm": 4.914327144622803, + "learning_rate": 3e-05, + "loss": 4.2673, + "step": 635 + }, + { + "epoch": 0.17282608695652174, + "grad_norm": 4.016328811645508, + "learning_rate": 3e-05, + "loss": 3.5927, + "step": 636 + }, + { + "epoch": 0.17309782608695654, + "grad_norm": 3.3188321590423584, + "learning_rate": 3e-05, + "loss": 2.9687, + "step": 637 + }, + { + "epoch": 0.1733695652173913, + "grad_norm": 3.4671499729156494, + "learning_rate": 3e-05, + "loss": 3.3144, + "step": 638 + }, + { + "epoch": 0.1736413043478261, + "grad_norm": 2.273728609085083, + "learning_rate": 3e-05, + "loss": 3.2453, + "step": 639 + }, + { + "epoch": 0.17391304347826086, + "grad_norm": 2.1210081577301025, + "learning_rate": 3e-05, + "loss": 3.0778, + "step": 640 + }, + { + "epoch": 0.17418478260869566, + "grad_norm": 3.6642115116119385, + "learning_rate": 3e-05, + "loss": 3.7503, + "step": 641 + }, + { + "epoch": 0.17445652173913043, + "grad_norm": 3.753002166748047, + "learning_rate": 3e-05, + "loss": 3.9504, + "step": 642 + }, + { + "epoch": 0.17472826086956522, + "grad_norm": 2.683023691177368, + "learning_rate": 3e-05, + "loss": 3.5103, + "step": 643 + }, + { + "epoch": 0.175, + "grad_norm": 2.7764129638671875, + "learning_rate": 3e-05, + "loss": 3.8473, + "step": 644 + }, + { + "epoch": 0.17527173913043478, + "grad_norm": 2.340895652770996, + "learning_rate": 3e-05, + "loss": 3.3381, + "step": 645 + }, + { + "epoch": 0.17554347826086958, + "grad_norm": 3.224985122680664, + "learning_rate": 3e-05, + "loss": 3.7218, + "step": 646 + }, + { + "epoch": 0.17581521739130435, + "grad_norm": 4.798543930053711, + "learning_rate": 3e-05, + "loss": 4.2621, + "step": 647 + }, + { + "epoch": 0.17608695652173914, + "grad_norm": 3.2494759559631348, + "learning_rate": 3e-05, + "loss": 3.4175, + "step": 648 + }, + { + "epoch": 0.1763586956521739, + "grad_norm": 1.9543368816375732, + "learning_rate": 3e-05, + "loss": 2.9133, + "step": 649 + }, + { + "epoch": 0.1766304347826087, + "grad_norm": 1.7534743547439575, + "learning_rate": 3e-05, + "loss": 3.2838, + "step": 650 + }, + { + "epoch": 0.17690217391304347, + "grad_norm": 1.8143278360366821, + "learning_rate": 3e-05, + "loss": 3.015, + "step": 651 + }, + { + "epoch": 0.17717391304347826, + "grad_norm": 3.4449610710144043, + "learning_rate": 3e-05, + "loss": 3.7868, + "step": 652 + }, + { + "epoch": 0.17744565217391303, + "grad_norm": 2.250290632247925, + "learning_rate": 3e-05, + "loss": 3.6572, + "step": 653 + }, + { + "epoch": 0.17771739130434783, + "grad_norm": 2.3306338787078857, + "learning_rate": 3e-05, + "loss": 3.6463, + "step": 654 + }, + { + "epoch": 0.17798913043478262, + "grad_norm": 1.6108341217041016, + "learning_rate": 3e-05, + "loss": 3.3977, + "step": 655 + }, + { + "epoch": 0.1782608695652174, + "grad_norm": 1.7349133491516113, + "learning_rate": 3e-05, + "loss": 3.5826, + "step": 656 + }, + { + "epoch": 0.17853260869565218, + "grad_norm": 2.5024263858795166, + "learning_rate": 3e-05, + "loss": 4.2271, + "step": 657 + }, + { + "epoch": 0.17880434782608695, + "grad_norm": 2.2973721027374268, + "learning_rate": 3e-05, + "loss": 3.8238, + "step": 658 + }, + { + "epoch": 0.17907608695652175, + "grad_norm": 2.167056083679199, + "learning_rate": 3e-05, + "loss": 3.5719, + "step": 659 + }, + { + "epoch": 0.1793478260869565, + "grad_norm": 1.8604989051818848, + "learning_rate": 3e-05, + "loss": 3.9091, + "step": 660 + }, + { + "epoch": 0.1796195652173913, + "grad_norm": 1.7601895332336426, + "learning_rate": 3e-05, + "loss": 3.8582, + "step": 661 + }, + { + "epoch": 0.17989130434782608, + "grad_norm": 1.5462661981582642, + "learning_rate": 3e-05, + "loss": 2.9604, + "step": 662 + }, + { + "epoch": 0.18016304347826087, + "grad_norm": 1.9906418323516846, + "learning_rate": 3e-05, + "loss": 3.8877, + "step": 663 + }, + { + "epoch": 0.18043478260869567, + "grad_norm": 2.1721136569976807, + "learning_rate": 3e-05, + "loss": 4.241, + "step": 664 + }, + { + "epoch": 0.18070652173913043, + "grad_norm": 1.7870231866836548, + "learning_rate": 3e-05, + "loss": 3.0484, + "step": 665 + }, + { + "epoch": 0.18097826086956523, + "grad_norm": 1.543925166130066, + "learning_rate": 3e-05, + "loss": 3.1115, + "step": 666 + }, + { + "epoch": 0.18125, + "grad_norm": 1.9159563779830933, + "learning_rate": 3e-05, + "loss": 3.3634, + "step": 667 + }, + { + "epoch": 0.1815217391304348, + "grad_norm": 2.8006904125213623, + "learning_rate": 3e-05, + "loss": 3.7725, + "step": 668 + }, + { + "epoch": 0.18179347826086956, + "grad_norm": 2.8094773292541504, + "learning_rate": 3e-05, + "loss": 3.5421, + "step": 669 + }, + { + "epoch": 0.18206521739130435, + "grad_norm": 2.0162885189056396, + "learning_rate": 3e-05, + "loss": 3.4805, + "step": 670 + }, + { + "epoch": 0.18233695652173912, + "grad_norm": 2.175633430480957, + "learning_rate": 3e-05, + "loss": 3.3937, + "step": 671 + }, + { + "epoch": 0.1826086956521739, + "grad_norm": 2.4267466068267822, + "learning_rate": 3e-05, + "loss": 3.1429, + "step": 672 + }, + { + "epoch": 0.1828804347826087, + "grad_norm": 4.546562194824219, + "learning_rate": 3e-05, + "loss": 4.1168, + "step": 673 + }, + { + "epoch": 0.18315217391304348, + "grad_norm": 3.3911025524139404, + "learning_rate": 3e-05, + "loss": 3.7966, + "step": 674 + }, + { + "epoch": 0.18342391304347827, + "grad_norm": 1.7894442081451416, + "learning_rate": 3e-05, + "loss": 3.8492, + "step": 675 + }, + { + "epoch": 0.18369565217391304, + "grad_norm": 2.7028391361236572, + "learning_rate": 3e-05, + "loss": 3.5555, + "step": 676 + }, + { + "epoch": 0.18396739130434783, + "grad_norm": 3.052639961242676, + "learning_rate": 3e-05, + "loss": 3.9254, + "step": 677 + }, + { + "epoch": 0.1842391304347826, + "grad_norm": 2.4930381774902344, + "learning_rate": 3e-05, + "loss": 3.4053, + "step": 678 + }, + { + "epoch": 0.1845108695652174, + "grad_norm": 1.9123059511184692, + "learning_rate": 3e-05, + "loss": 3.5101, + "step": 679 + }, + { + "epoch": 0.18478260869565216, + "grad_norm": 1.5995222330093384, + "learning_rate": 3e-05, + "loss": 3.1762, + "step": 680 + }, + { + "epoch": 0.18505434782608696, + "grad_norm": 4.272251129150391, + "learning_rate": 3e-05, + "loss": 4.397, + "step": 681 + }, + { + "epoch": 0.18532608695652175, + "grad_norm": 3.0316965579986572, + "learning_rate": 3e-05, + "loss": 3.8721, + "step": 682 + }, + { + "epoch": 0.18559782608695652, + "grad_norm": 2.0125246047973633, + "learning_rate": 3e-05, + "loss": 3.3782, + "step": 683 + }, + { + "epoch": 0.1858695652173913, + "grad_norm": 1.8079962730407715, + "learning_rate": 3e-05, + "loss": 3.4851, + "step": 684 + }, + { + "epoch": 0.18614130434782608, + "grad_norm": 1.8277610540390015, + "learning_rate": 3e-05, + "loss": 2.8756, + "step": 685 + }, + { + "epoch": 0.18641304347826088, + "grad_norm": 2.4962170124053955, + "learning_rate": 3e-05, + "loss": 4.2237, + "step": 686 + }, + { + "epoch": 0.18668478260869564, + "grad_norm": 1.3268901109695435, + "learning_rate": 3e-05, + "loss": 3.1177, + "step": 687 + }, + { + "epoch": 0.18695652173913044, + "grad_norm": 2.3978965282440186, + "learning_rate": 3e-05, + "loss": 3.583, + "step": 688 + }, + { + "epoch": 0.1872282608695652, + "grad_norm": 1.2693482637405396, + "learning_rate": 3e-05, + "loss": 3.3973, + "step": 689 + }, + { + "epoch": 0.1875, + "grad_norm": 2.0299911499023438, + "learning_rate": 3e-05, + "loss": 3.6627, + "step": 690 + }, + { + "epoch": 0.1877717391304348, + "grad_norm": 1.2860682010650635, + "learning_rate": 3e-05, + "loss": 3.2345, + "step": 691 + }, + { + "epoch": 0.18804347826086956, + "grad_norm": 2.264256477355957, + "learning_rate": 3e-05, + "loss": 4.2411, + "step": 692 + }, + { + "epoch": 0.18831521739130436, + "grad_norm": 1.5267597436904907, + "learning_rate": 3e-05, + "loss": 3.1942, + "step": 693 + }, + { + "epoch": 0.18858695652173912, + "grad_norm": 1.2684136629104614, + "learning_rate": 3e-05, + "loss": 3.192, + "step": 694 + }, + { + "epoch": 0.18885869565217392, + "grad_norm": 1.6705724000930786, + "learning_rate": 3e-05, + "loss": 3.6082, + "step": 695 + }, + { + "epoch": 0.1891304347826087, + "grad_norm": 2.0569725036621094, + "learning_rate": 3e-05, + "loss": 3.9019, + "step": 696 + }, + { + "epoch": 0.18940217391304348, + "grad_norm": 2.03778076171875, + "learning_rate": 3e-05, + "loss": 4.0118, + "step": 697 + }, + { + "epoch": 0.18967391304347825, + "grad_norm": 1.7906421422958374, + "learning_rate": 3e-05, + "loss": 3.7659, + "step": 698 + }, + { + "epoch": 0.18994565217391304, + "grad_norm": 1.6202226877212524, + "learning_rate": 3e-05, + "loss": 3.3228, + "step": 699 + }, + { + "epoch": 0.19021739130434784, + "grad_norm": 1.568548560142517, + "learning_rate": 3e-05, + "loss": 3.3992, + "step": 700 + }, + { + "epoch": 0.1904891304347826, + "grad_norm": 1.3486888408660889, + "learning_rate": 3e-05, + "loss": 2.7886, + "step": 701 + }, + { + "epoch": 0.1907608695652174, + "grad_norm": 2.013223171234131, + "learning_rate": 3e-05, + "loss": 3.365, + "step": 702 + }, + { + "epoch": 0.19103260869565217, + "grad_norm": 1.7130358219146729, + "learning_rate": 3e-05, + "loss": 3.7516, + "step": 703 + }, + { + "epoch": 0.19130434782608696, + "grad_norm": 1.7324903011322021, + "learning_rate": 3e-05, + "loss": 3.1275, + "step": 704 + }, + { + "epoch": 0.19157608695652173, + "grad_norm": 1.7227182388305664, + "learning_rate": 3e-05, + "loss": 3.1415, + "step": 705 + }, + { + "epoch": 0.19184782608695652, + "grad_norm": 1.3472079038619995, + "learning_rate": 3e-05, + "loss": 3.1949, + "step": 706 + }, + { + "epoch": 0.1921195652173913, + "grad_norm": 1.1573243141174316, + "learning_rate": 3e-05, + "loss": 2.8314, + "step": 707 + }, + { + "epoch": 0.1923913043478261, + "grad_norm": 1.6137231588363647, + "learning_rate": 3e-05, + "loss": 3.6169, + "step": 708 + }, + { + "epoch": 0.19266304347826088, + "grad_norm": 1.4710439443588257, + "learning_rate": 3e-05, + "loss": 3.4097, + "step": 709 + }, + { + "epoch": 0.19293478260869565, + "grad_norm": 2.347607374191284, + "learning_rate": 3e-05, + "loss": 4.1638, + "step": 710 + }, + { + "epoch": 0.19320652173913044, + "grad_norm": 1.4299064874649048, + "learning_rate": 3e-05, + "loss": 3.811, + "step": 711 + }, + { + "epoch": 0.1934782608695652, + "grad_norm": 1.5702636241912842, + "learning_rate": 3e-05, + "loss": 3.5396, + "step": 712 + }, + { + "epoch": 0.19375, + "grad_norm": 1.8417882919311523, + "learning_rate": 3e-05, + "loss": 3.5613, + "step": 713 + }, + { + "epoch": 0.19402173913043477, + "grad_norm": 1.4125804901123047, + "learning_rate": 3e-05, + "loss": 3.1808, + "step": 714 + }, + { + "epoch": 0.19429347826086957, + "grad_norm": 1.3733165264129639, + "learning_rate": 3e-05, + "loss": 3.3127, + "step": 715 + }, + { + "epoch": 0.19456521739130433, + "grad_norm": 2.1651833057403564, + "learning_rate": 3e-05, + "loss": 4.1732, + "step": 716 + }, + { + "epoch": 0.19483695652173913, + "grad_norm": 1.4096370935440063, + "learning_rate": 3e-05, + "loss": 3.5565, + "step": 717 + }, + { + "epoch": 0.19510869565217392, + "grad_norm": 1.7251945734024048, + "learning_rate": 3e-05, + "loss": 3.7342, + "step": 718 + }, + { + "epoch": 0.1953804347826087, + "grad_norm": 1.643470287322998, + "learning_rate": 3e-05, + "loss": 2.9678, + "step": 719 + }, + { + "epoch": 0.1956521739130435, + "grad_norm": 1.5827432870864868, + "learning_rate": 3e-05, + "loss": 3.1044, + "step": 720 + }, + { + "epoch": 0.19592391304347825, + "grad_norm": 1.4647548198699951, + "learning_rate": 3e-05, + "loss": 2.8413, + "step": 721 + }, + { + "epoch": 0.19619565217391305, + "grad_norm": 2.1561598777770996, + "learning_rate": 3e-05, + "loss": 3.4492, + "step": 722 + }, + { + "epoch": 0.19646739130434782, + "grad_norm": 1.6970568895339966, + "learning_rate": 3e-05, + "loss": 3.1848, + "step": 723 + }, + { + "epoch": 0.1967391304347826, + "grad_norm": 2.248117685317993, + "learning_rate": 3e-05, + "loss": 3.6296, + "step": 724 + }, + { + "epoch": 0.19701086956521738, + "grad_norm": 3.168339252471924, + "learning_rate": 3e-05, + "loss": 4.291, + "step": 725 + }, + { + "epoch": 0.19728260869565217, + "grad_norm": 3.802591323852539, + "learning_rate": 3e-05, + "loss": 3.9495, + "step": 726 + }, + { + "epoch": 0.19755434782608697, + "grad_norm": 3.848099708557129, + "learning_rate": 3e-05, + "loss": 3.6353, + "step": 727 + }, + { + "epoch": 0.19782608695652174, + "grad_norm": 3.534794569015503, + "learning_rate": 3e-05, + "loss": 3.9035, + "step": 728 + }, + { + "epoch": 0.19809782608695653, + "grad_norm": 1.7165888547897339, + "learning_rate": 3e-05, + "loss": 3.3266, + "step": 729 + }, + { + "epoch": 0.1983695652173913, + "grad_norm": 2.5262253284454346, + "learning_rate": 3e-05, + "loss": 3.6359, + "step": 730 + }, + { + "epoch": 0.1986413043478261, + "grad_norm": 2.5349607467651367, + "learning_rate": 3e-05, + "loss": 3.129, + "step": 731 + }, + { + "epoch": 0.19891304347826086, + "grad_norm": 5.680181503295898, + "learning_rate": 3e-05, + "loss": 4.8488, + "step": 732 + }, + { + "epoch": 0.19918478260869565, + "grad_norm": 3.0405654907226562, + "learning_rate": 3e-05, + "loss": 3.3313, + "step": 733 + }, + { + "epoch": 0.19945652173913042, + "grad_norm": 1.6890621185302734, + "learning_rate": 3e-05, + "loss": 3.2893, + "step": 734 + }, + { + "epoch": 0.19972826086956522, + "grad_norm": 2.204172372817993, + "learning_rate": 3e-05, + "loss": 3.5067, + "step": 735 + }, + { + "epoch": 0.2, + "grad_norm": 2.858398914337158, + "learning_rate": 3e-05, + "loss": 3.4036, + "step": 736 + }, + { + "epoch": 0.20027173913043478, + "grad_norm": 2.4763529300689697, + "learning_rate": 3e-05, + "loss": 3.5171, + "step": 737 + }, + { + "epoch": 0.20054347826086957, + "grad_norm": 2.957118034362793, + "learning_rate": 3e-05, + "loss": 4.1728, + "step": 738 + }, + { + "epoch": 0.20081521739130434, + "grad_norm": 1.8309071063995361, + "learning_rate": 3e-05, + "loss": 2.9663, + "step": 739 + }, + { + "epoch": 0.20108695652173914, + "grad_norm": 2.270033121109009, + "learning_rate": 3e-05, + "loss": 3.6115, + "step": 740 + }, + { + "epoch": 0.2013586956521739, + "grad_norm": 2.1604442596435547, + "learning_rate": 3e-05, + "loss": 3.2511, + "step": 741 + }, + { + "epoch": 0.2016304347826087, + "grad_norm": 3.5561349391937256, + "learning_rate": 3e-05, + "loss": 3.8886, + "step": 742 + }, + { + "epoch": 0.20190217391304346, + "grad_norm": 2.510963201522827, + "learning_rate": 3e-05, + "loss": 3.4812, + "step": 743 + }, + { + "epoch": 0.20217391304347826, + "grad_norm": 1.5206657648086548, + "learning_rate": 3e-05, + "loss": 3.0231, + "step": 744 + }, + { + "epoch": 0.20244565217391305, + "grad_norm": 1.645155906677246, + "learning_rate": 3e-05, + "loss": 3.0951, + "step": 745 + }, + { + "epoch": 0.20271739130434782, + "grad_norm": 2.075399160385132, + "learning_rate": 3e-05, + "loss": 3.4057, + "step": 746 + }, + { + "epoch": 0.20298913043478262, + "grad_norm": 2.36306095123291, + "learning_rate": 3e-05, + "loss": 3.6449, + "step": 747 + }, + { + "epoch": 0.20326086956521738, + "grad_norm": 1.9054253101348877, + "learning_rate": 3e-05, + "loss": 3.3983, + "step": 748 + }, + { + "epoch": 0.20353260869565218, + "grad_norm": 1.7470569610595703, + "learning_rate": 3e-05, + "loss": 3.7903, + "step": 749 + }, + { + "epoch": 0.20380434782608695, + "grad_norm": 1.6297615766525269, + "learning_rate": 3e-05, + "loss": 3.0304, + "step": 750 + }, + { + "epoch": 0.20407608695652174, + "grad_norm": 3.2830233573913574, + "learning_rate": 3e-05, + "loss": 4.2781, + "step": 751 + }, + { + "epoch": 0.20434782608695654, + "grad_norm": 1.361449122428894, + "learning_rate": 3e-05, + "loss": 3.2224, + "step": 752 + }, + { + "epoch": 0.2046195652173913, + "grad_norm": 1.4894417524337769, + "learning_rate": 3e-05, + "loss": 3.262, + "step": 753 + }, + { + "epoch": 0.2048913043478261, + "grad_norm": 1.7271361351013184, + "learning_rate": 3e-05, + "loss": 3.3017, + "step": 754 + }, + { + "epoch": 0.20516304347826086, + "grad_norm": 2.033215284347534, + "learning_rate": 3e-05, + "loss": 2.9264, + "step": 755 + }, + { + "epoch": 0.20543478260869566, + "grad_norm": 2.808903455734253, + "learning_rate": 3e-05, + "loss": 3.7223, + "step": 756 + }, + { + "epoch": 0.20570652173913043, + "grad_norm": 1.450661540031433, + "learning_rate": 3e-05, + "loss": 4.0346, + "step": 757 + }, + { + "epoch": 0.20597826086956522, + "grad_norm": 1.9796857833862305, + "learning_rate": 3e-05, + "loss": 3.2978, + "step": 758 + }, + { + "epoch": 0.20625, + "grad_norm": 2.9408633708953857, + "learning_rate": 3e-05, + "loss": 3.7247, + "step": 759 + }, + { + "epoch": 0.20652173913043478, + "grad_norm": 1.4558497667312622, + "learning_rate": 3e-05, + "loss": 3.0258, + "step": 760 + }, + { + "epoch": 0.20679347826086958, + "grad_norm": 1.1229232549667358, + "learning_rate": 3e-05, + "loss": 2.9328, + "step": 761 + }, + { + "epoch": 0.20706521739130435, + "grad_norm": 1.663891077041626, + "learning_rate": 3e-05, + "loss": 3.2935, + "step": 762 + }, + { + "epoch": 0.20733695652173914, + "grad_norm": 1.8489412069320679, + "learning_rate": 3e-05, + "loss": 2.8728, + "step": 763 + }, + { + "epoch": 0.2076086956521739, + "grad_norm": 1.359864592552185, + "learning_rate": 3e-05, + "loss": 3.6552, + "step": 764 + }, + { + "epoch": 0.2078804347826087, + "grad_norm": 1.5933263301849365, + "learning_rate": 3e-05, + "loss": 3.2384, + "step": 765 + }, + { + "epoch": 0.20815217391304347, + "grad_norm": 3.0126466751098633, + "learning_rate": 3e-05, + "loss": 3.4475, + "step": 766 + }, + { + "epoch": 0.20842391304347826, + "grad_norm": 2.608242988586426, + "learning_rate": 3e-05, + "loss": 3.2701, + "step": 767 + }, + { + "epoch": 0.20869565217391303, + "grad_norm": 2.7247979640960693, + "learning_rate": 3e-05, + "loss": 3.96, + "step": 768 + }, + { + "epoch": 0.20896739130434783, + "grad_norm": 1.8706130981445312, + "learning_rate": 3e-05, + "loss": 3.0889, + "step": 769 + }, + { + "epoch": 0.20923913043478262, + "grad_norm": 2.668093681335449, + "learning_rate": 3e-05, + "loss": 3.4351, + "step": 770 + }, + { + "epoch": 0.2095108695652174, + "grad_norm": 3.017181396484375, + "learning_rate": 3e-05, + "loss": 3.2463, + "step": 771 + }, + { + "epoch": 0.20978260869565218, + "grad_norm": 3.0812008380889893, + "learning_rate": 3e-05, + "loss": 3.3998, + "step": 772 + }, + { + "epoch": 0.21005434782608695, + "grad_norm": 2.7510030269622803, + "learning_rate": 3e-05, + "loss": 4.2353, + "step": 773 + }, + { + "epoch": 0.21032608695652175, + "grad_norm": 2.4355452060699463, + "learning_rate": 3e-05, + "loss": 3.8853, + "step": 774 + }, + { + "epoch": 0.2105978260869565, + "grad_norm": 4.041244983673096, + "learning_rate": 3e-05, + "loss": 4.0771, + "step": 775 + }, + { + "epoch": 0.2108695652173913, + "grad_norm": 3.403510332107544, + "learning_rate": 3e-05, + "loss": 3.6963, + "step": 776 + }, + { + "epoch": 0.21114130434782608, + "grad_norm": 3.216482400894165, + "learning_rate": 3e-05, + "loss": 3.6608, + "step": 777 + }, + { + "epoch": 0.21141304347826087, + "grad_norm": 2.5989348888397217, + "learning_rate": 3e-05, + "loss": 3.689, + "step": 778 + }, + { + "epoch": 0.21168478260869567, + "grad_norm": 2.0239574909210205, + "learning_rate": 3e-05, + "loss": 3.681, + "step": 779 + }, + { + "epoch": 0.21195652173913043, + "grad_norm": 3.3878705501556396, + "learning_rate": 3e-05, + "loss": 3.9301, + "step": 780 + }, + { + "epoch": 0.21222826086956523, + "grad_norm": 4.106480598449707, + "learning_rate": 3e-05, + "loss": 3.2929, + "step": 781 + }, + { + "epoch": 0.2125, + "grad_norm": 3.4818196296691895, + "learning_rate": 3e-05, + "loss": 2.8562, + "step": 782 + }, + { + "epoch": 0.2127717391304348, + "grad_norm": 3.4106242656707764, + "learning_rate": 3e-05, + "loss": 3.6078, + "step": 783 + }, + { + "epoch": 0.21304347826086956, + "grad_norm": 2.401315450668335, + "learning_rate": 3e-05, + "loss": 3.693, + "step": 784 + }, + { + "epoch": 0.21331521739130435, + "grad_norm": 3.092789649963379, + "learning_rate": 3e-05, + "loss": 3.6907, + "step": 785 + }, + { + "epoch": 0.21358695652173912, + "grad_norm": 2.8016254901885986, + "learning_rate": 3e-05, + "loss": 3.7602, + "step": 786 + }, + { + "epoch": 0.2138586956521739, + "grad_norm": 2.281668186187744, + "learning_rate": 3e-05, + "loss": 2.9451, + "step": 787 + }, + { + "epoch": 0.2141304347826087, + "grad_norm": 2.3392529487609863, + "learning_rate": 3e-05, + "loss": 3.1969, + "step": 788 + }, + { + "epoch": 0.21440217391304348, + "grad_norm": 1.7437257766723633, + "learning_rate": 3e-05, + "loss": 3.2468, + "step": 789 + }, + { + "epoch": 0.21467391304347827, + "grad_norm": 2.128056049346924, + "learning_rate": 3e-05, + "loss": 3.5476, + "step": 790 + }, + { + "epoch": 0.21494565217391304, + "grad_norm": 2.1912710666656494, + "learning_rate": 3e-05, + "loss": 3.7689, + "step": 791 + }, + { + "epoch": 0.21521739130434783, + "grad_norm": 2.5414609909057617, + "learning_rate": 3e-05, + "loss": 3.4014, + "step": 792 + }, + { + "epoch": 0.2154891304347826, + "grad_norm": 2.6355223655700684, + "learning_rate": 3e-05, + "loss": 3.3787, + "step": 793 + }, + { + "epoch": 0.2157608695652174, + "grad_norm": 2.1927592754364014, + "learning_rate": 3e-05, + "loss": 3.2381, + "step": 794 + }, + { + "epoch": 0.21603260869565216, + "grad_norm": 1.7231847047805786, + "learning_rate": 3e-05, + "loss": 3.4627, + "step": 795 + }, + { + "epoch": 0.21630434782608696, + "grad_norm": 1.8754600286483765, + "learning_rate": 3e-05, + "loss": 3.7409, + "step": 796 + }, + { + "epoch": 0.21657608695652175, + "grad_norm": 1.7413159608840942, + "learning_rate": 3e-05, + "loss": 3.6012, + "step": 797 + }, + { + "epoch": 0.21684782608695652, + "grad_norm": 1.9197379350662231, + "learning_rate": 3e-05, + "loss": 3.15, + "step": 798 + }, + { + "epoch": 0.2171195652173913, + "grad_norm": 2.100706100463867, + "learning_rate": 3e-05, + "loss": 2.8748, + "step": 799 + }, + { + "epoch": 0.21739130434782608, + "grad_norm": 2.7185757160186768, + "learning_rate": 3e-05, + "loss": 3.7859, + "step": 800 + }, + { + "epoch": 0.21766304347826088, + "grad_norm": 3.241926908493042, + "learning_rate": 3e-05, + "loss": 3.7198, + "step": 801 + }, + { + "epoch": 0.21793478260869564, + "grad_norm": 2.365151882171631, + "learning_rate": 3e-05, + "loss": 3.7607, + "step": 802 + }, + { + "epoch": 0.21820652173913044, + "grad_norm": 2.0234482288360596, + "learning_rate": 3e-05, + "loss": 3.1843, + "step": 803 + }, + { + "epoch": 0.2184782608695652, + "grad_norm": 3.084533929824829, + "learning_rate": 3e-05, + "loss": 4.1945, + "step": 804 + }, + { + "epoch": 0.21875, + "grad_norm": 2.927222490310669, + "learning_rate": 3e-05, + "loss": 3.3531, + "step": 805 + }, + { + "epoch": 0.2190217391304348, + "grad_norm": 2.6623260974884033, + "learning_rate": 3e-05, + "loss": 3.5994, + "step": 806 + }, + { + "epoch": 0.21929347826086956, + "grad_norm": 2.8344593048095703, + "learning_rate": 3e-05, + "loss": 3.6001, + "step": 807 + }, + { + "epoch": 0.21956521739130436, + "grad_norm": 1.8181391954421997, + "learning_rate": 3e-05, + "loss": 2.9434, + "step": 808 + }, + { + "epoch": 0.21983695652173912, + "grad_norm": 2.103512763977051, + "learning_rate": 3e-05, + "loss": 3.8034, + "step": 809 + }, + { + "epoch": 0.22010869565217392, + "grad_norm": 1.961503505706787, + "learning_rate": 3e-05, + "loss": 3.0883, + "step": 810 + }, + { + "epoch": 0.2203804347826087, + "grad_norm": 2.3394432067871094, + "learning_rate": 3e-05, + "loss": 2.8568, + "step": 811 + }, + { + "epoch": 0.22065217391304348, + "grad_norm": 2.6501801013946533, + "learning_rate": 3e-05, + "loss": 3.2469, + "step": 812 + }, + { + "epoch": 0.22092391304347825, + "grad_norm": 2.7663819789886475, + "learning_rate": 3e-05, + "loss": 3.3066, + "step": 813 + }, + { + "epoch": 0.22119565217391304, + "grad_norm": 2.4567878246307373, + "learning_rate": 3e-05, + "loss": 3.0416, + "step": 814 + }, + { + "epoch": 0.22146739130434784, + "grad_norm": 2.378188371658325, + "learning_rate": 3e-05, + "loss": 3.6954, + "step": 815 + }, + { + "epoch": 0.2217391304347826, + "grad_norm": 2.611842632293701, + "learning_rate": 3e-05, + "loss": 3.3243, + "step": 816 + }, + { + "epoch": 0.2220108695652174, + "grad_norm": 2.1804463863372803, + "learning_rate": 3e-05, + "loss": 2.9858, + "step": 817 + }, + { + "epoch": 0.22228260869565217, + "grad_norm": 2.03049373626709, + "learning_rate": 3e-05, + "loss": 2.787, + "step": 818 + }, + { + "epoch": 0.22255434782608696, + "grad_norm": 1.8013122081756592, + "learning_rate": 3e-05, + "loss": 2.9299, + "step": 819 + }, + { + "epoch": 0.22282608695652173, + "grad_norm": 2.039745569229126, + "learning_rate": 3e-05, + "loss": 3.2247, + "step": 820 + }, + { + "epoch": 0.22309782608695652, + "grad_norm": 1.6249157190322876, + "learning_rate": 3e-05, + "loss": 2.8942, + "step": 821 + }, + { + "epoch": 0.2233695652173913, + "grad_norm": 1.8277668952941895, + "learning_rate": 3e-05, + "loss": 3.5634, + "step": 822 + }, + { + "epoch": 0.2236413043478261, + "grad_norm": 2.2749369144439697, + "learning_rate": 3e-05, + "loss": 3.6253, + "step": 823 + }, + { + "epoch": 0.22391304347826088, + "grad_norm": 2.3847649097442627, + "learning_rate": 3e-05, + "loss": 3.3725, + "step": 824 + }, + { + "epoch": 0.22418478260869565, + "grad_norm": 1.9465465545654297, + "learning_rate": 3e-05, + "loss": 3.4556, + "step": 825 + }, + { + "epoch": 0.22445652173913044, + "grad_norm": 2.793731927871704, + "learning_rate": 3e-05, + "loss": 4.0801, + "step": 826 + }, + { + "epoch": 0.2247282608695652, + "grad_norm": 2.605842113494873, + "learning_rate": 3e-05, + "loss": 4.0396, + "step": 827 + }, + { + "epoch": 0.225, + "grad_norm": 2.200025796890259, + "learning_rate": 3e-05, + "loss": 3.7514, + "step": 828 + }, + { + "epoch": 0.22527173913043477, + "grad_norm": 1.8318731784820557, + "learning_rate": 3e-05, + "loss": 3.3619, + "step": 829 + }, + { + "epoch": 0.22554347826086957, + "grad_norm": 1.7559170722961426, + "learning_rate": 3e-05, + "loss": 3.3115, + "step": 830 + }, + { + "epoch": 0.22581521739130433, + "grad_norm": 1.7032020092010498, + "learning_rate": 3e-05, + "loss": 3.113, + "step": 831 + }, + { + "epoch": 0.22608695652173913, + "grad_norm": 1.8972278833389282, + "learning_rate": 3e-05, + "loss": 3.1965, + "step": 832 + }, + { + "epoch": 0.22635869565217392, + "grad_norm": 1.9449539184570312, + "learning_rate": 3e-05, + "loss": 3.2422, + "step": 833 + }, + { + "epoch": 0.2266304347826087, + "grad_norm": 1.6739164590835571, + "learning_rate": 3e-05, + "loss": 3.2255, + "step": 834 + }, + { + "epoch": 0.2269021739130435, + "grad_norm": 1.7071807384490967, + "learning_rate": 3e-05, + "loss": 3.4032, + "step": 835 + }, + { + "epoch": 0.22717391304347825, + "grad_norm": 3.5912764072418213, + "learning_rate": 3e-05, + "loss": 4.1757, + "step": 836 + }, + { + "epoch": 0.22744565217391305, + "grad_norm": 1.969997763633728, + "learning_rate": 3e-05, + "loss": 3.1138, + "step": 837 + }, + { + "epoch": 0.22771739130434782, + "grad_norm": 1.9781277179718018, + "learning_rate": 3e-05, + "loss": 3.5066, + "step": 838 + }, + { + "epoch": 0.2279891304347826, + "grad_norm": 1.6115202903747559, + "learning_rate": 3e-05, + "loss": 2.9759, + "step": 839 + }, + { + "epoch": 0.22826086956521738, + "grad_norm": 1.3473846912384033, + "learning_rate": 3e-05, + "loss": 3.0617, + "step": 840 + }, + { + "epoch": 0.22853260869565217, + "grad_norm": 1.6643739938735962, + "learning_rate": 3e-05, + "loss": 3.4268, + "step": 841 + }, + { + "epoch": 0.22880434782608697, + "grad_norm": 1.9249311685562134, + "learning_rate": 3e-05, + "loss": 3.6207, + "step": 842 + }, + { + "epoch": 0.22907608695652174, + "grad_norm": 2.079634666442871, + "learning_rate": 3e-05, + "loss": 3.534, + "step": 843 + }, + { + "epoch": 0.22934782608695653, + "grad_norm": 1.8367491960525513, + "learning_rate": 3e-05, + "loss": 3.9664, + "step": 844 + }, + { + "epoch": 0.2296195652173913, + "grad_norm": 1.7043336629867554, + "learning_rate": 3e-05, + "loss": 3.4275, + "step": 845 + }, + { + "epoch": 0.2298913043478261, + "grad_norm": 1.6460905075073242, + "learning_rate": 3e-05, + "loss": 3.1771, + "step": 846 + }, + { + "epoch": 0.23016304347826086, + "grad_norm": 2.0371439456939697, + "learning_rate": 3e-05, + "loss": 3.3044, + "step": 847 + }, + { + "epoch": 0.23043478260869565, + "grad_norm": 1.5710325241088867, + "learning_rate": 3e-05, + "loss": 3.0044, + "step": 848 + }, + { + "epoch": 0.23070652173913042, + "grad_norm": 1.6866766214370728, + "learning_rate": 3e-05, + "loss": 2.9916, + "step": 849 + }, + { + "epoch": 0.23097826086956522, + "grad_norm": 2.127194404602051, + "learning_rate": 3e-05, + "loss": 3.5753, + "step": 850 + }, + { + "epoch": 0.23125, + "grad_norm": 2.0354788303375244, + "learning_rate": 3e-05, + "loss": 2.8992, + "step": 851 + }, + { + "epoch": 0.23152173913043478, + "grad_norm": 1.5567448139190674, + "learning_rate": 3e-05, + "loss": 3.4486, + "step": 852 + }, + { + "epoch": 0.23179347826086957, + "grad_norm": 1.3115332126617432, + "learning_rate": 3e-05, + "loss": 3.1354, + "step": 853 + }, + { + "epoch": 0.23206521739130434, + "grad_norm": 1.3727189302444458, + "learning_rate": 3e-05, + "loss": 3.0521, + "step": 854 + }, + { + "epoch": 0.23233695652173914, + "grad_norm": 1.3423751592636108, + "learning_rate": 3e-05, + "loss": 2.908, + "step": 855 + }, + { + "epoch": 0.2326086956521739, + "grad_norm": 1.61993408203125, + "learning_rate": 3e-05, + "loss": 3.2873, + "step": 856 + }, + { + "epoch": 0.2328804347826087, + "grad_norm": 1.8327138423919678, + "learning_rate": 3e-05, + "loss": 3.6789, + "step": 857 + }, + { + "epoch": 0.23315217391304346, + "grad_norm": 1.626006007194519, + "learning_rate": 3e-05, + "loss": 3.0027, + "step": 858 + }, + { + "epoch": 0.23342391304347826, + "grad_norm": 2.236844539642334, + "learning_rate": 3e-05, + "loss": 4.083, + "step": 859 + }, + { + "epoch": 0.23369565217391305, + "grad_norm": 2.1321239471435547, + "learning_rate": 3e-05, + "loss": 3.455, + "step": 860 + }, + { + "epoch": 0.23396739130434782, + "grad_norm": 2.1679024696350098, + "learning_rate": 3e-05, + "loss": 3.8644, + "step": 861 + }, + { + "epoch": 0.23423913043478262, + "grad_norm": 1.9821666479110718, + "learning_rate": 3e-05, + "loss": 3.5024, + "step": 862 + }, + { + "epoch": 0.23451086956521738, + "grad_norm": 2.8352460861206055, + "learning_rate": 3e-05, + "loss": 4.0082, + "step": 863 + }, + { + "epoch": 0.23478260869565218, + "grad_norm": 2.2241246700286865, + "learning_rate": 3e-05, + "loss": 3.8031, + "step": 864 + }, + { + "epoch": 0.23505434782608695, + "grad_norm": 2.1706037521362305, + "learning_rate": 3e-05, + "loss": 4.0421, + "step": 865 + }, + { + "epoch": 0.23532608695652174, + "grad_norm": 1.695703148841858, + "learning_rate": 3e-05, + "loss": 2.9672, + "step": 866 + }, + { + "epoch": 0.23559782608695654, + "grad_norm": 1.994674563407898, + "learning_rate": 3e-05, + "loss": 3.3449, + "step": 867 + }, + { + "epoch": 0.2358695652173913, + "grad_norm": 2.692945957183838, + "learning_rate": 3e-05, + "loss": 4.1761, + "step": 868 + }, + { + "epoch": 0.2361413043478261, + "grad_norm": 2.5395121574401855, + "learning_rate": 3e-05, + "loss": 3.921, + "step": 869 + }, + { + "epoch": 0.23641304347826086, + "grad_norm": 1.991248607635498, + "learning_rate": 3e-05, + "loss": 2.9, + "step": 870 + }, + { + "epoch": 0.23668478260869566, + "grad_norm": 2.2419328689575195, + "learning_rate": 3e-05, + "loss": 3.501, + "step": 871 + }, + { + "epoch": 0.23695652173913043, + "grad_norm": 1.9551500082015991, + "learning_rate": 3e-05, + "loss": 3.8966, + "step": 872 + }, + { + "epoch": 0.23722826086956522, + "grad_norm": 1.9755557775497437, + "learning_rate": 3e-05, + "loss": 3.7852, + "step": 873 + }, + { + "epoch": 0.2375, + "grad_norm": 1.964316487312317, + "learning_rate": 3e-05, + "loss": 3.4841, + "step": 874 + }, + { + "epoch": 0.23777173913043478, + "grad_norm": 1.667006015777588, + "learning_rate": 3e-05, + "loss": 3.3479, + "step": 875 + }, + { + "epoch": 0.23804347826086958, + "grad_norm": 1.3252580165863037, + "learning_rate": 3e-05, + "loss": 3.1133, + "step": 876 + }, + { + "epoch": 0.23831521739130435, + "grad_norm": 1.904585361480713, + "learning_rate": 3e-05, + "loss": 3.7604, + "step": 877 + }, + { + "epoch": 0.23858695652173914, + "grad_norm": 1.581666350364685, + "learning_rate": 3e-05, + "loss": 3.3015, + "step": 878 + }, + { + "epoch": 0.2388586956521739, + "grad_norm": 1.2517045736312866, + "learning_rate": 3e-05, + "loss": 3.2687, + "step": 879 + }, + { + "epoch": 0.2391304347826087, + "grad_norm": 1.5604926347732544, + "learning_rate": 3e-05, + "loss": 3.0638, + "step": 880 + }, + { + "epoch": 0.23940217391304347, + "grad_norm": 1.9475339651107788, + "learning_rate": 3e-05, + "loss": 3.8968, + "step": 881 + }, + { + "epoch": 0.23967391304347826, + "grad_norm": 1.528891921043396, + "learning_rate": 3e-05, + "loss": 3.0379, + "step": 882 + }, + { + "epoch": 0.23994565217391303, + "grad_norm": 1.838294506072998, + "learning_rate": 3e-05, + "loss": 3.7645, + "step": 883 + }, + { + "epoch": 0.24021739130434783, + "grad_norm": 1.5220407247543335, + "learning_rate": 3e-05, + "loss": 3.9417, + "step": 884 + }, + { + "epoch": 0.24048913043478262, + "grad_norm": 1.4615991115570068, + "learning_rate": 3e-05, + "loss": 3.4286, + "step": 885 + }, + { + "epoch": 0.2407608695652174, + "grad_norm": 2.569244623184204, + "learning_rate": 3e-05, + "loss": 3.9001, + "step": 886 + }, + { + "epoch": 0.24103260869565218, + "grad_norm": 2.3717269897460938, + "learning_rate": 3e-05, + "loss": 4.2076, + "step": 887 + }, + { + "epoch": 0.24130434782608695, + "grad_norm": 1.8509140014648438, + "learning_rate": 3e-05, + "loss": 3.2101, + "step": 888 + }, + { + "epoch": 0.24157608695652175, + "grad_norm": 2.0046916007995605, + "learning_rate": 3e-05, + "loss": 2.9609, + "step": 889 + }, + { + "epoch": 0.2418478260869565, + "grad_norm": 2.2359912395477295, + "learning_rate": 3e-05, + "loss": 3.6416, + "step": 890 + }, + { + "epoch": 0.2421195652173913, + "grad_norm": 1.8625085353851318, + "learning_rate": 3e-05, + "loss": 3.3007, + "step": 891 + }, + { + "epoch": 0.24239130434782608, + "grad_norm": 2.0958662033081055, + "learning_rate": 3e-05, + "loss": 2.7472, + "step": 892 + }, + { + "epoch": 0.24266304347826087, + "grad_norm": 2.5501413345336914, + "learning_rate": 3e-05, + "loss": 3.0367, + "step": 893 + }, + { + "epoch": 0.24293478260869567, + "grad_norm": 3.020587921142578, + "learning_rate": 3e-05, + "loss": 3.7357, + "step": 894 + }, + { + "epoch": 0.24320652173913043, + "grad_norm": 2.0165812969207764, + "learning_rate": 3e-05, + "loss": 3.6235, + "step": 895 + }, + { + "epoch": 0.24347826086956523, + "grad_norm": 3.1172547340393066, + "learning_rate": 3e-05, + "loss": 4.1981, + "step": 896 + }, + { + "epoch": 0.24375, + "grad_norm": 2.381237268447876, + "learning_rate": 3e-05, + "loss": 3.104, + "step": 897 + }, + { + "epoch": 0.2440217391304348, + "grad_norm": 3.345118522644043, + "learning_rate": 3e-05, + "loss": 4.1412, + "step": 898 + }, + { + "epoch": 0.24429347826086956, + "grad_norm": 2.8453972339630127, + "learning_rate": 3e-05, + "loss": 3.3134, + "step": 899 + }, + { + "epoch": 0.24456521739130435, + "grad_norm": 1.9727962017059326, + "learning_rate": 3e-05, + "loss": 3.1409, + "step": 900 + }, + { + "epoch": 0.24483695652173912, + "grad_norm": 1.3924411535263062, + "learning_rate": 3e-05, + "loss": 3.2174, + "step": 901 + }, + { + "epoch": 0.2451086956521739, + "grad_norm": 1.8885339498519897, + "learning_rate": 3e-05, + "loss": 3.4122, + "step": 902 + }, + { + "epoch": 0.2453804347826087, + "grad_norm": 2.3274450302124023, + "learning_rate": 3e-05, + "loss": 3.2905, + "step": 903 + }, + { + "epoch": 0.24565217391304348, + "grad_norm": 1.877142310142517, + "learning_rate": 3e-05, + "loss": 3.4118, + "step": 904 + }, + { + "epoch": 0.24592391304347827, + "grad_norm": 1.8108350038528442, + "learning_rate": 3e-05, + "loss": 3.206, + "step": 905 + }, + { + "epoch": 0.24619565217391304, + "grad_norm": 1.6787278652191162, + "learning_rate": 3e-05, + "loss": 3.2236, + "step": 906 + }, + { + "epoch": 0.24646739130434783, + "grad_norm": 2.007154703140259, + "learning_rate": 3e-05, + "loss": 4.1134, + "step": 907 + }, + { + "epoch": 0.2467391304347826, + "grad_norm": 1.9029121398925781, + "learning_rate": 3e-05, + "loss": 3.2066, + "step": 908 + }, + { + "epoch": 0.2470108695652174, + "grad_norm": 1.5358937978744507, + "learning_rate": 3e-05, + "loss": 3.3406, + "step": 909 + }, + { + "epoch": 0.24728260869565216, + "grad_norm": 1.706233263015747, + "learning_rate": 3e-05, + "loss": 3.368, + "step": 910 + }, + { + "epoch": 0.24755434782608696, + "grad_norm": 2.283039093017578, + "learning_rate": 3e-05, + "loss": 4.1815, + "step": 911 + }, + { + "epoch": 0.24782608695652175, + "grad_norm": 1.6811976432800293, + "learning_rate": 3e-05, + "loss": 3.8339, + "step": 912 + }, + { + "epoch": 0.24809782608695652, + "grad_norm": 1.6838414669036865, + "learning_rate": 3e-05, + "loss": 3.6558, + "step": 913 + }, + { + "epoch": 0.2483695652173913, + "grad_norm": 1.451900839805603, + "learning_rate": 3e-05, + "loss": 3.2279, + "step": 914 + }, + { + "epoch": 0.24864130434782608, + "grad_norm": 2.3503222465515137, + "learning_rate": 3e-05, + "loss": 3.9484, + "step": 915 + }, + { + "epoch": 0.24891304347826088, + "grad_norm": 2.001974582672119, + "learning_rate": 3e-05, + "loss": 3.3028, + "step": 916 + }, + { + "epoch": 0.24918478260869564, + "grad_norm": 1.5036247968673706, + "learning_rate": 3e-05, + "loss": 3.2459, + "step": 917 + }, + { + "epoch": 0.24945652173913044, + "grad_norm": 2.033236026763916, + "learning_rate": 3e-05, + "loss": 3.4323, + "step": 918 + }, + { + "epoch": 0.2497282608695652, + "grad_norm": 1.4043093919754028, + "learning_rate": 3e-05, + "loss": 3.0965, + "step": 919 + }, + { + "epoch": 0.25, + "grad_norm": 1.203482985496521, + "learning_rate": 3e-05, + "loss": 3.1118, + "step": 920 + }, + { + "epoch": 0.25027173913043477, + "grad_norm": 1.4571930170059204, + "learning_rate": 3e-05, + "loss": 3.351, + "step": 921 + }, + { + "epoch": 0.2505434782608696, + "grad_norm": 0.9827572703361511, + "learning_rate": 3e-05, + "loss": 2.8122, + "step": 922 + }, + { + "epoch": 0.25081521739130436, + "grad_norm": 1.520359754562378, + "learning_rate": 3e-05, + "loss": 3.6678, + "step": 923 + }, + { + "epoch": 0.2510869565217391, + "grad_norm": 1.277106761932373, + "learning_rate": 3e-05, + "loss": 2.9916, + "step": 924 + }, + { + "epoch": 0.2513586956521739, + "grad_norm": 1.7534111738204956, + "learning_rate": 3e-05, + "loss": 3.9959, + "step": 925 + }, + { + "epoch": 0.2516304347826087, + "grad_norm": 1.031008005142212, + "learning_rate": 3e-05, + "loss": 2.9768, + "step": 926 + }, + { + "epoch": 0.2519021739130435, + "grad_norm": 1.1608198881149292, + "learning_rate": 3e-05, + "loss": 3.456, + "step": 927 + }, + { + "epoch": 0.25217391304347825, + "grad_norm": 1.4016661643981934, + "learning_rate": 3e-05, + "loss": 3.4423, + "step": 928 + }, + { + "epoch": 0.25244565217391307, + "grad_norm": 1.1200820207595825, + "learning_rate": 3e-05, + "loss": 2.897, + "step": 929 + }, + { + "epoch": 0.25271739130434784, + "grad_norm": 1.4193557500839233, + "learning_rate": 3e-05, + "loss": 3.7882, + "step": 930 + }, + { + "epoch": 0.2529891304347826, + "grad_norm": 1.2631326913833618, + "learning_rate": 3e-05, + "loss": 3.8642, + "step": 931 + }, + { + "epoch": 0.2532608695652174, + "grad_norm": 1.6645845174789429, + "learning_rate": 3e-05, + "loss": 3.7794, + "step": 932 + }, + { + "epoch": 0.2535326086956522, + "grad_norm": 1.7016172409057617, + "learning_rate": 3e-05, + "loss": 3.3343, + "step": 933 + }, + { + "epoch": 0.25380434782608696, + "grad_norm": 1.1917381286621094, + "learning_rate": 3e-05, + "loss": 3.3563, + "step": 934 + }, + { + "epoch": 0.25407608695652173, + "grad_norm": 1.4421312808990479, + "learning_rate": 3e-05, + "loss": 3.6694, + "step": 935 + }, + { + "epoch": 0.2543478260869565, + "grad_norm": 1.2393507957458496, + "learning_rate": 3e-05, + "loss": 2.8572, + "step": 936 + }, + { + "epoch": 0.2546195652173913, + "grad_norm": 1.4665427207946777, + "learning_rate": 3e-05, + "loss": 3.5944, + "step": 937 + }, + { + "epoch": 0.2548913043478261, + "grad_norm": 1.3151808977127075, + "learning_rate": 3e-05, + "loss": 3.1136, + "step": 938 + }, + { + "epoch": 0.25516304347826085, + "grad_norm": 1.3733018636703491, + "learning_rate": 3e-05, + "loss": 3.1559, + "step": 939 + }, + { + "epoch": 0.2554347826086957, + "grad_norm": 1.638494610786438, + "learning_rate": 3e-05, + "loss": 3.5616, + "step": 940 + }, + { + "epoch": 0.25570652173913044, + "grad_norm": 1.5186861753463745, + "learning_rate": 3e-05, + "loss": 3.7529, + "step": 941 + }, + { + "epoch": 0.2559782608695652, + "grad_norm": 1.7179397344589233, + "learning_rate": 3e-05, + "loss": 3.8634, + "step": 942 + }, + { + "epoch": 0.25625, + "grad_norm": 1.9366356134414673, + "learning_rate": 3e-05, + "loss": 3.785, + "step": 943 + }, + { + "epoch": 0.2565217391304348, + "grad_norm": 2.2738037109375, + "learning_rate": 3e-05, + "loss": 3.7442, + "step": 944 + }, + { + "epoch": 0.25679347826086957, + "grad_norm": 2.67051100730896, + "learning_rate": 3e-05, + "loss": 3.8186, + "step": 945 + }, + { + "epoch": 0.25706521739130433, + "grad_norm": 3.3741190433502197, + "learning_rate": 3e-05, + "loss": 4.2118, + "step": 946 + }, + { + "epoch": 0.25733695652173916, + "grad_norm": 1.6262043714523315, + "learning_rate": 3e-05, + "loss": 3.0208, + "step": 947 + }, + { + "epoch": 0.2576086956521739, + "grad_norm": 1.7342244386672974, + "learning_rate": 3e-05, + "loss": 3.1897, + "step": 948 + }, + { + "epoch": 0.2578804347826087, + "grad_norm": 2.0924177169799805, + "learning_rate": 3e-05, + "loss": 2.7324, + "step": 949 + }, + { + "epoch": 0.25815217391304346, + "grad_norm": 2.21530818939209, + "learning_rate": 3e-05, + "loss": 3.1098, + "step": 950 + }, + { + "epoch": 0.2584239130434783, + "grad_norm": 1.642945647239685, + "learning_rate": 3e-05, + "loss": 3.58, + "step": 951 + }, + { + "epoch": 0.25869565217391305, + "grad_norm": 1.4641486406326294, + "learning_rate": 3e-05, + "loss": 3.4771, + "step": 952 + }, + { + "epoch": 0.2589673913043478, + "grad_norm": 1.4831671714782715, + "learning_rate": 3e-05, + "loss": 3.0967, + "step": 953 + }, + { + "epoch": 0.2592391304347826, + "grad_norm": 1.605242133140564, + "learning_rate": 3e-05, + "loss": 3.3319, + "step": 954 + }, + { + "epoch": 0.2595108695652174, + "grad_norm": 2.312819480895996, + "learning_rate": 3e-05, + "loss": 4.214, + "step": 955 + }, + { + "epoch": 0.2597826086956522, + "grad_norm": 1.751943588256836, + "learning_rate": 3e-05, + "loss": 3.3116, + "step": 956 + }, + { + "epoch": 0.26005434782608694, + "grad_norm": 1.951372504234314, + "learning_rate": 3e-05, + "loss": 3.4478, + "step": 957 + }, + { + "epoch": 0.26032608695652176, + "grad_norm": 2.3981149196624756, + "learning_rate": 3e-05, + "loss": 3.4107, + "step": 958 + }, + { + "epoch": 0.26059782608695653, + "grad_norm": 2.448025703430176, + "learning_rate": 3e-05, + "loss": 4.0686, + "step": 959 + }, + { + "epoch": 0.2608695652173913, + "grad_norm": 2.1298911571502686, + "learning_rate": 3e-05, + "loss": 3.4676, + "step": 960 + }, + { + "epoch": 0.26114130434782606, + "grad_norm": 1.9788066148757935, + "learning_rate": 3e-05, + "loss": 3.4239, + "step": 961 + }, + { + "epoch": 0.2614130434782609, + "grad_norm": 1.6203289031982422, + "learning_rate": 3e-05, + "loss": 3.2623, + "step": 962 + }, + { + "epoch": 0.26168478260869565, + "grad_norm": 1.7709441184997559, + "learning_rate": 3e-05, + "loss": 3.5855, + "step": 963 + }, + { + "epoch": 0.2619565217391304, + "grad_norm": 1.4671415090560913, + "learning_rate": 3e-05, + "loss": 3.2081, + "step": 964 + }, + { + "epoch": 0.26222826086956524, + "grad_norm": 1.2123912572860718, + "learning_rate": 3e-05, + "loss": 3.0287, + "step": 965 + }, + { + "epoch": 0.2625, + "grad_norm": 1.9638895988464355, + "learning_rate": 3e-05, + "loss": 3.4101, + "step": 966 + }, + { + "epoch": 0.2627717391304348, + "grad_norm": 1.6470272541046143, + "learning_rate": 3e-05, + "loss": 3.1761, + "step": 967 + }, + { + "epoch": 0.26304347826086955, + "grad_norm": 1.7481547594070435, + "learning_rate": 3e-05, + "loss": 3.7102, + "step": 968 + }, + { + "epoch": 0.26331521739130437, + "grad_norm": 1.589220404624939, + "learning_rate": 3e-05, + "loss": 2.9376, + "step": 969 + }, + { + "epoch": 0.26358695652173914, + "grad_norm": 2.317216634750366, + "learning_rate": 3e-05, + "loss": 3.888, + "step": 970 + }, + { + "epoch": 0.2638586956521739, + "grad_norm": 1.9831353425979614, + "learning_rate": 3e-05, + "loss": 3.831, + "step": 971 + }, + { + "epoch": 0.26413043478260867, + "grad_norm": 2.201885223388672, + "learning_rate": 3e-05, + "loss": 3.5633, + "step": 972 + }, + { + "epoch": 0.2644021739130435, + "grad_norm": 1.755304217338562, + "learning_rate": 3e-05, + "loss": 3.9027, + "step": 973 + }, + { + "epoch": 0.26467391304347826, + "grad_norm": 1.7514090538024902, + "learning_rate": 3e-05, + "loss": 3.3256, + "step": 974 + }, + { + "epoch": 0.264945652173913, + "grad_norm": 2.1359901428222656, + "learning_rate": 3e-05, + "loss": 3.9904, + "step": 975 + }, + { + "epoch": 0.26521739130434785, + "grad_norm": 1.5768872499465942, + "learning_rate": 3e-05, + "loss": 3.3242, + "step": 976 + }, + { + "epoch": 0.2654891304347826, + "grad_norm": 1.8486355543136597, + "learning_rate": 3e-05, + "loss": 3.7133, + "step": 977 + }, + { + "epoch": 0.2657608695652174, + "grad_norm": 1.5333871841430664, + "learning_rate": 3e-05, + "loss": 3.4191, + "step": 978 + }, + { + "epoch": 0.26603260869565215, + "grad_norm": 1.6924866437911987, + "learning_rate": 3e-05, + "loss": 3.3763, + "step": 979 + }, + { + "epoch": 0.266304347826087, + "grad_norm": 2.1424450874328613, + "learning_rate": 3e-05, + "loss": 3.8962, + "step": 980 + }, + { + "epoch": 0.26657608695652174, + "grad_norm": 1.5438109636306763, + "learning_rate": 3e-05, + "loss": 3.2176, + "step": 981 + }, + { + "epoch": 0.2668478260869565, + "grad_norm": 1.5958982706069946, + "learning_rate": 3e-05, + "loss": 3.8708, + "step": 982 + }, + { + "epoch": 0.26711956521739133, + "grad_norm": 1.129310965538025, + "learning_rate": 3e-05, + "loss": 2.8674, + "step": 983 + }, + { + "epoch": 0.2673913043478261, + "grad_norm": 1.6280345916748047, + "learning_rate": 3e-05, + "loss": 4.165, + "step": 984 + }, + { + "epoch": 0.26766304347826086, + "grad_norm": 1.499475121498108, + "learning_rate": 3e-05, + "loss": 2.9953, + "step": 985 + }, + { + "epoch": 0.26793478260869563, + "grad_norm": 1.253732442855835, + "learning_rate": 3e-05, + "loss": 3.4718, + "step": 986 + }, + { + "epoch": 0.26820652173913045, + "grad_norm": 1.113315463066101, + "learning_rate": 3e-05, + "loss": 3.3101, + "step": 987 + }, + { + "epoch": 0.2684782608695652, + "grad_norm": 1.4947681427001953, + "learning_rate": 3e-05, + "loss": 3.217, + "step": 988 + }, + { + "epoch": 0.26875, + "grad_norm": 1.9959630966186523, + "learning_rate": 3e-05, + "loss": 4.0953, + "step": 989 + }, + { + "epoch": 0.26902173913043476, + "grad_norm": 1.6297112703323364, + "learning_rate": 3e-05, + "loss": 3.4108, + "step": 990 + }, + { + "epoch": 0.2692934782608696, + "grad_norm": 1.9028682708740234, + "learning_rate": 3e-05, + "loss": 4.1301, + "step": 991 + }, + { + "epoch": 0.26956521739130435, + "grad_norm": 1.6655460596084595, + "learning_rate": 3e-05, + "loss": 3.5511, + "step": 992 + }, + { + "epoch": 0.2698369565217391, + "grad_norm": 1.8980700969696045, + "learning_rate": 3e-05, + "loss": 3.5143, + "step": 993 + }, + { + "epoch": 0.27010869565217394, + "grad_norm": 1.5208625793457031, + "learning_rate": 3e-05, + "loss": 3.5514, + "step": 994 + }, + { + "epoch": 0.2703804347826087, + "grad_norm": 1.1632534265518188, + "learning_rate": 3e-05, + "loss": 3.2167, + "step": 995 + }, + { + "epoch": 0.27065217391304347, + "grad_norm": 2.094667673110962, + "learning_rate": 3e-05, + "loss": 3.0362, + "step": 996 + }, + { + "epoch": 0.27092391304347824, + "grad_norm": 1.4889837503433228, + "learning_rate": 3e-05, + "loss": 3.4279, + "step": 997 + }, + { + "epoch": 0.27119565217391306, + "grad_norm": 1.9965155124664307, + "learning_rate": 3e-05, + "loss": 3.774, + "step": 998 + }, + { + "epoch": 0.2714673913043478, + "grad_norm": 1.5410970449447632, + "learning_rate": 3e-05, + "loss": 3.2647, + "step": 999 + }, + { + "epoch": 0.2717391304347826, + "grad_norm": 1.9614113569259644, + "learning_rate": 3e-05, + "loss": 3.5648, + "step": 1000 + }, + { + "epoch": 0.2720108695652174, + "grad_norm": 1.3656377792358398, + "learning_rate": 3e-05, + "loss": 3.3263, + "step": 1001 + }, + { + "epoch": 0.2722826086956522, + "grad_norm": 2.057823419570923, + "learning_rate": 3e-05, + "loss": 3.1952, + "step": 1002 + }, + { + "epoch": 0.27255434782608695, + "grad_norm": 1.9812930822372437, + "learning_rate": 3e-05, + "loss": 3.5128, + "step": 1003 + }, + { + "epoch": 0.2728260869565217, + "grad_norm": 1.443973183631897, + "learning_rate": 3e-05, + "loss": 3.5108, + "step": 1004 + }, + { + "epoch": 0.27309782608695654, + "grad_norm": 1.1789878606796265, + "learning_rate": 3e-05, + "loss": 3.142, + "step": 1005 + }, + { + "epoch": 0.2733695652173913, + "grad_norm": 1.5677335262298584, + "learning_rate": 3e-05, + "loss": 3.0946, + "step": 1006 + }, + { + "epoch": 0.2736413043478261, + "grad_norm": 1.6985656023025513, + "learning_rate": 3e-05, + "loss": 3.8017, + "step": 1007 + }, + { + "epoch": 0.27391304347826084, + "grad_norm": 0.9420762658119202, + "learning_rate": 3e-05, + "loss": 2.8754, + "step": 1008 + }, + { + "epoch": 0.27418478260869567, + "grad_norm": 1.1210711002349854, + "learning_rate": 3e-05, + "loss": 3.1734, + "step": 1009 + }, + { + "epoch": 0.27445652173913043, + "grad_norm": 0.9131266474723816, + "learning_rate": 3e-05, + "loss": 2.5223, + "step": 1010 + }, + { + "epoch": 0.2747282608695652, + "grad_norm": 1.4077900648117065, + "learning_rate": 3e-05, + "loss": 3.2054, + "step": 1011 + }, + { + "epoch": 0.275, + "grad_norm": 1.4042190313339233, + "learning_rate": 3e-05, + "loss": 3.3916, + "step": 1012 + }, + { + "epoch": 0.2752717391304348, + "grad_norm": 1.1815489530563354, + "learning_rate": 3e-05, + "loss": 3.2108, + "step": 1013 + }, + { + "epoch": 0.27554347826086956, + "grad_norm": 1.007114052772522, + "learning_rate": 3e-05, + "loss": 2.8035, + "step": 1014 + }, + { + "epoch": 0.2758152173913043, + "grad_norm": 1.6630133390426636, + "learning_rate": 3e-05, + "loss": 3.5364, + "step": 1015 + }, + { + "epoch": 0.27608695652173915, + "grad_norm": 1.3151155710220337, + "learning_rate": 3e-05, + "loss": 3.8982, + "step": 1016 + }, + { + "epoch": 0.2763586956521739, + "grad_norm": 1.3191099166870117, + "learning_rate": 3e-05, + "loss": 3.3812, + "step": 1017 + }, + { + "epoch": 0.2766304347826087, + "grad_norm": 0.9553658366203308, + "learning_rate": 3e-05, + "loss": 3.2285, + "step": 1018 + }, + { + "epoch": 0.2769021739130435, + "grad_norm": 1.2714715003967285, + "learning_rate": 3e-05, + "loss": 4.1331, + "step": 1019 + }, + { + "epoch": 0.27717391304347827, + "grad_norm": 1.4878464937210083, + "learning_rate": 3e-05, + "loss": 3.5607, + "step": 1020 + }, + { + "epoch": 0.27744565217391304, + "grad_norm": 1.3467168807983398, + "learning_rate": 3e-05, + "loss": 3.4943, + "step": 1021 + }, + { + "epoch": 0.2777173913043478, + "grad_norm": 1.052966833114624, + "learning_rate": 3e-05, + "loss": 2.9399, + "step": 1022 + }, + { + "epoch": 0.2779891304347826, + "grad_norm": 1.1948199272155762, + "learning_rate": 3e-05, + "loss": 3.0152, + "step": 1023 + }, + { + "epoch": 0.2782608695652174, + "grad_norm": 1.1960163116455078, + "learning_rate": 3e-05, + "loss": 2.9338, + "step": 1024 + }, + { + "epoch": 0.27853260869565216, + "grad_norm": 1.2721116542816162, + "learning_rate": 3e-05, + "loss": 3.2096, + "step": 1025 + }, + { + "epoch": 0.27880434782608693, + "grad_norm": 1.620215654373169, + "learning_rate": 3e-05, + "loss": 3.3828, + "step": 1026 + }, + { + "epoch": 0.27907608695652175, + "grad_norm": 1.4918383359909058, + "learning_rate": 3e-05, + "loss": 3.5746, + "step": 1027 + }, + { + "epoch": 0.2793478260869565, + "grad_norm": 1.9811660051345825, + "learning_rate": 3e-05, + "loss": 4.0559, + "step": 1028 + }, + { + "epoch": 0.2796195652173913, + "grad_norm": 1.3812586069107056, + "learning_rate": 3e-05, + "loss": 3.1452, + "step": 1029 + }, + { + "epoch": 0.2798913043478261, + "grad_norm": 1.53925359249115, + "learning_rate": 3e-05, + "loss": 3.9887, + "step": 1030 + }, + { + "epoch": 0.2801630434782609, + "grad_norm": 2.2425036430358887, + "learning_rate": 3e-05, + "loss": 4.2516, + "step": 1031 + }, + { + "epoch": 0.28043478260869564, + "grad_norm": 1.2638776302337646, + "learning_rate": 3e-05, + "loss": 3.5207, + "step": 1032 + }, + { + "epoch": 0.2807065217391304, + "grad_norm": 1.1238954067230225, + "learning_rate": 3e-05, + "loss": 3.4072, + "step": 1033 + }, + { + "epoch": 0.28097826086956523, + "grad_norm": 1.2834945917129517, + "learning_rate": 3e-05, + "loss": 3.0585, + "step": 1034 + }, + { + "epoch": 0.28125, + "grad_norm": 1.5284252166748047, + "learning_rate": 3e-05, + "loss": 3.1547, + "step": 1035 + }, + { + "epoch": 0.28152173913043477, + "grad_norm": 1.8825403451919556, + "learning_rate": 3e-05, + "loss": 3.892, + "step": 1036 + }, + { + "epoch": 0.2817934782608696, + "grad_norm": 1.620376467704773, + "learning_rate": 3e-05, + "loss": 2.9044, + "step": 1037 + }, + { + "epoch": 0.28206521739130436, + "grad_norm": 2.1728601455688477, + "learning_rate": 3e-05, + "loss": 3.1266, + "step": 1038 + }, + { + "epoch": 0.2823369565217391, + "grad_norm": 2.090522050857544, + "learning_rate": 3e-05, + "loss": 4.4403, + "step": 1039 + }, + { + "epoch": 0.2826086956521739, + "grad_norm": 2.0362656116485596, + "learning_rate": 3e-05, + "loss": 3.6416, + "step": 1040 + }, + { + "epoch": 0.2828804347826087, + "grad_norm": 1.493579387664795, + "learning_rate": 3e-05, + "loss": 3.2111, + "step": 1041 + }, + { + "epoch": 0.2831521739130435, + "grad_norm": 1.2144675254821777, + "learning_rate": 3e-05, + "loss": 2.911, + "step": 1042 + }, + { + "epoch": 0.28342391304347825, + "grad_norm": 1.833510398864746, + "learning_rate": 3e-05, + "loss": 3.0022, + "step": 1043 + }, + { + "epoch": 0.28369565217391307, + "grad_norm": 2.0041656494140625, + "learning_rate": 3e-05, + "loss": 3.1412, + "step": 1044 + }, + { + "epoch": 0.28396739130434784, + "grad_norm": 1.2218122482299805, + "learning_rate": 3e-05, + "loss": 3.4581, + "step": 1045 + }, + { + "epoch": 0.2842391304347826, + "grad_norm": 1.6633751392364502, + "learning_rate": 3e-05, + "loss": 3.6047, + "step": 1046 + }, + { + "epoch": 0.2845108695652174, + "grad_norm": 2.4305849075317383, + "learning_rate": 3e-05, + "loss": 3.3797, + "step": 1047 + }, + { + "epoch": 0.2847826086956522, + "grad_norm": 2.309483289718628, + "learning_rate": 3e-05, + "loss": 3.6304, + "step": 1048 + }, + { + "epoch": 0.28505434782608696, + "grad_norm": 1.678916335105896, + "learning_rate": 3e-05, + "loss": 3.4073, + "step": 1049 + }, + { + "epoch": 0.28532608695652173, + "grad_norm": 1.9873173236846924, + "learning_rate": 3e-05, + "loss": 3.9298, + "step": 1050 + }, + { + "epoch": 0.2855978260869565, + "grad_norm": 2.085996389389038, + "learning_rate": 3e-05, + "loss": 4.4263, + "step": 1051 + }, + { + "epoch": 0.2858695652173913, + "grad_norm": 3.049558639526367, + "learning_rate": 3e-05, + "loss": 3.6507, + "step": 1052 + }, + { + "epoch": 0.2861413043478261, + "grad_norm": 1.4257125854492188, + "learning_rate": 3e-05, + "loss": 3.3861, + "step": 1053 + }, + { + "epoch": 0.28641304347826085, + "grad_norm": 1.5436605215072632, + "learning_rate": 3e-05, + "loss": 3.1414, + "step": 1054 + }, + { + "epoch": 0.2866847826086957, + "grad_norm": 3.0967252254486084, + "learning_rate": 3e-05, + "loss": 3.9067, + "step": 1055 + }, + { + "epoch": 0.28695652173913044, + "grad_norm": 1.6564278602600098, + "learning_rate": 3e-05, + "loss": 3.5354, + "step": 1056 + }, + { + "epoch": 0.2872282608695652, + "grad_norm": 2.498901605606079, + "learning_rate": 3e-05, + "loss": 3.3153, + "step": 1057 + }, + { + "epoch": 0.2875, + "grad_norm": 2.2996490001678467, + "learning_rate": 3e-05, + "loss": 3.4023, + "step": 1058 + }, + { + "epoch": 0.2877717391304348, + "grad_norm": 2.0972297191619873, + "learning_rate": 3e-05, + "loss": 3.038, + "step": 1059 + }, + { + "epoch": 0.28804347826086957, + "grad_norm": 1.221664547920227, + "learning_rate": 3e-05, + "loss": 3.0761, + "step": 1060 + }, + { + "epoch": 0.28831521739130433, + "grad_norm": 2.2344908714294434, + "learning_rate": 3e-05, + "loss": 2.8314, + "step": 1061 + }, + { + "epoch": 0.28858695652173916, + "grad_norm": 1.9931285381317139, + "learning_rate": 3e-05, + "loss": 3.3899, + "step": 1062 + }, + { + "epoch": 0.2888586956521739, + "grad_norm": 1.924612045288086, + "learning_rate": 3e-05, + "loss": 3.4548, + "step": 1063 + }, + { + "epoch": 0.2891304347826087, + "grad_norm": 1.9507229328155518, + "learning_rate": 3e-05, + "loss": 3.9302, + "step": 1064 + }, + { + "epoch": 0.28940217391304346, + "grad_norm": 1.8761188983917236, + "learning_rate": 3e-05, + "loss": 3.195, + "step": 1065 + }, + { + "epoch": 0.2896739130434783, + "grad_norm": 2.233293294906616, + "learning_rate": 3e-05, + "loss": 3.1672, + "step": 1066 + }, + { + "epoch": 0.28994565217391305, + "grad_norm": 1.7154324054718018, + "learning_rate": 3e-05, + "loss": 3.1413, + "step": 1067 + }, + { + "epoch": 0.2902173913043478, + "grad_norm": 1.8050798177719116, + "learning_rate": 3e-05, + "loss": 3.4341, + "step": 1068 + }, + { + "epoch": 0.2904891304347826, + "grad_norm": 1.838165044784546, + "learning_rate": 3e-05, + "loss": 3.3827, + "step": 1069 + }, + { + "epoch": 0.2907608695652174, + "grad_norm": 2.7229554653167725, + "learning_rate": 3e-05, + "loss": 3.3938, + "step": 1070 + }, + { + "epoch": 0.2910326086956522, + "grad_norm": 3.150635242462158, + "learning_rate": 3e-05, + "loss": 3.8396, + "step": 1071 + }, + { + "epoch": 0.29130434782608694, + "grad_norm": 2.2429847717285156, + "learning_rate": 3e-05, + "loss": 3.9437, + "step": 1072 + }, + { + "epoch": 0.29157608695652176, + "grad_norm": 1.305802345275879, + "learning_rate": 3e-05, + "loss": 3.5962, + "step": 1073 + }, + { + "epoch": 0.29184782608695653, + "grad_norm": 1.4509013891220093, + "learning_rate": 3e-05, + "loss": 3.2975, + "step": 1074 + }, + { + "epoch": 0.2921195652173913, + "grad_norm": 2.059382200241089, + "learning_rate": 3e-05, + "loss": 4.21, + "step": 1075 + }, + { + "epoch": 0.29239130434782606, + "grad_norm": 0.9824801087379456, + "learning_rate": 3e-05, + "loss": 2.6315, + "step": 1076 + }, + { + "epoch": 0.2926630434782609, + "grad_norm": 1.0867946147918701, + "learning_rate": 3e-05, + "loss": 2.6838, + "step": 1077 + }, + { + "epoch": 0.29293478260869565, + "grad_norm": 1.992459774017334, + "learning_rate": 3e-05, + "loss": 4.2482, + "step": 1078 + }, + { + "epoch": 0.2932065217391304, + "grad_norm": 1.6963497400283813, + "learning_rate": 3e-05, + "loss": 3.5684, + "step": 1079 + }, + { + "epoch": 0.29347826086956524, + "grad_norm": 1.2850168943405151, + "learning_rate": 3e-05, + "loss": 3.399, + "step": 1080 + }, + { + "epoch": 0.29375, + "grad_norm": 1.652712106704712, + "learning_rate": 3e-05, + "loss": 3.3213, + "step": 1081 + }, + { + "epoch": 0.2940217391304348, + "grad_norm": 1.733959436416626, + "learning_rate": 3e-05, + "loss": 3.1942, + "step": 1082 + }, + { + "epoch": 0.29429347826086955, + "grad_norm": 1.8719747066497803, + "learning_rate": 3e-05, + "loss": 3.8789, + "step": 1083 + }, + { + "epoch": 0.29456521739130437, + "grad_norm": 1.5426567792892456, + "learning_rate": 3e-05, + "loss": 3.1319, + "step": 1084 + }, + { + "epoch": 0.29483695652173914, + "grad_norm": 2.0286598205566406, + "learning_rate": 3e-05, + "loss": 3.9002, + "step": 1085 + }, + { + "epoch": 0.2951086956521739, + "grad_norm": 1.5031648874282837, + "learning_rate": 3e-05, + "loss": 3.3979, + "step": 1086 + }, + { + "epoch": 0.29538043478260867, + "grad_norm": 1.556738257408142, + "learning_rate": 3e-05, + "loss": 3.9843, + "step": 1087 + }, + { + "epoch": 0.2956521739130435, + "grad_norm": 1.8954527378082275, + "learning_rate": 3e-05, + "loss": 4.6858, + "step": 1088 + }, + { + "epoch": 0.29592391304347826, + "grad_norm": 1.7330039739608765, + "learning_rate": 3e-05, + "loss": 2.8431, + "step": 1089 + }, + { + "epoch": 0.296195652173913, + "grad_norm": 2.0785205364227295, + "learning_rate": 3e-05, + "loss": 3.6168, + "step": 1090 + }, + { + "epoch": 0.29646739130434785, + "grad_norm": 1.708628535270691, + "learning_rate": 3e-05, + "loss": 3.6387, + "step": 1091 + }, + { + "epoch": 0.2967391304347826, + "grad_norm": 2.5519232749938965, + "learning_rate": 3e-05, + "loss": 4.0809, + "step": 1092 + }, + { + "epoch": 0.2970108695652174, + "grad_norm": 1.5967369079589844, + "learning_rate": 3e-05, + "loss": 3.5908, + "step": 1093 + }, + { + "epoch": 0.29728260869565215, + "grad_norm": 1.345123052597046, + "learning_rate": 3e-05, + "loss": 3.297, + "step": 1094 + }, + { + "epoch": 0.297554347826087, + "grad_norm": 2.123605251312256, + "learning_rate": 3e-05, + "loss": 4.1267, + "step": 1095 + }, + { + "epoch": 0.29782608695652174, + "grad_norm": 1.6863003969192505, + "learning_rate": 3e-05, + "loss": 3.4321, + "step": 1096 + }, + { + "epoch": 0.2980978260869565, + "grad_norm": 2.0726163387298584, + "learning_rate": 3e-05, + "loss": 3.386, + "step": 1097 + }, + { + "epoch": 0.29836956521739133, + "grad_norm": 1.5322779417037964, + "learning_rate": 3e-05, + "loss": 2.9161, + "step": 1098 + }, + { + "epoch": 0.2986413043478261, + "grad_norm": 1.1131296157836914, + "learning_rate": 3e-05, + "loss": 2.7584, + "step": 1099 + }, + { + "epoch": 0.29891304347826086, + "grad_norm": 1.9430760145187378, + "learning_rate": 3e-05, + "loss": 3.4661, + "step": 1100 + }, + { + "epoch": 0.29918478260869563, + "grad_norm": 2.6745107173919678, + "learning_rate": 3e-05, + "loss": 3.0976, + "step": 1101 + }, + { + "epoch": 0.29945652173913045, + "grad_norm": 3.127723217010498, + "learning_rate": 3e-05, + "loss": 3.7013, + "step": 1102 + }, + { + "epoch": 0.2997282608695652, + "grad_norm": 1.894990086555481, + "learning_rate": 3e-05, + "loss": 3.4841, + "step": 1103 + }, + { + "epoch": 0.3, + "grad_norm": 1.9465398788452148, + "learning_rate": 3e-05, + "loss": 3.48, + "step": 1104 + }, + { + "epoch": 0.30027173913043476, + "grad_norm": 2.5459132194519043, + "learning_rate": 3e-05, + "loss": 3.239, + "step": 1105 + }, + { + "epoch": 0.3005434782608696, + "grad_norm": 2.382801055908203, + "learning_rate": 3e-05, + "loss": 3.2947, + "step": 1106 + }, + { + "epoch": 0.30081521739130435, + "grad_norm": 1.913734793663025, + "learning_rate": 3e-05, + "loss": 3.0162, + "step": 1107 + }, + { + "epoch": 0.3010869565217391, + "grad_norm": 1.1874632835388184, + "learning_rate": 3e-05, + "loss": 2.6723, + "step": 1108 + }, + { + "epoch": 0.30135869565217394, + "grad_norm": 1.6929433345794678, + "learning_rate": 3e-05, + "loss": 3.1864, + "step": 1109 + }, + { + "epoch": 0.3016304347826087, + "grad_norm": 2.1046361923217773, + "learning_rate": 3e-05, + "loss": 3.442, + "step": 1110 + }, + { + "epoch": 0.30190217391304347, + "grad_norm": 2.196091413497925, + "learning_rate": 3e-05, + "loss": 3.2791, + "step": 1111 + }, + { + "epoch": 0.30217391304347824, + "grad_norm": 2.6548757553100586, + "learning_rate": 3e-05, + "loss": 4.0965, + "step": 1112 + }, + { + "epoch": 0.30244565217391306, + "grad_norm": 1.7033164501190186, + "learning_rate": 3e-05, + "loss": 3.5031, + "step": 1113 + }, + { + "epoch": 0.3027173913043478, + "grad_norm": 1.5866395235061646, + "learning_rate": 3e-05, + "loss": 2.9015, + "step": 1114 + }, + { + "epoch": 0.3029891304347826, + "grad_norm": 2.0246646404266357, + "learning_rate": 3e-05, + "loss": 3.1981, + "step": 1115 + }, + { + "epoch": 0.3032608695652174, + "grad_norm": 2.0733206272125244, + "learning_rate": 3e-05, + "loss": 3.3524, + "step": 1116 + }, + { + "epoch": 0.3035326086956522, + "grad_norm": 1.6320792436599731, + "learning_rate": 3e-05, + "loss": 3.6667, + "step": 1117 + }, + { + "epoch": 0.30380434782608695, + "grad_norm": 1.4059102535247803, + "learning_rate": 3e-05, + "loss": 3.4399, + "step": 1118 + }, + { + "epoch": 0.3040760869565217, + "grad_norm": 1.6570987701416016, + "learning_rate": 3e-05, + "loss": 3.2745, + "step": 1119 + }, + { + "epoch": 0.30434782608695654, + "grad_norm": 1.9712703227996826, + "learning_rate": 3e-05, + "loss": 3.751, + "step": 1120 + }, + { + "epoch": 0.3046195652173913, + "grad_norm": 1.4697049856185913, + "learning_rate": 3e-05, + "loss": 3.4402, + "step": 1121 + }, + { + "epoch": 0.3048913043478261, + "grad_norm": 1.8021398782730103, + "learning_rate": 3e-05, + "loss": 3.8907, + "step": 1122 + }, + { + "epoch": 0.30516304347826084, + "grad_norm": 1.5529664754867554, + "learning_rate": 3e-05, + "loss": 3.7741, + "step": 1123 + }, + { + "epoch": 0.30543478260869567, + "grad_norm": 1.5865482091903687, + "learning_rate": 3e-05, + "loss": 3.7243, + "step": 1124 + }, + { + "epoch": 0.30570652173913043, + "grad_norm": 2.221069097518921, + "learning_rate": 3e-05, + "loss": 3.65, + "step": 1125 + }, + { + "epoch": 0.3059782608695652, + "grad_norm": 2.1728017330169678, + "learning_rate": 3e-05, + "loss": 3.0357, + "step": 1126 + }, + { + "epoch": 0.30625, + "grad_norm": 1.6755625009536743, + "learning_rate": 3e-05, + "loss": 3.2292, + "step": 1127 + }, + { + "epoch": 0.3065217391304348, + "grad_norm": 2.135910987854004, + "learning_rate": 3e-05, + "loss": 4.0064, + "step": 1128 + }, + { + "epoch": 0.30679347826086956, + "grad_norm": 1.7224137783050537, + "learning_rate": 3e-05, + "loss": 3.6076, + "step": 1129 + }, + { + "epoch": 0.3070652173913043, + "grad_norm": 1.5196242332458496, + "learning_rate": 3e-05, + "loss": 3.2132, + "step": 1130 + }, + { + "epoch": 0.30733695652173915, + "grad_norm": 2.016763210296631, + "learning_rate": 3e-05, + "loss": 4.1373, + "step": 1131 + }, + { + "epoch": 0.3076086956521739, + "grad_norm": 1.3778748512268066, + "learning_rate": 3e-05, + "loss": 3.4815, + "step": 1132 + }, + { + "epoch": 0.3078804347826087, + "grad_norm": 1.5161099433898926, + "learning_rate": 3e-05, + "loss": 3.2094, + "step": 1133 + }, + { + "epoch": 0.3081521739130435, + "grad_norm": 1.2191746234893799, + "learning_rate": 3e-05, + "loss": 3.0117, + "step": 1134 + }, + { + "epoch": 0.30842391304347827, + "grad_norm": 1.6270561218261719, + "learning_rate": 3e-05, + "loss": 3.8738, + "step": 1135 + }, + { + "epoch": 0.30869565217391304, + "grad_norm": 1.3209682703018188, + "learning_rate": 3e-05, + "loss": 3.1873, + "step": 1136 + }, + { + "epoch": 0.3089673913043478, + "grad_norm": 1.7960654497146606, + "learning_rate": 3e-05, + "loss": 3.9211, + "step": 1137 + }, + { + "epoch": 0.3092391304347826, + "grad_norm": 1.5645345449447632, + "learning_rate": 3e-05, + "loss": 3.8176, + "step": 1138 + }, + { + "epoch": 0.3095108695652174, + "grad_norm": 1.2274796962738037, + "learning_rate": 3e-05, + "loss": 3.0552, + "step": 1139 + }, + { + "epoch": 0.30978260869565216, + "grad_norm": 3.1637630462646484, + "learning_rate": 3e-05, + "loss": 4.0908, + "step": 1140 + }, + { + "epoch": 0.31005434782608693, + "grad_norm": 1.7452456951141357, + "learning_rate": 3e-05, + "loss": 3.2379, + "step": 1141 + }, + { + "epoch": 0.31032608695652175, + "grad_norm": 1.3035019636154175, + "learning_rate": 3e-05, + "loss": 3.1619, + "step": 1142 + }, + { + "epoch": 0.3105978260869565, + "grad_norm": 1.5721486806869507, + "learning_rate": 3e-05, + "loss": 3.3092, + "step": 1143 + }, + { + "epoch": 0.3108695652173913, + "grad_norm": 1.5714532136917114, + "learning_rate": 3e-05, + "loss": 3.7343, + "step": 1144 + }, + { + "epoch": 0.3111413043478261, + "grad_norm": 1.4483362436294556, + "learning_rate": 3e-05, + "loss": 3.5843, + "step": 1145 + }, + { + "epoch": 0.3114130434782609, + "grad_norm": 1.2886680364608765, + "learning_rate": 3e-05, + "loss": 3.4743, + "step": 1146 + }, + { + "epoch": 0.31168478260869564, + "grad_norm": 1.5609358549118042, + "learning_rate": 3e-05, + "loss": 3.7947, + "step": 1147 + }, + { + "epoch": 0.3119565217391304, + "grad_norm": 1.7903861999511719, + "learning_rate": 3e-05, + "loss": 3.3165, + "step": 1148 + }, + { + "epoch": 0.31222826086956523, + "grad_norm": 1.917932152748108, + "learning_rate": 3e-05, + "loss": 3.3985, + "step": 1149 + }, + { + "epoch": 0.3125, + "grad_norm": 1.5185085535049438, + "learning_rate": 3e-05, + "loss": 3.571, + "step": 1150 + }, + { + "epoch": 0.31277173913043477, + "grad_norm": 1.3498207330703735, + "learning_rate": 3e-05, + "loss": 3.3921, + "step": 1151 + }, + { + "epoch": 0.3130434782608696, + "grad_norm": 1.6622629165649414, + "learning_rate": 3e-05, + "loss": 3.4101, + "step": 1152 + }, + { + "epoch": 0.31331521739130436, + "grad_norm": 1.516265630722046, + "learning_rate": 3e-05, + "loss": 3.1228, + "step": 1153 + }, + { + "epoch": 0.3135869565217391, + "grad_norm": 1.5660970211029053, + "learning_rate": 3e-05, + "loss": 3.3751, + "step": 1154 + }, + { + "epoch": 0.3138586956521739, + "grad_norm": 1.8406776189804077, + "learning_rate": 3e-05, + "loss": 3.3181, + "step": 1155 + }, + { + "epoch": 0.3141304347826087, + "grad_norm": 1.8162765502929688, + "learning_rate": 3e-05, + "loss": 4.0117, + "step": 1156 + }, + { + "epoch": 0.3144021739130435, + "grad_norm": 1.2983214855194092, + "learning_rate": 3e-05, + "loss": 3.2412, + "step": 1157 + }, + { + "epoch": 0.31467391304347825, + "grad_norm": 1.1409294605255127, + "learning_rate": 3e-05, + "loss": 3.0329, + "step": 1158 + }, + { + "epoch": 0.31494565217391307, + "grad_norm": 1.754345178604126, + "learning_rate": 3e-05, + "loss": 3.6711, + "step": 1159 + }, + { + "epoch": 0.31521739130434784, + "grad_norm": 1.868580937385559, + "learning_rate": 3e-05, + "loss": 3.6684, + "step": 1160 + }, + { + "epoch": 0.3154891304347826, + "grad_norm": 1.8012367486953735, + "learning_rate": 3e-05, + "loss": 3.2884, + "step": 1161 + }, + { + "epoch": 0.3157608695652174, + "grad_norm": 1.7688064575195312, + "learning_rate": 3e-05, + "loss": 3.3512, + "step": 1162 + }, + { + "epoch": 0.3160326086956522, + "grad_norm": 1.5674185752868652, + "learning_rate": 3e-05, + "loss": 3.443, + "step": 1163 + }, + { + "epoch": 0.31630434782608696, + "grad_norm": 1.5843913555145264, + "learning_rate": 3e-05, + "loss": 3.8519, + "step": 1164 + }, + { + "epoch": 0.31657608695652173, + "grad_norm": 1.4522455930709839, + "learning_rate": 3e-05, + "loss": 3.123, + "step": 1165 + }, + { + "epoch": 0.3168478260869565, + "grad_norm": 1.5610785484313965, + "learning_rate": 3e-05, + "loss": 3.1963, + "step": 1166 + }, + { + "epoch": 0.3171195652173913, + "grad_norm": 1.4327051639556885, + "learning_rate": 3e-05, + "loss": 2.9382, + "step": 1167 + }, + { + "epoch": 0.3173913043478261, + "grad_norm": 1.6065129041671753, + "learning_rate": 3e-05, + "loss": 3.4928, + "step": 1168 + }, + { + "epoch": 0.31766304347826085, + "grad_norm": 1.9382498264312744, + "learning_rate": 3e-05, + "loss": 4.3174, + "step": 1169 + }, + { + "epoch": 0.3179347826086957, + "grad_norm": 1.164689064025879, + "learning_rate": 3e-05, + "loss": 3.0678, + "step": 1170 + }, + { + "epoch": 0.31820652173913044, + "grad_norm": 1.084055781364441, + "learning_rate": 3e-05, + "loss": 3.0878, + "step": 1171 + }, + { + "epoch": 0.3184782608695652, + "grad_norm": 1.4881168603897095, + "learning_rate": 3e-05, + "loss": 3.5084, + "step": 1172 + }, + { + "epoch": 0.31875, + "grad_norm": 1.5508403778076172, + "learning_rate": 3e-05, + "loss": 3.6359, + "step": 1173 + }, + { + "epoch": 0.3190217391304348, + "grad_norm": 1.7135038375854492, + "learning_rate": 3e-05, + "loss": 2.819, + "step": 1174 + }, + { + "epoch": 0.31929347826086957, + "grad_norm": 1.9360867738723755, + "learning_rate": 3e-05, + "loss": 3.711, + "step": 1175 + }, + { + "epoch": 0.31956521739130433, + "grad_norm": 1.3898394107818604, + "learning_rate": 3e-05, + "loss": 3.0356, + "step": 1176 + }, + { + "epoch": 0.31983695652173916, + "grad_norm": 1.37100088596344, + "learning_rate": 3e-05, + "loss": 2.9441, + "step": 1177 + }, + { + "epoch": 0.3201086956521739, + "grad_norm": 1.7066446542739868, + "learning_rate": 3e-05, + "loss": 3.8696, + "step": 1178 + }, + { + "epoch": 0.3203804347826087, + "grad_norm": 1.4085134267807007, + "learning_rate": 3e-05, + "loss": 3.4616, + "step": 1179 + }, + { + "epoch": 0.32065217391304346, + "grad_norm": 1.5345913171768188, + "learning_rate": 3e-05, + "loss": 3.6033, + "step": 1180 + }, + { + "epoch": 0.3209239130434783, + "grad_norm": 1.5240819454193115, + "learning_rate": 3e-05, + "loss": 3.1863, + "step": 1181 + }, + { + "epoch": 0.32119565217391305, + "grad_norm": 1.147559404373169, + "learning_rate": 3e-05, + "loss": 3.2533, + "step": 1182 + }, + { + "epoch": 0.3214673913043478, + "grad_norm": 1.3267796039581299, + "learning_rate": 3e-05, + "loss": 3.3342, + "step": 1183 + }, + { + "epoch": 0.3217391304347826, + "grad_norm": 1.567126989364624, + "learning_rate": 3e-05, + "loss": 3.5055, + "step": 1184 + }, + { + "epoch": 0.3220108695652174, + "grad_norm": 2.085658073425293, + "learning_rate": 3e-05, + "loss": 3.1383, + "step": 1185 + }, + { + "epoch": 0.3222826086956522, + "grad_norm": 1.7624791860580444, + "learning_rate": 3e-05, + "loss": 3.2409, + "step": 1186 + }, + { + "epoch": 0.32255434782608694, + "grad_norm": 1.4368826150894165, + "learning_rate": 3e-05, + "loss": 3.7424, + "step": 1187 + }, + { + "epoch": 0.32282608695652176, + "grad_norm": 1.5510128736495972, + "learning_rate": 3e-05, + "loss": 3.3895, + "step": 1188 + }, + { + "epoch": 0.32309782608695653, + "grad_norm": 1.471490740776062, + "learning_rate": 3e-05, + "loss": 3.6772, + "step": 1189 + }, + { + "epoch": 0.3233695652173913, + "grad_norm": 1.2040995359420776, + "learning_rate": 3e-05, + "loss": 3.3302, + "step": 1190 + }, + { + "epoch": 0.32364130434782606, + "grad_norm": 1.4731234312057495, + "learning_rate": 3e-05, + "loss": 3.3842, + "step": 1191 + }, + { + "epoch": 0.3239130434782609, + "grad_norm": 1.3904873132705688, + "learning_rate": 3e-05, + "loss": 3.8537, + "step": 1192 + }, + { + "epoch": 0.32418478260869565, + "grad_norm": 1.187960147857666, + "learning_rate": 3e-05, + "loss": 3.0656, + "step": 1193 + }, + { + "epoch": 0.3244565217391304, + "grad_norm": 1.1819937229156494, + "learning_rate": 3e-05, + "loss": 2.7234, + "step": 1194 + }, + { + "epoch": 0.32472826086956524, + "grad_norm": 1.6428884267807007, + "learning_rate": 3e-05, + "loss": 3.4446, + "step": 1195 + }, + { + "epoch": 0.325, + "grad_norm": 1.6877864599227905, + "learning_rate": 3e-05, + "loss": 3.2674, + "step": 1196 + }, + { + "epoch": 0.3252717391304348, + "grad_norm": 1.596166729927063, + "learning_rate": 3e-05, + "loss": 3.371, + "step": 1197 + }, + { + "epoch": 0.32554347826086955, + "grad_norm": 1.3886607885360718, + "learning_rate": 3e-05, + "loss": 3.3149, + "step": 1198 + }, + { + "epoch": 0.32581521739130437, + "grad_norm": 1.4295148849487305, + "learning_rate": 3e-05, + "loss": 3.0519, + "step": 1199 + }, + { + "epoch": 0.32608695652173914, + "grad_norm": 1.5317387580871582, + "learning_rate": 3e-05, + "loss": 3.3816, + "step": 1200 + }, + { + "epoch": 0.3263586956521739, + "grad_norm": 1.1524282693862915, + "learning_rate": 3e-05, + "loss": 3.2564, + "step": 1201 + }, + { + "epoch": 0.32663043478260867, + "grad_norm": 1.3997209072113037, + "learning_rate": 3e-05, + "loss": 3.3173, + "step": 1202 + }, + { + "epoch": 0.3269021739130435, + "grad_norm": 1.357279896736145, + "learning_rate": 3e-05, + "loss": 3.7316, + "step": 1203 + }, + { + "epoch": 0.32717391304347826, + "grad_norm": 1.1962755918502808, + "learning_rate": 3e-05, + "loss": 3.5203, + "step": 1204 + }, + { + "epoch": 0.327445652173913, + "grad_norm": 1.4180536270141602, + "learning_rate": 3e-05, + "loss": 3.7679, + "step": 1205 + }, + { + "epoch": 0.32771739130434785, + "grad_norm": 1.4997798204421997, + "learning_rate": 3e-05, + "loss": 3.9842, + "step": 1206 + }, + { + "epoch": 0.3279891304347826, + "grad_norm": 1.5224746465682983, + "learning_rate": 3e-05, + "loss": 3.723, + "step": 1207 + }, + { + "epoch": 0.3282608695652174, + "grad_norm": 1.4814677238464355, + "learning_rate": 3e-05, + "loss": 3.3444, + "step": 1208 + }, + { + "epoch": 0.32853260869565215, + "grad_norm": 1.8869359493255615, + "learning_rate": 3e-05, + "loss": 3.8663, + "step": 1209 + }, + { + "epoch": 0.328804347826087, + "grad_norm": 1.2305731773376465, + "learning_rate": 3e-05, + "loss": 3.1855, + "step": 1210 + }, + { + "epoch": 0.32907608695652174, + "grad_norm": 1.3822002410888672, + "learning_rate": 3e-05, + "loss": 3.9625, + "step": 1211 + }, + { + "epoch": 0.3293478260869565, + "grad_norm": 1.2895426750183105, + "learning_rate": 3e-05, + "loss": 2.9574, + "step": 1212 + }, + { + "epoch": 0.32961956521739133, + "grad_norm": 1.2724965810775757, + "learning_rate": 3e-05, + "loss": 3.2984, + "step": 1213 + }, + { + "epoch": 0.3298913043478261, + "grad_norm": 1.569366455078125, + "learning_rate": 3e-05, + "loss": 3.4928, + "step": 1214 + }, + { + "epoch": 0.33016304347826086, + "grad_norm": 1.538090467453003, + "learning_rate": 3e-05, + "loss": 3.8567, + "step": 1215 + }, + { + "epoch": 0.33043478260869563, + "grad_norm": 1.3484816551208496, + "learning_rate": 3e-05, + "loss": 3.1096, + "step": 1216 + }, + { + "epoch": 0.33070652173913045, + "grad_norm": 1.1450049877166748, + "learning_rate": 3e-05, + "loss": 2.773, + "step": 1217 + }, + { + "epoch": 0.3309782608695652, + "grad_norm": 1.5292000770568848, + "learning_rate": 3e-05, + "loss": 3.3972, + "step": 1218 + }, + { + "epoch": 0.33125, + "grad_norm": 1.3615137338638306, + "learning_rate": 3e-05, + "loss": 3.2462, + "step": 1219 + }, + { + "epoch": 0.33152173913043476, + "grad_norm": 1.7036170959472656, + "learning_rate": 3e-05, + "loss": 3.7338, + "step": 1220 + }, + { + "epoch": 0.3317934782608696, + "grad_norm": 1.7132785320281982, + "learning_rate": 3e-05, + "loss": 3.3378, + "step": 1221 + }, + { + "epoch": 0.33206521739130435, + "grad_norm": 1.9641152620315552, + "learning_rate": 3e-05, + "loss": 3.8121, + "step": 1222 + }, + { + "epoch": 0.3323369565217391, + "grad_norm": 1.3222547769546509, + "learning_rate": 3e-05, + "loss": 3.2911, + "step": 1223 + }, + { + "epoch": 0.33260869565217394, + "grad_norm": 1.3695377111434937, + "learning_rate": 3e-05, + "loss": 3.3004, + "step": 1224 + }, + { + "epoch": 0.3328804347826087, + "grad_norm": 1.8578277826309204, + "learning_rate": 3e-05, + "loss": 4.1831, + "step": 1225 + }, + { + "epoch": 0.33315217391304347, + "grad_norm": 1.7893558740615845, + "learning_rate": 3e-05, + "loss": 3.359, + "step": 1226 + }, + { + "epoch": 0.33342391304347824, + "grad_norm": 1.474753499031067, + "learning_rate": 3e-05, + "loss": 2.9094, + "step": 1227 + }, + { + "epoch": 0.33369565217391306, + "grad_norm": 1.973402738571167, + "learning_rate": 3e-05, + "loss": 4.0505, + "step": 1228 + }, + { + "epoch": 0.3339673913043478, + "grad_norm": 1.7613177299499512, + "learning_rate": 3e-05, + "loss": 3.3559, + "step": 1229 + }, + { + "epoch": 0.3342391304347826, + "grad_norm": 2.031470775604248, + "learning_rate": 3e-05, + "loss": 3.3066, + "step": 1230 + }, + { + "epoch": 0.3345108695652174, + "grad_norm": 2.1547582149505615, + "learning_rate": 3e-05, + "loss": 3.6043, + "step": 1231 + }, + { + "epoch": 0.3347826086956522, + "grad_norm": 2.2331910133361816, + "learning_rate": 3e-05, + "loss": 2.9871, + "step": 1232 + }, + { + "epoch": 0.33505434782608695, + "grad_norm": 2.156770944595337, + "learning_rate": 3e-05, + "loss": 3.6777, + "step": 1233 + }, + { + "epoch": 0.3353260869565217, + "grad_norm": 1.717960000038147, + "learning_rate": 3e-05, + "loss": 3.6268, + "step": 1234 + }, + { + "epoch": 0.33559782608695654, + "grad_norm": 1.649511694908142, + "learning_rate": 3e-05, + "loss": 3.3226, + "step": 1235 + }, + { + "epoch": 0.3358695652173913, + "grad_norm": 1.8623225688934326, + "learning_rate": 3e-05, + "loss": 3.9329, + "step": 1236 + }, + { + "epoch": 0.3361413043478261, + "grad_norm": 2.167483329772949, + "learning_rate": 3e-05, + "loss": 3.445, + "step": 1237 + }, + { + "epoch": 0.33641304347826084, + "grad_norm": 1.7240395545959473, + "learning_rate": 3e-05, + "loss": 3.493, + "step": 1238 + }, + { + "epoch": 0.33668478260869567, + "grad_norm": 1.3711011409759521, + "learning_rate": 3e-05, + "loss": 2.7254, + "step": 1239 + }, + { + "epoch": 0.33695652173913043, + "grad_norm": 2.091745615005493, + "learning_rate": 3e-05, + "loss": 4.0388, + "step": 1240 + }, + { + "epoch": 0.3372282608695652, + "grad_norm": 1.5960465669631958, + "learning_rate": 3e-05, + "loss": 3.5036, + "step": 1241 + }, + { + "epoch": 0.3375, + "grad_norm": 1.1491891145706177, + "learning_rate": 3e-05, + "loss": 3.2282, + "step": 1242 + }, + { + "epoch": 0.3377717391304348, + "grad_norm": 1.1655139923095703, + "learning_rate": 3e-05, + "loss": 3.3578, + "step": 1243 + }, + { + "epoch": 0.33804347826086956, + "grad_norm": 1.4522464275360107, + "learning_rate": 3e-05, + "loss": 3.4827, + "step": 1244 + }, + { + "epoch": 0.3383152173913043, + "grad_norm": 1.494310736656189, + "learning_rate": 3e-05, + "loss": 3.7874, + "step": 1245 + }, + { + "epoch": 0.33858695652173915, + "grad_norm": 1.98121976852417, + "learning_rate": 3e-05, + "loss": 3.2323, + "step": 1246 + }, + { + "epoch": 0.3388586956521739, + "grad_norm": 1.772510290145874, + "learning_rate": 3e-05, + "loss": 2.9783, + "step": 1247 + }, + { + "epoch": 0.3391304347826087, + "grad_norm": 1.3050928115844727, + "learning_rate": 3e-05, + "loss": 3.3902, + "step": 1248 + }, + { + "epoch": 0.3394021739130435, + "grad_norm": 1.590279221534729, + "learning_rate": 3e-05, + "loss": 3.0477, + "step": 1249 + }, + { + "epoch": 0.33967391304347827, + "grad_norm": 2.1541528701782227, + "learning_rate": 3e-05, + "loss": 3.5042, + "step": 1250 + }, + { + "epoch": 0.33994565217391304, + "grad_norm": 2.138265609741211, + "learning_rate": 3e-05, + "loss": 3.704, + "step": 1251 + }, + { + "epoch": 0.3402173913043478, + "grad_norm": 1.5365831851959229, + "learning_rate": 3e-05, + "loss": 3.3562, + "step": 1252 + }, + { + "epoch": 0.3404891304347826, + "grad_norm": 1.3827742338180542, + "learning_rate": 3e-05, + "loss": 3.2202, + "step": 1253 + }, + { + "epoch": 0.3407608695652174, + "grad_norm": 1.4450342655181885, + "learning_rate": 3e-05, + "loss": 3.4887, + "step": 1254 + }, + { + "epoch": 0.34103260869565216, + "grad_norm": 1.2322391271591187, + "learning_rate": 3e-05, + "loss": 2.7866, + "step": 1255 + }, + { + "epoch": 0.34130434782608693, + "grad_norm": 1.3767555952072144, + "learning_rate": 3e-05, + "loss": 2.8691, + "step": 1256 + }, + { + "epoch": 0.34157608695652175, + "grad_norm": 1.9902418851852417, + "learning_rate": 3e-05, + "loss": 3.5811, + "step": 1257 + }, + { + "epoch": 0.3418478260869565, + "grad_norm": 1.7038296461105347, + "learning_rate": 3e-05, + "loss": 3.363, + "step": 1258 + }, + { + "epoch": 0.3421195652173913, + "grad_norm": 1.82673978805542, + "learning_rate": 3e-05, + "loss": 3.0324, + "step": 1259 + }, + { + "epoch": 0.3423913043478261, + "grad_norm": 1.4924451112747192, + "learning_rate": 3e-05, + "loss": 3.5862, + "step": 1260 + }, + { + "epoch": 0.3426630434782609, + "grad_norm": 1.1290408372879028, + "learning_rate": 3e-05, + "loss": 3.0983, + "step": 1261 + }, + { + "epoch": 0.34293478260869564, + "grad_norm": 1.3652050495147705, + "learning_rate": 3e-05, + "loss": 3.2806, + "step": 1262 + }, + { + "epoch": 0.3432065217391304, + "grad_norm": 1.3618534803390503, + "learning_rate": 3e-05, + "loss": 3.0708, + "step": 1263 + }, + { + "epoch": 0.34347826086956523, + "grad_norm": 1.3616082668304443, + "learning_rate": 3e-05, + "loss": 2.8319, + "step": 1264 + }, + { + "epoch": 0.34375, + "grad_norm": 1.3398898839950562, + "learning_rate": 3e-05, + "loss": 3.1065, + "step": 1265 + }, + { + "epoch": 0.34402173913043477, + "grad_norm": 1.6687146425247192, + "learning_rate": 3e-05, + "loss": 3.3584, + "step": 1266 + }, + { + "epoch": 0.3442934782608696, + "grad_norm": 1.2948143482208252, + "learning_rate": 3e-05, + "loss": 3.2489, + "step": 1267 + }, + { + "epoch": 0.34456521739130436, + "grad_norm": 1.7158524990081787, + "learning_rate": 3e-05, + "loss": 3.3558, + "step": 1268 + }, + { + "epoch": 0.3448369565217391, + "grad_norm": 1.6835163831710815, + "learning_rate": 3e-05, + "loss": 2.9967, + "step": 1269 + }, + { + "epoch": 0.3451086956521739, + "grad_norm": 1.971177101135254, + "learning_rate": 3e-05, + "loss": 3.829, + "step": 1270 + }, + { + "epoch": 0.3453804347826087, + "grad_norm": 1.660911202430725, + "learning_rate": 3e-05, + "loss": 2.9975, + "step": 1271 + }, + { + "epoch": 0.3456521739130435, + "grad_norm": 1.8289529085159302, + "learning_rate": 3e-05, + "loss": 3.4466, + "step": 1272 + }, + { + "epoch": 0.34592391304347825, + "grad_norm": 1.9580225944519043, + "learning_rate": 3e-05, + "loss": 3.4431, + "step": 1273 + }, + { + "epoch": 0.34619565217391307, + "grad_norm": 1.9081965684890747, + "learning_rate": 3e-05, + "loss": 3.1295, + "step": 1274 + }, + { + "epoch": 0.34646739130434784, + "grad_norm": 2.8098397254943848, + "learning_rate": 3e-05, + "loss": 3.9265, + "step": 1275 + }, + { + "epoch": 0.3467391304347826, + "grad_norm": 1.8996559381484985, + "learning_rate": 3e-05, + "loss": 2.9118, + "step": 1276 + }, + { + "epoch": 0.3470108695652174, + "grad_norm": 1.8447978496551514, + "learning_rate": 3e-05, + "loss": 3.8908, + "step": 1277 + }, + { + "epoch": 0.3472826086956522, + "grad_norm": 1.450376033782959, + "learning_rate": 3e-05, + "loss": 3.1262, + "step": 1278 + }, + { + "epoch": 0.34755434782608696, + "grad_norm": 2.3735806941986084, + "learning_rate": 3e-05, + "loss": 3.4959, + "step": 1279 + }, + { + "epoch": 0.34782608695652173, + "grad_norm": 2.2202954292297363, + "learning_rate": 3e-05, + "loss": 3.2127, + "step": 1280 + }, + { + "epoch": 0.3480978260869565, + "grad_norm": 1.6313713788986206, + "learning_rate": 3e-05, + "loss": 2.7593, + "step": 1281 + }, + { + "epoch": 0.3483695652173913, + "grad_norm": 1.8261613845825195, + "learning_rate": 3e-05, + "loss": 3.5207, + "step": 1282 + }, + { + "epoch": 0.3486413043478261, + "grad_norm": 1.6101502180099487, + "learning_rate": 3e-05, + "loss": 3.1187, + "step": 1283 + }, + { + "epoch": 0.34891304347826085, + "grad_norm": 1.4293962717056274, + "learning_rate": 3e-05, + "loss": 3.3783, + "step": 1284 + }, + { + "epoch": 0.3491847826086957, + "grad_norm": 1.335721731185913, + "learning_rate": 3e-05, + "loss": 3.8365, + "step": 1285 + }, + { + "epoch": 0.34945652173913044, + "grad_norm": 1.5614105463027954, + "learning_rate": 3e-05, + "loss": 3.7661, + "step": 1286 + }, + { + "epoch": 0.3497282608695652, + "grad_norm": 1.38046133518219, + "learning_rate": 3e-05, + "loss": 3.1717, + "step": 1287 + }, + { + "epoch": 0.35, + "grad_norm": 1.3673638105392456, + "learning_rate": 3e-05, + "loss": 3.7002, + "step": 1288 + }, + { + "epoch": 0.3502717391304348, + "grad_norm": 1.2321789264678955, + "learning_rate": 3e-05, + "loss": 2.8849, + "step": 1289 + }, + { + "epoch": 0.35054347826086957, + "grad_norm": 1.3166979551315308, + "learning_rate": 3e-05, + "loss": 3.1094, + "step": 1290 + }, + { + "epoch": 0.35081521739130433, + "grad_norm": 1.5346661806106567, + "learning_rate": 3e-05, + "loss": 3.2696, + "step": 1291 + }, + { + "epoch": 0.35108695652173916, + "grad_norm": 1.550195336341858, + "learning_rate": 3e-05, + "loss": 3.3044, + "step": 1292 + }, + { + "epoch": 0.3513586956521739, + "grad_norm": 1.7930152416229248, + "learning_rate": 3e-05, + "loss": 3.7782, + "step": 1293 + }, + { + "epoch": 0.3516304347826087, + "grad_norm": 1.9929698705673218, + "learning_rate": 3e-05, + "loss": 3.12, + "step": 1294 + }, + { + "epoch": 0.35190217391304346, + "grad_norm": 1.244729995727539, + "learning_rate": 3e-05, + "loss": 2.9368, + "step": 1295 + }, + { + "epoch": 0.3521739130434783, + "grad_norm": 1.3026173114776611, + "learning_rate": 3e-05, + "loss": 3.1947, + "step": 1296 + }, + { + "epoch": 0.35244565217391305, + "grad_norm": 1.297813892364502, + "learning_rate": 3e-05, + "loss": 2.5978, + "step": 1297 + }, + { + "epoch": 0.3527173913043478, + "grad_norm": 1.4547146558761597, + "learning_rate": 3e-05, + "loss": 3.4649, + "step": 1298 + }, + { + "epoch": 0.3529891304347826, + "grad_norm": 1.2365776300430298, + "learning_rate": 3e-05, + "loss": 3.1037, + "step": 1299 + }, + { + "epoch": 0.3532608695652174, + "grad_norm": 1.2477210760116577, + "learning_rate": 3e-05, + "loss": 3.1045, + "step": 1300 + }, + { + "epoch": 0.3535326086956522, + "grad_norm": 1.4197901487350464, + "learning_rate": 3e-05, + "loss": 3.2411, + "step": 1301 + }, + { + "epoch": 0.35380434782608694, + "grad_norm": 1.1791913509368896, + "learning_rate": 3e-05, + "loss": 3.4775, + "step": 1302 + }, + { + "epoch": 0.35407608695652176, + "grad_norm": 1.1935291290283203, + "learning_rate": 3e-05, + "loss": 2.9038, + "step": 1303 + }, + { + "epoch": 0.35434782608695653, + "grad_norm": 1.515292763710022, + "learning_rate": 3e-05, + "loss": 3.6569, + "step": 1304 + }, + { + "epoch": 0.3546195652173913, + "grad_norm": 1.6459676027297974, + "learning_rate": 3e-05, + "loss": 3.638, + "step": 1305 + }, + { + "epoch": 0.35489130434782606, + "grad_norm": 1.2799147367477417, + "learning_rate": 3e-05, + "loss": 3.3, + "step": 1306 + }, + { + "epoch": 0.3551630434782609, + "grad_norm": 1.1415566205978394, + "learning_rate": 3e-05, + "loss": 2.7568, + "step": 1307 + }, + { + "epoch": 0.35543478260869565, + "grad_norm": 1.4160751104354858, + "learning_rate": 3e-05, + "loss": 3.5389, + "step": 1308 + }, + { + "epoch": 0.3557065217391304, + "grad_norm": 1.5589027404785156, + "learning_rate": 3e-05, + "loss": 3.5721, + "step": 1309 + }, + { + "epoch": 0.35597826086956524, + "grad_norm": 1.4486863613128662, + "learning_rate": 3e-05, + "loss": 3.5145, + "step": 1310 + }, + { + "epoch": 0.35625, + "grad_norm": 1.1204112768173218, + "learning_rate": 3e-05, + "loss": 3.2109, + "step": 1311 + }, + { + "epoch": 0.3565217391304348, + "grad_norm": 1.4160523414611816, + "learning_rate": 3e-05, + "loss": 3.9734, + "step": 1312 + }, + { + "epoch": 0.35679347826086955, + "grad_norm": 1.359877347946167, + "learning_rate": 3e-05, + "loss": 3.1197, + "step": 1313 + }, + { + "epoch": 0.35706521739130437, + "grad_norm": 1.3469575643539429, + "learning_rate": 3e-05, + "loss": 2.7031, + "step": 1314 + }, + { + "epoch": 0.35733695652173914, + "grad_norm": 1.877429485321045, + "learning_rate": 3e-05, + "loss": 3.6433, + "step": 1315 + }, + { + "epoch": 0.3576086956521739, + "grad_norm": 1.6881115436553955, + "learning_rate": 3e-05, + "loss": 3.8429, + "step": 1316 + }, + { + "epoch": 0.35788043478260867, + "grad_norm": 0.9885056018829346, + "learning_rate": 3e-05, + "loss": 2.9943, + "step": 1317 + }, + { + "epoch": 0.3581521739130435, + "grad_norm": 1.1146260499954224, + "learning_rate": 3e-05, + "loss": 3.0583, + "step": 1318 + }, + { + "epoch": 0.35842391304347826, + "grad_norm": 1.0930460691452026, + "learning_rate": 3e-05, + "loss": 3.1491, + "step": 1319 + }, + { + "epoch": 0.358695652173913, + "grad_norm": 1.7721790075302124, + "learning_rate": 3e-05, + "loss": 3.682, + "step": 1320 + }, + { + "epoch": 0.35896739130434785, + "grad_norm": 1.8233963251113892, + "learning_rate": 3e-05, + "loss": 4.0548, + "step": 1321 + }, + { + "epoch": 0.3592391304347826, + "grad_norm": 2.3196828365325928, + "learning_rate": 3e-05, + "loss": 3.8065, + "step": 1322 + }, + { + "epoch": 0.3595108695652174, + "grad_norm": 2.0975351333618164, + "learning_rate": 3e-05, + "loss": 3.3098, + "step": 1323 + }, + { + "epoch": 0.35978260869565215, + "grad_norm": 1.4240121841430664, + "learning_rate": 3e-05, + "loss": 3.327, + "step": 1324 + }, + { + "epoch": 0.360054347826087, + "grad_norm": 1.4401570558547974, + "learning_rate": 3e-05, + "loss": 2.9099, + "step": 1325 + }, + { + "epoch": 0.36032608695652174, + "grad_norm": 1.7879765033721924, + "learning_rate": 3e-05, + "loss": 2.9776, + "step": 1326 + }, + { + "epoch": 0.3605978260869565, + "grad_norm": 1.8819643259048462, + "learning_rate": 3e-05, + "loss": 3.3526, + "step": 1327 + }, + { + "epoch": 0.36086956521739133, + "grad_norm": 1.8750115633010864, + "learning_rate": 3e-05, + "loss": 3.7411, + "step": 1328 + }, + { + "epoch": 0.3611413043478261, + "grad_norm": 1.802832007408142, + "learning_rate": 3e-05, + "loss": 2.8997, + "step": 1329 + }, + { + "epoch": 0.36141304347826086, + "grad_norm": 1.9432928562164307, + "learning_rate": 3e-05, + "loss": 2.9711, + "step": 1330 + }, + { + "epoch": 0.36168478260869563, + "grad_norm": 2.222259998321533, + "learning_rate": 3e-05, + "loss": 3.5956, + "step": 1331 + }, + { + "epoch": 0.36195652173913045, + "grad_norm": 1.0173529386520386, + "learning_rate": 3e-05, + "loss": 3.1368, + "step": 1332 + }, + { + "epoch": 0.3622282608695652, + "grad_norm": 1.3840457201004028, + "learning_rate": 3e-05, + "loss": 3.3336, + "step": 1333 + }, + { + "epoch": 0.3625, + "grad_norm": 2.199570417404175, + "learning_rate": 3e-05, + "loss": 3.1846, + "step": 1334 + }, + { + "epoch": 0.36277173913043476, + "grad_norm": 2.3063576221466064, + "learning_rate": 3e-05, + "loss": 3.1996, + "step": 1335 + }, + { + "epoch": 0.3630434782608696, + "grad_norm": 2.0329015254974365, + "learning_rate": 3e-05, + "loss": 3.6631, + "step": 1336 + }, + { + "epoch": 0.36331521739130435, + "grad_norm": 0.9080404043197632, + "learning_rate": 3e-05, + "loss": 2.8746, + "step": 1337 + }, + { + "epoch": 0.3635869565217391, + "grad_norm": 2.173773765563965, + "learning_rate": 3e-05, + "loss": 3.8486, + "step": 1338 + }, + { + "epoch": 0.36385869565217394, + "grad_norm": 2.493887186050415, + "learning_rate": 3e-05, + "loss": 3.5707, + "step": 1339 + }, + { + "epoch": 0.3641304347826087, + "grad_norm": 1.8862491846084595, + "learning_rate": 3e-05, + "loss": 3.8585, + "step": 1340 + }, + { + "epoch": 0.36440217391304347, + "grad_norm": 1.23379647731781, + "learning_rate": 3e-05, + "loss": 3.5806, + "step": 1341 + }, + { + "epoch": 0.36467391304347824, + "grad_norm": 1.3684219121932983, + "learning_rate": 3e-05, + "loss": 3.1318, + "step": 1342 + }, + { + "epoch": 0.36494565217391306, + "grad_norm": 1.2184761762619019, + "learning_rate": 3e-05, + "loss": 3.4761, + "step": 1343 + }, + { + "epoch": 0.3652173913043478, + "grad_norm": 1.230040431022644, + "learning_rate": 3e-05, + "loss": 4.1053, + "step": 1344 + }, + { + "epoch": 0.3654891304347826, + "grad_norm": 1.4862629175186157, + "learning_rate": 3e-05, + "loss": 3.4211, + "step": 1345 + }, + { + "epoch": 0.3657608695652174, + "grad_norm": 1.2832411527633667, + "learning_rate": 3e-05, + "loss": 2.7821, + "step": 1346 + }, + { + "epoch": 0.3660326086956522, + "grad_norm": 1.1333979368209839, + "learning_rate": 3e-05, + "loss": 3.4129, + "step": 1347 + }, + { + "epoch": 0.36630434782608695, + "grad_norm": 1.9462121725082397, + "learning_rate": 3e-05, + "loss": 3.8102, + "step": 1348 + }, + { + "epoch": 0.3665760869565217, + "grad_norm": 1.314512848854065, + "learning_rate": 3e-05, + "loss": 3.5802, + "step": 1349 + }, + { + "epoch": 0.36684782608695654, + "grad_norm": 0.9899564981460571, + "learning_rate": 3e-05, + "loss": 3.2397, + "step": 1350 + }, + { + "epoch": 0.3671195652173913, + "grad_norm": 0.9129654765129089, + "learning_rate": 3e-05, + "loss": 3.0704, + "step": 1351 + }, + { + "epoch": 0.3673913043478261, + "grad_norm": 1.2658374309539795, + "learning_rate": 3e-05, + "loss": 2.9416, + "step": 1352 + }, + { + "epoch": 0.36766304347826084, + "grad_norm": 1.156530499458313, + "learning_rate": 3e-05, + "loss": 3.2254, + "step": 1353 + }, + { + "epoch": 0.36793478260869567, + "grad_norm": 1.4215689897537231, + "learning_rate": 3e-05, + "loss": 3.5928, + "step": 1354 + }, + { + "epoch": 0.36820652173913043, + "grad_norm": 1.4695452451705933, + "learning_rate": 3e-05, + "loss": 3.5298, + "step": 1355 + }, + { + "epoch": 0.3684782608695652, + "grad_norm": 1.185398817062378, + "learning_rate": 3e-05, + "loss": 2.9786, + "step": 1356 + }, + { + "epoch": 0.36875, + "grad_norm": 1.6181942224502563, + "learning_rate": 3e-05, + "loss": 3.8328, + "step": 1357 + }, + { + "epoch": 0.3690217391304348, + "grad_norm": 1.7642178535461426, + "learning_rate": 3e-05, + "loss": 3.6084, + "step": 1358 + }, + { + "epoch": 0.36929347826086956, + "grad_norm": 1.3343443870544434, + "learning_rate": 3e-05, + "loss": 4.0491, + "step": 1359 + }, + { + "epoch": 0.3695652173913043, + "grad_norm": 2.167877435684204, + "learning_rate": 3e-05, + "loss": 3.6741, + "step": 1360 + }, + { + "epoch": 0.36983695652173915, + "grad_norm": 2.0322484970092773, + "learning_rate": 3e-05, + "loss": 4.1298, + "step": 1361 + }, + { + "epoch": 0.3701086956521739, + "grad_norm": 1.1205507516860962, + "learning_rate": 3e-05, + "loss": 3.2478, + "step": 1362 + }, + { + "epoch": 0.3703804347826087, + "grad_norm": 1.4173895120620728, + "learning_rate": 3e-05, + "loss": 3.2024, + "step": 1363 + }, + { + "epoch": 0.3706521739130435, + "grad_norm": 1.4186296463012695, + "learning_rate": 3e-05, + "loss": 2.9463, + "step": 1364 + }, + { + "epoch": 0.37092391304347827, + "grad_norm": 1.4897836446762085, + "learning_rate": 3e-05, + "loss": 3.4138, + "step": 1365 + }, + { + "epoch": 0.37119565217391304, + "grad_norm": 0.9213505983352661, + "learning_rate": 3e-05, + "loss": 2.7726, + "step": 1366 + }, + { + "epoch": 0.3714673913043478, + "grad_norm": 1.4468334913253784, + "learning_rate": 3e-05, + "loss": 3.3552, + "step": 1367 + }, + { + "epoch": 0.3717391304347826, + "grad_norm": 1.6020596027374268, + "learning_rate": 3e-05, + "loss": 3.3098, + "step": 1368 + }, + { + "epoch": 0.3720108695652174, + "grad_norm": 1.2568944692611694, + "learning_rate": 3e-05, + "loss": 2.9896, + "step": 1369 + }, + { + "epoch": 0.37228260869565216, + "grad_norm": 1.2092691659927368, + "learning_rate": 3e-05, + "loss": 3.2886, + "step": 1370 + }, + { + "epoch": 0.37255434782608693, + "grad_norm": 1.0227844715118408, + "learning_rate": 3e-05, + "loss": 2.9555, + "step": 1371 + }, + { + "epoch": 0.37282608695652175, + "grad_norm": 1.400106430053711, + "learning_rate": 3e-05, + "loss": 3.2697, + "step": 1372 + }, + { + "epoch": 0.3730978260869565, + "grad_norm": 1.3000365495681763, + "learning_rate": 3e-05, + "loss": 3.1748, + "step": 1373 + }, + { + "epoch": 0.3733695652173913, + "grad_norm": 1.197072148323059, + "learning_rate": 3e-05, + "loss": 3.3346, + "step": 1374 + }, + { + "epoch": 0.3736413043478261, + "grad_norm": 1.1778289079666138, + "learning_rate": 3e-05, + "loss": 3.089, + "step": 1375 + }, + { + "epoch": 0.3739130434782609, + "grad_norm": 1.23442542552948, + "learning_rate": 3e-05, + "loss": 3.0972, + "step": 1376 + }, + { + "epoch": 0.37418478260869564, + "grad_norm": 1.135533094406128, + "learning_rate": 3e-05, + "loss": 2.9638, + "step": 1377 + }, + { + "epoch": 0.3744565217391304, + "grad_norm": 1.4869073629379272, + "learning_rate": 3e-05, + "loss": 3.7439, + "step": 1378 + }, + { + "epoch": 0.37472826086956523, + "grad_norm": 1.5528923273086548, + "learning_rate": 3e-05, + "loss": 2.8187, + "step": 1379 + }, + { + "epoch": 0.375, + "grad_norm": 1.0053822994232178, + "learning_rate": 3e-05, + "loss": 3.2647, + "step": 1380 + }, + { + "epoch": 0.37527173913043477, + "grad_norm": 1.2217366695404053, + "learning_rate": 3e-05, + "loss": 3.5844, + "step": 1381 + }, + { + "epoch": 0.3755434782608696, + "grad_norm": 1.245322823524475, + "learning_rate": 3e-05, + "loss": 2.9917, + "step": 1382 + }, + { + "epoch": 0.37581521739130436, + "grad_norm": 1.3003109693527222, + "learning_rate": 3e-05, + "loss": 3.7163, + "step": 1383 + }, + { + "epoch": 0.3760869565217391, + "grad_norm": 2.398679733276367, + "learning_rate": 3e-05, + "loss": 4.2985, + "step": 1384 + }, + { + "epoch": 0.3763586956521739, + "grad_norm": 1.2174817323684692, + "learning_rate": 3e-05, + "loss": 3.3311, + "step": 1385 + }, + { + "epoch": 0.3766304347826087, + "grad_norm": 1.2073396444320679, + "learning_rate": 3e-05, + "loss": 3.4313, + "step": 1386 + }, + { + "epoch": 0.3769021739130435, + "grad_norm": 1.2685997486114502, + "learning_rate": 3e-05, + "loss": 3.1358, + "step": 1387 + }, + { + "epoch": 0.37717391304347825, + "grad_norm": 1.3566850423812866, + "learning_rate": 3e-05, + "loss": 3.0956, + "step": 1388 + }, + { + "epoch": 0.37744565217391307, + "grad_norm": 1.6564099788665771, + "learning_rate": 3e-05, + "loss": 3.7791, + "step": 1389 + }, + { + "epoch": 0.37771739130434784, + "grad_norm": 1.3698561191558838, + "learning_rate": 3e-05, + "loss": 3.3446, + "step": 1390 + }, + { + "epoch": 0.3779891304347826, + "grad_norm": 1.2702010869979858, + "learning_rate": 3e-05, + "loss": 4.0951, + "step": 1391 + }, + { + "epoch": 0.3782608695652174, + "grad_norm": 1.4307619333267212, + "learning_rate": 3e-05, + "loss": 3.3821, + "step": 1392 + }, + { + "epoch": 0.3785326086956522, + "grad_norm": 1.207396149635315, + "learning_rate": 3e-05, + "loss": 3.6105, + "step": 1393 + }, + { + "epoch": 0.37880434782608696, + "grad_norm": 1.1929562091827393, + "learning_rate": 3e-05, + "loss": 2.8538, + "step": 1394 + }, + { + "epoch": 0.37907608695652173, + "grad_norm": 1.2328208684921265, + "learning_rate": 3e-05, + "loss": 3.0622, + "step": 1395 + }, + { + "epoch": 0.3793478260869565, + "grad_norm": 1.4275541305541992, + "learning_rate": 3e-05, + "loss": 3.4482, + "step": 1396 + }, + { + "epoch": 0.3796195652173913, + "grad_norm": 0.9579852223396301, + "learning_rate": 3e-05, + "loss": 2.8646, + "step": 1397 + }, + { + "epoch": 0.3798913043478261, + "grad_norm": 1.6573379039764404, + "learning_rate": 3e-05, + "loss": 3.4417, + "step": 1398 + }, + { + "epoch": 0.38016304347826085, + "grad_norm": 1.5388156175613403, + "learning_rate": 3e-05, + "loss": 3.2041, + "step": 1399 + }, + { + "epoch": 0.3804347826086957, + "grad_norm": 1.4631450176239014, + "learning_rate": 3e-05, + "loss": 3.8626, + "step": 1400 + }, + { + "epoch": 0.38070652173913044, + "grad_norm": 1.3570584058761597, + "learning_rate": 3e-05, + "loss": 3.441, + "step": 1401 + }, + { + "epoch": 0.3809782608695652, + "grad_norm": 1.8016875982284546, + "learning_rate": 3e-05, + "loss": 3.1616, + "step": 1402 + }, + { + "epoch": 0.38125, + "grad_norm": 1.4279495477676392, + "learning_rate": 3e-05, + "loss": 3.6129, + "step": 1403 + }, + { + "epoch": 0.3815217391304348, + "grad_norm": 1.3305083513259888, + "learning_rate": 3e-05, + "loss": 3.3956, + "step": 1404 + }, + { + "epoch": 0.38179347826086957, + "grad_norm": 1.7289437055587769, + "learning_rate": 3e-05, + "loss": 3.6499, + "step": 1405 + }, + { + "epoch": 0.38206521739130433, + "grad_norm": 1.75969398021698, + "learning_rate": 3e-05, + "loss": 3.104, + "step": 1406 + }, + { + "epoch": 0.38233695652173916, + "grad_norm": 1.545464277267456, + "learning_rate": 3e-05, + "loss": 4.1157, + "step": 1407 + }, + { + "epoch": 0.3826086956521739, + "grad_norm": 1.5299899578094482, + "learning_rate": 3e-05, + "loss": 2.9975, + "step": 1408 + }, + { + "epoch": 0.3828804347826087, + "grad_norm": 1.7116879224777222, + "learning_rate": 3e-05, + "loss": 3.0605, + "step": 1409 + }, + { + "epoch": 0.38315217391304346, + "grad_norm": 1.4728469848632812, + "learning_rate": 3e-05, + "loss": 3.2079, + "step": 1410 + }, + { + "epoch": 0.3834239130434783, + "grad_norm": 2.216775894165039, + "learning_rate": 3e-05, + "loss": 3.8937, + "step": 1411 + }, + { + "epoch": 0.38369565217391305, + "grad_norm": 1.6279774904251099, + "learning_rate": 3e-05, + "loss": 3.7957, + "step": 1412 + }, + { + "epoch": 0.3839673913043478, + "grad_norm": 0.9390795230865479, + "learning_rate": 3e-05, + "loss": 2.8752, + "step": 1413 + }, + { + "epoch": 0.3842391304347826, + "grad_norm": 1.3048573732376099, + "learning_rate": 3e-05, + "loss": 3.0132, + "step": 1414 + }, + { + "epoch": 0.3845108695652174, + "grad_norm": 1.132607340812683, + "learning_rate": 3e-05, + "loss": 2.965, + "step": 1415 + }, + { + "epoch": 0.3847826086956522, + "grad_norm": 1.6570241451263428, + "learning_rate": 3e-05, + "loss": 3.1474, + "step": 1416 + }, + { + "epoch": 0.38505434782608694, + "grad_norm": 1.5501550436019897, + "learning_rate": 3e-05, + "loss": 3.4049, + "step": 1417 + }, + { + "epoch": 0.38532608695652176, + "grad_norm": 1.1150649785995483, + "learning_rate": 3e-05, + "loss": 3.4114, + "step": 1418 + }, + { + "epoch": 0.38559782608695653, + "grad_norm": 1.3886046409606934, + "learning_rate": 3e-05, + "loss": 3.7872, + "step": 1419 + }, + { + "epoch": 0.3858695652173913, + "grad_norm": 1.178165078163147, + "learning_rate": 3e-05, + "loss": 3.1258, + "step": 1420 + }, + { + "epoch": 0.38614130434782606, + "grad_norm": 1.2557275295257568, + "learning_rate": 3e-05, + "loss": 3.4292, + "step": 1421 + }, + { + "epoch": 0.3864130434782609, + "grad_norm": 0.9806421995162964, + "learning_rate": 3e-05, + "loss": 2.8931, + "step": 1422 + }, + { + "epoch": 0.38668478260869565, + "grad_norm": 1.6238960027694702, + "learning_rate": 3e-05, + "loss": 3.5711, + "step": 1423 + }, + { + "epoch": 0.3869565217391304, + "grad_norm": 1.3578873872756958, + "learning_rate": 3e-05, + "loss": 2.9015, + "step": 1424 + }, + { + "epoch": 0.38722826086956524, + "grad_norm": 1.2328130006790161, + "learning_rate": 3e-05, + "loss": 3.2193, + "step": 1425 + }, + { + "epoch": 0.3875, + "grad_norm": 2.146702289581299, + "learning_rate": 3e-05, + "loss": 4.0931, + "step": 1426 + }, + { + "epoch": 0.3877717391304348, + "grad_norm": 1.9415775537490845, + "learning_rate": 3e-05, + "loss": 3.0855, + "step": 1427 + }, + { + "epoch": 0.38804347826086955, + "grad_norm": 1.7980471849441528, + "learning_rate": 3e-05, + "loss": 3.3307, + "step": 1428 + }, + { + "epoch": 0.38831521739130437, + "grad_norm": 1.429101824760437, + "learning_rate": 3e-05, + "loss": 3.1628, + "step": 1429 + }, + { + "epoch": 0.38858695652173914, + "grad_norm": 2.225107431411743, + "learning_rate": 3e-05, + "loss": 3.5771, + "step": 1430 + }, + { + "epoch": 0.3888586956521739, + "grad_norm": 2.46109938621521, + "learning_rate": 3e-05, + "loss": 3.7408, + "step": 1431 + }, + { + "epoch": 0.38913043478260867, + "grad_norm": 2.695655345916748, + "learning_rate": 3e-05, + "loss": 3.4776, + "step": 1432 + }, + { + "epoch": 0.3894021739130435, + "grad_norm": 2.3022725582122803, + "learning_rate": 3e-05, + "loss": 3.4633, + "step": 1433 + }, + { + "epoch": 0.38967391304347826, + "grad_norm": 1.5871728658676147, + "learning_rate": 3e-05, + "loss": 2.8911, + "step": 1434 + }, + { + "epoch": 0.389945652173913, + "grad_norm": 1.3922131061553955, + "learning_rate": 3e-05, + "loss": 2.7701, + "step": 1435 + }, + { + "epoch": 0.39021739130434785, + "grad_norm": 2.131774663925171, + "learning_rate": 3e-05, + "loss": 3.6811, + "step": 1436 + }, + { + "epoch": 0.3904891304347826, + "grad_norm": 1.8943536281585693, + "learning_rate": 3e-05, + "loss": 2.6957, + "step": 1437 + }, + { + "epoch": 0.3907608695652174, + "grad_norm": 1.7963645458221436, + "learning_rate": 3e-05, + "loss": 3.4574, + "step": 1438 + }, + { + "epoch": 0.39103260869565215, + "grad_norm": 1.613253116607666, + "learning_rate": 3e-05, + "loss": 3.5328, + "step": 1439 + }, + { + "epoch": 0.391304347826087, + "grad_norm": 1.8886051177978516, + "learning_rate": 3e-05, + "loss": 2.7514, + "step": 1440 + }, + { + "epoch": 0.39157608695652174, + "grad_norm": 2.6512999534606934, + "learning_rate": 3e-05, + "loss": 3.9306, + "step": 1441 + }, + { + "epoch": 0.3918478260869565, + "grad_norm": 1.6623197793960571, + "learning_rate": 3e-05, + "loss": 3.4983, + "step": 1442 + }, + { + "epoch": 0.39211956521739133, + "grad_norm": 1.2289677858352661, + "learning_rate": 3e-05, + "loss": 3.1967, + "step": 1443 + }, + { + "epoch": 0.3923913043478261, + "grad_norm": 0.9220126867294312, + "learning_rate": 3e-05, + "loss": 2.8288, + "step": 1444 + }, + { + "epoch": 0.39266304347826086, + "grad_norm": 1.853948950767517, + "learning_rate": 3e-05, + "loss": 3.2876, + "step": 1445 + }, + { + "epoch": 0.39293478260869563, + "grad_norm": 2.040001392364502, + "learning_rate": 3e-05, + "loss": 3.6805, + "step": 1446 + }, + { + "epoch": 0.39320652173913045, + "grad_norm": 1.4386035203933716, + "learning_rate": 3e-05, + "loss": 3.1173, + "step": 1447 + }, + { + "epoch": 0.3934782608695652, + "grad_norm": 1.48452889919281, + "learning_rate": 3e-05, + "loss": 3.3276, + "step": 1448 + }, + { + "epoch": 0.39375, + "grad_norm": 1.058858871459961, + "learning_rate": 3e-05, + "loss": 2.5619, + "step": 1449 + }, + { + "epoch": 0.39402173913043476, + "grad_norm": 2.131944179534912, + "learning_rate": 3e-05, + "loss": 3.7079, + "step": 1450 + }, + { + "epoch": 0.3942934782608696, + "grad_norm": 1.5783636569976807, + "learning_rate": 3e-05, + "loss": 3.5856, + "step": 1451 + }, + { + "epoch": 0.39456521739130435, + "grad_norm": 1.249036192893982, + "learning_rate": 3e-05, + "loss": 3.0789, + "step": 1452 + }, + { + "epoch": 0.3948369565217391, + "grad_norm": 1.6562423706054688, + "learning_rate": 3e-05, + "loss": 3.9025, + "step": 1453 + }, + { + "epoch": 0.39510869565217394, + "grad_norm": 1.6861488819122314, + "learning_rate": 3e-05, + "loss": 3.5286, + "step": 1454 + }, + { + "epoch": 0.3953804347826087, + "grad_norm": 1.503812313079834, + "learning_rate": 3e-05, + "loss": 3.0909, + "step": 1455 + }, + { + "epoch": 0.39565217391304347, + "grad_norm": 1.8524895906448364, + "learning_rate": 3e-05, + "loss": 3.4152, + "step": 1456 + }, + { + "epoch": 0.39592391304347824, + "grad_norm": 1.3550132513046265, + "learning_rate": 3e-05, + "loss": 3.5444, + "step": 1457 + }, + { + "epoch": 0.39619565217391306, + "grad_norm": 1.454416036605835, + "learning_rate": 3e-05, + "loss": 3.6969, + "step": 1458 + }, + { + "epoch": 0.3964673913043478, + "grad_norm": 1.5491242408752441, + "learning_rate": 3e-05, + "loss": 3.4632, + "step": 1459 + }, + { + "epoch": 0.3967391304347826, + "grad_norm": 1.177980661392212, + "learning_rate": 3e-05, + "loss": 3.2844, + "step": 1460 + }, + { + "epoch": 0.3970108695652174, + "grad_norm": 1.2985824346542358, + "learning_rate": 3e-05, + "loss": 3.6743, + "step": 1461 + }, + { + "epoch": 0.3972826086956522, + "grad_norm": 1.545882225036621, + "learning_rate": 3e-05, + "loss": 3.4353, + "step": 1462 + }, + { + "epoch": 0.39755434782608695, + "grad_norm": 1.3580163717269897, + "learning_rate": 3e-05, + "loss": 3.7926, + "step": 1463 + }, + { + "epoch": 0.3978260869565217, + "grad_norm": 1.4254475831985474, + "learning_rate": 3e-05, + "loss": 3.7358, + "step": 1464 + }, + { + "epoch": 0.39809782608695654, + "grad_norm": 1.2237005233764648, + "learning_rate": 3e-05, + "loss": 3.3169, + "step": 1465 + }, + { + "epoch": 0.3983695652173913, + "grad_norm": 1.1622488498687744, + "learning_rate": 3e-05, + "loss": 3.1261, + "step": 1466 + }, + { + "epoch": 0.3986413043478261, + "grad_norm": 0.9447959065437317, + "learning_rate": 3e-05, + "loss": 2.8757, + "step": 1467 + }, + { + "epoch": 0.39891304347826084, + "grad_norm": 1.5178827047348022, + "learning_rate": 3e-05, + "loss": 3.7686, + "step": 1468 + }, + { + "epoch": 0.39918478260869567, + "grad_norm": 1.2081400156021118, + "learning_rate": 3e-05, + "loss": 2.7978, + "step": 1469 + }, + { + "epoch": 0.39945652173913043, + "grad_norm": 1.3333414793014526, + "learning_rate": 3e-05, + "loss": 3.8005, + "step": 1470 + }, + { + "epoch": 0.3997282608695652, + "grad_norm": 1.2284449338912964, + "learning_rate": 3e-05, + "loss": 3.2384, + "step": 1471 + }, + { + "epoch": 0.4, + "grad_norm": 1.6261060237884521, + "learning_rate": 3e-05, + "loss": 3.6663, + "step": 1472 + }, + { + "epoch": 0.4002717391304348, + "grad_norm": 1.2063846588134766, + "learning_rate": 3e-05, + "loss": 3.2808, + "step": 1473 + }, + { + "epoch": 0.40054347826086956, + "grad_norm": 1.1668663024902344, + "learning_rate": 3e-05, + "loss": 3.6527, + "step": 1474 + }, + { + "epoch": 0.4008152173913043, + "grad_norm": 1.1682161092758179, + "learning_rate": 3e-05, + "loss": 3.0487, + "step": 1475 + }, + { + "epoch": 0.40108695652173915, + "grad_norm": 1.2479826211929321, + "learning_rate": 3e-05, + "loss": 3.8231, + "step": 1476 + }, + { + "epoch": 0.4013586956521739, + "grad_norm": 1.3903083801269531, + "learning_rate": 3e-05, + "loss": 3.3589, + "step": 1477 + }, + { + "epoch": 0.4016304347826087, + "grad_norm": 2.756995677947998, + "learning_rate": 3e-05, + "loss": 3.5956, + "step": 1478 + }, + { + "epoch": 0.4019021739130435, + "grad_norm": 1.1125861406326294, + "learning_rate": 3e-05, + "loss": 2.9482, + "step": 1479 + }, + { + "epoch": 0.40217391304347827, + "grad_norm": 1.7494776248931885, + "learning_rate": 3e-05, + "loss": 3.6373, + "step": 1480 + }, + { + "epoch": 0.40244565217391304, + "grad_norm": 1.7383618354797363, + "learning_rate": 3e-05, + "loss": 3.5196, + "step": 1481 + }, + { + "epoch": 0.4027173913043478, + "grad_norm": 2.0332143306732178, + "learning_rate": 3e-05, + "loss": 3.4659, + "step": 1482 + }, + { + "epoch": 0.4029891304347826, + "grad_norm": 1.7549830675125122, + "learning_rate": 3e-05, + "loss": 3.6027, + "step": 1483 + }, + { + "epoch": 0.4032608695652174, + "grad_norm": 1.4020575284957886, + "learning_rate": 3e-05, + "loss": 3.1451, + "step": 1484 + }, + { + "epoch": 0.40353260869565216, + "grad_norm": 1.134787917137146, + "learning_rate": 3e-05, + "loss": 2.6124, + "step": 1485 + }, + { + "epoch": 0.40380434782608693, + "grad_norm": 1.6569262742996216, + "learning_rate": 3e-05, + "loss": 3.8979, + "step": 1486 + }, + { + "epoch": 0.40407608695652175, + "grad_norm": 1.7637388706207275, + "learning_rate": 3e-05, + "loss": 3.8109, + "step": 1487 + }, + { + "epoch": 0.4043478260869565, + "grad_norm": 1.2442704439163208, + "learning_rate": 3e-05, + "loss": 3.6988, + "step": 1488 + }, + { + "epoch": 0.4046195652173913, + "grad_norm": 1.0942063331604004, + "learning_rate": 3e-05, + "loss": 3.0464, + "step": 1489 + }, + { + "epoch": 0.4048913043478261, + "grad_norm": 1.337386965751648, + "learning_rate": 3e-05, + "loss": 3.8127, + "step": 1490 + }, + { + "epoch": 0.4051630434782609, + "grad_norm": 1.7307548522949219, + "learning_rate": 3e-05, + "loss": 3.3854, + "step": 1491 + }, + { + "epoch": 0.40543478260869564, + "grad_norm": 3.1193931102752686, + "learning_rate": 3e-05, + "loss": 4.9053, + "step": 1492 + }, + { + "epoch": 0.4057065217391304, + "grad_norm": 1.5950829982757568, + "learning_rate": 3e-05, + "loss": 3.6687, + "step": 1493 + }, + { + "epoch": 0.40597826086956523, + "grad_norm": 1.6609991788864136, + "learning_rate": 3e-05, + "loss": 3.4985, + "step": 1494 + }, + { + "epoch": 0.40625, + "grad_norm": 1.824634075164795, + "learning_rate": 3e-05, + "loss": 2.7576, + "step": 1495 + }, + { + "epoch": 0.40652173913043477, + "grad_norm": 1.8175928592681885, + "learning_rate": 3e-05, + "loss": 3.7615, + "step": 1496 + }, + { + "epoch": 0.4067934782608696, + "grad_norm": 1.3160499334335327, + "learning_rate": 3e-05, + "loss": 3.1647, + "step": 1497 + }, + { + "epoch": 0.40706521739130436, + "grad_norm": 1.7744853496551514, + "learning_rate": 3e-05, + "loss": 3.0396, + "step": 1498 + }, + { + "epoch": 0.4073369565217391, + "grad_norm": 2.0241026878356934, + "learning_rate": 3e-05, + "loss": 3.8311, + "step": 1499 + }, + { + "epoch": 0.4076086956521739, + "grad_norm": 1.7789679765701294, + "learning_rate": 3e-05, + "loss": 2.9538, + "step": 1500 + }, + { + "epoch": 0.4078804347826087, + "grad_norm": 1.4469610452651978, + "learning_rate": 3e-05, + "loss": 3.3083, + "step": 1501 + }, + { + "epoch": 0.4081521739130435, + "grad_norm": 1.570556879043579, + "learning_rate": 3e-05, + "loss": 3.0793, + "step": 1502 + }, + { + "epoch": 0.40842391304347825, + "grad_norm": 1.3964000940322876, + "learning_rate": 3e-05, + "loss": 2.9603, + "step": 1503 + }, + { + "epoch": 0.40869565217391307, + "grad_norm": 1.1800280809402466, + "learning_rate": 3e-05, + "loss": 3.2305, + "step": 1504 + }, + { + "epoch": 0.40896739130434784, + "grad_norm": 1.762847661972046, + "learning_rate": 3e-05, + "loss": 3.7564, + "step": 1505 + }, + { + "epoch": 0.4092391304347826, + "grad_norm": 1.5233442783355713, + "learning_rate": 3e-05, + "loss": 3.2744, + "step": 1506 + }, + { + "epoch": 0.4095108695652174, + "grad_norm": 1.9167475700378418, + "learning_rate": 3e-05, + "loss": 3.6962, + "step": 1507 + }, + { + "epoch": 0.4097826086956522, + "grad_norm": 1.949634075164795, + "learning_rate": 3e-05, + "loss": 3.1084, + "step": 1508 + }, + { + "epoch": 0.41005434782608696, + "grad_norm": 1.9483072757720947, + "learning_rate": 3e-05, + "loss": 3.1895, + "step": 1509 + }, + { + "epoch": 0.41032608695652173, + "grad_norm": 1.368509292602539, + "learning_rate": 3e-05, + "loss": 2.6804, + "step": 1510 + }, + { + "epoch": 0.4105978260869565, + "grad_norm": 1.6476247310638428, + "learning_rate": 3e-05, + "loss": 3.1565, + "step": 1511 + }, + { + "epoch": 0.4108695652173913, + "grad_norm": 2.14841890335083, + "learning_rate": 3e-05, + "loss": 3.5723, + "step": 1512 + }, + { + "epoch": 0.4111413043478261, + "grad_norm": 2.1899356842041016, + "learning_rate": 3e-05, + "loss": 3.9324, + "step": 1513 + }, + { + "epoch": 0.41141304347826085, + "grad_norm": 1.971421480178833, + "learning_rate": 3e-05, + "loss": 3.3302, + "step": 1514 + }, + { + "epoch": 0.4116847826086957, + "grad_norm": 1.555216908454895, + "learning_rate": 3e-05, + "loss": 2.8559, + "step": 1515 + }, + { + "epoch": 0.41195652173913044, + "grad_norm": 2.025975227355957, + "learning_rate": 3e-05, + "loss": 3.8605, + "step": 1516 + }, + { + "epoch": 0.4122282608695652, + "grad_norm": 1.609168291091919, + "learning_rate": 3e-05, + "loss": 3.521, + "step": 1517 + }, + { + "epoch": 0.4125, + "grad_norm": 1.4011493921279907, + "learning_rate": 3e-05, + "loss": 2.905, + "step": 1518 + }, + { + "epoch": 0.4127717391304348, + "grad_norm": 1.813499927520752, + "learning_rate": 3e-05, + "loss": 3.9861, + "step": 1519 + }, + { + "epoch": 0.41304347826086957, + "grad_norm": 2.023961067199707, + "learning_rate": 3e-05, + "loss": 3.3814, + "step": 1520 + }, + { + "epoch": 0.41331521739130433, + "grad_norm": 1.9356021881103516, + "learning_rate": 3e-05, + "loss": 2.9743, + "step": 1521 + }, + { + "epoch": 0.41358695652173916, + "grad_norm": 1.6765419244766235, + "learning_rate": 3e-05, + "loss": 2.9504, + "step": 1522 + }, + { + "epoch": 0.4138586956521739, + "grad_norm": 1.954119324684143, + "learning_rate": 3e-05, + "loss": 3.7254, + "step": 1523 + }, + { + "epoch": 0.4141304347826087, + "grad_norm": 1.5637805461883545, + "learning_rate": 3e-05, + "loss": 3.8543, + "step": 1524 + }, + { + "epoch": 0.41440217391304346, + "grad_norm": 1.175187587738037, + "learning_rate": 3e-05, + "loss": 3.2871, + "step": 1525 + }, + { + "epoch": 0.4146739130434783, + "grad_norm": 1.3228040933609009, + "learning_rate": 3e-05, + "loss": 2.9662, + "step": 1526 + }, + { + "epoch": 0.41494565217391305, + "grad_norm": 1.4776461124420166, + "learning_rate": 3e-05, + "loss": 2.9617, + "step": 1527 + }, + { + "epoch": 0.4152173913043478, + "grad_norm": 1.5169886350631714, + "learning_rate": 3e-05, + "loss": 3.6288, + "step": 1528 + }, + { + "epoch": 0.4154891304347826, + "grad_norm": 1.4785473346710205, + "learning_rate": 3e-05, + "loss": 3.0807, + "step": 1529 + }, + { + "epoch": 0.4157608695652174, + "grad_norm": 1.4565258026123047, + "learning_rate": 3e-05, + "loss": 3.3099, + "step": 1530 + }, + { + "epoch": 0.4160326086956522, + "grad_norm": 1.2732715606689453, + "learning_rate": 3e-05, + "loss": 3.2948, + "step": 1531 + }, + { + "epoch": 0.41630434782608694, + "grad_norm": 1.0217174291610718, + "learning_rate": 3e-05, + "loss": 3.0084, + "step": 1532 + }, + { + "epoch": 0.41657608695652176, + "grad_norm": 1.2822792530059814, + "learning_rate": 3e-05, + "loss": 2.9284, + "step": 1533 + }, + { + "epoch": 0.41684782608695653, + "grad_norm": 1.641827940940857, + "learning_rate": 3e-05, + "loss": 3.4684, + "step": 1534 + }, + { + "epoch": 0.4171195652173913, + "grad_norm": 1.5296562910079956, + "learning_rate": 3e-05, + "loss": 2.5836, + "step": 1535 + }, + { + "epoch": 0.41739130434782606, + "grad_norm": 1.3570102453231812, + "learning_rate": 3e-05, + "loss": 3.1323, + "step": 1536 + }, + { + "epoch": 0.4176630434782609, + "grad_norm": 0.9901162981987, + "learning_rate": 3e-05, + "loss": 2.7749, + "step": 1537 + }, + { + "epoch": 0.41793478260869565, + "grad_norm": 1.2664462327957153, + "learning_rate": 3e-05, + "loss": 3.1213, + "step": 1538 + }, + { + "epoch": 0.4182065217391304, + "grad_norm": 1.682426929473877, + "learning_rate": 3e-05, + "loss": 3.8292, + "step": 1539 + }, + { + "epoch": 0.41847826086956524, + "grad_norm": 1.3441531658172607, + "learning_rate": 3e-05, + "loss": 2.8989, + "step": 1540 + }, + { + "epoch": 0.41875, + "grad_norm": 1.6413555145263672, + "learning_rate": 3e-05, + "loss": 3.4333, + "step": 1541 + }, + { + "epoch": 0.4190217391304348, + "grad_norm": 1.2601666450500488, + "learning_rate": 3e-05, + "loss": 3.2815, + "step": 1542 + }, + { + "epoch": 0.41929347826086955, + "grad_norm": 1.5582549571990967, + "learning_rate": 3e-05, + "loss": 3.5814, + "step": 1543 + }, + { + "epoch": 0.41956521739130437, + "grad_norm": 1.0220012664794922, + "learning_rate": 3e-05, + "loss": 2.7242, + "step": 1544 + }, + { + "epoch": 0.41983695652173914, + "grad_norm": 1.710802674293518, + "learning_rate": 3e-05, + "loss": 3.5271, + "step": 1545 + }, + { + "epoch": 0.4201086956521739, + "grad_norm": 2.3148810863494873, + "learning_rate": 3e-05, + "loss": 4.0931, + "step": 1546 + }, + { + "epoch": 0.42038043478260867, + "grad_norm": 1.4409066438674927, + "learning_rate": 3e-05, + "loss": 3.4806, + "step": 1547 + }, + { + "epoch": 0.4206521739130435, + "grad_norm": 1.4869251251220703, + "learning_rate": 3e-05, + "loss": 3.8039, + "step": 1548 + }, + { + "epoch": 0.42092391304347826, + "grad_norm": 2.2359511852264404, + "learning_rate": 3e-05, + "loss": 4.1842, + "step": 1549 + }, + { + "epoch": 0.421195652173913, + "grad_norm": 2.2962846755981445, + "learning_rate": 3e-05, + "loss": 3.8559, + "step": 1550 + }, + { + "epoch": 0.42146739130434785, + "grad_norm": 1.5050580501556396, + "learning_rate": 3e-05, + "loss": 3.6678, + "step": 1551 + }, + { + "epoch": 0.4217391304347826, + "grad_norm": 1.2851321697235107, + "learning_rate": 3e-05, + "loss": 3.4175, + "step": 1552 + }, + { + "epoch": 0.4220108695652174, + "grad_norm": 1.2671363353729248, + "learning_rate": 3e-05, + "loss": 3.0412, + "step": 1553 + }, + { + "epoch": 0.42228260869565215, + "grad_norm": 1.2936915159225464, + "learning_rate": 3e-05, + "loss": 3.0918, + "step": 1554 + }, + { + "epoch": 0.422554347826087, + "grad_norm": 1.4296749830245972, + "learning_rate": 3e-05, + "loss": 3.3257, + "step": 1555 + }, + { + "epoch": 0.42282608695652174, + "grad_norm": 1.2652101516723633, + "learning_rate": 3e-05, + "loss": 3.5451, + "step": 1556 + }, + { + "epoch": 0.4230978260869565, + "grad_norm": 1.4265552759170532, + "learning_rate": 3e-05, + "loss": 3.4145, + "step": 1557 + }, + { + "epoch": 0.42336956521739133, + "grad_norm": 1.3775297403335571, + "learning_rate": 3e-05, + "loss": 3.3248, + "step": 1558 + }, + { + "epoch": 0.4236413043478261, + "grad_norm": 1.8461849689483643, + "learning_rate": 3e-05, + "loss": 3.9212, + "step": 1559 + }, + { + "epoch": 0.42391304347826086, + "grad_norm": 1.412374496459961, + "learning_rate": 3e-05, + "loss": 3.4737, + "step": 1560 + }, + { + "epoch": 0.42418478260869563, + "grad_norm": 1.6272127628326416, + "learning_rate": 3e-05, + "loss": 3.4379, + "step": 1561 + }, + { + "epoch": 0.42445652173913045, + "grad_norm": 1.754745364189148, + "learning_rate": 3e-05, + "loss": 3.4164, + "step": 1562 + }, + { + "epoch": 0.4247282608695652, + "grad_norm": 1.2676855325698853, + "learning_rate": 3e-05, + "loss": 3.3106, + "step": 1563 + }, + { + "epoch": 0.425, + "grad_norm": 1.7627284526824951, + "learning_rate": 3e-05, + "loss": 3.841, + "step": 1564 + }, + { + "epoch": 0.42527173913043476, + "grad_norm": 1.143531084060669, + "learning_rate": 3e-05, + "loss": 2.82, + "step": 1565 + }, + { + "epoch": 0.4255434782608696, + "grad_norm": 1.0254485607147217, + "learning_rate": 3e-05, + "loss": 2.9556, + "step": 1566 + }, + { + "epoch": 0.42581521739130435, + "grad_norm": 1.3668675422668457, + "learning_rate": 3e-05, + "loss": 3.7105, + "step": 1567 + }, + { + "epoch": 0.4260869565217391, + "grad_norm": 2.208592176437378, + "learning_rate": 3e-05, + "loss": 3.4152, + "step": 1568 + }, + { + "epoch": 0.42635869565217394, + "grad_norm": 1.1941901445388794, + "learning_rate": 3e-05, + "loss": 3.3036, + "step": 1569 + }, + { + "epoch": 0.4266304347826087, + "grad_norm": 1.536787748336792, + "learning_rate": 3e-05, + "loss": 3.5194, + "step": 1570 + }, + { + "epoch": 0.42690217391304347, + "grad_norm": 1.4054354429244995, + "learning_rate": 3e-05, + "loss": 3.263, + "step": 1571 + }, + { + "epoch": 0.42717391304347824, + "grad_norm": 1.6471023559570312, + "learning_rate": 3e-05, + "loss": 3.4855, + "step": 1572 + }, + { + "epoch": 0.42744565217391306, + "grad_norm": 0.9407289028167725, + "learning_rate": 3e-05, + "loss": 3.1382, + "step": 1573 + }, + { + "epoch": 0.4277173913043478, + "grad_norm": 1.244971513748169, + "learning_rate": 3e-05, + "loss": 3.4904, + "step": 1574 + }, + { + "epoch": 0.4279891304347826, + "grad_norm": 1.305245041847229, + "learning_rate": 3e-05, + "loss": 3.3049, + "step": 1575 + }, + { + "epoch": 0.4282608695652174, + "grad_norm": 1.2542614936828613, + "learning_rate": 3e-05, + "loss": 3.4931, + "step": 1576 + }, + { + "epoch": 0.4285326086956522, + "grad_norm": 0.9759632349014282, + "learning_rate": 3e-05, + "loss": 3.4818, + "step": 1577 + }, + { + "epoch": 0.42880434782608695, + "grad_norm": 1.4484260082244873, + "learning_rate": 3e-05, + "loss": 3.2934, + "step": 1578 + }, + { + "epoch": 0.4290760869565217, + "grad_norm": 1.1898603439331055, + "learning_rate": 3e-05, + "loss": 3.4601, + "step": 1579 + }, + { + "epoch": 0.42934782608695654, + "grad_norm": 1.486471176147461, + "learning_rate": 3e-05, + "loss": 3.762, + "step": 1580 + }, + { + "epoch": 0.4296195652173913, + "grad_norm": 1.2455788850784302, + "learning_rate": 3e-05, + "loss": 3.721, + "step": 1581 + }, + { + "epoch": 0.4298913043478261, + "grad_norm": 1.1194945573806763, + "learning_rate": 3e-05, + "loss": 3.3573, + "step": 1582 + }, + { + "epoch": 0.43016304347826084, + "grad_norm": 1.666195273399353, + "learning_rate": 3e-05, + "loss": 3.531, + "step": 1583 + }, + { + "epoch": 0.43043478260869567, + "grad_norm": 1.192292332649231, + "learning_rate": 3e-05, + "loss": 3.1313, + "step": 1584 + }, + { + "epoch": 0.43070652173913043, + "grad_norm": 1.1080567836761475, + "learning_rate": 3e-05, + "loss": 3.2448, + "step": 1585 + }, + { + "epoch": 0.4309782608695652, + "grad_norm": 1.7732088565826416, + "learning_rate": 3e-05, + "loss": 3.4527, + "step": 1586 + }, + { + "epoch": 0.43125, + "grad_norm": 1.2465471029281616, + "learning_rate": 3e-05, + "loss": 2.8915, + "step": 1587 + }, + { + "epoch": 0.4315217391304348, + "grad_norm": 1.43854820728302, + "learning_rate": 3e-05, + "loss": 3.2366, + "step": 1588 + }, + { + "epoch": 0.43179347826086956, + "grad_norm": 1.3088682889938354, + "learning_rate": 3e-05, + "loss": 3.3654, + "step": 1589 + }, + { + "epoch": 0.4320652173913043, + "grad_norm": 1.0148249864578247, + "learning_rate": 3e-05, + "loss": 3.083, + "step": 1590 + }, + { + "epoch": 0.43233695652173915, + "grad_norm": 1.2219303846359253, + "learning_rate": 3e-05, + "loss": 3.0298, + "step": 1591 + }, + { + "epoch": 0.4326086956521739, + "grad_norm": 1.081207275390625, + "learning_rate": 3e-05, + "loss": 3.2116, + "step": 1592 + }, + { + "epoch": 0.4328804347826087, + "grad_norm": 1.515478491783142, + "learning_rate": 3e-05, + "loss": 3.4368, + "step": 1593 + }, + { + "epoch": 0.4331521739130435, + "grad_norm": 1.073887586593628, + "learning_rate": 3e-05, + "loss": 3.1832, + "step": 1594 + }, + { + "epoch": 0.43342391304347827, + "grad_norm": 1.1452287435531616, + "learning_rate": 3e-05, + "loss": 3.5617, + "step": 1595 + }, + { + "epoch": 0.43369565217391304, + "grad_norm": 1.512242078781128, + "learning_rate": 3e-05, + "loss": 3.4468, + "step": 1596 + }, + { + "epoch": 0.4339673913043478, + "grad_norm": 1.1862598657608032, + "learning_rate": 3e-05, + "loss": 3.3795, + "step": 1597 + }, + { + "epoch": 0.4342391304347826, + "grad_norm": 1.4696996212005615, + "learning_rate": 3e-05, + "loss": 3.5239, + "step": 1598 + }, + { + "epoch": 0.4345108695652174, + "grad_norm": 1.5985097885131836, + "learning_rate": 3e-05, + "loss": 3.8709, + "step": 1599 + }, + { + "epoch": 0.43478260869565216, + "grad_norm": 1.3134711980819702, + "learning_rate": 3e-05, + "loss": 3.4582, + "step": 1600 + }, + { + "epoch": 0.43505434782608693, + "grad_norm": 1.2656311988830566, + "learning_rate": 3e-05, + "loss": 3.7025, + "step": 1601 + }, + { + "epoch": 0.43532608695652175, + "grad_norm": 1.1780740022659302, + "learning_rate": 3e-05, + "loss": 3.3486, + "step": 1602 + }, + { + "epoch": 0.4355978260869565, + "grad_norm": 1.2048178911209106, + "learning_rate": 3e-05, + "loss": 3.3115, + "step": 1603 + }, + { + "epoch": 0.4358695652173913, + "grad_norm": 1.3589476346969604, + "learning_rate": 3e-05, + "loss": 3.2494, + "step": 1604 + }, + { + "epoch": 0.4361413043478261, + "grad_norm": 1.3171757459640503, + "learning_rate": 3e-05, + "loss": 2.8635, + "step": 1605 + }, + { + "epoch": 0.4364130434782609, + "grad_norm": 1.1358094215393066, + "learning_rate": 3e-05, + "loss": 2.9426, + "step": 1606 + }, + { + "epoch": 0.43668478260869564, + "grad_norm": 1.376583456993103, + "learning_rate": 3e-05, + "loss": 3.2611, + "step": 1607 + }, + { + "epoch": 0.4369565217391304, + "grad_norm": 1.7452890872955322, + "learning_rate": 3e-05, + "loss": 3.7445, + "step": 1608 + }, + { + "epoch": 0.43722826086956523, + "grad_norm": 1.715166687965393, + "learning_rate": 3e-05, + "loss": 3.4322, + "step": 1609 + }, + { + "epoch": 0.4375, + "grad_norm": 1.3614475727081299, + "learning_rate": 3e-05, + "loss": 3.5131, + "step": 1610 + }, + { + "epoch": 0.43777173913043477, + "grad_norm": 1.4811841249465942, + "learning_rate": 3e-05, + "loss": 3.714, + "step": 1611 + }, + { + "epoch": 0.4380434782608696, + "grad_norm": 1.071926236152649, + "learning_rate": 3e-05, + "loss": 3.0734, + "step": 1612 + }, + { + "epoch": 0.43831521739130436, + "grad_norm": 1.5618953704833984, + "learning_rate": 3e-05, + "loss": 3.876, + "step": 1613 + }, + { + "epoch": 0.4385869565217391, + "grad_norm": 1.1414976119995117, + "learning_rate": 3e-05, + "loss": 3.5318, + "step": 1614 + }, + { + "epoch": 0.4388586956521739, + "grad_norm": 1.0475927591323853, + "learning_rate": 3e-05, + "loss": 3.293, + "step": 1615 + }, + { + "epoch": 0.4391304347826087, + "grad_norm": 1.392609715461731, + "learning_rate": 3e-05, + "loss": 2.9974, + "step": 1616 + }, + { + "epoch": 0.4394021739130435, + "grad_norm": 0.906802237033844, + "learning_rate": 3e-05, + "loss": 2.9825, + "step": 1617 + }, + { + "epoch": 0.43967391304347825, + "grad_norm": 1.4019746780395508, + "learning_rate": 3e-05, + "loss": 3.8698, + "step": 1618 + }, + { + "epoch": 0.43994565217391307, + "grad_norm": 1.321829915046692, + "learning_rate": 3e-05, + "loss": 3.7109, + "step": 1619 + }, + { + "epoch": 0.44021739130434784, + "grad_norm": 1.0838751792907715, + "learning_rate": 3e-05, + "loss": 2.6205, + "step": 1620 + }, + { + "epoch": 0.4404891304347826, + "grad_norm": 2.383323907852173, + "learning_rate": 3e-05, + "loss": 4.2267, + "step": 1621 + }, + { + "epoch": 0.4407608695652174, + "grad_norm": 1.0130791664123535, + "learning_rate": 3e-05, + "loss": 2.788, + "step": 1622 + }, + { + "epoch": 0.4410326086956522, + "grad_norm": 1.3985093832015991, + "learning_rate": 3e-05, + "loss": 3.1002, + "step": 1623 + }, + { + "epoch": 0.44130434782608696, + "grad_norm": 1.3819383382797241, + "learning_rate": 3e-05, + "loss": 3.2219, + "step": 1624 + }, + { + "epoch": 0.44157608695652173, + "grad_norm": 1.4096386432647705, + "learning_rate": 3e-05, + "loss": 3.6886, + "step": 1625 + }, + { + "epoch": 0.4418478260869565, + "grad_norm": 1.4819231033325195, + "learning_rate": 3e-05, + "loss": 3.2191, + "step": 1626 + }, + { + "epoch": 0.4421195652173913, + "grad_norm": 1.3344846963882446, + "learning_rate": 3e-05, + "loss": 3.3927, + "step": 1627 + }, + { + "epoch": 0.4423913043478261, + "grad_norm": 1.469277024269104, + "learning_rate": 3e-05, + "loss": 3.4069, + "step": 1628 + }, + { + "epoch": 0.44266304347826085, + "grad_norm": 1.0877147912979126, + "learning_rate": 3e-05, + "loss": 2.9891, + "step": 1629 + }, + { + "epoch": 0.4429347826086957, + "grad_norm": 1.1775012016296387, + "learning_rate": 3e-05, + "loss": 3.3463, + "step": 1630 + }, + { + "epoch": 0.44320652173913044, + "grad_norm": 1.240936517715454, + "learning_rate": 3e-05, + "loss": 3.5751, + "step": 1631 + }, + { + "epoch": 0.4434782608695652, + "grad_norm": 0.877865195274353, + "learning_rate": 3e-05, + "loss": 2.7317, + "step": 1632 + }, + { + "epoch": 0.44375, + "grad_norm": 1.1812602281570435, + "learning_rate": 3e-05, + "loss": 3.0307, + "step": 1633 + }, + { + "epoch": 0.4440217391304348, + "grad_norm": 1.7113842964172363, + "learning_rate": 3e-05, + "loss": 4.1471, + "step": 1634 + }, + { + "epoch": 0.44429347826086957, + "grad_norm": 1.1820313930511475, + "learning_rate": 3e-05, + "loss": 3.362, + "step": 1635 + }, + { + "epoch": 0.44456521739130433, + "grad_norm": 1.3378487825393677, + "learning_rate": 3e-05, + "loss": 3.5815, + "step": 1636 + }, + { + "epoch": 0.44483695652173916, + "grad_norm": 0.8756458163261414, + "learning_rate": 3e-05, + "loss": 2.6882, + "step": 1637 + }, + { + "epoch": 0.4451086956521739, + "grad_norm": 1.020820140838623, + "learning_rate": 3e-05, + "loss": 3.1006, + "step": 1638 + }, + { + "epoch": 0.4453804347826087, + "grad_norm": 1.5075011253356934, + "learning_rate": 3e-05, + "loss": 4.0147, + "step": 1639 + }, + { + "epoch": 0.44565217391304346, + "grad_norm": 1.1874289512634277, + "learning_rate": 3e-05, + "loss": 3.0811, + "step": 1640 + }, + { + "epoch": 0.4459239130434783, + "grad_norm": 1.5765689611434937, + "learning_rate": 3e-05, + "loss": 3.4697, + "step": 1641 + }, + { + "epoch": 0.44619565217391305, + "grad_norm": 1.3812084197998047, + "learning_rate": 3e-05, + "loss": 3.6468, + "step": 1642 + }, + { + "epoch": 0.4464673913043478, + "grad_norm": 1.1477484703063965, + "learning_rate": 3e-05, + "loss": 3.2347, + "step": 1643 + }, + { + "epoch": 0.4467391304347826, + "grad_norm": 1.431220531463623, + "learning_rate": 3e-05, + "loss": 3.8073, + "step": 1644 + }, + { + "epoch": 0.4470108695652174, + "grad_norm": 1.0271748304367065, + "learning_rate": 3e-05, + "loss": 3.2165, + "step": 1645 + }, + { + "epoch": 0.4472826086956522, + "grad_norm": 1.2913216352462769, + "learning_rate": 3e-05, + "loss": 3.2766, + "step": 1646 + }, + { + "epoch": 0.44755434782608694, + "grad_norm": 1.341670274734497, + "learning_rate": 3e-05, + "loss": 3.8426, + "step": 1647 + }, + { + "epoch": 0.44782608695652176, + "grad_norm": 1.1871728897094727, + "learning_rate": 3e-05, + "loss": 3.1251, + "step": 1648 + }, + { + "epoch": 0.44809782608695653, + "grad_norm": 1.2151528596878052, + "learning_rate": 3e-05, + "loss": 3.3811, + "step": 1649 + }, + { + "epoch": 0.4483695652173913, + "grad_norm": 1.7120990753173828, + "learning_rate": 3e-05, + "loss": 3.7093, + "step": 1650 + }, + { + "epoch": 0.44864130434782606, + "grad_norm": 1.1526025533676147, + "learning_rate": 3e-05, + "loss": 3.1806, + "step": 1651 + }, + { + "epoch": 0.4489130434782609, + "grad_norm": 1.5061256885528564, + "learning_rate": 3e-05, + "loss": 3.7246, + "step": 1652 + }, + { + "epoch": 0.44918478260869565, + "grad_norm": 1.442798376083374, + "learning_rate": 3e-05, + "loss": 2.8975, + "step": 1653 + }, + { + "epoch": 0.4494565217391304, + "grad_norm": 1.5733429193496704, + "learning_rate": 3e-05, + "loss": 3.7047, + "step": 1654 + }, + { + "epoch": 0.44972826086956524, + "grad_norm": 1.5891181230545044, + "learning_rate": 3e-05, + "loss": 3.5587, + "step": 1655 + }, + { + "epoch": 0.45, + "grad_norm": 2.2529733180999756, + "learning_rate": 3e-05, + "loss": 3.079, + "step": 1656 + }, + { + "epoch": 0.4502717391304348, + "grad_norm": 2.397771120071411, + "learning_rate": 3e-05, + "loss": 3.2517, + "step": 1657 + }, + { + "epoch": 0.45054347826086955, + "grad_norm": 1.5895618200302124, + "learning_rate": 3e-05, + "loss": 3.289, + "step": 1658 + }, + { + "epoch": 0.45081521739130437, + "grad_norm": 1.3885931968688965, + "learning_rate": 3e-05, + "loss": 3.8401, + "step": 1659 + }, + { + "epoch": 0.45108695652173914, + "grad_norm": 1.848649263381958, + "learning_rate": 3e-05, + "loss": 3.2375, + "step": 1660 + }, + { + "epoch": 0.4513586956521739, + "grad_norm": 2.288536787033081, + "learning_rate": 3e-05, + "loss": 3.0737, + "step": 1661 + }, + { + "epoch": 0.45163043478260867, + "grad_norm": 1.5626355409622192, + "learning_rate": 3e-05, + "loss": 3.6127, + "step": 1662 + }, + { + "epoch": 0.4519021739130435, + "grad_norm": 1.2159314155578613, + "learning_rate": 3e-05, + "loss": 3.1681, + "step": 1663 + }, + { + "epoch": 0.45217391304347826, + "grad_norm": 2.7480335235595703, + "learning_rate": 3e-05, + "loss": 3.3753, + "step": 1664 + }, + { + "epoch": 0.452445652173913, + "grad_norm": 1.5772864818572998, + "learning_rate": 3e-05, + "loss": 3.4648, + "step": 1665 + }, + { + "epoch": 0.45271739130434785, + "grad_norm": 2.1393795013427734, + "learning_rate": 3e-05, + "loss": 4.067, + "step": 1666 + }, + { + "epoch": 0.4529891304347826, + "grad_norm": 1.48007071018219, + "learning_rate": 3e-05, + "loss": 3.0416, + "step": 1667 + }, + { + "epoch": 0.4532608695652174, + "grad_norm": 1.3958046436309814, + "learning_rate": 3e-05, + "loss": 2.9871, + "step": 1668 + }, + { + "epoch": 0.45353260869565215, + "grad_norm": 1.6530804634094238, + "learning_rate": 3e-05, + "loss": 3.1665, + "step": 1669 + }, + { + "epoch": 0.453804347826087, + "grad_norm": 1.5180555582046509, + "learning_rate": 3e-05, + "loss": 3.1545, + "step": 1670 + }, + { + "epoch": 0.45407608695652174, + "grad_norm": 1.3946945667266846, + "learning_rate": 3e-05, + "loss": 3.0606, + "step": 1671 + }, + { + "epoch": 0.4543478260869565, + "grad_norm": 1.655504584312439, + "learning_rate": 3e-05, + "loss": 3.5141, + "step": 1672 + }, + { + "epoch": 0.45461956521739133, + "grad_norm": 1.8746256828308105, + "learning_rate": 3e-05, + "loss": 3.2537, + "step": 1673 + }, + { + "epoch": 0.4548913043478261, + "grad_norm": 1.6152852773666382, + "learning_rate": 3e-05, + "loss": 3.1725, + "step": 1674 + }, + { + "epoch": 0.45516304347826086, + "grad_norm": 2.030758857727051, + "learning_rate": 3e-05, + "loss": 4.1911, + "step": 1675 + }, + { + "epoch": 0.45543478260869563, + "grad_norm": 1.573011040687561, + "learning_rate": 3e-05, + "loss": 2.7004, + "step": 1676 + }, + { + "epoch": 0.45570652173913045, + "grad_norm": 1.7782936096191406, + "learning_rate": 3e-05, + "loss": 3.4129, + "step": 1677 + }, + { + "epoch": 0.4559782608695652, + "grad_norm": 1.6364409923553467, + "learning_rate": 3e-05, + "loss": 3.7828, + "step": 1678 + }, + { + "epoch": 0.45625, + "grad_norm": 1.095005750656128, + "learning_rate": 3e-05, + "loss": 2.9575, + "step": 1679 + }, + { + "epoch": 0.45652173913043476, + "grad_norm": 1.4323863983154297, + "learning_rate": 3e-05, + "loss": 2.9855, + "step": 1680 + }, + { + "epoch": 0.4567934782608696, + "grad_norm": 1.8281795978546143, + "learning_rate": 3e-05, + "loss": 3.2274, + "step": 1681 + }, + { + "epoch": 0.45706521739130435, + "grad_norm": 1.553568959236145, + "learning_rate": 3e-05, + "loss": 3.4245, + "step": 1682 + }, + { + "epoch": 0.4573369565217391, + "grad_norm": 1.229791522026062, + "learning_rate": 3e-05, + "loss": 2.8749, + "step": 1683 + }, + { + "epoch": 0.45760869565217394, + "grad_norm": 1.4264166355133057, + "learning_rate": 3e-05, + "loss": 3.0987, + "step": 1684 + }, + { + "epoch": 0.4578804347826087, + "grad_norm": 1.4009737968444824, + "learning_rate": 3e-05, + "loss": 3.2024, + "step": 1685 + }, + { + "epoch": 0.45815217391304347, + "grad_norm": 1.2461766004562378, + "learning_rate": 3e-05, + "loss": 3.1414, + "step": 1686 + }, + { + "epoch": 0.45842391304347824, + "grad_norm": 1.1765496730804443, + "learning_rate": 3e-05, + "loss": 3.7237, + "step": 1687 + }, + { + "epoch": 0.45869565217391306, + "grad_norm": 1.2187762260437012, + "learning_rate": 3e-05, + "loss": 3.5997, + "step": 1688 + }, + { + "epoch": 0.4589673913043478, + "grad_norm": 2.117086410522461, + "learning_rate": 3e-05, + "loss": 3.5497, + "step": 1689 + }, + { + "epoch": 0.4592391304347826, + "grad_norm": 1.2980027198791504, + "learning_rate": 3e-05, + "loss": 3.6001, + "step": 1690 + }, + { + "epoch": 0.4595108695652174, + "grad_norm": 1.1031181812286377, + "learning_rate": 3e-05, + "loss": 2.814, + "step": 1691 + }, + { + "epoch": 0.4597826086956522, + "grad_norm": 1.201127052307129, + "learning_rate": 3e-05, + "loss": 3.4627, + "step": 1692 + }, + { + "epoch": 0.46005434782608695, + "grad_norm": 1.9975850582122803, + "learning_rate": 3e-05, + "loss": 3.8157, + "step": 1693 + }, + { + "epoch": 0.4603260869565217, + "grad_norm": 1.4874250888824463, + "learning_rate": 3e-05, + "loss": 3.5901, + "step": 1694 + }, + { + "epoch": 0.46059782608695654, + "grad_norm": 1.9747381210327148, + "learning_rate": 3e-05, + "loss": 3.8642, + "step": 1695 + }, + { + "epoch": 0.4608695652173913, + "grad_norm": 1.405049443244934, + "learning_rate": 3e-05, + "loss": 3.3367, + "step": 1696 + }, + { + "epoch": 0.4611413043478261, + "grad_norm": 1.4599591493606567, + "learning_rate": 3e-05, + "loss": 3.6499, + "step": 1697 + }, + { + "epoch": 0.46141304347826084, + "grad_norm": 1.6565916538238525, + "learning_rate": 3e-05, + "loss": 3.5368, + "step": 1698 + }, + { + "epoch": 0.46168478260869567, + "grad_norm": 1.327285647392273, + "learning_rate": 3e-05, + "loss": 2.7457, + "step": 1699 + }, + { + "epoch": 0.46195652173913043, + "grad_norm": 1.317681074142456, + "learning_rate": 3e-05, + "loss": 3.3931, + "step": 1700 + }, + { + "epoch": 0.4622282608695652, + "grad_norm": 1.7210628986358643, + "learning_rate": 3e-05, + "loss": 4.0996, + "step": 1701 + }, + { + "epoch": 0.4625, + "grad_norm": 1.4575893878936768, + "learning_rate": 3e-05, + "loss": 3.5176, + "step": 1702 + }, + { + "epoch": 0.4627717391304348, + "grad_norm": 0.8817341327667236, + "learning_rate": 3e-05, + "loss": 2.9138, + "step": 1703 + }, + { + "epoch": 0.46304347826086956, + "grad_norm": 1.6268318891525269, + "learning_rate": 3e-05, + "loss": 3.3346, + "step": 1704 + }, + { + "epoch": 0.4633152173913043, + "grad_norm": 1.281497836112976, + "learning_rate": 3e-05, + "loss": 2.776, + "step": 1705 + }, + { + "epoch": 0.46358695652173915, + "grad_norm": 1.6672241687774658, + "learning_rate": 3e-05, + "loss": 3.5205, + "step": 1706 + }, + { + "epoch": 0.4638586956521739, + "grad_norm": 1.4224839210510254, + "learning_rate": 3e-05, + "loss": 3.7986, + "step": 1707 + }, + { + "epoch": 0.4641304347826087, + "grad_norm": 1.1987775564193726, + "learning_rate": 3e-05, + "loss": 3.2364, + "step": 1708 + }, + { + "epoch": 0.4644021739130435, + "grad_norm": 1.4384889602661133, + "learning_rate": 3e-05, + "loss": 3.6339, + "step": 1709 + }, + { + "epoch": 0.46467391304347827, + "grad_norm": 1.1359198093414307, + "learning_rate": 3e-05, + "loss": 3.2827, + "step": 1710 + }, + { + "epoch": 0.46494565217391304, + "grad_norm": 1.3200854063034058, + "learning_rate": 3e-05, + "loss": 3.2209, + "step": 1711 + }, + { + "epoch": 0.4652173913043478, + "grad_norm": 1.613688588142395, + "learning_rate": 3e-05, + "loss": 3.7578, + "step": 1712 + }, + { + "epoch": 0.4654891304347826, + "grad_norm": 1.7866336107254028, + "learning_rate": 3e-05, + "loss": 3.9267, + "step": 1713 + }, + { + "epoch": 0.4657608695652174, + "grad_norm": 2.0430612564086914, + "learning_rate": 3e-05, + "loss": 3.6663, + "step": 1714 + }, + { + "epoch": 0.46603260869565216, + "grad_norm": 1.848181128501892, + "learning_rate": 3e-05, + "loss": 3.3196, + "step": 1715 + }, + { + "epoch": 0.46630434782608693, + "grad_norm": 1.4686863422393799, + "learning_rate": 3e-05, + "loss": 3.352, + "step": 1716 + }, + { + "epoch": 0.46657608695652175, + "grad_norm": 1.1621712446212769, + "learning_rate": 3e-05, + "loss": 3.1954, + "step": 1717 + }, + { + "epoch": 0.4668478260869565, + "grad_norm": 1.2042642831802368, + "learning_rate": 3e-05, + "loss": 2.7059, + "step": 1718 + }, + { + "epoch": 0.4671195652173913, + "grad_norm": 2.4052507877349854, + "learning_rate": 3e-05, + "loss": 3.6323, + "step": 1719 + }, + { + "epoch": 0.4673913043478261, + "grad_norm": 0.8568060994148254, + "learning_rate": 3e-05, + "loss": 2.5241, + "step": 1720 + }, + { + "epoch": 0.4676630434782609, + "grad_norm": 0.9677852988243103, + "learning_rate": 3e-05, + "loss": 2.5949, + "step": 1721 + }, + { + "epoch": 0.46793478260869564, + "grad_norm": 1.2894710302352905, + "learning_rate": 3e-05, + "loss": 2.989, + "step": 1722 + }, + { + "epoch": 0.4682065217391304, + "grad_norm": 1.5006341934204102, + "learning_rate": 3e-05, + "loss": 3.6919, + "step": 1723 + }, + { + "epoch": 0.46847826086956523, + "grad_norm": 1.2537362575531006, + "learning_rate": 3e-05, + "loss": 3.3816, + "step": 1724 + }, + { + "epoch": 0.46875, + "grad_norm": 1.333892822265625, + "learning_rate": 3e-05, + "loss": 3.3992, + "step": 1725 + }, + { + "epoch": 0.46902173913043477, + "grad_norm": 1.1863936185836792, + "learning_rate": 3e-05, + "loss": 3.1168, + "step": 1726 + }, + { + "epoch": 0.4692934782608696, + "grad_norm": 1.2577567100524902, + "learning_rate": 3e-05, + "loss": 3.4444, + "step": 1727 + }, + { + "epoch": 0.46956521739130436, + "grad_norm": 1.2792398929595947, + "learning_rate": 3e-05, + "loss": 2.8464, + "step": 1728 + }, + { + "epoch": 0.4698369565217391, + "grad_norm": 2.50295090675354, + "learning_rate": 3e-05, + "loss": 3.2523, + "step": 1729 + }, + { + "epoch": 0.4701086956521739, + "grad_norm": 1.842974066734314, + "learning_rate": 3e-05, + "loss": 3.3193, + "step": 1730 + }, + { + "epoch": 0.4703804347826087, + "grad_norm": 1.1899735927581787, + "learning_rate": 3e-05, + "loss": 3.4586, + "step": 1731 + }, + { + "epoch": 0.4706521739130435, + "grad_norm": 1.7522422075271606, + "learning_rate": 3e-05, + "loss": 3.44, + "step": 1732 + }, + { + "epoch": 0.47092391304347825, + "grad_norm": 2.0558695793151855, + "learning_rate": 3e-05, + "loss": 3.3303, + "step": 1733 + }, + { + "epoch": 0.47119565217391307, + "grad_norm": 1.5954774618148804, + "learning_rate": 3e-05, + "loss": 3.3366, + "step": 1734 + }, + { + "epoch": 0.47146739130434784, + "grad_norm": 1.4976911544799805, + "learning_rate": 3e-05, + "loss": 3.7435, + "step": 1735 + }, + { + "epoch": 0.4717391304347826, + "grad_norm": 1.4843627214431763, + "learning_rate": 3e-05, + "loss": 3.0997, + "step": 1736 + }, + { + "epoch": 0.4720108695652174, + "grad_norm": 1.7294570207595825, + "learning_rate": 3e-05, + "loss": 3.4683, + "step": 1737 + }, + { + "epoch": 0.4722826086956522, + "grad_norm": 1.7918407917022705, + "learning_rate": 3e-05, + "loss": 3.5634, + "step": 1738 + }, + { + "epoch": 0.47255434782608696, + "grad_norm": 1.458388090133667, + "learning_rate": 3e-05, + "loss": 3.724, + "step": 1739 + }, + { + "epoch": 0.47282608695652173, + "grad_norm": 1.7183854579925537, + "learning_rate": 3e-05, + "loss": 3.7768, + "step": 1740 + }, + { + "epoch": 0.4730978260869565, + "grad_norm": 1.6272084712982178, + "learning_rate": 3e-05, + "loss": 3.1361, + "step": 1741 + }, + { + "epoch": 0.4733695652173913, + "grad_norm": 1.2989143133163452, + "learning_rate": 3e-05, + "loss": 3.1645, + "step": 1742 + }, + { + "epoch": 0.4736413043478261, + "grad_norm": 1.1352357864379883, + "learning_rate": 3e-05, + "loss": 2.9178, + "step": 1743 + }, + { + "epoch": 0.47391304347826085, + "grad_norm": 1.6457839012145996, + "learning_rate": 3e-05, + "loss": 3.8608, + "step": 1744 + }, + { + "epoch": 0.4741847826086957, + "grad_norm": 1.3702744245529175, + "learning_rate": 3e-05, + "loss": 2.8434, + "step": 1745 + }, + { + "epoch": 0.47445652173913044, + "grad_norm": 1.2821577787399292, + "learning_rate": 3e-05, + "loss": 3.0827, + "step": 1746 + }, + { + "epoch": 0.4747282608695652, + "grad_norm": 1.4484645128250122, + "learning_rate": 3e-05, + "loss": 3.222, + "step": 1747 + }, + { + "epoch": 0.475, + "grad_norm": 1.7668582201004028, + "learning_rate": 3e-05, + "loss": 3.2441, + "step": 1748 + }, + { + "epoch": 0.4752717391304348, + "grad_norm": 1.764919400215149, + "learning_rate": 3e-05, + "loss": 3.7584, + "step": 1749 + }, + { + "epoch": 0.47554347826086957, + "grad_norm": 1.640993356704712, + "learning_rate": 3e-05, + "loss": 3.3807, + "step": 1750 + }, + { + "epoch": 0.47581521739130433, + "grad_norm": 1.6005651950836182, + "learning_rate": 3e-05, + "loss": 4.2452, + "step": 1751 + }, + { + "epoch": 0.47608695652173916, + "grad_norm": 1.421935796737671, + "learning_rate": 3e-05, + "loss": 3.0033, + "step": 1752 + }, + { + "epoch": 0.4763586956521739, + "grad_norm": 1.7603181600570679, + "learning_rate": 3e-05, + "loss": 3.2585, + "step": 1753 + }, + { + "epoch": 0.4766304347826087, + "grad_norm": 1.378020167350769, + "learning_rate": 3e-05, + "loss": 3.0198, + "step": 1754 + }, + { + "epoch": 0.47690217391304346, + "grad_norm": 1.5630027055740356, + "learning_rate": 3e-05, + "loss": 3.0409, + "step": 1755 + }, + { + "epoch": 0.4771739130434783, + "grad_norm": 1.1110471487045288, + "learning_rate": 3e-05, + "loss": 3.2521, + "step": 1756 + }, + { + "epoch": 0.47744565217391305, + "grad_norm": 1.403874397277832, + "learning_rate": 3e-05, + "loss": 2.9834, + "step": 1757 + }, + { + "epoch": 0.4777173913043478, + "grad_norm": 1.2942219972610474, + "learning_rate": 3e-05, + "loss": 3.1724, + "step": 1758 + }, + { + "epoch": 0.4779891304347826, + "grad_norm": 1.152891993522644, + "learning_rate": 3e-05, + "loss": 3.2869, + "step": 1759 + }, + { + "epoch": 0.4782608695652174, + "grad_norm": 1.1676504611968994, + "learning_rate": 3e-05, + "loss": 3.1582, + "step": 1760 + }, + { + "epoch": 0.4785326086956522, + "grad_norm": 1.151185154914856, + "learning_rate": 3e-05, + "loss": 3.3322, + "step": 1761 + }, + { + "epoch": 0.47880434782608694, + "grad_norm": 1.4192887544631958, + "learning_rate": 3e-05, + "loss": 3.3659, + "step": 1762 + }, + { + "epoch": 0.47907608695652176, + "grad_norm": 1.5391749143600464, + "learning_rate": 3e-05, + "loss": 3.2556, + "step": 1763 + }, + { + "epoch": 0.47934782608695653, + "grad_norm": 1.4667121171951294, + "learning_rate": 3e-05, + "loss": 3.8051, + "step": 1764 + }, + { + "epoch": 0.4796195652173913, + "grad_norm": 1.3416023254394531, + "learning_rate": 3e-05, + "loss": 3.4739, + "step": 1765 + }, + { + "epoch": 0.47989130434782606, + "grad_norm": 1.7379714250564575, + "learning_rate": 3e-05, + "loss": 3.2991, + "step": 1766 + }, + { + "epoch": 0.4801630434782609, + "grad_norm": 1.6660994291305542, + "learning_rate": 3e-05, + "loss": 3.1885, + "step": 1767 + }, + { + "epoch": 0.48043478260869565, + "grad_norm": 1.9886304140090942, + "learning_rate": 3e-05, + "loss": 3.3295, + "step": 1768 + }, + { + "epoch": 0.4807065217391304, + "grad_norm": 2.345217227935791, + "learning_rate": 3e-05, + "loss": 3.6796, + "step": 1769 + }, + { + "epoch": 0.48097826086956524, + "grad_norm": 1.6064239740371704, + "learning_rate": 3e-05, + "loss": 2.977, + "step": 1770 + }, + { + "epoch": 0.48125, + "grad_norm": 1.7928065061569214, + "learning_rate": 3e-05, + "loss": 3.5848, + "step": 1771 + }, + { + "epoch": 0.4815217391304348, + "grad_norm": 1.5981751680374146, + "learning_rate": 3e-05, + "loss": 3.1049, + "step": 1772 + }, + { + "epoch": 0.48179347826086955, + "grad_norm": 1.6446112394332886, + "learning_rate": 3e-05, + "loss": 2.9694, + "step": 1773 + }, + { + "epoch": 0.48206521739130437, + "grad_norm": 1.8991427421569824, + "learning_rate": 3e-05, + "loss": 3.3139, + "step": 1774 + }, + { + "epoch": 0.48233695652173914, + "grad_norm": 2.3716299533843994, + "learning_rate": 3e-05, + "loss": 3.9361, + "step": 1775 + }, + { + "epoch": 0.4826086956521739, + "grad_norm": 1.8573355674743652, + "learning_rate": 3e-05, + "loss": 3.1394, + "step": 1776 + }, + { + "epoch": 0.48288043478260867, + "grad_norm": 1.8625982999801636, + "learning_rate": 3e-05, + "loss": 3.8017, + "step": 1777 + }, + { + "epoch": 0.4831521739130435, + "grad_norm": 2.2275497913360596, + "learning_rate": 3e-05, + "loss": 3.4565, + "step": 1778 + }, + { + "epoch": 0.48342391304347826, + "grad_norm": 1.9856013059616089, + "learning_rate": 3e-05, + "loss": 3.5517, + "step": 1779 + }, + { + "epoch": 0.483695652173913, + "grad_norm": 1.3510781526565552, + "learning_rate": 3e-05, + "loss": 3.1976, + "step": 1780 + }, + { + "epoch": 0.48396739130434785, + "grad_norm": 1.0482627153396606, + "learning_rate": 3e-05, + "loss": 3.0025, + "step": 1781 + }, + { + "epoch": 0.4842391304347826, + "grad_norm": 1.3749173879623413, + "learning_rate": 3e-05, + "loss": 3.2774, + "step": 1782 + }, + { + "epoch": 0.4845108695652174, + "grad_norm": 1.405247449874878, + "learning_rate": 3e-05, + "loss": 3.4099, + "step": 1783 + }, + { + "epoch": 0.48478260869565215, + "grad_norm": 1.2630794048309326, + "learning_rate": 3e-05, + "loss": 3.7092, + "step": 1784 + }, + { + "epoch": 0.485054347826087, + "grad_norm": 1.1290115118026733, + "learning_rate": 3e-05, + "loss": 3.432, + "step": 1785 + }, + { + "epoch": 0.48532608695652174, + "grad_norm": 1.2523903846740723, + "learning_rate": 3e-05, + "loss": 3.1431, + "step": 1786 + }, + { + "epoch": 0.4855978260869565, + "grad_norm": 1.218867540359497, + "learning_rate": 3e-05, + "loss": 3.4881, + "step": 1787 + }, + { + "epoch": 0.48586956521739133, + "grad_norm": 1.6144737005233765, + "learning_rate": 3e-05, + "loss": 4.0674, + "step": 1788 + }, + { + "epoch": 0.4861413043478261, + "grad_norm": 1.4506441354751587, + "learning_rate": 3e-05, + "loss": 3.5498, + "step": 1789 + }, + { + "epoch": 0.48641304347826086, + "grad_norm": 1.5907642841339111, + "learning_rate": 3e-05, + "loss": 3.1341, + "step": 1790 + }, + { + "epoch": 0.48668478260869563, + "grad_norm": 1.18953275680542, + "learning_rate": 3e-05, + "loss": 3.0503, + "step": 1791 + }, + { + "epoch": 0.48695652173913045, + "grad_norm": 1.2186987400054932, + "learning_rate": 3e-05, + "loss": 2.7508, + "step": 1792 + }, + { + "epoch": 0.4872282608695652, + "grad_norm": 1.111574649810791, + "learning_rate": 3e-05, + "loss": 2.8275, + "step": 1793 + }, + { + "epoch": 0.4875, + "grad_norm": 1.275896430015564, + "learning_rate": 3e-05, + "loss": 3.4576, + "step": 1794 + }, + { + "epoch": 0.48777173913043476, + "grad_norm": 1.182626485824585, + "learning_rate": 3e-05, + "loss": 2.9678, + "step": 1795 + }, + { + "epoch": 0.4880434782608696, + "grad_norm": 1.098120927810669, + "learning_rate": 3e-05, + "loss": 3.1544, + "step": 1796 + }, + { + "epoch": 0.48831521739130435, + "grad_norm": 0.9524686932563782, + "learning_rate": 3e-05, + "loss": 3.0739, + "step": 1797 + }, + { + "epoch": 0.4885869565217391, + "grad_norm": 1.302868366241455, + "learning_rate": 3e-05, + "loss": 3.3036, + "step": 1798 + }, + { + "epoch": 0.48885869565217394, + "grad_norm": 1.124603271484375, + "learning_rate": 3e-05, + "loss": 3.2132, + "step": 1799 + }, + { + "epoch": 0.4891304347826087, + "grad_norm": 1.3126977682113647, + "learning_rate": 3e-05, + "loss": 3.6832, + "step": 1800 + }, + { + "epoch": 0.48940217391304347, + "grad_norm": 1.436482310295105, + "learning_rate": 3e-05, + "loss": 3.9607, + "step": 1801 + }, + { + "epoch": 0.48967391304347824, + "grad_norm": 1.6329675912857056, + "learning_rate": 3e-05, + "loss": 2.6751, + "step": 1802 + }, + { + "epoch": 0.48994565217391306, + "grad_norm": 1.2873584032058716, + "learning_rate": 3e-05, + "loss": 3.3793, + "step": 1803 + }, + { + "epoch": 0.4902173913043478, + "grad_norm": 1.0539244413375854, + "learning_rate": 3e-05, + "loss": 3.0761, + "step": 1804 + }, + { + "epoch": 0.4904891304347826, + "grad_norm": 1.7557803392410278, + "learning_rate": 3e-05, + "loss": 3.0488, + "step": 1805 + }, + { + "epoch": 0.4907608695652174, + "grad_norm": 1.8407940864562988, + "learning_rate": 3e-05, + "loss": 3.1567, + "step": 1806 + }, + { + "epoch": 0.4910326086956522, + "grad_norm": 1.3472580909729004, + "learning_rate": 3e-05, + "loss": 3.2331, + "step": 1807 + }, + { + "epoch": 0.49130434782608695, + "grad_norm": 1.093360424041748, + "learning_rate": 3e-05, + "loss": 3.1883, + "step": 1808 + }, + { + "epoch": 0.4915760869565217, + "grad_norm": 1.650109887123108, + "learning_rate": 3e-05, + "loss": 3.7661, + "step": 1809 + }, + { + "epoch": 0.49184782608695654, + "grad_norm": 1.454372525215149, + "learning_rate": 3e-05, + "loss": 3.1205, + "step": 1810 + }, + { + "epoch": 0.4921195652173913, + "grad_norm": 1.4370765686035156, + "learning_rate": 3e-05, + "loss": 3.8767, + "step": 1811 + }, + { + "epoch": 0.4923913043478261, + "grad_norm": 2.6512532234191895, + "learning_rate": 3e-05, + "loss": 4.0366, + "step": 1812 + }, + { + "epoch": 0.49266304347826084, + "grad_norm": 2.1347718238830566, + "learning_rate": 3e-05, + "loss": 3.5459, + "step": 1813 + }, + { + "epoch": 0.49293478260869567, + "grad_norm": 1.5831135511398315, + "learning_rate": 3e-05, + "loss": 3.4828, + "step": 1814 + }, + { + "epoch": 0.49320652173913043, + "grad_norm": 1.0648609399795532, + "learning_rate": 3e-05, + "loss": 3.0868, + "step": 1815 + }, + { + "epoch": 0.4934782608695652, + "grad_norm": 1.467032551765442, + "learning_rate": 3e-05, + "loss": 3.6537, + "step": 1816 + }, + { + "epoch": 0.49375, + "grad_norm": 1.2471240758895874, + "learning_rate": 3e-05, + "loss": 3.2671, + "step": 1817 + }, + { + "epoch": 0.4940217391304348, + "grad_norm": 0.9745358228683472, + "learning_rate": 3e-05, + "loss": 3.4916, + "step": 1818 + }, + { + "epoch": 0.49429347826086956, + "grad_norm": 1.3291053771972656, + "learning_rate": 3e-05, + "loss": 3.7441, + "step": 1819 + }, + { + "epoch": 0.4945652173913043, + "grad_norm": 1.3420668840408325, + "learning_rate": 3e-05, + "loss": 3.3106, + "step": 1820 + }, + { + "epoch": 0.49483695652173915, + "grad_norm": 1.0097370147705078, + "learning_rate": 3e-05, + "loss": 3.1215, + "step": 1821 + }, + { + "epoch": 0.4951086956521739, + "grad_norm": 1.1169931888580322, + "learning_rate": 3e-05, + "loss": 2.9595, + "step": 1822 + }, + { + "epoch": 0.4953804347826087, + "grad_norm": 1.1370187997817993, + "learning_rate": 3e-05, + "loss": 3.5224, + "step": 1823 + }, + { + "epoch": 0.4956521739130435, + "grad_norm": 1.5470448732376099, + "learning_rate": 3e-05, + "loss": 3.9269, + "step": 1824 + }, + { + "epoch": 0.49592391304347827, + "grad_norm": 1.000895380973816, + "learning_rate": 3e-05, + "loss": 2.9649, + "step": 1825 + }, + { + "epoch": 0.49619565217391304, + "grad_norm": 1.2407017946243286, + "learning_rate": 3e-05, + "loss": 3.5568, + "step": 1826 + }, + { + "epoch": 0.4964673913043478, + "grad_norm": 1.226485013961792, + "learning_rate": 3e-05, + "loss": 2.5796, + "step": 1827 + }, + { + "epoch": 0.4967391304347826, + "grad_norm": 1.1970787048339844, + "learning_rate": 3e-05, + "loss": 2.8772, + "step": 1828 + }, + { + "epoch": 0.4970108695652174, + "grad_norm": 1.2727575302124023, + "learning_rate": 3e-05, + "loss": 3.6501, + "step": 1829 + }, + { + "epoch": 0.49728260869565216, + "grad_norm": 1.2288799285888672, + "learning_rate": 3e-05, + "loss": 3.0078, + "step": 1830 + }, + { + "epoch": 0.49755434782608693, + "grad_norm": 1.688232660293579, + "learning_rate": 3e-05, + "loss": 3.2542, + "step": 1831 + }, + { + "epoch": 0.49782608695652175, + "grad_norm": 1.6907511949539185, + "learning_rate": 3e-05, + "loss": 3.6193, + "step": 1832 + }, + { + "epoch": 0.4980978260869565, + "grad_norm": 1.1391464471817017, + "learning_rate": 3e-05, + "loss": 3.3688, + "step": 1833 + }, + { + "epoch": 0.4983695652173913, + "grad_norm": 1.2691822052001953, + "learning_rate": 3e-05, + "loss": 3.2763, + "step": 1834 + }, + { + "epoch": 0.4986413043478261, + "grad_norm": 1.1176376342773438, + "learning_rate": 3e-05, + "loss": 3.0038, + "step": 1835 + }, + { + "epoch": 0.4989130434782609, + "grad_norm": 0.9173082709312439, + "learning_rate": 3e-05, + "loss": 2.7976, + "step": 1836 + }, + { + "epoch": 0.49918478260869564, + "grad_norm": 1.0989121198654175, + "learning_rate": 3e-05, + "loss": 3.0254, + "step": 1837 + }, + { + "epoch": 0.4994565217391304, + "grad_norm": 1.2885278463363647, + "learning_rate": 3e-05, + "loss": 3.2596, + "step": 1838 + }, + { + "epoch": 0.49972826086956523, + "grad_norm": 1.1294320821762085, + "learning_rate": 3e-05, + "loss": 3.0833, + "step": 1839 + }, + { + "epoch": 0.5, + "grad_norm": 0.9139266014099121, + "learning_rate": 3e-05, + "loss": 3.043, + "step": 1840 + }, + { + "epoch": 0.5002717391304348, + "grad_norm": 0.9999741315841675, + "learning_rate": 3e-05, + "loss": 2.6746, + "step": 1841 + }, + { + "epoch": 0.5005434782608695, + "grad_norm": 2.3439218997955322, + "learning_rate": 3e-05, + "loss": 4.0552, + "step": 1842 + }, + { + "epoch": 0.5008152173913043, + "grad_norm": 1.3185632228851318, + "learning_rate": 3e-05, + "loss": 3.4046, + "step": 1843 + }, + { + "epoch": 0.5010869565217392, + "grad_norm": 1.4254107475280762, + "learning_rate": 3e-05, + "loss": 3.5323, + "step": 1844 + }, + { + "epoch": 0.501358695652174, + "grad_norm": 1.0706696510314941, + "learning_rate": 3e-05, + "loss": 3.0672, + "step": 1845 + }, + { + "epoch": 0.5016304347826087, + "grad_norm": 1.316158413887024, + "learning_rate": 3e-05, + "loss": 3.3709, + "step": 1846 + }, + { + "epoch": 0.5019021739130435, + "grad_norm": 1.2480392456054688, + "learning_rate": 3e-05, + "loss": 3.673, + "step": 1847 + }, + { + "epoch": 0.5021739130434782, + "grad_norm": 1.596869707107544, + "learning_rate": 3e-05, + "loss": 3.3651, + "step": 1848 + }, + { + "epoch": 0.502445652173913, + "grad_norm": 1.6036261320114136, + "learning_rate": 3e-05, + "loss": 3.68, + "step": 1849 + }, + { + "epoch": 0.5027173913043478, + "grad_norm": 1.2420178651809692, + "learning_rate": 3e-05, + "loss": 2.9173, + "step": 1850 + }, + { + "epoch": 0.5029891304347827, + "grad_norm": 1.588339924812317, + "learning_rate": 3e-05, + "loss": 3.7931, + "step": 1851 + }, + { + "epoch": 0.5032608695652174, + "grad_norm": 1.2808369398117065, + "learning_rate": 3e-05, + "loss": 3.05, + "step": 1852 + }, + { + "epoch": 0.5035326086956522, + "grad_norm": 1.234026551246643, + "learning_rate": 3e-05, + "loss": 3.3401, + "step": 1853 + }, + { + "epoch": 0.503804347826087, + "grad_norm": 1.6384221315383911, + "learning_rate": 3e-05, + "loss": 3.5597, + "step": 1854 + }, + { + "epoch": 0.5040760869565217, + "grad_norm": 1.8604109287261963, + "learning_rate": 3e-05, + "loss": 3.6333, + "step": 1855 + }, + { + "epoch": 0.5043478260869565, + "grad_norm": 1.4895154237747192, + "learning_rate": 3e-05, + "loss": 3.4255, + "step": 1856 + }, + { + "epoch": 0.5046195652173913, + "grad_norm": 1.3410334587097168, + "learning_rate": 3e-05, + "loss": 2.7753, + "step": 1857 + }, + { + "epoch": 0.5048913043478261, + "grad_norm": 1.446807861328125, + "learning_rate": 3e-05, + "loss": 4.1229, + "step": 1858 + }, + { + "epoch": 0.5051630434782609, + "grad_norm": 1.188097596168518, + "learning_rate": 3e-05, + "loss": 3.2846, + "step": 1859 + }, + { + "epoch": 0.5054347826086957, + "grad_norm": 1.2120987176895142, + "learning_rate": 3e-05, + "loss": 3.147, + "step": 1860 + }, + { + "epoch": 0.5057065217391304, + "grad_norm": 1.5114277601242065, + "learning_rate": 3e-05, + "loss": 3.0952, + "step": 1861 + }, + { + "epoch": 0.5059782608695652, + "grad_norm": 1.360628366470337, + "learning_rate": 3e-05, + "loss": 2.907, + "step": 1862 + }, + { + "epoch": 0.50625, + "grad_norm": 1.4083906412124634, + "learning_rate": 3e-05, + "loss": 3.4222, + "step": 1863 + }, + { + "epoch": 0.5065217391304347, + "grad_norm": 1.245968222618103, + "learning_rate": 3e-05, + "loss": 3.5063, + "step": 1864 + }, + { + "epoch": 0.5067934782608695, + "grad_norm": 1.3542323112487793, + "learning_rate": 3e-05, + "loss": 3.981, + "step": 1865 + }, + { + "epoch": 0.5070652173913044, + "grad_norm": 2.4021403789520264, + "learning_rate": 3e-05, + "loss": 3.9367, + "step": 1866 + }, + { + "epoch": 0.5073369565217392, + "grad_norm": 1.4897315502166748, + "learning_rate": 3e-05, + "loss": 2.7825, + "step": 1867 + }, + { + "epoch": 0.5076086956521739, + "grad_norm": 1.7353469133377075, + "learning_rate": 3e-05, + "loss": 3.5709, + "step": 1868 + }, + { + "epoch": 0.5078804347826087, + "grad_norm": 1.216436505317688, + "learning_rate": 3e-05, + "loss": 3.1314, + "step": 1869 + }, + { + "epoch": 0.5081521739130435, + "grad_norm": 1.2940754890441895, + "learning_rate": 3e-05, + "loss": 3.3858, + "step": 1870 + }, + { + "epoch": 0.5084239130434782, + "grad_norm": 0.9794735312461853, + "learning_rate": 3e-05, + "loss": 2.823, + "step": 1871 + }, + { + "epoch": 0.508695652173913, + "grad_norm": 1.4511076211929321, + "learning_rate": 3e-05, + "loss": 4.0647, + "step": 1872 + }, + { + "epoch": 0.5089673913043479, + "grad_norm": 1.559597373008728, + "learning_rate": 3e-05, + "loss": 3.093, + "step": 1873 + }, + { + "epoch": 0.5092391304347826, + "grad_norm": 1.7877241373062134, + "learning_rate": 3e-05, + "loss": 3.742, + "step": 1874 + }, + { + "epoch": 0.5095108695652174, + "grad_norm": 1.585822582244873, + "learning_rate": 3e-05, + "loss": 3.631, + "step": 1875 + }, + { + "epoch": 0.5097826086956522, + "grad_norm": 1.284661054611206, + "learning_rate": 3e-05, + "loss": 3.228, + "step": 1876 + }, + { + "epoch": 0.5100543478260869, + "grad_norm": 1.295775294303894, + "learning_rate": 3e-05, + "loss": 3.7922, + "step": 1877 + }, + { + "epoch": 0.5103260869565217, + "grad_norm": 1.2768347263336182, + "learning_rate": 3e-05, + "loss": 3.0369, + "step": 1878 + }, + { + "epoch": 0.5105978260869565, + "grad_norm": 1.4773716926574707, + "learning_rate": 3e-05, + "loss": 3.3407, + "step": 1879 + }, + { + "epoch": 0.5108695652173914, + "grad_norm": 1.2089797258377075, + "learning_rate": 3e-05, + "loss": 3.0556, + "step": 1880 + }, + { + "epoch": 0.5111413043478261, + "grad_norm": 1.4668608903884888, + "learning_rate": 3e-05, + "loss": 3.4665, + "step": 1881 + }, + { + "epoch": 0.5114130434782609, + "grad_norm": 1.998399257659912, + "learning_rate": 3e-05, + "loss": 3.8042, + "step": 1882 + }, + { + "epoch": 0.5116847826086957, + "grad_norm": 1.448439598083496, + "learning_rate": 3e-05, + "loss": 3.7303, + "step": 1883 + }, + { + "epoch": 0.5119565217391304, + "grad_norm": 0.9362940192222595, + "learning_rate": 3e-05, + "loss": 2.665, + "step": 1884 + }, + { + "epoch": 0.5122282608695652, + "grad_norm": 1.6077910661697388, + "learning_rate": 3e-05, + "loss": 3.4826, + "step": 1885 + }, + { + "epoch": 0.5125, + "grad_norm": 1.5588127374649048, + "learning_rate": 3e-05, + "loss": 2.837, + "step": 1886 + }, + { + "epoch": 0.5127717391304348, + "grad_norm": 1.473461389541626, + "learning_rate": 3e-05, + "loss": 3.2511, + "step": 1887 + }, + { + "epoch": 0.5130434782608696, + "grad_norm": 1.6276698112487793, + "learning_rate": 3e-05, + "loss": 3.5821, + "step": 1888 + }, + { + "epoch": 0.5133152173913044, + "grad_norm": 1.6201423406600952, + "learning_rate": 3e-05, + "loss": 3.195, + "step": 1889 + }, + { + "epoch": 0.5135869565217391, + "grad_norm": 1.6411088705062866, + "learning_rate": 3e-05, + "loss": 3.1111, + "step": 1890 + }, + { + "epoch": 0.5138586956521739, + "grad_norm": 1.2973185777664185, + "learning_rate": 3e-05, + "loss": 2.6324, + "step": 1891 + }, + { + "epoch": 0.5141304347826087, + "grad_norm": 2.3508846759796143, + "learning_rate": 3e-05, + "loss": 3.7621, + "step": 1892 + }, + { + "epoch": 0.5144021739130434, + "grad_norm": 1.718246340751648, + "learning_rate": 3e-05, + "loss": 3.759, + "step": 1893 + }, + { + "epoch": 0.5146739130434783, + "grad_norm": 1.5278515815734863, + "learning_rate": 3e-05, + "loss": 3.2879, + "step": 1894 + }, + { + "epoch": 0.5149456521739131, + "grad_norm": 1.308929681777954, + "learning_rate": 3e-05, + "loss": 3.1075, + "step": 1895 + }, + { + "epoch": 0.5152173913043478, + "grad_norm": 1.40326988697052, + "learning_rate": 3e-05, + "loss": 3.5211, + "step": 1896 + }, + { + "epoch": 0.5154891304347826, + "grad_norm": 1.5129557847976685, + "learning_rate": 3e-05, + "loss": 3.2549, + "step": 1897 + }, + { + "epoch": 0.5157608695652174, + "grad_norm": 1.1121063232421875, + "learning_rate": 3e-05, + "loss": 2.9461, + "step": 1898 + }, + { + "epoch": 0.5160326086956522, + "grad_norm": 1.4302382469177246, + "learning_rate": 3e-05, + "loss": 3.627, + "step": 1899 + }, + { + "epoch": 0.5163043478260869, + "grad_norm": 1.6184173822402954, + "learning_rate": 3e-05, + "loss": 2.9157, + "step": 1900 + }, + { + "epoch": 0.5165760869565217, + "grad_norm": 1.1715192794799805, + "learning_rate": 3e-05, + "loss": 2.9425, + "step": 1901 + }, + { + "epoch": 0.5168478260869566, + "grad_norm": 2.065067768096924, + "learning_rate": 3e-05, + "loss": 3.3198, + "step": 1902 + }, + { + "epoch": 0.5171195652173913, + "grad_norm": 1.6942062377929688, + "learning_rate": 3e-05, + "loss": 3.4998, + "step": 1903 + }, + { + "epoch": 0.5173913043478261, + "grad_norm": 1.4067577123641968, + "learning_rate": 3e-05, + "loss": 3.501, + "step": 1904 + }, + { + "epoch": 0.5176630434782609, + "grad_norm": 1.5794352293014526, + "learning_rate": 3e-05, + "loss": 3.1307, + "step": 1905 + }, + { + "epoch": 0.5179347826086956, + "grad_norm": 1.5043370723724365, + "learning_rate": 3e-05, + "loss": 3.8714, + "step": 1906 + }, + { + "epoch": 0.5182065217391304, + "grad_norm": 1.2884905338287354, + "learning_rate": 3e-05, + "loss": 3.4459, + "step": 1907 + }, + { + "epoch": 0.5184782608695652, + "grad_norm": 1.3515671491622925, + "learning_rate": 3e-05, + "loss": 3.5685, + "step": 1908 + }, + { + "epoch": 0.51875, + "grad_norm": 1.2115263938903809, + "learning_rate": 3e-05, + "loss": 3.2234, + "step": 1909 + }, + { + "epoch": 0.5190217391304348, + "grad_norm": 1.3323924541473389, + "learning_rate": 3e-05, + "loss": 3.42, + "step": 1910 + }, + { + "epoch": 0.5192934782608696, + "grad_norm": 1.2837159633636475, + "learning_rate": 3e-05, + "loss": 3.0186, + "step": 1911 + }, + { + "epoch": 0.5195652173913043, + "grad_norm": 1.1012318134307861, + "learning_rate": 3e-05, + "loss": 2.9993, + "step": 1912 + }, + { + "epoch": 0.5198369565217391, + "grad_norm": 1.253250002861023, + "learning_rate": 3e-05, + "loss": 3.371, + "step": 1913 + }, + { + "epoch": 0.5201086956521739, + "grad_norm": 0.9867514967918396, + "learning_rate": 3e-05, + "loss": 2.7761, + "step": 1914 + }, + { + "epoch": 0.5203804347826086, + "grad_norm": 1.0671703815460205, + "learning_rate": 3e-05, + "loss": 3.5288, + "step": 1915 + }, + { + "epoch": 0.5206521739130435, + "grad_norm": 0.9514213800430298, + "learning_rate": 3e-05, + "loss": 2.7831, + "step": 1916 + }, + { + "epoch": 0.5209239130434783, + "grad_norm": 1.3238492012023926, + "learning_rate": 3e-05, + "loss": 3.1874, + "step": 1917 + }, + { + "epoch": 0.5211956521739131, + "grad_norm": 1.5408774614334106, + "learning_rate": 3e-05, + "loss": 3.1997, + "step": 1918 + }, + { + "epoch": 0.5214673913043478, + "grad_norm": 1.1520451307296753, + "learning_rate": 3e-05, + "loss": 3.1287, + "step": 1919 + }, + { + "epoch": 0.5217391304347826, + "grad_norm": 1.5356744527816772, + "learning_rate": 3e-05, + "loss": 3.5281, + "step": 1920 + }, + { + "epoch": 0.5220108695652174, + "grad_norm": 1.0727752447128296, + "learning_rate": 3e-05, + "loss": 3.0281, + "step": 1921 + }, + { + "epoch": 0.5222826086956521, + "grad_norm": 1.4737282991409302, + "learning_rate": 3e-05, + "loss": 4.0763, + "step": 1922 + }, + { + "epoch": 0.522554347826087, + "grad_norm": 1.2212142944335938, + "learning_rate": 3e-05, + "loss": 3.6583, + "step": 1923 + }, + { + "epoch": 0.5228260869565218, + "grad_norm": 1.42307710647583, + "learning_rate": 3e-05, + "loss": 3.9908, + "step": 1924 + }, + { + "epoch": 0.5230978260869565, + "grad_norm": 1.2455291748046875, + "learning_rate": 3e-05, + "loss": 2.7686, + "step": 1925 + }, + { + "epoch": 0.5233695652173913, + "grad_norm": 1.1880292892456055, + "learning_rate": 3e-05, + "loss": 2.8165, + "step": 1926 + }, + { + "epoch": 0.5236413043478261, + "grad_norm": 1.329270362854004, + "learning_rate": 3e-05, + "loss": 3.421, + "step": 1927 + }, + { + "epoch": 0.5239130434782608, + "grad_norm": 1.8960901498794556, + "learning_rate": 3e-05, + "loss": 4.0437, + "step": 1928 + }, + { + "epoch": 0.5241847826086956, + "grad_norm": 1.223769187927246, + "learning_rate": 3e-05, + "loss": 3.3202, + "step": 1929 + }, + { + "epoch": 0.5244565217391305, + "grad_norm": 1.5670809745788574, + "learning_rate": 3e-05, + "loss": 3.6805, + "step": 1930 + }, + { + "epoch": 0.5247282608695653, + "grad_norm": 1.9276920557022095, + "learning_rate": 3e-05, + "loss": 3.4752, + "step": 1931 + }, + { + "epoch": 0.525, + "grad_norm": 1.7917790412902832, + "learning_rate": 3e-05, + "loss": 3.4745, + "step": 1932 + }, + { + "epoch": 0.5252717391304348, + "grad_norm": 1.1239231824874878, + "learning_rate": 3e-05, + "loss": 2.9622, + "step": 1933 + }, + { + "epoch": 0.5255434782608696, + "grad_norm": 1.2177050113677979, + "learning_rate": 3e-05, + "loss": 2.691, + "step": 1934 + }, + { + "epoch": 0.5258152173913043, + "grad_norm": 1.1526474952697754, + "learning_rate": 3e-05, + "loss": 2.7313, + "step": 1935 + }, + { + "epoch": 0.5260869565217391, + "grad_norm": 1.3403104543685913, + "learning_rate": 3e-05, + "loss": 3.215, + "step": 1936 + }, + { + "epoch": 0.5263586956521739, + "grad_norm": 1.1823958158493042, + "learning_rate": 3e-05, + "loss": 3.0382, + "step": 1937 + }, + { + "epoch": 0.5266304347826087, + "grad_norm": 1.2578941583633423, + "learning_rate": 3e-05, + "loss": 2.9962, + "step": 1938 + }, + { + "epoch": 0.5269021739130435, + "grad_norm": 1.2342561483383179, + "learning_rate": 3e-05, + "loss": 3.6362, + "step": 1939 + }, + { + "epoch": 0.5271739130434783, + "grad_norm": 1.0608114004135132, + "learning_rate": 3e-05, + "loss": 3.2318, + "step": 1940 + }, + { + "epoch": 0.527445652173913, + "grad_norm": 1.2502753734588623, + "learning_rate": 3e-05, + "loss": 3.1968, + "step": 1941 + }, + { + "epoch": 0.5277173913043478, + "grad_norm": 0.9138715267181396, + "learning_rate": 3e-05, + "loss": 3.0353, + "step": 1942 + }, + { + "epoch": 0.5279891304347826, + "grad_norm": 1.257987380027771, + "learning_rate": 3e-05, + "loss": 2.5984, + "step": 1943 + }, + { + "epoch": 0.5282608695652173, + "grad_norm": 1.5247408151626587, + "learning_rate": 3e-05, + "loss": 3.4343, + "step": 1944 + }, + { + "epoch": 0.5285326086956522, + "grad_norm": 1.1517291069030762, + "learning_rate": 3e-05, + "loss": 2.7655, + "step": 1945 + }, + { + "epoch": 0.528804347826087, + "grad_norm": 1.5516606569290161, + "learning_rate": 3e-05, + "loss": 3.9976, + "step": 1946 + }, + { + "epoch": 0.5290760869565218, + "grad_norm": 1.6808713674545288, + "learning_rate": 3e-05, + "loss": 3.6858, + "step": 1947 + }, + { + "epoch": 0.5293478260869565, + "grad_norm": 1.0476493835449219, + "learning_rate": 3e-05, + "loss": 2.8933, + "step": 1948 + }, + { + "epoch": 0.5296195652173913, + "grad_norm": 1.7297437191009521, + "learning_rate": 3e-05, + "loss": 3.1754, + "step": 1949 + }, + { + "epoch": 0.529891304347826, + "grad_norm": 1.4908910989761353, + "learning_rate": 3e-05, + "loss": 3.9229, + "step": 1950 + }, + { + "epoch": 0.5301630434782608, + "grad_norm": 1.3128184080123901, + "learning_rate": 3e-05, + "loss": 3.5215, + "step": 1951 + }, + { + "epoch": 0.5304347826086957, + "grad_norm": 1.7025582790374756, + "learning_rate": 3e-05, + "loss": 3.5113, + "step": 1952 + }, + { + "epoch": 0.5307065217391305, + "grad_norm": 1.4770921468734741, + "learning_rate": 3e-05, + "loss": 3.0171, + "step": 1953 + }, + { + "epoch": 0.5309782608695652, + "grad_norm": 1.6477012634277344, + "learning_rate": 3e-05, + "loss": 3.5914, + "step": 1954 + }, + { + "epoch": 0.53125, + "grad_norm": 1.5680129528045654, + "learning_rate": 3e-05, + "loss": 3.6162, + "step": 1955 + }, + { + "epoch": 0.5315217391304348, + "grad_norm": 2.2198116779327393, + "learning_rate": 3e-05, + "loss": 3.2198, + "step": 1956 + }, + { + "epoch": 0.5317934782608695, + "grad_norm": 1.9690414667129517, + "learning_rate": 3e-05, + "loss": 3.0914, + "step": 1957 + }, + { + "epoch": 0.5320652173913043, + "grad_norm": 1.4856075048446655, + "learning_rate": 3e-05, + "loss": 3.206, + "step": 1958 + }, + { + "epoch": 0.5323369565217392, + "grad_norm": 1.1297136545181274, + "learning_rate": 3e-05, + "loss": 2.8479, + "step": 1959 + }, + { + "epoch": 0.532608695652174, + "grad_norm": 1.859328269958496, + "learning_rate": 3e-05, + "loss": 3.3152, + "step": 1960 + }, + { + "epoch": 0.5328804347826087, + "grad_norm": 1.939894199371338, + "learning_rate": 3e-05, + "loss": 3.2557, + "step": 1961 + }, + { + "epoch": 0.5331521739130435, + "grad_norm": 2.148735523223877, + "learning_rate": 3e-05, + "loss": 3.6686, + "step": 1962 + }, + { + "epoch": 0.5334239130434782, + "grad_norm": 1.4653475284576416, + "learning_rate": 3e-05, + "loss": 3.8618, + "step": 1963 + }, + { + "epoch": 0.533695652173913, + "grad_norm": 2.1480512619018555, + "learning_rate": 3e-05, + "loss": 3.4809, + "step": 1964 + }, + { + "epoch": 0.5339673913043478, + "grad_norm": 2.2585768699645996, + "learning_rate": 3e-05, + "loss": 3.1685, + "step": 1965 + }, + { + "epoch": 0.5342391304347827, + "grad_norm": 2.143669605255127, + "learning_rate": 3e-05, + "loss": 3.8206, + "step": 1966 + }, + { + "epoch": 0.5345108695652174, + "grad_norm": 1.5735530853271484, + "learning_rate": 3e-05, + "loss": 3.5483, + "step": 1967 + }, + { + "epoch": 0.5347826086956522, + "grad_norm": 1.5427213907241821, + "learning_rate": 3e-05, + "loss": 3.469, + "step": 1968 + }, + { + "epoch": 0.535054347826087, + "grad_norm": 1.7961771488189697, + "learning_rate": 3e-05, + "loss": 3.2229, + "step": 1969 + }, + { + "epoch": 0.5353260869565217, + "grad_norm": 1.5997658967971802, + "learning_rate": 3e-05, + "loss": 3.6575, + "step": 1970 + }, + { + "epoch": 0.5355978260869565, + "grad_norm": 1.2979601621627808, + "learning_rate": 3e-05, + "loss": 3.7012, + "step": 1971 + }, + { + "epoch": 0.5358695652173913, + "grad_norm": 1.5162733793258667, + "learning_rate": 3e-05, + "loss": 3.1471, + "step": 1972 + }, + { + "epoch": 0.5361413043478261, + "grad_norm": 1.5043525695800781, + "learning_rate": 3e-05, + "loss": 2.8777, + "step": 1973 + }, + { + "epoch": 0.5364130434782609, + "grad_norm": 1.1250759363174438, + "learning_rate": 3e-05, + "loss": 2.9937, + "step": 1974 + }, + { + "epoch": 0.5366847826086957, + "grad_norm": 1.2006231546401978, + "learning_rate": 3e-05, + "loss": 3.5737, + "step": 1975 + }, + { + "epoch": 0.5369565217391304, + "grad_norm": 0.9950990080833435, + "learning_rate": 3e-05, + "loss": 2.9763, + "step": 1976 + }, + { + "epoch": 0.5372282608695652, + "grad_norm": 1.2111644744873047, + "learning_rate": 3e-05, + "loss": 3.5871, + "step": 1977 + }, + { + "epoch": 0.5375, + "grad_norm": 1.4334638118743896, + "learning_rate": 3e-05, + "loss": 3.645, + "step": 1978 + }, + { + "epoch": 0.5377717391304347, + "grad_norm": 1.376451849937439, + "learning_rate": 3e-05, + "loss": 3.5722, + "step": 1979 + }, + { + "epoch": 0.5380434782608695, + "grad_norm": 1.1852439641952515, + "learning_rate": 3e-05, + "loss": 3.2733, + "step": 1980 + }, + { + "epoch": 0.5383152173913044, + "grad_norm": 1.208321452140808, + "learning_rate": 3e-05, + "loss": 3.4223, + "step": 1981 + }, + { + "epoch": 0.5385869565217392, + "grad_norm": 1.2566808462142944, + "learning_rate": 3e-05, + "loss": 3.227, + "step": 1982 + }, + { + "epoch": 0.5388586956521739, + "grad_norm": 1.1662538051605225, + "learning_rate": 3e-05, + "loss": 3.9731, + "step": 1983 + }, + { + "epoch": 0.5391304347826087, + "grad_norm": 0.9991174340248108, + "learning_rate": 3e-05, + "loss": 3.2856, + "step": 1984 + }, + { + "epoch": 0.5394021739130435, + "grad_norm": 1.6792123317718506, + "learning_rate": 3e-05, + "loss": 4.0315, + "step": 1985 + }, + { + "epoch": 0.5396739130434782, + "grad_norm": 1.3554564714431763, + "learning_rate": 3e-05, + "loss": 2.7884, + "step": 1986 + }, + { + "epoch": 0.539945652173913, + "grad_norm": 2.0495409965515137, + "learning_rate": 3e-05, + "loss": 4.0834, + "step": 1987 + }, + { + "epoch": 0.5402173913043479, + "grad_norm": 1.3076777458190918, + "learning_rate": 3e-05, + "loss": 2.9288, + "step": 1988 + }, + { + "epoch": 0.5404891304347826, + "grad_norm": 1.2214155197143555, + "learning_rate": 3e-05, + "loss": 2.9785, + "step": 1989 + }, + { + "epoch": 0.5407608695652174, + "grad_norm": 1.6654658317565918, + "learning_rate": 3e-05, + "loss": 3.6795, + "step": 1990 + }, + { + "epoch": 0.5410326086956522, + "grad_norm": 1.224678874015808, + "learning_rate": 3e-05, + "loss": 2.9736, + "step": 1991 + }, + { + "epoch": 0.5413043478260869, + "grad_norm": 1.128450870513916, + "learning_rate": 3e-05, + "loss": 3.0371, + "step": 1992 + }, + { + "epoch": 0.5415760869565217, + "grad_norm": 1.1452875137329102, + "learning_rate": 3e-05, + "loss": 2.989, + "step": 1993 + }, + { + "epoch": 0.5418478260869565, + "grad_norm": 1.2154463529586792, + "learning_rate": 3e-05, + "loss": 2.9404, + "step": 1994 + }, + { + "epoch": 0.5421195652173914, + "grad_norm": 1.1099878549575806, + "learning_rate": 3e-05, + "loss": 3.1758, + "step": 1995 + }, + { + "epoch": 0.5423913043478261, + "grad_norm": 0.9395751953125, + "learning_rate": 3e-05, + "loss": 2.9722, + "step": 1996 + }, + { + "epoch": 0.5426630434782609, + "grad_norm": 1.1049753427505493, + "learning_rate": 3e-05, + "loss": 3.0526, + "step": 1997 + }, + { + "epoch": 0.5429347826086957, + "grad_norm": 1.45954430103302, + "learning_rate": 3e-05, + "loss": 3.7409, + "step": 1998 + }, + { + "epoch": 0.5432065217391304, + "grad_norm": 1.2864134311676025, + "learning_rate": 3e-05, + "loss": 3.4088, + "step": 1999 + }, + { + "epoch": 0.5434782608695652, + "grad_norm": 1.2165127992630005, + "learning_rate": 3e-05, + "loss": 3.2604, + "step": 2000 + }, + { + "epoch": 0.54375, + "grad_norm": 1.272796869277954, + "learning_rate": 3e-05, + "loss": 3.6574, + "step": 2001 + }, + { + "epoch": 0.5440217391304348, + "grad_norm": 1.3245208263397217, + "learning_rate": 3e-05, + "loss": 3.1406, + "step": 2002 + }, + { + "epoch": 0.5442934782608696, + "grad_norm": 1.4110616445541382, + "learning_rate": 3e-05, + "loss": 2.8426, + "step": 2003 + }, + { + "epoch": 0.5445652173913044, + "grad_norm": 1.7510126829147339, + "learning_rate": 3e-05, + "loss": 3.926, + "step": 2004 + }, + { + "epoch": 0.5448369565217391, + "grad_norm": 1.226434588432312, + "learning_rate": 3e-05, + "loss": 3.5155, + "step": 2005 + }, + { + "epoch": 0.5451086956521739, + "grad_norm": 1.5343637466430664, + "learning_rate": 3e-05, + "loss": 2.7504, + "step": 2006 + }, + { + "epoch": 0.5453804347826087, + "grad_norm": 1.3316307067871094, + "learning_rate": 3e-05, + "loss": 3.0848, + "step": 2007 + }, + { + "epoch": 0.5456521739130434, + "grad_norm": 1.2327113151550293, + "learning_rate": 3e-05, + "loss": 3.0306, + "step": 2008 + }, + { + "epoch": 0.5459239130434783, + "grad_norm": 1.180682897567749, + "learning_rate": 3e-05, + "loss": 3.0591, + "step": 2009 + }, + { + "epoch": 0.5461956521739131, + "grad_norm": 1.3480994701385498, + "learning_rate": 3e-05, + "loss": 3.3818, + "step": 2010 + }, + { + "epoch": 0.5464673913043478, + "grad_norm": 1.246237874031067, + "learning_rate": 3e-05, + "loss": 3.3416, + "step": 2011 + }, + { + "epoch": 0.5467391304347826, + "grad_norm": 0.8937397003173828, + "learning_rate": 3e-05, + "loss": 3.0593, + "step": 2012 + }, + { + "epoch": 0.5470108695652174, + "grad_norm": 1.3019452095031738, + "learning_rate": 3e-05, + "loss": 3.2755, + "step": 2013 + }, + { + "epoch": 0.5472826086956522, + "grad_norm": 1.408234715461731, + "learning_rate": 3e-05, + "loss": 3.4283, + "step": 2014 + }, + { + "epoch": 0.5475543478260869, + "grad_norm": 1.5709846019744873, + "learning_rate": 3e-05, + "loss": 3.2933, + "step": 2015 + }, + { + "epoch": 0.5478260869565217, + "grad_norm": 1.6085715293884277, + "learning_rate": 3e-05, + "loss": 3.2472, + "step": 2016 + }, + { + "epoch": 0.5480978260869566, + "grad_norm": 1.2559678554534912, + "learning_rate": 3e-05, + "loss": 2.8599, + "step": 2017 + }, + { + "epoch": 0.5483695652173913, + "grad_norm": 1.272606372833252, + "learning_rate": 3e-05, + "loss": 3.7284, + "step": 2018 + }, + { + "epoch": 0.5486413043478261, + "grad_norm": 1.098386287689209, + "learning_rate": 3e-05, + "loss": 3.0488, + "step": 2019 + }, + { + "epoch": 0.5489130434782609, + "grad_norm": 1.707297444343567, + "learning_rate": 3e-05, + "loss": 3.122, + "step": 2020 + }, + { + "epoch": 0.5491847826086956, + "grad_norm": 1.7398796081542969, + "learning_rate": 3e-05, + "loss": 3.9982, + "step": 2021 + }, + { + "epoch": 0.5494565217391304, + "grad_norm": 1.1505324840545654, + "learning_rate": 3e-05, + "loss": 3.6328, + "step": 2022 + }, + { + "epoch": 0.5497282608695652, + "grad_norm": 0.978933572769165, + "learning_rate": 3e-05, + "loss": 2.8985, + "step": 2023 + }, + { + "epoch": 0.55, + "grad_norm": 1.2565099000930786, + "learning_rate": 3e-05, + "loss": 3.1667, + "step": 2024 + }, + { + "epoch": 0.5502717391304348, + "grad_norm": 1.2417691946029663, + "learning_rate": 3e-05, + "loss": 3.5036, + "step": 2025 + }, + { + "epoch": 0.5505434782608696, + "grad_norm": 1.2185314893722534, + "learning_rate": 3e-05, + "loss": 3.1685, + "step": 2026 + }, + { + "epoch": 0.5508152173913043, + "grad_norm": 1.4619181156158447, + "learning_rate": 3e-05, + "loss": 4.2336, + "step": 2027 + }, + { + "epoch": 0.5510869565217391, + "grad_norm": 1.2455370426177979, + "learning_rate": 3e-05, + "loss": 3.6007, + "step": 2028 + }, + { + "epoch": 0.5513586956521739, + "grad_norm": 1.4226831197738647, + "learning_rate": 3e-05, + "loss": 3.7129, + "step": 2029 + }, + { + "epoch": 0.5516304347826086, + "grad_norm": 1.2961459159851074, + "learning_rate": 3e-05, + "loss": 3.2041, + "step": 2030 + }, + { + "epoch": 0.5519021739130435, + "grad_norm": 1.259477138519287, + "learning_rate": 3e-05, + "loss": 3.1745, + "step": 2031 + }, + { + "epoch": 0.5521739130434783, + "grad_norm": 1.691286325454712, + "learning_rate": 3e-05, + "loss": 3.5502, + "step": 2032 + }, + { + "epoch": 0.5524456521739131, + "grad_norm": 1.3523718118667603, + "learning_rate": 3e-05, + "loss": 3.4055, + "step": 2033 + }, + { + "epoch": 0.5527173913043478, + "grad_norm": 1.2605832815170288, + "learning_rate": 3e-05, + "loss": 2.7262, + "step": 2034 + }, + { + "epoch": 0.5529891304347826, + "grad_norm": 1.440855860710144, + "learning_rate": 3e-05, + "loss": 2.9385, + "step": 2035 + }, + { + "epoch": 0.5532608695652174, + "grad_norm": 2.026885986328125, + "learning_rate": 3e-05, + "loss": 4.3039, + "step": 2036 + }, + { + "epoch": 0.5535326086956521, + "grad_norm": 1.1887251138687134, + "learning_rate": 3e-05, + "loss": 2.8882, + "step": 2037 + }, + { + "epoch": 0.553804347826087, + "grad_norm": 1.322061538696289, + "learning_rate": 3e-05, + "loss": 3.2703, + "step": 2038 + }, + { + "epoch": 0.5540760869565218, + "grad_norm": 1.3264415264129639, + "learning_rate": 3e-05, + "loss": 3.2199, + "step": 2039 + }, + { + "epoch": 0.5543478260869565, + "grad_norm": 1.6366621255874634, + "learning_rate": 3e-05, + "loss": 3.6402, + "step": 2040 + }, + { + "epoch": 0.5546195652173913, + "grad_norm": 1.592647910118103, + "learning_rate": 3e-05, + "loss": 3.2358, + "step": 2041 + }, + { + "epoch": 0.5548913043478261, + "grad_norm": 1.4537502527236938, + "learning_rate": 3e-05, + "loss": 3.7269, + "step": 2042 + }, + { + "epoch": 0.5551630434782608, + "grad_norm": 1.3483185768127441, + "learning_rate": 3e-05, + "loss": 3.2456, + "step": 2043 + }, + { + "epoch": 0.5554347826086956, + "grad_norm": 1.271835207939148, + "learning_rate": 3e-05, + "loss": 3.5684, + "step": 2044 + }, + { + "epoch": 0.5557065217391305, + "grad_norm": 1.2505134344100952, + "learning_rate": 3e-05, + "loss": 3.2272, + "step": 2045 + }, + { + "epoch": 0.5559782608695653, + "grad_norm": 1.3669381141662598, + "learning_rate": 3e-05, + "loss": 2.8729, + "step": 2046 + }, + { + "epoch": 0.55625, + "grad_norm": 1.1156485080718994, + "learning_rate": 3e-05, + "loss": 3.003, + "step": 2047 + }, + { + "epoch": 0.5565217391304348, + "grad_norm": 1.458754301071167, + "learning_rate": 3e-05, + "loss": 3.0448, + "step": 2048 + }, + { + "epoch": 0.5567934782608696, + "grad_norm": 1.1323773860931396, + "learning_rate": 3e-05, + "loss": 3.1986, + "step": 2049 + }, + { + "epoch": 0.5570652173913043, + "grad_norm": 1.1786192655563354, + "learning_rate": 3e-05, + "loss": 3.6913, + "step": 2050 + }, + { + "epoch": 0.5573369565217391, + "grad_norm": 1.2700916528701782, + "learning_rate": 3e-05, + "loss": 2.9963, + "step": 2051 + }, + { + "epoch": 0.5576086956521739, + "grad_norm": 1.21451735496521, + "learning_rate": 3e-05, + "loss": 3.0128, + "step": 2052 + }, + { + "epoch": 0.5578804347826087, + "grad_norm": 1.5701782703399658, + "learning_rate": 3e-05, + "loss": 2.8858, + "step": 2053 + }, + { + "epoch": 0.5581521739130435, + "grad_norm": 1.7219531536102295, + "learning_rate": 3e-05, + "loss": 2.8876, + "step": 2054 + }, + { + "epoch": 0.5584239130434783, + "grad_norm": 1.694474220275879, + "learning_rate": 3e-05, + "loss": 3.1169, + "step": 2055 + }, + { + "epoch": 0.558695652173913, + "grad_norm": 1.684638500213623, + "learning_rate": 3e-05, + "loss": 3.9718, + "step": 2056 + }, + { + "epoch": 0.5589673913043478, + "grad_norm": 1.8228230476379395, + "learning_rate": 3e-05, + "loss": 3.0664, + "step": 2057 + }, + { + "epoch": 0.5592391304347826, + "grad_norm": 1.8484151363372803, + "learning_rate": 3e-05, + "loss": 3.148, + "step": 2058 + }, + { + "epoch": 0.5595108695652173, + "grad_norm": 1.6520676612854004, + "learning_rate": 3e-05, + "loss": 3.6549, + "step": 2059 + }, + { + "epoch": 0.5597826086956522, + "grad_norm": 1.4072345495224, + "learning_rate": 3e-05, + "loss": 2.7737, + "step": 2060 + }, + { + "epoch": 0.560054347826087, + "grad_norm": 1.459381341934204, + "learning_rate": 3e-05, + "loss": 3.2269, + "step": 2061 + }, + { + "epoch": 0.5603260869565218, + "grad_norm": 1.6240458488464355, + "learning_rate": 3e-05, + "loss": 3.2216, + "step": 2062 + }, + { + "epoch": 0.5605978260869565, + "grad_norm": 1.3474336862564087, + "learning_rate": 3e-05, + "loss": 3.191, + "step": 2063 + }, + { + "epoch": 0.5608695652173913, + "grad_norm": 1.2936657667160034, + "learning_rate": 3e-05, + "loss": 3.1124, + "step": 2064 + }, + { + "epoch": 0.561141304347826, + "grad_norm": 1.4614254236221313, + "learning_rate": 3e-05, + "loss": 3.5676, + "step": 2065 + }, + { + "epoch": 0.5614130434782608, + "grad_norm": 1.4121429920196533, + "learning_rate": 3e-05, + "loss": 2.9918, + "step": 2066 + }, + { + "epoch": 0.5616847826086957, + "grad_norm": 1.8185652494430542, + "learning_rate": 3e-05, + "loss": 3.2873, + "step": 2067 + }, + { + "epoch": 0.5619565217391305, + "grad_norm": 1.7184209823608398, + "learning_rate": 3e-05, + "loss": 3.3629, + "step": 2068 + }, + { + "epoch": 0.5622282608695652, + "grad_norm": 1.2696248292922974, + "learning_rate": 3e-05, + "loss": 2.8962, + "step": 2069 + }, + { + "epoch": 0.5625, + "grad_norm": 1.2504596710205078, + "learning_rate": 3e-05, + "loss": 3.111, + "step": 2070 + }, + { + "epoch": 0.5627717391304348, + "grad_norm": 1.2336483001708984, + "learning_rate": 3e-05, + "loss": 3.1268, + "step": 2071 + }, + { + "epoch": 0.5630434782608695, + "grad_norm": 1.420301914215088, + "learning_rate": 3e-05, + "loss": 3.0844, + "step": 2072 + }, + { + "epoch": 0.5633152173913043, + "grad_norm": 1.4836974143981934, + "learning_rate": 3e-05, + "loss": 3.305, + "step": 2073 + }, + { + "epoch": 0.5635869565217392, + "grad_norm": 1.4846380949020386, + "learning_rate": 3e-05, + "loss": 2.9409, + "step": 2074 + }, + { + "epoch": 0.563858695652174, + "grad_norm": 1.5037497282028198, + "learning_rate": 3e-05, + "loss": 3.691, + "step": 2075 + }, + { + "epoch": 0.5641304347826087, + "grad_norm": 1.1565651893615723, + "learning_rate": 3e-05, + "loss": 3.678, + "step": 2076 + }, + { + "epoch": 0.5644021739130435, + "grad_norm": 1.3678765296936035, + "learning_rate": 3e-05, + "loss": 3.7943, + "step": 2077 + }, + { + "epoch": 0.5646739130434782, + "grad_norm": 1.6229230165481567, + "learning_rate": 3e-05, + "loss": 3.6795, + "step": 2078 + }, + { + "epoch": 0.564945652173913, + "grad_norm": 1.3019323348999023, + "learning_rate": 3e-05, + "loss": 3.2213, + "step": 2079 + }, + { + "epoch": 0.5652173913043478, + "grad_norm": 1.5816229581832886, + "learning_rate": 3e-05, + "loss": 3.9553, + "step": 2080 + }, + { + "epoch": 0.5654891304347827, + "grad_norm": 1.5422357320785522, + "learning_rate": 3e-05, + "loss": 4.4725, + "step": 2081 + }, + { + "epoch": 0.5657608695652174, + "grad_norm": 1.4155722856521606, + "learning_rate": 3e-05, + "loss": 3.6369, + "step": 2082 + }, + { + "epoch": 0.5660326086956522, + "grad_norm": 1.52280592918396, + "learning_rate": 3e-05, + "loss": 3.205, + "step": 2083 + }, + { + "epoch": 0.566304347826087, + "grad_norm": 1.4019720554351807, + "learning_rate": 3e-05, + "loss": 3.5101, + "step": 2084 + }, + { + "epoch": 0.5665760869565217, + "grad_norm": 1.2434489727020264, + "learning_rate": 3e-05, + "loss": 3.5674, + "step": 2085 + }, + { + "epoch": 0.5668478260869565, + "grad_norm": 1.4792324304580688, + "learning_rate": 3e-05, + "loss": 2.8302, + "step": 2086 + }, + { + "epoch": 0.5671195652173913, + "grad_norm": 2.289886236190796, + "learning_rate": 3e-05, + "loss": 3.8648, + "step": 2087 + }, + { + "epoch": 0.5673913043478261, + "grad_norm": 1.3174787759780884, + "learning_rate": 3e-05, + "loss": 3.2811, + "step": 2088 + }, + { + "epoch": 0.5676630434782609, + "grad_norm": 1.2132010459899902, + "learning_rate": 3e-05, + "loss": 3.6173, + "step": 2089 + }, + { + "epoch": 0.5679347826086957, + "grad_norm": 1.528756856918335, + "learning_rate": 3e-05, + "loss": 3.5117, + "step": 2090 + }, + { + "epoch": 0.5682065217391304, + "grad_norm": 1.3807978630065918, + "learning_rate": 3e-05, + "loss": 3.0206, + "step": 2091 + }, + { + "epoch": 0.5684782608695652, + "grad_norm": 1.0497171878814697, + "learning_rate": 3e-05, + "loss": 3.2278, + "step": 2092 + }, + { + "epoch": 0.56875, + "grad_norm": 1.6013227701187134, + "learning_rate": 3e-05, + "loss": 3.4125, + "step": 2093 + }, + { + "epoch": 0.5690217391304347, + "grad_norm": 0.9287932515144348, + "learning_rate": 3e-05, + "loss": 3.1837, + "step": 2094 + }, + { + "epoch": 0.5692934782608695, + "grad_norm": 1.1170742511749268, + "learning_rate": 3e-05, + "loss": 3.5987, + "step": 2095 + }, + { + "epoch": 0.5695652173913044, + "grad_norm": 0.9773249626159668, + "learning_rate": 3e-05, + "loss": 3.4219, + "step": 2096 + }, + { + "epoch": 0.5698369565217392, + "grad_norm": 0.9633996486663818, + "learning_rate": 3e-05, + "loss": 3.3764, + "step": 2097 + }, + { + "epoch": 0.5701086956521739, + "grad_norm": 1.2055789232254028, + "learning_rate": 3e-05, + "loss": 3.3378, + "step": 2098 + }, + { + "epoch": 0.5703804347826087, + "grad_norm": 0.9966376423835754, + "learning_rate": 3e-05, + "loss": 2.8607, + "step": 2099 + }, + { + "epoch": 0.5706521739130435, + "grad_norm": 1.2070221900939941, + "learning_rate": 3e-05, + "loss": 2.7183, + "step": 2100 + }, + { + "epoch": 0.5709239130434782, + "grad_norm": 1.4068926572799683, + "learning_rate": 3e-05, + "loss": 3.8481, + "step": 2101 + }, + { + "epoch": 0.571195652173913, + "grad_norm": 1.5947731733322144, + "learning_rate": 3e-05, + "loss": 2.9881, + "step": 2102 + }, + { + "epoch": 0.5714673913043479, + "grad_norm": 1.2691866159439087, + "learning_rate": 3e-05, + "loss": 3.4337, + "step": 2103 + }, + { + "epoch": 0.5717391304347826, + "grad_norm": 1.850817322731018, + "learning_rate": 3e-05, + "loss": 3.4459, + "step": 2104 + }, + { + "epoch": 0.5720108695652174, + "grad_norm": 1.5024229288101196, + "learning_rate": 3e-05, + "loss": 3.1217, + "step": 2105 + }, + { + "epoch": 0.5722826086956522, + "grad_norm": 1.167541742324829, + "learning_rate": 3e-05, + "loss": 3.584, + "step": 2106 + }, + { + "epoch": 0.5725543478260869, + "grad_norm": 1.89936363697052, + "learning_rate": 3e-05, + "loss": 3.7818, + "step": 2107 + }, + { + "epoch": 0.5728260869565217, + "grad_norm": 1.5929173231124878, + "learning_rate": 3e-05, + "loss": 3.4213, + "step": 2108 + }, + { + "epoch": 0.5730978260869565, + "grad_norm": 1.122673749923706, + "learning_rate": 3e-05, + "loss": 3.6769, + "step": 2109 + }, + { + "epoch": 0.5733695652173914, + "grad_norm": 1.0892047882080078, + "learning_rate": 3e-05, + "loss": 3.2629, + "step": 2110 + }, + { + "epoch": 0.5736413043478261, + "grad_norm": 0.9661053419113159, + "learning_rate": 3e-05, + "loss": 3.0303, + "step": 2111 + }, + { + "epoch": 0.5739130434782609, + "grad_norm": 1.2852338552474976, + "learning_rate": 3e-05, + "loss": 3.6586, + "step": 2112 + }, + { + "epoch": 0.5741847826086957, + "grad_norm": 1.211471676826477, + "learning_rate": 3e-05, + "loss": 3.1448, + "step": 2113 + }, + { + "epoch": 0.5744565217391304, + "grad_norm": 1.1296757459640503, + "learning_rate": 3e-05, + "loss": 2.9875, + "step": 2114 + }, + { + "epoch": 0.5747282608695652, + "grad_norm": 1.3020541667938232, + "learning_rate": 3e-05, + "loss": 3.111, + "step": 2115 + }, + { + "epoch": 0.575, + "grad_norm": 1.0208520889282227, + "learning_rate": 3e-05, + "loss": 2.7882, + "step": 2116 + }, + { + "epoch": 0.5752717391304348, + "grad_norm": 1.4555299282073975, + "learning_rate": 3e-05, + "loss": 3.3721, + "step": 2117 + }, + { + "epoch": 0.5755434782608696, + "grad_norm": 2.195035934448242, + "learning_rate": 3e-05, + "loss": 4.1556, + "step": 2118 + }, + { + "epoch": 0.5758152173913044, + "grad_norm": 1.4949889183044434, + "learning_rate": 3e-05, + "loss": 2.6564, + "step": 2119 + }, + { + "epoch": 0.5760869565217391, + "grad_norm": 1.1324836015701294, + "learning_rate": 3e-05, + "loss": 3.1073, + "step": 2120 + }, + { + "epoch": 0.5763586956521739, + "grad_norm": 1.3076951503753662, + "learning_rate": 3e-05, + "loss": 3.6178, + "step": 2121 + }, + { + "epoch": 0.5766304347826087, + "grad_norm": 1.322385311126709, + "learning_rate": 3e-05, + "loss": 3.9284, + "step": 2122 + }, + { + "epoch": 0.5769021739130434, + "grad_norm": 1.567959189414978, + "learning_rate": 3e-05, + "loss": 4.0229, + "step": 2123 + }, + { + "epoch": 0.5771739130434783, + "grad_norm": 1.4464517831802368, + "learning_rate": 3e-05, + "loss": 3.3669, + "step": 2124 + }, + { + "epoch": 0.5774456521739131, + "grad_norm": 1.5971894264221191, + "learning_rate": 3e-05, + "loss": 4.1792, + "step": 2125 + }, + { + "epoch": 0.5777173913043478, + "grad_norm": 1.4091331958770752, + "learning_rate": 3e-05, + "loss": 3.3196, + "step": 2126 + }, + { + "epoch": 0.5779891304347826, + "grad_norm": 1.4950571060180664, + "learning_rate": 3e-05, + "loss": 3.6238, + "step": 2127 + }, + { + "epoch": 0.5782608695652174, + "grad_norm": 1.0344115495681763, + "learning_rate": 3e-05, + "loss": 2.8622, + "step": 2128 + }, + { + "epoch": 0.5785326086956522, + "grad_norm": 1.1362147331237793, + "learning_rate": 3e-05, + "loss": 2.7986, + "step": 2129 + }, + { + "epoch": 0.5788043478260869, + "grad_norm": 1.1396723985671997, + "learning_rate": 3e-05, + "loss": 2.7951, + "step": 2130 + }, + { + "epoch": 0.5790760869565217, + "grad_norm": 1.355010747909546, + "learning_rate": 3e-05, + "loss": 3.1708, + "step": 2131 + }, + { + "epoch": 0.5793478260869566, + "grad_norm": 1.5592600107192993, + "learning_rate": 3e-05, + "loss": 3.5758, + "step": 2132 + }, + { + "epoch": 0.5796195652173913, + "grad_norm": 1.56844961643219, + "learning_rate": 3e-05, + "loss": 3.4122, + "step": 2133 + }, + { + "epoch": 0.5798913043478261, + "grad_norm": 1.3906667232513428, + "learning_rate": 3e-05, + "loss": 3.7639, + "step": 2134 + }, + { + "epoch": 0.5801630434782609, + "grad_norm": 1.131856918334961, + "learning_rate": 3e-05, + "loss": 3.3024, + "step": 2135 + }, + { + "epoch": 0.5804347826086956, + "grad_norm": 1.2349239587783813, + "learning_rate": 3e-05, + "loss": 3.2338, + "step": 2136 + }, + { + "epoch": 0.5807065217391304, + "grad_norm": 1.320646047592163, + "learning_rate": 3e-05, + "loss": 3.2089, + "step": 2137 + }, + { + "epoch": 0.5809782608695652, + "grad_norm": 1.0962297916412354, + "learning_rate": 3e-05, + "loss": 2.9312, + "step": 2138 + }, + { + "epoch": 0.58125, + "grad_norm": 1.2109733819961548, + "learning_rate": 3e-05, + "loss": 2.6762, + "step": 2139 + }, + { + "epoch": 0.5815217391304348, + "grad_norm": 1.3556448221206665, + "learning_rate": 3e-05, + "loss": 3.6022, + "step": 2140 + }, + { + "epoch": 0.5817934782608696, + "grad_norm": 1.216602087020874, + "learning_rate": 3e-05, + "loss": 3.9107, + "step": 2141 + }, + { + "epoch": 0.5820652173913043, + "grad_norm": 1.2305996417999268, + "learning_rate": 3e-05, + "loss": 3.2447, + "step": 2142 + }, + { + "epoch": 0.5823369565217391, + "grad_norm": 1.2739956378936768, + "learning_rate": 3e-05, + "loss": 3.4232, + "step": 2143 + }, + { + "epoch": 0.5826086956521739, + "grad_norm": 1.2404019832611084, + "learning_rate": 3e-05, + "loss": 3.467, + "step": 2144 + }, + { + "epoch": 0.5828804347826086, + "grad_norm": 1.3793765306472778, + "learning_rate": 3e-05, + "loss": 3.4627, + "step": 2145 + }, + { + "epoch": 0.5831521739130435, + "grad_norm": 1.6223127841949463, + "learning_rate": 3e-05, + "loss": 3.2725, + "step": 2146 + }, + { + "epoch": 0.5834239130434783, + "grad_norm": 1.055362343788147, + "learning_rate": 3e-05, + "loss": 3.2517, + "step": 2147 + }, + { + "epoch": 0.5836956521739131, + "grad_norm": 1.2135682106018066, + "learning_rate": 3e-05, + "loss": 3.0423, + "step": 2148 + }, + { + "epoch": 0.5839673913043478, + "grad_norm": 1.7276960611343384, + "learning_rate": 3e-05, + "loss": 4.2823, + "step": 2149 + }, + { + "epoch": 0.5842391304347826, + "grad_norm": 1.1459511518478394, + "learning_rate": 3e-05, + "loss": 3.129, + "step": 2150 + }, + { + "epoch": 0.5845108695652174, + "grad_norm": 1.7933846712112427, + "learning_rate": 3e-05, + "loss": 3.6957, + "step": 2151 + }, + { + "epoch": 0.5847826086956521, + "grad_norm": 1.3888658285140991, + "learning_rate": 3e-05, + "loss": 4.2824, + "step": 2152 + }, + { + "epoch": 0.585054347826087, + "grad_norm": 1.18975031375885, + "learning_rate": 3e-05, + "loss": 2.9272, + "step": 2153 + }, + { + "epoch": 0.5853260869565218, + "grad_norm": 1.0910744667053223, + "learning_rate": 3e-05, + "loss": 3.4617, + "step": 2154 + }, + { + "epoch": 0.5855978260869565, + "grad_norm": 1.3888293504714966, + "learning_rate": 3e-05, + "loss": 2.9697, + "step": 2155 + }, + { + "epoch": 0.5858695652173913, + "grad_norm": 1.1816877126693726, + "learning_rate": 3e-05, + "loss": 3.5301, + "step": 2156 + }, + { + "epoch": 0.5861413043478261, + "grad_norm": 1.2659231424331665, + "learning_rate": 3e-05, + "loss": 3.5646, + "step": 2157 + }, + { + "epoch": 0.5864130434782608, + "grad_norm": 1.5958353281021118, + "learning_rate": 3e-05, + "loss": 3.378, + "step": 2158 + }, + { + "epoch": 0.5866847826086956, + "grad_norm": 1.2585597038269043, + "learning_rate": 3e-05, + "loss": 3.0523, + "step": 2159 + }, + { + "epoch": 0.5869565217391305, + "grad_norm": 1.0975054502487183, + "learning_rate": 3e-05, + "loss": 3.0933, + "step": 2160 + }, + { + "epoch": 0.5872282608695653, + "grad_norm": 1.5233417749404907, + "learning_rate": 3e-05, + "loss": 3.2665, + "step": 2161 + }, + { + "epoch": 0.5875, + "grad_norm": 2.02248215675354, + "learning_rate": 3e-05, + "loss": 3.754, + "step": 2162 + }, + { + "epoch": 0.5877717391304348, + "grad_norm": 0.9959807395935059, + "learning_rate": 3e-05, + "loss": 2.9276, + "step": 2163 + }, + { + "epoch": 0.5880434782608696, + "grad_norm": 1.1323211193084717, + "learning_rate": 3e-05, + "loss": 3.0789, + "step": 2164 + }, + { + "epoch": 0.5883152173913043, + "grad_norm": 1.0242100954055786, + "learning_rate": 3e-05, + "loss": 3.3039, + "step": 2165 + }, + { + "epoch": 0.5885869565217391, + "grad_norm": 1.0082273483276367, + "learning_rate": 3e-05, + "loss": 3.3631, + "step": 2166 + }, + { + "epoch": 0.5888586956521739, + "grad_norm": 1.1713404655456543, + "learning_rate": 3e-05, + "loss": 2.7448, + "step": 2167 + }, + { + "epoch": 0.5891304347826087, + "grad_norm": 1.1318339109420776, + "learning_rate": 3e-05, + "loss": 3.3552, + "step": 2168 + }, + { + "epoch": 0.5894021739130435, + "grad_norm": 1.3056354522705078, + "learning_rate": 3e-05, + "loss": 3.8349, + "step": 2169 + }, + { + "epoch": 0.5896739130434783, + "grad_norm": 1.1058902740478516, + "learning_rate": 3e-05, + "loss": 2.9931, + "step": 2170 + }, + { + "epoch": 0.589945652173913, + "grad_norm": 1.3379881381988525, + "learning_rate": 3e-05, + "loss": 3.0709, + "step": 2171 + }, + { + "epoch": 0.5902173913043478, + "grad_norm": 1.3458565473556519, + "learning_rate": 3e-05, + "loss": 3.5521, + "step": 2172 + }, + { + "epoch": 0.5904891304347826, + "grad_norm": 1.2749676704406738, + "learning_rate": 3e-05, + "loss": 3.3535, + "step": 2173 + }, + { + "epoch": 0.5907608695652173, + "grad_norm": 1.3268214464187622, + "learning_rate": 3e-05, + "loss": 3.6918, + "step": 2174 + }, + { + "epoch": 0.5910326086956522, + "grad_norm": 1.4857879877090454, + "learning_rate": 3e-05, + "loss": 3.6046, + "step": 2175 + }, + { + "epoch": 0.591304347826087, + "grad_norm": 1.009945034980774, + "learning_rate": 3e-05, + "loss": 3.435, + "step": 2176 + }, + { + "epoch": 0.5915760869565218, + "grad_norm": 1.1983013153076172, + "learning_rate": 3e-05, + "loss": 3.5871, + "step": 2177 + }, + { + "epoch": 0.5918478260869565, + "grad_norm": 0.9718300700187683, + "learning_rate": 3e-05, + "loss": 3.3072, + "step": 2178 + }, + { + "epoch": 0.5921195652173913, + "grad_norm": 1.101165771484375, + "learning_rate": 3e-05, + "loss": 3.2276, + "step": 2179 + }, + { + "epoch": 0.592391304347826, + "grad_norm": 1.012749433517456, + "learning_rate": 3e-05, + "loss": 3.1112, + "step": 2180 + }, + { + "epoch": 0.5926630434782608, + "grad_norm": 0.9688409566879272, + "learning_rate": 3e-05, + "loss": 3.0918, + "step": 2181 + }, + { + "epoch": 0.5929347826086957, + "grad_norm": 0.9854469895362854, + "learning_rate": 3e-05, + "loss": 3.1777, + "step": 2182 + }, + { + "epoch": 0.5932065217391305, + "grad_norm": 0.9278433918952942, + "learning_rate": 3e-05, + "loss": 3.1989, + "step": 2183 + }, + { + "epoch": 0.5934782608695652, + "grad_norm": 0.983065664768219, + "learning_rate": 3e-05, + "loss": 3.3761, + "step": 2184 + }, + { + "epoch": 0.59375, + "grad_norm": 1.473644733428955, + "learning_rate": 3e-05, + "loss": 3.6095, + "step": 2185 + }, + { + "epoch": 0.5940217391304348, + "grad_norm": 1.2839797735214233, + "learning_rate": 3e-05, + "loss": 2.9857, + "step": 2186 + }, + { + "epoch": 0.5942934782608695, + "grad_norm": 1.7509822845458984, + "learning_rate": 3e-05, + "loss": 3.7711, + "step": 2187 + }, + { + "epoch": 0.5945652173913043, + "grad_norm": 1.171937346458435, + "learning_rate": 3e-05, + "loss": 3.6335, + "step": 2188 + }, + { + "epoch": 0.5948369565217392, + "grad_norm": 1.4512766599655151, + "learning_rate": 3e-05, + "loss": 2.9629, + "step": 2189 + }, + { + "epoch": 0.595108695652174, + "grad_norm": 1.0719918012619019, + "learning_rate": 3e-05, + "loss": 2.8943, + "step": 2190 + }, + { + "epoch": 0.5953804347826087, + "grad_norm": 1.11574387550354, + "learning_rate": 3e-05, + "loss": 3.7779, + "step": 2191 + }, + { + "epoch": 0.5956521739130435, + "grad_norm": 1.6633427143096924, + "learning_rate": 3e-05, + "loss": 3.7736, + "step": 2192 + }, + { + "epoch": 0.5959239130434782, + "grad_norm": 1.1903231143951416, + "learning_rate": 3e-05, + "loss": 2.9944, + "step": 2193 + }, + { + "epoch": 0.596195652173913, + "grad_norm": 1.2331459522247314, + "learning_rate": 3e-05, + "loss": 3.639, + "step": 2194 + }, + { + "epoch": 0.5964673913043478, + "grad_norm": 1.8560791015625, + "learning_rate": 3e-05, + "loss": 4.0684, + "step": 2195 + }, + { + "epoch": 0.5967391304347827, + "grad_norm": 1.2733112573623657, + "learning_rate": 3e-05, + "loss": 3.1905, + "step": 2196 + }, + { + "epoch": 0.5970108695652174, + "grad_norm": 1.156823992729187, + "learning_rate": 3e-05, + "loss": 3.3605, + "step": 2197 + }, + { + "epoch": 0.5972826086956522, + "grad_norm": 1.0290706157684326, + "learning_rate": 3e-05, + "loss": 3.2214, + "step": 2198 + }, + { + "epoch": 0.597554347826087, + "grad_norm": 1.1795909404754639, + "learning_rate": 3e-05, + "loss": 2.9137, + "step": 2199 + }, + { + "epoch": 0.5978260869565217, + "grad_norm": 1.3558768033981323, + "learning_rate": 3e-05, + "loss": 3.6129, + "step": 2200 + }, + { + "epoch": 0.5980978260869565, + "grad_norm": 0.9279618859291077, + "learning_rate": 3e-05, + "loss": 2.5655, + "step": 2201 + }, + { + "epoch": 0.5983695652173913, + "grad_norm": 1.2650532722473145, + "learning_rate": 3e-05, + "loss": 2.8838, + "step": 2202 + }, + { + "epoch": 0.5986413043478261, + "grad_norm": 1.2165402173995972, + "learning_rate": 3e-05, + "loss": 3.5675, + "step": 2203 + }, + { + "epoch": 0.5989130434782609, + "grad_norm": 1.1845487356185913, + "learning_rate": 3e-05, + "loss": 2.9909, + "step": 2204 + }, + { + "epoch": 0.5991847826086957, + "grad_norm": 1.1865789890289307, + "learning_rate": 3e-05, + "loss": 3.6638, + "step": 2205 + }, + { + "epoch": 0.5994565217391304, + "grad_norm": 0.9880366921424866, + "learning_rate": 3e-05, + "loss": 3.023, + "step": 2206 + }, + { + "epoch": 0.5997282608695652, + "grad_norm": 1.1242929697036743, + "learning_rate": 3e-05, + "loss": 3.1062, + "step": 2207 + }, + { + "epoch": 0.6, + "grad_norm": 1.1712896823883057, + "learning_rate": 3e-05, + "loss": 3.029, + "step": 2208 + }, + { + "epoch": 0.6002717391304347, + "grad_norm": 1.5733115673065186, + "learning_rate": 3e-05, + "loss": 3.7851, + "step": 2209 + }, + { + "epoch": 0.6005434782608695, + "grad_norm": 1.2813576459884644, + "learning_rate": 3e-05, + "loss": 3.4965, + "step": 2210 + }, + { + "epoch": 0.6008152173913044, + "grad_norm": 1.3263084888458252, + "learning_rate": 3e-05, + "loss": 4.1356, + "step": 2211 + }, + { + "epoch": 0.6010869565217392, + "grad_norm": 1.1176859140396118, + "learning_rate": 3e-05, + "loss": 3.4671, + "step": 2212 + }, + { + "epoch": 0.6013586956521739, + "grad_norm": 1.2956467866897583, + "learning_rate": 3e-05, + "loss": 3.8482, + "step": 2213 + }, + { + "epoch": 0.6016304347826087, + "grad_norm": 1.8947595357894897, + "learning_rate": 3e-05, + "loss": 3.2963, + "step": 2214 + }, + { + "epoch": 0.6019021739130435, + "grad_norm": 1.3561947345733643, + "learning_rate": 3e-05, + "loss": 3.1028, + "step": 2215 + }, + { + "epoch": 0.6021739130434782, + "grad_norm": 1.1945207118988037, + "learning_rate": 3e-05, + "loss": 3.2949, + "step": 2216 + }, + { + "epoch": 0.602445652173913, + "grad_norm": 1.3198587894439697, + "learning_rate": 3e-05, + "loss": 3.5389, + "step": 2217 + }, + { + "epoch": 0.6027173913043479, + "grad_norm": 1.6522905826568604, + "learning_rate": 3e-05, + "loss": 3.2054, + "step": 2218 + }, + { + "epoch": 0.6029891304347826, + "grad_norm": 1.2735859155654907, + "learning_rate": 3e-05, + "loss": 3.4955, + "step": 2219 + }, + { + "epoch": 0.6032608695652174, + "grad_norm": 1.2329673767089844, + "learning_rate": 3e-05, + "loss": 3.7713, + "step": 2220 + }, + { + "epoch": 0.6035326086956522, + "grad_norm": 1.1366288661956787, + "learning_rate": 3e-05, + "loss": 3.333, + "step": 2221 + }, + { + "epoch": 0.6038043478260869, + "grad_norm": 1.0295037031173706, + "learning_rate": 3e-05, + "loss": 3.2991, + "step": 2222 + }, + { + "epoch": 0.6040760869565217, + "grad_norm": 1.0760974884033203, + "learning_rate": 3e-05, + "loss": 3.6266, + "step": 2223 + }, + { + "epoch": 0.6043478260869565, + "grad_norm": 1.0044188499450684, + "learning_rate": 3e-05, + "loss": 3.1854, + "step": 2224 + }, + { + "epoch": 0.6046195652173914, + "grad_norm": 0.79954993724823, + "learning_rate": 3e-05, + "loss": 2.5833, + "step": 2225 + }, + { + "epoch": 0.6048913043478261, + "grad_norm": 1.3376729488372803, + "learning_rate": 3e-05, + "loss": 3.8299, + "step": 2226 + }, + { + "epoch": 0.6051630434782609, + "grad_norm": 1.3323661088943481, + "learning_rate": 3e-05, + "loss": 3.5493, + "step": 2227 + }, + { + "epoch": 0.6054347826086957, + "grad_norm": 0.9946414828300476, + "learning_rate": 3e-05, + "loss": 3.0806, + "step": 2228 + }, + { + "epoch": 0.6057065217391304, + "grad_norm": 1.486099123954773, + "learning_rate": 3e-05, + "loss": 3.5046, + "step": 2229 + }, + { + "epoch": 0.6059782608695652, + "grad_norm": 1.4763749837875366, + "learning_rate": 3e-05, + "loss": 3.2249, + "step": 2230 + }, + { + "epoch": 0.60625, + "grad_norm": 1.1273298263549805, + "learning_rate": 3e-05, + "loss": 3.097, + "step": 2231 + }, + { + "epoch": 0.6065217391304348, + "grad_norm": 1.273410677909851, + "learning_rate": 3e-05, + "loss": 3.3963, + "step": 2232 + }, + { + "epoch": 0.6067934782608696, + "grad_norm": 1.6567277908325195, + "learning_rate": 3e-05, + "loss": 3.4074, + "step": 2233 + }, + { + "epoch": 0.6070652173913044, + "grad_norm": 1.272204875946045, + "learning_rate": 3e-05, + "loss": 3.5062, + "step": 2234 + }, + { + "epoch": 0.6073369565217391, + "grad_norm": 0.9623717665672302, + "learning_rate": 3e-05, + "loss": 3.348, + "step": 2235 + }, + { + "epoch": 0.6076086956521739, + "grad_norm": 1.7306931018829346, + "learning_rate": 3e-05, + "loss": 3.5976, + "step": 2236 + }, + { + "epoch": 0.6078804347826087, + "grad_norm": 1.2716137170791626, + "learning_rate": 3e-05, + "loss": 3.8244, + "step": 2237 + }, + { + "epoch": 0.6081521739130434, + "grad_norm": 1.3230034112930298, + "learning_rate": 3e-05, + "loss": 3.4558, + "step": 2238 + }, + { + "epoch": 0.6084239130434783, + "grad_norm": 1.733973503112793, + "learning_rate": 3e-05, + "loss": 3.0608, + "step": 2239 + }, + { + "epoch": 0.6086956521739131, + "grad_norm": 1.2597112655639648, + "learning_rate": 3e-05, + "loss": 2.8695, + "step": 2240 + }, + { + "epoch": 0.6089673913043478, + "grad_norm": 1.033078670501709, + "learning_rate": 3e-05, + "loss": 2.528, + "step": 2241 + }, + { + "epoch": 0.6092391304347826, + "grad_norm": 1.4419554471969604, + "learning_rate": 3e-05, + "loss": 3.5569, + "step": 2242 + }, + { + "epoch": 0.6095108695652174, + "grad_norm": 1.376265287399292, + "learning_rate": 3e-05, + "loss": 3.4915, + "step": 2243 + }, + { + "epoch": 0.6097826086956522, + "grad_norm": 1.1557855606079102, + "learning_rate": 3e-05, + "loss": 3.4314, + "step": 2244 + }, + { + "epoch": 0.6100543478260869, + "grad_norm": 1.3031344413757324, + "learning_rate": 3e-05, + "loss": 3.5682, + "step": 2245 + }, + { + "epoch": 0.6103260869565217, + "grad_norm": 1.2761789560317993, + "learning_rate": 3e-05, + "loss": 3.4629, + "step": 2246 + }, + { + "epoch": 0.6105978260869566, + "grad_norm": 1.1963956356048584, + "learning_rate": 3e-05, + "loss": 3.2122, + "step": 2247 + }, + { + "epoch": 0.6108695652173913, + "grad_norm": 1.357904314994812, + "learning_rate": 3e-05, + "loss": 3.0868, + "step": 2248 + }, + { + "epoch": 0.6111413043478261, + "grad_norm": 1.3014525175094604, + "learning_rate": 3e-05, + "loss": 2.9693, + "step": 2249 + }, + { + "epoch": 0.6114130434782609, + "grad_norm": 1.3314582109451294, + "learning_rate": 3e-05, + "loss": 3.3714, + "step": 2250 + }, + { + "epoch": 0.6116847826086956, + "grad_norm": 1.3849120140075684, + "learning_rate": 3e-05, + "loss": 2.7755, + "step": 2251 + }, + { + "epoch": 0.6119565217391304, + "grad_norm": 0.9548225402832031, + "learning_rate": 3e-05, + "loss": 3.0885, + "step": 2252 + }, + { + "epoch": 0.6122282608695652, + "grad_norm": 1.4415459632873535, + "learning_rate": 3e-05, + "loss": 3.8783, + "step": 2253 + }, + { + "epoch": 0.6125, + "grad_norm": 1.7835477590560913, + "learning_rate": 3e-05, + "loss": 3.6217, + "step": 2254 + }, + { + "epoch": 0.6127717391304348, + "grad_norm": 1.4011732339859009, + "learning_rate": 3e-05, + "loss": 3.9328, + "step": 2255 + }, + { + "epoch": 0.6130434782608696, + "grad_norm": 1.0685231685638428, + "learning_rate": 3e-05, + "loss": 3.2949, + "step": 2256 + }, + { + "epoch": 0.6133152173913043, + "grad_norm": 1.2185416221618652, + "learning_rate": 3e-05, + "loss": 3.5132, + "step": 2257 + }, + { + "epoch": 0.6135869565217391, + "grad_norm": 1.0910768508911133, + "learning_rate": 3e-05, + "loss": 3.0972, + "step": 2258 + }, + { + "epoch": 0.6138586956521739, + "grad_norm": 1.0861088037490845, + "learning_rate": 3e-05, + "loss": 3.3816, + "step": 2259 + }, + { + "epoch": 0.6141304347826086, + "grad_norm": 1.0227428674697876, + "learning_rate": 3e-05, + "loss": 3.1733, + "step": 2260 + }, + { + "epoch": 0.6144021739130435, + "grad_norm": 1.3343944549560547, + "learning_rate": 3e-05, + "loss": 3.2618, + "step": 2261 + }, + { + "epoch": 0.6146739130434783, + "grad_norm": 1.3029199838638306, + "learning_rate": 3e-05, + "loss": 3.4307, + "step": 2262 + }, + { + "epoch": 0.6149456521739131, + "grad_norm": 1.152552843093872, + "learning_rate": 3e-05, + "loss": 3.4878, + "step": 2263 + }, + { + "epoch": 0.6152173913043478, + "grad_norm": 1.5189865827560425, + "learning_rate": 3e-05, + "loss": 3.4252, + "step": 2264 + }, + { + "epoch": 0.6154891304347826, + "grad_norm": 1.1164250373840332, + "learning_rate": 3e-05, + "loss": 3.5027, + "step": 2265 + }, + { + "epoch": 0.6157608695652174, + "grad_norm": 0.9906913638114929, + "learning_rate": 3e-05, + "loss": 2.9876, + "step": 2266 + }, + { + "epoch": 0.6160326086956521, + "grad_norm": 1.1681506633758545, + "learning_rate": 3e-05, + "loss": 3.2722, + "step": 2267 + }, + { + "epoch": 0.616304347826087, + "grad_norm": 1.183396816253662, + "learning_rate": 3e-05, + "loss": 3.2224, + "step": 2268 + }, + { + "epoch": 0.6165760869565218, + "grad_norm": 1.0938737392425537, + "learning_rate": 3e-05, + "loss": 3.2871, + "step": 2269 + }, + { + "epoch": 0.6168478260869565, + "grad_norm": 1.5161495208740234, + "learning_rate": 3e-05, + "loss": 3.7379, + "step": 2270 + }, + { + "epoch": 0.6171195652173913, + "grad_norm": 1.2433841228485107, + "learning_rate": 3e-05, + "loss": 2.9899, + "step": 2271 + }, + { + "epoch": 0.6173913043478261, + "grad_norm": 1.092914342880249, + "learning_rate": 3e-05, + "loss": 2.7804, + "step": 2272 + }, + { + "epoch": 0.6176630434782608, + "grad_norm": 1.2224652767181396, + "learning_rate": 3e-05, + "loss": 3.628, + "step": 2273 + }, + { + "epoch": 0.6179347826086956, + "grad_norm": 1.3725236654281616, + "learning_rate": 3e-05, + "loss": 3.6826, + "step": 2274 + }, + { + "epoch": 0.6182065217391305, + "grad_norm": 1.0092295408248901, + "learning_rate": 3e-05, + "loss": 2.9988, + "step": 2275 + }, + { + "epoch": 0.6184782608695653, + "grad_norm": 1.259446144104004, + "learning_rate": 3e-05, + "loss": 3.4234, + "step": 2276 + }, + { + "epoch": 0.61875, + "grad_norm": 1.220897912979126, + "learning_rate": 3e-05, + "loss": 3.2565, + "step": 2277 + }, + { + "epoch": 0.6190217391304348, + "grad_norm": 0.9857214093208313, + "learning_rate": 3e-05, + "loss": 3.2311, + "step": 2278 + }, + { + "epoch": 0.6192934782608696, + "grad_norm": 1.256536841392517, + "learning_rate": 3e-05, + "loss": 3.4595, + "step": 2279 + }, + { + "epoch": 0.6195652173913043, + "grad_norm": 1.0857690572738647, + "learning_rate": 3e-05, + "loss": 2.8696, + "step": 2280 + }, + { + "epoch": 0.6198369565217391, + "grad_norm": 0.9030542969703674, + "learning_rate": 3e-05, + "loss": 2.931, + "step": 2281 + }, + { + "epoch": 0.6201086956521739, + "grad_norm": 1.245595097541809, + "learning_rate": 3e-05, + "loss": 3.0565, + "step": 2282 + }, + { + "epoch": 0.6203804347826087, + "grad_norm": 1.130018949508667, + "learning_rate": 3e-05, + "loss": 3.6569, + "step": 2283 + }, + { + "epoch": 0.6206521739130435, + "grad_norm": 1.1791630983352661, + "learning_rate": 3e-05, + "loss": 2.9299, + "step": 2284 + }, + { + "epoch": 0.6209239130434783, + "grad_norm": 1.6365814208984375, + "learning_rate": 3e-05, + "loss": 3.7201, + "step": 2285 + }, + { + "epoch": 0.621195652173913, + "grad_norm": 1.023141622543335, + "learning_rate": 3e-05, + "loss": 3.4364, + "step": 2286 + }, + { + "epoch": 0.6214673913043478, + "grad_norm": 1.3121358156204224, + "learning_rate": 3e-05, + "loss": 3.4811, + "step": 2287 + }, + { + "epoch": 0.6217391304347826, + "grad_norm": 1.009194254875183, + "learning_rate": 3e-05, + "loss": 3.0576, + "step": 2288 + }, + { + "epoch": 0.6220108695652173, + "grad_norm": 1.0715727806091309, + "learning_rate": 3e-05, + "loss": 3.193, + "step": 2289 + }, + { + "epoch": 0.6222826086956522, + "grad_norm": 1.0925318002700806, + "learning_rate": 3e-05, + "loss": 3.1963, + "step": 2290 + }, + { + "epoch": 0.622554347826087, + "grad_norm": 1.6540546417236328, + "learning_rate": 3e-05, + "loss": 3.7859, + "step": 2291 + }, + { + "epoch": 0.6228260869565218, + "grad_norm": 1.0642199516296387, + "learning_rate": 3e-05, + "loss": 3.8112, + "step": 2292 + }, + { + "epoch": 0.6230978260869565, + "grad_norm": 1.2560967206954956, + "learning_rate": 3e-05, + "loss": 3.528, + "step": 2293 + }, + { + "epoch": 0.6233695652173913, + "grad_norm": 1.2816760540008545, + "learning_rate": 3e-05, + "loss": 3.1706, + "step": 2294 + }, + { + "epoch": 0.623641304347826, + "grad_norm": 0.9735274314880371, + "learning_rate": 3e-05, + "loss": 3.2275, + "step": 2295 + }, + { + "epoch": 0.6239130434782608, + "grad_norm": 1.1143537759780884, + "learning_rate": 3e-05, + "loss": 3.2471, + "step": 2296 + }, + { + "epoch": 0.6241847826086957, + "grad_norm": 1.2723908424377441, + "learning_rate": 3e-05, + "loss": 2.8132, + "step": 2297 + }, + { + "epoch": 0.6244565217391305, + "grad_norm": 1.269571304321289, + "learning_rate": 3e-05, + "loss": 3.2307, + "step": 2298 + }, + { + "epoch": 0.6247282608695652, + "grad_norm": 1.0772123336791992, + "learning_rate": 3e-05, + "loss": 3.2448, + "step": 2299 + }, + { + "epoch": 0.625, + "grad_norm": 1.5265069007873535, + "learning_rate": 3e-05, + "loss": 3.2314, + "step": 2300 + }, + { + "epoch": 0.6252717391304348, + "grad_norm": 1.133500576019287, + "learning_rate": 3e-05, + "loss": 3.1259, + "step": 2301 + }, + { + "epoch": 0.6255434782608695, + "grad_norm": 1.0310444831848145, + "learning_rate": 3e-05, + "loss": 3.1374, + "step": 2302 + }, + { + "epoch": 0.6258152173913043, + "grad_norm": 1.789737582206726, + "learning_rate": 3e-05, + "loss": 3.657, + "step": 2303 + }, + { + "epoch": 0.6260869565217392, + "grad_norm": 1.028426170349121, + "learning_rate": 3e-05, + "loss": 3.0235, + "step": 2304 + }, + { + "epoch": 0.626358695652174, + "grad_norm": 1.0289982557296753, + "learning_rate": 3e-05, + "loss": 3.0216, + "step": 2305 + }, + { + "epoch": 0.6266304347826087, + "grad_norm": 1.2136385440826416, + "learning_rate": 3e-05, + "loss": 3.2615, + "step": 2306 + }, + { + "epoch": 0.6269021739130435, + "grad_norm": 1.1237839460372925, + "learning_rate": 3e-05, + "loss": 3.1519, + "step": 2307 + }, + { + "epoch": 0.6271739130434782, + "grad_norm": 1.0916386842727661, + "learning_rate": 3e-05, + "loss": 3.404, + "step": 2308 + }, + { + "epoch": 0.627445652173913, + "grad_norm": 1.1526830196380615, + "learning_rate": 3e-05, + "loss": 3.671, + "step": 2309 + }, + { + "epoch": 0.6277173913043478, + "grad_norm": 1.0757122039794922, + "learning_rate": 3e-05, + "loss": 3.069, + "step": 2310 + }, + { + "epoch": 0.6279891304347827, + "grad_norm": 1.0339561700820923, + "learning_rate": 3e-05, + "loss": 2.9702, + "step": 2311 + }, + { + "epoch": 0.6282608695652174, + "grad_norm": 1.2000614404678345, + "learning_rate": 3e-05, + "loss": 2.9536, + "step": 2312 + }, + { + "epoch": 0.6285326086956522, + "grad_norm": 1.3104127645492554, + "learning_rate": 3e-05, + "loss": 3.1653, + "step": 2313 + }, + { + "epoch": 0.628804347826087, + "grad_norm": 1.2230424880981445, + "learning_rate": 3e-05, + "loss": 3.464, + "step": 2314 + }, + { + "epoch": 0.6290760869565217, + "grad_norm": 1.258482813835144, + "learning_rate": 3e-05, + "loss": 3.4273, + "step": 2315 + }, + { + "epoch": 0.6293478260869565, + "grad_norm": 1.1824918985366821, + "learning_rate": 3e-05, + "loss": 3.5311, + "step": 2316 + }, + { + "epoch": 0.6296195652173913, + "grad_norm": 0.9511439800262451, + "learning_rate": 3e-05, + "loss": 2.979, + "step": 2317 + }, + { + "epoch": 0.6298913043478261, + "grad_norm": 1.0357087850570679, + "learning_rate": 3e-05, + "loss": 2.8735, + "step": 2318 + }, + { + "epoch": 0.6301630434782609, + "grad_norm": 1.392822504043579, + "learning_rate": 3e-05, + "loss": 3.3224, + "step": 2319 + }, + { + "epoch": 0.6304347826086957, + "grad_norm": 1.4284133911132812, + "learning_rate": 3e-05, + "loss": 3.8771, + "step": 2320 + }, + { + "epoch": 0.6307065217391304, + "grad_norm": 1.2343987226486206, + "learning_rate": 3e-05, + "loss": 3.4731, + "step": 2321 + }, + { + "epoch": 0.6309782608695652, + "grad_norm": 1.3236418962478638, + "learning_rate": 3e-05, + "loss": 3.0388, + "step": 2322 + }, + { + "epoch": 0.63125, + "grad_norm": 1.332114577293396, + "learning_rate": 3e-05, + "loss": 3.1262, + "step": 2323 + }, + { + "epoch": 0.6315217391304347, + "grad_norm": 1.0655372142791748, + "learning_rate": 3e-05, + "loss": 3.2109, + "step": 2324 + }, + { + "epoch": 0.6317934782608695, + "grad_norm": 1.0099713802337646, + "learning_rate": 3e-05, + "loss": 3.2809, + "step": 2325 + }, + { + "epoch": 0.6320652173913044, + "grad_norm": 1.0744123458862305, + "learning_rate": 3e-05, + "loss": 2.862, + "step": 2326 + }, + { + "epoch": 0.6323369565217392, + "grad_norm": 1.051435112953186, + "learning_rate": 3e-05, + "loss": 2.8705, + "step": 2327 + }, + { + "epoch": 0.6326086956521739, + "grad_norm": 1.400758147239685, + "learning_rate": 3e-05, + "loss": 3.7773, + "step": 2328 + }, + { + "epoch": 0.6328804347826087, + "grad_norm": 1.1519436836242676, + "learning_rate": 3e-05, + "loss": 2.9399, + "step": 2329 + }, + { + "epoch": 0.6331521739130435, + "grad_norm": 1.264049768447876, + "learning_rate": 3e-05, + "loss": 3.7764, + "step": 2330 + }, + { + "epoch": 0.6334239130434782, + "grad_norm": 1.0223599672317505, + "learning_rate": 3e-05, + "loss": 2.9021, + "step": 2331 + }, + { + "epoch": 0.633695652173913, + "grad_norm": 1.255913257598877, + "learning_rate": 3e-05, + "loss": 3.6456, + "step": 2332 + }, + { + "epoch": 0.6339673913043479, + "grad_norm": 1.7114654779434204, + "learning_rate": 3e-05, + "loss": 4.0404, + "step": 2333 + }, + { + "epoch": 0.6342391304347826, + "grad_norm": 2.052917718887329, + "learning_rate": 3e-05, + "loss": 4.1781, + "step": 2334 + }, + { + "epoch": 0.6345108695652174, + "grad_norm": 1.376478672027588, + "learning_rate": 3e-05, + "loss": 3.4097, + "step": 2335 + }, + { + "epoch": 0.6347826086956522, + "grad_norm": 1.3323421478271484, + "learning_rate": 3e-05, + "loss": 3.0153, + "step": 2336 + }, + { + "epoch": 0.6350543478260869, + "grad_norm": 1.7771837711334229, + "learning_rate": 3e-05, + "loss": 3.7671, + "step": 2337 + }, + { + "epoch": 0.6353260869565217, + "grad_norm": 2.060122489929199, + "learning_rate": 3e-05, + "loss": 3.5854, + "step": 2338 + }, + { + "epoch": 0.6355978260869565, + "grad_norm": 1.3856672048568726, + "learning_rate": 3e-05, + "loss": 3.0497, + "step": 2339 + }, + { + "epoch": 0.6358695652173914, + "grad_norm": 1.3146904706954956, + "learning_rate": 3e-05, + "loss": 3.1533, + "step": 2340 + }, + { + "epoch": 0.6361413043478261, + "grad_norm": 1.4848661422729492, + "learning_rate": 3e-05, + "loss": 3.5613, + "step": 2341 + }, + { + "epoch": 0.6364130434782609, + "grad_norm": 1.4037156105041504, + "learning_rate": 3e-05, + "loss": 3.8576, + "step": 2342 + }, + { + "epoch": 0.6366847826086957, + "grad_norm": 1.288294792175293, + "learning_rate": 3e-05, + "loss": 3.5626, + "step": 2343 + }, + { + "epoch": 0.6369565217391304, + "grad_norm": 1.080474853515625, + "learning_rate": 3e-05, + "loss": 3.2761, + "step": 2344 + }, + { + "epoch": 0.6372282608695652, + "grad_norm": 1.1792863607406616, + "learning_rate": 3e-05, + "loss": 2.9453, + "step": 2345 + }, + { + "epoch": 0.6375, + "grad_norm": 1.4226300716400146, + "learning_rate": 3e-05, + "loss": 2.8206, + "step": 2346 + }, + { + "epoch": 0.6377717391304348, + "grad_norm": 1.5433789491653442, + "learning_rate": 3e-05, + "loss": 3.4793, + "step": 2347 + }, + { + "epoch": 0.6380434782608696, + "grad_norm": 1.0570566654205322, + "learning_rate": 3e-05, + "loss": 3.0796, + "step": 2348 + }, + { + "epoch": 0.6383152173913044, + "grad_norm": 1.4246288537979126, + "learning_rate": 3e-05, + "loss": 3.259, + "step": 2349 + }, + { + "epoch": 0.6385869565217391, + "grad_norm": 2.0091404914855957, + "learning_rate": 3e-05, + "loss": 3.2464, + "step": 2350 + }, + { + "epoch": 0.6388586956521739, + "grad_norm": 1.7661538124084473, + "learning_rate": 3e-05, + "loss": 3.0348, + "step": 2351 + }, + { + "epoch": 0.6391304347826087, + "grad_norm": 1.6644985675811768, + "learning_rate": 3e-05, + "loss": 3.2656, + "step": 2352 + }, + { + "epoch": 0.6394021739130434, + "grad_norm": 1.192597508430481, + "learning_rate": 3e-05, + "loss": 3.2846, + "step": 2353 + }, + { + "epoch": 0.6396739130434783, + "grad_norm": 1.234485149383545, + "learning_rate": 3e-05, + "loss": 3.2553, + "step": 2354 + }, + { + "epoch": 0.6399456521739131, + "grad_norm": 1.2841020822525024, + "learning_rate": 3e-05, + "loss": 2.8951, + "step": 2355 + }, + { + "epoch": 0.6402173913043478, + "grad_norm": 1.119943618774414, + "learning_rate": 3e-05, + "loss": 2.7422, + "step": 2356 + }, + { + "epoch": 0.6404891304347826, + "grad_norm": 1.5436328649520874, + "learning_rate": 3e-05, + "loss": 3.9712, + "step": 2357 + }, + { + "epoch": 0.6407608695652174, + "grad_norm": 1.1272965669631958, + "learning_rate": 3e-05, + "loss": 3.1857, + "step": 2358 + }, + { + "epoch": 0.6410326086956522, + "grad_norm": 1.340259075164795, + "learning_rate": 3e-05, + "loss": 3.7488, + "step": 2359 + }, + { + "epoch": 0.6413043478260869, + "grad_norm": 1.486802577972412, + "learning_rate": 3e-05, + "loss": 3.8155, + "step": 2360 + }, + { + "epoch": 0.6415760869565217, + "grad_norm": 0.9581712484359741, + "learning_rate": 3e-05, + "loss": 3.117, + "step": 2361 + }, + { + "epoch": 0.6418478260869566, + "grad_norm": 1.012144923210144, + "learning_rate": 3e-05, + "loss": 3.0744, + "step": 2362 + }, + { + "epoch": 0.6421195652173913, + "grad_norm": 1.1179242134094238, + "learning_rate": 3e-05, + "loss": 3.2825, + "step": 2363 + }, + { + "epoch": 0.6423913043478261, + "grad_norm": 1.3681845664978027, + "learning_rate": 3e-05, + "loss": 3.1833, + "step": 2364 + }, + { + "epoch": 0.6426630434782609, + "grad_norm": 1.192631721496582, + "learning_rate": 3e-05, + "loss": 3.3842, + "step": 2365 + }, + { + "epoch": 0.6429347826086956, + "grad_norm": 1.2317907810211182, + "learning_rate": 3e-05, + "loss": 3.1083, + "step": 2366 + }, + { + "epoch": 0.6432065217391304, + "grad_norm": 1.0752184391021729, + "learning_rate": 3e-05, + "loss": 3.2455, + "step": 2367 + }, + { + "epoch": 0.6434782608695652, + "grad_norm": 1.035239338874817, + "learning_rate": 3e-05, + "loss": 2.9436, + "step": 2368 + }, + { + "epoch": 0.64375, + "grad_norm": 1.3296703100204468, + "learning_rate": 3e-05, + "loss": 3.2708, + "step": 2369 + }, + { + "epoch": 0.6440217391304348, + "grad_norm": 1.2996970415115356, + "learning_rate": 3e-05, + "loss": 3.8312, + "step": 2370 + }, + { + "epoch": 0.6442934782608696, + "grad_norm": 0.9831019639968872, + "learning_rate": 3e-05, + "loss": 3.2478, + "step": 2371 + }, + { + "epoch": 0.6445652173913043, + "grad_norm": 0.9067811965942383, + "learning_rate": 3e-05, + "loss": 3.1556, + "step": 2372 + }, + { + "epoch": 0.6448369565217391, + "grad_norm": 1.2001372575759888, + "learning_rate": 3e-05, + "loss": 3.7225, + "step": 2373 + }, + { + "epoch": 0.6451086956521739, + "grad_norm": 1.1935441493988037, + "learning_rate": 3e-05, + "loss": 3.3109, + "step": 2374 + }, + { + "epoch": 0.6453804347826086, + "grad_norm": 0.9141666889190674, + "learning_rate": 3e-05, + "loss": 3.0955, + "step": 2375 + }, + { + "epoch": 0.6456521739130435, + "grad_norm": 1.0943185091018677, + "learning_rate": 3e-05, + "loss": 3.3094, + "step": 2376 + }, + { + "epoch": 0.6459239130434783, + "grad_norm": 1.1921093463897705, + "learning_rate": 3e-05, + "loss": 3.3715, + "step": 2377 + }, + { + "epoch": 0.6461956521739131, + "grad_norm": 1.0566505193710327, + "learning_rate": 3e-05, + "loss": 3.1073, + "step": 2378 + }, + { + "epoch": 0.6464673913043478, + "grad_norm": 1.5691936016082764, + "learning_rate": 3e-05, + "loss": 3.4321, + "step": 2379 + }, + { + "epoch": 0.6467391304347826, + "grad_norm": 1.041152000427246, + "learning_rate": 3e-05, + "loss": 2.6529, + "step": 2380 + }, + { + "epoch": 0.6470108695652174, + "grad_norm": 1.2687933444976807, + "learning_rate": 3e-05, + "loss": 2.9547, + "step": 2381 + }, + { + "epoch": 0.6472826086956521, + "grad_norm": 1.074779748916626, + "learning_rate": 3e-05, + "loss": 3.0635, + "step": 2382 + }, + { + "epoch": 0.647554347826087, + "grad_norm": 1.0904431343078613, + "learning_rate": 3e-05, + "loss": 2.8928, + "step": 2383 + }, + { + "epoch": 0.6478260869565218, + "grad_norm": 1.3329259157180786, + "learning_rate": 3e-05, + "loss": 3.6509, + "step": 2384 + }, + { + "epoch": 0.6480978260869565, + "grad_norm": 1.2522311210632324, + "learning_rate": 3e-05, + "loss": 3.6357, + "step": 2385 + }, + { + "epoch": 0.6483695652173913, + "grad_norm": 1.2754337787628174, + "learning_rate": 3e-05, + "loss": 3.7422, + "step": 2386 + }, + { + "epoch": 0.6486413043478261, + "grad_norm": 1.338080883026123, + "learning_rate": 3e-05, + "loss": 3.3859, + "step": 2387 + }, + { + "epoch": 0.6489130434782608, + "grad_norm": 1.1422319412231445, + "learning_rate": 3e-05, + "loss": 3.8218, + "step": 2388 + }, + { + "epoch": 0.6491847826086956, + "grad_norm": 1.647072196006775, + "learning_rate": 3e-05, + "loss": 3.6864, + "step": 2389 + }, + { + "epoch": 0.6494565217391305, + "grad_norm": 1.1428362131118774, + "learning_rate": 3e-05, + "loss": 2.8359, + "step": 2390 + }, + { + "epoch": 0.6497282608695653, + "grad_norm": 1.1382925510406494, + "learning_rate": 3e-05, + "loss": 3.1163, + "step": 2391 + }, + { + "epoch": 0.65, + "grad_norm": 1.0694186687469482, + "learning_rate": 3e-05, + "loss": 3.1842, + "step": 2392 + }, + { + "epoch": 0.6502717391304348, + "grad_norm": 1.413272738456726, + "learning_rate": 3e-05, + "loss": 3.7837, + "step": 2393 + }, + { + "epoch": 0.6505434782608696, + "grad_norm": 1.353119134902954, + "learning_rate": 3e-05, + "loss": 3.2845, + "step": 2394 + }, + { + "epoch": 0.6508152173913043, + "grad_norm": 1.542140245437622, + "learning_rate": 3e-05, + "loss": 3.4894, + "step": 2395 + }, + { + "epoch": 0.6510869565217391, + "grad_norm": 1.6058558225631714, + "learning_rate": 3e-05, + "loss": 3.4778, + "step": 2396 + }, + { + "epoch": 0.6513586956521739, + "grad_norm": 1.719277262687683, + "learning_rate": 3e-05, + "loss": 3.5328, + "step": 2397 + }, + { + "epoch": 0.6516304347826087, + "grad_norm": 1.7350586652755737, + "learning_rate": 3e-05, + "loss": 3.5851, + "step": 2398 + }, + { + "epoch": 0.6519021739130435, + "grad_norm": 1.4615671634674072, + "learning_rate": 3e-05, + "loss": 4.099, + "step": 2399 + }, + { + "epoch": 0.6521739130434783, + "grad_norm": 1.194122314453125, + "learning_rate": 3e-05, + "loss": 3.353, + "step": 2400 + }, + { + "epoch": 0.652445652173913, + "grad_norm": 1.4127830266952515, + "learning_rate": 3e-05, + "loss": 3.8697, + "step": 2401 + }, + { + "epoch": 0.6527173913043478, + "grad_norm": 1.6707136631011963, + "learning_rate": 3e-05, + "loss": 3.9621, + "step": 2402 + }, + { + "epoch": 0.6529891304347826, + "grad_norm": 1.4606815576553345, + "learning_rate": 3e-05, + "loss": 3.7677, + "step": 2403 + }, + { + "epoch": 0.6532608695652173, + "grad_norm": 1.2474243640899658, + "learning_rate": 3e-05, + "loss": 3.5581, + "step": 2404 + }, + { + "epoch": 0.6535326086956522, + "grad_norm": 1.1824653148651123, + "learning_rate": 3e-05, + "loss": 2.9751, + "step": 2405 + }, + { + "epoch": 0.653804347826087, + "grad_norm": 1.22133469581604, + "learning_rate": 3e-05, + "loss": 3.3986, + "step": 2406 + }, + { + "epoch": 0.6540760869565218, + "grad_norm": 1.160557746887207, + "learning_rate": 3e-05, + "loss": 3.2008, + "step": 2407 + }, + { + "epoch": 0.6543478260869565, + "grad_norm": 1.2002159357070923, + "learning_rate": 3e-05, + "loss": 3.6029, + "step": 2408 + }, + { + "epoch": 0.6546195652173913, + "grad_norm": 1.1110812425613403, + "learning_rate": 3e-05, + "loss": 3.2051, + "step": 2409 + }, + { + "epoch": 0.654891304347826, + "grad_norm": 0.9706496596336365, + "learning_rate": 3e-05, + "loss": 2.9527, + "step": 2410 + }, + { + "epoch": 0.6551630434782608, + "grad_norm": 1.0990405082702637, + "learning_rate": 3e-05, + "loss": 2.6553, + "step": 2411 + }, + { + "epoch": 0.6554347826086957, + "grad_norm": 1.065306544303894, + "learning_rate": 3e-05, + "loss": 3.3728, + "step": 2412 + }, + { + "epoch": 0.6557065217391305, + "grad_norm": 0.956051766872406, + "learning_rate": 3e-05, + "loss": 2.9055, + "step": 2413 + }, + { + "epoch": 0.6559782608695652, + "grad_norm": 1.1310175657272339, + "learning_rate": 3e-05, + "loss": 3.8269, + "step": 2414 + }, + { + "epoch": 0.65625, + "grad_norm": 0.9513333439826965, + "learning_rate": 3e-05, + "loss": 3.2811, + "step": 2415 + }, + { + "epoch": 0.6565217391304348, + "grad_norm": 1.301697015762329, + "learning_rate": 3e-05, + "loss": 3.5579, + "step": 2416 + }, + { + "epoch": 0.6567934782608695, + "grad_norm": 0.9810994267463684, + "learning_rate": 3e-05, + "loss": 2.9914, + "step": 2417 + }, + { + "epoch": 0.6570652173913043, + "grad_norm": 1.3848271369934082, + "learning_rate": 3e-05, + "loss": 3.8109, + "step": 2418 + }, + { + "epoch": 0.6573369565217392, + "grad_norm": 1.169350504875183, + "learning_rate": 3e-05, + "loss": 2.8291, + "step": 2419 + }, + { + "epoch": 0.657608695652174, + "grad_norm": 1.3478142023086548, + "learning_rate": 3e-05, + "loss": 3.5609, + "step": 2420 + }, + { + "epoch": 0.6578804347826087, + "grad_norm": 1.1918085813522339, + "learning_rate": 3e-05, + "loss": 3.5339, + "step": 2421 + }, + { + "epoch": 0.6581521739130435, + "grad_norm": 1.4257116317749023, + "learning_rate": 3e-05, + "loss": 3.4636, + "step": 2422 + }, + { + "epoch": 0.6584239130434782, + "grad_norm": 1.2269481420516968, + "learning_rate": 3e-05, + "loss": 3.2741, + "step": 2423 + }, + { + "epoch": 0.658695652173913, + "grad_norm": 1.2066056728363037, + "learning_rate": 3e-05, + "loss": 3.2177, + "step": 2424 + }, + { + "epoch": 0.6589673913043478, + "grad_norm": 1.3328955173492432, + "learning_rate": 3e-05, + "loss": 3.527, + "step": 2425 + }, + { + "epoch": 0.6592391304347827, + "grad_norm": 1.1746604442596436, + "learning_rate": 3e-05, + "loss": 3.1194, + "step": 2426 + }, + { + "epoch": 0.6595108695652174, + "grad_norm": 1.028620958328247, + "learning_rate": 3e-05, + "loss": 3.1119, + "step": 2427 + }, + { + "epoch": 0.6597826086956522, + "grad_norm": 1.0080275535583496, + "learning_rate": 3e-05, + "loss": 2.7994, + "step": 2428 + }, + { + "epoch": 0.660054347826087, + "grad_norm": 0.8145027756690979, + "learning_rate": 3e-05, + "loss": 2.9428, + "step": 2429 + }, + { + "epoch": 0.6603260869565217, + "grad_norm": 1.0437175035476685, + "learning_rate": 3e-05, + "loss": 2.9878, + "step": 2430 + }, + { + "epoch": 0.6605978260869565, + "grad_norm": 1.130467414855957, + "learning_rate": 3e-05, + "loss": 3.7007, + "step": 2431 + }, + { + "epoch": 0.6608695652173913, + "grad_norm": 1.0865567922592163, + "learning_rate": 3e-05, + "loss": 3.1175, + "step": 2432 + }, + { + "epoch": 0.6611413043478261, + "grad_norm": 1.1606049537658691, + "learning_rate": 3e-05, + "loss": 3.3582, + "step": 2433 + }, + { + "epoch": 0.6614130434782609, + "grad_norm": 0.9023473262786865, + "learning_rate": 3e-05, + "loss": 2.9861, + "step": 2434 + }, + { + "epoch": 0.6616847826086957, + "grad_norm": 0.856779932975769, + "learning_rate": 3e-05, + "loss": 3.0402, + "step": 2435 + }, + { + "epoch": 0.6619565217391304, + "grad_norm": 1.0253337621688843, + "learning_rate": 3e-05, + "loss": 3.0909, + "step": 2436 + }, + { + "epoch": 0.6622282608695652, + "grad_norm": 1.0018707513809204, + "learning_rate": 3e-05, + "loss": 2.9996, + "step": 2437 + }, + { + "epoch": 0.6625, + "grad_norm": 1.0722147226333618, + "learning_rate": 3e-05, + "loss": 3.3102, + "step": 2438 + }, + { + "epoch": 0.6627717391304347, + "grad_norm": 1.0798076391220093, + "learning_rate": 3e-05, + "loss": 2.9907, + "step": 2439 + }, + { + "epoch": 0.6630434782608695, + "grad_norm": 1.1593507528305054, + "learning_rate": 3e-05, + "loss": 3.3741, + "step": 2440 + }, + { + "epoch": 0.6633152173913044, + "grad_norm": 1.0947400331497192, + "learning_rate": 3e-05, + "loss": 3.0087, + "step": 2441 + }, + { + "epoch": 0.6635869565217392, + "grad_norm": 1.2882115840911865, + "learning_rate": 3e-05, + "loss": 3.1978, + "step": 2442 + }, + { + "epoch": 0.6638586956521739, + "grad_norm": 1.2732001543045044, + "learning_rate": 3e-05, + "loss": 3.8371, + "step": 2443 + }, + { + "epoch": 0.6641304347826087, + "grad_norm": 0.9186998605728149, + "learning_rate": 3e-05, + "loss": 2.643, + "step": 2444 + }, + { + "epoch": 0.6644021739130435, + "grad_norm": 1.0889346599578857, + "learning_rate": 3e-05, + "loss": 2.9247, + "step": 2445 + }, + { + "epoch": 0.6646739130434782, + "grad_norm": 1.094017505645752, + "learning_rate": 3e-05, + "loss": 3.1954, + "step": 2446 + }, + { + "epoch": 0.664945652173913, + "grad_norm": 0.9859854578971863, + "learning_rate": 3e-05, + "loss": 2.7335, + "step": 2447 + }, + { + "epoch": 0.6652173913043479, + "grad_norm": 1.8823328018188477, + "learning_rate": 3e-05, + "loss": 3.4934, + "step": 2448 + }, + { + "epoch": 0.6654891304347826, + "grad_norm": 0.9667043685913086, + "learning_rate": 3e-05, + "loss": 3.0711, + "step": 2449 + }, + { + "epoch": 0.6657608695652174, + "grad_norm": 1.0037490129470825, + "learning_rate": 3e-05, + "loss": 3.1836, + "step": 2450 + }, + { + "epoch": 0.6660326086956522, + "grad_norm": 1.0636874437332153, + "learning_rate": 3e-05, + "loss": 3.3269, + "step": 2451 + }, + { + "epoch": 0.6663043478260869, + "grad_norm": 1.02604079246521, + "learning_rate": 3e-05, + "loss": 3.4928, + "step": 2452 + }, + { + "epoch": 0.6665760869565217, + "grad_norm": 1.050881028175354, + "learning_rate": 3e-05, + "loss": 3.3983, + "step": 2453 + }, + { + "epoch": 0.6668478260869565, + "grad_norm": 1.1651184558868408, + "learning_rate": 3e-05, + "loss": 3.573, + "step": 2454 + }, + { + "epoch": 0.6671195652173914, + "grad_norm": 0.9464408755302429, + "learning_rate": 3e-05, + "loss": 3.4415, + "step": 2455 + }, + { + "epoch": 0.6673913043478261, + "grad_norm": 0.9894149899482727, + "learning_rate": 3e-05, + "loss": 3.4124, + "step": 2456 + }, + { + "epoch": 0.6676630434782609, + "grad_norm": 1.1918611526489258, + "learning_rate": 3e-05, + "loss": 3.1563, + "step": 2457 + }, + { + "epoch": 0.6679347826086957, + "grad_norm": 1.1251368522644043, + "learning_rate": 3e-05, + "loss": 3.0224, + "step": 2458 + }, + { + "epoch": 0.6682065217391304, + "grad_norm": 0.9575350880622864, + "learning_rate": 3e-05, + "loss": 3.1342, + "step": 2459 + }, + { + "epoch": 0.6684782608695652, + "grad_norm": 1.1674573421478271, + "learning_rate": 3e-05, + "loss": 3.5287, + "step": 2460 + }, + { + "epoch": 0.66875, + "grad_norm": 1.3104429244995117, + "learning_rate": 3e-05, + "loss": 3.3732, + "step": 2461 + }, + { + "epoch": 0.6690217391304348, + "grad_norm": 1.0205787420272827, + "learning_rate": 3e-05, + "loss": 3.3107, + "step": 2462 + }, + { + "epoch": 0.6692934782608696, + "grad_norm": 1.1491562128067017, + "learning_rate": 3e-05, + "loss": 3.1578, + "step": 2463 + }, + { + "epoch": 0.6695652173913044, + "grad_norm": 1.2550066709518433, + "learning_rate": 3e-05, + "loss": 2.9625, + "step": 2464 + }, + { + "epoch": 0.6698369565217391, + "grad_norm": 1.396528720855713, + "learning_rate": 3e-05, + "loss": 3.4526, + "step": 2465 + }, + { + "epoch": 0.6701086956521739, + "grad_norm": 1.3484258651733398, + "learning_rate": 3e-05, + "loss": 3.0718, + "step": 2466 + }, + { + "epoch": 0.6703804347826087, + "grad_norm": 1.2837543487548828, + "learning_rate": 3e-05, + "loss": 3.9352, + "step": 2467 + }, + { + "epoch": 0.6706521739130434, + "grad_norm": 1.126395583152771, + "learning_rate": 3e-05, + "loss": 3.2975, + "step": 2468 + }, + { + "epoch": 0.6709239130434783, + "grad_norm": 1.0975403785705566, + "learning_rate": 3e-05, + "loss": 2.9911, + "step": 2469 + }, + { + "epoch": 0.6711956521739131, + "grad_norm": 1.260852336883545, + "learning_rate": 3e-05, + "loss": 3.4067, + "step": 2470 + }, + { + "epoch": 0.6714673913043478, + "grad_norm": 1.3457390069961548, + "learning_rate": 3e-05, + "loss": 3.3277, + "step": 2471 + }, + { + "epoch": 0.6717391304347826, + "grad_norm": 1.6091570854187012, + "learning_rate": 3e-05, + "loss": 3.8467, + "step": 2472 + }, + { + "epoch": 0.6720108695652174, + "grad_norm": 1.3068939447402954, + "learning_rate": 3e-05, + "loss": 3.5756, + "step": 2473 + }, + { + "epoch": 0.6722826086956522, + "grad_norm": 1.532850742340088, + "learning_rate": 3e-05, + "loss": 3.6188, + "step": 2474 + }, + { + "epoch": 0.6725543478260869, + "grad_norm": 1.179315447807312, + "learning_rate": 3e-05, + "loss": 3.3408, + "step": 2475 + }, + { + "epoch": 0.6728260869565217, + "grad_norm": 1.278639793395996, + "learning_rate": 3e-05, + "loss": 3.9328, + "step": 2476 + }, + { + "epoch": 0.6730978260869566, + "grad_norm": 0.9712943434715271, + "learning_rate": 3e-05, + "loss": 2.7786, + "step": 2477 + }, + { + "epoch": 0.6733695652173913, + "grad_norm": 1.1749228239059448, + "learning_rate": 3e-05, + "loss": 2.7182, + "step": 2478 + }, + { + "epoch": 0.6736413043478261, + "grad_norm": 1.3901921510696411, + "learning_rate": 3e-05, + "loss": 3.2455, + "step": 2479 + }, + { + "epoch": 0.6739130434782609, + "grad_norm": 1.3088897466659546, + "learning_rate": 3e-05, + "loss": 3.4134, + "step": 2480 + }, + { + "epoch": 0.6741847826086956, + "grad_norm": 1.2072702646255493, + "learning_rate": 3e-05, + "loss": 3.7264, + "step": 2481 + }, + { + "epoch": 0.6744565217391304, + "grad_norm": 1.3006356954574585, + "learning_rate": 3e-05, + "loss": 3.1591, + "step": 2482 + }, + { + "epoch": 0.6747282608695652, + "grad_norm": 2.3080663681030273, + "learning_rate": 3e-05, + "loss": 3.6425, + "step": 2483 + }, + { + "epoch": 0.675, + "grad_norm": 1.3426471948623657, + "learning_rate": 3e-05, + "loss": 2.9936, + "step": 2484 + }, + { + "epoch": 0.6752717391304348, + "grad_norm": 1.1594306230545044, + "learning_rate": 3e-05, + "loss": 3.3481, + "step": 2485 + }, + { + "epoch": 0.6755434782608696, + "grad_norm": 1.0936838388442993, + "learning_rate": 3e-05, + "loss": 3.0749, + "step": 2486 + }, + { + "epoch": 0.6758152173913043, + "grad_norm": 1.3417373895645142, + "learning_rate": 3e-05, + "loss": 3.2727, + "step": 2487 + }, + { + "epoch": 0.6760869565217391, + "grad_norm": 1.6536362171173096, + "learning_rate": 3e-05, + "loss": 3.3898, + "step": 2488 + }, + { + "epoch": 0.6763586956521739, + "grad_norm": 1.2044962644577026, + "learning_rate": 3e-05, + "loss": 3.06, + "step": 2489 + }, + { + "epoch": 0.6766304347826086, + "grad_norm": 1.6772176027297974, + "learning_rate": 3e-05, + "loss": 3.6963, + "step": 2490 + }, + { + "epoch": 0.6769021739130435, + "grad_norm": 1.2547991275787354, + "learning_rate": 3e-05, + "loss": 2.9258, + "step": 2491 + }, + { + "epoch": 0.6771739130434783, + "grad_norm": 1.2318508625030518, + "learning_rate": 3e-05, + "loss": 3.4305, + "step": 2492 + }, + { + "epoch": 0.6774456521739131, + "grad_norm": 1.4965567588806152, + "learning_rate": 3e-05, + "loss": 3.4216, + "step": 2493 + }, + { + "epoch": 0.6777173913043478, + "grad_norm": 1.6711788177490234, + "learning_rate": 3e-05, + "loss": 2.8026, + "step": 2494 + }, + { + "epoch": 0.6779891304347826, + "grad_norm": 1.285457968711853, + "learning_rate": 3e-05, + "loss": 2.9369, + "step": 2495 + }, + { + "epoch": 0.6782608695652174, + "grad_norm": 1.1391198635101318, + "learning_rate": 3e-05, + "loss": 2.6831, + "step": 2496 + }, + { + "epoch": 0.6785326086956521, + "grad_norm": 1.4873690605163574, + "learning_rate": 3e-05, + "loss": 3.5054, + "step": 2497 + }, + { + "epoch": 0.678804347826087, + "grad_norm": 1.5623383522033691, + "learning_rate": 3e-05, + "loss": 3.5831, + "step": 2498 + }, + { + "epoch": 0.6790760869565218, + "grad_norm": 1.0868808031082153, + "learning_rate": 3e-05, + "loss": 3.5378, + "step": 2499 + }, + { + "epoch": 0.6793478260869565, + "grad_norm": 1.5260769128799438, + "learning_rate": 3e-05, + "loss": 3.6442, + "step": 2500 + }, + { + "epoch": 0.6796195652173913, + "grad_norm": 1.5094234943389893, + "learning_rate": 3e-05, + "loss": 2.8721, + "step": 2501 + }, + { + "epoch": 0.6798913043478261, + "grad_norm": 1.03408944606781, + "learning_rate": 3e-05, + "loss": 3.2489, + "step": 2502 + }, + { + "epoch": 0.6801630434782608, + "grad_norm": 1.5139715671539307, + "learning_rate": 3e-05, + "loss": 3.1632, + "step": 2503 + }, + { + "epoch": 0.6804347826086956, + "grad_norm": 0.8952838778495789, + "learning_rate": 3e-05, + "loss": 2.678, + "step": 2504 + }, + { + "epoch": 0.6807065217391305, + "grad_norm": 0.886978268623352, + "learning_rate": 3e-05, + "loss": 2.9663, + "step": 2505 + }, + { + "epoch": 0.6809782608695653, + "grad_norm": 1.4015369415283203, + "learning_rate": 3e-05, + "loss": 3.4536, + "step": 2506 + }, + { + "epoch": 0.68125, + "grad_norm": 1.1252248287200928, + "learning_rate": 3e-05, + "loss": 3.1302, + "step": 2507 + }, + { + "epoch": 0.6815217391304348, + "grad_norm": 1.2304083108901978, + "learning_rate": 3e-05, + "loss": 3.5204, + "step": 2508 + }, + { + "epoch": 0.6817934782608696, + "grad_norm": 1.2945035696029663, + "learning_rate": 3e-05, + "loss": 3.7562, + "step": 2509 + }, + { + "epoch": 0.6820652173913043, + "grad_norm": 1.1755324602127075, + "learning_rate": 3e-05, + "loss": 2.787, + "step": 2510 + }, + { + "epoch": 0.6823369565217391, + "grad_norm": 1.413987159729004, + "learning_rate": 3e-05, + "loss": 3.5305, + "step": 2511 + }, + { + "epoch": 0.6826086956521739, + "grad_norm": 1.0807994604110718, + "learning_rate": 3e-05, + "loss": 3.2698, + "step": 2512 + }, + { + "epoch": 0.6828804347826087, + "grad_norm": 1.5065345764160156, + "learning_rate": 3e-05, + "loss": 3.3808, + "step": 2513 + }, + { + "epoch": 0.6831521739130435, + "grad_norm": 1.2206026315689087, + "learning_rate": 3e-05, + "loss": 2.9492, + "step": 2514 + }, + { + "epoch": 0.6834239130434783, + "grad_norm": 1.1851751804351807, + "learning_rate": 3e-05, + "loss": 3.425, + "step": 2515 + }, + { + "epoch": 0.683695652173913, + "grad_norm": 1.1727910041809082, + "learning_rate": 3e-05, + "loss": 3.6993, + "step": 2516 + }, + { + "epoch": 0.6839673913043478, + "grad_norm": 1.5769599676132202, + "learning_rate": 3e-05, + "loss": 3.3359, + "step": 2517 + }, + { + "epoch": 0.6842391304347826, + "grad_norm": 1.5909152030944824, + "learning_rate": 3e-05, + "loss": 3.1606, + "step": 2518 + }, + { + "epoch": 0.6845108695652173, + "grad_norm": 1.094890832901001, + "learning_rate": 3e-05, + "loss": 3.2958, + "step": 2519 + }, + { + "epoch": 0.6847826086956522, + "grad_norm": 1.5951069593429565, + "learning_rate": 3e-05, + "loss": 3.6897, + "step": 2520 + }, + { + "epoch": 0.685054347826087, + "grad_norm": 1.3260122537612915, + "learning_rate": 3e-05, + "loss": 3.5941, + "step": 2521 + }, + { + "epoch": 0.6853260869565218, + "grad_norm": 1.431256651878357, + "learning_rate": 3e-05, + "loss": 2.8581, + "step": 2522 + }, + { + "epoch": 0.6855978260869565, + "grad_norm": 1.3032617568969727, + "learning_rate": 3e-05, + "loss": 3.3358, + "step": 2523 + }, + { + "epoch": 0.6858695652173913, + "grad_norm": 1.0501641035079956, + "learning_rate": 3e-05, + "loss": 2.9872, + "step": 2524 + }, + { + "epoch": 0.686141304347826, + "grad_norm": 1.1217620372772217, + "learning_rate": 3e-05, + "loss": 3.1064, + "step": 2525 + }, + { + "epoch": 0.6864130434782608, + "grad_norm": 1.0204346179962158, + "learning_rate": 3e-05, + "loss": 3.13, + "step": 2526 + }, + { + "epoch": 0.6866847826086957, + "grad_norm": 1.479447603225708, + "learning_rate": 3e-05, + "loss": 3.3739, + "step": 2527 + }, + { + "epoch": 0.6869565217391305, + "grad_norm": 1.3281017541885376, + "learning_rate": 3e-05, + "loss": 3.1517, + "step": 2528 + }, + { + "epoch": 0.6872282608695652, + "grad_norm": 1.358286738395691, + "learning_rate": 3e-05, + "loss": 3.215, + "step": 2529 + }, + { + "epoch": 0.6875, + "grad_norm": 1.169724464416504, + "learning_rate": 3e-05, + "loss": 2.9665, + "step": 2530 + }, + { + "epoch": 0.6877717391304348, + "grad_norm": 1.373205542564392, + "learning_rate": 3e-05, + "loss": 3.3161, + "step": 2531 + }, + { + "epoch": 0.6880434782608695, + "grad_norm": 1.0364346504211426, + "learning_rate": 3e-05, + "loss": 3.135, + "step": 2532 + }, + { + "epoch": 0.6883152173913043, + "grad_norm": 1.2910298109054565, + "learning_rate": 3e-05, + "loss": 3.2629, + "step": 2533 + }, + { + "epoch": 0.6885869565217392, + "grad_norm": 1.7945650815963745, + "learning_rate": 3e-05, + "loss": 3.5541, + "step": 2534 + }, + { + "epoch": 0.688858695652174, + "grad_norm": 1.3837659358978271, + "learning_rate": 3e-05, + "loss": 3.5515, + "step": 2535 + }, + { + "epoch": 0.6891304347826087, + "grad_norm": 1.3154077529907227, + "learning_rate": 3e-05, + "loss": 3.6571, + "step": 2536 + }, + { + "epoch": 0.6894021739130435, + "grad_norm": 1.3466911315917969, + "learning_rate": 3e-05, + "loss": 3.0786, + "step": 2537 + }, + { + "epoch": 0.6896739130434782, + "grad_norm": 1.3170411586761475, + "learning_rate": 3e-05, + "loss": 2.9075, + "step": 2538 + }, + { + "epoch": 0.689945652173913, + "grad_norm": 1.1219418048858643, + "learning_rate": 3e-05, + "loss": 3.1026, + "step": 2539 + }, + { + "epoch": 0.6902173913043478, + "grad_norm": 1.15764319896698, + "learning_rate": 3e-05, + "loss": 3.45, + "step": 2540 + }, + { + "epoch": 0.6904891304347827, + "grad_norm": 1.196807861328125, + "learning_rate": 3e-05, + "loss": 3.185, + "step": 2541 + }, + { + "epoch": 0.6907608695652174, + "grad_norm": 1.187324047088623, + "learning_rate": 3e-05, + "loss": 3.4707, + "step": 2542 + }, + { + "epoch": 0.6910326086956522, + "grad_norm": 1.0559121370315552, + "learning_rate": 3e-05, + "loss": 2.9871, + "step": 2543 + }, + { + "epoch": 0.691304347826087, + "grad_norm": 1.1243149042129517, + "learning_rate": 3e-05, + "loss": 3.5086, + "step": 2544 + }, + { + "epoch": 0.6915760869565217, + "grad_norm": 1.1120831966400146, + "learning_rate": 3e-05, + "loss": 3.1197, + "step": 2545 + }, + { + "epoch": 0.6918478260869565, + "grad_norm": 1.294692039489746, + "learning_rate": 3e-05, + "loss": 3.9801, + "step": 2546 + }, + { + "epoch": 0.6921195652173913, + "grad_norm": 1.3371599912643433, + "learning_rate": 3e-05, + "loss": 3.8999, + "step": 2547 + }, + { + "epoch": 0.6923913043478261, + "grad_norm": 1.219671368598938, + "learning_rate": 3e-05, + "loss": 3.3713, + "step": 2548 + }, + { + "epoch": 0.6926630434782609, + "grad_norm": 0.9593479037284851, + "learning_rate": 3e-05, + "loss": 2.8705, + "step": 2549 + }, + { + "epoch": 0.6929347826086957, + "grad_norm": 1.2361105680465698, + "learning_rate": 3e-05, + "loss": 3.6656, + "step": 2550 + }, + { + "epoch": 0.6932065217391304, + "grad_norm": 0.9579165577888489, + "learning_rate": 3e-05, + "loss": 2.9437, + "step": 2551 + }, + { + "epoch": 0.6934782608695652, + "grad_norm": 1.0255228281021118, + "learning_rate": 3e-05, + "loss": 3.0341, + "step": 2552 + }, + { + "epoch": 0.69375, + "grad_norm": 1.2201510667800903, + "learning_rate": 3e-05, + "loss": 3.6407, + "step": 2553 + }, + { + "epoch": 0.6940217391304347, + "grad_norm": 1.1329454183578491, + "learning_rate": 3e-05, + "loss": 3.4387, + "step": 2554 + }, + { + "epoch": 0.6942934782608695, + "grad_norm": 1.0753947496414185, + "learning_rate": 3e-05, + "loss": 3.1663, + "step": 2555 + }, + { + "epoch": 0.6945652173913044, + "grad_norm": 1.8046010732650757, + "learning_rate": 3e-05, + "loss": 3.624, + "step": 2556 + }, + { + "epoch": 0.6948369565217392, + "grad_norm": 1.0631359815597534, + "learning_rate": 3e-05, + "loss": 3.2537, + "step": 2557 + }, + { + "epoch": 0.6951086956521739, + "grad_norm": 1.0694079399108887, + "learning_rate": 3e-05, + "loss": 2.8613, + "step": 2558 + }, + { + "epoch": 0.6953804347826087, + "grad_norm": 1.2790333032608032, + "learning_rate": 3e-05, + "loss": 3.0732, + "step": 2559 + }, + { + "epoch": 0.6956521739130435, + "grad_norm": 1.253812551498413, + "learning_rate": 3e-05, + "loss": 3.0818, + "step": 2560 + }, + { + "epoch": 0.6959239130434782, + "grad_norm": 1.2076342105865479, + "learning_rate": 3e-05, + "loss": 2.7011, + "step": 2561 + }, + { + "epoch": 0.696195652173913, + "grad_norm": 1.2318416833877563, + "learning_rate": 3e-05, + "loss": 2.7658, + "step": 2562 + }, + { + "epoch": 0.6964673913043479, + "grad_norm": 1.442963719367981, + "learning_rate": 3e-05, + "loss": 3.7564, + "step": 2563 + }, + { + "epoch": 0.6967391304347826, + "grad_norm": 1.3947499990463257, + "learning_rate": 3e-05, + "loss": 3.4476, + "step": 2564 + }, + { + "epoch": 0.6970108695652174, + "grad_norm": 1.8033784627914429, + "learning_rate": 3e-05, + "loss": 3.7798, + "step": 2565 + }, + { + "epoch": 0.6972826086956522, + "grad_norm": 1.3678721189498901, + "learning_rate": 3e-05, + "loss": 3.3788, + "step": 2566 + }, + { + "epoch": 0.6975543478260869, + "grad_norm": 1.0561424493789673, + "learning_rate": 3e-05, + "loss": 3.1749, + "step": 2567 + }, + { + "epoch": 0.6978260869565217, + "grad_norm": 1.2744587659835815, + "learning_rate": 3e-05, + "loss": 3.3584, + "step": 2568 + }, + { + "epoch": 0.6980978260869565, + "grad_norm": 1.2167631387710571, + "learning_rate": 3e-05, + "loss": 3.4043, + "step": 2569 + }, + { + "epoch": 0.6983695652173914, + "grad_norm": 1.1461526155471802, + "learning_rate": 3e-05, + "loss": 3.381, + "step": 2570 + }, + { + "epoch": 0.6986413043478261, + "grad_norm": 1.173651933670044, + "learning_rate": 3e-05, + "loss": 2.8328, + "step": 2571 + }, + { + "epoch": 0.6989130434782609, + "grad_norm": 1.3456881046295166, + "learning_rate": 3e-05, + "loss": 3.9301, + "step": 2572 + }, + { + "epoch": 0.6991847826086957, + "grad_norm": 1.6022355556488037, + "learning_rate": 3e-05, + "loss": 4.0052, + "step": 2573 + }, + { + "epoch": 0.6994565217391304, + "grad_norm": 1.4133738279342651, + "learning_rate": 3e-05, + "loss": 3.6391, + "step": 2574 + }, + { + "epoch": 0.6997282608695652, + "grad_norm": 1.2311711311340332, + "learning_rate": 3e-05, + "loss": 3.6895, + "step": 2575 + }, + { + "epoch": 0.7, + "grad_norm": 1.0733444690704346, + "learning_rate": 3e-05, + "loss": 2.685, + "step": 2576 + }, + { + "epoch": 0.7002717391304348, + "grad_norm": 1.596375823020935, + "learning_rate": 3e-05, + "loss": 3.2422, + "step": 2577 + }, + { + "epoch": 0.7005434782608696, + "grad_norm": 1.098339319229126, + "learning_rate": 3e-05, + "loss": 3.0258, + "step": 2578 + }, + { + "epoch": 0.7008152173913044, + "grad_norm": 1.215551495552063, + "learning_rate": 3e-05, + "loss": 3.6583, + "step": 2579 + }, + { + "epoch": 0.7010869565217391, + "grad_norm": 1.175678014755249, + "learning_rate": 3e-05, + "loss": 2.9082, + "step": 2580 + }, + { + "epoch": 0.7013586956521739, + "grad_norm": 1.285050868988037, + "learning_rate": 3e-05, + "loss": 3.0373, + "step": 2581 + }, + { + "epoch": 0.7016304347826087, + "grad_norm": 1.0280994176864624, + "learning_rate": 3e-05, + "loss": 3.027, + "step": 2582 + }, + { + "epoch": 0.7019021739130434, + "grad_norm": 1.1611984968185425, + "learning_rate": 3e-05, + "loss": 2.9426, + "step": 2583 + }, + { + "epoch": 0.7021739130434783, + "grad_norm": 1.475341558456421, + "learning_rate": 3e-05, + "loss": 3.366, + "step": 2584 + }, + { + "epoch": 0.7024456521739131, + "grad_norm": 1.6377277374267578, + "learning_rate": 3e-05, + "loss": 3.4074, + "step": 2585 + }, + { + "epoch": 0.7027173913043478, + "grad_norm": 1.4452388286590576, + "learning_rate": 3e-05, + "loss": 2.9369, + "step": 2586 + }, + { + "epoch": 0.7029891304347826, + "grad_norm": 1.533266305923462, + "learning_rate": 3e-05, + "loss": 3.9352, + "step": 2587 + }, + { + "epoch": 0.7032608695652174, + "grad_norm": 1.6853458881378174, + "learning_rate": 3e-05, + "loss": 3.4523, + "step": 2588 + }, + { + "epoch": 0.7035326086956522, + "grad_norm": 1.3552666902542114, + "learning_rate": 3e-05, + "loss": 2.7867, + "step": 2589 + }, + { + "epoch": 0.7038043478260869, + "grad_norm": 1.5494905710220337, + "learning_rate": 3e-05, + "loss": 3.4699, + "step": 2590 + }, + { + "epoch": 0.7040760869565217, + "grad_norm": 1.74055016040802, + "learning_rate": 3e-05, + "loss": 3.3874, + "step": 2591 + }, + { + "epoch": 0.7043478260869566, + "grad_norm": 1.250354290008545, + "learning_rate": 3e-05, + "loss": 3.1795, + "step": 2592 + }, + { + "epoch": 0.7046195652173913, + "grad_norm": 1.302441954612732, + "learning_rate": 3e-05, + "loss": 3.9465, + "step": 2593 + }, + { + "epoch": 0.7048913043478261, + "grad_norm": 1.7971372604370117, + "learning_rate": 3e-05, + "loss": 3.6084, + "step": 2594 + }, + { + "epoch": 0.7051630434782609, + "grad_norm": 1.188605546951294, + "learning_rate": 3e-05, + "loss": 3.6716, + "step": 2595 + }, + { + "epoch": 0.7054347826086956, + "grad_norm": 0.9954262971878052, + "learning_rate": 3e-05, + "loss": 3.1982, + "step": 2596 + }, + { + "epoch": 0.7057065217391304, + "grad_norm": 1.174482822418213, + "learning_rate": 3e-05, + "loss": 3.1708, + "step": 2597 + }, + { + "epoch": 0.7059782608695652, + "grad_norm": 1.039506435394287, + "learning_rate": 3e-05, + "loss": 3.1857, + "step": 2598 + }, + { + "epoch": 0.70625, + "grad_norm": 0.9666784405708313, + "learning_rate": 3e-05, + "loss": 3.6005, + "step": 2599 + }, + { + "epoch": 0.7065217391304348, + "grad_norm": 0.921465277671814, + "learning_rate": 3e-05, + "loss": 3.0565, + "step": 2600 + }, + { + "epoch": 0.7067934782608696, + "grad_norm": 1.018864393234253, + "learning_rate": 3e-05, + "loss": 2.7895, + "step": 2601 + }, + { + "epoch": 0.7070652173913043, + "grad_norm": 1.0360243320465088, + "learning_rate": 3e-05, + "loss": 3.125, + "step": 2602 + }, + { + "epoch": 0.7073369565217391, + "grad_norm": 0.9393570423126221, + "learning_rate": 3e-05, + "loss": 3.1148, + "step": 2603 + }, + { + "epoch": 0.7076086956521739, + "grad_norm": 1.1420108079910278, + "learning_rate": 3e-05, + "loss": 3.1415, + "step": 2604 + }, + { + "epoch": 0.7078804347826086, + "grad_norm": 1.1641017198562622, + "learning_rate": 3e-05, + "loss": 3.0936, + "step": 2605 + }, + { + "epoch": 0.7081521739130435, + "grad_norm": 1.6107722520828247, + "learning_rate": 3e-05, + "loss": 4.061, + "step": 2606 + }, + { + "epoch": 0.7084239130434783, + "grad_norm": 1.2036430835723877, + "learning_rate": 3e-05, + "loss": 3.1156, + "step": 2607 + }, + { + "epoch": 0.7086956521739131, + "grad_norm": 1.2490811347961426, + "learning_rate": 3e-05, + "loss": 3.0078, + "step": 2608 + }, + { + "epoch": 0.7089673913043478, + "grad_norm": 1.0128928422927856, + "learning_rate": 3e-05, + "loss": 2.8499, + "step": 2609 + }, + { + "epoch": 0.7092391304347826, + "grad_norm": 1.363193154335022, + "learning_rate": 3e-05, + "loss": 3.9375, + "step": 2610 + }, + { + "epoch": 0.7095108695652174, + "grad_norm": 0.8343818187713623, + "learning_rate": 3e-05, + "loss": 3.1519, + "step": 2611 + }, + { + "epoch": 0.7097826086956521, + "grad_norm": 1.0047922134399414, + "learning_rate": 3e-05, + "loss": 2.8608, + "step": 2612 + }, + { + "epoch": 0.710054347826087, + "grad_norm": 1.18521249294281, + "learning_rate": 3e-05, + "loss": 3.7399, + "step": 2613 + }, + { + "epoch": 0.7103260869565218, + "grad_norm": 1.3591742515563965, + "learning_rate": 3e-05, + "loss": 3.5885, + "step": 2614 + }, + { + "epoch": 0.7105978260869565, + "grad_norm": 1.1936426162719727, + "learning_rate": 3e-05, + "loss": 3.7192, + "step": 2615 + }, + { + "epoch": 0.7108695652173913, + "grad_norm": 1.114531397819519, + "learning_rate": 3e-05, + "loss": 3.22, + "step": 2616 + }, + { + "epoch": 0.7111413043478261, + "grad_norm": 1.1929512023925781, + "learning_rate": 3e-05, + "loss": 2.751, + "step": 2617 + }, + { + "epoch": 0.7114130434782608, + "grad_norm": 1.3775954246520996, + "learning_rate": 3e-05, + "loss": 3.6234, + "step": 2618 + }, + { + "epoch": 0.7116847826086956, + "grad_norm": 1.313889980316162, + "learning_rate": 3e-05, + "loss": 3.2312, + "step": 2619 + }, + { + "epoch": 0.7119565217391305, + "grad_norm": 1.4679901599884033, + "learning_rate": 3e-05, + "loss": 3.2006, + "step": 2620 + }, + { + "epoch": 0.7122282608695653, + "grad_norm": 0.9835330843925476, + "learning_rate": 3e-05, + "loss": 3.2365, + "step": 2621 + }, + { + "epoch": 0.7125, + "grad_norm": 1.9934345483779907, + "learning_rate": 3e-05, + "loss": 4.4249, + "step": 2622 + }, + { + "epoch": 0.7127717391304348, + "grad_norm": 1.2943999767303467, + "learning_rate": 3e-05, + "loss": 3.517, + "step": 2623 + }, + { + "epoch": 0.7130434782608696, + "grad_norm": 1.1526103019714355, + "learning_rate": 3e-05, + "loss": 3.6404, + "step": 2624 + }, + { + "epoch": 0.7133152173913043, + "grad_norm": 1.3137702941894531, + "learning_rate": 3e-05, + "loss": 3.305, + "step": 2625 + }, + { + "epoch": 0.7135869565217391, + "grad_norm": 1.2102550268173218, + "learning_rate": 3e-05, + "loss": 3.0545, + "step": 2626 + }, + { + "epoch": 0.7138586956521739, + "grad_norm": 1.0959724187850952, + "learning_rate": 3e-05, + "loss": 3.0383, + "step": 2627 + }, + { + "epoch": 0.7141304347826087, + "grad_norm": 1.7450772523880005, + "learning_rate": 3e-05, + "loss": 3.5891, + "step": 2628 + }, + { + "epoch": 0.7144021739130435, + "grad_norm": 1.3901242017745972, + "learning_rate": 3e-05, + "loss": 3.5405, + "step": 2629 + }, + { + "epoch": 0.7146739130434783, + "grad_norm": 1.2243647575378418, + "learning_rate": 3e-05, + "loss": 3.4307, + "step": 2630 + }, + { + "epoch": 0.714945652173913, + "grad_norm": 2.3953914642333984, + "learning_rate": 3e-05, + "loss": 4.3341, + "step": 2631 + }, + { + "epoch": 0.7152173913043478, + "grad_norm": 1.8682715892791748, + "learning_rate": 3e-05, + "loss": 3.0704, + "step": 2632 + }, + { + "epoch": 0.7154891304347826, + "grad_norm": 1.4622515439987183, + "learning_rate": 3e-05, + "loss": 3.3376, + "step": 2633 + }, + { + "epoch": 0.7157608695652173, + "grad_norm": 1.2892646789550781, + "learning_rate": 3e-05, + "loss": 3.312, + "step": 2634 + }, + { + "epoch": 0.7160326086956522, + "grad_norm": 1.4950429201126099, + "learning_rate": 3e-05, + "loss": 3.3923, + "step": 2635 + }, + { + "epoch": 0.716304347826087, + "grad_norm": 1.380149245262146, + "learning_rate": 3e-05, + "loss": 3.5931, + "step": 2636 + }, + { + "epoch": 0.7165760869565218, + "grad_norm": 1.433959722518921, + "learning_rate": 3e-05, + "loss": 2.9042, + "step": 2637 + }, + { + "epoch": 0.7168478260869565, + "grad_norm": 1.2683796882629395, + "learning_rate": 3e-05, + "loss": 3.3012, + "step": 2638 + }, + { + "epoch": 0.7171195652173913, + "grad_norm": 1.0387176275253296, + "learning_rate": 3e-05, + "loss": 2.9117, + "step": 2639 + }, + { + "epoch": 0.717391304347826, + "grad_norm": 1.5096675157546997, + "learning_rate": 3e-05, + "loss": 3.125, + "step": 2640 + }, + { + "epoch": 0.7176630434782608, + "grad_norm": 1.1815725564956665, + "learning_rate": 3e-05, + "loss": 2.9721, + "step": 2641 + }, + { + "epoch": 0.7179347826086957, + "grad_norm": 1.1765364408493042, + "learning_rate": 3e-05, + "loss": 3.3088, + "step": 2642 + }, + { + "epoch": 0.7182065217391305, + "grad_norm": 1.087518572807312, + "learning_rate": 3e-05, + "loss": 3.1549, + "step": 2643 + }, + { + "epoch": 0.7184782608695652, + "grad_norm": 1.1884835958480835, + "learning_rate": 3e-05, + "loss": 3.2079, + "step": 2644 + }, + { + "epoch": 0.71875, + "grad_norm": 1.1198290586471558, + "learning_rate": 3e-05, + "loss": 2.9726, + "step": 2645 + }, + { + "epoch": 0.7190217391304348, + "grad_norm": 1.2286672592163086, + "learning_rate": 3e-05, + "loss": 3.2783, + "step": 2646 + }, + { + "epoch": 0.7192934782608695, + "grad_norm": 1.3148912191390991, + "learning_rate": 3e-05, + "loss": 3.914, + "step": 2647 + }, + { + "epoch": 0.7195652173913043, + "grad_norm": 1.1227190494537354, + "learning_rate": 3e-05, + "loss": 3.1269, + "step": 2648 + }, + { + "epoch": 0.7198369565217392, + "grad_norm": 1.3971885442733765, + "learning_rate": 3e-05, + "loss": 3.3385, + "step": 2649 + }, + { + "epoch": 0.720108695652174, + "grad_norm": 1.0572166442871094, + "learning_rate": 3e-05, + "loss": 2.8119, + "step": 2650 + }, + { + "epoch": 0.7203804347826087, + "grad_norm": 1.3299329280853271, + "learning_rate": 3e-05, + "loss": 3.5842, + "step": 2651 + }, + { + "epoch": 0.7206521739130435, + "grad_norm": 1.2632654905319214, + "learning_rate": 3e-05, + "loss": 2.8276, + "step": 2652 + }, + { + "epoch": 0.7209239130434782, + "grad_norm": 1.2086924314498901, + "learning_rate": 3e-05, + "loss": 3.3622, + "step": 2653 + }, + { + "epoch": 0.721195652173913, + "grad_norm": 1.1867645978927612, + "learning_rate": 3e-05, + "loss": 3.1258, + "step": 2654 + }, + { + "epoch": 0.7214673913043478, + "grad_norm": 1.123406171798706, + "learning_rate": 3e-05, + "loss": 3.3561, + "step": 2655 + }, + { + "epoch": 0.7217391304347827, + "grad_norm": 1.3678390979766846, + "learning_rate": 3e-05, + "loss": 3.4183, + "step": 2656 + }, + { + "epoch": 0.7220108695652174, + "grad_norm": 1.21256685256958, + "learning_rate": 3e-05, + "loss": 3.5207, + "step": 2657 + }, + { + "epoch": 0.7222826086956522, + "grad_norm": 1.1133040189743042, + "learning_rate": 3e-05, + "loss": 2.8218, + "step": 2658 + }, + { + "epoch": 0.722554347826087, + "grad_norm": 1.0846513509750366, + "learning_rate": 3e-05, + "loss": 2.9519, + "step": 2659 + }, + { + "epoch": 0.7228260869565217, + "grad_norm": 1.0870620012283325, + "learning_rate": 3e-05, + "loss": 2.8142, + "step": 2660 + }, + { + "epoch": 0.7230978260869565, + "grad_norm": 1.2407505512237549, + "learning_rate": 3e-05, + "loss": 3.0965, + "step": 2661 + }, + { + "epoch": 0.7233695652173913, + "grad_norm": 1.0677931308746338, + "learning_rate": 3e-05, + "loss": 2.841, + "step": 2662 + }, + { + "epoch": 0.7236413043478261, + "grad_norm": 1.1866375207901, + "learning_rate": 3e-05, + "loss": 3.2199, + "step": 2663 + }, + { + "epoch": 0.7239130434782609, + "grad_norm": 1.4401320219039917, + "learning_rate": 3e-05, + "loss": 4.1391, + "step": 2664 + }, + { + "epoch": 0.7241847826086957, + "grad_norm": 1.8525235652923584, + "learning_rate": 3e-05, + "loss": 2.954, + "step": 2665 + }, + { + "epoch": 0.7244565217391304, + "grad_norm": 1.4307048320770264, + "learning_rate": 3e-05, + "loss": 3.6121, + "step": 2666 + }, + { + "epoch": 0.7247282608695652, + "grad_norm": 1.4132686853408813, + "learning_rate": 3e-05, + "loss": 3.2657, + "step": 2667 + }, + { + "epoch": 0.725, + "grad_norm": 1.1710587739944458, + "learning_rate": 3e-05, + "loss": 3.5075, + "step": 2668 + }, + { + "epoch": 0.7252717391304347, + "grad_norm": 1.4402532577514648, + "learning_rate": 3e-05, + "loss": 3.5403, + "step": 2669 + }, + { + "epoch": 0.7255434782608695, + "grad_norm": 1.0295389890670776, + "learning_rate": 3e-05, + "loss": 3.2313, + "step": 2670 + }, + { + "epoch": 0.7258152173913044, + "grad_norm": 1.9836853742599487, + "learning_rate": 3e-05, + "loss": 3.7605, + "step": 2671 + }, + { + "epoch": 0.7260869565217392, + "grad_norm": 1.313320279121399, + "learning_rate": 3e-05, + "loss": 2.892, + "step": 2672 + }, + { + "epoch": 0.7263586956521739, + "grad_norm": 1.227516770362854, + "learning_rate": 3e-05, + "loss": 3.4072, + "step": 2673 + }, + { + "epoch": 0.7266304347826087, + "grad_norm": 1.343525767326355, + "learning_rate": 3e-05, + "loss": 3.1093, + "step": 2674 + }, + { + "epoch": 0.7269021739130435, + "grad_norm": 1.6472978591918945, + "learning_rate": 3e-05, + "loss": 3.2588, + "step": 2675 + }, + { + "epoch": 0.7271739130434782, + "grad_norm": 1.313114047050476, + "learning_rate": 3e-05, + "loss": 3.0819, + "step": 2676 + }, + { + "epoch": 0.727445652173913, + "grad_norm": 1.3546169996261597, + "learning_rate": 3e-05, + "loss": 3.3884, + "step": 2677 + }, + { + "epoch": 0.7277173913043479, + "grad_norm": 1.4308792352676392, + "learning_rate": 3e-05, + "loss": 3.2228, + "step": 2678 + }, + { + "epoch": 0.7279891304347826, + "grad_norm": 1.2303006649017334, + "learning_rate": 3e-05, + "loss": 2.9464, + "step": 2679 + }, + { + "epoch": 0.7282608695652174, + "grad_norm": 2.12964129447937, + "learning_rate": 3e-05, + "loss": 4.1515, + "step": 2680 + }, + { + "epoch": 0.7285326086956522, + "grad_norm": 1.1187695264816284, + "learning_rate": 3e-05, + "loss": 3.2245, + "step": 2681 + }, + { + "epoch": 0.7288043478260869, + "grad_norm": 1.440588355064392, + "learning_rate": 3e-05, + "loss": 4.0875, + "step": 2682 + }, + { + "epoch": 0.7290760869565217, + "grad_norm": 1.2928602695465088, + "learning_rate": 3e-05, + "loss": 3.4652, + "step": 2683 + }, + { + "epoch": 0.7293478260869565, + "grad_norm": 1.1873844861984253, + "learning_rate": 3e-05, + "loss": 2.8352, + "step": 2684 + }, + { + "epoch": 0.7296195652173914, + "grad_norm": 1.1879850625991821, + "learning_rate": 3e-05, + "loss": 2.8956, + "step": 2685 + }, + { + "epoch": 0.7298913043478261, + "grad_norm": 1.206374168395996, + "learning_rate": 3e-05, + "loss": 3.3039, + "step": 2686 + }, + { + "epoch": 0.7301630434782609, + "grad_norm": 1.1491773128509521, + "learning_rate": 3e-05, + "loss": 3.0314, + "step": 2687 + }, + { + "epoch": 0.7304347826086957, + "grad_norm": 1.3866368532180786, + "learning_rate": 3e-05, + "loss": 3.055, + "step": 2688 + }, + { + "epoch": 0.7307065217391304, + "grad_norm": 1.5674668550491333, + "learning_rate": 3e-05, + "loss": 3.1808, + "step": 2689 + }, + { + "epoch": 0.7309782608695652, + "grad_norm": 1.562813639640808, + "learning_rate": 3e-05, + "loss": 3.2706, + "step": 2690 + }, + { + "epoch": 0.73125, + "grad_norm": 1.6883511543273926, + "learning_rate": 3e-05, + "loss": 3.2192, + "step": 2691 + }, + { + "epoch": 0.7315217391304348, + "grad_norm": 1.1254132986068726, + "learning_rate": 3e-05, + "loss": 3.1529, + "step": 2692 + }, + { + "epoch": 0.7317934782608696, + "grad_norm": 1.0081984996795654, + "learning_rate": 3e-05, + "loss": 3.0864, + "step": 2693 + }, + { + "epoch": 0.7320652173913044, + "grad_norm": 1.453481674194336, + "learning_rate": 3e-05, + "loss": 3.2169, + "step": 2694 + }, + { + "epoch": 0.7323369565217391, + "grad_norm": 1.3893227577209473, + "learning_rate": 3e-05, + "loss": 3.3761, + "step": 2695 + }, + { + "epoch": 0.7326086956521739, + "grad_norm": 1.7900173664093018, + "learning_rate": 3e-05, + "loss": 3.4638, + "step": 2696 + }, + { + "epoch": 0.7328804347826087, + "grad_norm": 1.1031737327575684, + "learning_rate": 3e-05, + "loss": 2.6962, + "step": 2697 + }, + { + "epoch": 0.7331521739130434, + "grad_norm": 1.3984167575836182, + "learning_rate": 3e-05, + "loss": 3.3052, + "step": 2698 + }, + { + "epoch": 0.7334239130434783, + "grad_norm": 1.6182535886764526, + "learning_rate": 3e-05, + "loss": 3.4213, + "step": 2699 + }, + { + "epoch": 0.7336956521739131, + "grad_norm": 1.7064732313156128, + "learning_rate": 3e-05, + "loss": 4.0381, + "step": 2700 + }, + { + "epoch": 0.7339673913043478, + "grad_norm": 1.357704520225525, + "learning_rate": 3e-05, + "loss": 3.6925, + "step": 2701 + }, + { + "epoch": 0.7342391304347826, + "grad_norm": 1.6365474462509155, + "learning_rate": 3e-05, + "loss": 3.2891, + "step": 2702 + }, + { + "epoch": 0.7345108695652174, + "grad_norm": 1.8574589490890503, + "learning_rate": 3e-05, + "loss": 3.3637, + "step": 2703 + }, + { + "epoch": 0.7347826086956522, + "grad_norm": 1.3467528820037842, + "learning_rate": 3e-05, + "loss": 3.1417, + "step": 2704 + }, + { + "epoch": 0.7350543478260869, + "grad_norm": 1.4658093452453613, + "learning_rate": 3e-05, + "loss": 3.3401, + "step": 2705 + }, + { + "epoch": 0.7353260869565217, + "grad_norm": 1.2771729230880737, + "learning_rate": 3e-05, + "loss": 3.3731, + "step": 2706 + }, + { + "epoch": 0.7355978260869566, + "grad_norm": 1.0597422122955322, + "learning_rate": 3e-05, + "loss": 3.1315, + "step": 2707 + }, + { + "epoch": 0.7358695652173913, + "grad_norm": 1.1597671508789062, + "learning_rate": 3e-05, + "loss": 2.6791, + "step": 2708 + }, + { + "epoch": 0.7361413043478261, + "grad_norm": 1.7157272100448608, + "learning_rate": 3e-05, + "loss": 3.8816, + "step": 2709 + }, + { + "epoch": 0.7364130434782609, + "grad_norm": 1.3123220205307007, + "learning_rate": 3e-05, + "loss": 3.6255, + "step": 2710 + }, + { + "epoch": 0.7366847826086956, + "grad_norm": 1.2993775606155396, + "learning_rate": 3e-05, + "loss": 3.6012, + "step": 2711 + }, + { + "epoch": 0.7369565217391304, + "grad_norm": 1.0929709672927856, + "learning_rate": 3e-05, + "loss": 3.3147, + "step": 2712 + }, + { + "epoch": 0.7372282608695652, + "grad_norm": 1.2103468179702759, + "learning_rate": 3e-05, + "loss": 3.3148, + "step": 2713 + }, + { + "epoch": 0.7375, + "grad_norm": 0.9965810179710388, + "learning_rate": 3e-05, + "loss": 3.2745, + "step": 2714 + }, + { + "epoch": 0.7377717391304348, + "grad_norm": 1.7804696559906006, + "learning_rate": 3e-05, + "loss": 3.6753, + "step": 2715 + }, + { + "epoch": 0.7380434782608696, + "grad_norm": 1.2880659103393555, + "learning_rate": 3e-05, + "loss": 3.4252, + "step": 2716 + }, + { + "epoch": 0.7383152173913043, + "grad_norm": 0.9781152606010437, + "learning_rate": 3e-05, + "loss": 2.9325, + "step": 2717 + }, + { + "epoch": 0.7385869565217391, + "grad_norm": 1.724921464920044, + "learning_rate": 3e-05, + "loss": 3.6383, + "step": 2718 + }, + { + "epoch": 0.7388586956521739, + "grad_norm": 1.3509818315505981, + "learning_rate": 3e-05, + "loss": 3.2334, + "step": 2719 + }, + { + "epoch": 0.7391304347826086, + "grad_norm": 1.2037723064422607, + "learning_rate": 3e-05, + "loss": 3.242, + "step": 2720 + }, + { + "epoch": 0.7394021739130435, + "grad_norm": 1.337058424949646, + "learning_rate": 3e-05, + "loss": 3.5805, + "step": 2721 + }, + { + "epoch": 0.7396739130434783, + "grad_norm": 1.3748770952224731, + "learning_rate": 3e-05, + "loss": 2.9626, + "step": 2722 + }, + { + "epoch": 0.7399456521739131, + "grad_norm": 1.3539892435073853, + "learning_rate": 3e-05, + "loss": 3.0298, + "step": 2723 + }, + { + "epoch": 0.7402173913043478, + "grad_norm": 0.979083776473999, + "learning_rate": 3e-05, + "loss": 2.7295, + "step": 2724 + }, + { + "epoch": 0.7404891304347826, + "grad_norm": 1.7237200736999512, + "learning_rate": 3e-05, + "loss": 3.5329, + "step": 2725 + }, + { + "epoch": 0.7407608695652174, + "grad_norm": 1.4173531532287598, + "learning_rate": 3e-05, + "loss": 3.5202, + "step": 2726 + }, + { + "epoch": 0.7410326086956521, + "grad_norm": 1.3420840501785278, + "learning_rate": 3e-05, + "loss": 3.3239, + "step": 2727 + }, + { + "epoch": 0.741304347826087, + "grad_norm": 1.3799097537994385, + "learning_rate": 3e-05, + "loss": 3.4807, + "step": 2728 + }, + { + "epoch": 0.7415760869565218, + "grad_norm": 1.2215850353240967, + "learning_rate": 3e-05, + "loss": 3.5144, + "step": 2729 + }, + { + "epoch": 0.7418478260869565, + "grad_norm": 1.2649718523025513, + "learning_rate": 3e-05, + "loss": 3.4898, + "step": 2730 + }, + { + "epoch": 0.7421195652173913, + "grad_norm": 1.1176096200942993, + "learning_rate": 3e-05, + "loss": 3.4977, + "step": 2731 + }, + { + "epoch": 0.7423913043478261, + "grad_norm": 1.0005884170532227, + "learning_rate": 3e-05, + "loss": 3.1848, + "step": 2732 + }, + { + "epoch": 0.7426630434782608, + "grad_norm": 1.6306564807891846, + "learning_rate": 3e-05, + "loss": 3.8107, + "step": 2733 + }, + { + "epoch": 0.7429347826086956, + "grad_norm": 1.6788973808288574, + "learning_rate": 3e-05, + "loss": 3.985, + "step": 2734 + }, + { + "epoch": 0.7432065217391305, + "grad_norm": 1.16837739944458, + "learning_rate": 3e-05, + "loss": 2.8648, + "step": 2735 + }, + { + "epoch": 0.7434782608695653, + "grad_norm": 1.324316382408142, + "learning_rate": 3e-05, + "loss": 3.5445, + "step": 2736 + }, + { + "epoch": 0.74375, + "grad_norm": 1.2631101608276367, + "learning_rate": 3e-05, + "loss": 3.2311, + "step": 2737 + }, + { + "epoch": 0.7440217391304348, + "grad_norm": 1.030339241027832, + "learning_rate": 3e-05, + "loss": 2.8981, + "step": 2738 + }, + { + "epoch": 0.7442934782608696, + "grad_norm": 1.5260175466537476, + "learning_rate": 3e-05, + "loss": 3.2335, + "step": 2739 + }, + { + "epoch": 0.7445652173913043, + "grad_norm": 1.610879898071289, + "learning_rate": 3e-05, + "loss": 3.3666, + "step": 2740 + }, + { + "epoch": 0.7448369565217391, + "grad_norm": 1.290809988975525, + "learning_rate": 3e-05, + "loss": 2.9172, + "step": 2741 + }, + { + "epoch": 0.7451086956521739, + "grad_norm": 1.2643457651138306, + "learning_rate": 3e-05, + "loss": 3.1168, + "step": 2742 + }, + { + "epoch": 0.7453804347826087, + "grad_norm": 1.2268829345703125, + "learning_rate": 3e-05, + "loss": 3.2188, + "step": 2743 + }, + { + "epoch": 0.7456521739130435, + "grad_norm": 1.022884726524353, + "learning_rate": 3e-05, + "loss": 2.884, + "step": 2744 + }, + { + "epoch": 0.7459239130434783, + "grad_norm": 1.9909472465515137, + "learning_rate": 3e-05, + "loss": 4.1557, + "step": 2745 + }, + { + "epoch": 0.746195652173913, + "grad_norm": 2.2413318157196045, + "learning_rate": 3e-05, + "loss": 4.0246, + "step": 2746 + }, + { + "epoch": 0.7464673913043478, + "grad_norm": 1.0311791896820068, + "learning_rate": 3e-05, + "loss": 2.8403, + "step": 2747 + }, + { + "epoch": 0.7467391304347826, + "grad_norm": 1.0483710765838623, + "learning_rate": 3e-05, + "loss": 2.7621, + "step": 2748 + }, + { + "epoch": 0.7470108695652173, + "grad_norm": 0.9712722897529602, + "learning_rate": 3e-05, + "loss": 2.7713, + "step": 2749 + }, + { + "epoch": 0.7472826086956522, + "grad_norm": 1.5642272233963013, + "learning_rate": 3e-05, + "loss": 3.2916, + "step": 2750 + }, + { + "epoch": 0.747554347826087, + "grad_norm": 1.0362199544906616, + "learning_rate": 3e-05, + "loss": 3.0884, + "step": 2751 + }, + { + "epoch": 0.7478260869565218, + "grad_norm": 1.2575013637542725, + "learning_rate": 3e-05, + "loss": 3.3603, + "step": 2752 + }, + { + "epoch": 0.7480978260869565, + "grad_norm": 1.2161589860916138, + "learning_rate": 3e-05, + "loss": 3.3827, + "step": 2753 + }, + { + "epoch": 0.7483695652173913, + "grad_norm": 1.0151880979537964, + "learning_rate": 3e-05, + "loss": 2.9711, + "step": 2754 + }, + { + "epoch": 0.748641304347826, + "grad_norm": 1.3773422241210938, + "learning_rate": 3e-05, + "loss": 3.9046, + "step": 2755 + }, + { + "epoch": 0.7489130434782608, + "grad_norm": 0.8108665943145752, + "learning_rate": 3e-05, + "loss": 2.9052, + "step": 2756 + }, + { + "epoch": 0.7491847826086957, + "grad_norm": 0.8243381381034851, + "learning_rate": 3e-05, + "loss": 2.8174, + "step": 2757 + }, + { + "epoch": 0.7494565217391305, + "grad_norm": 1.3421036005020142, + "learning_rate": 3e-05, + "loss": 3.3869, + "step": 2758 + }, + { + "epoch": 0.7497282608695652, + "grad_norm": 1.1268045902252197, + "learning_rate": 3e-05, + "loss": 3.4369, + "step": 2759 + }, + { + "epoch": 0.75, + "grad_norm": 0.982552707195282, + "learning_rate": 3e-05, + "loss": 3.0077, + "step": 2760 + }, + { + "epoch": 0.7502717391304348, + "grad_norm": 1.548680305480957, + "learning_rate": 3e-05, + "loss": 2.9879, + "step": 2761 + }, + { + "epoch": 0.7505434782608695, + "grad_norm": 1.1556179523468018, + "learning_rate": 3e-05, + "loss": 2.9654, + "step": 2762 + }, + { + "epoch": 0.7508152173913043, + "grad_norm": 1.310678482055664, + "learning_rate": 3e-05, + "loss": 3.6287, + "step": 2763 + }, + { + "epoch": 0.7510869565217392, + "grad_norm": 1.4762641191482544, + "learning_rate": 3e-05, + "loss": 3.6961, + "step": 2764 + }, + { + "epoch": 0.751358695652174, + "grad_norm": 1.3265812397003174, + "learning_rate": 3e-05, + "loss": 3.2219, + "step": 2765 + }, + { + "epoch": 0.7516304347826087, + "grad_norm": 1.2346985340118408, + "learning_rate": 3e-05, + "loss": 3.259, + "step": 2766 + }, + { + "epoch": 0.7519021739130435, + "grad_norm": 0.9740216135978699, + "learning_rate": 3e-05, + "loss": 2.854, + "step": 2767 + }, + { + "epoch": 0.7521739130434782, + "grad_norm": 1.2616466283798218, + "learning_rate": 3e-05, + "loss": 2.9113, + "step": 2768 + }, + { + "epoch": 0.752445652173913, + "grad_norm": 1.4515007734298706, + "learning_rate": 3e-05, + "loss": 4.4309, + "step": 2769 + }, + { + "epoch": 0.7527173913043478, + "grad_norm": 1.2112640142440796, + "learning_rate": 3e-05, + "loss": 3.1847, + "step": 2770 + }, + { + "epoch": 0.7529891304347827, + "grad_norm": 1.123841643333435, + "learning_rate": 3e-05, + "loss": 3.3345, + "step": 2771 + }, + { + "epoch": 0.7532608695652174, + "grad_norm": 1.1541368961334229, + "learning_rate": 3e-05, + "loss": 2.8234, + "step": 2772 + }, + { + "epoch": 0.7535326086956522, + "grad_norm": 1.0209699869155884, + "learning_rate": 3e-05, + "loss": 3.1726, + "step": 2773 + }, + { + "epoch": 0.753804347826087, + "grad_norm": 1.2185944318771362, + "learning_rate": 3e-05, + "loss": 3.8311, + "step": 2774 + }, + { + "epoch": 0.7540760869565217, + "grad_norm": 1.553025722503662, + "learning_rate": 3e-05, + "loss": 3.2896, + "step": 2775 + }, + { + "epoch": 0.7543478260869565, + "grad_norm": 1.5103728771209717, + "learning_rate": 3e-05, + "loss": 3.0133, + "step": 2776 + }, + { + "epoch": 0.7546195652173913, + "grad_norm": 1.0733397006988525, + "learning_rate": 3e-05, + "loss": 3.0122, + "step": 2777 + }, + { + "epoch": 0.7548913043478261, + "grad_norm": 1.4341579675674438, + "learning_rate": 3e-05, + "loss": 3.5646, + "step": 2778 + }, + { + "epoch": 0.7551630434782609, + "grad_norm": 1.6829897165298462, + "learning_rate": 3e-05, + "loss": 3.6338, + "step": 2779 + }, + { + "epoch": 0.7554347826086957, + "grad_norm": 1.0816172361373901, + "learning_rate": 3e-05, + "loss": 3.1679, + "step": 2780 + }, + { + "epoch": 0.7557065217391304, + "grad_norm": 1.3089653253555298, + "learning_rate": 3e-05, + "loss": 3.6569, + "step": 2781 + }, + { + "epoch": 0.7559782608695652, + "grad_norm": 1.329774022102356, + "learning_rate": 3e-05, + "loss": 3.0589, + "step": 2782 + }, + { + "epoch": 0.75625, + "grad_norm": 1.5232934951782227, + "learning_rate": 3e-05, + "loss": 3.5649, + "step": 2783 + }, + { + "epoch": 0.7565217391304347, + "grad_norm": 1.501591444015503, + "learning_rate": 3e-05, + "loss": 3.7119, + "step": 2784 + }, + { + "epoch": 0.7567934782608695, + "grad_norm": 0.89922696352005, + "learning_rate": 3e-05, + "loss": 2.9755, + "step": 2785 + }, + { + "epoch": 0.7570652173913044, + "grad_norm": 1.0538735389709473, + "learning_rate": 3e-05, + "loss": 3.4823, + "step": 2786 + }, + { + "epoch": 0.7573369565217392, + "grad_norm": 0.8761924505233765, + "learning_rate": 3e-05, + "loss": 3.1571, + "step": 2787 + }, + { + "epoch": 0.7576086956521739, + "grad_norm": 1.2133067846298218, + "learning_rate": 3e-05, + "loss": 3.3668, + "step": 2788 + }, + { + "epoch": 0.7578804347826087, + "grad_norm": 1.0436729192733765, + "learning_rate": 3e-05, + "loss": 3.338, + "step": 2789 + }, + { + "epoch": 0.7581521739130435, + "grad_norm": 1.0040472745895386, + "learning_rate": 3e-05, + "loss": 3.314, + "step": 2790 + }, + { + "epoch": 0.7584239130434782, + "grad_norm": 1.019578456878662, + "learning_rate": 3e-05, + "loss": 3.3716, + "step": 2791 + }, + { + "epoch": 0.758695652173913, + "grad_norm": 1.209367036819458, + "learning_rate": 3e-05, + "loss": 3.7626, + "step": 2792 + }, + { + "epoch": 0.7589673913043479, + "grad_norm": 1.0667492151260376, + "learning_rate": 3e-05, + "loss": 3.4984, + "step": 2793 + }, + { + "epoch": 0.7592391304347826, + "grad_norm": 1.200366497039795, + "learning_rate": 3e-05, + "loss": 3.228, + "step": 2794 + }, + { + "epoch": 0.7595108695652174, + "grad_norm": 1.2726914882659912, + "learning_rate": 3e-05, + "loss": 2.9393, + "step": 2795 + }, + { + "epoch": 0.7597826086956522, + "grad_norm": 1.244170904159546, + "learning_rate": 3e-05, + "loss": 2.9552, + "step": 2796 + }, + { + "epoch": 0.7600543478260869, + "grad_norm": 1.113420844078064, + "learning_rate": 3e-05, + "loss": 3.1033, + "step": 2797 + }, + { + "epoch": 0.7603260869565217, + "grad_norm": 1.9228954315185547, + "learning_rate": 3e-05, + "loss": 3.7877, + "step": 2798 + }, + { + "epoch": 0.7605978260869565, + "grad_norm": 1.1893031597137451, + "learning_rate": 3e-05, + "loss": 3.0852, + "step": 2799 + }, + { + "epoch": 0.7608695652173914, + "grad_norm": 1.524493932723999, + "learning_rate": 3e-05, + "loss": 4.1925, + "step": 2800 + }, + { + "epoch": 0.7611413043478261, + "grad_norm": 1.5600111484527588, + "learning_rate": 3e-05, + "loss": 3.4646, + "step": 2801 + }, + { + "epoch": 0.7614130434782609, + "grad_norm": 0.9608951210975647, + "learning_rate": 3e-05, + "loss": 2.8357, + "step": 2802 + }, + { + "epoch": 0.7616847826086957, + "grad_norm": 1.2598545551300049, + "learning_rate": 3e-05, + "loss": 3.2298, + "step": 2803 + }, + { + "epoch": 0.7619565217391304, + "grad_norm": 1.4729243516921997, + "learning_rate": 3e-05, + "loss": 3.2004, + "step": 2804 + }, + { + "epoch": 0.7622282608695652, + "grad_norm": 1.04935622215271, + "learning_rate": 3e-05, + "loss": 2.9192, + "step": 2805 + }, + { + "epoch": 0.7625, + "grad_norm": 1.1317492723464966, + "learning_rate": 3e-05, + "loss": 3.5182, + "step": 2806 + }, + { + "epoch": 0.7627717391304348, + "grad_norm": 0.8651010394096375, + "learning_rate": 3e-05, + "loss": 2.7038, + "step": 2807 + }, + { + "epoch": 0.7630434782608696, + "grad_norm": 1.4115127325057983, + "learning_rate": 3e-05, + "loss": 3.225, + "step": 2808 + }, + { + "epoch": 0.7633152173913044, + "grad_norm": 1.0244847536087036, + "learning_rate": 3e-05, + "loss": 3.1254, + "step": 2809 + }, + { + "epoch": 0.7635869565217391, + "grad_norm": 1.4117242097854614, + "learning_rate": 3e-05, + "loss": 3.4427, + "step": 2810 + }, + { + "epoch": 0.7638586956521739, + "grad_norm": 1.3557825088500977, + "learning_rate": 3e-05, + "loss": 3.2669, + "step": 2811 + }, + { + "epoch": 0.7641304347826087, + "grad_norm": 1.2034603357315063, + "learning_rate": 3e-05, + "loss": 3.77, + "step": 2812 + }, + { + "epoch": 0.7644021739130434, + "grad_norm": 1.2670207023620605, + "learning_rate": 3e-05, + "loss": 3.1149, + "step": 2813 + }, + { + "epoch": 0.7646739130434783, + "grad_norm": 0.9181430339813232, + "learning_rate": 3e-05, + "loss": 3.0427, + "step": 2814 + }, + { + "epoch": 0.7649456521739131, + "grad_norm": 1.2132922410964966, + "learning_rate": 3e-05, + "loss": 3.4703, + "step": 2815 + }, + { + "epoch": 0.7652173913043478, + "grad_norm": 0.9925717115402222, + "learning_rate": 3e-05, + "loss": 3.1396, + "step": 2816 + }, + { + "epoch": 0.7654891304347826, + "grad_norm": 1.0748741626739502, + "learning_rate": 3e-05, + "loss": 2.8827, + "step": 2817 + }, + { + "epoch": 0.7657608695652174, + "grad_norm": 1.0484018325805664, + "learning_rate": 3e-05, + "loss": 3.1837, + "step": 2818 + }, + { + "epoch": 0.7660326086956522, + "grad_norm": 1.1004911661148071, + "learning_rate": 3e-05, + "loss": 3.135, + "step": 2819 + }, + { + "epoch": 0.7663043478260869, + "grad_norm": 1.3505610227584839, + "learning_rate": 3e-05, + "loss": 3.3415, + "step": 2820 + }, + { + "epoch": 0.7665760869565217, + "grad_norm": 1.1511911153793335, + "learning_rate": 3e-05, + "loss": 3.3407, + "step": 2821 + }, + { + "epoch": 0.7668478260869566, + "grad_norm": 1.2488864660263062, + "learning_rate": 3e-05, + "loss": 3.0697, + "step": 2822 + }, + { + "epoch": 0.7671195652173913, + "grad_norm": 0.970241129398346, + "learning_rate": 3e-05, + "loss": 2.6801, + "step": 2823 + }, + { + "epoch": 0.7673913043478261, + "grad_norm": 1.3036609888076782, + "learning_rate": 3e-05, + "loss": 3.7613, + "step": 2824 + }, + { + "epoch": 0.7676630434782609, + "grad_norm": 1.1077839136123657, + "learning_rate": 3e-05, + "loss": 3.829, + "step": 2825 + }, + { + "epoch": 0.7679347826086956, + "grad_norm": 1.1794723272323608, + "learning_rate": 3e-05, + "loss": 3.1147, + "step": 2826 + }, + { + "epoch": 0.7682065217391304, + "grad_norm": 1.460880994796753, + "learning_rate": 3e-05, + "loss": 3.7514, + "step": 2827 + }, + { + "epoch": 0.7684782608695652, + "grad_norm": 1.2111482620239258, + "learning_rate": 3e-05, + "loss": 3.5524, + "step": 2828 + }, + { + "epoch": 0.76875, + "grad_norm": 1.426535725593567, + "learning_rate": 3e-05, + "loss": 4.0174, + "step": 2829 + }, + { + "epoch": 0.7690217391304348, + "grad_norm": 1.4052951335906982, + "learning_rate": 3e-05, + "loss": 3.3475, + "step": 2830 + }, + { + "epoch": 0.7692934782608696, + "grad_norm": 0.9852263331413269, + "learning_rate": 3e-05, + "loss": 2.9218, + "step": 2831 + }, + { + "epoch": 0.7695652173913043, + "grad_norm": 1.2342568635940552, + "learning_rate": 3e-05, + "loss": 3.1937, + "step": 2832 + }, + { + "epoch": 0.7698369565217391, + "grad_norm": 2.569143772125244, + "learning_rate": 3e-05, + "loss": 3.8627, + "step": 2833 + }, + { + "epoch": 0.7701086956521739, + "grad_norm": 1.07598876953125, + "learning_rate": 3e-05, + "loss": 2.8528, + "step": 2834 + }, + { + "epoch": 0.7703804347826086, + "grad_norm": 1.0930018424987793, + "learning_rate": 3e-05, + "loss": 3.6242, + "step": 2835 + }, + { + "epoch": 0.7706521739130435, + "grad_norm": 1.571661353111267, + "learning_rate": 3e-05, + "loss": 3.2027, + "step": 2836 + }, + { + "epoch": 0.7709239130434783, + "grad_norm": 1.623907446861267, + "learning_rate": 3e-05, + "loss": 3.9722, + "step": 2837 + }, + { + "epoch": 0.7711956521739131, + "grad_norm": 1.4214915037155151, + "learning_rate": 3e-05, + "loss": 3.483, + "step": 2838 + }, + { + "epoch": 0.7714673913043478, + "grad_norm": 1.1177326440811157, + "learning_rate": 3e-05, + "loss": 3.2388, + "step": 2839 + }, + { + "epoch": 0.7717391304347826, + "grad_norm": 1.2823110818862915, + "learning_rate": 3e-05, + "loss": 3.2644, + "step": 2840 + }, + { + "epoch": 0.7720108695652174, + "grad_norm": 1.169673204421997, + "learning_rate": 3e-05, + "loss": 3.8629, + "step": 2841 + }, + { + "epoch": 0.7722826086956521, + "grad_norm": 1.5543997287750244, + "learning_rate": 3e-05, + "loss": 3.6887, + "step": 2842 + }, + { + "epoch": 0.772554347826087, + "grad_norm": 0.8735353946685791, + "learning_rate": 3e-05, + "loss": 2.7277, + "step": 2843 + }, + { + "epoch": 0.7728260869565218, + "grad_norm": 0.8694882988929749, + "learning_rate": 3e-05, + "loss": 2.9491, + "step": 2844 + }, + { + "epoch": 0.7730978260869565, + "grad_norm": 1.2447638511657715, + "learning_rate": 3e-05, + "loss": 3.473, + "step": 2845 + }, + { + "epoch": 0.7733695652173913, + "grad_norm": 0.869049072265625, + "learning_rate": 3e-05, + "loss": 3.0047, + "step": 2846 + }, + { + "epoch": 0.7736413043478261, + "grad_norm": 1.1336791515350342, + "learning_rate": 3e-05, + "loss": 3.5196, + "step": 2847 + }, + { + "epoch": 0.7739130434782608, + "grad_norm": 1.0776828527450562, + "learning_rate": 3e-05, + "loss": 3.2469, + "step": 2848 + }, + { + "epoch": 0.7741847826086956, + "grad_norm": 1.1838096380233765, + "learning_rate": 3e-05, + "loss": 3.5876, + "step": 2849 + }, + { + "epoch": 0.7744565217391305, + "grad_norm": 0.9228270649909973, + "learning_rate": 3e-05, + "loss": 3.0038, + "step": 2850 + }, + { + "epoch": 0.7747282608695653, + "grad_norm": 0.9622864127159119, + "learning_rate": 3e-05, + "loss": 2.7338, + "step": 2851 + }, + { + "epoch": 0.775, + "grad_norm": 1.3739656209945679, + "learning_rate": 3e-05, + "loss": 3.3167, + "step": 2852 + }, + { + "epoch": 0.7752717391304348, + "grad_norm": 1.2962409257888794, + "learning_rate": 3e-05, + "loss": 2.9095, + "step": 2853 + }, + { + "epoch": 0.7755434782608696, + "grad_norm": 1.3018954992294312, + "learning_rate": 3e-05, + "loss": 3.4756, + "step": 2854 + }, + { + "epoch": 0.7758152173913043, + "grad_norm": 1.2301543951034546, + "learning_rate": 3e-05, + "loss": 3.658, + "step": 2855 + }, + { + "epoch": 0.7760869565217391, + "grad_norm": 0.9971374869346619, + "learning_rate": 3e-05, + "loss": 2.8921, + "step": 2856 + }, + { + "epoch": 0.7763586956521739, + "grad_norm": 1.0504611730575562, + "learning_rate": 3e-05, + "loss": 2.907, + "step": 2857 + }, + { + "epoch": 0.7766304347826087, + "grad_norm": 0.9376839399337769, + "learning_rate": 3e-05, + "loss": 2.9993, + "step": 2858 + }, + { + "epoch": 0.7769021739130435, + "grad_norm": 1.1995012760162354, + "learning_rate": 3e-05, + "loss": 3.505, + "step": 2859 + }, + { + "epoch": 0.7771739130434783, + "grad_norm": 1.2391998767852783, + "learning_rate": 3e-05, + "loss": 3.5183, + "step": 2860 + }, + { + "epoch": 0.777445652173913, + "grad_norm": 1.0271795988082886, + "learning_rate": 3e-05, + "loss": 2.764, + "step": 2861 + }, + { + "epoch": 0.7777173913043478, + "grad_norm": 1.1367555856704712, + "learning_rate": 3e-05, + "loss": 3.4225, + "step": 2862 + }, + { + "epoch": 0.7779891304347826, + "grad_norm": 1.258692979812622, + "learning_rate": 3e-05, + "loss": 3.9128, + "step": 2863 + }, + { + "epoch": 0.7782608695652173, + "grad_norm": 1.050280213356018, + "learning_rate": 3e-05, + "loss": 3.1644, + "step": 2864 + }, + { + "epoch": 0.7785326086956522, + "grad_norm": 1.4861702919006348, + "learning_rate": 3e-05, + "loss": 3.651, + "step": 2865 + }, + { + "epoch": 0.778804347826087, + "grad_norm": 1.9443000555038452, + "learning_rate": 3e-05, + "loss": 3.1064, + "step": 2866 + }, + { + "epoch": 0.7790760869565218, + "grad_norm": 1.2355011701583862, + "learning_rate": 3e-05, + "loss": 3.2686, + "step": 2867 + }, + { + "epoch": 0.7793478260869565, + "grad_norm": 1.06037175655365, + "learning_rate": 3e-05, + "loss": 3.4073, + "step": 2868 + }, + { + "epoch": 0.7796195652173913, + "grad_norm": 1.3563976287841797, + "learning_rate": 3e-05, + "loss": 3.9419, + "step": 2869 + }, + { + "epoch": 0.779891304347826, + "grad_norm": 1.411892294883728, + "learning_rate": 3e-05, + "loss": 3.4305, + "step": 2870 + }, + { + "epoch": 0.7801630434782608, + "grad_norm": 2.1324245929718018, + "learning_rate": 3e-05, + "loss": 3.5014, + "step": 2871 + }, + { + "epoch": 0.7804347826086957, + "grad_norm": 1.3413655757904053, + "learning_rate": 3e-05, + "loss": 3.4645, + "step": 2872 + }, + { + "epoch": 0.7807065217391305, + "grad_norm": 1.2617512941360474, + "learning_rate": 3e-05, + "loss": 3.1992, + "step": 2873 + }, + { + "epoch": 0.7809782608695652, + "grad_norm": 1.1358786821365356, + "learning_rate": 3e-05, + "loss": 3.223, + "step": 2874 + }, + { + "epoch": 0.78125, + "grad_norm": 1.284077525138855, + "learning_rate": 3e-05, + "loss": 3.2483, + "step": 2875 + }, + { + "epoch": 0.7815217391304348, + "grad_norm": 2.2102537155151367, + "learning_rate": 3e-05, + "loss": 3.3587, + "step": 2876 + }, + { + "epoch": 0.7817934782608695, + "grad_norm": 1.2199740409851074, + "learning_rate": 3e-05, + "loss": 3.0096, + "step": 2877 + }, + { + "epoch": 0.7820652173913043, + "grad_norm": 1.0134555101394653, + "learning_rate": 3e-05, + "loss": 3.0385, + "step": 2878 + }, + { + "epoch": 0.7823369565217392, + "grad_norm": 0.9114274978637695, + "learning_rate": 3e-05, + "loss": 2.7133, + "step": 2879 + }, + { + "epoch": 0.782608695652174, + "grad_norm": 1.0371575355529785, + "learning_rate": 3e-05, + "loss": 3.1674, + "step": 2880 + }, + { + "epoch": 0.7828804347826087, + "grad_norm": 1.2628506422042847, + "learning_rate": 3e-05, + "loss": 3.5175, + "step": 2881 + }, + { + "epoch": 0.7831521739130435, + "grad_norm": 1.2910369634628296, + "learning_rate": 3e-05, + "loss": 3.1921, + "step": 2882 + }, + { + "epoch": 0.7834239130434782, + "grad_norm": 1.460605263710022, + "learning_rate": 3e-05, + "loss": 3.2101, + "step": 2883 + }, + { + "epoch": 0.783695652173913, + "grad_norm": 1.4789795875549316, + "learning_rate": 3e-05, + "loss": 3.5643, + "step": 2884 + }, + { + "epoch": 0.7839673913043478, + "grad_norm": 1.3687986135482788, + "learning_rate": 3e-05, + "loss": 3.5187, + "step": 2885 + }, + { + "epoch": 0.7842391304347827, + "grad_norm": 1.4191471338272095, + "learning_rate": 3e-05, + "loss": 3.8682, + "step": 2886 + }, + { + "epoch": 0.7845108695652174, + "grad_norm": 1.3148168325424194, + "learning_rate": 3e-05, + "loss": 3.0195, + "step": 2887 + }, + { + "epoch": 0.7847826086956522, + "grad_norm": 1.15743088722229, + "learning_rate": 3e-05, + "loss": 2.9205, + "step": 2888 + }, + { + "epoch": 0.785054347826087, + "grad_norm": 1.167377233505249, + "learning_rate": 3e-05, + "loss": 3.3717, + "step": 2889 + }, + { + "epoch": 0.7853260869565217, + "grad_norm": 1.369253158569336, + "learning_rate": 3e-05, + "loss": 3.3943, + "step": 2890 + }, + { + "epoch": 0.7855978260869565, + "grad_norm": 1.3081823587417603, + "learning_rate": 3e-05, + "loss": 3.1458, + "step": 2891 + }, + { + "epoch": 0.7858695652173913, + "grad_norm": 1.0637850761413574, + "learning_rate": 3e-05, + "loss": 3.5853, + "step": 2892 + }, + { + "epoch": 0.7861413043478261, + "grad_norm": 1.325531005859375, + "learning_rate": 3e-05, + "loss": 3.6246, + "step": 2893 + }, + { + "epoch": 0.7864130434782609, + "grad_norm": 1.2909876108169556, + "learning_rate": 3e-05, + "loss": 2.7209, + "step": 2894 + }, + { + "epoch": 0.7866847826086957, + "grad_norm": 1.072770357131958, + "learning_rate": 3e-05, + "loss": 2.7364, + "step": 2895 + }, + { + "epoch": 0.7869565217391304, + "grad_norm": 1.2623767852783203, + "learning_rate": 3e-05, + "loss": 3.193, + "step": 2896 + }, + { + "epoch": 0.7872282608695652, + "grad_norm": 1.262104868888855, + "learning_rate": 3e-05, + "loss": 3.4054, + "step": 2897 + }, + { + "epoch": 0.7875, + "grad_norm": 0.9568426012992859, + "learning_rate": 3e-05, + "loss": 3.0418, + "step": 2898 + }, + { + "epoch": 0.7877717391304347, + "grad_norm": 1.0304343700408936, + "learning_rate": 3e-05, + "loss": 3.1849, + "step": 2899 + }, + { + "epoch": 0.7880434782608695, + "grad_norm": 1.1511303186416626, + "learning_rate": 3e-05, + "loss": 3.3258, + "step": 2900 + }, + { + "epoch": 0.7883152173913044, + "grad_norm": 0.9511053562164307, + "learning_rate": 3e-05, + "loss": 2.9477, + "step": 2901 + }, + { + "epoch": 0.7885869565217392, + "grad_norm": 1.1826517581939697, + "learning_rate": 3e-05, + "loss": 2.9749, + "step": 2902 + }, + { + "epoch": 0.7888586956521739, + "grad_norm": 0.9371961355209351, + "learning_rate": 3e-05, + "loss": 3.0142, + "step": 2903 + }, + { + "epoch": 0.7891304347826087, + "grad_norm": 1.3846261501312256, + "learning_rate": 3e-05, + "loss": 3.3297, + "step": 2904 + }, + { + "epoch": 0.7894021739130435, + "grad_norm": 1.09294593334198, + "learning_rate": 3e-05, + "loss": 3.1767, + "step": 2905 + }, + { + "epoch": 0.7896739130434782, + "grad_norm": 1.326731562614441, + "learning_rate": 3e-05, + "loss": 3.8467, + "step": 2906 + }, + { + "epoch": 0.789945652173913, + "grad_norm": 1.1063183546066284, + "learning_rate": 3e-05, + "loss": 3.0861, + "step": 2907 + }, + { + "epoch": 0.7902173913043479, + "grad_norm": 1.174807071685791, + "learning_rate": 3e-05, + "loss": 3.0249, + "step": 2908 + }, + { + "epoch": 0.7904891304347826, + "grad_norm": 0.9046750068664551, + "learning_rate": 3e-05, + "loss": 3.0051, + "step": 2909 + }, + { + "epoch": 0.7907608695652174, + "grad_norm": 2.0610415935516357, + "learning_rate": 3e-05, + "loss": 3.9009, + "step": 2910 + }, + { + "epoch": 0.7910326086956522, + "grad_norm": 1.4573287963867188, + "learning_rate": 3e-05, + "loss": 3.1832, + "step": 2911 + }, + { + "epoch": 0.7913043478260869, + "grad_norm": 1.1538214683532715, + "learning_rate": 3e-05, + "loss": 2.7802, + "step": 2912 + }, + { + "epoch": 0.7915760869565217, + "grad_norm": 1.0094918012619019, + "learning_rate": 3e-05, + "loss": 3.0695, + "step": 2913 + }, + { + "epoch": 0.7918478260869565, + "grad_norm": 1.1341625452041626, + "learning_rate": 3e-05, + "loss": 3.4901, + "step": 2914 + }, + { + "epoch": 0.7921195652173914, + "grad_norm": 1.0618107318878174, + "learning_rate": 3e-05, + "loss": 3.1211, + "step": 2915 + }, + { + "epoch": 0.7923913043478261, + "grad_norm": 1.4313887357711792, + "learning_rate": 3e-05, + "loss": 3.8207, + "step": 2916 + }, + { + "epoch": 0.7926630434782609, + "grad_norm": 1.4587217569351196, + "learning_rate": 3e-05, + "loss": 3.5706, + "step": 2917 + }, + { + "epoch": 0.7929347826086957, + "grad_norm": 1.4848532676696777, + "learning_rate": 3e-05, + "loss": 3.6406, + "step": 2918 + }, + { + "epoch": 0.7932065217391304, + "grad_norm": 1.006595253944397, + "learning_rate": 3e-05, + "loss": 3.0846, + "step": 2919 + }, + { + "epoch": 0.7934782608695652, + "grad_norm": 1.184501051902771, + "learning_rate": 3e-05, + "loss": 3.3719, + "step": 2920 + }, + { + "epoch": 0.79375, + "grad_norm": 1.3015398979187012, + "learning_rate": 3e-05, + "loss": 2.9707, + "step": 2921 + }, + { + "epoch": 0.7940217391304348, + "grad_norm": 1.001300573348999, + "learning_rate": 3e-05, + "loss": 2.9032, + "step": 2922 + }, + { + "epoch": 0.7942934782608696, + "grad_norm": 1.4572199583053589, + "learning_rate": 3e-05, + "loss": 3.7695, + "step": 2923 + }, + { + "epoch": 0.7945652173913044, + "grad_norm": 1.3405942916870117, + "learning_rate": 3e-05, + "loss": 2.8765, + "step": 2924 + }, + { + "epoch": 0.7948369565217391, + "grad_norm": 1.6628684997558594, + "learning_rate": 3e-05, + "loss": 3.6848, + "step": 2925 + }, + { + "epoch": 0.7951086956521739, + "grad_norm": 1.3848426342010498, + "learning_rate": 3e-05, + "loss": 3.3296, + "step": 2926 + }, + { + "epoch": 0.7953804347826087, + "grad_norm": 1.5421236753463745, + "learning_rate": 3e-05, + "loss": 3.363, + "step": 2927 + }, + { + "epoch": 0.7956521739130434, + "grad_norm": 1.3111261129379272, + "learning_rate": 3e-05, + "loss": 2.8346, + "step": 2928 + }, + { + "epoch": 0.7959239130434783, + "grad_norm": 1.0846003293991089, + "learning_rate": 3e-05, + "loss": 3.3071, + "step": 2929 + }, + { + "epoch": 0.7961956521739131, + "grad_norm": 0.9776524305343628, + "learning_rate": 3e-05, + "loss": 2.9533, + "step": 2930 + }, + { + "epoch": 0.7964673913043478, + "grad_norm": 1.3840489387512207, + "learning_rate": 3e-05, + "loss": 3.4739, + "step": 2931 + }, + { + "epoch": 0.7967391304347826, + "grad_norm": 1.2059979438781738, + "learning_rate": 3e-05, + "loss": 2.9474, + "step": 2932 + }, + { + "epoch": 0.7970108695652174, + "grad_norm": 1.2110824584960938, + "learning_rate": 3e-05, + "loss": 3.2581, + "step": 2933 + }, + { + "epoch": 0.7972826086956522, + "grad_norm": 1.0525647401809692, + "learning_rate": 3e-05, + "loss": 2.9519, + "step": 2934 + }, + { + "epoch": 0.7975543478260869, + "grad_norm": 1.2633483409881592, + "learning_rate": 3e-05, + "loss": 3.1686, + "step": 2935 + }, + { + "epoch": 0.7978260869565217, + "grad_norm": 0.8925679922103882, + "learning_rate": 3e-05, + "loss": 2.8761, + "step": 2936 + }, + { + "epoch": 0.7980978260869566, + "grad_norm": 0.9802154898643494, + "learning_rate": 3e-05, + "loss": 3.2334, + "step": 2937 + }, + { + "epoch": 0.7983695652173913, + "grad_norm": 1.3679046630859375, + "learning_rate": 3e-05, + "loss": 3.2836, + "step": 2938 + }, + { + "epoch": 0.7986413043478261, + "grad_norm": 1.0334969758987427, + "learning_rate": 3e-05, + "loss": 3.1355, + "step": 2939 + }, + { + "epoch": 0.7989130434782609, + "grad_norm": 1.0484052896499634, + "learning_rate": 3e-05, + "loss": 3.1652, + "step": 2940 + }, + { + "epoch": 0.7991847826086956, + "grad_norm": 1.0501008033752441, + "learning_rate": 3e-05, + "loss": 2.9963, + "step": 2941 + }, + { + "epoch": 0.7994565217391304, + "grad_norm": 1.2090955972671509, + "learning_rate": 3e-05, + "loss": 3.3553, + "step": 2942 + }, + { + "epoch": 0.7997282608695652, + "grad_norm": 1.1819127798080444, + "learning_rate": 3e-05, + "loss": 3.1212, + "step": 2943 + }, + { + "epoch": 0.8, + "grad_norm": 1.1051278114318848, + "learning_rate": 3e-05, + "loss": 3.2308, + "step": 2944 + }, + { + "epoch": 0.8002717391304348, + "grad_norm": 0.9563131928443909, + "learning_rate": 3e-05, + "loss": 2.7561, + "step": 2945 + }, + { + "epoch": 0.8005434782608696, + "grad_norm": 1.1029945611953735, + "learning_rate": 3e-05, + "loss": 3.6245, + "step": 2946 + }, + { + "epoch": 0.8008152173913043, + "grad_norm": 1.3487423658370972, + "learning_rate": 3e-05, + "loss": 3.4707, + "step": 2947 + }, + { + "epoch": 0.8010869565217391, + "grad_norm": 1.7481412887573242, + "learning_rate": 3e-05, + "loss": 3.6706, + "step": 2948 + }, + { + "epoch": 0.8013586956521739, + "grad_norm": 1.0554581880569458, + "learning_rate": 3e-05, + "loss": 3.0406, + "step": 2949 + }, + { + "epoch": 0.8016304347826086, + "grad_norm": 1.0953891277313232, + "learning_rate": 3e-05, + "loss": 2.7938, + "step": 2950 + }, + { + "epoch": 0.8019021739130435, + "grad_norm": 1.3923141956329346, + "learning_rate": 3e-05, + "loss": 3.0691, + "step": 2951 + }, + { + "epoch": 0.8021739130434783, + "grad_norm": 1.1058765649795532, + "learning_rate": 3e-05, + "loss": 2.9356, + "step": 2952 + }, + { + "epoch": 0.8024456521739131, + "grad_norm": 1.1275757551193237, + "learning_rate": 3e-05, + "loss": 2.8867, + "step": 2953 + }, + { + "epoch": 0.8027173913043478, + "grad_norm": 1.2130517959594727, + "learning_rate": 3e-05, + "loss": 3.1133, + "step": 2954 + }, + { + "epoch": 0.8029891304347826, + "grad_norm": 1.6094809770584106, + "learning_rate": 3e-05, + "loss": 3.8152, + "step": 2955 + }, + { + "epoch": 0.8032608695652174, + "grad_norm": 0.9870564341545105, + "learning_rate": 3e-05, + "loss": 2.721, + "step": 2956 + }, + { + "epoch": 0.8035326086956521, + "grad_norm": 1.2996047735214233, + "learning_rate": 3e-05, + "loss": 3.6848, + "step": 2957 + }, + { + "epoch": 0.803804347826087, + "grad_norm": 1.2500205039978027, + "learning_rate": 3e-05, + "loss": 3.3352, + "step": 2958 + }, + { + "epoch": 0.8040760869565218, + "grad_norm": 1.3211798667907715, + "learning_rate": 3e-05, + "loss": 3.7284, + "step": 2959 + }, + { + "epoch": 0.8043478260869565, + "grad_norm": 1.514905571937561, + "learning_rate": 3e-05, + "loss": 3.2425, + "step": 2960 + }, + { + "epoch": 0.8046195652173913, + "grad_norm": 1.2107200622558594, + "learning_rate": 3e-05, + "loss": 3.4771, + "step": 2961 + }, + { + "epoch": 0.8048913043478261, + "grad_norm": 1.707463264465332, + "learning_rate": 3e-05, + "loss": 3.6518, + "step": 2962 + }, + { + "epoch": 0.8051630434782608, + "grad_norm": 0.9443992376327515, + "learning_rate": 3e-05, + "loss": 2.8471, + "step": 2963 + }, + { + "epoch": 0.8054347826086956, + "grad_norm": 1.3588606119155884, + "learning_rate": 3e-05, + "loss": 3.3993, + "step": 2964 + }, + { + "epoch": 0.8057065217391305, + "grad_norm": 1.2722697257995605, + "learning_rate": 3e-05, + "loss": 3.2316, + "step": 2965 + }, + { + "epoch": 0.8059782608695653, + "grad_norm": 1.6636009216308594, + "learning_rate": 3e-05, + "loss": 3.5648, + "step": 2966 + }, + { + "epoch": 0.80625, + "grad_norm": 1.3926972150802612, + "learning_rate": 3e-05, + "loss": 3.6403, + "step": 2967 + }, + { + "epoch": 0.8065217391304348, + "grad_norm": 1.2489240169525146, + "learning_rate": 3e-05, + "loss": 3.6428, + "step": 2968 + }, + { + "epoch": 0.8067934782608696, + "grad_norm": 1.3040508031845093, + "learning_rate": 3e-05, + "loss": 3.3698, + "step": 2969 + }, + { + "epoch": 0.8070652173913043, + "grad_norm": 1.024529218673706, + "learning_rate": 3e-05, + "loss": 3.1765, + "step": 2970 + }, + { + "epoch": 0.8073369565217391, + "grad_norm": 1.127133846282959, + "learning_rate": 3e-05, + "loss": 3.4167, + "step": 2971 + }, + { + "epoch": 0.8076086956521739, + "grad_norm": 1.5928646326065063, + "learning_rate": 3e-05, + "loss": 3.2939, + "step": 2972 + }, + { + "epoch": 0.8078804347826087, + "grad_norm": 1.099817156791687, + "learning_rate": 3e-05, + "loss": 3.3166, + "step": 2973 + }, + { + "epoch": 0.8081521739130435, + "grad_norm": 0.9949479699134827, + "learning_rate": 3e-05, + "loss": 3.4637, + "step": 2974 + }, + { + "epoch": 0.8084239130434783, + "grad_norm": 1.3012653589248657, + "learning_rate": 3e-05, + "loss": 3.3314, + "step": 2975 + }, + { + "epoch": 0.808695652173913, + "grad_norm": 1.1788009405136108, + "learning_rate": 3e-05, + "loss": 2.9917, + "step": 2976 + }, + { + "epoch": 0.8089673913043478, + "grad_norm": 1.5235824584960938, + "learning_rate": 3e-05, + "loss": 3.0784, + "step": 2977 + }, + { + "epoch": 0.8092391304347826, + "grad_norm": 1.4393073320388794, + "learning_rate": 3e-05, + "loss": 3.8917, + "step": 2978 + }, + { + "epoch": 0.8095108695652173, + "grad_norm": 1.432785987854004, + "learning_rate": 3e-05, + "loss": 3.8753, + "step": 2979 + }, + { + "epoch": 0.8097826086956522, + "grad_norm": 1.4102272987365723, + "learning_rate": 3e-05, + "loss": 3.0096, + "step": 2980 + }, + { + "epoch": 0.810054347826087, + "grad_norm": 1.5349915027618408, + "learning_rate": 3e-05, + "loss": 3.2429, + "step": 2981 + }, + { + "epoch": 0.8103260869565218, + "grad_norm": 0.9741056561470032, + "learning_rate": 3e-05, + "loss": 3.2216, + "step": 2982 + }, + { + "epoch": 0.8105978260869565, + "grad_norm": 1.3699191808700562, + "learning_rate": 3e-05, + "loss": 2.8594, + "step": 2983 + }, + { + "epoch": 0.8108695652173913, + "grad_norm": 1.0808976888656616, + "learning_rate": 3e-05, + "loss": 3.1742, + "step": 2984 + }, + { + "epoch": 0.811141304347826, + "grad_norm": 0.9032256007194519, + "learning_rate": 3e-05, + "loss": 3.0112, + "step": 2985 + }, + { + "epoch": 0.8114130434782608, + "grad_norm": 1.119470477104187, + "learning_rate": 3e-05, + "loss": 3.1507, + "step": 2986 + }, + { + "epoch": 0.8116847826086957, + "grad_norm": 1.3185955286026, + "learning_rate": 3e-05, + "loss": 3.7403, + "step": 2987 + }, + { + "epoch": 0.8119565217391305, + "grad_norm": 1.283615231513977, + "learning_rate": 3e-05, + "loss": 3.2814, + "step": 2988 + }, + { + "epoch": 0.8122282608695652, + "grad_norm": 1.2075817584991455, + "learning_rate": 3e-05, + "loss": 3.5796, + "step": 2989 + }, + { + "epoch": 0.8125, + "grad_norm": 1.1944336891174316, + "learning_rate": 3e-05, + "loss": 3.3505, + "step": 2990 + }, + { + "epoch": 0.8127717391304348, + "grad_norm": 1.1111395359039307, + "learning_rate": 3e-05, + "loss": 3.2605, + "step": 2991 + }, + { + "epoch": 0.8130434782608695, + "grad_norm": 1.2699017524719238, + "learning_rate": 3e-05, + "loss": 3.4753, + "step": 2992 + }, + { + "epoch": 0.8133152173913043, + "grad_norm": 1.1877753734588623, + "learning_rate": 3e-05, + "loss": 3.0174, + "step": 2993 + }, + { + "epoch": 0.8135869565217392, + "grad_norm": 1.198492407798767, + "learning_rate": 3e-05, + "loss": 3.4792, + "step": 2994 + }, + { + "epoch": 0.813858695652174, + "grad_norm": 1.2535799741744995, + "learning_rate": 3e-05, + "loss": 3.4246, + "step": 2995 + }, + { + "epoch": 0.8141304347826087, + "grad_norm": 1.0552626848220825, + "learning_rate": 3e-05, + "loss": 2.9214, + "step": 2996 + }, + { + "epoch": 0.8144021739130435, + "grad_norm": 1.1369494199752808, + "learning_rate": 3e-05, + "loss": 2.9704, + "step": 2997 + }, + { + "epoch": 0.8146739130434782, + "grad_norm": 1.04477858543396, + "learning_rate": 3e-05, + "loss": 2.916, + "step": 2998 + }, + { + "epoch": 0.814945652173913, + "grad_norm": 0.9620361328125, + "learning_rate": 3e-05, + "loss": 2.6156, + "step": 2999 + }, + { + "epoch": 0.8152173913043478, + "grad_norm": 1.0974937677383423, + "learning_rate": 3e-05, + "loss": 3.362, + "step": 3000 + } + ], + "logging_steps": 1, + "max_steps": 3680, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 1000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.262770368118784e+18, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}