{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 390, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.002564102564102564, "grad_norm": 1.828752040863037, "learning_rate": 0.0, "loss": 1.0674, "step": 1 }, { "epoch": 0.005128205128205128, "grad_norm": 2.0600810050964355, "learning_rate": 1e-05, "loss": 1.0127, "step": 2 }, { "epoch": 0.007692307692307693, "grad_norm": 1.7609747648239136, "learning_rate": 1e-05, "loss": 0.8945, "step": 3 }, { "epoch": 0.010256410256410256, "grad_norm": 2.0786192417144775, "learning_rate": 9.974226804123713e-06, "loss": 1.0947, "step": 4 }, { "epoch": 0.01282051282051282, "grad_norm": 1.8740049600601196, "learning_rate": 9.948453608247423e-06, "loss": 1.0859, "step": 5 }, { "epoch": 0.015384615384615385, "grad_norm": 1.8995977640151978, "learning_rate": 9.922680412371136e-06, "loss": 0.9326, "step": 6 }, { "epoch": 0.017948717948717947, "grad_norm": 1.807106375694275, "learning_rate": 9.896907216494846e-06, "loss": 0.9839, "step": 7 }, { "epoch": 0.020512820512820513, "grad_norm": 2.0719048976898193, "learning_rate": 9.871134020618558e-06, "loss": 1.1914, "step": 8 }, { "epoch": 0.023076923076923078, "grad_norm": 1.714697241783142, "learning_rate": 9.84536082474227e-06, "loss": 0.9263, "step": 9 }, { "epoch": 0.02564102564102564, "grad_norm": 1.7930278778076172, "learning_rate": 9.819587628865979e-06, "loss": 0.9546, "step": 10 }, { "epoch": 0.028205128205128206, "grad_norm": 1.8702346086502075, "learning_rate": 9.793814432989691e-06, "loss": 1.0645, "step": 11 }, { "epoch": 0.03076923076923077, "grad_norm": 1.7376536130905151, "learning_rate": 9.768041237113403e-06, "loss": 1.0166, "step": 12 }, { "epoch": 0.03333333333333333, "grad_norm": 1.804457426071167, "learning_rate": 9.742268041237114e-06, "loss": 0.8545, "step": 13 }, { "epoch": 0.035897435897435895, "grad_norm": 1.8418951034545898, "learning_rate": 9.716494845360826e-06, "loss": 1.1855, "step": 14 }, { "epoch": 0.038461538461538464, "grad_norm": 3.295741558074951, "learning_rate": 9.690721649484536e-06, "loss": 1.1631, "step": 15 }, { "epoch": 0.041025641025641026, "grad_norm": 2.6767563819885254, "learning_rate": 9.664948453608248e-06, "loss": 1.0205, "step": 16 }, { "epoch": 0.04358974358974359, "grad_norm": 2.088998556137085, "learning_rate": 9.63917525773196e-06, "loss": 0.9512, "step": 17 }, { "epoch": 0.046153846153846156, "grad_norm": 3.8526885509490967, "learning_rate": 9.613402061855671e-06, "loss": 1.2227, "step": 18 }, { "epoch": 0.04871794871794872, "grad_norm": 2.5801403522491455, "learning_rate": 9.587628865979383e-06, "loss": 0.9424, "step": 19 }, { "epoch": 0.05128205128205128, "grad_norm": 2.2087137699127197, "learning_rate": 9.561855670103093e-06, "loss": 1.3262, "step": 20 }, { "epoch": 0.05384615384615385, "grad_norm": 1.9106372594833374, "learning_rate": 9.536082474226806e-06, "loss": 0.9004, "step": 21 }, { "epoch": 0.05641025641025641, "grad_norm": 1.7588039636611938, "learning_rate": 9.510309278350516e-06, "loss": 0.9482, "step": 22 }, { "epoch": 0.05897435897435897, "grad_norm": 2.0292000770568848, "learning_rate": 9.484536082474226e-06, "loss": 1.0371, "step": 23 }, { "epoch": 0.06153846153846154, "grad_norm": 1.8534419536590576, "learning_rate": 9.458762886597939e-06, "loss": 1.0303, "step": 24 }, { "epoch": 0.0641025641025641, "grad_norm": 2.4608726501464844, "learning_rate": 9.43298969072165e-06, "loss": 1.4395, "step": 25 }, { "epoch": 0.06666666666666667, "grad_norm": 2.119417905807495, "learning_rate": 9.407216494845361e-06, "loss": 1.1143, "step": 26 }, { "epoch": 0.06923076923076923, "grad_norm": 2.145531177520752, "learning_rate": 9.381443298969073e-06, "loss": 0.9443, "step": 27 }, { "epoch": 0.07179487179487179, "grad_norm": 4.710203170776367, "learning_rate": 9.355670103092784e-06, "loss": 1.3311, "step": 28 }, { "epoch": 0.07435897435897436, "grad_norm": 3.2553186416625977, "learning_rate": 9.329896907216496e-06, "loss": 1.123, "step": 29 }, { "epoch": 0.07692307692307693, "grad_norm": 2.38273024559021, "learning_rate": 9.304123711340208e-06, "loss": 1.124, "step": 30 }, { "epoch": 0.07948717948717948, "grad_norm": 3.6473400592803955, "learning_rate": 9.278350515463918e-06, "loss": 1.2568, "step": 31 }, { "epoch": 0.08205128205128205, "grad_norm": 2.2762739658355713, "learning_rate": 9.25257731958763e-06, "loss": 1.0898, "step": 32 }, { "epoch": 0.08461538461538462, "grad_norm": 1.870347499847412, "learning_rate": 9.226804123711341e-06, "loss": 0.8887, "step": 33 }, { "epoch": 0.08717948717948718, "grad_norm": 3.9631059169769287, "learning_rate": 9.201030927835051e-06, "loss": 0.937, "step": 34 }, { "epoch": 0.08974358974358974, "grad_norm": 2.2279770374298096, "learning_rate": 9.175257731958764e-06, "loss": 0.9785, "step": 35 }, { "epoch": 0.09230769230769231, "grad_norm": 2.2554802894592285, "learning_rate": 9.149484536082474e-06, "loss": 1.0762, "step": 36 }, { "epoch": 0.09487179487179487, "grad_norm": 2.188344955444336, "learning_rate": 9.123711340206186e-06, "loss": 1.1318, "step": 37 }, { "epoch": 0.09743589743589744, "grad_norm": 2.1649510860443115, "learning_rate": 9.097938144329898e-06, "loss": 0.7827, "step": 38 }, { "epoch": 0.1, "grad_norm": 2.4596219062805176, "learning_rate": 9.072164948453609e-06, "loss": 0.8247, "step": 39 }, { "epoch": 0.10256410256410256, "grad_norm": 2.487870931625366, "learning_rate": 9.04639175257732e-06, "loss": 1.0635, "step": 40 }, { "epoch": 0.10512820512820513, "grad_norm": 1.9071369171142578, "learning_rate": 9.020618556701031e-06, "loss": 0.9946, "step": 41 }, { "epoch": 0.1076923076923077, "grad_norm": 2.219597101211548, "learning_rate": 8.994845360824743e-06, "loss": 1.1309, "step": 42 }, { "epoch": 0.11025641025641025, "grad_norm": 2.2724695205688477, "learning_rate": 8.969072164948455e-06, "loss": 1.2637, "step": 43 }, { "epoch": 0.11282051282051282, "grad_norm": 2.331315755844116, "learning_rate": 8.943298969072166e-06, "loss": 1.1611, "step": 44 }, { "epoch": 0.11538461538461539, "grad_norm": 2.036611318588257, "learning_rate": 8.917525773195878e-06, "loss": 0.9233, "step": 45 }, { "epoch": 0.11794871794871795, "grad_norm": 1.9779893159866333, "learning_rate": 8.891752577319588e-06, "loss": 0.8071, "step": 46 }, { "epoch": 0.12051282051282051, "grad_norm": 2.511869192123413, "learning_rate": 8.865979381443299e-06, "loss": 0.9165, "step": 47 }, { "epoch": 0.12307692307692308, "grad_norm": 1.9824575185775757, "learning_rate": 8.840206185567011e-06, "loss": 0.9712, "step": 48 }, { "epoch": 0.12564102564102564, "grad_norm": 2.224874973297119, "learning_rate": 8.814432989690721e-06, "loss": 0.8857, "step": 49 }, { "epoch": 0.1282051282051282, "grad_norm": 2.290484666824341, "learning_rate": 8.788659793814434e-06, "loss": 1.0908, "step": 50 }, { "epoch": 0.13076923076923078, "grad_norm": 2.009584903717041, "learning_rate": 8.762886597938146e-06, "loss": 0.833, "step": 51 }, { "epoch": 0.13333333333333333, "grad_norm": 4.70264196395874, "learning_rate": 8.737113402061856e-06, "loss": 0.998, "step": 52 }, { "epoch": 0.1358974358974359, "grad_norm": 3.110806465148926, "learning_rate": 8.711340206185568e-06, "loss": 0.9189, "step": 53 }, { "epoch": 0.13846153846153847, "grad_norm": 1.902529239654541, "learning_rate": 8.685567010309279e-06, "loss": 1.0293, "step": 54 }, { "epoch": 0.14102564102564102, "grad_norm": 2.1648287773132324, "learning_rate": 8.65979381443299e-06, "loss": 1.0488, "step": 55 }, { "epoch": 0.14358974358974358, "grad_norm": 1.7489367723464966, "learning_rate": 8.634020618556703e-06, "loss": 0.8091, "step": 56 }, { "epoch": 0.14615384615384616, "grad_norm": 1.9431246519088745, "learning_rate": 8.608247422680413e-06, "loss": 1.0049, "step": 57 }, { "epoch": 0.14871794871794872, "grad_norm": 2.2493724822998047, "learning_rate": 8.582474226804124e-06, "loss": 0.9321, "step": 58 }, { "epoch": 0.15128205128205127, "grad_norm": 2.1866893768310547, "learning_rate": 8.556701030927836e-06, "loss": 1.2656, "step": 59 }, { "epoch": 0.15384615384615385, "grad_norm": 2.024313449859619, "learning_rate": 8.530927835051546e-06, "loss": 1.0801, "step": 60 }, { "epoch": 0.1564102564102564, "grad_norm": 1.77311110496521, "learning_rate": 8.505154639175259e-06, "loss": 0.9946, "step": 61 }, { "epoch": 0.15897435897435896, "grad_norm": 1.6627540588378906, "learning_rate": 8.479381443298969e-06, "loss": 0.9702, "step": 62 }, { "epoch": 0.16153846153846155, "grad_norm": 1.9997308254241943, "learning_rate": 8.453608247422681e-06, "loss": 0.9487, "step": 63 }, { "epoch": 0.1641025641025641, "grad_norm": 1.6049851179122925, "learning_rate": 8.427835051546393e-06, "loss": 0.9512, "step": 64 }, { "epoch": 0.16666666666666666, "grad_norm": 1.8791325092315674, "learning_rate": 8.402061855670104e-06, "loss": 0.9355, "step": 65 }, { "epoch": 0.16923076923076924, "grad_norm": 2.173393964767456, "learning_rate": 8.376288659793816e-06, "loss": 1.1992, "step": 66 }, { "epoch": 0.1717948717948718, "grad_norm": 2.0367043018341064, "learning_rate": 8.350515463917526e-06, "loss": 1.1426, "step": 67 }, { "epoch": 0.17435897435897435, "grad_norm": 1.7792794704437256, "learning_rate": 8.324742268041238e-06, "loss": 0.9097, "step": 68 }, { "epoch": 0.17692307692307693, "grad_norm": 1.9103853702545166, "learning_rate": 8.29896907216495e-06, "loss": 1.0225, "step": 69 }, { "epoch": 0.1794871794871795, "grad_norm": 2.0614025592803955, "learning_rate": 8.27319587628866e-06, "loss": 1.2012, "step": 70 }, { "epoch": 0.18205128205128204, "grad_norm": 1.5830662250518799, "learning_rate": 8.247422680412371e-06, "loss": 0.7656, "step": 71 }, { "epoch": 0.18461538461538463, "grad_norm": 1.8411104679107666, "learning_rate": 8.221649484536083e-06, "loss": 1.1162, "step": 72 }, { "epoch": 0.18717948717948718, "grad_norm": 1.954744815826416, "learning_rate": 8.195876288659794e-06, "loss": 1.2031, "step": 73 }, { "epoch": 0.18974358974358974, "grad_norm": 2.1274709701538086, "learning_rate": 8.170103092783506e-06, "loss": 1.1641, "step": 74 }, { "epoch": 0.19230769230769232, "grad_norm": 1.8388986587524414, "learning_rate": 8.144329896907216e-06, "loss": 0.9517, "step": 75 }, { "epoch": 0.19487179487179487, "grad_norm": 2.0911478996276855, "learning_rate": 8.118556701030929e-06, "loss": 1.0801, "step": 76 }, { "epoch": 0.19743589743589743, "grad_norm": 1.9565143585205078, "learning_rate": 8.09278350515464e-06, "loss": 0.9478, "step": 77 }, { "epoch": 0.2, "grad_norm": 1.8666242361068726, "learning_rate": 8.067010309278351e-06, "loss": 1.3125, "step": 78 }, { "epoch": 0.20256410256410257, "grad_norm": 1.9826719760894775, "learning_rate": 8.041237113402063e-06, "loss": 0.9492, "step": 79 }, { "epoch": 0.20512820512820512, "grad_norm": 2.062222719192505, "learning_rate": 8.015463917525774e-06, "loss": 1.0195, "step": 80 }, { "epoch": 0.2076923076923077, "grad_norm": 2.2730824947357178, "learning_rate": 7.989690721649486e-06, "loss": 0.9248, "step": 81 }, { "epoch": 0.21025641025641026, "grad_norm": 3.432387113571167, "learning_rate": 7.963917525773196e-06, "loss": 0.9678, "step": 82 }, { "epoch": 0.2128205128205128, "grad_norm": 2.1514482498168945, "learning_rate": 7.938144329896907e-06, "loss": 0.7881, "step": 83 }, { "epoch": 0.2153846153846154, "grad_norm": 1.6954137086868286, "learning_rate": 7.912371134020619e-06, "loss": 0.8833, "step": 84 }, { "epoch": 0.21794871794871795, "grad_norm": 1.9222341775894165, "learning_rate": 7.886597938144331e-06, "loss": 0.9917, "step": 85 }, { "epoch": 0.2205128205128205, "grad_norm": 1.9066567420959473, "learning_rate": 7.860824742268041e-06, "loss": 0.9507, "step": 86 }, { "epoch": 0.2230769230769231, "grad_norm": 1.9370355606079102, "learning_rate": 7.835051546391754e-06, "loss": 0.9814, "step": 87 }, { "epoch": 0.22564102564102564, "grad_norm": 1.8217931985855103, "learning_rate": 7.809278350515464e-06, "loss": 1.0186, "step": 88 }, { "epoch": 0.2282051282051282, "grad_norm": 1.6907788515090942, "learning_rate": 7.783505154639176e-06, "loss": 0.9624, "step": 89 }, { "epoch": 0.23076923076923078, "grad_norm": 1.6538673639297485, "learning_rate": 7.757731958762888e-06, "loss": 0.8237, "step": 90 }, { "epoch": 0.23333333333333334, "grad_norm": 1.9128988981246948, "learning_rate": 7.731958762886599e-06, "loss": 0.9175, "step": 91 }, { "epoch": 0.2358974358974359, "grad_norm": 3.1747217178344727, "learning_rate": 7.70618556701031e-06, "loss": 0.9775, "step": 92 }, { "epoch": 0.23846153846153847, "grad_norm": 1.6946772336959839, "learning_rate": 7.680412371134021e-06, "loss": 0.8823, "step": 93 }, { "epoch": 0.24102564102564103, "grad_norm": 2.1718826293945312, "learning_rate": 7.654639175257732e-06, "loss": 1.1543, "step": 94 }, { "epoch": 0.24358974358974358, "grad_norm": 2.0842125415802, "learning_rate": 7.628865979381444e-06, "loss": 0.9058, "step": 95 }, { "epoch": 0.24615384615384617, "grad_norm": 1.8005638122558594, "learning_rate": 7.603092783505155e-06, "loss": 0.9146, "step": 96 }, { "epoch": 0.24871794871794872, "grad_norm": 1.6228526830673218, "learning_rate": 7.577319587628866e-06, "loss": 0.7095, "step": 97 }, { "epoch": 0.2512820512820513, "grad_norm": 2.022739887237549, "learning_rate": 7.551546391752578e-06, "loss": 0.9229, "step": 98 }, { "epoch": 0.25384615384615383, "grad_norm": 1.7935630083084106, "learning_rate": 7.525773195876289e-06, "loss": 0.7896, "step": 99 }, { "epoch": 0.2564102564102564, "grad_norm": 1.747321367263794, "learning_rate": 7.500000000000001e-06, "loss": 0.9551, "step": 100 }, { "epoch": 0.258974358974359, "grad_norm": 1.8076218366622925, "learning_rate": 7.474226804123712e-06, "loss": 0.8394, "step": 101 }, { "epoch": 0.26153846153846155, "grad_norm": 1.730443000793457, "learning_rate": 7.448453608247424e-06, "loss": 0.9062, "step": 102 }, { "epoch": 0.2641025641025641, "grad_norm": 2.290266275405884, "learning_rate": 7.422680412371135e-06, "loss": 1.248, "step": 103 }, { "epoch": 0.26666666666666666, "grad_norm": 1.772620677947998, "learning_rate": 7.396907216494846e-06, "loss": 0.8296, "step": 104 }, { "epoch": 0.2692307692307692, "grad_norm": 1.7270923852920532, "learning_rate": 7.3711340206185574e-06, "loss": 0.8735, "step": 105 }, { "epoch": 0.2717948717948718, "grad_norm": 3.0547115802764893, "learning_rate": 7.3453608247422696e-06, "loss": 0.9385, "step": 106 }, { "epoch": 0.2743589743589744, "grad_norm": 2.1224873065948486, "learning_rate": 7.319587628865979e-06, "loss": 1.0918, "step": 107 }, { "epoch": 0.27692307692307694, "grad_norm": 2.175182819366455, "learning_rate": 7.293814432989691e-06, "loss": 0.9707, "step": 108 }, { "epoch": 0.2794871794871795, "grad_norm": 1.8396246433258057, "learning_rate": 7.2680412371134026e-06, "loss": 0.9609, "step": 109 }, { "epoch": 0.28205128205128205, "grad_norm": 1.8551801443099976, "learning_rate": 7.242268041237114e-06, "loss": 0.8916, "step": 110 }, { "epoch": 0.2846153846153846, "grad_norm": 1.960680603981018, "learning_rate": 7.216494845360825e-06, "loss": 1.0352, "step": 111 }, { "epoch": 0.28717948717948716, "grad_norm": 1.8434001207351685, "learning_rate": 7.190721649484536e-06, "loss": 1.0576, "step": 112 }, { "epoch": 0.28974358974358977, "grad_norm": 2.5577752590179443, "learning_rate": 7.164948453608248e-06, "loss": 1.0762, "step": 113 }, { "epoch": 0.2923076923076923, "grad_norm": 1.8725932836532593, "learning_rate": 7.13917525773196e-06, "loss": 0.79, "step": 114 }, { "epoch": 0.2948717948717949, "grad_norm": 2.0246224403381348, "learning_rate": 7.113402061855671e-06, "loss": 0.9058, "step": 115 }, { "epoch": 0.29743589743589743, "grad_norm": 1.8787579536437988, "learning_rate": 7.087628865979382e-06, "loss": 0.8569, "step": 116 }, { "epoch": 0.3, "grad_norm": 1.6573154926300049, "learning_rate": 7.061855670103094e-06, "loss": 0.8091, "step": 117 }, { "epoch": 0.30256410256410254, "grad_norm": 1.7937979698181152, "learning_rate": 7.036082474226805e-06, "loss": 0.978, "step": 118 }, { "epoch": 0.30512820512820515, "grad_norm": 1.9054224491119385, "learning_rate": 7.010309278350515e-06, "loss": 1.0635, "step": 119 }, { "epoch": 0.3076923076923077, "grad_norm": 1.6725414991378784, "learning_rate": 6.984536082474227e-06, "loss": 1.0312, "step": 120 }, { "epoch": 0.31025641025641026, "grad_norm": 1.8967286348342896, "learning_rate": 6.958762886597939e-06, "loss": 0.9746, "step": 121 }, { "epoch": 0.3128205128205128, "grad_norm": 2.158597230911255, "learning_rate": 6.93298969072165e-06, "loss": 1.1309, "step": 122 }, { "epoch": 0.3153846153846154, "grad_norm": 2.2662479877471924, "learning_rate": 6.907216494845361e-06, "loss": 1.0674, "step": 123 }, { "epoch": 0.31794871794871793, "grad_norm": 1.9628630876541138, "learning_rate": 6.881443298969073e-06, "loss": 0.9644, "step": 124 }, { "epoch": 0.32051282051282054, "grad_norm": 1.7974278926849365, "learning_rate": 6.855670103092784e-06, "loss": 0.8877, "step": 125 }, { "epoch": 0.3230769230769231, "grad_norm": 1.7548365592956543, "learning_rate": 6.829896907216495e-06, "loss": 0.9204, "step": 126 }, { "epoch": 0.32564102564102565, "grad_norm": 1.9516572952270508, "learning_rate": 6.804123711340207e-06, "loss": 1.1602, "step": 127 }, { "epoch": 0.3282051282051282, "grad_norm": 2.1989433765411377, "learning_rate": 6.778350515463919e-06, "loss": 1.1133, "step": 128 }, { "epoch": 0.33076923076923076, "grad_norm": 1.800710916519165, "learning_rate": 6.75257731958763e-06, "loss": 0.791, "step": 129 }, { "epoch": 0.3333333333333333, "grad_norm": 3.385741710662842, "learning_rate": 6.726804123711341e-06, "loss": 1.1758, "step": 130 }, { "epoch": 0.33589743589743587, "grad_norm": 2.0922465324401855, "learning_rate": 6.701030927835052e-06, "loss": 0.9214, "step": 131 }, { "epoch": 0.3384615384615385, "grad_norm": 1.573805332183838, "learning_rate": 6.675257731958763e-06, "loss": 0.7715, "step": 132 }, { "epoch": 0.34102564102564104, "grad_norm": 1.7901231050491333, "learning_rate": 6.649484536082474e-06, "loss": 0.9443, "step": 133 }, { "epoch": 0.3435897435897436, "grad_norm": 1.6376028060913086, "learning_rate": 6.623711340206186e-06, "loss": 0.9854, "step": 134 }, { "epoch": 0.34615384615384615, "grad_norm": 2.0153403282165527, "learning_rate": 6.597938144329898e-06, "loss": 1.2373, "step": 135 }, { "epoch": 0.3487179487179487, "grad_norm": 1.819011926651001, "learning_rate": 6.572164948453609e-06, "loss": 0.9048, "step": 136 }, { "epoch": 0.35128205128205126, "grad_norm": 1.623148798942566, "learning_rate": 6.54639175257732e-06, "loss": 0.8262, "step": 137 }, { "epoch": 0.35384615384615387, "grad_norm": 2.2499585151672363, "learning_rate": 6.520618556701031e-06, "loss": 1.1152, "step": 138 }, { "epoch": 0.3564102564102564, "grad_norm": 1.9235867261886597, "learning_rate": 6.494845360824743e-06, "loss": 0.9067, "step": 139 }, { "epoch": 0.358974358974359, "grad_norm": 1.9163473844528198, "learning_rate": 6.469072164948455e-06, "loss": 1.0469, "step": 140 }, { "epoch": 0.36153846153846153, "grad_norm": 2.092963457107544, "learning_rate": 6.443298969072166e-06, "loss": 1.0352, "step": 141 }, { "epoch": 0.3641025641025641, "grad_norm": 9.061576843261719, "learning_rate": 6.417525773195877e-06, "loss": 0.833, "step": 142 }, { "epoch": 0.36666666666666664, "grad_norm": 2.0472512245178223, "learning_rate": 6.391752577319588e-06, "loss": 1.043, "step": 143 }, { "epoch": 0.36923076923076925, "grad_norm": 1.9477477073669434, "learning_rate": 6.365979381443299e-06, "loss": 0.9644, "step": 144 }, { "epoch": 0.3717948717948718, "grad_norm": 1.9295361042022705, "learning_rate": 6.34020618556701e-06, "loss": 0.9746, "step": 145 }, { "epoch": 0.37435897435897436, "grad_norm": 1.9540655612945557, "learning_rate": 6.314432989690722e-06, "loss": 0.9375, "step": 146 }, { "epoch": 0.3769230769230769, "grad_norm": 1.9495689868927002, "learning_rate": 6.288659793814433e-06, "loss": 1.123, "step": 147 }, { "epoch": 0.37948717948717947, "grad_norm": 1.8979523181915283, "learning_rate": 6.262886597938145e-06, "loss": 1.0762, "step": 148 }, { "epoch": 0.382051282051282, "grad_norm": 2.086167573928833, "learning_rate": 6.237113402061856e-06, "loss": 0.8975, "step": 149 }, { "epoch": 0.38461538461538464, "grad_norm": 1.9562528133392334, "learning_rate": 6.211340206185568e-06, "loss": 0.9429, "step": 150 }, { "epoch": 0.3871794871794872, "grad_norm": 2.3442628383636475, "learning_rate": 6.185567010309279e-06, "loss": 0.9844, "step": 151 }, { "epoch": 0.38974358974358975, "grad_norm": 1.8352185487747192, "learning_rate": 6.15979381443299e-06, "loss": 1.1045, "step": 152 }, { "epoch": 0.3923076923076923, "grad_norm": 3.1961944103240967, "learning_rate": 6.134020618556702e-06, "loss": 1.0596, "step": 153 }, { "epoch": 0.39487179487179486, "grad_norm": 1.8429200649261475, "learning_rate": 6.108247422680414e-06, "loss": 0.9365, "step": 154 }, { "epoch": 0.3974358974358974, "grad_norm": 1.929801106452942, "learning_rate": 6.082474226804124e-06, "loss": 0.9883, "step": 155 }, { "epoch": 0.4, "grad_norm": 1.670444369316101, "learning_rate": 6.056701030927835e-06, "loss": 0.7524, "step": 156 }, { "epoch": 0.4025641025641026, "grad_norm": 2.022891044616699, "learning_rate": 6.030927835051547e-06, "loss": 1.2197, "step": 157 }, { "epoch": 0.40512820512820513, "grad_norm": 1.7056248188018799, "learning_rate": 6.005154639175258e-06, "loss": 0.9121, "step": 158 }, { "epoch": 0.4076923076923077, "grad_norm": 2.0686631202697754, "learning_rate": 5.979381443298969e-06, "loss": 1.1865, "step": 159 }, { "epoch": 0.41025641025641024, "grad_norm": 1.8868600130081177, "learning_rate": 5.9536082474226805e-06, "loss": 0.9658, "step": 160 }, { "epoch": 0.4128205128205128, "grad_norm": 2.3580541610717773, "learning_rate": 5.927835051546393e-06, "loss": 0.9297, "step": 161 }, { "epoch": 0.4153846153846154, "grad_norm": 2.071708917617798, "learning_rate": 5.902061855670104e-06, "loss": 1.0693, "step": 162 }, { "epoch": 0.41794871794871796, "grad_norm": 1.8569782972335815, "learning_rate": 5.876288659793815e-06, "loss": 0.9438, "step": 163 }, { "epoch": 0.4205128205128205, "grad_norm": 1.7496881484985352, "learning_rate": 5.8505154639175264e-06, "loss": 0.8574, "step": 164 }, { "epoch": 0.4230769230769231, "grad_norm": 1.825770378112793, "learning_rate": 5.824742268041238e-06, "loss": 0.9678, "step": 165 }, { "epoch": 0.4256410256410256, "grad_norm": 2.156632423400879, "learning_rate": 5.79896907216495e-06, "loss": 1.2031, "step": 166 }, { "epoch": 0.4282051282051282, "grad_norm": 1.6969801187515259, "learning_rate": 5.7731958762886594e-06, "loss": 0.8359, "step": 167 }, { "epoch": 0.4307692307692308, "grad_norm": 2.4445745944976807, "learning_rate": 5.7474226804123716e-06, "loss": 1.0293, "step": 168 }, { "epoch": 0.43333333333333335, "grad_norm": 1.9905025959014893, "learning_rate": 5.721649484536083e-06, "loss": 1.0791, "step": 169 }, { "epoch": 0.4358974358974359, "grad_norm": 1.9443804025650024, "learning_rate": 5.695876288659794e-06, "loss": 0.9702, "step": 170 }, { "epoch": 0.43846153846153846, "grad_norm": 2.2694649696350098, "learning_rate": 5.670103092783505e-06, "loss": 1.1914, "step": 171 }, { "epoch": 0.441025641025641, "grad_norm": 2.1340649127960205, "learning_rate": 5.644329896907217e-06, "loss": 1.0371, "step": 172 }, { "epoch": 0.44358974358974357, "grad_norm": 1.9536010026931763, "learning_rate": 5.618556701030928e-06, "loss": 0.9888, "step": 173 }, { "epoch": 0.4461538461538462, "grad_norm": 1.837241530418396, "learning_rate": 5.59278350515464e-06, "loss": 0.9033, "step": 174 }, { "epoch": 0.44871794871794873, "grad_norm": 1.8256818056106567, "learning_rate": 5.567010309278351e-06, "loss": 1.0117, "step": 175 }, { "epoch": 0.4512820512820513, "grad_norm": 2.1817359924316406, "learning_rate": 5.541237113402063e-06, "loss": 0.8506, "step": 176 }, { "epoch": 0.45384615384615384, "grad_norm": 1.7470033168792725, "learning_rate": 5.515463917525774e-06, "loss": 0.9473, "step": 177 }, { "epoch": 0.4564102564102564, "grad_norm": 1.8381420373916626, "learning_rate": 5.489690721649485e-06, "loss": 0.8828, "step": 178 }, { "epoch": 0.45897435897435895, "grad_norm": 1.9407833814620972, "learning_rate": 5.463917525773196e-06, "loss": 1.3193, "step": 179 }, { "epoch": 0.46153846153846156, "grad_norm": 1.9037673473358154, "learning_rate": 5.438144329896907e-06, "loss": 0.916, "step": 180 }, { "epoch": 0.4641025641025641, "grad_norm": 1.8702125549316406, "learning_rate": 5.412371134020619e-06, "loss": 1.0039, "step": 181 }, { "epoch": 0.4666666666666667, "grad_norm": 2.0988097190856934, "learning_rate": 5.38659793814433e-06, "loss": 1.0225, "step": 182 }, { "epoch": 0.46923076923076923, "grad_norm": 1.9512522220611572, "learning_rate": 5.360824742268042e-06, "loss": 0.9854, "step": 183 }, { "epoch": 0.4717948717948718, "grad_norm": 1.8114992380142212, "learning_rate": 5.335051546391753e-06, "loss": 0.8008, "step": 184 }, { "epoch": 0.47435897435897434, "grad_norm": 1.7766282558441162, "learning_rate": 5.309278350515464e-06, "loss": 0.9795, "step": 185 }, { "epoch": 0.47692307692307695, "grad_norm": 2.319395065307617, "learning_rate": 5.2835051546391755e-06, "loss": 1.1572, "step": 186 }, { "epoch": 0.4794871794871795, "grad_norm": 1.7959656715393066, "learning_rate": 5.257731958762888e-06, "loss": 0.8994, "step": 187 }, { "epoch": 0.48205128205128206, "grad_norm": 1.7262598276138306, "learning_rate": 5.231958762886599e-06, "loss": 1.0264, "step": 188 }, { "epoch": 0.4846153846153846, "grad_norm": 1.9442336559295654, "learning_rate": 5.20618556701031e-06, "loss": 1.0674, "step": 189 }, { "epoch": 0.48717948717948717, "grad_norm": 1.7376888990402222, "learning_rate": 5.1804123711340214e-06, "loss": 0.8032, "step": 190 }, { "epoch": 0.4897435897435897, "grad_norm": 1.5488858222961426, "learning_rate": 5.154639175257732e-06, "loss": 0.8101, "step": 191 }, { "epoch": 0.49230769230769234, "grad_norm": 1.9175901412963867, "learning_rate": 5.128865979381443e-06, "loss": 0.8418, "step": 192 }, { "epoch": 0.4948717948717949, "grad_norm": 2.069321393966675, "learning_rate": 5.1030927835051544e-06, "loss": 1.1826, "step": 193 }, { "epoch": 0.49743589743589745, "grad_norm": 1.918543815612793, "learning_rate": 5.077319587628866e-06, "loss": 1.0059, "step": 194 }, { "epoch": 0.5, "grad_norm": 1.7345309257507324, "learning_rate": 5.051546391752578e-06, "loss": 0.9028, "step": 195 }, { "epoch": 0.5025641025641026, "grad_norm": 1.7581024169921875, "learning_rate": 5.025773195876289e-06, "loss": 1.0, "step": 196 }, { "epoch": 0.5051282051282051, "grad_norm": 2.021634817123413, "learning_rate": 5e-06, "loss": 0.8677, "step": 197 }, { "epoch": 0.5076923076923077, "grad_norm": 2.0879619121551514, "learning_rate": 4.974226804123712e-06, "loss": 1.1279, "step": 198 }, { "epoch": 0.5102564102564102, "grad_norm": 2.040804386138916, "learning_rate": 4.948453608247423e-06, "loss": 1.2178, "step": 199 }, { "epoch": 0.5128205128205128, "grad_norm": 2.0652830600738525, "learning_rate": 4.922680412371135e-06, "loss": 0.873, "step": 200 }, { "epoch": 0.5153846153846153, "grad_norm": 1.8770358562469482, "learning_rate": 4.8969072164948455e-06, "loss": 0.7212, "step": 201 }, { "epoch": 0.517948717948718, "grad_norm": 1.8745349645614624, "learning_rate": 4.871134020618557e-06, "loss": 1.126, "step": 202 }, { "epoch": 0.5205128205128206, "grad_norm": 1.8858857154846191, "learning_rate": 4.845360824742268e-06, "loss": 1.0771, "step": 203 }, { "epoch": 0.5230769230769231, "grad_norm": 2.110069513320923, "learning_rate": 4.81958762886598e-06, "loss": 1.2754, "step": 204 }, { "epoch": 0.5256410256410257, "grad_norm": 4.52376127243042, "learning_rate": 4.7938144329896915e-06, "loss": 1.1182, "step": 205 }, { "epoch": 0.5282051282051282, "grad_norm": 1.6591471433639526, "learning_rate": 4.768041237113403e-06, "loss": 0.8276, "step": 206 }, { "epoch": 0.5307692307692308, "grad_norm": 1.9472140073776245, "learning_rate": 4.742268041237113e-06, "loss": 1.0273, "step": 207 }, { "epoch": 0.5333333333333333, "grad_norm": 1.8485890626907349, "learning_rate": 4.716494845360825e-06, "loss": 0.813, "step": 208 }, { "epoch": 0.5358974358974359, "grad_norm": 1.7967491149902344, "learning_rate": 4.690721649484537e-06, "loss": 0.9946, "step": 209 }, { "epoch": 0.5384615384615384, "grad_norm": 1.7534973621368408, "learning_rate": 4.664948453608248e-06, "loss": 0.7993, "step": 210 }, { "epoch": 0.541025641025641, "grad_norm": 1.8620507717132568, "learning_rate": 4.639175257731959e-06, "loss": 0.9341, "step": 211 }, { "epoch": 0.5435897435897435, "grad_norm": 1.8512839078903198, "learning_rate": 4.6134020618556705e-06, "loss": 0.8013, "step": 212 }, { "epoch": 0.5461538461538461, "grad_norm": 2.151174306869507, "learning_rate": 4.587628865979382e-06, "loss": 1.2969, "step": 213 }, { "epoch": 0.5487179487179488, "grad_norm": 1.9421318769454956, "learning_rate": 4.561855670103093e-06, "loss": 0.9375, "step": 214 }, { "epoch": 0.5512820512820513, "grad_norm": 1.6634801626205444, "learning_rate": 4.536082474226804e-06, "loss": 0.7603, "step": 215 }, { "epoch": 0.5538461538461539, "grad_norm": 1.8529914617538452, "learning_rate": 4.510309278350516e-06, "loss": 0.8545, "step": 216 }, { "epoch": 0.5564102564102564, "grad_norm": 1.8996697664260864, "learning_rate": 4.484536082474228e-06, "loss": 0.9492, "step": 217 }, { "epoch": 0.558974358974359, "grad_norm": 1.793915033340454, "learning_rate": 4.458762886597939e-06, "loss": 0.6826, "step": 218 }, { "epoch": 0.5615384615384615, "grad_norm": 2.973825454711914, "learning_rate": 4.4329896907216494e-06, "loss": 0.9023, "step": 219 }, { "epoch": 0.5641025641025641, "grad_norm": 1.786086082458496, "learning_rate": 4.407216494845361e-06, "loss": 0.9502, "step": 220 }, { "epoch": 0.5666666666666667, "grad_norm": 2.263026714324951, "learning_rate": 4.381443298969073e-06, "loss": 0.9326, "step": 221 }, { "epoch": 0.5692307692307692, "grad_norm": 2.35546612739563, "learning_rate": 4.355670103092784e-06, "loss": 1.1348, "step": 222 }, { "epoch": 0.5717948717948718, "grad_norm": 1.6735016107559204, "learning_rate": 4.329896907216495e-06, "loss": 0.9204, "step": 223 }, { "epoch": 0.5743589743589743, "grad_norm": 2.4395010471343994, "learning_rate": 4.304123711340207e-06, "loss": 1.1689, "step": 224 }, { "epoch": 0.5769230769230769, "grad_norm": 1.8484439849853516, "learning_rate": 4.278350515463918e-06, "loss": 0.9751, "step": 225 }, { "epoch": 0.5794871794871795, "grad_norm": 2.61309552192688, "learning_rate": 4.252577319587629e-06, "loss": 0.9272, "step": 226 }, { "epoch": 0.5820512820512821, "grad_norm": 1.8143435716629028, "learning_rate": 4.2268041237113405e-06, "loss": 1.0508, "step": 227 }, { "epoch": 0.5846153846153846, "grad_norm": 1.9212270975112915, "learning_rate": 4.201030927835052e-06, "loss": 0.8813, "step": 228 }, { "epoch": 0.5871794871794872, "grad_norm": 2.3576104640960693, "learning_rate": 4.175257731958763e-06, "loss": 0.98, "step": 229 }, { "epoch": 0.5897435897435898, "grad_norm": 2.2258710861206055, "learning_rate": 4.149484536082475e-06, "loss": 0.9766, "step": 230 }, { "epoch": 0.5923076923076923, "grad_norm": 1.8784958124160767, "learning_rate": 4.123711340206186e-06, "loss": 0.9575, "step": 231 }, { "epoch": 0.5948717948717949, "grad_norm": 1.7360793352127075, "learning_rate": 4.097938144329897e-06, "loss": 0.9014, "step": 232 }, { "epoch": 0.5974358974358974, "grad_norm": 2.718904733657837, "learning_rate": 4.072164948453608e-06, "loss": 0.8682, "step": 233 }, { "epoch": 0.6, "grad_norm": 3.4316585063934326, "learning_rate": 4.04639175257732e-06, "loss": 0.8096, "step": 234 }, { "epoch": 0.6025641025641025, "grad_norm": 1.9802138805389404, "learning_rate": 4.020618556701032e-06, "loss": 1.1055, "step": 235 }, { "epoch": 0.6051282051282051, "grad_norm": 2.3054115772247314, "learning_rate": 3.994845360824743e-06, "loss": 0.8696, "step": 236 }, { "epoch": 0.6076923076923076, "grad_norm": 2.1070103645324707, "learning_rate": 3.969072164948453e-06, "loss": 0.9907, "step": 237 }, { "epoch": 0.6102564102564103, "grad_norm": 1.71999192237854, "learning_rate": 3.9432989690721655e-06, "loss": 1.0342, "step": 238 }, { "epoch": 0.6128205128205129, "grad_norm": 1.7554974555969238, "learning_rate": 3.917525773195877e-06, "loss": 0.9712, "step": 239 }, { "epoch": 0.6153846153846154, "grad_norm": 1.8095436096191406, "learning_rate": 3.891752577319588e-06, "loss": 1.0586, "step": 240 }, { "epoch": 0.617948717948718, "grad_norm": 2.331573963165283, "learning_rate": 3.865979381443299e-06, "loss": 1.1729, "step": 241 }, { "epoch": 0.6205128205128205, "grad_norm": 1.8929247856140137, "learning_rate": 3.840206185567011e-06, "loss": 1.0576, "step": 242 }, { "epoch": 0.6230769230769231, "grad_norm": 1.6739861965179443, "learning_rate": 3.814432989690722e-06, "loss": 0.8008, "step": 243 }, { "epoch": 0.6256410256410256, "grad_norm": 1.6616827249526978, "learning_rate": 3.788659793814433e-06, "loss": 0.876, "step": 244 }, { "epoch": 0.6282051282051282, "grad_norm": 7.3671746253967285, "learning_rate": 3.7628865979381445e-06, "loss": 1.0293, "step": 245 }, { "epoch": 0.6307692307692307, "grad_norm": 1.9617962837219238, "learning_rate": 3.737113402061856e-06, "loss": 1.0117, "step": 246 }, { "epoch": 0.6333333333333333, "grad_norm": 1.3733716011047363, "learning_rate": 3.7113402061855674e-06, "loss": 0.6865, "step": 247 }, { "epoch": 0.6358974358974359, "grad_norm": 1.6532529592514038, "learning_rate": 3.6855670103092787e-06, "loss": 0.9604, "step": 248 }, { "epoch": 0.6384615384615384, "grad_norm": 1.8402736186981201, "learning_rate": 3.6597938144329896e-06, "loss": 0.916, "step": 249 }, { "epoch": 0.6410256410256411, "grad_norm": 1.7600955963134766, "learning_rate": 3.6340206185567013e-06, "loss": 0.9678, "step": 250 }, { "epoch": 0.6435897435897436, "grad_norm": 1.6891589164733887, "learning_rate": 3.6082474226804126e-06, "loss": 0.793, "step": 251 }, { "epoch": 0.6461538461538462, "grad_norm": 1.8142110109329224, "learning_rate": 3.582474226804124e-06, "loss": 1.0518, "step": 252 }, { "epoch": 0.6487179487179487, "grad_norm": 2.1421236991882324, "learning_rate": 3.5567010309278356e-06, "loss": 0.9414, "step": 253 }, { "epoch": 0.6512820512820513, "grad_norm": 5.681818962097168, "learning_rate": 3.530927835051547e-06, "loss": 0.9951, "step": 254 }, { "epoch": 0.6538461538461539, "grad_norm": 2.093968391418457, "learning_rate": 3.5051546391752577e-06, "loss": 0.7842, "step": 255 }, { "epoch": 0.6564102564102564, "grad_norm": 1.8707085847854614, "learning_rate": 3.4793814432989694e-06, "loss": 0.9858, "step": 256 }, { "epoch": 0.658974358974359, "grad_norm": 1.9755574464797974, "learning_rate": 3.4536082474226807e-06, "loss": 0.9136, "step": 257 }, { "epoch": 0.6615384615384615, "grad_norm": 2.0634946823120117, "learning_rate": 3.427835051546392e-06, "loss": 0.7593, "step": 258 }, { "epoch": 0.6641025641025641, "grad_norm": 1.842301845550537, "learning_rate": 3.4020618556701037e-06, "loss": 0.8062, "step": 259 }, { "epoch": 0.6666666666666666, "grad_norm": 1.857254981994629, "learning_rate": 3.376288659793815e-06, "loss": 1.2021, "step": 260 }, { "epoch": 0.6692307692307692, "grad_norm": 1.9125348329544067, "learning_rate": 3.350515463917526e-06, "loss": 1.084, "step": 261 }, { "epoch": 0.6717948717948717, "grad_norm": 1.8903979063034058, "learning_rate": 3.324742268041237e-06, "loss": 0.7666, "step": 262 }, { "epoch": 0.6743589743589744, "grad_norm": 1.981501579284668, "learning_rate": 3.298969072164949e-06, "loss": 1.0508, "step": 263 }, { "epoch": 0.676923076923077, "grad_norm": 1.5312262773513794, "learning_rate": 3.27319587628866e-06, "loss": 0.7231, "step": 264 }, { "epoch": 0.6794871794871795, "grad_norm": 1.6376142501831055, "learning_rate": 3.2474226804123714e-06, "loss": 0.8013, "step": 265 }, { "epoch": 0.6820512820512821, "grad_norm": 1.5917991399765015, "learning_rate": 3.221649484536083e-06, "loss": 0.8042, "step": 266 }, { "epoch": 0.6846153846153846, "grad_norm": 1.497527837753296, "learning_rate": 3.195876288659794e-06, "loss": 0.7031, "step": 267 }, { "epoch": 0.6871794871794872, "grad_norm": 1.692023515701294, "learning_rate": 3.170103092783505e-06, "loss": 0.6646, "step": 268 }, { "epoch": 0.6897435897435897, "grad_norm": 1.862017273902893, "learning_rate": 3.1443298969072165e-06, "loss": 0.8765, "step": 269 }, { "epoch": 0.6923076923076923, "grad_norm": 2.128854513168335, "learning_rate": 3.118556701030928e-06, "loss": 0.7183, "step": 270 }, { "epoch": 0.6948717948717948, "grad_norm": 1.7785433530807495, "learning_rate": 3.0927835051546395e-06, "loss": 0.9966, "step": 271 }, { "epoch": 0.6974358974358974, "grad_norm": 1.8306283950805664, "learning_rate": 3.067010309278351e-06, "loss": 0.7773, "step": 272 }, { "epoch": 0.7, "grad_norm": 1.5251365900039673, "learning_rate": 3.041237113402062e-06, "loss": 0.7046, "step": 273 }, { "epoch": 0.7025641025641025, "grad_norm": 2.0568642616271973, "learning_rate": 3.0154639175257733e-06, "loss": 1.0459, "step": 274 }, { "epoch": 0.7051282051282052, "grad_norm": 1.6169660091400146, "learning_rate": 2.9896907216494846e-06, "loss": 0.7451, "step": 275 }, { "epoch": 0.7076923076923077, "grad_norm": 1.6986936330795288, "learning_rate": 2.9639175257731963e-06, "loss": 0.8735, "step": 276 }, { "epoch": 0.7102564102564103, "grad_norm": 1.6713733673095703, "learning_rate": 2.9381443298969076e-06, "loss": 0.8862, "step": 277 }, { "epoch": 0.7128205128205128, "grad_norm": 1.7637569904327393, "learning_rate": 2.912371134020619e-06, "loss": 0.9609, "step": 278 }, { "epoch": 0.7153846153846154, "grad_norm": 2.076930284500122, "learning_rate": 2.8865979381443297e-06, "loss": 1.1377, "step": 279 }, { "epoch": 0.717948717948718, "grad_norm": 1.972031831741333, "learning_rate": 2.8608247422680414e-06, "loss": 0.9395, "step": 280 }, { "epoch": 0.7205128205128205, "grad_norm": 1.8608795404434204, "learning_rate": 2.8350515463917527e-06, "loss": 1.0654, "step": 281 }, { "epoch": 0.7230769230769231, "grad_norm": 3.467540979385376, "learning_rate": 2.809278350515464e-06, "loss": 0.7603, "step": 282 }, { "epoch": 0.7256410256410256, "grad_norm": 1.765555739402771, "learning_rate": 2.7835051546391757e-06, "loss": 0.6758, "step": 283 }, { "epoch": 0.7282051282051282, "grad_norm": 1.6693044900894165, "learning_rate": 2.757731958762887e-06, "loss": 0.8433, "step": 284 }, { "epoch": 0.7307692307692307, "grad_norm": 1.9119174480438232, "learning_rate": 2.731958762886598e-06, "loss": 0.8184, "step": 285 }, { "epoch": 0.7333333333333333, "grad_norm": 2.043612241744995, "learning_rate": 2.7061855670103095e-06, "loss": 1.0049, "step": 286 }, { "epoch": 0.735897435897436, "grad_norm": 1.919756531715393, "learning_rate": 2.680412371134021e-06, "loss": 0.9229, "step": 287 }, { "epoch": 0.7384615384615385, "grad_norm": 1.8900898694992065, "learning_rate": 2.654639175257732e-06, "loss": 1.0078, "step": 288 }, { "epoch": 0.7410256410256411, "grad_norm": 1.793310284614563, "learning_rate": 2.628865979381444e-06, "loss": 1.0752, "step": 289 }, { "epoch": 0.7435897435897436, "grad_norm": 2.732642889022827, "learning_rate": 2.603092783505155e-06, "loss": 0.9229, "step": 290 }, { "epoch": 0.7461538461538462, "grad_norm": 1.845354676246643, "learning_rate": 2.577319587628866e-06, "loss": 0.8286, "step": 291 }, { "epoch": 0.7487179487179487, "grad_norm": 1.9272360801696777, "learning_rate": 2.5515463917525772e-06, "loss": 1.0518, "step": 292 }, { "epoch": 0.7512820512820513, "grad_norm": 2.1714913845062256, "learning_rate": 2.525773195876289e-06, "loss": 0.9043, "step": 293 }, { "epoch": 0.7538461538461538, "grad_norm": 1.7473788261413574, "learning_rate": 2.5e-06, "loss": 0.7793, "step": 294 }, { "epoch": 0.7564102564102564, "grad_norm": 1.9058947563171387, "learning_rate": 2.4742268041237115e-06, "loss": 0.8643, "step": 295 }, { "epoch": 0.7589743589743589, "grad_norm": 2.064485788345337, "learning_rate": 2.4484536082474228e-06, "loss": 1.1475, "step": 296 }, { "epoch": 0.7615384615384615, "grad_norm": 2.100177526473999, "learning_rate": 2.422680412371134e-06, "loss": 0.853, "step": 297 }, { "epoch": 0.764102564102564, "grad_norm": 2.2327897548675537, "learning_rate": 2.3969072164948458e-06, "loss": 1.2168, "step": 298 }, { "epoch": 0.7666666666666667, "grad_norm": 1.9748061895370483, "learning_rate": 2.3711340206185566e-06, "loss": 1.0713, "step": 299 }, { "epoch": 0.7692307692307693, "grad_norm": 1.987243890762329, "learning_rate": 2.3453608247422683e-06, "loss": 1.0508, "step": 300 }, { "epoch": 0.7717948717948718, "grad_norm": 2.191959857940674, "learning_rate": 2.3195876288659796e-06, "loss": 0.9326, "step": 301 }, { "epoch": 0.7743589743589744, "grad_norm": 2.242733955383301, "learning_rate": 2.293814432989691e-06, "loss": 0.7515, "step": 302 }, { "epoch": 0.7769230769230769, "grad_norm": 1.6882741451263428, "learning_rate": 2.268041237113402e-06, "loss": 0.7534, "step": 303 }, { "epoch": 0.7794871794871795, "grad_norm": 2.7890372276306152, "learning_rate": 2.242268041237114e-06, "loss": 0.9082, "step": 304 }, { "epoch": 0.782051282051282, "grad_norm": 2.7322838306427, "learning_rate": 2.2164948453608247e-06, "loss": 1.0332, "step": 305 }, { "epoch": 0.7846153846153846, "grad_norm": 1.9101502895355225, "learning_rate": 2.1907216494845364e-06, "loss": 1.0088, "step": 306 }, { "epoch": 0.7871794871794872, "grad_norm": 1.7795121669769287, "learning_rate": 2.1649484536082477e-06, "loss": 0.9082, "step": 307 }, { "epoch": 0.7897435897435897, "grad_norm": 1.8345370292663574, "learning_rate": 2.139175257731959e-06, "loss": 1.0576, "step": 308 }, { "epoch": 0.7923076923076923, "grad_norm": 2.1167471408843994, "learning_rate": 2.1134020618556703e-06, "loss": 0.918, "step": 309 }, { "epoch": 0.7948717948717948, "grad_norm": 1.7261496782302856, "learning_rate": 2.0876288659793816e-06, "loss": 0.8975, "step": 310 }, { "epoch": 0.7974358974358975, "grad_norm": 1.793904185295105, "learning_rate": 2.061855670103093e-06, "loss": 0.8916, "step": 311 }, { "epoch": 0.8, "grad_norm": 1.9415448904037476, "learning_rate": 2.036082474226804e-06, "loss": 0.8379, "step": 312 }, { "epoch": 0.8025641025641026, "grad_norm": 1.8090825080871582, "learning_rate": 2.010309278350516e-06, "loss": 0.8784, "step": 313 }, { "epoch": 0.8051282051282052, "grad_norm": 1.775429368019104, "learning_rate": 1.9845360824742267e-06, "loss": 0.8643, "step": 314 }, { "epoch": 0.8076923076923077, "grad_norm": 1.848243236541748, "learning_rate": 1.9587628865979384e-06, "loss": 1.1084, "step": 315 }, { "epoch": 0.8102564102564103, "grad_norm": 1.7999858856201172, "learning_rate": 1.9329896907216497e-06, "loss": 0.9321, "step": 316 }, { "epoch": 0.8128205128205128, "grad_norm": 2.8370208740234375, "learning_rate": 1.907216494845361e-06, "loss": 0.9883, "step": 317 }, { "epoch": 0.8153846153846154, "grad_norm": 2.2633893489837646, "learning_rate": 1.8814432989690722e-06, "loss": 1.0166, "step": 318 }, { "epoch": 0.8179487179487179, "grad_norm": 1.9850558042526245, "learning_rate": 1.8556701030927837e-06, "loss": 1.0283, "step": 319 }, { "epoch": 0.8205128205128205, "grad_norm": 1.904843807220459, "learning_rate": 1.8298969072164948e-06, "loss": 0.9829, "step": 320 }, { "epoch": 0.823076923076923, "grad_norm": 1.7778921127319336, "learning_rate": 1.8041237113402063e-06, "loss": 0.9229, "step": 321 }, { "epoch": 0.8256410256410256, "grad_norm": 1.792877197265625, "learning_rate": 1.7783505154639178e-06, "loss": 0.876, "step": 322 }, { "epoch": 0.8282051282051283, "grad_norm": 1.8950697183609009, "learning_rate": 1.7525773195876288e-06, "loss": 0.915, "step": 323 }, { "epoch": 0.8307692307692308, "grad_norm": 1.911402940750122, "learning_rate": 1.7268041237113403e-06, "loss": 1.043, "step": 324 }, { "epoch": 0.8333333333333334, "grad_norm": 2.1029365062713623, "learning_rate": 1.7010309278350518e-06, "loss": 0.9243, "step": 325 }, { "epoch": 0.8358974358974359, "grad_norm": 2.052480936050415, "learning_rate": 1.675257731958763e-06, "loss": 1.2256, "step": 326 }, { "epoch": 0.8384615384615385, "grad_norm": 1.7406338453292847, "learning_rate": 1.6494845360824744e-06, "loss": 0.8804, "step": 327 }, { "epoch": 0.841025641025641, "grad_norm": 1.8102054595947266, "learning_rate": 1.6237113402061857e-06, "loss": 0.854, "step": 328 }, { "epoch": 0.8435897435897436, "grad_norm": 1.9868026971817017, "learning_rate": 1.597938144329897e-06, "loss": 0.8491, "step": 329 }, { "epoch": 0.8461538461538461, "grad_norm": 1.774075984954834, "learning_rate": 1.5721649484536082e-06, "loss": 0.8457, "step": 330 }, { "epoch": 0.8487179487179487, "grad_norm": 1.8672053813934326, "learning_rate": 1.5463917525773197e-06, "loss": 1.0176, "step": 331 }, { "epoch": 0.8512820512820513, "grad_norm": 1.5850166082382202, "learning_rate": 1.520618556701031e-06, "loss": 0.6777, "step": 332 }, { "epoch": 0.8538461538461538, "grad_norm": 1.6745821237564087, "learning_rate": 1.4948453608247423e-06, "loss": 0.939, "step": 333 }, { "epoch": 0.8564102564102564, "grad_norm": 1.8641244173049927, "learning_rate": 1.4690721649484538e-06, "loss": 1.0, "step": 334 }, { "epoch": 0.8589743589743589, "grad_norm": 1.6371408700942993, "learning_rate": 1.4432989690721649e-06, "loss": 0.8037, "step": 335 }, { "epoch": 0.8615384615384616, "grad_norm": 2.0685510635375977, "learning_rate": 1.4175257731958764e-06, "loss": 0.9287, "step": 336 }, { "epoch": 0.8641025641025641, "grad_norm": 1.8242462873458862, "learning_rate": 1.3917525773195878e-06, "loss": 0.8877, "step": 337 }, { "epoch": 0.8666666666666667, "grad_norm": 2.243830442428589, "learning_rate": 1.365979381443299e-06, "loss": 0.7764, "step": 338 }, { "epoch": 0.8692307692307693, "grad_norm": 2.188321590423584, "learning_rate": 1.3402061855670104e-06, "loss": 1.125, "step": 339 }, { "epoch": 0.8717948717948718, "grad_norm": 2.412721633911133, "learning_rate": 1.314432989690722e-06, "loss": 1.1016, "step": 340 }, { "epoch": 0.8743589743589744, "grad_norm": 2.149219512939453, "learning_rate": 1.288659793814433e-06, "loss": 1.1104, "step": 341 }, { "epoch": 0.8769230769230769, "grad_norm": 1.603258490562439, "learning_rate": 1.2628865979381445e-06, "loss": 0.7148, "step": 342 }, { "epoch": 0.8794871794871795, "grad_norm": 1.7832096815109253, "learning_rate": 1.2371134020618557e-06, "loss": 0.9902, "step": 343 }, { "epoch": 0.882051282051282, "grad_norm": 1.8631786108016968, "learning_rate": 1.211340206185567e-06, "loss": 0.8569, "step": 344 }, { "epoch": 0.8846153846153846, "grad_norm": 1.6944712400436401, "learning_rate": 1.1855670103092783e-06, "loss": 0.7397, "step": 345 }, { "epoch": 0.8871794871794871, "grad_norm": 1.842955231666565, "learning_rate": 1.1597938144329898e-06, "loss": 0.8457, "step": 346 }, { "epoch": 0.8897435897435897, "grad_norm": 2.5506865978240967, "learning_rate": 1.134020618556701e-06, "loss": 0.9238, "step": 347 }, { "epoch": 0.8923076923076924, "grad_norm": 2.157381534576416, "learning_rate": 1.1082474226804124e-06, "loss": 1.1855, "step": 348 }, { "epoch": 0.8948717948717949, "grad_norm": 1.835581660270691, "learning_rate": 1.0824742268041239e-06, "loss": 0.9375, "step": 349 }, { "epoch": 0.8974358974358975, "grad_norm": 2.591947078704834, "learning_rate": 1.0567010309278351e-06, "loss": 1.0088, "step": 350 }, { "epoch": 0.9, "grad_norm": 1.9977459907531738, "learning_rate": 1.0309278350515464e-06, "loss": 1.0576, "step": 351 }, { "epoch": 0.9025641025641026, "grad_norm": 1.7029050588607788, "learning_rate": 1.005154639175258e-06, "loss": 0.8643, "step": 352 }, { "epoch": 0.9051282051282051, "grad_norm": 1.6559293270111084, "learning_rate": 9.793814432989692e-07, "loss": 0.8652, "step": 353 }, { "epoch": 0.9076923076923077, "grad_norm": 2.0177323818206787, "learning_rate": 9.536082474226805e-07, "loss": 1.0361, "step": 354 }, { "epoch": 0.9102564102564102, "grad_norm": 1.6591455936431885, "learning_rate": 9.278350515463919e-07, "loss": 0.8872, "step": 355 }, { "epoch": 0.9128205128205128, "grad_norm": 2.092712640762329, "learning_rate": 9.020618556701031e-07, "loss": 0.9644, "step": 356 }, { "epoch": 0.9153846153846154, "grad_norm": 1.7917448282241821, "learning_rate": 8.762886597938144e-07, "loss": 0.9399, "step": 357 }, { "epoch": 0.9179487179487179, "grad_norm": 1.9611492156982422, "learning_rate": 8.505154639175259e-07, "loss": 1.292, "step": 358 }, { "epoch": 0.9205128205128205, "grad_norm": 1.7951692342758179, "learning_rate": 8.247422680412372e-07, "loss": 0.9551, "step": 359 }, { "epoch": 0.9230769230769231, "grad_norm": 1.8831959962844849, "learning_rate": 7.989690721649485e-07, "loss": 1.1973, "step": 360 }, { "epoch": 0.9256410256410257, "grad_norm": 1.9080712795257568, "learning_rate": 7.731958762886599e-07, "loss": 0.917, "step": 361 }, { "epoch": 0.9282051282051282, "grad_norm": 2.041674852371216, "learning_rate": 7.474226804123711e-07, "loss": 0.9111, "step": 362 }, { "epoch": 0.9307692307692308, "grad_norm": 1.8067349195480347, "learning_rate": 7.216494845360824e-07, "loss": 0.7915, "step": 363 }, { "epoch": 0.9333333333333333, "grad_norm": 1.733258843421936, "learning_rate": 6.958762886597939e-07, "loss": 0.8726, "step": 364 }, { "epoch": 0.9358974358974359, "grad_norm": 2.1133229732513428, "learning_rate": 6.701030927835052e-07, "loss": 0.9028, "step": 365 }, { "epoch": 0.9384615384615385, "grad_norm": 1.8070833683013916, "learning_rate": 6.443298969072165e-07, "loss": 0.8071, "step": 366 }, { "epoch": 0.941025641025641, "grad_norm": 1.610521912574768, "learning_rate": 6.185567010309279e-07, "loss": 0.686, "step": 367 }, { "epoch": 0.9435897435897436, "grad_norm": 1.8273773193359375, "learning_rate": 5.927835051546392e-07, "loss": 1.042, "step": 368 }, { "epoch": 0.9461538461538461, "grad_norm": 1.6408607959747314, "learning_rate": 5.670103092783505e-07, "loss": 0.8584, "step": 369 }, { "epoch": 0.9487179487179487, "grad_norm": 1.950465440750122, "learning_rate": 5.412371134020619e-07, "loss": 0.9629, "step": 370 }, { "epoch": 0.9512820512820512, "grad_norm": 1.6559863090515137, "learning_rate": 5.154639175257732e-07, "loss": 0.7119, "step": 371 }, { "epoch": 0.9538461538461539, "grad_norm": 1.497727394104004, "learning_rate": 4.896907216494846e-07, "loss": 0.71, "step": 372 }, { "epoch": 0.9564102564102565, "grad_norm": 1.7516483068466187, "learning_rate": 4.6391752577319593e-07, "loss": 0.7275, "step": 373 }, { "epoch": 0.958974358974359, "grad_norm": 1.8231966495513916, "learning_rate": 4.381443298969072e-07, "loss": 1.0664, "step": 374 }, { "epoch": 0.9615384615384616, "grad_norm": 2.928769826889038, "learning_rate": 4.123711340206186e-07, "loss": 0.7983, "step": 375 }, { "epoch": 0.9641025641025641, "grad_norm": 1.5636661052703857, "learning_rate": 3.8659793814432993e-07, "loss": 0.6655, "step": 376 }, { "epoch": 0.9666666666666667, "grad_norm": 1.7315055131912231, "learning_rate": 3.608247422680412e-07, "loss": 0.8994, "step": 377 }, { "epoch": 0.9692307692307692, "grad_norm": 1.863347053527832, "learning_rate": 3.350515463917526e-07, "loss": 0.9229, "step": 378 }, { "epoch": 0.9717948717948718, "grad_norm": 1.6344797611236572, "learning_rate": 3.0927835051546394e-07, "loss": 0.915, "step": 379 }, { "epoch": 0.9743589743589743, "grad_norm": 1.7815521955490112, "learning_rate": 2.8350515463917527e-07, "loss": 1.002, "step": 380 }, { "epoch": 0.9769230769230769, "grad_norm": 1.689065933227539, "learning_rate": 2.577319587628866e-07, "loss": 0.8452, "step": 381 }, { "epoch": 0.9794871794871794, "grad_norm": 1.8653761148452759, "learning_rate": 2.3195876288659797e-07, "loss": 0.9282, "step": 382 }, { "epoch": 0.982051282051282, "grad_norm": 1.7302168607711792, "learning_rate": 2.061855670103093e-07, "loss": 0.6318, "step": 383 }, { "epoch": 0.9846153846153847, "grad_norm": 1.8644577264785767, "learning_rate": 1.804123711340206e-07, "loss": 0.8018, "step": 384 }, { "epoch": 0.9871794871794872, "grad_norm": 1.725438117980957, "learning_rate": 1.5463917525773197e-07, "loss": 0.9722, "step": 385 }, { "epoch": 0.9897435897435898, "grad_norm": 2.536639928817749, "learning_rate": 1.288659793814433e-07, "loss": 1.0117, "step": 386 }, { "epoch": 0.9923076923076923, "grad_norm": 6.378511905670166, "learning_rate": 1.0309278350515465e-07, "loss": 1.0488, "step": 387 }, { "epoch": 0.9948717948717949, "grad_norm": 1.8261665105819702, "learning_rate": 7.731958762886598e-08, "loss": 0.9814, "step": 388 }, { "epoch": 0.9974358974358974, "grad_norm": 1.732452630996704, "learning_rate": 5.1546391752577325e-08, "loss": 0.7505, "step": 389 }, { "epoch": 1.0, "grad_norm": 1.9269561767578125, "learning_rate": 2.5773195876288662e-08, "loss": 1.0332, "step": 390 } ], "logging_steps": 1.0, "max_steps": 390, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }