| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 390, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.002564102564102564, | |
| "grad_norm": 1.828752040863037, | |
| "learning_rate": 0.0, | |
| "loss": 1.0674, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.005128205128205128, | |
| "grad_norm": 2.0600810050964355, | |
| "learning_rate": 1e-05, | |
| "loss": 1.0127, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.007692307692307693, | |
| "grad_norm": 1.7609747648239136, | |
| "learning_rate": 1e-05, | |
| "loss": 0.8945, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.010256410256410256, | |
| "grad_norm": 2.0786192417144775, | |
| "learning_rate": 9.974226804123713e-06, | |
| "loss": 1.0947, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.01282051282051282, | |
| "grad_norm": 1.8740049600601196, | |
| "learning_rate": 9.948453608247423e-06, | |
| "loss": 1.0859, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.015384615384615385, | |
| "grad_norm": 1.8995977640151978, | |
| "learning_rate": 9.922680412371136e-06, | |
| "loss": 0.9326, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.017948717948717947, | |
| "grad_norm": 1.807106375694275, | |
| "learning_rate": 9.896907216494846e-06, | |
| "loss": 0.9839, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.020512820512820513, | |
| "grad_norm": 2.0719048976898193, | |
| "learning_rate": 9.871134020618558e-06, | |
| "loss": 1.1914, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.023076923076923078, | |
| "grad_norm": 1.714697241783142, | |
| "learning_rate": 9.84536082474227e-06, | |
| "loss": 0.9263, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.02564102564102564, | |
| "grad_norm": 1.7930278778076172, | |
| "learning_rate": 9.819587628865979e-06, | |
| "loss": 0.9546, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.028205128205128206, | |
| "grad_norm": 1.8702346086502075, | |
| "learning_rate": 9.793814432989691e-06, | |
| "loss": 1.0645, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.03076923076923077, | |
| "grad_norm": 1.7376536130905151, | |
| "learning_rate": 9.768041237113403e-06, | |
| "loss": 1.0166, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.03333333333333333, | |
| "grad_norm": 1.804457426071167, | |
| "learning_rate": 9.742268041237114e-06, | |
| "loss": 0.8545, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.035897435897435895, | |
| "grad_norm": 1.8418951034545898, | |
| "learning_rate": 9.716494845360826e-06, | |
| "loss": 1.1855, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.038461538461538464, | |
| "grad_norm": 3.295741558074951, | |
| "learning_rate": 9.690721649484536e-06, | |
| "loss": 1.1631, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.041025641025641026, | |
| "grad_norm": 2.6767563819885254, | |
| "learning_rate": 9.664948453608248e-06, | |
| "loss": 1.0205, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.04358974358974359, | |
| "grad_norm": 2.088998556137085, | |
| "learning_rate": 9.63917525773196e-06, | |
| "loss": 0.9512, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.046153846153846156, | |
| "grad_norm": 3.8526885509490967, | |
| "learning_rate": 9.613402061855671e-06, | |
| "loss": 1.2227, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.04871794871794872, | |
| "grad_norm": 2.5801403522491455, | |
| "learning_rate": 9.587628865979383e-06, | |
| "loss": 0.9424, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.05128205128205128, | |
| "grad_norm": 2.2087137699127197, | |
| "learning_rate": 9.561855670103093e-06, | |
| "loss": 1.3262, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.05384615384615385, | |
| "grad_norm": 1.9106372594833374, | |
| "learning_rate": 9.536082474226806e-06, | |
| "loss": 0.9004, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.05641025641025641, | |
| "grad_norm": 1.7588039636611938, | |
| "learning_rate": 9.510309278350516e-06, | |
| "loss": 0.9482, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.05897435897435897, | |
| "grad_norm": 2.0292000770568848, | |
| "learning_rate": 9.484536082474226e-06, | |
| "loss": 1.0371, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.06153846153846154, | |
| "grad_norm": 1.8534419536590576, | |
| "learning_rate": 9.458762886597939e-06, | |
| "loss": 1.0303, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.0641025641025641, | |
| "grad_norm": 2.4608726501464844, | |
| "learning_rate": 9.43298969072165e-06, | |
| "loss": 1.4395, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.06666666666666667, | |
| "grad_norm": 2.119417905807495, | |
| "learning_rate": 9.407216494845361e-06, | |
| "loss": 1.1143, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.06923076923076923, | |
| "grad_norm": 2.145531177520752, | |
| "learning_rate": 9.381443298969073e-06, | |
| "loss": 0.9443, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.07179487179487179, | |
| "grad_norm": 4.710203170776367, | |
| "learning_rate": 9.355670103092784e-06, | |
| "loss": 1.3311, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.07435897435897436, | |
| "grad_norm": 3.2553186416625977, | |
| "learning_rate": 9.329896907216496e-06, | |
| "loss": 1.123, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.07692307692307693, | |
| "grad_norm": 2.38273024559021, | |
| "learning_rate": 9.304123711340208e-06, | |
| "loss": 1.124, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.07948717948717948, | |
| "grad_norm": 3.6473400592803955, | |
| "learning_rate": 9.278350515463918e-06, | |
| "loss": 1.2568, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.08205128205128205, | |
| "grad_norm": 2.2762739658355713, | |
| "learning_rate": 9.25257731958763e-06, | |
| "loss": 1.0898, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.08461538461538462, | |
| "grad_norm": 1.870347499847412, | |
| "learning_rate": 9.226804123711341e-06, | |
| "loss": 0.8887, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.08717948717948718, | |
| "grad_norm": 3.9631059169769287, | |
| "learning_rate": 9.201030927835051e-06, | |
| "loss": 0.937, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.08974358974358974, | |
| "grad_norm": 2.2279770374298096, | |
| "learning_rate": 9.175257731958764e-06, | |
| "loss": 0.9785, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.09230769230769231, | |
| "grad_norm": 2.2554802894592285, | |
| "learning_rate": 9.149484536082474e-06, | |
| "loss": 1.0762, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.09487179487179487, | |
| "grad_norm": 2.188344955444336, | |
| "learning_rate": 9.123711340206186e-06, | |
| "loss": 1.1318, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.09743589743589744, | |
| "grad_norm": 2.1649510860443115, | |
| "learning_rate": 9.097938144329898e-06, | |
| "loss": 0.7827, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 2.4596219062805176, | |
| "learning_rate": 9.072164948453609e-06, | |
| "loss": 0.8247, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.10256410256410256, | |
| "grad_norm": 2.487870931625366, | |
| "learning_rate": 9.04639175257732e-06, | |
| "loss": 1.0635, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.10512820512820513, | |
| "grad_norm": 1.9071369171142578, | |
| "learning_rate": 9.020618556701031e-06, | |
| "loss": 0.9946, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.1076923076923077, | |
| "grad_norm": 2.219597101211548, | |
| "learning_rate": 8.994845360824743e-06, | |
| "loss": 1.1309, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.11025641025641025, | |
| "grad_norm": 2.2724695205688477, | |
| "learning_rate": 8.969072164948455e-06, | |
| "loss": 1.2637, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.11282051282051282, | |
| "grad_norm": 2.331315755844116, | |
| "learning_rate": 8.943298969072166e-06, | |
| "loss": 1.1611, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.11538461538461539, | |
| "grad_norm": 2.036611318588257, | |
| "learning_rate": 8.917525773195878e-06, | |
| "loss": 0.9233, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.11794871794871795, | |
| "grad_norm": 1.9779893159866333, | |
| "learning_rate": 8.891752577319588e-06, | |
| "loss": 0.8071, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.12051282051282051, | |
| "grad_norm": 2.511869192123413, | |
| "learning_rate": 8.865979381443299e-06, | |
| "loss": 0.9165, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.12307692307692308, | |
| "grad_norm": 1.9824575185775757, | |
| "learning_rate": 8.840206185567011e-06, | |
| "loss": 0.9712, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.12564102564102564, | |
| "grad_norm": 2.224874973297119, | |
| "learning_rate": 8.814432989690721e-06, | |
| "loss": 0.8857, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.1282051282051282, | |
| "grad_norm": 2.290484666824341, | |
| "learning_rate": 8.788659793814434e-06, | |
| "loss": 1.0908, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.13076923076923078, | |
| "grad_norm": 2.009584903717041, | |
| "learning_rate": 8.762886597938146e-06, | |
| "loss": 0.833, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.13333333333333333, | |
| "grad_norm": 4.70264196395874, | |
| "learning_rate": 8.737113402061856e-06, | |
| "loss": 0.998, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.1358974358974359, | |
| "grad_norm": 3.110806465148926, | |
| "learning_rate": 8.711340206185568e-06, | |
| "loss": 0.9189, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.13846153846153847, | |
| "grad_norm": 1.902529239654541, | |
| "learning_rate": 8.685567010309279e-06, | |
| "loss": 1.0293, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.14102564102564102, | |
| "grad_norm": 2.1648287773132324, | |
| "learning_rate": 8.65979381443299e-06, | |
| "loss": 1.0488, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.14358974358974358, | |
| "grad_norm": 1.7489367723464966, | |
| "learning_rate": 8.634020618556703e-06, | |
| "loss": 0.8091, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.14615384615384616, | |
| "grad_norm": 1.9431246519088745, | |
| "learning_rate": 8.608247422680413e-06, | |
| "loss": 1.0049, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.14871794871794872, | |
| "grad_norm": 2.2493724822998047, | |
| "learning_rate": 8.582474226804124e-06, | |
| "loss": 0.9321, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.15128205128205127, | |
| "grad_norm": 2.1866893768310547, | |
| "learning_rate": 8.556701030927836e-06, | |
| "loss": 1.2656, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.15384615384615385, | |
| "grad_norm": 2.024313449859619, | |
| "learning_rate": 8.530927835051546e-06, | |
| "loss": 1.0801, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.1564102564102564, | |
| "grad_norm": 1.77311110496521, | |
| "learning_rate": 8.505154639175259e-06, | |
| "loss": 0.9946, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.15897435897435896, | |
| "grad_norm": 1.6627540588378906, | |
| "learning_rate": 8.479381443298969e-06, | |
| "loss": 0.9702, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.16153846153846155, | |
| "grad_norm": 1.9997308254241943, | |
| "learning_rate": 8.453608247422681e-06, | |
| "loss": 0.9487, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.1641025641025641, | |
| "grad_norm": 1.6049851179122925, | |
| "learning_rate": 8.427835051546393e-06, | |
| "loss": 0.9512, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.16666666666666666, | |
| "grad_norm": 1.8791325092315674, | |
| "learning_rate": 8.402061855670104e-06, | |
| "loss": 0.9355, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.16923076923076924, | |
| "grad_norm": 2.173393964767456, | |
| "learning_rate": 8.376288659793816e-06, | |
| "loss": 1.1992, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.1717948717948718, | |
| "grad_norm": 2.0367043018341064, | |
| "learning_rate": 8.350515463917526e-06, | |
| "loss": 1.1426, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.17435897435897435, | |
| "grad_norm": 1.7792794704437256, | |
| "learning_rate": 8.324742268041238e-06, | |
| "loss": 0.9097, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.17692307692307693, | |
| "grad_norm": 1.9103853702545166, | |
| "learning_rate": 8.29896907216495e-06, | |
| "loss": 1.0225, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.1794871794871795, | |
| "grad_norm": 2.0614025592803955, | |
| "learning_rate": 8.27319587628866e-06, | |
| "loss": 1.2012, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.18205128205128204, | |
| "grad_norm": 1.5830662250518799, | |
| "learning_rate": 8.247422680412371e-06, | |
| "loss": 0.7656, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.18461538461538463, | |
| "grad_norm": 1.8411104679107666, | |
| "learning_rate": 8.221649484536083e-06, | |
| "loss": 1.1162, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.18717948717948718, | |
| "grad_norm": 1.954744815826416, | |
| "learning_rate": 8.195876288659794e-06, | |
| "loss": 1.2031, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.18974358974358974, | |
| "grad_norm": 2.1274709701538086, | |
| "learning_rate": 8.170103092783506e-06, | |
| "loss": 1.1641, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.19230769230769232, | |
| "grad_norm": 1.8388986587524414, | |
| "learning_rate": 8.144329896907216e-06, | |
| "loss": 0.9517, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.19487179487179487, | |
| "grad_norm": 2.0911478996276855, | |
| "learning_rate": 8.118556701030929e-06, | |
| "loss": 1.0801, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.19743589743589743, | |
| "grad_norm": 1.9565143585205078, | |
| "learning_rate": 8.09278350515464e-06, | |
| "loss": 0.9478, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 1.8666242361068726, | |
| "learning_rate": 8.067010309278351e-06, | |
| "loss": 1.3125, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.20256410256410257, | |
| "grad_norm": 1.9826719760894775, | |
| "learning_rate": 8.041237113402063e-06, | |
| "loss": 0.9492, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.20512820512820512, | |
| "grad_norm": 2.062222719192505, | |
| "learning_rate": 8.015463917525774e-06, | |
| "loss": 1.0195, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.2076923076923077, | |
| "grad_norm": 2.2730824947357178, | |
| "learning_rate": 7.989690721649486e-06, | |
| "loss": 0.9248, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.21025641025641026, | |
| "grad_norm": 3.432387113571167, | |
| "learning_rate": 7.963917525773196e-06, | |
| "loss": 0.9678, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.2128205128205128, | |
| "grad_norm": 2.1514482498168945, | |
| "learning_rate": 7.938144329896907e-06, | |
| "loss": 0.7881, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.2153846153846154, | |
| "grad_norm": 1.6954137086868286, | |
| "learning_rate": 7.912371134020619e-06, | |
| "loss": 0.8833, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.21794871794871795, | |
| "grad_norm": 1.9222341775894165, | |
| "learning_rate": 7.886597938144331e-06, | |
| "loss": 0.9917, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.2205128205128205, | |
| "grad_norm": 1.9066567420959473, | |
| "learning_rate": 7.860824742268041e-06, | |
| "loss": 0.9507, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.2230769230769231, | |
| "grad_norm": 1.9370355606079102, | |
| "learning_rate": 7.835051546391754e-06, | |
| "loss": 0.9814, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.22564102564102564, | |
| "grad_norm": 1.8217931985855103, | |
| "learning_rate": 7.809278350515464e-06, | |
| "loss": 1.0186, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.2282051282051282, | |
| "grad_norm": 1.6907788515090942, | |
| "learning_rate": 7.783505154639176e-06, | |
| "loss": 0.9624, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.23076923076923078, | |
| "grad_norm": 1.6538673639297485, | |
| "learning_rate": 7.757731958762888e-06, | |
| "loss": 0.8237, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.23333333333333334, | |
| "grad_norm": 1.9128988981246948, | |
| "learning_rate": 7.731958762886599e-06, | |
| "loss": 0.9175, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.2358974358974359, | |
| "grad_norm": 3.1747217178344727, | |
| "learning_rate": 7.70618556701031e-06, | |
| "loss": 0.9775, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.23846153846153847, | |
| "grad_norm": 1.6946772336959839, | |
| "learning_rate": 7.680412371134021e-06, | |
| "loss": 0.8823, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.24102564102564103, | |
| "grad_norm": 2.1718826293945312, | |
| "learning_rate": 7.654639175257732e-06, | |
| "loss": 1.1543, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.24358974358974358, | |
| "grad_norm": 2.0842125415802, | |
| "learning_rate": 7.628865979381444e-06, | |
| "loss": 0.9058, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.24615384615384617, | |
| "grad_norm": 1.8005638122558594, | |
| "learning_rate": 7.603092783505155e-06, | |
| "loss": 0.9146, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.24871794871794872, | |
| "grad_norm": 1.6228526830673218, | |
| "learning_rate": 7.577319587628866e-06, | |
| "loss": 0.7095, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.2512820512820513, | |
| "grad_norm": 2.022739887237549, | |
| "learning_rate": 7.551546391752578e-06, | |
| "loss": 0.9229, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.25384615384615383, | |
| "grad_norm": 1.7935630083084106, | |
| "learning_rate": 7.525773195876289e-06, | |
| "loss": 0.7896, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.2564102564102564, | |
| "grad_norm": 1.747321367263794, | |
| "learning_rate": 7.500000000000001e-06, | |
| "loss": 0.9551, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.258974358974359, | |
| "grad_norm": 1.8076218366622925, | |
| "learning_rate": 7.474226804123712e-06, | |
| "loss": 0.8394, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.26153846153846155, | |
| "grad_norm": 1.730443000793457, | |
| "learning_rate": 7.448453608247424e-06, | |
| "loss": 0.9062, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.2641025641025641, | |
| "grad_norm": 2.290266275405884, | |
| "learning_rate": 7.422680412371135e-06, | |
| "loss": 1.248, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.26666666666666666, | |
| "grad_norm": 1.772620677947998, | |
| "learning_rate": 7.396907216494846e-06, | |
| "loss": 0.8296, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.2692307692307692, | |
| "grad_norm": 1.7270923852920532, | |
| "learning_rate": 7.3711340206185574e-06, | |
| "loss": 0.8735, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.2717948717948718, | |
| "grad_norm": 3.0547115802764893, | |
| "learning_rate": 7.3453608247422696e-06, | |
| "loss": 0.9385, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.2743589743589744, | |
| "grad_norm": 2.1224873065948486, | |
| "learning_rate": 7.319587628865979e-06, | |
| "loss": 1.0918, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.27692307692307694, | |
| "grad_norm": 2.175182819366455, | |
| "learning_rate": 7.293814432989691e-06, | |
| "loss": 0.9707, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.2794871794871795, | |
| "grad_norm": 1.8396246433258057, | |
| "learning_rate": 7.2680412371134026e-06, | |
| "loss": 0.9609, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.28205128205128205, | |
| "grad_norm": 1.8551801443099976, | |
| "learning_rate": 7.242268041237114e-06, | |
| "loss": 0.8916, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.2846153846153846, | |
| "grad_norm": 1.960680603981018, | |
| "learning_rate": 7.216494845360825e-06, | |
| "loss": 1.0352, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.28717948717948716, | |
| "grad_norm": 1.8434001207351685, | |
| "learning_rate": 7.190721649484536e-06, | |
| "loss": 1.0576, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.28974358974358977, | |
| "grad_norm": 2.5577752590179443, | |
| "learning_rate": 7.164948453608248e-06, | |
| "loss": 1.0762, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.2923076923076923, | |
| "grad_norm": 1.8725932836532593, | |
| "learning_rate": 7.13917525773196e-06, | |
| "loss": 0.79, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.2948717948717949, | |
| "grad_norm": 2.0246224403381348, | |
| "learning_rate": 7.113402061855671e-06, | |
| "loss": 0.9058, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.29743589743589743, | |
| "grad_norm": 1.8787579536437988, | |
| "learning_rate": 7.087628865979382e-06, | |
| "loss": 0.8569, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 1.6573154926300049, | |
| "learning_rate": 7.061855670103094e-06, | |
| "loss": 0.8091, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.30256410256410254, | |
| "grad_norm": 1.7937979698181152, | |
| "learning_rate": 7.036082474226805e-06, | |
| "loss": 0.978, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.30512820512820515, | |
| "grad_norm": 1.9054224491119385, | |
| "learning_rate": 7.010309278350515e-06, | |
| "loss": 1.0635, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.3076923076923077, | |
| "grad_norm": 1.6725414991378784, | |
| "learning_rate": 6.984536082474227e-06, | |
| "loss": 1.0312, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.31025641025641026, | |
| "grad_norm": 1.8967286348342896, | |
| "learning_rate": 6.958762886597939e-06, | |
| "loss": 0.9746, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.3128205128205128, | |
| "grad_norm": 2.158597230911255, | |
| "learning_rate": 6.93298969072165e-06, | |
| "loss": 1.1309, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.3153846153846154, | |
| "grad_norm": 2.2662479877471924, | |
| "learning_rate": 6.907216494845361e-06, | |
| "loss": 1.0674, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.31794871794871793, | |
| "grad_norm": 1.9628630876541138, | |
| "learning_rate": 6.881443298969073e-06, | |
| "loss": 0.9644, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.32051282051282054, | |
| "grad_norm": 1.7974278926849365, | |
| "learning_rate": 6.855670103092784e-06, | |
| "loss": 0.8877, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.3230769230769231, | |
| "grad_norm": 1.7548365592956543, | |
| "learning_rate": 6.829896907216495e-06, | |
| "loss": 0.9204, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.32564102564102565, | |
| "grad_norm": 1.9516572952270508, | |
| "learning_rate": 6.804123711340207e-06, | |
| "loss": 1.1602, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.3282051282051282, | |
| "grad_norm": 2.1989433765411377, | |
| "learning_rate": 6.778350515463919e-06, | |
| "loss": 1.1133, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.33076923076923076, | |
| "grad_norm": 1.800710916519165, | |
| "learning_rate": 6.75257731958763e-06, | |
| "loss": 0.791, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.3333333333333333, | |
| "grad_norm": 3.385741710662842, | |
| "learning_rate": 6.726804123711341e-06, | |
| "loss": 1.1758, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.33589743589743587, | |
| "grad_norm": 2.0922465324401855, | |
| "learning_rate": 6.701030927835052e-06, | |
| "loss": 0.9214, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.3384615384615385, | |
| "grad_norm": 1.573805332183838, | |
| "learning_rate": 6.675257731958763e-06, | |
| "loss": 0.7715, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.34102564102564104, | |
| "grad_norm": 1.7901231050491333, | |
| "learning_rate": 6.649484536082474e-06, | |
| "loss": 0.9443, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.3435897435897436, | |
| "grad_norm": 1.6376028060913086, | |
| "learning_rate": 6.623711340206186e-06, | |
| "loss": 0.9854, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.34615384615384615, | |
| "grad_norm": 2.0153403282165527, | |
| "learning_rate": 6.597938144329898e-06, | |
| "loss": 1.2373, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.3487179487179487, | |
| "grad_norm": 1.819011926651001, | |
| "learning_rate": 6.572164948453609e-06, | |
| "loss": 0.9048, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.35128205128205126, | |
| "grad_norm": 1.623148798942566, | |
| "learning_rate": 6.54639175257732e-06, | |
| "loss": 0.8262, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.35384615384615387, | |
| "grad_norm": 2.2499585151672363, | |
| "learning_rate": 6.520618556701031e-06, | |
| "loss": 1.1152, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.3564102564102564, | |
| "grad_norm": 1.9235867261886597, | |
| "learning_rate": 6.494845360824743e-06, | |
| "loss": 0.9067, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.358974358974359, | |
| "grad_norm": 1.9163473844528198, | |
| "learning_rate": 6.469072164948455e-06, | |
| "loss": 1.0469, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.36153846153846153, | |
| "grad_norm": 2.092963457107544, | |
| "learning_rate": 6.443298969072166e-06, | |
| "loss": 1.0352, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.3641025641025641, | |
| "grad_norm": 9.061576843261719, | |
| "learning_rate": 6.417525773195877e-06, | |
| "loss": 0.833, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.36666666666666664, | |
| "grad_norm": 2.0472512245178223, | |
| "learning_rate": 6.391752577319588e-06, | |
| "loss": 1.043, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.36923076923076925, | |
| "grad_norm": 1.9477477073669434, | |
| "learning_rate": 6.365979381443299e-06, | |
| "loss": 0.9644, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.3717948717948718, | |
| "grad_norm": 1.9295361042022705, | |
| "learning_rate": 6.34020618556701e-06, | |
| "loss": 0.9746, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.37435897435897436, | |
| "grad_norm": 1.9540655612945557, | |
| "learning_rate": 6.314432989690722e-06, | |
| "loss": 0.9375, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.3769230769230769, | |
| "grad_norm": 1.9495689868927002, | |
| "learning_rate": 6.288659793814433e-06, | |
| "loss": 1.123, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.37948717948717947, | |
| "grad_norm": 1.8979523181915283, | |
| "learning_rate": 6.262886597938145e-06, | |
| "loss": 1.0762, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.382051282051282, | |
| "grad_norm": 2.086167573928833, | |
| "learning_rate": 6.237113402061856e-06, | |
| "loss": 0.8975, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.38461538461538464, | |
| "grad_norm": 1.9562528133392334, | |
| "learning_rate": 6.211340206185568e-06, | |
| "loss": 0.9429, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.3871794871794872, | |
| "grad_norm": 2.3442628383636475, | |
| "learning_rate": 6.185567010309279e-06, | |
| "loss": 0.9844, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.38974358974358975, | |
| "grad_norm": 1.8352185487747192, | |
| "learning_rate": 6.15979381443299e-06, | |
| "loss": 1.1045, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.3923076923076923, | |
| "grad_norm": 3.1961944103240967, | |
| "learning_rate": 6.134020618556702e-06, | |
| "loss": 1.0596, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.39487179487179486, | |
| "grad_norm": 1.8429200649261475, | |
| "learning_rate": 6.108247422680414e-06, | |
| "loss": 0.9365, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.3974358974358974, | |
| "grad_norm": 1.929801106452942, | |
| "learning_rate": 6.082474226804124e-06, | |
| "loss": 0.9883, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 1.670444369316101, | |
| "learning_rate": 6.056701030927835e-06, | |
| "loss": 0.7524, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.4025641025641026, | |
| "grad_norm": 2.022891044616699, | |
| "learning_rate": 6.030927835051547e-06, | |
| "loss": 1.2197, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.40512820512820513, | |
| "grad_norm": 1.7056248188018799, | |
| "learning_rate": 6.005154639175258e-06, | |
| "loss": 0.9121, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.4076923076923077, | |
| "grad_norm": 2.0686631202697754, | |
| "learning_rate": 5.979381443298969e-06, | |
| "loss": 1.1865, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.41025641025641024, | |
| "grad_norm": 1.8868600130081177, | |
| "learning_rate": 5.9536082474226805e-06, | |
| "loss": 0.9658, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.4128205128205128, | |
| "grad_norm": 2.3580541610717773, | |
| "learning_rate": 5.927835051546393e-06, | |
| "loss": 0.9297, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.4153846153846154, | |
| "grad_norm": 2.071708917617798, | |
| "learning_rate": 5.902061855670104e-06, | |
| "loss": 1.0693, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.41794871794871796, | |
| "grad_norm": 1.8569782972335815, | |
| "learning_rate": 5.876288659793815e-06, | |
| "loss": 0.9438, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.4205128205128205, | |
| "grad_norm": 1.7496881484985352, | |
| "learning_rate": 5.8505154639175264e-06, | |
| "loss": 0.8574, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.4230769230769231, | |
| "grad_norm": 1.825770378112793, | |
| "learning_rate": 5.824742268041238e-06, | |
| "loss": 0.9678, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.4256410256410256, | |
| "grad_norm": 2.156632423400879, | |
| "learning_rate": 5.79896907216495e-06, | |
| "loss": 1.2031, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.4282051282051282, | |
| "grad_norm": 1.6969801187515259, | |
| "learning_rate": 5.7731958762886594e-06, | |
| "loss": 0.8359, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.4307692307692308, | |
| "grad_norm": 2.4445745944976807, | |
| "learning_rate": 5.7474226804123716e-06, | |
| "loss": 1.0293, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.43333333333333335, | |
| "grad_norm": 1.9905025959014893, | |
| "learning_rate": 5.721649484536083e-06, | |
| "loss": 1.0791, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.4358974358974359, | |
| "grad_norm": 1.9443804025650024, | |
| "learning_rate": 5.695876288659794e-06, | |
| "loss": 0.9702, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.43846153846153846, | |
| "grad_norm": 2.2694649696350098, | |
| "learning_rate": 5.670103092783505e-06, | |
| "loss": 1.1914, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.441025641025641, | |
| "grad_norm": 2.1340649127960205, | |
| "learning_rate": 5.644329896907217e-06, | |
| "loss": 1.0371, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.44358974358974357, | |
| "grad_norm": 1.9536010026931763, | |
| "learning_rate": 5.618556701030928e-06, | |
| "loss": 0.9888, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.4461538461538462, | |
| "grad_norm": 1.837241530418396, | |
| "learning_rate": 5.59278350515464e-06, | |
| "loss": 0.9033, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.44871794871794873, | |
| "grad_norm": 1.8256818056106567, | |
| "learning_rate": 5.567010309278351e-06, | |
| "loss": 1.0117, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.4512820512820513, | |
| "grad_norm": 2.1817359924316406, | |
| "learning_rate": 5.541237113402063e-06, | |
| "loss": 0.8506, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.45384615384615384, | |
| "grad_norm": 1.7470033168792725, | |
| "learning_rate": 5.515463917525774e-06, | |
| "loss": 0.9473, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.4564102564102564, | |
| "grad_norm": 1.8381420373916626, | |
| "learning_rate": 5.489690721649485e-06, | |
| "loss": 0.8828, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.45897435897435895, | |
| "grad_norm": 1.9407833814620972, | |
| "learning_rate": 5.463917525773196e-06, | |
| "loss": 1.3193, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.46153846153846156, | |
| "grad_norm": 1.9037673473358154, | |
| "learning_rate": 5.438144329896907e-06, | |
| "loss": 0.916, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.4641025641025641, | |
| "grad_norm": 1.8702125549316406, | |
| "learning_rate": 5.412371134020619e-06, | |
| "loss": 1.0039, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.4666666666666667, | |
| "grad_norm": 2.0988097190856934, | |
| "learning_rate": 5.38659793814433e-06, | |
| "loss": 1.0225, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.46923076923076923, | |
| "grad_norm": 1.9512522220611572, | |
| "learning_rate": 5.360824742268042e-06, | |
| "loss": 0.9854, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.4717948717948718, | |
| "grad_norm": 1.8114992380142212, | |
| "learning_rate": 5.335051546391753e-06, | |
| "loss": 0.8008, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.47435897435897434, | |
| "grad_norm": 1.7766282558441162, | |
| "learning_rate": 5.309278350515464e-06, | |
| "loss": 0.9795, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.47692307692307695, | |
| "grad_norm": 2.319395065307617, | |
| "learning_rate": 5.2835051546391755e-06, | |
| "loss": 1.1572, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.4794871794871795, | |
| "grad_norm": 1.7959656715393066, | |
| "learning_rate": 5.257731958762888e-06, | |
| "loss": 0.8994, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.48205128205128206, | |
| "grad_norm": 1.7262598276138306, | |
| "learning_rate": 5.231958762886599e-06, | |
| "loss": 1.0264, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.4846153846153846, | |
| "grad_norm": 1.9442336559295654, | |
| "learning_rate": 5.20618556701031e-06, | |
| "loss": 1.0674, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.48717948717948717, | |
| "grad_norm": 1.7376888990402222, | |
| "learning_rate": 5.1804123711340214e-06, | |
| "loss": 0.8032, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.4897435897435897, | |
| "grad_norm": 1.5488858222961426, | |
| "learning_rate": 5.154639175257732e-06, | |
| "loss": 0.8101, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.49230769230769234, | |
| "grad_norm": 1.9175901412963867, | |
| "learning_rate": 5.128865979381443e-06, | |
| "loss": 0.8418, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.4948717948717949, | |
| "grad_norm": 2.069321393966675, | |
| "learning_rate": 5.1030927835051544e-06, | |
| "loss": 1.1826, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.49743589743589745, | |
| "grad_norm": 1.918543815612793, | |
| "learning_rate": 5.077319587628866e-06, | |
| "loss": 1.0059, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 1.7345309257507324, | |
| "learning_rate": 5.051546391752578e-06, | |
| "loss": 0.9028, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.5025641025641026, | |
| "grad_norm": 1.7581024169921875, | |
| "learning_rate": 5.025773195876289e-06, | |
| "loss": 1.0, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.5051282051282051, | |
| "grad_norm": 2.021634817123413, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8677, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.5076923076923077, | |
| "grad_norm": 2.0879619121551514, | |
| "learning_rate": 4.974226804123712e-06, | |
| "loss": 1.1279, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.5102564102564102, | |
| "grad_norm": 2.040804386138916, | |
| "learning_rate": 4.948453608247423e-06, | |
| "loss": 1.2178, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.5128205128205128, | |
| "grad_norm": 2.0652830600738525, | |
| "learning_rate": 4.922680412371135e-06, | |
| "loss": 0.873, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.5153846153846153, | |
| "grad_norm": 1.8770358562469482, | |
| "learning_rate": 4.8969072164948455e-06, | |
| "loss": 0.7212, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.517948717948718, | |
| "grad_norm": 1.8745349645614624, | |
| "learning_rate": 4.871134020618557e-06, | |
| "loss": 1.126, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.5205128205128206, | |
| "grad_norm": 1.8858857154846191, | |
| "learning_rate": 4.845360824742268e-06, | |
| "loss": 1.0771, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.5230769230769231, | |
| "grad_norm": 2.110069513320923, | |
| "learning_rate": 4.81958762886598e-06, | |
| "loss": 1.2754, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.5256410256410257, | |
| "grad_norm": 4.52376127243042, | |
| "learning_rate": 4.7938144329896915e-06, | |
| "loss": 1.1182, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.5282051282051282, | |
| "grad_norm": 1.6591471433639526, | |
| "learning_rate": 4.768041237113403e-06, | |
| "loss": 0.8276, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.5307692307692308, | |
| "grad_norm": 1.9472140073776245, | |
| "learning_rate": 4.742268041237113e-06, | |
| "loss": 1.0273, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.5333333333333333, | |
| "grad_norm": 1.8485890626907349, | |
| "learning_rate": 4.716494845360825e-06, | |
| "loss": 0.813, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.5358974358974359, | |
| "grad_norm": 1.7967491149902344, | |
| "learning_rate": 4.690721649484537e-06, | |
| "loss": 0.9946, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.5384615384615384, | |
| "grad_norm": 1.7534973621368408, | |
| "learning_rate": 4.664948453608248e-06, | |
| "loss": 0.7993, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.541025641025641, | |
| "grad_norm": 1.8620507717132568, | |
| "learning_rate": 4.639175257731959e-06, | |
| "loss": 0.9341, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.5435897435897435, | |
| "grad_norm": 1.8512839078903198, | |
| "learning_rate": 4.6134020618556705e-06, | |
| "loss": 0.8013, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.5461538461538461, | |
| "grad_norm": 2.151174306869507, | |
| "learning_rate": 4.587628865979382e-06, | |
| "loss": 1.2969, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.5487179487179488, | |
| "grad_norm": 1.9421318769454956, | |
| "learning_rate": 4.561855670103093e-06, | |
| "loss": 0.9375, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.5512820512820513, | |
| "grad_norm": 1.6634801626205444, | |
| "learning_rate": 4.536082474226804e-06, | |
| "loss": 0.7603, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.5538461538461539, | |
| "grad_norm": 1.8529914617538452, | |
| "learning_rate": 4.510309278350516e-06, | |
| "loss": 0.8545, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.5564102564102564, | |
| "grad_norm": 1.8996697664260864, | |
| "learning_rate": 4.484536082474228e-06, | |
| "loss": 0.9492, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.558974358974359, | |
| "grad_norm": 1.793915033340454, | |
| "learning_rate": 4.458762886597939e-06, | |
| "loss": 0.6826, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.5615384615384615, | |
| "grad_norm": 2.973825454711914, | |
| "learning_rate": 4.4329896907216494e-06, | |
| "loss": 0.9023, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.5641025641025641, | |
| "grad_norm": 1.786086082458496, | |
| "learning_rate": 4.407216494845361e-06, | |
| "loss": 0.9502, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.5666666666666667, | |
| "grad_norm": 2.263026714324951, | |
| "learning_rate": 4.381443298969073e-06, | |
| "loss": 0.9326, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.5692307692307692, | |
| "grad_norm": 2.35546612739563, | |
| "learning_rate": 4.355670103092784e-06, | |
| "loss": 1.1348, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.5717948717948718, | |
| "grad_norm": 1.6735016107559204, | |
| "learning_rate": 4.329896907216495e-06, | |
| "loss": 0.9204, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.5743589743589743, | |
| "grad_norm": 2.4395010471343994, | |
| "learning_rate": 4.304123711340207e-06, | |
| "loss": 1.1689, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.5769230769230769, | |
| "grad_norm": 1.8484439849853516, | |
| "learning_rate": 4.278350515463918e-06, | |
| "loss": 0.9751, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.5794871794871795, | |
| "grad_norm": 2.61309552192688, | |
| "learning_rate": 4.252577319587629e-06, | |
| "loss": 0.9272, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.5820512820512821, | |
| "grad_norm": 1.8143435716629028, | |
| "learning_rate": 4.2268041237113405e-06, | |
| "loss": 1.0508, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.5846153846153846, | |
| "grad_norm": 1.9212270975112915, | |
| "learning_rate": 4.201030927835052e-06, | |
| "loss": 0.8813, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.5871794871794872, | |
| "grad_norm": 2.3576104640960693, | |
| "learning_rate": 4.175257731958763e-06, | |
| "loss": 0.98, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.5897435897435898, | |
| "grad_norm": 2.2258710861206055, | |
| "learning_rate": 4.149484536082475e-06, | |
| "loss": 0.9766, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.5923076923076923, | |
| "grad_norm": 1.8784958124160767, | |
| "learning_rate": 4.123711340206186e-06, | |
| "loss": 0.9575, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.5948717948717949, | |
| "grad_norm": 1.7360793352127075, | |
| "learning_rate": 4.097938144329897e-06, | |
| "loss": 0.9014, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.5974358974358974, | |
| "grad_norm": 2.718904733657837, | |
| "learning_rate": 4.072164948453608e-06, | |
| "loss": 0.8682, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 3.4316585063934326, | |
| "learning_rate": 4.04639175257732e-06, | |
| "loss": 0.8096, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.6025641025641025, | |
| "grad_norm": 1.9802138805389404, | |
| "learning_rate": 4.020618556701032e-06, | |
| "loss": 1.1055, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.6051282051282051, | |
| "grad_norm": 2.3054115772247314, | |
| "learning_rate": 3.994845360824743e-06, | |
| "loss": 0.8696, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.6076923076923076, | |
| "grad_norm": 2.1070103645324707, | |
| "learning_rate": 3.969072164948453e-06, | |
| "loss": 0.9907, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.6102564102564103, | |
| "grad_norm": 1.71999192237854, | |
| "learning_rate": 3.9432989690721655e-06, | |
| "loss": 1.0342, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.6128205128205129, | |
| "grad_norm": 1.7554974555969238, | |
| "learning_rate": 3.917525773195877e-06, | |
| "loss": 0.9712, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.6153846153846154, | |
| "grad_norm": 1.8095436096191406, | |
| "learning_rate": 3.891752577319588e-06, | |
| "loss": 1.0586, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.617948717948718, | |
| "grad_norm": 2.331573963165283, | |
| "learning_rate": 3.865979381443299e-06, | |
| "loss": 1.1729, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.6205128205128205, | |
| "grad_norm": 1.8929247856140137, | |
| "learning_rate": 3.840206185567011e-06, | |
| "loss": 1.0576, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.6230769230769231, | |
| "grad_norm": 1.6739861965179443, | |
| "learning_rate": 3.814432989690722e-06, | |
| "loss": 0.8008, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.6256410256410256, | |
| "grad_norm": 1.6616827249526978, | |
| "learning_rate": 3.788659793814433e-06, | |
| "loss": 0.876, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.6282051282051282, | |
| "grad_norm": 7.3671746253967285, | |
| "learning_rate": 3.7628865979381445e-06, | |
| "loss": 1.0293, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.6307692307692307, | |
| "grad_norm": 1.9617962837219238, | |
| "learning_rate": 3.737113402061856e-06, | |
| "loss": 1.0117, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.6333333333333333, | |
| "grad_norm": 1.3733716011047363, | |
| "learning_rate": 3.7113402061855674e-06, | |
| "loss": 0.6865, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.6358974358974359, | |
| "grad_norm": 1.6532529592514038, | |
| "learning_rate": 3.6855670103092787e-06, | |
| "loss": 0.9604, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.6384615384615384, | |
| "grad_norm": 1.8402736186981201, | |
| "learning_rate": 3.6597938144329896e-06, | |
| "loss": 0.916, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.6410256410256411, | |
| "grad_norm": 1.7600955963134766, | |
| "learning_rate": 3.6340206185567013e-06, | |
| "loss": 0.9678, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.6435897435897436, | |
| "grad_norm": 1.6891589164733887, | |
| "learning_rate": 3.6082474226804126e-06, | |
| "loss": 0.793, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.6461538461538462, | |
| "grad_norm": 1.8142110109329224, | |
| "learning_rate": 3.582474226804124e-06, | |
| "loss": 1.0518, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.6487179487179487, | |
| "grad_norm": 2.1421236991882324, | |
| "learning_rate": 3.5567010309278356e-06, | |
| "loss": 0.9414, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.6512820512820513, | |
| "grad_norm": 5.681818962097168, | |
| "learning_rate": 3.530927835051547e-06, | |
| "loss": 0.9951, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.6538461538461539, | |
| "grad_norm": 2.093968391418457, | |
| "learning_rate": 3.5051546391752577e-06, | |
| "loss": 0.7842, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.6564102564102564, | |
| "grad_norm": 1.8707085847854614, | |
| "learning_rate": 3.4793814432989694e-06, | |
| "loss": 0.9858, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.658974358974359, | |
| "grad_norm": 1.9755574464797974, | |
| "learning_rate": 3.4536082474226807e-06, | |
| "loss": 0.9136, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.6615384615384615, | |
| "grad_norm": 2.0634946823120117, | |
| "learning_rate": 3.427835051546392e-06, | |
| "loss": 0.7593, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.6641025641025641, | |
| "grad_norm": 1.842301845550537, | |
| "learning_rate": 3.4020618556701037e-06, | |
| "loss": 0.8062, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.6666666666666666, | |
| "grad_norm": 1.857254981994629, | |
| "learning_rate": 3.376288659793815e-06, | |
| "loss": 1.2021, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.6692307692307692, | |
| "grad_norm": 1.9125348329544067, | |
| "learning_rate": 3.350515463917526e-06, | |
| "loss": 1.084, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.6717948717948717, | |
| "grad_norm": 1.8903979063034058, | |
| "learning_rate": 3.324742268041237e-06, | |
| "loss": 0.7666, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.6743589743589744, | |
| "grad_norm": 1.981501579284668, | |
| "learning_rate": 3.298969072164949e-06, | |
| "loss": 1.0508, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.676923076923077, | |
| "grad_norm": 1.5312262773513794, | |
| "learning_rate": 3.27319587628866e-06, | |
| "loss": 0.7231, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.6794871794871795, | |
| "grad_norm": 1.6376142501831055, | |
| "learning_rate": 3.2474226804123714e-06, | |
| "loss": 0.8013, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.6820512820512821, | |
| "grad_norm": 1.5917991399765015, | |
| "learning_rate": 3.221649484536083e-06, | |
| "loss": 0.8042, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.6846153846153846, | |
| "grad_norm": 1.497527837753296, | |
| "learning_rate": 3.195876288659794e-06, | |
| "loss": 0.7031, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.6871794871794872, | |
| "grad_norm": 1.692023515701294, | |
| "learning_rate": 3.170103092783505e-06, | |
| "loss": 0.6646, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.6897435897435897, | |
| "grad_norm": 1.862017273902893, | |
| "learning_rate": 3.1443298969072165e-06, | |
| "loss": 0.8765, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.6923076923076923, | |
| "grad_norm": 2.128854513168335, | |
| "learning_rate": 3.118556701030928e-06, | |
| "loss": 0.7183, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.6948717948717948, | |
| "grad_norm": 1.7785433530807495, | |
| "learning_rate": 3.0927835051546395e-06, | |
| "loss": 0.9966, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.6974358974358974, | |
| "grad_norm": 1.8306283950805664, | |
| "learning_rate": 3.067010309278351e-06, | |
| "loss": 0.7773, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 1.5251365900039673, | |
| "learning_rate": 3.041237113402062e-06, | |
| "loss": 0.7046, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.7025641025641025, | |
| "grad_norm": 2.0568642616271973, | |
| "learning_rate": 3.0154639175257733e-06, | |
| "loss": 1.0459, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.7051282051282052, | |
| "grad_norm": 1.6169660091400146, | |
| "learning_rate": 2.9896907216494846e-06, | |
| "loss": 0.7451, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.7076923076923077, | |
| "grad_norm": 1.6986936330795288, | |
| "learning_rate": 2.9639175257731963e-06, | |
| "loss": 0.8735, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.7102564102564103, | |
| "grad_norm": 1.6713733673095703, | |
| "learning_rate": 2.9381443298969076e-06, | |
| "loss": 0.8862, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.7128205128205128, | |
| "grad_norm": 1.7637569904327393, | |
| "learning_rate": 2.912371134020619e-06, | |
| "loss": 0.9609, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.7153846153846154, | |
| "grad_norm": 2.076930284500122, | |
| "learning_rate": 2.8865979381443297e-06, | |
| "loss": 1.1377, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.717948717948718, | |
| "grad_norm": 1.972031831741333, | |
| "learning_rate": 2.8608247422680414e-06, | |
| "loss": 0.9395, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.7205128205128205, | |
| "grad_norm": 1.8608795404434204, | |
| "learning_rate": 2.8350515463917527e-06, | |
| "loss": 1.0654, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.7230769230769231, | |
| "grad_norm": 3.467540979385376, | |
| "learning_rate": 2.809278350515464e-06, | |
| "loss": 0.7603, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.7256410256410256, | |
| "grad_norm": 1.765555739402771, | |
| "learning_rate": 2.7835051546391757e-06, | |
| "loss": 0.6758, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.7282051282051282, | |
| "grad_norm": 1.6693044900894165, | |
| "learning_rate": 2.757731958762887e-06, | |
| "loss": 0.8433, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.7307692307692307, | |
| "grad_norm": 1.9119174480438232, | |
| "learning_rate": 2.731958762886598e-06, | |
| "loss": 0.8184, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.7333333333333333, | |
| "grad_norm": 2.043612241744995, | |
| "learning_rate": 2.7061855670103095e-06, | |
| "loss": 1.0049, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.735897435897436, | |
| "grad_norm": 1.919756531715393, | |
| "learning_rate": 2.680412371134021e-06, | |
| "loss": 0.9229, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.7384615384615385, | |
| "grad_norm": 1.8900898694992065, | |
| "learning_rate": 2.654639175257732e-06, | |
| "loss": 1.0078, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.7410256410256411, | |
| "grad_norm": 1.793310284614563, | |
| "learning_rate": 2.628865979381444e-06, | |
| "loss": 1.0752, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.7435897435897436, | |
| "grad_norm": 2.732642889022827, | |
| "learning_rate": 2.603092783505155e-06, | |
| "loss": 0.9229, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.7461538461538462, | |
| "grad_norm": 1.845354676246643, | |
| "learning_rate": 2.577319587628866e-06, | |
| "loss": 0.8286, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.7487179487179487, | |
| "grad_norm": 1.9272360801696777, | |
| "learning_rate": 2.5515463917525772e-06, | |
| "loss": 1.0518, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.7512820512820513, | |
| "grad_norm": 2.1714913845062256, | |
| "learning_rate": 2.525773195876289e-06, | |
| "loss": 0.9043, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.7538461538461538, | |
| "grad_norm": 1.7473788261413574, | |
| "learning_rate": 2.5e-06, | |
| "loss": 0.7793, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.7564102564102564, | |
| "grad_norm": 1.9058947563171387, | |
| "learning_rate": 2.4742268041237115e-06, | |
| "loss": 0.8643, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.7589743589743589, | |
| "grad_norm": 2.064485788345337, | |
| "learning_rate": 2.4484536082474228e-06, | |
| "loss": 1.1475, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.7615384615384615, | |
| "grad_norm": 2.100177526473999, | |
| "learning_rate": 2.422680412371134e-06, | |
| "loss": 0.853, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.764102564102564, | |
| "grad_norm": 2.2327897548675537, | |
| "learning_rate": 2.3969072164948458e-06, | |
| "loss": 1.2168, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.7666666666666667, | |
| "grad_norm": 1.9748061895370483, | |
| "learning_rate": 2.3711340206185566e-06, | |
| "loss": 1.0713, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.7692307692307693, | |
| "grad_norm": 1.987243890762329, | |
| "learning_rate": 2.3453608247422683e-06, | |
| "loss": 1.0508, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.7717948717948718, | |
| "grad_norm": 2.191959857940674, | |
| "learning_rate": 2.3195876288659796e-06, | |
| "loss": 0.9326, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.7743589743589744, | |
| "grad_norm": 2.242733955383301, | |
| "learning_rate": 2.293814432989691e-06, | |
| "loss": 0.7515, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.7769230769230769, | |
| "grad_norm": 1.6882741451263428, | |
| "learning_rate": 2.268041237113402e-06, | |
| "loss": 0.7534, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.7794871794871795, | |
| "grad_norm": 2.7890372276306152, | |
| "learning_rate": 2.242268041237114e-06, | |
| "loss": 0.9082, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.782051282051282, | |
| "grad_norm": 2.7322838306427, | |
| "learning_rate": 2.2164948453608247e-06, | |
| "loss": 1.0332, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.7846153846153846, | |
| "grad_norm": 1.9101502895355225, | |
| "learning_rate": 2.1907216494845364e-06, | |
| "loss": 1.0088, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.7871794871794872, | |
| "grad_norm": 1.7795121669769287, | |
| "learning_rate": 2.1649484536082477e-06, | |
| "loss": 0.9082, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.7897435897435897, | |
| "grad_norm": 1.8345370292663574, | |
| "learning_rate": 2.139175257731959e-06, | |
| "loss": 1.0576, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.7923076923076923, | |
| "grad_norm": 2.1167471408843994, | |
| "learning_rate": 2.1134020618556703e-06, | |
| "loss": 0.918, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.7948717948717948, | |
| "grad_norm": 1.7261496782302856, | |
| "learning_rate": 2.0876288659793816e-06, | |
| "loss": 0.8975, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.7974358974358975, | |
| "grad_norm": 1.793904185295105, | |
| "learning_rate": 2.061855670103093e-06, | |
| "loss": 0.8916, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 1.9415448904037476, | |
| "learning_rate": 2.036082474226804e-06, | |
| "loss": 0.8379, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.8025641025641026, | |
| "grad_norm": 1.8090825080871582, | |
| "learning_rate": 2.010309278350516e-06, | |
| "loss": 0.8784, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.8051282051282052, | |
| "grad_norm": 1.775429368019104, | |
| "learning_rate": 1.9845360824742267e-06, | |
| "loss": 0.8643, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.8076923076923077, | |
| "grad_norm": 1.848243236541748, | |
| "learning_rate": 1.9587628865979384e-06, | |
| "loss": 1.1084, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.8102564102564103, | |
| "grad_norm": 1.7999858856201172, | |
| "learning_rate": 1.9329896907216497e-06, | |
| "loss": 0.9321, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.8128205128205128, | |
| "grad_norm": 2.8370208740234375, | |
| "learning_rate": 1.907216494845361e-06, | |
| "loss": 0.9883, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.8153846153846154, | |
| "grad_norm": 2.2633893489837646, | |
| "learning_rate": 1.8814432989690722e-06, | |
| "loss": 1.0166, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.8179487179487179, | |
| "grad_norm": 1.9850558042526245, | |
| "learning_rate": 1.8556701030927837e-06, | |
| "loss": 1.0283, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.8205128205128205, | |
| "grad_norm": 1.904843807220459, | |
| "learning_rate": 1.8298969072164948e-06, | |
| "loss": 0.9829, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.823076923076923, | |
| "grad_norm": 1.7778921127319336, | |
| "learning_rate": 1.8041237113402063e-06, | |
| "loss": 0.9229, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.8256410256410256, | |
| "grad_norm": 1.792877197265625, | |
| "learning_rate": 1.7783505154639178e-06, | |
| "loss": 0.876, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.8282051282051283, | |
| "grad_norm": 1.8950697183609009, | |
| "learning_rate": 1.7525773195876288e-06, | |
| "loss": 0.915, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.8307692307692308, | |
| "grad_norm": 1.911402940750122, | |
| "learning_rate": 1.7268041237113403e-06, | |
| "loss": 1.043, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.8333333333333334, | |
| "grad_norm": 2.1029365062713623, | |
| "learning_rate": 1.7010309278350518e-06, | |
| "loss": 0.9243, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.8358974358974359, | |
| "grad_norm": 2.052480936050415, | |
| "learning_rate": 1.675257731958763e-06, | |
| "loss": 1.2256, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.8384615384615385, | |
| "grad_norm": 1.7406338453292847, | |
| "learning_rate": 1.6494845360824744e-06, | |
| "loss": 0.8804, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.841025641025641, | |
| "grad_norm": 1.8102054595947266, | |
| "learning_rate": 1.6237113402061857e-06, | |
| "loss": 0.854, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.8435897435897436, | |
| "grad_norm": 1.9868026971817017, | |
| "learning_rate": 1.597938144329897e-06, | |
| "loss": 0.8491, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.8461538461538461, | |
| "grad_norm": 1.774075984954834, | |
| "learning_rate": 1.5721649484536082e-06, | |
| "loss": 0.8457, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.8487179487179487, | |
| "grad_norm": 1.8672053813934326, | |
| "learning_rate": 1.5463917525773197e-06, | |
| "loss": 1.0176, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.8512820512820513, | |
| "grad_norm": 1.5850166082382202, | |
| "learning_rate": 1.520618556701031e-06, | |
| "loss": 0.6777, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.8538461538461538, | |
| "grad_norm": 1.6745821237564087, | |
| "learning_rate": 1.4948453608247423e-06, | |
| "loss": 0.939, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.8564102564102564, | |
| "grad_norm": 1.8641244173049927, | |
| "learning_rate": 1.4690721649484538e-06, | |
| "loss": 1.0, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.8589743589743589, | |
| "grad_norm": 1.6371408700942993, | |
| "learning_rate": 1.4432989690721649e-06, | |
| "loss": 0.8037, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.8615384615384616, | |
| "grad_norm": 2.0685510635375977, | |
| "learning_rate": 1.4175257731958764e-06, | |
| "loss": 0.9287, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.8641025641025641, | |
| "grad_norm": 1.8242462873458862, | |
| "learning_rate": 1.3917525773195878e-06, | |
| "loss": 0.8877, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.8666666666666667, | |
| "grad_norm": 2.243830442428589, | |
| "learning_rate": 1.365979381443299e-06, | |
| "loss": 0.7764, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.8692307692307693, | |
| "grad_norm": 2.188321590423584, | |
| "learning_rate": 1.3402061855670104e-06, | |
| "loss": 1.125, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.8717948717948718, | |
| "grad_norm": 2.412721633911133, | |
| "learning_rate": 1.314432989690722e-06, | |
| "loss": 1.1016, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.8743589743589744, | |
| "grad_norm": 2.149219512939453, | |
| "learning_rate": 1.288659793814433e-06, | |
| "loss": 1.1104, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.8769230769230769, | |
| "grad_norm": 1.603258490562439, | |
| "learning_rate": 1.2628865979381445e-06, | |
| "loss": 0.7148, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.8794871794871795, | |
| "grad_norm": 1.7832096815109253, | |
| "learning_rate": 1.2371134020618557e-06, | |
| "loss": 0.9902, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.882051282051282, | |
| "grad_norm": 1.8631786108016968, | |
| "learning_rate": 1.211340206185567e-06, | |
| "loss": 0.8569, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.8846153846153846, | |
| "grad_norm": 1.6944712400436401, | |
| "learning_rate": 1.1855670103092783e-06, | |
| "loss": 0.7397, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.8871794871794871, | |
| "grad_norm": 1.842955231666565, | |
| "learning_rate": 1.1597938144329898e-06, | |
| "loss": 0.8457, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.8897435897435897, | |
| "grad_norm": 2.5506865978240967, | |
| "learning_rate": 1.134020618556701e-06, | |
| "loss": 0.9238, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.8923076923076924, | |
| "grad_norm": 2.157381534576416, | |
| "learning_rate": 1.1082474226804124e-06, | |
| "loss": 1.1855, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.8948717948717949, | |
| "grad_norm": 1.835581660270691, | |
| "learning_rate": 1.0824742268041239e-06, | |
| "loss": 0.9375, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.8974358974358975, | |
| "grad_norm": 2.591947078704834, | |
| "learning_rate": 1.0567010309278351e-06, | |
| "loss": 1.0088, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 1.9977459907531738, | |
| "learning_rate": 1.0309278350515464e-06, | |
| "loss": 1.0576, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.9025641025641026, | |
| "grad_norm": 1.7029050588607788, | |
| "learning_rate": 1.005154639175258e-06, | |
| "loss": 0.8643, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.9051282051282051, | |
| "grad_norm": 1.6559293270111084, | |
| "learning_rate": 9.793814432989692e-07, | |
| "loss": 0.8652, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.9076923076923077, | |
| "grad_norm": 2.0177323818206787, | |
| "learning_rate": 9.536082474226805e-07, | |
| "loss": 1.0361, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.9102564102564102, | |
| "grad_norm": 1.6591455936431885, | |
| "learning_rate": 9.278350515463919e-07, | |
| "loss": 0.8872, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.9128205128205128, | |
| "grad_norm": 2.092712640762329, | |
| "learning_rate": 9.020618556701031e-07, | |
| "loss": 0.9644, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.9153846153846154, | |
| "grad_norm": 1.7917448282241821, | |
| "learning_rate": 8.762886597938144e-07, | |
| "loss": 0.9399, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.9179487179487179, | |
| "grad_norm": 1.9611492156982422, | |
| "learning_rate": 8.505154639175259e-07, | |
| "loss": 1.292, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.9205128205128205, | |
| "grad_norm": 1.7951692342758179, | |
| "learning_rate": 8.247422680412372e-07, | |
| "loss": 0.9551, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.9230769230769231, | |
| "grad_norm": 1.8831959962844849, | |
| "learning_rate": 7.989690721649485e-07, | |
| "loss": 1.1973, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.9256410256410257, | |
| "grad_norm": 1.9080712795257568, | |
| "learning_rate": 7.731958762886599e-07, | |
| "loss": 0.917, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.9282051282051282, | |
| "grad_norm": 2.041674852371216, | |
| "learning_rate": 7.474226804123711e-07, | |
| "loss": 0.9111, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.9307692307692308, | |
| "grad_norm": 1.8067349195480347, | |
| "learning_rate": 7.216494845360824e-07, | |
| "loss": 0.7915, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.9333333333333333, | |
| "grad_norm": 1.733258843421936, | |
| "learning_rate": 6.958762886597939e-07, | |
| "loss": 0.8726, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.9358974358974359, | |
| "grad_norm": 2.1133229732513428, | |
| "learning_rate": 6.701030927835052e-07, | |
| "loss": 0.9028, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.9384615384615385, | |
| "grad_norm": 1.8070833683013916, | |
| "learning_rate": 6.443298969072165e-07, | |
| "loss": 0.8071, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.941025641025641, | |
| "grad_norm": 1.610521912574768, | |
| "learning_rate": 6.185567010309279e-07, | |
| "loss": 0.686, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.9435897435897436, | |
| "grad_norm": 1.8273773193359375, | |
| "learning_rate": 5.927835051546392e-07, | |
| "loss": 1.042, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.9461538461538461, | |
| "grad_norm": 1.6408607959747314, | |
| "learning_rate": 5.670103092783505e-07, | |
| "loss": 0.8584, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.9487179487179487, | |
| "grad_norm": 1.950465440750122, | |
| "learning_rate": 5.412371134020619e-07, | |
| "loss": 0.9629, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.9512820512820512, | |
| "grad_norm": 1.6559863090515137, | |
| "learning_rate": 5.154639175257732e-07, | |
| "loss": 0.7119, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.9538461538461539, | |
| "grad_norm": 1.497727394104004, | |
| "learning_rate": 4.896907216494846e-07, | |
| "loss": 0.71, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.9564102564102565, | |
| "grad_norm": 1.7516483068466187, | |
| "learning_rate": 4.6391752577319593e-07, | |
| "loss": 0.7275, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.958974358974359, | |
| "grad_norm": 1.8231966495513916, | |
| "learning_rate": 4.381443298969072e-07, | |
| "loss": 1.0664, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.9615384615384616, | |
| "grad_norm": 2.928769826889038, | |
| "learning_rate": 4.123711340206186e-07, | |
| "loss": 0.7983, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.9641025641025641, | |
| "grad_norm": 1.5636661052703857, | |
| "learning_rate": 3.8659793814432993e-07, | |
| "loss": 0.6655, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.9666666666666667, | |
| "grad_norm": 1.7315055131912231, | |
| "learning_rate": 3.608247422680412e-07, | |
| "loss": 0.8994, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.9692307692307692, | |
| "grad_norm": 1.863347053527832, | |
| "learning_rate": 3.350515463917526e-07, | |
| "loss": 0.9229, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.9717948717948718, | |
| "grad_norm": 1.6344797611236572, | |
| "learning_rate": 3.0927835051546394e-07, | |
| "loss": 0.915, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.9743589743589743, | |
| "grad_norm": 1.7815521955490112, | |
| "learning_rate": 2.8350515463917527e-07, | |
| "loss": 1.002, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.9769230769230769, | |
| "grad_norm": 1.689065933227539, | |
| "learning_rate": 2.577319587628866e-07, | |
| "loss": 0.8452, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.9794871794871794, | |
| "grad_norm": 1.8653761148452759, | |
| "learning_rate": 2.3195876288659797e-07, | |
| "loss": 0.9282, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.982051282051282, | |
| "grad_norm": 1.7302168607711792, | |
| "learning_rate": 2.061855670103093e-07, | |
| "loss": 0.6318, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.9846153846153847, | |
| "grad_norm": 1.8644577264785767, | |
| "learning_rate": 1.804123711340206e-07, | |
| "loss": 0.8018, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.9871794871794872, | |
| "grad_norm": 1.725438117980957, | |
| "learning_rate": 1.5463917525773197e-07, | |
| "loss": 0.9722, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.9897435897435898, | |
| "grad_norm": 2.536639928817749, | |
| "learning_rate": 1.288659793814433e-07, | |
| "loss": 1.0117, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.9923076923076923, | |
| "grad_norm": 6.378511905670166, | |
| "learning_rate": 1.0309278350515465e-07, | |
| "loss": 1.0488, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.9948717948717949, | |
| "grad_norm": 1.8261665105819702, | |
| "learning_rate": 7.731958762886598e-08, | |
| "loss": 0.9814, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.9974358974358974, | |
| "grad_norm": 1.732452630996704, | |
| "learning_rate": 5.1546391752577325e-08, | |
| "loss": 0.7505, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 1.9269561767578125, | |
| "learning_rate": 2.5773195876288662e-08, | |
| "loss": 1.0332, | |
| "step": 390 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 390, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 1000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |