| { | |
| "best_metric": 0.9733333333333334, | |
| "best_model_checkpoint": "D:\\HDD Tiziana\\Output-prova_melanoma\\checkpoint-94", | |
| "epoch": 40.0, | |
| "global_step": 1880, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 7.978723404255319e-08, | |
| "loss": 0.7312, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.5957446808510638e-07, | |
| "loss": 0.396, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 2.3936170212765954e-07, | |
| "loss": 0.629, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 3.1914893617021275e-07, | |
| "loss": 0.4933, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 3.9893617021276597e-07, | |
| "loss": 0.5513, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.787234042553191e-07, | |
| "loss": 0.3559, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 5.585106382978723e-07, | |
| "loss": 0.9017, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 6.382978723404255e-07, | |
| "loss": 0.6033, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 7.180851063829788e-07, | |
| "loss": 0.4169, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 7.978723404255319e-07, | |
| "loss": 0.5239, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 8.776595744680851e-07, | |
| "loss": 0.2203, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.574468085106382e-07, | |
| "loss": 0.3341, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.0372340425531915e-06, | |
| "loss": 0.5486, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.1170212765957447e-06, | |
| "loss": 0.8415, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.196808510638298e-06, | |
| "loss": 0.5249, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.276595744680851e-06, | |
| "loss": 0.8753, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.3563829787234044e-06, | |
| "loss": 0.669, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.4361702127659576e-06, | |
| "loss": 0.5461, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.5159574468085105e-06, | |
| "loss": 0.4987, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.5957446808510639e-06, | |
| "loss": 0.5869, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.675531914893617e-06, | |
| "loss": 0.7961, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.7553191489361702e-06, | |
| "loss": 0.2292, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.8351063829787234e-06, | |
| "loss": 0.5078, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.9148936170212763e-06, | |
| "loss": 0.455, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 1.9946808510638297e-06, | |
| "loss": 0.5097, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 2.074468085106383e-06, | |
| "loss": 0.4172, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 2.1542553191489364e-06, | |
| "loss": 0.6054, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 2.2340425531914894e-06, | |
| "loss": 0.4193, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 2.3138297872340428e-06, | |
| "loss": 0.1209, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 2.393617021276596e-06, | |
| "loss": 0.3814, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 2.4734042553191487e-06, | |
| "loss": 0.525, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 2.553191489361702e-06, | |
| "loss": 0.3607, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 2.6329787234042554e-06, | |
| "loss": 0.5259, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 2.7127659574468088e-06, | |
| "loss": 0.3994, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 2.7925531914893617e-06, | |
| "loss": 0.3272, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 2.872340425531915e-06, | |
| "loss": 0.322, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 2.952127659574468e-06, | |
| "loss": 0.6201, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.031914893617021e-06, | |
| "loss": 0.5613, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.1117021276595744e-06, | |
| "loss": 0.4053, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.1914893617021277e-06, | |
| "loss": 0.4774, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.271276595744681e-06, | |
| "loss": 0.3988, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.351063829787234e-06, | |
| "loss": 0.415, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.4308510638297874e-06, | |
| "loss": 0.3347, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.5106382978723404e-06, | |
| "loss": 0.502, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 3.5904255319148938e-06, | |
| "loss": 0.6566, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.6702127659574467e-06, | |
| "loss": 0.2601, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.75e-06, | |
| "loss": 0.4501, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_accuracy": 0.9666666666666667, | |
| "eval_loss": 0.20941703021526337, | |
| "eval_runtime": 26.0888, | |
| "eval_samples_per_second": 5.75, | |
| "eval_steps_per_second": 0.383, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.829787234042553e-06, | |
| "loss": 0.2295, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.909574468085106e-06, | |
| "loss": 0.45, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.989361702127659e-06, | |
| "loss": 0.2109, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 4.069148936170212e-06, | |
| "loss": 0.4221, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 4.148936170212766e-06, | |
| "loss": 0.7513, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 4.228723404255319e-06, | |
| "loss": 0.6137, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 4.308510638297873e-06, | |
| "loss": 0.638, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 4.388297872340426e-06, | |
| "loss": 0.4911, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 4.468085106382979e-06, | |
| "loss": 0.1663, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 4.5478723404255326e-06, | |
| "loss": 0.2347, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 4.6276595744680855e-06, | |
| "loss": 0.5612, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 4.7074468085106385e-06, | |
| "loss": 0.3035, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 4.787234042553192e-06, | |
| "loss": 0.6025, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 4.867021276595744e-06, | |
| "loss": 0.531, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 4.946808510638297e-06, | |
| "loss": 0.3869, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 5.026595744680851e-06, | |
| "loss": 0.3956, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 5.106382978723404e-06, | |
| "loss": 0.2386, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 5.186170212765958e-06, | |
| "loss": 0.4691, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 5.265957446808511e-06, | |
| "loss": 0.2354, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 5.345744680851064e-06, | |
| "loss": 0.3984, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 5.4255319148936176e-06, | |
| "loss": 0.5805, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 5.5053191489361705e-06, | |
| "loss": 0.129, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 5.5851063829787235e-06, | |
| "loss": 0.6375, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 5.664893617021277e-06, | |
| "loss": 0.5601, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 5.74468085106383e-06, | |
| "loss": 0.4343, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 5.824468085106383e-06, | |
| "loss": 0.6349, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 5.904255319148936e-06, | |
| "loss": 0.4034, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 5.984042553191489e-06, | |
| "loss": 0.2161, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 6.063829787234042e-06, | |
| "loss": 0.2297, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 6.143617021276596e-06, | |
| "loss": 0.5965, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 6.223404255319149e-06, | |
| "loss": 0.3742, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 6.3031914893617025e-06, | |
| "loss": 0.2334, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 6.3829787234042555e-06, | |
| "loss": 0.3378, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 6.4627659574468084e-06, | |
| "loss": 0.5329, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 6.542553191489362e-06, | |
| "loss": 0.2968, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 6.622340425531915e-06, | |
| "loss": 0.3714, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 6.702127659574468e-06, | |
| "loss": 0.2825, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 6.781914893617022e-06, | |
| "loss": 0.1291, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 6.861702127659575e-06, | |
| "loss": 0.6279, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 6.941489361702128e-06, | |
| "loss": 0.4305, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 7.021276595744681e-06, | |
| "loss": 0.3262, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 7.101063829787234e-06, | |
| "loss": 0.2074, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 7.1808510638297875e-06, | |
| "loss": 0.2801, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 7.2606382978723405e-06, | |
| "loss": 0.2295, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 7.3404255319148934e-06, | |
| "loss": 0.5025, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 7.420212765957447e-06, | |
| "loss": 0.1845, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 7.5e-06, | |
| "loss": 0.5554, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_accuracy": 0.9733333333333334, | |
| "eval_loss": 0.2009524255990982, | |
| "eval_runtime": 25.4517, | |
| "eval_samples_per_second": 5.894, | |
| "eval_steps_per_second": 0.393, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 7.579787234042554e-06, | |
| "loss": 0.3345, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 7.659574468085105e-06, | |
| "loss": 0.3124, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 7.739361702127659e-06, | |
| "loss": 0.2996, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 7.819148936170213e-06, | |
| "loss": 0.3655, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 7.898936170212765e-06, | |
| "loss": 0.2631, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 7.978723404255319e-06, | |
| "loss": 0.3411, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 8.058510638297873e-06, | |
| "loss": 0.1479, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 8.138297872340425e-06, | |
| "loss": 0.1513, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 8.218085106382978e-06, | |
| "loss": 0.2663, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 8.297872340425532e-06, | |
| "loss": 0.3175, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 8.377659574468084e-06, | |
| "loss": 0.2903, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 8.457446808510638e-06, | |
| "loss": 0.2651, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 8.537234042553192e-06, | |
| "loss": 0.3939, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 8.617021276595746e-06, | |
| "loss": 0.2158, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 8.696808510638298e-06, | |
| "loss": 0.3097, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 8.776595744680852e-06, | |
| "loss": 0.4003, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 8.856382978723405e-06, | |
| "loss": 0.377, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 8.936170212765958e-06, | |
| "loss": 0.1615, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 9.015957446808511e-06, | |
| "loss": 0.4294, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 9.095744680851065e-06, | |
| "loss": 0.4334, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 9.175531914893617e-06, | |
| "loss": 0.142, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 9.255319148936171e-06, | |
| "loss": 0.2035, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 9.335106382978725e-06, | |
| "loss": 0.4099, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 9.414893617021277e-06, | |
| "loss": 0.3669, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 9.49468085106383e-06, | |
| "loss": 0.3185, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 9.574468085106385e-06, | |
| "loss": 0.3532, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 9.654255319148937e-06, | |
| "loss": 0.2366, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 9.734042553191489e-06, | |
| "loss": 0.1627, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 9.813829787234043e-06, | |
| "loss": 0.3432, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 9.893617021276595e-06, | |
| "loss": 0.1567, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 9.973404255319148e-06, | |
| "loss": 0.3308, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 1.0053191489361702e-05, | |
| "loss": 0.1848, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 1.0132978723404254e-05, | |
| "loss": 0.4657, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 1.0212765957446808e-05, | |
| "loss": 0.3165, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 1.0292553191489362e-05, | |
| "loss": 0.2542, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 1.0372340425531916e-05, | |
| "loss": 0.3977, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 1.0452127659574468e-05, | |
| "loss": 0.0958, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 1.0531914893617022e-05, | |
| "loss": 0.2821, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 1.0611702127659575e-05, | |
| "loss": 0.3806, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 1.0691489361702128e-05, | |
| "loss": 0.5306, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 1.0771276595744681e-05, | |
| "loss": 0.4589, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 1.0851063829787235e-05, | |
| "loss": 0.1802, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 1.0930851063829787e-05, | |
| "loss": 0.2594, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.1010638297872341e-05, | |
| "loss": 0.2643, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 1.1090425531914895e-05, | |
| "loss": 0.3489, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 1.1170212765957447e-05, | |
| "loss": 0.2528, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 1.125e-05, | |
| "loss": 0.5299, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_accuracy": 0.9733333333333334, | |
| "eval_loss": 0.15951746702194214, | |
| "eval_runtime": 25.5312, | |
| "eval_samples_per_second": 5.875, | |
| "eval_steps_per_second": 0.392, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 3.02, | |
| "learning_rate": 1.1329787234042555e-05, | |
| "loss": 0.1793, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "learning_rate": 1.1409574468085107e-05, | |
| "loss": 0.1429, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 3.06, | |
| "learning_rate": 1.148936170212766e-05, | |
| "loss": 0.1413, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 1.1569148936170214e-05, | |
| "loss": 0.3876, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 3.11, | |
| "learning_rate": 1.1648936170212766e-05, | |
| "loss": 0.2127, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 3.13, | |
| "learning_rate": 1.172872340425532e-05, | |
| "loss": 0.2292, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 3.15, | |
| "learning_rate": 1.1808510638297872e-05, | |
| "loss": 0.2352, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 3.17, | |
| "learning_rate": 1.1888297872340424e-05, | |
| "loss": 0.3663, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 3.19, | |
| "learning_rate": 1.1968085106382978e-05, | |
| "loss": 0.3093, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 1.2047872340425532e-05, | |
| "loss": 0.1904, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 3.23, | |
| "learning_rate": 1.2127659574468084e-05, | |
| "loss": 0.2457, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 3.26, | |
| "learning_rate": 1.2207446808510638e-05, | |
| "loss": 0.2975, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 3.28, | |
| "learning_rate": 1.2287234042553192e-05, | |
| "loss": 0.4788, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 3.3, | |
| "learning_rate": 1.2367021276595745e-05, | |
| "loss": 0.2399, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 3.32, | |
| "learning_rate": 1.2446808510638298e-05, | |
| "loss": 0.145, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 3.34, | |
| "learning_rate": 1.2526595744680851e-05, | |
| "loss": 0.3222, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "learning_rate": 1.2606382978723405e-05, | |
| "loss": 0.2309, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 3.38, | |
| "learning_rate": 1.2686170212765957e-05, | |
| "loss": 0.5458, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 1.2765957446808511e-05, | |
| "loss": 0.1605, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "learning_rate": 1.2845744680851065e-05, | |
| "loss": 0.2338, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 3.45, | |
| "learning_rate": 1.2925531914893617e-05, | |
| "loss": 0.3412, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "learning_rate": 1.300531914893617e-05, | |
| "loss": 0.2987, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 1.3085106382978724e-05, | |
| "loss": 0.3609, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "learning_rate": 1.3164893617021277e-05, | |
| "loss": 0.241, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "learning_rate": 1.324468085106383e-05, | |
| "loss": 0.2317, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "learning_rate": 1.3324468085106384e-05, | |
| "loss": 0.2111, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 1.3404255319148936e-05, | |
| "loss": 0.3572, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 1.348404255319149e-05, | |
| "loss": 0.2214, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 1.3563829787234044e-05, | |
| "loss": 0.2236, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 3.64, | |
| "learning_rate": 1.3643617021276596e-05, | |
| "loss": 0.4217, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 3.66, | |
| "learning_rate": 1.372340425531915e-05, | |
| "loss": 0.2764, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "learning_rate": 1.3803191489361704e-05, | |
| "loss": 0.3796, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 3.7, | |
| "learning_rate": 1.3882978723404256e-05, | |
| "loss": 0.3644, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 3.72, | |
| "learning_rate": 1.3962765957446808e-05, | |
| "loss": 0.2486, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 1.4042553191489362e-05, | |
| "loss": 0.2654, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 3.77, | |
| "learning_rate": 1.4122340425531915e-05, | |
| "loss": 0.1885, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 3.79, | |
| "learning_rate": 1.4202127659574467e-05, | |
| "loss": 0.3417, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "learning_rate": 1.4281914893617021e-05, | |
| "loss": 0.4061, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 3.83, | |
| "learning_rate": 1.4361702127659575e-05, | |
| "loss": 0.1553, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 3.85, | |
| "learning_rate": 1.4441489361702127e-05, | |
| "loss": 0.1909, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 1.4521276595744681e-05, | |
| "loss": 0.1842, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 3.89, | |
| "learning_rate": 1.4601063829787235e-05, | |
| "loss": 0.2601, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 3.91, | |
| "learning_rate": 1.4680851063829787e-05, | |
| "loss": 0.3434, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 3.94, | |
| "learning_rate": 1.476063829787234e-05, | |
| "loss": 0.1287, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "learning_rate": 1.4840425531914894e-05, | |
| "loss": 0.348, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 1.4920212765957447e-05, | |
| "loss": 0.2517, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 1.5e-05, | |
| "loss": 0.0854, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_accuracy": 0.9666666666666667, | |
| "eval_loss": 0.15290077030658722, | |
| "eval_runtime": 25.2794, | |
| "eval_samples_per_second": 5.934, | |
| "eval_steps_per_second": 0.396, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 1.499113475177305e-05, | |
| "loss": 0.2641, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 4.04, | |
| "learning_rate": 1.49822695035461e-05, | |
| "loss": 0.1313, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 4.06, | |
| "learning_rate": 1.497340425531915e-05, | |
| "loss": 0.2177, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 4.09, | |
| "learning_rate": 1.4964539007092199e-05, | |
| "loss": 0.3097, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 4.11, | |
| "learning_rate": 1.4955673758865248e-05, | |
| "loss": 0.2605, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 4.13, | |
| "learning_rate": 1.4946808510638299e-05, | |
| "loss": 0.2271, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 4.15, | |
| "learning_rate": 1.4937943262411348e-05, | |
| "loss": 0.2107, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 4.17, | |
| "learning_rate": 1.4929078014184397e-05, | |
| "loss": 0.136, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 4.19, | |
| "learning_rate": 1.4920212765957447e-05, | |
| "loss": 0.2588, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 4.21, | |
| "learning_rate": 1.4911347517730497e-05, | |
| "loss": 0.3398, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 4.23, | |
| "learning_rate": 1.4902482269503547e-05, | |
| "loss": 0.3291, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 4.26, | |
| "learning_rate": 1.4893617021276596e-05, | |
| "loss": 0.2509, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 4.28, | |
| "learning_rate": 1.4884751773049645e-05, | |
| "loss": 0.2343, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 4.3, | |
| "learning_rate": 1.4875886524822696e-05, | |
| "loss": 0.1709, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 4.32, | |
| "learning_rate": 1.4867021276595745e-05, | |
| "loss": 0.1216, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 4.34, | |
| "learning_rate": 1.4858156028368794e-05, | |
| "loss": 0.2355, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 4.36, | |
| "learning_rate": 1.4849290780141844e-05, | |
| "loss": 0.1234, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 4.38, | |
| "learning_rate": 1.4840425531914894e-05, | |
| "loss": 0.0835, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 1.4831560283687944e-05, | |
| "loss": 0.1898, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 4.43, | |
| "learning_rate": 1.4822695035460993e-05, | |
| "loss": 0.1608, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 4.45, | |
| "learning_rate": 1.4813829787234042e-05, | |
| "loss": 0.2743, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 4.47, | |
| "learning_rate": 1.4804964539007093e-05, | |
| "loss": 0.223, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 4.49, | |
| "learning_rate": 1.4796099290780142e-05, | |
| "loss": 0.1756, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 4.51, | |
| "learning_rate": 1.4787234042553191e-05, | |
| "loss": 0.1692, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 4.53, | |
| "learning_rate": 1.477836879432624e-05, | |
| "loss": 0.2501, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 4.55, | |
| "learning_rate": 1.4769503546099291e-05, | |
| "loss": 0.1627, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 4.57, | |
| "learning_rate": 1.476063829787234e-05, | |
| "loss": 0.3239, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 4.6, | |
| "learning_rate": 1.475177304964539e-05, | |
| "loss": 0.161, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 4.62, | |
| "learning_rate": 1.4742907801418439e-05, | |
| "loss": 0.1908, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 4.64, | |
| "learning_rate": 1.473404255319149e-05, | |
| "loss": 0.1159, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 4.66, | |
| "learning_rate": 1.472517730496454e-05, | |
| "loss": 0.131, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 4.68, | |
| "learning_rate": 1.4716312056737588e-05, | |
| "loss": 0.3526, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 4.7, | |
| "learning_rate": 1.470744680851064e-05, | |
| "loss": 0.1874, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 4.72, | |
| "learning_rate": 1.4698581560283688e-05, | |
| "loss": 0.2784, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 4.74, | |
| "learning_rate": 1.4689716312056738e-05, | |
| "loss": 0.217, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 4.77, | |
| "learning_rate": 1.4680851063829787e-05, | |
| "loss": 0.1852, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 4.79, | |
| "learning_rate": 1.4671985815602838e-05, | |
| "loss": 0.202, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 4.81, | |
| "learning_rate": 1.4663120567375887e-05, | |
| "loss": 0.3029, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 4.83, | |
| "learning_rate": 1.4654255319148936e-05, | |
| "loss": 0.2096, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 4.85, | |
| "learning_rate": 1.4645390070921985e-05, | |
| "loss": 0.2709, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 4.87, | |
| "learning_rate": 1.4636524822695036e-05, | |
| "loss": 0.1944, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 4.89, | |
| "learning_rate": 1.4627659574468085e-05, | |
| "loss": 0.2461, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 4.91, | |
| "learning_rate": 1.4618794326241135e-05, | |
| "loss": 0.1252, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 4.94, | |
| "learning_rate": 1.4609929078014184e-05, | |
| "loss": 0.4034, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 4.96, | |
| "learning_rate": 1.4601063829787235e-05, | |
| "loss": 0.219, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 4.98, | |
| "learning_rate": 1.4592198581560284e-05, | |
| "loss": 0.3115, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "learning_rate": 1.4583333333333333e-05, | |
| "loss": 0.2766, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_accuracy": 0.9666666666666667, | |
| "eval_loss": 0.14656034111976624, | |
| "eval_runtime": 25.2742, | |
| "eval_samples_per_second": 5.935, | |
| "eval_steps_per_second": 0.396, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 5.02, | |
| "learning_rate": 1.4574468085106382e-05, | |
| "loss": 0.2147, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 5.04, | |
| "learning_rate": 1.4565602836879433e-05, | |
| "loss": 0.2162, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 5.06, | |
| "learning_rate": 1.4556737588652482e-05, | |
| "loss": 0.3667, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 5.09, | |
| "learning_rate": 1.4547872340425532e-05, | |
| "loss": 0.1173, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 5.11, | |
| "learning_rate": 1.4539007092198581e-05, | |
| "loss": 0.1122, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 5.13, | |
| "learning_rate": 1.4530141843971632e-05, | |
| "loss": 0.2879, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 5.15, | |
| "learning_rate": 1.4521276595744681e-05, | |
| "loss": 0.1473, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 5.17, | |
| "learning_rate": 1.451241134751773e-05, | |
| "loss": 0.2639, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 5.19, | |
| "learning_rate": 1.450354609929078e-05, | |
| "loss": 0.1501, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 5.21, | |
| "learning_rate": 1.449468085106383e-05, | |
| "loss": 0.1987, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 5.23, | |
| "learning_rate": 1.448581560283688e-05, | |
| "loss": 0.1389, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 5.26, | |
| "learning_rate": 1.4476950354609929e-05, | |
| "loss": 0.2079, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 5.28, | |
| "learning_rate": 1.4468085106382978e-05, | |
| "loss": 0.1424, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 5.3, | |
| "learning_rate": 1.4459219858156029e-05, | |
| "loss": 0.1236, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 5.32, | |
| "learning_rate": 1.4450354609929078e-05, | |
| "loss": 0.1392, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 5.34, | |
| "learning_rate": 1.4441489361702127e-05, | |
| "loss": 0.3196, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 5.36, | |
| "learning_rate": 1.4432624113475176e-05, | |
| "loss": 0.2522, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 5.38, | |
| "learning_rate": 1.4423758865248227e-05, | |
| "loss": 0.1305, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 5.4, | |
| "learning_rate": 1.4414893617021276e-05, | |
| "loss": 0.1663, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 5.43, | |
| "learning_rate": 1.4406028368794326e-05, | |
| "loss": 0.2762, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 5.45, | |
| "learning_rate": 1.4397163120567377e-05, | |
| "loss": 0.1044, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 5.47, | |
| "learning_rate": 1.4388297872340426e-05, | |
| "loss": 0.2071, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 5.49, | |
| "learning_rate": 1.4379432624113475e-05, | |
| "loss": 0.1581, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 5.51, | |
| "learning_rate": 1.4370567375886524e-05, | |
| "loss": 0.1798, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 5.53, | |
| "learning_rate": 1.4361702127659575e-05, | |
| "loss": 0.0981, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 5.55, | |
| "learning_rate": 1.4352836879432624e-05, | |
| "loss": 0.1568, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 5.57, | |
| "learning_rate": 1.4343971631205673e-05, | |
| "loss": 0.1833, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 5.6, | |
| "learning_rate": 1.4335106382978723e-05, | |
| "loss": 0.1484, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 5.62, | |
| "learning_rate": 1.4326241134751774e-05, | |
| "loss": 0.1887, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 5.64, | |
| "learning_rate": 1.4317375886524823e-05, | |
| "loss": 0.1635, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 5.66, | |
| "learning_rate": 1.4308510638297872e-05, | |
| "loss": 0.13, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 5.68, | |
| "learning_rate": 1.4299645390070921e-05, | |
| "loss": 0.2125, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 5.7, | |
| "learning_rate": 1.4290780141843972e-05, | |
| "loss": 0.2993, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 5.72, | |
| "learning_rate": 1.4281914893617021e-05, | |
| "loss": 0.2421, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 5.74, | |
| "learning_rate": 1.427304964539007e-05, | |
| "loss": 0.1027, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 5.77, | |
| "learning_rate": 1.426418439716312e-05, | |
| "loss": 0.2153, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 5.79, | |
| "learning_rate": 1.425531914893617e-05, | |
| "loss": 0.2365, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 5.81, | |
| "learning_rate": 1.424645390070922e-05, | |
| "loss": 0.0801, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 5.83, | |
| "learning_rate": 1.4237588652482269e-05, | |
| "loss": 0.2231, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 5.85, | |
| "learning_rate": 1.4228723404255318e-05, | |
| "loss": 0.0905, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 5.87, | |
| "learning_rate": 1.4219858156028369e-05, | |
| "loss": 0.2794, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 5.89, | |
| "learning_rate": 1.4210992907801418e-05, | |
| "loss": 0.1411, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 5.91, | |
| "learning_rate": 1.4202127659574467e-05, | |
| "loss": 0.1906, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 5.94, | |
| "learning_rate": 1.4193262411347517e-05, | |
| "loss": 0.1982, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 5.96, | |
| "learning_rate": 1.4184397163120568e-05, | |
| "loss": 0.1509, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 5.98, | |
| "learning_rate": 1.4175531914893617e-05, | |
| "loss": 0.1661, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "learning_rate": 1.4166666666666666e-05, | |
| "loss": 0.3158, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "eval_accuracy": 0.96, | |
| "eval_loss": 0.19158737361431122, | |
| "eval_runtime": 25.1976, | |
| "eval_samples_per_second": 5.953, | |
| "eval_steps_per_second": 0.397, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 6.02, | |
| "learning_rate": 1.4157801418439715e-05, | |
| "loss": 0.2351, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 6.04, | |
| "learning_rate": 1.4148936170212766e-05, | |
| "loss": 0.2252, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 6.06, | |
| "learning_rate": 1.4140070921985815e-05, | |
| "loss": 0.1855, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 6.09, | |
| "learning_rate": 1.4131205673758864e-05, | |
| "loss": 0.3494, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 6.11, | |
| "learning_rate": 1.4122340425531915e-05, | |
| "loss": 0.2276, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 6.13, | |
| "learning_rate": 1.4113475177304965e-05, | |
| "loss": 0.2324, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 6.15, | |
| "learning_rate": 1.4104609929078014e-05, | |
| "loss": 0.3129, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 6.17, | |
| "learning_rate": 1.4095744680851063e-05, | |
| "loss": 0.1555, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 6.19, | |
| "learning_rate": 1.4086879432624114e-05, | |
| "loss": 0.1816, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 6.21, | |
| "learning_rate": 1.4078014184397163e-05, | |
| "loss": 0.3329, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 6.23, | |
| "learning_rate": 1.4069148936170212e-05, | |
| "loss": 0.1315, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 6.26, | |
| "learning_rate": 1.4060283687943261e-05, | |
| "loss": 0.2297, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 6.28, | |
| "learning_rate": 1.4051418439716312e-05, | |
| "loss": 0.1636, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 6.3, | |
| "learning_rate": 1.4042553191489362e-05, | |
| "loss": 0.1921, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 6.32, | |
| "learning_rate": 1.403368794326241e-05, | |
| "loss": 0.193, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 6.34, | |
| "learning_rate": 1.402482269503546e-05, | |
| "loss": 0.1744, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 6.36, | |
| "learning_rate": 1.4015957446808511e-05, | |
| "loss": 0.2316, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 6.38, | |
| "learning_rate": 1.400709219858156e-05, | |
| "loss": 0.2121, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 6.4, | |
| "learning_rate": 1.399822695035461e-05, | |
| "loss": 0.1489, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 6.43, | |
| "learning_rate": 1.3989361702127658e-05, | |
| "loss": 0.1928, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 6.45, | |
| "learning_rate": 1.398049645390071e-05, | |
| "loss": 0.24, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 6.47, | |
| "learning_rate": 1.3971631205673759e-05, | |
| "loss": 0.1377, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 6.49, | |
| "learning_rate": 1.3962765957446808e-05, | |
| "loss": 0.1917, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 6.51, | |
| "learning_rate": 1.3953900709219857e-05, | |
| "loss": 0.2619, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 6.53, | |
| "learning_rate": 1.3945035460992908e-05, | |
| "loss": 0.2478, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 6.55, | |
| "learning_rate": 1.3936170212765957e-05, | |
| "loss": 0.1657, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 6.57, | |
| "learning_rate": 1.3927304964539008e-05, | |
| "loss": 0.143, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 6.6, | |
| "learning_rate": 1.3918439716312057e-05, | |
| "loss": 0.1996, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 6.62, | |
| "learning_rate": 1.3909574468085108e-05, | |
| "loss": 0.0984, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 6.64, | |
| "learning_rate": 1.3900709219858157e-05, | |
| "loss": 0.1056, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 6.66, | |
| "learning_rate": 1.3891843971631206e-05, | |
| "loss": 0.237, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 6.68, | |
| "learning_rate": 1.3882978723404256e-05, | |
| "loss": 0.0908, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 6.7, | |
| "learning_rate": 1.3874113475177307e-05, | |
| "loss": 0.0977, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 6.72, | |
| "learning_rate": 1.3865248226950356e-05, | |
| "loss": 0.0932, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 6.74, | |
| "learning_rate": 1.3856382978723405e-05, | |
| "loss": 0.1781, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 6.77, | |
| "learning_rate": 1.3847517730496454e-05, | |
| "loss": 0.2247, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 6.79, | |
| "learning_rate": 1.3838652482269505e-05, | |
| "loss": 0.1969, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 6.81, | |
| "learning_rate": 1.3829787234042554e-05, | |
| "loss": 0.2142, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 6.83, | |
| "learning_rate": 1.3820921985815603e-05, | |
| "loss": 0.3049, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 6.85, | |
| "learning_rate": 1.3812056737588654e-05, | |
| "loss": 0.1278, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 6.87, | |
| "learning_rate": 1.3803191489361704e-05, | |
| "loss": 0.138, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 6.89, | |
| "learning_rate": 1.3794326241134753e-05, | |
| "loss": 0.1893, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 6.91, | |
| "learning_rate": 1.3785460992907802e-05, | |
| "loss": 0.2039, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 6.94, | |
| "learning_rate": 1.3776595744680853e-05, | |
| "loss": 0.2903, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 6.96, | |
| "learning_rate": 1.3767730496453902e-05, | |
| "loss": 0.3838, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 6.98, | |
| "learning_rate": 1.3758865248226951e-05, | |
| "loss": 0.2625, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "learning_rate": 1.375e-05, | |
| "loss": 0.1322, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "eval_accuracy": 0.9733333333333334, | |
| "eval_loss": 0.19238106906414032, | |
| "eval_runtime": 25.1474, | |
| "eval_samples_per_second": 5.965, | |
| "eval_steps_per_second": 0.398, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 7.02, | |
| "learning_rate": 1.3741134751773051e-05, | |
| "loss": 0.0685, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 7.04, | |
| "learning_rate": 1.37322695035461e-05, | |
| "loss": 0.1413, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 7.06, | |
| "learning_rate": 1.372340425531915e-05, | |
| "loss": 0.1239, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 7.09, | |
| "learning_rate": 1.3714539007092199e-05, | |
| "loss": 0.0993, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 7.11, | |
| "learning_rate": 1.370567375886525e-05, | |
| "loss": 0.1668, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 7.13, | |
| "learning_rate": 1.3696808510638299e-05, | |
| "loss": 0.2324, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 7.15, | |
| "learning_rate": 1.3687943262411348e-05, | |
| "loss": 0.0858, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 7.17, | |
| "learning_rate": 1.3679078014184397e-05, | |
| "loss": 0.1086, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 7.19, | |
| "learning_rate": 1.3670212765957448e-05, | |
| "loss": 0.1754, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 7.21, | |
| "learning_rate": 1.3661347517730498e-05, | |
| "loss": 0.149, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 7.23, | |
| "learning_rate": 1.3652482269503547e-05, | |
| "loss": 0.2183, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 7.26, | |
| "learning_rate": 1.3643617021276596e-05, | |
| "loss": 0.1504, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 7.28, | |
| "learning_rate": 1.3634751773049647e-05, | |
| "loss": 0.1531, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 7.3, | |
| "learning_rate": 1.3625886524822696e-05, | |
| "loss": 0.0925, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 7.32, | |
| "learning_rate": 1.3617021276595745e-05, | |
| "loss": 0.1442, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 7.34, | |
| "learning_rate": 1.3608156028368794e-05, | |
| "loss": 0.1971, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 7.36, | |
| "learning_rate": 1.3599290780141845e-05, | |
| "loss": 0.1356, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 7.38, | |
| "learning_rate": 1.3590425531914895e-05, | |
| "loss": 0.0792, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 7.4, | |
| "learning_rate": 1.3581560283687944e-05, | |
| "loss": 0.0933, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 7.43, | |
| "learning_rate": 1.3572695035460993e-05, | |
| "loss": 0.1055, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 7.45, | |
| "learning_rate": 1.3563829787234044e-05, | |
| "loss": 0.2178, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 7.47, | |
| "learning_rate": 1.3554964539007093e-05, | |
| "loss": 0.1075, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 7.49, | |
| "learning_rate": 1.3546099290780142e-05, | |
| "loss": 0.2189, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 7.51, | |
| "learning_rate": 1.3537234042553193e-05, | |
| "loss": 0.1785, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 7.53, | |
| "learning_rate": 1.3528368794326242e-05, | |
| "loss": 0.1363, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 7.55, | |
| "learning_rate": 1.3519503546099292e-05, | |
| "loss": 0.1779, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 7.57, | |
| "learning_rate": 1.351063829787234e-05, | |
| "loss": 0.138, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 7.6, | |
| "learning_rate": 1.3501773049645392e-05, | |
| "loss": 0.1659, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 7.62, | |
| "learning_rate": 1.3492907801418441e-05, | |
| "loss": 0.0913, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 7.64, | |
| "learning_rate": 1.348404255319149e-05, | |
| "loss": 0.2239, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 7.66, | |
| "learning_rate": 1.347517730496454e-05, | |
| "loss": 0.1361, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 7.68, | |
| "learning_rate": 1.346631205673759e-05, | |
| "loss": 0.1915, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 7.7, | |
| "learning_rate": 1.345744680851064e-05, | |
| "loss": 0.2005, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 7.72, | |
| "learning_rate": 1.3448581560283689e-05, | |
| "loss": 0.1947, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 7.74, | |
| "learning_rate": 1.3439716312056738e-05, | |
| "loss": 0.2874, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 7.77, | |
| "learning_rate": 1.3430851063829789e-05, | |
| "loss": 0.0842, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 7.79, | |
| "learning_rate": 1.3421985815602838e-05, | |
| "loss": 0.2964, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 7.81, | |
| "learning_rate": 1.3413120567375887e-05, | |
| "loss": 0.178, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 7.83, | |
| "learning_rate": 1.3404255319148936e-05, | |
| "loss": 0.1066, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 7.85, | |
| "learning_rate": 1.3395390070921987e-05, | |
| "loss": 0.1075, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 7.87, | |
| "learning_rate": 1.3386524822695036e-05, | |
| "loss": 0.2262, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 7.89, | |
| "learning_rate": 1.3377659574468086e-05, | |
| "loss": 0.18, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 7.91, | |
| "learning_rate": 1.3368794326241135e-05, | |
| "loss": 0.1103, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 7.94, | |
| "learning_rate": 1.3359929078014186e-05, | |
| "loss": 0.1307, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 7.96, | |
| "learning_rate": 1.3351063829787235e-05, | |
| "loss": 0.1084, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 7.98, | |
| "learning_rate": 1.3342198581560284e-05, | |
| "loss": 0.1343, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "learning_rate": 1.3333333333333333e-05, | |
| "loss": 0.065, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "eval_accuracy": 0.9533333333333334, | |
| "eval_loss": 0.19049762189388275, | |
| "eval_runtime": 28.9286, | |
| "eval_samples_per_second": 5.185, | |
| "eval_steps_per_second": 0.346, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 8.02, | |
| "learning_rate": 1.3324468085106384e-05, | |
| "loss": 0.2245, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 8.04, | |
| "learning_rate": 1.3315602836879433e-05, | |
| "loss": 0.178, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 8.06, | |
| "learning_rate": 1.3306737588652483e-05, | |
| "loss": 0.238, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 8.09, | |
| "learning_rate": 1.3297872340425532e-05, | |
| "loss": 0.1443, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 8.11, | |
| "learning_rate": 1.3289007092198583e-05, | |
| "loss": 0.142, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 8.13, | |
| "learning_rate": 1.3280141843971632e-05, | |
| "loss": 0.0795, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 8.15, | |
| "learning_rate": 1.3271276595744681e-05, | |
| "loss": 0.1447, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 8.17, | |
| "learning_rate": 1.3262411347517732e-05, | |
| "loss": 0.2268, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 8.19, | |
| "learning_rate": 1.3253546099290781e-05, | |
| "loss": 0.2079, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 8.21, | |
| "learning_rate": 1.324468085106383e-05, | |
| "loss": 0.1702, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 8.23, | |
| "learning_rate": 1.323581560283688e-05, | |
| "loss": 0.279, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 8.26, | |
| "learning_rate": 1.322695035460993e-05, | |
| "loss": 0.1325, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 8.28, | |
| "learning_rate": 1.321808510638298e-05, | |
| "loss": 0.1672, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 8.3, | |
| "learning_rate": 1.3209219858156029e-05, | |
| "loss": 0.1856, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 8.32, | |
| "learning_rate": 1.3200354609929078e-05, | |
| "loss": 0.1052, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 8.34, | |
| "learning_rate": 1.3191489361702129e-05, | |
| "loss": 0.132, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 8.36, | |
| "learning_rate": 1.3182624113475178e-05, | |
| "loss": 0.1395, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 8.38, | |
| "learning_rate": 1.3173758865248227e-05, | |
| "loss": 0.1697, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 8.4, | |
| "learning_rate": 1.3164893617021277e-05, | |
| "loss": 0.1875, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 8.43, | |
| "learning_rate": 1.3156028368794327e-05, | |
| "loss": 0.1638, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 8.45, | |
| "learning_rate": 1.3147163120567377e-05, | |
| "loss": 0.1102, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 8.47, | |
| "learning_rate": 1.3138297872340426e-05, | |
| "loss": 0.3231, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 8.49, | |
| "learning_rate": 1.3129432624113475e-05, | |
| "loss": 0.1934, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 8.51, | |
| "learning_rate": 1.3120567375886526e-05, | |
| "loss": 0.2385, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 8.53, | |
| "learning_rate": 1.3111702127659575e-05, | |
| "loss": 0.3054, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 8.55, | |
| "learning_rate": 1.3102836879432624e-05, | |
| "loss": 0.1647, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 8.57, | |
| "learning_rate": 1.3093971631205674e-05, | |
| "loss": 0.1654, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 8.6, | |
| "learning_rate": 1.3085106382978724e-05, | |
| "loss": 0.057, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 8.62, | |
| "learning_rate": 1.3076241134751774e-05, | |
| "loss": 0.2054, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 8.64, | |
| "learning_rate": 1.3067375886524823e-05, | |
| "loss": 0.0559, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 8.66, | |
| "learning_rate": 1.3058510638297872e-05, | |
| "loss": 0.1806, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 8.68, | |
| "learning_rate": 1.3049645390070923e-05, | |
| "loss": 0.2999, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 8.7, | |
| "learning_rate": 1.3040780141843972e-05, | |
| "loss": 0.1729, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 8.72, | |
| "learning_rate": 1.3031914893617021e-05, | |
| "loss": 0.2028, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 8.74, | |
| "learning_rate": 1.302304964539007e-05, | |
| "loss": 0.1871, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 8.77, | |
| "learning_rate": 1.3014184397163121e-05, | |
| "loss": 0.1262, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 8.79, | |
| "learning_rate": 1.300531914893617e-05, | |
| "loss": 0.1197, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 8.81, | |
| "learning_rate": 1.299645390070922e-05, | |
| "loss": 0.1675, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 8.83, | |
| "learning_rate": 1.2987588652482269e-05, | |
| "loss": 0.1084, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 8.85, | |
| "learning_rate": 1.297872340425532e-05, | |
| "loss": 0.2089, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 8.87, | |
| "learning_rate": 1.296985815602837e-05, | |
| "loss": 0.1522, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 8.89, | |
| "learning_rate": 1.2960992907801418e-05, | |
| "loss": 0.1595, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 8.91, | |
| "learning_rate": 1.295212765957447e-05, | |
| "loss": 0.0795, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 8.94, | |
| "learning_rate": 1.2943262411347518e-05, | |
| "loss": 0.1784, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 8.96, | |
| "learning_rate": 1.2934397163120568e-05, | |
| "loss": 0.1539, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 8.98, | |
| "learning_rate": 1.2925531914893617e-05, | |
| "loss": 0.1593, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "learning_rate": 1.2916666666666668e-05, | |
| "loss": 0.1565, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "eval_accuracy": 0.9466666666666667, | |
| "eval_loss": 0.20253299176692963, | |
| "eval_runtime": 28.6412, | |
| "eval_samples_per_second": 5.237, | |
| "eval_steps_per_second": 0.349, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 9.02, | |
| "learning_rate": 1.2907801418439717e-05, | |
| "loss": 0.2258, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 9.04, | |
| "learning_rate": 1.2898936170212766e-05, | |
| "loss": 0.2443, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 9.06, | |
| "learning_rate": 1.2890070921985815e-05, | |
| "loss": 0.1435, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 9.09, | |
| "learning_rate": 1.2881205673758866e-05, | |
| "loss": 0.1073, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 9.11, | |
| "learning_rate": 1.2872340425531915e-05, | |
| "loss": 0.231, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 9.13, | |
| "learning_rate": 1.2863475177304965e-05, | |
| "loss": 0.0673, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 9.15, | |
| "learning_rate": 1.2854609929078014e-05, | |
| "loss": 0.0983, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 9.17, | |
| "learning_rate": 1.2845744680851065e-05, | |
| "loss": 0.1906, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 9.19, | |
| "learning_rate": 1.2836879432624114e-05, | |
| "loss": 0.276, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 9.21, | |
| "learning_rate": 1.2828014184397163e-05, | |
| "loss": 0.2517, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 9.23, | |
| "learning_rate": 1.2819148936170212e-05, | |
| "loss": 0.2482, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 9.26, | |
| "learning_rate": 1.2810283687943263e-05, | |
| "loss": 0.0767, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 9.28, | |
| "learning_rate": 1.2801418439716312e-05, | |
| "loss": 0.1493, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 9.3, | |
| "learning_rate": 1.2792553191489362e-05, | |
| "loss": 0.1712, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 9.32, | |
| "learning_rate": 1.2783687943262411e-05, | |
| "loss": 0.134, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 9.34, | |
| "learning_rate": 1.2774822695035462e-05, | |
| "loss": 0.1837, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 9.36, | |
| "learning_rate": 1.2765957446808511e-05, | |
| "loss": 0.0289, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 9.38, | |
| "learning_rate": 1.275709219858156e-05, | |
| "loss": 0.3248, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 9.4, | |
| "learning_rate": 1.274822695035461e-05, | |
| "loss": 0.196, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 9.43, | |
| "learning_rate": 1.273936170212766e-05, | |
| "loss": 0.0722, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 9.45, | |
| "learning_rate": 1.273049645390071e-05, | |
| "loss": 0.089, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 9.47, | |
| "learning_rate": 1.2721631205673759e-05, | |
| "loss": 0.137, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 9.49, | |
| "learning_rate": 1.2712765957446808e-05, | |
| "loss": 0.2285, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 9.51, | |
| "learning_rate": 1.2703900709219859e-05, | |
| "loss": 0.1613, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 9.53, | |
| "learning_rate": 1.2695035460992908e-05, | |
| "loss": 0.133, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 9.55, | |
| "learning_rate": 1.2686170212765957e-05, | |
| "loss": 0.1625, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 9.57, | |
| "learning_rate": 1.2677304964539008e-05, | |
| "loss": 0.172, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 9.6, | |
| "learning_rate": 1.2668439716312057e-05, | |
| "loss": 0.2983, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 9.62, | |
| "learning_rate": 1.2659574468085106e-05, | |
| "loss": 0.2477, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 9.64, | |
| "learning_rate": 1.2650709219858156e-05, | |
| "loss": 0.2189, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 9.66, | |
| "learning_rate": 1.2641843971631207e-05, | |
| "loss": 0.207, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 9.68, | |
| "learning_rate": 1.2632978723404256e-05, | |
| "loss": 0.3683, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 9.7, | |
| "learning_rate": 1.2624113475177305e-05, | |
| "loss": 0.2673, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 9.72, | |
| "learning_rate": 1.2615248226950354e-05, | |
| "loss": 0.2192, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 9.74, | |
| "learning_rate": 1.2606382978723405e-05, | |
| "loss": 0.1778, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 9.77, | |
| "learning_rate": 1.2597517730496454e-05, | |
| "loss": 0.1843, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 9.79, | |
| "learning_rate": 1.2588652482269504e-05, | |
| "loss": 0.2419, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 9.81, | |
| "learning_rate": 1.2579787234042553e-05, | |
| "loss": 0.1821, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 9.83, | |
| "learning_rate": 1.2570921985815604e-05, | |
| "loss": 0.3082, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 9.85, | |
| "learning_rate": 1.2562056737588653e-05, | |
| "loss": 0.1197, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 9.87, | |
| "learning_rate": 1.2553191489361702e-05, | |
| "loss": 0.1454, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 9.89, | |
| "learning_rate": 1.2544326241134751e-05, | |
| "loss": 0.106, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 9.91, | |
| "learning_rate": 1.2535460992907802e-05, | |
| "loss": 0.1494, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 9.94, | |
| "learning_rate": 1.2526595744680851e-05, | |
| "loss": 0.1983, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 9.96, | |
| "learning_rate": 1.25177304964539e-05, | |
| "loss": 0.2089, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 9.98, | |
| "learning_rate": 1.250886524822695e-05, | |
| "loss": 0.1771, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "learning_rate": 1.25e-05, | |
| "loss": 0.1296, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "eval_accuracy": 0.9333333333333333, | |
| "eval_loss": 0.23667007684707642, | |
| "eval_runtime": 28.4672, | |
| "eval_samples_per_second": 5.269, | |
| "eval_steps_per_second": 0.351, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 10.02, | |
| "learning_rate": 1.249113475177305e-05, | |
| "loss": 0.1104, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 10.04, | |
| "learning_rate": 1.2482269503546099e-05, | |
| "loss": 0.1753, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 10.06, | |
| "learning_rate": 1.2473404255319148e-05, | |
| "loss": 0.1356, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 10.09, | |
| "learning_rate": 1.2464539007092199e-05, | |
| "loss": 0.0924, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 10.11, | |
| "learning_rate": 1.2455673758865248e-05, | |
| "loss": 0.1512, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 10.13, | |
| "learning_rate": 1.2446808510638298e-05, | |
| "loss": 0.1471, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 10.15, | |
| "learning_rate": 1.2437943262411347e-05, | |
| "loss": 0.1833, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 10.17, | |
| "learning_rate": 1.2429078014184398e-05, | |
| "loss": 0.2836, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 10.19, | |
| "learning_rate": 1.2420212765957447e-05, | |
| "loss": 0.1223, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 10.21, | |
| "learning_rate": 1.2411347517730496e-05, | |
| "loss": 0.148, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 10.23, | |
| "learning_rate": 1.2402482269503547e-05, | |
| "loss": 0.271, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 10.26, | |
| "learning_rate": 1.2393617021276596e-05, | |
| "loss": 0.1023, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 10.28, | |
| "learning_rate": 1.2384751773049645e-05, | |
| "loss": 0.123, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 10.3, | |
| "learning_rate": 1.2375886524822695e-05, | |
| "loss": 0.1744, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 10.32, | |
| "learning_rate": 1.2367021276595745e-05, | |
| "loss": 0.0978, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 10.34, | |
| "learning_rate": 1.2358156028368795e-05, | |
| "loss": 0.2175, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 10.36, | |
| "learning_rate": 1.2349290780141844e-05, | |
| "loss": 0.1131, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 10.38, | |
| "learning_rate": 1.2340425531914893e-05, | |
| "loss": 0.0598, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 10.4, | |
| "learning_rate": 1.2331560283687944e-05, | |
| "loss": 0.1394, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 10.43, | |
| "learning_rate": 1.2322695035460993e-05, | |
| "loss": 0.1909, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 10.45, | |
| "learning_rate": 1.2313829787234042e-05, | |
| "loss": 0.0563, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 10.47, | |
| "learning_rate": 1.2304964539007092e-05, | |
| "loss": 0.2545, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 10.49, | |
| "learning_rate": 1.2296099290780142e-05, | |
| "loss": 0.0948, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 10.51, | |
| "learning_rate": 1.2287234042553192e-05, | |
| "loss": 0.2104, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 10.53, | |
| "learning_rate": 1.227836879432624e-05, | |
| "loss": 0.0963, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 10.55, | |
| "learning_rate": 1.226950354609929e-05, | |
| "loss": 0.2494, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 10.57, | |
| "learning_rate": 1.2260638297872341e-05, | |
| "loss": 0.148, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 10.6, | |
| "learning_rate": 1.225177304964539e-05, | |
| "loss": 0.2809, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 10.62, | |
| "learning_rate": 1.224290780141844e-05, | |
| "loss": 0.1528, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 10.64, | |
| "learning_rate": 1.2234042553191489e-05, | |
| "loss": 0.1734, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 10.66, | |
| "learning_rate": 1.222517730496454e-05, | |
| "loss": 0.1837, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 10.68, | |
| "learning_rate": 1.2216312056737589e-05, | |
| "loss": 0.0968, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 10.7, | |
| "learning_rate": 1.2207446808510638e-05, | |
| "loss": 0.2506, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 10.72, | |
| "learning_rate": 1.2198581560283687e-05, | |
| "loss": 0.0743, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 10.74, | |
| "learning_rate": 1.2189716312056738e-05, | |
| "loss": 0.173, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 10.77, | |
| "learning_rate": 1.2180851063829787e-05, | |
| "loss": 0.0552, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 10.79, | |
| "learning_rate": 1.2171985815602836e-05, | |
| "loss": 0.2641, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 10.81, | |
| "learning_rate": 1.2163120567375886e-05, | |
| "loss": 0.2161, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 10.83, | |
| "learning_rate": 1.2154255319148936e-05, | |
| "loss": 0.1642, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 10.85, | |
| "learning_rate": 1.2145390070921986e-05, | |
| "loss": 0.0607, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 10.87, | |
| "learning_rate": 1.2136524822695035e-05, | |
| "loss": 0.1426, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 10.89, | |
| "learning_rate": 1.2127659574468084e-05, | |
| "loss": 0.0666, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 10.91, | |
| "learning_rate": 1.2118794326241135e-05, | |
| "loss": 0.1533, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 10.94, | |
| "learning_rate": 1.2109929078014184e-05, | |
| "loss": 0.0975, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 10.96, | |
| "learning_rate": 1.2101063829787233e-05, | |
| "loss": 0.1948, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 10.98, | |
| "learning_rate": 1.2092198581560284e-05, | |
| "loss": 0.2029, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "learning_rate": 1.2083333333333333e-05, | |
| "loss": 0.2448, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "eval_accuracy": 0.94, | |
| "eval_loss": 0.22552601993083954, | |
| "eval_runtime": 30.6905, | |
| "eval_samples_per_second": 4.888, | |
| "eval_steps_per_second": 0.326, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 11.02, | |
| "learning_rate": 1.2074468085106383e-05, | |
| "loss": 0.2155, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 11.04, | |
| "learning_rate": 1.2065602836879432e-05, | |
| "loss": 0.1177, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 11.06, | |
| "learning_rate": 1.2056737588652483e-05, | |
| "loss": 0.2541, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 11.09, | |
| "learning_rate": 1.2047872340425532e-05, | |
| "loss": 0.1512, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 11.11, | |
| "learning_rate": 1.2039007092198581e-05, | |
| "loss": 0.114, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 11.13, | |
| "learning_rate": 1.203014184397163e-05, | |
| "loss": 0.2651, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 11.15, | |
| "learning_rate": 1.2021276595744681e-05, | |
| "loss": 0.1714, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 11.17, | |
| "learning_rate": 1.201241134751773e-05, | |
| "loss": 0.1603, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 11.19, | |
| "learning_rate": 1.200354609929078e-05, | |
| "loss": 0.1275, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 11.21, | |
| "learning_rate": 1.1994680851063829e-05, | |
| "loss": 0.1761, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 11.23, | |
| "learning_rate": 1.198581560283688e-05, | |
| "loss": 0.2852, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 11.26, | |
| "learning_rate": 1.1976950354609929e-05, | |
| "loss": 0.134, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 11.28, | |
| "learning_rate": 1.1968085106382978e-05, | |
| "loss": 0.1283, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 11.3, | |
| "learning_rate": 1.1959219858156027e-05, | |
| "loss": 0.2067, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 11.32, | |
| "learning_rate": 1.1950354609929078e-05, | |
| "loss": 0.1087, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 11.34, | |
| "learning_rate": 1.1941489361702127e-05, | |
| "loss": 0.1526, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 11.36, | |
| "learning_rate": 1.1932624113475177e-05, | |
| "loss": 0.1642, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 11.38, | |
| "learning_rate": 1.1923758865248226e-05, | |
| "loss": 0.1011, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 11.4, | |
| "learning_rate": 1.1914893617021277e-05, | |
| "loss": 0.0663, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 11.43, | |
| "learning_rate": 1.1906028368794326e-05, | |
| "loss": 0.1544, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 11.45, | |
| "learning_rate": 1.1897163120567375e-05, | |
| "loss": 0.0456, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 11.47, | |
| "learning_rate": 1.1888297872340424e-05, | |
| "loss": 0.1029, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 11.49, | |
| "learning_rate": 1.1879432624113475e-05, | |
| "loss": 0.1739, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 11.51, | |
| "learning_rate": 1.1870567375886524e-05, | |
| "loss": 0.1684, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 11.53, | |
| "learning_rate": 1.1861702127659574e-05, | |
| "loss": 0.1037, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 11.55, | |
| "learning_rate": 1.1852836879432623e-05, | |
| "loss": 0.0316, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 11.57, | |
| "learning_rate": 1.1843971631205674e-05, | |
| "loss": 0.087, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 11.6, | |
| "learning_rate": 1.1835106382978723e-05, | |
| "loss": 0.2221, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 11.62, | |
| "learning_rate": 1.1826241134751772e-05, | |
| "loss": 0.2113, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 11.64, | |
| "learning_rate": 1.1817375886524823e-05, | |
| "loss": 0.0396, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 11.66, | |
| "learning_rate": 1.1808510638297872e-05, | |
| "loss": 0.1843, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 11.68, | |
| "learning_rate": 1.1799645390070921e-05, | |
| "loss": 0.1738, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 11.7, | |
| "learning_rate": 1.1790780141843972e-05, | |
| "loss": 0.0559, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 11.72, | |
| "learning_rate": 1.1781914893617022e-05, | |
| "loss": 0.1944, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 11.74, | |
| "learning_rate": 1.1773049645390072e-05, | |
| "loss": 0.142, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 11.77, | |
| "learning_rate": 1.1764184397163122e-05, | |
| "loss": 0.1181, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 11.79, | |
| "learning_rate": 1.175531914893617e-05, | |
| "loss": 0.1508, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 11.81, | |
| "learning_rate": 1.1746453900709222e-05, | |
| "loss": 0.1853, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 11.83, | |
| "learning_rate": 1.1737588652482271e-05, | |
| "loss": 0.1372, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 11.85, | |
| "learning_rate": 1.172872340425532e-05, | |
| "loss": 0.1763, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 11.87, | |
| "learning_rate": 1.171985815602837e-05, | |
| "loss": 0.232, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 11.89, | |
| "learning_rate": 1.171099290780142e-05, | |
| "loss": 0.1057, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 11.91, | |
| "learning_rate": 1.170212765957447e-05, | |
| "loss": 0.2012, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 11.94, | |
| "learning_rate": 1.1693262411347519e-05, | |
| "loss": 0.1429, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 11.96, | |
| "learning_rate": 1.1684397163120568e-05, | |
| "loss": 0.1406, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 11.98, | |
| "learning_rate": 1.1675531914893619e-05, | |
| "loss": 0.0789, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "learning_rate": 1.1666666666666668e-05, | |
| "loss": 0.067, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "eval_accuracy": 0.94, | |
| "eval_loss": 0.23146581649780273, | |
| "eval_runtime": 29.2663, | |
| "eval_samples_per_second": 5.125, | |
| "eval_steps_per_second": 0.342, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 12.02, | |
| "learning_rate": 1.1657801418439717e-05, | |
| "loss": 0.0925, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 12.04, | |
| "learning_rate": 1.1648936170212766e-05, | |
| "loss": 0.1319, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 12.06, | |
| "learning_rate": 1.1640070921985817e-05, | |
| "loss": 0.1403, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 12.09, | |
| "learning_rate": 1.1631205673758866e-05, | |
| "loss": 0.0693, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 12.11, | |
| "learning_rate": 1.1622340425531916e-05, | |
| "loss": 0.1105, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 12.13, | |
| "learning_rate": 1.1613475177304965e-05, | |
| "loss": 0.0723, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 12.15, | |
| "learning_rate": 1.1604609929078016e-05, | |
| "loss": 0.2769, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 12.17, | |
| "learning_rate": 1.1595744680851065e-05, | |
| "loss": 0.0758, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 12.19, | |
| "learning_rate": 1.1586879432624114e-05, | |
| "loss": 0.1333, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 12.21, | |
| "learning_rate": 1.1578014184397163e-05, | |
| "loss": 0.1997, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 12.23, | |
| "learning_rate": 1.1569148936170214e-05, | |
| "loss": 0.1312, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 12.26, | |
| "learning_rate": 1.1560283687943263e-05, | |
| "loss": 0.1629, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 12.28, | |
| "learning_rate": 1.1551418439716313e-05, | |
| "loss": 0.1287, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 12.3, | |
| "learning_rate": 1.1542553191489362e-05, | |
| "loss": 0.2013, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 12.32, | |
| "learning_rate": 1.1533687943262413e-05, | |
| "loss": 0.0576, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 12.34, | |
| "learning_rate": 1.1524822695035462e-05, | |
| "loss": 0.2348, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 12.36, | |
| "learning_rate": 1.1515957446808511e-05, | |
| "loss": 0.1205, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 12.38, | |
| "learning_rate": 1.1507092198581562e-05, | |
| "loss": 0.0334, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 12.4, | |
| "learning_rate": 1.1498226950354611e-05, | |
| "loss": 0.0996, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 12.43, | |
| "learning_rate": 1.148936170212766e-05, | |
| "loss": 0.1054, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 12.45, | |
| "learning_rate": 1.148049645390071e-05, | |
| "loss": 0.2659, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 12.47, | |
| "learning_rate": 1.147163120567376e-05, | |
| "loss": 0.1041, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 12.49, | |
| "learning_rate": 1.146276595744681e-05, | |
| "loss": 0.2094, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 12.51, | |
| "learning_rate": 1.1453900709219859e-05, | |
| "loss": 0.2323, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 12.53, | |
| "learning_rate": 1.1445035460992908e-05, | |
| "loss": 0.1112, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 12.55, | |
| "learning_rate": 1.1436170212765959e-05, | |
| "loss": 0.1401, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 12.57, | |
| "learning_rate": 1.1427304964539008e-05, | |
| "loss": 0.0838, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 12.6, | |
| "learning_rate": 1.1418439716312057e-05, | |
| "loss": 0.1064, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 12.62, | |
| "learning_rate": 1.1409574468085107e-05, | |
| "loss": 0.2386, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 12.64, | |
| "learning_rate": 1.1400709219858158e-05, | |
| "loss": 0.1132, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 12.66, | |
| "learning_rate": 1.1391843971631207e-05, | |
| "loss": 0.1151, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 12.68, | |
| "learning_rate": 1.1382978723404256e-05, | |
| "loss": 0.1186, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 12.7, | |
| "learning_rate": 1.1374113475177305e-05, | |
| "loss": 0.1311, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 12.72, | |
| "learning_rate": 1.1365248226950356e-05, | |
| "loss": 0.0708, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 12.74, | |
| "learning_rate": 1.1356382978723405e-05, | |
| "loss": 0.176, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 12.77, | |
| "learning_rate": 1.1347517730496454e-05, | |
| "loss": 0.2388, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 12.79, | |
| "learning_rate": 1.1338652482269504e-05, | |
| "loss": 0.1255, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 12.81, | |
| "learning_rate": 1.1329787234042555e-05, | |
| "loss": 0.2945, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 12.83, | |
| "learning_rate": 1.1320921985815604e-05, | |
| "loss": 0.2043, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 12.85, | |
| "learning_rate": 1.1312056737588653e-05, | |
| "loss": 0.178, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 12.87, | |
| "learning_rate": 1.1303191489361702e-05, | |
| "loss": 0.1145, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 12.89, | |
| "learning_rate": 1.1294326241134753e-05, | |
| "loss": 0.1003, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 12.91, | |
| "learning_rate": 1.1285460992907802e-05, | |
| "loss": 0.2194, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 12.94, | |
| "learning_rate": 1.1276595744680851e-05, | |
| "loss": 0.2483, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 12.96, | |
| "learning_rate": 1.12677304964539e-05, | |
| "loss": 0.2026, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 12.98, | |
| "learning_rate": 1.1258865248226952e-05, | |
| "loss": 0.1588, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "learning_rate": 1.125e-05, | |
| "loss": 0.0764, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "eval_accuracy": 0.9466666666666667, | |
| "eval_loss": 0.2478657066822052, | |
| "eval_runtime": 27.3223, | |
| "eval_samples_per_second": 5.49, | |
| "eval_steps_per_second": 0.366, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 13.02, | |
| "learning_rate": 1.124113475177305e-05, | |
| "loss": 0.1586, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 13.04, | |
| "learning_rate": 1.12322695035461e-05, | |
| "loss": 0.1786, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 13.06, | |
| "learning_rate": 1.122340425531915e-05, | |
| "loss": 0.1282, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 13.09, | |
| "learning_rate": 1.12145390070922e-05, | |
| "loss": 0.1282, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 13.11, | |
| "learning_rate": 1.1205673758865248e-05, | |
| "loss": 0.0984, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 13.13, | |
| "learning_rate": 1.11968085106383e-05, | |
| "loss": 0.1625, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 13.15, | |
| "learning_rate": 1.1187943262411349e-05, | |
| "loss": 0.1567, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 13.17, | |
| "learning_rate": 1.1179078014184398e-05, | |
| "loss": 0.1898, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 13.19, | |
| "learning_rate": 1.1170212765957447e-05, | |
| "loss": 0.2715, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 13.21, | |
| "learning_rate": 1.1161347517730498e-05, | |
| "loss": 0.1214, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 13.23, | |
| "learning_rate": 1.1152482269503547e-05, | |
| "loss": 0.2001, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 13.26, | |
| "learning_rate": 1.1143617021276596e-05, | |
| "loss": 0.3139, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 13.28, | |
| "learning_rate": 1.1134751773049645e-05, | |
| "loss": 0.1541, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 13.3, | |
| "learning_rate": 1.1125886524822696e-05, | |
| "loss": 0.2244, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 13.32, | |
| "learning_rate": 1.1117021276595746e-05, | |
| "loss": 0.1197, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 13.34, | |
| "learning_rate": 1.1108156028368795e-05, | |
| "loss": 0.1103, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 13.36, | |
| "learning_rate": 1.1099290780141844e-05, | |
| "loss": 0.1947, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 13.38, | |
| "learning_rate": 1.1090425531914895e-05, | |
| "loss": 0.1468, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 13.4, | |
| "learning_rate": 1.1081560283687944e-05, | |
| "loss": 0.1434, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 13.43, | |
| "learning_rate": 1.1072695035460993e-05, | |
| "loss": 0.2457, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 13.45, | |
| "learning_rate": 1.1063829787234042e-05, | |
| "loss": 0.1257, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 13.47, | |
| "learning_rate": 1.1054964539007093e-05, | |
| "loss": 0.1687, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 13.49, | |
| "learning_rate": 1.1046099290780143e-05, | |
| "loss": 0.0841, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 13.51, | |
| "learning_rate": 1.1037234042553192e-05, | |
| "loss": 0.1573, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 13.53, | |
| "learning_rate": 1.1028368794326241e-05, | |
| "loss": 0.0723, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 13.55, | |
| "learning_rate": 1.1019503546099292e-05, | |
| "loss": 0.253, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 13.57, | |
| "learning_rate": 1.1010638297872341e-05, | |
| "loss": 0.1882, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 13.6, | |
| "learning_rate": 1.100177304964539e-05, | |
| "loss": 0.1305, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 13.62, | |
| "learning_rate": 1.099290780141844e-05, | |
| "loss": 0.0719, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 13.64, | |
| "learning_rate": 1.098404255319149e-05, | |
| "loss": 0.2096, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 13.66, | |
| "learning_rate": 1.097517730496454e-05, | |
| "loss": 0.0952, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 13.68, | |
| "learning_rate": 1.0966312056737589e-05, | |
| "loss": 0.1869, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 13.7, | |
| "learning_rate": 1.095744680851064e-05, | |
| "loss": 0.1948, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 13.72, | |
| "learning_rate": 1.0948581560283689e-05, | |
| "loss": 0.1136, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 13.74, | |
| "learning_rate": 1.0939716312056738e-05, | |
| "loss": 0.2405, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 13.77, | |
| "learning_rate": 1.0930851063829787e-05, | |
| "loss": 0.066, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 13.79, | |
| "learning_rate": 1.0921985815602838e-05, | |
| "loss": 0.0909, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 13.81, | |
| "learning_rate": 1.0913120567375887e-05, | |
| "loss": 0.1452, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 13.83, | |
| "learning_rate": 1.0904255319148937e-05, | |
| "loss": 0.075, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 13.85, | |
| "learning_rate": 1.0895390070921986e-05, | |
| "loss": 0.0925, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 13.87, | |
| "learning_rate": 1.0886524822695037e-05, | |
| "loss": 0.0521, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 13.89, | |
| "learning_rate": 1.0877659574468086e-05, | |
| "loss": 0.1644, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 13.91, | |
| "learning_rate": 1.0868794326241135e-05, | |
| "loss": 0.1039, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 13.94, | |
| "learning_rate": 1.0859929078014184e-05, | |
| "loss": 0.0912, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 13.96, | |
| "learning_rate": 1.0851063829787235e-05, | |
| "loss": 0.116, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 13.98, | |
| "learning_rate": 1.0842198581560284e-05, | |
| "loss": 0.2023, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "learning_rate": 1.0833333333333334e-05, | |
| "loss": 0.1472, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "eval_accuracy": 0.9333333333333333, | |
| "eval_loss": 0.2599130868911743, | |
| "eval_runtime": 29.5623, | |
| "eval_samples_per_second": 5.074, | |
| "eval_steps_per_second": 0.338, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 14.02, | |
| "learning_rate": 1.0824468085106383e-05, | |
| "loss": 0.1322, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 14.04, | |
| "learning_rate": 1.0815602836879434e-05, | |
| "loss": 0.125, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 14.06, | |
| "learning_rate": 1.0806737588652483e-05, | |
| "loss": 0.1188, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 14.09, | |
| "learning_rate": 1.0797872340425532e-05, | |
| "loss": 0.118, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 14.11, | |
| "learning_rate": 1.0789007092198581e-05, | |
| "loss": 0.1424, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 14.13, | |
| "learning_rate": 1.0780141843971632e-05, | |
| "loss": 0.1038, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 14.15, | |
| "learning_rate": 1.0771276595744681e-05, | |
| "loss": 0.2191, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 14.17, | |
| "learning_rate": 1.076241134751773e-05, | |
| "loss": 0.1955, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 14.19, | |
| "learning_rate": 1.075354609929078e-05, | |
| "loss": 0.2242, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 14.21, | |
| "learning_rate": 1.074468085106383e-05, | |
| "loss": 0.139, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 14.23, | |
| "learning_rate": 1.073581560283688e-05, | |
| "loss": 0.2557, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 14.26, | |
| "learning_rate": 1.0726950354609929e-05, | |
| "loss": 0.2158, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 14.28, | |
| "learning_rate": 1.0718085106382978e-05, | |
| "loss": 0.2972, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 14.3, | |
| "learning_rate": 1.0709219858156029e-05, | |
| "loss": 0.2584, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 14.32, | |
| "learning_rate": 1.0700354609929078e-05, | |
| "loss": 0.0944, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 14.34, | |
| "learning_rate": 1.0691489361702128e-05, | |
| "loss": 0.18, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 14.36, | |
| "learning_rate": 1.0682624113475177e-05, | |
| "loss": 0.0587, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 14.38, | |
| "learning_rate": 1.0673758865248228e-05, | |
| "loss": 0.0762, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 14.4, | |
| "learning_rate": 1.0664893617021277e-05, | |
| "loss": 0.0868, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 14.43, | |
| "learning_rate": 1.0656028368794326e-05, | |
| "loss": 0.1355, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 14.45, | |
| "learning_rate": 1.0647163120567377e-05, | |
| "loss": 0.0821, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 14.47, | |
| "learning_rate": 1.0638297872340426e-05, | |
| "loss": 0.1032, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 14.49, | |
| "learning_rate": 1.0629432624113475e-05, | |
| "loss": 0.1368, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 14.51, | |
| "learning_rate": 1.0620567375886525e-05, | |
| "loss": 0.092, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 14.53, | |
| "learning_rate": 1.0611702127659575e-05, | |
| "loss": 0.1153, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 14.55, | |
| "learning_rate": 1.0602836879432625e-05, | |
| "loss": 0.215, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 14.57, | |
| "learning_rate": 1.0593971631205674e-05, | |
| "loss": 0.1163, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 14.6, | |
| "learning_rate": 1.0585106382978723e-05, | |
| "loss": 0.1171, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 14.62, | |
| "learning_rate": 1.0576241134751774e-05, | |
| "loss": 0.0918, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 14.64, | |
| "learning_rate": 1.0567375886524823e-05, | |
| "loss": 0.3054, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 14.66, | |
| "learning_rate": 1.0558510638297872e-05, | |
| "loss": 0.1806, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 14.68, | |
| "learning_rate": 1.0549645390070922e-05, | |
| "loss": 0.0383, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 14.7, | |
| "learning_rate": 1.0540780141843972e-05, | |
| "loss": 0.1518, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 14.72, | |
| "learning_rate": 1.0531914893617022e-05, | |
| "loss": 0.1575, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 14.74, | |
| "learning_rate": 1.052304964539007e-05, | |
| "loss": 0.0929, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 14.77, | |
| "learning_rate": 1.051418439716312e-05, | |
| "loss": 0.0418, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 14.79, | |
| "learning_rate": 1.0505319148936171e-05, | |
| "loss": 0.0758, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 14.81, | |
| "learning_rate": 1.049645390070922e-05, | |
| "loss": 0.041, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 14.83, | |
| "learning_rate": 1.048758865248227e-05, | |
| "loss": 0.128, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 14.85, | |
| "learning_rate": 1.0478723404255319e-05, | |
| "loss": 0.2844, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 14.87, | |
| "learning_rate": 1.046985815602837e-05, | |
| "loss": 0.1974, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 14.89, | |
| "learning_rate": 1.0460992907801419e-05, | |
| "loss": 0.1484, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 14.91, | |
| "learning_rate": 1.0452127659574468e-05, | |
| "loss": 0.1317, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 14.94, | |
| "learning_rate": 1.0443262411347517e-05, | |
| "loss": 0.1426, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 14.96, | |
| "learning_rate": 1.0434397163120568e-05, | |
| "loss": 0.1753, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 14.98, | |
| "learning_rate": 1.0425531914893617e-05, | |
| "loss": 0.1863, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "learning_rate": 1.0416666666666666e-05, | |
| "loss": 0.0483, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "eval_accuracy": 0.9533333333333334, | |
| "eval_loss": 0.1910819709300995, | |
| "eval_runtime": 25.3686, | |
| "eval_samples_per_second": 5.913, | |
| "eval_steps_per_second": 0.394, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 15.02, | |
| "learning_rate": 1.0407801418439716e-05, | |
| "loss": 0.2788, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 15.04, | |
| "learning_rate": 1.0398936170212766e-05, | |
| "loss": 0.1394, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 15.06, | |
| "learning_rate": 1.0390070921985816e-05, | |
| "loss": 0.1538, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 15.09, | |
| "learning_rate": 1.0381205673758865e-05, | |
| "loss": 0.0613, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 15.11, | |
| "learning_rate": 1.0372340425531916e-05, | |
| "loss": 0.131, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 15.13, | |
| "learning_rate": 1.0363475177304965e-05, | |
| "loss": 0.0771, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 15.15, | |
| "learning_rate": 1.0354609929078014e-05, | |
| "loss": 0.246, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 15.17, | |
| "learning_rate": 1.0345744680851063e-05, | |
| "loss": 0.0305, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 15.19, | |
| "learning_rate": 1.0336879432624114e-05, | |
| "loss": 0.1816, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 15.21, | |
| "learning_rate": 1.0328014184397163e-05, | |
| "loss": 0.0717, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 15.23, | |
| "learning_rate": 1.0319148936170213e-05, | |
| "loss": 0.1006, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 15.26, | |
| "learning_rate": 1.0310283687943262e-05, | |
| "loss": 0.2086, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 15.28, | |
| "learning_rate": 1.0301418439716313e-05, | |
| "loss": 0.0609, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 15.3, | |
| "learning_rate": 1.0292553191489362e-05, | |
| "loss": 0.2033, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 15.32, | |
| "learning_rate": 1.0283687943262411e-05, | |
| "loss": 0.0658, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 15.34, | |
| "learning_rate": 1.027482269503546e-05, | |
| "loss": 0.2492, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 15.36, | |
| "learning_rate": 1.0265957446808511e-05, | |
| "loss": 0.1693, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 15.38, | |
| "learning_rate": 1.025709219858156e-05, | |
| "loss": 0.1319, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 15.4, | |
| "learning_rate": 1.024822695035461e-05, | |
| "loss": 0.2362, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 15.43, | |
| "learning_rate": 1.0239361702127659e-05, | |
| "loss": 0.1193, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 15.45, | |
| "learning_rate": 1.023049645390071e-05, | |
| "loss": 0.2145, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 15.47, | |
| "learning_rate": 1.0221631205673759e-05, | |
| "loss": 0.1524, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 15.49, | |
| "learning_rate": 1.0212765957446808e-05, | |
| "loss": 0.105, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 15.51, | |
| "learning_rate": 1.0203900709219857e-05, | |
| "loss": 0.0452, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 15.53, | |
| "learning_rate": 1.0195035460992908e-05, | |
| "loss": 0.0767, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 15.55, | |
| "learning_rate": 1.0186170212765957e-05, | |
| "loss": 0.0617, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 15.57, | |
| "learning_rate": 1.0177304964539007e-05, | |
| "loss": 0.1381, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 15.6, | |
| "learning_rate": 1.0168439716312056e-05, | |
| "loss": 0.218, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 15.62, | |
| "learning_rate": 1.0159574468085107e-05, | |
| "loss": 0.2145, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 15.64, | |
| "learning_rate": 1.0150709219858156e-05, | |
| "loss": 0.1117, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 15.66, | |
| "learning_rate": 1.0141843971631205e-05, | |
| "loss": 0.0753, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 15.68, | |
| "learning_rate": 1.0132978723404254e-05, | |
| "loss": 0.1559, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 15.7, | |
| "learning_rate": 1.0124113475177305e-05, | |
| "loss": 0.0548, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 15.72, | |
| "learning_rate": 1.0115248226950354e-05, | |
| "loss": 0.1088, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 15.74, | |
| "learning_rate": 1.0106382978723404e-05, | |
| "loss": 0.0457, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 15.77, | |
| "learning_rate": 1.0097517730496453e-05, | |
| "loss": 0.1416, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 15.79, | |
| "learning_rate": 1.0088652482269504e-05, | |
| "loss": 0.1492, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 15.81, | |
| "learning_rate": 1.0079787234042553e-05, | |
| "loss": 0.0635, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 15.83, | |
| "learning_rate": 1.0070921985815602e-05, | |
| "loss": 0.1284, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 15.85, | |
| "learning_rate": 1.0062056737588653e-05, | |
| "loss": 0.1256, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 15.87, | |
| "learning_rate": 1.0053191489361702e-05, | |
| "loss": 0.0626, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 15.89, | |
| "learning_rate": 1.0044326241134751e-05, | |
| "loss": 0.2339, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 15.91, | |
| "learning_rate": 1.00354609929078e-05, | |
| "loss": 0.1178, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 15.94, | |
| "learning_rate": 1.0026595744680852e-05, | |
| "loss": 0.0929, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 15.96, | |
| "learning_rate": 1.00177304964539e-05, | |
| "loss": 0.1118, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 15.98, | |
| "learning_rate": 1.000886524822695e-05, | |
| "loss": 0.1599, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "learning_rate": 9.999999999999999e-06, | |
| "loss": 0.0961, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "eval_accuracy": 0.9533333333333334, | |
| "eval_loss": 0.18685631453990936, | |
| "eval_runtime": 28.6184, | |
| "eval_samples_per_second": 5.241, | |
| "eval_steps_per_second": 0.349, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 16.02, | |
| "learning_rate": 9.99113475177305e-06, | |
| "loss": 0.191, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 16.04, | |
| "learning_rate": 9.9822695035461e-06, | |
| "loss": 0.1426, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 16.06, | |
| "learning_rate": 9.973404255319148e-06, | |
| "loss": 0.1199, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 16.09, | |
| "learning_rate": 9.964539007092198e-06, | |
| "loss": 0.2577, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 16.11, | |
| "learning_rate": 9.955673758865249e-06, | |
| "loss": 0.0697, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 16.13, | |
| "learning_rate": 9.946808510638298e-06, | |
| "loss": 0.1362, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 16.15, | |
| "learning_rate": 9.937943262411347e-06, | |
| "loss": 0.0386, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 16.17, | |
| "learning_rate": 9.929078014184396e-06, | |
| "loss": 0.1211, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 16.19, | |
| "learning_rate": 9.920212765957447e-06, | |
| "loss": 0.1575, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 16.21, | |
| "learning_rate": 9.911347517730496e-06, | |
| "loss": 0.1798, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 16.23, | |
| "learning_rate": 9.902482269503545e-06, | |
| "loss": 0.1533, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 16.26, | |
| "learning_rate": 9.893617021276595e-06, | |
| "loss": 0.0647, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 16.28, | |
| "learning_rate": 9.884751773049646e-06, | |
| "loss": 0.1768, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 16.3, | |
| "learning_rate": 9.875886524822695e-06, | |
| "loss": 0.1946, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 16.32, | |
| "learning_rate": 9.867021276595744e-06, | |
| "loss": 0.0584, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 16.34, | |
| "learning_rate": 9.858156028368793e-06, | |
| "loss": 0.1867, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 16.36, | |
| "learning_rate": 9.849290780141844e-06, | |
| "loss": 0.1036, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 16.38, | |
| "learning_rate": 9.840425531914893e-06, | |
| "loss": 0.1536, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 16.4, | |
| "learning_rate": 9.831560283687942e-06, | |
| "loss": 0.1061, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 16.43, | |
| "learning_rate": 9.822695035460992e-06, | |
| "loss": 0.0839, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 16.45, | |
| "learning_rate": 9.813829787234043e-06, | |
| "loss": 0.2435, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 16.47, | |
| "learning_rate": 9.804964539007092e-06, | |
| "loss": 0.136, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 16.49, | |
| "learning_rate": 9.796099290780141e-06, | |
| "loss": 0.231, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 16.51, | |
| "learning_rate": 9.787234042553192e-06, | |
| "loss": 0.0978, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 16.53, | |
| "learning_rate": 9.778368794326241e-06, | |
| "loss": 0.1245, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 16.55, | |
| "learning_rate": 9.76950354609929e-06, | |
| "loss": 0.1219, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 16.57, | |
| "learning_rate": 9.76063829787234e-06, | |
| "loss": 0.1246, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 16.6, | |
| "learning_rate": 9.75177304964539e-06, | |
| "loss": 0.0997, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 16.62, | |
| "learning_rate": 9.74290780141844e-06, | |
| "loss": 0.1179, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 16.64, | |
| "learning_rate": 9.734042553191489e-06, | |
| "loss": 0.0973, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 16.66, | |
| "learning_rate": 9.725177304964538e-06, | |
| "loss": 0.0716, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 16.68, | |
| "learning_rate": 9.716312056737589e-06, | |
| "loss": 0.141, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 16.7, | |
| "learning_rate": 9.707446808510638e-06, | |
| "loss": 0.2365, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 16.72, | |
| "learning_rate": 9.698581560283687e-06, | |
| "loss": 0.1254, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 16.74, | |
| "learning_rate": 9.689716312056736e-06, | |
| "loss": 0.1511, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 16.77, | |
| "learning_rate": 9.680851063829787e-06, | |
| "loss": 0.1153, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 16.79, | |
| "learning_rate": 9.671985815602837e-06, | |
| "loss": 0.1348, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 16.81, | |
| "learning_rate": 9.663120567375886e-06, | |
| "loss": 0.1255, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 16.83, | |
| "learning_rate": 9.654255319148937e-06, | |
| "loss": 0.0599, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 16.85, | |
| "learning_rate": 9.645390070921986e-06, | |
| "loss": 0.2116, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 16.87, | |
| "learning_rate": 9.636524822695037e-06, | |
| "loss": 0.1076, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 16.89, | |
| "learning_rate": 9.627659574468086e-06, | |
| "loss": 0.0725, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 16.91, | |
| "learning_rate": 9.618794326241135e-06, | |
| "loss": 0.1247, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 16.94, | |
| "learning_rate": 9.609929078014186e-06, | |
| "loss": 0.1186, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 16.96, | |
| "learning_rate": 9.601063829787235e-06, | |
| "loss": 0.0888, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 16.98, | |
| "learning_rate": 9.592198581560284e-06, | |
| "loss": 0.1309, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "learning_rate": 9.583333333333334e-06, | |
| "loss": 0.1146, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "eval_accuracy": 0.9333333333333333, | |
| "eval_loss": 0.23552052676677704, | |
| "eval_runtime": 29.0233, | |
| "eval_samples_per_second": 5.168, | |
| "eval_steps_per_second": 0.345, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 17.02, | |
| "learning_rate": 9.574468085106385e-06, | |
| "loss": 0.2751, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 17.04, | |
| "learning_rate": 9.565602836879434e-06, | |
| "loss": 0.0695, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 17.06, | |
| "learning_rate": 9.556737588652483e-06, | |
| "loss": 0.0593, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 17.09, | |
| "learning_rate": 9.547872340425532e-06, | |
| "loss": 0.1915, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 17.11, | |
| "learning_rate": 9.539007092198583e-06, | |
| "loss": 0.2488, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 17.13, | |
| "learning_rate": 9.530141843971632e-06, | |
| "loss": 0.0657, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 17.15, | |
| "learning_rate": 9.521276595744681e-06, | |
| "loss": 0.0752, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 17.17, | |
| "learning_rate": 9.512411347517732e-06, | |
| "loss": 0.0984, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 17.19, | |
| "learning_rate": 9.503546099290782e-06, | |
| "loss": 0.1647, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 17.21, | |
| "learning_rate": 9.49468085106383e-06, | |
| "loss": 0.0986, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 17.23, | |
| "learning_rate": 9.48581560283688e-06, | |
| "loss": 0.0473, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 17.26, | |
| "learning_rate": 9.47695035460993e-06, | |
| "loss": 0.1171, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 17.28, | |
| "learning_rate": 9.46808510638298e-06, | |
| "loss": 0.1064, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 17.3, | |
| "learning_rate": 9.45921985815603e-06, | |
| "loss": 0.1417, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 17.32, | |
| "learning_rate": 9.450354609929078e-06, | |
| "loss": 0.1133, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 17.34, | |
| "learning_rate": 9.44148936170213e-06, | |
| "loss": 0.182, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 17.36, | |
| "learning_rate": 9.432624113475179e-06, | |
| "loss": 0.2498, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 17.38, | |
| "learning_rate": 9.423758865248228e-06, | |
| "loss": 0.0781, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 17.4, | |
| "learning_rate": 9.414893617021277e-06, | |
| "loss": 0.1122, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 17.43, | |
| "learning_rate": 9.406028368794328e-06, | |
| "loss": 0.1203, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 17.45, | |
| "learning_rate": 9.397163120567377e-06, | |
| "loss": 0.1266, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 17.47, | |
| "learning_rate": 9.388297872340426e-06, | |
| "loss": 0.0705, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 17.49, | |
| "learning_rate": 9.379432624113475e-06, | |
| "loss": 0.1507, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 17.51, | |
| "learning_rate": 9.370567375886526e-06, | |
| "loss": 0.1685, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 17.53, | |
| "learning_rate": 9.361702127659576e-06, | |
| "loss": 0.0584, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 17.55, | |
| "learning_rate": 9.352836879432625e-06, | |
| "loss": 0.0447, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 17.57, | |
| "learning_rate": 9.343971631205674e-06, | |
| "loss": 0.0902, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 17.6, | |
| "learning_rate": 9.335106382978725e-06, | |
| "loss": 0.2375, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 17.62, | |
| "learning_rate": 9.326241134751774e-06, | |
| "loss": 0.0756, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 17.64, | |
| "learning_rate": 9.317375886524823e-06, | |
| "loss": 0.0775, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 17.66, | |
| "learning_rate": 9.308510638297872e-06, | |
| "loss": 0.236, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 17.68, | |
| "learning_rate": 9.299645390070923e-06, | |
| "loss": 0.0382, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 17.7, | |
| "learning_rate": 9.290780141843973e-06, | |
| "loss": 0.1955, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 17.72, | |
| "learning_rate": 9.281914893617022e-06, | |
| "loss": 0.1331, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 17.74, | |
| "learning_rate": 9.273049645390071e-06, | |
| "loss": 0.1168, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 17.77, | |
| "learning_rate": 9.264184397163122e-06, | |
| "loss": 0.2426, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 17.79, | |
| "learning_rate": 9.255319148936171e-06, | |
| "loss": 0.1497, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 17.81, | |
| "learning_rate": 9.24645390070922e-06, | |
| "loss": 0.2129, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 17.83, | |
| "learning_rate": 9.23758865248227e-06, | |
| "loss": 0.1372, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 17.85, | |
| "learning_rate": 9.22872340425532e-06, | |
| "loss": 0.1677, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 17.87, | |
| "learning_rate": 9.21985815602837e-06, | |
| "loss": 0.1427, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 17.89, | |
| "learning_rate": 9.210992907801419e-06, | |
| "loss": 0.0805, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 17.91, | |
| "learning_rate": 9.20212765957447e-06, | |
| "loss": 0.276, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 17.94, | |
| "learning_rate": 9.193262411347519e-06, | |
| "loss": 0.0914, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 17.96, | |
| "learning_rate": 9.184397163120568e-06, | |
| "loss": 0.1804, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 17.98, | |
| "learning_rate": 9.175531914893617e-06, | |
| "loss": 0.0819, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "learning_rate": 9.166666666666668e-06, | |
| "loss": 0.2117, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "eval_accuracy": 0.94, | |
| "eval_loss": 0.19297966361045837, | |
| "eval_runtime": 27.2448, | |
| "eval_samples_per_second": 5.506, | |
| "eval_steps_per_second": 0.367, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 18.02, | |
| "learning_rate": 9.157801418439717e-06, | |
| "loss": 0.1613, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 18.04, | |
| "learning_rate": 9.148936170212767e-06, | |
| "loss": 0.0973, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 18.06, | |
| "learning_rate": 9.140070921985816e-06, | |
| "loss": 0.1676, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 18.09, | |
| "learning_rate": 9.131205673758867e-06, | |
| "loss": 0.0888, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 18.11, | |
| "learning_rate": 9.122340425531916e-06, | |
| "loss": 0.2338, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 18.13, | |
| "learning_rate": 9.113475177304965e-06, | |
| "loss": 0.2096, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 18.15, | |
| "learning_rate": 9.104609929078014e-06, | |
| "loss": 0.1234, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 18.17, | |
| "learning_rate": 9.095744680851065e-06, | |
| "loss": 0.1363, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 18.19, | |
| "learning_rate": 9.086879432624114e-06, | |
| "loss": 0.0499, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 18.21, | |
| "learning_rate": 9.078014184397164e-06, | |
| "loss": 0.152, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 18.23, | |
| "learning_rate": 9.069148936170213e-06, | |
| "loss": 0.1698, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 18.26, | |
| "learning_rate": 9.060283687943264e-06, | |
| "loss": 0.04, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 18.28, | |
| "learning_rate": 9.051418439716313e-06, | |
| "loss": 0.2551, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 18.3, | |
| "learning_rate": 9.042553191489362e-06, | |
| "loss": 0.1289, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 18.32, | |
| "learning_rate": 9.033687943262411e-06, | |
| "loss": 0.0864, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 18.34, | |
| "learning_rate": 9.024822695035462e-06, | |
| "loss": 0.1777, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 18.36, | |
| "learning_rate": 9.015957446808511e-06, | |
| "loss": 0.0665, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 18.38, | |
| "learning_rate": 9.00709219858156e-06, | |
| "loss": 0.0968, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 18.4, | |
| "learning_rate": 8.99822695035461e-06, | |
| "loss": 0.1676, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 18.43, | |
| "learning_rate": 8.98936170212766e-06, | |
| "loss": 0.1994, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 18.45, | |
| "learning_rate": 8.98049645390071e-06, | |
| "loss": 0.1528, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 18.47, | |
| "learning_rate": 8.971631205673759e-06, | |
| "loss": 0.1279, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 18.49, | |
| "learning_rate": 8.962765957446808e-06, | |
| "loss": 0.198, | |
| "step": 869 | |
| }, | |
| { | |
| "epoch": 18.51, | |
| "learning_rate": 8.953900709219859e-06, | |
| "loss": 0.148, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 18.53, | |
| "learning_rate": 8.945035460992908e-06, | |
| "loss": 0.0873, | |
| "step": 871 | |
| }, | |
| { | |
| "epoch": 18.55, | |
| "learning_rate": 8.936170212765958e-06, | |
| "loss": 0.0946, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 18.57, | |
| "learning_rate": 8.927304964539008e-06, | |
| "loss": 0.1208, | |
| "step": 873 | |
| }, | |
| { | |
| "epoch": 18.6, | |
| "learning_rate": 8.918439716312058e-06, | |
| "loss": 0.1184, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 18.62, | |
| "learning_rate": 8.909574468085107e-06, | |
| "loss": 0.3374, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 18.64, | |
| "learning_rate": 8.900709219858156e-06, | |
| "loss": 0.1455, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 18.66, | |
| "learning_rate": 8.891843971631207e-06, | |
| "loss": 0.1696, | |
| "step": 877 | |
| }, | |
| { | |
| "epoch": 18.68, | |
| "learning_rate": 8.882978723404256e-06, | |
| "loss": 0.1199, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 18.7, | |
| "learning_rate": 8.874113475177305e-06, | |
| "loss": 0.2619, | |
| "step": 879 | |
| }, | |
| { | |
| "epoch": 18.72, | |
| "learning_rate": 8.865248226950355e-06, | |
| "loss": 0.1963, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 18.74, | |
| "learning_rate": 8.856382978723405e-06, | |
| "loss": 0.1393, | |
| "step": 881 | |
| }, | |
| { | |
| "epoch": 18.77, | |
| "learning_rate": 8.847517730496455e-06, | |
| "loss": 0.0775, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 18.79, | |
| "learning_rate": 8.838652482269504e-06, | |
| "loss": 0.0844, | |
| "step": 883 | |
| }, | |
| { | |
| "epoch": 18.81, | |
| "learning_rate": 8.829787234042553e-06, | |
| "loss": 0.1409, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 18.83, | |
| "learning_rate": 8.820921985815604e-06, | |
| "loss": 0.131, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 18.85, | |
| "learning_rate": 8.812056737588653e-06, | |
| "loss": 0.077, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 18.87, | |
| "learning_rate": 8.803191489361702e-06, | |
| "loss": 0.1306, | |
| "step": 887 | |
| }, | |
| { | |
| "epoch": 18.89, | |
| "learning_rate": 8.794326241134752e-06, | |
| "loss": 0.1103, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 18.91, | |
| "learning_rate": 8.785460992907802e-06, | |
| "loss": 0.1299, | |
| "step": 889 | |
| }, | |
| { | |
| "epoch": 18.94, | |
| "learning_rate": 8.776595744680852e-06, | |
| "loss": 0.1182, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 18.96, | |
| "learning_rate": 8.7677304964539e-06, | |
| "loss": 0.1094, | |
| "step": 891 | |
| }, | |
| { | |
| "epoch": 18.98, | |
| "learning_rate": 8.75886524822695e-06, | |
| "loss": 0.0879, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "learning_rate": 8.750000000000001e-06, | |
| "loss": 0.2859, | |
| "step": 893 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "eval_accuracy": 0.9466666666666667, | |
| "eval_loss": 0.19023513793945312, | |
| "eval_runtime": 26.6196, | |
| "eval_samples_per_second": 5.635, | |
| "eval_steps_per_second": 0.376, | |
| "step": 893 | |
| }, | |
| { | |
| "epoch": 19.02, | |
| "learning_rate": 8.74113475177305e-06, | |
| "loss": 0.1001, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 19.04, | |
| "learning_rate": 8.7322695035461e-06, | |
| "loss": 0.2251, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 19.06, | |
| "learning_rate": 8.723404255319149e-06, | |
| "loss": 0.1941, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 19.09, | |
| "learning_rate": 8.7145390070922e-06, | |
| "loss": 0.1163, | |
| "step": 897 | |
| }, | |
| { | |
| "epoch": 19.11, | |
| "learning_rate": 8.705673758865249e-06, | |
| "loss": 0.1035, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 19.13, | |
| "learning_rate": 8.696808510638298e-06, | |
| "loss": 0.1709, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 19.15, | |
| "learning_rate": 8.687943262411347e-06, | |
| "loss": 0.0757, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 19.17, | |
| "learning_rate": 8.679078014184398e-06, | |
| "loss": 0.1465, | |
| "step": 901 | |
| }, | |
| { | |
| "epoch": 19.19, | |
| "learning_rate": 8.670212765957447e-06, | |
| "loss": 0.1137, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 19.21, | |
| "learning_rate": 8.661347517730496e-06, | |
| "loss": 0.0592, | |
| "step": 903 | |
| }, | |
| { | |
| "epoch": 19.23, | |
| "learning_rate": 8.652482269503547e-06, | |
| "loss": 0.0811, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 19.26, | |
| "learning_rate": 8.643617021276596e-06, | |
| "loss": 0.1585, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 19.28, | |
| "learning_rate": 8.634751773049646e-06, | |
| "loss": 0.0809, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 19.3, | |
| "learning_rate": 8.625886524822695e-06, | |
| "loss": 0.1348, | |
| "step": 907 | |
| }, | |
| { | |
| "epoch": 19.32, | |
| "learning_rate": 8.617021276595746e-06, | |
| "loss": 0.1253, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 19.34, | |
| "learning_rate": 8.608156028368795e-06, | |
| "loss": 0.0707, | |
| "step": 909 | |
| }, | |
| { | |
| "epoch": 19.36, | |
| "learning_rate": 8.599290780141844e-06, | |
| "loss": 0.1951, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 19.38, | |
| "learning_rate": 8.590425531914893e-06, | |
| "loss": 0.1861, | |
| "step": 911 | |
| }, | |
| { | |
| "epoch": 19.4, | |
| "learning_rate": 8.581560283687944e-06, | |
| "loss": 0.2975, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 19.43, | |
| "learning_rate": 8.572695035460993e-06, | |
| "loss": 0.1971, | |
| "step": 913 | |
| }, | |
| { | |
| "epoch": 19.45, | |
| "learning_rate": 8.563829787234043e-06, | |
| "loss": 0.1056, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 19.47, | |
| "learning_rate": 8.554964539007092e-06, | |
| "loss": 0.0632, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 19.49, | |
| "learning_rate": 8.546099290780143e-06, | |
| "loss": 0.0823, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 19.51, | |
| "learning_rate": 8.537234042553192e-06, | |
| "loss": 0.1916, | |
| "step": 917 | |
| }, | |
| { | |
| "epoch": 19.53, | |
| "learning_rate": 8.528368794326241e-06, | |
| "loss": 0.0352, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 19.55, | |
| "learning_rate": 8.51950354609929e-06, | |
| "loss": 0.0729, | |
| "step": 919 | |
| }, | |
| { | |
| "epoch": 19.57, | |
| "learning_rate": 8.510638297872341e-06, | |
| "loss": 0.0513, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 19.6, | |
| "learning_rate": 8.50177304964539e-06, | |
| "loss": 0.1326, | |
| "step": 921 | |
| }, | |
| { | |
| "epoch": 19.62, | |
| "learning_rate": 8.49290780141844e-06, | |
| "loss": 0.0672, | |
| "step": 922 | |
| }, | |
| { | |
| "epoch": 19.64, | |
| "learning_rate": 8.484042553191489e-06, | |
| "loss": 0.0707, | |
| "step": 923 | |
| }, | |
| { | |
| "epoch": 19.66, | |
| "learning_rate": 8.47517730496454e-06, | |
| "loss": 0.1507, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 19.68, | |
| "learning_rate": 8.466312056737589e-06, | |
| "loss": 0.1689, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 19.7, | |
| "learning_rate": 8.457446808510638e-06, | |
| "loss": 0.3677, | |
| "step": 926 | |
| }, | |
| { | |
| "epoch": 19.72, | |
| "learning_rate": 8.448581560283687e-06, | |
| "loss": 0.1072, | |
| "step": 927 | |
| }, | |
| { | |
| "epoch": 19.74, | |
| "learning_rate": 8.439716312056738e-06, | |
| "loss": 0.2191, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 19.77, | |
| "learning_rate": 8.430851063829787e-06, | |
| "loss": 0.2591, | |
| "step": 929 | |
| }, | |
| { | |
| "epoch": 19.79, | |
| "learning_rate": 8.421985815602837e-06, | |
| "loss": 0.1031, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 19.81, | |
| "learning_rate": 8.413120567375886e-06, | |
| "loss": 0.0508, | |
| "step": 931 | |
| }, | |
| { | |
| "epoch": 19.83, | |
| "learning_rate": 8.404255319148937e-06, | |
| "loss": 0.109, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 19.85, | |
| "learning_rate": 8.395390070921986e-06, | |
| "loss": 0.0551, | |
| "step": 933 | |
| }, | |
| { | |
| "epoch": 19.87, | |
| "learning_rate": 8.386524822695035e-06, | |
| "loss": 0.0721, | |
| "step": 934 | |
| }, | |
| { | |
| "epoch": 19.89, | |
| "learning_rate": 8.377659574468084e-06, | |
| "loss": 0.2265, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 19.91, | |
| "learning_rate": 8.368794326241135e-06, | |
| "loss": 0.1058, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 19.94, | |
| "learning_rate": 8.359929078014184e-06, | |
| "loss": 0.0766, | |
| "step": 937 | |
| }, | |
| { | |
| "epoch": 19.96, | |
| "learning_rate": 8.351063829787234e-06, | |
| "loss": 0.0913, | |
| "step": 938 | |
| }, | |
| { | |
| "epoch": 19.98, | |
| "learning_rate": 8.342198581560285e-06, | |
| "loss": 0.182, | |
| "step": 939 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "learning_rate": 8.333333333333334e-06, | |
| "loss": 0.0798, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "eval_accuracy": 0.9333333333333333, | |
| "eval_loss": 0.24364662170410156, | |
| "eval_runtime": 27.243, | |
| "eval_samples_per_second": 5.506, | |
| "eval_steps_per_second": 0.367, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 20.02, | |
| "learning_rate": 8.324468085106383e-06, | |
| "loss": 0.1908, | |
| "step": 941 | |
| }, | |
| { | |
| "epoch": 20.04, | |
| "learning_rate": 8.315602836879432e-06, | |
| "loss": 0.1042, | |
| "step": 942 | |
| }, | |
| { | |
| "epoch": 20.06, | |
| "learning_rate": 8.306737588652483e-06, | |
| "loss": 0.1064, | |
| "step": 943 | |
| }, | |
| { | |
| "epoch": 20.09, | |
| "learning_rate": 8.297872340425532e-06, | |
| "loss": 0.0455, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 20.11, | |
| "learning_rate": 8.289007092198581e-06, | |
| "loss": 0.0944, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 20.13, | |
| "learning_rate": 8.28014184397163e-06, | |
| "loss": 0.1079, | |
| "step": 946 | |
| }, | |
| { | |
| "epoch": 20.15, | |
| "learning_rate": 8.271276595744682e-06, | |
| "loss": 0.0871, | |
| "step": 947 | |
| }, | |
| { | |
| "epoch": 20.17, | |
| "learning_rate": 8.26241134751773e-06, | |
| "loss": 0.1214, | |
| "step": 948 | |
| }, | |
| { | |
| "epoch": 20.19, | |
| "learning_rate": 8.25354609929078e-06, | |
| "loss": 0.1297, | |
| "step": 949 | |
| }, | |
| { | |
| "epoch": 20.21, | |
| "learning_rate": 8.244680851063829e-06, | |
| "loss": 0.1767, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 20.23, | |
| "learning_rate": 8.23581560283688e-06, | |
| "loss": 0.1746, | |
| "step": 951 | |
| }, | |
| { | |
| "epoch": 20.26, | |
| "learning_rate": 8.22695035460993e-06, | |
| "loss": 0.0846, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 20.28, | |
| "learning_rate": 8.218085106382978e-06, | |
| "loss": 0.1596, | |
| "step": 953 | |
| }, | |
| { | |
| "epoch": 20.3, | |
| "learning_rate": 8.209219858156028e-06, | |
| "loss": 0.2268, | |
| "step": 954 | |
| }, | |
| { | |
| "epoch": 20.32, | |
| "learning_rate": 8.200354609929079e-06, | |
| "loss": 0.0493, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 20.34, | |
| "learning_rate": 8.191489361702128e-06, | |
| "loss": 0.1771, | |
| "step": 956 | |
| }, | |
| { | |
| "epoch": 20.36, | |
| "learning_rate": 8.182624113475177e-06, | |
| "loss": 0.0831, | |
| "step": 957 | |
| }, | |
| { | |
| "epoch": 20.38, | |
| "learning_rate": 8.173758865248226e-06, | |
| "loss": 0.0961, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 20.4, | |
| "learning_rate": 8.164893617021277e-06, | |
| "loss": 0.1269, | |
| "step": 959 | |
| }, | |
| { | |
| "epoch": 20.43, | |
| "learning_rate": 8.156028368794326e-06, | |
| "loss": 0.1635, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 20.45, | |
| "learning_rate": 8.147163120567375e-06, | |
| "loss": 0.0715, | |
| "step": 961 | |
| }, | |
| { | |
| "epoch": 20.47, | |
| "learning_rate": 8.138297872340425e-06, | |
| "loss": 0.1465, | |
| "step": 962 | |
| }, | |
| { | |
| "epoch": 20.49, | |
| "learning_rate": 8.129432624113476e-06, | |
| "loss": 0.198, | |
| "step": 963 | |
| }, | |
| { | |
| "epoch": 20.51, | |
| "learning_rate": 8.120567375886525e-06, | |
| "loss": 0.1373, | |
| "step": 964 | |
| }, | |
| { | |
| "epoch": 20.53, | |
| "learning_rate": 8.111702127659574e-06, | |
| "loss": 0.0983, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 20.55, | |
| "learning_rate": 8.102836879432623e-06, | |
| "loss": 0.1683, | |
| "step": 966 | |
| }, | |
| { | |
| "epoch": 20.57, | |
| "learning_rate": 8.093971631205674e-06, | |
| "loss": 0.13, | |
| "step": 967 | |
| }, | |
| { | |
| "epoch": 20.6, | |
| "learning_rate": 8.085106382978723e-06, | |
| "loss": 0.2037, | |
| "step": 968 | |
| }, | |
| { | |
| "epoch": 20.62, | |
| "learning_rate": 8.076241134751772e-06, | |
| "loss": 0.1004, | |
| "step": 969 | |
| }, | |
| { | |
| "epoch": 20.64, | |
| "learning_rate": 8.067375886524823e-06, | |
| "loss": 0.2872, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 20.66, | |
| "learning_rate": 8.058510638297873e-06, | |
| "loss": 0.0504, | |
| "step": 971 | |
| }, | |
| { | |
| "epoch": 20.68, | |
| "learning_rate": 8.049645390070922e-06, | |
| "loss": 0.0568, | |
| "step": 972 | |
| }, | |
| { | |
| "epoch": 20.7, | |
| "learning_rate": 8.040780141843971e-06, | |
| "loss": 0.1555, | |
| "step": 973 | |
| }, | |
| { | |
| "epoch": 20.72, | |
| "learning_rate": 8.031914893617022e-06, | |
| "loss": 0.1105, | |
| "step": 974 | |
| }, | |
| { | |
| "epoch": 20.74, | |
| "learning_rate": 8.023049645390071e-06, | |
| "loss": 0.0817, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 20.77, | |
| "learning_rate": 8.01418439716312e-06, | |
| "loss": 0.0548, | |
| "step": 976 | |
| }, | |
| { | |
| "epoch": 20.79, | |
| "learning_rate": 8.00531914893617e-06, | |
| "loss": 0.1093, | |
| "step": 977 | |
| }, | |
| { | |
| "epoch": 20.81, | |
| "learning_rate": 7.99645390070922e-06, | |
| "loss": 0.0515, | |
| "step": 978 | |
| }, | |
| { | |
| "epoch": 20.83, | |
| "learning_rate": 7.98758865248227e-06, | |
| "loss": 0.1279, | |
| "step": 979 | |
| }, | |
| { | |
| "epoch": 20.85, | |
| "learning_rate": 7.978723404255319e-06, | |
| "loss": 0.0415, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 20.87, | |
| "learning_rate": 7.969858156028368e-06, | |
| "loss": 0.0788, | |
| "step": 981 | |
| }, | |
| { | |
| "epoch": 20.89, | |
| "learning_rate": 7.960992907801419e-06, | |
| "loss": 0.1698, | |
| "step": 982 | |
| }, | |
| { | |
| "epoch": 20.91, | |
| "learning_rate": 7.952127659574468e-06, | |
| "loss": 0.0951, | |
| "step": 983 | |
| }, | |
| { | |
| "epoch": 20.94, | |
| "learning_rate": 7.943262411347517e-06, | |
| "loss": 0.1008, | |
| "step": 984 | |
| }, | |
| { | |
| "epoch": 20.96, | |
| "learning_rate": 7.934397163120566e-06, | |
| "loss": 0.0619, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 20.98, | |
| "learning_rate": 7.925531914893617e-06, | |
| "loss": 0.028, | |
| "step": 986 | |
| }, | |
| { | |
| "epoch": 21.0, | |
| "learning_rate": 7.916666666666667e-06, | |
| "loss": 0.16, | |
| "step": 987 | |
| }, | |
| { | |
| "epoch": 21.0, | |
| "eval_accuracy": 0.94, | |
| "eval_loss": 0.23405639827251434, | |
| "eval_runtime": 27.4409, | |
| "eval_samples_per_second": 5.466, | |
| "eval_steps_per_second": 0.364, | |
| "step": 987 | |
| }, | |
| { | |
| "epoch": 21.02, | |
| "learning_rate": 7.907801418439716e-06, | |
| "loss": 0.1253, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 21.04, | |
| "learning_rate": 7.898936170212765e-06, | |
| "loss": 0.0658, | |
| "step": 989 | |
| }, | |
| { | |
| "epoch": 21.06, | |
| "learning_rate": 7.890070921985816e-06, | |
| "loss": 0.0527, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 21.09, | |
| "learning_rate": 7.881205673758865e-06, | |
| "loss": 0.0214, | |
| "step": 991 | |
| }, | |
| { | |
| "epoch": 21.11, | |
| "learning_rate": 7.872340425531914e-06, | |
| "loss": 0.1719, | |
| "step": 992 | |
| }, | |
| { | |
| "epoch": 21.13, | |
| "learning_rate": 7.863475177304963e-06, | |
| "loss": 0.1384, | |
| "step": 993 | |
| }, | |
| { | |
| "epoch": 21.15, | |
| "learning_rate": 7.854609929078014e-06, | |
| "loss": 0.0788, | |
| "step": 994 | |
| }, | |
| { | |
| "epoch": 21.17, | |
| "learning_rate": 7.845744680851064e-06, | |
| "loss": 0.1318, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 21.19, | |
| "learning_rate": 7.836879432624113e-06, | |
| "loss": 0.0928, | |
| "step": 996 | |
| }, | |
| { | |
| "epoch": 21.21, | |
| "learning_rate": 7.828014184397162e-06, | |
| "loss": 0.0363, | |
| "step": 997 | |
| }, | |
| { | |
| "epoch": 21.23, | |
| "learning_rate": 7.819148936170213e-06, | |
| "loss": 0.2698, | |
| "step": 998 | |
| }, | |
| { | |
| "epoch": 21.26, | |
| "learning_rate": 7.810283687943262e-06, | |
| "loss": 0.0741, | |
| "step": 999 | |
| }, | |
| { | |
| "epoch": 21.28, | |
| "learning_rate": 7.801418439716311e-06, | |
| "loss": 0.0691, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 21.3, | |
| "learning_rate": 7.79255319148936e-06, | |
| "loss": 0.1175, | |
| "step": 1001 | |
| }, | |
| { | |
| "epoch": 21.32, | |
| "learning_rate": 7.783687943262411e-06, | |
| "loss": 0.1296, | |
| "step": 1002 | |
| }, | |
| { | |
| "epoch": 21.34, | |
| "learning_rate": 7.77482269503546e-06, | |
| "loss": 0.0324, | |
| "step": 1003 | |
| }, | |
| { | |
| "epoch": 21.36, | |
| "learning_rate": 7.76595744680851e-06, | |
| "loss": 0.195, | |
| "step": 1004 | |
| }, | |
| { | |
| "epoch": 21.38, | |
| "learning_rate": 7.75709219858156e-06, | |
| "loss": 0.1613, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 21.4, | |
| "learning_rate": 7.74822695035461e-06, | |
| "loss": 0.1307, | |
| "step": 1006 | |
| }, | |
| { | |
| "epoch": 21.43, | |
| "learning_rate": 7.739361702127659e-06, | |
| "loss": 0.3244, | |
| "step": 1007 | |
| }, | |
| { | |
| "epoch": 21.45, | |
| "learning_rate": 7.730496453900708e-06, | |
| "loss": 0.069, | |
| "step": 1008 | |
| }, | |
| { | |
| "epoch": 21.47, | |
| "learning_rate": 7.721631205673759e-06, | |
| "loss": 0.1258, | |
| "step": 1009 | |
| }, | |
| { | |
| "epoch": 21.49, | |
| "learning_rate": 7.712765957446808e-06, | |
| "loss": 0.0236, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 21.51, | |
| "learning_rate": 7.703900709219858e-06, | |
| "loss": 0.0359, | |
| "step": 1011 | |
| }, | |
| { | |
| "epoch": 21.53, | |
| "learning_rate": 7.695035460992907e-06, | |
| "loss": 0.0437, | |
| "step": 1012 | |
| }, | |
| { | |
| "epoch": 21.55, | |
| "learning_rate": 7.686170212765958e-06, | |
| "loss": 0.1325, | |
| "step": 1013 | |
| }, | |
| { | |
| "epoch": 21.57, | |
| "learning_rate": 7.677304964539007e-06, | |
| "loss": 0.1379, | |
| "step": 1014 | |
| }, | |
| { | |
| "epoch": 21.6, | |
| "learning_rate": 7.668439716312056e-06, | |
| "loss": 0.2367, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 21.62, | |
| "learning_rate": 7.659574468085105e-06, | |
| "loss": 0.1327, | |
| "step": 1016 | |
| }, | |
| { | |
| "epoch": 21.64, | |
| "learning_rate": 7.650709219858156e-06, | |
| "loss": 0.0986, | |
| "step": 1017 | |
| }, | |
| { | |
| "epoch": 21.66, | |
| "learning_rate": 7.641843971631205e-06, | |
| "loss": 0.2345, | |
| "step": 1018 | |
| }, | |
| { | |
| "epoch": 21.68, | |
| "learning_rate": 7.632978723404255e-06, | |
| "loss": 0.1944, | |
| "step": 1019 | |
| }, | |
| { | |
| "epoch": 21.7, | |
| "learning_rate": 7.624113475177305e-06, | |
| "loss": 0.0799, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 21.72, | |
| "learning_rate": 7.6152482269503555e-06, | |
| "loss": 0.1019, | |
| "step": 1021 | |
| }, | |
| { | |
| "epoch": 21.74, | |
| "learning_rate": 7.606382978723405e-06, | |
| "loss": 0.0466, | |
| "step": 1022 | |
| }, | |
| { | |
| "epoch": 21.77, | |
| "learning_rate": 7.597517730496454e-06, | |
| "loss": 0.118, | |
| "step": 1023 | |
| }, | |
| { | |
| "epoch": 21.79, | |
| "learning_rate": 7.588652482269503e-06, | |
| "loss": 0.117, | |
| "step": 1024 | |
| }, | |
| { | |
| "epoch": 21.81, | |
| "learning_rate": 7.579787234042554e-06, | |
| "loss": 0.1552, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 21.83, | |
| "learning_rate": 7.570921985815603e-06, | |
| "loss": 0.1053, | |
| "step": 1026 | |
| }, | |
| { | |
| "epoch": 21.85, | |
| "learning_rate": 7.562056737588652e-06, | |
| "loss": 0.0909, | |
| "step": 1027 | |
| }, | |
| { | |
| "epoch": 21.87, | |
| "learning_rate": 7.553191489361702e-06, | |
| "loss": 0.1486, | |
| "step": 1028 | |
| }, | |
| { | |
| "epoch": 21.89, | |
| "learning_rate": 7.5443262411347525e-06, | |
| "loss": 0.1332, | |
| "step": 1029 | |
| }, | |
| { | |
| "epoch": 21.91, | |
| "learning_rate": 7.535460992907802e-06, | |
| "loss": 0.142, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 21.94, | |
| "learning_rate": 7.526595744680851e-06, | |
| "loss": 0.0985, | |
| "step": 1031 | |
| }, | |
| { | |
| "epoch": 21.96, | |
| "learning_rate": 7.5177304964539e-06, | |
| "loss": 0.0757, | |
| "step": 1032 | |
| }, | |
| { | |
| "epoch": 21.98, | |
| "learning_rate": 7.508865248226951e-06, | |
| "loss": 0.1582, | |
| "step": 1033 | |
| }, | |
| { | |
| "epoch": 22.0, | |
| "learning_rate": 7.5e-06, | |
| "loss": 0.1968, | |
| "step": 1034 | |
| }, | |
| { | |
| "epoch": 22.0, | |
| "eval_accuracy": 0.9066666666666666, | |
| "eval_loss": 0.35522255301475525, | |
| "eval_runtime": 27.5305, | |
| "eval_samples_per_second": 5.449, | |
| "eval_steps_per_second": 0.363, | |
| "step": 1034 | |
| }, | |
| { | |
| "epoch": 22.02, | |
| "learning_rate": 7.49113475177305e-06, | |
| "loss": 0.1003, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 22.04, | |
| "learning_rate": 7.4822695035460994e-06, | |
| "loss": 0.0561, | |
| "step": 1036 | |
| }, | |
| { | |
| "epoch": 22.06, | |
| "learning_rate": 7.4734042553191495e-06, | |
| "loss": 0.2077, | |
| "step": 1037 | |
| }, | |
| { | |
| "epoch": 22.09, | |
| "learning_rate": 7.464539007092199e-06, | |
| "loss": 0.0335, | |
| "step": 1038 | |
| }, | |
| { | |
| "epoch": 22.11, | |
| "learning_rate": 7.455673758865249e-06, | |
| "loss": 0.0902, | |
| "step": 1039 | |
| }, | |
| { | |
| "epoch": 22.13, | |
| "learning_rate": 7.446808510638298e-06, | |
| "loss": 0.2083, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 22.15, | |
| "learning_rate": 7.437943262411348e-06, | |
| "loss": 0.1213, | |
| "step": 1041 | |
| }, | |
| { | |
| "epoch": 22.17, | |
| "learning_rate": 7.429078014184397e-06, | |
| "loss": 0.1161, | |
| "step": 1042 | |
| }, | |
| { | |
| "epoch": 22.19, | |
| "learning_rate": 7.420212765957447e-06, | |
| "loss": 0.1175, | |
| "step": 1043 | |
| }, | |
| { | |
| "epoch": 22.21, | |
| "learning_rate": 7.4113475177304964e-06, | |
| "loss": 0.0772, | |
| "step": 1044 | |
| }, | |
| { | |
| "epoch": 22.23, | |
| "learning_rate": 7.4024822695035465e-06, | |
| "loss": 0.0766, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 22.26, | |
| "learning_rate": 7.393617021276596e-06, | |
| "loss": 0.1255, | |
| "step": 1046 | |
| }, | |
| { | |
| "epoch": 22.28, | |
| "learning_rate": 7.384751773049646e-06, | |
| "loss": 0.1484, | |
| "step": 1047 | |
| }, | |
| { | |
| "epoch": 22.3, | |
| "learning_rate": 7.375886524822695e-06, | |
| "loss": 0.0519, | |
| "step": 1048 | |
| }, | |
| { | |
| "epoch": 22.32, | |
| "learning_rate": 7.367021276595745e-06, | |
| "loss": 0.1121, | |
| "step": 1049 | |
| }, | |
| { | |
| "epoch": 22.34, | |
| "learning_rate": 7.358156028368794e-06, | |
| "loss": 0.1695, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 22.36, | |
| "learning_rate": 7.349290780141844e-06, | |
| "loss": 0.0618, | |
| "step": 1051 | |
| }, | |
| { | |
| "epoch": 22.38, | |
| "learning_rate": 7.3404255319148934e-06, | |
| "loss": 0.2333, | |
| "step": 1052 | |
| }, | |
| { | |
| "epoch": 22.4, | |
| "learning_rate": 7.3315602836879435e-06, | |
| "loss": 0.0709, | |
| "step": 1053 | |
| }, | |
| { | |
| "epoch": 22.43, | |
| "learning_rate": 7.322695035460993e-06, | |
| "loss": 0.0558, | |
| "step": 1054 | |
| }, | |
| { | |
| "epoch": 22.45, | |
| "learning_rate": 7.313829787234043e-06, | |
| "loss": 0.0425, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 22.47, | |
| "learning_rate": 7.304964539007092e-06, | |
| "loss": 0.1995, | |
| "step": 1056 | |
| }, | |
| { | |
| "epoch": 22.49, | |
| "learning_rate": 7.296099290780142e-06, | |
| "loss": 0.0658, | |
| "step": 1057 | |
| }, | |
| { | |
| "epoch": 22.51, | |
| "learning_rate": 7.287234042553191e-06, | |
| "loss": 0.1148, | |
| "step": 1058 | |
| }, | |
| { | |
| "epoch": 22.53, | |
| "learning_rate": 7.278368794326241e-06, | |
| "loss": 0.1157, | |
| "step": 1059 | |
| }, | |
| { | |
| "epoch": 22.55, | |
| "learning_rate": 7.2695035460992904e-06, | |
| "loss": 0.1226, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 22.57, | |
| "learning_rate": 7.2606382978723405e-06, | |
| "loss": 0.126, | |
| "step": 1061 | |
| }, | |
| { | |
| "epoch": 22.6, | |
| "learning_rate": 7.25177304964539e-06, | |
| "loss": 0.1343, | |
| "step": 1062 | |
| }, | |
| { | |
| "epoch": 22.62, | |
| "learning_rate": 7.24290780141844e-06, | |
| "loss": 0.1293, | |
| "step": 1063 | |
| }, | |
| { | |
| "epoch": 22.64, | |
| "learning_rate": 7.234042553191489e-06, | |
| "loss": 0.0695, | |
| "step": 1064 | |
| }, | |
| { | |
| "epoch": 22.66, | |
| "learning_rate": 7.225177304964539e-06, | |
| "loss": 0.0675, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 22.68, | |
| "learning_rate": 7.216312056737588e-06, | |
| "loss": 0.0854, | |
| "step": 1066 | |
| }, | |
| { | |
| "epoch": 22.7, | |
| "learning_rate": 7.207446808510638e-06, | |
| "loss": 0.1393, | |
| "step": 1067 | |
| }, | |
| { | |
| "epoch": 22.72, | |
| "learning_rate": 7.198581560283688e-06, | |
| "loss": 0.0844, | |
| "step": 1068 | |
| }, | |
| { | |
| "epoch": 22.74, | |
| "learning_rate": 7.1897163120567375e-06, | |
| "loss": 0.0994, | |
| "step": 1069 | |
| }, | |
| { | |
| "epoch": 22.77, | |
| "learning_rate": 7.1808510638297875e-06, | |
| "loss": 0.1126, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 22.79, | |
| "learning_rate": 7.171985815602837e-06, | |
| "loss": 0.1795, | |
| "step": 1071 | |
| }, | |
| { | |
| "epoch": 22.81, | |
| "learning_rate": 7.163120567375887e-06, | |
| "loss": 0.0939, | |
| "step": 1072 | |
| }, | |
| { | |
| "epoch": 22.83, | |
| "learning_rate": 7.154255319148936e-06, | |
| "loss": 0.1778, | |
| "step": 1073 | |
| }, | |
| { | |
| "epoch": 22.85, | |
| "learning_rate": 7.145390070921986e-06, | |
| "loss": 0.1106, | |
| "step": 1074 | |
| }, | |
| { | |
| "epoch": 22.87, | |
| "learning_rate": 7.136524822695035e-06, | |
| "loss": 0.2092, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 22.89, | |
| "learning_rate": 7.127659574468085e-06, | |
| "loss": 0.0609, | |
| "step": 1076 | |
| }, | |
| { | |
| "epoch": 22.91, | |
| "learning_rate": 7.1187943262411345e-06, | |
| "loss": 0.0783, | |
| "step": 1077 | |
| }, | |
| { | |
| "epoch": 22.94, | |
| "learning_rate": 7.1099290780141845e-06, | |
| "loss": 0.0797, | |
| "step": 1078 | |
| }, | |
| { | |
| "epoch": 22.96, | |
| "learning_rate": 7.101063829787234e-06, | |
| "loss": 0.0383, | |
| "step": 1079 | |
| }, | |
| { | |
| "epoch": 22.98, | |
| "learning_rate": 7.092198581560284e-06, | |
| "loss": 0.1757, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 23.0, | |
| "learning_rate": 7.083333333333333e-06, | |
| "loss": 0.1049, | |
| "step": 1081 | |
| }, | |
| { | |
| "epoch": 23.0, | |
| "eval_accuracy": 0.9266666666666666, | |
| "eval_loss": 0.25412723422050476, | |
| "eval_runtime": 27.9754, | |
| "eval_samples_per_second": 5.362, | |
| "eval_steps_per_second": 0.357, | |
| "step": 1081 | |
| }, | |
| { | |
| "epoch": 23.02, | |
| "learning_rate": 7.074468085106383e-06, | |
| "loss": 0.2182, | |
| "step": 1082 | |
| }, | |
| { | |
| "epoch": 23.04, | |
| "learning_rate": 7.065602836879432e-06, | |
| "loss": 0.0919, | |
| "step": 1083 | |
| }, | |
| { | |
| "epoch": 23.06, | |
| "learning_rate": 7.056737588652482e-06, | |
| "loss": 0.1797, | |
| "step": 1084 | |
| }, | |
| { | |
| "epoch": 23.09, | |
| "learning_rate": 7.0478723404255315e-06, | |
| "loss": 0.1538, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 23.11, | |
| "learning_rate": 7.0390070921985815e-06, | |
| "loss": 0.1178, | |
| "step": 1086 | |
| }, | |
| { | |
| "epoch": 23.13, | |
| "learning_rate": 7.030141843971631e-06, | |
| "loss": 0.0575, | |
| "step": 1087 | |
| }, | |
| { | |
| "epoch": 23.15, | |
| "learning_rate": 7.021276595744681e-06, | |
| "loss": 0.1962, | |
| "step": 1088 | |
| }, | |
| { | |
| "epoch": 23.17, | |
| "learning_rate": 7.01241134751773e-06, | |
| "loss": 0.1583, | |
| "step": 1089 | |
| }, | |
| { | |
| "epoch": 23.19, | |
| "learning_rate": 7.00354609929078e-06, | |
| "loss": 0.1805, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 23.21, | |
| "learning_rate": 6.994680851063829e-06, | |
| "loss": 0.1405, | |
| "step": 1091 | |
| }, | |
| { | |
| "epoch": 23.23, | |
| "learning_rate": 6.985815602836879e-06, | |
| "loss": 0.0913, | |
| "step": 1092 | |
| }, | |
| { | |
| "epoch": 23.26, | |
| "learning_rate": 6.9769503546099285e-06, | |
| "loss": 0.089, | |
| "step": 1093 | |
| }, | |
| { | |
| "epoch": 23.28, | |
| "learning_rate": 6.9680851063829785e-06, | |
| "loss": 0.0425, | |
| "step": 1094 | |
| }, | |
| { | |
| "epoch": 23.3, | |
| "learning_rate": 6.959219858156029e-06, | |
| "loss": 0.1275, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 23.32, | |
| "learning_rate": 6.950354609929079e-06, | |
| "loss": 0.0567, | |
| "step": 1096 | |
| }, | |
| { | |
| "epoch": 23.34, | |
| "learning_rate": 6.941489361702128e-06, | |
| "loss": 0.1149, | |
| "step": 1097 | |
| }, | |
| { | |
| "epoch": 23.36, | |
| "learning_rate": 6.932624113475178e-06, | |
| "loss": 0.2122, | |
| "step": 1098 | |
| }, | |
| { | |
| "epoch": 23.38, | |
| "learning_rate": 6.923758865248227e-06, | |
| "loss": 0.1219, | |
| "step": 1099 | |
| }, | |
| { | |
| "epoch": 23.4, | |
| "learning_rate": 6.914893617021277e-06, | |
| "loss": 0.0891, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 23.43, | |
| "learning_rate": 6.906028368794327e-06, | |
| "loss": 0.1381, | |
| "step": 1101 | |
| }, | |
| { | |
| "epoch": 23.45, | |
| "learning_rate": 6.897163120567376e-06, | |
| "loss": 0.0776, | |
| "step": 1102 | |
| }, | |
| { | |
| "epoch": 23.47, | |
| "learning_rate": 6.8882978723404264e-06, | |
| "loss": 0.0681, | |
| "step": 1103 | |
| }, | |
| { | |
| "epoch": 23.49, | |
| "learning_rate": 6.879432624113476e-06, | |
| "loss": 0.0745, | |
| "step": 1104 | |
| }, | |
| { | |
| "epoch": 23.51, | |
| "learning_rate": 6.870567375886526e-06, | |
| "loss": 0.0307, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 23.53, | |
| "learning_rate": 6.861702127659575e-06, | |
| "loss": 0.099, | |
| "step": 1106 | |
| }, | |
| { | |
| "epoch": 23.55, | |
| "learning_rate": 6.852836879432625e-06, | |
| "loss": 0.1409, | |
| "step": 1107 | |
| }, | |
| { | |
| "epoch": 23.57, | |
| "learning_rate": 6.843971631205674e-06, | |
| "loss": 0.1254, | |
| "step": 1108 | |
| }, | |
| { | |
| "epoch": 23.6, | |
| "learning_rate": 6.835106382978724e-06, | |
| "loss": 0.1417, | |
| "step": 1109 | |
| }, | |
| { | |
| "epoch": 23.62, | |
| "learning_rate": 6.826241134751773e-06, | |
| "loss": 0.0496, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 23.64, | |
| "learning_rate": 6.8173758865248234e-06, | |
| "loss": 0.1601, | |
| "step": 1111 | |
| }, | |
| { | |
| "epoch": 23.66, | |
| "learning_rate": 6.808510638297873e-06, | |
| "loss": 0.1412, | |
| "step": 1112 | |
| }, | |
| { | |
| "epoch": 23.68, | |
| "learning_rate": 6.799645390070923e-06, | |
| "loss": 0.0645, | |
| "step": 1113 | |
| }, | |
| { | |
| "epoch": 23.7, | |
| "learning_rate": 6.790780141843972e-06, | |
| "loss": 0.1808, | |
| "step": 1114 | |
| }, | |
| { | |
| "epoch": 23.72, | |
| "learning_rate": 6.781914893617022e-06, | |
| "loss": 0.0929, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 23.74, | |
| "learning_rate": 6.773049645390071e-06, | |
| "loss": 0.196, | |
| "step": 1116 | |
| }, | |
| { | |
| "epoch": 23.77, | |
| "learning_rate": 6.764184397163121e-06, | |
| "loss": 0.0641, | |
| "step": 1117 | |
| }, | |
| { | |
| "epoch": 23.79, | |
| "learning_rate": 6.75531914893617e-06, | |
| "loss": 0.1109, | |
| "step": 1118 | |
| }, | |
| { | |
| "epoch": 23.81, | |
| "learning_rate": 6.7464539007092204e-06, | |
| "loss": 0.0288, | |
| "step": 1119 | |
| }, | |
| { | |
| "epoch": 23.83, | |
| "learning_rate": 6.73758865248227e-06, | |
| "loss": 0.173, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 23.85, | |
| "learning_rate": 6.72872340425532e-06, | |
| "loss": 0.0472, | |
| "step": 1121 | |
| }, | |
| { | |
| "epoch": 23.87, | |
| "learning_rate": 6.719858156028369e-06, | |
| "loss": 0.0941, | |
| "step": 1122 | |
| }, | |
| { | |
| "epoch": 23.89, | |
| "learning_rate": 6.710992907801419e-06, | |
| "loss": 0.0552, | |
| "step": 1123 | |
| }, | |
| { | |
| "epoch": 23.91, | |
| "learning_rate": 6.702127659574468e-06, | |
| "loss": 0.0598, | |
| "step": 1124 | |
| }, | |
| { | |
| "epoch": 23.94, | |
| "learning_rate": 6.693262411347518e-06, | |
| "loss": 0.0584, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 23.96, | |
| "learning_rate": 6.684397163120567e-06, | |
| "loss": 0.0546, | |
| "step": 1126 | |
| }, | |
| { | |
| "epoch": 23.98, | |
| "learning_rate": 6.6755319148936174e-06, | |
| "loss": 0.1478, | |
| "step": 1127 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "learning_rate": 6.666666666666667e-06, | |
| "loss": 0.1102, | |
| "step": 1128 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "eval_accuracy": 0.9466666666666667, | |
| "eval_loss": 0.18386472761631012, | |
| "eval_runtime": 28.367, | |
| "eval_samples_per_second": 5.288, | |
| "eval_steps_per_second": 0.353, | |
| "step": 1128 | |
| }, | |
| { | |
| "epoch": 24.02, | |
| "learning_rate": 6.657801418439717e-06, | |
| "loss": 0.1522, | |
| "step": 1129 | |
| }, | |
| { | |
| "epoch": 24.04, | |
| "learning_rate": 6.648936170212766e-06, | |
| "loss": 0.1308, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 24.06, | |
| "learning_rate": 6.640070921985816e-06, | |
| "loss": 0.1576, | |
| "step": 1131 | |
| }, | |
| { | |
| "epoch": 24.09, | |
| "learning_rate": 6.631205673758866e-06, | |
| "loss": 0.0709, | |
| "step": 1132 | |
| }, | |
| { | |
| "epoch": 24.11, | |
| "learning_rate": 6.622340425531915e-06, | |
| "loss": 0.1029, | |
| "step": 1133 | |
| }, | |
| { | |
| "epoch": 24.13, | |
| "learning_rate": 6.613475177304965e-06, | |
| "loss": 0.0299, | |
| "step": 1134 | |
| }, | |
| { | |
| "epoch": 24.15, | |
| "learning_rate": 6.6046099290780144e-06, | |
| "loss": 0.1249, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 24.17, | |
| "learning_rate": 6.5957446808510645e-06, | |
| "loss": 0.159, | |
| "step": 1136 | |
| }, | |
| { | |
| "epoch": 24.19, | |
| "learning_rate": 6.586879432624114e-06, | |
| "loss": 0.1365, | |
| "step": 1137 | |
| }, | |
| { | |
| "epoch": 24.21, | |
| "learning_rate": 6.578014184397164e-06, | |
| "loss": 0.0701, | |
| "step": 1138 | |
| }, | |
| { | |
| "epoch": 24.23, | |
| "learning_rate": 6.569148936170213e-06, | |
| "loss": 0.1008, | |
| "step": 1139 | |
| }, | |
| { | |
| "epoch": 24.26, | |
| "learning_rate": 6.560283687943263e-06, | |
| "loss": 0.0866, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 24.28, | |
| "learning_rate": 6.551418439716312e-06, | |
| "loss": 0.0619, | |
| "step": 1141 | |
| }, | |
| { | |
| "epoch": 24.3, | |
| "learning_rate": 6.542553191489362e-06, | |
| "loss": 0.0512, | |
| "step": 1142 | |
| }, | |
| { | |
| "epoch": 24.32, | |
| "learning_rate": 6.5336879432624114e-06, | |
| "loss": 0.0182, | |
| "step": 1143 | |
| }, | |
| { | |
| "epoch": 24.34, | |
| "learning_rate": 6.5248226950354615e-06, | |
| "loss": 0.091, | |
| "step": 1144 | |
| }, | |
| { | |
| "epoch": 24.36, | |
| "learning_rate": 6.515957446808511e-06, | |
| "loss": 0.1767, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 24.38, | |
| "learning_rate": 6.507092198581561e-06, | |
| "loss": 0.119, | |
| "step": 1146 | |
| }, | |
| { | |
| "epoch": 24.4, | |
| "learning_rate": 6.49822695035461e-06, | |
| "loss": 0.0602, | |
| "step": 1147 | |
| }, | |
| { | |
| "epoch": 24.43, | |
| "learning_rate": 6.48936170212766e-06, | |
| "loss": 0.126, | |
| "step": 1148 | |
| }, | |
| { | |
| "epoch": 24.45, | |
| "learning_rate": 6.480496453900709e-06, | |
| "loss": 0.1149, | |
| "step": 1149 | |
| }, | |
| { | |
| "epoch": 24.47, | |
| "learning_rate": 6.471631205673759e-06, | |
| "loss": 0.0793, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 24.49, | |
| "learning_rate": 6.4627659574468084e-06, | |
| "loss": 0.0914, | |
| "step": 1151 | |
| }, | |
| { | |
| "epoch": 24.51, | |
| "learning_rate": 6.4539007092198585e-06, | |
| "loss": 0.2547, | |
| "step": 1152 | |
| }, | |
| { | |
| "epoch": 24.53, | |
| "learning_rate": 6.445035460992908e-06, | |
| "loss": 0.0852, | |
| "step": 1153 | |
| }, | |
| { | |
| "epoch": 24.55, | |
| "learning_rate": 6.436170212765958e-06, | |
| "loss": 0.1063, | |
| "step": 1154 | |
| }, | |
| { | |
| "epoch": 24.57, | |
| "learning_rate": 6.427304964539007e-06, | |
| "loss": 0.1311, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 24.6, | |
| "learning_rate": 6.418439716312057e-06, | |
| "loss": 0.1015, | |
| "step": 1156 | |
| }, | |
| { | |
| "epoch": 24.62, | |
| "learning_rate": 6.409574468085106e-06, | |
| "loss": 0.1207, | |
| "step": 1157 | |
| }, | |
| { | |
| "epoch": 24.64, | |
| "learning_rate": 6.400709219858156e-06, | |
| "loss": 0.1437, | |
| "step": 1158 | |
| }, | |
| { | |
| "epoch": 24.66, | |
| "learning_rate": 6.3918439716312055e-06, | |
| "loss": 0.2054, | |
| "step": 1159 | |
| }, | |
| { | |
| "epoch": 24.68, | |
| "learning_rate": 6.3829787234042555e-06, | |
| "loss": 0.0767, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 24.7, | |
| "learning_rate": 6.374113475177305e-06, | |
| "loss": 0.051, | |
| "step": 1161 | |
| }, | |
| { | |
| "epoch": 24.72, | |
| "learning_rate": 6.365248226950355e-06, | |
| "loss": 0.1696, | |
| "step": 1162 | |
| }, | |
| { | |
| "epoch": 24.74, | |
| "learning_rate": 6.356382978723404e-06, | |
| "loss": 0.0566, | |
| "step": 1163 | |
| }, | |
| { | |
| "epoch": 24.77, | |
| "learning_rate": 6.347517730496454e-06, | |
| "loss": 0.1097, | |
| "step": 1164 | |
| }, | |
| { | |
| "epoch": 24.79, | |
| "learning_rate": 6.338652482269504e-06, | |
| "loss": 0.1762, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 24.81, | |
| "learning_rate": 6.329787234042553e-06, | |
| "loss": 0.1866, | |
| "step": 1166 | |
| }, | |
| { | |
| "epoch": 24.83, | |
| "learning_rate": 6.320921985815603e-06, | |
| "loss": 0.1638, | |
| "step": 1167 | |
| }, | |
| { | |
| "epoch": 24.85, | |
| "learning_rate": 6.3120567375886525e-06, | |
| "loss": 0.0754, | |
| "step": 1168 | |
| }, | |
| { | |
| "epoch": 24.87, | |
| "learning_rate": 6.3031914893617025e-06, | |
| "loss": 0.0812, | |
| "step": 1169 | |
| }, | |
| { | |
| "epoch": 24.89, | |
| "learning_rate": 6.294326241134752e-06, | |
| "loss": 0.1629, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 24.91, | |
| "learning_rate": 6.285460992907802e-06, | |
| "loss": 0.1143, | |
| "step": 1171 | |
| }, | |
| { | |
| "epoch": 24.94, | |
| "learning_rate": 6.276595744680851e-06, | |
| "loss": 0.0937, | |
| "step": 1172 | |
| }, | |
| { | |
| "epoch": 24.96, | |
| "learning_rate": 6.267730496453901e-06, | |
| "loss": 0.0565, | |
| "step": 1173 | |
| }, | |
| { | |
| "epoch": 24.98, | |
| "learning_rate": 6.25886524822695e-06, | |
| "loss": 0.0355, | |
| "step": 1174 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "learning_rate": 6.25e-06, | |
| "loss": 0.3039, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "eval_accuracy": 0.9333333333333333, | |
| "eval_loss": 0.22690236568450928, | |
| "eval_runtime": 28.007, | |
| "eval_samples_per_second": 5.356, | |
| "eval_steps_per_second": 0.357, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 25.02, | |
| "learning_rate": 6.2411347517730495e-06, | |
| "loss": 0.073, | |
| "step": 1176 | |
| }, | |
| { | |
| "epoch": 25.04, | |
| "learning_rate": 6.2322695035460995e-06, | |
| "loss": 0.0825, | |
| "step": 1177 | |
| }, | |
| { | |
| "epoch": 25.06, | |
| "learning_rate": 6.223404255319149e-06, | |
| "loss": 0.073, | |
| "step": 1178 | |
| }, | |
| { | |
| "epoch": 25.09, | |
| "learning_rate": 6.214539007092199e-06, | |
| "loss": 0.1091, | |
| "step": 1179 | |
| }, | |
| { | |
| "epoch": 25.11, | |
| "learning_rate": 6.205673758865248e-06, | |
| "loss": 0.1012, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 25.13, | |
| "learning_rate": 6.196808510638298e-06, | |
| "loss": 0.0724, | |
| "step": 1181 | |
| }, | |
| { | |
| "epoch": 25.15, | |
| "learning_rate": 6.187943262411347e-06, | |
| "loss": 0.0507, | |
| "step": 1182 | |
| }, | |
| { | |
| "epoch": 25.17, | |
| "learning_rate": 6.179078014184397e-06, | |
| "loss": 0.1333, | |
| "step": 1183 | |
| }, | |
| { | |
| "epoch": 25.19, | |
| "learning_rate": 6.1702127659574465e-06, | |
| "loss": 0.071, | |
| "step": 1184 | |
| }, | |
| { | |
| "epoch": 25.21, | |
| "learning_rate": 6.1613475177304966e-06, | |
| "loss": 0.0801, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 25.23, | |
| "learning_rate": 6.152482269503546e-06, | |
| "loss": 0.0461, | |
| "step": 1186 | |
| }, | |
| { | |
| "epoch": 25.26, | |
| "learning_rate": 6.143617021276596e-06, | |
| "loss": 0.0574, | |
| "step": 1187 | |
| }, | |
| { | |
| "epoch": 25.28, | |
| "learning_rate": 6.134751773049645e-06, | |
| "loss": 0.0471, | |
| "step": 1188 | |
| }, | |
| { | |
| "epoch": 25.3, | |
| "learning_rate": 6.125886524822695e-06, | |
| "loss": 0.0473, | |
| "step": 1189 | |
| }, | |
| { | |
| "epoch": 25.32, | |
| "learning_rate": 6.117021276595744e-06, | |
| "loss": 0.0988, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 25.34, | |
| "learning_rate": 6.108156028368794e-06, | |
| "loss": 0.0829, | |
| "step": 1191 | |
| }, | |
| { | |
| "epoch": 25.36, | |
| "learning_rate": 6.0992907801418435e-06, | |
| "loss": 0.1398, | |
| "step": 1192 | |
| }, | |
| { | |
| "epoch": 25.38, | |
| "learning_rate": 6.0904255319148936e-06, | |
| "loss": 0.0833, | |
| "step": 1193 | |
| }, | |
| { | |
| "epoch": 25.4, | |
| "learning_rate": 6.081560283687943e-06, | |
| "loss": 0.1067, | |
| "step": 1194 | |
| }, | |
| { | |
| "epoch": 25.43, | |
| "learning_rate": 6.072695035460993e-06, | |
| "loss": 0.0681, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 25.45, | |
| "learning_rate": 6.063829787234042e-06, | |
| "loss": 0.0417, | |
| "step": 1196 | |
| }, | |
| { | |
| "epoch": 25.47, | |
| "learning_rate": 6.054964539007092e-06, | |
| "loss": 0.03, | |
| "step": 1197 | |
| }, | |
| { | |
| "epoch": 25.49, | |
| "learning_rate": 6.046099290780142e-06, | |
| "loss": 0.1714, | |
| "step": 1198 | |
| }, | |
| { | |
| "epoch": 25.51, | |
| "learning_rate": 6.037234042553191e-06, | |
| "loss": 0.0912, | |
| "step": 1199 | |
| }, | |
| { | |
| "epoch": 25.53, | |
| "learning_rate": 6.028368794326241e-06, | |
| "loss": 0.1877, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 25.55, | |
| "learning_rate": 6.0195035460992906e-06, | |
| "loss": 0.1312, | |
| "step": 1201 | |
| }, | |
| { | |
| "epoch": 25.57, | |
| "learning_rate": 6.010638297872341e-06, | |
| "loss": 0.0524, | |
| "step": 1202 | |
| }, | |
| { | |
| "epoch": 25.6, | |
| "learning_rate": 6.00177304964539e-06, | |
| "loss": 0.0931, | |
| "step": 1203 | |
| }, | |
| { | |
| "epoch": 25.62, | |
| "learning_rate": 5.99290780141844e-06, | |
| "loss": 0.1839, | |
| "step": 1204 | |
| }, | |
| { | |
| "epoch": 25.64, | |
| "learning_rate": 5.984042553191489e-06, | |
| "loss": 0.0787, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 25.66, | |
| "learning_rate": 5.975177304964539e-06, | |
| "loss": 0.1313, | |
| "step": 1206 | |
| }, | |
| { | |
| "epoch": 25.68, | |
| "learning_rate": 5.966312056737588e-06, | |
| "loss": 0.1401, | |
| "step": 1207 | |
| }, | |
| { | |
| "epoch": 25.7, | |
| "learning_rate": 5.957446808510638e-06, | |
| "loss": 0.1008, | |
| "step": 1208 | |
| }, | |
| { | |
| "epoch": 25.72, | |
| "learning_rate": 5.9485815602836876e-06, | |
| "loss": 0.1085, | |
| "step": 1209 | |
| }, | |
| { | |
| "epoch": 25.74, | |
| "learning_rate": 5.939716312056738e-06, | |
| "loss": 0.099, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 25.77, | |
| "learning_rate": 5.930851063829787e-06, | |
| "loss": 0.2044, | |
| "step": 1211 | |
| }, | |
| { | |
| "epoch": 25.79, | |
| "learning_rate": 5.921985815602837e-06, | |
| "loss": 0.0578, | |
| "step": 1212 | |
| }, | |
| { | |
| "epoch": 25.81, | |
| "learning_rate": 5.913120567375886e-06, | |
| "loss": 0.1401, | |
| "step": 1213 | |
| }, | |
| { | |
| "epoch": 25.83, | |
| "learning_rate": 5.904255319148936e-06, | |
| "loss": 0.0713, | |
| "step": 1214 | |
| }, | |
| { | |
| "epoch": 25.85, | |
| "learning_rate": 5.895390070921986e-06, | |
| "loss": 0.1164, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 25.87, | |
| "learning_rate": 5.886524822695036e-06, | |
| "loss": 0.1817, | |
| "step": 1216 | |
| }, | |
| { | |
| "epoch": 25.89, | |
| "learning_rate": 5.877659574468085e-06, | |
| "loss": 0.1177, | |
| "step": 1217 | |
| }, | |
| { | |
| "epoch": 25.91, | |
| "learning_rate": 5.8687943262411355e-06, | |
| "loss": 0.0854, | |
| "step": 1218 | |
| }, | |
| { | |
| "epoch": 25.94, | |
| "learning_rate": 5.859929078014185e-06, | |
| "loss": 0.1344, | |
| "step": 1219 | |
| }, | |
| { | |
| "epoch": 25.96, | |
| "learning_rate": 5.851063829787235e-06, | |
| "loss": 0.1374, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 25.98, | |
| "learning_rate": 5.842198581560284e-06, | |
| "loss": 0.0723, | |
| "step": 1221 | |
| }, | |
| { | |
| "epoch": 26.0, | |
| "learning_rate": 5.833333333333334e-06, | |
| "loss": 0.1188, | |
| "step": 1222 | |
| }, | |
| { | |
| "epoch": 26.0, | |
| "eval_accuracy": 0.9533333333333334, | |
| "eval_loss": 0.2063450664281845, | |
| "eval_runtime": 25.0393, | |
| "eval_samples_per_second": 5.991, | |
| "eval_steps_per_second": 0.399, | |
| "step": 1222 | |
| }, | |
| { | |
| "epoch": 26.02, | |
| "learning_rate": 5.824468085106383e-06, | |
| "loss": 0.0595, | |
| "step": 1223 | |
| }, | |
| { | |
| "epoch": 26.04, | |
| "learning_rate": 5.815602836879433e-06, | |
| "loss": 0.0651, | |
| "step": 1224 | |
| }, | |
| { | |
| "epoch": 26.06, | |
| "learning_rate": 5.806737588652482e-06, | |
| "loss": 0.1662, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 26.09, | |
| "learning_rate": 5.7978723404255325e-06, | |
| "loss": 0.0509, | |
| "step": 1226 | |
| }, | |
| { | |
| "epoch": 26.11, | |
| "learning_rate": 5.789007092198582e-06, | |
| "loss": 0.0349, | |
| "step": 1227 | |
| }, | |
| { | |
| "epoch": 26.13, | |
| "learning_rate": 5.780141843971632e-06, | |
| "loss": 0.1645, | |
| "step": 1228 | |
| }, | |
| { | |
| "epoch": 26.15, | |
| "learning_rate": 5.771276595744681e-06, | |
| "loss": 0.0612, | |
| "step": 1229 | |
| }, | |
| { | |
| "epoch": 26.17, | |
| "learning_rate": 5.762411347517731e-06, | |
| "loss": 0.1428, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 26.19, | |
| "learning_rate": 5.753546099290781e-06, | |
| "loss": 0.2387, | |
| "step": 1231 | |
| }, | |
| { | |
| "epoch": 26.21, | |
| "learning_rate": 5.74468085106383e-06, | |
| "loss": 0.0854, | |
| "step": 1232 | |
| }, | |
| { | |
| "epoch": 26.23, | |
| "learning_rate": 5.73581560283688e-06, | |
| "loss": 0.1895, | |
| "step": 1233 | |
| }, | |
| { | |
| "epoch": 26.26, | |
| "learning_rate": 5.7269503546099295e-06, | |
| "loss": 0.0503, | |
| "step": 1234 | |
| }, | |
| { | |
| "epoch": 26.28, | |
| "learning_rate": 5.7180851063829795e-06, | |
| "loss": 0.0948, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 26.3, | |
| "learning_rate": 5.709219858156029e-06, | |
| "loss": 0.15, | |
| "step": 1236 | |
| }, | |
| { | |
| "epoch": 26.32, | |
| "learning_rate": 5.700354609929079e-06, | |
| "loss": 0.0873, | |
| "step": 1237 | |
| }, | |
| { | |
| "epoch": 26.34, | |
| "learning_rate": 5.691489361702128e-06, | |
| "loss": 0.1345, | |
| "step": 1238 | |
| }, | |
| { | |
| "epoch": 26.36, | |
| "learning_rate": 5.682624113475178e-06, | |
| "loss": 0.1343, | |
| "step": 1239 | |
| }, | |
| { | |
| "epoch": 26.38, | |
| "learning_rate": 5.673758865248227e-06, | |
| "loss": 0.1756, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 26.4, | |
| "learning_rate": 5.664893617021277e-06, | |
| "loss": 0.1113, | |
| "step": 1241 | |
| }, | |
| { | |
| "epoch": 26.43, | |
| "learning_rate": 5.6560283687943265e-06, | |
| "loss": 0.1345, | |
| "step": 1242 | |
| }, | |
| { | |
| "epoch": 26.45, | |
| "learning_rate": 5.6471631205673765e-06, | |
| "loss": 0.1905, | |
| "step": 1243 | |
| }, | |
| { | |
| "epoch": 26.47, | |
| "learning_rate": 5.638297872340426e-06, | |
| "loss": 0.1241, | |
| "step": 1244 | |
| }, | |
| { | |
| "epoch": 26.49, | |
| "learning_rate": 5.629432624113476e-06, | |
| "loss": 0.0772, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 26.51, | |
| "learning_rate": 5.620567375886525e-06, | |
| "loss": 0.1829, | |
| "step": 1246 | |
| }, | |
| { | |
| "epoch": 26.53, | |
| "learning_rate": 5.611702127659575e-06, | |
| "loss": 0.0687, | |
| "step": 1247 | |
| }, | |
| { | |
| "epoch": 26.55, | |
| "learning_rate": 5.602836879432624e-06, | |
| "loss": 0.0398, | |
| "step": 1248 | |
| }, | |
| { | |
| "epoch": 26.57, | |
| "learning_rate": 5.593971631205674e-06, | |
| "loss": 0.2204, | |
| "step": 1249 | |
| }, | |
| { | |
| "epoch": 26.6, | |
| "learning_rate": 5.5851063829787235e-06, | |
| "loss": 0.176, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 26.62, | |
| "learning_rate": 5.5762411347517735e-06, | |
| "loss": 0.1709, | |
| "step": 1251 | |
| }, | |
| { | |
| "epoch": 26.64, | |
| "learning_rate": 5.567375886524823e-06, | |
| "loss": 0.1042, | |
| "step": 1252 | |
| }, | |
| { | |
| "epoch": 26.66, | |
| "learning_rate": 5.558510638297873e-06, | |
| "loss": 0.1964, | |
| "step": 1253 | |
| }, | |
| { | |
| "epoch": 26.68, | |
| "learning_rate": 5.549645390070922e-06, | |
| "loss": 0.0956, | |
| "step": 1254 | |
| }, | |
| { | |
| "epoch": 26.7, | |
| "learning_rate": 5.540780141843972e-06, | |
| "loss": 0.118, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 26.72, | |
| "learning_rate": 5.531914893617021e-06, | |
| "loss": 0.0913, | |
| "step": 1256 | |
| }, | |
| { | |
| "epoch": 26.74, | |
| "learning_rate": 5.523049645390071e-06, | |
| "loss": 0.0838, | |
| "step": 1257 | |
| }, | |
| { | |
| "epoch": 26.77, | |
| "learning_rate": 5.5141843971631205e-06, | |
| "loss": 0.0875, | |
| "step": 1258 | |
| }, | |
| { | |
| "epoch": 26.79, | |
| "learning_rate": 5.5053191489361705e-06, | |
| "loss": 0.1726, | |
| "step": 1259 | |
| }, | |
| { | |
| "epoch": 26.81, | |
| "learning_rate": 5.49645390070922e-06, | |
| "loss": 0.1131, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 26.83, | |
| "learning_rate": 5.48758865248227e-06, | |
| "loss": 0.0656, | |
| "step": 1261 | |
| }, | |
| { | |
| "epoch": 26.85, | |
| "learning_rate": 5.47872340425532e-06, | |
| "loss": 0.0378, | |
| "step": 1262 | |
| }, | |
| { | |
| "epoch": 26.87, | |
| "learning_rate": 5.469858156028369e-06, | |
| "loss": 0.1447, | |
| "step": 1263 | |
| }, | |
| { | |
| "epoch": 26.89, | |
| "learning_rate": 5.460992907801419e-06, | |
| "loss": 0.1445, | |
| "step": 1264 | |
| }, | |
| { | |
| "epoch": 26.91, | |
| "learning_rate": 5.452127659574468e-06, | |
| "loss": 0.0846, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 26.94, | |
| "learning_rate": 5.443262411347518e-06, | |
| "loss": 0.0831, | |
| "step": 1266 | |
| }, | |
| { | |
| "epoch": 26.96, | |
| "learning_rate": 5.4343971631205675e-06, | |
| "loss": 0.1197, | |
| "step": 1267 | |
| }, | |
| { | |
| "epoch": 26.98, | |
| "learning_rate": 5.4255319148936176e-06, | |
| "loss": 0.0863, | |
| "step": 1268 | |
| }, | |
| { | |
| "epoch": 27.0, | |
| "learning_rate": 5.416666666666667e-06, | |
| "loss": 0.2008, | |
| "step": 1269 | |
| }, | |
| { | |
| "epoch": 27.0, | |
| "eval_accuracy": 0.94, | |
| "eval_loss": 0.19716215133666992, | |
| "eval_runtime": 24.9234, | |
| "eval_samples_per_second": 6.018, | |
| "eval_steps_per_second": 0.401, | |
| "step": 1269 | |
| }, | |
| { | |
| "epoch": 27.02, | |
| "learning_rate": 5.407801418439717e-06, | |
| "loss": 0.1059, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 27.04, | |
| "learning_rate": 5.398936170212766e-06, | |
| "loss": 0.0287, | |
| "step": 1271 | |
| }, | |
| { | |
| "epoch": 27.06, | |
| "learning_rate": 5.390070921985816e-06, | |
| "loss": 0.0672, | |
| "step": 1272 | |
| }, | |
| { | |
| "epoch": 27.09, | |
| "learning_rate": 5.381205673758865e-06, | |
| "loss": 0.0791, | |
| "step": 1273 | |
| }, | |
| { | |
| "epoch": 27.11, | |
| "learning_rate": 5.372340425531915e-06, | |
| "loss": 0.1095, | |
| "step": 1274 | |
| }, | |
| { | |
| "epoch": 27.13, | |
| "learning_rate": 5.3634751773049645e-06, | |
| "loss": 0.1848, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 27.15, | |
| "learning_rate": 5.3546099290780146e-06, | |
| "loss": 0.0421, | |
| "step": 1276 | |
| }, | |
| { | |
| "epoch": 27.17, | |
| "learning_rate": 5.345744680851064e-06, | |
| "loss": 0.1635, | |
| "step": 1277 | |
| }, | |
| { | |
| "epoch": 27.19, | |
| "learning_rate": 5.336879432624114e-06, | |
| "loss": 0.0495, | |
| "step": 1278 | |
| }, | |
| { | |
| "epoch": 27.21, | |
| "learning_rate": 5.328014184397163e-06, | |
| "loss": 0.0947, | |
| "step": 1279 | |
| }, | |
| { | |
| "epoch": 27.23, | |
| "learning_rate": 5.319148936170213e-06, | |
| "loss": 0.0758, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 27.26, | |
| "learning_rate": 5.310283687943262e-06, | |
| "loss": 0.0598, | |
| "step": 1281 | |
| }, | |
| { | |
| "epoch": 27.28, | |
| "learning_rate": 5.301418439716312e-06, | |
| "loss": 0.1515, | |
| "step": 1282 | |
| }, | |
| { | |
| "epoch": 27.3, | |
| "learning_rate": 5.2925531914893615e-06, | |
| "loss": 0.1175, | |
| "step": 1283 | |
| }, | |
| { | |
| "epoch": 27.32, | |
| "learning_rate": 5.2836879432624116e-06, | |
| "loss": 0.0454, | |
| "step": 1284 | |
| }, | |
| { | |
| "epoch": 27.34, | |
| "learning_rate": 5.274822695035461e-06, | |
| "loss": 0.1373, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 27.36, | |
| "learning_rate": 5.265957446808511e-06, | |
| "loss": 0.159, | |
| "step": 1286 | |
| }, | |
| { | |
| "epoch": 27.38, | |
| "learning_rate": 5.25709219858156e-06, | |
| "loss": 0.1784, | |
| "step": 1287 | |
| }, | |
| { | |
| "epoch": 27.4, | |
| "learning_rate": 5.24822695035461e-06, | |
| "loss": 0.1493, | |
| "step": 1288 | |
| }, | |
| { | |
| "epoch": 27.43, | |
| "learning_rate": 5.239361702127659e-06, | |
| "loss": 0.0678, | |
| "step": 1289 | |
| }, | |
| { | |
| "epoch": 27.45, | |
| "learning_rate": 5.230496453900709e-06, | |
| "loss": 0.1112, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 27.47, | |
| "learning_rate": 5.2216312056737585e-06, | |
| "loss": 0.1168, | |
| "step": 1291 | |
| }, | |
| { | |
| "epoch": 27.49, | |
| "learning_rate": 5.2127659574468086e-06, | |
| "loss": 0.029, | |
| "step": 1292 | |
| }, | |
| { | |
| "epoch": 27.51, | |
| "learning_rate": 5.203900709219858e-06, | |
| "loss": 0.0683, | |
| "step": 1293 | |
| }, | |
| { | |
| "epoch": 27.53, | |
| "learning_rate": 5.195035460992908e-06, | |
| "loss": 0.0563, | |
| "step": 1294 | |
| }, | |
| { | |
| "epoch": 27.55, | |
| "learning_rate": 5.186170212765958e-06, | |
| "loss": 0.0939, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 27.57, | |
| "learning_rate": 5.177304964539007e-06, | |
| "loss": 0.0459, | |
| "step": 1296 | |
| }, | |
| { | |
| "epoch": 27.6, | |
| "learning_rate": 5.168439716312057e-06, | |
| "loss": 0.1148, | |
| "step": 1297 | |
| }, | |
| { | |
| "epoch": 27.62, | |
| "learning_rate": 5.159574468085106e-06, | |
| "loss": 0.1735, | |
| "step": 1298 | |
| }, | |
| { | |
| "epoch": 27.64, | |
| "learning_rate": 5.150709219858156e-06, | |
| "loss": 0.0884, | |
| "step": 1299 | |
| }, | |
| { | |
| "epoch": 27.66, | |
| "learning_rate": 5.1418439716312056e-06, | |
| "loss": 0.1077, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 27.68, | |
| "learning_rate": 5.132978723404256e-06, | |
| "loss": 0.1192, | |
| "step": 1301 | |
| }, | |
| { | |
| "epoch": 27.7, | |
| "learning_rate": 5.124113475177305e-06, | |
| "loss": 0.076, | |
| "step": 1302 | |
| }, | |
| { | |
| "epoch": 27.72, | |
| "learning_rate": 5.115248226950355e-06, | |
| "loss": 0.1373, | |
| "step": 1303 | |
| }, | |
| { | |
| "epoch": 27.74, | |
| "learning_rate": 5.106382978723404e-06, | |
| "loss": 0.0539, | |
| "step": 1304 | |
| }, | |
| { | |
| "epoch": 27.77, | |
| "learning_rate": 5.097517730496454e-06, | |
| "loss": 0.1357, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 27.79, | |
| "learning_rate": 5.088652482269503e-06, | |
| "loss": 0.1443, | |
| "step": 1306 | |
| }, | |
| { | |
| "epoch": 27.81, | |
| "learning_rate": 5.079787234042553e-06, | |
| "loss": 0.1314, | |
| "step": 1307 | |
| }, | |
| { | |
| "epoch": 27.83, | |
| "learning_rate": 5.0709219858156026e-06, | |
| "loss": 0.1095, | |
| "step": 1308 | |
| }, | |
| { | |
| "epoch": 27.85, | |
| "learning_rate": 5.062056737588653e-06, | |
| "loss": 0.0325, | |
| "step": 1309 | |
| }, | |
| { | |
| "epoch": 27.87, | |
| "learning_rate": 5.053191489361702e-06, | |
| "loss": 0.1123, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 27.89, | |
| "learning_rate": 5.044326241134752e-06, | |
| "loss": 0.1057, | |
| "step": 1311 | |
| }, | |
| { | |
| "epoch": 27.91, | |
| "learning_rate": 5.035460992907801e-06, | |
| "loss": 0.1266, | |
| "step": 1312 | |
| }, | |
| { | |
| "epoch": 27.94, | |
| "learning_rate": 5.026595744680851e-06, | |
| "loss": 0.0745, | |
| "step": 1313 | |
| }, | |
| { | |
| "epoch": 27.96, | |
| "learning_rate": 5.0177304964539e-06, | |
| "loss": 0.0895, | |
| "step": 1314 | |
| }, | |
| { | |
| "epoch": 27.98, | |
| "learning_rate": 5.00886524822695e-06, | |
| "loss": 0.1719, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "learning_rate": 4.9999999999999996e-06, | |
| "loss": 0.1113, | |
| "step": 1316 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "eval_accuracy": 0.94, | |
| "eval_loss": 0.21567092835903168, | |
| "eval_runtime": 25.2004, | |
| "eval_samples_per_second": 5.952, | |
| "eval_steps_per_second": 0.397, | |
| "step": 1316 | |
| }, | |
| { | |
| "epoch": 28.02, | |
| "learning_rate": 4.99113475177305e-06, | |
| "loss": 0.0545, | |
| "step": 1317 | |
| }, | |
| { | |
| "epoch": 28.04, | |
| "learning_rate": 4.982269503546099e-06, | |
| "loss": 0.1276, | |
| "step": 1318 | |
| }, | |
| { | |
| "epoch": 28.06, | |
| "learning_rate": 4.973404255319149e-06, | |
| "loss": 0.142, | |
| "step": 1319 | |
| }, | |
| { | |
| "epoch": 28.09, | |
| "learning_rate": 4.964539007092198e-06, | |
| "loss": 0.1783, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 28.11, | |
| "learning_rate": 4.955673758865248e-06, | |
| "loss": 0.0231, | |
| "step": 1321 | |
| }, | |
| { | |
| "epoch": 28.13, | |
| "learning_rate": 4.946808510638297e-06, | |
| "loss": 0.1805, | |
| "step": 1322 | |
| }, | |
| { | |
| "epoch": 28.15, | |
| "learning_rate": 4.937943262411347e-06, | |
| "loss": 0.1121, | |
| "step": 1323 | |
| }, | |
| { | |
| "epoch": 28.17, | |
| "learning_rate": 4.9290780141843966e-06, | |
| "loss": 0.1239, | |
| "step": 1324 | |
| }, | |
| { | |
| "epoch": 28.19, | |
| "learning_rate": 4.920212765957447e-06, | |
| "loss": 0.0837, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 28.21, | |
| "learning_rate": 4.911347517730496e-06, | |
| "loss": 0.0545, | |
| "step": 1326 | |
| }, | |
| { | |
| "epoch": 28.23, | |
| "learning_rate": 4.902482269503546e-06, | |
| "loss": 0.0922, | |
| "step": 1327 | |
| }, | |
| { | |
| "epoch": 28.26, | |
| "learning_rate": 4.893617021276596e-06, | |
| "loss": 0.1383, | |
| "step": 1328 | |
| }, | |
| { | |
| "epoch": 28.28, | |
| "learning_rate": 4.884751773049645e-06, | |
| "loss": 0.0128, | |
| "step": 1329 | |
| }, | |
| { | |
| "epoch": 28.3, | |
| "learning_rate": 4.875886524822695e-06, | |
| "loss": 0.1164, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 28.32, | |
| "learning_rate": 4.867021276595744e-06, | |
| "loss": 0.1099, | |
| "step": 1331 | |
| }, | |
| { | |
| "epoch": 28.34, | |
| "learning_rate": 4.858156028368794e-06, | |
| "loss": 0.072, | |
| "step": 1332 | |
| }, | |
| { | |
| "epoch": 28.36, | |
| "learning_rate": 4.849290780141844e-06, | |
| "loss": 0.1136, | |
| "step": 1333 | |
| }, | |
| { | |
| "epoch": 28.38, | |
| "learning_rate": 4.840425531914894e-06, | |
| "loss": 0.1904, | |
| "step": 1334 | |
| }, | |
| { | |
| "epoch": 28.4, | |
| "learning_rate": 4.831560283687943e-06, | |
| "loss": 0.1201, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 28.43, | |
| "learning_rate": 4.822695035460993e-06, | |
| "loss": 0.2126, | |
| "step": 1336 | |
| }, | |
| { | |
| "epoch": 28.45, | |
| "learning_rate": 4.813829787234043e-06, | |
| "loss": 0.0973, | |
| "step": 1337 | |
| }, | |
| { | |
| "epoch": 28.47, | |
| "learning_rate": 4.804964539007093e-06, | |
| "loss": 0.0243, | |
| "step": 1338 | |
| }, | |
| { | |
| "epoch": 28.49, | |
| "learning_rate": 4.796099290780142e-06, | |
| "loss": 0.0855, | |
| "step": 1339 | |
| }, | |
| { | |
| "epoch": 28.51, | |
| "learning_rate": 4.787234042553192e-06, | |
| "loss": 0.1365, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 28.53, | |
| "learning_rate": 4.7783687943262415e-06, | |
| "loss": 0.0964, | |
| "step": 1341 | |
| }, | |
| { | |
| "epoch": 28.55, | |
| "learning_rate": 4.7695035460992915e-06, | |
| "loss": 0.2668, | |
| "step": 1342 | |
| }, | |
| { | |
| "epoch": 28.57, | |
| "learning_rate": 4.760638297872341e-06, | |
| "loss": 0.029, | |
| "step": 1343 | |
| }, | |
| { | |
| "epoch": 28.6, | |
| "learning_rate": 4.751773049645391e-06, | |
| "loss": 0.0855, | |
| "step": 1344 | |
| }, | |
| { | |
| "epoch": 28.62, | |
| "learning_rate": 4.74290780141844e-06, | |
| "loss": 0.052, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 28.64, | |
| "learning_rate": 4.73404255319149e-06, | |
| "loss": 0.0962, | |
| "step": 1346 | |
| }, | |
| { | |
| "epoch": 28.66, | |
| "learning_rate": 4.725177304964539e-06, | |
| "loss": 0.1003, | |
| "step": 1347 | |
| }, | |
| { | |
| "epoch": 28.68, | |
| "learning_rate": 4.716312056737589e-06, | |
| "loss": 0.0227, | |
| "step": 1348 | |
| }, | |
| { | |
| "epoch": 28.7, | |
| "learning_rate": 4.7074468085106385e-06, | |
| "loss": 0.0667, | |
| "step": 1349 | |
| }, | |
| { | |
| "epoch": 28.72, | |
| "learning_rate": 4.6985815602836885e-06, | |
| "loss": 0.0559, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 28.74, | |
| "learning_rate": 4.689716312056738e-06, | |
| "loss": 0.1089, | |
| "step": 1351 | |
| }, | |
| { | |
| "epoch": 28.77, | |
| "learning_rate": 4.680851063829788e-06, | |
| "loss": 0.0473, | |
| "step": 1352 | |
| }, | |
| { | |
| "epoch": 28.79, | |
| "learning_rate": 4.671985815602837e-06, | |
| "loss": 0.0902, | |
| "step": 1353 | |
| }, | |
| { | |
| "epoch": 28.81, | |
| "learning_rate": 4.663120567375887e-06, | |
| "loss": 0.0864, | |
| "step": 1354 | |
| }, | |
| { | |
| "epoch": 28.83, | |
| "learning_rate": 4.654255319148936e-06, | |
| "loss": 0.0814, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 28.85, | |
| "learning_rate": 4.645390070921986e-06, | |
| "loss": 0.0436, | |
| "step": 1356 | |
| }, | |
| { | |
| "epoch": 28.87, | |
| "learning_rate": 4.6365248226950355e-06, | |
| "loss": 0.1101, | |
| "step": 1357 | |
| }, | |
| { | |
| "epoch": 28.89, | |
| "learning_rate": 4.6276595744680855e-06, | |
| "loss": 0.078, | |
| "step": 1358 | |
| }, | |
| { | |
| "epoch": 28.91, | |
| "learning_rate": 4.618794326241135e-06, | |
| "loss": 0.1061, | |
| "step": 1359 | |
| }, | |
| { | |
| "epoch": 28.94, | |
| "learning_rate": 4.609929078014185e-06, | |
| "loss": 0.0854, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 28.96, | |
| "learning_rate": 4.601063829787235e-06, | |
| "loss": 0.2126, | |
| "step": 1361 | |
| }, | |
| { | |
| "epoch": 28.98, | |
| "learning_rate": 4.592198581560284e-06, | |
| "loss": 0.0875, | |
| "step": 1362 | |
| }, | |
| { | |
| "epoch": 29.0, | |
| "learning_rate": 4.583333333333334e-06, | |
| "loss": 0.1377, | |
| "step": 1363 | |
| }, | |
| { | |
| "epoch": 29.0, | |
| "eval_accuracy": 0.9533333333333334, | |
| "eval_loss": 0.20305314660072327, | |
| "eval_runtime": 24.9056, | |
| "eval_samples_per_second": 6.023, | |
| "eval_steps_per_second": 0.402, | |
| "step": 1363 | |
| }, | |
| { | |
| "epoch": 29.02, | |
| "learning_rate": 4.574468085106383e-06, | |
| "loss": 0.1088, | |
| "step": 1364 | |
| }, | |
| { | |
| "epoch": 29.04, | |
| "learning_rate": 4.565602836879433e-06, | |
| "loss": 0.0499, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 29.06, | |
| "learning_rate": 4.5567375886524825e-06, | |
| "loss": 0.0858, | |
| "step": 1366 | |
| }, | |
| { | |
| "epoch": 29.09, | |
| "learning_rate": 4.5478723404255326e-06, | |
| "loss": 0.0853, | |
| "step": 1367 | |
| }, | |
| { | |
| "epoch": 29.11, | |
| "learning_rate": 4.539007092198582e-06, | |
| "loss": 0.0361, | |
| "step": 1368 | |
| }, | |
| { | |
| "epoch": 29.13, | |
| "learning_rate": 4.530141843971632e-06, | |
| "loss": 0.0441, | |
| "step": 1369 | |
| }, | |
| { | |
| "epoch": 29.15, | |
| "learning_rate": 4.521276595744681e-06, | |
| "loss": 0.2215, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 29.17, | |
| "learning_rate": 4.512411347517731e-06, | |
| "loss": 0.11, | |
| "step": 1371 | |
| }, | |
| { | |
| "epoch": 29.19, | |
| "learning_rate": 4.50354609929078e-06, | |
| "loss": 0.16, | |
| "step": 1372 | |
| }, | |
| { | |
| "epoch": 29.21, | |
| "learning_rate": 4.49468085106383e-06, | |
| "loss": 0.04, | |
| "step": 1373 | |
| }, | |
| { | |
| "epoch": 29.23, | |
| "learning_rate": 4.4858156028368795e-06, | |
| "loss": 0.0873, | |
| "step": 1374 | |
| }, | |
| { | |
| "epoch": 29.26, | |
| "learning_rate": 4.4769503546099296e-06, | |
| "loss": 0.2033, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 29.28, | |
| "learning_rate": 4.468085106382979e-06, | |
| "loss": 0.3217, | |
| "step": 1376 | |
| }, | |
| { | |
| "epoch": 29.3, | |
| "learning_rate": 4.459219858156029e-06, | |
| "loss": 0.143, | |
| "step": 1377 | |
| }, | |
| { | |
| "epoch": 29.32, | |
| "learning_rate": 4.450354609929078e-06, | |
| "loss": 0.0845, | |
| "step": 1378 | |
| }, | |
| { | |
| "epoch": 29.34, | |
| "learning_rate": 4.441489361702128e-06, | |
| "loss": 0.0656, | |
| "step": 1379 | |
| }, | |
| { | |
| "epoch": 29.36, | |
| "learning_rate": 4.432624113475177e-06, | |
| "loss": 0.1179, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 29.38, | |
| "learning_rate": 4.423758865248227e-06, | |
| "loss": 0.1502, | |
| "step": 1381 | |
| }, | |
| { | |
| "epoch": 29.4, | |
| "learning_rate": 4.4148936170212765e-06, | |
| "loss": 0.0664, | |
| "step": 1382 | |
| }, | |
| { | |
| "epoch": 29.43, | |
| "learning_rate": 4.4060283687943266e-06, | |
| "loss": 0.0914, | |
| "step": 1383 | |
| }, | |
| { | |
| "epoch": 29.45, | |
| "learning_rate": 4.397163120567376e-06, | |
| "loss": 0.0128, | |
| "step": 1384 | |
| }, | |
| { | |
| "epoch": 29.47, | |
| "learning_rate": 4.388297872340426e-06, | |
| "loss": 0.0647, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 29.49, | |
| "learning_rate": 4.379432624113475e-06, | |
| "loss": 0.0131, | |
| "step": 1386 | |
| }, | |
| { | |
| "epoch": 29.51, | |
| "learning_rate": 4.370567375886525e-06, | |
| "loss": 0.0549, | |
| "step": 1387 | |
| }, | |
| { | |
| "epoch": 29.53, | |
| "learning_rate": 4.361702127659574e-06, | |
| "loss": 0.0973, | |
| "step": 1388 | |
| }, | |
| { | |
| "epoch": 29.55, | |
| "learning_rate": 4.352836879432624e-06, | |
| "loss": 0.1432, | |
| "step": 1389 | |
| }, | |
| { | |
| "epoch": 29.57, | |
| "learning_rate": 4.3439716312056735e-06, | |
| "loss": 0.0981, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 29.6, | |
| "learning_rate": 4.3351063829787236e-06, | |
| "loss": 0.0645, | |
| "step": 1391 | |
| }, | |
| { | |
| "epoch": 29.62, | |
| "learning_rate": 4.326241134751774e-06, | |
| "loss": 0.1165, | |
| "step": 1392 | |
| }, | |
| { | |
| "epoch": 29.64, | |
| "learning_rate": 4.317375886524823e-06, | |
| "loss": 0.0702, | |
| "step": 1393 | |
| }, | |
| { | |
| "epoch": 29.66, | |
| "learning_rate": 4.308510638297873e-06, | |
| "loss": 0.1262, | |
| "step": 1394 | |
| }, | |
| { | |
| "epoch": 29.68, | |
| "learning_rate": 4.299645390070922e-06, | |
| "loss": 0.0236, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 29.7, | |
| "learning_rate": 4.290780141843972e-06, | |
| "loss": 0.1002, | |
| "step": 1396 | |
| }, | |
| { | |
| "epoch": 29.72, | |
| "learning_rate": 4.281914893617021e-06, | |
| "loss": 0.0965, | |
| "step": 1397 | |
| }, | |
| { | |
| "epoch": 29.74, | |
| "learning_rate": 4.273049645390071e-06, | |
| "loss": 0.0942, | |
| "step": 1398 | |
| }, | |
| { | |
| "epoch": 29.77, | |
| "learning_rate": 4.2641843971631206e-06, | |
| "loss": 0.1379, | |
| "step": 1399 | |
| }, | |
| { | |
| "epoch": 29.79, | |
| "learning_rate": 4.255319148936171e-06, | |
| "loss": 0.27, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 29.81, | |
| "learning_rate": 4.24645390070922e-06, | |
| "loss": 0.0278, | |
| "step": 1401 | |
| }, | |
| { | |
| "epoch": 29.83, | |
| "learning_rate": 4.23758865248227e-06, | |
| "loss": 0.0529, | |
| "step": 1402 | |
| }, | |
| { | |
| "epoch": 29.85, | |
| "learning_rate": 4.228723404255319e-06, | |
| "loss": 0.0813, | |
| "step": 1403 | |
| }, | |
| { | |
| "epoch": 29.87, | |
| "learning_rate": 4.219858156028369e-06, | |
| "loss": 0.0319, | |
| "step": 1404 | |
| }, | |
| { | |
| "epoch": 29.89, | |
| "learning_rate": 4.210992907801418e-06, | |
| "loss": 0.0125, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 29.91, | |
| "learning_rate": 4.202127659574468e-06, | |
| "loss": 0.0405, | |
| "step": 1406 | |
| }, | |
| { | |
| "epoch": 29.94, | |
| "learning_rate": 4.1932624113475176e-06, | |
| "loss": 0.0992, | |
| "step": 1407 | |
| }, | |
| { | |
| "epoch": 29.96, | |
| "learning_rate": 4.184397163120568e-06, | |
| "loss": 0.1068, | |
| "step": 1408 | |
| }, | |
| { | |
| "epoch": 29.98, | |
| "learning_rate": 4.175531914893617e-06, | |
| "loss": 0.1696, | |
| "step": 1409 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "learning_rate": 4.166666666666667e-06, | |
| "loss": 0.042, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "eval_accuracy": 0.9533333333333334, | |
| "eval_loss": 0.21241797506809235, | |
| "eval_runtime": 24.895, | |
| "eval_samples_per_second": 6.025, | |
| "eval_steps_per_second": 0.402, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 30.02, | |
| "learning_rate": 4.157801418439716e-06, | |
| "loss": 0.1096, | |
| "step": 1411 | |
| }, | |
| { | |
| "epoch": 30.04, | |
| "learning_rate": 4.148936170212766e-06, | |
| "loss": 0.0599, | |
| "step": 1412 | |
| }, | |
| { | |
| "epoch": 30.06, | |
| "learning_rate": 4.140070921985815e-06, | |
| "loss": 0.0318, | |
| "step": 1413 | |
| }, | |
| { | |
| "epoch": 30.09, | |
| "learning_rate": 4.131205673758865e-06, | |
| "loss": 0.089, | |
| "step": 1414 | |
| }, | |
| { | |
| "epoch": 30.11, | |
| "learning_rate": 4.1223404255319146e-06, | |
| "loss": 0.048, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 30.13, | |
| "learning_rate": 4.113475177304965e-06, | |
| "loss": 0.2738, | |
| "step": 1416 | |
| }, | |
| { | |
| "epoch": 30.15, | |
| "learning_rate": 4.104609929078014e-06, | |
| "loss": 0.0628, | |
| "step": 1417 | |
| }, | |
| { | |
| "epoch": 30.17, | |
| "learning_rate": 4.095744680851064e-06, | |
| "loss": 0.0444, | |
| "step": 1418 | |
| }, | |
| { | |
| "epoch": 30.19, | |
| "learning_rate": 4.086879432624113e-06, | |
| "loss": 0.118, | |
| "step": 1419 | |
| }, | |
| { | |
| "epoch": 30.21, | |
| "learning_rate": 4.078014184397163e-06, | |
| "loss": 0.1514, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 30.23, | |
| "learning_rate": 4.069148936170212e-06, | |
| "loss": 0.0415, | |
| "step": 1421 | |
| }, | |
| { | |
| "epoch": 30.26, | |
| "learning_rate": 4.060283687943262e-06, | |
| "loss": 0.099, | |
| "step": 1422 | |
| }, | |
| { | |
| "epoch": 30.28, | |
| "learning_rate": 4.0514184397163116e-06, | |
| "loss": 0.0484, | |
| "step": 1423 | |
| }, | |
| { | |
| "epoch": 30.3, | |
| "learning_rate": 4.042553191489362e-06, | |
| "loss": 0.1565, | |
| "step": 1424 | |
| }, | |
| { | |
| "epoch": 30.32, | |
| "learning_rate": 4.033687943262412e-06, | |
| "loss": 0.1495, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 30.34, | |
| "learning_rate": 4.024822695035461e-06, | |
| "loss": 0.1206, | |
| "step": 1426 | |
| }, | |
| { | |
| "epoch": 30.36, | |
| "learning_rate": 4.015957446808511e-06, | |
| "loss": 0.1405, | |
| "step": 1427 | |
| }, | |
| { | |
| "epoch": 30.38, | |
| "learning_rate": 4.00709219858156e-06, | |
| "loss": 0.1195, | |
| "step": 1428 | |
| }, | |
| { | |
| "epoch": 30.4, | |
| "learning_rate": 3.99822695035461e-06, | |
| "loss": 0.0926, | |
| "step": 1429 | |
| }, | |
| { | |
| "epoch": 30.43, | |
| "learning_rate": 3.989361702127659e-06, | |
| "loss": 0.0197, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 30.45, | |
| "learning_rate": 3.980496453900709e-06, | |
| "loss": 0.2193, | |
| "step": 1431 | |
| }, | |
| { | |
| "epoch": 30.47, | |
| "learning_rate": 3.971631205673759e-06, | |
| "loss": 0.1347, | |
| "step": 1432 | |
| }, | |
| { | |
| "epoch": 30.49, | |
| "learning_rate": 3.962765957446809e-06, | |
| "loss": 0.0587, | |
| "step": 1433 | |
| }, | |
| { | |
| "epoch": 30.51, | |
| "learning_rate": 3.953900709219858e-06, | |
| "loss": 0.1008, | |
| "step": 1434 | |
| }, | |
| { | |
| "epoch": 30.53, | |
| "learning_rate": 3.945035460992908e-06, | |
| "loss": 0.1743, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 30.55, | |
| "learning_rate": 3.936170212765957e-06, | |
| "loss": 0.0516, | |
| "step": 1436 | |
| }, | |
| { | |
| "epoch": 30.57, | |
| "learning_rate": 3.927304964539007e-06, | |
| "loss": 0.1391, | |
| "step": 1437 | |
| }, | |
| { | |
| "epoch": 30.6, | |
| "learning_rate": 3.918439716312056e-06, | |
| "loss": 0.0334, | |
| "step": 1438 | |
| }, | |
| { | |
| "epoch": 30.62, | |
| "learning_rate": 3.909574468085106e-06, | |
| "loss": 0.1353, | |
| "step": 1439 | |
| }, | |
| { | |
| "epoch": 30.64, | |
| "learning_rate": 3.900709219858156e-06, | |
| "loss": 0.0955, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 30.66, | |
| "learning_rate": 3.891843971631206e-06, | |
| "loss": 0.0462, | |
| "step": 1441 | |
| }, | |
| { | |
| "epoch": 30.68, | |
| "learning_rate": 3.882978723404255e-06, | |
| "loss": 0.0787, | |
| "step": 1442 | |
| }, | |
| { | |
| "epoch": 30.7, | |
| "learning_rate": 3.874113475177305e-06, | |
| "loss": 0.1467, | |
| "step": 1443 | |
| }, | |
| { | |
| "epoch": 30.72, | |
| "learning_rate": 3.865248226950354e-06, | |
| "loss": 0.1909, | |
| "step": 1444 | |
| }, | |
| { | |
| "epoch": 30.74, | |
| "learning_rate": 3.856382978723404e-06, | |
| "loss": 0.0564, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 30.77, | |
| "learning_rate": 3.847517730496453e-06, | |
| "loss": 0.1135, | |
| "step": 1446 | |
| }, | |
| { | |
| "epoch": 30.79, | |
| "learning_rate": 3.838652482269503e-06, | |
| "loss": 0.1333, | |
| "step": 1447 | |
| }, | |
| { | |
| "epoch": 30.81, | |
| "learning_rate": 3.829787234042553e-06, | |
| "loss": 0.0705, | |
| "step": 1448 | |
| }, | |
| { | |
| "epoch": 30.83, | |
| "learning_rate": 3.820921985815603e-06, | |
| "loss": 0.1499, | |
| "step": 1449 | |
| }, | |
| { | |
| "epoch": 30.85, | |
| "learning_rate": 3.8120567375886523e-06, | |
| "loss": 0.1633, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 30.87, | |
| "learning_rate": 3.8031914893617023e-06, | |
| "loss": 0.0913, | |
| "step": 1451 | |
| }, | |
| { | |
| "epoch": 30.89, | |
| "learning_rate": 3.7943262411347515e-06, | |
| "loss": 0.1711, | |
| "step": 1452 | |
| }, | |
| { | |
| "epoch": 30.91, | |
| "learning_rate": 3.7854609929078016e-06, | |
| "loss": 0.133, | |
| "step": 1453 | |
| }, | |
| { | |
| "epoch": 30.94, | |
| "learning_rate": 3.776595744680851e-06, | |
| "loss": 0.1334, | |
| "step": 1454 | |
| }, | |
| { | |
| "epoch": 30.96, | |
| "learning_rate": 3.767730496453901e-06, | |
| "loss": 0.1475, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 30.98, | |
| "learning_rate": 3.75886524822695e-06, | |
| "loss": 0.0488, | |
| "step": 1456 | |
| }, | |
| { | |
| "epoch": 31.0, | |
| "learning_rate": 3.75e-06, | |
| "loss": 0.0841, | |
| "step": 1457 | |
| }, | |
| { | |
| "epoch": 31.0, | |
| "eval_accuracy": 0.94, | |
| "eval_loss": 0.21738062798976898, | |
| "eval_runtime": 25.8747, | |
| "eval_samples_per_second": 5.797, | |
| "eval_steps_per_second": 0.386, | |
| "step": 1457 | |
| }, | |
| { | |
| "epoch": 31.02, | |
| "learning_rate": 3.7411347517730497e-06, | |
| "loss": 0.0664, | |
| "step": 1458 | |
| }, | |
| { | |
| "epoch": 31.04, | |
| "learning_rate": 3.7322695035460993e-06, | |
| "loss": 0.1141, | |
| "step": 1459 | |
| }, | |
| { | |
| "epoch": 31.06, | |
| "learning_rate": 3.723404255319149e-06, | |
| "loss": 0.0353, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 31.09, | |
| "learning_rate": 3.7145390070921986e-06, | |
| "loss": 0.1147, | |
| "step": 1461 | |
| }, | |
| { | |
| "epoch": 31.11, | |
| "learning_rate": 3.7056737588652482e-06, | |
| "loss": 0.0726, | |
| "step": 1462 | |
| }, | |
| { | |
| "epoch": 31.13, | |
| "learning_rate": 3.696808510638298e-06, | |
| "loss": 0.0422, | |
| "step": 1463 | |
| }, | |
| { | |
| "epoch": 31.15, | |
| "learning_rate": 3.6879432624113475e-06, | |
| "loss": 0.0441, | |
| "step": 1464 | |
| }, | |
| { | |
| "epoch": 31.17, | |
| "learning_rate": 3.679078014184397e-06, | |
| "loss": 0.1242, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 31.19, | |
| "learning_rate": 3.6702127659574467e-06, | |
| "loss": 0.0626, | |
| "step": 1466 | |
| }, | |
| { | |
| "epoch": 31.21, | |
| "learning_rate": 3.6613475177304963e-06, | |
| "loss": 0.0856, | |
| "step": 1467 | |
| }, | |
| { | |
| "epoch": 31.23, | |
| "learning_rate": 3.652482269503546e-06, | |
| "loss": 0.1611, | |
| "step": 1468 | |
| }, | |
| { | |
| "epoch": 31.26, | |
| "learning_rate": 3.6436170212765956e-06, | |
| "loss": 0.1375, | |
| "step": 1469 | |
| }, | |
| { | |
| "epoch": 31.28, | |
| "learning_rate": 3.6347517730496452e-06, | |
| "loss": 0.2441, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 31.3, | |
| "learning_rate": 3.625886524822695e-06, | |
| "loss": 0.1371, | |
| "step": 1471 | |
| }, | |
| { | |
| "epoch": 31.32, | |
| "learning_rate": 3.6170212765957445e-06, | |
| "loss": 0.07, | |
| "step": 1472 | |
| }, | |
| { | |
| "epoch": 31.34, | |
| "learning_rate": 3.608156028368794e-06, | |
| "loss": 0.1211, | |
| "step": 1473 | |
| }, | |
| { | |
| "epoch": 31.36, | |
| "learning_rate": 3.599290780141844e-06, | |
| "loss": 0.2036, | |
| "step": 1474 | |
| }, | |
| { | |
| "epoch": 31.38, | |
| "learning_rate": 3.5904255319148938e-06, | |
| "loss": 0.0995, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 31.4, | |
| "learning_rate": 3.5815602836879434e-06, | |
| "loss": 0.0771, | |
| "step": 1476 | |
| }, | |
| { | |
| "epoch": 31.43, | |
| "learning_rate": 3.572695035460993e-06, | |
| "loss": 0.1251, | |
| "step": 1477 | |
| }, | |
| { | |
| "epoch": 31.45, | |
| "learning_rate": 3.5638297872340426e-06, | |
| "loss": 0.1222, | |
| "step": 1478 | |
| }, | |
| { | |
| "epoch": 31.47, | |
| "learning_rate": 3.5549645390070923e-06, | |
| "loss": 0.1565, | |
| "step": 1479 | |
| }, | |
| { | |
| "epoch": 31.49, | |
| "learning_rate": 3.546099290780142e-06, | |
| "loss": 0.1217, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 31.51, | |
| "learning_rate": 3.5372340425531915e-06, | |
| "loss": 0.0987, | |
| "step": 1481 | |
| }, | |
| { | |
| "epoch": 31.53, | |
| "learning_rate": 3.528368794326241e-06, | |
| "loss": 0.0806, | |
| "step": 1482 | |
| }, | |
| { | |
| "epoch": 31.55, | |
| "learning_rate": 3.5195035460992908e-06, | |
| "loss": 0.0882, | |
| "step": 1483 | |
| }, | |
| { | |
| "epoch": 31.57, | |
| "learning_rate": 3.5106382978723404e-06, | |
| "loss": 0.1107, | |
| "step": 1484 | |
| }, | |
| { | |
| "epoch": 31.6, | |
| "learning_rate": 3.50177304964539e-06, | |
| "loss": 0.1911, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 31.62, | |
| "learning_rate": 3.4929078014184396e-06, | |
| "loss": 0.048, | |
| "step": 1486 | |
| }, | |
| { | |
| "epoch": 31.64, | |
| "learning_rate": 3.4840425531914893e-06, | |
| "loss": 0.0313, | |
| "step": 1487 | |
| }, | |
| { | |
| "epoch": 31.66, | |
| "learning_rate": 3.4751773049645393e-06, | |
| "loss": 0.0384, | |
| "step": 1488 | |
| }, | |
| { | |
| "epoch": 31.68, | |
| "learning_rate": 3.466312056737589e-06, | |
| "loss": 0.0721, | |
| "step": 1489 | |
| }, | |
| { | |
| "epoch": 31.7, | |
| "learning_rate": 3.4574468085106386e-06, | |
| "loss": 0.1006, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 31.72, | |
| "learning_rate": 3.448581560283688e-06, | |
| "loss": 0.0557, | |
| "step": 1491 | |
| }, | |
| { | |
| "epoch": 31.74, | |
| "learning_rate": 3.439716312056738e-06, | |
| "loss": 0.0741, | |
| "step": 1492 | |
| }, | |
| { | |
| "epoch": 31.77, | |
| "learning_rate": 3.4308510638297874e-06, | |
| "loss": 0.1769, | |
| "step": 1493 | |
| }, | |
| { | |
| "epoch": 31.79, | |
| "learning_rate": 3.421985815602837e-06, | |
| "loss": 0.0727, | |
| "step": 1494 | |
| }, | |
| { | |
| "epoch": 31.81, | |
| "learning_rate": 3.4131205673758867e-06, | |
| "loss": 0.0057, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 31.83, | |
| "learning_rate": 3.4042553191489363e-06, | |
| "loss": 0.1173, | |
| "step": 1496 | |
| }, | |
| { | |
| "epoch": 31.85, | |
| "learning_rate": 3.395390070921986e-06, | |
| "loss": 0.0231, | |
| "step": 1497 | |
| }, | |
| { | |
| "epoch": 31.87, | |
| "learning_rate": 3.3865248226950356e-06, | |
| "loss": 0.0421, | |
| "step": 1498 | |
| }, | |
| { | |
| "epoch": 31.89, | |
| "learning_rate": 3.377659574468085e-06, | |
| "loss": 0.1633, | |
| "step": 1499 | |
| }, | |
| { | |
| "epoch": 31.91, | |
| "learning_rate": 3.368794326241135e-06, | |
| "loss": 0.0855, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 31.94, | |
| "learning_rate": 3.3599290780141844e-06, | |
| "loss": 0.0353, | |
| "step": 1501 | |
| }, | |
| { | |
| "epoch": 31.96, | |
| "learning_rate": 3.351063829787234e-06, | |
| "loss": 0.0322, | |
| "step": 1502 | |
| }, | |
| { | |
| "epoch": 31.98, | |
| "learning_rate": 3.3421985815602837e-06, | |
| "loss": 0.0854, | |
| "step": 1503 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "learning_rate": 3.3333333333333333e-06, | |
| "loss": 0.046, | |
| "step": 1504 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "eval_accuracy": 0.9466666666666667, | |
| "eval_loss": 0.21360725164413452, | |
| "eval_runtime": 25.1101, | |
| "eval_samples_per_second": 5.974, | |
| "eval_steps_per_second": 0.398, | |
| "step": 1504 | |
| }, | |
| { | |
| "epoch": 32.02, | |
| "learning_rate": 3.324468085106383e-06, | |
| "loss": 0.0687, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 32.04, | |
| "learning_rate": 3.315602836879433e-06, | |
| "loss": 0.2089, | |
| "step": 1506 | |
| }, | |
| { | |
| "epoch": 32.06, | |
| "learning_rate": 3.3067375886524826e-06, | |
| "loss": 0.0812, | |
| "step": 1507 | |
| }, | |
| { | |
| "epoch": 32.09, | |
| "learning_rate": 3.2978723404255322e-06, | |
| "loss": 0.0519, | |
| "step": 1508 | |
| }, | |
| { | |
| "epoch": 32.11, | |
| "learning_rate": 3.289007092198582e-06, | |
| "loss": 0.0688, | |
| "step": 1509 | |
| }, | |
| { | |
| "epoch": 32.13, | |
| "learning_rate": 3.2801418439716315e-06, | |
| "loss": 0.0278, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 32.15, | |
| "learning_rate": 3.271276595744681e-06, | |
| "loss": 0.0498, | |
| "step": 1511 | |
| }, | |
| { | |
| "epoch": 32.17, | |
| "learning_rate": 3.2624113475177307e-06, | |
| "loss": 0.0746, | |
| "step": 1512 | |
| }, | |
| { | |
| "epoch": 32.19, | |
| "learning_rate": 3.2535460992907804e-06, | |
| "loss": 0.0767, | |
| "step": 1513 | |
| }, | |
| { | |
| "epoch": 32.21, | |
| "learning_rate": 3.24468085106383e-06, | |
| "loss": 0.0501, | |
| "step": 1514 | |
| }, | |
| { | |
| "epoch": 32.23, | |
| "learning_rate": 3.2358156028368796e-06, | |
| "loss": 0.2516, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 32.26, | |
| "learning_rate": 3.2269503546099292e-06, | |
| "loss": 0.1483, | |
| "step": 1516 | |
| }, | |
| { | |
| "epoch": 32.28, | |
| "learning_rate": 3.218085106382979e-06, | |
| "loss": 0.0206, | |
| "step": 1517 | |
| }, | |
| { | |
| "epoch": 32.3, | |
| "learning_rate": 3.2092198581560285e-06, | |
| "loss": 0.0938, | |
| "step": 1518 | |
| }, | |
| { | |
| "epoch": 32.32, | |
| "learning_rate": 3.200354609929078e-06, | |
| "loss": 0.1567, | |
| "step": 1519 | |
| }, | |
| { | |
| "epoch": 32.34, | |
| "learning_rate": 3.1914893617021277e-06, | |
| "loss": 0.0868, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 32.36, | |
| "learning_rate": 3.1826241134751774e-06, | |
| "loss": 0.0506, | |
| "step": 1521 | |
| }, | |
| { | |
| "epoch": 32.38, | |
| "learning_rate": 3.173758865248227e-06, | |
| "loss": 0.1771, | |
| "step": 1522 | |
| }, | |
| { | |
| "epoch": 32.4, | |
| "learning_rate": 3.1648936170212766e-06, | |
| "loss": 0.1371, | |
| "step": 1523 | |
| }, | |
| { | |
| "epoch": 32.43, | |
| "learning_rate": 3.1560283687943263e-06, | |
| "loss": 0.1133, | |
| "step": 1524 | |
| }, | |
| { | |
| "epoch": 32.45, | |
| "learning_rate": 3.147163120567376e-06, | |
| "loss": 0.1236, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 32.47, | |
| "learning_rate": 3.1382978723404255e-06, | |
| "loss": 0.1276, | |
| "step": 1526 | |
| }, | |
| { | |
| "epoch": 32.49, | |
| "learning_rate": 3.129432624113475e-06, | |
| "loss": 0.0744, | |
| "step": 1527 | |
| }, | |
| { | |
| "epoch": 32.51, | |
| "learning_rate": 3.1205673758865248e-06, | |
| "loss": 0.1463, | |
| "step": 1528 | |
| }, | |
| { | |
| "epoch": 32.53, | |
| "learning_rate": 3.1117021276595744e-06, | |
| "loss": 0.0959, | |
| "step": 1529 | |
| }, | |
| { | |
| "epoch": 32.55, | |
| "learning_rate": 3.102836879432624e-06, | |
| "loss": 0.0573, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 32.57, | |
| "learning_rate": 3.0939716312056736e-06, | |
| "loss": 0.0558, | |
| "step": 1531 | |
| }, | |
| { | |
| "epoch": 32.6, | |
| "learning_rate": 3.0851063829787233e-06, | |
| "loss": 0.1299, | |
| "step": 1532 | |
| }, | |
| { | |
| "epoch": 32.62, | |
| "learning_rate": 3.076241134751773e-06, | |
| "loss": 0.0473, | |
| "step": 1533 | |
| }, | |
| { | |
| "epoch": 32.64, | |
| "learning_rate": 3.0673758865248225e-06, | |
| "loss": 0.0549, | |
| "step": 1534 | |
| }, | |
| { | |
| "epoch": 32.66, | |
| "learning_rate": 3.058510638297872e-06, | |
| "loss": 0.0473, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 32.68, | |
| "learning_rate": 3.0496453900709218e-06, | |
| "loss": 0.0846, | |
| "step": 1536 | |
| }, | |
| { | |
| "epoch": 32.7, | |
| "learning_rate": 3.0407801418439714e-06, | |
| "loss": 0.1144, | |
| "step": 1537 | |
| }, | |
| { | |
| "epoch": 32.72, | |
| "learning_rate": 3.031914893617021e-06, | |
| "loss": 0.1049, | |
| "step": 1538 | |
| }, | |
| { | |
| "epoch": 32.74, | |
| "learning_rate": 3.023049645390071e-06, | |
| "loss": 0.088, | |
| "step": 1539 | |
| }, | |
| { | |
| "epoch": 32.77, | |
| "learning_rate": 3.0141843971631207e-06, | |
| "loss": 0.0717, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 32.79, | |
| "learning_rate": 3.0053191489361703e-06, | |
| "loss": 0.0461, | |
| "step": 1541 | |
| }, | |
| { | |
| "epoch": 32.81, | |
| "learning_rate": 2.99645390070922e-06, | |
| "loss": 0.1567, | |
| "step": 1542 | |
| }, | |
| { | |
| "epoch": 32.83, | |
| "learning_rate": 2.9875886524822696e-06, | |
| "loss": 0.1877, | |
| "step": 1543 | |
| }, | |
| { | |
| "epoch": 32.85, | |
| "learning_rate": 2.978723404255319e-06, | |
| "loss": 0.0267, | |
| "step": 1544 | |
| }, | |
| { | |
| "epoch": 32.87, | |
| "learning_rate": 2.969858156028369e-06, | |
| "loss": 0.0864, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 32.89, | |
| "learning_rate": 2.9609929078014184e-06, | |
| "loss": 0.1528, | |
| "step": 1546 | |
| }, | |
| { | |
| "epoch": 32.91, | |
| "learning_rate": 2.952127659574468e-06, | |
| "loss": 0.1722, | |
| "step": 1547 | |
| }, | |
| { | |
| "epoch": 32.94, | |
| "learning_rate": 2.943262411347518e-06, | |
| "loss": 0.054, | |
| "step": 1548 | |
| }, | |
| { | |
| "epoch": 32.96, | |
| "learning_rate": 2.9343971631205677e-06, | |
| "loss": 0.0905, | |
| "step": 1549 | |
| }, | |
| { | |
| "epoch": 32.98, | |
| "learning_rate": 2.9255319148936174e-06, | |
| "loss": 0.0576, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 33.0, | |
| "learning_rate": 2.916666666666667e-06, | |
| "loss": 0.1309, | |
| "step": 1551 | |
| }, | |
| { | |
| "epoch": 33.0, | |
| "eval_accuracy": 0.96, | |
| "eval_loss": 0.19805511832237244, | |
| "eval_runtime": 25.1333, | |
| "eval_samples_per_second": 5.968, | |
| "eval_steps_per_second": 0.398, | |
| "step": 1551 | |
| }, | |
| { | |
| "epoch": 33.02, | |
| "learning_rate": 2.9078014184397166e-06, | |
| "loss": 0.0531, | |
| "step": 1552 | |
| }, | |
| { | |
| "epoch": 33.04, | |
| "learning_rate": 2.8989361702127662e-06, | |
| "loss": 0.0631, | |
| "step": 1553 | |
| }, | |
| { | |
| "epoch": 33.06, | |
| "learning_rate": 2.890070921985816e-06, | |
| "loss": 0.0427, | |
| "step": 1554 | |
| }, | |
| { | |
| "epoch": 33.09, | |
| "learning_rate": 2.8812056737588655e-06, | |
| "loss": 0.0778, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 33.11, | |
| "learning_rate": 2.872340425531915e-06, | |
| "loss": 0.0829, | |
| "step": 1556 | |
| }, | |
| { | |
| "epoch": 33.13, | |
| "learning_rate": 2.8634751773049647e-06, | |
| "loss": 0.0889, | |
| "step": 1557 | |
| }, | |
| { | |
| "epoch": 33.15, | |
| "learning_rate": 2.8546099290780144e-06, | |
| "loss": 0.0416, | |
| "step": 1558 | |
| }, | |
| { | |
| "epoch": 33.17, | |
| "learning_rate": 2.845744680851064e-06, | |
| "loss": 0.0386, | |
| "step": 1559 | |
| }, | |
| { | |
| "epoch": 33.19, | |
| "learning_rate": 2.8368794326241136e-06, | |
| "loss": 0.0282, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 33.21, | |
| "learning_rate": 2.8280141843971632e-06, | |
| "loss": 0.0201, | |
| "step": 1561 | |
| }, | |
| { | |
| "epoch": 33.23, | |
| "learning_rate": 2.819148936170213e-06, | |
| "loss": 0.0299, | |
| "step": 1562 | |
| }, | |
| { | |
| "epoch": 33.26, | |
| "learning_rate": 2.8102836879432625e-06, | |
| "loss": 0.1303, | |
| "step": 1563 | |
| }, | |
| { | |
| "epoch": 33.28, | |
| "learning_rate": 2.801418439716312e-06, | |
| "loss": 0.0738, | |
| "step": 1564 | |
| }, | |
| { | |
| "epoch": 33.3, | |
| "learning_rate": 2.7925531914893617e-06, | |
| "loss": 0.0461, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 33.32, | |
| "learning_rate": 2.7836879432624114e-06, | |
| "loss": 0.074, | |
| "step": 1566 | |
| }, | |
| { | |
| "epoch": 33.34, | |
| "learning_rate": 2.774822695035461e-06, | |
| "loss": 0.1712, | |
| "step": 1567 | |
| }, | |
| { | |
| "epoch": 33.36, | |
| "learning_rate": 2.7659574468085106e-06, | |
| "loss": 0.1368, | |
| "step": 1568 | |
| }, | |
| { | |
| "epoch": 33.38, | |
| "learning_rate": 2.7570921985815602e-06, | |
| "loss": 0.1132, | |
| "step": 1569 | |
| }, | |
| { | |
| "epoch": 33.4, | |
| "learning_rate": 2.74822695035461e-06, | |
| "loss": 0.0383, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 33.43, | |
| "learning_rate": 2.73936170212766e-06, | |
| "loss": 0.019, | |
| "step": 1571 | |
| }, | |
| { | |
| "epoch": 33.45, | |
| "learning_rate": 2.7304964539007095e-06, | |
| "loss": 0.0532, | |
| "step": 1572 | |
| }, | |
| { | |
| "epoch": 33.47, | |
| "learning_rate": 2.721631205673759e-06, | |
| "loss": 0.0973, | |
| "step": 1573 | |
| }, | |
| { | |
| "epoch": 33.49, | |
| "learning_rate": 2.7127659574468088e-06, | |
| "loss": 0.055, | |
| "step": 1574 | |
| }, | |
| { | |
| "epoch": 33.51, | |
| "learning_rate": 2.7039007092198584e-06, | |
| "loss": 0.0862, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 33.53, | |
| "learning_rate": 2.695035460992908e-06, | |
| "loss": 0.0797, | |
| "step": 1576 | |
| }, | |
| { | |
| "epoch": 33.55, | |
| "learning_rate": 2.6861702127659577e-06, | |
| "loss": 0.1177, | |
| "step": 1577 | |
| }, | |
| { | |
| "epoch": 33.57, | |
| "learning_rate": 2.6773049645390073e-06, | |
| "loss": 0.1177, | |
| "step": 1578 | |
| }, | |
| { | |
| "epoch": 33.6, | |
| "learning_rate": 2.668439716312057e-06, | |
| "loss": 0.1047, | |
| "step": 1579 | |
| }, | |
| { | |
| "epoch": 33.62, | |
| "learning_rate": 2.6595744680851065e-06, | |
| "loss": 0.1341, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 33.64, | |
| "learning_rate": 2.650709219858156e-06, | |
| "loss": 0.0862, | |
| "step": 1581 | |
| }, | |
| { | |
| "epoch": 33.66, | |
| "learning_rate": 2.6418439716312058e-06, | |
| "loss": 0.1746, | |
| "step": 1582 | |
| }, | |
| { | |
| "epoch": 33.68, | |
| "learning_rate": 2.6329787234042554e-06, | |
| "loss": 0.0576, | |
| "step": 1583 | |
| }, | |
| { | |
| "epoch": 33.7, | |
| "learning_rate": 2.624113475177305e-06, | |
| "loss": 0.0415, | |
| "step": 1584 | |
| }, | |
| { | |
| "epoch": 33.72, | |
| "learning_rate": 2.6152482269503547e-06, | |
| "loss": 0.154, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 33.74, | |
| "learning_rate": 2.6063829787234043e-06, | |
| "loss": 0.065, | |
| "step": 1586 | |
| }, | |
| { | |
| "epoch": 33.77, | |
| "learning_rate": 2.597517730496454e-06, | |
| "loss": 0.0943, | |
| "step": 1587 | |
| }, | |
| { | |
| "epoch": 33.79, | |
| "learning_rate": 2.5886524822695035e-06, | |
| "loss": 0.126, | |
| "step": 1588 | |
| }, | |
| { | |
| "epoch": 33.81, | |
| "learning_rate": 2.579787234042553e-06, | |
| "loss": 0.0141, | |
| "step": 1589 | |
| }, | |
| { | |
| "epoch": 33.83, | |
| "learning_rate": 2.5709219858156028e-06, | |
| "loss": 0.082, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 33.85, | |
| "learning_rate": 2.5620567375886524e-06, | |
| "loss": 0.0736, | |
| "step": 1591 | |
| }, | |
| { | |
| "epoch": 33.87, | |
| "learning_rate": 2.553191489361702e-06, | |
| "loss": 0.1645, | |
| "step": 1592 | |
| }, | |
| { | |
| "epoch": 33.89, | |
| "learning_rate": 2.5443262411347517e-06, | |
| "loss": 0.0474, | |
| "step": 1593 | |
| }, | |
| { | |
| "epoch": 33.91, | |
| "learning_rate": 2.5354609929078013e-06, | |
| "loss": 0.0855, | |
| "step": 1594 | |
| }, | |
| { | |
| "epoch": 33.94, | |
| "learning_rate": 2.526595744680851e-06, | |
| "loss": 0.0796, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 33.96, | |
| "learning_rate": 2.5177304964539005e-06, | |
| "loss": 0.0348, | |
| "step": 1596 | |
| }, | |
| { | |
| "epoch": 33.98, | |
| "learning_rate": 2.50886524822695e-06, | |
| "loss": 0.0807, | |
| "step": 1597 | |
| }, | |
| { | |
| "epoch": 34.0, | |
| "learning_rate": 2.4999999999999998e-06, | |
| "loss": 0.1207, | |
| "step": 1598 | |
| }, | |
| { | |
| "epoch": 34.0, | |
| "eval_accuracy": 0.94, | |
| "eval_loss": 0.2333659827709198, | |
| "eval_runtime": 29.1238, | |
| "eval_samples_per_second": 5.15, | |
| "eval_steps_per_second": 0.343, | |
| "step": 1598 | |
| }, | |
| { | |
| "epoch": 34.02, | |
| "learning_rate": 2.4911347517730494e-06, | |
| "loss": 0.0307, | |
| "step": 1599 | |
| }, | |
| { | |
| "epoch": 34.04, | |
| "learning_rate": 2.482269503546099e-06, | |
| "loss": 0.0486, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 34.06, | |
| "learning_rate": 2.4734042553191487e-06, | |
| "loss": 0.1469, | |
| "step": 1601 | |
| }, | |
| { | |
| "epoch": 34.09, | |
| "learning_rate": 2.4645390070921983e-06, | |
| "loss": 0.0683, | |
| "step": 1602 | |
| }, | |
| { | |
| "epoch": 34.11, | |
| "learning_rate": 2.455673758865248e-06, | |
| "loss": 0.0456, | |
| "step": 1603 | |
| }, | |
| { | |
| "epoch": 34.13, | |
| "learning_rate": 2.446808510638298e-06, | |
| "loss": 0.0481, | |
| "step": 1604 | |
| }, | |
| { | |
| "epoch": 34.15, | |
| "learning_rate": 2.4379432624113476e-06, | |
| "loss": 0.0878, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 34.17, | |
| "learning_rate": 2.429078014184397e-06, | |
| "loss": 0.0619, | |
| "step": 1606 | |
| }, | |
| { | |
| "epoch": 34.19, | |
| "learning_rate": 2.420212765957447e-06, | |
| "loss": 0.0872, | |
| "step": 1607 | |
| }, | |
| { | |
| "epoch": 34.21, | |
| "learning_rate": 2.4113475177304965e-06, | |
| "loss": 0.1275, | |
| "step": 1608 | |
| }, | |
| { | |
| "epoch": 34.23, | |
| "learning_rate": 2.4024822695035465e-06, | |
| "loss": 0.0803, | |
| "step": 1609 | |
| }, | |
| { | |
| "epoch": 34.26, | |
| "learning_rate": 2.393617021276596e-06, | |
| "loss": 0.095, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 34.28, | |
| "learning_rate": 2.3847517730496458e-06, | |
| "loss": 0.0251, | |
| "step": 1611 | |
| }, | |
| { | |
| "epoch": 34.3, | |
| "learning_rate": 2.3758865248226954e-06, | |
| "loss": 0.2184, | |
| "step": 1612 | |
| }, | |
| { | |
| "epoch": 34.32, | |
| "learning_rate": 2.367021276595745e-06, | |
| "loss": 0.0209, | |
| "step": 1613 | |
| }, | |
| { | |
| "epoch": 34.34, | |
| "learning_rate": 2.3581560283687946e-06, | |
| "loss": 0.0397, | |
| "step": 1614 | |
| }, | |
| { | |
| "epoch": 34.36, | |
| "learning_rate": 2.3492907801418443e-06, | |
| "loss": 0.0804, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 34.38, | |
| "learning_rate": 2.340425531914894e-06, | |
| "loss": 0.0868, | |
| "step": 1616 | |
| }, | |
| { | |
| "epoch": 34.4, | |
| "learning_rate": 2.3315602836879435e-06, | |
| "loss": 0.0615, | |
| "step": 1617 | |
| }, | |
| { | |
| "epoch": 34.43, | |
| "learning_rate": 2.322695035460993e-06, | |
| "loss": 0.1223, | |
| "step": 1618 | |
| }, | |
| { | |
| "epoch": 34.45, | |
| "learning_rate": 2.3138297872340428e-06, | |
| "loss": 0.0914, | |
| "step": 1619 | |
| }, | |
| { | |
| "epoch": 34.47, | |
| "learning_rate": 2.3049645390070924e-06, | |
| "loss": 0.0738, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 34.49, | |
| "learning_rate": 2.296099290780142e-06, | |
| "loss": 0.0971, | |
| "step": 1621 | |
| }, | |
| { | |
| "epoch": 34.51, | |
| "learning_rate": 2.2872340425531916e-06, | |
| "loss": 0.1186, | |
| "step": 1622 | |
| }, | |
| { | |
| "epoch": 34.53, | |
| "learning_rate": 2.2783687943262413e-06, | |
| "loss": 0.0699, | |
| "step": 1623 | |
| }, | |
| { | |
| "epoch": 34.55, | |
| "learning_rate": 2.269503546099291e-06, | |
| "loss": 0.0873, | |
| "step": 1624 | |
| }, | |
| { | |
| "epoch": 34.57, | |
| "learning_rate": 2.2606382978723405e-06, | |
| "loss": 0.1232, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 34.6, | |
| "learning_rate": 2.25177304964539e-06, | |
| "loss": 0.068, | |
| "step": 1626 | |
| }, | |
| { | |
| "epoch": 34.62, | |
| "learning_rate": 2.2429078014184398e-06, | |
| "loss": 0.0647, | |
| "step": 1627 | |
| }, | |
| { | |
| "epoch": 34.64, | |
| "learning_rate": 2.2340425531914894e-06, | |
| "loss": 0.0732, | |
| "step": 1628 | |
| }, | |
| { | |
| "epoch": 34.66, | |
| "learning_rate": 2.225177304964539e-06, | |
| "loss": 0.12, | |
| "step": 1629 | |
| }, | |
| { | |
| "epoch": 34.68, | |
| "learning_rate": 2.2163120567375886e-06, | |
| "loss": 0.0824, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 34.7, | |
| "learning_rate": 2.2074468085106383e-06, | |
| "loss": 0.1158, | |
| "step": 1631 | |
| }, | |
| { | |
| "epoch": 34.72, | |
| "learning_rate": 2.198581560283688e-06, | |
| "loss": 0.0284, | |
| "step": 1632 | |
| }, | |
| { | |
| "epoch": 34.74, | |
| "learning_rate": 2.1897163120567375e-06, | |
| "loss": 0.0685, | |
| "step": 1633 | |
| }, | |
| { | |
| "epoch": 34.77, | |
| "learning_rate": 2.180851063829787e-06, | |
| "loss": 0.2221, | |
| "step": 1634 | |
| }, | |
| { | |
| "epoch": 34.79, | |
| "learning_rate": 2.1719858156028368e-06, | |
| "loss": 0.096, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 34.81, | |
| "learning_rate": 2.163120567375887e-06, | |
| "loss": 0.1044, | |
| "step": 1636 | |
| }, | |
| { | |
| "epoch": 34.83, | |
| "learning_rate": 2.1542553191489364e-06, | |
| "loss": 0.0807, | |
| "step": 1637 | |
| }, | |
| { | |
| "epoch": 34.85, | |
| "learning_rate": 2.145390070921986e-06, | |
| "loss": 0.1314, | |
| "step": 1638 | |
| }, | |
| { | |
| "epoch": 34.87, | |
| "learning_rate": 2.1365248226950357e-06, | |
| "loss": 0.0367, | |
| "step": 1639 | |
| }, | |
| { | |
| "epoch": 34.89, | |
| "learning_rate": 2.1276595744680853e-06, | |
| "loss": 0.0181, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 34.91, | |
| "learning_rate": 2.118794326241135e-06, | |
| "loss": 0.1259, | |
| "step": 1641 | |
| }, | |
| { | |
| "epoch": 34.94, | |
| "learning_rate": 2.1099290780141846e-06, | |
| "loss": 0.1429, | |
| "step": 1642 | |
| }, | |
| { | |
| "epoch": 34.96, | |
| "learning_rate": 2.101063829787234e-06, | |
| "loss": 0.0901, | |
| "step": 1643 | |
| }, | |
| { | |
| "epoch": 34.98, | |
| "learning_rate": 2.092198581560284e-06, | |
| "loss": 0.0922, | |
| "step": 1644 | |
| }, | |
| { | |
| "epoch": 35.0, | |
| "learning_rate": 2.0833333333333334e-06, | |
| "loss": 0.1216, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 35.0, | |
| "eval_accuracy": 0.94, | |
| "eval_loss": 0.22382324934005737, | |
| "eval_runtime": 28.116, | |
| "eval_samples_per_second": 5.335, | |
| "eval_steps_per_second": 0.356, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 35.02, | |
| "learning_rate": 2.074468085106383e-06, | |
| "loss": 0.0721, | |
| "step": 1646 | |
| }, | |
| { | |
| "epoch": 35.04, | |
| "learning_rate": 2.0656028368794327e-06, | |
| "loss": 0.0281, | |
| "step": 1647 | |
| }, | |
| { | |
| "epoch": 35.06, | |
| "learning_rate": 2.0567375886524823e-06, | |
| "loss": 0.0774, | |
| "step": 1648 | |
| }, | |
| { | |
| "epoch": 35.09, | |
| "learning_rate": 2.047872340425532e-06, | |
| "loss": 0.2029, | |
| "step": 1649 | |
| }, | |
| { | |
| "epoch": 35.11, | |
| "learning_rate": 2.0390070921985816e-06, | |
| "loss": 0.0738, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 35.13, | |
| "learning_rate": 2.030141843971631e-06, | |
| "loss": 0.0871, | |
| "step": 1651 | |
| }, | |
| { | |
| "epoch": 35.15, | |
| "learning_rate": 2.021276595744681e-06, | |
| "loss": 0.0792, | |
| "step": 1652 | |
| }, | |
| { | |
| "epoch": 35.17, | |
| "learning_rate": 2.0124113475177304e-06, | |
| "loss": 0.0803, | |
| "step": 1653 | |
| }, | |
| { | |
| "epoch": 35.19, | |
| "learning_rate": 2.00354609929078e-06, | |
| "loss": 0.0468, | |
| "step": 1654 | |
| }, | |
| { | |
| "epoch": 35.21, | |
| "learning_rate": 1.9946808510638297e-06, | |
| "loss": 0.0937, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 35.23, | |
| "learning_rate": 1.9858156028368793e-06, | |
| "loss": 0.0399, | |
| "step": 1656 | |
| }, | |
| { | |
| "epoch": 35.26, | |
| "learning_rate": 1.976950354609929e-06, | |
| "loss": 0.1084, | |
| "step": 1657 | |
| }, | |
| { | |
| "epoch": 35.28, | |
| "learning_rate": 1.9680851063829786e-06, | |
| "loss": 0.0878, | |
| "step": 1658 | |
| }, | |
| { | |
| "epoch": 35.3, | |
| "learning_rate": 1.959219858156028e-06, | |
| "loss": 0.1537, | |
| "step": 1659 | |
| }, | |
| { | |
| "epoch": 35.32, | |
| "learning_rate": 1.950354609929078e-06, | |
| "loss": 0.0481, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 35.34, | |
| "learning_rate": 1.9414893617021274e-06, | |
| "loss": 0.0863, | |
| "step": 1661 | |
| }, | |
| { | |
| "epoch": 35.36, | |
| "learning_rate": 1.932624113475177e-06, | |
| "loss": 0.0959, | |
| "step": 1662 | |
| }, | |
| { | |
| "epoch": 35.38, | |
| "learning_rate": 1.9237588652482267e-06, | |
| "loss": 0.1063, | |
| "step": 1663 | |
| }, | |
| { | |
| "epoch": 35.4, | |
| "learning_rate": 1.9148936170212763e-06, | |
| "loss": 0.1392, | |
| "step": 1664 | |
| }, | |
| { | |
| "epoch": 35.43, | |
| "learning_rate": 1.9060283687943261e-06, | |
| "loss": 0.0916, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 35.45, | |
| "learning_rate": 1.8971631205673758e-06, | |
| "loss": 0.1215, | |
| "step": 1666 | |
| }, | |
| { | |
| "epoch": 35.47, | |
| "learning_rate": 1.8882978723404254e-06, | |
| "loss": 0.192, | |
| "step": 1667 | |
| }, | |
| { | |
| "epoch": 35.49, | |
| "learning_rate": 1.879432624113475e-06, | |
| "loss": 0.1034, | |
| "step": 1668 | |
| }, | |
| { | |
| "epoch": 35.51, | |
| "learning_rate": 1.8705673758865249e-06, | |
| "loss": 0.0834, | |
| "step": 1669 | |
| }, | |
| { | |
| "epoch": 35.53, | |
| "learning_rate": 1.8617021276595745e-06, | |
| "loss": 0.1271, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 35.55, | |
| "learning_rate": 1.8528368794326241e-06, | |
| "loss": 0.1128, | |
| "step": 1671 | |
| }, | |
| { | |
| "epoch": 35.57, | |
| "learning_rate": 1.8439716312056737e-06, | |
| "loss": 0.1149, | |
| "step": 1672 | |
| }, | |
| { | |
| "epoch": 35.6, | |
| "learning_rate": 1.8351063829787234e-06, | |
| "loss": 0.0779, | |
| "step": 1673 | |
| }, | |
| { | |
| "epoch": 35.62, | |
| "learning_rate": 1.826241134751773e-06, | |
| "loss": 0.0762, | |
| "step": 1674 | |
| }, | |
| { | |
| "epoch": 35.64, | |
| "learning_rate": 1.8173758865248226e-06, | |
| "loss": 0.0166, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 35.66, | |
| "learning_rate": 1.8085106382978722e-06, | |
| "loss": 0.199, | |
| "step": 1676 | |
| }, | |
| { | |
| "epoch": 35.68, | |
| "learning_rate": 1.799645390070922e-06, | |
| "loss": 0.0426, | |
| "step": 1677 | |
| }, | |
| { | |
| "epoch": 35.7, | |
| "learning_rate": 1.7907801418439717e-06, | |
| "loss": 0.0876, | |
| "step": 1678 | |
| }, | |
| { | |
| "epoch": 35.72, | |
| "learning_rate": 1.7819148936170213e-06, | |
| "loss": 0.0849, | |
| "step": 1679 | |
| }, | |
| { | |
| "epoch": 35.74, | |
| "learning_rate": 1.773049645390071e-06, | |
| "loss": 0.0969, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 35.77, | |
| "learning_rate": 1.7641843971631206e-06, | |
| "loss": 0.0856, | |
| "step": 1681 | |
| }, | |
| { | |
| "epoch": 35.79, | |
| "learning_rate": 1.7553191489361702e-06, | |
| "loss": 0.095, | |
| "step": 1682 | |
| }, | |
| { | |
| "epoch": 35.81, | |
| "learning_rate": 1.7464539007092198e-06, | |
| "loss": 0.029, | |
| "step": 1683 | |
| }, | |
| { | |
| "epoch": 35.83, | |
| "learning_rate": 1.7375886524822697e-06, | |
| "loss": 0.1419, | |
| "step": 1684 | |
| }, | |
| { | |
| "epoch": 35.85, | |
| "learning_rate": 1.7287234042553193e-06, | |
| "loss": 0.0612, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 35.87, | |
| "learning_rate": 1.719858156028369e-06, | |
| "loss": 0.2745, | |
| "step": 1686 | |
| }, | |
| { | |
| "epoch": 35.89, | |
| "learning_rate": 1.7109929078014185e-06, | |
| "loss": 0.0193, | |
| "step": 1687 | |
| }, | |
| { | |
| "epoch": 35.91, | |
| "learning_rate": 1.7021276595744682e-06, | |
| "loss": 0.0455, | |
| "step": 1688 | |
| }, | |
| { | |
| "epoch": 35.94, | |
| "learning_rate": 1.6932624113475178e-06, | |
| "loss": 0.045, | |
| "step": 1689 | |
| }, | |
| { | |
| "epoch": 35.96, | |
| "learning_rate": 1.6843971631205674e-06, | |
| "loss": 0.1071, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 35.98, | |
| "learning_rate": 1.675531914893617e-06, | |
| "loss": 0.1577, | |
| "step": 1691 | |
| }, | |
| { | |
| "epoch": 36.0, | |
| "learning_rate": 1.6666666666666667e-06, | |
| "loss": 0.0518, | |
| "step": 1692 | |
| }, | |
| { | |
| "epoch": 36.0, | |
| "eval_accuracy": 0.9466666666666667, | |
| "eval_loss": 0.24408310651779175, | |
| "eval_runtime": 28.7317, | |
| "eval_samples_per_second": 5.221, | |
| "eval_steps_per_second": 0.348, | |
| "step": 1692 | |
| }, | |
| { | |
| "epoch": 36.02, | |
| "learning_rate": 1.6578014184397165e-06, | |
| "loss": 0.1965, | |
| "step": 1693 | |
| }, | |
| { | |
| "epoch": 36.04, | |
| "learning_rate": 1.6489361702127661e-06, | |
| "loss": 0.0258, | |
| "step": 1694 | |
| }, | |
| { | |
| "epoch": 36.06, | |
| "learning_rate": 1.6400709219858157e-06, | |
| "loss": 0.0449, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 36.09, | |
| "learning_rate": 1.6312056737588654e-06, | |
| "loss": 0.0754, | |
| "step": 1696 | |
| }, | |
| { | |
| "epoch": 36.11, | |
| "learning_rate": 1.622340425531915e-06, | |
| "loss": 0.1434, | |
| "step": 1697 | |
| }, | |
| { | |
| "epoch": 36.13, | |
| "learning_rate": 1.6134751773049646e-06, | |
| "loss": 0.1049, | |
| "step": 1698 | |
| }, | |
| { | |
| "epoch": 36.15, | |
| "learning_rate": 1.6046099290780142e-06, | |
| "loss": 0.1433, | |
| "step": 1699 | |
| }, | |
| { | |
| "epoch": 36.17, | |
| "learning_rate": 1.5957446808510639e-06, | |
| "loss": 0.177, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 36.19, | |
| "learning_rate": 1.5868794326241135e-06, | |
| "loss": 0.0584, | |
| "step": 1701 | |
| }, | |
| { | |
| "epoch": 36.21, | |
| "learning_rate": 1.5780141843971631e-06, | |
| "loss": 0.1839, | |
| "step": 1702 | |
| }, | |
| { | |
| "epoch": 36.23, | |
| "learning_rate": 1.5691489361702128e-06, | |
| "loss": 0.1316, | |
| "step": 1703 | |
| }, | |
| { | |
| "epoch": 36.26, | |
| "learning_rate": 1.5602836879432624e-06, | |
| "loss": 0.0747, | |
| "step": 1704 | |
| }, | |
| { | |
| "epoch": 36.28, | |
| "learning_rate": 1.551418439716312e-06, | |
| "loss": 0.0689, | |
| "step": 1705 | |
| }, | |
| { | |
| "epoch": 36.3, | |
| "learning_rate": 1.5425531914893616e-06, | |
| "loss": 0.0287, | |
| "step": 1706 | |
| }, | |
| { | |
| "epoch": 36.32, | |
| "learning_rate": 1.5336879432624113e-06, | |
| "loss": 0.0757, | |
| "step": 1707 | |
| }, | |
| { | |
| "epoch": 36.34, | |
| "learning_rate": 1.5248226950354609e-06, | |
| "loss": 0.0624, | |
| "step": 1708 | |
| }, | |
| { | |
| "epoch": 36.36, | |
| "learning_rate": 1.5159574468085105e-06, | |
| "loss": 0.0264, | |
| "step": 1709 | |
| }, | |
| { | |
| "epoch": 36.38, | |
| "learning_rate": 1.5070921985815603e-06, | |
| "loss": 0.1256, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 36.4, | |
| "learning_rate": 1.49822695035461e-06, | |
| "loss": 0.1497, | |
| "step": 1711 | |
| }, | |
| { | |
| "epoch": 36.43, | |
| "learning_rate": 1.4893617021276596e-06, | |
| "loss": 0.0761, | |
| "step": 1712 | |
| }, | |
| { | |
| "epoch": 36.45, | |
| "learning_rate": 1.4804964539007092e-06, | |
| "loss": 0.103, | |
| "step": 1713 | |
| }, | |
| { | |
| "epoch": 36.47, | |
| "learning_rate": 1.471631205673759e-06, | |
| "loss": 0.0468, | |
| "step": 1714 | |
| }, | |
| { | |
| "epoch": 36.49, | |
| "learning_rate": 1.4627659574468087e-06, | |
| "loss": 0.2117, | |
| "step": 1715 | |
| }, | |
| { | |
| "epoch": 36.51, | |
| "learning_rate": 1.4539007092198583e-06, | |
| "loss": 0.1024, | |
| "step": 1716 | |
| }, | |
| { | |
| "epoch": 36.53, | |
| "learning_rate": 1.445035460992908e-06, | |
| "loss": 0.0467, | |
| "step": 1717 | |
| }, | |
| { | |
| "epoch": 36.55, | |
| "learning_rate": 1.4361702127659576e-06, | |
| "loss": 0.1682, | |
| "step": 1718 | |
| }, | |
| { | |
| "epoch": 36.57, | |
| "learning_rate": 1.4273049645390072e-06, | |
| "loss": 0.0626, | |
| "step": 1719 | |
| }, | |
| { | |
| "epoch": 36.6, | |
| "learning_rate": 1.4184397163120568e-06, | |
| "loss": 0.0401, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 36.62, | |
| "learning_rate": 1.4095744680851064e-06, | |
| "loss": 0.0554, | |
| "step": 1721 | |
| }, | |
| { | |
| "epoch": 36.64, | |
| "learning_rate": 1.400709219858156e-06, | |
| "loss": 0.0421, | |
| "step": 1722 | |
| }, | |
| { | |
| "epoch": 36.66, | |
| "learning_rate": 1.3918439716312057e-06, | |
| "loss": 0.2075, | |
| "step": 1723 | |
| }, | |
| { | |
| "epoch": 36.68, | |
| "learning_rate": 1.3829787234042553e-06, | |
| "loss": 0.1145, | |
| "step": 1724 | |
| }, | |
| { | |
| "epoch": 36.7, | |
| "learning_rate": 1.374113475177305e-06, | |
| "loss": 0.1192, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 36.72, | |
| "learning_rate": 1.3652482269503548e-06, | |
| "loss": 0.0628, | |
| "step": 1726 | |
| }, | |
| { | |
| "epoch": 36.74, | |
| "learning_rate": 1.3563829787234044e-06, | |
| "loss": 0.0904, | |
| "step": 1727 | |
| }, | |
| { | |
| "epoch": 36.77, | |
| "learning_rate": 1.347517730496454e-06, | |
| "loss": 0.1315, | |
| "step": 1728 | |
| }, | |
| { | |
| "epoch": 36.79, | |
| "learning_rate": 1.3386524822695036e-06, | |
| "loss": 0.0175, | |
| "step": 1729 | |
| }, | |
| { | |
| "epoch": 36.81, | |
| "learning_rate": 1.3297872340425533e-06, | |
| "loss": 0.0282, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 36.83, | |
| "learning_rate": 1.3209219858156029e-06, | |
| "loss": 0.0701, | |
| "step": 1731 | |
| }, | |
| { | |
| "epoch": 36.85, | |
| "learning_rate": 1.3120567375886525e-06, | |
| "loss": 0.0863, | |
| "step": 1732 | |
| }, | |
| { | |
| "epoch": 36.87, | |
| "learning_rate": 1.3031914893617021e-06, | |
| "loss": 0.1144, | |
| "step": 1733 | |
| }, | |
| { | |
| "epoch": 36.89, | |
| "learning_rate": 1.2943262411347518e-06, | |
| "loss": 0.0315, | |
| "step": 1734 | |
| }, | |
| { | |
| "epoch": 36.91, | |
| "learning_rate": 1.2854609929078014e-06, | |
| "loss": 0.0457, | |
| "step": 1735 | |
| }, | |
| { | |
| "epoch": 36.94, | |
| "learning_rate": 1.276595744680851e-06, | |
| "loss": 0.0654, | |
| "step": 1736 | |
| }, | |
| { | |
| "epoch": 36.96, | |
| "learning_rate": 1.2677304964539006e-06, | |
| "loss": 0.1998, | |
| "step": 1737 | |
| }, | |
| { | |
| "epoch": 36.98, | |
| "learning_rate": 1.2588652482269503e-06, | |
| "loss": 0.1427, | |
| "step": 1738 | |
| }, | |
| { | |
| "epoch": 37.0, | |
| "learning_rate": 1.2499999999999999e-06, | |
| "loss": 0.0852, | |
| "step": 1739 | |
| }, | |
| { | |
| "epoch": 37.0, | |
| "eval_accuracy": 0.9466666666666667, | |
| "eval_loss": 0.22427058219909668, | |
| "eval_runtime": 28.9467, | |
| "eval_samples_per_second": 5.182, | |
| "eval_steps_per_second": 0.345, | |
| "step": 1739 | |
| }, | |
| { | |
| "epoch": 37.02, | |
| "learning_rate": 1.2411347517730495e-06, | |
| "loss": 0.1184, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 37.04, | |
| "learning_rate": 1.2322695035460991e-06, | |
| "loss": 0.1621, | |
| "step": 1741 | |
| }, | |
| { | |
| "epoch": 37.06, | |
| "learning_rate": 1.223404255319149e-06, | |
| "loss": 0.1252, | |
| "step": 1742 | |
| }, | |
| { | |
| "epoch": 37.09, | |
| "learning_rate": 1.2145390070921986e-06, | |
| "loss": 0.0422, | |
| "step": 1743 | |
| }, | |
| { | |
| "epoch": 37.11, | |
| "learning_rate": 1.2056737588652482e-06, | |
| "loss": 0.1138, | |
| "step": 1744 | |
| }, | |
| { | |
| "epoch": 37.13, | |
| "learning_rate": 1.196808510638298e-06, | |
| "loss": 0.0775, | |
| "step": 1745 | |
| }, | |
| { | |
| "epoch": 37.15, | |
| "learning_rate": 1.1879432624113477e-06, | |
| "loss": 0.0961, | |
| "step": 1746 | |
| }, | |
| { | |
| "epoch": 37.17, | |
| "learning_rate": 1.1790780141843973e-06, | |
| "loss": 0.0561, | |
| "step": 1747 | |
| }, | |
| { | |
| "epoch": 37.19, | |
| "learning_rate": 1.170212765957447e-06, | |
| "loss": 0.0263, | |
| "step": 1748 | |
| }, | |
| { | |
| "epoch": 37.21, | |
| "learning_rate": 1.1613475177304966e-06, | |
| "loss": 0.1198, | |
| "step": 1749 | |
| }, | |
| { | |
| "epoch": 37.23, | |
| "learning_rate": 1.1524822695035462e-06, | |
| "loss": 0.0766, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 37.26, | |
| "learning_rate": 1.1436170212765958e-06, | |
| "loss": 0.0385, | |
| "step": 1751 | |
| }, | |
| { | |
| "epoch": 37.28, | |
| "learning_rate": 1.1347517730496454e-06, | |
| "loss": 0.0692, | |
| "step": 1752 | |
| }, | |
| { | |
| "epoch": 37.3, | |
| "learning_rate": 1.125886524822695e-06, | |
| "loss": 0.0426, | |
| "step": 1753 | |
| }, | |
| { | |
| "epoch": 37.32, | |
| "learning_rate": 1.1170212765957447e-06, | |
| "loss": 0.1437, | |
| "step": 1754 | |
| }, | |
| { | |
| "epoch": 37.34, | |
| "learning_rate": 1.1081560283687943e-06, | |
| "loss": 0.0387, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 37.36, | |
| "learning_rate": 1.099290780141844e-06, | |
| "loss": 0.0531, | |
| "step": 1756 | |
| }, | |
| { | |
| "epoch": 37.38, | |
| "learning_rate": 1.0904255319148936e-06, | |
| "loss": 0.0201, | |
| "step": 1757 | |
| }, | |
| { | |
| "epoch": 37.4, | |
| "learning_rate": 1.0815602836879434e-06, | |
| "loss": 0.1054, | |
| "step": 1758 | |
| }, | |
| { | |
| "epoch": 37.43, | |
| "learning_rate": 1.072695035460993e-06, | |
| "loss": 0.159, | |
| "step": 1759 | |
| }, | |
| { | |
| "epoch": 37.45, | |
| "learning_rate": 1.0638297872340427e-06, | |
| "loss": 0.1181, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 37.47, | |
| "learning_rate": 1.0549645390070923e-06, | |
| "loss": 0.07, | |
| "step": 1761 | |
| }, | |
| { | |
| "epoch": 37.49, | |
| "learning_rate": 1.046099290780142e-06, | |
| "loss": 0.1666, | |
| "step": 1762 | |
| }, | |
| { | |
| "epoch": 37.51, | |
| "learning_rate": 1.0372340425531915e-06, | |
| "loss": 0.1269, | |
| "step": 1763 | |
| }, | |
| { | |
| "epoch": 37.53, | |
| "learning_rate": 1.0283687943262412e-06, | |
| "loss": 0.0373, | |
| "step": 1764 | |
| }, | |
| { | |
| "epoch": 37.55, | |
| "learning_rate": 1.0195035460992908e-06, | |
| "loss": 0.0359, | |
| "step": 1765 | |
| }, | |
| { | |
| "epoch": 37.57, | |
| "learning_rate": 1.0106382978723404e-06, | |
| "loss": 0.1171, | |
| "step": 1766 | |
| }, | |
| { | |
| "epoch": 37.6, | |
| "learning_rate": 1.00177304964539e-06, | |
| "loss": 0.0265, | |
| "step": 1767 | |
| }, | |
| { | |
| "epoch": 37.62, | |
| "learning_rate": 9.929078014184397e-07, | |
| "loss": 0.135, | |
| "step": 1768 | |
| }, | |
| { | |
| "epoch": 37.64, | |
| "learning_rate": 9.840425531914893e-07, | |
| "loss": 0.0969, | |
| "step": 1769 | |
| }, | |
| { | |
| "epoch": 37.66, | |
| "learning_rate": 9.75177304964539e-07, | |
| "loss": 0.0939, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 37.68, | |
| "learning_rate": 9.663120567375885e-07, | |
| "loss": 0.2363, | |
| "step": 1771 | |
| }, | |
| { | |
| "epoch": 37.7, | |
| "learning_rate": 9.574468085106382e-07, | |
| "loss": 0.1923, | |
| "step": 1772 | |
| }, | |
| { | |
| "epoch": 37.72, | |
| "learning_rate": 9.485815602836879e-07, | |
| "loss": 0.21, | |
| "step": 1773 | |
| }, | |
| { | |
| "epoch": 37.74, | |
| "learning_rate": 9.397163120567375e-07, | |
| "loss": 0.1848, | |
| "step": 1774 | |
| }, | |
| { | |
| "epoch": 37.77, | |
| "learning_rate": 9.308510638297872e-07, | |
| "loss": 0.1188, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 37.79, | |
| "learning_rate": 9.219858156028369e-07, | |
| "loss": 0.0461, | |
| "step": 1776 | |
| }, | |
| { | |
| "epoch": 37.81, | |
| "learning_rate": 9.131205673758865e-07, | |
| "loss": 0.1132, | |
| "step": 1777 | |
| }, | |
| { | |
| "epoch": 37.83, | |
| "learning_rate": 9.042553191489361e-07, | |
| "loss": 0.0514, | |
| "step": 1778 | |
| }, | |
| { | |
| "epoch": 37.85, | |
| "learning_rate": 8.953900709219858e-07, | |
| "loss": 0.1249, | |
| "step": 1779 | |
| }, | |
| { | |
| "epoch": 37.87, | |
| "learning_rate": 8.865248226950355e-07, | |
| "loss": 0.159, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 37.89, | |
| "learning_rate": 8.776595744680851e-07, | |
| "loss": 0.0433, | |
| "step": 1781 | |
| }, | |
| { | |
| "epoch": 37.91, | |
| "learning_rate": 8.687943262411348e-07, | |
| "loss": 0.0761, | |
| "step": 1782 | |
| }, | |
| { | |
| "epoch": 37.94, | |
| "learning_rate": 8.599290780141845e-07, | |
| "loss": 0.1045, | |
| "step": 1783 | |
| }, | |
| { | |
| "epoch": 37.96, | |
| "learning_rate": 8.510638297872341e-07, | |
| "loss": 0.1362, | |
| "step": 1784 | |
| }, | |
| { | |
| "epoch": 37.98, | |
| "learning_rate": 8.421985815602837e-07, | |
| "loss": 0.1994, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 38.0, | |
| "learning_rate": 8.333333333333333e-07, | |
| "loss": 0.0853, | |
| "step": 1786 | |
| }, | |
| { | |
| "epoch": 38.0, | |
| "eval_accuracy": 0.9533333333333334, | |
| "eval_loss": 0.20281197130680084, | |
| "eval_runtime": 29.4548, | |
| "eval_samples_per_second": 5.093, | |
| "eval_steps_per_second": 0.34, | |
| "step": 1786 | |
| }, | |
| { | |
| "epoch": 38.02, | |
| "learning_rate": 8.244680851063831e-07, | |
| "loss": 0.1527, | |
| "step": 1787 | |
| }, | |
| { | |
| "epoch": 38.04, | |
| "learning_rate": 8.156028368794327e-07, | |
| "loss": 0.0754, | |
| "step": 1788 | |
| }, | |
| { | |
| "epoch": 38.06, | |
| "learning_rate": 8.067375886524823e-07, | |
| "loss": 0.0793, | |
| "step": 1789 | |
| }, | |
| { | |
| "epoch": 38.09, | |
| "learning_rate": 7.978723404255319e-07, | |
| "loss": 0.1325, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 38.11, | |
| "learning_rate": 7.890070921985816e-07, | |
| "loss": 0.0721, | |
| "step": 1791 | |
| }, | |
| { | |
| "epoch": 38.13, | |
| "learning_rate": 7.801418439716312e-07, | |
| "loss": 0.015, | |
| "step": 1792 | |
| }, | |
| { | |
| "epoch": 38.15, | |
| "learning_rate": 7.712765957446808e-07, | |
| "loss": 0.0819, | |
| "step": 1793 | |
| }, | |
| { | |
| "epoch": 38.17, | |
| "learning_rate": 7.624113475177304e-07, | |
| "loss": 0.0472, | |
| "step": 1794 | |
| }, | |
| { | |
| "epoch": 38.19, | |
| "learning_rate": 7.535460992907802e-07, | |
| "loss": 0.0759, | |
| "step": 1795 | |
| }, | |
| { | |
| "epoch": 38.21, | |
| "learning_rate": 7.446808510638298e-07, | |
| "loss": 0.111, | |
| "step": 1796 | |
| }, | |
| { | |
| "epoch": 38.23, | |
| "learning_rate": 7.358156028368795e-07, | |
| "loss": 0.1027, | |
| "step": 1797 | |
| }, | |
| { | |
| "epoch": 38.26, | |
| "learning_rate": 7.269503546099292e-07, | |
| "loss": 0.0063, | |
| "step": 1798 | |
| }, | |
| { | |
| "epoch": 38.28, | |
| "learning_rate": 7.180851063829788e-07, | |
| "loss": 0.1085, | |
| "step": 1799 | |
| }, | |
| { | |
| "epoch": 38.3, | |
| "learning_rate": 7.092198581560284e-07, | |
| "loss": 0.0932, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 38.32, | |
| "learning_rate": 7.00354609929078e-07, | |
| "loss": 0.0965, | |
| "step": 1801 | |
| }, | |
| { | |
| "epoch": 38.34, | |
| "learning_rate": 6.914893617021277e-07, | |
| "loss": 0.0285, | |
| "step": 1802 | |
| }, | |
| { | |
| "epoch": 38.36, | |
| "learning_rate": 6.826241134751774e-07, | |
| "loss": 0.161, | |
| "step": 1803 | |
| }, | |
| { | |
| "epoch": 38.38, | |
| "learning_rate": 6.73758865248227e-07, | |
| "loss": 0.0664, | |
| "step": 1804 | |
| }, | |
| { | |
| "epoch": 38.4, | |
| "learning_rate": 6.648936170212766e-07, | |
| "loss": 0.0357, | |
| "step": 1805 | |
| }, | |
| { | |
| "epoch": 38.43, | |
| "learning_rate": 6.560283687943263e-07, | |
| "loss": 0.1009, | |
| "step": 1806 | |
| }, | |
| { | |
| "epoch": 38.45, | |
| "learning_rate": 6.471631205673759e-07, | |
| "loss": 0.1221, | |
| "step": 1807 | |
| }, | |
| { | |
| "epoch": 38.47, | |
| "learning_rate": 6.382978723404255e-07, | |
| "loss": 0.1135, | |
| "step": 1808 | |
| }, | |
| { | |
| "epoch": 38.49, | |
| "learning_rate": 6.294326241134751e-07, | |
| "loss": 0.1504, | |
| "step": 1809 | |
| }, | |
| { | |
| "epoch": 38.51, | |
| "learning_rate": 6.205673758865248e-07, | |
| "loss": 0.0496, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 38.53, | |
| "learning_rate": 6.117021276595745e-07, | |
| "loss": 0.0672, | |
| "step": 1811 | |
| }, | |
| { | |
| "epoch": 38.55, | |
| "learning_rate": 6.028368794326241e-07, | |
| "loss": 0.0953, | |
| "step": 1812 | |
| }, | |
| { | |
| "epoch": 38.57, | |
| "learning_rate": 5.939716312056738e-07, | |
| "loss": 0.0978, | |
| "step": 1813 | |
| }, | |
| { | |
| "epoch": 38.6, | |
| "learning_rate": 5.851063829787235e-07, | |
| "loss": 0.0641, | |
| "step": 1814 | |
| }, | |
| { | |
| "epoch": 38.62, | |
| "learning_rate": 5.762411347517731e-07, | |
| "loss": 0.0412, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 38.64, | |
| "learning_rate": 5.673758865248227e-07, | |
| "loss": 0.1155, | |
| "step": 1816 | |
| }, | |
| { | |
| "epoch": 38.66, | |
| "learning_rate": 5.585106382978723e-07, | |
| "loss": 0.0475, | |
| "step": 1817 | |
| }, | |
| { | |
| "epoch": 38.68, | |
| "learning_rate": 5.49645390070922e-07, | |
| "loss": 0.0519, | |
| "step": 1818 | |
| }, | |
| { | |
| "epoch": 38.7, | |
| "learning_rate": 5.407801418439717e-07, | |
| "loss": 0.0969, | |
| "step": 1819 | |
| }, | |
| { | |
| "epoch": 38.72, | |
| "learning_rate": 5.319148936170213e-07, | |
| "loss": 0.0496, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 38.74, | |
| "learning_rate": 5.23049645390071e-07, | |
| "loss": 0.1084, | |
| "step": 1821 | |
| }, | |
| { | |
| "epoch": 38.77, | |
| "learning_rate": 5.141843971631206e-07, | |
| "loss": 0.199, | |
| "step": 1822 | |
| }, | |
| { | |
| "epoch": 38.79, | |
| "learning_rate": 5.053191489361702e-07, | |
| "loss": 0.1801, | |
| "step": 1823 | |
| }, | |
| { | |
| "epoch": 38.81, | |
| "learning_rate": 4.964539007092198e-07, | |
| "loss": 0.0379, | |
| "step": 1824 | |
| }, | |
| { | |
| "epoch": 38.83, | |
| "learning_rate": 4.875886524822695e-07, | |
| "loss": 0.1444, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 38.85, | |
| "learning_rate": 4.787234042553191e-07, | |
| "loss": 0.1048, | |
| "step": 1826 | |
| }, | |
| { | |
| "epoch": 38.87, | |
| "learning_rate": 4.6985815602836876e-07, | |
| "loss": 0.0957, | |
| "step": 1827 | |
| }, | |
| { | |
| "epoch": 38.89, | |
| "learning_rate": 4.6099290780141843e-07, | |
| "loss": 0.0649, | |
| "step": 1828 | |
| }, | |
| { | |
| "epoch": 38.91, | |
| "learning_rate": 4.5212765957446806e-07, | |
| "loss": 0.0467, | |
| "step": 1829 | |
| }, | |
| { | |
| "epoch": 38.94, | |
| "learning_rate": 4.4326241134751774e-07, | |
| "loss": 0.0251, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 38.96, | |
| "learning_rate": 4.343971631205674e-07, | |
| "loss": 0.1594, | |
| "step": 1831 | |
| }, | |
| { | |
| "epoch": 38.98, | |
| "learning_rate": 4.2553191489361704e-07, | |
| "loss": 0.0435, | |
| "step": 1832 | |
| }, | |
| { | |
| "epoch": 39.0, | |
| "learning_rate": 4.1666666666666667e-07, | |
| "loss": 0.055, | |
| "step": 1833 | |
| }, | |
| { | |
| "epoch": 39.0, | |
| "eval_accuracy": 0.9466666666666667, | |
| "eval_loss": 0.21239179372787476, | |
| "eval_runtime": 27.5219, | |
| "eval_samples_per_second": 5.45, | |
| "eval_steps_per_second": 0.363, | |
| "step": 1833 | |
| }, | |
| { | |
| "epoch": 39.02, | |
| "learning_rate": 4.0780141843971634e-07, | |
| "loss": 0.0096, | |
| "step": 1834 | |
| }, | |
| { | |
| "epoch": 39.04, | |
| "learning_rate": 3.9893617021276597e-07, | |
| "loss": 0.2012, | |
| "step": 1835 | |
| }, | |
| { | |
| "epoch": 39.06, | |
| "learning_rate": 3.900709219858156e-07, | |
| "loss": 0.1289, | |
| "step": 1836 | |
| }, | |
| { | |
| "epoch": 39.09, | |
| "learning_rate": 3.812056737588652e-07, | |
| "loss": 0.0745, | |
| "step": 1837 | |
| }, | |
| { | |
| "epoch": 39.11, | |
| "learning_rate": 3.723404255319149e-07, | |
| "loss": 0.0766, | |
| "step": 1838 | |
| }, | |
| { | |
| "epoch": 39.13, | |
| "learning_rate": 3.634751773049646e-07, | |
| "loss": 0.0595, | |
| "step": 1839 | |
| }, | |
| { | |
| "epoch": 39.15, | |
| "learning_rate": 3.546099290780142e-07, | |
| "loss": 0.155, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 39.17, | |
| "learning_rate": 3.457446808510638e-07, | |
| "loss": 0.0118, | |
| "step": 1841 | |
| }, | |
| { | |
| "epoch": 39.19, | |
| "learning_rate": 3.368794326241135e-07, | |
| "loss": 0.176, | |
| "step": 1842 | |
| }, | |
| { | |
| "epoch": 39.21, | |
| "learning_rate": 3.2801418439716313e-07, | |
| "loss": 0.0267, | |
| "step": 1843 | |
| }, | |
| { | |
| "epoch": 39.23, | |
| "learning_rate": 3.1914893617021275e-07, | |
| "loss": 0.0417, | |
| "step": 1844 | |
| }, | |
| { | |
| "epoch": 39.26, | |
| "learning_rate": 3.102836879432624e-07, | |
| "loss": 0.0437, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 39.28, | |
| "learning_rate": 3.0141843971631206e-07, | |
| "loss": 0.0737, | |
| "step": 1846 | |
| }, | |
| { | |
| "epoch": 39.3, | |
| "learning_rate": 2.9255319148936174e-07, | |
| "loss": 0.0216, | |
| "step": 1847 | |
| }, | |
| { | |
| "epoch": 39.32, | |
| "learning_rate": 2.8368794326241136e-07, | |
| "loss": 0.1002, | |
| "step": 1848 | |
| }, | |
| { | |
| "epoch": 39.34, | |
| "learning_rate": 2.74822695035461e-07, | |
| "loss": 0.0893, | |
| "step": 1849 | |
| }, | |
| { | |
| "epoch": 39.36, | |
| "learning_rate": 2.6595744680851066e-07, | |
| "loss": 0.1174, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 39.38, | |
| "learning_rate": 2.570921985815603e-07, | |
| "loss": 0.0869, | |
| "step": 1851 | |
| }, | |
| { | |
| "epoch": 39.4, | |
| "learning_rate": 2.482269503546099e-07, | |
| "loss": 0.0384, | |
| "step": 1852 | |
| }, | |
| { | |
| "epoch": 39.43, | |
| "learning_rate": 2.3936170212765954e-07, | |
| "loss": 0.155, | |
| "step": 1853 | |
| }, | |
| { | |
| "epoch": 39.45, | |
| "learning_rate": 2.3049645390070922e-07, | |
| "loss": 0.136, | |
| "step": 1854 | |
| }, | |
| { | |
| "epoch": 39.47, | |
| "learning_rate": 2.2163120567375887e-07, | |
| "loss": 0.129, | |
| "step": 1855 | |
| }, | |
| { | |
| "epoch": 39.49, | |
| "learning_rate": 2.1276595744680852e-07, | |
| "loss": 0.0625, | |
| "step": 1856 | |
| }, | |
| { | |
| "epoch": 39.51, | |
| "learning_rate": 2.0390070921985817e-07, | |
| "loss": 0.1097, | |
| "step": 1857 | |
| }, | |
| { | |
| "epoch": 39.53, | |
| "learning_rate": 1.950354609929078e-07, | |
| "loss": 0.1263, | |
| "step": 1858 | |
| }, | |
| { | |
| "epoch": 39.55, | |
| "learning_rate": 1.8617021276595745e-07, | |
| "loss": 0.1646, | |
| "step": 1859 | |
| }, | |
| { | |
| "epoch": 39.57, | |
| "learning_rate": 1.773049645390071e-07, | |
| "loss": 0.0497, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 39.6, | |
| "learning_rate": 1.6843971631205675e-07, | |
| "loss": 0.1346, | |
| "step": 1861 | |
| }, | |
| { | |
| "epoch": 39.62, | |
| "learning_rate": 1.5957446808510638e-07, | |
| "loss": 0.1142, | |
| "step": 1862 | |
| }, | |
| { | |
| "epoch": 39.64, | |
| "learning_rate": 1.5070921985815603e-07, | |
| "loss": 0.1872, | |
| "step": 1863 | |
| }, | |
| { | |
| "epoch": 39.66, | |
| "learning_rate": 1.4184397163120568e-07, | |
| "loss": 0.1338, | |
| "step": 1864 | |
| }, | |
| { | |
| "epoch": 39.68, | |
| "learning_rate": 1.3297872340425533e-07, | |
| "loss": 0.1156, | |
| "step": 1865 | |
| }, | |
| { | |
| "epoch": 39.7, | |
| "learning_rate": 1.2411347517730496e-07, | |
| "loss": 0.0192, | |
| "step": 1866 | |
| }, | |
| { | |
| "epoch": 39.72, | |
| "learning_rate": 1.1524822695035461e-07, | |
| "loss": 0.0808, | |
| "step": 1867 | |
| }, | |
| { | |
| "epoch": 39.74, | |
| "learning_rate": 1.0638297872340426e-07, | |
| "loss": 0.0551, | |
| "step": 1868 | |
| }, | |
| { | |
| "epoch": 39.77, | |
| "learning_rate": 9.75177304964539e-08, | |
| "loss": 0.1356, | |
| "step": 1869 | |
| }, | |
| { | |
| "epoch": 39.79, | |
| "learning_rate": 8.865248226950355e-08, | |
| "loss": 0.171, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 39.81, | |
| "learning_rate": 7.978723404255319e-08, | |
| "loss": 0.0596, | |
| "step": 1871 | |
| }, | |
| { | |
| "epoch": 39.83, | |
| "learning_rate": 7.092198581560284e-08, | |
| "loss": 0.0834, | |
| "step": 1872 | |
| }, | |
| { | |
| "epoch": 39.85, | |
| "learning_rate": 6.205673758865248e-08, | |
| "loss": 0.0668, | |
| "step": 1873 | |
| }, | |
| { | |
| "epoch": 39.87, | |
| "learning_rate": 5.319148936170213e-08, | |
| "loss": 0.1565, | |
| "step": 1874 | |
| }, | |
| { | |
| "epoch": 39.89, | |
| "learning_rate": 4.4326241134751775e-08, | |
| "loss": 0.1287, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 39.91, | |
| "learning_rate": 3.546099290780142e-08, | |
| "loss": 0.0424, | |
| "step": 1876 | |
| }, | |
| { | |
| "epoch": 39.94, | |
| "learning_rate": 2.6595744680851065e-08, | |
| "loss": 0.0173, | |
| "step": 1877 | |
| }, | |
| { | |
| "epoch": 39.96, | |
| "learning_rate": 1.773049645390071e-08, | |
| "loss": 0.0677, | |
| "step": 1878 | |
| }, | |
| { | |
| "epoch": 39.98, | |
| "learning_rate": 8.865248226950355e-09, | |
| "loss": 0.0839, | |
| "step": 1879 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "learning_rate": 0.0, | |
| "loss": 0.0646, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "eval_accuracy": 0.9466666666666667, | |
| "eval_loss": 0.22022071480751038, | |
| "eval_runtime": 28.5866, | |
| "eval_samples_per_second": 5.247, | |
| "eval_steps_per_second": 0.35, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "step": 1880, | |
| "total_flos": 9.727364893598024e+19, | |
| "train_loss": 0.15093487496487795, | |
| "train_runtime": 14666.0739, | |
| "train_samples_per_second": 8.177, | |
| "train_steps_per_second": 0.128 | |
| } | |
| ], | |
| "max_steps": 1880, | |
| "num_train_epochs": 40, | |
| "total_flos": 9.727364893598024e+19, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |