nemo_nano_100k / trainer_state.json
ryanmarten's picture
Upload model
7ca99c0 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 5.0,
"eval_steps": 500,
"global_step": 305,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01639344262295082,
"grad_norm": 5.58112545754218,
"learning_rate": 2.580645161290323e-06,
"loss": 0.9597,
"step": 1
},
{
"epoch": 0.03278688524590164,
"grad_norm": 5.5621177659627685,
"learning_rate": 5.161290322580646e-06,
"loss": 0.9618,
"step": 2
},
{
"epoch": 0.04918032786885246,
"grad_norm": 5.2002830302431375,
"learning_rate": 7.741935483870968e-06,
"loss": 0.9528,
"step": 3
},
{
"epoch": 0.06557377049180328,
"grad_norm": 2.429288619179718,
"learning_rate": 1.0322580645161291e-05,
"loss": 0.8743,
"step": 4
},
{
"epoch": 0.08196721311475409,
"grad_norm": 2.303682448858842,
"learning_rate": 1.2903225806451613e-05,
"loss": 0.8371,
"step": 5
},
{
"epoch": 0.09836065573770492,
"grad_norm": 3.7903444701868207,
"learning_rate": 1.5483870967741936e-05,
"loss": 0.8396,
"step": 6
},
{
"epoch": 0.11475409836065574,
"grad_norm": 4.276574856653076,
"learning_rate": 1.806451612903226e-05,
"loss": 0.8132,
"step": 7
},
{
"epoch": 0.13114754098360656,
"grad_norm": 3.2100580662432443,
"learning_rate": 2.0645161290322582e-05,
"loss": 0.8073,
"step": 8
},
{
"epoch": 0.14754098360655737,
"grad_norm": 2.400368523717908,
"learning_rate": 2.3225806451612906e-05,
"loss": 0.7532,
"step": 9
},
{
"epoch": 0.16393442622950818,
"grad_norm": 2.107929978755255,
"learning_rate": 2.5806451612903226e-05,
"loss": 0.7392,
"step": 10
},
{
"epoch": 0.18032786885245902,
"grad_norm": 1.706605478358681,
"learning_rate": 2.8387096774193552e-05,
"loss": 0.7263,
"step": 11
},
{
"epoch": 0.19672131147540983,
"grad_norm": 1.4882561058595487,
"learning_rate": 3.096774193548387e-05,
"loss": 0.709,
"step": 12
},
{
"epoch": 0.21311475409836064,
"grad_norm": 1.2623324140261931,
"learning_rate": 3.3548387096774195e-05,
"loss": 0.6926,
"step": 13
},
{
"epoch": 0.22950819672131148,
"grad_norm": 1.0042413251607623,
"learning_rate": 3.612903225806452e-05,
"loss": 0.6882,
"step": 14
},
{
"epoch": 0.2459016393442623,
"grad_norm": 1.2054555229493642,
"learning_rate": 3.870967741935484e-05,
"loss": 0.6705,
"step": 15
},
{
"epoch": 0.26229508196721313,
"grad_norm": 1.404103110949475,
"learning_rate": 4.1290322580645165e-05,
"loss": 0.6702,
"step": 16
},
{
"epoch": 0.2786885245901639,
"grad_norm": 1.4484944472411994,
"learning_rate": 4.387096774193548e-05,
"loss": 0.6655,
"step": 17
},
{
"epoch": 0.29508196721311475,
"grad_norm": 0.9845835703253549,
"learning_rate": 4.645161290322581e-05,
"loss": 0.6602,
"step": 18
},
{
"epoch": 0.3114754098360656,
"grad_norm": 1.8228708849031587,
"learning_rate": 4.9032258064516135e-05,
"loss": 0.6586,
"step": 19
},
{
"epoch": 0.32786885245901637,
"grad_norm": 1.0073875744998761,
"learning_rate": 5.161290322580645e-05,
"loss": 0.6283,
"step": 20
},
{
"epoch": 0.3442622950819672,
"grad_norm": 2.1477016621670657,
"learning_rate": 5.4193548387096774e-05,
"loss": 0.645,
"step": 21
},
{
"epoch": 0.36065573770491804,
"grad_norm": 1.4144920525799276,
"learning_rate": 5.6774193548387104e-05,
"loss": 0.6402,
"step": 22
},
{
"epoch": 0.3770491803278688,
"grad_norm": 1.7423289786096712,
"learning_rate": 5.935483870967743e-05,
"loss": 0.6422,
"step": 23
},
{
"epoch": 0.39344262295081966,
"grad_norm": 1.5173523064275696,
"learning_rate": 6.193548387096774e-05,
"loss": 0.6368,
"step": 24
},
{
"epoch": 0.4098360655737705,
"grad_norm": 1.400015090802773,
"learning_rate": 6.451612903225807e-05,
"loss": 0.6331,
"step": 25
},
{
"epoch": 0.4262295081967213,
"grad_norm": 1.4981731922098314,
"learning_rate": 6.709677419354839e-05,
"loss": 0.6154,
"step": 26
},
{
"epoch": 0.4426229508196721,
"grad_norm": 1.3910445077945457,
"learning_rate": 6.967741935483871e-05,
"loss": 0.6248,
"step": 27
},
{
"epoch": 0.45901639344262296,
"grad_norm": 2.287234175350141,
"learning_rate": 7.225806451612904e-05,
"loss": 0.6323,
"step": 28
},
{
"epoch": 0.47540983606557374,
"grad_norm": 1.8671369685906136,
"learning_rate": 7.483870967741936e-05,
"loss": 0.6233,
"step": 29
},
{
"epoch": 0.4918032786885246,
"grad_norm": 1.465283599887322,
"learning_rate": 7.741935483870968e-05,
"loss": 0.6156,
"step": 30
},
{
"epoch": 0.5081967213114754,
"grad_norm": 2.013369833154031,
"learning_rate": 8e-05,
"loss": 0.6138,
"step": 31
},
{
"epoch": 0.5245901639344263,
"grad_norm": 1.4445848293844874,
"learning_rate": 7.99973707985831e-05,
"loss": 0.6199,
"step": 32
},
{
"epoch": 0.5409836065573771,
"grad_norm": 2.008002806928902,
"learning_rate": 7.998948353996737e-05,
"loss": 0.6296,
"step": 33
},
{
"epoch": 0.5573770491803278,
"grad_norm": 1.6006692484547813,
"learning_rate": 7.99763392610124e-05,
"loss": 0.6103,
"step": 34
},
{
"epoch": 0.5737704918032787,
"grad_norm": 1.7952820693638532,
"learning_rate": 7.995793968966601e-05,
"loss": 0.6189,
"step": 35
},
{
"epoch": 0.5901639344262295,
"grad_norm": 1.260255034241788,
"learning_rate": 7.993428724473719e-05,
"loss": 0.604,
"step": 36
},
{
"epoch": 0.6065573770491803,
"grad_norm": 2.918515643648626,
"learning_rate": 7.990538503557801e-05,
"loss": 0.6133,
"step": 37
},
{
"epoch": 0.6229508196721312,
"grad_norm": 1.5936203819891444,
"learning_rate": 7.987123686167492e-05,
"loss": 0.6041,
"step": 38
},
{
"epoch": 0.639344262295082,
"grad_norm": 3.246485277477806,
"learning_rate": 7.98318472121493e-05,
"loss": 0.6172,
"step": 39
},
{
"epoch": 0.6557377049180327,
"grad_norm": 2.818701756393196,
"learning_rate": 7.978722126516727e-05,
"loss": 0.6167,
"step": 40
},
{
"epoch": 0.6721311475409836,
"grad_norm": 1.7221453757855867,
"learning_rate": 7.973736488725896e-05,
"loss": 0.5973,
"step": 41
},
{
"epoch": 0.6885245901639344,
"grad_norm": 1.802719546933628,
"learning_rate": 7.968228463254735e-05,
"loss": 0.6016,
"step": 42
},
{
"epoch": 0.7049180327868853,
"grad_norm": 1.471860307081128,
"learning_rate": 7.962198774188664e-05,
"loss": 0.5971,
"step": 43
},
{
"epoch": 0.7213114754098361,
"grad_norm": 1.2373741203327389,
"learning_rate": 7.955648214191034e-05,
"loss": 0.5889,
"step": 44
},
{
"epoch": 0.7377049180327869,
"grad_norm": 1.261142820035152,
"learning_rate": 7.948577644398924e-05,
"loss": 0.5849,
"step": 45
},
{
"epoch": 0.7540983606557377,
"grad_norm": 2.0142144188297166,
"learning_rate": 7.940987994309944e-05,
"loss": 0.5871,
"step": 46
},
{
"epoch": 0.7704918032786885,
"grad_norm": 1.1871765895293158,
"learning_rate": 7.932880261660031e-05,
"loss": 0.577,
"step": 47
},
{
"epoch": 0.7868852459016393,
"grad_norm": 1.2630851815058795,
"learning_rate": 7.924255512292291e-05,
"loss": 0.5856,
"step": 48
},
{
"epoch": 0.8032786885245902,
"grad_norm": 1.6849343569275448,
"learning_rate": 7.91511488001689e-05,
"loss": 0.5931,
"step": 49
},
{
"epoch": 0.819672131147541,
"grad_norm": 0.9859431187761973,
"learning_rate": 7.905459566461993e-05,
"loss": 0.571,
"step": 50
},
{
"epoch": 0.8360655737704918,
"grad_norm": 1.7623450712427522,
"learning_rate": 7.895290840915803e-05,
"loss": 0.5915,
"step": 51
},
{
"epoch": 0.8524590163934426,
"grad_norm": 1.2069052100241535,
"learning_rate": 7.884610040159704e-05,
"loss": 0.5879,
"step": 52
},
{
"epoch": 0.8688524590163934,
"grad_norm": 1.2236642927441974,
"learning_rate": 7.873418568292517e-05,
"loss": 0.5779,
"step": 53
},
{
"epoch": 0.8852459016393442,
"grad_norm": 5.906328363997742,
"learning_rate": 7.861717896545928e-05,
"loss": 0.6193,
"step": 54
},
{
"epoch": 0.9016393442622951,
"grad_norm": 2.8517859472775013,
"learning_rate": 7.849509563091073e-05,
"loss": 0.6092,
"step": 55
},
{
"epoch": 0.9180327868852459,
"grad_norm": 1.6402739036234257,
"learning_rate": 7.836795172836335e-05,
"loss": 0.5895,
"step": 56
},
{
"epoch": 0.9344262295081968,
"grad_norm": 1.071478048704771,
"learning_rate": 7.823576397216354e-05,
"loss": 0.5821,
"step": 57
},
{
"epoch": 0.9508196721311475,
"grad_norm": 2.6529573722112563,
"learning_rate": 7.809854973972311e-05,
"loss": 0.5988,
"step": 58
},
{
"epoch": 0.9672131147540983,
"grad_norm": 2.1179614996092426,
"learning_rate": 7.795632706923478e-05,
"loss": 0.6166,
"step": 59
},
{
"epoch": 0.9836065573770492,
"grad_norm": 1.9008009613755237,
"learning_rate": 7.78091146573009e-05,
"loss": 0.5992,
"step": 60
},
{
"epoch": 1.0,
"grad_norm": 1.894221127540697,
"learning_rate": 7.765693185647557e-05,
"loss": 0.5826,
"step": 61
},
{
"epoch": 1.0163934426229508,
"grad_norm": 1.708613207423494,
"learning_rate": 7.749979867272055e-05,
"loss": 0.5692,
"step": 62
},
{
"epoch": 1.0327868852459017,
"grad_norm": 1.8649041648138827,
"learning_rate": 7.733773576277533e-05,
"loss": 0.5729,
"step": 63
},
{
"epoch": 1.0491803278688525,
"grad_norm": 1.41612070780247,
"learning_rate": 7.71707644314415e-05,
"loss": 0.5579,
"step": 64
},
{
"epoch": 1.0655737704918034,
"grad_norm": 1.9035698103601548,
"learning_rate": 7.699890662878214e-05,
"loss": 0.5545,
"step": 65
},
{
"epoch": 1.0819672131147542,
"grad_norm": 4.94355544144476,
"learning_rate": 7.682218494723615e-05,
"loss": 0.5625,
"step": 66
},
{
"epoch": 1.098360655737705,
"grad_norm": 1.4759904398074293,
"learning_rate": 7.66406226186483e-05,
"loss": 0.5791,
"step": 67
},
{
"epoch": 1.1147540983606556,
"grad_norm": 1.5331088373110235,
"learning_rate": 7.645424351121519e-05,
"loss": 0.5604,
"step": 68
},
{
"epoch": 1.1311475409836065,
"grad_norm": 1.289579106246813,
"learning_rate": 7.626307212634745e-05,
"loss": 0.5585,
"step": 69
},
{
"epoch": 1.1475409836065573,
"grad_norm": 1.3594377937672748,
"learning_rate": 7.606713359544891e-05,
"loss": 0.5563,
"step": 70
},
{
"epoch": 1.1639344262295082,
"grad_norm": 1.4718812627084297,
"learning_rate": 7.586645367661272e-05,
"loss": 0.5581,
"step": 71
},
{
"epoch": 1.180327868852459,
"grad_norm": 0.9356040160597947,
"learning_rate": 7.566105875123523e-05,
"loss": 0.5361,
"step": 72
},
{
"epoch": 1.1967213114754098,
"grad_norm": 0.801036937648957,
"learning_rate": 7.545097582054786e-05,
"loss": 0.555,
"step": 73
},
{
"epoch": 1.2131147540983607,
"grad_norm": 0.9230784392329119,
"learning_rate": 7.523623250206758e-05,
"loss": 0.5514,
"step": 74
},
{
"epoch": 1.2295081967213115,
"grad_norm": 1.27446883289555,
"learning_rate": 7.501685702596625e-05,
"loss": 0.5514,
"step": 75
},
{
"epoch": 1.2459016393442623,
"grad_norm": 1.089059775020075,
"learning_rate": 7.47928782313595e-05,
"loss": 0.5494,
"step": 76
},
{
"epoch": 1.2622950819672132,
"grad_norm": 0.9532311404835904,
"learning_rate": 7.456432556251552e-05,
"loss": 0.5369,
"step": 77
},
{
"epoch": 1.278688524590164,
"grad_norm": 1.1037696208490315,
"learning_rate": 7.433122906498436e-05,
"loss": 0.5454,
"step": 78
},
{
"epoch": 1.2950819672131146,
"grad_norm": 1.0044629833456924,
"learning_rate": 7.409361938164811e-05,
"loss": 0.5536,
"step": 79
},
{
"epoch": 1.3114754098360657,
"grad_norm": 1.139817816380103,
"learning_rate": 7.385152774869257e-05,
"loss": 0.5451,
"step": 80
},
{
"epoch": 1.3278688524590163,
"grad_norm": 0.9956186270275268,
"learning_rate": 7.360498599150095e-05,
"loss": 0.5444,
"step": 81
},
{
"epoch": 1.3442622950819672,
"grad_norm": 1.4777189178807848,
"learning_rate": 7.335402652047011e-05,
"loss": 0.5482,
"step": 82
},
{
"epoch": 1.360655737704918,
"grad_norm": 1.0684258214405637,
"learning_rate": 7.309868232674992e-05,
"loss": 0.5425,
"step": 83
},
{
"epoch": 1.3770491803278688,
"grad_norm": 1.045469949816082,
"learning_rate": 7.283898697790615e-05,
"loss": 0.5326,
"step": 84
},
{
"epoch": 1.3934426229508197,
"grad_norm": 1.2884574146396586,
"learning_rate": 7.257497461350778e-05,
"loss": 0.541,
"step": 85
},
{
"epoch": 1.4098360655737705,
"grad_norm": 1.2291748562831601,
"learning_rate": 7.23066799406389e-05,
"loss": 0.5463,
"step": 86
},
{
"epoch": 1.4262295081967213,
"grad_norm": 0.8801073224983168,
"learning_rate": 7.203413822933626e-05,
"loss": 0.5362,
"step": 87
},
{
"epoch": 1.4426229508196722,
"grad_norm": 0.9677124160991732,
"learning_rate": 7.17573853079525e-05,
"loss": 0.5405,
"step": 88
},
{
"epoch": 1.459016393442623,
"grad_norm": 1.3289631307779222,
"learning_rate": 7.14764575584463e-05,
"loss": 0.5393,
"step": 89
},
{
"epoch": 1.4754098360655736,
"grad_norm": 0.6274914847275181,
"learning_rate": 7.119139191159949e-05,
"loss": 0.5329,
"step": 90
},
{
"epoch": 1.4918032786885247,
"grad_norm": 0.5931193088659492,
"learning_rate": 7.090222584216221e-05,
"loss": 0.5369,
"step": 91
},
{
"epoch": 1.5081967213114753,
"grad_norm": 1.0213786001116576,
"learning_rate": 7.060899736392643e-05,
"loss": 0.5404,
"step": 92
},
{
"epoch": 1.5245901639344264,
"grad_norm": 1.2526087387777403,
"learning_rate": 7.031174502472868e-05,
"loss": 0.5445,
"step": 93
},
{
"epoch": 1.540983606557377,
"grad_norm": 0.9943578918521142,
"learning_rate": 7.001050790138252e-05,
"loss": 0.5315,
"step": 94
},
{
"epoch": 1.5573770491803278,
"grad_norm": 1.2583383738970864,
"learning_rate": 6.970532559454152e-05,
"loss": 0.5369,
"step": 95
},
{
"epoch": 1.5737704918032787,
"grad_norm": 0.8624167954684502,
"learning_rate": 6.939623822349339e-05,
"loss": 0.5212,
"step": 96
},
{
"epoch": 1.5901639344262295,
"grad_norm": 0.8714228266723913,
"learning_rate": 6.908328642088579e-05,
"loss": 0.5325,
"step": 97
},
{
"epoch": 1.6065573770491803,
"grad_norm": 0.9559074422174284,
"learning_rate": 6.87665113273849e-05,
"loss": 0.5354,
"step": 98
},
{
"epoch": 1.6229508196721312,
"grad_norm": 1.020705004388792,
"learning_rate": 6.844595458626691e-05,
"loss": 0.5305,
"step": 99
},
{
"epoch": 1.639344262295082,
"grad_norm": 1.2718848911597436,
"learning_rate": 6.812165833794375e-05,
"loss": 0.534,
"step": 100
},
{
"epoch": 1.6557377049180326,
"grad_norm": 0.8646369335110455,
"learning_rate": 6.779366521442318e-05,
"loss": 0.5298,
"step": 101
},
{
"epoch": 1.6721311475409837,
"grad_norm": 1.0431466383974872,
"learning_rate": 6.746201833370447e-05,
"loss": 0.5258,
"step": 102
},
{
"epoch": 1.6885245901639343,
"grad_norm": 1.2838018594279825,
"learning_rate": 6.712676129411005e-05,
"loss": 0.5363,
"step": 103
},
{
"epoch": 1.7049180327868854,
"grad_norm": 0.7735313109171527,
"learning_rate": 6.678793816855408e-05,
"loss": 0.5256,
"step": 104
},
{
"epoch": 1.721311475409836,
"grad_norm": 1.1292517886407496,
"learning_rate": 6.644559349874869e-05,
"loss": 0.5366,
"step": 105
},
{
"epoch": 1.737704918032787,
"grad_norm": 1.1330330672560118,
"learning_rate": 6.609977228934838e-05,
"loss": 0.522,
"step": 106
},
{
"epoch": 1.7540983606557377,
"grad_norm": 0.8960440091301521,
"learning_rate": 6.575052000203386e-05,
"loss": 0.5321,
"step": 107
},
{
"epoch": 1.7704918032786885,
"grad_norm": 0.8001970911153957,
"learning_rate": 6.539788254953557e-05,
"loss": 0.5292,
"step": 108
},
{
"epoch": 1.7868852459016393,
"grad_norm": 0.826768977068413,
"learning_rate": 6.504190628959798e-05,
"loss": 0.5266,
"step": 109
},
{
"epoch": 1.8032786885245902,
"grad_norm": 1.047081785970805,
"learning_rate": 6.468263801888546e-05,
"loss": 0.5249,
"step": 110
},
{
"epoch": 1.819672131147541,
"grad_norm": 1.2237949792369915,
"learning_rate": 6.432012496683032e-05,
"loss": 0.5252,
"step": 111
},
{
"epoch": 1.8360655737704918,
"grad_norm": 0.5326723853060618,
"learning_rate": 6.395441478942404e-05,
"loss": 0.5286,
"step": 112
},
{
"epoch": 1.8524590163934427,
"grad_norm": 0.8184475093441959,
"learning_rate": 6.358555556295252e-05,
"loss": 0.5271,
"step": 113
},
{
"epoch": 1.8688524590163933,
"grad_norm": 1.0030170729677295,
"learning_rate": 6.321359577767574e-05,
"loss": 0.5355,
"step": 114
},
{
"epoch": 1.8852459016393444,
"grad_norm": 0.8438447158790536,
"learning_rate": 6.283858433145345e-05,
"loss": 0.5256,
"step": 115
},
{
"epoch": 1.901639344262295,
"grad_norm": 0.6775475091945453,
"learning_rate": 6.246057052331693e-05,
"loss": 0.524,
"step": 116
},
{
"epoch": 1.918032786885246,
"grad_norm": 0.5277047106995992,
"learning_rate": 6.207960404698818e-05,
"loss": 0.5243,
"step": 117
},
{
"epoch": 1.9344262295081966,
"grad_norm": 0.5241956608413472,
"learning_rate": 6.169573498434717e-05,
"loss": 0.5257,
"step": 118
},
{
"epoch": 1.9508196721311475,
"grad_norm": 0.6035995248402194,
"learning_rate": 6.130901379884808e-05,
"loss": 0.5202,
"step": 119
},
{
"epoch": 1.9672131147540983,
"grad_norm": 0.779564589475285,
"learning_rate": 6.091949132888534e-05,
"loss": 0.527,
"step": 120
},
{
"epoch": 1.9836065573770492,
"grad_norm": 1.022288473196731,
"learning_rate": 6.052721878111045e-05,
"loss": 0.5281,
"step": 121
},
{
"epoch": 2.0,
"grad_norm": 1.2764091263174406,
"learning_rate": 6.0132247723700324e-05,
"loss": 0.5128,
"step": 122
},
{
"epoch": 2.0163934426229506,
"grad_norm": 0.7708344672192273,
"learning_rate": 5.973463007957817e-05,
"loss": 0.4906,
"step": 123
},
{
"epoch": 2.0327868852459017,
"grad_norm": 0.7977537199838536,
"learning_rate": 5.933441811958764e-05,
"loss": 0.4939,
"step": 124
},
{
"epoch": 2.0491803278688523,
"grad_norm": 1.0793702753193162,
"learning_rate": 5.8931664455621345e-05,
"loss": 0.4952,
"step": 125
},
{
"epoch": 2.0655737704918034,
"grad_norm": 0.8986778009783392,
"learning_rate": 5.8526422033704494e-05,
"loss": 0.4892,
"step": 126
},
{
"epoch": 2.081967213114754,
"grad_norm": 1.030605766210909,
"learning_rate": 5.811874412703457e-05,
"loss": 0.4903,
"step": 127
},
{
"epoch": 2.098360655737705,
"grad_norm": 1.0660924489319643,
"learning_rate": 5.770868432897809e-05,
"loss": 0.4874,
"step": 128
},
{
"epoch": 2.1147540983606556,
"grad_norm": 0.940262946228767,
"learning_rate": 5.729629654602513e-05,
"loss": 0.4829,
"step": 129
},
{
"epoch": 2.1311475409836067,
"grad_norm": 0.9888397367489762,
"learning_rate": 5.6881634990702855e-05,
"loss": 0.4918,
"step": 130
},
{
"epoch": 2.1475409836065573,
"grad_norm": 0.8484560909641439,
"learning_rate": 5.646475417444873e-05,
"loss": 0.4901,
"step": 131
},
{
"epoch": 2.1639344262295084,
"grad_norm": 0.521192253144461,
"learning_rate": 5.6045708900444364e-05,
"loss": 0.4949,
"step": 132
},
{
"epoch": 2.180327868852459,
"grad_norm": 0.771777999740359,
"learning_rate": 5.5624554256411175e-05,
"loss": 0.4811,
"step": 133
},
{
"epoch": 2.19672131147541,
"grad_norm": 0.8227491310955285,
"learning_rate": 5.520134560736852e-05,
"loss": 0.4885,
"step": 134
},
{
"epoch": 2.2131147540983607,
"grad_norm": 0.5992373916210076,
"learning_rate": 5.477613858835536e-05,
"loss": 0.4872,
"step": 135
},
{
"epoch": 2.2295081967213113,
"grad_norm": 0.5052047355472387,
"learning_rate": 5.434898909711656e-05,
"loss": 0.4805,
"step": 136
},
{
"epoch": 2.2459016393442623,
"grad_norm": 0.7157168717143663,
"learning_rate": 5.3919953286754477e-05,
"loss": 0.4825,
"step": 137
},
{
"epoch": 2.262295081967213,
"grad_norm": 0.5410573992811272,
"learning_rate": 5.348908755834716e-05,
"loss": 0.4834,
"step": 138
},
{
"epoch": 2.278688524590164,
"grad_norm": 0.4166121333458553,
"learning_rate": 5.305644855353378e-05,
"loss": 0.4833,
"step": 139
},
{
"epoch": 2.2950819672131146,
"grad_norm": 0.5233638295240889,
"learning_rate": 5.262209314706856e-05,
"loss": 0.4844,
"step": 140
},
{
"epoch": 2.3114754098360657,
"grad_norm": 0.4090985726833338,
"learning_rate": 5.218607843934401e-05,
"loss": 0.487,
"step": 141
},
{
"epoch": 2.3278688524590163,
"grad_norm": 0.363420954496579,
"learning_rate": 5.1748461748884476e-05,
"loss": 0.4774,
"step": 142
},
{
"epoch": 2.3442622950819674,
"grad_norm": 0.4690278843251207,
"learning_rate": 5.130930060481113e-05,
"loss": 0.4811,
"step": 143
},
{
"epoch": 2.360655737704918,
"grad_norm": 0.46496866291592687,
"learning_rate": 5.0868652739279045e-05,
"loss": 0.4799,
"step": 144
},
{
"epoch": 2.3770491803278686,
"grad_norm": 0.4067588783945542,
"learning_rate": 5.042657607988787e-05,
"loss": 0.4884,
"step": 145
},
{
"epoch": 2.3934426229508197,
"grad_norm": 0.3249832363624369,
"learning_rate": 4.9983128742066553e-05,
"loss": 0.4819,
"step": 146
},
{
"epoch": 2.4098360655737707,
"grad_norm": 0.3155267675853147,
"learning_rate": 4.953836902143354e-05,
"loss": 0.4817,
"step": 147
},
{
"epoch": 2.4262295081967213,
"grad_norm": 0.30276002082546505,
"learning_rate": 4.909235538613321e-05,
"loss": 0.4817,
"step": 148
},
{
"epoch": 2.442622950819672,
"grad_norm": 0.3578858890853383,
"learning_rate": 4.864514646914968e-05,
"loss": 0.4855,
"step": 149
},
{
"epoch": 2.459016393442623,
"grad_norm": 0.28701615085250826,
"learning_rate": 4.819680106059884e-05,
"loss": 0.4829,
"step": 150
},
{
"epoch": 2.4754098360655736,
"grad_norm": 0.37224620973373557,
"learning_rate": 4.7747378099999866e-05,
"loss": 0.491,
"step": 151
},
{
"epoch": 2.4918032786885247,
"grad_norm": 0.34012579144317995,
"learning_rate": 4.729693666852699e-05,
"loss": 0.4884,
"step": 152
},
{
"epoch": 2.5081967213114753,
"grad_norm": 0.3774404172171865,
"learning_rate": 4.684553598124272e-05,
"loss": 0.4851,
"step": 153
},
{
"epoch": 2.5245901639344264,
"grad_norm": 0.3549426068916513,
"learning_rate": 4.639323537931339e-05,
"loss": 0.4819,
"step": 154
},
{
"epoch": 2.540983606557377,
"grad_norm": 0.39031877291102196,
"learning_rate": 4.5940094322208156e-05,
"loss": 0.4905,
"step": 155
},
{
"epoch": 2.557377049180328,
"grad_norm": 0.295463066655204,
"learning_rate": 4.5486172379882495e-05,
"loss": 0.4765,
"step": 156
},
{
"epoch": 2.5737704918032787,
"grad_norm": 0.29781720638326936,
"learning_rate": 4.5031529224947106e-05,
"loss": 0.4852,
"step": 157
},
{
"epoch": 2.5901639344262293,
"grad_norm": 0.2852567559187638,
"learning_rate": 4.4576224624823334e-05,
"loss": 0.4836,
"step": 158
},
{
"epoch": 2.6065573770491803,
"grad_norm": 0.239161233813967,
"learning_rate": 4.412031843388618e-05,
"loss": 0.4889,
"step": 159
},
{
"epoch": 2.6229508196721314,
"grad_norm": 0.37163233149883945,
"learning_rate": 4.36638705855958e-05,
"loss": 0.481,
"step": 160
},
{
"epoch": 2.639344262295082,
"grad_norm": 0.35373235092972494,
"learning_rate": 4.320694108461868e-05,
"loss": 0.4738,
"step": 161
},
{
"epoch": 2.6557377049180326,
"grad_norm": 0.3602393611325291,
"learning_rate": 4.2749589998939366e-05,
"loss": 0.4783,
"step": 162
},
{
"epoch": 2.6721311475409837,
"grad_norm": 0.33028586314976544,
"learning_rate": 4.2291877451964e-05,
"loss": 0.4807,
"step": 163
},
{
"epoch": 2.6885245901639343,
"grad_norm": 0.27454726727065487,
"learning_rate": 4.183386361461643e-05,
"loss": 0.4754,
"step": 164
},
{
"epoch": 2.7049180327868854,
"grad_norm": 0.2201982880819997,
"learning_rate": 4.137560869742816e-05,
"loss": 0.4816,
"step": 165
},
{
"epoch": 2.721311475409836,
"grad_norm": 0.24024736204090635,
"learning_rate": 4.0917172942623074e-05,
"loss": 0.477,
"step": 166
},
{
"epoch": 2.737704918032787,
"grad_norm": 0.2774042453950074,
"learning_rate": 4.045861661619796e-05,
"loss": 0.4863,
"step": 167
},
{
"epoch": 2.7540983606557377,
"grad_norm": 0.25031478188990913,
"learning_rate": 4e-05,
"loss": 0.4806,
"step": 168
},
{
"epoch": 2.7704918032786887,
"grad_norm": 0.3146343902104566,
"learning_rate": 3.954138338380204e-05,
"loss": 0.4819,
"step": 169
},
{
"epoch": 2.7868852459016393,
"grad_norm": 0.29476137692022925,
"learning_rate": 3.908282705737694e-05,
"loss": 0.4829,
"step": 170
},
{
"epoch": 2.80327868852459,
"grad_norm": 0.2781979864605831,
"learning_rate": 3.862439130257184e-05,
"loss": 0.4842,
"step": 171
},
{
"epoch": 2.819672131147541,
"grad_norm": 0.2812486506160709,
"learning_rate": 3.8166136385383585e-05,
"loss": 0.4848,
"step": 172
},
{
"epoch": 2.836065573770492,
"grad_norm": 0.2443137176068297,
"learning_rate": 3.770812254803601e-05,
"loss": 0.4814,
"step": 173
},
{
"epoch": 2.8524590163934427,
"grad_norm": 0.2139764225453824,
"learning_rate": 3.725041000106065e-05,
"loss": 0.4793,
"step": 174
},
{
"epoch": 2.8688524590163933,
"grad_norm": 0.2771342287766635,
"learning_rate": 3.679305891538133e-05,
"loss": 0.466,
"step": 175
},
{
"epoch": 2.8852459016393444,
"grad_norm": 0.26719263531966153,
"learning_rate": 3.6336129414404214e-05,
"loss": 0.4818,
"step": 176
},
{
"epoch": 2.901639344262295,
"grad_norm": 0.21705964839463288,
"learning_rate": 3.5879681566113836e-05,
"loss": 0.4841,
"step": 177
},
{
"epoch": 2.918032786885246,
"grad_norm": 0.20964241906287742,
"learning_rate": 3.5423775375176686e-05,
"loss": 0.4857,
"step": 178
},
{
"epoch": 2.9344262295081966,
"grad_norm": 0.19494840468268942,
"learning_rate": 3.4968470775052914e-05,
"loss": 0.475,
"step": 179
},
{
"epoch": 2.9508196721311473,
"grad_norm": 0.19411352485607267,
"learning_rate": 3.451382762011753e-05,
"loss": 0.4765,
"step": 180
},
{
"epoch": 2.9672131147540983,
"grad_norm": 0.2389081408757777,
"learning_rate": 3.405990567779185e-05,
"loss": 0.4843,
"step": 181
},
{
"epoch": 2.9836065573770494,
"grad_norm": 0.19939844528891387,
"learning_rate": 3.3606764620686614e-05,
"loss": 0.4807,
"step": 182
},
{
"epoch": 3.0,
"grad_norm": 0.23959467510788,
"learning_rate": 3.3154464018757284e-05,
"loss": 0.483,
"step": 183
},
{
"epoch": 3.0163934426229506,
"grad_norm": 0.20927947022485016,
"learning_rate": 3.270306333147301e-05,
"loss": 0.4458,
"step": 184
},
{
"epoch": 3.0327868852459017,
"grad_norm": 0.2064164028547752,
"learning_rate": 3.225262190000015e-05,
"loss": 0.46,
"step": 185
},
{
"epoch": 3.0491803278688523,
"grad_norm": 0.22391411463107186,
"learning_rate": 3.180319893940117e-05,
"loss": 0.4488,
"step": 186
},
{
"epoch": 3.0655737704918034,
"grad_norm": 0.21718594350542614,
"learning_rate": 3.135485353085033e-05,
"loss": 0.4487,
"step": 187
},
{
"epoch": 3.081967213114754,
"grad_norm": 0.2467699050215421,
"learning_rate": 3.0907644613866795e-05,
"loss": 0.4466,
"step": 188
},
{
"epoch": 3.098360655737705,
"grad_norm": 0.23078659838474802,
"learning_rate": 3.0461630978566473e-05,
"loss": 0.4423,
"step": 189
},
{
"epoch": 3.1147540983606556,
"grad_norm": 0.2370592376095119,
"learning_rate": 3.0016871257933457e-05,
"loss": 0.4438,
"step": 190
},
{
"epoch": 3.1311475409836067,
"grad_norm": 0.21570100983782378,
"learning_rate": 2.9573423920112155e-05,
"loss": 0.4455,
"step": 191
},
{
"epoch": 3.1475409836065573,
"grad_norm": 0.21407395891268235,
"learning_rate": 2.9131347260720965e-05,
"loss": 0.4463,
"step": 192
},
{
"epoch": 3.1639344262295084,
"grad_norm": 0.20617008048319857,
"learning_rate": 2.8690699395188873e-05,
"loss": 0.4465,
"step": 193
},
{
"epoch": 3.180327868852459,
"grad_norm": 0.2125529545129492,
"learning_rate": 2.825153825111553e-05,
"loss": 0.445,
"step": 194
},
{
"epoch": 3.19672131147541,
"grad_norm": 0.2002948390041686,
"learning_rate": 2.7813921560656003e-05,
"loss": 0.4472,
"step": 195
},
{
"epoch": 3.2131147540983607,
"grad_norm": 0.19399060354592243,
"learning_rate": 2.737790685293145e-05,
"loss": 0.4384,
"step": 196
},
{
"epoch": 3.2295081967213113,
"grad_norm": 0.20413317551543775,
"learning_rate": 2.694355144646622e-05,
"loss": 0.4474,
"step": 197
},
{
"epoch": 3.2459016393442623,
"grad_norm": 0.178524571287303,
"learning_rate": 2.6510912441652855e-05,
"loss": 0.4476,
"step": 198
},
{
"epoch": 3.262295081967213,
"grad_norm": 0.21495762020528217,
"learning_rate": 2.6080046713245527e-05,
"loss": 0.4481,
"step": 199
},
{
"epoch": 3.278688524590164,
"grad_norm": 0.17779692741231456,
"learning_rate": 2.5651010902883457e-05,
"loss": 0.4447,
"step": 200
},
{
"epoch": 3.2950819672131146,
"grad_norm": 0.19536488579919992,
"learning_rate": 2.5223861411644647e-05,
"loss": 0.4406,
"step": 201
},
{
"epoch": 3.3114754098360657,
"grad_norm": 0.21534013434506138,
"learning_rate": 2.4798654392631502e-05,
"loss": 0.443,
"step": 202
},
{
"epoch": 3.3278688524590163,
"grad_norm": 0.17678888902216489,
"learning_rate": 2.4375445743588832e-05,
"loss": 0.4481,
"step": 203
},
{
"epoch": 3.3442622950819674,
"grad_norm": 0.21881378852102576,
"learning_rate": 2.3954291099555652e-05,
"loss": 0.4484,
"step": 204
},
{
"epoch": 3.360655737704918,
"grad_norm": 0.1911371536452111,
"learning_rate": 2.3535245825551285e-05,
"loss": 0.4407,
"step": 205
},
{
"epoch": 3.3770491803278686,
"grad_norm": 0.1822153835411314,
"learning_rate": 2.3118365009297145e-05,
"loss": 0.4474,
"step": 206
},
{
"epoch": 3.3934426229508197,
"grad_norm": 0.1967826759425408,
"learning_rate": 2.2703703453974883e-05,
"loss": 0.4466,
"step": 207
},
{
"epoch": 3.4098360655737707,
"grad_norm": 0.1787783354685364,
"learning_rate": 2.2291315671021922e-05,
"loss": 0.4417,
"step": 208
},
{
"epoch": 3.4262295081967213,
"grad_norm": 0.18368967889880147,
"learning_rate": 2.1881255872965434e-05,
"loss": 0.4438,
"step": 209
},
{
"epoch": 3.442622950819672,
"grad_norm": 0.16138571200714233,
"learning_rate": 2.1473577966295506e-05,
"loss": 0.4479,
"step": 210
},
{
"epoch": 3.459016393442623,
"grad_norm": 0.15851838029612553,
"learning_rate": 2.1068335544378672e-05,
"loss": 0.4511,
"step": 211
},
{
"epoch": 3.4754098360655736,
"grad_norm": 0.15179774543882815,
"learning_rate": 2.0665581880412366e-05,
"loss": 0.4447,
"step": 212
},
{
"epoch": 3.4918032786885247,
"grad_norm": 0.15563700688196921,
"learning_rate": 2.026536992042184e-05,
"loss": 0.4456,
"step": 213
},
{
"epoch": 3.5081967213114753,
"grad_norm": 0.14825148109958275,
"learning_rate": 1.9867752276299686e-05,
"loss": 0.4472,
"step": 214
},
{
"epoch": 3.5245901639344264,
"grad_norm": 0.12432896908408256,
"learning_rate": 1.9472781218889565e-05,
"loss": 0.4491,
"step": 215
},
{
"epoch": 3.540983606557377,
"grad_norm": 0.14986131261136762,
"learning_rate": 1.908050867111468e-05,
"loss": 0.447,
"step": 216
},
{
"epoch": 3.557377049180328,
"grad_norm": 0.1280996425182516,
"learning_rate": 1.869098620115194e-05,
"loss": 0.4459,
"step": 217
},
{
"epoch": 3.5737704918032787,
"grad_norm": 0.15181510097669454,
"learning_rate": 1.8304265015652835e-05,
"loss": 0.444,
"step": 218
},
{
"epoch": 3.5901639344262293,
"grad_norm": 0.1402130102169863,
"learning_rate": 1.792039595301183e-05,
"loss": 0.4521,
"step": 219
},
{
"epoch": 3.6065573770491803,
"grad_norm": 0.16291621368769718,
"learning_rate": 1.7539429476683088e-05,
"loss": 0.447,
"step": 220
},
{
"epoch": 3.6229508196721314,
"grad_norm": 0.13943515675692342,
"learning_rate": 1.7161415668546557e-05,
"loss": 0.4411,
"step": 221
},
{
"epoch": 3.639344262295082,
"grad_norm": 0.1419363610523589,
"learning_rate": 1.6786404222324278e-05,
"loss": 0.442,
"step": 222
},
{
"epoch": 3.6557377049180326,
"grad_norm": 0.13518707011783945,
"learning_rate": 1.641444443704749e-05,
"loss": 0.444,
"step": 223
},
{
"epoch": 3.6721311475409837,
"grad_norm": 0.12940286831893916,
"learning_rate": 1.6045585210575955e-05,
"loss": 0.448,
"step": 224
},
{
"epoch": 3.6885245901639343,
"grad_norm": 0.15612865445371793,
"learning_rate": 1.56798750331697e-05,
"loss": 0.4439,
"step": 225
},
{
"epoch": 3.7049180327868854,
"grad_norm": 0.13228999124818513,
"learning_rate": 1.531736198111456e-05,
"loss": 0.4402,
"step": 226
},
{
"epoch": 3.721311475409836,
"grad_norm": 0.14034136382445142,
"learning_rate": 1.4958093710402036e-05,
"loss": 0.4394,
"step": 227
},
{
"epoch": 3.737704918032787,
"grad_norm": 0.14905661892566827,
"learning_rate": 1.4602117450464453e-05,
"loss": 0.4481,
"step": 228
},
{
"epoch": 3.7540983606557377,
"grad_norm": 0.16177536977164272,
"learning_rate": 1.4249479997966145e-05,
"loss": 0.4457,
"step": 229
},
{
"epoch": 3.7704918032786887,
"grad_norm": 0.12677113397589643,
"learning_rate": 1.3900227710651647e-05,
"loss": 0.4445,
"step": 230
},
{
"epoch": 3.7868852459016393,
"grad_norm": 0.17192511217282086,
"learning_rate": 1.3554406501251332e-05,
"loss": 0.4459,
"step": 231
},
{
"epoch": 3.80327868852459,
"grad_norm": 0.11920529678713093,
"learning_rate": 1.3212061831445913e-05,
"loss": 0.4382,
"step": 232
},
{
"epoch": 3.819672131147541,
"grad_norm": 0.12712529888330631,
"learning_rate": 1.2873238705889968e-05,
"loss": 0.441,
"step": 233
},
{
"epoch": 3.836065573770492,
"grad_norm": 0.12199439925977827,
"learning_rate": 1.2537981666295536e-05,
"loss": 0.442,
"step": 234
},
{
"epoch": 3.8524590163934427,
"grad_norm": 0.13371713843483143,
"learning_rate": 1.2206334785576822e-05,
"loss": 0.4381,
"step": 235
},
{
"epoch": 3.8688524590163933,
"grad_norm": 0.11420461356422432,
"learning_rate": 1.1878341662056259e-05,
"loss": 0.4442,
"step": 236
},
{
"epoch": 3.8852459016393444,
"grad_norm": 0.11704524464444101,
"learning_rate": 1.1554045413733097e-05,
"loss": 0.4435,
"step": 237
},
{
"epoch": 3.901639344262295,
"grad_norm": 0.11799803709109379,
"learning_rate": 1.123348867261512e-05,
"loss": 0.4444,
"step": 238
},
{
"epoch": 3.918032786885246,
"grad_norm": 0.12095672068157937,
"learning_rate": 1.0916713579114218e-05,
"loss": 0.4425,
"step": 239
},
{
"epoch": 3.9344262295081966,
"grad_norm": 0.10579439481053957,
"learning_rate": 1.0603761776506616e-05,
"loss": 0.4448,
"step": 240
},
{
"epoch": 3.9508196721311473,
"grad_norm": 0.12927283153010521,
"learning_rate": 1.029467440545849e-05,
"loss": 0.444,
"step": 241
},
{
"epoch": 3.9672131147540983,
"grad_norm": 0.11979578396485492,
"learning_rate": 9.989492098617494e-06,
"loss": 0.4462,
"step": 242
},
{
"epoch": 3.9836065573770494,
"grad_norm": 0.11270512581429702,
"learning_rate": 9.688254975271327e-06,
"loss": 0.4439,
"step": 243
},
{
"epoch": 4.0,
"grad_norm": 0.12212819186751334,
"learning_rate": 9.391002636073589e-06,
"loss": 0.427,
"step": 244
},
{
"epoch": 4.016393442622951,
"grad_norm": 0.15550079584648527,
"learning_rate": 9.0977741578378e-06,
"loss": 0.4269,
"step": 245
},
{
"epoch": 4.032786885245901,
"grad_norm": 0.13688670983262652,
"learning_rate": 8.808608088400522e-06,
"loss": 0.4269,
"step": 246
},
{
"epoch": 4.049180327868853,
"grad_norm": 0.11888941041824816,
"learning_rate": 8.523542441553712e-06,
"loss": 0.4212,
"step": 247
},
{
"epoch": 4.065573770491803,
"grad_norm": 0.13424666760515866,
"learning_rate": 8.242614692047507e-06,
"loss": 0.4288,
"step": 248
},
{
"epoch": 4.081967213114754,
"grad_norm": 0.15658242982421974,
"learning_rate": 7.96586177066375e-06,
"loss": 0.4234,
"step": 249
},
{
"epoch": 4.098360655737705,
"grad_norm": 0.14858315425443683,
"learning_rate": 7.693320059361107e-06,
"loss": 0.4236,
"step": 250
},
{
"epoch": 4.114754098360656,
"grad_norm": 0.1432177280213629,
"learning_rate": 7.425025386492231e-06,
"loss": 0.4324,
"step": 251
},
{
"epoch": 4.131147540983607,
"grad_norm": 0.1363328212537908,
"learning_rate": 7.161013022093861e-06,
"loss": 0.4305,
"step": 252
},
{
"epoch": 4.147540983606557,
"grad_norm": 0.13831059143305544,
"learning_rate": 6.901317673250089e-06,
"loss": 0.4198,
"step": 253
},
{
"epoch": 4.163934426229508,
"grad_norm": 0.1479605646882516,
"learning_rate": 6.645973479529892e-06,
"loss": 0.4275,
"step": 254
},
{
"epoch": 4.180327868852459,
"grad_norm": 0.12133684415312865,
"learning_rate": 6.395014008499067e-06,
"loss": 0.4286,
"step": 255
},
{
"epoch": 4.19672131147541,
"grad_norm": 0.11552099846590175,
"learning_rate": 6.148472251307436e-06,
"loss": 0.4249,
"step": 256
},
{
"epoch": 4.213114754098361,
"grad_norm": 0.1286344544278559,
"learning_rate": 5.906380618351893e-06,
"loss": 0.4266,
"step": 257
},
{
"epoch": 4.229508196721311,
"grad_norm": 0.12137839353942781,
"learning_rate": 5.668770935015641e-06,
"loss": 0.4193,
"step": 258
},
{
"epoch": 4.245901639344262,
"grad_norm": 0.10879456442085193,
"learning_rate": 5.435674437484491e-06,
"loss": 0.4213,
"step": 259
},
{
"epoch": 4.262295081967213,
"grad_norm": 0.10654250511384006,
"learning_rate": 5.207121768640519e-06,
"loss": 0.415,
"step": 260
},
{
"epoch": 4.278688524590164,
"grad_norm": 0.11634254635354951,
"learning_rate": 4.983142974033764e-06,
"loss": 0.4278,
"step": 261
},
{
"epoch": 4.295081967213115,
"grad_norm": 0.1188048854614297,
"learning_rate": 4.763767497932419e-06,
"loss": 0.4186,
"step": 262
},
{
"epoch": 4.311475409836065,
"grad_norm": 0.10499559306706431,
"learning_rate": 4.549024179452151e-06,
"loss": 0.4192,
"step": 263
},
{
"epoch": 4.327868852459017,
"grad_norm": 0.10258020778548946,
"learning_rate": 4.338941248764781e-06,
"loss": 0.4221,
"step": 264
},
{
"epoch": 4.344262295081967,
"grad_norm": 0.10174308704204628,
"learning_rate": 4.133546323387281e-06,
"loss": 0.4291,
"step": 265
},
{
"epoch": 4.360655737704918,
"grad_norm": 0.09609276817159343,
"learning_rate": 3.932866404551092e-06,
"loss": 0.4183,
"step": 266
},
{
"epoch": 4.377049180327869,
"grad_norm": 0.11045174230896027,
"learning_rate": 3.7369278736525584e-06,
"loss": 0.4237,
"step": 267
},
{
"epoch": 4.39344262295082,
"grad_norm": 0.09919053798520305,
"learning_rate": 3.545756488784826e-06,
"loss": 0.4247,
"step": 268
},
{
"epoch": 4.409836065573771,
"grad_norm": 0.10170621462280902,
"learning_rate": 3.359377381351707e-06,
"loss": 0.4162,
"step": 269
},
{
"epoch": 4.426229508196721,
"grad_norm": 0.09599940271102718,
"learning_rate": 3.1778150527638664e-06,
"loss": 0.425,
"step": 270
},
{
"epoch": 4.442622950819672,
"grad_norm": 0.09230359083631827,
"learning_rate": 3.0010933712178735e-06,
"loss": 0.425,
"step": 271
},
{
"epoch": 4.459016393442623,
"grad_norm": 0.09161534788091766,
"learning_rate": 2.8292355685585147e-06,
"loss": 0.431,
"step": 272
},
{
"epoch": 4.475409836065574,
"grad_norm": 0.09181661135879722,
"learning_rate": 2.6622642372246874e-06,
"loss": 0.4202,
"step": 273
},
{
"epoch": 4.491803278688525,
"grad_norm": 0.09201200891820971,
"learning_rate": 2.5002013272794654e-06,
"loss": 0.4199,
"step": 274
},
{
"epoch": 4.508196721311475,
"grad_norm": 0.09965795343039312,
"learning_rate": 2.3430681435244386e-06,
"loss": 0.4268,
"step": 275
},
{
"epoch": 4.524590163934426,
"grad_norm": 0.09118520040005619,
"learning_rate": 2.1908853426990983e-06,
"loss": 0.4251,
"step": 276
},
{
"epoch": 4.540983606557377,
"grad_norm": 0.0928501159245484,
"learning_rate": 2.0436729307652213e-06,
"loss": 0.4213,
"step": 277
},
{
"epoch": 4.557377049180328,
"grad_norm": 0.09183472059851619,
"learning_rate": 1.9014502602769003e-06,
"loss": 0.4245,
"step": 278
},
{
"epoch": 4.573770491803279,
"grad_norm": 0.08782980102005061,
"learning_rate": 1.7642360278364767e-06,
"loss": 0.4233,
"step": 279
},
{
"epoch": 4.590163934426229,
"grad_norm": 0.09375666869538167,
"learning_rate": 1.6320482716366726e-06,
"loss": 0.425,
"step": 280
},
{
"epoch": 4.60655737704918,
"grad_norm": 0.08785996758165288,
"learning_rate": 1.5049043690892729e-06,
"loss": 0.4145,
"step": 281
},
{
"epoch": 4.622950819672131,
"grad_norm": 0.09413477632311608,
"learning_rate": 1.3828210345407268e-06,
"loss": 0.4216,
"step": 282
},
{
"epoch": 4.639344262295082,
"grad_norm": 0.09447484104691518,
"learning_rate": 1.2658143170748383e-06,
"loss": 0.4201,
"step": 283
},
{
"epoch": 4.655737704918033,
"grad_norm": 0.09229151843804058,
"learning_rate": 1.1538995984029654e-06,
"loss": 0.4163,
"step": 284
},
{
"epoch": 4.672131147540983,
"grad_norm": 0.08568358284035806,
"learning_rate": 1.047091590841971e-06,
"loss": 0.4263,
"step": 285
},
{
"epoch": 4.688524590163935,
"grad_norm": 0.09023433045860603,
"learning_rate": 9.454043353800801e-07,
"loss": 0.4221,
"step": 286
},
{
"epoch": 4.704918032786885,
"grad_norm": 0.0844734735476981,
"learning_rate": 8.488511998311133e-07,
"loss": 0.422,
"step": 287
},
{
"epoch": 4.721311475409836,
"grad_norm": 0.08564054109423068,
"learning_rate": 7.574448770771003e-07,
"loss": 0.4226,
"step": 288
},
{
"epoch": 4.737704918032787,
"grad_norm": 0.08653160576715448,
"learning_rate": 6.711973833997088e-07,
"loss": 0.4179,
"step": 289
},
{
"epoch": 4.754098360655737,
"grad_norm": 0.08383004695421666,
"learning_rate": 5.90120056900565e-07,
"loss": 0.415,
"step": 290
},
{
"epoch": 4.770491803278689,
"grad_norm": 0.08834395449932624,
"learning_rate": 5.142235560107623e-07,
"loss": 0.4263,
"step": 291
},
{
"epoch": 4.786885245901639,
"grad_norm": 0.08320061631305875,
"learning_rate": 4.4351785808967485e-07,
"loss": 0.4272,
"step": 292
},
{
"epoch": 4.80327868852459,
"grad_norm": 0.08304223152666458,
"learning_rate": 3.780122581133672e-07,
"loss": 0.4223,
"step": 293
},
{
"epoch": 4.8196721311475414,
"grad_norm": 0.08927719387363141,
"learning_rate": 3.177153674526556e-07,
"loss": 0.4262,
"step": 294
},
{
"epoch": 4.836065573770492,
"grad_norm": 0.08297799431829674,
"learning_rate": 2.626351127410498e-07,
"loss": 0.4223,
"step": 295
},
{
"epoch": 4.852459016393443,
"grad_norm": 0.08318865729803561,
"learning_rate": 2.1277873483274148e-07,
"loss": 0.4259,
"step": 296
},
{
"epoch": 4.868852459016393,
"grad_norm": 0.0828464753884451,
"learning_rate": 1.6815278785069944e-07,
"loss": 0.4235,
"step": 297
},
{
"epoch": 4.885245901639344,
"grad_norm": 0.08134276640655788,
"learning_rate": 1.2876313832508314e-07,
"loss": 0.4193,
"step": 298
},
{
"epoch": 4.901639344262295,
"grad_norm": 0.08124195226061579,
"learning_rate": 9.461496442199735e-08,
"loss": 0.417,
"step": 299
},
{
"epoch": 4.918032786885246,
"grad_norm": 0.08525708385313023,
"learning_rate": 6.571275526281451e-08,
"loss": 0.4301,
"step": 300
},
{
"epoch": 4.934426229508197,
"grad_norm": 0.08475437381287206,
"learning_rate": 4.206031033399338e-08,
"loss": 0.4309,
"step": 301
},
{
"epoch": 4.950819672131147,
"grad_norm": 0.08111110985986,
"learning_rate": 2.366073898761645e-08,
"loss": 0.4143,
"step": 302
},
{
"epoch": 4.967213114754099,
"grad_norm": 0.08973114402396704,
"learning_rate": 1.0516460032641284e-08,
"loss": 0.4301,
"step": 303
},
{
"epoch": 4.983606557377049,
"grad_norm": 0.07995200669593748,
"learning_rate": 2.6292014169104764e-09,
"loss": 0.418,
"step": 304
},
{
"epoch": 5.0,
"grad_norm": 0.0879869455035565,
"learning_rate": 0.0,
"loss": 0.4146,
"step": 305
},
{
"epoch": 5.0,
"step": 305,
"total_flos": 5116723388743680.0,
"train_loss": 0.5099891699728418,
"train_runtime": 4463.8352,
"train_samples_per_second": 34.825,
"train_steps_per_second": 0.068
}
],
"logging_steps": 1,
"max_steps": 305,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5116723388743680.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}