v0.8.8-adapter / trainer_state.json
gotzmann's picture
..
9d15054
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.1873840445269017,
"eval_steps": 500,
"global_step": 160,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01,
"grad_norm": 0.01736689789892289,
"learning_rate": 7.142857142857143e-06,
"loss": 1.3968,
"step": 1
},
{
"epoch": 0.01,
"grad_norm": 0.017027077372532427,
"learning_rate": 1.4285714285714285e-05,
"loss": 1.4007,
"step": 2
},
{
"epoch": 0.02,
"grad_norm": 0.016201064682232102,
"learning_rate": 2.1428571428571428e-05,
"loss": 1.4031,
"step": 3
},
{
"epoch": 0.03,
"grad_norm": 0.017052502856808767,
"learning_rate": 2.857142857142857e-05,
"loss": 1.4268,
"step": 4
},
{
"epoch": 0.04,
"grad_norm": 0.017589796386440718,
"learning_rate": 3.571428571428572e-05,
"loss": 1.4357,
"step": 5
},
{
"epoch": 0.04,
"grad_norm": 0.018257751591257577,
"learning_rate": 4.2857142857142856e-05,
"loss": 1.4047,
"step": 6
},
{
"epoch": 0.05,
"grad_norm": 0.018295865167676045,
"learning_rate": 5e-05,
"loss": 1.4105,
"step": 7
},
{
"epoch": 0.06,
"grad_norm": 0.02117599813759531,
"learning_rate": 5.714285714285714e-05,
"loss": 1.3693,
"step": 8
},
{
"epoch": 0.07,
"grad_norm": 0.02205957316713292,
"learning_rate": 6.428571428571429e-05,
"loss": 1.3524,
"step": 9
},
{
"epoch": 0.07,
"grad_norm": 0.023657076230798574,
"learning_rate": 7.142857142857143e-05,
"loss": 1.3913,
"step": 10
},
{
"epoch": 0.08,
"grad_norm": 0.027315806443124473,
"learning_rate": 7.857142857142858e-05,
"loss": 1.4036,
"step": 11
},
{
"epoch": 0.09,
"grad_norm": 0.028918483559492007,
"learning_rate": 8.571428571428571e-05,
"loss": 1.4478,
"step": 12
},
{
"epoch": 0.1,
"grad_norm": 0.03141992210034351,
"learning_rate": 9.285714285714286e-05,
"loss": 1.3865,
"step": 13
},
{
"epoch": 0.1,
"grad_norm": 0.02716522203126749,
"learning_rate": 0.0001,
"loss": 1.3674,
"step": 14
},
{
"epoch": 0.11,
"grad_norm": 0.04349952540870549,
"learning_rate": 9.999617556940085e-05,
"loss": 1.3423,
"step": 15
},
{
"epoch": 0.12,
"grad_norm": 0.02315330665074004,
"learning_rate": 9.998470286265416e-05,
"loss": 1.3017,
"step": 16
},
{
"epoch": 0.13,
"grad_norm": 0.02981766810064017,
"learning_rate": 9.996558363482277e-05,
"loss": 1.3578,
"step": 17
},
{
"epoch": 0.13,
"grad_norm": 0.018058552912596896,
"learning_rate": 9.993882081071306e-05,
"loss": 1.3604,
"step": 18
},
{
"epoch": 0.14,
"grad_norm": 0.01617412339760896,
"learning_rate": 9.99044184844276e-05,
"loss": 1.3599,
"step": 19
},
{
"epoch": 0.15,
"grad_norm": 0.01571914020346384,
"learning_rate": 9.986238191873874e-05,
"loss": 1.3314,
"step": 20
},
{
"epoch": 0.16,
"grad_norm": 0.01794066179347792,
"learning_rate": 9.98127175442836e-05,
"loss": 1.3244,
"step": 21
},
{
"epoch": 0.16,
"grad_norm": 0.020268083274328504,
"learning_rate": 9.975543295858035e-05,
"loss": 1.3502,
"step": 22
},
{
"epoch": 0.17,
"grad_norm": 0.022181555162566497,
"learning_rate": 9.969053692486583e-05,
"loss": 1.3378,
"step": 23
},
{
"epoch": 0.18,
"grad_norm": 0.021942253446061308,
"learning_rate": 9.961803937075516e-05,
"loss": 1.2991,
"step": 24
},
{
"epoch": 0.19,
"grad_norm": 0.022098681290951168,
"learning_rate": 9.953795138672291e-05,
"loss": 1.3299,
"step": 25
},
{
"epoch": 0.19,
"grad_norm": 0.019480865843562824,
"learning_rate": 9.945028522440653e-05,
"loss": 1.2898,
"step": 26
},
{
"epoch": 0.2,
"grad_norm": 0.019522365809044647,
"learning_rate": 9.93550542947322e-05,
"loss": 1.3358,
"step": 27
},
{
"epoch": 0.21,
"grad_norm": 0.01626195754263703,
"learning_rate": 9.925227316586316e-05,
"loss": 1.3236,
"step": 28
},
{
"epoch": 0.22,
"grad_norm": 0.015208995645064876,
"learning_rate": 9.91419575609712e-05,
"loss": 1.3095,
"step": 29
},
{
"epoch": 0.22,
"grad_norm": 0.014245114044458424,
"learning_rate": 9.902412435583128e-05,
"loss": 1.2887,
"step": 30
},
{
"epoch": 0.23,
"grad_norm": 0.018282343094477974,
"learning_rate": 9.889879157624002e-05,
"loss": 1.3037,
"step": 31
},
{
"epoch": 0.24,
"grad_norm": 0.013227523223555335,
"learning_rate": 9.876597839525814e-05,
"loss": 1.291,
"step": 32
},
{
"epoch": 0.24,
"grad_norm": 0.01347818214113649,
"learning_rate": 9.862570513027735e-05,
"loss": 1.351,
"step": 33
},
{
"epoch": 0.25,
"grad_norm": 0.013651640512312836,
"learning_rate": 9.847799323991234e-05,
"loss": 1.3239,
"step": 34
},
{
"epoch": 0.26,
"grad_norm": 0.013176778064524027,
"learning_rate": 9.832286532071802e-05,
"loss": 1.3285,
"step": 35
},
{
"epoch": 0.27,
"grad_norm": 0.01339867538916516,
"learning_rate": 9.816034510373286e-05,
"loss": 1.3151,
"step": 36
},
{
"epoch": 0.27,
"grad_norm": 0.013215819889291318,
"learning_rate": 9.799045745084847e-05,
"loss": 1.3401,
"step": 37
},
{
"epoch": 0.28,
"grad_norm": 0.013158439789624273,
"learning_rate": 9.781322835100638e-05,
"loss": 1.3366,
"step": 38
},
{
"epoch": 0.29,
"grad_norm": 0.012039494551754166,
"learning_rate": 9.76286849162223e-05,
"loss": 1.3187,
"step": 39
},
{
"epoch": 0.3,
"grad_norm": 0.012156687939496454,
"learning_rate": 9.743685537743856e-05,
"loss": 1.308,
"step": 40
},
{
"epoch": 0.3,
"grad_norm": 0.015724501188273832,
"learning_rate": 9.72377690802055e-05,
"loss": 1.2725,
"step": 41
},
{
"epoch": 0.31,
"grad_norm": 0.01236228699037047,
"learning_rate": 9.70314564801922e-05,
"loss": 1.2936,
"step": 42
},
{
"epoch": 0.32,
"grad_norm": 0.011472953053820326,
"learning_rate": 9.681794913852746e-05,
"loss": 1.2899,
"step": 43
},
{
"epoch": 0.33,
"grad_norm": 0.012026745769038633,
"learning_rate": 9.659727971697174e-05,
"loss": 1.2681,
"step": 44
},
{
"epoch": 0.33,
"grad_norm": 0.011343760163365182,
"learning_rate": 9.636948197292052e-05,
"loss": 1.2824,
"step": 45
},
{
"epoch": 0.34,
"grad_norm": 0.011947389090492217,
"learning_rate": 9.613459075424034e-05,
"loss": 1.3101,
"step": 46
},
{
"epoch": 0.35,
"grad_norm": 0.011797898208483941,
"learning_rate": 9.589264199393776e-05,
"loss": 1.3065,
"step": 47
},
{
"epoch": 0.36,
"grad_norm": 0.012117930154062621,
"learning_rate": 9.564367270466247e-05,
"loss": 1.3168,
"step": 48
},
{
"epoch": 0.36,
"grad_norm": 0.012075709345252717,
"learning_rate": 9.538772097304521e-05,
"loss": 1.3017,
"step": 49
},
{
"epoch": 0.37,
"grad_norm": 0.016464363967779214,
"learning_rate": 9.512482595387132e-05,
"loss": 1.3192,
"step": 50
},
{
"epoch": 0.38,
"grad_norm": 0.01258114656504482,
"learning_rate": 9.485502786409107e-05,
"loss": 1.3028,
"step": 51
},
{
"epoch": 0.39,
"grad_norm": 0.015912293882306635,
"learning_rate": 9.457836797666722e-05,
"loss": 1.3188,
"step": 52
},
{
"epoch": 0.39,
"grad_norm": 0.013021737938041136,
"learning_rate": 9.429488861426137e-05,
"loss": 1.3306,
"step": 53
},
{
"epoch": 0.4,
"grad_norm": 0.010972354267984178,
"learning_rate": 9.400463314275943e-05,
"loss": 1.2738,
"step": 54
},
{
"epoch": 0.41,
"grad_norm": 0.015962110281111695,
"learning_rate": 9.370764596463763e-05,
"loss": 1.2998,
"step": 55
},
{
"epoch": 0.42,
"grad_norm": 0.012606746770630168,
"learning_rate": 9.340397251217009e-05,
"loss": 1.301,
"step": 56
},
{
"epoch": 0.42,
"grad_norm": 0.01343997346580131,
"learning_rate": 9.309365924047853e-05,
"loss": 1.2821,
"step": 57
},
{
"epoch": 0.43,
"grad_norm": 0.01999816247440519,
"learning_rate": 9.277675362042581e-05,
"loss": 1.2801,
"step": 58
},
{
"epoch": 0.44,
"grad_norm": 0.011840965361160609,
"learning_rate": 9.245330413135395e-05,
"loss": 1.2646,
"step": 59
},
{
"epoch": 0.45,
"grad_norm": 0.012345942063074974,
"learning_rate": 9.212336025366788e-05,
"loss": 1.2654,
"step": 60
},
{
"epoch": 0.45,
"grad_norm": 0.011357300497689855,
"learning_rate": 9.178697246126607e-05,
"loss": 1.2286,
"step": 61
},
{
"epoch": 0.46,
"grad_norm": 0.0119333070271641,
"learning_rate": 9.144419221381919e-05,
"loss": 1.269,
"step": 62
},
{
"epoch": 0.47,
"grad_norm": 0.019987714588407004,
"learning_rate": 9.109507194889792e-05,
"loss": 1.3249,
"step": 63
},
{
"epoch": 0.47,
"grad_norm": 0.012629696476057141,
"learning_rate": 9.073966507395122e-05,
"loss": 1.2814,
"step": 64
},
{
"epoch": 0.48,
"grad_norm": 0.01514642919032275,
"learning_rate": 9.03780259581362e-05,
"loss": 1.2987,
"step": 65
},
{
"epoch": 0.49,
"grad_norm": 0.01217033214667514,
"learning_rate": 9.001020992400087e-05,
"loss": 1.3174,
"step": 66
},
{
"epoch": 0.5,
"grad_norm": 0.013313442038870889,
"learning_rate": 8.963627323902104e-05,
"loss": 1.3361,
"step": 67
},
{
"epoch": 0.5,
"grad_norm": 0.013140512018867165,
"learning_rate": 8.925627310699275e-05,
"loss": 1.2949,
"step": 68
},
{
"epoch": 0.51,
"grad_norm": 0.012629795148567,
"learning_rate": 8.887026765928129e-05,
"loss": 1.313,
"step": 69
},
{
"epoch": 0.52,
"grad_norm": 0.01267262432517559,
"learning_rate": 8.84783159459285e-05,
"loss": 1.2777,
"step": 70
},
{
"epoch": 0.53,
"grad_norm": 0.012491865926420912,
"learning_rate": 8.808047792661939e-05,
"loss": 1.2867,
"step": 71
},
{
"epoch": 0.53,
"grad_norm": 0.07419724215245581,
"learning_rate": 8.767681446150977e-05,
"loss": 1.2957,
"step": 72
},
{
"epoch": 0.54,
"grad_norm": 0.013685509971746855,
"learning_rate": 8.726738730191595e-05,
"loss": 1.3099,
"step": 73
},
{
"epoch": 0.55,
"grad_norm": 0.01255329844529139,
"learning_rate": 8.68522590808682e-05,
"loss": 1.3064,
"step": 74
},
{
"epoch": 0.56,
"grad_norm": 0.013555284699966613,
"learning_rate": 8.643149330352937e-05,
"loss": 1.3343,
"step": 75
},
{
"epoch": 0.56,
"grad_norm": 0.012428652847928174,
"learning_rate": 8.600515433748003e-05,
"loss": 1.2939,
"step": 76
},
{
"epoch": 0.57,
"grad_norm": 0.013730881730960102,
"learning_rate": 8.557330740287166e-05,
"loss": 1.3471,
"step": 77
},
{
"epoch": 0.58,
"grad_norm": 0.012536875741319477,
"learning_rate": 8.51360185624495e-05,
"loss": 1.2874,
"step": 78
},
{
"epoch": 0.59,
"grad_norm": 0.01315736800211036,
"learning_rate": 8.469335471144646e-05,
"loss": 1.3271,
"step": 79
},
{
"epoch": 0.59,
"grad_norm": 0.012790573071406599,
"learning_rate": 8.424538356734957e-05,
"loss": 1.2776,
"step": 80
},
{
"epoch": 0.6,
"grad_norm": 0.012774514721889157,
"learning_rate": 8.379217365954089e-05,
"loss": 1.3239,
"step": 81
},
{
"epoch": 0.61,
"grad_norm": 0.013206124918576702,
"learning_rate": 8.333379431881397e-05,
"loss": 1.3169,
"step": 82
},
{
"epoch": 0.62,
"grad_norm": 0.012150938166283229,
"learning_rate": 8.287031566676791e-05,
"loss": 1.3474,
"step": 83
},
{
"epoch": 0.62,
"grad_norm": 0.01296507972827761,
"learning_rate": 8.240180860508027e-05,
"loss": 1.2432,
"step": 84
},
{
"epoch": 0.63,
"grad_norm": 0.012431111381985924,
"learning_rate": 8.192834480466071e-05,
"loss": 1.2638,
"step": 85
},
{
"epoch": 0.64,
"grad_norm": 0.013206165711432155,
"learning_rate": 8.144999669468714e-05,
"loss": 1.2272,
"step": 86
},
{
"epoch": 0.65,
"grad_norm": 0.013107051117932556,
"learning_rate": 8.096683745152544e-05,
"loss": 1.3318,
"step": 87
},
{
"epoch": 0.65,
"grad_norm": 0.02417096924036142,
"learning_rate": 8.047894098753541e-05,
"loss": 1.3244,
"step": 88
},
{
"epoch": 0.66,
"grad_norm": 0.012897246540957289,
"learning_rate": 7.998638193976366e-05,
"loss": 1.2507,
"step": 89
},
{
"epoch": 0.67,
"grad_norm": 0.013675475740782888,
"learning_rate": 7.948923565852598e-05,
"loss": 1.2856,
"step": 90
},
{
"epoch": 0.68,
"grad_norm": 0.013091972232698349,
"learning_rate": 7.898757819588037e-05,
"loss": 1.286,
"step": 91
},
{
"epoch": 0.68,
"grad_norm": 0.013533547626105997,
"learning_rate": 7.848148629399285e-05,
"loss": 1.2917,
"step": 92
},
{
"epoch": 0.69,
"grad_norm": 0.013985718493269256,
"learning_rate": 7.797103737339766e-05,
"loss": 1.2814,
"step": 93
},
{
"epoch": 0.7,
"grad_norm": 0.012376190644291445,
"learning_rate": 7.745630952115363e-05,
"loss": 1.25,
"step": 94
},
{
"epoch": 0.71,
"grad_norm": 0.012580525752519058,
"learning_rate": 7.693738147889868e-05,
"loss": 1.2387,
"step": 95
},
{
"epoch": 0.71,
"grad_norm": 0.01280307621246117,
"learning_rate": 7.641433263080418e-05,
"loss": 1.2729,
"step": 96
},
{
"epoch": 0.72,
"grad_norm": 0.01346417935213005,
"learning_rate": 7.58872429914309e-05,
"loss": 1.2846,
"step": 97
},
{
"epoch": 0.73,
"grad_norm": 0.012689348395905247,
"learning_rate": 7.535619319348866e-05,
"loss": 1.3234,
"step": 98
},
{
"epoch": 0.73,
"grad_norm": 0.014176549248225716,
"learning_rate": 7.482126447550132e-05,
"loss": 1.3098,
"step": 99
},
{
"epoch": 0.74,
"grad_norm": 0.012665629411883726,
"learning_rate": 7.428253866937918e-05,
"loss": 1.3076,
"step": 100
},
{
"epoch": 0.75,
"grad_norm": 0.012301371470685241,
"learning_rate": 7.374009818790057e-05,
"loss": 1.2843,
"step": 101
},
{
"epoch": 0.76,
"grad_norm": 0.013326801939436429,
"learning_rate": 7.319402601210448e-05,
"loss": 1.2821,
"step": 102
},
{
"epoch": 0.76,
"grad_norm": 0.012305830268215724,
"learning_rate": 7.264440567859645e-05,
"loss": 1.3097,
"step": 103
},
{
"epoch": 0.77,
"grad_norm": 0.015048797529506962,
"learning_rate": 7.209132126676934e-05,
"loss": 1.2876,
"step": 104
},
{
"epoch": 0.78,
"grad_norm": 0.01836099524577789,
"learning_rate": 7.153485738594112e-05,
"loss": 1.2855,
"step": 105
},
{
"epoch": 0.79,
"grad_norm": 0.01395973727861205,
"learning_rate": 7.097509916241145e-05,
"loss": 1.3342,
"step": 106
},
{
"epoch": 0.79,
"grad_norm": 0.012705698786448718,
"learning_rate": 7.041213222643952e-05,
"loss": 1.2649,
"step": 107
},
{
"epoch": 0.8,
"grad_norm": 0.012865924289881111,
"learning_rate": 6.984604269914436e-05,
"loss": 1.2805,
"step": 108
},
{
"epoch": 0.81,
"grad_norm": 0.013770157843605553,
"learning_rate": 6.927691717933039e-05,
"loss": 1.3429,
"step": 109
},
{
"epoch": 0.82,
"grad_norm": 0.013491603805534848,
"learning_rate": 6.870484273023968e-05,
"loss": 1.3288,
"step": 110
},
{
"epoch": 0.82,
"grad_norm": 0.023074917576336243,
"learning_rate": 6.812990686623335e-05,
"loss": 1.3031,
"step": 111
},
{
"epoch": 0.83,
"grad_norm": 0.01360635130244037,
"learning_rate": 6.755219753940388e-05,
"loss": 1.2731,
"step": 112
},
{
"epoch": 0.84,
"grad_norm": 0.015779011261624655,
"learning_rate": 6.697180312612034e-05,
"loss": 1.3228,
"step": 113
},
{
"epoch": 0.85,
"grad_norm": 0.0191668822690555,
"learning_rate": 6.638881241350884e-05,
"loss": 1.2764,
"step": 114
},
{
"epoch": 0.85,
"grad_norm": 0.01432796902850521,
"learning_rate": 6.580331458587022e-05,
"loss": 1.2711,
"step": 115
},
{
"epoch": 0.86,
"grad_norm": 0.013611387104348411,
"learning_rate": 6.52153992110368e-05,
"loss": 1.3004,
"step": 116
},
{
"epoch": 0.87,
"grad_norm": 0.01399516730154587,
"learning_rate": 6.462515622667056e-05,
"loss": 1.2873,
"step": 117
},
{
"epoch": 0.88,
"grad_norm": 0.012770900961118808,
"learning_rate": 6.403267592650465e-05,
"loss": 1.2504,
"step": 118
},
{
"epoch": 0.88,
"grad_norm": 0.015945892933358158,
"learning_rate": 6.343804894653072e-05,
"loss": 1.2729,
"step": 119
},
{
"epoch": 0.89,
"grad_norm": 0.013482605970181992,
"learning_rate": 6.28413662511334e-05,
"loss": 1.2511,
"step": 120
},
{
"epoch": 0.9,
"grad_norm": 0.01378724352547337,
"learning_rate": 6.224271911917507e-05,
"loss": 1.2795,
"step": 121
},
{
"epoch": 0.91,
"grad_norm": 0.013380712527366646,
"learning_rate": 6.164219913003207e-05,
"loss": 1.2899,
"step": 122
},
{
"epoch": 0.91,
"grad_norm": 0.013150310617423817,
"learning_rate": 6.10398981495853e-05,
"loss": 1.2691,
"step": 123
},
{
"epoch": 0.92,
"grad_norm": 0.013440245678550199,
"learning_rate": 6.043590831616677e-05,
"loss": 1.292,
"step": 124
},
{
"epoch": 0.93,
"grad_norm": 0.014263259675451731,
"learning_rate": 5.9830322026464435e-05,
"loss": 1.3098,
"step": 125
},
{
"epoch": 0.94,
"grad_norm": 0.013569190009703554,
"learning_rate": 5.92232319213878e-05,
"loss": 1.2721,
"step": 126
},
{
"epoch": 0.94,
"grad_norm": 0.013956420143396445,
"learning_rate": 5.861473087189584e-05,
"loss": 1.3348,
"step": 127
},
{
"epoch": 0.95,
"grad_norm": 0.013219085418619908,
"learning_rate": 5.800491196478989e-05,
"loss": 1.2506,
"step": 128
},
{
"epoch": 0.96,
"grad_norm": 0.01391932503516543,
"learning_rate": 5.739386848847346e-05,
"loss": 1.2505,
"step": 129
},
{
"epoch": 0.96,
"grad_norm": 0.013286230114586647,
"learning_rate": 5.6781693918681275e-05,
"loss": 1.2736,
"step": 130
},
{
"epoch": 0.97,
"grad_norm": 0.014026096614615548,
"learning_rate": 5.616848190417965e-05,
"loss": 1.2942,
"step": 131
},
{
"epoch": 0.98,
"grad_norm": 0.013600317168386526,
"learning_rate": 5.5554326252440235e-05,
"loss": 1.2915,
"step": 132
},
{
"epoch": 0.99,
"grad_norm": 0.013507822076023955,
"learning_rate": 5.4939320915289714e-05,
"loss": 1.3112,
"step": 133
},
{
"epoch": 0.99,
"grad_norm": 0.014000478245483857,
"learning_rate": 5.432355997453729e-05,
"loss": 1.2811,
"step": 134
},
{
"epoch": 1.0,
"grad_norm": 0.01420975223503783,
"learning_rate": 5.370713762758232e-05,
"loss": 1.3203,
"step": 135
},
{
"epoch": 1.01,
"grad_norm": 0.013270421327306004,
"learning_rate": 5.3090148173004216e-05,
"loss": 1.2095,
"step": 136
},
{
"epoch": 1.02,
"grad_norm": 0.013841368832750186,
"learning_rate": 5.247268599613696e-05,
"loss": 1.2439,
"step": 137
},
{
"epoch": 1.02,
"grad_norm": 0.013599294458212107,
"learning_rate": 5.185484555463026e-05,
"loss": 1.2876,
"step": 138
},
{
"epoch": 1.03,
"grad_norm": 0.01403933789982391,
"learning_rate": 5.123672136399975e-05,
"loss": 1.2487,
"step": 139
},
{
"epoch": 1.04,
"grad_norm": 0.013785304842120809,
"learning_rate": 5.0618407983168146e-05,
"loss": 1.2908,
"step": 140
},
{
"epoch": 1.05,
"grad_norm": 0.02499232225633767,
"learning_rate": 5e-05,
"loss": 1.2542,
"step": 141
},
{
"epoch": 1.05,
"grad_norm": 0.013748729590013059,
"learning_rate": 4.938159201683186e-05,
"loss": 1.3158,
"step": 142
},
{
"epoch": 1.06,
"grad_norm": 0.015244834626064248,
"learning_rate": 4.876327863600025e-05,
"loss": 1.2813,
"step": 143
},
{
"epoch": 1.07,
"grad_norm": 0.013074690579273199,
"learning_rate": 4.814515444536974e-05,
"loss": 1.2589,
"step": 144
},
{
"epoch": 1.08,
"grad_norm": 0.013640971548809892,
"learning_rate": 4.752731400386306e-05,
"loss": 1.3409,
"step": 145
},
{
"epoch": 1.08,
"grad_norm": 0.013857785253189963,
"learning_rate": 4.69098518269958e-05,
"loss": 1.2884,
"step": 146
},
{
"epoch": 1.09,
"grad_norm": 0.01439159312691957,
"learning_rate": 4.62928623724177e-05,
"loss": 1.3038,
"step": 147
},
{
"epoch": 1.1,
"grad_norm": 0.01369785011946924,
"learning_rate": 4.567644002546273e-05,
"loss": 1.297,
"step": 148
},
{
"epoch": 1.11,
"grad_norm": 0.016775596557033395,
"learning_rate": 4.506067908471029e-05,
"loss": 1.2421,
"step": 149
},
{
"epoch": 1.11,
"grad_norm": 0.017413598587797105,
"learning_rate": 4.4445673747559776e-05,
"loss": 1.3068,
"step": 150
},
{
"epoch": 1.12,
"grad_norm": 0.015992833931677785,
"learning_rate": 4.383151809582035e-05,
"loss": 1.2794,
"step": 151
},
{
"epoch": 1.13,
"grad_norm": 0.01579232038410006,
"learning_rate": 4.3218306081318716e-05,
"loss": 1.3043,
"step": 152
},
{
"epoch": 1.14,
"grad_norm": 0.026770388446340592,
"learning_rate": 4.2606131511526555e-05,
"loss": 1.2837,
"step": 153
},
{
"epoch": 1.14,
"grad_norm": 0.031691067064066485,
"learning_rate": 4.1995088035210124e-05,
"loss": 1.2788,
"step": 154
},
{
"epoch": 1.15,
"grad_norm": 0.0162069083852305,
"learning_rate": 4.138526912810418e-05,
"loss": 1.2895,
"step": 155
},
{
"epoch": 1.16,
"grad_norm": 0.015730822077210757,
"learning_rate": 4.077676807861221e-05,
"loss": 1.3028,
"step": 156
},
{
"epoch": 1.17,
"grad_norm": 0.017380121781761425,
"learning_rate": 4.016967797353558e-05,
"loss": 1.2553,
"step": 157
},
{
"epoch": 1.17,
"grad_norm": 0.014829204054633923,
"learning_rate": 3.956409168383325e-05,
"loss": 1.2804,
"step": 158
},
{
"epoch": 1.18,
"grad_norm": 0.014816527490603743,
"learning_rate": 3.89601018504147e-05,
"loss": 1.2691,
"step": 159
},
{
"epoch": 1.19,
"grad_norm": 0.014742205868113261,
"learning_rate": 3.835780086996794e-05,
"loss": 1.3161,
"step": 160
}
],
"logging_steps": 1.0,
"max_steps": 268,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 20,
"total_flos": 2473123236544512.0,
"train_batch_size": 3,
"trial_name": null,
"trial_params": null
}