model_c2285a5d / checkpoint-296 /trainer_state.json
ugaoo's picture
Upload folder using huggingface_hub
f07c2ee verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.9551569506726456,
"eval_steps": 500,
"global_step": 296,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.013452914798206279,
"grad_norm": 95.4926986694336,
"learning_rate": 5.0000000000000004e-08,
"loss": 12.2856,
"step": 1
},
{
"epoch": 0.026905829596412557,
"grad_norm": 93.69285583496094,
"learning_rate": 1.0000000000000001e-07,
"loss": 12.2383,
"step": 2
},
{
"epoch": 0.04035874439461883,
"grad_norm": 95.09840393066406,
"learning_rate": 1.5000000000000002e-07,
"loss": 12.1293,
"step": 3
},
{
"epoch": 0.053811659192825115,
"grad_norm": 95.04216766357422,
"learning_rate": 2.0000000000000002e-07,
"loss": 12.1453,
"step": 4
},
{
"epoch": 0.06726457399103139,
"grad_norm": 93.44210052490234,
"learning_rate": 2.5000000000000004e-07,
"loss": 12.165,
"step": 5
},
{
"epoch": 0.08071748878923767,
"grad_norm": 93.28514862060547,
"learning_rate": 3.0000000000000004e-07,
"loss": 12.063,
"step": 6
},
{
"epoch": 0.09417040358744394,
"grad_norm": 95.3654556274414,
"learning_rate": 3.5000000000000004e-07,
"loss": 11.9676,
"step": 7
},
{
"epoch": 0.10762331838565023,
"grad_norm": 96.05154418945312,
"learning_rate": 4.0000000000000003e-07,
"loss": 12.0911,
"step": 8
},
{
"epoch": 0.1210762331838565,
"grad_norm": 97.69881439208984,
"learning_rate": 4.5000000000000003e-07,
"loss": 12.0717,
"step": 9
},
{
"epoch": 0.13452914798206278,
"grad_norm": 95.55254364013672,
"learning_rate": 5.000000000000001e-07,
"loss": 11.9752,
"step": 10
},
{
"epoch": 0.14798206278026907,
"grad_norm": 95.9182357788086,
"learning_rate": 5.5e-07,
"loss": 11.9413,
"step": 11
},
{
"epoch": 0.16143497757847533,
"grad_norm": 95.40771484375,
"learning_rate": 6.000000000000001e-07,
"loss": 11.7523,
"step": 12
},
{
"epoch": 0.17488789237668162,
"grad_norm": 94.40055847167969,
"learning_rate": 6.5e-07,
"loss": 11.6384,
"step": 13
},
{
"epoch": 0.18834080717488788,
"grad_norm": 93.58352661132812,
"learning_rate": 7.000000000000001e-07,
"loss": 11.4293,
"step": 14
},
{
"epoch": 0.20179372197309417,
"grad_norm": 94.48737335205078,
"learning_rate": 7.5e-07,
"loss": 11.1445,
"step": 15
},
{
"epoch": 0.21524663677130046,
"grad_norm": 92.44265747070312,
"learning_rate": 8.000000000000001e-07,
"loss": 10.7705,
"step": 16
},
{
"epoch": 0.22869955156950672,
"grad_norm": 90.97422790527344,
"learning_rate": 8.500000000000001e-07,
"loss": 10.3754,
"step": 17
},
{
"epoch": 0.242152466367713,
"grad_norm": 88.54856872558594,
"learning_rate": 9.000000000000001e-07,
"loss": 10.0019,
"step": 18
},
{
"epoch": 0.2556053811659193,
"grad_norm": 88.39138793945312,
"learning_rate": 9.500000000000001e-07,
"loss": 9.5093,
"step": 19
},
{
"epoch": 0.26905829596412556,
"grad_norm": 86.55109405517578,
"learning_rate": 1.0000000000000002e-06,
"loss": 9.2342,
"step": 20
},
{
"epoch": 0.2825112107623318,
"grad_norm": 80.62335205078125,
"learning_rate": 1.0500000000000001e-06,
"loss": 8.597,
"step": 21
},
{
"epoch": 0.29596412556053814,
"grad_norm": 73.67768859863281,
"learning_rate": 1.1e-06,
"loss": 8.1122,
"step": 22
},
{
"epoch": 0.3094170403587444,
"grad_norm": 64.57353210449219,
"learning_rate": 1.1500000000000002e-06,
"loss": 7.6455,
"step": 23
},
{
"epoch": 0.32286995515695066,
"grad_norm": 55.2818603515625,
"learning_rate": 1.2000000000000002e-06,
"loss": 7.2493,
"step": 24
},
{
"epoch": 0.336322869955157,
"grad_norm": 48.274452209472656,
"learning_rate": 1.25e-06,
"loss": 7.0377,
"step": 25
},
{
"epoch": 0.34977578475336324,
"grad_norm": 42.7370491027832,
"learning_rate": 1.3e-06,
"loss": 6.5782,
"step": 26
},
{
"epoch": 0.3632286995515695,
"grad_norm": 39.297462463378906,
"learning_rate": 1.3500000000000002e-06,
"loss": 6.2558,
"step": 27
},
{
"epoch": 0.37668161434977576,
"grad_norm": 37.91667938232422,
"learning_rate": 1.4000000000000001e-06,
"loss": 5.9809,
"step": 28
},
{
"epoch": 0.3901345291479821,
"grad_norm": 37.87322998046875,
"learning_rate": 1.45e-06,
"loss": 5.7268,
"step": 29
},
{
"epoch": 0.40358744394618834,
"grad_norm": 36.48906707763672,
"learning_rate": 1.5e-06,
"loss": 5.449,
"step": 30
},
{
"epoch": 0.4170403587443946,
"grad_norm": 36.38510513305664,
"learning_rate": 1.5500000000000002e-06,
"loss": 5.1884,
"step": 31
},
{
"epoch": 0.4304932735426009,
"grad_norm": 35.656829833984375,
"learning_rate": 1.6000000000000001e-06,
"loss": 4.899,
"step": 32
},
{
"epoch": 0.4439461883408072,
"grad_norm": 34.09960174560547,
"learning_rate": 1.6500000000000003e-06,
"loss": 4.5842,
"step": 33
},
{
"epoch": 0.45739910313901344,
"grad_norm": 32.74240493774414,
"learning_rate": 1.7000000000000002e-06,
"loss": 4.3009,
"step": 34
},
{
"epoch": 0.47085201793721976,
"grad_norm": 31.867507934570312,
"learning_rate": 1.75e-06,
"loss": 3.9865,
"step": 35
},
{
"epoch": 0.484304932735426,
"grad_norm": 30.741374969482422,
"learning_rate": 1.8000000000000001e-06,
"loss": 3.6916,
"step": 36
},
{
"epoch": 0.4977578475336323,
"grad_norm": 27.8775577545166,
"learning_rate": 1.85e-06,
"loss": 3.3719,
"step": 37
},
{
"epoch": 0.5112107623318386,
"grad_norm": 25.97083282470703,
"learning_rate": 1.9000000000000002e-06,
"loss": 3.0907,
"step": 38
},
{
"epoch": 0.5246636771300448,
"grad_norm": 23.62006950378418,
"learning_rate": 1.9500000000000004e-06,
"loss": 2.8336,
"step": 39
},
{
"epoch": 0.5381165919282511,
"grad_norm": 23.80520248413086,
"learning_rate": 2.0000000000000003e-06,
"loss": 2.5717,
"step": 40
},
{
"epoch": 0.5515695067264574,
"grad_norm": 25.32924461364746,
"learning_rate": 2.05e-06,
"loss": 2.3658,
"step": 41
},
{
"epoch": 0.5650224215246636,
"grad_norm": 26.20570182800293,
"learning_rate": 2.1000000000000002e-06,
"loss": 2.2443,
"step": 42
},
{
"epoch": 0.57847533632287,
"grad_norm": 24.581693649291992,
"learning_rate": 2.15e-06,
"loss": 1.926,
"step": 43
},
{
"epoch": 0.5919282511210763,
"grad_norm": 24.414310455322266,
"learning_rate": 2.2e-06,
"loss": 1.7034,
"step": 44
},
{
"epoch": 0.6053811659192825,
"grad_norm": 22.691083908081055,
"learning_rate": 2.25e-06,
"loss": 1.4857,
"step": 45
},
{
"epoch": 0.6188340807174888,
"grad_norm": 20.669803619384766,
"learning_rate": 2.3000000000000004e-06,
"loss": 1.2415,
"step": 46
},
{
"epoch": 0.6322869955156951,
"grad_norm": 20.149641036987305,
"learning_rate": 2.35e-06,
"loss": 0.997,
"step": 47
},
{
"epoch": 0.6457399103139013,
"grad_norm": 18.632596969604492,
"learning_rate": 2.4000000000000003e-06,
"loss": 0.7552,
"step": 48
},
{
"epoch": 0.6591928251121076,
"grad_norm": 16.93793296813965,
"learning_rate": 2.4500000000000003e-06,
"loss": 0.5883,
"step": 49
},
{
"epoch": 0.672645739910314,
"grad_norm": 14.432519912719727,
"learning_rate": 2.5e-06,
"loss": 0.4382,
"step": 50
},
{
"epoch": 0.6860986547085202,
"grad_norm": 11.829660415649414,
"learning_rate": 2.55e-06,
"loss": 0.2983,
"step": 51
},
{
"epoch": 0.6995515695067265,
"grad_norm": 8.680500030517578,
"learning_rate": 2.6e-06,
"loss": 0.1988,
"step": 52
},
{
"epoch": 0.7130044843049327,
"grad_norm": 6.53156852722168,
"learning_rate": 2.6500000000000005e-06,
"loss": 0.1589,
"step": 53
},
{
"epoch": 0.726457399103139,
"grad_norm": 2.9756624698638916,
"learning_rate": 2.7000000000000004e-06,
"loss": 0.0686,
"step": 54
},
{
"epoch": 0.7399103139013453,
"grad_norm": 5.545580863952637,
"learning_rate": 2.7500000000000004e-06,
"loss": 0.0865,
"step": 55
},
{
"epoch": 0.7533632286995515,
"grad_norm": 4.045405387878418,
"learning_rate": 2.8000000000000003e-06,
"loss": 0.0949,
"step": 56
},
{
"epoch": 0.7668161434977578,
"grad_norm": 1.6688120365142822,
"learning_rate": 2.85e-06,
"loss": 0.0396,
"step": 57
},
{
"epoch": 0.7802690582959642,
"grad_norm": 2.4520657062530518,
"learning_rate": 2.9e-06,
"loss": 0.0439,
"step": 58
},
{
"epoch": 0.7937219730941704,
"grad_norm": 2.608729600906372,
"learning_rate": 2.95e-06,
"loss": 0.057,
"step": 59
},
{
"epoch": 0.8071748878923767,
"grad_norm": 2.365234851837158,
"learning_rate": 3e-06,
"loss": 0.0547,
"step": 60
},
{
"epoch": 0.820627802690583,
"grad_norm": 0.787550687789917,
"learning_rate": 3.05e-06,
"loss": 0.0209,
"step": 61
},
{
"epoch": 0.8340807174887892,
"grad_norm": 0.7686442732810974,
"learning_rate": 3.1000000000000004e-06,
"loss": 0.0221,
"step": 62
},
{
"epoch": 0.8475336322869955,
"grad_norm": 1.2510555982589722,
"learning_rate": 3.1500000000000003e-06,
"loss": 0.0165,
"step": 63
},
{
"epoch": 0.8609865470852018,
"grad_norm": 0.8923770189285278,
"learning_rate": 3.2000000000000003e-06,
"loss": 0.0187,
"step": 64
},
{
"epoch": 0.874439461883408,
"grad_norm": 0.8052615523338318,
"learning_rate": 3.2500000000000002e-06,
"loss": 0.0266,
"step": 65
},
{
"epoch": 0.8878923766816144,
"grad_norm": 0.6710303425788879,
"learning_rate": 3.3000000000000006e-06,
"loss": 0.0154,
"step": 66
},
{
"epoch": 0.9013452914798207,
"grad_norm": 0.5213025212287903,
"learning_rate": 3.3500000000000005e-06,
"loss": 0.0085,
"step": 67
},
{
"epoch": 0.9147982062780269,
"grad_norm": 0.5758580565452576,
"learning_rate": 3.4000000000000005e-06,
"loss": 0.0133,
"step": 68
},
{
"epoch": 0.9282511210762332,
"grad_norm": 0.6828752160072327,
"learning_rate": 3.45e-06,
"loss": 0.0186,
"step": 69
},
{
"epoch": 0.9417040358744395,
"grad_norm": 0.6814988255500793,
"learning_rate": 3.5e-06,
"loss": 0.0215,
"step": 70
},
{
"epoch": 0.9551569506726457,
"grad_norm": 0.718296229839325,
"learning_rate": 3.5500000000000003e-06,
"loss": 0.0204,
"step": 71
},
{
"epoch": 0.968609865470852,
"grad_norm": 0.7816944122314453,
"learning_rate": 3.6000000000000003e-06,
"loss": 0.0184,
"step": 72
},
{
"epoch": 0.9820627802690582,
"grad_norm": 0.6058817505836487,
"learning_rate": 3.65e-06,
"loss": 0.0179,
"step": 73
},
{
"epoch": 0.9955156950672646,
"grad_norm": 1.0496101379394531,
"learning_rate": 3.7e-06,
"loss": 0.032,
"step": 74
},
{
"epoch": 1.0,
"grad_norm": 1.0496101379394531,
"learning_rate": 3.7500000000000005e-06,
"loss": 0.0137,
"step": 75
},
{
"epoch": 1.0134529147982063,
"grad_norm": 1.598720669746399,
"learning_rate": 3.8000000000000005e-06,
"loss": 0.0211,
"step": 76
},
{
"epoch": 1.0269058295964126,
"grad_norm": 0.7792187333106995,
"learning_rate": 3.85e-06,
"loss": 0.0213,
"step": 77
},
{
"epoch": 1.0403587443946187,
"grad_norm": 0.7717252373695374,
"learning_rate": 3.900000000000001e-06,
"loss": 0.0151,
"step": 78
},
{
"epoch": 1.053811659192825,
"grad_norm": 0.39334648847579956,
"learning_rate": 3.95e-06,
"loss": 0.0186,
"step": 79
},
{
"epoch": 1.0672645739910314,
"grad_norm": 0.9775457382202148,
"learning_rate": 4.000000000000001e-06,
"loss": 0.0043,
"step": 80
},
{
"epoch": 1.0807174887892377,
"grad_norm": 1.2425150871276855,
"learning_rate": 4.05e-06,
"loss": 0.0179,
"step": 81
},
{
"epoch": 1.094170403587444,
"grad_norm": 0.3884654939174652,
"learning_rate": 4.1e-06,
"loss": 0.008,
"step": 82
},
{
"epoch": 1.1076233183856503,
"grad_norm": 0.4746466875076294,
"learning_rate": 4.15e-06,
"loss": 0.0144,
"step": 83
},
{
"epoch": 1.1210762331838564,
"grad_norm": 0.6812214255332947,
"learning_rate": 4.2000000000000004e-06,
"loss": 0.0072,
"step": 84
},
{
"epoch": 1.1345291479820627,
"grad_norm": 0.5414469838142395,
"learning_rate": 4.25e-06,
"loss": 0.016,
"step": 85
},
{
"epoch": 1.147982062780269,
"grad_norm": 0.8709024786949158,
"learning_rate": 4.3e-06,
"loss": 0.0154,
"step": 86
},
{
"epoch": 1.1614349775784754,
"grad_norm": 0.3181096613407135,
"learning_rate": 4.350000000000001e-06,
"loss": 0.0039,
"step": 87
},
{
"epoch": 1.1748878923766817,
"grad_norm": 0.2581265866756439,
"learning_rate": 4.4e-06,
"loss": 0.0014,
"step": 88
},
{
"epoch": 1.188340807174888,
"grad_norm": 0.10870776325464249,
"learning_rate": 4.450000000000001e-06,
"loss": 0.0088,
"step": 89
},
{
"epoch": 1.201793721973094,
"grad_norm": 0.45314452052116394,
"learning_rate": 4.5e-06,
"loss": 0.0019,
"step": 90
},
{
"epoch": 1.2152466367713004,
"grad_norm": 0.7328381538391113,
"learning_rate": 4.5500000000000005e-06,
"loss": 0.0082,
"step": 91
},
{
"epoch": 1.2286995515695067,
"grad_norm": 0.6641069650650024,
"learning_rate": 4.600000000000001e-06,
"loss": 0.0075,
"step": 92
},
{
"epoch": 1.242152466367713,
"grad_norm": 0.27984222769737244,
"learning_rate": 4.65e-06,
"loss": 0.0009,
"step": 93
},
{
"epoch": 1.2556053811659194,
"grad_norm": 0.8341127634048462,
"learning_rate": 4.7e-06,
"loss": 0.0204,
"step": 94
},
{
"epoch": 1.2690582959641254,
"grad_norm": 1.0140557289123535,
"learning_rate": 4.75e-06,
"loss": 0.0228,
"step": 95
},
{
"epoch": 1.2825112107623318,
"grad_norm": 0.9439787268638611,
"learning_rate": 4.800000000000001e-06,
"loss": 0.027,
"step": 96
},
{
"epoch": 1.295964125560538,
"grad_norm": 1.3867762088775635,
"learning_rate": 4.85e-06,
"loss": 0.0223,
"step": 97
},
{
"epoch": 1.3094170403587444,
"grad_norm": 0.4747941493988037,
"learning_rate": 4.9000000000000005e-06,
"loss": 0.0048,
"step": 98
},
{
"epoch": 1.3228699551569507,
"grad_norm": 0.5673424601554871,
"learning_rate": 4.95e-06,
"loss": 0.0033,
"step": 99
},
{
"epoch": 1.336322869955157,
"grad_norm": 0.24510182440280914,
"learning_rate": 5e-06,
"loss": 0.0048,
"step": 100
},
{
"epoch": 1.3497757847533634,
"grad_norm": 0.6084151268005371,
"learning_rate": 4.99989574668946e-06,
"loss": 0.032,
"step": 101
},
{
"epoch": 1.3632286995515694,
"grad_norm": 0.606272280216217,
"learning_rate": 4.999582995452842e-06,
"loss": 0.0058,
"step": 102
},
{
"epoch": 1.3766816143497758,
"grad_norm": 0.2270481288433075,
"learning_rate": 4.999061772374426e-06,
"loss": 0.0043,
"step": 103
},
{
"epoch": 1.390134529147982,
"grad_norm": 0.2832431495189667,
"learning_rate": 4.998332120925598e-06,
"loss": 0.0079,
"step": 104
},
{
"epoch": 1.4035874439461884,
"grad_norm": 0.44565466046333313,
"learning_rate": 4.9973941019612235e-06,
"loss": 0.017,
"step": 105
},
{
"epoch": 1.4170403587443947,
"grad_norm": 0.3518397808074951,
"learning_rate": 4.996247793714565e-06,
"loss": 0.0057,
"step": 106
},
{
"epoch": 1.4304932735426008,
"grad_norm": 0.41977858543395996,
"learning_rate": 4.994893291790768e-06,
"loss": 0.0056,
"step": 107
},
{
"epoch": 1.4439461883408071,
"grad_norm": 0.22567600011825562,
"learning_rate": 4.993330709158879e-06,
"loss": 0.0067,
"step": 108
},
{
"epoch": 1.4573991031390134,
"grad_norm": 0.2184637039899826,
"learning_rate": 4.9915601761424304e-06,
"loss": 0.0043,
"step": 109
},
{
"epoch": 1.4708520179372198,
"grad_norm": 0.24834735691547394,
"learning_rate": 4.989581840408562e-06,
"loss": 0.0034,
"step": 110
},
{
"epoch": 1.484304932735426,
"grad_norm": 0.13165591657161713,
"learning_rate": 4.987395866955716e-06,
"loss": 0.0041,
"step": 111
},
{
"epoch": 1.4977578475336322,
"grad_norm": 0.27152982354164124,
"learning_rate": 4.9850024380998655e-06,
"loss": 0.0034,
"step": 112
},
{
"epoch": 1.5112107623318387,
"grad_norm": 0.2793160676956177,
"learning_rate": 4.982401753459317e-06,
"loss": 0.0049,
"step": 113
},
{
"epoch": 1.5246636771300448,
"grad_norm": 0.06674113124608994,
"learning_rate": 4.979594029938058e-06,
"loss": 0.0034,
"step": 114
},
{
"epoch": 1.5381165919282511,
"grad_norm": 0.2601087689399719,
"learning_rate": 4.976579501707665e-06,
"loss": 0.0025,
"step": 115
},
{
"epoch": 1.5515695067264574,
"grad_norm": 0.17075951397418976,
"learning_rate": 4.973358420187776e-06,
"loss": 0.0005,
"step": 116
},
{
"epoch": 1.5650224215246635,
"grad_norm": 0.11211276799440384,
"learning_rate": 4.969931054025122e-06,
"loss": 0.0059,
"step": 117
},
{
"epoch": 1.57847533632287,
"grad_norm": 0.5140838027000427,
"learning_rate": 4.966297689071117e-06,
"loss": 0.018,
"step": 118
},
{
"epoch": 1.5919282511210762,
"grad_norm": 0.5356995463371277,
"learning_rate": 4.962458628358021e-06,
"loss": 0.0007,
"step": 119
},
{
"epoch": 1.6053811659192825,
"grad_norm": 0.3967442214488983,
"learning_rate": 4.958414192073665e-06,
"loss": 0.014,
"step": 120
},
{
"epoch": 1.6188340807174888,
"grad_norm": 0.4567921757698059,
"learning_rate": 4.954164717534748e-06,
"loss": 0.0027,
"step": 121
},
{
"epoch": 1.6322869955156951,
"grad_norm": 0.4405613839626312,
"learning_rate": 4.949710559158699e-06,
"loss": 0.0135,
"step": 122
},
{
"epoch": 1.6457399103139014,
"grad_norm": 0.7164422869682312,
"learning_rate": 4.945052088434123e-06,
"loss": 0.0207,
"step": 123
},
{
"epoch": 1.6591928251121075,
"grad_norm": 0.47173142433166504,
"learning_rate": 4.940189693889819e-06,
"loss": 0.0019,
"step": 124
},
{
"epoch": 1.672645739910314,
"grad_norm": 0.4606887996196747,
"learning_rate": 4.9351237810623655e-06,
"loss": 0.0125,
"step": 125
},
{
"epoch": 1.6860986547085202,
"grad_norm": 0.7209401726722717,
"learning_rate": 4.929854772462312e-06,
"loss": 0.012,
"step": 126
},
{
"epoch": 1.6995515695067265,
"grad_norm": 1.7617985010147095,
"learning_rate": 4.924383107538929e-06,
"loss": 0.0079,
"step": 127
},
{
"epoch": 1.7130044843049328,
"grad_norm": 1.0861084461212158,
"learning_rate": 4.918709242643563e-06,
"loss": 0.0091,
"step": 128
},
{
"epoch": 1.726457399103139,
"grad_norm": 0.11263061314821243,
"learning_rate": 4.9128336509915746e-06,
"loss": 0.0006,
"step": 129
},
{
"epoch": 1.7399103139013454,
"grad_norm": 0.3411642909049988,
"learning_rate": 4.906756822622865e-06,
"loss": 0.0037,
"step": 130
},
{
"epoch": 1.7533632286995515,
"grad_norm": 0.1550491899251938,
"learning_rate": 4.900479264361017e-06,
"loss": 0.0056,
"step": 131
},
{
"epoch": 1.7668161434977578,
"grad_norm": 0.8448930382728577,
"learning_rate": 4.894001499771015e-06,
"loss": 0.0135,
"step": 132
},
{
"epoch": 1.7802690582959642,
"grad_norm": 0.5105615258216858,
"learning_rate": 4.887324069115582e-06,
"loss": 0.009,
"step": 133
},
{
"epoch": 1.7937219730941703,
"grad_norm": 0.5042226910591125,
"learning_rate": 4.880447529310118e-06,
"loss": 0.0022,
"step": 134
},
{
"epoch": 1.8071748878923768,
"grad_norm": 0.15797697007656097,
"learning_rate": 4.873372453876255e-06,
"loss": 0.0073,
"step": 135
},
{
"epoch": 1.8206278026905829,
"grad_norm": 0.47805944085121155,
"learning_rate": 4.866099432894023e-06,
"loss": 0.0084,
"step": 136
},
{
"epoch": 1.8340807174887892,
"grad_norm": 0.3370952904224396,
"learning_rate": 4.858629072952635e-06,
"loss": 0.0037,
"step": 137
},
{
"epoch": 1.8475336322869955,
"grad_norm": 0.30983835458755493,
"learning_rate": 4.850961997099892e-06,
"loss": 0.0024,
"step": 138
},
{
"epoch": 1.8609865470852018,
"grad_norm": 0.2800588011741638,
"learning_rate": 4.843098844790228e-06,
"loss": 0.0032,
"step": 139
},
{
"epoch": 1.8744394618834082,
"grad_norm": 0.2037343531847,
"learning_rate": 4.835040271831371e-06,
"loss": 0.0016,
"step": 140
},
{
"epoch": 1.8878923766816142,
"grad_norm": 0.45981553196907043,
"learning_rate": 4.826786950329646e-06,
"loss": 0.0035,
"step": 141
},
{
"epoch": 1.9013452914798208,
"grad_norm": 0.17092454433441162,
"learning_rate": 4.818339568633926e-06,
"loss": 0.0068,
"step": 142
},
{
"epoch": 1.9147982062780269,
"grad_norm": 0.5339077711105347,
"learning_rate": 4.809698831278217e-06,
"loss": 0.0026,
"step": 143
},
{
"epoch": 1.9282511210762332,
"grad_norm": 0.5208529829978943,
"learning_rate": 4.800865458922899e-06,
"loss": 0.0047,
"step": 144
},
{
"epoch": 1.9417040358744395,
"grad_norm": 0.7488933801651001,
"learning_rate": 4.79184018829462e-06,
"loss": 0.0096,
"step": 145
},
{
"epoch": 1.9551569506726456,
"grad_norm": 0.5727106332778931,
"learning_rate": 4.782623772124854e-06,
"loss": 0.001,
"step": 146
},
{
"epoch": 1.9686098654708521,
"grad_norm": 0.3813195526599884,
"learning_rate": 4.77321697908712e-06,
"loss": 0.0085,
"step": 147
},
{
"epoch": 1.9820627802690582,
"grad_norm": 0.5406109094619751,
"learning_rate": 4.763620593732867e-06,
"loss": 0.0017,
"step": 148
},
{
"epoch": 1.9955156950672646,
"grad_norm": 0.391985148191452,
"learning_rate": 4.7538354164260515e-06,
"loss": 0.0019,
"step": 149
},
{
"epoch": 2.0,
"grad_norm": 0.391985148191452,
"learning_rate": 4.743862263276376e-06,
"loss": 0.0003,
"step": 150
},
{
"epoch": 2.013452914798206,
"grad_norm": 0.10133524239063263,
"learning_rate": 4.733701966071226e-06,
"loss": 0.0012,
"step": 151
},
{
"epoch": 2.0269058295964126,
"grad_norm": 0.2032414674758911,
"learning_rate": 4.723355372206297e-06,
"loss": 0.0001,
"step": 152
},
{
"epoch": 2.0403587443946187,
"grad_norm": 0.00796876847743988,
"learning_rate": 4.712823344614921e-06,
"loss": 0.0025,
"step": 153
},
{
"epoch": 2.0538116591928253,
"grad_norm": 0.36200040578842163,
"learning_rate": 4.702106761696091e-06,
"loss": 0.0003,
"step": 154
},
{
"epoch": 2.0672645739910314,
"grad_norm": 0.19646115601062775,
"learning_rate": 4.691206517241205e-06,
"loss": 0.0009,
"step": 155
},
{
"epoch": 2.0807174887892375,
"grad_norm": 0.06766581535339355,
"learning_rate": 4.68012352035952e-06,
"loss": 0.0001,
"step": 156
},
{
"epoch": 2.094170403587444,
"grad_norm": 0.00797713827341795,
"learning_rate": 4.668858695402326e-06,
"loss": 0.0001,
"step": 157
},
{
"epoch": 2.10762331838565,
"grad_norm": 0.020394539460539818,
"learning_rate": 4.657412981885862e-06,
"loss": 0.0002,
"step": 158
},
{
"epoch": 2.1210762331838566,
"grad_norm": 0.017584379762411118,
"learning_rate": 4.645787334412945e-06,
"loss": 0.0002,
"step": 159
},
{
"epoch": 2.1345291479820627,
"grad_norm": 0.07967082411050797,
"learning_rate": 4.633982722593367e-06,
"loss": 0.0003,
"step": 160
},
{
"epoch": 2.1479820627802693,
"grad_norm": 0.01606675237417221,
"learning_rate": 4.622000130963015e-06,
"loss": 0.0003,
"step": 161
},
{
"epoch": 2.1614349775784754,
"grad_norm": 0.06052660569548607,
"learning_rate": 4.6098405589017685e-06,
"loss": 0.0007,
"step": 162
},
{
"epoch": 2.1748878923766815,
"grad_norm": 0.07991409301757812,
"learning_rate": 4.597505020550138e-06,
"loss": 0.0002,
"step": 163
},
{
"epoch": 2.188340807174888,
"grad_norm": 0.04991272836923599,
"learning_rate": 4.584994544724695e-06,
"loss": 0.0001,
"step": 164
},
{
"epoch": 2.201793721973094,
"grad_norm": 0.027111921459436417,
"learning_rate": 4.572310174832255e-06,
"loss": 0.0001,
"step": 165
},
{
"epoch": 2.2152466367713006,
"grad_norm": 0.006566982250660658,
"learning_rate": 4.5594529687828615e-06,
"loss": 0.0,
"step": 166
},
{
"epoch": 2.2286995515695067,
"grad_norm": 0.002437079790979624,
"learning_rate": 4.546423998901549e-06,
"loss": 0.0004,
"step": 167
},
{
"epoch": 2.242152466367713,
"grad_norm": 0.08434150367975235,
"learning_rate": 4.533224351838914e-06,
"loss": 0.0001,
"step": 168
},
{
"epoch": 2.2556053811659194,
"grad_norm": 0.013094129040837288,
"learning_rate": 4.519855128480478e-06,
"loss": 0.0001,
"step": 169
},
{
"epoch": 2.2690582959641254,
"grad_norm": 0.002770340768620372,
"learning_rate": 4.5063174438548775e-06,
"loss": 0.0001,
"step": 170
},
{
"epoch": 2.282511210762332,
"grad_norm": 0.021480072289705276,
"learning_rate": 4.492612427040864e-06,
"loss": 0.0001,
"step": 171
},
{
"epoch": 2.295964125560538,
"grad_norm": 0.0028862387407571077,
"learning_rate": 4.478741221073136e-06,
"loss": 0.0001,
"step": 172
},
{
"epoch": 2.3094170403587446,
"grad_norm": 0.01969303749501705,
"learning_rate": 4.464704982847008e-06,
"loss": 0.0,
"step": 173
},
{
"epoch": 2.3228699551569507,
"grad_norm": 0.00244798487983644,
"learning_rate": 4.450504883021923e-06,
"loss": 0.0,
"step": 174
},
{
"epoch": 2.336322869955157,
"grad_norm": 0.0036638586316257715,
"learning_rate": 4.436142105923814e-06,
"loss": 0.0009,
"step": 175
},
{
"epoch": 2.3497757847533634,
"grad_norm": 0.4378701448440552,
"learning_rate": 4.4216178494463305e-06,
"loss": 0.0003,
"step": 176
},
{
"epoch": 2.3632286995515694,
"grad_norm": 0.004919757135212421,
"learning_rate": 4.406933324950929e-06,
"loss": 0.0,
"step": 177
},
{
"epoch": 2.376681614349776,
"grad_norm": 0.0061010573990643024,
"learning_rate": 4.392089757165841e-06,
"loss": 0.0001,
"step": 178
},
{
"epoch": 2.390134529147982,
"grad_norm": 0.00750540429726243,
"learning_rate": 4.377088384083935e-06,
"loss": 0.0004,
"step": 179
},
{
"epoch": 2.403587443946188,
"grad_norm": 0.18055735528469086,
"learning_rate": 4.361930456859455e-06,
"loss": 0.0001,
"step": 180
},
{
"epoch": 2.4170403587443947,
"grad_norm": 0.00540179992094636,
"learning_rate": 4.346617239703676e-06,
"loss": 0.0017,
"step": 181
},
{
"epoch": 2.430493273542601,
"grad_norm": 0.6051300764083862,
"learning_rate": 4.3311500097794655e-06,
"loss": 0.0002,
"step": 182
},
{
"epoch": 2.4439461883408073,
"grad_norm": 0.04465539753437042,
"learning_rate": 4.315530057094763e-06,
"loss": 0.0001,
"step": 183
},
{
"epoch": 2.4573991031390134,
"grad_norm": 0.009125534445047379,
"learning_rate": 4.2997586843949905e-06,
"loss": 0.0001,
"step": 184
},
{
"epoch": 2.4708520179372195,
"grad_norm": 0.009754106402397156,
"learning_rate": 4.2838372070544e-06,
"loss": 0.0001,
"step": 185
},
{
"epoch": 2.484304932735426,
"grad_norm": 0.03795509785413742,
"learning_rate": 4.267766952966369e-06,
"loss": 0.0006,
"step": 186
},
{
"epoch": 2.497757847533632,
"grad_norm": 0.19477951526641846,
"learning_rate": 4.25154926243265e-06,
"loss": 0.0013,
"step": 187
},
{
"epoch": 2.5112107623318387,
"grad_norm": 0.27783000469207764,
"learning_rate": 4.2351854880515856e-06,
"loss": 0.0001,
"step": 188
},
{
"epoch": 2.524663677130045,
"grad_norm": 0.030597640201449394,
"learning_rate": 4.218676994605295e-06,
"loss": 0.0,
"step": 189
},
{
"epoch": 2.538116591928251,
"grad_norm": 0.019212787970900536,
"learning_rate": 4.202025158945855e-06,
"loss": 0.0001,
"step": 190
},
{
"epoch": 2.5515695067264574,
"grad_norm": 0.017139675095677376,
"learning_rate": 4.185231369880461e-06,
"loss": 0.0001,
"step": 191
},
{
"epoch": 2.5650224215246635,
"grad_norm": 0.0013709627091884613,
"learning_rate": 4.168297028055599e-06,
"loss": 0.0002,
"step": 192
},
{
"epoch": 2.57847533632287,
"grad_norm": 0.07277967780828476,
"learning_rate": 4.151223545840225e-06,
"loss": 0.0,
"step": 193
},
{
"epoch": 2.591928251121076,
"grad_norm": 0.0014003290561959147,
"learning_rate": 4.134012347207974e-06,
"loss": 0.0001,
"step": 194
},
{
"epoch": 2.6053811659192823,
"grad_norm": 0.04370618611574173,
"learning_rate": 4.116664867618395e-06,
"loss": 0.0004,
"step": 195
},
{
"epoch": 2.618834080717489,
"grad_norm": 0.24697266519069672,
"learning_rate": 4.099182553897228e-06,
"loss": 0.0,
"step": 196
},
{
"epoch": 2.6322869955156953,
"grad_norm": 0.0013013904681429267,
"learning_rate": 4.081566864115741e-06,
"loss": 0.0,
"step": 197
},
{
"epoch": 2.6457399103139014,
"grad_norm": 0.001239327946677804,
"learning_rate": 4.063819267469114e-06,
"loss": 0.0,
"step": 198
},
{
"epoch": 2.6591928251121075,
"grad_norm": 0.007082940544933081,
"learning_rate": 4.04594124415391e-06,
"loss": 0.0,
"step": 199
},
{
"epoch": 2.672645739910314,
"grad_norm": 0.00556205864995718,
"learning_rate": 4.027934285244624e-06,
"loss": 0.0001,
"step": 200
},
{
"epoch": 2.68609865470852,
"grad_norm": 0.025378312915563583,
"learning_rate": 4.009799892569317e-06,
"loss": 0.0001,
"step": 201
},
{
"epoch": 2.6995515695067267,
"grad_norm": 0.006344004534184933,
"learning_rate": 3.991539578584368e-06,
"loss": 0.0,
"step": 202
},
{
"epoch": 2.713004484304933,
"grad_norm": 0.0016575426561757922,
"learning_rate": 3.973154866248323e-06,
"loss": 0.0,
"step": 203
},
{
"epoch": 2.726457399103139,
"grad_norm": 0.0020909749437123537,
"learning_rate": 3.9546472888948825e-06,
"loss": 0.0,
"step": 204
},
{
"epoch": 2.7399103139013454,
"grad_norm": 0.004855802282691002,
"learning_rate": 3.936018390105013e-06,
"loss": 0.0,
"step": 205
},
{
"epoch": 2.7533632286995515,
"grad_norm": 0.0082467095926404,
"learning_rate": 3.917269723578212e-06,
"loss": 0.0,
"step": 206
},
{
"epoch": 2.766816143497758,
"grad_norm": 0.0013344286708161235,
"learning_rate": 3.898402853002921e-06,
"loss": 0.0,
"step": 207
},
{
"epoch": 2.780269058295964,
"grad_norm": 0.0034061160404235125,
"learning_rate": 3.879419351926115e-06,
"loss": 0.0,
"step": 208
},
{
"epoch": 2.7937219730941703,
"grad_norm": 0.011912211775779724,
"learning_rate": 3.86032080362206e-06,
"loss": 0.0,
"step": 209
},
{
"epoch": 2.807174887892377,
"grad_norm": 0.00134057376999408,
"learning_rate": 3.841108800960264e-06,
"loss": 0.0,
"step": 210
},
{
"epoch": 2.820627802690583,
"grad_norm": 0.0016560767544433475,
"learning_rate": 3.8217849462726334e-06,
"loss": 0.0,
"step": 211
},
{
"epoch": 2.8340807174887894,
"grad_norm": 0.006285225041210651,
"learning_rate": 3.802350851219826e-06,
"loss": 0.0,
"step": 212
},
{
"epoch": 2.8475336322869955,
"grad_norm": 0.0018534021219238639,
"learning_rate": 3.7828081366568388e-06,
"loss": 0.0,
"step": 213
},
{
"epoch": 2.8609865470852016,
"grad_norm": 0.0041048116981983185,
"learning_rate": 3.763158432497824e-06,
"loss": 0.0,
"step": 214
},
{
"epoch": 2.874439461883408,
"grad_norm": 0.0023216214030981064,
"learning_rate": 3.743403377580149e-06,
"loss": 0.0,
"step": 215
},
{
"epoch": 2.8878923766816142,
"grad_norm": 0.000994804548099637,
"learning_rate": 3.723544619527714e-06,
"loss": 0.0,
"step": 216
},
{
"epoch": 2.901345291479821,
"grad_norm": 0.003168656025081873,
"learning_rate": 3.703583814613536e-06,
"loss": 0.0,
"step": 217
},
{
"epoch": 2.914798206278027,
"grad_norm": 0.0013909138506278396,
"learning_rate": 3.6835226276216087e-06,
"loss": 0.0,
"step": 218
},
{
"epoch": 2.928251121076233,
"grad_norm": 0.008506865240633488,
"learning_rate": 3.663362731708059e-06,
"loss": 0.0,
"step": 219
},
{
"epoch": 2.9417040358744395,
"grad_norm": 0.0017819993663579226,
"learning_rate": 3.6431058082615966e-06,
"loss": 0.0,
"step": 220
},
{
"epoch": 2.9551569506726456,
"grad_norm": 0.0014212332898750901,
"learning_rate": 3.6227535467632873e-06,
"loss": 0.0,
"step": 221
},
{
"epoch": 2.968609865470852,
"grad_norm": 0.003146026050671935,
"learning_rate": 3.6023076446456415e-06,
"loss": 0.0,
"step": 222
},
{
"epoch": 2.9820627802690582,
"grad_norm": 0.0012437553377822042,
"learning_rate": 3.581769807151044e-06,
"loss": 0.0,
"step": 223
},
{
"epoch": 2.9955156950672643,
"grad_norm": 0.017311880365014076,
"learning_rate": 3.561141747189538e-06,
"loss": 0.0001,
"step": 224
},
{
"epoch": 3.0,
"grad_norm": 0.012313771061599255,
"learning_rate": 3.5404251851959537e-06,
"loss": 0.0,
"step": 225
},
{
"epoch": 3.013452914798206,
"grad_norm": 0.0043451120145618916,
"learning_rate": 3.519621848986428e-06,
"loss": 0.0,
"step": 226
},
{
"epoch": 3.0269058295964126,
"grad_norm": 0.002464739605784416,
"learning_rate": 3.498733473614298e-06,
"loss": 0.0,
"step": 227
},
{
"epoch": 3.0403587443946187,
"grad_norm": 0.0019529856508597732,
"learning_rate": 3.47776180122539e-06,
"loss": 0.0,
"step": 228
},
{
"epoch": 3.0538116591928253,
"grad_norm": 0.0013189928140491247,
"learning_rate": 3.4567085809127247e-06,
"loss": 0.0,
"step": 229
},
{
"epoch": 3.0672645739910314,
"grad_norm": 0.0021529668010771275,
"learning_rate": 3.435575568570633e-06,
"loss": 0.0,
"step": 230
},
{
"epoch": 3.0807174887892375,
"grad_norm": 0.007290184032171965,
"learning_rate": 3.4143645267483144e-06,
"loss": 0.0,
"step": 231
},
{
"epoch": 3.094170403587444,
"grad_norm": 0.001345694880001247,
"learning_rate": 3.393077224502832e-06,
"loss": 0.0,
"step": 232
},
{
"epoch": 3.10762331838565,
"grad_norm": 0.004343180451542139,
"learning_rate": 3.3717154372515716e-06,
"loss": 0.0,
"step": 233
},
{
"epoch": 3.1210762331838566,
"grad_norm": 0.0010816323338076472,
"learning_rate": 3.350280946624166e-06,
"loss": 0.0,
"step": 234
},
{
"epoch": 3.1345291479820627,
"grad_norm": 0.003842687699943781,
"learning_rate": 3.3287755403139007e-06,
"loss": 0.0,
"step": 235
},
{
"epoch": 3.1479820627802693,
"grad_norm": 0.0013102421071380377,
"learning_rate": 3.3072010119286156e-06,
"loss": 0.0,
"step": 236
},
{
"epoch": 3.1614349775784754,
"grad_norm": 0.0014620574656873941,
"learning_rate": 3.2855591608411203e-06,
"loss": 0.0,
"step": 237
},
{
"epoch": 3.1748878923766815,
"grad_norm": 0.0007613594643771648,
"learning_rate": 3.2638517920391095e-06,
"loss": 0.0,
"step": 238
},
{
"epoch": 3.188340807174888,
"grad_norm": 0.001927120960317552,
"learning_rate": 3.2420807159746333e-06,
"loss": 0.0,
"step": 239
},
{
"epoch": 3.201793721973094,
"grad_norm": 0.0047145108692348,
"learning_rate": 3.2202477484130947e-06,
"loss": 0.0,
"step": 240
},
{
"epoch": 3.2152466367713006,
"grad_norm": 0.001525243278592825,
"learning_rate": 3.1983547102818104e-06,
"loss": 0.0,
"step": 241
},
{
"epoch": 3.2286995515695067,
"grad_norm": 0.0008274471038021147,
"learning_rate": 3.1764034275181436e-06,
"loss": 0.0,
"step": 242
},
{
"epoch": 3.242152466367713,
"grad_norm": 0.0009357924573123455,
"learning_rate": 3.1543957309172136e-06,
"loss": 0.0,
"step": 243
},
{
"epoch": 3.2556053811659194,
"grad_norm": 0.0016225146828219295,
"learning_rate": 3.132333455979202e-06,
"loss": 0.0,
"step": 244
},
{
"epoch": 3.2690582959641254,
"grad_norm": 0.013506707735359669,
"learning_rate": 3.1102184427562696e-06,
"loss": 0.0001,
"step": 245
},
{
"epoch": 3.282511210762332,
"grad_norm": 0.0010258476249873638,
"learning_rate": 3.0880525356990898e-06,
"loss": 0.0,
"step": 246
},
{
"epoch": 3.295964125560538,
"grad_norm": 0.006462691817432642,
"learning_rate": 3.0658375835030148e-06,
"loss": 0.0,
"step": 247
},
{
"epoch": 3.3094170403587446,
"grad_norm": 0.0011441456153988838,
"learning_rate": 3.043575438953893e-06,
"loss": 0.0,
"step": 248
},
{
"epoch": 3.3228699551569507,
"grad_norm": 0.000751970277633518,
"learning_rate": 3.02126795877354e-06,
"loss": 0.0,
"step": 249
},
{
"epoch": 3.336322869955157,
"grad_norm": 0.006623424123972654,
"learning_rate": 2.9989170034648823e-06,
"loss": 0.0,
"step": 250
},
{
"epoch": 3.3497757847533634,
"grad_norm": 0.0021056546829640865,
"learning_rate": 2.9765244371567873e-06,
"loss": 0.0,
"step": 251
},
{
"epoch": 3.3632286995515694,
"grad_norm": 0.003343602642416954,
"learning_rate": 2.9540921274485913e-06,
"loss": 0.0,
"step": 252
},
{
"epoch": 3.376681614349776,
"grad_norm": 0.0032404386438429356,
"learning_rate": 2.9316219452543342e-06,
"loss": 0.0,
"step": 253
},
{
"epoch": 3.390134529147982,
"grad_norm": 0.02200383134186268,
"learning_rate": 2.9091157646467205e-06,
"loss": 0.0001,
"step": 254
},
{
"epoch": 3.403587443946188,
"grad_norm": 0.0015861240681260824,
"learning_rate": 2.886575462700821e-06,
"loss": 0.0,
"step": 255
},
{
"epoch": 3.4170403587443947,
"grad_norm": 0.0023104625288397074,
"learning_rate": 2.864002919337513e-06,
"loss": 0.0,
"step": 256
},
{
"epoch": 3.430493273542601,
"grad_norm": 0.0009963945485651493,
"learning_rate": 2.8414000171666952e-06,
"loss": 0.0,
"step": 257
},
{
"epoch": 3.4439461883408073,
"grad_norm": 0.006071928422898054,
"learning_rate": 2.81876864133027e-06,
"loss": 0.0,
"step": 258
},
{
"epoch": 3.4573991031390134,
"grad_norm": 0.003969075623899698,
"learning_rate": 2.7961106793449217e-06,
"loss": 0.0,
"step": 259
},
{
"epoch": 3.4708520179372195,
"grad_norm": 0.0010353871621191502,
"learning_rate": 2.773428020944687e-06,
"loss": 0.0,
"step": 260
},
{
"epoch": 3.484304932735426,
"grad_norm": 0.003665305208414793,
"learning_rate": 2.7507225579233487e-06,
"loss": 0.0,
"step": 261
},
{
"epoch": 3.497757847533632,
"grad_norm": 0.003311133710667491,
"learning_rate": 2.727996183976659e-06,
"loss": 0.0,
"step": 262
},
{
"epoch": 3.5112107623318387,
"grad_norm": 0.004625072702765465,
"learning_rate": 2.705250794544393e-06,
"loss": 0.0,
"step": 263
},
{
"epoch": 3.524663677130045,
"grad_norm": 0.0023010042496025562,
"learning_rate": 2.682488286652269e-06,
"loss": 0.0,
"step": 264
},
{
"epoch": 3.538116591928251,
"grad_norm": 0.0008537416579201818,
"learning_rate": 2.6597105587537307e-06,
"loss": 0.0,
"step": 265
},
{
"epoch": 3.5515695067264574,
"grad_norm": 0.0014414336765184999,
"learning_rate": 2.6369195105716087e-06,
"loss": 0.0,
"step": 266
},
{
"epoch": 3.5650224215246635,
"grad_norm": 0.003060834715142846,
"learning_rate": 2.614117042939685e-06,
"loss": 0.0,
"step": 267
},
{
"epoch": 3.57847533632287,
"grad_norm": 0.002594274003058672,
"learning_rate": 2.591305057644148e-06,
"loss": 0.0,
"step": 268
},
{
"epoch": 3.591928251121076,
"grad_norm": 0.0009433673694729805,
"learning_rate": 2.5684854572649876e-06,
"loss": 0.0,
"step": 269
},
{
"epoch": 3.6053811659192823,
"grad_norm": 0.000922717503271997,
"learning_rate": 2.5456601450173123e-06,
"loss": 0.0,
"step": 270
},
{
"epoch": 3.618834080717489,
"grad_norm": 0.0013827934162691236,
"learning_rate": 2.522831024592615e-06,
"loss": 0.0,
"step": 271
},
{
"epoch": 3.6322869955156953,
"grad_norm": 0.0007690931670367718,
"learning_rate": 2.5e-06,
"loss": 0.0,
"step": 272
},
{
"epoch": 3.6457399103139014,
"grad_norm": 0.002565343165770173,
"learning_rate": 2.4771689754073856e-06,
"loss": 0.0,
"step": 273
},
{
"epoch": 3.6591928251121075,
"grad_norm": 0.0013943302910774946,
"learning_rate": 2.454339854982688e-06,
"loss": 0.0,
"step": 274
},
{
"epoch": 3.672645739910314,
"grad_norm": 0.0006668591522611678,
"learning_rate": 2.4315145427350132e-06,
"loss": 0.0,
"step": 275
},
{
"epoch": 3.68609865470852,
"grad_norm": 0.0025589216966181993,
"learning_rate": 2.408694942355853e-06,
"loss": 0.0,
"step": 276
},
{
"epoch": 3.6995515695067267,
"grad_norm": 0.001420054235495627,
"learning_rate": 2.3858829570603157e-06,
"loss": 0.0,
"step": 277
},
{
"epoch": 3.713004484304933,
"grad_norm": 0.0014220779994502664,
"learning_rate": 2.363080489428391e-06,
"loss": 0.0,
"step": 278
},
{
"epoch": 3.726457399103139,
"grad_norm": 0.0009400748531334102,
"learning_rate": 2.3402894412462697e-06,
"loss": 0.0,
"step": 279
},
{
"epoch": 3.7399103139013454,
"grad_norm": 0.0006919830339029431,
"learning_rate": 2.317511713347731e-06,
"loss": 0.0,
"step": 280
},
{
"epoch": 3.7533632286995515,
"grad_norm": 0.0011200032895430923,
"learning_rate": 2.2947492054556075e-06,
"loss": 0.0,
"step": 281
},
{
"epoch": 3.766816143497758,
"grad_norm": 0.0008829529979266226,
"learning_rate": 2.272003816023341e-06,
"loss": 0.0,
"step": 282
},
{
"epoch": 3.780269058295964,
"grad_norm": 0.0009032113594003022,
"learning_rate": 2.2492774420766517e-06,
"loss": 0.0,
"step": 283
},
{
"epoch": 3.7937219730941703,
"grad_norm": 0.0012500348966568708,
"learning_rate": 2.2265719790553147e-06,
"loss": 0.0,
"step": 284
},
{
"epoch": 3.807174887892377,
"grad_norm": 0.0008029688615351915,
"learning_rate": 2.20388932065508e-06,
"loss": 0.0,
"step": 285
},
{
"epoch": 3.820627802690583,
"grad_norm": 0.0015973382396623492,
"learning_rate": 2.1812313586697307e-06,
"loss": 0.0,
"step": 286
},
{
"epoch": 3.8340807174887894,
"grad_norm": 0.0021750489249825478,
"learning_rate": 2.1585999828333065e-06,
"loss": 0.0,
"step": 287
},
{
"epoch": 3.8475336322869955,
"grad_norm": 0.004780885297805071,
"learning_rate": 2.1359970806624886e-06,
"loss": 0.0,
"step": 288
},
{
"epoch": 3.8609865470852016,
"grad_norm": 0.0007556549389846623,
"learning_rate": 2.11342453729918e-06,
"loss": 0.0,
"step": 289
},
{
"epoch": 3.874439461883408,
"grad_norm": 0.0006669393624179065,
"learning_rate": 2.0908842353532803e-06,
"loss": 0.0,
"step": 290
},
{
"epoch": 3.8878923766816142,
"grad_norm": 0.001046078628860414,
"learning_rate": 2.0683780547456666e-06,
"loss": 0.0,
"step": 291
},
{
"epoch": 3.901345291479821,
"grad_norm": 0.004534538835287094,
"learning_rate": 2.045907872551409e-06,
"loss": 0.0,
"step": 292
},
{
"epoch": 3.914798206278027,
"grad_norm": 0.0008524219738319516,
"learning_rate": 2.0234755628432135e-06,
"loss": 0.0,
"step": 293
},
{
"epoch": 3.928251121076233,
"grad_norm": 0.0006630662246607244,
"learning_rate": 2.0010829965351185e-06,
"loss": 0.0,
"step": 294
},
{
"epoch": 3.9417040358744395,
"grad_norm": 0.008820832706987858,
"learning_rate": 1.978732041226461e-06,
"loss": 0.0,
"step": 295
},
{
"epoch": 3.9551569506726456,
"grad_norm": 0.0007411285769194365,
"learning_rate": 1.956424561046108e-06,
"loss": 0.0,
"step": 296
}
],
"logging_steps": 1,
"max_steps": 444,
"num_input_tokens_seen": 0,
"num_train_epochs": 6,
"save_steps": 74,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 6.586020432455926e+17,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}