MNLP_M3_rag_model / checkpoint-2800 /trainer_state.json
qchapp's picture
Upload folder using huggingface_hub
0218dff verified
{
"best_global_step": 2800,
"best_metric": 0.94057297706604,
"best_model_checkpoint": "models/MNLP_M3_rag_model_test/checkpoint-2800",
"epoch": 2.923237597911227,
"eval_steps": 200,
"global_step": 2800,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.010443864229765013,
"grad_norm": 10.195783615112305,
"learning_rate": 9.000000000000001e-07,
"loss": 3.4273,
"step": 10
},
{
"epoch": 0.020887728459530026,
"grad_norm": 7.810600280761719,
"learning_rate": 1.9000000000000002e-06,
"loss": 3.2543,
"step": 20
},
{
"epoch": 0.031331592689295036,
"grad_norm": 5.423489570617676,
"learning_rate": 2.9e-06,
"loss": 3.0848,
"step": 30
},
{
"epoch": 0.04177545691906005,
"grad_norm": 6.003882884979248,
"learning_rate": 3.900000000000001e-06,
"loss": 3.0378,
"step": 40
},
{
"epoch": 0.05221932114882506,
"grad_norm": 5.395635604858398,
"learning_rate": 4.9000000000000005e-06,
"loss": 2.935,
"step": 50
},
{
"epoch": 0.06266318537859007,
"grad_norm": 5.4613823890686035,
"learning_rate": 5.9e-06,
"loss": 2.8095,
"step": 60
},
{
"epoch": 0.0731070496083551,
"grad_norm": 5.638515472412109,
"learning_rate": 6.9e-06,
"loss": 2.8053,
"step": 70
},
{
"epoch": 0.0835509138381201,
"grad_norm": 5.723353385925293,
"learning_rate": 7.9e-06,
"loss": 2.7206,
"step": 80
},
{
"epoch": 0.09399477806788512,
"grad_norm": 6.676548480987549,
"learning_rate": 8.900000000000001e-06,
"loss": 2.6614,
"step": 90
},
{
"epoch": 0.10443864229765012,
"grad_norm": 6.09738302230835,
"learning_rate": 9.9e-06,
"loss": 2.4909,
"step": 100
},
{
"epoch": 0.11488250652741515,
"grad_norm": 6.762812614440918,
"learning_rate": 9.967520750631542e-06,
"loss": 2.4897,
"step": 110
},
{
"epoch": 0.12532637075718014,
"grad_norm": 6.7795796394348145,
"learning_rate": 9.931432695777698e-06,
"loss": 2.3994,
"step": 120
},
{
"epoch": 0.13577023498694518,
"grad_norm": 6.7266669273376465,
"learning_rate": 9.895344640923855e-06,
"loss": 2.395,
"step": 130
},
{
"epoch": 0.1462140992167102,
"grad_norm": 6.685121536254883,
"learning_rate": 9.859256586070011e-06,
"loss": 2.3435,
"step": 140
},
{
"epoch": 0.1566579634464752,
"grad_norm": 7.3826985359191895,
"learning_rate": 9.823168531216168e-06,
"loss": 2.1847,
"step": 150
},
{
"epoch": 0.1671018276762402,
"grad_norm": 6.167830467224121,
"learning_rate": 9.787080476362326e-06,
"loss": 2.2339,
"step": 160
},
{
"epoch": 0.17754569190600522,
"grad_norm": 9.586319923400879,
"learning_rate": 9.750992421508482e-06,
"loss": 2.1565,
"step": 170
},
{
"epoch": 0.18798955613577023,
"grad_norm": 9.096244812011719,
"learning_rate": 9.714904366654639e-06,
"loss": 2.2398,
"step": 180
},
{
"epoch": 0.19843342036553524,
"grad_norm": 9.149866104125977,
"learning_rate": 9.68242511728618e-06,
"loss": 2.0402,
"step": 190
},
{
"epoch": 0.20887728459530025,
"grad_norm": 7.662464141845703,
"learning_rate": 9.646337062432336e-06,
"loss": 1.853,
"step": 200
},
{
"epoch": 0.20887728459530025,
"eval_loss": 1.9208216667175293,
"eval_runtime": 23.2773,
"eval_samples_per_second": 36.559,
"eval_steps_per_second": 4.597,
"step": 200
},
{
"epoch": 0.2193211488250653,
"grad_norm": 7.436556339263916,
"learning_rate": 9.610249007578492e-06,
"loss": 1.9207,
"step": 210
},
{
"epoch": 0.2297650130548303,
"grad_norm": 7.291353225708008,
"learning_rate": 9.574160952724649e-06,
"loss": 1.9513,
"step": 220
},
{
"epoch": 0.2402088772845953,
"grad_norm": 7.357730865478516,
"learning_rate": 9.538072897870805e-06,
"loss": 1.8785,
"step": 230
},
{
"epoch": 0.2506527415143603,
"grad_norm": 7.417892932891846,
"learning_rate": 9.501984843016962e-06,
"loss": 1.8304,
"step": 240
},
{
"epoch": 0.26109660574412535,
"grad_norm": 7.092027187347412,
"learning_rate": 9.46589678816312e-06,
"loss": 1.7244,
"step": 250
},
{
"epoch": 0.27154046997389036,
"grad_norm": 8.140602111816406,
"learning_rate": 9.429808733309276e-06,
"loss": 1.9412,
"step": 260
},
{
"epoch": 0.2819843342036554,
"grad_norm": 6.528462886810303,
"learning_rate": 9.393720678455433e-06,
"loss": 1.7654,
"step": 270
},
{
"epoch": 0.2924281984334204,
"grad_norm": 6.523935794830322,
"learning_rate": 9.35763262360159e-06,
"loss": 1.8401,
"step": 280
},
{
"epoch": 0.3028720626631854,
"grad_norm": 6.85488224029541,
"learning_rate": 9.321544568747746e-06,
"loss": 1.8402,
"step": 290
},
{
"epoch": 0.3133159268929504,
"grad_norm": 6.1414875984191895,
"learning_rate": 9.285456513893902e-06,
"loss": 1.682,
"step": 300
},
{
"epoch": 0.3237597911227154,
"grad_norm": 7.549343109130859,
"learning_rate": 9.249368459040059e-06,
"loss": 1.5124,
"step": 310
},
{
"epoch": 0.3342036553524804,
"grad_norm": 7.781056880950928,
"learning_rate": 9.213280404186215e-06,
"loss": 1.6474,
"step": 320
},
{
"epoch": 0.34464751958224543,
"grad_norm": 6.674058437347412,
"learning_rate": 9.177192349332372e-06,
"loss": 1.6971,
"step": 330
},
{
"epoch": 0.35509138381201044,
"grad_norm": 8.93964958190918,
"learning_rate": 9.14110429447853e-06,
"loss": 1.5545,
"step": 340
},
{
"epoch": 0.36553524804177545,
"grad_norm": 7.935058116912842,
"learning_rate": 9.105016239624686e-06,
"loss": 1.6633,
"step": 350
},
{
"epoch": 0.37597911227154046,
"grad_norm": 6.446653842926025,
"learning_rate": 9.068928184770843e-06,
"loss": 1.6209,
"step": 360
},
{
"epoch": 0.38642297650130547,
"grad_norm": 9.611429214477539,
"learning_rate": 9.036448935402382e-06,
"loss": 1.4615,
"step": 370
},
{
"epoch": 0.3968668407310705,
"grad_norm": 8.988059043884277,
"learning_rate": 9.000360880548538e-06,
"loss": 1.5815,
"step": 380
},
{
"epoch": 0.4073107049608355,
"grad_norm": 7.07716178894043,
"learning_rate": 8.964272825694695e-06,
"loss": 1.6719,
"step": 390
},
{
"epoch": 0.4177545691906005,
"grad_norm": 6.789717674255371,
"learning_rate": 8.928184770840851e-06,
"loss": 1.4354,
"step": 400
},
{
"epoch": 0.4177545691906005,
"eval_loss": 1.4963077306747437,
"eval_runtime": 23.1593,
"eval_samples_per_second": 36.746,
"eval_steps_per_second": 4.62,
"step": 400
},
{
"epoch": 0.4281984334203655,
"grad_norm": 6.676051139831543,
"learning_rate": 8.89209671598701e-06,
"loss": 1.695,
"step": 410
},
{
"epoch": 0.4386422976501306,
"grad_norm": 6.711126327514648,
"learning_rate": 8.856008661133166e-06,
"loss": 1.6562,
"step": 420
},
{
"epoch": 0.4490861618798956,
"grad_norm": 7.0567731857299805,
"learning_rate": 8.819920606279322e-06,
"loss": 1.6646,
"step": 430
},
{
"epoch": 0.4595300261096606,
"grad_norm": 6.944768905639648,
"learning_rate": 8.783832551425479e-06,
"loss": 1.4864,
"step": 440
},
{
"epoch": 0.4699738903394256,
"grad_norm": 6.025828838348389,
"learning_rate": 8.747744496571635e-06,
"loss": 1.4261,
"step": 450
},
{
"epoch": 0.4804177545691906,
"grad_norm": 7.318626880645752,
"learning_rate": 8.711656441717792e-06,
"loss": 1.4183,
"step": 460
},
{
"epoch": 0.4908616187989556,
"grad_norm": 9.207475662231445,
"learning_rate": 8.675568386863948e-06,
"loss": 1.6308,
"step": 470
},
{
"epoch": 0.5013054830287206,
"grad_norm": 7.152507781982422,
"learning_rate": 8.639480332010105e-06,
"loss": 1.5118,
"step": 480
},
{
"epoch": 0.5117493472584856,
"grad_norm": 6.89570426940918,
"learning_rate": 8.603392277156261e-06,
"loss": 1.6088,
"step": 490
},
{
"epoch": 0.5221932114882507,
"grad_norm": 8.016292572021484,
"learning_rate": 8.567304222302419e-06,
"loss": 1.6461,
"step": 500
},
{
"epoch": 0.5326370757180157,
"grad_norm": 5.700518608093262,
"learning_rate": 8.531216167448576e-06,
"loss": 1.5497,
"step": 510
},
{
"epoch": 0.5430809399477807,
"grad_norm": 7.127388000488281,
"learning_rate": 8.495128112594732e-06,
"loss": 1.4443,
"step": 520
},
{
"epoch": 0.5535248041775457,
"grad_norm": 9.201033592224121,
"learning_rate": 8.459040057740888e-06,
"loss": 1.3384,
"step": 530
},
{
"epoch": 0.5639686684073107,
"grad_norm": 9.565522193908691,
"learning_rate": 8.422952002887045e-06,
"loss": 1.3732,
"step": 540
},
{
"epoch": 0.5744125326370757,
"grad_norm": 6.29095458984375,
"learning_rate": 8.386863948033201e-06,
"loss": 1.5768,
"step": 550
},
{
"epoch": 0.5848563968668408,
"grad_norm": 7.279628753662109,
"learning_rate": 8.350775893179358e-06,
"loss": 1.5237,
"step": 560
},
{
"epoch": 0.5953002610966057,
"grad_norm": 5.866252899169922,
"learning_rate": 8.314687838325514e-06,
"loss": 1.491,
"step": 570
},
{
"epoch": 0.6057441253263708,
"grad_norm": 6.8628621101379395,
"learning_rate": 8.27859978347167e-06,
"loss": 1.3991,
"step": 580
},
{
"epoch": 0.6161879895561357,
"grad_norm": 5.9102654457092285,
"learning_rate": 8.242511728617829e-06,
"loss": 1.3842,
"step": 590
},
{
"epoch": 0.6266318537859008,
"grad_norm": 6.6509199142456055,
"learning_rate": 8.206423673763985e-06,
"loss": 1.4015,
"step": 600
},
{
"epoch": 0.6266318537859008,
"eval_loss": 1.3116086721420288,
"eval_runtime": 23.131,
"eval_samples_per_second": 36.79,
"eval_steps_per_second": 4.626,
"step": 600
},
{
"epoch": 0.6370757180156658,
"grad_norm": 10.004154205322266,
"learning_rate": 8.170335618910142e-06,
"loss": 1.4046,
"step": 610
},
{
"epoch": 0.6475195822454308,
"grad_norm": 6.851074695587158,
"learning_rate": 8.134247564056298e-06,
"loss": 1.3358,
"step": 620
},
{
"epoch": 0.6579634464751958,
"grad_norm": 6.676290512084961,
"learning_rate": 8.098159509202455e-06,
"loss": 1.2649,
"step": 630
},
{
"epoch": 0.6684073107049608,
"grad_norm": 6.854121685028076,
"learning_rate": 8.062071454348611e-06,
"loss": 1.4563,
"step": 640
},
{
"epoch": 0.6788511749347258,
"grad_norm": 7.104318141937256,
"learning_rate": 8.025983399494768e-06,
"loss": 1.2667,
"step": 650
},
{
"epoch": 0.6892950391644909,
"grad_norm": 6.3163957595825195,
"learning_rate": 7.989895344640924e-06,
"loss": 1.4532,
"step": 660
},
{
"epoch": 0.6997389033942559,
"grad_norm": 7.5210652351379395,
"learning_rate": 7.95380728978708e-06,
"loss": 1.179,
"step": 670
},
{
"epoch": 0.7101827676240209,
"grad_norm": 7.099525451660156,
"learning_rate": 7.917719234933237e-06,
"loss": 1.2406,
"step": 680
},
{
"epoch": 0.720626631853786,
"grad_norm": 6.814334392547607,
"learning_rate": 7.881631180079395e-06,
"loss": 1.1178,
"step": 690
},
{
"epoch": 0.7310704960835509,
"grad_norm": 6.895037651062012,
"learning_rate": 7.845543125225551e-06,
"loss": 1.3291,
"step": 700
},
{
"epoch": 0.741514360313316,
"grad_norm": 6.383569717407227,
"learning_rate": 7.809455070371708e-06,
"loss": 1.2887,
"step": 710
},
{
"epoch": 0.7519582245430809,
"grad_norm": 8.13758373260498,
"learning_rate": 7.773367015517864e-06,
"loss": 1.2576,
"step": 720
},
{
"epoch": 0.762402088772846,
"grad_norm": 5.2739362716674805,
"learning_rate": 7.73727896066402e-06,
"loss": 1.2835,
"step": 730
},
{
"epoch": 0.7728459530026109,
"grad_norm": 7.7005934715271,
"learning_rate": 7.701190905810177e-06,
"loss": 1.2633,
"step": 740
},
{
"epoch": 0.783289817232376,
"grad_norm": 6.738245964050293,
"learning_rate": 7.665102850956334e-06,
"loss": 1.2129,
"step": 750
},
{
"epoch": 0.793733681462141,
"grad_norm": 5.725705146789551,
"learning_rate": 7.629014796102491e-06,
"loss": 1.2501,
"step": 760
},
{
"epoch": 0.804177545691906,
"grad_norm": 6.256880760192871,
"learning_rate": 7.5929267412486475e-06,
"loss": 1.2567,
"step": 770
},
{
"epoch": 0.814621409921671,
"grad_norm": 6.835810661315918,
"learning_rate": 7.556838686394804e-06,
"loss": 1.2408,
"step": 780
},
{
"epoch": 0.825065274151436,
"grad_norm": 5.811525821685791,
"learning_rate": 7.52075063154096e-06,
"loss": 1.2409,
"step": 790
},
{
"epoch": 0.835509138381201,
"grad_norm": 8.554996490478516,
"learning_rate": 7.484662576687118e-06,
"loss": 1.2575,
"step": 800
},
{
"epoch": 0.835509138381201,
"eval_loss": 1.2006852626800537,
"eval_runtime": 23.1272,
"eval_samples_per_second": 36.796,
"eval_steps_per_second": 4.627,
"step": 800
},
{
"epoch": 0.8459530026109661,
"grad_norm": 7.091418266296387,
"learning_rate": 7.448574521833274e-06,
"loss": 1.0256,
"step": 810
},
{
"epoch": 0.856396866840731,
"grad_norm": 7.5542683601379395,
"learning_rate": 7.4124864669794306e-06,
"loss": 1.0605,
"step": 820
},
{
"epoch": 0.8668407310704961,
"grad_norm": 7.177179336547852,
"learning_rate": 7.376398412125587e-06,
"loss": 1.1747,
"step": 830
},
{
"epoch": 0.8772845953002611,
"grad_norm": 6.226783752441406,
"learning_rate": 7.3403103572717434e-06,
"loss": 1.1518,
"step": 840
},
{
"epoch": 0.8877284595300261,
"grad_norm": 7.062104225158691,
"learning_rate": 7.304222302417901e-06,
"loss": 1.2322,
"step": 850
},
{
"epoch": 0.8981723237597912,
"grad_norm": 8.756392478942871,
"learning_rate": 7.268134247564057e-06,
"loss": 1.271,
"step": 860
},
{
"epoch": 0.9086161879895561,
"grad_norm": 6.364953994750977,
"learning_rate": 7.232046192710214e-06,
"loss": 1.1853,
"step": 870
},
{
"epoch": 0.9190600522193212,
"grad_norm": 6.629589080810547,
"learning_rate": 7.19595813785637e-06,
"loss": 1.2279,
"step": 880
},
{
"epoch": 0.9295039164490861,
"grad_norm": 5.4719977378845215,
"learning_rate": 7.1598700830025265e-06,
"loss": 1.0789,
"step": 890
},
{
"epoch": 0.9399477806788512,
"grad_norm": 5.824141025543213,
"learning_rate": 7.123782028148684e-06,
"loss": 1.0435,
"step": 900
},
{
"epoch": 0.9503916449086162,
"grad_norm": 9.214339256286621,
"learning_rate": 7.08769397329484e-06,
"loss": 1.0518,
"step": 910
},
{
"epoch": 0.9608355091383812,
"grad_norm": 6.3007941246032715,
"learning_rate": 7.051605918440997e-06,
"loss": 1.0297,
"step": 920
},
{
"epoch": 0.9712793733681462,
"grad_norm": 5.431386947631836,
"learning_rate": 7.015517863587153e-06,
"loss": 1.1292,
"step": 930
},
{
"epoch": 0.9817232375979112,
"grad_norm": 8.941716194152832,
"learning_rate": 6.9794298087333105e-06,
"loss": 1.1617,
"step": 940
},
{
"epoch": 0.9921671018276762,
"grad_norm": 8.926597595214844,
"learning_rate": 6.943341753879467e-06,
"loss": 1.3497,
"step": 950
},
{
"epoch": 1.002088772845953,
"grad_norm": 6.556091785430908,
"learning_rate": 6.907253699025623e-06,
"loss": 1.176,
"step": 960
},
{
"epoch": 1.012532637075718,
"grad_norm": 12.022808074951172,
"learning_rate": 6.87116564417178e-06,
"loss": 0.9251,
"step": 970
},
{
"epoch": 1.022976501305483,
"grad_norm": 7.0122389793396,
"learning_rate": 6.835077589317936e-06,
"loss": 0.9231,
"step": 980
},
{
"epoch": 1.033420365535248,
"grad_norm": 6.355090141296387,
"learning_rate": 6.7989895344640936e-06,
"loss": 1.0371,
"step": 990
},
{
"epoch": 1.0438642297650131,
"grad_norm": 7.36827278137207,
"learning_rate": 6.76290147961025e-06,
"loss": 1.0666,
"step": 1000
},
{
"epoch": 1.0438642297650131,
"eval_loss": 1.1358416080474854,
"eval_runtime": 23.1354,
"eval_samples_per_second": 36.783,
"eval_steps_per_second": 4.625,
"step": 1000
},
{
"epoch": 1.054308093994778,
"grad_norm": 5.1998772621154785,
"learning_rate": 6.7268134247564065e-06,
"loss": 0.847,
"step": 1010
},
{
"epoch": 1.064751958224543,
"grad_norm": 6.828693866729736,
"learning_rate": 6.690725369902563e-06,
"loss": 1.0755,
"step": 1020
},
{
"epoch": 1.0751958224543081,
"grad_norm": 6.408750534057617,
"learning_rate": 6.65463731504872e-06,
"loss": 0.9039,
"step": 1030
},
{
"epoch": 1.0856396866840732,
"grad_norm": 7.197463035583496,
"learning_rate": 6.618549260194877e-06,
"loss": 0.959,
"step": 1040
},
{
"epoch": 1.096083550913838,
"grad_norm": 5.504906177520752,
"learning_rate": 6.582461205341033e-06,
"loss": 0.9467,
"step": 1050
},
{
"epoch": 1.106527415143603,
"grad_norm": 7.88711404800415,
"learning_rate": 6.5463731504871896e-06,
"loss": 1.0635,
"step": 1060
},
{
"epoch": 1.1169712793733682,
"grad_norm": 5.1561055183410645,
"learning_rate": 6.510285095633346e-06,
"loss": 0.9635,
"step": 1070
},
{
"epoch": 1.1274151436031332,
"grad_norm": 5.088303565979004,
"learning_rate": 6.474197040779503e-06,
"loss": 0.9605,
"step": 1080
},
{
"epoch": 1.137859007832898,
"grad_norm": 4.595157146453857,
"learning_rate": 6.43810898592566e-06,
"loss": 1.0494,
"step": 1090
},
{
"epoch": 1.1483028720626631,
"grad_norm": 6.682077884674072,
"learning_rate": 6.402020931071816e-06,
"loss": 0.888,
"step": 1100
},
{
"epoch": 1.1587467362924282,
"grad_norm": 9.08073902130127,
"learning_rate": 6.365932876217973e-06,
"loss": 1.1239,
"step": 1110
},
{
"epoch": 1.1691906005221933,
"grad_norm": 7.480021953582764,
"learning_rate": 6.329844821364129e-06,
"loss": 1.0337,
"step": 1120
},
{
"epoch": 1.1796344647519583,
"grad_norm": 5.750502109527588,
"learning_rate": 6.293756766510286e-06,
"loss": 1.132,
"step": 1130
},
{
"epoch": 1.1900783289817232,
"grad_norm": 6.151831150054932,
"learning_rate": 6.257668711656443e-06,
"loss": 0.9985,
"step": 1140
},
{
"epoch": 1.2005221932114882,
"grad_norm": 6.567698001861572,
"learning_rate": 6.221580656802599e-06,
"loss": 1.1177,
"step": 1150
},
{
"epoch": 1.2109660574412533,
"grad_norm": 6.1963887214660645,
"learning_rate": 6.185492601948756e-06,
"loss": 0.9328,
"step": 1160
},
{
"epoch": 1.2214099216710184,
"grad_norm": 5.044003009796143,
"learning_rate": 6.149404547094913e-06,
"loss": 0.8986,
"step": 1170
},
{
"epoch": 1.2318537859007832,
"grad_norm": 6.648619651794434,
"learning_rate": 6.113316492241068e-06,
"loss": 0.9905,
"step": 1180
},
{
"epoch": 1.2422976501305483,
"grad_norm": 9.110396385192871,
"learning_rate": 6.077228437387225e-06,
"loss": 1.0126,
"step": 1190
},
{
"epoch": 1.2527415143603133,
"grad_norm": 8.707374572753906,
"learning_rate": 6.0411403825333815e-06,
"loss": 0.8611,
"step": 1200
},
{
"epoch": 1.2527415143603133,
"eval_loss": 1.0788438320159912,
"eval_runtime": 23.1258,
"eval_samples_per_second": 36.799,
"eval_steps_per_second": 4.627,
"step": 1200
},
{
"epoch": 1.2631853785900784,
"grad_norm": 5.457468509674072,
"learning_rate": 6.005052327679538e-06,
"loss": 0.8327,
"step": 1210
},
{
"epoch": 1.2736292428198435,
"grad_norm": 6.050765037536621,
"learning_rate": 5.9689642728256944e-06,
"loss": 1.0782,
"step": 1220
},
{
"epoch": 1.2840731070496083,
"grad_norm": 6.254447937011719,
"learning_rate": 5.932876217971852e-06,
"loss": 0.9388,
"step": 1230
},
{
"epoch": 1.2945169712793734,
"grad_norm": 5.181304931640625,
"learning_rate": 5.896788163118008e-06,
"loss": 0.9711,
"step": 1240
},
{
"epoch": 1.3049608355091384,
"grad_norm": 6.832638263702393,
"learning_rate": 5.860700108264165e-06,
"loss": 1.0333,
"step": 1250
},
{
"epoch": 1.3154046997389033,
"grad_norm": 8.406023025512695,
"learning_rate": 5.824612053410321e-06,
"loss": 1.0902,
"step": 1260
},
{
"epoch": 1.3258485639686683,
"grad_norm": 6.346268653869629,
"learning_rate": 5.7885239985564775e-06,
"loss": 0.8651,
"step": 1270
},
{
"epoch": 1.3362924281984334,
"grad_norm": 8.447615623474121,
"learning_rate": 5.752435943702635e-06,
"loss": 1.046,
"step": 1280
},
{
"epoch": 1.3467362924281985,
"grad_norm": 8.351264953613281,
"learning_rate": 5.716347888848791e-06,
"loss": 1.029,
"step": 1290
},
{
"epoch": 1.3571801566579635,
"grad_norm": 6.036417007446289,
"learning_rate": 5.680259833994948e-06,
"loss": 1.0089,
"step": 1300
},
{
"epoch": 1.3676240208877284,
"grad_norm": 5.646811485290527,
"learning_rate": 5.644171779141104e-06,
"loss": 0.9346,
"step": 1310
},
{
"epoch": 1.3780678851174935,
"grad_norm": 6.4004950523376465,
"learning_rate": 5.608083724287261e-06,
"loss": 0.9489,
"step": 1320
},
{
"epoch": 1.3885117493472585,
"grad_norm": 5.746732234954834,
"learning_rate": 5.571995669433418e-06,
"loss": 0.9356,
"step": 1330
},
{
"epoch": 1.3989556135770234,
"grad_norm": 5.3405632972717285,
"learning_rate": 5.535907614579574e-06,
"loss": 0.7099,
"step": 1340
},
{
"epoch": 1.4093994778067884,
"grad_norm": 6.424510955810547,
"learning_rate": 5.499819559725731e-06,
"loss": 0.6955,
"step": 1350
},
{
"epoch": 1.4198433420365535,
"grad_norm": 5.751818656921387,
"learning_rate": 5.463731504871887e-06,
"loss": 0.9094,
"step": 1360
},
{
"epoch": 1.4302872062663186,
"grad_norm": 7.787084579467773,
"learning_rate": 5.4276434500180445e-06,
"loss": 0.9618,
"step": 1370
},
{
"epoch": 1.4407310704960836,
"grad_norm": 6.467785835266113,
"learning_rate": 5.391555395164201e-06,
"loss": 0.8549,
"step": 1380
},
{
"epoch": 1.4511749347258487,
"grad_norm": 6.860886573791504,
"learning_rate": 5.3554673403103574e-06,
"loss": 0.8342,
"step": 1390
},
{
"epoch": 1.4616187989556135,
"grad_norm": 5.627669811248779,
"learning_rate": 5.319379285456514e-06,
"loss": 0.9012,
"step": 1400
},
{
"epoch": 1.4616187989556135,
"eval_loss": 1.0375572443008423,
"eval_runtime": 23.1318,
"eval_samples_per_second": 36.789,
"eval_steps_per_second": 4.626,
"step": 1400
},
{
"epoch": 1.4720626631853786,
"grad_norm": 7.144196033477783,
"learning_rate": 5.28329123060267e-06,
"loss": 0.9949,
"step": 1410
},
{
"epoch": 1.4825065274151437,
"grad_norm": 6.208961486816406,
"learning_rate": 5.247203175748828e-06,
"loss": 1.0719,
"step": 1420
},
{
"epoch": 1.4929503916449085,
"grad_norm": 7.110988140106201,
"learning_rate": 5.211115120894984e-06,
"loss": 0.9162,
"step": 1430
},
{
"epoch": 1.5033942558746736,
"grad_norm": 6.903599739074707,
"learning_rate": 5.1750270660411405e-06,
"loss": 0.9974,
"step": 1440
},
{
"epoch": 1.5138381201044386,
"grad_norm": 5.059232234954834,
"learning_rate": 5.138939011187297e-06,
"loss": 0.9602,
"step": 1450
},
{
"epoch": 1.5242819843342037,
"grad_norm": 3.6045143604278564,
"learning_rate": 5.1028509563334534e-06,
"loss": 0.8448,
"step": 1460
},
{
"epoch": 1.5347258485639688,
"grad_norm": 8.628995895385742,
"learning_rate": 5.066762901479611e-06,
"loss": 0.914,
"step": 1470
},
{
"epoch": 1.5451697127937338,
"grad_norm": 7.227733612060547,
"learning_rate": 5.030674846625767e-06,
"loss": 0.9206,
"step": 1480
},
{
"epoch": 1.5556135770234987,
"grad_norm": 7.930648326873779,
"learning_rate": 4.994586791771924e-06,
"loss": 1.1221,
"step": 1490
},
{
"epoch": 1.5660574412532637,
"grad_norm": 6.340338706970215,
"learning_rate": 4.95849873691808e-06,
"loss": 0.8298,
"step": 1500
},
{
"epoch": 1.5765013054830286,
"grad_norm": 5.558096408843994,
"learning_rate": 4.922410682064237e-06,
"loss": 0.9457,
"step": 1510
},
{
"epoch": 1.5869451697127936,
"grad_norm": 7.608903408050537,
"learning_rate": 4.886322627210394e-06,
"loss": 0.9164,
"step": 1520
},
{
"epoch": 1.5973890339425587,
"grad_norm": 9.46885871887207,
"learning_rate": 4.85023457235655e-06,
"loss": 0.9498,
"step": 1530
},
{
"epoch": 1.6078328981723238,
"grad_norm": 7.6691107749938965,
"learning_rate": 4.814146517502707e-06,
"loss": 0.9764,
"step": 1540
},
{
"epoch": 1.6182767624020888,
"grad_norm": 8.231538772583008,
"learning_rate": 4.778058462648863e-06,
"loss": 0.8314,
"step": 1550
},
{
"epoch": 1.628720626631854,
"grad_norm": 5.868668556213379,
"learning_rate": 4.7419704077950205e-06,
"loss": 0.9127,
"step": 1560
},
{
"epoch": 1.6391644908616188,
"grad_norm": 7.236291885375977,
"learning_rate": 4.705882352941177e-06,
"loss": 0.9485,
"step": 1570
},
{
"epoch": 1.6496083550913838,
"grad_norm": 6.871162414550781,
"learning_rate": 4.669794298087333e-06,
"loss": 1.0603,
"step": 1580
},
{
"epoch": 1.6600522193211487,
"grad_norm": 6.303982734680176,
"learning_rate": 4.63370624323349e-06,
"loss": 0.945,
"step": 1590
},
{
"epoch": 1.6704960835509137,
"grad_norm": 7.167915344238281,
"learning_rate": 4.597618188379647e-06,
"loss": 0.7351,
"step": 1600
},
{
"epoch": 1.6704960835509137,
"eval_loss": 1.0095878839492798,
"eval_runtime": 23.1917,
"eval_samples_per_second": 36.694,
"eval_steps_per_second": 4.614,
"step": 1600
},
{
"epoch": 1.6809399477806788,
"grad_norm": 6.626043796539307,
"learning_rate": 4.5615301335258035e-06,
"loss": 0.9369,
"step": 1610
},
{
"epoch": 1.6913838120104439,
"grad_norm": 6.645303726196289,
"learning_rate": 4.52544207867196e-06,
"loss": 0.8509,
"step": 1620
},
{
"epoch": 1.701827676240209,
"grad_norm": 7.132906913757324,
"learning_rate": 4.4893540238181164e-06,
"loss": 0.9522,
"step": 1630
},
{
"epoch": 1.712271540469974,
"grad_norm": 6.155941009521484,
"learning_rate": 4.453265968964273e-06,
"loss": 0.9662,
"step": 1640
},
{
"epoch": 1.722715404699739,
"grad_norm": 5.0147705078125,
"learning_rate": 4.41717791411043e-06,
"loss": 0.8755,
"step": 1650
},
{
"epoch": 1.733159268929504,
"grad_norm": 7.039682388305664,
"learning_rate": 4.381089859256587e-06,
"loss": 0.8761,
"step": 1660
},
{
"epoch": 1.743603133159269,
"grad_norm": 8.536678314208984,
"learning_rate": 4.345001804402743e-06,
"loss": 0.8054,
"step": 1670
},
{
"epoch": 1.7540469973890338,
"grad_norm": 6.860482215881348,
"learning_rate": 4.3089137495488995e-06,
"loss": 0.9095,
"step": 1680
},
{
"epoch": 1.7644908616187989,
"grad_norm": 6.3796563148498535,
"learning_rate": 4.272825694695056e-06,
"loss": 1.08,
"step": 1690
},
{
"epoch": 1.774934725848564,
"grad_norm": 11.013219833374023,
"learning_rate": 4.236737639841213e-06,
"loss": 0.834,
"step": 1700
},
{
"epoch": 1.785378590078329,
"grad_norm": 6.231220722198486,
"learning_rate": 4.20064958498737e-06,
"loss": 0.8488,
"step": 1710
},
{
"epoch": 1.795822454308094,
"grad_norm": 7.019144535064697,
"learning_rate": 4.164561530133526e-06,
"loss": 0.7922,
"step": 1720
},
{
"epoch": 1.8062663185378591,
"grad_norm": 6.586852550506592,
"learning_rate": 4.128473475279683e-06,
"loss": 0.8326,
"step": 1730
},
{
"epoch": 1.816710182767624,
"grad_norm": 6.34022855758667,
"learning_rate": 4.09238542042584e-06,
"loss": 0.906,
"step": 1740
},
{
"epoch": 1.827154046997389,
"grad_norm": 7.541686058044434,
"learning_rate": 4.056297365571996e-06,
"loss": 0.9029,
"step": 1750
},
{
"epoch": 1.837597911227154,
"grad_norm": 5.867885589599609,
"learning_rate": 4.020209310718153e-06,
"loss": 0.8072,
"step": 1760
},
{
"epoch": 1.848041775456919,
"grad_norm": 4.805484771728516,
"learning_rate": 3.984121255864309e-06,
"loss": 0.9158,
"step": 1770
},
{
"epoch": 1.858485639686684,
"grad_norm": 5.949447154998779,
"learning_rate": 3.948033201010466e-06,
"loss": 0.8116,
"step": 1780
},
{
"epoch": 1.868929503916449,
"grad_norm": 7.641289234161377,
"learning_rate": 3.911945146156623e-06,
"loss": 0.8824,
"step": 1790
},
{
"epoch": 1.8793733681462141,
"grad_norm": 5.530484199523926,
"learning_rate": 3.8758570913027795e-06,
"loss": 0.9264,
"step": 1800
},
{
"epoch": 1.8793733681462141,
"eval_loss": 0.9786838293075562,
"eval_runtime": 23.1205,
"eval_samples_per_second": 36.807,
"eval_steps_per_second": 4.628,
"step": 1800
},
{
"epoch": 1.8898172323759792,
"grad_norm": 5.981470584869385,
"learning_rate": 3.839769036448936e-06,
"loss": 0.7929,
"step": 1810
},
{
"epoch": 1.9002610966057443,
"grad_norm": 8.019466400146484,
"learning_rate": 3.8036809815950928e-06,
"loss": 0.7616,
"step": 1820
},
{
"epoch": 1.9107049608355091,
"grad_norm": 7.649405479431152,
"learning_rate": 3.7675929267412492e-06,
"loss": 0.8661,
"step": 1830
},
{
"epoch": 1.9211488250652742,
"grad_norm": 8.259872436523438,
"learning_rate": 3.7315048718874057e-06,
"loss": 1.0303,
"step": 1840
},
{
"epoch": 1.931592689295039,
"grad_norm": 5.947850227355957,
"learning_rate": 3.695416817033562e-06,
"loss": 0.9326,
"step": 1850
},
{
"epoch": 1.942036553524804,
"grad_norm": 5.193607330322266,
"learning_rate": 3.6593287621797186e-06,
"loss": 0.8527,
"step": 1860
},
{
"epoch": 1.9524804177545692,
"grad_norm": 6.412321090698242,
"learning_rate": 3.623240707325875e-06,
"loss": 0.8768,
"step": 1870
},
{
"epoch": 1.9629242819843342,
"grad_norm": 6.859325408935547,
"learning_rate": 3.587152652472032e-06,
"loss": 0.8131,
"step": 1880
},
{
"epoch": 1.9733681462140993,
"grad_norm": 4.910820484161377,
"learning_rate": 3.5510645976181883e-06,
"loss": 0.7232,
"step": 1890
},
{
"epoch": 1.9838120104438643,
"grad_norm": 6.052480220794678,
"learning_rate": 3.5149765427643452e-06,
"loss": 0.7567,
"step": 1900
},
{
"epoch": 1.9942558746736292,
"grad_norm": 6.609030246734619,
"learning_rate": 3.4788884879105017e-06,
"loss": 0.8219,
"step": 1910
},
{
"epoch": 2.004177545691906,
"grad_norm": 4.539740562438965,
"learning_rate": 3.4428004330566585e-06,
"loss": 0.8498,
"step": 1920
},
{
"epoch": 2.014621409921671,
"grad_norm": 5.981167316436768,
"learning_rate": 3.406712378202815e-06,
"loss": 0.7024,
"step": 1930
},
{
"epoch": 2.025065274151436,
"grad_norm": 6.600665092468262,
"learning_rate": 3.3706243233489714e-06,
"loss": 0.7467,
"step": 1940
},
{
"epoch": 2.035509138381201,
"grad_norm": 5.560609340667725,
"learning_rate": 3.3345362684951283e-06,
"loss": 0.7434,
"step": 1950
},
{
"epoch": 2.045953002610966,
"grad_norm": 4.679533004760742,
"learning_rate": 3.2984482136412848e-06,
"loss": 0.7125,
"step": 1960
},
{
"epoch": 2.056396866840731,
"grad_norm": 7.177086353302002,
"learning_rate": 3.2623601587874416e-06,
"loss": 0.6745,
"step": 1970
},
{
"epoch": 2.066840731070496,
"grad_norm": 6.860986709594727,
"learning_rate": 3.226272103933598e-06,
"loss": 0.6814,
"step": 1980
},
{
"epoch": 2.077284595300261,
"grad_norm": 8.40719223022461,
"learning_rate": 3.190184049079755e-06,
"loss": 0.7077,
"step": 1990
},
{
"epoch": 2.0877284595300263,
"grad_norm": 5.830367088317871,
"learning_rate": 3.1540959942259114e-06,
"loss": 0.8435,
"step": 2000
},
{
"epoch": 2.0877284595300263,
"eval_loss": 0.9753687381744385,
"eval_runtime": 23.1964,
"eval_samples_per_second": 36.687,
"eval_steps_per_second": 4.613,
"step": 2000
},
{
"epoch": 2.0981723237597913,
"grad_norm": 6.964289665222168,
"learning_rate": 3.118007939372068e-06,
"loss": 0.6964,
"step": 2010
},
{
"epoch": 2.108616187989556,
"grad_norm": 6.764989852905273,
"learning_rate": 3.0819198845182247e-06,
"loss": 0.7107,
"step": 2020
},
{
"epoch": 2.119060052219321,
"grad_norm": 7.992194652557373,
"learning_rate": 3.045831829664381e-06,
"loss": 0.7146,
"step": 2030
},
{
"epoch": 2.129503916449086,
"grad_norm": 7.102138996124268,
"learning_rate": 3.009743774810538e-06,
"loss": 0.686,
"step": 2040
},
{
"epoch": 2.139947780678851,
"grad_norm": 7.446751117706299,
"learning_rate": 2.9736557199566945e-06,
"loss": 0.7681,
"step": 2050
},
{
"epoch": 2.1503916449086162,
"grad_norm": 7.091776371002197,
"learning_rate": 2.9375676651028514e-06,
"loss": 0.7674,
"step": 2060
},
{
"epoch": 2.1608355091383813,
"grad_norm": 7.994192123413086,
"learning_rate": 2.901479610249008e-06,
"loss": 0.7187,
"step": 2070
},
{
"epoch": 2.1712793733681464,
"grad_norm": 4.8329386711120605,
"learning_rate": 2.8653915553951643e-06,
"loss": 0.7501,
"step": 2080
},
{
"epoch": 2.1817232375979114,
"grad_norm": 6.802753925323486,
"learning_rate": 2.829303500541321e-06,
"loss": 0.5658,
"step": 2090
},
{
"epoch": 2.192167101827676,
"grad_norm": 7.07351541519165,
"learning_rate": 2.7932154456874776e-06,
"loss": 0.8106,
"step": 2100
},
{
"epoch": 2.202610966057441,
"grad_norm": 6.761138916015625,
"learning_rate": 2.7571273908336344e-06,
"loss": 0.7434,
"step": 2110
},
{
"epoch": 2.213054830287206,
"grad_norm": 4.220724582672119,
"learning_rate": 2.721039335979791e-06,
"loss": 0.7526,
"step": 2120
},
{
"epoch": 2.2234986945169712,
"grad_norm": 6.14243745803833,
"learning_rate": 2.6849512811259478e-06,
"loss": 0.6486,
"step": 2130
},
{
"epoch": 2.2339425587467363,
"grad_norm": 8.640827178955078,
"learning_rate": 2.6488632262721042e-06,
"loss": 0.5949,
"step": 2140
},
{
"epoch": 2.2443864229765014,
"grad_norm": 6.576625823974609,
"learning_rate": 2.6127751714182607e-06,
"loss": 0.6923,
"step": 2150
},
{
"epoch": 2.2548302872062664,
"grad_norm": 6.136504173278809,
"learning_rate": 2.5766871165644175e-06,
"loss": 0.6255,
"step": 2160
},
{
"epoch": 2.2652741514360315,
"grad_norm": 7.2910847663879395,
"learning_rate": 2.540599061710574e-06,
"loss": 0.6843,
"step": 2170
},
{
"epoch": 2.275718015665796,
"grad_norm": 6.936916351318359,
"learning_rate": 2.504511006856731e-06,
"loss": 0.6751,
"step": 2180
},
{
"epoch": 2.286161879895561,
"grad_norm": 5.899853229522705,
"learning_rate": 2.4684229520028873e-06,
"loss": 0.6584,
"step": 2190
},
{
"epoch": 2.2966057441253263,
"grad_norm": 4.891731262207031,
"learning_rate": 2.4323348971490438e-06,
"loss": 0.6373,
"step": 2200
},
{
"epoch": 2.2966057441253263,
"eval_loss": 0.958043098449707,
"eval_runtime": 23.1881,
"eval_samples_per_second": 36.7,
"eval_steps_per_second": 4.614,
"step": 2200
},
{
"epoch": 2.3070496083550913,
"grad_norm": 6.206886291503906,
"learning_rate": 2.3962468422952e-06,
"loss": 0.6725,
"step": 2210
},
{
"epoch": 2.3174934725848564,
"grad_norm": 4.663551330566406,
"learning_rate": 2.360158787441357e-06,
"loss": 0.5578,
"step": 2220
},
{
"epoch": 2.3279373368146214,
"grad_norm": 6.175649166107178,
"learning_rate": 2.3240707325875135e-06,
"loss": 0.6835,
"step": 2230
},
{
"epoch": 2.3383812010443865,
"grad_norm": 6.676774501800537,
"learning_rate": 2.2879826777336704e-06,
"loss": 0.7966,
"step": 2240
},
{
"epoch": 2.3488250652741516,
"grad_norm": 8.847614288330078,
"learning_rate": 2.251894622879827e-06,
"loss": 0.7679,
"step": 2250
},
{
"epoch": 2.3592689295039166,
"grad_norm": 6.491757869720459,
"learning_rate": 2.2158065680259837e-06,
"loss": 0.6274,
"step": 2260
},
{
"epoch": 2.3697127937336813,
"grad_norm": 6.540876388549805,
"learning_rate": 2.17971851317214e-06,
"loss": 0.674,
"step": 2270
},
{
"epoch": 2.3801566579634463,
"grad_norm": 7.067712306976318,
"learning_rate": 2.1436304583182966e-06,
"loss": 0.6716,
"step": 2280
},
{
"epoch": 2.3906005221932114,
"grad_norm": 4.959332466125488,
"learning_rate": 2.1075424034644535e-06,
"loss": 0.6729,
"step": 2290
},
{
"epoch": 2.4010443864229765,
"grad_norm": 4.016025066375732,
"learning_rate": 2.07145434861061e-06,
"loss": 0.6358,
"step": 2300
},
{
"epoch": 2.4114882506527415,
"grad_norm": 4.044537544250488,
"learning_rate": 2.035366293756767e-06,
"loss": 0.6867,
"step": 2310
},
{
"epoch": 2.4219321148825066,
"grad_norm": 4.88841438293457,
"learning_rate": 1.9992782389029233e-06,
"loss": 0.7561,
"step": 2320
},
{
"epoch": 2.4323759791122717,
"grad_norm": 7.33749532699585,
"learning_rate": 1.96319018404908e-06,
"loss": 0.6579,
"step": 2330
},
{
"epoch": 2.4428198433420367,
"grad_norm": 6.818521976470947,
"learning_rate": 1.9271021291952366e-06,
"loss": 0.7322,
"step": 2340
},
{
"epoch": 2.453263707571802,
"grad_norm": 5.549405097961426,
"learning_rate": 1.8910140743413932e-06,
"loss": 0.634,
"step": 2350
},
{
"epoch": 2.4637075718015664,
"grad_norm": 6.154874801635742,
"learning_rate": 1.85492601948755e-06,
"loss": 0.6044,
"step": 2360
},
{
"epoch": 2.4741514360313315,
"grad_norm": 5.5303521156311035,
"learning_rate": 1.8188379646337066e-06,
"loss": 0.6833,
"step": 2370
},
{
"epoch": 2.4845953002610965,
"grad_norm": 6.135169982910156,
"learning_rate": 1.7827499097798632e-06,
"loss": 0.7765,
"step": 2380
},
{
"epoch": 2.4950391644908616,
"grad_norm": 7.397289752960205,
"learning_rate": 1.7466618549260197e-06,
"loss": 0.6748,
"step": 2390
},
{
"epoch": 2.5054830287206267,
"grad_norm": 5.909689426422119,
"learning_rate": 1.7105738000721763e-06,
"loss": 0.6791,
"step": 2400
},
{
"epoch": 2.5054830287206267,
"eval_loss": 0.955007791519165,
"eval_runtime": 23.2476,
"eval_samples_per_second": 36.606,
"eval_steps_per_second": 4.603,
"step": 2400
},
{
"epoch": 2.5159268929503917,
"grad_norm": 6.320558547973633,
"learning_rate": 1.6744857452183328e-06,
"loss": 0.6219,
"step": 2410
},
{
"epoch": 2.526370757180157,
"grad_norm": 7.978168487548828,
"learning_rate": 1.6383976903644894e-06,
"loss": 0.8254,
"step": 2420
},
{
"epoch": 2.5368146214099214,
"grad_norm": 6.5808210372924805,
"learning_rate": 1.602309635510646e-06,
"loss": 0.7243,
"step": 2430
},
{
"epoch": 2.547258485639687,
"grad_norm": 4.769480228424072,
"learning_rate": 1.5662215806568025e-06,
"loss": 0.7497,
"step": 2440
},
{
"epoch": 2.5577023498694516,
"grad_norm": 5.738780975341797,
"learning_rate": 1.5301335258029592e-06,
"loss": 0.7048,
"step": 2450
},
{
"epoch": 2.5681462140992166,
"grad_norm": 5.658013343811035,
"learning_rate": 1.4940454709491159e-06,
"loss": 0.739,
"step": 2460
},
{
"epoch": 2.5785900783289817,
"grad_norm": 6.587325096130371,
"learning_rate": 1.4579574160952725e-06,
"loss": 0.7076,
"step": 2470
},
{
"epoch": 2.5890339425587467,
"grad_norm": 5.956645965576172,
"learning_rate": 1.4218693612414292e-06,
"loss": 0.6372,
"step": 2480
},
{
"epoch": 2.599477806788512,
"grad_norm": 5.966655731201172,
"learning_rate": 1.3857813063875859e-06,
"loss": 0.6934,
"step": 2490
},
{
"epoch": 2.609921671018277,
"grad_norm": 5.313653945922852,
"learning_rate": 1.3496932515337425e-06,
"loss": 0.7163,
"step": 2500
},
{
"epoch": 2.620365535248042,
"grad_norm": 6.935596466064453,
"learning_rate": 1.313605196679899e-06,
"loss": 0.7104,
"step": 2510
},
{
"epoch": 2.6308093994778066,
"grad_norm": 4.822442054748535,
"learning_rate": 1.2775171418260556e-06,
"loss": 0.6378,
"step": 2520
},
{
"epoch": 2.641253263707572,
"grad_norm": 5.288422107696533,
"learning_rate": 1.2414290869722123e-06,
"loss": 0.6463,
"step": 2530
},
{
"epoch": 2.6516971279373367,
"grad_norm": 6.668851852416992,
"learning_rate": 1.205341032118369e-06,
"loss": 0.7505,
"step": 2540
},
{
"epoch": 2.6621409921671018,
"grad_norm": 5.71054220199585,
"learning_rate": 1.1692529772645256e-06,
"loss": 0.5856,
"step": 2550
},
{
"epoch": 2.672584856396867,
"grad_norm": 6.284550666809082,
"learning_rate": 1.1331649224106823e-06,
"loss": 0.8122,
"step": 2560
},
{
"epoch": 2.683028720626632,
"grad_norm": 8.781463623046875,
"learning_rate": 1.097076867556839e-06,
"loss": 0.7063,
"step": 2570
},
{
"epoch": 2.693472584856397,
"grad_norm": 7.29454231262207,
"learning_rate": 1.0609888127029954e-06,
"loss": 0.7429,
"step": 2580
},
{
"epoch": 2.703916449086162,
"grad_norm": 5.689371109008789,
"learning_rate": 1.024900757849152e-06,
"loss": 0.6658,
"step": 2590
},
{
"epoch": 2.714360313315927,
"grad_norm": 7.286506175994873,
"learning_rate": 9.888127029953087e-07,
"loss": 0.748,
"step": 2600
},
{
"epoch": 2.714360313315927,
"eval_loss": 0.9431054592132568,
"eval_runtime": 23.1747,
"eval_samples_per_second": 36.721,
"eval_steps_per_second": 4.617,
"step": 2600
},
{
"epoch": 2.7248041775456917,
"grad_norm": 5.635782241821289,
"learning_rate": 9.527246481414652e-07,
"loss": 0.673,
"step": 2610
},
{
"epoch": 2.7352480417754568,
"grad_norm": 5.282413959503174,
"learning_rate": 9.166365932876219e-07,
"loss": 0.9037,
"step": 2620
},
{
"epoch": 2.745691906005222,
"grad_norm": 7.922749042510986,
"learning_rate": 8.805485384337785e-07,
"loss": 0.6862,
"step": 2630
},
{
"epoch": 2.756135770234987,
"grad_norm": 5.463962078094482,
"learning_rate": 8.444604835799351e-07,
"loss": 0.6,
"step": 2640
},
{
"epoch": 2.766579634464752,
"grad_norm": 8.0007963180542,
"learning_rate": 8.083724287260918e-07,
"loss": 0.6688,
"step": 2650
},
{
"epoch": 2.777023498694517,
"grad_norm": 7.617900371551514,
"learning_rate": 7.722843738722483e-07,
"loss": 0.7277,
"step": 2660
},
{
"epoch": 2.787467362924282,
"grad_norm": 5.969784259796143,
"learning_rate": 7.36196319018405e-07,
"loss": 0.7085,
"step": 2670
},
{
"epoch": 2.7979112271540467,
"grad_norm": 5.169407367706299,
"learning_rate": 7.001082641645617e-07,
"loss": 0.7344,
"step": 2680
},
{
"epoch": 2.8083550913838122,
"grad_norm": 8.009687423706055,
"learning_rate": 6.640202093107181e-07,
"loss": 0.6457,
"step": 2690
},
{
"epoch": 2.818798955613577,
"grad_norm": 7.4137187004089355,
"learning_rate": 6.279321544568748e-07,
"loss": 0.6416,
"step": 2700
},
{
"epoch": 2.829242819843342,
"grad_norm": 5.35453462600708,
"learning_rate": 5.918440996030314e-07,
"loss": 0.7867,
"step": 2710
},
{
"epoch": 2.839686684073107,
"grad_norm": 7.469908237457275,
"learning_rate": 5.557560447491881e-07,
"loss": 0.7955,
"step": 2720
},
{
"epoch": 2.850130548302872,
"grad_norm": 6.326605319976807,
"learning_rate": 5.196679898953446e-07,
"loss": 0.6995,
"step": 2730
},
{
"epoch": 2.860574412532637,
"grad_norm": 7.096553802490234,
"learning_rate": 4.835799350415013e-07,
"loss": 0.6625,
"step": 2740
},
{
"epoch": 2.871018276762402,
"grad_norm": 5.128674507141113,
"learning_rate": 4.474918801876579e-07,
"loss": 0.6791,
"step": 2750
},
{
"epoch": 2.8814621409921672,
"grad_norm": 6.457350254058838,
"learning_rate": 4.1140382533381457e-07,
"loss": 0.7586,
"step": 2760
},
{
"epoch": 2.891906005221932,
"grad_norm": 6.231655597686768,
"learning_rate": 3.753157704799711e-07,
"loss": 0.677,
"step": 2770
},
{
"epoch": 2.9023498694516974,
"grad_norm": 6.412544250488281,
"learning_rate": 3.392277156261278e-07,
"loss": 0.7449,
"step": 2780
},
{
"epoch": 2.912793733681462,
"grad_norm": 3.725374698638916,
"learning_rate": 3.031396607722844e-07,
"loss": 0.5804,
"step": 2790
},
{
"epoch": 2.923237597911227,
"grad_norm": 5.284286975860596,
"learning_rate": 2.67051605918441e-07,
"loss": 0.571,
"step": 2800
},
{
"epoch": 2.923237597911227,
"eval_loss": 0.94057297706604,
"eval_runtime": 23.1841,
"eval_samples_per_second": 36.706,
"eval_steps_per_second": 4.615,
"step": 2800
}
],
"logging_steps": 10,
"max_steps": 2871,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 200,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3.029626779598848e+16,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}