lora32b / trainer_state.json
yis77's picture
Upload folder using huggingface_hub
0f3163f verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.18467220683287167,
"eval_steps": 500,
"global_step": 300,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0030778701138811943,
"grad_norm": 0.47482776641845703,
"learning_rate": 4.9999916942231176e-05,
"loss": 1.1992,
"num_input_tokens_seen": 42384,
"step": 5,
"train_runtime": 19.849,
"train_tokens_per_second": 2135.325
},
{
"epoch": 0.006155740227762389,
"grad_norm": 0.24747319519519806,
"learning_rate": 4.999957952099119e-05,
"loss": 1.0247,
"num_input_tokens_seen": 86000,
"step": 10,
"train_runtime": 39.0811,
"train_tokens_per_second": 2200.554
},
{
"epoch": 0.009233610341643583,
"grad_norm": 0.10441822558641434,
"learning_rate": 4.9998982548669973e-05,
"loss": 0.8631,
"num_input_tokens_seen": 124784,
"step": 15,
"train_runtime": 57.1682,
"train_tokens_per_second": 2182.751
},
{
"epoch": 0.012311480455524777,
"grad_norm": 0.11058320850133896,
"learning_rate": 4.999812603146542e-05,
"loss": 0.9173,
"num_input_tokens_seen": 165952,
"step": 20,
"train_runtime": 75.5733,
"train_tokens_per_second": 2195.909
},
{
"epoch": 0.015389350569405972,
"grad_norm": 0.09879708290100098,
"learning_rate": 4.9997009978270085e-05,
"loss": 0.8722,
"num_input_tokens_seen": 201584,
"step": 25,
"train_runtime": 92.5588,
"train_tokens_per_second": 2177.903
},
{
"epoch": 0.018467220683287166,
"grad_norm": 0.11391397565603256,
"learning_rate": 4.999563440067108e-05,
"loss": 0.866,
"num_input_tokens_seen": 240656,
"step": 30,
"train_runtime": 111.1373,
"train_tokens_per_second": 2165.393
},
{
"epoch": 0.02154509079716836,
"grad_norm": 0.09722574800252914,
"learning_rate": 4.9993999312949936e-05,
"loss": 0.853,
"num_input_tokens_seen": 287248,
"step": 35,
"train_runtime": 130.6684,
"train_tokens_per_second": 2198.297
},
{
"epoch": 0.024622960911049555,
"grad_norm": 0.12133017182350159,
"learning_rate": 4.99921047320825e-05,
"loss": 0.88,
"num_input_tokens_seen": 328448,
"step": 40,
"train_runtime": 149.193,
"train_tokens_per_second": 2201.498
},
{
"epoch": 0.027700831024930747,
"grad_norm": 0.12174953520298004,
"learning_rate": 4.9989950677738726e-05,
"loss": 0.8503,
"num_input_tokens_seen": 370512,
"step": 45,
"train_runtime": 168.2236,
"train_tokens_per_second": 2202.497
},
{
"epoch": 0.030778701138811943,
"grad_norm": 0.12440948933362961,
"learning_rate": 4.9987537172282466e-05,
"loss": 0.8707,
"num_input_tokens_seen": 409936,
"step": 50,
"train_runtime": 186.7223,
"train_tokens_per_second": 2195.432
},
{
"epoch": 0.03385657125269314,
"grad_norm": 0.12797962129116058,
"learning_rate": 4.998486424077127e-05,
"loss": 0.8475,
"num_input_tokens_seen": 450928,
"step": 55,
"train_runtime": 205.1107,
"train_tokens_per_second": 2198.462
},
{
"epoch": 0.03693444136657433,
"grad_norm": 0.09793120622634888,
"learning_rate": 4.9981931910956086e-05,
"loss": 0.8409,
"num_input_tokens_seen": 491152,
"step": 60,
"train_runtime": 223.3671,
"train_tokens_per_second": 2198.856
},
{
"epoch": 0.040012311480455524,
"grad_norm": 0.10957301408052444,
"learning_rate": 4.9978740213281005e-05,
"loss": 0.8524,
"num_input_tokens_seen": 532160,
"step": 65,
"train_runtime": 241.5682,
"train_tokens_per_second": 2202.939
},
{
"epoch": 0.04309018159433672,
"grad_norm": 0.12283050268888474,
"learning_rate": 4.9975289180882936e-05,
"loss": 0.8944,
"num_input_tokens_seen": 573392,
"step": 70,
"train_runtime": 259.8567,
"train_tokens_per_second": 2206.57
},
{
"epoch": 0.046168051708217916,
"grad_norm": 0.1338459849357605,
"learning_rate": 4.9971578849591225e-05,
"loss": 0.8478,
"num_input_tokens_seen": 613488,
"step": 75,
"train_runtime": 278.0456,
"train_tokens_per_second": 2206.429
},
{
"epoch": 0.04924592182209911,
"grad_norm": 0.10969562828540802,
"learning_rate": 4.996760925792737e-05,
"loss": 0.8331,
"num_input_tokens_seen": 656480,
"step": 80,
"train_runtime": 296.3508,
"train_tokens_per_second": 2215.212
},
{
"epoch": 0.0523237919359803,
"grad_norm": 0.1411936730146408,
"learning_rate": 4.996338044710452e-05,
"loss": 0.8306,
"num_input_tokens_seen": 698272,
"step": 85,
"train_runtime": 315.096,
"train_tokens_per_second": 2216.061
},
{
"epoch": 0.055401662049861494,
"grad_norm": 0.1435951143503189,
"learning_rate": 4.995889246102711e-05,
"loss": 0.8318,
"num_input_tokens_seen": 735376,
"step": 90,
"train_runtime": 332.4774,
"train_tokens_per_second": 2211.807
},
{
"epoch": 0.05847953216374269,
"grad_norm": 0.15133805572986603,
"learning_rate": 4.995414534629039e-05,
"loss": 0.8674,
"num_input_tokens_seen": 772544,
"step": 95,
"train_runtime": 350.9252,
"train_tokens_per_second": 2201.449
},
{
"epoch": 0.061557402277623886,
"grad_norm": 0.10615513473749161,
"learning_rate": 4.9949139152179944e-05,
"loss": 0.901,
"num_input_tokens_seen": 814928,
"step": 100,
"train_runtime": 369.5452,
"train_tokens_per_second": 2205.219
},
{
"epoch": 0.06463527239150507,
"grad_norm": 0.10441838204860687,
"learning_rate": 4.994387393067117e-05,
"loss": 0.8522,
"num_input_tokens_seen": 855104,
"step": 105,
"train_runtime": 388.7482,
"train_tokens_per_second": 2199.635
},
{
"epoch": 0.06771314250538628,
"grad_norm": 0.1405644714832306,
"learning_rate": 4.993834973642874e-05,
"loss": 0.8614,
"num_input_tokens_seen": 896272,
"step": 110,
"train_runtime": 406.8129,
"train_tokens_per_second": 2203.155
},
{
"epoch": 0.07079101261926747,
"grad_norm": 0.11736899614334106,
"learning_rate": 4.993256662680604e-05,
"loss": 0.8371,
"num_input_tokens_seen": 940864,
"step": 115,
"train_runtime": 425.836,
"train_tokens_per_second": 2209.452
},
{
"epoch": 0.07386888273314866,
"grad_norm": 0.1269819736480713,
"learning_rate": 4.992652466184458e-05,
"loss": 0.9,
"num_input_tokens_seen": 983360,
"step": 120,
"train_runtime": 444.5802,
"train_tokens_per_second": 2211.884
},
{
"epoch": 0.07694675284702986,
"grad_norm": 0.10944321751594543,
"learning_rate": 4.992022390427335e-05,
"loss": 0.8629,
"num_input_tokens_seen": 1025120,
"step": 125,
"train_runtime": 463.3805,
"train_tokens_per_second": 2212.264
},
{
"epoch": 0.08002462296091105,
"grad_norm": 0.1438760608434677,
"learning_rate": 4.99136644195082e-05,
"loss": 0.8797,
"num_input_tokens_seen": 1065360,
"step": 130,
"train_runtime": 480.9152,
"train_tokens_per_second": 2215.276
},
{
"epoch": 0.08310249307479224,
"grad_norm": 0.13859188556671143,
"learning_rate": 4.9906846275651125e-05,
"loss": 0.8317,
"num_input_tokens_seen": 1106112,
"step": 135,
"train_runtime": 498.8557,
"train_tokens_per_second": 2217.299
},
{
"epoch": 0.08618036318867343,
"grad_norm": 0.1412304937839508,
"learning_rate": 4.989976954348958e-05,
"loss": 0.8422,
"num_input_tokens_seen": 1149632,
"step": 140,
"train_runtime": 517.2481,
"train_tokens_per_second": 2222.593
},
{
"epoch": 0.08925823330255463,
"grad_norm": 0.1513959765434265,
"learning_rate": 4.989243429649573e-05,
"loss": 0.8403,
"num_input_tokens_seen": 1187824,
"step": 145,
"train_runtime": 535.2196,
"train_tokens_per_second": 2219.321
},
{
"epoch": 0.09233610341643583,
"grad_norm": 0.13685107231140137,
"learning_rate": 4.988484061082572e-05,
"loss": 0.8961,
"num_input_tokens_seen": 1229088,
"step": 150,
"train_runtime": 553.4777,
"train_tokens_per_second": 2220.664
},
{
"epoch": 0.09541397353031703,
"grad_norm": 0.12999333441257477,
"learning_rate": 4.987698856531884e-05,
"loss": 0.814,
"num_input_tokens_seen": 1273056,
"step": 155,
"train_runtime": 572.9621,
"train_tokens_per_second": 2221.885
},
{
"epoch": 0.09849184364419822,
"grad_norm": 0.1481194943189621,
"learning_rate": 4.986887824149674e-05,
"loss": 0.8542,
"num_input_tokens_seen": 1310288,
"step": 160,
"train_runtime": 589.8405,
"train_tokens_per_second": 2221.427
},
{
"epoch": 0.10156971375807941,
"grad_norm": 0.1396579146385193,
"learning_rate": 4.9860509723562573e-05,
"loss": 0.8722,
"num_input_tokens_seen": 1351312,
"step": 165,
"train_runtime": 607.8908,
"train_tokens_per_second": 2222.952
},
{
"epoch": 0.1046475838719606,
"grad_norm": 0.12688206136226654,
"learning_rate": 4.985188309840012e-05,
"loss": 0.8587,
"num_input_tokens_seen": 1396336,
"step": 170,
"train_runtime": 626.8902,
"train_tokens_per_second": 2227.401
},
{
"epoch": 0.1077254539858418,
"grad_norm": 0.13825534284114838,
"learning_rate": 4.984299845557287e-05,
"loss": 0.8014,
"num_input_tokens_seen": 1437984,
"step": 175,
"train_runtime": 645.2717,
"train_tokens_per_second": 2228.494
},
{
"epoch": 0.11080332409972299,
"grad_norm": 0.12385495752096176,
"learning_rate": 4.983385588732312e-05,
"loss": 0.8385,
"num_input_tokens_seen": 1475360,
"step": 180,
"train_runtime": 662.6866,
"train_tokens_per_second": 2226.332
},
{
"epoch": 0.11388119421360418,
"grad_norm": 0.12246419489383698,
"learning_rate": 4.982445548857102e-05,
"loss": 0.8372,
"num_input_tokens_seen": 1513152,
"step": 185,
"train_runtime": 680.3316,
"train_tokens_per_second": 2224.139
},
{
"epoch": 0.11695906432748537,
"grad_norm": 0.12281708419322968,
"learning_rate": 4.981479735691354e-05,
"loss": 0.8446,
"num_input_tokens_seen": 1556256,
"step": 190,
"train_runtime": 698.7652,
"train_tokens_per_second": 2227.152
},
{
"epoch": 0.12003693444136658,
"grad_norm": 0.13810615241527557,
"learning_rate": 4.980488159262353e-05,
"loss": 0.8141,
"num_input_tokens_seen": 1598832,
"step": 195,
"train_runtime": 717.4389,
"train_tokens_per_second": 2228.527
},
{
"epoch": 0.12311480455524777,
"grad_norm": 0.15435761213302612,
"learning_rate": 4.97947082986486e-05,
"loss": 0.8398,
"num_input_tokens_seen": 1641712,
"step": 200,
"train_runtime": 736.4476,
"train_tokens_per_second": 2229.231
},
{
"epoch": 0.12619267466912895,
"grad_norm": 0.15266934037208557,
"learning_rate": 4.978427758061009e-05,
"loss": 0.8714,
"num_input_tokens_seen": 1684544,
"step": 205,
"train_runtime": 756.7894,
"train_tokens_per_second": 2225.908
},
{
"epoch": 0.12927054478301014,
"grad_norm": 0.13250574469566345,
"learning_rate": 4.977358954680202e-05,
"loss": 0.8116,
"num_input_tokens_seen": 1723120,
"step": 210,
"train_runtime": 774.2541,
"train_tokens_per_second": 2225.523
},
{
"epoch": 0.13234841489689136,
"grad_norm": 0.11579591035842896,
"learning_rate": 4.9762644308189865e-05,
"loss": 0.8204,
"num_input_tokens_seen": 1768096,
"step": 215,
"train_runtime": 793.0535,
"train_tokens_per_second": 2229.479
},
{
"epoch": 0.13542628501077256,
"grad_norm": 0.12654677033424377,
"learning_rate": 4.9751441978409484e-05,
"loss": 0.8334,
"num_input_tokens_seen": 1808176,
"step": 220,
"train_runtime": 810.905,
"train_tokens_per_second": 2229.825
},
{
"epoch": 0.13850415512465375,
"grad_norm": 0.15607228875160217,
"learning_rate": 4.97399826737659e-05,
"loss": 0.8177,
"num_input_tokens_seen": 1848480,
"step": 225,
"train_runtime": 829.3126,
"train_tokens_per_second": 2228.93
},
{
"epoch": 0.14158202523853494,
"grad_norm": 0.1535269021987915,
"learning_rate": 4.972826651323211e-05,
"loss": 0.807,
"num_input_tokens_seen": 1889168,
"step": 230,
"train_runtime": 847.0449,
"train_tokens_per_second": 2230.304
},
{
"epoch": 0.14465989535241613,
"grad_norm": 0.13663041591644287,
"learning_rate": 4.971629361844785e-05,
"loss": 0.8557,
"num_input_tokens_seen": 1930784,
"step": 235,
"train_runtime": 865.4457,
"train_tokens_per_second": 2230.971
},
{
"epoch": 0.14773776546629733,
"grad_norm": 0.1719619184732437,
"learning_rate": 4.9704064113718326e-05,
"loss": 0.8436,
"num_input_tokens_seen": 1974176,
"step": 240,
"train_runtime": 883.9962,
"train_tokens_per_second": 2233.24
},
{
"epoch": 0.15081563558017852,
"grad_norm": 0.15830345451831818,
"learning_rate": 4.9691578126012905e-05,
"loss": 0.853,
"num_input_tokens_seen": 2013008,
"step": 245,
"train_runtime": 902.1681,
"train_tokens_per_second": 2231.3
},
{
"epoch": 0.1538935056940597,
"grad_norm": 0.14472956955432892,
"learning_rate": 4.967883578496385e-05,
"loss": 0.874,
"num_input_tokens_seen": 2053328,
"step": 250,
"train_runtime": 919.7822,
"train_tokens_per_second": 2232.407
},
{
"epoch": 0.1569713758079409,
"grad_norm": 0.13329900801181793,
"learning_rate": 4.966583722286491e-05,
"loss": 0.8333,
"num_input_tokens_seen": 2094992,
"step": 255,
"train_runtime": 937.9537,
"train_tokens_per_second": 2233.577
},
{
"epoch": 0.1600492459218221,
"grad_norm": 0.1551639437675476,
"learning_rate": 4.9652582574669995e-05,
"loss": 0.864,
"num_input_tokens_seen": 2135376,
"step": 260,
"train_runtime": 955.8955,
"train_tokens_per_second": 2233.901
},
{
"epoch": 0.1631271160357033,
"grad_norm": 0.12460612505674362,
"learning_rate": 4.963907197799174e-05,
"loss": 0.8321,
"num_input_tokens_seen": 2176624,
"step": 265,
"train_runtime": 974.3283,
"train_tokens_per_second": 2233.974
},
{
"epoch": 0.16620498614958448,
"grad_norm": 0.13981439173221588,
"learning_rate": 4.9625305573100115e-05,
"loss": 0.8349,
"num_input_tokens_seen": 2220496,
"step": 270,
"train_runtime": 992.5551,
"train_tokens_per_second": 2237.151
},
{
"epoch": 0.16928285626346568,
"grad_norm": 0.15424658358097076,
"learning_rate": 4.961128350292094e-05,
"loss": 0.8248,
"num_input_tokens_seen": 2263808,
"step": 275,
"train_runtime": 1010.9218,
"train_tokens_per_second": 2239.35
},
{
"epoch": 0.17236072637734687,
"grad_norm": 0.14288848638534546,
"learning_rate": 4.959700591303438e-05,
"loss": 0.8488,
"num_input_tokens_seen": 2311696,
"step": 280,
"train_runtime": 1030.8143,
"train_tokens_per_second": 2242.592
},
{
"epoch": 0.17543859649122806,
"grad_norm": 0.15639737248420715,
"learning_rate": 4.95824729516735e-05,
"loss": 0.8602,
"num_input_tokens_seen": 2353200,
"step": 285,
"train_runtime": 1049.2401,
"train_tokens_per_second": 2242.766
},
{
"epoch": 0.17851646660510925,
"grad_norm": 0.1268603354692459,
"learning_rate": 4.9567684769722664e-05,
"loss": 0.8987,
"num_input_tokens_seen": 2396000,
"step": 290,
"train_runtime": 1067.7324,
"train_tokens_per_second": 2244.008
},
{
"epoch": 0.18159433671899045,
"grad_norm": 0.14782211184501648,
"learning_rate": 4.9552641520716014e-05,
"loss": 0.7794,
"num_input_tokens_seen": 2438032,
"step": 295,
"train_runtime": 1085.9257,
"train_tokens_per_second": 2245.119
},
{
"epoch": 0.18467220683287167,
"grad_norm": 0.15259359776973724,
"learning_rate": 4.953734336083583e-05,
"loss": 0.8648,
"num_input_tokens_seen": 2478048,
"step": 300,
"train_runtime": 1103.5538,
"train_tokens_per_second": 2245.516
}
],
"logging_steps": 5,
"max_steps": 4875,
"num_input_tokens_seen": 2478048,
"num_train_epochs": 3,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 4.765645728959693e+17,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}