marocknewAI1 / trainer_state.json
iko-01's picture
رفع النموذج النهائي مع جميع الملفات للتجربة أو إعادة التدريب
e1e9081 verified
{
"best_global_step": 6000,
"best_metric": 1.4575997591018677,
"best_model_checkpoint": "/content/marocAI-finetuned/checkpoint-2821/checkpoint-6000",
"epoch": 7.0,
"eval_steps": 500,
"global_step": 6062,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.4618937644341801,
"grad_norm": 1.0584869384765625,
"learning_rate": 1.695852534562212e-05,
"loss": 2.9443,
"step": 100
},
{
"epoch": 0.9237875288683602,
"grad_norm": 1.051153302192688,
"learning_rate": 1.3886328725038403e-05,
"loss": 2.3079,
"step": 200
},
{
"epoch": 1.3833718244803694,
"grad_norm": 0.9572842121124268,
"learning_rate": 1.0814132104454686e-05,
"loss": 2.2004,
"step": 300
},
{
"epoch": 1.8452655889145495,
"grad_norm": 0.7911381125450134,
"learning_rate": 7.741935483870968e-06,
"loss": 2.0987,
"step": 400
},
{
"epoch": 2.304849884526559,
"grad_norm": 1.0212643146514893,
"learning_rate": 4.669738863287251e-06,
"loss": 2.0724,
"step": 500
},
{
"epoch": 2.304849884526559,
"eval_loss": 1.7146347761154175,
"eval_runtime": 3.6485,
"eval_samples_per_second": 50.157,
"eval_steps_per_second": 12.608,
"step": 500
},
{
"epoch": 2.766743648960739,
"grad_norm": 1.0736603736877441,
"learning_rate": 1.5975422427035332e-06,
"loss": 2.0376,
"step": 600
},
{
"epoch": 3.2263279445727484,
"grad_norm": 0.960295557975769,
"learning_rate": 9.262672811059909e-06,
"loss": 2.0108,
"step": 700
},
{
"epoch": 3.6882217090069283,
"grad_norm": 0.8508874773979187,
"learning_rate": 7.726574500768049e-06,
"loss": 2.0039,
"step": 800
},
{
"epoch": 4.147806004618937,
"grad_norm": 0.9244301319122314,
"learning_rate": 6.1904761904761914e-06,
"loss": 2.0071,
"step": 900
},
{
"epoch": 4.609699769053118,
"grad_norm": 0.8824007511138916,
"learning_rate": 4.654377880184332e-06,
"loss": 1.9621,
"step": 1000
},
{
"epoch": 4.609699769053118,
"eval_loss": 1.6397035121917725,
"eval_runtime": 3.6176,
"eval_samples_per_second": 50.586,
"eval_steps_per_second": 12.716,
"step": 1000
},
{
"epoch": 5.069284064665127,
"grad_norm": 0.8873424530029297,
"learning_rate": 3.1182795698924735e-06,
"loss": 1.9262,
"step": 1100
},
{
"epoch": 5.531177829099307,
"grad_norm": 0.8406194448471069,
"learning_rate": 1.5821812596006145e-06,
"loss": 1.9235,
"step": 1200
},
{
"epoch": 5.993071593533488,
"grad_norm": 0.8757996559143066,
"learning_rate": 4.608294930875576e-08,
"loss": 1.9402,
"step": 1300
},
{
"epoch": 6.452655889145497,
"grad_norm": 0.8397168517112732,
"learning_rate": 1.0081531371853954e-05,
"loss": 1.9086,
"step": 1400
},
{
"epoch": 6.914549653579677,
"grad_norm": 0.8500417470932007,
"learning_rate": 9.372562920950018e-06,
"loss": 1.9123,
"step": 1500
},
{
"epoch": 6.914549653579677,
"eval_loss": 1.5952860116958618,
"eval_runtime": 3.6287,
"eval_samples_per_second": 50.432,
"eval_steps_per_second": 12.677,
"step": 1500
},
{
"epoch": 7.374133949191686,
"grad_norm": 0.8373268246650696,
"learning_rate": 8.663594470046084e-06,
"loss": 1.8666,
"step": 1600
},
{
"epoch": 7.836027713625866,
"grad_norm": 0.8652852773666382,
"learning_rate": 7.954626019142148e-06,
"loss": 1.8377,
"step": 1700
},
{
"epoch": 8.295612009237875,
"grad_norm": 0.8815492391586304,
"learning_rate": 7.245657568238214e-06,
"loss": 1.8395,
"step": 1800
},
{
"epoch": 8.757505773672055,
"grad_norm": 0.8464174270629883,
"learning_rate": 6.53668911733428e-06,
"loss": 1.8285,
"step": 1900
},
{
"epoch": 9.217090069284065,
"grad_norm": 0.818134069442749,
"learning_rate": 5.827720666430344e-06,
"loss": 1.8244,
"step": 2000
},
{
"epoch": 9.217090069284065,
"eval_loss": 1.5474414825439453,
"eval_runtime": 3.7454,
"eval_samples_per_second": 48.86,
"eval_steps_per_second": 12.282,
"step": 2000
},
{
"epoch": 9.678983833718245,
"grad_norm": 0.8570533394813538,
"learning_rate": 5.118752215526409e-06,
"loss": 1.8044,
"step": 2100
},
{
"epoch": 10.138568129330254,
"grad_norm": 0.8370910286903381,
"learning_rate": 4.409783764622475e-06,
"loss": 1.7769,
"step": 2200
},
{
"epoch": 10.600461893764434,
"grad_norm": 0.8950196504592896,
"learning_rate": 3.70081531371854e-06,
"loss": 1.7914,
"step": 2300
},
{
"epoch": 11.060046189376443,
"grad_norm": 0.7839242815971375,
"learning_rate": 2.9918468628146054e-06,
"loss": 1.7767,
"step": 2400
},
{
"epoch": 11.521939953810623,
"grad_norm": 0.8934968709945679,
"learning_rate": 2.28287841191067e-06,
"loss": 1.7807,
"step": 2500
},
{
"epoch": 11.521939953810623,
"eval_loss": 1.5255564451217651,
"eval_runtime": 3.6222,
"eval_samples_per_second": 50.522,
"eval_steps_per_second": 12.7,
"step": 2500
},
{
"epoch": 11.983833718244803,
"grad_norm": 0.9384580254554749,
"learning_rate": 1.5739099610067355e-06,
"loss": 1.7571,
"step": 2600
},
{
"epoch": 12.443418013856814,
"grad_norm": 0.838097333908081,
"learning_rate": 8.649415101028006e-07,
"loss": 1.7684,
"step": 2700
},
{
"epoch": 12.905311778290994,
"grad_norm": 0.807694673538208,
"learning_rate": 1.5597305919886567e-07,
"loss": 1.7643,
"step": 2800
},
{
"epoch": 3.348729792147806,
"grad_norm": 1.560190200805664,
"learning_rate": 1.0610533378061055e-05,
"loss": 1.7522,
"step": 2900
},
{
"epoch": 3.464203233256351,
"grad_norm": 1.5953682661056519,
"learning_rate": 1.0275075478027507e-05,
"loss": 1.7661,
"step": 3000
},
{
"epoch": 3.464203233256351,
"eval_loss": 1.5233831405639648,
"eval_runtime": 3.6827,
"eval_samples_per_second": 49.692,
"eval_steps_per_second": 12.491,
"step": 3000
},
{
"epoch": 3.579676674364896,
"grad_norm": 1.5370293855667114,
"learning_rate": 9.939617577993964e-06,
"loss": 1.7759,
"step": 3100
},
{
"epoch": 3.695150115473441,
"grad_norm": 1.470035195350647,
"learning_rate": 9.604159677960416e-06,
"loss": 1.7367,
"step": 3200
},
{
"epoch": 3.8106235565819864,
"grad_norm": 1.5553827285766602,
"learning_rate": 9.26870177792687e-06,
"loss": 1.7614,
"step": 3300
},
{
"epoch": 3.9260969976905313,
"grad_norm": 1.7353712320327759,
"learning_rate": 8.933243877893324e-06,
"loss": 1.7301,
"step": 3400
},
{
"epoch": 4.041570438799076,
"grad_norm": 1.8023557662963867,
"learning_rate": 8.59778597785978e-06,
"loss": 1.7797,
"step": 3500
},
{
"epoch": 4.041570438799076,
"eval_loss": 1.5006413459777832,
"eval_runtime": 3.6376,
"eval_samples_per_second": 50.308,
"eval_steps_per_second": 12.646,
"step": 3500
},
{
"epoch": 4.157043879907621,
"grad_norm": 1.4707646369934082,
"learning_rate": 8.262328077826235e-06,
"loss": 1.7436,
"step": 3600
},
{
"epoch": 4.272517321016166,
"grad_norm": 1.8790034055709839,
"learning_rate": 7.926870177792688e-06,
"loss": 1.7703,
"step": 3700
},
{
"epoch": 4.387990762124711,
"grad_norm": 1.7127020359039307,
"learning_rate": 7.591412277759142e-06,
"loss": 1.7297,
"step": 3800
},
{
"epoch": 4.503464203233256,
"grad_norm": 1.575723648071289,
"learning_rate": 7.255954377725596e-06,
"loss": 1.6714,
"step": 3900
},
{
"epoch": 4.618937644341801,
"grad_norm": 1.6164535284042358,
"learning_rate": 6.92049647769205e-06,
"loss": 1.6929,
"step": 4000
},
{
"epoch": 4.618937644341801,
"eval_loss": 1.4842592477798462,
"eval_runtime": 3.769,
"eval_samples_per_second": 48.554,
"eval_steps_per_second": 12.205,
"step": 4000
},
{
"epoch": 4.734411085450346,
"grad_norm": 1.4100682735443115,
"learning_rate": 6.585038577658505e-06,
"loss": 1.7042,
"step": 4100
},
{
"epoch": 4.849884526558892,
"grad_norm": 1.533423900604248,
"learning_rate": 6.249580677624959e-06,
"loss": 1.6724,
"step": 4200
},
{
"epoch": 4.965357967667437,
"grad_norm": 1.7051324844360352,
"learning_rate": 5.9141227775914126e-06,
"loss": 1.6685,
"step": 4300
},
{
"epoch": 5.080831408775982,
"grad_norm": 1.5413386821746826,
"learning_rate": 5.578664877557867e-06,
"loss": 1.6784,
"step": 4400
},
{
"epoch": 5.196304849884527,
"grad_norm": 1.7100459337234497,
"learning_rate": 5.243206977524321e-06,
"loss": 1.6237,
"step": 4500
},
{
"epoch": 5.196304849884527,
"eval_loss": 1.4744157791137695,
"eval_runtime": 3.6566,
"eval_samples_per_second": 50.047,
"eval_steps_per_second": 12.58,
"step": 4500
},
{
"epoch": 5.311778290993072,
"grad_norm": 1.8523712158203125,
"learning_rate": 4.907749077490776e-06,
"loss": 1.7114,
"step": 4600
},
{
"epoch": 5.427251732101617,
"grad_norm": 1.578623652458191,
"learning_rate": 4.572291177457229e-06,
"loss": 1.6396,
"step": 4700
},
{
"epoch": 5.542725173210162,
"grad_norm": 1.6069693565368652,
"learning_rate": 4.2368332774236835e-06,
"loss": 1.6942,
"step": 4800
},
{
"epoch": 5.658198614318707,
"grad_norm": 1.2671387195587158,
"learning_rate": 3.901375377390138e-06,
"loss": 1.6999,
"step": 4900
},
{
"epoch": 5.773672055427252,
"grad_norm": 1.5692400932312012,
"learning_rate": 3.5659174773565918e-06,
"loss": 1.6395,
"step": 5000
},
{
"epoch": 5.773672055427252,
"eval_loss": 1.464585542678833,
"eval_runtime": 3.6703,
"eval_samples_per_second": 49.859,
"eval_steps_per_second": 12.533,
"step": 5000
},
{
"epoch": 5.8891454965357966,
"grad_norm": 1.6615720987319946,
"learning_rate": 3.230459577323046e-06,
"loss": 1.6839,
"step": 5100
},
{
"epoch": 6.0046189376443415,
"grad_norm": 1.6811989545822144,
"learning_rate": 2.8950016772895005e-06,
"loss": 1.7058,
"step": 5200
},
{
"epoch": 6.1200923787528865,
"grad_norm": 1.569676399230957,
"learning_rate": 2.5595437772559544e-06,
"loss": 1.6426,
"step": 5300
},
{
"epoch": 6.235565819861431,
"grad_norm": 1.753227949142456,
"learning_rate": 2.2240858772224088e-06,
"loss": 1.646,
"step": 5400
},
{
"epoch": 6.351039260969977,
"grad_norm": 1.3525090217590332,
"learning_rate": 1.888627977188863e-06,
"loss": 1.6184,
"step": 5500
},
{
"epoch": 6.351039260969977,
"eval_loss": 1.459729790687561,
"eval_runtime": 3.6443,
"eval_samples_per_second": 50.216,
"eval_steps_per_second": 12.623,
"step": 5500
},
{
"epoch": 6.466512702078522,
"grad_norm": 1.6190038919448853,
"learning_rate": 1.553170077155317e-06,
"loss": 1.6743,
"step": 5600
},
{
"epoch": 6.581986143187067,
"grad_norm": 1.5956363677978516,
"learning_rate": 1.2177121771217714e-06,
"loss": 1.6474,
"step": 5700
},
{
"epoch": 6.697459584295612,
"grad_norm": 1.6586838960647583,
"learning_rate": 8.822542770882254e-07,
"loss": 1.7052,
"step": 5800
},
{
"epoch": 6.812933025404157,
"grad_norm": 1.3021780252456665,
"learning_rate": 5.467963770546797e-07,
"loss": 1.6333,
"step": 5900
},
{
"epoch": 6.928406466512702,
"grad_norm": 1.7689718008041382,
"learning_rate": 2.1133847702113386e-07,
"loss": 1.6687,
"step": 6000
},
{
"epoch": 6.928406466512702,
"eval_loss": 1.4575997591018677,
"eval_runtime": 3.7418,
"eval_samples_per_second": 48.907,
"eval_steps_per_second": 12.294,
"step": 6000
}
],
"logging_steps": 100,
"max_steps": 6062,
"num_input_tokens_seen": 0,
"num_train_epochs": 7,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.5149450723328e+16,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}