| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.173964149280047, | |
| "eval_steps": 500, | |
| "global_step": 500, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.023508668821627974, | |
| "grad_norm": 84.37700653076172, | |
| "learning_rate": 7.031250000000001e-06, | |
| "loss": 115.85, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.04701733764325595, | |
| "grad_norm": 106.4520034790039, | |
| "learning_rate": 1.484375e-05, | |
| "loss": 116.5781, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.07052600646488393, | |
| "grad_norm": 130.89254760742188, | |
| "learning_rate": 2.2656250000000002e-05, | |
| "loss": 103.3281, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.0940346752865119, | |
| "grad_norm": 37.089256286621094, | |
| "learning_rate": 3.0468750000000002e-05, | |
| "loss": 60.0406, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.11754334410813988, | |
| "grad_norm": 44.051109313964844, | |
| "learning_rate": 3.828125e-05, | |
| "loss": 39.75, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.14105201292976785, | |
| "grad_norm": 43.561981201171875, | |
| "learning_rate": 4.609375e-05, | |
| "loss": 24.7984, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.16456068175139582, | |
| "grad_norm": 34.622520446777344, | |
| "learning_rate": 5.3906250000000006e-05, | |
| "loss": 11.9723, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.1880693505730238, | |
| "grad_norm": 22.26849937438965, | |
| "learning_rate": 6.171875e-05, | |
| "loss": 5.4578, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.2115780193946518, | |
| "grad_norm": 9.059814453125, | |
| "learning_rate": 6.953125e-05, | |
| "loss": 3.992, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.23508668821627976, | |
| "grad_norm": 4.332883834838867, | |
| "learning_rate": 7.734375e-05, | |
| "loss": 3.1058, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.2585953570379077, | |
| "grad_norm": 6.844908237457275, | |
| "learning_rate": 8.515625e-05, | |
| "loss": 2.6654, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.2821040258595357, | |
| "grad_norm": 4.539117813110352, | |
| "learning_rate": 9.296875e-05, | |
| "loss": 2.422, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.3056126946811637, | |
| "grad_norm": 3.098025321960449, | |
| "learning_rate": 9.999981342914437e-05, | |
| "loss": 2.1623, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.32912136350279164, | |
| "grad_norm": 1.3452321290969849, | |
| "learning_rate": 9.997742661115932e-05, | |
| "loss": 2.1429, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.35263003232441964, | |
| "grad_norm": 0.37226417660713196, | |
| "learning_rate": 9.991774476447404e-05, | |
| "loss": 2.1347, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.3761387011460476, | |
| "grad_norm": 3.523218870162964, | |
| "learning_rate": 9.982081242591919e-05, | |
| "loss": 2.0512, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.3996473699676756, | |
| "grad_norm": 0.33414289355278015, | |
| "learning_rate": 9.968670193003843e-05, | |
| "loss": 2.047, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.4231560387893036, | |
| "grad_norm": 0.24106919765472412, | |
| "learning_rate": 9.951551335510978e-05, | |
| "loss": 2.1046, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.4466647076109315, | |
| "grad_norm": 0.328752726316452, | |
| "learning_rate": 9.930737444846331e-05, | |
| "loss": 2.1397, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.4701733764325595, | |
| "grad_norm": 1.1409285068511963, | |
| "learning_rate": 9.906244053115143e-05, | |
| "loss": 2.1377, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.49368204525418746, | |
| "grad_norm": 0.0690290704369545, | |
| "learning_rate": 9.87808943820424e-05, | |
| "loss": 2.0992, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.5171907140758154, | |
| "grad_norm": 0.20527280867099762, | |
| "learning_rate": 9.846294610142398e-05, | |
| "loss": 2.1339, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.5406993828974435, | |
| "grad_norm": 0.3991909623146057, | |
| "learning_rate": 9.810883295421864e-05, | |
| "loss": 2.1176, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.5642080517190714, | |
| "grad_norm": 0.08950258791446686, | |
| "learning_rate": 9.771881919292765e-05, | |
| "loss": 2.0895, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.5877167205406993, | |
| "grad_norm": 0.17848969995975494, | |
| "learning_rate": 9.729319586043591e-05, | |
| "loss": 2.0877, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.6112253893623274, | |
| "grad_norm": 0.30160897970199585, | |
| "learning_rate": 9.683228057282483e-05, | |
| "loss": 2.0648, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.6347340581839553, | |
| "grad_norm": 0.11883804202079773, | |
| "learning_rate": 9.63364172823554e-05, | |
| "loss": 2.0591, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.6582427270055833, | |
| "grad_norm": 0.1619461327791214, | |
| "learning_rate": 9.580597602079802e-05, | |
| "loss": 2.1386, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.6817513958272113, | |
| "grad_norm": 0.11473017930984497, | |
| "learning_rate": 9.524135262330098e-05, | |
| "loss": 2.1028, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.7052600646488393, | |
| "grad_norm": 0.059143248945474625, | |
| "learning_rate": 9.464296843300342e-05, | |
| "loss": 2.0881, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.7287687334704672, | |
| "grad_norm": 0.1624162495136261, | |
| "learning_rate": 9.401126998661328e-05, | |
| "loss": 2.0971, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.7522774022920952, | |
| "grad_norm": 0.15460653603076935, | |
| "learning_rate": 9.334672868118491e-05, | |
| "loss": 2.1154, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.7757860711137232, | |
| "grad_norm": 0.4213317334651947, | |
| "learning_rate": 9.26498404223449e-05, | |
| "loss": 2.1113, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.7992947399353512, | |
| "grad_norm": 0.14529550075531006, | |
| "learning_rate": 9.192112525422868e-05, | |
| "loss": 2.0867, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.8228034087569791, | |
| "grad_norm": 0.06652400642633438, | |
| "learning_rate": 9.116112697140418e-05, | |
| "loss": 2.0807, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.8463120775786072, | |
| "grad_norm": 0.12574820220470428, | |
| "learning_rate": 9.037041271307188e-05, | |
| "loss": 2.1711, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.8698207464002351, | |
| "grad_norm": 0.10869150608778, | |
| "learning_rate": 8.954957253984426e-05, | |
| "loss": 2.1252, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.893329415221863, | |
| "grad_norm": 0.143524169921875, | |
| "learning_rate": 8.869921899342056e-05, | |
| "loss": 2.0468, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.916838084043491, | |
| "grad_norm": 0.27682727575302124, | |
| "learning_rate": 8.781998663948513e-05, | |
| "loss": 2.1023, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.940346752865119, | |
| "grad_norm": 0.08671136200428009, | |
| "learning_rate": 8.691253159417074e-05, | |
| "loss": 2.0991, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.963855421686747, | |
| "grad_norm": 0.14743109047412872, | |
| "learning_rate": 8.597753103444016e-05, | |
| "loss": 2.0858, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.9873640905083749, | |
| "grad_norm": 0.14369799196720123, | |
| "learning_rate": 8.501568269275126e-05, | |
| "loss": 2.1057, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.0094034675286512, | |
| "grad_norm": 0.09220755100250244, | |
| "learning_rate": 8.40277043363831e-05, | |
| "loss": 1.9405, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.0329121363502791, | |
| "grad_norm": 0.25697803497314453, | |
| "learning_rate": 8.301433323181076e-05, | |
| "loss": 2.1221, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.056420805171907, | |
| "grad_norm": 0.1321459412574768, | |
| "learning_rate": 8.19763255945298e-05, | |
| "loss": 2.1576, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.079929473993535, | |
| "grad_norm": 0.11310122162103653, | |
| "learning_rate": 8.091445602473972e-05, | |
| "loss": 2.0483, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.1034381428151632, | |
| "grad_norm": 0.08170254528522491, | |
| "learning_rate": 7.982951692930829e-05, | |
| "loss": 2.1367, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.1269468116367911, | |
| "grad_norm": 0.022492246702313423, | |
| "learning_rate": 7.87223179304479e-05, | |
| "loss": 2.1435, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.150455480458419, | |
| "grad_norm": 0.015893638134002686, | |
| "learning_rate": 7.759368526154509e-05, | |
| "loss": 2.0666, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.173964149280047, | |
| "grad_norm": 0.053727954626083374, | |
| "learning_rate": 7.644446115059425e-05, | |
| "loss": 2.0886, | |
| "step": 500 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 1278, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 8.002988656862822e+16, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |