| { | |
| "best_metric": 0.12418080866336823, | |
| "best_model_checkpoint": "/content/drive/MyDrive/ml2/ML_project_2_course/weights1/checkpoint-700", | |
| "epoch": 4.0, | |
| "eval_steps": 100, | |
| "global_step": 788, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 1.3563565015792847, | |
| "learning_rate": 0.00019746192893401016, | |
| "loss": 0.6637, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 1.3717408180236816, | |
| "learning_rate": 0.00019492385786802033, | |
| "loss": 0.5792, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 1.0099543333053589, | |
| "learning_rate": 0.00019238578680203048, | |
| "loss": 0.5253, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 5.920238494873047, | |
| "learning_rate": 0.0001898477157360406, | |
| "loss": 0.3762, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 3.513320207595825, | |
| "learning_rate": 0.00018730964467005077, | |
| "loss": 0.3987, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 1.2757742404937744, | |
| "learning_rate": 0.00018477157360406092, | |
| "loss": 0.3336, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 1.0954958200454712, | |
| "learning_rate": 0.00018223350253807107, | |
| "loss": 0.4492, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 1.1074856519699097, | |
| "learning_rate": 0.00017969543147208124, | |
| "loss": 0.3453, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 2.97078800201416, | |
| "learning_rate": 0.0001771573604060914, | |
| "loss": 0.3103, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 1.3375837802886963, | |
| "learning_rate": 0.0001746192893401015, | |
| "loss": 0.2034, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "eval_f1": 0.8307086614173228, | |
| "eval_loss": 0.5720954537391663, | |
| "eval_runtime": 5.4754, | |
| "eval_samples_per_second": 77.255, | |
| "eval_steps_per_second": 9.68, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 1.1420457363128662, | |
| "learning_rate": 0.00017208121827411168, | |
| "loss": 0.3871, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 1.5014958381652832, | |
| "learning_rate": 0.00016954314720812183, | |
| "loss": 0.242, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 2.7499921321868896, | |
| "learning_rate": 0.00016700507614213198, | |
| "loss": 0.205, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 1.3920105695724487, | |
| "learning_rate": 0.00016446700507614215, | |
| "loss": 0.2539, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 2.2138984203338623, | |
| "learning_rate": 0.0001619289340101523, | |
| "loss": 0.3, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.4511910080909729, | |
| "learning_rate": 0.00015939086294416242, | |
| "loss": 0.1906, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 3.9145667552948, | |
| "learning_rate": 0.0001568527918781726, | |
| "loss": 0.2873, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 7.277167797088623, | |
| "learning_rate": 0.00015431472081218274, | |
| "loss": 0.1775, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 3.8661720752716064, | |
| "learning_rate": 0.0001517766497461929, | |
| "loss": 0.1682, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "grad_norm": 0.29035553336143494, | |
| "learning_rate": 0.00014923857868020306, | |
| "loss": 0.201, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "eval_f1": 0.9207048458149779, | |
| "eval_loss": 0.2542865574359894, | |
| "eval_runtime": 5.3252, | |
| "eval_samples_per_second": 79.434, | |
| "eval_steps_per_second": 9.953, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "grad_norm": 7.454772472381592, | |
| "learning_rate": 0.0001467005076142132, | |
| "loss": 0.219, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "grad_norm": 2.4295766353607178, | |
| "learning_rate": 0.00014416243654822336, | |
| "loss": 0.1328, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "grad_norm": 2.6288206577301025, | |
| "learning_rate": 0.0001416243654822335, | |
| "loss": 0.1695, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "grad_norm": 1.8671033382415771, | |
| "learning_rate": 0.00013908629441624365, | |
| "loss": 0.1648, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "grad_norm": 0.6234340667724609, | |
| "learning_rate": 0.0001365482233502538, | |
| "loss": 0.1318, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "grad_norm": 1.7282599210739136, | |
| "learning_rate": 0.00013401015228426397, | |
| "loss": 0.1646, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "grad_norm": 8.562129020690918, | |
| "learning_rate": 0.00013147208121827412, | |
| "loss": 0.0765, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "grad_norm": 0.10765184462070465, | |
| "learning_rate": 0.00012893401015228427, | |
| "loss": 0.1424, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "grad_norm": 1.785663366317749, | |
| "learning_rate": 0.00012639593908629442, | |
| "loss": 0.0838, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "grad_norm": 2.627047538757324, | |
| "learning_rate": 0.00012385786802030456, | |
| "loss": 0.2025, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "eval_f1": 0.9406392694063928, | |
| "eval_loss": 0.21563898026943207, | |
| "eval_runtime": 5.5211, | |
| "eval_samples_per_second": 76.615, | |
| "eval_steps_per_second": 9.599, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "grad_norm": 0.36638501286506653, | |
| "learning_rate": 0.00012131979695431472, | |
| "loss": 0.1436, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "grad_norm": 2.350069761276245, | |
| "learning_rate": 0.00011878172588832489, | |
| "loss": 0.0701, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "grad_norm": 1.4287878274917603, | |
| "learning_rate": 0.00011624365482233503, | |
| "loss": 0.0843, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "grad_norm": 0.32531386613845825, | |
| "learning_rate": 0.0001137055837563452, | |
| "loss": 0.1415, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "grad_norm": 0.5714915990829468, | |
| "learning_rate": 0.00011116751269035533, | |
| "loss": 0.0751, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "grad_norm": 4.953071117401123, | |
| "learning_rate": 0.00010862944162436547, | |
| "loss": 0.126, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "grad_norm": 4.382538318634033, | |
| "learning_rate": 0.00010609137055837564, | |
| "loss": 0.1279, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "grad_norm": 2.4561574459075928, | |
| "learning_rate": 0.0001035532994923858, | |
| "loss": 0.0932, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "grad_norm": 1.9325863122940063, | |
| "learning_rate": 0.00010101522842639594, | |
| "loss": 0.1361, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "grad_norm": 0.1038457602262497, | |
| "learning_rate": 9.847715736040609e-05, | |
| "loss": 0.0564, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "eval_f1": 0.9452954048140045, | |
| "eval_loss": 0.1903565376996994, | |
| "eval_runtime": 6.5595, | |
| "eval_samples_per_second": 64.486, | |
| "eval_steps_per_second": 8.08, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "grad_norm": 0.12188208103179932, | |
| "learning_rate": 9.593908629441625e-05, | |
| "loss": 0.0947, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "grad_norm": 0.8643466234207153, | |
| "learning_rate": 9.34010152284264e-05, | |
| "loss": 0.058, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "grad_norm": 0.06271246075630188, | |
| "learning_rate": 9.086294416243655e-05, | |
| "loss": 0.065, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "grad_norm": 0.593055784702301, | |
| "learning_rate": 8.83248730964467e-05, | |
| "loss": 0.0585, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "grad_norm": 0.20611844956874847, | |
| "learning_rate": 8.578680203045685e-05, | |
| "loss": 0.0695, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "grad_norm": 0.09795264154672623, | |
| "learning_rate": 8.324873096446701e-05, | |
| "loss": 0.0351, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "grad_norm": 0.04693546146154404, | |
| "learning_rate": 8.071065989847716e-05, | |
| "loss": 0.0367, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "grad_norm": 0.09200535714626312, | |
| "learning_rate": 7.817258883248731e-05, | |
| "loss": 0.0266, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "grad_norm": 0.134303480386734, | |
| "learning_rate": 7.563451776649747e-05, | |
| "loss": 0.0306, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "grad_norm": 0.041602715849876404, | |
| "learning_rate": 7.309644670050762e-05, | |
| "loss": 0.0262, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "eval_f1": 0.9495412844036697, | |
| "eval_loss": 0.21375472843647003, | |
| "eval_runtime": 5.9305, | |
| "eval_samples_per_second": 71.326, | |
| "eval_steps_per_second": 8.937, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "grad_norm": 5.353679656982422, | |
| "learning_rate": 7.055837563451776e-05, | |
| "loss": 0.0615, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "grad_norm": 7.85841703414917, | |
| "learning_rate": 6.802030456852793e-05, | |
| "loss": 0.0927, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "grad_norm": 0.5664938688278198, | |
| "learning_rate": 6.548223350253807e-05, | |
| "loss": 0.1466, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "grad_norm": 0.04644864797592163, | |
| "learning_rate": 6.294416243654822e-05, | |
| "loss": 0.0148, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "grad_norm": 0.043112464249134064, | |
| "learning_rate": 6.040609137055838e-05, | |
| "loss": 0.041, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "grad_norm": 0.9218811392784119, | |
| "learning_rate": 5.786802030456853e-05, | |
| "loss": 0.0998, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "grad_norm": 0.09559272974729538, | |
| "learning_rate": 5.532994923857868e-05, | |
| "loss": 0.0447, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "grad_norm": 0.04861769080162048, | |
| "learning_rate": 5.2791878172588836e-05, | |
| "loss": 0.0296, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "grad_norm": 0.13610342144966125, | |
| "learning_rate": 5.0253807106598984e-05, | |
| "loss": 0.037, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "grad_norm": 0.048563599586486816, | |
| "learning_rate": 4.771573604060914e-05, | |
| "loss": 0.0222, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "eval_f1": 0.9508196721311475, | |
| "eval_loss": 0.15546846389770508, | |
| "eval_runtime": 5.4832, | |
| "eval_samples_per_second": 77.145, | |
| "eval_steps_per_second": 9.666, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 3.1, | |
| "grad_norm": 0.05200362578034401, | |
| "learning_rate": 4.517766497461929e-05, | |
| "loss": 0.0118, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 3.15, | |
| "grad_norm": 1.2061017751693726, | |
| "learning_rate": 4.2639593908629446e-05, | |
| "loss": 0.0469, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "grad_norm": 0.05518200248479843, | |
| "learning_rate": 4.010152284263959e-05, | |
| "loss": 0.0228, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 3.25, | |
| "grad_norm": 0.19791868329048157, | |
| "learning_rate": 3.756345177664975e-05, | |
| "loss": 0.0184, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 3.3, | |
| "grad_norm": 0.10160644352436066, | |
| "learning_rate": 3.50253807106599e-05, | |
| "loss": 0.0267, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 3.35, | |
| "grad_norm": 0.04751751944422722, | |
| "learning_rate": 3.248730964467005e-05, | |
| "loss": 0.0096, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "grad_norm": 0.04644055664539337, | |
| "learning_rate": 2.9949238578680206e-05, | |
| "loss": 0.0367, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 3.45, | |
| "grad_norm": 0.060281168669462204, | |
| "learning_rate": 2.7411167512690357e-05, | |
| "loss": 0.0068, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 3.5, | |
| "grad_norm": 0.034235160797834396, | |
| "learning_rate": 2.4873096446700507e-05, | |
| "loss": 0.0398, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "grad_norm": 0.3630073368549347, | |
| "learning_rate": 2.233502538071066e-05, | |
| "loss": 0.0069, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "eval_f1": 0.963302752293578, | |
| "eval_loss": 0.12418080866336823, | |
| "eval_runtime": 6.5914, | |
| "eval_samples_per_second": 64.175, | |
| "eval_steps_per_second": 8.041, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "grad_norm": 0.040880296379327774, | |
| "learning_rate": 1.9796954314720815e-05, | |
| "loss": 0.0045, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 3.65, | |
| "grad_norm": 0.5524014830589294, | |
| "learning_rate": 1.7258883248730966e-05, | |
| "loss": 0.04, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "grad_norm": 0.03459872677922249, | |
| "learning_rate": 1.4720812182741117e-05, | |
| "loss": 0.0131, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "grad_norm": 2.4353725910186768, | |
| "learning_rate": 1.218274111675127e-05, | |
| "loss": 0.0114, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "grad_norm": 0.049515530467033386, | |
| "learning_rate": 9.644670050761421e-06, | |
| "loss": 0.0163, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 3.86, | |
| "grad_norm": 0.033101875334978104, | |
| "learning_rate": 7.106598984771575e-06, | |
| "loss": 0.0108, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 3.91, | |
| "grad_norm": 0.08629802614450455, | |
| "learning_rate": 4.568527918781726e-06, | |
| "loss": 0.0435, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "grad_norm": 0.09244952350854874, | |
| "learning_rate": 2.030456852791878e-06, | |
| "loss": 0.0195, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "step": 788, | |
| "total_flos": 9.74539261393109e+17, | |
| "train_loss": 0.13476083650806833, | |
| "train_runtime": 481.1459, | |
| "train_samples_per_second": 26.138, | |
| "train_steps_per_second": 1.638 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 788, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 4, | |
| "save_steps": 100, | |
| "total_flos": 9.74539261393109e+17, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |