{ "best_metric": 0.951232302045097, "best_model_checkpoint": "/home/avramit/classifier/04122025/classifier/binary/neodictabert-finetuned-binary-041225/checkpoint-1000", "epoch": 0.9992323719706108, "eval_steps": 500, "global_step": 1139, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.008772891764447855, "grad_norm": 18.548049926757812, "learning_rate": 4.0000000000000003e-07, "loss": 0.712, "step": 10 }, { "epoch": 0.01754578352889571, "grad_norm": 27.947410583496094, "learning_rate": 8.000000000000001e-07, "loss": 0.7286, "step": 20 }, { "epoch": 0.026318675293343568, "grad_norm": 15.611265182495117, "learning_rate": 1.2000000000000002e-06, "loss": 0.6691, "step": 30 }, { "epoch": 0.03509156705779142, "grad_norm": 14.706952095031738, "learning_rate": 1.6000000000000001e-06, "loss": 0.6678, "step": 40 }, { "epoch": 0.04386445882223928, "grad_norm": 20.6095027923584, "learning_rate": 2.0000000000000003e-06, "loss": 0.6755, "step": 50 }, { "epoch": 0.052637350586687136, "grad_norm": 15.785083770751953, "learning_rate": 2.4000000000000003e-06, "loss": 0.6245, "step": 60 }, { "epoch": 0.06141024235113499, "grad_norm": 16.40140724182129, "learning_rate": 2.8000000000000003e-06, "loss": 0.6065, "step": 70 }, { "epoch": 0.07018313411558284, "grad_norm": 16.08489990234375, "learning_rate": 3.2000000000000003e-06, "loss": 0.5906, "step": 80 }, { "epoch": 0.0789560258800307, "grad_norm": 10.633989334106445, "learning_rate": 3.6000000000000003e-06, "loss": 0.5696, "step": 90 }, { "epoch": 0.08772891764447856, "grad_norm": 13.910264015197754, "learning_rate": 4.000000000000001e-06, "loss": 0.5952, "step": 100 }, { "epoch": 0.09650180940892641, "grad_norm": 16.332483291625977, "learning_rate": 4.4e-06, "loss": 0.5808, "step": 110 }, { "epoch": 0.10527470117337427, "grad_norm": 12.419142723083496, "learning_rate": 4.800000000000001e-06, "loss": 0.564, "step": 120 }, { "epoch": 0.11404759293782213, "grad_norm": 14.58597183227539, "learning_rate": 5.2e-06, "loss": 0.6409, "step": 130 }, { "epoch": 0.12282048470226999, "grad_norm": 9.651647567749023, "learning_rate": 5.600000000000001e-06, "loss": 0.5049, "step": 140 }, { "epoch": 0.13159337646671784, "grad_norm": 23.112552642822266, "learning_rate": 6e-06, "loss": 0.6125, "step": 150 }, { "epoch": 0.1403662682311657, "grad_norm": 14.535902976989746, "learning_rate": 6.4000000000000006e-06, "loss": 0.5401, "step": 160 }, { "epoch": 0.14913915999561356, "grad_norm": 19.689529418945312, "learning_rate": 6.800000000000001e-06, "loss": 0.5239, "step": 170 }, { "epoch": 0.1579120517600614, "grad_norm": 27.909992218017578, "learning_rate": 7.2000000000000005e-06, "loss": 0.5906, "step": 180 }, { "epoch": 0.16668494352450927, "grad_norm": 10.908613204956055, "learning_rate": 7.600000000000001e-06, "loss": 0.5276, "step": 190 }, { "epoch": 0.17545783528895711, "grad_norm": 11.81062126159668, "learning_rate": 8.000000000000001e-06, "loss": 0.5355, "step": 200 }, { "epoch": 0.18423072705340499, "grad_norm": 16.26637077331543, "learning_rate": 8.400000000000001e-06, "loss": 0.5265, "step": 210 }, { "epoch": 0.19300361881785283, "grad_norm": 9.670770645141602, "learning_rate": 8.8e-06, "loss": 0.5206, "step": 220 }, { "epoch": 0.2017765105823007, "grad_norm": 14.58901309967041, "learning_rate": 9.200000000000002e-06, "loss": 0.5791, "step": 230 }, { "epoch": 0.21054940234674854, "grad_norm": 8.426600456237793, "learning_rate": 9.600000000000001e-06, "loss": 0.5681, "step": 240 }, { "epoch": 0.21932229411119641, "grad_norm": 10.079155921936035, "learning_rate": 1e-05, "loss": 0.6321, "step": 250 }, { "epoch": 0.22809518587564426, "grad_norm": 11.706363677978516, "learning_rate": 1.04e-05, "loss": 0.5798, "step": 260 }, { "epoch": 0.23686807764009213, "grad_norm": 17.979509353637695, "learning_rate": 1.0800000000000002e-05, "loss": 0.5672, "step": 270 }, { "epoch": 0.24564096940453997, "grad_norm": 10.209049224853516, "learning_rate": 1.1200000000000001e-05, "loss": 0.4833, "step": 280 }, { "epoch": 0.25441386116898784, "grad_norm": 15.83713150024414, "learning_rate": 1.16e-05, "loss": 0.5465, "step": 290 }, { "epoch": 0.2631867529334357, "grad_norm": 14.48042106628418, "learning_rate": 1.2e-05, "loss": 0.5025, "step": 300 }, { "epoch": 0.27195964469788353, "grad_norm": 14.64489459991455, "learning_rate": 1.2400000000000002e-05, "loss": 0.4883, "step": 310 }, { "epoch": 0.2807325364623314, "grad_norm": 19.78868865966797, "learning_rate": 1.2800000000000001e-05, "loss": 0.5079, "step": 320 }, { "epoch": 0.28950542822677927, "grad_norm": 11.523838996887207, "learning_rate": 1.3200000000000002e-05, "loss": 0.5315, "step": 330 }, { "epoch": 0.2982783199912271, "grad_norm": 7.584911346435547, "learning_rate": 1.3600000000000002e-05, "loss": 0.5383, "step": 340 }, { "epoch": 0.30705121175567496, "grad_norm": 10.454729080200195, "learning_rate": 1.4e-05, "loss": 0.6835, "step": 350 }, { "epoch": 0.3158241035201228, "grad_norm": 10.022806167602539, "learning_rate": 1.4400000000000001e-05, "loss": 0.625, "step": 360 }, { "epoch": 0.3245969952845707, "grad_norm": 7.622408866882324, "learning_rate": 1.48e-05, "loss": 0.6109, "step": 370 }, { "epoch": 0.33336988704901854, "grad_norm": 6.975420951843262, "learning_rate": 1.5200000000000002e-05, "loss": 0.5333, "step": 380 }, { "epoch": 0.3421427788134664, "grad_norm": 10.684267044067383, "learning_rate": 1.5600000000000003e-05, "loss": 0.4575, "step": 390 }, { "epoch": 0.35091567057791423, "grad_norm": 13.571067810058594, "learning_rate": 1.6000000000000003e-05, "loss": 0.5544, "step": 400 }, { "epoch": 0.35968856234236213, "grad_norm": 7.205085277557373, "learning_rate": 1.64e-05, "loss": 0.6155, "step": 410 }, { "epoch": 0.36846145410680997, "grad_norm": 13.097399711608887, "learning_rate": 1.6800000000000002e-05, "loss": 0.5048, "step": 420 }, { "epoch": 0.3772343458712578, "grad_norm": 16.442916870117188, "learning_rate": 1.72e-05, "loss": 0.4976, "step": 430 }, { "epoch": 0.38600723763570566, "grad_norm": 6.001971244812012, "learning_rate": 1.76e-05, "loss": 0.4145, "step": 440 }, { "epoch": 0.3947801294001535, "grad_norm": 2.986736297607422, "learning_rate": 1.8e-05, "loss": 0.5126, "step": 450 }, { "epoch": 0.4035530211646014, "grad_norm": 9.169132232666016, "learning_rate": 1.8400000000000003e-05, "loss": 0.5626, "step": 460 }, { "epoch": 0.41232591292904924, "grad_norm": 29.881427764892578, "learning_rate": 1.88e-05, "loss": 0.5039, "step": 470 }, { "epoch": 0.4210988046934971, "grad_norm": 14.136128425598145, "learning_rate": 1.9200000000000003e-05, "loss": 0.4363, "step": 480 }, { "epoch": 0.42987169645794493, "grad_norm": 10.021966934204102, "learning_rate": 1.9600000000000002e-05, "loss": 0.4673, "step": 490 }, { "epoch": 0.43864458822239283, "grad_norm": 14.060495376586914, "learning_rate": 2e-05, "loss": 0.3646, "step": 500 }, { "epoch": 0.43864458822239283, "eval_accuracy": 0.8244752121482805, "eval_f1": 0.77834179357022, "eval_loss": 0.4058316648006439, "eval_precision": 0.8146399055489965, "eval_recall": 0.7451403887688985, "eval_runtime": 561.5176, "eval_samples_per_second": 3.987, "eval_steps_per_second": 0.499, "step": 500 }, { "epoch": 0.44741747998684067, "grad_norm": 60.427162170410156, "learning_rate": 1.968701095461659e-05, "loss": 0.5038, "step": 510 }, { "epoch": 0.4561903717512885, "grad_norm": 10.006261825561523, "learning_rate": 1.9374021909233177e-05, "loss": 0.4059, "step": 520 }, { "epoch": 0.46496326351573636, "grad_norm": 7.567866802215576, "learning_rate": 1.9061032863849767e-05, "loss": 0.4314, "step": 530 }, { "epoch": 0.47373615528018426, "grad_norm": 12.703323364257812, "learning_rate": 1.8748043818466356e-05, "loss": 0.2459, "step": 540 }, { "epoch": 0.4825090470446321, "grad_norm": 3.947957754135132, "learning_rate": 1.8435054773082942e-05, "loss": 0.1668, "step": 550 }, { "epoch": 0.49128193880907994, "grad_norm": 9.213306427001953, "learning_rate": 1.8122065727699532e-05, "loss": 0.2323, "step": 560 }, { "epoch": 0.5000548305735278, "grad_norm": 0.25514841079711914, "learning_rate": 1.780907668231612e-05, "loss": 0.1923, "step": 570 }, { "epoch": 0.5088277223379757, "grad_norm": 50.12841796875, "learning_rate": 1.7496087636932707e-05, "loss": 0.4251, "step": 580 }, { "epoch": 0.5176006141024235, "grad_norm": 6.860497951507568, "learning_rate": 1.7183098591549297e-05, "loss": 0.3019, "step": 590 }, { "epoch": 0.5263735058668714, "grad_norm": 19.863380432128906, "learning_rate": 1.6870109546165886e-05, "loss": 0.1773, "step": 600 }, { "epoch": 0.5351463976313192, "grad_norm": 0.8949945569038391, "learning_rate": 1.6557120500782473e-05, "loss": 0.1894, "step": 610 }, { "epoch": 0.5439192893957671, "grad_norm": 17.815959930419922, "learning_rate": 1.6244131455399062e-05, "loss": 0.2715, "step": 620 }, { "epoch": 0.5526921811602149, "grad_norm": 12.630502700805664, "learning_rate": 1.5931142410015648e-05, "loss": 0.0829, "step": 630 }, { "epoch": 0.5614650729246627, "grad_norm": 10.575263977050781, "learning_rate": 1.5618153364632238e-05, "loss": 0.1798, "step": 640 }, { "epoch": 0.5702379646891107, "grad_norm": 18.799726486206055, "learning_rate": 1.5305164319248827e-05, "loss": 0.2497, "step": 650 }, { "epoch": 0.5790108564535585, "grad_norm": 0.6230038404464722, "learning_rate": 1.4992175273865417e-05, "loss": 0.1408, "step": 660 }, { "epoch": 0.5877837482180064, "grad_norm": 0.14638830721378326, "learning_rate": 1.4679186228482005e-05, "loss": 0.2073, "step": 670 }, { "epoch": 0.5965566399824542, "grad_norm": 11.391288757324219, "learning_rate": 1.4366197183098594e-05, "loss": 0.2668, "step": 680 }, { "epoch": 0.6053295317469021, "grad_norm": 21.246145248413086, "learning_rate": 1.4053208137715182e-05, "loss": 0.157, "step": 690 }, { "epoch": 0.6141024235113499, "grad_norm": 9.763635635375977, "learning_rate": 1.374021909233177e-05, "loss": 0.2703, "step": 700 }, { "epoch": 0.6228753152757978, "grad_norm": 5.701905727386475, "learning_rate": 1.342723004694836e-05, "loss": 0.1502, "step": 710 }, { "epoch": 0.6316482070402456, "grad_norm": 15.160579681396484, "learning_rate": 1.3114241001564947e-05, "loss": 0.1193, "step": 720 }, { "epoch": 0.6404210988046934, "grad_norm": 0.9100450873374939, "learning_rate": 1.2801251956181535e-05, "loss": 0.1526, "step": 730 }, { "epoch": 0.6491939905691414, "grad_norm": 1.2482988834381104, "learning_rate": 1.2488262910798124e-05, "loss": 0.0877, "step": 740 }, { "epoch": 0.6579668823335892, "grad_norm": 33.362972259521484, "learning_rate": 1.2175273865414712e-05, "loss": 0.2326, "step": 750 }, { "epoch": 0.6667397740980371, "grad_norm": 18.053966522216797, "learning_rate": 1.18622848200313e-05, "loss": 0.1339, "step": 760 }, { "epoch": 0.6755126658624849, "grad_norm": 0.13135869801044464, "learning_rate": 1.1549295774647888e-05, "loss": 0.111, "step": 770 }, { "epoch": 0.6842855576269328, "grad_norm": 0.08466053754091263, "learning_rate": 1.1236306729264477e-05, "loss": 0.1719, "step": 780 }, { "epoch": 0.6930584493913806, "grad_norm": 10.705062866210938, "learning_rate": 1.0923317683881065e-05, "loss": 0.271, "step": 790 }, { "epoch": 0.7018313411558285, "grad_norm": 21.25899314880371, "learning_rate": 1.0610328638497653e-05, "loss": 0.0995, "step": 800 }, { "epoch": 0.7106042329202763, "grad_norm": 4.999286651611328, "learning_rate": 1.0297339593114242e-05, "loss": 0.0837, "step": 810 }, { "epoch": 0.7193771246847243, "grad_norm": 8.578291893005371, "learning_rate": 9.98435054773083e-06, "loss": 0.1303, "step": 820 }, { "epoch": 0.7281500164491721, "grad_norm": 14.489535331726074, "learning_rate": 9.671361502347418e-06, "loss": 0.0931, "step": 830 }, { "epoch": 0.7369229082136199, "grad_norm": 0.21449220180511475, "learning_rate": 9.358372456964007e-06, "loss": 0.151, "step": 840 }, { "epoch": 0.7456957999780678, "grad_norm": 6.524932384490967, "learning_rate": 9.045383411580595e-06, "loss": 0.0509, "step": 850 }, { "epoch": 0.7544686917425156, "grad_norm": 12.212717056274414, "learning_rate": 8.732394366197183e-06, "loss": 0.2197, "step": 860 }, { "epoch": 0.7632415835069635, "grad_norm": 7.826491832733154, "learning_rate": 8.419405320813773e-06, "loss": 0.1044, "step": 870 }, { "epoch": 0.7720144752714113, "grad_norm": 0.8386934995651245, "learning_rate": 8.10641627543036e-06, "loss": 0.1518, "step": 880 }, { "epoch": 0.7807873670358592, "grad_norm": 0.19109545648097992, "learning_rate": 7.79342723004695e-06, "loss": 0.1311, "step": 890 }, { "epoch": 0.789560258800307, "grad_norm": 1.3321959972381592, "learning_rate": 7.480438184663538e-06, "loss": 0.0714, "step": 900 }, { "epoch": 0.798333150564755, "grad_norm": 8.981656074523926, "learning_rate": 7.167449139280126e-06, "loss": 0.1203, "step": 910 }, { "epoch": 0.8071060423292028, "grad_norm": 0.2177647352218628, "learning_rate": 6.854460093896714e-06, "loss": 0.087, "step": 920 }, { "epoch": 0.8158789340936506, "grad_norm": 1.4992311000823975, "learning_rate": 6.541471048513303e-06, "loss": 0.1357, "step": 930 }, { "epoch": 0.8246518258580985, "grad_norm": 2.294389486312866, "learning_rate": 6.228482003129891e-06, "loss": 0.0661, "step": 940 }, { "epoch": 0.8334247176225463, "grad_norm": 0.09688606858253479, "learning_rate": 5.915492957746479e-06, "loss": 0.0544, "step": 950 }, { "epoch": 0.8421976093869942, "grad_norm": 0.15929299592971802, "learning_rate": 5.602503912363068e-06, "loss": 0.07, "step": 960 }, { "epoch": 0.850970501151442, "grad_norm": 0.24913333356380463, "learning_rate": 5.289514866979656e-06, "loss": 0.0736, "step": 970 }, { "epoch": 0.8597433929158899, "grad_norm": 23.65605926513672, "learning_rate": 4.976525821596244e-06, "loss": 0.0595, "step": 980 }, { "epoch": 0.8685162846803377, "grad_norm": 14.373787879943848, "learning_rate": 4.663536776212833e-06, "loss": 0.1203, "step": 990 }, { "epoch": 0.8772891764447857, "grad_norm": 1.452081322669983, "learning_rate": 4.350547730829422e-06, "loss": 0.0877, "step": 1000 }, { "epoch": 0.8772891764447857, "eval_accuracy": 0.9584635998213488, "eval_f1": 0.951232302045097, "eval_loss": 0.1738564521074295, "eval_precision": 0.9245667686034659, "eval_recall": 0.9794816414686826, "eval_runtime": 479.7572, "eval_samples_per_second": 4.667, "eval_steps_per_second": 0.584, "step": 1000 }, { "epoch": 0.8860620682092335, "grad_norm": 7.37061882019043, "learning_rate": 4.0375586854460095e-06, "loss": 0.0067, "step": 1010 }, { "epoch": 0.8948349599736813, "grad_norm": 2.164485216140747, "learning_rate": 3.724569640062598e-06, "loss": 0.0256, "step": 1020 }, { "epoch": 0.9036078517381292, "grad_norm": 20.99627685546875, "learning_rate": 3.4115805946791864e-06, "loss": 0.1531, "step": 1030 }, { "epoch": 0.912380743502577, "grad_norm": 10.877192497253418, "learning_rate": 3.0985915492957746e-06, "loss": 0.1619, "step": 1040 }, { "epoch": 0.9211536352670249, "grad_norm": 0.06543366611003876, "learning_rate": 2.7856025039123637e-06, "loss": 0.0532, "step": 1050 }, { "epoch": 0.9299265270314727, "grad_norm": 1.2488692998886108, "learning_rate": 2.4726134585289515e-06, "loss": 0.0916, "step": 1060 }, { "epoch": 0.9386994187959206, "grad_norm": 0.08849219232797623, "learning_rate": 2.15962441314554e-06, "loss": 0.0663, "step": 1070 }, { "epoch": 0.9474723105603685, "grad_norm": 82.8930892944336, "learning_rate": 1.8466353677621286e-06, "loss": 0.1457, "step": 1080 }, { "epoch": 0.9562452023248164, "grad_norm": 11.937994956970215, "learning_rate": 1.5336463223787168e-06, "loss": 0.0481, "step": 1090 }, { "epoch": 0.9650180940892642, "grad_norm": 45.752872467041016, "learning_rate": 1.2206572769953053e-06, "loss": 0.1291, "step": 1100 }, { "epoch": 0.973790985853712, "grad_norm": 0.06413407623767853, "learning_rate": 9.076682316118937e-07, "loss": 0.0267, "step": 1110 }, { "epoch": 0.9825638776181599, "grad_norm": 11.743976593017578, "learning_rate": 5.94679186228482e-07, "loss": 0.1039, "step": 1120 }, { "epoch": 0.9913367693826077, "grad_norm": 1.0850673913955688, "learning_rate": 2.8169014084507043e-07, "loss": 0.0433, "step": 1130 }, { "epoch": 0.9992323719706108, "step": 1139, "total_flos": 1.186312889546834e+17, "train_loss": 0.3326558042530222, "train_runtime": 10607.6074, "train_samples_per_second": 1.719, "train_steps_per_second": 0.107 } ], "logging_steps": 10, "max_steps": 1139, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.186312889546834e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }