{ "best_metric": 0.03188331797719002, "best_model_checkpoint": "saves/psy-course/MentaLLaMA-chat-7B/train/fold6/checkpoint-1900", "epoch": 4.995305164319249, "eval_steps": 50, "global_step": 3325, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.015023474178403756, "grad_norm": 1.6643978357315063, "learning_rate": 3.003003003003003e-06, "loss": 1.6617, "step": 10 }, { "epoch": 0.03004694835680751, "grad_norm": 1.970920443534851, "learning_rate": 6.006006006006006e-06, "loss": 1.7717, "step": 20 }, { "epoch": 0.04507042253521127, "grad_norm": 2.2089600563049316, "learning_rate": 9.00900900900901e-06, "loss": 1.6755, "step": 30 }, { "epoch": 0.06009389671361502, "grad_norm": 3.677905797958374, "learning_rate": 1.2012012012012012e-05, "loss": 1.3133, "step": 40 }, { "epoch": 0.07511737089201878, "grad_norm": 1.0802251100540161, "learning_rate": 1.5015015015015016e-05, "loss": 0.8292, "step": 50 }, { "epoch": 0.07511737089201878, "eval_loss": 0.6473308205604553, "eval_runtime": 207.0843, "eval_samples_per_second": 5.717, "eval_steps_per_second": 5.717, "step": 50 }, { "epoch": 0.09014084507042254, "grad_norm": 0.8090672492980957, "learning_rate": 1.801801801801802e-05, "loss": 0.6816, "step": 60 }, { "epoch": 0.10516431924882629, "grad_norm": 0.7934019565582275, "learning_rate": 2.102102102102102e-05, "loss": 0.4595, "step": 70 }, { "epoch": 0.12018779342723004, "grad_norm": 0.7469967007637024, "learning_rate": 2.4024024024024024e-05, "loss": 0.2868, "step": 80 }, { "epoch": 0.1352112676056338, "grad_norm": 0.6989622712135315, "learning_rate": 2.702702702702703e-05, "loss": 0.1691, "step": 90 }, { "epoch": 0.15023474178403756, "grad_norm": 0.714993417263031, "learning_rate": 3.0030030030030033e-05, "loss": 0.1595, "step": 100 }, { "epoch": 0.15023474178403756, "eval_loss": 0.11690635979175568, "eval_runtime": 208.0685, "eval_samples_per_second": 5.69, "eval_steps_per_second": 5.69, "step": 100 }, { "epoch": 0.1652582159624413, "grad_norm": 0.6915186643600464, "learning_rate": 3.3033033033033035e-05, "loss": 0.1358, "step": 110 }, { "epoch": 0.18028169014084508, "grad_norm": 0.7759954333305359, "learning_rate": 3.603603603603604e-05, "loss": 0.1063, "step": 120 }, { "epoch": 0.19530516431924883, "grad_norm": 0.8066917061805725, "learning_rate": 3.903903903903904e-05, "loss": 0.0996, "step": 130 }, { "epoch": 0.21032863849765257, "grad_norm": 0.4547995924949646, "learning_rate": 4.204204204204204e-05, "loss": 0.1047, "step": 140 }, { "epoch": 0.22535211267605634, "grad_norm": 0.7537536025047302, "learning_rate": 4.5045045045045046e-05, "loss": 0.0933, "step": 150 }, { "epoch": 0.22535211267605634, "eval_loss": 0.07271095365285873, "eval_runtime": 207.977, "eval_samples_per_second": 5.693, "eval_steps_per_second": 5.693, "step": 150 }, { "epoch": 0.2403755868544601, "grad_norm": 0.652043879032135, "learning_rate": 4.804804804804805e-05, "loss": 0.0726, "step": 160 }, { "epoch": 0.25539906103286386, "grad_norm": 0.7913267016410828, "learning_rate": 5.105105105105106e-05, "loss": 0.0503, "step": 170 }, { "epoch": 0.2704225352112676, "grad_norm": 0.6111043095588684, "learning_rate": 5.405405405405406e-05, "loss": 0.0682, "step": 180 }, { "epoch": 0.28544600938967135, "grad_norm": 0.6898893713951111, "learning_rate": 5.705705705705706e-05, "loss": 0.0651, "step": 190 }, { "epoch": 0.3004694835680751, "grad_norm": 0.5223749876022339, "learning_rate": 6.0060060060060066e-05, "loss": 0.0512, "step": 200 }, { "epoch": 0.3004694835680751, "eval_loss": 0.05808287858963013, "eval_runtime": 207.9535, "eval_samples_per_second": 5.694, "eval_steps_per_second": 5.694, "step": 200 }, { "epoch": 0.3154929577464789, "grad_norm": 0.7966965436935425, "learning_rate": 6.306306306306306e-05, "loss": 0.0562, "step": 210 }, { "epoch": 0.3305164319248826, "grad_norm": 0.4759034514427185, "learning_rate": 6.606606606606607e-05, "loss": 0.0591, "step": 220 }, { "epoch": 0.3455399061032864, "grad_norm": 0.421832412481308, "learning_rate": 6.906906906906907e-05, "loss": 0.0504, "step": 230 }, { "epoch": 0.36056338028169016, "grad_norm": 0.6044343113899231, "learning_rate": 7.207207207207208e-05, "loss": 0.0609, "step": 240 }, { "epoch": 0.3755868544600939, "grad_norm": 0.702263355255127, "learning_rate": 7.507507507507507e-05, "loss": 0.0619, "step": 250 }, { "epoch": 0.3755868544600939, "eval_loss": 0.04737601429224014, "eval_runtime": 207.9514, "eval_samples_per_second": 5.694, "eval_steps_per_second": 5.694, "step": 250 }, { "epoch": 0.39061032863849765, "grad_norm": 0.4086180627346039, "learning_rate": 7.807807807807808e-05, "loss": 0.0473, "step": 260 }, { "epoch": 0.4056338028169014, "grad_norm": 0.3927551209926605, "learning_rate": 8.108108108108109e-05, "loss": 0.0552, "step": 270 }, { "epoch": 0.42065727699530514, "grad_norm": 0.26502013206481934, "learning_rate": 8.408408408408409e-05, "loss": 0.0567, "step": 280 }, { "epoch": 0.4356807511737089, "grad_norm": 0.8767386674880981, "learning_rate": 8.70870870870871e-05, "loss": 0.0572, "step": 290 }, { "epoch": 0.4507042253521127, "grad_norm": 0.4874660074710846, "learning_rate": 9.009009009009009e-05, "loss": 0.0395, "step": 300 }, { "epoch": 0.4507042253521127, "eval_loss": 0.04596845433115959, "eval_runtime": 207.5439, "eval_samples_per_second": 5.705, "eval_steps_per_second": 5.705, "step": 300 }, { "epoch": 0.46572769953051646, "grad_norm": 0.43807005882263184, "learning_rate": 9.30930930930931e-05, "loss": 0.0444, "step": 310 }, { "epoch": 0.4807511737089202, "grad_norm": 0.7260940074920654, "learning_rate": 9.60960960960961e-05, "loss": 0.063, "step": 320 }, { "epoch": 0.49577464788732395, "grad_norm": 0.40789633989334106, "learning_rate": 9.90990990990991e-05, "loss": 0.0497, "step": 330 }, { "epoch": 0.5107981220657277, "grad_norm": 0.5078976154327393, "learning_rate": 9.999864944989638e-05, "loss": 0.058, "step": 340 }, { "epoch": 0.5258215962441315, "grad_norm": 0.3737334907054901, "learning_rate": 9.999203468625017e-05, "loss": 0.0476, "step": 350 }, { "epoch": 0.5258215962441315, "eval_loss": 0.045401681214571, "eval_runtime": 207.0191, "eval_samples_per_second": 5.719, "eval_steps_per_second": 5.719, "step": 350 }, { "epoch": 0.5408450704225352, "grad_norm": 0.5752517580986023, "learning_rate": 9.997990837719421e-05, "loss": 0.055, "step": 360 }, { "epoch": 0.5558685446009389, "grad_norm": 0.28732818365097046, "learning_rate": 9.996227185963554e-05, "loss": 0.0412, "step": 370 }, { "epoch": 0.5708920187793427, "grad_norm": 0.41930803656578064, "learning_rate": 9.993912707797329e-05, "loss": 0.0448, "step": 380 }, { "epoch": 0.5859154929577465, "grad_norm": 0.35693687200546265, "learning_rate": 9.99104765838842e-05, "loss": 0.0423, "step": 390 }, { "epoch": 0.6009389671361502, "grad_norm": 0.11278364062309265, "learning_rate": 9.987632353604151e-05, "loss": 0.0444, "step": 400 }, { "epoch": 0.6009389671361502, "eval_loss": 0.0407099612057209, "eval_runtime": 207.012, "eval_samples_per_second": 5.719, "eval_steps_per_second": 5.719, "step": 400 }, { "epoch": 0.615962441314554, "grad_norm": 0.41261032223701477, "learning_rate": 9.98366716997665e-05, "loss": 0.0478, "step": 410 }, { "epoch": 0.6309859154929578, "grad_norm": 0.4349944293498993, "learning_rate": 9.979152544661354e-05, "loss": 0.0381, "step": 420 }, { "epoch": 0.6460093896713615, "grad_norm": 0.609887957572937, "learning_rate": 9.974088975388802e-05, "loss": 0.0435, "step": 430 }, { "epoch": 0.6610328638497652, "grad_norm": 0.37665075063705444, "learning_rate": 9.968477020409766e-05, "loss": 0.0378, "step": 440 }, { "epoch": 0.676056338028169, "grad_norm": 0.30656006932258606, "learning_rate": 9.962317298433705e-05, "loss": 0.0543, "step": 450 }, { "epoch": 0.676056338028169, "eval_loss": 0.042538370937108994, "eval_runtime": 207.26, "eval_samples_per_second": 5.713, "eval_steps_per_second": 5.713, "step": 450 }, { "epoch": 0.6910798122065728, "grad_norm": 0.2620968520641327, "learning_rate": 9.955610488560551e-05, "loss": 0.0529, "step": 460 }, { "epoch": 0.7061032863849765, "grad_norm": 0.4962412118911743, "learning_rate": 9.948357330205842e-05, "loss": 0.0488, "step": 470 }, { "epoch": 0.7211267605633803, "grad_norm": 0.11893105506896973, "learning_rate": 9.940558623019201e-05, "loss": 0.0282, "step": 480 }, { "epoch": 0.7361502347417841, "grad_norm": 0.24164587259292603, "learning_rate": 9.932215226796172e-05, "loss": 0.0377, "step": 490 }, { "epoch": 0.7511737089201878, "grad_norm": 0.23037177324295044, "learning_rate": 9.923328061383435e-05, "loss": 0.0454, "step": 500 }, { "epoch": 0.7511737089201878, "eval_loss": 0.0371524840593338, "eval_runtime": 207.3218, "eval_samples_per_second": 5.711, "eval_steps_per_second": 5.711, "step": 500 }, { "epoch": 0.7661971830985915, "grad_norm": 0.1442870795726776, "learning_rate": 9.913898106577393e-05, "loss": 0.0382, "step": 510 }, { "epoch": 0.7812206572769953, "grad_norm": 0.2532236576080322, "learning_rate": 9.903926402016153e-05, "loss": 0.0344, "step": 520 }, { "epoch": 0.7962441314553991, "grad_norm": 0.14402006566524506, "learning_rate": 9.893414047064897e-05, "loss": 0.0398, "step": 530 }, { "epoch": 0.8112676056338028, "grad_norm": 0.2404378205537796, "learning_rate": 9.88236220069469e-05, "loss": 0.0381, "step": 540 }, { "epoch": 0.8262910798122066, "grad_norm": 0.27102166414260864, "learning_rate": 9.870772081354705e-05, "loss": 0.0562, "step": 550 }, { "epoch": 0.8262910798122066, "eval_loss": 0.03766689449548721, "eval_runtime": 207.5601, "eval_samples_per_second": 5.704, "eval_steps_per_second": 5.704, "step": 550 }, { "epoch": 0.8413145539906103, "grad_norm": 0.2679292857646942, "learning_rate": 9.858644966837878e-05, "loss": 0.0406, "step": 560 }, { "epoch": 0.856338028169014, "grad_norm": 0.0837322548031807, "learning_rate": 9.845982194140051e-05, "loss": 0.0381, "step": 570 }, { "epoch": 0.8713615023474178, "grad_norm": 0.1658477932214737, "learning_rate": 9.832785159312559e-05, "loss": 0.0404, "step": 580 }, { "epoch": 0.8863849765258216, "grad_norm": 0.2504975199699402, "learning_rate": 9.819055317308317e-05, "loss": 0.0365, "step": 590 }, { "epoch": 0.9014084507042254, "grad_norm": 0.24881432950496674, "learning_rate": 9.804794181821422e-05, "loss": 0.0336, "step": 600 }, { "epoch": 0.9014084507042254, "eval_loss": 0.036089226603507996, "eval_runtime": 208.6413, "eval_samples_per_second": 5.675, "eval_steps_per_second": 5.675, "step": 600 }, { "epoch": 0.9164319248826291, "grad_norm": 0.20400799810886383, "learning_rate": 9.790003325120261e-05, "loss": 0.0375, "step": 610 }, { "epoch": 0.9314553990610329, "grad_norm": 0.4139688313007355, "learning_rate": 9.774684377874178e-05, "loss": 0.0466, "step": 620 }, { "epoch": 0.9464788732394366, "grad_norm": 0.302133172750473, "learning_rate": 9.758839028973692e-05, "loss": 0.0308, "step": 630 }, { "epoch": 0.9615023474178404, "grad_norm": 0.26886606216430664, "learning_rate": 9.742469025344298e-05, "loss": 0.0416, "step": 640 }, { "epoch": 0.9765258215962441, "grad_norm": 0.22003549337387085, "learning_rate": 9.725576171753874e-05, "loss": 0.0494, "step": 650 }, { "epoch": 0.9765258215962441, "eval_loss": 0.03678149729967117, "eval_runtime": 209.7916, "eval_samples_per_second": 5.644, "eval_steps_per_second": 5.644, "step": 650 }, { "epoch": 0.9915492957746479, "grad_norm": 0.686739981174469, "learning_rate": 9.708162330613708e-05, "loss": 0.0367, "step": 660 }, { "epoch": 1.0065727699530516, "grad_norm": 0.1825464963912964, "learning_rate": 9.690229421773167e-05, "loss": 0.0369, "step": 670 }, { "epoch": 1.0215962441314554, "grad_norm": 0.442874014377594, "learning_rate": 9.67177942230804e-05, "loss": 0.0312, "step": 680 }, { "epoch": 1.036619718309859, "grad_norm": 0.45542624592781067, "learning_rate": 9.652814366302568e-05, "loss": 0.0392, "step": 690 }, { "epoch": 1.051643192488263, "grad_norm": 0.16704720258712769, "learning_rate": 9.633336344625185e-05, "loss": 0.0354, "step": 700 }, { "epoch": 1.051643192488263, "eval_loss": 0.0386400930583477, "eval_runtime": 210.8244, "eval_samples_per_second": 5.616, "eval_steps_per_second": 5.616, "step": 700 }, { "epoch": 1.0666666666666667, "grad_norm": 0.2750037610530853, "learning_rate": 9.61334750469801e-05, "loss": 0.0282, "step": 710 }, { "epoch": 1.0816901408450703, "grad_norm": 0.5251227021217346, "learning_rate": 9.592850050260089e-05, "loss": 0.0349, "step": 720 }, { "epoch": 1.0967136150234742, "grad_norm": 0.13910022377967834, "learning_rate": 9.571846241124446e-05, "loss": 0.0291, "step": 730 }, { "epoch": 1.1117370892018779, "grad_norm": 0.17034609615802765, "learning_rate": 9.55033839292893e-05, "loss": 0.0257, "step": 740 }, { "epoch": 1.1267605633802817, "grad_norm": 0.4655674397945404, "learning_rate": 9.52832887688093e-05, "loss": 0.029, "step": 750 }, { "epoch": 1.1267605633802817, "eval_loss": 0.037642158567905426, "eval_runtime": 212.6123, "eval_samples_per_second": 5.569, "eval_steps_per_second": 5.569, "step": 750 }, { "epoch": 1.1417840375586854, "grad_norm": 0.33910053968429565, "learning_rate": 9.50582011949595e-05, "loss": 0.0329, "step": 760 }, { "epoch": 1.1568075117370893, "grad_norm": 0.1620519608259201, "learning_rate": 9.482814602330084e-05, "loss": 0.0328, "step": 770 }, { "epoch": 1.171830985915493, "grad_norm": 0.3181590735912323, "learning_rate": 9.459314861706435e-05, "loss": 0.0363, "step": 780 }, { "epoch": 1.1868544600938966, "grad_norm": 0.3142762780189514, "learning_rate": 9.435323488435488e-05, "loss": 0.0279, "step": 790 }, { "epoch": 1.2018779342723005, "grad_norm": 0.34480199217796326, "learning_rate": 9.410843127529473e-05, "loss": 0.0301, "step": 800 }, { "epoch": 1.2018779342723005, "eval_loss": 0.03519855812191963, "eval_runtime": 213.9318, "eval_samples_per_second": 5.534, "eval_steps_per_second": 5.534, "step": 800 }, { "epoch": 1.2169014084507042, "grad_norm": 0.3214387893676758, "learning_rate": 9.385876477910765e-05, "loss": 0.0317, "step": 810 }, { "epoch": 1.231924882629108, "grad_norm": 0.20507577061653137, "learning_rate": 9.360426292114314e-05, "loss": 0.0291, "step": 820 }, { "epoch": 1.2469483568075117, "grad_norm": 0.2585867941379547, "learning_rate": 9.334495375984212e-05, "loss": 0.0279, "step": 830 }, { "epoch": 1.2619718309859156, "grad_norm": 0.21939821541309357, "learning_rate": 9.30808658836432e-05, "loss": 0.0293, "step": 840 }, { "epoch": 1.2769953051643192, "grad_norm": 0.42434459924697876, "learning_rate": 9.281202840783108e-05, "loss": 0.0321, "step": 850 }, { "epoch": 1.2769953051643192, "eval_loss": 0.034056928008794785, "eval_runtime": 215.0317, "eval_samples_per_second": 5.506, "eval_steps_per_second": 5.506, "step": 850 }, { "epoch": 1.292018779342723, "grad_norm": 0.22999003529548645, "learning_rate": 9.253847097132655e-05, "loss": 0.0243, "step": 860 }, { "epoch": 1.3070422535211268, "grad_norm": 0.1908729076385498, "learning_rate": 9.226022373341882e-05, "loss": 0.0247, "step": 870 }, { "epoch": 1.3220657276995305, "grad_norm": 0.16463080048561096, "learning_rate": 9.19773173704406e-05, "loss": 0.0245, "step": 880 }, { "epoch": 1.3370892018779343, "grad_norm": 0.5442132949829102, "learning_rate": 9.168978307238594e-05, "loss": 0.032, "step": 890 }, { "epoch": 1.352112676056338, "grad_norm": 0.11218319833278656, "learning_rate": 9.13976525394717e-05, "loss": 0.0271, "step": 900 }, { "epoch": 1.352112676056338, "eval_loss": 0.03430046886205673, "eval_runtime": 215.5861, "eval_samples_per_second": 5.492, "eval_steps_per_second": 5.492, "step": 900 }, { "epoch": 1.3671361502347419, "grad_norm": 0.13387706875801086, "learning_rate": 9.110095797864263e-05, "loss": 0.0224, "step": 910 }, { "epoch": 1.3821596244131455, "grad_norm": 0.20326955616474152, "learning_rate": 9.079973210002051e-05, "loss": 0.0318, "step": 920 }, { "epoch": 1.3971830985915492, "grad_norm": 0.2666754424571991, "learning_rate": 9.049400811329807e-05, "loss": 0.0322, "step": 930 }, { "epoch": 1.412206572769953, "grad_norm": 0.29467517137527466, "learning_rate": 9.01838197240775e-05, "loss": 0.0305, "step": 940 }, { "epoch": 1.4272300469483568, "grad_norm": 0.27433130145072937, "learning_rate": 8.986920113015461e-05, "loss": 0.0351, "step": 950 }, { "epoch": 1.4272300469483568, "eval_loss": 0.03298981115221977, "eval_runtime": 216.0148, "eval_samples_per_second": 5.481, "eval_steps_per_second": 5.481, "step": 950 }, { "epoch": 1.4422535211267606, "grad_norm": 0.23651893436908722, "learning_rate": 8.955018701774846e-05, "loss": 0.0262, "step": 960 }, { "epoch": 1.4572769953051643, "grad_norm": 0.2778637111186981, "learning_rate": 8.922681255767731e-05, "loss": 0.0347, "step": 970 }, { "epoch": 1.4723004694835682, "grad_norm": 0.18652495741844177, "learning_rate": 8.889911340148112e-05, "loss": 0.0308, "step": 980 }, { "epoch": 1.4873239436619718, "grad_norm": 0.17595553398132324, "learning_rate": 8.856712567749095e-05, "loss": 0.0278, "step": 990 }, { "epoch": 1.5023474178403755, "grad_norm": 0.4016423523426056, "learning_rate": 8.82308859868459e-05, "loss": 0.0244, "step": 1000 }, { "epoch": 1.5023474178403755, "eval_loss": 0.033029552549123764, "eval_runtime": 217.6657, "eval_samples_per_second": 5.44, "eval_steps_per_second": 5.44, "step": 1000 }, { "epoch": 1.5173708920187794, "grad_norm": 0.2475142925977707, "learning_rate": 8.789043139945795e-05, "loss": 0.0298, "step": 1010 }, { "epoch": 1.532394366197183, "grad_norm": 0.2059515416622162, "learning_rate": 8.754579944992491e-05, "loss": 0.036, "step": 1020 }, { "epoch": 1.5474178403755867, "grad_norm": 0.16493824124336243, "learning_rate": 8.719702813339248e-05, "loss": 0.0288, "step": 1030 }, { "epoch": 1.5624413145539906, "grad_norm": 0.20951515436172485, "learning_rate": 8.684415590136518e-05, "loss": 0.0277, "step": 1040 }, { "epoch": 1.5774647887323945, "grad_norm": 0.0918218195438385, "learning_rate": 8.648722165746722e-05, "loss": 0.0277, "step": 1050 }, { "epoch": 1.5774647887323945, "eval_loss": 0.034053802490234375, "eval_runtime": 219.1262, "eval_samples_per_second": 5.403, "eval_steps_per_second": 5.403, "step": 1050 }, { "epoch": 1.5924882629107981, "grad_norm": 0.09690333157777786, "learning_rate": 8.61262647531534e-05, "loss": 0.0251, "step": 1060 }, { "epoch": 1.6075117370892018, "grad_norm": 0.19759806990623474, "learning_rate": 8.576132498337068e-05, "loss": 0.0362, "step": 1070 }, { "epoch": 1.6225352112676057, "grad_norm": 0.2693615257740021, "learning_rate": 8.539244258217088e-05, "loss": 0.0319, "step": 1080 }, { "epoch": 1.6375586854460094, "grad_norm": 0.20964917540550232, "learning_rate": 8.501965821827485e-05, "loss": 0.032, "step": 1090 }, { "epoch": 1.652582159624413, "grad_norm": 0.2406897395849228, "learning_rate": 8.464301299058892e-05, "loss": 0.0231, "step": 1100 }, { "epoch": 1.652582159624413, "eval_loss": 0.034014929085969925, "eval_runtime": 219.9896, "eval_samples_per_second": 5.382, "eval_steps_per_second": 5.382, "step": 1100 }, { "epoch": 1.667605633802817, "grad_norm": 0.25754356384277344, "learning_rate": 8.426254842367374e-05, "loss": 0.0258, "step": 1110 }, { "epoch": 1.6826291079812208, "grad_norm": 0.2006315290927887, "learning_rate": 8.387830646316623e-05, "loss": 0.0334, "step": 1120 }, { "epoch": 1.6976525821596244, "grad_norm": 0.24149750173091888, "learning_rate": 8.349032947115525e-05, "loss": 0.0249, "step": 1130 }, { "epoch": 1.712676056338028, "grad_norm": 0.30418938398361206, "learning_rate": 8.309866022151107e-05, "loss": 0.0382, "step": 1140 }, { "epoch": 1.727699530516432, "grad_norm": 0.18970516324043274, "learning_rate": 8.270334189516983e-05, "loss": 0.0261, "step": 1150 }, { "epoch": 1.727699530516432, "eval_loss": 0.03269151970744133, "eval_runtime": 220.3747, "eval_samples_per_second": 5.373, "eval_steps_per_second": 5.373, "step": 1150 }, { "epoch": 1.7427230046948357, "grad_norm": 0.11434226483106613, "learning_rate": 8.230441807537277e-05, "loss": 0.0393, "step": 1160 }, { "epoch": 1.7577464788732393, "grad_norm": 0.18536998331546783, "learning_rate": 8.190193274286122e-05, "loss": 0.026, "step": 1170 }, { "epoch": 1.7727699530516432, "grad_norm": 0.1809949427843094, "learning_rate": 8.149593027102789e-05, "loss": 0.0295, "step": 1180 }, { "epoch": 1.787793427230047, "grad_norm": 0.2668783664703369, "learning_rate": 8.108645542102469e-05, "loss": 0.0255, "step": 1190 }, { "epoch": 1.8028169014084507, "grad_norm": 0.19622312486171722, "learning_rate": 8.067355333682798e-05, "loss": 0.0297, "step": 1200 }, { "epoch": 1.8028169014084507, "eval_loss": 0.03482316806912422, "eval_runtime": 220.3204, "eval_samples_per_second": 5.374, "eval_steps_per_second": 5.374, "step": 1200 }, { "epoch": 1.8178403755868544, "grad_norm": 0.28984394669532776, "learning_rate": 8.025726954026138e-05, "loss": 0.0228, "step": 1210 }, { "epoch": 1.8328638497652583, "grad_norm": 0.2442556470632553, "learning_rate": 7.983764992597716e-05, "loss": 0.0272, "step": 1220 }, { "epoch": 1.847887323943662, "grad_norm": 0.33816081285476685, "learning_rate": 7.94147407563964e-05, "loss": 0.0296, "step": 1230 }, { "epoch": 1.8629107981220656, "grad_norm": 0.3621574938297272, "learning_rate": 7.89885886566086e-05, "loss": 0.0333, "step": 1240 }, { "epoch": 1.8779342723004695, "grad_norm": 0.27948886156082153, "learning_rate": 7.855924060923141e-05, "loss": 0.027, "step": 1250 }, { "epoch": 1.8779342723004695, "eval_loss": 0.03339933604001999, "eval_runtime": 220.8261, "eval_samples_per_second": 5.362, "eval_steps_per_second": 5.362, "step": 1250 }, { "epoch": 1.8929577464788734, "grad_norm": 0.19844475388526917, "learning_rate": 7.812674394923077e-05, "loss": 0.03, "step": 1260 }, { "epoch": 1.907981220657277, "grad_norm": 0.17690996825695038, "learning_rate": 7.769114635870231e-05, "loss": 0.0275, "step": 1270 }, { "epoch": 1.9230046948356807, "grad_norm": 0.26047441363334656, "learning_rate": 7.725249586161463e-05, "loss": 0.0294, "step": 1280 }, { "epoch": 1.9380281690140846, "grad_norm": 0.16973556578159332, "learning_rate": 7.68108408185145e-05, "loss": 0.031, "step": 1290 }, { "epoch": 1.9530516431924883, "grad_norm": 0.18663664162158966, "learning_rate": 7.636622992119536e-05, "loss": 0.0417, "step": 1300 }, { "epoch": 1.9530516431924883, "eval_loss": 0.03484388440847397, "eval_runtime": 221.5077, "eval_samples_per_second": 5.345, "eval_steps_per_second": 5.345, "step": 1300 }, { "epoch": 1.968075117370892, "grad_norm": 0.2905583083629608, "learning_rate": 7.591871218732902e-05, "loss": 0.0329, "step": 1310 }, { "epoch": 1.9830985915492958, "grad_norm": 0.18613983690738678, "learning_rate": 7.54683369550616e-05, "loss": 0.0364, "step": 1320 }, { "epoch": 1.9981220657276997, "grad_norm": 0.1276523619890213, "learning_rate": 7.501515387757404e-05, "loss": 0.0269, "step": 1330 }, { "epoch": 2.013145539906103, "grad_norm": 0.22376187145709991, "learning_rate": 7.455921291760796e-05, "loss": 0.0313, "step": 1340 }, { "epoch": 2.028169014084507, "grad_norm": 0.09134189784526825, "learning_rate": 7.410056434195725e-05, "loss": 0.0173, "step": 1350 }, { "epoch": 2.028169014084507, "eval_loss": 0.03278001397848129, "eval_runtime": 221.4183, "eval_samples_per_second": 5.347, "eval_steps_per_second": 5.347, "step": 1350 }, { "epoch": 2.043192488262911, "grad_norm": 0.13673311471939087, "learning_rate": 7.363925871592629e-05, "loss": 0.0205, "step": 1360 }, { "epoch": 2.0582159624413148, "grad_norm": 0.161691814661026, "learning_rate": 7.317534689775528e-05, "loss": 0.0201, "step": 1370 }, { "epoch": 2.073239436619718, "grad_norm": 0.1258237212896347, "learning_rate": 7.270888003301304e-05, "loss": 0.0159, "step": 1380 }, { "epoch": 2.088262910798122, "grad_norm": 0.29005277156829834, "learning_rate": 7.22399095489584e-05, "loss": 0.0222, "step": 1390 }, { "epoch": 2.103286384976526, "grad_norm": 0.11337249726057053, "learning_rate": 7.176848714887042e-05, "loss": 0.0207, "step": 1400 }, { "epoch": 2.103286384976526, "eval_loss": 0.03226166218519211, "eval_runtime": 221.8527, "eval_samples_per_second": 5.337, "eval_steps_per_second": 5.337, "step": 1400 }, { "epoch": 2.1183098591549294, "grad_norm": 0.16560769081115723, "learning_rate": 7.129466480634806e-05, "loss": 0.0225, "step": 1410 }, { "epoch": 2.1333333333333333, "grad_norm": 0.3505042493343353, "learning_rate": 7.081849475958042e-05, "loss": 0.0187, "step": 1420 }, { "epoch": 2.148356807511737, "grad_norm": 0.16660155355930328, "learning_rate": 7.034002950558723e-05, "loss": 0.0185, "step": 1430 }, { "epoch": 2.1633802816901406, "grad_norm": 0.13852401077747345, "learning_rate": 6.985932179443144e-05, "loss": 0.0192, "step": 1440 }, { "epoch": 2.1784037558685445, "grad_norm": 0.21416670083999634, "learning_rate": 6.937642462340342e-05, "loss": 0.0223, "step": 1450 }, { "epoch": 2.1784037558685445, "eval_loss": 0.03254189342260361, "eval_runtime": 222.3283, "eval_samples_per_second": 5.325, "eval_steps_per_second": 5.325, "step": 1450 }, { "epoch": 2.1934272300469484, "grad_norm": 0.1465296745300293, "learning_rate": 6.889139123117817e-05, "loss": 0.0191, "step": 1460 }, { "epoch": 2.2084507042253523, "grad_norm": 0.19770440459251404, "learning_rate": 6.840427509194575e-05, "loss": 0.0209, "step": 1470 }, { "epoch": 2.2234741784037557, "grad_norm": 0.22087684273719788, "learning_rate": 6.791512990951597e-05, "loss": 0.0149, "step": 1480 }, { "epoch": 2.2384976525821596, "grad_norm": 0.34373584389686584, "learning_rate": 6.74240096113975e-05, "loss": 0.0211, "step": 1490 }, { "epoch": 2.2535211267605635, "grad_norm": 0.12940266728401184, "learning_rate": 6.693096834285256e-05, "loss": 0.0107, "step": 1500 }, { "epoch": 2.2535211267605635, "eval_loss": 0.035872478038072586, "eval_runtime": 223.0953, "eval_samples_per_second": 5.307, "eval_steps_per_second": 5.307, "step": 1500 }, { "epoch": 2.2685446009389674, "grad_norm": 0.39470982551574707, "learning_rate": 6.643606046092732e-05, "loss": 0.0226, "step": 1510 }, { "epoch": 2.283568075117371, "grad_norm": 0.22834685444831848, "learning_rate": 6.593934052845929e-05, "loss": 0.0253, "step": 1520 }, { "epoch": 2.2985915492957747, "grad_norm": 0.41707947850227356, "learning_rate": 6.544086330806181e-05, "loss": 0.0205, "step": 1530 }, { "epoch": 2.3136150234741786, "grad_norm": 0.1707378327846527, "learning_rate": 6.494068375608646e-05, "loss": 0.0197, "step": 1540 }, { "epoch": 2.328638497652582, "grad_norm": 0.17165277898311615, "learning_rate": 6.443885701656432e-05, "loss": 0.0182, "step": 1550 }, { "epoch": 2.328638497652582, "eval_loss": 0.03318268060684204, "eval_runtime": 223.282, "eval_samples_per_second": 5.303, "eval_steps_per_second": 5.303, "step": 1550 }, { "epoch": 2.343661971830986, "grad_norm": 0.14184924960136414, "learning_rate": 6.393543841512632e-05, "loss": 0.02, "step": 1560 }, { "epoch": 2.35868544600939, "grad_norm": 0.31733861565589905, "learning_rate": 6.343048345290386e-05, "loss": 0.0144, "step": 1570 }, { "epoch": 2.3737089201877932, "grad_norm": 0.16232426464557648, "learning_rate": 6.292404780040961e-05, "loss": 0.023, "step": 1580 }, { "epoch": 2.388732394366197, "grad_norm": 0.18042533099651337, "learning_rate": 6.241618729140018e-05, "loss": 0.0157, "step": 1590 }, { "epoch": 2.403755868544601, "grad_norm": 0.4602644741535187, "learning_rate": 6.190695791672042e-05, "loss": 0.0187, "step": 1600 }, { "epoch": 2.403755868544601, "eval_loss": 0.032255951315164566, "eval_runtime": 222.5999, "eval_samples_per_second": 5.319, "eval_steps_per_second": 5.319, "step": 1600 }, { "epoch": 2.418779342723005, "grad_norm": 0.0981522649526596, "learning_rate": 6.139641581813052e-05, "loss": 0.0183, "step": 1610 }, { "epoch": 2.4338028169014083, "grad_norm": 0.1687208116054535, "learning_rate": 6.088461728211642e-05, "loss": 0.0177, "step": 1620 }, { "epoch": 2.448826291079812, "grad_norm": 0.19458502531051636, "learning_rate": 6.0371618733684474e-05, "loss": 0.0117, "step": 1630 }, { "epoch": 2.463849765258216, "grad_norm": 0.12449292838573456, "learning_rate": 5.9857476730140485e-05, "loss": 0.0175, "step": 1640 }, { "epoch": 2.4788732394366195, "grad_norm": 0.20214103162288666, "learning_rate": 5.9342247954854466e-05, "loss": 0.018, "step": 1650 }, { "epoch": 2.4788732394366195, "eval_loss": 0.03270378336310387, "eval_runtime": 221.3125, "eval_samples_per_second": 5.35, "eval_steps_per_second": 5.35, "step": 1650 }, { "epoch": 2.4938967136150234, "grad_norm": 0.5592941641807556, "learning_rate": 5.8825989211011335e-05, "loss": 0.022, "step": 1660 }, { "epoch": 2.5089201877934273, "grad_norm": 0.20930063724517822, "learning_rate": 5.830875741534852e-05, "loss": 0.0159, "step": 1670 }, { "epoch": 2.523943661971831, "grad_norm": 0.24804644286632538, "learning_rate": 5.7790609591880826e-05, "loss": 0.0239, "step": 1680 }, { "epoch": 2.5389671361502346, "grad_norm": 0.17634353041648865, "learning_rate": 5.727160286561386e-05, "loss": 0.0224, "step": 1690 }, { "epoch": 2.5539906103286385, "grad_norm": 0.23160183429718018, "learning_rate": 5.675179445624581e-05, "loss": 0.0205, "step": 1700 }, { "epoch": 2.5539906103286385, "eval_loss": 0.03496889770030975, "eval_runtime": 221.4535, "eval_samples_per_second": 5.346, "eval_steps_per_second": 5.346, "step": 1700 }, { "epoch": 2.5690140845070424, "grad_norm": 0.09332422912120819, "learning_rate": 5.62312416718593e-05, "loss": 0.0164, "step": 1710 }, { "epoch": 2.584037558685446, "grad_norm": 0.09970489144325256, "learning_rate": 5.5710001902603116e-05, "loss": 0.0257, "step": 1720 }, { "epoch": 2.5990610328638497, "grad_norm": 0.17599700391292572, "learning_rate": 5.5188132614365094e-05, "loss": 0.0198, "step": 1730 }, { "epoch": 2.6140845070422536, "grad_norm": 0.11823566257953644, "learning_rate": 5.4665691342436565e-05, "loss": 0.0152, "step": 1740 }, { "epoch": 2.629107981220657, "grad_norm": 0.31226423382759094, "learning_rate": 5.414273568516919e-05, "loss": 0.0182, "step": 1750 }, { "epoch": 2.629107981220657, "eval_loss": 0.032272376120090485, "eval_runtime": 221.7599, "eval_samples_per_second": 5.339, "eval_steps_per_second": 5.339, "step": 1750 }, { "epoch": 2.644131455399061, "grad_norm": 0.12536101043224335, "learning_rate": 5.361932329762481e-05, "loss": 0.0197, "step": 1760 }, { "epoch": 2.659154929577465, "grad_norm": 0.11700065433979034, "learning_rate": 5.309551188521914e-05, "loss": 0.0197, "step": 1770 }, { "epoch": 2.6741784037558687, "grad_norm": 0.30704525113105774, "learning_rate": 5.2571359197359704e-05, "loss": 0.0161, "step": 1780 }, { "epoch": 2.6892018779342726, "grad_norm": 0.06998223811388016, "learning_rate": 5.2046923021079175e-05, "loss": 0.0151, "step": 1790 }, { "epoch": 2.704225352112676, "grad_norm": 0.09223470091819763, "learning_rate": 5.1522261174664346e-05, "loss": 0.0202, "step": 1800 }, { "epoch": 2.704225352112676, "eval_loss": 0.032467469573020935, "eval_runtime": 221.8494, "eval_samples_per_second": 5.337, "eval_steps_per_second": 5.337, "step": 1800 }, { "epoch": 2.71924882629108, "grad_norm": 0.3536123037338257, "learning_rate": 5.0997431501281835e-05, "loss": 0.0262, "step": 1810 }, { "epoch": 2.7342723004694838, "grad_norm": 0.2920973300933838, "learning_rate": 5.0472491862600915e-05, "loss": 0.0202, "step": 1820 }, { "epoch": 2.749295774647887, "grad_norm": 0.23328785598278046, "learning_rate": 4.994750013241435e-05, "loss": 0.022, "step": 1830 }, { "epoch": 2.764319248826291, "grad_norm": 0.13965994119644165, "learning_rate": 4.9422514190257974e-05, "loss": 0.0201, "step": 1840 }, { "epoch": 2.779342723004695, "grad_norm": 0.12157478928565979, "learning_rate": 4.88975919150294e-05, "loss": 0.0218, "step": 1850 }, { "epoch": 2.779342723004695, "eval_loss": 0.03232384845614433, "eval_runtime": 221.306, "eval_samples_per_second": 5.35, "eval_steps_per_second": 5.35, "step": 1850 }, { "epoch": 2.7943661971830984, "grad_norm": 0.17367467284202576, "learning_rate": 4.83727911786071e-05, "loss": 0.022, "step": 1860 }, { "epoch": 2.8093896713615023, "grad_norm": 0.30343097448349, "learning_rate": 4.7848169839470145e-05, "loss": 0.0224, "step": 1870 }, { "epoch": 2.824413145539906, "grad_norm": 0.12888950109481812, "learning_rate": 4.7323785736319244e-05, "loss": 0.021, "step": 1880 }, { "epoch": 2.8394366197183096, "grad_norm": 0.15008972585201263, "learning_rate": 4.679969668170024e-05, "loss": 0.0232, "step": 1890 }, { "epoch": 2.8544600938967135, "grad_norm": 0.10453104227781296, "learning_rate": 4.627596045563031e-05, "loss": 0.0179, "step": 1900 }, { "epoch": 2.8544600938967135, "eval_loss": 0.03188331797719002, "eval_runtime": 221.3325, "eval_samples_per_second": 5.349, "eval_steps_per_second": 5.349, "step": 1900 }, { "epoch": 2.8694835680751174, "grad_norm": 0.20146887004375458, "learning_rate": 4.575263479922783e-05, "loss": 0.0147, "step": 1910 }, { "epoch": 2.8845070422535213, "grad_norm": 0.10789839178323746, "learning_rate": 4.522977740834651e-05, "loss": 0.0118, "step": 1920 }, { "epoch": 2.8995305164319247, "grad_norm": 0.1628926396369934, "learning_rate": 4.4707445927214456e-05, "loss": 0.0126, "step": 1930 }, { "epoch": 2.9145539906103286, "grad_norm": 0.14242199063301086, "learning_rate": 4.4185697942079115e-05, "loss": 0.0199, "step": 1940 }, { "epoch": 2.9295774647887325, "grad_norm": 0.15721061825752258, "learning_rate": 4.366459097485832e-05, "loss": 0.0213, "step": 1950 }, { "epoch": 2.9295774647887325, "eval_loss": 0.03299366310238838, "eval_runtime": 221.4895, "eval_samples_per_second": 5.346, "eval_steps_per_second": 5.346, "step": 1950 }, { "epoch": 2.9446009389671364, "grad_norm": 0.2670339047908783, "learning_rate": 4.314418247679866e-05, "loss": 0.0231, "step": 1960 }, { "epoch": 2.95962441314554, "grad_norm": 0.18777424097061157, "learning_rate": 4.26245298221416e-05, "loss": 0.0141, "step": 1970 }, { "epoch": 2.9746478873239437, "grad_norm": 0.08959437906742096, "learning_rate": 4.2105690301798014e-05, "loss": 0.0143, "step": 1980 }, { "epoch": 2.9896713615023476, "grad_norm": 0.23774972558021545, "learning_rate": 4.158772111703194e-05, "loss": 0.0167, "step": 1990 }, { "epoch": 3.004694835680751, "grad_norm": 0.025070570409297943, "learning_rate": 4.107067937315429e-05, "loss": 0.0104, "step": 2000 }, { "epoch": 3.004694835680751, "eval_loss": 0.03279737010598183, "eval_runtime": 222.0291, "eval_samples_per_second": 5.333, "eval_steps_per_second": 5.333, "step": 2000 }, { "epoch": 3.019718309859155, "grad_norm": 0.12388351559638977, "learning_rate": 4.055462207322698e-05, "loss": 0.01, "step": 2010 }, { "epoch": 3.034741784037559, "grad_norm": 0.024688435718417168, "learning_rate": 4.003960611177855e-05, "loss": 0.0101, "step": 2020 }, { "epoch": 3.0497652582159622, "grad_norm": 0.24434754252433777, "learning_rate": 3.952568826853152e-05, "loss": 0.0094, "step": 2030 }, { "epoch": 3.064788732394366, "grad_norm": 0.05700072646141052, "learning_rate": 3.901292520214256e-05, "loss": 0.0068, "step": 2040 }, { "epoch": 3.07981220657277, "grad_norm": 0.6166403889656067, "learning_rate": 3.850137344395598e-05, "loss": 0.0097, "step": 2050 }, { "epoch": 3.07981220657277, "eval_loss": 0.035948336124420166, "eval_runtime": 221.527, "eval_samples_per_second": 5.345, "eval_steps_per_second": 5.345, "step": 2050 }, { "epoch": 3.094835680751174, "grad_norm": 0.2868582308292389, "learning_rate": 3.799108939177118e-05, "loss": 0.0076, "step": 2060 }, { "epoch": 3.1098591549295773, "grad_norm": 0.201212540268898, "learning_rate": 3.7482129303624934e-05, "loss": 0.0148, "step": 2070 }, { "epoch": 3.124882629107981, "grad_norm": 0.512485921382904, "learning_rate": 3.697454929158901e-05, "loss": 0.0115, "step": 2080 }, { "epoch": 3.139906103286385, "grad_norm": 0.23441621661186218, "learning_rate": 3.6468405315583854e-05, "loss": 0.0147, "step": 2090 }, { "epoch": 3.1549295774647885, "grad_norm": 0.20293329656124115, "learning_rate": 3.59637531772092e-05, "loss": 0.0103, "step": 2100 }, { "epoch": 3.1549295774647885, "eval_loss": 0.0363353006541729, "eval_runtime": 221.0841, "eval_samples_per_second": 5.355, "eval_steps_per_second": 5.355, "step": 2100 }, { "epoch": 3.1699530516431924, "grad_norm": 0.2901355028152466, "learning_rate": 3.546064851359192e-05, "loss": 0.0082, "step": 2110 }, { "epoch": 3.1849765258215963, "grad_norm": 0.11046724766492844, "learning_rate": 3.495914679125212e-05, "loss": 0.0097, "step": 2120 }, { "epoch": 3.2, "grad_norm": 0.07721813023090363, "learning_rate": 3.445930329998819e-05, "loss": 0.0103, "step": 2130 }, { "epoch": 3.2150234741784036, "grad_norm": 0.26007962226867676, "learning_rate": 3.396117314678097e-05, "loss": 0.0142, "step": 2140 }, { "epoch": 3.2300469483568075, "grad_norm": 0.4068683087825775, "learning_rate": 3.3464811249718474e-05, "loss": 0.0131, "step": 2150 }, { "epoch": 3.2300469483568075, "eval_loss": 0.03587425872683525, "eval_runtime": 221.6748, "eval_samples_per_second": 5.341, "eval_steps_per_second": 5.341, "step": 2150 }, { "epoch": 3.2450704225352114, "grad_norm": 0.18533091247081757, "learning_rate": 3.297027233194114e-05, "loss": 0.0126, "step": 2160 }, { "epoch": 3.260093896713615, "grad_norm": 0.10798130184412003, "learning_rate": 3.2477610915608704e-05, "loss": 0.0081, "step": 2170 }, { "epoch": 3.2751173708920187, "grad_norm": 0.19456113874912262, "learning_rate": 3.1986881315889315e-05, "loss": 0.0091, "step": 2180 }, { "epoch": 3.2901408450704226, "grad_norm": 0.26297393441200256, "learning_rate": 3.149813763497124e-05, "loss": 0.0155, "step": 2190 }, { "epoch": 3.3051643192488265, "grad_norm": 0.26686403155326843, "learning_rate": 3.101143375609818e-05, "loss": 0.0149, "step": 2200 }, { "epoch": 3.3051643192488265, "eval_loss": 0.036220405250787735, "eval_runtime": 221.9816, "eval_samples_per_second": 5.334, "eval_steps_per_second": 5.334, "step": 2200 }, { "epoch": 3.32018779342723, "grad_norm": 0.3246915638446808, "learning_rate": 3.0526823337628915e-05, "loss": 0.0129, "step": 2210 }, { "epoch": 3.335211267605634, "grad_norm": 0.4109603464603424, "learning_rate": 3.004435980712129e-05, "loss": 0.0098, "step": 2220 }, { "epoch": 3.3502347417840377, "grad_norm": 0.4035777747631073, "learning_rate": 2.9564096355442116e-05, "loss": 0.0128, "step": 2230 }, { "epoch": 3.365258215962441, "grad_norm": 0.1403612494468689, "learning_rate": 2.9086085930902824e-05, "loss": 0.0137, "step": 2240 }, { "epoch": 3.380281690140845, "grad_norm": 0.21675720810890198, "learning_rate": 2.8610381233422058e-05, "loss": 0.0083, "step": 2250 }, { "epoch": 3.380281690140845, "eval_loss": 0.03651123493909836, "eval_runtime": 222.0068, "eval_samples_per_second": 5.333, "eval_steps_per_second": 5.333, "step": 2250 }, { "epoch": 3.395305164319249, "grad_norm": 0.30155134201049805, "learning_rate": 2.8137034708715592e-05, "loss": 0.0098, "step": 2260 }, { "epoch": 3.4103286384976528, "grad_norm": 0.05509026348590851, "learning_rate": 2.7666098542514273e-05, "loss": 0.0096, "step": 2270 }, { "epoch": 3.425352112676056, "grad_norm": 0.25339803099632263, "learning_rate": 2.719762465481055e-05, "loss": 0.0123, "step": 2280 }, { "epoch": 3.44037558685446, "grad_norm": 0.052862461656332016, "learning_rate": 2.6731664694134473e-05, "loss": 0.0118, "step": 2290 }, { "epoch": 3.455399061032864, "grad_norm": 0.16749408841133118, "learning_rate": 2.6268270031859476e-05, "loss": 0.0115, "step": 2300 }, { "epoch": 3.455399061032864, "eval_loss": 0.035897351801395416, "eval_runtime": 221.6066, "eval_samples_per_second": 5.343, "eval_steps_per_second": 5.343, "step": 2300 }, { "epoch": 3.4704225352112674, "grad_norm": 0.28338125348091125, "learning_rate": 2.580749175653877e-05, "loss": 0.0097, "step": 2310 }, { "epoch": 3.4854460093896713, "grad_norm": 0.36096012592315674, "learning_rate": 2.5349380668272905e-05, "loss": 0.0122, "step": 2320 }, { "epoch": 3.500469483568075, "grad_norm": 0.20076704025268555, "learning_rate": 2.489398727310908e-05, "loss": 0.0096, "step": 2330 }, { "epoch": 3.5154929577464786, "grad_norm": 0.28074732422828674, "learning_rate": 2.4441361777473066e-05, "loss": 0.0088, "step": 2340 }, { "epoch": 3.5305164319248825, "grad_norm": 0.5027658343315125, "learning_rate": 2.3991554082633912e-05, "loss": 0.0111, "step": 2350 }, { "epoch": 3.5305164319248825, "eval_loss": 0.038713712245225906, "eval_runtime": 221.2043, "eval_samples_per_second": 5.353, "eval_steps_per_second": 5.353, "step": 2350 }, { "epoch": 3.5455399061032864, "grad_norm": 0.11489449441432953, "learning_rate": 2.354461377920239e-05, "loss": 0.009, "step": 2360 }, { "epoch": 3.5605633802816903, "grad_norm": 0.42795881628990173, "learning_rate": 2.3100590141663807e-05, "loss": 0.0123, "step": 2370 }, { "epoch": 3.575586854460094, "grad_norm": 0.34739652276039124, "learning_rate": 2.265953212294551e-05, "loss": 0.0143, "step": 2380 }, { "epoch": 3.5906103286384976, "grad_norm": 0.5267793536186218, "learning_rate": 2.2221488349019903e-05, "loss": 0.015, "step": 2390 }, { "epoch": 3.6056338028169015, "grad_norm": 0.2632468342781067, "learning_rate": 2.1786507113543457e-05, "loss": 0.0094, "step": 2400 }, { "epoch": 3.6056338028169015, "eval_loss": 0.03762707859277725, "eval_runtime": 222.4043, "eval_samples_per_second": 5.324, "eval_steps_per_second": 5.324, "step": 2400 }, { "epoch": 3.6206572769953054, "grad_norm": 0.15667402744293213, "learning_rate": 2.1354636372532523e-05, "loss": 0.0149, "step": 2410 }, { "epoch": 3.635680751173709, "grad_norm": 0.22469666600227356, "learning_rate": 2.092592373907617e-05, "loss": 0.0106, "step": 2420 }, { "epoch": 3.6507042253521127, "grad_norm": 0.2074466347694397, "learning_rate": 2.0500416478086932e-05, "loss": 0.0079, "step": 2430 }, { "epoch": 3.6657276995305166, "grad_norm": 0.21211783587932587, "learning_rate": 2.0078161501089954e-05, "loss": 0.0106, "step": 2440 }, { "epoch": 3.68075117370892, "grad_norm": 0.13224627077579498, "learning_rate": 1.9659205361050982e-05, "loss": 0.0051, "step": 2450 }, { "epoch": 3.68075117370892, "eval_loss": 0.03760899230837822, "eval_runtime": 222.6155, "eval_samples_per_second": 5.319, "eval_steps_per_second": 5.319, "step": 2450 }, { "epoch": 3.695774647887324, "grad_norm": 0.07832111418247223, "learning_rate": 1.924359424724408e-05, "loss": 0.0143, "step": 2460 }, { "epoch": 3.710798122065728, "grad_norm": 0.34279316663742065, "learning_rate": 1.8831373980159296e-05, "loss": 0.0129, "step": 2470 }, { "epoch": 3.7258215962441312, "grad_norm": 0.13393940031528473, "learning_rate": 1.8422590006450947e-05, "loss": 0.0062, "step": 2480 }, { "epoch": 3.740845070422535, "grad_norm": 0.05350172892212868, "learning_rate": 1.801728739392731e-05, "loss": 0.0087, "step": 2490 }, { "epoch": 3.755868544600939, "grad_norm": 0.0817003846168518, "learning_rate": 1.7615510826581904e-05, "loss": 0.0053, "step": 2500 }, { "epoch": 3.755868544600939, "eval_loss": 0.03752273693680763, "eval_runtime": 222.372, "eval_samples_per_second": 5.324, "eval_steps_per_second": 5.324, "step": 2500 }, { "epoch": 3.770892018779343, "grad_norm": 0.21435320377349854, "learning_rate": 1.7217304599667146e-05, "loss": 0.0111, "step": 2510 }, { "epoch": 3.7859154929577463, "grad_norm": 0.15997354686260223, "learning_rate": 1.6822712614810893e-05, "loss": 0.0101, "step": 2520 }, { "epoch": 3.80093896713615, "grad_norm": 0.182434543967247, "learning_rate": 1.643177837517631e-05, "loss": 0.0093, "step": 2530 }, { "epoch": 3.815962441314554, "grad_norm": 0.22743846476078033, "learning_rate": 1.6044544980665767e-05, "loss": 0.0125, "step": 2540 }, { "epoch": 3.830985915492958, "grad_norm": 0.33443447947502136, "learning_rate": 1.5661055123169126e-05, "loss": 0.0078, "step": 2550 }, { "epoch": 3.830985915492958, "eval_loss": 0.037725865840911865, "eval_runtime": 221.8726, "eval_samples_per_second": 5.336, "eval_steps_per_second": 5.336, "step": 2550 }, { "epoch": 3.8460093896713614, "grad_norm": 0.5252001285552979, "learning_rate": 1.5281351081856974e-05, "loss": 0.0118, "step": 2560 }, { "epoch": 3.8610328638497653, "grad_norm": 0.17551793158054352, "learning_rate": 1.4905474718519491e-05, "loss": 0.007, "step": 2570 }, { "epoch": 3.876056338028169, "grad_norm": 0.1480369120836258, "learning_rate": 1.453346747295119e-05, "loss": 0.0141, "step": 2580 }, { "epoch": 3.8910798122065726, "grad_norm": 0.19130997359752655, "learning_rate": 1.4165370358382274e-05, "loss": 0.0093, "step": 2590 }, { "epoch": 3.9061032863849765, "grad_norm": 0.1769220381975174, "learning_rate": 1.3801223956956994e-05, "loss": 0.0105, "step": 2600 }, { "epoch": 3.9061032863849765, "eval_loss": 0.0372486487030983, "eval_runtime": 221.9339, "eval_samples_per_second": 5.335, "eval_steps_per_second": 5.335, "step": 2600 }, { "epoch": 3.9211267605633804, "grad_norm": 0.11678332090377808, "learning_rate": 1.344106841525946e-05, "loss": 0.0141, "step": 2610 }, { "epoch": 3.936150234741784, "grad_norm": 0.24193385243415833, "learning_rate": 1.3084943439887659e-05, "loss": 0.0085, "step": 2620 }, { "epoch": 3.9511737089201877, "grad_norm": 0.10747163742780685, "learning_rate": 1.273288829307579e-05, "loss": 0.0054, "step": 2630 }, { "epoch": 3.9661971830985916, "grad_norm": 0.28762826323509216, "learning_rate": 1.2384941788365622e-05, "loss": 0.0078, "step": 2640 }, { "epoch": 3.981220657276995, "grad_norm": 0.11588622629642487, "learning_rate": 1.2041142286327477e-05, "loss": 0.0105, "step": 2650 }, { "epoch": 3.981220657276995, "eval_loss": 0.03705061599612236, "eval_runtime": 222.758, "eval_samples_per_second": 5.315, "eval_steps_per_second": 5.315, "step": 2650 }, { "epoch": 3.996244131455399, "grad_norm": 0.2062222808599472, "learning_rate": 1.170152769033095e-05, "loss": 0.0085, "step": 2660 }, { "epoch": 4.011267605633803, "grad_norm": 0.15399198234081268, "learning_rate": 1.1366135442366127e-05, "loss": 0.0057, "step": 2670 }, { "epoch": 4.026291079812206, "grad_norm": 0.12908077239990234, "learning_rate": 1.103500251891571e-05, "loss": 0.0074, "step": 2680 }, { "epoch": 4.041314553990611, "grad_norm": 0.06553179025650024, "learning_rate": 1.0708165426878325e-05, "loss": 0.0049, "step": 2690 }, { "epoch": 4.056338028169014, "grad_norm": 0.12032724171876907, "learning_rate": 1.0385660199543812e-05, "loss": 0.0064, "step": 2700 }, { "epoch": 4.056338028169014, "eval_loss": 0.038154859095811844, "eval_runtime": 220.5195, "eval_samples_per_second": 5.369, "eval_steps_per_second": 5.369, "step": 2700 }, { "epoch": 4.0713615023474174, "grad_norm": 0.06930918991565704, "learning_rate": 1.0067522392620537e-05, "loss": 0.0052, "step": 2710 }, { "epoch": 4.086384976525822, "grad_norm": 0.10133420675992966, "learning_rate": 9.753787080315385e-06, "loss": 0.0075, "step": 2720 }, { "epoch": 4.101408450704225, "grad_norm": 0.01755693554878235, "learning_rate": 9.444488851467042e-06, "loss": 0.0059, "step": 2730 }, { "epoch": 4.1164319248826295, "grad_norm": 0.1588931679725647, "learning_rate": 9.139661805732435e-06, "loss": 0.0042, "step": 2740 }, { "epoch": 4.131455399061033, "grad_norm": 0.39881432056427, "learning_rate": 8.839339549827397e-06, "loss": 0.0048, "step": 2750 }, { "epoch": 4.131455399061033, "eval_loss": 0.03982308879494667, "eval_runtime": 221.9797, "eval_samples_per_second": 5.334, "eval_steps_per_second": 5.334, "step": 2750 }, { "epoch": 4.146478873239436, "grad_norm": 0.07863107323646545, "learning_rate": 8.543555193821634e-06, "loss": 0.0075, "step": 2760 }, { "epoch": 4.161502347417841, "grad_norm": 0.25240081548690796, "learning_rate": 8.252341347488251e-06, "loss": 0.0043, "step": 2770 }, { "epoch": 4.176525821596244, "grad_norm": 0.22023944556713104, "learning_rate": 7.965730116708681e-06, "loss": 0.008, "step": 2780 }, { "epoch": 4.191549295774648, "grad_norm": 0.010785081423819065, "learning_rate": 7.68375309993304e-06, "loss": 0.0031, "step": 2790 }, { "epoch": 4.206572769953052, "grad_norm": 0.11607765406370163, "learning_rate": 7.406441384696372e-06, "loss": 0.0065, "step": 2800 }, { "epoch": 4.206572769953052, "eval_loss": 0.04066622629761696, "eval_runtime": 221.9129, "eval_samples_per_second": 5.335, "eval_steps_per_second": 5.335, "step": 2800 }, { "epoch": 4.221596244131455, "grad_norm": 0.05429501459002495, "learning_rate": 7.133825544191464e-06, "loss": 0.0095, "step": 2810 }, { "epoch": 4.236619718309859, "grad_norm": 0.09971319139003754, "learning_rate": 6.865935633897996e-06, "loss": 0.004, "step": 2820 }, { "epoch": 4.251643192488263, "grad_norm": 0.09597835689783096, "learning_rate": 6.602801188269081e-06, "loss": 0.0064, "step": 2830 }, { "epoch": 4.266666666666667, "grad_norm": 0.1952216774225235, "learning_rate": 6.344451217475183e-06, "loss": 0.0047, "step": 2840 }, { "epoch": 4.28169014084507, "grad_norm": 0.0561475045979023, "learning_rate": 6.090914204205655e-06, "loss": 0.0031, "step": 2850 }, { "epoch": 4.28169014084507, "eval_loss": 0.04169252887368202, "eval_runtime": 222.1513, "eval_samples_per_second": 5.33, "eval_steps_per_second": 5.33, "step": 2850 }, { "epoch": 4.296713615023474, "grad_norm": 0.2749840021133423, "learning_rate": 5.842218100528679e-06, "loss": 0.0052, "step": 2860 }, { "epoch": 4.311737089201878, "grad_norm": 0.22904397547245026, "learning_rate": 5.598390324809555e-06, "loss": 0.0046, "step": 2870 }, { "epoch": 4.326760563380281, "grad_norm": 0.26261669397354126, "learning_rate": 5.359457758687841e-06, "loss": 0.0089, "step": 2880 }, { "epoch": 4.341784037558686, "grad_norm": 0.21348492801189423, "learning_rate": 5.125446744113743e-06, "loss": 0.0057, "step": 2890 }, { "epoch": 4.356807511737089, "grad_norm": 0.0984061136841774, "learning_rate": 4.896383080443934e-06, "loss": 0.0028, "step": 2900 }, { "epoch": 4.356807511737089, "eval_loss": 0.041960135102272034, "eval_runtime": 222.6261, "eval_samples_per_second": 5.318, "eval_steps_per_second": 5.318, "step": 2900 }, { "epoch": 4.371830985915493, "grad_norm": 0.2619991898536682, "learning_rate": 4.672292021597174e-06, "loss": 0.0056, "step": 2910 }, { "epoch": 4.386854460093897, "grad_norm": 0.16088174283504486, "learning_rate": 4.4531982732702145e-06, "loss": 0.0064, "step": 2920 }, { "epoch": 4.4018779342723, "grad_norm": 0.3044176995754242, "learning_rate": 4.239125990213883e-06, "loss": 0.0039, "step": 2930 }, { "epoch": 4.416901408450705, "grad_norm": 0.06394259631633759, "learning_rate": 4.030098773570174e-06, "loss": 0.0016, "step": 2940 }, { "epoch": 4.431924882629108, "grad_norm": 0.15494537353515625, "learning_rate": 3.826139668270234e-06, "loss": 0.0043, "step": 2950 }, { "epoch": 4.431924882629108, "eval_loss": 0.042107198387384415, "eval_runtime": 222.671, "eval_samples_per_second": 5.317, "eval_steps_per_second": 5.317, "step": 2950 }, { "epoch": 4.446948356807511, "grad_norm": 0.1741316169500351, "learning_rate": 3.6272711604936504e-06, "loss": 0.0049, "step": 2960 }, { "epoch": 4.461971830985916, "grad_norm": 0.08207106590270996, "learning_rate": 3.433515175189428e-06, "loss": 0.0053, "step": 2970 }, { "epoch": 4.476995305164319, "grad_norm": 0.09723722189664841, "learning_rate": 3.2448930736588e-06, "loss": 0.0077, "step": 2980 }, { "epoch": 4.492018779342723, "grad_norm": 0.03407386317849159, "learning_rate": 3.061425651200117e-06, "loss": 0.0038, "step": 2990 }, { "epoch": 4.507042253521127, "grad_norm": 0.21065545082092285, "learning_rate": 2.883133134816296e-06, "loss": 0.0048, "step": 3000 }, { "epoch": 4.507042253521127, "eval_loss": 0.04240112751722336, "eval_runtime": 222.0883, "eval_samples_per_second": 5.331, "eval_steps_per_second": 5.331, "step": 3000 }, { "epoch": 4.52206572769953, "grad_norm": 0.06989271938800812, "learning_rate": 2.7100351809847326e-06, "loss": 0.0046, "step": 3010 }, { "epoch": 4.537089201877935, "grad_norm": 0.07983122766017914, "learning_rate": 2.542150873490251e-06, "loss": 0.0028, "step": 3020 }, { "epoch": 4.552112676056338, "grad_norm": 0.27960672974586487, "learning_rate": 2.3794987213211383e-06, "loss": 0.0055, "step": 3030 }, { "epoch": 4.567136150234742, "grad_norm": 0.2628881633281708, "learning_rate": 2.222096656628547e-06, "loss": 0.0035, "step": 3040 }, { "epoch": 4.582159624413146, "grad_norm": 0.13972419500350952, "learning_rate": 2.0699620327495174e-06, "loss": 0.0038, "step": 3050 }, { "epoch": 4.582159624413146, "eval_loss": 0.04280637204647064, "eval_runtime": 222.2913, "eval_samples_per_second": 5.326, "eval_steps_per_second": 5.326, "step": 3050 }, { "epoch": 4.597183098591549, "grad_norm": 0.017372630536556244, "learning_rate": 1.9231116222937996e-06, "loss": 0.0025, "step": 3060 }, { "epoch": 4.612206572769953, "grad_norm": 0.14479027688503265, "learning_rate": 1.7815616152946523e-06, "loss": 0.0045, "step": 3070 }, { "epoch": 4.627230046948357, "grad_norm": 0.03770596161484718, "learning_rate": 1.6453276174240195e-06, "loss": 0.004, "step": 3080 }, { "epoch": 4.642253521126761, "grad_norm": 0.19395627081394196, "learning_rate": 1.5144246482719114e-06, "loss": 0.0043, "step": 3090 }, { "epoch": 4.657276995305164, "grad_norm": 0.05098709091544151, "learning_rate": 1.3888671396905805e-06, "loss": 0.0041, "step": 3100 }, { "epoch": 4.657276995305164, "eval_loss": 0.0430205836892128, "eval_runtime": 222.6399, "eval_samples_per_second": 5.318, "eval_steps_per_second": 5.318, "step": 3100 }, { "epoch": 4.672300469483568, "grad_norm": 0.3063965439796448, "learning_rate": 1.2686689342034431e-06, "loss": 0.0082, "step": 3110 }, { "epoch": 4.687323943661972, "grad_norm": 0.21486127376556396, "learning_rate": 1.1538432834789227e-06, "loss": 0.0054, "step": 3120 }, { "epoch": 4.702347417840375, "grad_norm": 0.3423399329185486, "learning_rate": 1.044402846869491e-06, "loss": 0.0049, "step": 3130 }, { "epoch": 4.71737089201878, "grad_norm": 0.15926554799079895, "learning_rate": 9.403596900160073e-07, "loss": 0.0034, "step": 3140 }, { "epoch": 4.732394366197183, "grad_norm": 0.24265912175178528, "learning_rate": 8.417252835174749e-07, "loss": 0.0066, "step": 3150 }, { "epoch": 4.732394366197183, "eval_loss": 0.043101657181978226, "eval_runtime": 223.1112, "eval_samples_per_second": 5.307, "eval_steps_per_second": 5.307, "step": 3150 }, { "epoch": 4.7474178403755865, "grad_norm": 0.13980786502361298, "learning_rate": 7.48510501666455e-07, "loss": 0.0073, "step": 3160 }, { "epoch": 4.762441314553991, "grad_norm": 0.14822153747081757, "learning_rate": 6.607256212501578e-07, "loss": 0.0067, "step": 3170 }, { "epoch": 4.777464788732394, "grad_norm": 0.2598550021648407, "learning_rate": 5.783803204174654e-07, "loss": 0.0054, "step": 3180 }, { "epoch": 4.792488262910798, "grad_norm": 0.145918071269989, "learning_rate": 5.014836776119358e-07, "loss": 0.0051, "step": 3190 }, { "epoch": 4.807511737089202, "grad_norm": 0.152046799659729, "learning_rate": 4.300441705708924e-07, "loss": 0.003, "step": 3200 }, { "epoch": 4.807511737089202, "eval_loss": 0.043016187846660614, "eval_runtime": 222.4587, "eval_samples_per_second": 5.322, "eval_steps_per_second": 5.322, "step": 3200 }, { "epoch": 4.822535211267605, "grad_norm": 0.1859809011220932, "learning_rate": 3.6406967539078796e-07, "loss": 0.0042, "step": 3210 }, { "epoch": 4.83755868544601, "grad_norm": 0.10558822751045227, "learning_rate": 3.0356746565887715e-07, "loss": 0.0056, "step": 3220 }, { "epoch": 4.852582159624413, "grad_norm": 0.3504423201084137, "learning_rate": 2.485442116513026e-07, "loss": 0.0048, "step": 3230 }, { "epoch": 4.867605633802817, "grad_norm": 0.21484680473804474, "learning_rate": 1.9900597959770507e-07, "loss": 0.0079, "step": 3240 }, { "epoch": 4.882629107981221, "grad_norm": 0.07226985692977905, "learning_rate": 1.5495823101245866e-07, "loss": 0.0031, "step": 3250 }, { "epoch": 4.882629107981221, "eval_loss": 0.042939841747283936, "eval_runtime": 222.1162, "eval_samples_per_second": 5.331, "eval_steps_per_second": 5.331, "step": 3250 }, { "epoch": 4.897652582159624, "grad_norm": 0.0429568886756897, "learning_rate": 1.164058220925135e-07, "loss": 0.0043, "step": 3260 }, { "epoch": 4.912676056338028, "grad_norm": 0.3218914568424225, "learning_rate": 8.335300318201844e-08, "loss": 0.0074, "step": 3270 }, { "epoch": 4.927699530516432, "grad_norm": 0.10779738426208496, "learning_rate": 5.5803418303745917e-08, "loss": 0.005, "step": 3280 }, { "epoch": 4.942723004694836, "grad_norm": 0.0708744004368782, "learning_rate": 3.3760104757313284e-08, "loss": 0.003, "step": 3290 }, { "epoch": 4.957746478873239, "grad_norm": 0.19295717775821686, "learning_rate": 1.7225492784345156e-08, "loss": 0.0046, "step": 3300 }, { "epoch": 4.957746478873239, "eval_loss": 0.042976681143045425, "eval_runtime": 222.2614, "eval_samples_per_second": 5.327, "eval_steps_per_second": 5.327, "step": 3300 }, { "epoch": 4.972769953051643, "grad_norm": 0.10675567388534546, "learning_rate": 6.201405300532148e-09, "loss": 0.0058, "step": 3310 }, { "epoch": 4.987793427230047, "grad_norm": 0.09640572220087051, "learning_rate": 6.890576946805282e-10, "loss": 0.0043, "step": 3320 }, { "epoch": 4.995305164319249, "step": 3325, "total_flos": 8.987524405535048e+17, "train_loss": 0.049058449685125426, "train_runtime": 45792.1032, "train_samples_per_second": 1.163, "train_steps_per_second": 0.073 } ], "logging_steps": 10, "max_steps": 3325, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 50, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.987524405535048e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }