| { | |
| "best_global_step": 1800, | |
| "best_metric": 2.285778284072876, | |
| "best_model_checkpoint": "models/mini_stage2/checkpoint-1800", | |
| "epoch": 0.8933002481389578, | |
| "eval_steps": 300, | |
| "global_step": 1800, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.004962779156327543, | |
| "grad_norm": 7.886381149291992, | |
| "learning_rate": 1.3366336633663365e-06, | |
| "loss": 4.6656, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.009925558312655087, | |
| "grad_norm": 5.6866021156311035, | |
| "learning_rate": 2.821782178217822e-06, | |
| "loss": 4.4733, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.01488833746898263, | |
| "grad_norm": 4.5250091552734375, | |
| "learning_rate": 4.306930693069306e-06, | |
| "loss": 4.0093, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.019851116625310174, | |
| "grad_norm": 3.3117995262145996, | |
| "learning_rate": 5.792079207920792e-06, | |
| "loss": 3.9259, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.02481389578163772, | |
| "grad_norm": 3.5785279273986816, | |
| "learning_rate": 7.277227722772277e-06, | |
| "loss": 3.8315, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.02977667493796526, | |
| "grad_norm": 5.32008695602417, | |
| "learning_rate": 8.762376237623762e-06, | |
| "loss": 3.673, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.034739454094292806, | |
| "grad_norm": 4.776355266571045, | |
| "learning_rate": 1.0247524752475248e-05, | |
| "loss": 3.5076, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.03970223325062035, | |
| "grad_norm": 4.523674488067627, | |
| "learning_rate": 1.1732673267326734e-05, | |
| "loss": 3.4416, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.04466501240694789, | |
| "grad_norm": 4.444321155548096, | |
| "learning_rate": 1.3217821782178218e-05, | |
| "loss": 3.4362, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.04962779156327544, | |
| "grad_norm": 4.364451885223389, | |
| "learning_rate": 1.4702970297029704e-05, | |
| "loss": 3.3934, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.05459057071960298, | |
| "grad_norm": 5.098319053649902, | |
| "learning_rate": 1.618811881188119e-05, | |
| "loss": 3.3343, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.05955334987593052, | |
| "grad_norm": 4.762153148651123, | |
| "learning_rate": 1.7673267326732672e-05, | |
| "loss": 3.3018, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.06451612903225806, | |
| "grad_norm": 4.8034563064575195, | |
| "learning_rate": 1.915841584158416e-05, | |
| "loss": 3.2882, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.06947890818858561, | |
| "grad_norm": 4.955429553985596, | |
| "learning_rate": 2.0643564356435643e-05, | |
| "loss": 3.3027, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.07444168734491315, | |
| "grad_norm": 4.338839530944824, | |
| "learning_rate": 2.212871287128713e-05, | |
| "loss": 3.2177, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.0794044665012407, | |
| "grad_norm": 4.854152202606201, | |
| "learning_rate": 2.3613861386138615e-05, | |
| "loss": 3.2708, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.08436724565756824, | |
| "grad_norm": 4.855740070343018, | |
| "learning_rate": 2.5099009900990097e-05, | |
| "loss": 3.2645, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.08933002481389578, | |
| "grad_norm": 5.82074499130249, | |
| "learning_rate": 2.6584158415841586e-05, | |
| "loss": 3.1939, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.09429280397022333, | |
| "grad_norm": 5.284363746643066, | |
| "learning_rate": 2.8069306930693072e-05, | |
| "loss": 3.0575, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.09925558312655088, | |
| "grad_norm": 5.663958549499512, | |
| "learning_rate": 2.9554455445544555e-05, | |
| "loss": 3.0799, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.10421836228287841, | |
| "grad_norm": 5.67520809173584, | |
| "learning_rate": 2.9884169884169887e-05, | |
| "loss": 3.0824, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.10918114143920596, | |
| "grad_norm": 5.432056903839111, | |
| "learning_rate": 2.9718698290126862e-05, | |
| "loss": 3.0693, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.1141439205955335, | |
| "grad_norm": 5.387454032897949, | |
| "learning_rate": 2.9553226696083837e-05, | |
| "loss": 3.1014, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.11910669975186104, | |
| "grad_norm": 5.515647888183594, | |
| "learning_rate": 2.9387755102040816e-05, | |
| "loss": 3.0458, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.12406947890818859, | |
| "grad_norm": 5.962894916534424, | |
| "learning_rate": 2.9222283507997795e-05, | |
| "loss": 3.04, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.12903225806451613, | |
| "grad_norm": 5.549206733703613, | |
| "learning_rate": 2.9056811913954774e-05, | |
| "loss": 3.0311, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.13399503722084366, | |
| "grad_norm": 4.67368221282959, | |
| "learning_rate": 2.889134031991175e-05, | |
| "loss": 2.9778, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.13895781637717122, | |
| "grad_norm": 5.445274829864502, | |
| "learning_rate": 2.8725868725868724e-05, | |
| "loss": 3.0701, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.14392059553349876, | |
| "grad_norm": 5.643553256988525, | |
| "learning_rate": 2.8560397131825703e-05, | |
| "loss": 2.9039, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.1488833746898263, | |
| "grad_norm": 5.453239440917969, | |
| "learning_rate": 2.8394925537782682e-05, | |
| "loss": 3.0449, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.1488833746898263, | |
| "eval_loss": 2.5684616565704346, | |
| "eval_runtime": 119.1131, | |
| "eval_samples_per_second": 142.251, | |
| "eval_steps_per_second": 4.45, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.15384615384615385, | |
| "grad_norm": 5.800920009613037, | |
| "learning_rate": 2.822945394373966e-05, | |
| "loss": 2.8896, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.1588089330024814, | |
| "grad_norm": 4.8736724853515625, | |
| "learning_rate": 2.8063982349696636e-05, | |
| "loss": 3.0527, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.16377171215880892, | |
| "grad_norm": 4.498941898345947, | |
| "learning_rate": 2.7898510755653615e-05, | |
| "loss": 3.0153, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.1687344913151365, | |
| "grad_norm": 5.518968105316162, | |
| "learning_rate": 2.773303916161059e-05, | |
| "loss": 2.869, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.17369727047146402, | |
| "grad_norm": 5.431845188140869, | |
| "learning_rate": 2.756756756756757e-05, | |
| "loss": 2.9678, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.17866004962779156, | |
| "grad_norm": 5.1342973709106445, | |
| "learning_rate": 2.7402095973524544e-05, | |
| "loss": 2.9756, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.18362282878411912, | |
| "grad_norm": 4.8498101234436035, | |
| "learning_rate": 2.7236624379481523e-05, | |
| "loss": 2.9348, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.18858560794044665, | |
| "grad_norm": 5.223404407501221, | |
| "learning_rate": 2.70711527854385e-05, | |
| "loss": 2.9967, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.1935483870967742, | |
| "grad_norm": 5.078680038452148, | |
| "learning_rate": 2.690568119139548e-05, | |
| "loss": 2.8953, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.19851116625310175, | |
| "grad_norm": 4.336000442504883, | |
| "learning_rate": 2.6740209597352456e-05, | |
| "loss": 2.9546, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.20347394540942929, | |
| "grad_norm": 5.194223880767822, | |
| "learning_rate": 2.657473800330943e-05, | |
| "loss": 2.9919, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.20843672456575682, | |
| "grad_norm": 4.870718002319336, | |
| "learning_rate": 2.640926640926641e-05, | |
| "loss": 2.8487, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.21339950372208435, | |
| "grad_norm": 5.526561260223389, | |
| "learning_rate": 2.624379481522339e-05, | |
| "loss": 2.7609, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.21836228287841192, | |
| "grad_norm": 5.372407913208008, | |
| "learning_rate": 2.6078323221180364e-05, | |
| "loss": 2.9126, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.22332506203473945, | |
| "grad_norm": 4.934298992156982, | |
| "learning_rate": 2.5912851627137343e-05, | |
| "loss": 2.8991, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.228287841191067, | |
| "grad_norm": 4.916905403137207, | |
| "learning_rate": 2.574738003309432e-05, | |
| "loss": 2.9272, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.23325062034739455, | |
| "grad_norm": 5.660480976104736, | |
| "learning_rate": 2.5581908439051297e-05, | |
| "loss": 2.9084, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.23821339950372208, | |
| "grad_norm": 4.987634658813477, | |
| "learning_rate": 2.5416436845008272e-05, | |
| "loss": 2.7963, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.24317617866004962, | |
| "grad_norm": 6.746623992919922, | |
| "learning_rate": 2.525096525096525e-05, | |
| "loss": 2.822, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.24813895781637718, | |
| "grad_norm": 4.658477783203125, | |
| "learning_rate": 2.508549365692223e-05, | |
| "loss": 2.9376, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.2531017369727047, | |
| "grad_norm": 5.2974162101745605, | |
| "learning_rate": 2.492002206287921e-05, | |
| "loss": 2.8969, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.25806451612903225, | |
| "grad_norm": 5.682817459106445, | |
| "learning_rate": 2.4754550468836184e-05, | |
| "loss": 2.7745, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.2630272952853598, | |
| "grad_norm": 5.85469388961792, | |
| "learning_rate": 2.458907887479316e-05, | |
| "loss": 2.8103, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.2679900744416873, | |
| "grad_norm": 4.600194454193115, | |
| "learning_rate": 2.4423607280750138e-05, | |
| "loss": 2.8189, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.2729528535980149, | |
| "grad_norm": 6.168639659881592, | |
| "learning_rate": 2.4258135686707117e-05, | |
| "loss": 2.8322, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.27791563275434245, | |
| "grad_norm": 4.763863563537598, | |
| "learning_rate": 2.4092664092664092e-05, | |
| "loss": 2.7627, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.28287841191067, | |
| "grad_norm": 5.0977091789245605, | |
| "learning_rate": 2.392719249862107e-05, | |
| "loss": 2.7796, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.2878411910669975, | |
| "grad_norm": 4.786524772644043, | |
| "learning_rate": 2.376172090457805e-05, | |
| "loss": 2.8515, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.29280397022332505, | |
| "grad_norm": 5.386171817779541, | |
| "learning_rate": 2.3596249310535025e-05, | |
| "loss": 2.8758, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.2977667493796526, | |
| "grad_norm": 4.96185827255249, | |
| "learning_rate": 2.3430777716492e-05, | |
| "loss": 2.7963, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.2977667493796526, | |
| "eval_loss": 2.4141688346862793, | |
| "eval_runtime": 112.2744, | |
| "eval_samples_per_second": 150.916, | |
| "eval_steps_per_second": 4.721, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.3027295285359802, | |
| "grad_norm": 7.280220985412598, | |
| "learning_rate": 2.326530612244898e-05, | |
| "loss": 2.8259, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.3076923076923077, | |
| "grad_norm": 5.671541690826416, | |
| "learning_rate": 2.3099834528405958e-05, | |
| "loss": 2.829, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.31265508684863524, | |
| "grad_norm": 5.268543720245361, | |
| "learning_rate": 2.2934362934362936e-05, | |
| "loss": 2.7699, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.3176178660049628, | |
| "grad_norm": 4.945257186889648, | |
| "learning_rate": 2.2768891340319915e-05, | |
| "loss": 2.7311, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.3225806451612903, | |
| "grad_norm": 5.198265075683594, | |
| "learning_rate": 2.2603419746276887e-05, | |
| "loss": 2.735, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.32754342431761785, | |
| "grad_norm": 4.624203205108643, | |
| "learning_rate": 2.2437948152233866e-05, | |
| "loss": 2.7306, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.3325062034739454, | |
| "grad_norm": 5.143148899078369, | |
| "learning_rate": 2.2272476558190845e-05, | |
| "loss": 2.7467, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.337468982630273, | |
| "grad_norm": 4.643360614776611, | |
| "learning_rate": 2.2107004964147823e-05, | |
| "loss": 2.7494, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.3424317617866005, | |
| "grad_norm": 5.404439926147461, | |
| "learning_rate": 2.19415333701048e-05, | |
| "loss": 2.7386, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.34739454094292804, | |
| "grad_norm": 4.590571403503418, | |
| "learning_rate": 2.1776061776061778e-05, | |
| "loss": 2.8513, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.3523573200992556, | |
| "grad_norm": 5.185126781463623, | |
| "learning_rate": 2.1610590182018756e-05, | |
| "loss": 2.673, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.3573200992555831, | |
| "grad_norm": 5.4743547439575195, | |
| "learning_rate": 2.144511858797573e-05, | |
| "loss": 2.8101, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.36228287841191065, | |
| "grad_norm": 5.463687896728516, | |
| "learning_rate": 2.1279646993932707e-05, | |
| "loss": 2.7527, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.36724565756823824, | |
| "grad_norm": 6.102997303009033, | |
| "learning_rate": 2.1114175399889686e-05, | |
| "loss": 2.7213, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.37220843672456577, | |
| "grad_norm": 5.1998724937438965, | |
| "learning_rate": 2.0948703805846664e-05, | |
| "loss": 2.753, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.3771712158808933, | |
| "grad_norm": 5.181605815887451, | |
| "learning_rate": 2.0783232211803643e-05, | |
| "loss": 2.8034, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.38213399503722084, | |
| "grad_norm": 5.865420818328857, | |
| "learning_rate": 2.061776061776062e-05, | |
| "loss": 2.8288, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.3870967741935484, | |
| "grad_norm": 4.965085029602051, | |
| "learning_rate": 2.0452289023717594e-05, | |
| "loss": 2.613, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.3920595533498759, | |
| "grad_norm": 5.534601211547852, | |
| "learning_rate": 2.0286817429674573e-05, | |
| "loss": 2.7315, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.3970223325062035, | |
| "grad_norm": 5.220632076263428, | |
| "learning_rate": 2.012134583563155e-05, | |
| "loss": 2.8077, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.40198511166253104, | |
| "grad_norm": 5.059537410736084, | |
| "learning_rate": 1.9955874241588527e-05, | |
| "loss": 2.7442, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.40694789081885857, | |
| "grad_norm": 5.325795650482178, | |
| "learning_rate": 1.9790402647545506e-05, | |
| "loss": 2.7351, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.4119106699751861, | |
| "grad_norm": 8.455648422241211, | |
| "learning_rate": 1.9624931053502484e-05, | |
| "loss": 2.7643, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.41687344913151364, | |
| "grad_norm": 5.344241619110107, | |
| "learning_rate": 1.945945945945946e-05, | |
| "loss": 2.8984, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.4218362282878412, | |
| "grad_norm": 4.816617965698242, | |
| "learning_rate": 1.9293987865416435e-05, | |
| "loss": 2.7377, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.4267990074441687, | |
| "grad_norm": 5.334592819213867, | |
| "learning_rate": 1.9128516271373414e-05, | |
| "loss": 2.7021, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.4317617866004963, | |
| "grad_norm": 5.0712385177612305, | |
| "learning_rate": 1.8963044677330393e-05, | |
| "loss": 2.6756, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.43672456575682383, | |
| "grad_norm": 5.225603103637695, | |
| "learning_rate": 1.879757308328737e-05, | |
| "loss": 2.7852, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.44168734491315137, | |
| "grad_norm": 5.172723770141602, | |
| "learning_rate": 1.8632101489244347e-05, | |
| "loss": 2.7531, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.4466501240694789, | |
| "grad_norm": 4.887333869934082, | |
| "learning_rate": 1.8466629895201322e-05, | |
| "loss": 2.6636, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.4466501240694789, | |
| "eval_loss": 2.3456013202667236, | |
| "eval_runtime": 113.8402, | |
| "eval_samples_per_second": 148.84, | |
| "eval_steps_per_second": 4.656, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.45161290322580644, | |
| "grad_norm": 5.249894142150879, | |
| "learning_rate": 1.83011583011583e-05, | |
| "loss": 2.7089, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.456575682382134, | |
| "grad_norm": 4.984340667724609, | |
| "learning_rate": 1.813568670711528e-05, | |
| "loss": 2.8029, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.46153846153846156, | |
| "grad_norm": 5.297105312347412, | |
| "learning_rate": 1.7970215113072255e-05, | |
| "loss": 2.721, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.4665012406947891, | |
| "grad_norm": 5.074361801147461, | |
| "learning_rate": 1.7804743519029234e-05, | |
| "loss": 2.5606, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.47146401985111663, | |
| "grad_norm": 5.639915943145752, | |
| "learning_rate": 1.7639271924986212e-05, | |
| "loss": 2.6397, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.47642679900744417, | |
| "grad_norm": 5.445567607879639, | |
| "learning_rate": 1.7473800330943188e-05, | |
| "loss": 2.6563, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.4813895781637717, | |
| "grad_norm": 4.462475299835205, | |
| "learning_rate": 1.7308328736900166e-05, | |
| "loss": 2.7163, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.48635235732009924, | |
| "grad_norm": 5.833073139190674, | |
| "learning_rate": 1.7142857142857142e-05, | |
| "loss": 2.6225, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.4913151364764268, | |
| "grad_norm": 5.428512096405029, | |
| "learning_rate": 1.697738554881412e-05, | |
| "loss": 2.645, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.49627791563275436, | |
| "grad_norm": 4.768510818481445, | |
| "learning_rate": 1.68119139547711e-05, | |
| "loss": 2.6576, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.5012406947890818, | |
| "grad_norm": 4.692178726196289, | |
| "learning_rate": 1.6646442360728078e-05, | |
| "loss": 2.7019, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.5062034739454094, | |
| "grad_norm": 5.51165771484375, | |
| "learning_rate": 1.6480970766685053e-05, | |
| "loss": 2.7195, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.511166253101737, | |
| "grad_norm": 4.9710516929626465, | |
| "learning_rate": 1.631549917264203e-05, | |
| "loss": 2.7242, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.5161290322580645, | |
| "grad_norm": 5.166830062866211, | |
| "learning_rate": 1.6150027578599008e-05, | |
| "loss": 2.6729, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.5210918114143921, | |
| "grad_norm": 4.945868015289307, | |
| "learning_rate": 1.5984555984555986e-05, | |
| "loss": 2.7637, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.5260545905707196, | |
| "grad_norm": 5.7991943359375, | |
| "learning_rate": 1.581908439051296e-05, | |
| "loss": 2.677, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.5310173697270472, | |
| "grad_norm": 5.255971431732178, | |
| "learning_rate": 1.565361279646994e-05, | |
| "loss": 2.7018, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.5359801488833746, | |
| "grad_norm": 5.910732746124268, | |
| "learning_rate": 1.548814120242692e-05, | |
| "loss": 2.6469, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.5409429280397022, | |
| "grad_norm": 4.5612616539001465, | |
| "learning_rate": 1.5322669608383895e-05, | |
| "loss": 2.7186, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.5459057071960298, | |
| "grad_norm": 5.111081600189209, | |
| "learning_rate": 1.515719801434087e-05, | |
| "loss": 2.6728, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.5508684863523573, | |
| "grad_norm": 5.450465202331543, | |
| "learning_rate": 1.4991726420297849e-05, | |
| "loss": 2.6694, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.5558312655086849, | |
| "grad_norm": 4.850940704345703, | |
| "learning_rate": 1.4826254826254827e-05, | |
| "loss": 2.7839, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.5607940446650124, | |
| "grad_norm": 5.036434173583984, | |
| "learning_rate": 1.4660783232211803e-05, | |
| "loss": 2.5834, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.56575682382134, | |
| "grad_norm": 40.256370544433594, | |
| "learning_rate": 1.4495311638168781e-05, | |
| "loss": 2.6905, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.5707196029776674, | |
| "grad_norm": 5.939204692840576, | |
| "learning_rate": 1.4329840044125759e-05, | |
| "loss": 2.7223, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.575682382133995, | |
| "grad_norm": 5.288106441497803, | |
| "learning_rate": 1.4164368450082736e-05, | |
| "loss": 2.7235, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.5806451612903226, | |
| "grad_norm": 5.540660858154297, | |
| "learning_rate": 1.3998896856039713e-05, | |
| "loss": 2.636, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.5856079404466501, | |
| "grad_norm": 5.182808876037598, | |
| "learning_rate": 1.3833425261996691e-05, | |
| "loss": 2.6314, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.5905707196029777, | |
| "grad_norm": 4.9914140701293945, | |
| "learning_rate": 1.3667953667953668e-05, | |
| "loss": 2.5941, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.5955334987593052, | |
| "grad_norm": 4.8229475021362305, | |
| "learning_rate": 1.3502482073910646e-05, | |
| "loss": 2.7827, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.5955334987593052, | |
| "eval_loss": 2.2910733222961426, | |
| "eval_runtime": 115.8769, | |
| "eval_samples_per_second": 146.224, | |
| "eval_steps_per_second": 4.574, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.6004962779156328, | |
| "grad_norm": 5.556408882141113, | |
| "learning_rate": 1.3337010479867623e-05, | |
| "loss": 2.6104, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.6054590570719603, | |
| "grad_norm": 6.0620551109313965, | |
| "learning_rate": 1.3171538885824601e-05, | |
| "loss": 2.6148, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.6104218362282878, | |
| "grad_norm": 5.285867214202881, | |
| "learning_rate": 1.3006067291781577e-05, | |
| "loss": 2.6355, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.6153846153846154, | |
| "grad_norm": 5.499397277832031, | |
| "learning_rate": 1.2840595697738555e-05, | |
| "loss": 2.6269, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.6203473945409429, | |
| "grad_norm": 4.932219505310059, | |
| "learning_rate": 1.2675124103695532e-05, | |
| "loss": 2.6003, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.6253101736972705, | |
| "grad_norm": 5.04417610168457, | |
| "learning_rate": 1.250965250965251e-05, | |
| "loss": 2.6256, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.630272952853598, | |
| "grad_norm": 5.0753679275512695, | |
| "learning_rate": 1.2344180915609488e-05, | |
| "loss": 2.6326, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.6352357320099256, | |
| "grad_norm": 5.003329753875732, | |
| "learning_rate": 1.2178709321566465e-05, | |
| "loss": 2.681, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.6401985111662531, | |
| "grad_norm": 5.094631671905518, | |
| "learning_rate": 1.2013237727523442e-05, | |
| "loss": 2.5776, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.6451612903225806, | |
| "grad_norm": 4.915186405181885, | |
| "learning_rate": 1.184776613348042e-05, | |
| "loss": 2.7528, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.6501240694789082, | |
| "grad_norm": 5.3234076499938965, | |
| "learning_rate": 1.1682294539437398e-05, | |
| "loss": 2.6076, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.6550868486352357, | |
| "grad_norm": 6.5297064781188965, | |
| "learning_rate": 1.1516822945394374e-05, | |
| "loss": 2.5784, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.6600496277915633, | |
| "grad_norm": 5.549312591552734, | |
| "learning_rate": 1.1351351351351352e-05, | |
| "loss": 2.6064, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.6650124069478908, | |
| "grad_norm": 4.589717388153076, | |
| "learning_rate": 1.118587975730833e-05, | |
| "loss": 2.5757, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.6699751861042184, | |
| "grad_norm": 6.217405796051025, | |
| "learning_rate": 1.1020408163265306e-05, | |
| "loss": 2.5851, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.674937965260546, | |
| "grad_norm": 5.602436542510986, | |
| "learning_rate": 1.0854936569222283e-05, | |
| "loss": 2.6007, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.6799007444168734, | |
| "grad_norm": 4.934968948364258, | |
| "learning_rate": 1.0689464975179262e-05, | |
| "loss": 2.5674, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.684863523573201, | |
| "grad_norm": 5.8763322830200195, | |
| "learning_rate": 1.0523993381136238e-05, | |
| "loss": 2.6984, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.6898263027295285, | |
| "grad_norm": 5.077223777770996, | |
| "learning_rate": 1.0358521787093216e-05, | |
| "loss": 2.6202, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.6947890818858561, | |
| "grad_norm": 6.423431873321533, | |
| "learning_rate": 1.0193050193050193e-05, | |
| "loss": 2.6729, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.6997518610421837, | |
| "grad_norm": 5.683679103851318, | |
| "learning_rate": 1.002757859900717e-05, | |
| "loss": 2.6683, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.7047146401985112, | |
| "grad_norm": 5.217991828918457, | |
| "learning_rate": 9.862107004964148e-06, | |
| "loss": 2.6355, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.7096774193548387, | |
| "grad_norm": 5.0821332931518555, | |
| "learning_rate": 9.696635410921126e-06, | |
| "loss": 2.6033, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.7146401985111662, | |
| "grad_norm": 5.337713718414307, | |
| "learning_rate": 9.531163816878102e-06, | |
| "loss": 2.6834, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.7196029776674938, | |
| "grad_norm": 5.425897598266602, | |
| "learning_rate": 9.36569222283508e-06, | |
| "loss": 2.6597, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.7245657568238213, | |
| "grad_norm": 4.878344535827637, | |
| "learning_rate": 9.200220628792057e-06, | |
| "loss": 2.6298, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.7295285359801489, | |
| "grad_norm": 5.754893779754639, | |
| "learning_rate": 9.034749034749034e-06, | |
| "loss": 2.6232, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.7344913151364765, | |
| "grad_norm": 4.991476535797119, | |
| "learning_rate": 8.869277440706012e-06, | |
| "loss": 2.5672, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.739454094292804, | |
| "grad_norm": 6.236905097961426, | |
| "learning_rate": 8.70380584666299e-06, | |
| "loss": 2.5139, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.7444168734491315, | |
| "grad_norm": 4.892019748687744, | |
| "learning_rate": 8.538334252619966e-06, | |
| "loss": 2.6248, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.7444168734491315, | |
| "eval_loss": 2.3090243339538574, | |
| "eval_runtime": 114.529, | |
| "eval_samples_per_second": 147.945, | |
| "eval_steps_per_second": 4.628, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.749379652605459, | |
| "grad_norm": 5.293705463409424, | |
| "learning_rate": 8.372862658576944e-06, | |
| "loss": 2.6417, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.7543424317617866, | |
| "grad_norm": 6.500164031982422, | |
| "learning_rate": 8.207391064533921e-06, | |
| "loss": 2.6197, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.7593052109181141, | |
| "grad_norm": 5.456136226654053, | |
| "learning_rate": 8.0419194704909e-06, | |
| "loss": 2.6911, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.7642679900744417, | |
| "grad_norm": 5.395550727844238, | |
| "learning_rate": 7.876447876447876e-06, | |
| "loss": 2.5542, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.7692307692307693, | |
| "grad_norm": 5.446476936340332, | |
| "learning_rate": 7.710976282404854e-06, | |
| "loss": 2.6584, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.7741935483870968, | |
| "grad_norm": 5.699354648590088, | |
| "learning_rate": 7.5455046883618305e-06, | |
| "loss": 2.6182, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.7791563275434243, | |
| "grad_norm": 5.835291862487793, | |
| "learning_rate": 7.3800330943188084e-06, | |
| "loss": 2.6301, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.7841191066997518, | |
| "grad_norm": 5.066339015960693, | |
| "learning_rate": 7.214561500275786e-06, | |
| "loss": 2.5629, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.7890818858560794, | |
| "grad_norm": 6.113178253173828, | |
| "learning_rate": 7.049089906232763e-06, | |
| "loss": 2.5965, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.794044665012407, | |
| "grad_norm": 5.544576168060303, | |
| "learning_rate": 6.8836183121897404e-06, | |
| "loss": 2.5722, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.7990074441687345, | |
| "grad_norm": 5.560300350189209, | |
| "learning_rate": 6.718146718146718e-06, | |
| "loss": 2.5835, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.8039702233250621, | |
| "grad_norm": 5.354392051696777, | |
| "learning_rate": 6.552675124103695e-06, | |
| "loss": 2.5901, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.8089330024813896, | |
| "grad_norm": 5.256809234619141, | |
| "learning_rate": 6.387203530060673e-06, | |
| "loss": 2.6055, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.8138957816377171, | |
| "grad_norm": 6.089339256286621, | |
| "learning_rate": 6.22173193601765e-06, | |
| "loss": 2.6019, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.8188585607940446, | |
| "grad_norm": 4.713355541229248, | |
| "learning_rate": 6.056260341974627e-06, | |
| "loss": 2.6421, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.8238213399503722, | |
| "grad_norm": 6.513923168182373, | |
| "learning_rate": 5.890788747931605e-06, | |
| "loss": 2.6049, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.8287841191066998, | |
| "grad_norm": 5.37587833404541, | |
| "learning_rate": 5.725317153888582e-06, | |
| "loss": 2.5351, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.8337468982630273, | |
| "grad_norm": 6.346861362457275, | |
| "learning_rate": 5.5598455598455594e-06, | |
| "loss": 2.6158, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.8387096774193549, | |
| "grad_norm": 5.206553936004639, | |
| "learning_rate": 5.394373965802537e-06, | |
| "loss": 2.5994, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.8436724565756824, | |
| "grad_norm": 6.567471504211426, | |
| "learning_rate": 5.228902371759514e-06, | |
| "loss": 2.5816, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.8486352357320099, | |
| "grad_norm": 5.689151287078857, | |
| "learning_rate": 5.0634307777164914e-06, | |
| "loss": 2.5848, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.8535980148883374, | |
| "grad_norm": 5.288684844970703, | |
| "learning_rate": 4.897959183673469e-06, | |
| "loss": 2.6138, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.858560794044665, | |
| "grad_norm": 5.238705635070801, | |
| "learning_rate": 4.732487589630447e-06, | |
| "loss": 2.5811, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.8635235732009926, | |
| "grad_norm": 5.114046573638916, | |
| "learning_rate": 4.567015995587424e-06, | |
| "loss": 2.5933, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.8684863523573201, | |
| "grad_norm": 6.13997220993042, | |
| "learning_rate": 4.401544401544402e-06, | |
| "loss": 2.5869, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.8734491315136477, | |
| "grad_norm": 5.258871555328369, | |
| "learning_rate": 4.236072807501379e-06, | |
| "loss": 2.5464, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.8784119106699751, | |
| "grad_norm": 5.823874473571777, | |
| "learning_rate": 4.070601213458356e-06, | |
| "loss": 2.6842, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.8833746898263027, | |
| "grad_norm": 5.351442337036133, | |
| "learning_rate": 3.905129619415334e-06, | |
| "loss": 2.6312, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.8883374689826302, | |
| "grad_norm": 5.9153947830200195, | |
| "learning_rate": 3.739658025372311e-06, | |
| "loss": 2.5621, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.8933002481389578, | |
| "grad_norm": 5.032203197479248, | |
| "learning_rate": 3.5741864313292883e-06, | |
| "loss": 2.6103, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.8933002481389578, | |
| "eval_loss": 2.285778284072876, | |
| "eval_runtime": 108.9864, | |
| "eval_samples_per_second": 155.469, | |
| "eval_steps_per_second": 4.863, | |
| "step": 1800 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 2015, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 300, | |
| "stateful_callbacks": { | |
| "EarlyStoppingCallback": { | |
| "args": { | |
| "early_stopping_patience": 5, | |
| "early_stopping_threshold": 0.0 | |
| }, | |
| "attributes": { | |
| "early_stopping_patience_counter": 0 | |
| } | |
| }, | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 64, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |