| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.9997197702115734, | |
| "eval_steps": 50, | |
| "global_step": 892, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.02241838307412078, | |
| "grad_norm": 11.57065531644996, | |
| "learning_rate": 1.9775784753363228e-06, | |
| "loss": 0.2054, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.04483676614824156, | |
| "grad_norm": 10.25086765406488, | |
| "learning_rate": 1.9551569506726456e-06, | |
| "loss": 0.1492, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.06725514922236234, | |
| "grad_norm": 7.822767237983853, | |
| "learning_rate": 1.9327354260089685e-06, | |
| "loss": 0.152, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.08967353229648312, | |
| "grad_norm": 8.157708200242721, | |
| "learning_rate": 1.9103139013452914e-06, | |
| "loss": 0.1472, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.1120919153706039, | |
| "grad_norm": 12.8257632132892, | |
| "learning_rate": 1.8878923766816142e-06, | |
| "loss": 0.1356, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.1120919153706039, | |
| "eval_loss": 0.1209246814250946, | |
| "eval_runtime": 477.8616, | |
| "eval_samples_per_second": 6.288, | |
| "eval_steps_per_second": 0.787, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.13451029844472467, | |
| "grad_norm": 8.90816614814061, | |
| "learning_rate": 1.865470852017937e-06, | |
| "loss": 0.1281, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.15692868151884545, | |
| "grad_norm": 9.628880600516572, | |
| "learning_rate": 1.84304932735426e-06, | |
| "loss": 0.1381, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.17934706459296623, | |
| "grad_norm": 8.899868902286775, | |
| "learning_rate": 1.8206278026905828e-06, | |
| "loss": 0.1324, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.201765447667087, | |
| "grad_norm": 130.29750942761115, | |
| "learning_rate": 1.798206278026906e-06, | |
| "loss": 0.1277, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.2241838307412078, | |
| "grad_norm": 6.4686996638845145, | |
| "learning_rate": 1.7757847533632286e-06, | |
| "loss": 0.1221, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.2241838307412078, | |
| "eval_loss": 0.11111436039209366, | |
| "eval_runtime": 477.824, | |
| "eval_samples_per_second": 6.289, | |
| "eval_steps_per_second": 0.787, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.24660221381532857, | |
| "grad_norm": 8.356076121379255, | |
| "learning_rate": 1.7533632286995514e-06, | |
| "loss": 0.1198, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.26902059688944935, | |
| "grad_norm": 6.743428651631778, | |
| "learning_rate": 1.7309417040358743e-06, | |
| "loss": 0.1076, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.29143897996357016, | |
| "grad_norm": 6.8645809938665465, | |
| "learning_rate": 1.7085201793721974e-06, | |
| "loss": 0.1187, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.3138573630376909, | |
| "grad_norm": 5.999842993813071, | |
| "learning_rate": 1.68609865470852e-06, | |
| "loss": 0.1182, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.3362757461118117, | |
| "grad_norm": 7.7478230858078305, | |
| "learning_rate": 1.6636771300448429e-06, | |
| "loss": 0.1127, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.3362757461118117, | |
| "eval_loss": 0.10507839918136597, | |
| "eval_runtime": 477.8167, | |
| "eval_samples_per_second": 6.289, | |
| "eval_steps_per_second": 0.787, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.35869412918593246, | |
| "grad_norm": 6.564504796683449, | |
| "learning_rate": 1.641255605381166e-06, | |
| "loss": 0.1144, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.38111251226005327, | |
| "grad_norm": 20.75971255482363, | |
| "learning_rate": 1.6188340807174888e-06, | |
| "loss": 0.1078, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.403530895334174, | |
| "grad_norm": 6.913685427209001, | |
| "learning_rate": 1.5964125560538115e-06, | |
| "loss": 0.1189, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.42594927840829483, | |
| "grad_norm": 11.143255383687194, | |
| "learning_rate": 1.5739910313901345e-06, | |
| "loss": 0.1189, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.4483676614824156, | |
| "grad_norm": 4.747347623410045, | |
| "learning_rate": 1.5515695067264574e-06, | |
| "loss": 0.1072, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.4483676614824156, | |
| "eval_loss": 0.09710206091403961, | |
| "eval_runtime": 477.8045, | |
| "eval_samples_per_second": 6.289, | |
| "eval_steps_per_second": 0.787, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.4707860445565364, | |
| "grad_norm": 5.512284850848945, | |
| "learning_rate": 1.5291479820627803e-06, | |
| "loss": 0.1136, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.49320442763065714, | |
| "grad_norm": 17.278200474599828, | |
| "learning_rate": 1.506726457399103e-06, | |
| "loss": 0.1094, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.5156228107047779, | |
| "grad_norm": 46.879816629413476, | |
| "learning_rate": 1.484304932735426e-06, | |
| "loss": 0.1091, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.5380411937788987, | |
| "grad_norm": 8.493451669290872, | |
| "learning_rate": 1.4618834080717489e-06, | |
| "loss": 0.1019, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.5604595768530195, | |
| "grad_norm": 9.286929395356434, | |
| "learning_rate": 1.4394618834080715e-06, | |
| "loss": 0.1075, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.5604595768530195, | |
| "eval_loss": 0.09229769557714462, | |
| "eval_runtime": 477.6411, | |
| "eval_samples_per_second": 6.291, | |
| "eval_steps_per_second": 0.787, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.5828779599271403, | |
| "grad_norm": 7.559299755990858, | |
| "learning_rate": 1.4170403587443946e-06, | |
| "loss": 0.0985, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.605296343001261, | |
| "grad_norm": 3.9026812483393387, | |
| "learning_rate": 1.3946188340807175e-06, | |
| "loss": 0.0896, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.6277147260753818, | |
| "grad_norm": 8.335653074155596, | |
| "learning_rate": 1.3721973094170403e-06, | |
| "loss": 0.0893, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.6501331091495026, | |
| "grad_norm": 6.897239653871222, | |
| "learning_rate": 1.349775784753363e-06, | |
| "loss": 0.0914, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.6725514922236234, | |
| "grad_norm": 29.648594106464095, | |
| "learning_rate": 1.327354260089686e-06, | |
| "loss": 0.0972, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.6725514922236234, | |
| "eval_loss": 0.08544214069843292, | |
| "eval_runtime": 477.8106, | |
| "eval_samples_per_second": 6.289, | |
| "eval_steps_per_second": 0.787, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.6949698752977441, | |
| "grad_norm": 3.969809179016744, | |
| "learning_rate": 1.304932735426009e-06, | |
| "loss": 0.089, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.7173882583718649, | |
| "grad_norm": 5.34402617506164, | |
| "learning_rate": 1.2825112107623318e-06, | |
| "loss": 0.0867, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.7398066414459857, | |
| "grad_norm": 5.506697389437645, | |
| "learning_rate": 1.2600896860986546e-06, | |
| "loss": 0.0803, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.7622250245201065, | |
| "grad_norm": 8.056193416126446, | |
| "learning_rate": 1.2376681614349775e-06, | |
| "loss": 0.0917, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.7846434075942272, | |
| "grad_norm": 4.7612681555615, | |
| "learning_rate": 1.2152466367713004e-06, | |
| "loss": 0.089, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.7846434075942272, | |
| "eval_loss": 0.07997283339500427, | |
| "eval_runtime": 477.8824, | |
| "eval_samples_per_second": 6.288, | |
| "eval_steps_per_second": 0.787, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.807061790668348, | |
| "grad_norm": 9.702063628974917, | |
| "learning_rate": 1.1928251121076232e-06, | |
| "loss": 0.0886, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.8294801737424689, | |
| "grad_norm": 4.937196401368392, | |
| "learning_rate": 1.170403587443946e-06, | |
| "loss": 0.0805, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.8518985568165897, | |
| "grad_norm": 6.8769697579900555, | |
| "learning_rate": 1.147982062780269e-06, | |
| "loss": 0.0754, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.8743169398907104, | |
| "grad_norm": 10.270492248069893, | |
| "learning_rate": 1.1255605381165918e-06, | |
| "loss": 0.0855, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.8967353229648312, | |
| "grad_norm": 8.129704483519474, | |
| "learning_rate": 1.103139013452915e-06, | |
| "loss": 0.0898, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.8967353229648312, | |
| "eval_loss": 0.0756540298461914, | |
| "eval_runtime": 477.7284, | |
| "eval_samples_per_second": 6.29, | |
| "eval_steps_per_second": 0.787, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.919153706038952, | |
| "grad_norm": 7.785886269800572, | |
| "learning_rate": 1.0807174887892376e-06, | |
| "loss": 0.0827, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.9415720891130728, | |
| "grad_norm": 7.119394503978009, | |
| "learning_rate": 1.0582959641255604e-06, | |
| "loss": 0.0862, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.9639904721871935, | |
| "grad_norm": 5.06264160032142, | |
| "learning_rate": 1.0358744394618835e-06, | |
| "loss": 0.078, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.9864088552613143, | |
| "grad_norm": 6.034467278570911, | |
| "learning_rate": 1.0134529147982064e-06, | |
| "loss": 0.0745, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.008827238335435, | |
| "grad_norm": 6.107563557049032, | |
| "learning_rate": 9.91031390134529e-07, | |
| "loss": 0.0635, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.008827238335435, | |
| "eval_loss": 0.07114721089601517, | |
| "eval_runtime": 477.6398, | |
| "eval_samples_per_second": 6.291, | |
| "eval_steps_per_second": 0.787, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.0312456214095558, | |
| "grad_norm": 8.472989007522791, | |
| "learning_rate": 9.68609865470852e-07, | |
| "loss": 0.0687, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.0536640044836767, | |
| "grad_norm": 5.080789399491274, | |
| "learning_rate": 9.461883408071749e-07, | |
| "loss": 0.0635, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.0760823875577974, | |
| "grad_norm": 8.06976721278842, | |
| "learning_rate": 9.237668161434977e-07, | |
| "loss": 0.0674, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.098500770631918, | |
| "grad_norm": 9.00400557879772, | |
| "learning_rate": 9.013452914798207e-07, | |
| "loss": 0.0623, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.120919153706039, | |
| "grad_norm": 10.419684143484417, | |
| "learning_rate": 8.789237668161434e-07, | |
| "loss": 0.0742, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.120919153706039, | |
| "eval_loss": 0.06726241111755371, | |
| "eval_runtime": 478.2051, | |
| "eval_samples_per_second": 6.284, | |
| "eval_steps_per_second": 0.786, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.1433375367801597, | |
| "grad_norm": 5.834457752004053, | |
| "learning_rate": 8.565022421524663e-07, | |
| "loss": 0.0698, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.1657559198542806, | |
| "grad_norm": 5.920575213015667, | |
| "learning_rate": 8.340807174887892e-07, | |
| "loss": 0.061, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.1881743029284013, | |
| "grad_norm": 11.4664445988588, | |
| "learning_rate": 8.11659192825112e-07, | |
| "loss": 0.0612, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.210592686002522, | |
| "grad_norm": 5.074793088570693, | |
| "learning_rate": 7.892376681614349e-07, | |
| "loss": 0.0582, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.233011069076643, | |
| "grad_norm": 4.964609158362768, | |
| "learning_rate": 7.668161434977578e-07, | |
| "loss": 0.0627, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.233011069076643, | |
| "eval_loss": 0.06428983807563782, | |
| "eval_runtime": 477.8813, | |
| "eval_samples_per_second": 6.288, | |
| "eval_steps_per_second": 0.787, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.2554294521507636, | |
| "grad_norm": 5.775433133659037, | |
| "learning_rate": 7.443946188340807e-07, | |
| "loss": 0.0613, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.2778478352248843, | |
| "grad_norm": 7.504088139984216, | |
| "learning_rate": 7.219730941704035e-07, | |
| "loss": 0.0625, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.3002662182990052, | |
| "grad_norm": 12.423874958836857, | |
| "learning_rate": 6.995515695067265e-07, | |
| "loss": 0.0541, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.322684601373126, | |
| "grad_norm": 13.037612748873672, | |
| "learning_rate": 6.771300448430492e-07, | |
| "loss": 0.0638, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.3451029844472466, | |
| "grad_norm": 7.643633731840427, | |
| "learning_rate": 6.547085201793722e-07, | |
| "loss": 0.0528, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.3451029844472466, | |
| "eval_loss": 0.06060384213924408, | |
| "eval_runtime": 477.628, | |
| "eval_samples_per_second": 6.292, | |
| "eval_steps_per_second": 0.787, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.3675213675213675, | |
| "grad_norm": 9.627024940339941, | |
| "learning_rate": 6.322869955156951e-07, | |
| "loss": 0.0569, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.3899397505954882, | |
| "grad_norm": 14.780208722543843, | |
| "learning_rate": 6.098654708520179e-07, | |
| "loss": 0.0504, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.4123581336696092, | |
| "grad_norm": 7.999021557044027, | |
| "learning_rate": 5.874439461883408e-07, | |
| "loss": 0.0609, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.4347765167437299, | |
| "grad_norm": 8.443684543866313, | |
| "learning_rate": 5.650224215246636e-07, | |
| "loss": 0.0531, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.4571948998178508, | |
| "grad_norm": 3.5781417864085086, | |
| "learning_rate": 5.426008968609865e-07, | |
| "loss": 0.0542, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.4571948998178508, | |
| "eval_loss": 0.057715680450201035, | |
| "eval_runtime": 477.9099, | |
| "eval_samples_per_second": 6.288, | |
| "eval_steps_per_second": 0.787, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.4796132828919715, | |
| "grad_norm": 6.394299603034078, | |
| "learning_rate": 5.201793721973094e-07, | |
| "loss": 0.0549, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.5020316659660922, | |
| "grad_norm": 6.251887881398542, | |
| "learning_rate": 4.977578475336322e-07, | |
| "loss": 0.0536, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.524450049040213, | |
| "grad_norm": 4.45363233623469, | |
| "learning_rate": 4.753363228699551e-07, | |
| "loss": 0.0519, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.5468684321143338, | |
| "grad_norm": 7.236100371063919, | |
| "learning_rate": 4.5291479820627797e-07, | |
| "loss": 0.0516, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.5692868151884545, | |
| "grad_norm": 11.475278894303377, | |
| "learning_rate": 4.304932735426009e-07, | |
| "loss": 0.0463, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.5692868151884545, | |
| "eval_loss": 0.05437139794230461, | |
| "eval_runtime": 477.9054, | |
| "eval_samples_per_second": 6.288, | |
| "eval_steps_per_second": 0.787, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.5917051982625754, | |
| "grad_norm": 6.676408323109306, | |
| "learning_rate": 4.0807174887892375e-07, | |
| "loss": 0.0532, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.614123581336696, | |
| "grad_norm": 4.965760634983937, | |
| "learning_rate": 3.856502242152466e-07, | |
| "loss": 0.0545, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.6365419644108168, | |
| "grad_norm": 27.900762279082834, | |
| "learning_rate": 3.632286995515695e-07, | |
| "loss": 0.0544, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.6589603474849377, | |
| "grad_norm": 5.498723625213236, | |
| "learning_rate": 3.4080717488789235e-07, | |
| "loss": 0.0518, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.6813787305590584, | |
| "grad_norm": 5.5744858185141775, | |
| "learning_rate": 3.183856502242152e-07, | |
| "loss": 0.0463, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.6813787305590584, | |
| "eval_loss": 0.05197111889719963, | |
| "eval_runtime": 477.8827, | |
| "eval_samples_per_second": 6.288, | |
| "eval_steps_per_second": 0.787, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.703797113633179, | |
| "grad_norm": 5.741468990313337, | |
| "learning_rate": 2.9596412556053813e-07, | |
| "loss": 0.0454, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.7262154967073, | |
| "grad_norm": 9.331526709026493, | |
| "learning_rate": 2.73542600896861e-07, | |
| "loss": 0.0448, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.748633879781421, | |
| "grad_norm": 8.801924259928679, | |
| "learning_rate": 2.5112107623318386e-07, | |
| "loss": 0.0491, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.7710522628555414, | |
| "grad_norm": 3.9216347240361435, | |
| "learning_rate": 2.2869955156950672e-07, | |
| "loss": 0.0456, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.7934706459296623, | |
| "grad_norm": 5.775711795911055, | |
| "learning_rate": 2.062780269058296e-07, | |
| "loss": 0.0434, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.7934706459296623, | |
| "eval_loss": 0.04976892098784447, | |
| "eval_runtime": 477.8498, | |
| "eval_samples_per_second": 6.289, | |
| "eval_steps_per_second": 0.787, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.8158890290037832, | |
| "grad_norm": 15.832420424237657, | |
| "learning_rate": 1.8385650224215245e-07, | |
| "loss": 0.0549, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.838307412077904, | |
| "grad_norm": 11.331927257096979, | |
| "learning_rate": 1.6143497757847531e-07, | |
| "loss": 0.0479, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.8607257951520246, | |
| "grad_norm": 3.4779552694261846, | |
| "learning_rate": 1.390134529147982e-07, | |
| "loss": 0.045, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.8831441782261455, | |
| "grad_norm": 14.096117466941674, | |
| "learning_rate": 1.1659192825112107e-07, | |
| "loss": 0.0468, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.9055625613002662, | |
| "grad_norm": 4.258023342595773, | |
| "learning_rate": 9.417040358744393e-08, | |
| "loss": 0.043, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.9055625613002662, | |
| "eval_loss": 0.047411367297172546, | |
| "eval_runtime": 478.0538, | |
| "eval_samples_per_second": 6.286, | |
| "eval_steps_per_second": 0.787, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.927980944374387, | |
| "grad_norm": 4.472323710821738, | |
| "learning_rate": 7.174887892376681e-08, | |
| "loss": 0.0485, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.9503993274485079, | |
| "grad_norm": 8.238550394871776, | |
| "learning_rate": 4.932735426008968e-08, | |
| "loss": 0.0461, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.9728177105226286, | |
| "grad_norm": 3.661996756017377, | |
| "learning_rate": 2.6905829596412556e-08, | |
| "loss": 0.0414, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.9952360935967492, | |
| "grad_norm": 4.240752961770031, | |
| "learning_rate": 4.484304932735426e-09, | |
| "loss": 0.0496, | |
| "step": 890 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 892, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 360, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.1364495346169283e+18, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |