| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.9634684865515857, |
| "eval_steps": 500, |
| "global_step": 2400, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0, |
| "learning_rate": 2.666666666666667e-06, |
| "loss": 42.4441, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 5.333333333333334e-06, |
| "loss": 37.1075, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 8.000000000000001e-06, |
| "loss": 27.8987, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 1.0666666666666667e-05, |
| "loss": 24.6867, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 1.3333333333333333e-05, |
| "loss": 22.6003, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 1.6000000000000003e-05, |
| "loss": 21.1455, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 1.866666666666667e-05, |
| "loss": 21.6205, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 1.9999788644103418e-05, |
| "loss": 20.1194, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 1.999809785053594e-05, |
| "loss": 20.1889, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 1.9994716549285312e-05, |
| "loss": 19.7878, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 1.9989645312071867e-05, |
| "loss": 19.5043, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.9982884996355248e-05, |
| "loss": 19.2639, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.9974436745189444e-05, |
| "loss": 19.0332, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.9964301987029523e-05, |
| "loss": 19.3774, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.9952482435490094e-05, |
| "loss": 18.8272, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.9938980089055565e-05, |
| "loss": 18.9226, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.992379723074224e-05, |
| "loss": 19.2136, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.9906936427712295e-05, |
| "loss": 18.6037, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.9888400530839713e-05, |
| "loss": 18.6965, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.986819267422826e-05, |
| "loss": 18.4695, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.9846316274681547e-05, |
| "loss": 18.5388, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.982277503112531e-05, |
| "loss": 18.6583, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.979757292398201e-05, |
| "loss": 18.1656, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.977071421449776e-05, |
| "loss": 18.5414, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.9742203444021878e-05, |
| "loss": 18.3158, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.9712045433238972e-05, |
| "loss": 18.0978, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.9680245281353894e-05, |
| "loss": 18.0812, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.9646808365229506e-05, |
| "loss": 17.8044, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.961174033847757e-05, |
| "loss": 17.9748, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9575047130502813e-05, |
| "loss": 17.7447, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.953673494550037e-05, |
| "loss": 18.2907, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.949681026140674e-05, |
| "loss": 18.1314, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.9455279828804526e-05, |
| "loss": 17.9227, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9412150669780952e-05, |
| "loss": 18.4178, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.936743007674063e-05, |
| "loss": 17.8701, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9321125611172468e-05, |
| "loss": 18.1603, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.92732451023712e-05, |
| "loss": 17.8314, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.9223796646113567e-05, |
| "loss": 17.5233, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.9172788603289453e-05, |
| "loss": 18.2041, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.9120229598488218e-05, |
| "loss": 18.1653, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.9066128518540408e-05, |
| "loss": 17.6848, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.9010494511015164e-05, |
| "loss": 18.1238, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.8953336982673506e-05, |
| "loss": 17.956, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.8894665597877824e-05, |
| "loss": 17.699, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.8834490276957788e-05, |
| "loss": 17.6947, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.8772821194533e-05, |
| "loss": 17.6464, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.8709668777792633e-05, |
| "loss": 17.89, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.8645043704732367e-05, |
| "loss": 17.6851, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.8578956902348945e-05, |
| "loss": 17.3849, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.851141954479256e-05, |
| "loss": 17.8666, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.844244305147755e-05, |
| "loss": 17.4788, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.8372039085151537e-05, |
| "loss": 17.4202, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.830021954992345e-05, |
| "loss": 17.2379, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.8226996589250775e-05, |
| "loss": 17.5357, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.8152382583886272e-05, |
| "loss": 17.9992, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.8076390149784622e-05, |
| "loss": 17.7149, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.7999032135969265e-05, |
| "loss": 17.5005, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.7920321622359876e-05, |
| "loss": 17.4384, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.784027191756075e-05, |
| "loss": 17.0999, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.7758896556610547e-05, |
| "loss": 17.203, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.7676209298693765e-05, |
| "loss": 17.3599, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.759222412481428e-05, |
| "loss": 17.7387, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.75069552354314e-05, |
| "loss": 17.4499, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.7420417048058816e-05, |
| "loss": 17.396, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.7332624194826847e-05, |
| "loss": 17.441, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.7243591520008384e-05, |
| "loss": 17.0808, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.7153334077508983e-05, |
| "loss": 17.1419, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.7061867128321524e-05, |
| "loss": 17.3086, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.6969206137945797e-05, |
| "loss": 17.3577, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.6875366773773604e-05, |
| "loss": 17.6731, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.678036490243962e-05, |
| "loss": 17.3216, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.6684216587138647e-05, |
| "loss": 17.3844, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.658693808490959e-05, |
| "loss": 17.4103, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.6488545843886677e-05, |
| "loss": 17.0814, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.6389056500518343e-05, |
| "loss": 17.3081, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.6288486876754314e-05, |
| "loss": 17.148, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.618685397720128e-05, |
| "loss": 16.8685, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.6084174986247738e-05, |
| "loss": 17.4187, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.598046726515836e-05, |
| "loss": 17.4885, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.5875748349138533e-05, |
| "loss": 17.2386, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.5770035944369456e-05, |
| "loss": 16.9695, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.5663347925014302e-05, |
| "loss": 17.3497, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.5555702330196024e-05, |
| "loss": 16.9592, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.5447117360947244e-05, |
| "loss": 17.5106, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.5337611377132757e-05, |
| "loss": 17.2786, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.522720289434521e-05, |
| "loss": 17.4009, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.511591058077441e-05, |
| "loss": 17.2044, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.500375325405087e-05, |
| "loss": 17.1285, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.489074987806406e-05, |
| "loss": 17.3726, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.477691955975594e-05, |
| "loss": 17.1684, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.46622815458903e-05, |
| "loss": 17.2715, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.454685521979846e-05, |
| "loss": 17.3633, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.443066009810188e-05, |
| "loss": 17.1012, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.4313715827412243e-05, |
| "loss": 16.7467, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.4196042181009525e-05, |
| "loss": 16.741, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.4077659055498695e-05, |
| "loss": 17.0569, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.3958586467445532e-05, |
| "loss": 17.0277, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.383884454999215e-05, |
| "loss": 16.8944, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.3718453549452843e-05, |
| "loss": 17.1323, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.3597433821890787e-05, |
| "loss": 16.9379, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.3475805829676149e-05, |
| "loss": 16.9814, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.3353590138026273e-05, |
| "loss": 17.1369, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.323080741152845e-05, |
| "loss": 16.8629, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.3107478410645875e-05, |
| "loss": 16.8607, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.2983623988207432e-05, |
| "loss": 16.6911, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.28592650858818e-05, |
| "loss": 16.9764, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.2734422730636617e-05, |
| "loss": 16.8255, |
| "step": 1070 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.2609118031183144e-05, |
| "loss": 16.8564, |
| "step": 1080 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.2483372174407155e-05, |
| "loss": 16.9249, |
| "step": 1090 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.2357206421786611e-05, |
| "loss": 16.8046, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.2230642105796674e-05, |
| "loss": 17.0138, |
| "step": 1110 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.2103700626302784e-05, |
| "loss": 16.8029, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.197640344694228e-05, |
| "loss": 16.7554, |
| "step": 1130 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.1848772091495287e-05, |
| "loss": 16.62, |
| "step": 1140 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.1720828140245393e-05, |
| "loss": 16.8458, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.1592593226330802e-05, |
| "loss": 17.1714, |
| "step": 1160 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.1464089032086547e-05, |
| "loss": 17.0085, |
| "step": 1170 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.1335337285378359e-05, |
| "loss": 16.8423, |
| "step": 1180 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.1206359755928865e-05, |
| "loss": 16.8152, |
| "step": 1190 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.1077178251636702e-05, |
| "loss": 16.619, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.0947814614889174e-05, |
| "loss": 16.7087, |
| "step": 1210 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.0818290718869068e-05, |
| "loss": 16.7935, |
| "step": 1220 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.0688628463856287e-05, |
| "loss": 17.4941, |
| "step": 1230 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.055884977352487e-05, |
| "loss": 17.0523, |
| "step": 1240 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.0428976591236082e-05, |
| "loss": 16.7451, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.029903087632817e-05, |
| "loss": 16.8536, |
| "step": 1260 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.0169034600403404e-05, |
| "loss": 16.6058, |
| "step": 1270 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.003900974361306e-05, |
| "loss": 16.6489, |
| "step": 1280 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 9.908978290940948e-06, |
| "loss": 16.989, |
| "step": 1290 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 9.778962228486138e-06, |
| "loss": 16.885, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 9.648983539745468e-06, |
| "loss": 16.9012, |
| "step": 1310 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 9.51906420189652e-06, |
| "loss": 16.8678, |
| "step": 1320 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 9.38922618208166e-06, |
| "loss": 16.7718, |
| "step": 1330 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 9.259491433693751e-06, |
| "loss": 16.7869, |
| "step": 1340 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 9.129881892664232e-06, |
| "loss": 16.9432, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 9.00041947375411e-06, |
| "loss": 16.8835, |
| "step": 1360 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 8.871126066848552e-06, |
| "loss": 16.8081, |
| "step": 1370 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 8.742023533255677e-06, |
| "loss": 16.6375, |
| "step": 1380 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 8.613133702010196e-06, |
| "loss": 16.7379, |
| "step": 1390 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 8.484478366182472e-06, |
| "loss": 16.5849, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 8.356079279193703e-06, |
| "loss": 16.6192, |
| "step": 1410 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 8.227958151137773e-06, |
| "loss": 16.4591, |
| "step": 1420 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 8.10013664511047e-06, |
| "loss": 16.5343, |
| "step": 1430 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 7.9726363735466e-06, |
| "loss": 16.9542, |
| "step": 1440 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 7.84547889456571e-06, |
| "loss": 16.8205, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 7.718685708326965e-06, |
| "loss": 16.5807, |
| "step": 1460 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 7.592278253393859e-06, |
| "loss": 16.5174, |
| "step": 1470 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 7.466277903109291e-06, |
| "loss": 16.6076, |
| "step": 1480 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 7.340705961981722e-06, |
| "loss": 16.7884, |
| "step": 1490 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 7.215583662082939e-06, |
| "loss": 16.8359, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 7.090932159458067e-06, |
| "loss": 16.5588, |
| "step": 1510 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 6.966772530548448e-06, |
| "loss": 16.6256, |
| "step": 1520 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 6.843125768627983e-06, |
| "loss": 16.7764, |
| "step": 1530 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 6.720012780253509e-06, |
| "loss": 16.9245, |
| "step": 1540 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 6.597454381729873e-06, |
| "loss": 16.5384, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 6.475471295590248e-06, |
| "loss": 16.6298, |
| "step": 1560 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 6.354084147092296e-06, |
| "loss": 16.752, |
| "step": 1570 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 6.2333134607308e-06, |
| "loss": 16.6684, |
| "step": 1580 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 6.113179656767319e-06, |
| "loss": 16.4827, |
| "step": 1590 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 5.993703047777468e-06, |
| "loss": 16.5697, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 5.874903835216417e-06, |
| "loss": 16.4795, |
| "step": 1610 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 5.756802106003148e-06, |
| "loss": 16.8255, |
| "step": 1620 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 5.639417829124132e-06, |
| "loss": 16.6714, |
| "step": 1630 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 5.52277085225688e-06, |
| "loss": 16.7242, |
| "step": 1640 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 5.406880898414069e-06, |
| "loss": 16.4696, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 5.291767562608705e-06, |
| "loss": 16.5084, |
| "step": 1660 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 5.177450308540928e-06, |
| "loss": 16.5646, |
| "step": 1670 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 5.0639484653070535e-06, |
| "loss": 16.4026, |
| "step": 1680 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 4.95128122413135e-06, |
| "loss": 16.3032, |
| "step": 1690 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 4.83946763512111e-06, |
| "loss": 16.5088, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 4.7285266040456255e-06, |
| "loss": 16.3905, |
| "step": 1710 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 4.618476889139538e-06, |
| "loss": 16.4207, |
| "step": 1720 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 4.50933709793113e-06, |
| "loss": 16.5308, |
| "step": 1730 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 4.401125684096124e-06, |
| "loss": 16.521, |
| "step": 1740 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 4.293860944337482e-06, |
| "loss": 16.3123, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 4.1875610152917225e-06, |
| "loss": 16.6079, |
| "step": 1760 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 4.082243870462362e-06, |
| "loss": 16.5056, |
| "step": 1770 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 3.977927317180864e-06, |
| "loss": 16.4315, |
| "step": 1780 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 3.87462899359575e-06, |
| "loss": 16.1708, |
| "step": 1790 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 3.772366365690283e-06, |
| "loss": 16.3817, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 3.6711567243292547e-06, |
| "loss": 16.5455, |
| "step": 1810 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 3.5710171823354145e-06, |
| "loss": 16.3721, |
| "step": 1820 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 3.4719646715959777e-06, |
| "loss": 16.4659, |
| "step": 1830 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 3.3740159401997173e-06, |
| "loss": 16.5392, |
| "step": 1840 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 3.2771875496051743e-06, |
| "loss": 16.6646, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 3.181495871840379e-06, |
| "loss": 16.3682, |
| "step": 1860 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 3.0869570867346167e-06, |
| "loss": 16.5935, |
| "step": 1870 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 2.9935871791827166e-06, |
| "loss": 16.539, |
| "step": 1880 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 2.9014019364422606e-06, |
| "loss": 16.1143, |
| "step": 1890 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 2.8104169454642293e-06, |
| "loss": 16.5375, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 2.7206475902575225e-06, |
| "loss": 16.3927, |
| "step": 1910 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 2.6321090492877823e-06, |
| "loss": 16.3639, |
| "step": 1920 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 2.544816292910962e-06, |
| "loss": 16.3929, |
| "step": 1930 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 2.458784080842127e-06, |
| "loss": 16.3984, |
| "step": 1940 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 2.3740269596597943e-06, |
| "loss": 16.3938, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 2.2905592603463888e-06, |
| "loss": 16.3884, |
| "step": 1960 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 2.2083950958651103e-06, |
| "loss": 16.3289, |
| "step": 1970 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 2.1275483587736577e-06, |
| "loss": 16.3735, |
| "step": 1980 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 2.048032718875255e-06, |
| "loss": 16.4073, |
| "step": 1990 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 1.969861620907314e-06, |
| "loss": 16.4259, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.8930482822681473e-06, |
| "loss": 16.3312, |
| "step": 2010 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.8176056907821482e-06, |
| "loss": 16.2809, |
| "step": 2020 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.7435466025037684e-06, |
| "loss": 16.2617, |
| "step": 2030 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.6708835395606704e-06, |
| "loss": 16.6299, |
| "step": 2040 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.5996287880364736e-06, |
| "loss": 16.2559, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.5297943958933748e-06, |
| "loss": 16.4145, |
| "step": 2060 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.4613921709350342e-06, |
| "loss": 16.3823, |
| "step": 2070 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.3944336788100976e-06, |
| "loss": 16.3725, |
| "step": 2080 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.3289302410566318e-06, |
| "loss": 16.2138, |
| "step": 2090 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.2648929331878423e-06, |
| "loss": 16.2873, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.202332582819402e-06, |
| "loss": 16.4605, |
| "step": 2110 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.141259767838675e-06, |
| "loss": 16.1904, |
| "step": 2120 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.0816848146161895e-06, |
| "loss": 16.4713, |
| "step": 2130 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.0236177962596173e-06, |
| "loss": 16.143, |
| "step": 2140 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 9.670685309105786e-07, |
| "loss": 16.2049, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 9.120465800845723e-07, |
| "loss": 16.3906, |
| "step": 2160 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 8.585612470542859e-07, |
| "loss": 16.1819, |
| "step": 2170 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 8.06621575276556e-07, |
| "loss": 16.2495, |
| "step": 2180 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 7.562363468632949e-07, |
| "loss": 16.3325, |
| "step": 2190 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 7.074140810965724e-07, |
| "loss": 16.3265, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 6.601630329881525e-07, |
| "loss": 16.3694, |
| "step": 2210 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 6.14491191883716e-07, |
| "loss": 16.2883, |
| "step": 2220 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 5.704062801119947e-07, |
| "loss": 16.2978, |
| "step": 2230 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 5.279157516790545e-07, |
| "loss": 16.2589, |
| "step": 2240 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 4.870267910079618e-07, |
| "loss": 16.6149, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 4.4774631172400663e-07, |
| "loss": 16.2879, |
| "step": 2260 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 4.100809554857343e-07, |
| "loss": 16.2822, |
| "step": 2270 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 3.740370908619528e-07, |
| "loss": 16.3097, |
| "step": 2280 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 3.396208122549194e-07, |
| "loss": 16.3474, |
| "step": 2290 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 3.0683793886986943e-07, |
| "loss": 16.2617, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 2.7569401373110595e-07, |
| "loss": 16.1619, |
| "step": 2310 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 2.461943027447555e-07, |
| "loss": 16.2584, |
| "step": 2320 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 2.1834379380839655e-07, |
| "loss": 16.1723, |
| "step": 2330 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.921471959676957e-07, |
| "loss": 16.2365, |
| "step": 2340 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.676089386201818e-07, |
| "loss": 16.2976, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.4473317076631355e-07, |
| "loss": 16.3046, |
| "step": 2360 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.2352376030795753e-07, |
| "loss": 16.2992, |
| "step": 2370 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 1.0398429339438353e-07, |
| "loss": 16.2474, |
| "step": 2380 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 8.61180738159173e-08, |
| "loss": 16.2852, |
| "step": 2390 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 6.992812244532188e-08, |
| "loss": 16.3501, |
| "step": 2400 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 2491, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 100, |
| "total_flos": 0.0, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|