| { | |
| "best_metric": 1.0115959644317627, | |
| "best_model_checkpoint": "checkpoints/instrucode/text_models/llama-2-13b/checkpoint-6200", | |
| "epoch": 4.9916006719462445, | |
| "eval_steps": 200, | |
| "global_step": 15600, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1e-05, | |
| "loss": 1.8592, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2e-05, | |
| "loss": 1.7835, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3e-05, | |
| "loss": 1.7661, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4e-05, | |
| "loss": 1.5736, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 5e-05, | |
| "loss": 1.3416, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 6e-05, | |
| "loss": 1.1455, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 7e-05, | |
| "loss": 1.0841, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 7.900000000000001e-05, | |
| "loss": 1.0352, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 8.900000000000001e-05, | |
| "loss": 1.0574, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 9.900000000000001e-05, | |
| "loss": 1.0802, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.994202898550726e-05, | |
| "loss": 1.027, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.987761674718196e-05, | |
| "loss": 1.0086, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.981320450885668e-05, | |
| "loss": 1.0188, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.974879227053141e-05, | |
| "loss": 1.0513, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.968438003220612e-05, | |
| "loss": 1.074, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.962640901771337e-05, | |
| "loss": 1.0432, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 9.956199677938809e-05, | |
| "loss": 1.0586, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 9.949758454106281e-05, | |
| "loss": 1.0195, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 9.943961352657005e-05, | |
| "loss": 1.1137, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 9.937520128824477e-05, | |
| "loss": 1.012, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "eval_loss": 1.050466537475586, | |
| "eval_runtime": 186.7925, | |
| "eval_samples_per_second": 10.707, | |
| "eval_steps_per_second": 1.338, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 9.931078904991949e-05, | |
| "loss": 1.0544, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 9.924637681159421e-05, | |
| "loss": 1.0796, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 9.918196457326893e-05, | |
| "loss": 1.0681, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 9.911755233494364e-05, | |
| "loss": 1.066, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 9.905314009661836e-05, | |
| "loss": 1.0483, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 9.898872785829308e-05, | |
| "loss": 1.0917, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.89243156199678e-05, | |
| "loss": 1.0391, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.885990338164252e-05, | |
| "loss": 1.023, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.879549114331724e-05, | |
| "loss": 1.072, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 9.873107890499196e-05, | |
| "loss": 1.038, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 9.866666666666668e-05, | |
| "loss": 1.0242, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 9.86022544283414e-05, | |
| "loss": 1.0045, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 9.85378421900161e-05, | |
| "loss": 1.0362, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 9.847342995169082e-05, | |
| "loss": 1.0272, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 9.840901771336555e-05, | |
| "loss": 0.9938, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 9.834460547504027e-05, | |
| "loss": 1.0338, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 9.828019323671497e-05, | |
| "loss": 1.0146, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 9.821578099838969e-05, | |
| "loss": 1.0324, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 9.815136876006442e-05, | |
| "loss": 1.0159, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 9.808695652173914e-05, | |
| "loss": 1.0283, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "eval_loss": 1.041812777519226, | |
| "eval_runtime": 187.3424, | |
| "eval_samples_per_second": 10.676, | |
| "eval_steps_per_second": 1.334, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 9.802254428341385e-05, | |
| "loss": 1.0189, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 9.795813204508857e-05, | |
| "loss": 1.0623, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.789371980676329e-05, | |
| "loss": 1.0266, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.7829307568438e-05, | |
| "loss": 1.0436, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.776489533011272e-05, | |
| "loss": 1.0393, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.770048309178744e-05, | |
| "loss": 1.0452, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.763607085346216e-05, | |
| "loss": 1.06, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.757165861513688e-05, | |
| "loss": 1.0115, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 9.75072463768116e-05, | |
| "loss": 1.0523, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 9.744283413848632e-05, | |
| "loss": 1.028, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 9.737842190016104e-05, | |
| "loss": 1.0984, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 9.731400966183575e-05, | |
| "loss": 1.0312, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 9.724959742351047e-05, | |
| "loss": 1.0687, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 9.718518518518519e-05, | |
| "loss": 1.0485, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 9.712077294685991e-05, | |
| "loss": 1.0423, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 9.705636070853462e-05, | |
| "loss": 1.0005, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 9.699194847020935e-05, | |
| "loss": 1.0217, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 9.692753623188407e-05, | |
| "loss": 1.0233, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 9.686312399355879e-05, | |
| "loss": 1.0381, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 9.679871175523349e-05, | |
| "loss": 1.0045, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "eval_loss": 1.0341911315917969, | |
| "eval_runtime": 190.2864, | |
| "eval_samples_per_second": 10.51, | |
| "eval_steps_per_second": 1.314, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.673429951690822e-05, | |
| "loss": 1.0345, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.666988727858294e-05, | |
| "loss": 1.0507, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.660547504025766e-05, | |
| "loss": 1.0261, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.654106280193237e-05, | |
| "loss": 1.0063, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 9.647665056360708e-05, | |
| "loss": 1.011, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 9.641223832528182e-05, | |
| "loss": 0.9829, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 9.634782608695652e-05, | |
| "loss": 1.0007, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 9.628341384863124e-05, | |
| "loss": 1.0159, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 9.621900161030596e-05, | |
| "loss": 1.0307, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 9.615458937198069e-05, | |
| "loss": 0.9949, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 9.60901771336554e-05, | |
| "loss": 1.0362, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 9.602576489533011e-05, | |
| "loss": 1.0397, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 9.596135265700483e-05, | |
| "loss": 1.0709, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.589694041867955e-05, | |
| "loss": 1.0271, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.583252818035427e-05, | |
| "loss": 1.0336, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.576811594202899e-05, | |
| "loss": 0.9954, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 9.570370370370371e-05, | |
| "loss": 0.9927, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 9.563929146537843e-05, | |
| "loss": 1.0313, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 9.557487922705315e-05, | |
| "loss": 1.0671, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.551046698872786e-05, | |
| "loss": 1.0038, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "eval_loss": 1.0327339172363281, | |
| "eval_runtime": 187.9065, | |
| "eval_samples_per_second": 10.644, | |
| "eval_steps_per_second": 1.33, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.544605475040258e-05, | |
| "loss": 1.0013, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.53816425120773e-05, | |
| "loss": 0.9574, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 9.531723027375202e-05, | |
| "loss": 1.0208, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 9.525281803542674e-05, | |
| "loss": 1.0067, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 9.518840579710146e-05, | |
| "loss": 1.0408, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.512399355877616e-05, | |
| "loss": 1.0299, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.505958132045088e-05, | |
| "loss": 1.0195, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.499516908212561e-05, | |
| "loss": 1.0095, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.493075684380033e-05, | |
| "loss": 1.0174, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.486634460547504e-05, | |
| "loss": 1.0238, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.480193236714976e-05, | |
| "loss": 1.0506, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.473752012882449e-05, | |
| "loss": 1.0308, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.467310789049921e-05, | |
| "loss": 1.018, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.460869565217391e-05, | |
| "loss": 1.058, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.454428341384863e-05, | |
| "loss": 1.0226, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 9.447987117552335e-05, | |
| "loss": 1.0569, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 9.441545893719808e-05, | |
| "loss": 0.988, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 9.435104669887279e-05, | |
| "loss": 1.0036, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 9.42866344605475e-05, | |
| "loss": 1.012, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 9.422222222222223e-05, | |
| "loss": 0.9998, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "eval_loss": 1.0286028385162354, | |
| "eval_runtime": 186.8358, | |
| "eval_samples_per_second": 10.705, | |
| "eval_steps_per_second": 1.338, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 9.415780998389694e-05, | |
| "loss": 1.0794, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.409339774557166e-05, | |
| "loss": 1.0211, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.402898550724638e-05, | |
| "loss": 1.039, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.39645732689211e-05, | |
| "loss": 1.0337, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 9.390016103059582e-05, | |
| "loss": 1.0546, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 9.383574879227054e-05, | |
| "loss": 1.0126, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 9.377133655394526e-05, | |
| "loss": 1.01, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 9.370692431561997e-05, | |
| "loss": 1.005, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 9.364251207729468e-05, | |
| "loss": 1.0742, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 9.357809983896941e-05, | |
| "loss": 1.0131, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.351368760064413e-05, | |
| "loss": 1.0148, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.344927536231885e-05, | |
| "loss": 1.0298, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.338486312399355e-05, | |
| "loss": 1.0235, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.332045088566829e-05, | |
| "loss": 1.0461, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 9.3256038647343e-05, | |
| "loss": 1.0328, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 9.319162640901772e-05, | |
| "loss": 1.0336, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 9.312721417069243e-05, | |
| "loss": 1.0651, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.306280193236715e-05, | |
| "loss": 0.9966, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.299838969404188e-05, | |
| "loss": 1.0428, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.293397745571659e-05, | |
| "loss": 1.0506, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "eval_loss": 1.0272533893585205, | |
| "eval_runtime": 193.4811, | |
| "eval_samples_per_second": 10.337, | |
| "eval_steps_per_second": 1.292, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.28695652173913e-05, | |
| "loss": 1.0431, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.280515297906602e-05, | |
| "loss": 0.9844, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.274074074074076e-05, | |
| "loss": 1.0488, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 9.267632850241546e-05, | |
| "loss": 0.9965, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 9.261191626409018e-05, | |
| "loss": 1.0397, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 9.25475040257649e-05, | |
| "loss": 1.0347, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 9.248309178743962e-05, | |
| "loss": 0.9955, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 9.241867954911434e-05, | |
| "loss": 0.9883, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 9.235426731078905e-05, | |
| "loss": 0.9947, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.228985507246377e-05, | |
| "loss": 1.0275, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.222544283413849e-05, | |
| "loss": 0.9823, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.216103059581321e-05, | |
| "loss": 1.0227, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 9.209661835748793e-05, | |
| "loss": 1.0172, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 9.203220611916265e-05, | |
| "loss": 0.9689, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 9.196779388083737e-05, | |
| "loss": 1.0471, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.190338164251207e-05, | |
| "loss": 1.0608, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.18389694041868e-05, | |
| "loss": 1.0556, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.177455716586152e-05, | |
| "loss": 0.9823, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.171014492753624e-05, | |
| "loss": 1.0325, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 9.164573268921095e-05, | |
| "loss": 1.0067, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "eval_loss": 1.0259579420089722, | |
| "eval_runtime": 187.0176, | |
| "eval_samples_per_second": 10.694, | |
| "eval_steps_per_second": 1.337, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 9.158132045088568e-05, | |
| "loss": 1.0186, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 9.15169082125604e-05, | |
| "loss": 1.0293, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 9.14524959742351e-05, | |
| "loss": 1.0173, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 9.138808373590982e-05, | |
| "loss": 1.064, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 9.132367149758455e-05, | |
| "loss": 1.0415, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 9.125925925925927e-05, | |
| "loss": 1.0465, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 9.119484702093398e-05, | |
| "loss": 1.0284, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 9.11304347826087e-05, | |
| "loss": 1.036, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 9.106602254428341e-05, | |
| "loss": 1.0351, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 9.100161030595815e-05, | |
| "loss": 1.0446, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 9.093719806763285e-05, | |
| "loss": 1.0019, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 9.087278582930757e-05, | |
| "loss": 1.018, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 9.080837359098229e-05, | |
| "loss": 1.0221, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 9.074396135265701e-05, | |
| "loss": 1.047, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 9.067954911433173e-05, | |
| "loss": 1.0105, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 9.061513687600645e-05, | |
| "loss": 1.0245, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 9.055072463768116e-05, | |
| "loss": 1.0513, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 9.048631239935588e-05, | |
| "loss": 1.0033, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 9.04219001610306e-05, | |
| "loss": 0.9867, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 9.035748792270532e-05, | |
| "loss": 1.0351, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "eval_loss": 1.0244308710098267, | |
| "eval_runtime": 186.9957, | |
| "eval_samples_per_second": 10.695, | |
| "eval_steps_per_second": 1.337, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.029307568438004e-05, | |
| "loss": 1.0109, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.022866344605474e-05, | |
| "loss": 1.0464, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.016425120772948e-05, | |
| "loss": 1.0204, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.00998389694042e-05, | |
| "loss": 1.0605, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.003542673107891e-05, | |
| "loss": 1.0209, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 8.997101449275362e-05, | |
| "loss": 1.0588, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 8.990660225442834e-05, | |
| "loss": 1.0094, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 8.984219001610307e-05, | |
| "loss": 1.1152, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 8.977777777777779e-05, | |
| "loss": 1.0059, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 8.97133655394525e-05, | |
| "loss": 1.042, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 8.964895330112721e-05, | |
| "loss": 0.9825, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 8.958454106280194e-05, | |
| "loss": 1.0188, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 8.952012882447666e-05, | |
| "loss": 1.0384, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 8.945571658615137e-05, | |
| "loss": 1.0454, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 8.939130434782609e-05, | |
| "loss": 1.0252, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 8.932689210950082e-05, | |
| "loss": 1.0093, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 8.926247987117552e-05, | |
| "loss": 1.0079, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 8.919806763285024e-05, | |
| "loss": 1.0162, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 8.913365539452496e-05, | |
| "loss": 1.0445, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 8.906924315619968e-05, | |
| "loss": 1.0151, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "eval_loss": 1.022465467453003, | |
| "eval_runtime": 187.4046, | |
| "eval_samples_per_second": 10.672, | |
| "eval_steps_per_second": 1.334, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 8.90048309178744e-05, | |
| "loss": 1.0026, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 8.894041867954912e-05, | |
| "loss": 1.0338, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 8.887600644122384e-05, | |
| "loss": 1.0004, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 8.881159420289856e-05, | |
| "loss": 1.0013, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 8.874718196457327e-05, | |
| "loss": 0.9698, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 8.868276972624799e-05, | |
| "loss": 1.0308, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 8.861835748792271e-05, | |
| "loss": 0.9898, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 8.855394524959743e-05, | |
| "loss": 1.0491, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 8.848953301127214e-05, | |
| "loss": 1.0242, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 8.842512077294687e-05, | |
| "loss": 1.0419, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 8.836070853462159e-05, | |
| "loss": 1.0346, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 8.82962962962963e-05, | |
| "loss": 0.9888, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 8.823188405797101e-05, | |
| "loss": 1.0115, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 8.816747181964574e-05, | |
| "loss": 1.0193, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 8.810305958132046e-05, | |
| "loss": 1.0633, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 8.803864734299517e-05, | |
| "loss": 1.048, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 8.797423510466989e-05, | |
| "loss": 1.0345, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 8.79098228663446e-05, | |
| "loss": 1.0477, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 8.784541062801934e-05, | |
| "loss": 1.0232, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 8.778099838969404e-05, | |
| "loss": 1.0614, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "eval_loss": 1.0219651460647583, | |
| "eval_runtime": 187.5045, | |
| "eval_samples_per_second": 10.666, | |
| "eval_steps_per_second": 1.333, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 8.771658615136876e-05, | |
| "loss": 1.0489, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 8.765217391304348e-05, | |
| "loss": 0.9984, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 8.758776167471821e-05, | |
| "loss": 1.0021, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 8.752334943639292e-05, | |
| "loss": 1.0382, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 8.745893719806763e-05, | |
| "loss": 1.0086, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 8.739452495974235e-05, | |
| "loss": 1.0211, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 8.733011272141707e-05, | |
| "loss": 0.9975, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 8.726570048309179e-05, | |
| "loss": 1.0468, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 8.720128824476651e-05, | |
| "loss": 1.0646, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 8.713687600644123e-05, | |
| "loss": 1.0311, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 8.707246376811595e-05, | |
| "loss": 1.0253, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 8.700805152979067e-05, | |
| "loss": 1.0511, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 8.694363929146538e-05, | |
| "loss": 1.0652, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 8.68792270531401e-05, | |
| "loss": 1.0397, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 8.681481481481482e-05, | |
| "loss": 1.0262, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 8.675040257648954e-05, | |
| "loss": 1.0374, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 8.668599033816426e-05, | |
| "loss": 1.0446, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 8.662157809983898e-05, | |
| "loss": 1.0278, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 8.655716586151368e-05, | |
| "loss": 0.9984, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 8.64927536231884e-05, | |
| "loss": 1.0373, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "eval_loss": 1.020651936531067, | |
| "eval_runtime": 187.3402, | |
| "eval_samples_per_second": 10.676, | |
| "eval_steps_per_second": 1.334, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 8.642834138486313e-05, | |
| "loss": 1.0451, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 8.636392914653785e-05, | |
| "loss": 1.0249, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 8.629951690821256e-05, | |
| "loss": 0.9943, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 8.623510466988728e-05, | |
| "loss": 1.0084, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 8.617069243156201e-05, | |
| "loss": 1.0046, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 8.610628019323673e-05, | |
| "loss": 0.982, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 8.604186795491143e-05, | |
| "loss": 1.0438, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 8.597745571658615e-05, | |
| "loss": 1.032, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 8.591304347826087e-05, | |
| "loss": 1.0066, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 8.584863123993559e-05, | |
| "loss": 1.0412, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 8.578421900161031e-05, | |
| "loss": 0.9981, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 8.571980676328503e-05, | |
| "loss": 1.0111, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 8.565539452495974e-05, | |
| "loss": 1.0057, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 8.559098228663446e-05, | |
| "loss": 1.022, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 8.552657004830918e-05, | |
| "loss": 0.9963, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 8.54621578099839e-05, | |
| "loss": 1.0382, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 8.539774557165862e-05, | |
| "loss": 1.0083, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 8.533333333333334e-05, | |
| "loss": 1.0126, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 8.526892109500806e-05, | |
| "loss": 0.9937, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 8.520450885668278e-05, | |
| "loss": 1.0266, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "eval_loss": 1.0203664302825928, | |
| "eval_runtime": 188.9153, | |
| "eval_samples_per_second": 10.587, | |
| "eval_steps_per_second": 1.323, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 8.51400966183575e-05, | |
| "loss": 1.0608, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 8.50756843800322e-05, | |
| "loss": 1.0238, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 8.501127214170693e-05, | |
| "loss": 1.0271, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 8.494685990338165e-05, | |
| "loss": 1.0275, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 8.488244766505637e-05, | |
| "loss": 1.0453, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 8.481803542673107e-05, | |
| "loss": 1.045, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 8.475362318840581e-05, | |
| "loss": 1.0463, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 8.468921095008053e-05, | |
| "loss": 1.0199, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 8.462479871175524e-05, | |
| "loss": 0.9889, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 8.456038647342995e-05, | |
| "loss": 0.9973, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 8.449597423510467e-05, | |
| "loss": 1.041, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 8.44315619967794e-05, | |
| "loss": 1.0169, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 8.43671497584541e-05, | |
| "loss": 1.0221, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 8.430273752012882e-05, | |
| "loss": 1.0357, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 8.423832528180354e-05, | |
| "loss": 0.9808, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 8.417391304347828e-05, | |
| "loss": 1.0055, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 8.410950080515298e-05, | |
| "loss": 1.0113, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.40450885668277e-05, | |
| "loss": 1.0418, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.398067632850242e-05, | |
| "loss": 1.0252, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 8.391626409017714e-05, | |
| "loss": 1.0141, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "eval_loss": 1.0183026790618896, | |
| "eval_runtime": 188.3677, | |
| "eval_samples_per_second": 10.618, | |
| "eval_steps_per_second": 1.327, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.385185185185186e-05, | |
| "loss": 0.9696, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.378743961352657e-05, | |
| "loss": 1.0527, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.372302737520129e-05, | |
| "loss": 1.0197, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.365861513687601e-05, | |
| "loss": 1.0035, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 8.359420289855073e-05, | |
| "loss": 1.0426, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 8.352979066022545e-05, | |
| "loss": 1.0199, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 8.346537842190017e-05, | |
| "loss": 1.0231, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 8.340096618357489e-05, | |
| "loss": 1.0153, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 8.33365539452496e-05, | |
| "loss": 0.9991, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 8.327214170692432e-05, | |
| "loss": 1.0103, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 8.320772946859904e-05, | |
| "loss": 0.9782, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 8.314331723027375e-05, | |
| "loss": 1.0254, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 8.307890499194847e-05, | |
| "loss": 1.0162, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 8.30144927536232e-05, | |
| "loss": 0.9674, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 8.295008051529792e-05, | |
| "loss": 1.0262, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 8.288566827697262e-05, | |
| "loss": 1.0294, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 8.282125603864734e-05, | |
| "loss": 1.0064, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 8.275684380032207e-05, | |
| "loss": 0.9869, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 8.269243156199679e-05, | |
| "loss": 1.0191, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 8.26280193236715e-05, | |
| "loss": 1.0347, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "eval_loss": 1.0175596475601196, | |
| "eval_runtime": 187.2212, | |
| "eval_samples_per_second": 10.683, | |
| "eval_steps_per_second": 1.335, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 8.256360708534622e-05, | |
| "loss": 1.0247, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 8.249919484702093e-05, | |
| "loss": 1.016, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 8.243478260869565e-05, | |
| "loss": 1.0039, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 8.237037037037037e-05, | |
| "loss": 1.0306, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 8.230595813204509e-05, | |
| "loss": 1.0314, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 8.224154589371981e-05, | |
| "loss": 0.9795, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 8.217713365539453e-05, | |
| "loss": 1.0489, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 8.211272141706925e-05, | |
| "loss": 1.0023, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 8.204830917874397e-05, | |
| "loss": 0.9973, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 8.198389694041868e-05, | |
| "loss": 1.0281, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 8.19194847020934e-05, | |
| "loss": 1.0256, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 8.185507246376812e-05, | |
| "loss": 1.015, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 8.179066022544284e-05, | |
| "loss": 0.981, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 8.172624798711756e-05, | |
| "loss": 1.0001, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 8.166183574879226e-05, | |
| "loss": 1.0324, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 8.1597423510467e-05, | |
| "loss": 0.9589, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 8.153301127214171e-05, | |
| "loss": 0.994, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 8.146859903381643e-05, | |
| "loss": 1.0238, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 8.140418679549114e-05, | |
| "loss": 1.0347, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 8.133977455716587e-05, | |
| "loss": 1.0248, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "eval_loss": 1.0177215337753296, | |
| "eval_runtime": 187.3372, | |
| "eval_samples_per_second": 10.676, | |
| "eval_steps_per_second": 1.334, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 8.127536231884059e-05, | |
| "loss": 1.0349, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 8.121095008051531e-05, | |
| "loss": 1.0264, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 8.114653784219001e-05, | |
| "loss": 0.9841, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 8.108212560386473e-05, | |
| "loss": 1.0262, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 8.101771336553946e-05, | |
| "loss": 1.0255, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 8.095330112721417e-05, | |
| "loss": 0.9856, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 8.088888888888889e-05, | |
| "loss": 1.0685, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 8.082447665056361e-05, | |
| "loss": 1.0094, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 8.076006441223834e-05, | |
| "loss": 1.022, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 8.069565217391304e-05, | |
| "loss": 0.9743, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 8.063123993558776e-05, | |
| "loss": 1.0228, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 8.056682769726248e-05, | |
| "loss": 1.0398, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 8.05024154589372e-05, | |
| "loss": 1.0022, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 8.043800322061192e-05, | |
| "loss": 1.0136, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 8.038003220611917e-05, | |
| "loss": 0.9806, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 8.031561996779389e-05, | |
| "loss": 1.0222, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 8.02512077294686e-05, | |
| "loss": 0.9584, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 8.018679549114332e-05, | |
| "loss": 0.9746, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 8.012238325281803e-05, | |
| "loss": 1.0131, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 8.005797101449277e-05, | |
| "loss": 0.977, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "eval_loss": 1.0170557498931885, | |
| "eval_runtime": 187.1066, | |
| "eval_samples_per_second": 10.689, | |
| "eval_steps_per_second": 1.336, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 7.999355877616747e-05, | |
| "loss": 1.0382, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 7.992914653784219e-05, | |
| "loss": 0.958, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 7.986473429951691e-05, | |
| "loss": 1.0106, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 7.980032206119164e-05, | |
| "loss": 1.0137, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 7.973590982286635e-05, | |
| "loss": 0.9941, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 7.967149758454107e-05, | |
| "loss": 1.0061, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 7.960708534621578e-05, | |
| "loss": 1.0035, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 7.95426731078905e-05, | |
| "loss": 1.0023, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 7.947826086956522e-05, | |
| "loss": 0.9759, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 7.941384863123994e-05, | |
| "loss": 0.9744, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 7.934943639291466e-05, | |
| "loss": 1.0105, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 7.928502415458938e-05, | |
| "loss": 0.9845, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 7.92206119162641e-05, | |
| "loss": 1.0166, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 7.915619967793881e-05, | |
| "loss": 0.9917, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 7.909178743961353e-05, | |
| "loss": 0.994, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 7.902737520128824e-05, | |
| "loss": 1.0173, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 7.896296296296297e-05, | |
| "loss": 1.0043, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 7.889855072463769e-05, | |
| "loss": 1.0058, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 7.883413848631241e-05, | |
| "loss": 1.0035, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 7.876972624798711e-05, | |
| "loss": 0.9418, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "eval_loss": 1.017248511314392, | |
| "eval_runtime": 187.1833, | |
| "eval_samples_per_second": 10.685, | |
| "eval_steps_per_second": 1.336, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 7.870531400966183e-05, | |
| "loss": 1.0129, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 7.864090177133656e-05, | |
| "loss": 0.9637, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 7.857648953301128e-05, | |
| "loss": 0.9824, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 7.851207729468599e-05, | |
| "loss": 0.9758, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 7.844766505636071e-05, | |
| "loss": 0.9935, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 7.838325281803544e-05, | |
| "loss": 1.0135, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 7.831884057971014e-05, | |
| "loss": 1.0353, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 7.825442834138486e-05, | |
| "loss": 0.9855, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 7.819001610305958e-05, | |
| "loss": 0.9416, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 7.81256038647343e-05, | |
| "loss": 0.9333, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 7.806119162640902e-05, | |
| "loss": 0.9864, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 7.799677938808374e-05, | |
| "loss": 1.0263, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 7.793236714975846e-05, | |
| "loss": 0.9814, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 7.786795491143318e-05, | |
| "loss": 1.011, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 7.78035426731079e-05, | |
| "loss": 1.0106, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 7.773913043478261e-05, | |
| "loss": 0.9994, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 7.767471819645733e-05, | |
| "loss": 0.993, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 7.761030595813205e-05, | |
| "loss": 0.9515, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 7.754589371980677e-05, | |
| "loss": 1.0204, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 7.748148148148149e-05, | |
| "loss": 1.0048, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "eval_loss": 1.016575574874878, | |
| "eval_runtime": 187.3536, | |
| "eval_samples_per_second": 10.675, | |
| "eval_steps_per_second": 1.334, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 7.74170692431562e-05, | |
| "loss": 0.9734, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 7.735265700483092e-05, | |
| "loss": 0.9742, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 7.728824476650563e-05, | |
| "loss": 1.0187, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 7.722383252818036e-05, | |
| "loss": 1.0016, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 7.715942028985508e-05, | |
| "loss": 0.9926, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 7.70950080515298e-05, | |
| "loss": 0.9847, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 7.70305958132045e-05, | |
| "loss": 1.0032, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 7.696618357487924e-05, | |
| "loss": 1.0086, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 7.690177133655396e-05, | |
| "loss": 0.9906, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 7.683735909822866e-05, | |
| "loss": 0.9989, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 7.677294685990338e-05, | |
| "loss": 0.9741, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 7.67085346215781e-05, | |
| "loss": 1.0101, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 7.664412238325283e-05, | |
| "loss": 1.012, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 7.657971014492754e-05, | |
| "loss": 0.9715, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 7.651529790660225e-05, | |
| "loss": 1.028, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 7.645088566827697e-05, | |
| "loss": 0.9875, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 7.63864734299517e-05, | |
| "loss": 0.9787, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 7.632206119162641e-05, | |
| "loss": 0.9776, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 7.625764895330113e-05, | |
| "loss": 1.0093, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 7.619323671497585e-05, | |
| "loss": 0.9882, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "eval_loss": 1.0168739557266235, | |
| "eval_runtime": 187.0506, | |
| "eval_samples_per_second": 10.692, | |
| "eval_steps_per_second": 1.337, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 7.612882447665057e-05, | |
| "loss": 1.0298, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 7.606441223832529e-05, | |
| "loss": 1.0028, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 7.6e-05, | |
| "loss": 1.0006, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 7.593558776167472e-05, | |
| "loss": 1.037, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 7.587117552334944e-05, | |
| "loss": 1.0014, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 7.580676328502416e-05, | |
| "loss": 0.9677, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 7.574235104669888e-05, | |
| "loss": 1.0172, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 7.56779388083736e-05, | |
| "loss": 1.0142, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 7.561352657004832e-05, | |
| "loss": 1.0057, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 7.554911433172304e-05, | |
| "loss": 0.9965, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 7.548470209339775e-05, | |
| "loss": 1.003, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 7.542028985507247e-05, | |
| "loss": 0.9872, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 7.535587761674718e-05, | |
| "loss": 0.9755, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 7.52914653784219e-05, | |
| "loss": 0.9956, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 7.522705314009663e-05, | |
| "loss": 0.9961, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 7.516264090177135e-05, | |
| "loss": 0.9804, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 7.509822866344605e-05, | |
| "loss": 0.983, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 7.503381642512077e-05, | |
| "loss": 1.0013, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 7.49694041867955e-05, | |
| "loss": 0.9991, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 7.490499194847022e-05, | |
| "loss": 1.0178, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "eval_loss": 1.0156399011611938, | |
| "eval_runtime": 187.0787, | |
| "eval_samples_per_second": 10.691, | |
| "eval_steps_per_second": 1.336, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 7.484057971014493e-05, | |
| "loss": 1.0248, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 7.477616747181965e-05, | |
| "loss": 0.9947, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 7.471175523349436e-05, | |
| "loss": 0.9559, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 7.464734299516908e-05, | |
| "loss": 1.0175, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 7.45829307568438e-05, | |
| "loss": 0.9885, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 7.451851851851852e-05, | |
| "loss": 0.9748, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 7.445410628019324e-05, | |
| "loss": 0.9759, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 7.438969404186796e-05, | |
| "loss": 1.0137, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 7.432528180354268e-05, | |
| "loss": 0.9924, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 7.42608695652174e-05, | |
| "loss": 0.9475, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 7.419645732689211e-05, | |
| "loss": 0.9703, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 7.413204508856682e-05, | |
| "loss": 0.9994, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 7.406763285024155e-05, | |
| "loss": 0.9504, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 7.400322061191627e-05, | |
| "loss": 1.0119, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 7.393880837359099e-05, | |
| "loss": 0.9706, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 7.38743961352657e-05, | |
| "loss": 1.0015, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 7.380998389694043e-05, | |
| "loss": 0.9691, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 7.374557165861515e-05, | |
| "loss": 0.952, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 7.368115942028986e-05, | |
| "loss": 1.0084, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 7.361674718196457e-05, | |
| "loss": 1.0056, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "eval_loss": 1.0150794982910156, | |
| "eval_runtime": 187.7211, | |
| "eval_samples_per_second": 10.654, | |
| "eval_steps_per_second": 1.332, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 7.355233494363929e-05, | |
| "loss": 0.996, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 7.348792270531402e-05, | |
| "loss": 1.0038, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 7.342351046698873e-05, | |
| "loss": 0.9989, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 7.335909822866344e-05, | |
| "loss": 1.0098, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 7.329468599033816e-05, | |
| "loss": 0.975, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 7.32302737520129e-05, | |
| "loss": 0.9879, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 7.31658615136876e-05, | |
| "loss": 1.0246, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 7.310144927536232e-05, | |
| "loss": 0.9887, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 7.303703703703704e-05, | |
| "loss": 0.9639, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 7.297262479871177e-05, | |
| "loss": 0.9882, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 7.290821256038647e-05, | |
| "loss": 1.003, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 7.28438003220612e-05, | |
| "loss": 0.9804, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 7.277938808373591e-05, | |
| "loss": 0.999, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 7.271497584541063e-05, | |
| "loss": 0.9829, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 7.265056360708535e-05, | |
| "loss": 1.0062, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 7.258615136876007e-05, | |
| "loss": 1.0131, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 7.252173913043479e-05, | |
| "loss": 0.9988, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 7.24573268921095e-05, | |
| "loss": 1.018, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 7.239291465378422e-05, | |
| "loss": 1.0007, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 7.232850241545894e-05, | |
| "loss": 1.0071, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "eval_loss": 1.0151050090789795, | |
| "eval_runtime": 187.3911, | |
| "eval_samples_per_second": 10.673, | |
| "eval_steps_per_second": 1.334, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 7.226409017713366e-05, | |
| "loss": 1.0332, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 7.219967793880838e-05, | |
| "loss": 0.9705, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 7.213526570048309e-05, | |
| "loss": 1.0056, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 7.207085346215782e-05, | |
| "loss": 0.9856, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 7.200644122383254e-05, | |
| "loss": 0.9867, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 7.194202898550724e-05, | |
| "loss": 1.0127, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 7.187761674718196e-05, | |
| "loss": 0.9923, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 7.181320450885669e-05, | |
| "loss": 0.9255, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 7.174879227053141e-05, | |
| "loss": 1.0205, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 7.168438003220612e-05, | |
| "loss": 0.9925, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 7.161996779388084e-05, | |
| "loss": 0.9856, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 7.155555555555555e-05, | |
| "loss": 1.0136, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 7.149114331723029e-05, | |
| "loss": 1.0018, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 7.142673107890499e-05, | |
| "loss": 0.9814, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 7.136231884057971e-05, | |
| "loss": 0.9785, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 7.129790660225443e-05, | |
| "loss": 1.0153, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 7.123349436392915e-05, | |
| "loss": 1.0266, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 7.116908212560387e-05, | |
| "loss": 1.0527, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 7.110466988727859e-05, | |
| "loss": 0.9867, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 7.10402576489533e-05, | |
| "loss": 1.0175, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "eval_loss": 1.0150611400604248, | |
| "eval_runtime": 187.508, | |
| "eval_samples_per_second": 10.666, | |
| "eval_steps_per_second": 1.333, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 7.097584541062802e-05, | |
| "loss": 1.0153, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 7.091143317230274e-05, | |
| "loss": 0.9565, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 7.084702093397746e-05, | |
| "loss": 0.9942, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 7.078260869565218e-05, | |
| "loss": 1.0096, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 7.07181964573269e-05, | |
| "loss": 0.9973, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 7.065378421900162e-05, | |
| "loss": 0.9457, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 7.058937198067633e-05, | |
| "loss": 0.9938, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 7.052495974235105e-05, | |
| "loss": 0.9798, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 7.046054750402576e-05, | |
| "loss": 0.9603, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 7.039613526570049e-05, | |
| "loss": 0.975, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 7.033172302737521e-05, | |
| "loss": 0.998, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 7.026731078904993e-05, | |
| "loss": 1.0167, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 7.020289855072463e-05, | |
| "loss": 0.9682, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 7.013848631239935e-05, | |
| "loss": 0.9827, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 7.007407407407408e-05, | |
| "loss": 0.9866, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 7.00096618357488e-05, | |
| "loss": 1.0018, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 6.994524959742351e-05, | |
| "loss": 1.0173, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 6.988083735909823e-05, | |
| "loss": 1.0103, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 6.981642512077296e-05, | |
| "loss": 0.9881, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 6.975201288244766e-05, | |
| "loss": 1.0108, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "eval_loss": 1.014174461364746, | |
| "eval_runtime": 187.5385, | |
| "eval_samples_per_second": 10.664, | |
| "eval_steps_per_second": 1.333, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 6.968760064412238e-05, | |
| "loss": 0.9903, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 6.96231884057971e-05, | |
| "loss": 1.0169, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 6.955877616747182e-05, | |
| "loss": 0.9837, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 6.949436392914654e-05, | |
| "loss": 0.9573, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 6.942995169082126e-05, | |
| "loss": 0.9814, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 6.936553945249598e-05, | |
| "loss": 0.9829, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 6.93011272141707e-05, | |
| "loss": 0.9956, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 6.923671497584541e-05, | |
| "loss": 1.0234, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 6.917230273752013e-05, | |
| "loss": 0.986, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 6.910789049919485e-05, | |
| "loss": 0.9829, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 6.904347826086957e-05, | |
| "loss": 1.0251, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 6.897906602254429e-05, | |
| "loss": 0.9879, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 6.891465378421901e-05, | |
| "loss": 1.039, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 6.885024154589373e-05, | |
| "loss": 0.9752, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 6.878582930756844e-05, | |
| "loss": 1.0013, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 6.872141706924315e-05, | |
| "loss": 0.9558, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 6.865700483091788e-05, | |
| "loss": 1.0102, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 6.85925925925926e-05, | |
| "loss": 0.9936, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 6.85281803542673e-05, | |
| "loss": 0.9864, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 6.846376811594202e-05, | |
| "loss": 1.0204, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "eval_loss": 1.013562798500061, | |
| "eval_runtime": 187.3848, | |
| "eval_samples_per_second": 10.673, | |
| "eval_steps_per_second": 1.334, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 6.839935587761676e-05, | |
| "loss": 0.9885, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 6.833494363929148e-05, | |
| "loss": 0.9857, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 6.827053140096618e-05, | |
| "loss": 0.9764, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 6.82061191626409e-05, | |
| "loss": 0.9809, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 6.814170692431562e-05, | |
| "loss": 0.9855, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 6.807729468599035e-05, | |
| "loss": 1.0134, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 6.801288244766506e-05, | |
| "loss": 0.9927, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 6.794847020933977e-05, | |
| "loss": 1.0337, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 6.788405797101449e-05, | |
| "loss": 0.9835, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 6.781964573268923e-05, | |
| "loss": 1.0772, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 6.775523349436393e-05, | |
| "loss": 0.9874, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 6.769082125603865e-05, | |
| "loss": 0.9956, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 6.762640901771337e-05, | |
| "loss": 1.014, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 6.756199677938809e-05, | |
| "loss": 0.9957, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 6.74975845410628e-05, | |
| "loss": 0.9861, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 6.743317230273752e-05, | |
| "loss": 0.9824, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 6.736876006441224e-05, | |
| "loss": 1.0184, | |
| "step": 5170 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 6.730434782608696e-05, | |
| "loss": 1.0121, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 6.723993558776168e-05, | |
| "loss": 0.9957, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 6.71755233494364e-05, | |
| "loss": 1.0209, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "eval_loss": 1.0133100748062134, | |
| "eval_runtime": 187.3315, | |
| "eval_samples_per_second": 10.676, | |
| "eval_steps_per_second": 1.335, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 6.711111111111112e-05, | |
| "loss": 0.9808, | |
| "step": 5210 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 6.704669887278582e-05, | |
| "loss": 1.0036, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 6.698228663446055e-05, | |
| "loss": 0.9727, | |
| "step": 5230 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 6.691787439613527e-05, | |
| "loss": 1.0178, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 6.685346215780999e-05, | |
| "loss": 0.976, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 6.67890499194847e-05, | |
| "loss": 0.9991, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 6.672463768115942e-05, | |
| "loss": 0.9956, | |
| "step": 5270 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 6.666022544283415e-05, | |
| "loss": 0.9791, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 6.659581320450887e-05, | |
| "loss": 1.0081, | |
| "step": 5290 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 6.653140096618357e-05, | |
| "loss": 1.0244, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 6.646698872785829e-05, | |
| "loss": 1.0103, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 6.640257648953302e-05, | |
| "loss": 1.0579, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 6.633816425120773e-05, | |
| "loss": 0.9774, | |
| "step": 5330 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 6.627375201288245e-05, | |
| "loss": 0.9376, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 6.620933977455717e-05, | |
| "loss": 1.006, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 6.614492753623188e-05, | |
| "loss": 0.9889, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 6.60805152979066e-05, | |
| "loss": 1.0201, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 6.601610305958132e-05, | |
| "loss": 1.0034, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 6.595169082125604e-05, | |
| "loss": 1.0556, | |
| "step": 5390 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 6.588727858293076e-05, | |
| "loss": 1.0124, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "eval_loss": 1.012602686882019, | |
| "eval_runtime": 187.4458, | |
| "eval_samples_per_second": 10.67, | |
| "eval_steps_per_second": 1.334, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 6.582286634460548e-05, | |
| "loss": 0.9938, | |
| "step": 5410 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 6.57584541062802e-05, | |
| "loss": 0.9905, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 6.569404186795492e-05, | |
| "loss": 0.9896, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 6.562962962962963e-05, | |
| "loss": 1.0039, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 6.556521739130434e-05, | |
| "loss": 1.0021, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 6.550080515297907e-05, | |
| "loss": 0.99, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 6.543639291465379e-05, | |
| "loss": 0.9826, | |
| "step": 5470 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 6.537198067632851e-05, | |
| "loss": 0.9841, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 6.530756843800321e-05, | |
| "loss": 0.9749, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 6.524315619967795e-05, | |
| "loss": 0.9905, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 6.517874396135267e-05, | |
| "loss": 1.0376, | |
| "step": 5510 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 6.511433172302738e-05, | |
| "loss": 1.0513, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 6.504991948470209e-05, | |
| "loss": 1.0093, | |
| "step": 5530 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 6.498550724637682e-05, | |
| "loss": 1.0263, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 6.492109500805154e-05, | |
| "loss": 0.9843, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 6.485668276972625e-05, | |
| "loss": 1.0204, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 6.479227053140096e-05, | |
| "loss": 0.9995, | |
| "step": 5570 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 6.472785829307568e-05, | |
| "loss": 1.0205, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 6.466344605475041e-05, | |
| "loss": 0.9466, | |
| "step": 5590 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 6.459903381642512e-05, | |
| "loss": 1.0259, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "eval_loss": 1.0134350061416626, | |
| "eval_runtime": 187.4026, | |
| "eval_samples_per_second": 10.672, | |
| "eval_steps_per_second": 1.334, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 6.453462157809984e-05, | |
| "loss": 0.9799, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 6.447020933977456e-05, | |
| "loss": 1.014, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 6.440579710144929e-05, | |
| "loss": 1.0418, | |
| "step": 5630 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 6.4341384863124e-05, | |
| "loss": 1.0168, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 6.427697262479871e-05, | |
| "loss": 0.948, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 6.421256038647343e-05, | |
| "loss": 1.0076, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 6.414814814814815e-05, | |
| "loss": 1.0181, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 6.408373590982287e-05, | |
| "loss": 1.0225, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 6.401932367149759e-05, | |
| "loss": 1.0063, | |
| "step": 5690 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 6.395491143317231e-05, | |
| "loss": 1.0238, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 6.389049919484703e-05, | |
| "loss": 0.9782, | |
| "step": 5710 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 6.382608695652174e-05, | |
| "loss": 1.0378, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 6.376167471819646e-05, | |
| "loss": 0.954, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 6.369726247987118e-05, | |
| "loss": 1.004, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 6.36328502415459e-05, | |
| "loss": 0.9934, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 6.35684380032206e-05, | |
| "loss": 0.9874, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 6.350402576489534e-05, | |
| "loss": 0.9677, | |
| "step": 5770 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 6.343961352657006e-05, | |
| "loss": 0.9833, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 6.337520128824476e-05, | |
| "loss": 0.9972, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 6.331078904991948e-05, | |
| "loss": 1.0289, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "eval_loss": 1.0126157999038696, | |
| "eval_runtime": 187.3366, | |
| "eval_samples_per_second": 10.676, | |
| "eval_steps_per_second": 1.334, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 6.324637681159421e-05, | |
| "loss": 0.9907, | |
| "step": 5810 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 6.318196457326893e-05, | |
| "loss": 0.9942, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 6.311755233494364e-05, | |
| "loss": 0.9764, | |
| "step": 5830 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 6.305314009661836e-05, | |
| "loss": 0.9942, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 6.298872785829309e-05, | |
| "loss": 1.0397, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 6.29243156199678e-05, | |
| "loss": 1.0182, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 6.285990338164251e-05, | |
| "loss": 0.9443, | |
| "step": 5870 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 6.279549114331723e-05, | |
| "loss": 0.973, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 6.273107890499195e-05, | |
| "loss": 0.951, | |
| "step": 5890 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 6.266666666666667e-05, | |
| "loss": 0.9824, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 6.260225442834139e-05, | |
| "loss": 1.0014, | |
| "step": 5910 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 6.25378421900161e-05, | |
| "loss": 1.0029, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 6.247342995169082e-05, | |
| "loss": 1.0166, | |
| "step": 5930 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 6.240901771336554e-05, | |
| "loss": 1.0011, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 6.234460547504026e-05, | |
| "loss": 1.0156, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 6.228019323671498e-05, | |
| "loss": 1.0166, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 6.22157809983897e-05, | |
| "loss": 0.9613, | |
| "step": 5970 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 6.21513687600644e-05, | |
| "loss": 0.9551, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 6.208695652173914e-05, | |
| "loss": 0.9628, | |
| "step": 5990 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 6.202254428341385e-05, | |
| "loss": 0.9675, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "eval_loss": 1.012001633644104, | |
| "eval_runtime": 187.1833, | |
| "eval_samples_per_second": 10.685, | |
| "eval_steps_per_second": 1.336, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 6.195813204508857e-05, | |
| "loss": 0.9864, | |
| "step": 6010 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 6.189371980676328e-05, | |
| "loss": 0.9786, | |
| "step": 6020 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 6.182930756843801e-05, | |
| "loss": 0.9732, | |
| "step": 6030 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 6.176489533011273e-05, | |
| "loss": 0.9915, | |
| "step": 6040 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 6.170048309178745e-05, | |
| "loss": 0.967, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 6.163607085346215e-05, | |
| "loss": 1.0035, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 6.157165861513687e-05, | |
| "loss": 1.0008, | |
| "step": 6070 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 6.15072463768116e-05, | |
| "loss": 0.9626, | |
| "step": 6080 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 6.144283413848631e-05, | |
| "loss": 1.0128, | |
| "step": 6090 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 6.137842190016103e-05, | |
| "loss": 0.9504, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 6.131400966183575e-05, | |
| "loss": 1.004, | |
| "step": 6110 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 6.124959742351048e-05, | |
| "loss": 1.0037, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 6.118518518518518e-05, | |
| "loss": 1.0124, | |
| "step": 6130 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 6.11207729468599e-05, | |
| "loss": 1.0248, | |
| "step": 6140 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 6.105636070853462e-05, | |
| "loss": 0.9643, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 6.099194847020935e-05, | |
| "loss": 0.9443, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 6.0927536231884066e-05, | |
| "loss": 1.0231, | |
| "step": 6170 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 6.086312399355878e-05, | |
| "loss": 1.0228, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 6.0798711755233496e-05, | |
| "loss": 1.0145, | |
| "step": 6190 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 6.073429951690821e-05, | |
| "loss": 1.0081, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "eval_loss": 1.0115959644317627, | |
| "eval_runtime": 187.1705, | |
| "eval_samples_per_second": 10.685, | |
| "eval_steps_per_second": 1.336, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 6.066988727858294e-05, | |
| "loss": 0.9809, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 6.060547504025765e-05, | |
| "loss": 0.9942, | |
| "step": 6220 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 6.054106280193237e-05, | |
| "loss": 1.0487, | |
| "step": 6230 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 6.047665056360708e-05, | |
| "loss": 0.9674, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 6.041223832528181e-05, | |
| "loss": 0.9941, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 6.034782608695653e-05, | |
| "loss": 0.98, | |
| "step": 6260 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 6.0283413848631246e-05, | |
| "loss": 0.9749, | |
| "step": 6270 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 6.021900161030596e-05, | |
| "loss": 0.9085, | |
| "step": 6280 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 6.015458937198068e-05, | |
| "loss": 0.9835, | |
| "step": 6290 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 6.00901771336554e-05, | |
| "loss": 0.9691, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 6.0025764895330114e-05, | |
| "loss": 0.9643, | |
| "step": 6310 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 5.996135265700483e-05, | |
| "loss": 0.9596, | |
| "step": 6320 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 5.9896940418679545e-05, | |
| "loss": 0.9899, | |
| "step": 6330 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 5.983252818035428e-05, | |
| "loss": 0.974, | |
| "step": 6340 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 5.976811594202899e-05, | |
| "loss": 0.9449, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 5.970370370370371e-05, | |
| "loss": 0.9808, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 5.963929146537842e-05, | |
| "loss": 0.9546, | |
| "step": 6370 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 5.957487922705314e-05, | |
| "loss": 0.9845, | |
| "step": 6380 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 5.9510466988727864e-05, | |
| "loss": 0.9513, | |
| "step": 6390 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 5.944605475040258e-05, | |
| "loss": 0.9811, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "eval_loss": 1.0142450332641602, | |
| "eval_runtime": 187.1141, | |
| "eval_samples_per_second": 10.689, | |
| "eval_steps_per_second": 1.336, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 5.9381642512077294e-05, | |
| "loss": 0.9986, | |
| "step": 6410 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 5.931723027375201e-05, | |
| "loss": 0.9792, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 5.925281803542674e-05, | |
| "loss": 0.962, | |
| "step": 6430 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 5.918840579710146e-05, | |
| "loss": 0.971, | |
| "step": 6440 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 5.912399355877617e-05, | |
| "loss": 0.9833, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 5.905958132045089e-05, | |
| "loss": 0.9825, | |
| "step": 6460 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 5.899516908212561e-05, | |
| "loss": 0.9646, | |
| "step": 6470 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 5.8930756843800325e-05, | |
| "loss": 0.9543, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 5.8866344605475044e-05, | |
| "loss": 1.0055, | |
| "step": 6490 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 5.8801932367149756e-05, | |
| "loss": 0.9524, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 5.8737520128824475e-05, | |
| "loss": 0.9463, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 5.86731078904992e-05, | |
| "loss": 0.9943, | |
| "step": 6520 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 5.860869565217392e-05, | |
| "loss": 0.9899, | |
| "step": 6530 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 5.854428341384863e-05, | |
| "loss": 1.0231, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 5.847987117552335e-05, | |
| "loss": 0.9453, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 5.8415458937198075e-05, | |
| "loss": 0.9887, | |
| "step": 6560 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 5.8351046698872794e-05, | |
| "loss": 0.9487, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 5.8286634460547505e-05, | |
| "loss": 0.9814, | |
| "step": 6580 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 5.8222222222222224e-05, | |
| "loss": 0.9312, | |
| "step": 6590 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 5.8157809983896936e-05, | |
| "loss": 0.9525, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "eval_loss": 1.0148019790649414, | |
| "eval_runtime": 187.2863, | |
| "eval_samples_per_second": 10.679, | |
| "eval_steps_per_second": 1.335, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 5.809339774557166e-05, | |
| "loss": 0.9827, | |
| "step": 6610 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 5.802898550724638e-05, | |
| "loss": 1.0076, | |
| "step": 6620 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 5.79645732689211e-05, | |
| "loss": 0.9929, | |
| "step": 6630 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 5.790016103059581e-05, | |
| "loss": 0.9358, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 5.7835748792270536e-05, | |
| "loss": 0.9382, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 5.7771336553945255e-05, | |
| "loss": 0.9894, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 5.770692431561997e-05, | |
| "loss": 0.9757, | |
| "step": 6670 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 5.7642512077294686e-05, | |
| "loss": 0.9541, | |
| "step": 6680 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 5.7578099838969404e-05, | |
| "loss": 0.9451, | |
| "step": 6690 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 5.751368760064413e-05, | |
| "loss": 1.0106, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 5.744927536231884e-05, | |
| "loss": 0.9709, | |
| "step": 6710 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 5.738486312399356e-05, | |
| "loss": 0.9139, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 5.732045088566827e-05, | |
| "loss": 0.9994, | |
| "step": 6730 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 5.7256038647343005e-05, | |
| "loss": 0.9361, | |
| "step": 6740 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 5.719162640901772e-05, | |
| "loss": 0.9798, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 5.7127214170692435e-05, | |
| "loss": 0.9802, | |
| "step": 6760 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 5.706280193236715e-05, | |
| "loss": 0.9239, | |
| "step": 6770 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 5.699838969404187e-05, | |
| "loss": 0.9608, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 5.693397745571659e-05, | |
| "loss": 0.9406, | |
| "step": 6790 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 5.686956521739131e-05, | |
| "loss": 0.9859, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "eval_loss": 1.0146571397781372, | |
| "eval_runtime": 187.2691, | |
| "eval_samples_per_second": 10.68, | |
| "eval_steps_per_second": 1.335, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 5.680515297906602e-05, | |
| "loss": 0.9655, | |
| "step": 6810 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 5.674074074074074e-05, | |
| "loss": 1.0016, | |
| "step": 6820 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 5.6676328502415466e-05, | |
| "loss": 0.9549, | |
| "step": 6830 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 5.661191626409018e-05, | |
| "loss": 0.9516, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 5.65475040257649e-05, | |
| "loss": 0.9834, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 5.6483091787439616e-05, | |
| "loss": 0.9359, | |
| "step": 6860 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 5.641867954911434e-05, | |
| "loss": 0.9511, | |
| "step": 6870 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 5.635426731078905e-05, | |
| "loss": 0.9848, | |
| "step": 6880 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 5.628985507246377e-05, | |
| "loss": 0.9289, | |
| "step": 6890 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 5.6225442834138484e-05, | |
| "loss": 0.9918, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 5.61610305958132e-05, | |
| "loss": 1.0145, | |
| "step": 6910 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 5.609661835748793e-05, | |
| "loss": 0.9003, | |
| "step": 6920 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 5.6032206119162647e-05, | |
| "loss": 0.9625, | |
| "step": 6930 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 5.596779388083736e-05, | |
| "loss": 0.98, | |
| "step": 6940 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 5.590338164251208e-05, | |
| "loss": 0.9497, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 5.58389694041868e-05, | |
| "loss": 0.9297, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 5.577455716586152e-05, | |
| "loss": 0.9802, | |
| "step": 6970 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 5.571014492753623e-05, | |
| "loss": 0.9524, | |
| "step": 6980 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 5.564573268921095e-05, | |
| "loss": 0.9421, | |
| "step": 6990 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 5.5581320450885664e-05, | |
| "loss": 1.0126, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "eval_loss": 1.0157631635665894, | |
| "eval_runtime": 187.2785, | |
| "eval_samples_per_second": 10.679, | |
| "eval_steps_per_second": 1.335, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 5.551690821256039e-05, | |
| "loss": 0.9211, | |
| "step": 7010 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 5.545249597423511e-05, | |
| "loss": 0.9814, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 5.538808373590983e-05, | |
| "loss": 0.9696, | |
| "step": 7030 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 5.532367149758454e-05, | |
| "loss": 0.9691, | |
| "step": 7040 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 5.5259259259259264e-05, | |
| "loss": 0.9357, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 5.519484702093398e-05, | |
| "loss": 0.9451, | |
| "step": 7060 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 5.5130434782608695e-05, | |
| "loss": 0.968, | |
| "step": 7070 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 5.5066022544283414e-05, | |
| "loss": 0.946, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 5.500161030595814e-05, | |
| "loss": 0.961, | |
| "step": 7090 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 5.493719806763286e-05, | |
| "loss": 0.9741, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 5.487278582930757e-05, | |
| "loss": 0.9341, | |
| "step": 7110 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 5.480837359098229e-05, | |
| "loss": 0.9488, | |
| "step": 7120 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 5.4743961352657e-05, | |
| "loss": 0.9791, | |
| "step": 7130 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 5.467954911433173e-05, | |
| "loss": 0.9957, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 5.4615136876006445e-05, | |
| "loss": 0.978, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 5.455072463768116e-05, | |
| "loss": 0.9768, | |
| "step": 7160 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 5.4486312399355875e-05, | |
| "loss": 0.9585, | |
| "step": 7170 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 5.44219001610306e-05, | |
| "loss": 0.9307, | |
| "step": 7180 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 5.435748792270532e-05, | |
| "loss": 1.0232, | |
| "step": 7190 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 5.429307568438004e-05, | |
| "loss": 0.9892, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "eval_loss": 1.0147501230239868, | |
| "eval_runtime": 187.244, | |
| "eval_samples_per_second": 10.681, | |
| "eval_steps_per_second": 1.335, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 5.422866344605475e-05, | |
| "loss": 0.9983, | |
| "step": 7210 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 5.416425120772947e-05, | |
| "loss": 0.9572, | |
| "step": 7220 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 5.4099838969404194e-05, | |
| "loss": 0.9815, | |
| "step": 7230 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 5.4035426731078906e-05, | |
| "loss": 0.9854, | |
| "step": 7240 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 5.3971014492753625e-05, | |
| "loss": 0.9947, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 5.390660225442834e-05, | |
| "loss": 0.9893, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 5.384219001610307e-05, | |
| "loss": 0.9645, | |
| "step": 7270 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 5.377777777777778e-05, | |
| "loss": 0.9776, | |
| "step": 7280 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 5.37133655394525e-05, | |
| "loss": 0.9974, | |
| "step": 7290 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 5.364895330112721e-05, | |
| "loss": 0.9439, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 5.358454106280193e-05, | |
| "loss": 0.996, | |
| "step": 7310 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 5.3520128824476656e-05, | |
| "loss": 0.8993, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 5.3455716586151374e-05, | |
| "loss": 0.984, | |
| "step": 7330 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 5.3391304347826086e-05, | |
| "loss": 0.9444, | |
| "step": 7340 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 5.3326892109500805e-05, | |
| "loss": 0.9773, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 5.326247987117553e-05, | |
| "loss": 0.9527, | |
| "step": 7360 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 5.319806763285025e-05, | |
| "loss": 0.9611, | |
| "step": 7370 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 5.313365539452496e-05, | |
| "loss": 0.9423, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 5.306924315619968e-05, | |
| "loss": 0.9939, | |
| "step": 7390 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 5.3004830917874405e-05, | |
| "loss": 0.9833, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "eval_loss": 1.0141487121582031, | |
| "eval_runtime": 187.1541, | |
| "eval_samples_per_second": 10.686, | |
| "eval_steps_per_second": 1.336, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 5.294041867954912e-05, | |
| "loss": 0.9883, | |
| "step": 7410 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 5.2876006441223836e-05, | |
| "loss": 0.9374, | |
| "step": 7420 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 5.281159420289855e-05, | |
| "loss": 0.9671, | |
| "step": 7430 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 5.2747181964573267e-05, | |
| "loss": 0.9635, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 5.268276972624799e-05, | |
| "loss": 0.9736, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 5.261835748792271e-05, | |
| "loss": 0.9731, | |
| "step": 7460 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 5.255394524959742e-05, | |
| "loss": 0.9757, | |
| "step": 7470 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 5.248953301127214e-05, | |
| "loss": 0.9522, | |
| "step": 7480 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 5.242512077294687e-05, | |
| "loss": 0.965, | |
| "step": 7490 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 5.2360708534621586e-05, | |
| "loss": 0.928, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 5.22962962962963e-05, | |
| "loss": 0.9651, | |
| "step": 7510 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 5.2231884057971016e-05, | |
| "loss": 0.9773, | |
| "step": 7520 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 5.216747181964573e-05, | |
| "loss": 0.9923, | |
| "step": 7530 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 5.2103059581320454e-05, | |
| "loss": 0.9781, | |
| "step": 7540 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 5.203864734299517e-05, | |
| "loss": 0.9443, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 5.197423510466989e-05, | |
| "loss": 0.9917, | |
| "step": 7560 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 5.19098228663446e-05, | |
| "loss": 0.9616, | |
| "step": 7570 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 5.184541062801933e-05, | |
| "loss": 0.9467, | |
| "step": 7580 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 5.178099838969405e-05, | |
| "loss": 0.9879, | |
| "step": 7590 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 5.171658615136876e-05, | |
| "loss": 1.0006, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "eval_loss": 1.0145756006240845, | |
| "eval_runtime": 186.84, | |
| "eval_samples_per_second": 10.704, | |
| "eval_steps_per_second": 1.338, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 5.165217391304348e-05, | |
| "loss": 0.9786, | |
| "step": 7610 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 5.1587761674718197e-05, | |
| "loss": 1.0002, | |
| "step": 7620 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 5.152334943639292e-05, | |
| "loss": 0.9648, | |
| "step": 7630 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 5.1458937198067634e-05, | |
| "loss": 0.9719, | |
| "step": 7640 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 5.139452495974235e-05, | |
| "loss": 0.9647, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 5.1330112721417065e-05, | |
| "loss": 0.9868, | |
| "step": 7660 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 5.12657004830918e-05, | |
| "loss": 0.9521, | |
| "step": 7670 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 5.120128824476651e-05, | |
| "loss": 0.9742, | |
| "step": 7680 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 5.113687600644123e-05, | |
| "loss": 0.9976, | |
| "step": 7690 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 5.107246376811594e-05, | |
| "loss": 0.9621, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 5.1008051529790665e-05, | |
| "loss": 0.9312, | |
| "step": 7710 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 5.0943639291465384e-05, | |
| "loss": 0.9685, | |
| "step": 7720 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 5.08792270531401e-05, | |
| "loss": 1.0045, | |
| "step": 7730 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 5.0814814814814814e-05, | |
| "loss": 0.9582, | |
| "step": 7740 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 5.075040257648953e-05, | |
| "loss": 0.9393, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 5.068599033816426e-05, | |
| "loss": 0.94, | |
| "step": 7760 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 5.062157809983897e-05, | |
| "loss": 0.9547, | |
| "step": 7770 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 5.055716586151369e-05, | |
| "loss": 0.978, | |
| "step": 7780 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 5.049275362318841e-05, | |
| "loss": 0.993, | |
| "step": 7790 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 5.042834138486313e-05, | |
| "loss": 0.998, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "eval_loss": 1.015630841255188, | |
| "eval_runtime": 187.1282, | |
| "eval_samples_per_second": 10.688, | |
| "eval_steps_per_second": 1.336, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 5.0363929146537845e-05, | |
| "loss": 0.9846, | |
| "step": 7810 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 5.0299516908212564e-05, | |
| "loss": 0.9118, | |
| "step": 7820 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 5.0235104669887276e-05, | |
| "loss": 1.0075, | |
| "step": 7830 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 5.0170692431561994e-05, | |
| "loss": 0.9677, | |
| "step": 7840 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 5.010628019323672e-05, | |
| "loss": 0.9867, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 5.004186795491144e-05, | |
| "loss": 0.9946, | |
| "step": 7860 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 4.997745571658615e-05, | |
| "loss": 0.9275, | |
| "step": 7870 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 4.9913043478260876e-05, | |
| "loss": 0.9904, | |
| "step": 7880 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 4.985507246376812e-05, | |
| "loss": 0.9669, | |
| "step": 7890 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 4.9790660225442835e-05, | |
| "loss": 0.9631, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 4.9726247987117554e-05, | |
| "loss": 0.9708, | |
| "step": 7910 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 4.966183574879227e-05, | |
| "loss": 0.9634, | |
| "step": 7920 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 4.959742351046699e-05, | |
| "loss": 0.9495, | |
| "step": 7930 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 4.953301127214171e-05, | |
| "loss": 0.9572, | |
| "step": 7940 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 4.946859903381643e-05, | |
| "loss": 0.9966, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 4.940418679549115e-05, | |
| "loss": 0.9309, | |
| "step": 7960 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 4.9339774557165866e-05, | |
| "loss": 0.9672, | |
| "step": 7970 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 4.9275362318840584e-05, | |
| "loss": 0.9842, | |
| "step": 7980 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 4.92109500805153e-05, | |
| "loss": 0.9883, | |
| "step": 7990 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 4.9146537842190015e-05, | |
| "loss": 0.9761, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "eval_loss": 1.0148636102676392, | |
| "eval_runtime": 187.0058, | |
| "eval_samples_per_second": 10.695, | |
| "eval_steps_per_second": 1.337, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 4.9082125603864734e-05, | |
| "loss": 0.9617, | |
| "step": 8010 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 4.901771336553945e-05, | |
| "loss": 0.9516, | |
| "step": 8020 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 4.895330112721417e-05, | |
| "loss": 0.9722, | |
| "step": 8030 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 4.888888888888889e-05, | |
| "loss": 0.9616, | |
| "step": 8040 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 4.882447665056361e-05, | |
| "loss": 0.9258, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 4.876006441223833e-05, | |
| "loss": 0.981, | |
| "step": 8060 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 4.8695652173913046e-05, | |
| "loss": 0.9976, | |
| "step": 8070 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 4.8631239935587765e-05, | |
| "loss": 0.9362, | |
| "step": 8080 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 4.8566827697262483e-05, | |
| "loss": 0.9574, | |
| "step": 8090 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 4.85024154589372e-05, | |
| "loss": 0.9829, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 4.8438003220611914e-05, | |
| "loss": 0.9812, | |
| "step": 8110 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 4.837359098228664e-05, | |
| "loss": 0.9608, | |
| "step": 8120 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 4.830917874396135e-05, | |
| "loss": 0.9526, | |
| "step": 8130 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 4.824476650563608e-05, | |
| "loss": 0.9605, | |
| "step": 8140 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 4.818035426731079e-05, | |
| "loss": 0.974, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 4.8115942028985514e-05, | |
| "loss": 0.9996, | |
| "step": 8160 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 4.8051529790660226e-05, | |
| "loss": 0.9518, | |
| "step": 8170 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 4.7987117552334945e-05, | |
| "loss": 1.0006, | |
| "step": 8180 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 4.7922705314009664e-05, | |
| "loss": 0.9218, | |
| "step": 8190 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 4.785829307568438e-05, | |
| "loss": 0.9336, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "eval_loss": 1.0153106451034546, | |
| "eval_runtime": 187.0257, | |
| "eval_samples_per_second": 10.694, | |
| "eval_steps_per_second": 1.337, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 4.77938808373591e-05, | |
| "loss": 0.981, | |
| "step": 8210 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 4.772946859903382e-05, | |
| "loss": 0.9565, | |
| "step": 8220 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 4.766505636070854e-05, | |
| "loss": 1.0248, | |
| "step": 8230 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 4.760064412238325e-05, | |
| "loss": 0.9458, | |
| "step": 8240 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 4.7536231884057976e-05, | |
| "loss": 0.9753, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 4.747181964573269e-05, | |
| "loss": 0.9396, | |
| "step": 8260 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 4.740740740740741e-05, | |
| "loss": 0.9538, | |
| "step": 8270 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 4.7342995169082125e-05, | |
| "loss": 0.9652, | |
| "step": 8280 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 4.727858293075685e-05, | |
| "loss": 0.9746, | |
| "step": 8290 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 4.721417069243156e-05, | |
| "loss": 0.9441, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 4.714975845410628e-05, | |
| "loss": 0.9899, | |
| "step": 8310 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 4.7085346215781e-05, | |
| "loss": 0.9278, | |
| "step": 8320 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 4.702093397745572e-05, | |
| "loss": 0.9441, | |
| "step": 8330 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 4.695652173913044e-05, | |
| "loss": 0.9635, | |
| "step": 8340 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 4.6892109500805156e-05, | |
| "loss": 0.9947, | |
| "step": 8350 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 4.6827697262479875e-05, | |
| "loss": 0.9693, | |
| "step": 8360 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 4.6763285024154594e-05, | |
| "loss": 0.9755, | |
| "step": 8370 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 4.669887278582931e-05, | |
| "loss": 0.9791, | |
| "step": 8380 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 4.6634460547504024e-05, | |
| "loss": 0.9476, | |
| "step": 8390 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 4.657004830917875e-05, | |
| "loss": 1.0156, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "eval_loss": 1.0150635242462158, | |
| "eval_runtime": 187.2093, | |
| "eval_samples_per_second": 10.683, | |
| "eval_steps_per_second": 1.335, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 4.650563607085346e-05, | |
| "loss": 0.952, | |
| "step": 8410 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 4.644122383252818e-05, | |
| "loss": 0.95, | |
| "step": 8420 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 4.63768115942029e-05, | |
| "loss": 0.9887, | |
| "step": 8430 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 4.631239935587762e-05, | |
| "loss": 0.9511, | |
| "step": 8440 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 4.6247987117552336e-05, | |
| "loss": 0.9746, | |
| "step": 8450 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.6183574879227055e-05, | |
| "loss": 0.9705, | |
| "step": 8460 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.6119162640901774e-05, | |
| "loss": 1.0017, | |
| "step": 8470 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.605475040257649e-05, | |
| "loss": 0.9205, | |
| "step": 8480 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.599033816425121e-05, | |
| "loss": 0.9397, | |
| "step": 8490 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.592592592592593e-05, | |
| "loss": 0.9969, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.586151368760065e-05, | |
| "loss": 0.9735, | |
| "step": 8510 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.579710144927537e-05, | |
| "loss": 0.9382, | |
| "step": 8520 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.573268921095008e-05, | |
| "loss": 0.9465, | |
| "step": 8530 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.5668276972624805e-05, | |
| "loss": 0.9855, | |
| "step": 8540 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 4.560386473429952e-05, | |
| "loss": 0.9988, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 4.5539452495974235e-05, | |
| "loss": 0.9556, | |
| "step": 8560 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 4.5475040257648954e-05, | |
| "loss": 0.9487, | |
| "step": 8570 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.541062801932367e-05, | |
| "loss": 0.9595, | |
| "step": 8580 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.534621578099839e-05, | |
| "loss": 0.9676, | |
| "step": 8590 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.528180354267311e-05, | |
| "loss": 0.9626, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "eval_loss": 1.0142388343811035, | |
| "eval_runtime": 187.0049, | |
| "eval_samples_per_second": 10.695, | |
| "eval_steps_per_second": 1.337, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.521739130434783e-05, | |
| "loss": 0.9475, | |
| "step": 8610 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.515297906602254e-05, | |
| "loss": 0.9718, | |
| "step": 8620 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.5088566827697266e-05, | |
| "loss": 1.0018, | |
| "step": 8630 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.502415458937198e-05, | |
| "loss": 0.9209, | |
| "step": 8640 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.4959742351046704e-05, | |
| "loss": 1.0049, | |
| "step": 8650 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.4895330112721416e-05, | |
| "loss": 0.9694, | |
| "step": 8660 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.483091787439614e-05, | |
| "loss": 0.9897, | |
| "step": 8670 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.476650563607085e-05, | |
| "loss": 0.9217, | |
| "step": 8680 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.470209339774558e-05, | |
| "loss": 0.938, | |
| "step": 8690 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.463768115942029e-05, | |
| "loss": 0.96, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.4573268921095016e-05, | |
| "loss": 0.9891, | |
| "step": 8710 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.450885668276973e-05, | |
| "loss": 0.99, | |
| "step": 8720 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.4444444444444447e-05, | |
| "loss": 0.9262, | |
| "step": 8730 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.4380032206119165e-05, | |
| "loss": 0.9725, | |
| "step": 8740 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.4315619967793884e-05, | |
| "loss": 0.9849, | |
| "step": 8750 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.42512077294686e-05, | |
| "loss": 0.9607, | |
| "step": 8760 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 4.4186795491143315e-05, | |
| "loss": 0.972, | |
| "step": 8770 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 4.412238325281804e-05, | |
| "loss": 0.9642, | |
| "step": 8780 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 4.405797101449275e-05, | |
| "loss": 0.9878, | |
| "step": 8790 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 4.399355877616748e-05, | |
| "loss": 0.9221, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "eval_loss": 1.013732671737671, | |
| "eval_runtime": 187.1182, | |
| "eval_samples_per_second": 10.688, | |
| "eval_steps_per_second": 1.336, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 4.392914653784219e-05, | |
| "loss": 0.9452, | |
| "step": 8810 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 4.3864734299516915e-05, | |
| "loss": 0.9913, | |
| "step": 8820 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 4.380032206119163e-05, | |
| "loss": 0.9799, | |
| "step": 8830 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 4.3735909822866346e-05, | |
| "loss": 0.9718, | |
| "step": 8840 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 4.3671497584541064e-05, | |
| "loss": 0.9797, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 4.360708534621578e-05, | |
| "loss": 0.9495, | |
| "step": 8860 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 4.35426731078905e-05, | |
| "loss": 0.9621, | |
| "step": 8870 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 4.347826086956522e-05, | |
| "loss": 0.9725, | |
| "step": 8880 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 4.341384863123994e-05, | |
| "loss": 0.9885, | |
| "step": 8890 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 4.334943639291466e-05, | |
| "loss": 0.9793, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 4.3285024154589376e-05, | |
| "loss": 0.9414, | |
| "step": 8910 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 4.3220611916264095e-05, | |
| "loss": 0.967, | |
| "step": 8920 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 4.315619967793881e-05, | |
| "loss": 0.9524, | |
| "step": 8930 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 4.3091787439613526e-05, | |
| "loss": 0.946, | |
| "step": 8940 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 4.3027375201288245e-05, | |
| "loss": 0.9985, | |
| "step": 8950 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 4.296296296296296e-05, | |
| "loss": 0.9851, | |
| "step": 8960 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 4.289855072463768e-05, | |
| "loss": 0.9708, | |
| "step": 8970 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 4.28341384863124e-05, | |
| "loss": 1.0027, | |
| "step": 8980 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 4.276972624798712e-05, | |
| "loss": 0.9281, | |
| "step": 8990 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 4.270531400966184e-05, | |
| "loss": 0.9813, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "eval_loss": 1.0139975547790527, | |
| "eval_runtime": 187.2523, | |
| "eval_samples_per_second": 10.681, | |
| "eval_steps_per_second": 1.335, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 4.264090177133656e-05, | |
| "loss": 0.9759, | |
| "step": 9010 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 4.2576489533011275e-05, | |
| "loss": 0.9611, | |
| "step": 9020 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 4.2512077294685994e-05, | |
| "loss": 0.9823, | |
| "step": 9030 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 4.2447665056360706e-05, | |
| "loss": 1.007, | |
| "step": 9040 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 4.238325281803543e-05, | |
| "loss": 0.981, | |
| "step": 9050 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 4.2318840579710143e-05, | |
| "loss": 0.948, | |
| "step": 9060 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 4.225442834138487e-05, | |
| "loss": 0.9801, | |
| "step": 9070 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.219001610305958e-05, | |
| "loss": 0.9826, | |
| "step": 9080 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.2125603864734306e-05, | |
| "loss": 0.9682, | |
| "step": 9090 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.206119162640902e-05, | |
| "loss": 0.9964, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.199677938808374e-05, | |
| "loss": 0.9624, | |
| "step": 9110 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 4.1932367149758456e-05, | |
| "loss": 0.9617, | |
| "step": 9120 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 4.1867954911433174e-05, | |
| "loss": 0.9658, | |
| "step": 9130 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 4.180354267310789e-05, | |
| "loss": 0.9926, | |
| "step": 9140 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 4.1739130434782605e-05, | |
| "loss": 0.9645, | |
| "step": 9150 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 4.167471819645733e-05, | |
| "loss": 0.9907, | |
| "step": 9160 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 4.161030595813204e-05, | |
| "loss": 0.9988, | |
| "step": 9170 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 4.154589371980677e-05, | |
| "loss": 0.9527, | |
| "step": 9180 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 4.148148148148148e-05, | |
| "loss": 0.9935, | |
| "step": 9190 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 4.1417069243156205e-05, | |
| "loss": 0.9716, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "eval_loss": 1.013387680053711, | |
| "eval_runtime": 185.6135, | |
| "eval_samples_per_second": 10.775, | |
| "eval_steps_per_second": 1.347, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 4.135265700483092e-05, | |
| "loss": 0.9595, | |
| "step": 9210 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 4.128824476650564e-05, | |
| "loss": 0.9971, | |
| "step": 9220 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 4.1223832528180355e-05, | |
| "loss": 0.9503, | |
| "step": 9230 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 4.115942028985507e-05, | |
| "loss": 1.0049, | |
| "step": 9240 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 4.109500805152979e-05, | |
| "loss": 0.9448, | |
| "step": 9250 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 4.103059581320451e-05, | |
| "loss": 0.9209, | |
| "step": 9260 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 4.096618357487923e-05, | |
| "loss": 0.9852, | |
| "step": 9270 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 4.090177133655395e-05, | |
| "loss": 0.9536, | |
| "step": 9280 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 4.083735909822867e-05, | |
| "loss": 0.9979, | |
| "step": 9290 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 4.0772946859903386e-05, | |
| "loss": 0.976, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 4.0708534621578104e-05, | |
| "loss": 0.9783, | |
| "step": 9310 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 4.0644122383252816e-05, | |
| "loss": 1.0021, | |
| "step": 9320 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 4.057971014492754e-05, | |
| "loss": 0.9696, | |
| "step": 9330 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 4.0515297906602254e-05, | |
| "loss": 0.9533, | |
| "step": 9340 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 4.045088566827697e-05, | |
| "loss": 0.9493, | |
| "step": 9350 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 4.038647342995169e-05, | |
| "loss": 0.9244, | |
| "step": 9360 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 4.032206119162641e-05, | |
| "loss": 0.9857, | |
| "step": 9370 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 4.025764895330113e-05, | |
| "loss": 0.9501, | |
| "step": 9380 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 4.019323671497585e-05, | |
| "loss": 0.9155, | |
| "step": 9390 | |
| }, | |
| { | |
| "epoch": 3.01, | |
| "learning_rate": 4.0128824476650566e-05, | |
| "loss": 0.9645, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 3.01, | |
| "eval_loss": 1.0159056186676025, | |
| "eval_runtime": 185.5657, | |
| "eval_samples_per_second": 10.778, | |
| "eval_steps_per_second": 1.347, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 3.01, | |
| "learning_rate": 4.0064412238325285e-05, | |
| "loss": 0.9724, | |
| "step": 9410 | |
| }, | |
| { | |
| "epoch": 3.01, | |
| "learning_rate": 4e-05, | |
| "loss": 0.9588, | |
| "step": 9420 | |
| }, | |
| { | |
| "epoch": 3.02, | |
| "learning_rate": 3.993558776167472e-05, | |
| "loss": 0.9152, | |
| "step": 9430 | |
| }, | |
| { | |
| "epoch": 3.02, | |
| "learning_rate": 3.987117552334944e-05, | |
| "loss": 0.9943, | |
| "step": 9440 | |
| }, | |
| { | |
| "epoch": 3.02, | |
| "learning_rate": 3.980676328502416e-05, | |
| "loss": 0.9171, | |
| "step": 9450 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "learning_rate": 3.974235104669887e-05, | |
| "loss": 0.9249, | |
| "step": 9460 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "learning_rate": 3.96779388083736e-05, | |
| "loss": 0.9021, | |
| "step": 9470 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "learning_rate": 3.961352657004831e-05, | |
| "loss": 0.9527, | |
| "step": 9480 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "learning_rate": 3.954911433172303e-05, | |
| "loss": 0.9116, | |
| "step": 9490 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "learning_rate": 3.9484702093397746e-05, | |
| "loss": 0.8851, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "learning_rate": 3.9420289855072465e-05, | |
| "loss": 0.9686, | |
| "step": 9510 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "learning_rate": 3.9355877616747184e-05, | |
| "loss": 0.9302, | |
| "step": 9520 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "learning_rate": 3.92914653784219e-05, | |
| "loss": 0.9466, | |
| "step": 9530 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "learning_rate": 3.922705314009662e-05, | |
| "loss": 0.9057, | |
| "step": 9540 | |
| }, | |
| { | |
| "epoch": 3.06, | |
| "learning_rate": 3.916264090177133e-05, | |
| "loss": 0.9495, | |
| "step": 9550 | |
| }, | |
| { | |
| "epoch": 3.06, | |
| "learning_rate": 3.909822866344606e-05, | |
| "loss": 0.9586, | |
| "step": 9560 | |
| }, | |
| { | |
| "epoch": 3.06, | |
| "learning_rate": 3.903381642512077e-05, | |
| "loss": 0.9013, | |
| "step": 9570 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "learning_rate": 3.8969404186795496e-05, | |
| "loss": 0.9334, | |
| "step": 9580 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "learning_rate": 3.890499194847021e-05, | |
| "loss": 0.9388, | |
| "step": 9590 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "learning_rate": 3.884057971014493e-05, | |
| "loss": 0.9922, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "eval_loss": 1.0189166069030762, | |
| "eval_runtime": 185.5091, | |
| "eval_samples_per_second": 10.781, | |
| "eval_steps_per_second": 1.348, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "learning_rate": 3.8776167471819645e-05, | |
| "loss": 0.9233, | |
| "step": 9610 | |
| }, | |
| { | |
| "epoch": 3.08, | |
| "learning_rate": 3.871175523349437e-05, | |
| "loss": 0.9634, | |
| "step": 9620 | |
| }, | |
| { | |
| "epoch": 3.08, | |
| "learning_rate": 3.864734299516908e-05, | |
| "loss": 0.9194, | |
| "step": 9630 | |
| }, | |
| { | |
| "epoch": 3.08, | |
| "learning_rate": 3.858293075684381e-05, | |
| "loss": 0.9787, | |
| "step": 9640 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 3.851851851851852e-05, | |
| "loss": 0.9738, | |
| "step": 9650 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 3.845410628019324e-05, | |
| "loss": 0.974, | |
| "step": 9660 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 3.838969404186796e-05, | |
| "loss": 0.9409, | |
| "step": 9670 | |
| }, | |
| { | |
| "epoch": 3.1, | |
| "learning_rate": 3.8325281803542676e-05, | |
| "loss": 0.9536, | |
| "step": 9680 | |
| }, | |
| { | |
| "epoch": 3.1, | |
| "learning_rate": 3.8260869565217395e-05, | |
| "loss": 0.9244, | |
| "step": 9690 | |
| }, | |
| { | |
| "epoch": 3.1, | |
| "learning_rate": 3.819645732689211e-05, | |
| "loss": 0.9427, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 3.11, | |
| "learning_rate": 3.813204508856683e-05, | |
| "loss": 0.9354, | |
| "step": 9710 | |
| }, | |
| { | |
| "epoch": 3.11, | |
| "learning_rate": 3.8067632850241544e-05, | |
| "loss": 0.9299, | |
| "step": 9720 | |
| }, | |
| { | |
| "epoch": 3.11, | |
| "learning_rate": 3.800322061191627e-05, | |
| "loss": 0.9283, | |
| "step": 9730 | |
| }, | |
| { | |
| "epoch": 3.12, | |
| "learning_rate": 3.793880837359098e-05, | |
| "loss": 0.9511, | |
| "step": 9740 | |
| }, | |
| { | |
| "epoch": 3.12, | |
| "learning_rate": 3.787439613526571e-05, | |
| "loss": 0.9484, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 3.12, | |
| "learning_rate": 3.780998389694042e-05, | |
| "loss": 0.9257, | |
| "step": 9760 | |
| }, | |
| { | |
| "epoch": 3.13, | |
| "learning_rate": 3.774557165861514e-05, | |
| "loss": 0.9735, | |
| "step": 9770 | |
| }, | |
| { | |
| "epoch": 3.13, | |
| "learning_rate": 3.7681159420289856e-05, | |
| "loss": 0.9484, | |
| "step": 9780 | |
| }, | |
| { | |
| "epoch": 3.13, | |
| "learning_rate": 3.7616747181964575e-05, | |
| "loss": 0.9602, | |
| "step": 9790 | |
| }, | |
| { | |
| "epoch": 3.14, | |
| "learning_rate": 3.7552334943639294e-05, | |
| "loss": 0.9633, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 3.14, | |
| "eval_loss": 1.0187861919403076, | |
| "eval_runtime": 185.4008, | |
| "eval_samples_per_second": 10.787, | |
| "eval_steps_per_second": 1.348, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 3.14, | |
| "learning_rate": 3.748792270531401e-05, | |
| "loss": 0.9493, | |
| "step": 9810 | |
| }, | |
| { | |
| "epoch": 3.14, | |
| "learning_rate": 3.742351046698873e-05, | |
| "loss": 0.9233, | |
| "step": 9820 | |
| }, | |
| { | |
| "epoch": 3.15, | |
| "learning_rate": 3.735909822866345e-05, | |
| "loss": 0.9254, | |
| "step": 9830 | |
| }, | |
| { | |
| "epoch": 3.15, | |
| "learning_rate": 3.729468599033817e-05, | |
| "loss": 0.9369, | |
| "step": 9840 | |
| }, | |
| { | |
| "epoch": 3.15, | |
| "learning_rate": 3.723027375201289e-05, | |
| "loss": 0.9657, | |
| "step": 9850 | |
| }, | |
| { | |
| "epoch": 3.15, | |
| "learning_rate": 3.71658615136876e-05, | |
| "loss": 0.9604, | |
| "step": 9860 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "learning_rate": 3.710144927536232e-05, | |
| "loss": 0.9487, | |
| "step": 9870 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "learning_rate": 3.7037037037037037e-05, | |
| "loss": 0.9339, | |
| "step": 9880 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "learning_rate": 3.6972624798711755e-05, | |
| "loss": 0.9448, | |
| "step": 9890 | |
| }, | |
| { | |
| "epoch": 3.17, | |
| "learning_rate": 3.6908212560386474e-05, | |
| "loss": 0.9705, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 3.17, | |
| "learning_rate": 3.684380032206119e-05, | |
| "loss": 0.9557, | |
| "step": 9910 | |
| }, | |
| { | |
| "epoch": 3.17, | |
| "learning_rate": 3.677938808373591e-05, | |
| "loss": 0.9182, | |
| "step": 9920 | |
| }, | |
| { | |
| "epoch": 3.18, | |
| "learning_rate": 3.671497584541063e-05, | |
| "loss": 0.9427, | |
| "step": 9930 | |
| }, | |
| { | |
| "epoch": 3.18, | |
| "learning_rate": 3.665056360708535e-05, | |
| "loss": 0.9685, | |
| "step": 9940 | |
| }, | |
| { | |
| "epoch": 3.18, | |
| "learning_rate": 3.658615136876007e-05, | |
| "loss": 0.9145, | |
| "step": 9950 | |
| }, | |
| { | |
| "epoch": 3.19, | |
| "learning_rate": 3.6521739130434786e-05, | |
| "loss": 0.9437, | |
| "step": 9960 | |
| }, | |
| { | |
| "epoch": 3.19, | |
| "learning_rate": 3.64573268921095e-05, | |
| "loss": 0.9433, | |
| "step": 9970 | |
| }, | |
| { | |
| "epoch": 3.19, | |
| "learning_rate": 3.6392914653784224e-05, | |
| "loss": 0.9535, | |
| "step": 9980 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "learning_rate": 3.6328502415458936e-05, | |
| "loss": 0.9413, | |
| "step": 9990 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "learning_rate": 3.626409017713366e-05, | |
| "loss": 0.9191, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "eval_loss": 1.021185278892517, | |
| "eval_runtime": 185.5678, | |
| "eval_samples_per_second": 10.778, | |
| "eval_steps_per_second": 1.347, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "learning_rate": 3.619967793880837e-05, | |
| "loss": 0.9523, | |
| "step": 10010 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 3.61352657004831e-05, | |
| "loss": 0.9134, | |
| "step": 10020 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 3.607085346215781e-05, | |
| "loss": 0.9614, | |
| "step": 10030 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 3.600644122383253e-05, | |
| "loss": 1.0293, | |
| "step": 10040 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "learning_rate": 3.594202898550725e-05, | |
| "loss": 0.9625, | |
| "step": 10050 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "learning_rate": 3.5877616747181966e-05, | |
| "loss": 0.9045, | |
| "step": 10060 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "learning_rate": 3.5813204508856685e-05, | |
| "loss": 0.9246, | |
| "step": 10070 | |
| }, | |
| { | |
| "epoch": 3.23, | |
| "learning_rate": 3.57487922705314e-05, | |
| "loss": 0.8946, | |
| "step": 10080 | |
| }, | |
| { | |
| "epoch": 3.23, | |
| "learning_rate": 3.568438003220612e-05, | |
| "loss": 0.9478, | |
| "step": 10090 | |
| }, | |
| { | |
| "epoch": 3.23, | |
| "learning_rate": 3.5619967793880834e-05, | |
| "loss": 0.9524, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 3.23, | |
| "learning_rate": 3.555555555555556e-05, | |
| "loss": 0.9536, | |
| "step": 10110 | |
| }, | |
| { | |
| "epoch": 3.24, | |
| "learning_rate": 3.549114331723027e-05, | |
| "loss": 0.967, | |
| "step": 10120 | |
| }, | |
| { | |
| "epoch": 3.24, | |
| "learning_rate": 3.5426731078905e-05, | |
| "loss": 0.9211, | |
| "step": 10130 | |
| }, | |
| { | |
| "epoch": 3.24, | |
| "learning_rate": 3.536231884057971e-05, | |
| "loss": 0.9923, | |
| "step": 10140 | |
| }, | |
| { | |
| "epoch": 3.25, | |
| "learning_rate": 3.5297906602254435e-05, | |
| "loss": 0.9603, | |
| "step": 10150 | |
| }, | |
| { | |
| "epoch": 3.25, | |
| "learning_rate": 3.523349436392915e-05, | |
| "loss": 0.9683, | |
| "step": 10160 | |
| }, | |
| { | |
| "epoch": 3.25, | |
| "learning_rate": 3.5169082125603865e-05, | |
| "loss": 0.9934, | |
| "step": 10170 | |
| }, | |
| { | |
| "epoch": 3.26, | |
| "learning_rate": 3.5104669887278584e-05, | |
| "loss": 0.9405, | |
| "step": 10180 | |
| }, | |
| { | |
| "epoch": 3.26, | |
| "learning_rate": 3.50402576489533e-05, | |
| "loss": 0.965, | |
| "step": 10190 | |
| }, | |
| { | |
| "epoch": 3.26, | |
| "learning_rate": 3.497584541062802e-05, | |
| "loss": 0.9408, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 3.26, | |
| "eval_loss": 1.0209527015686035, | |
| "eval_runtime": 185.3872, | |
| "eval_samples_per_second": 10.788, | |
| "eval_steps_per_second": 1.349, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 3.27, | |
| "learning_rate": 3.491143317230274e-05, | |
| "loss": 0.9095, | |
| "step": 10210 | |
| }, | |
| { | |
| "epoch": 3.27, | |
| "learning_rate": 3.484702093397746e-05, | |
| "loss": 0.9343, | |
| "step": 10220 | |
| }, | |
| { | |
| "epoch": 3.27, | |
| "learning_rate": 3.478260869565218e-05, | |
| "loss": 0.9609, | |
| "step": 10230 | |
| }, | |
| { | |
| "epoch": 3.28, | |
| "learning_rate": 3.4718196457326896e-05, | |
| "loss": 0.9451, | |
| "step": 10240 | |
| }, | |
| { | |
| "epoch": 3.28, | |
| "learning_rate": 3.465378421900161e-05, | |
| "loss": 0.983, | |
| "step": 10250 | |
| }, | |
| { | |
| "epoch": 3.28, | |
| "learning_rate": 3.4589371980676334e-05, | |
| "loss": 0.9515, | |
| "step": 10260 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 3.4524959742351046e-05, | |
| "loss": 0.9411, | |
| "step": 10270 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 3.4460547504025764e-05, | |
| "loss": 0.9106, | |
| "step": 10280 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 3.439613526570048e-05, | |
| "loss": 0.9859, | |
| "step": 10290 | |
| }, | |
| { | |
| "epoch": 3.3, | |
| "learning_rate": 3.43317230273752e-05, | |
| "loss": 0.9778, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 3.3, | |
| "learning_rate": 3.426731078904992e-05, | |
| "loss": 0.9504, | |
| "step": 10310 | |
| }, | |
| { | |
| "epoch": 3.3, | |
| "learning_rate": 3.420289855072464e-05, | |
| "loss": 0.9215, | |
| "step": 10320 | |
| }, | |
| { | |
| "epoch": 3.31, | |
| "learning_rate": 3.413848631239936e-05, | |
| "loss": 0.9195, | |
| "step": 10330 | |
| }, | |
| { | |
| "epoch": 3.31, | |
| "learning_rate": 3.4074074074074077e-05, | |
| "loss": 0.9375, | |
| "step": 10340 | |
| }, | |
| { | |
| "epoch": 3.31, | |
| "learning_rate": 3.4009661835748795e-05, | |
| "loss": 0.9505, | |
| "step": 10350 | |
| }, | |
| { | |
| "epoch": 3.31, | |
| "learning_rate": 3.3945249597423514e-05, | |
| "loss": 0.9539, | |
| "step": 10360 | |
| }, | |
| { | |
| "epoch": 3.32, | |
| "learning_rate": 3.3880837359098226e-05, | |
| "loss": 0.9822, | |
| "step": 10370 | |
| }, | |
| { | |
| "epoch": 3.32, | |
| "learning_rate": 3.381642512077295e-05, | |
| "loss": 0.9272, | |
| "step": 10380 | |
| }, | |
| { | |
| "epoch": 3.32, | |
| "learning_rate": 3.375201288244766e-05, | |
| "loss": 0.9588, | |
| "step": 10390 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 3.368760064412239e-05, | |
| "loss": 0.919, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "eval_loss": 1.0204192399978638, | |
| "eval_runtime": 185.6273, | |
| "eval_samples_per_second": 10.774, | |
| "eval_steps_per_second": 1.347, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 3.36231884057971e-05, | |
| "loss": 0.9416, | |
| "step": 10410 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 3.355877616747182e-05, | |
| "loss": 0.9138, | |
| "step": 10420 | |
| }, | |
| { | |
| "epoch": 3.34, | |
| "learning_rate": 3.349436392914654e-05, | |
| "loss": 0.9681, | |
| "step": 10430 | |
| }, | |
| { | |
| "epoch": 3.34, | |
| "learning_rate": 3.342995169082126e-05, | |
| "loss": 0.9824, | |
| "step": 10440 | |
| }, | |
| { | |
| "epoch": 3.34, | |
| "learning_rate": 3.3365539452495976e-05, | |
| "loss": 0.9876, | |
| "step": 10450 | |
| }, | |
| { | |
| "epoch": 3.35, | |
| "learning_rate": 3.3301127214170694e-05, | |
| "loss": 0.9239, | |
| "step": 10460 | |
| }, | |
| { | |
| "epoch": 3.35, | |
| "learning_rate": 3.323671497584541e-05, | |
| "loss": 0.9314, | |
| "step": 10470 | |
| }, | |
| { | |
| "epoch": 3.35, | |
| "learning_rate": 3.3172302737520125e-05, | |
| "loss": 0.9426, | |
| "step": 10480 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "learning_rate": 3.310789049919485e-05, | |
| "loss": 0.9282, | |
| "step": 10490 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "learning_rate": 3.304347826086956e-05, | |
| "loss": 0.9742, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "learning_rate": 3.297906602254429e-05, | |
| "loss": 0.9916, | |
| "step": 10510 | |
| }, | |
| { | |
| "epoch": 3.37, | |
| "learning_rate": 3.2914653784219e-05, | |
| "loss": 0.9346, | |
| "step": 10520 | |
| }, | |
| { | |
| "epoch": 3.37, | |
| "learning_rate": 3.2850241545893725e-05, | |
| "loss": 0.9036, | |
| "step": 10530 | |
| }, | |
| { | |
| "epoch": 3.37, | |
| "learning_rate": 3.278582930756844e-05, | |
| "loss": 0.9029, | |
| "step": 10540 | |
| }, | |
| { | |
| "epoch": 3.38, | |
| "learning_rate": 3.272141706924316e-05, | |
| "loss": 0.9495, | |
| "step": 10550 | |
| }, | |
| { | |
| "epoch": 3.38, | |
| "learning_rate": 3.2657004830917875e-05, | |
| "loss": 0.9441, | |
| "step": 10560 | |
| }, | |
| { | |
| "epoch": 3.38, | |
| "learning_rate": 3.25925925925926e-05, | |
| "loss": 0.9757, | |
| "step": 10570 | |
| }, | |
| { | |
| "epoch": 3.39, | |
| "learning_rate": 3.252818035426731e-05, | |
| "loss": 0.9667, | |
| "step": 10580 | |
| }, | |
| { | |
| "epoch": 3.39, | |
| "learning_rate": 3.246376811594203e-05, | |
| "loss": 0.9603, | |
| "step": 10590 | |
| }, | |
| { | |
| "epoch": 3.39, | |
| "learning_rate": 3.239935587761675e-05, | |
| "loss": 0.9362, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 3.39, | |
| "eval_loss": 1.0204658508300781, | |
| "eval_runtime": 185.6227, | |
| "eval_samples_per_second": 10.775, | |
| "eval_steps_per_second": 1.347, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 3.39, | |
| "learning_rate": 3.233494363929147e-05, | |
| "loss": 0.915, | |
| "step": 10610 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 3.227053140096619e-05, | |
| "loss": 0.9603, | |
| "step": 10620 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 3.22061191626409e-05, | |
| "loss": 0.91, | |
| "step": 10630 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 3.2141706924315624e-05, | |
| "loss": 0.9411, | |
| "step": 10640 | |
| }, | |
| { | |
| "epoch": 3.41, | |
| "learning_rate": 3.2077294685990336e-05, | |
| "loss": 0.9688, | |
| "step": 10650 | |
| }, | |
| { | |
| "epoch": 3.41, | |
| "learning_rate": 3.201288244766506e-05, | |
| "loss": 0.9213, | |
| "step": 10660 | |
| }, | |
| { | |
| "epoch": 3.41, | |
| "learning_rate": 3.1948470209339774e-05, | |
| "loss": 0.9911, | |
| "step": 10670 | |
| }, | |
| { | |
| "epoch": 3.42, | |
| "learning_rate": 3.188405797101449e-05, | |
| "loss": 0.9298, | |
| "step": 10680 | |
| }, | |
| { | |
| "epoch": 3.42, | |
| "learning_rate": 3.181964573268921e-05, | |
| "loss": 0.9331, | |
| "step": 10690 | |
| }, | |
| { | |
| "epoch": 3.42, | |
| "learning_rate": 3.175523349436393e-05, | |
| "loss": 0.938, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "learning_rate": 3.169082125603865e-05, | |
| "loss": 0.97, | |
| "step": 10710 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "learning_rate": 3.162640901771337e-05, | |
| "loss": 0.9349, | |
| "step": 10720 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "learning_rate": 3.1561996779388086e-05, | |
| "loss": 0.9143, | |
| "step": 10730 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 3.1497584541062804e-05, | |
| "loss": 0.986, | |
| "step": 10740 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 3.143317230273752e-05, | |
| "loss": 0.9483, | |
| "step": 10750 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 3.136876006441224e-05, | |
| "loss": 0.9623, | |
| "step": 10760 | |
| }, | |
| { | |
| "epoch": 3.45, | |
| "learning_rate": 3.130434782608696e-05, | |
| "loss": 0.9497, | |
| "step": 10770 | |
| }, | |
| { | |
| "epoch": 3.45, | |
| "learning_rate": 3.123993558776168e-05, | |
| "loss": 0.928, | |
| "step": 10780 | |
| }, | |
| { | |
| "epoch": 3.45, | |
| "learning_rate": 3.117552334943639e-05, | |
| "loss": 0.9597, | |
| "step": 10790 | |
| }, | |
| { | |
| "epoch": 3.46, | |
| "learning_rate": 3.111111111111111e-05, | |
| "loss": 0.9228, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 3.46, | |
| "eval_loss": 1.019952416419983, | |
| "eval_runtime": 185.7809, | |
| "eval_samples_per_second": 10.765, | |
| "eval_steps_per_second": 1.346, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 3.46, | |
| "learning_rate": 3.104669887278583e-05, | |
| "loss": 0.9535, | |
| "step": 10810 | |
| }, | |
| { | |
| "epoch": 3.46, | |
| "learning_rate": 3.098228663446055e-05, | |
| "loss": 0.9234, | |
| "step": 10820 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "learning_rate": 3.0917874396135266e-05, | |
| "loss": 0.968, | |
| "step": 10830 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "learning_rate": 3.0853462157809985e-05, | |
| "loss": 0.9003, | |
| "step": 10840 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "learning_rate": 3.0789049919484703e-05, | |
| "loss": 0.9745, | |
| "step": 10850 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "learning_rate": 3.072463768115942e-05, | |
| "loss": 0.9016, | |
| "step": 10860 | |
| }, | |
| { | |
| "epoch": 3.48, | |
| "learning_rate": 3.066022544283414e-05, | |
| "loss": 0.9605, | |
| "step": 10870 | |
| }, | |
| { | |
| "epoch": 3.48, | |
| "learning_rate": 3.059581320450886e-05, | |
| "loss": 0.8885, | |
| "step": 10880 | |
| }, | |
| { | |
| "epoch": 3.48, | |
| "learning_rate": 3.053140096618358e-05, | |
| "loss": 0.962, | |
| "step": 10890 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 3.0466988727858294e-05, | |
| "loss": 0.9208, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 3.0402576489533012e-05, | |
| "loss": 0.9032, | |
| "step": 10910 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 3.0338164251207728e-05, | |
| "loss": 0.9875, | |
| "step": 10920 | |
| }, | |
| { | |
| "epoch": 3.5, | |
| "learning_rate": 3.027375201288245e-05, | |
| "loss": 0.9408, | |
| "step": 10930 | |
| }, | |
| { | |
| "epoch": 3.5, | |
| "learning_rate": 3.0209339774557165e-05, | |
| "loss": 0.9283, | |
| "step": 10940 | |
| }, | |
| { | |
| "epoch": 3.5, | |
| "learning_rate": 3.0144927536231887e-05, | |
| "loss": 0.9668, | |
| "step": 10950 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "learning_rate": 3.0080515297906602e-05, | |
| "loss": 0.9202, | |
| "step": 10960 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "learning_rate": 3.0016103059581324e-05, | |
| "loss": 0.9675, | |
| "step": 10970 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "learning_rate": 2.995169082125604e-05, | |
| "loss": 0.9549, | |
| "step": 10980 | |
| }, | |
| { | |
| "epoch": 3.52, | |
| "learning_rate": 2.9887278582930755e-05, | |
| "loss": 0.9618, | |
| "step": 10990 | |
| }, | |
| { | |
| "epoch": 3.52, | |
| "learning_rate": 2.9822866344605477e-05, | |
| "loss": 0.9135, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 3.52, | |
| "eval_loss": 1.0204010009765625, | |
| "eval_runtime": 185.6196, | |
| "eval_samples_per_second": 10.775, | |
| "eval_steps_per_second": 1.347, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 3.52, | |
| "learning_rate": 2.9758454106280193e-05, | |
| "loss": 0.9539, | |
| "step": 11010 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "learning_rate": 2.9694041867954915e-05, | |
| "loss": 0.964, | |
| "step": 11020 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "learning_rate": 2.962962962962963e-05, | |
| "loss": 0.9265, | |
| "step": 11030 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "learning_rate": 2.9565217391304352e-05, | |
| "loss": 0.9267, | |
| "step": 11040 | |
| }, | |
| { | |
| "epoch": 3.54, | |
| "learning_rate": 2.9500805152979067e-05, | |
| "loss": 0.9282, | |
| "step": 11050 | |
| }, | |
| { | |
| "epoch": 3.54, | |
| "learning_rate": 2.9436392914653786e-05, | |
| "loss": 0.9555, | |
| "step": 11060 | |
| }, | |
| { | |
| "epoch": 3.54, | |
| "learning_rate": 2.9371980676328505e-05, | |
| "loss": 0.925, | |
| "step": 11070 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "learning_rate": 2.9307568438003223e-05, | |
| "loss": 0.9174, | |
| "step": 11080 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "learning_rate": 2.924315619967794e-05, | |
| "loss": 0.9491, | |
| "step": 11090 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "learning_rate": 2.9178743961352657e-05, | |
| "loss": 0.9035, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "learning_rate": 2.9114331723027376e-05, | |
| "loss": 0.9239, | |
| "step": 11110 | |
| }, | |
| { | |
| "epoch": 3.56, | |
| "learning_rate": 2.904991948470209e-05, | |
| "loss": 0.9158, | |
| "step": 11120 | |
| }, | |
| { | |
| "epoch": 3.56, | |
| "learning_rate": 2.8985507246376814e-05, | |
| "loss": 0.9734, | |
| "step": 11130 | |
| }, | |
| { | |
| "epoch": 3.56, | |
| "learning_rate": 2.892109500805153e-05, | |
| "loss": 0.9419, | |
| "step": 11140 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 2.885668276972625e-05, | |
| "loss": 0.9582, | |
| "step": 11150 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 2.8792270531400966e-05, | |
| "loss": 0.9283, | |
| "step": 11160 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 2.872785829307569e-05, | |
| "loss": 0.932, | |
| "step": 11170 | |
| }, | |
| { | |
| "epoch": 3.58, | |
| "learning_rate": 2.8663446054750404e-05, | |
| "loss": 0.9349, | |
| "step": 11180 | |
| }, | |
| { | |
| "epoch": 3.58, | |
| "learning_rate": 2.8599033816425126e-05, | |
| "loss": 0.9378, | |
| "step": 11190 | |
| }, | |
| { | |
| "epoch": 3.58, | |
| "learning_rate": 2.853462157809984e-05, | |
| "loss": 0.9915, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 3.58, | |
| "eval_loss": 1.0208972692489624, | |
| "eval_runtime": 185.8582, | |
| "eval_samples_per_second": 10.761, | |
| "eval_steps_per_second": 1.345, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 3.59, | |
| "learning_rate": 2.8470209339774556e-05, | |
| "loss": 0.957, | |
| "step": 11210 | |
| }, | |
| { | |
| "epoch": 3.59, | |
| "learning_rate": 2.840579710144928e-05, | |
| "loss": 0.9514, | |
| "step": 11220 | |
| }, | |
| { | |
| "epoch": 3.59, | |
| "learning_rate": 2.8341384863123994e-05, | |
| "loss": 0.9931, | |
| "step": 11230 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 2.8276972624798716e-05, | |
| "loss": 0.9304, | |
| "step": 11240 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 2.821256038647343e-05, | |
| "loss": 0.9281, | |
| "step": 11250 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 2.814814814814815e-05, | |
| "loss": 0.9726, | |
| "step": 11260 | |
| }, | |
| { | |
| "epoch": 3.61, | |
| "learning_rate": 2.8083735909822865e-05, | |
| "loss": 0.9539, | |
| "step": 11270 | |
| }, | |
| { | |
| "epoch": 3.61, | |
| "learning_rate": 2.8019323671497587e-05, | |
| "loss": 0.959, | |
| "step": 11280 | |
| }, | |
| { | |
| "epoch": 3.61, | |
| "learning_rate": 2.7954911433172303e-05, | |
| "loss": 0.9409, | |
| "step": 11290 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 2.7890499194847018e-05, | |
| "loss": 0.8763, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 2.782608695652174e-05, | |
| "loss": 0.9532, | |
| "step": 11310 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 2.7761674718196455e-05, | |
| "loss": 0.9698, | |
| "step": 11320 | |
| }, | |
| { | |
| "epoch": 3.63, | |
| "learning_rate": 2.7697262479871177e-05, | |
| "loss": 0.9248, | |
| "step": 11330 | |
| }, | |
| { | |
| "epoch": 3.63, | |
| "learning_rate": 2.7632850241545893e-05, | |
| "loss": 0.9537, | |
| "step": 11340 | |
| }, | |
| { | |
| "epoch": 3.63, | |
| "learning_rate": 2.7568438003220615e-05, | |
| "loss": 0.8953, | |
| "step": 11350 | |
| }, | |
| { | |
| "epoch": 3.63, | |
| "learning_rate": 2.750402576489533e-05, | |
| "loss": 0.9163, | |
| "step": 11360 | |
| }, | |
| { | |
| "epoch": 3.64, | |
| "learning_rate": 2.7439613526570052e-05, | |
| "loss": 0.9223, | |
| "step": 11370 | |
| }, | |
| { | |
| "epoch": 3.64, | |
| "learning_rate": 2.7375201288244768e-05, | |
| "loss": 0.9459, | |
| "step": 11380 | |
| }, | |
| { | |
| "epoch": 3.64, | |
| "learning_rate": 2.731078904991949e-05, | |
| "loss": 0.9296, | |
| "step": 11390 | |
| }, | |
| { | |
| "epoch": 3.65, | |
| "learning_rate": 2.7246376811594205e-05, | |
| "loss": 0.9182, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 3.65, | |
| "eval_loss": 1.020736813545227, | |
| "eval_runtime": 185.7004, | |
| "eval_samples_per_second": 10.77, | |
| "eval_steps_per_second": 1.346, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 3.65, | |
| "learning_rate": 2.718196457326892e-05, | |
| "loss": 0.8978, | |
| "step": 11410 | |
| }, | |
| { | |
| "epoch": 3.65, | |
| "learning_rate": 2.7117552334943642e-05, | |
| "loss": 0.952, | |
| "step": 11420 | |
| }, | |
| { | |
| "epoch": 3.66, | |
| "learning_rate": 2.7053140096618358e-05, | |
| "loss": 0.9485, | |
| "step": 11430 | |
| }, | |
| { | |
| "epoch": 3.66, | |
| "learning_rate": 2.6988727858293076e-05, | |
| "loss": 0.9477, | |
| "step": 11440 | |
| }, | |
| { | |
| "epoch": 3.66, | |
| "learning_rate": 2.6924315619967795e-05, | |
| "loss": 0.9488, | |
| "step": 11450 | |
| }, | |
| { | |
| "epoch": 3.67, | |
| "learning_rate": 2.6859903381642514e-05, | |
| "loss": 0.9239, | |
| "step": 11460 | |
| }, | |
| { | |
| "epoch": 3.67, | |
| "learning_rate": 2.679549114331723e-05, | |
| "loss": 0.9253, | |
| "step": 11470 | |
| }, | |
| { | |
| "epoch": 3.67, | |
| "learning_rate": 2.673107890499195e-05, | |
| "loss": 0.9307, | |
| "step": 11480 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "learning_rate": 2.6666666666666667e-05, | |
| "loss": 0.9309, | |
| "step": 11490 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "learning_rate": 2.660225442834139e-05, | |
| "loss": 0.9163, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "learning_rate": 2.6537842190016104e-05, | |
| "loss": 0.9616, | |
| "step": 11510 | |
| }, | |
| { | |
| "epoch": 3.69, | |
| "learning_rate": 2.647342995169082e-05, | |
| "loss": 0.9303, | |
| "step": 11520 | |
| }, | |
| { | |
| "epoch": 3.69, | |
| "learning_rate": 2.640901771336554e-05, | |
| "loss": 0.9434, | |
| "step": 11530 | |
| }, | |
| { | |
| "epoch": 3.69, | |
| "learning_rate": 2.6344605475040257e-05, | |
| "loss": 0.9441, | |
| "step": 11540 | |
| }, | |
| { | |
| "epoch": 3.7, | |
| "learning_rate": 2.628019323671498e-05, | |
| "loss": 0.9193, | |
| "step": 11550 | |
| }, | |
| { | |
| "epoch": 3.7, | |
| "learning_rate": 2.6215780998389694e-05, | |
| "loss": 0.9246, | |
| "step": 11560 | |
| }, | |
| { | |
| "epoch": 3.7, | |
| "learning_rate": 2.6151368760064416e-05, | |
| "loss": 0.9663, | |
| "step": 11570 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "learning_rate": 2.608695652173913e-05, | |
| "loss": 0.9627, | |
| "step": 11580 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "learning_rate": 2.6022544283413854e-05, | |
| "loss": 0.9495, | |
| "step": 11590 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "learning_rate": 2.595813204508857e-05, | |
| "loss": 0.9912, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "eval_loss": 1.0197750329971313, | |
| "eval_runtime": 185.734, | |
| "eval_samples_per_second": 10.768, | |
| "eval_steps_per_second": 1.346, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "learning_rate": 2.5893719806763284e-05, | |
| "loss": 0.9793, | |
| "step": 11610 | |
| }, | |
| { | |
| "epoch": 3.72, | |
| "learning_rate": 2.5829307568438006e-05, | |
| "loss": 0.9954, | |
| "step": 11620 | |
| }, | |
| { | |
| "epoch": 3.72, | |
| "learning_rate": 2.576489533011272e-05, | |
| "loss": 0.9784, | |
| "step": 11630 | |
| }, | |
| { | |
| "epoch": 3.72, | |
| "learning_rate": 2.570048309178744e-05, | |
| "loss": 0.9772, | |
| "step": 11640 | |
| }, | |
| { | |
| "epoch": 3.73, | |
| "learning_rate": 2.563607085346216e-05, | |
| "loss": 0.9081, | |
| "step": 11650 | |
| }, | |
| { | |
| "epoch": 3.73, | |
| "learning_rate": 2.5571658615136878e-05, | |
| "loss": 1.0142, | |
| "step": 11660 | |
| }, | |
| { | |
| "epoch": 3.73, | |
| "learning_rate": 2.5507246376811593e-05, | |
| "loss": 0.9344, | |
| "step": 11670 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 2.5442834138486315e-05, | |
| "loss": 0.898, | |
| "step": 11680 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 2.537842190016103e-05, | |
| "loss": 0.9452, | |
| "step": 11690 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 2.5314009661835753e-05, | |
| "loss": 0.9549, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 3.75, | |
| "learning_rate": 2.5249597423510468e-05, | |
| "loss": 0.9725, | |
| "step": 11710 | |
| }, | |
| { | |
| "epoch": 3.75, | |
| "learning_rate": 2.5185185185185183e-05, | |
| "loss": 0.9805, | |
| "step": 11720 | |
| }, | |
| { | |
| "epoch": 3.75, | |
| "learning_rate": 2.5120772946859905e-05, | |
| "loss": 0.943, | |
| "step": 11730 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "learning_rate": 2.505636070853462e-05, | |
| "loss": 0.9829, | |
| "step": 11740 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "learning_rate": 2.4991948470209343e-05, | |
| "loss": 0.8923, | |
| "step": 11750 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "learning_rate": 2.492753623188406e-05, | |
| "loss": 0.941, | |
| "step": 11760 | |
| }, | |
| { | |
| "epoch": 3.77, | |
| "learning_rate": 2.4863123993558777e-05, | |
| "loss": 0.9365, | |
| "step": 11770 | |
| }, | |
| { | |
| "epoch": 3.77, | |
| "learning_rate": 2.4798711755233495e-05, | |
| "loss": 0.9625, | |
| "step": 11780 | |
| }, | |
| { | |
| "epoch": 3.77, | |
| "learning_rate": 2.4734299516908214e-05, | |
| "loss": 0.9093, | |
| "step": 11790 | |
| }, | |
| { | |
| "epoch": 3.78, | |
| "learning_rate": 2.4669887278582933e-05, | |
| "loss": 0.9494, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 3.78, | |
| "eval_loss": 1.0202767848968506, | |
| "eval_runtime": 185.7202, | |
| "eval_samples_per_second": 10.769, | |
| "eval_steps_per_second": 1.346, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 3.78, | |
| "learning_rate": 2.460547504025765e-05, | |
| "loss": 0.9672, | |
| "step": 11810 | |
| }, | |
| { | |
| "epoch": 3.78, | |
| "learning_rate": 2.4541062801932367e-05, | |
| "loss": 0.9556, | |
| "step": 11820 | |
| }, | |
| { | |
| "epoch": 3.79, | |
| "learning_rate": 2.4476650563607086e-05, | |
| "loss": 0.9633, | |
| "step": 11830 | |
| }, | |
| { | |
| "epoch": 3.79, | |
| "learning_rate": 2.4412238325281804e-05, | |
| "loss": 0.9554, | |
| "step": 11840 | |
| }, | |
| { | |
| "epoch": 3.79, | |
| "learning_rate": 2.4347826086956523e-05, | |
| "loss": 0.9376, | |
| "step": 11850 | |
| }, | |
| { | |
| "epoch": 3.79, | |
| "learning_rate": 2.4283413848631242e-05, | |
| "loss": 0.945, | |
| "step": 11860 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 2.4219001610305957e-05, | |
| "loss": 0.9306, | |
| "step": 11870 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 2.4154589371980676e-05, | |
| "loss": 0.9107, | |
| "step": 11880 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 2.4090177133655394e-05, | |
| "loss": 0.9122, | |
| "step": 11890 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "learning_rate": 2.4025764895330113e-05, | |
| "loss": 0.9185, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "learning_rate": 2.3961352657004832e-05, | |
| "loss": 0.9245, | |
| "step": 11910 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "learning_rate": 2.389694041867955e-05, | |
| "loss": 0.978, | |
| "step": 11920 | |
| }, | |
| { | |
| "epoch": 3.82, | |
| "learning_rate": 2.383252818035427e-05, | |
| "loss": 0.915, | |
| "step": 11930 | |
| }, | |
| { | |
| "epoch": 3.82, | |
| "learning_rate": 2.3768115942028988e-05, | |
| "loss": 0.9275, | |
| "step": 11940 | |
| }, | |
| { | |
| "epoch": 3.82, | |
| "learning_rate": 2.3703703703703707e-05, | |
| "loss": 0.8838, | |
| "step": 11950 | |
| }, | |
| { | |
| "epoch": 3.83, | |
| "learning_rate": 2.3639291465378425e-05, | |
| "loss": 0.9676, | |
| "step": 11960 | |
| }, | |
| { | |
| "epoch": 3.83, | |
| "learning_rate": 2.357487922705314e-05, | |
| "loss": 0.9168, | |
| "step": 11970 | |
| }, | |
| { | |
| "epoch": 3.83, | |
| "learning_rate": 2.351046698872786e-05, | |
| "loss": 0.9476, | |
| "step": 11980 | |
| }, | |
| { | |
| "epoch": 3.84, | |
| "learning_rate": 2.3446054750402578e-05, | |
| "loss": 0.9697, | |
| "step": 11990 | |
| }, | |
| { | |
| "epoch": 3.84, | |
| "learning_rate": 2.3381642512077297e-05, | |
| "loss": 0.9911, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 3.84, | |
| "eval_loss": 1.0203664302825928, | |
| "eval_runtime": 185.728, | |
| "eval_samples_per_second": 10.768, | |
| "eval_steps_per_second": 1.346, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 3.84, | |
| "learning_rate": 2.3317230273752012e-05, | |
| "loss": 0.9203, | |
| "step": 12010 | |
| }, | |
| { | |
| "epoch": 3.85, | |
| "learning_rate": 2.325281803542673e-05, | |
| "loss": 0.9131, | |
| "step": 12020 | |
| }, | |
| { | |
| "epoch": 3.85, | |
| "learning_rate": 2.318840579710145e-05, | |
| "loss": 0.9434, | |
| "step": 12030 | |
| }, | |
| { | |
| "epoch": 3.85, | |
| "learning_rate": 2.3123993558776168e-05, | |
| "loss": 0.9394, | |
| "step": 12040 | |
| }, | |
| { | |
| "epoch": 3.86, | |
| "learning_rate": 2.3059581320450887e-05, | |
| "loss": 0.9288, | |
| "step": 12050 | |
| }, | |
| { | |
| "epoch": 3.86, | |
| "learning_rate": 2.2995169082125606e-05, | |
| "loss": 0.8921, | |
| "step": 12060 | |
| }, | |
| { | |
| "epoch": 3.86, | |
| "learning_rate": 2.2930756843800324e-05, | |
| "loss": 0.9627, | |
| "step": 12070 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 2.286634460547504e-05, | |
| "loss": 0.9223, | |
| "step": 12080 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 2.280193236714976e-05, | |
| "loss": 0.9458, | |
| "step": 12090 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 2.2737520128824477e-05, | |
| "loss": 0.9444, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 2.2673107890499196e-05, | |
| "loss": 0.9566, | |
| "step": 12110 | |
| }, | |
| { | |
| "epoch": 3.88, | |
| "learning_rate": 2.2608695652173914e-05, | |
| "loss": 0.9318, | |
| "step": 12120 | |
| }, | |
| { | |
| "epoch": 3.88, | |
| "learning_rate": 2.2544283413848633e-05, | |
| "loss": 0.9321, | |
| "step": 12130 | |
| }, | |
| { | |
| "epoch": 3.88, | |
| "learning_rate": 2.2479871175523352e-05, | |
| "loss": 0.916, | |
| "step": 12140 | |
| }, | |
| { | |
| "epoch": 3.89, | |
| "learning_rate": 2.241545893719807e-05, | |
| "loss": 0.9448, | |
| "step": 12150 | |
| }, | |
| { | |
| "epoch": 3.89, | |
| "learning_rate": 2.235104669887279e-05, | |
| "loss": 0.9434, | |
| "step": 12160 | |
| }, | |
| { | |
| "epoch": 3.89, | |
| "learning_rate": 2.2286634460547508e-05, | |
| "loss": 0.9573, | |
| "step": 12170 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 2.2222222222222223e-05, | |
| "loss": 0.9395, | |
| "step": 12180 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 2.2157809983896942e-05, | |
| "loss": 0.9259, | |
| "step": 12190 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 2.2093397745571657e-05, | |
| "loss": 0.9804, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "eval_loss": 1.020531177520752, | |
| "eval_runtime": 185.969, | |
| "eval_samples_per_second": 10.754, | |
| "eval_steps_per_second": 1.344, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 3.91, | |
| "learning_rate": 2.2028985507246376e-05, | |
| "loss": 0.9442, | |
| "step": 12210 | |
| }, | |
| { | |
| "epoch": 3.91, | |
| "learning_rate": 2.1964573268921095e-05, | |
| "loss": 0.9528, | |
| "step": 12220 | |
| }, | |
| { | |
| "epoch": 3.91, | |
| "learning_rate": 2.1900161030595813e-05, | |
| "loss": 0.9734, | |
| "step": 12230 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "learning_rate": 2.1835748792270532e-05, | |
| "loss": 0.9398, | |
| "step": 12240 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "learning_rate": 2.177133655394525e-05, | |
| "loss": 0.9482, | |
| "step": 12250 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "learning_rate": 2.170692431561997e-05, | |
| "loss": 0.9219, | |
| "step": 12260 | |
| }, | |
| { | |
| "epoch": 3.93, | |
| "learning_rate": 2.1642512077294688e-05, | |
| "loss": 0.9085, | |
| "step": 12270 | |
| }, | |
| { | |
| "epoch": 3.93, | |
| "learning_rate": 2.1578099838969404e-05, | |
| "loss": 0.9246, | |
| "step": 12280 | |
| }, | |
| { | |
| "epoch": 3.93, | |
| "learning_rate": 2.1513687600644122e-05, | |
| "loss": 0.9608, | |
| "step": 12290 | |
| }, | |
| { | |
| "epoch": 3.94, | |
| "learning_rate": 2.144927536231884e-05, | |
| "loss": 0.9587, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 3.94, | |
| "learning_rate": 2.138486312399356e-05, | |
| "loss": 0.9486, | |
| "step": 12310 | |
| }, | |
| { | |
| "epoch": 3.94, | |
| "learning_rate": 2.132045088566828e-05, | |
| "loss": 0.9677, | |
| "step": 12320 | |
| }, | |
| { | |
| "epoch": 3.95, | |
| "learning_rate": 2.1256038647342997e-05, | |
| "loss": 0.9397, | |
| "step": 12330 | |
| }, | |
| { | |
| "epoch": 3.95, | |
| "learning_rate": 2.1191626409017716e-05, | |
| "loss": 0.9501, | |
| "step": 12340 | |
| }, | |
| { | |
| "epoch": 3.95, | |
| "learning_rate": 2.1127214170692434e-05, | |
| "loss": 0.9385, | |
| "step": 12350 | |
| }, | |
| { | |
| "epoch": 3.95, | |
| "learning_rate": 2.1062801932367153e-05, | |
| "loss": 0.9012, | |
| "step": 12360 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "learning_rate": 2.099838969404187e-05, | |
| "loss": 0.9114, | |
| "step": 12370 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "learning_rate": 2.0933977455716587e-05, | |
| "loss": 0.897, | |
| "step": 12380 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "learning_rate": 2.0869565217391303e-05, | |
| "loss": 0.9286, | |
| "step": 12390 | |
| }, | |
| { | |
| "epoch": 3.97, | |
| "learning_rate": 2.080515297906602e-05, | |
| "loss": 0.9204, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 3.97, | |
| "eval_loss": 1.0194604396820068, | |
| "eval_runtime": 185.8824, | |
| "eval_samples_per_second": 10.759, | |
| "eval_steps_per_second": 1.345, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 3.97, | |
| "learning_rate": 2.074074074074074e-05, | |
| "loss": 0.9409, | |
| "step": 12410 | |
| }, | |
| { | |
| "epoch": 3.97, | |
| "learning_rate": 2.067632850241546e-05, | |
| "loss": 0.9488, | |
| "step": 12420 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 2.0611916264090177e-05, | |
| "loss": 0.9582, | |
| "step": 12430 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 2.0547504025764896e-05, | |
| "loss": 0.9216, | |
| "step": 12440 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 2.0483091787439615e-05, | |
| "loss": 0.9372, | |
| "step": 12450 | |
| }, | |
| { | |
| "epoch": 3.99, | |
| "learning_rate": 2.0418679549114333e-05, | |
| "loss": 0.9349, | |
| "step": 12460 | |
| }, | |
| { | |
| "epoch": 3.99, | |
| "learning_rate": 2.0354267310789052e-05, | |
| "loss": 0.9378, | |
| "step": 12470 | |
| }, | |
| { | |
| "epoch": 3.99, | |
| "learning_rate": 2.028985507246377e-05, | |
| "loss": 0.8969, | |
| "step": 12480 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 2.0225442834138486e-05, | |
| "loss": 0.932, | |
| "step": 12490 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 2.0161030595813205e-05, | |
| "loss": 0.9217, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 2.0096618357487924e-05, | |
| "loss": 0.9195, | |
| "step": 12510 | |
| }, | |
| { | |
| "epoch": 4.01, | |
| "learning_rate": 2.0032206119162642e-05, | |
| "loss": 0.8911, | |
| "step": 12520 | |
| }, | |
| { | |
| "epoch": 4.01, | |
| "learning_rate": 1.996779388083736e-05, | |
| "loss": 0.9525, | |
| "step": 12530 | |
| }, | |
| { | |
| "epoch": 4.01, | |
| "learning_rate": 1.990338164251208e-05, | |
| "loss": 0.8723, | |
| "step": 12540 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 1.98389694041868e-05, | |
| "loss": 0.9319, | |
| "step": 12550 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 1.9774557165861514e-05, | |
| "loss": 0.9389, | |
| "step": 12560 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 1.9710144927536232e-05, | |
| "loss": 0.951, | |
| "step": 12570 | |
| }, | |
| { | |
| "epoch": 4.03, | |
| "learning_rate": 1.964573268921095e-05, | |
| "loss": 0.94, | |
| "step": 12580 | |
| }, | |
| { | |
| "epoch": 4.03, | |
| "learning_rate": 1.9581320450885666e-05, | |
| "loss": 0.8881, | |
| "step": 12590 | |
| }, | |
| { | |
| "epoch": 4.03, | |
| "learning_rate": 1.9516908212560385e-05, | |
| "loss": 0.9259, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 4.03, | |
| "eval_loss": 1.0242584943771362, | |
| "eval_runtime": 185.6505, | |
| "eval_samples_per_second": 10.773, | |
| "eval_steps_per_second": 1.347, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 4.03, | |
| "learning_rate": 1.9452495974235104e-05, | |
| "loss": 0.9263, | |
| "step": 12610 | |
| }, | |
| { | |
| "epoch": 4.04, | |
| "learning_rate": 1.9388083735909823e-05, | |
| "loss": 0.9086, | |
| "step": 12620 | |
| }, | |
| { | |
| "epoch": 4.04, | |
| "learning_rate": 1.932367149758454e-05, | |
| "loss": 0.9166, | |
| "step": 12630 | |
| }, | |
| { | |
| "epoch": 4.04, | |
| "learning_rate": 1.925925925925926e-05, | |
| "loss": 0.8954, | |
| "step": 12640 | |
| }, | |
| { | |
| "epoch": 4.05, | |
| "learning_rate": 1.919484702093398e-05, | |
| "loss": 0.9404, | |
| "step": 12650 | |
| }, | |
| { | |
| "epoch": 4.05, | |
| "learning_rate": 1.9130434782608697e-05, | |
| "loss": 0.9082, | |
| "step": 12660 | |
| }, | |
| { | |
| "epoch": 4.05, | |
| "learning_rate": 1.9066022544283416e-05, | |
| "loss": 0.8876, | |
| "step": 12670 | |
| }, | |
| { | |
| "epoch": 4.06, | |
| "learning_rate": 1.9001610305958135e-05, | |
| "loss": 0.9323, | |
| "step": 12680 | |
| }, | |
| { | |
| "epoch": 4.06, | |
| "learning_rate": 1.8937198067632853e-05, | |
| "loss": 0.9473, | |
| "step": 12690 | |
| }, | |
| { | |
| "epoch": 4.06, | |
| "learning_rate": 1.887278582930757e-05, | |
| "loss": 0.924, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 4.07, | |
| "learning_rate": 1.8808373590982287e-05, | |
| "loss": 0.9408, | |
| "step": 12710 | |
| }, | |
| { | |
| "epoch": 4.07, | |
| "learning_rate": 1.8743961352657006e-05, | |
| "loss": 0.9696, | |
| "step": 12720 | |
| }, | |
| { | |
| "epoch": 4.07, | |
| "learning_rate": 1.8679549114331725e-05, | |
| "loss": 0.8972, | |
| "step": 12730 | |
| }, | |
| { | |
| "epoch": 4.08, | |
| "learning_rate": 1.8615136876006444e-05, | |
| "loss": 0.9433, | |
| "step": 12740 | |
| }, | |
| { | |
| "epoch": 4.08, | |
| "learning_rate": 1.855072463768116e-05, | |
| "loss": 0.885, | |
| "step": 12750 | |
| }, | |
| { | |
| "epoch": 4.08, | |
| "learning_rate": 1.8486312399355878e-05, | |
| "loss": 0.8641, | |
| "step": 12760 | |
| }, | |
| { | |
| "epoch": 4.09, | |
| "learning_rate": 1.8421900161030596e-05, | |
| "loss": 0.9207, | |
| "step": 12770 | |
| }, | |
| { | |
| "epoch": 4.09, | |
| "learning_rate": 1.8357487922705315e-05, | |
| "loss": 0.9599, | |
| "step": 12780 | |
| }, | |
| { | |
| "epoch": 4.09, | |
| "learning_rate": 1.8293075684380034e-05, | |
| "loss": 0.933, | |
| "step": 12790 | |
| }, | |
| { | |
| "epoch": 4.1, | |
| "learning_rate": 1.822866344605475e-05, | |
| "loss": 0.8847, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 4.1, | |
| "eval_loss": 1.0249016284942627, | |
| "eval_runtime": 185.1711, | |
| "eval_samples_per_second": 10.801, | |
| "eval_steps_per_second": 1.35, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 4.1, | |
| "learning_rate": 1.8164251207729468e-05, | |
| "loss": 0.9367, | |
| "step": 12810 | |
| }, | |
| { | |
| "epoch": 4.1, | |
| "learning_rate": 1.8099838969404186e-05, | |
| "loss": 0.8933, | |
| "step": 12820 | |
| }, | |
| { | |
| "epoch": 4.11, | |
| "learning_rate": 1.8035426731078905e-05, | |
| "loss": 0.9401, | |
| "step": 12830 | |
| }, | |
| { | |
| "epoch": 4.11, | |
| "learning_rate": 1.7971014492753624e-05, | |
| "loss": 0.8991, | |
| "step": 12840 | |
| }, | |
| { | |
| "epoch": 4.11, | |
| "learning_rate": 1.7906602254428343e-05, | |
| "loss": 0.918, | |
| "step": 12850 | |
| }, | |
| { | |
| "epoch": 4.11, | |
| "learning_rate": 1.784219001610306e-05, | |
| "loss": 0.8747, | |
| "step": 12860 | |
| }, | |
| { | |
| "epoch": 4.12, | |
| "learning_rate": 1.777777777777778e-05, | |
| "loss": 0.912, | |
| "step": 12870 | |
| }, | |
| { | |
| "epoch": 4.12, | |
| "learning_rate": 1.77133655394525e-05, | |
| "loss": 0.9363, | |
| "step": 12880 | |
| }, | |
| { | |
| "epoch": 4.12, | |
| "learning_rate": 1.7648953301127217e-05, | |
| "loss": 0.9562, | |
| "step": 12890 | |
| }, | |
| { | |
| "epoch": 4.13, | |
| "learning_rate": 1.7584541062801933e-05, | |
| "loss": 0.8962, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 4.13, | |
| "learning_rate": 1.752012882447665e-05, | |
| "loss": 0.9182, | |
| "step": 12910 | |
| }, | |
| { | |
| "epoch": 4.13, | |
| "learning_rate": 1.745571658615137e-05, | |
| "loss": 0.8928, | |
| "step": 12920 | |
| }, | |
| { | |
| "epoch": 4.14, | |
| "learning_rate": 1.739130434782609e-05, | |
| "loss": 0.9068, | |
| "step": 12930 | |
| }, | |
| { | |
| "epoch": 4.14, | |
| "learning_rate": 1.7326892109500804e-05, | |
| "loss": 0.9521, | |
| "step": 12940 | |
| }, | |
| { | |
| "epoch": 4.14, | |
| "learning_rate": 1.7262479871175523e-05, | |
| "loss": 0.9492, | |
| "step": 12950 | |
| }, | |
| { | |
| "epoch": 4.15, | |
| "learning_rate": 1.719806763285024e-05, | |
| "loss": 0.9432, | |
| "step": 12960 | |
| }, | |
| { | |
| "epoch": 4.15, | |
| "learning_rate": 1.713365539452496e-05, | |
| "loss": 0.9103, | |
| "step": 12970 | |
| }, | |
| { | |
| "epoch": 4.15, | |
| "learning_rate": 1.706924315619968e-05, | |
| "loss": 0.8698, | |
| "step": 12980 | |
| }, | |
| { | |
| "epoch": 4.16, | |
| "learning_rate": 1.7004830917874398e-05, | |
| "loss": 0.9173, | |
| "step": 12990 | |
| }, | |
| { | |
| "epoch": 4.16, | |
| "learning_rate": 1.6940418679549113e-05, | |
| "loss": 0.9322, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 4.16, | |
| "eval_loss": 1.0260157585144043, | |
| "eval_runtime": 185.2834, | |
| "eval_samples_per_second": 10.794, | |
| "eval_steps_per_second": 1.349, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 4.16, | |
| "learning_rate": 1.687600644122383e-05, | |
| "loss": 0.9101, | |
| "step": 13010 | |
| }, | |
| { | |
| "epoch": 4.17, | |
| "learning_rate": 1.681159420289855e-05, | |
| "loss": 0.9106, | |
| "step": 13020 | |
| }, | |
| { | |
| "epoch": 4.17, | |
| "learning_rate": 1.674718196457327e-05, | |
| "loss": 0.9305, | |
| "step": 13030 | |
| }, | |
| { | |
| "epoch": 4.17, | |
| "learning_rate": 1.6682769726247988e-05, | |
| "loss": 0.921, | |
| "step": 13040 | |
| }, | |
| { | |
| "epoch": 4.18, | |
| "learning_rate": 1.6618357487922706e-05, | |
| "loss": 0.9068, | |
| "step": 13050 | |
| }, | |
| { | |
| "epoch": 4.18, | |
| "learning_rate": 1.6553945249597425e-05, | |
| "loss": 0.9289, | |
| "step": 13060 | |
| }, | |
| { | |
| "epoch": 4.18, | |
| "learning_rate": 1.6489533011272144e-05, | |
| "loss": 0.9705, | |
| "step": 13070 | |
| }, | |
| { | |
| "epoch": 4.19, | |
| "learning_rate": 1.6425120772946863e-05, | |
| "loss": 0.904, | |
| "step": 13080 | |
| }, | |
| { | |
| "epoch": 4.19, | |
| "learning_rate": 1.636070853462158e-05, | |
| "loss": 0.928, | |
| "step": 13090 | |
| }, | |
| { | |
| "epoch": 4.19, | |
| "learning_rate": 1.62962962962963e-05, | |
| "loss": 0.8668, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 4.19, | |
| "learning_rate": 1.6231884057971015e-05, | |
| "loss": 0.9068, | |
| "step": 13110 | |
| }, | |
| { | |
| "epoch": 4.2, | |
| "learning_rate": 1.6167471819645734e-05, | |
| "loss": 0.9461, | |
| "step": 13120 | |
| }, | |
| { | |
| "epoch": 4.2, | |
| "learning_rate": 1.610305958132045e-05, | |
| "loss": 0.9201, | |
| "step": 13130 | |
| }, | |
| { | |
| "epoch": 4.2, | |
| "learning_rate": 1.6038647342995168e-05, | |
| "loss": 0.9456, | |
| "step": 13140 | |
| }, | |
| { | |
| "epoch": 4.21, | |
| "learning_rate": 1.5974235104669887e-05, | |
| "loss": 0.93, | |
| "step": 13150 | |
| }, | |
| { | |
| "epoch": 4.21, | |
| "learning_rate": 1.5909822866344605e-05, | |
| "loss": 0.9325, | |
| "step": 13160 | |
| }, | |
| { | |
| "epoch": 4.21, | |
| "learning_rate": 1.5845410628019324e-05, | |
| "loss": 0.9258, | |
| "step": 13170 | |
| }, | |
| { | |
| "epoch": 4.22, | |
| "learning_rate": 1.5780998389694043e-05, | |
| "loss": 0.9383, | |
| "step": 13180 | |
| }, | |
| { | |
| "epoch": 4.22, | |
| "learning_rate": 1.571658615136876e-05, | |
| "loss": 0.9102, | |
| "step": 13190 | |
| }, | |
| { | |
| "epoch": 4.22, | |
| "learning_rate": 1.565217391304348e-05, | |
| "loss": 0.9468, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 4.22, | |
| "eval_loss": 1.0262330770492554, | |
| "eval_runtime": 185.7469, | |
| "eval_samples_per_second": 10.767, | |
| "eval_steps_per_second": 1.346, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 4.23, | |
| "learning_rate": 1.5587761674718196e-05, | |
| "loss": 0.8922, | |
| "step": 13210 | |
| }, | |
| { | |
| "epoch": 4.23, | |
| "learning_rate": 1.5523349436392914e-05, | |
| "loss": 0.8682, | |
| "step": 13220 | |
| }, | |
| { | |
| "epoch": 4.23, | |
| "learning_rate": 1.5458937198067633e-05, | |
| "loss": 0.9325, | |
| "step": 13230 | |
| }, | |
| { | |
| "epoch": 4.24, | |
| "learning_rate": 1.5394524959742352e-05, | |
| "loss": 0.9187, | |
| "step": 13240 | |
| }, | |
| { | |
| "epoch": 4.24, | |
| "learning_rate": 1.533011272141707e-05, | |
| "loss": 0.9496, | |
| "step": 13250 | |
| }, | |
| { | |
| "epoch": 4.24, | |
| "learning_rate": 1.526570048309179e-05, | |
| "loss": 0.8942, | |
| "step": 13260 | |
| }, | |
| { | |
| "epoch": 4.25, | |
| "learning_rate": 1.5201288244766506e-05, | |
| "loss": 0.899, | |
| "step": 13270 | |
| }, | |
| { | |
| "epoch": 4.25, | |
| "learning_rate": 1.5136876006441225e-05, | |
| "loss": 0.923, | |
| "step": 13280 | |
| }, | |
| { | |
| "epoch": 4.25, | |
| "learning_rate": 1.5072463768115944e-05, | |
| "loss": 0.9093, | |
| "step": 13290 | |
| }, | |
| { | |
| "epoch": 4.26, | |
| "learning_rate": 1.5008051529790662e-05, | |
| "loss": 0.9187, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 4.26, | |
| "learning_rate": 1.4943639291465378e-05, | |
| "loss": 0.9003, | |
| "step": 13310 | |
| }, | |
| { | |
| "epoch": 4.26, | |
| "learning_rate": 1.4879227053140096e-05, | |
| "loss": 0.9492, | |
| "step": 13320 | |
| }, | |
| { | |
| "epoch": 4.27, | |
| "learning_rate": 1.4814814814814815e-05, | |
| "loss": 0.8683, | |
| "step": 13330 | |
| }, | |
| { | |
| "epoch": 4.27, | |
| "learning_rate": 1.4750402576489534e-05, | |
| "loss": 0.921, | |
| "step": 13340 | |
| }, | |
| { | |
| "epoch": 4.27, | |
| "learning_rate": 1.4685990338164252e-05, | |
| "loss": 0.9394, | |
| "step": 13350 | |
| }, | |
| { | |
| "epoch": 4.27, | |
| "learning_rate": 1.462157809983897e-05, | |
| "loss": 0.914, | |
| "step": 13360 | |
| }, | |
| { | |
| "epoch": 4.28, | |
| "learning_rate": 1.4557165861513688e-05, | |
| "loss": 0.9091, | |
| "step": 13370 | |
| }, | |
| { | |
| "epoch": 4.28, | |
| "learning_rate": 1.4492753623188407e-05, | |
| "loss": 0.9052, | |
| "step": 13380 | |
| }, | |
| { | |
| "epoch": 4.28, | |
| "learning_rate": 1.4428341384863125e-05, | |
| "loss": 0.9106, | |
| "step": 13390 | |
| }, | |
| { | |
| "epoch": 4.29, | |
| "learning_rate": 1.4363929146537844e-05, | |
| "loss": 0.9246, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 4.29, | |
| "eval_loss": 1.0260928869247437, | |
| "eval_runtime": 185.7325, | |
| "eval_samples_per_second": 10.768, | |
| "eval_steps_per_second": 1.346, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 4.29, | |
| "learning_rate": 1.4299516908212563e-05, | |
| "loss": 0.9285, | |
| "step": 13410 | |
| }, | |
| { | |
| "epoch": 4.29, | |
| "learning_rate": 1.4235104669887278e-05, | |
| "loss": 0.9403, | |
| "step": 13420 | |
| }, | |
| { | |
| "epoch": 4.3, | |
| "learning_rate": 1.4170692431561997e-05, | |
| "loss": 0.9398, | |
| "step": 13430 | |
| }, | |
| { | |
| "epoch": 4.3, | |
| "learning_rate": 1.4106280193236716e-05, | |
| "loss": 0.9141, | |
| "step": 13440 | |
| }, | |
| { | |
| "epoch": 4.3, | |
| "learning_rate": 1.4041867954911433e-05, | |
| "loss": 0.918, | |
| "step": 13450 | |
| }, | |
| { | |
| "epoch": 4.31, | |
| "learning_rate": 1.3977455716586151e-05, | |
| "loss": 0.9358, | |
| "step": 13460 | |
| }, | |
| { | |
| "epoch": 4.31, | |
| "learning_rate": 1.391304347826087e-05, | |
| "loss": 0.9295, | |
| "step": 13470 | |
| }, | |
| { | |
| "epoch": 4.31, | |
| "learning_rate": 1.3848631239935589e-05, | |
| "loss": 0.9361, | |
| "step": 13480 | |
| }, | |
| { | |
| "epoch": 4.32, | |
| "learning_rate": 1.3784219001610307e-05, | |
| "loss": 0.9246, | |
| "step": 13490 | |
| }, | |
| { | |
| "epoch": 4.32, | |
| "learning_rate": 1.3719806763285026e-05, | |
| "loss": 0.9352, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 4.32, | |
| "learning_rate": 1.3655394524959745e-05, | |
| "loss": 0.9274, | |
| "step": 13510 | |
| }, | |
| { | |
| "epoch": 4.33, | |
| "learning_rate": 1.359098228663446e-05, | |
| "loss": 0.903, | |
| "step": 13520 | |
| }, | |
| { | |
| "epoch": 4.33, | |
| "learning_rate": 1.3526570048309179e-05, | |
| "loss": 0.9416, | |
| "step": 13530 | |
| }, | |
| { | |
| "epoch": 4.33, | |
| "learning_rate": 1.3462157809983898e-05, | |
| "loss": 0.9134, | |
| "step": 13540 | |
| }, | |
| { | |
| "epoch": 4.34, | |
| "learning_rate": 1.3397745571658615e-05, | |
| "loss": 0.9292, | |
| "step": 13550 | |
| }, | |
| { | |
| "epoch": 4.34, | |
| "learning_rate": 1.3333333333333333e-05, | |
| "loss": 0.9274, | |
| "step": 13560 | |
| }, | |
| { | |
| "epoch": 4.34, | |
| "learning_rate": 1.3268921095008052e-05, | |
| "loss": 0.9214, | |
| "step": 13570 | |
| }, | |
| { | |
| "epoch": 4.35, | |
| "learning_rate": 1.320450885668277e-05, | |
| "loss": 0.9906, | |
| "step": 13580 | |
| }, | |
| { | |
| "epoch": 4.35, | |
| "learning_rate": 1.314009661835749e-05, | |
| "loss": 0.903, | |
| "step": 13590 | |
| }, | |
| { | |
| "epoch": 4.35, | |
| "learning_rate": 1.3075684380032208e-05, | |
| "loss": 0.9098, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 4.35, | |
| "eval_loss": 1.0272597074508667, | |
| "eval_runtime": 185.8178, | |
| "eval_samples_per_second": 10.763, | |
| "eval_steps_per_second": 1.345, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 4.35, | |
| "learning_rate": 1.3011272141706927e-05, | |
| "loss": 0.9239, | |
| "step": 13610 | |
| }, | |
| { | |
| "epoch": 4.36, | |
| "learning_rate": 1.2946859903381642e-05, | |
| "loss": 0.9003, | |
| "step": 13620 | |
| }, | |
| { | |
| "epoch": 4.36, | |
| "learning_rate": 1.288244766505636e-05, | |
| "loss": 0.9239, | |
| "step": 13630 | |
| }, | |
| { | |
| "epoch": 4.36, | |
| "learning_rate": 1.281803542673108e-05, | |
| "loss": 0.868, | |
| "step": 13640 | |
| }, | |
| { | |
| "epoch": 4.37, | |
| "learning_rate": 1.2753623188405797e-05, | |
| "loss": 0.9009, | |
| "step": 13650 | |
| }, | |
| { | |
| "epoch": 4.37, | |
| "learning_rate": 1.2689210950080515e-05, | |
| "loss": 0.8691, | |
| "step": 13660 | |
| }, | |
| { | |
| "epoch": 4.37, | |
| "learning_rate": 1.2624798711755234e-05, | |
| "loss": 0.9485, | |
| "step": 13670 | |
| }, | |
| { | |
| "epoch": 4.38, | |
| "learning_rate": 1.2560386473429953e-05, | |
| "loss": 0.9548, | |
| "step": 13680 | |
| }, | |
| { | |
| "epoch": 4.38, | |
| "learning_rate": 1.2495974235104671e-05, | |
| "loss": 0.9139, | |
| "step": 13690 | |
| }, | |
| { | |
| "epoch": 4.38, | |
| "learning_rate": 1.2431561996779388e-05, | |
| "loss": 0.8865, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 4.39, | |
| "learning_rate": 1.2367149758454107e-05, | |
| "loss": 0.9589, | |
| "step": 13710 | |
| }, | |
| { | |
| "epoch": 4.39, | |
| "learning_rate": 1.2302737520128826e-05, | |
| "loss": 0.9178, | |
| "step": 13720 | |
| }, | |
| { | |
| "epoch": 4.39, | |
| "learning_rate": 1.2238325281803543e-05, | |
| "loss": 0.8904, | |
| "step": 13730 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 1.2173913043478261e-05, | |
| "loss": 0.9689, | |
| "step": 13740 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 1.2109500805152979e-05, | |
| "loss": 0.9332, | |
| "step": 13750 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 1.2045088566827697e-05, | |
| "loss": 0.8886, | |
| "step": 13760 | |
| }, | |
| { | |
| "epoch": 4.41, | |
| "learning_rate": 1.1980676328502416e-05, | |
| "loss": 0.9236, | |
| "step": 13770 | |
| }, | |
| { | |
| "epoch": 4.41, | |
| "learning_rate": 1.1916264090177135e-05, | |
| "loss": 0.9277, | |
| "step": 13780 | |
| }, | |
| { | |
| "epoch": 4.41, | |
| "learning_rate": 1.1851851851851853e-05, | |
| "loss": 0.9584, | |
| "step": 13790 | |
| }, | |
| { | |
| "epoch": 4.42, | |
| "learning_rate": 1.178743961352657e-05, | |
| "loss": 0.9438, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 4.42, | |
| "eval_loss": 1.0258300304412842, | |
| "eval_runtime": 185.6037, | |
| "eval_samples_per_second": 10.776, | |
| "eval_steps_per_second": 1.347, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 4.42, | |
| "learning_rate": 1.1723027375201289e-05, | |
| "loss": 0.8793, | |
| "step": 13810 | |
| }, | |
| { | |
| "epoch": 4.42, | |
| "learning_rate": 1.1658615136876006e-05, | |
| "loss": 0.9195, | |
| "step": 13820 | |
| }, | |
| { | |
| "epoch": 4.43, | |
| "learning_rate": 1.1594202898550725e-05, | |
| "loss": 0.9315, | |
| "step": 13830 | |
| }, | |
| { | |
| "epoch": 4.43, | |
| "learning_rate": 1.1529790660225443e-05, | |
| "loss": 0.9155, | |
| "step": 13840 | |
| }, | |
| { | |
| "epoch": 4.43, | |
| "learning_rate": 1.1465378421900162e-05, | |
| "loss": 0.9861, | |
| "step": 13850 | |
| }, | |
| { | |
| "epoch": 4.43, | |
| "learning_rate": 1.140096618357488e-05, | |
| "loss": 0.9285, | |
| "step": 13860 | |
| }, | |
| { | |
| "epoch": 4.44, | |
| "learning_rate": 1.1336553945249598e-05, | |
| "loss": 0.9304, | |
| "step": 13870 | |
| }, | |
| { | |
| "epoch": 4.44, | |
| "learning_rate": 1.1272141706924317e-05, | |
| "loss": 0.8874, | |
| "step": 13880 | |
| }, | |
| { | |
| "epoch": 4.44, | |
| "learning_rate": 1.1207729468599035e-05, | |
| "loss": 0.9123, | |
| "step": 13890 | |
| }, | |
| { | |
| "epoch": 4.45, | |
| "learning_rate": 1.1143317230273754e-05, | |
| "loss": 0.9032, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 4.45, | |
| "learning_rate": 1.1078904991948471e-05, | |
| "loss": 0.9069, | |
| "step": 13910 | |
| }, | |
| { | |
| "epoch": 4.45, | |
| "learning_rate": 1.1014492753623188e-05, | |
| "loss": 0.9081, | |
| "step": 13920 | |
| }, | |
| { | |
| "epoch": 4.46, | |
| "learning_rate": 1.0950080515297907e-05, | |
| "loss": 0.9451, | |
| "step": 13930 | |
| }, | |
| { | |
| "epoch": 4.46, | |
| "learning_rate": 1.0885668276972625e-05, | |
| "loss": 0.8822, | |
| "step": 13940 | |
| }, | |
| { | |
| "epoch": 4.46, | |
| "learning_rate": 1.0821256038647344e-05, | |
| "loss": 0.9095, | |
| "step": 13950 | |
| }, | |
| { | |
| "epoch": 4.47, | |
| "learning_rate": 1.0756843800322061e-05, | |
| "loss": 0.9091, | |
| "step": 13960 | |
| }, | |
| { | |
| "epoch": 4.47, | |
| "learning_rate": 1.069243156199678e-05, | |
| "loss": 0.9613, | |
| "step": 13970 | |
| }, | |
| { | |
| "epoch": 4.47, | |
| "learning_rate": 1.0628019323671499e-05, | |
| "loss": 0.9071, | |
| "step": 13980 | |
| }, | |
| { | |
| "epoch": 4.48, | |
| "learning_rate": 1.0563607085346217e-05, | |
| "loss": 0.9167, | |
| "step": 13990 | |
| }, | |
| { | |
| "epoch": 4.48, | |
| "learning_rate": 1.0499194847020934e-05, | |
| "loss": 0.9141, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 4.48, | |
| "eval_loss": 1.0271040201187134, | |
| "eval_runtime": 185.513, | |
| "eval_samples_per_second": 10.781, | |
| "eval_steps_per_second": 1.348, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 4.48, | |
| "learning_rate": 1.0434782608695651e-05, | |
| "loss": 0.9496, | |
| "step": 14010 | |
| }, | |
| { | |
| "epoch": 4.49, | |
| "learning_rate": 1.037037037037037e-05, | |
| "loss": 0.9135, | |
| "step": 14020 | |
| }, | |
| { | |
| "epoch": 4.49, | |
| "learning_rate": 1.0305958132045089e-05, | |
| "loss": 0.93, | |
| "step": 14030 | |
| }, | |
| { | |
| "epoch": 4.49, | |
| "learning_rate": 1.0241545893719807e-05, | |
| "loss": 0.9371, | |
| "step": 14040 | |
| }, | |
| { | |
| "epoch": 4.5, | |
| "learning_rate": 1.0177133655394526e-05, | |
| "loss": 0.8959, | |
| "step": 14050 | |
| }, | |
| { | |
| "epoch": 4.5, | |
| "learning_rate": 1.0112721417069243e-05, | |
| "loss": 0.9462, | |
| "step": 14060 | |
| }, | |
| { | |
| "epoch": 4.5, | |
| "learning_rate": 1.0048309178743962e-05, | |
| "loss": 0.9088, | |
| "step": 14070 | |
| }, | |
| { | |
| "epoch": 4.51, | |
| "learning_rate": 9.98389694041868e-06, | |
| "loss": 0.94, | |
| "step": 14080 | |
| }, | |
| { | |
| "epoch": 4.51, | |
| "learning_rate": 9.9194847020934e-06, | |
| "loss": 0.9294, | |
| "step": 14090 | |
| }, | |
| { | |
| "epoch": 4.51, | |
| "learning_rate": 9.855072463768116e-06, | |
| "loss": 0.9061, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 4.51, | |
| "learning_rate": 9.790660225442833e-06, | |
| "loss": 0.9217, | |
| "step": 14110 | |
| }, | |
| { | |
| "epoch": 4.52, | |
| "learning_rate": 9.726247987117552e-06, | |
| "loss": 0.9119, | |
| "step": 14120 | |
| }, | |
| { | |
| "epoch": 4.52, | |
| "learning_rate": 9.66183574879227e-06, | |
| "loss": 0.9177, | |
| "step": 14130 | |
| }, | |
| { | |
| "epoch": 4.52, | |
| "learning_rate": 9.59742351046699e-06, | |
| "loss": 0.9344, | |
| "step": 14140 | |
| }, | |
| { | |
| "epoch": 4.53, | |
| "learning_rate": 9.533011272141708e-06, | |
| "loss": 0.9323, | |
| "step": 14150 | |
| }, | |
| { | |
| "epoch": 4.53, | |
| "learning_rate": 9.468599033816427e-06, | |
| "loss": 0.887, | |
| "step": 14160 | |
| }, | |
| { | |
| "epoch": 4.53, | |
| "learning_rate": 9.404186795491144e-06, | |
| "loss": 0.9849, | |
| "step": 14170 | |
| }, | |
| { | |
| "epoch": 4.54, | |
| "learning_rate": 9.339774557165862e-06, | |
| "loss": 0.8887, | |
| "step": 14180 | |
| }, | |
| { | |
| "epoch": 4.54, | |
| "learning_rate": 9.27536231884058e-06, | |
| "loss": 0.8912, | |
| "step": 14190 | |
| }, | |
| { | |
| "epoch": 4.54, | |
| "learning_rate": 9.210950080515298e-06, | |
| "loss": 0.901, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 4.54, | |
| "eval_loss": 1.0271906852722168, | |
| "eval_runtime": 185.5026, | |
| "eval_samples_per_second": 10.782, | |
| "eval_steps_per_second": 1.348, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 4.55, | |
| "learning_rate": 9.146537842190017e-06, | |
| "loss": 0.9378, | |
| "step": 14210 | |
| }, | |
| { | |
| "epoch": 4.55, | |
| "learning_rate": 9.082125603864734e-06, | |
| "loss": 0.9211, | |
| "step": 14220 | |
| }, | |
| { | |
| "epoch": 4.55, | |
| "learning_rate": 9.017713365539453e-06, | |
| "loss": 0.9268, | |
| "step": 14230 | |
| }, | |
| { | |
| "epoch": 4.56, | |
| "learning_rate": 8.953301127214171e-06, | |
| "loss": 0.9014, | |
| "step": 14240 | |
| }, | |
| { | |
| "epoch": 4.56, | |
| "learning_rate": 8.88888888888889e-06, | |
| "loss": 1.0102, | |
| "step": 14250 | |
| }, | |
| { | |
| "epoch": 4.56, | |
| "learning_rate": 8.824476650563609e-06, | |
| "loss": 0.8908, | |
| "step": 14260 | |
| }, | |
| { | |
| "epoch": 4.57, | |
| "learning_rate": 8.760064412238326e-06, | |
| "loss": 0.9336, | |
| "step": 14270 | |
| }, | |
| { | |
| "epoch": 4.57, | |
| "learning_rate": 8.695652173913044e-06, | |
| "loss": 0.9673, | |
| "step": 14280 | |
| }, | |
| { | |
| "epoch": 4.57, | |
| "learning_rate": 8.631239935587761e-06, | |
| "loss": 0.9396, | |
| "step": 14290 | |
| }, | |
| { | |
| "epoch": 4.58, | |
| "learning_rate": 8.56682769726248e-06, | |
| "loss": 0.9446, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 4.58, | |
| "learning_rate": 8.502415458937199e-06, | |
| "loss": 0.9203, | |
| "step": 14310 | |
| }, | |
| { | |
| "epoch": 4.58, | |
| "learning_rate": 8.438003220611916e-06, | |
| "loss": 0.8645, | |
| "step": 14320 | |
| }, | |
| { | |
| "epoch": 4.59, | |
| "learning_rate": 8.373590982286635e-06, | |
| "loss": 0.9053, | |
| "step": 14330 | |
| }, | |
| { | |
| "epoch": 4.59, | |
| "learning_rate": 8.309178743961353e-06, | |
| "loss": 0.9713, | |
| "step": 14340 | |
| }, | |
| { | |
| "epoch": 4.59, | |
| "learning_rate": 8.244766505636072e-06, | |
| "loss": 0.9441, | |
| "step": 14350 | |
| }, | |
| { | |
| "epoch": 4.59, | |
| "learning_rate": 8.18035426731079e-06, | |
| "loss": 0.916, | |
| "step": 14360 | |
| }, | |
| { | |
| "epoch": 4.6, | |
| "learning_rate": 8.115942028985508e-06, | |
| "loss": 0.9249, | |
| "step": 14370 | |
| }, | |
| { | |
| "epoch": 4.6, | |
| "learning_rate": 8.051529790660225e-06, | |
| "loss": 0.924, | |
| "step": 14380 | |
| }, | |
| { | |
| "epoch": 4.6, | |
| "learning_rate": 7.987117552334943e-06, | |
| "loss": 0.9267, | |
| "step": 14390 | |
| }, | |
| { | |
| "epoch": 4.61, | |
| "learning_rate": 7.922705314009662e-06, | |
| "loss": 0.9236, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 4.61, | |
| "eval_loss": 1.026214599609375, | |
| "eval_runtime": 185.9209, | |
| "eval_samples_per_second": 10.757, | |
| "eval_steps_per_second": 1.345, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 4.61, | |
| "learning_rate": 7.85829307568438e-06, | |
| "loss": 0.9255, | |
| "step": 14410 | |
| }, | |
| { | |
| "epoch": 4.61, | |
| "learning_rate": 7.793880837359098e-06, | |
| "loss": 0.9468, | |
| "step": 14420 | |
| }, | |
| { | |
| "epoch": 4.62, | |
| "learning_rate": 7.729468599033817e-06, | |
| "loss": 0.9403, | |
| "step": 14430 | |
| }, | |
| { | |
| "epoch": 4.62, | |
| "learning_rate": 7.665056360708535e-06, | |
| "loss": 0.9465, | |
| "step": 14440 | |
| }, | |
| { | |
| "epoch": 4.62, | |
| "learning_rate": 7.600644122383253e-06, | |
| "loss": 0.8782, | |
| "step": 14450 | |
| }, | |
| { | |
| "epoch": 4.63, | |
| "learning_rate": 7.536231884057972e-06, | |
| "loss": 0.9377, | |
| "step": 14460 | |
| }, | |
| { | |
| "epoch": 4.63, | |
| "learning_rate": 7.471819645732689e-06, | |
| "loss": 0.941, | |
| "step": 14470 | |
| }, | |
| { | |
| "epoch": 4.63, | |
| "learning_rate": 7.4074074074074075e-06, | |
| "loss": 0.926, | |
| "step": 14480 | |
| }, | |
| { | |
| "epoch": 4.64, | |
| "learning_rate": 7.342995169082126e-06, | |
| "loss": 0.9418, | |
| "step": 14490 | |
| }, | |
| { | |
| "epoch": 4.64, | |
| "learning_rate": 7.278582930756844e-06, | |
| "loss": 0.9115, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 4.64, | |
| "learning_rate": 7.214170692431563e-06, | |
| "loss": 0.9632, | |
| "step": 14510 | |
| }, | |
| { | |
| "epoch": 4.65, | |
| "learning_rate": 7.1497584541062814e-06, | |
| "loss": 0.9224, | |
| "step": 14520 | |
| }, | |
| { | |
| "epoch": 4.65, | |
| "learning_rate": 7.0853462157809985e-06, | |
| "loss": 0.9383, | |
| "step": 14530 | |
| }, | |
| { | |
| "epoch": 4.65, | |
| "learning_rate": 7.020933977455716e-06, | |
| "loss": 0.9366, | |
| "step": 14540 | |
| }, | |
| { | |
| "epoch": 4.66, | |
| "learning_rate": 6.956521739130435e-06, | |
| "loss": 0.933, | |
| "step": 14550 | |
| }, | |
| { | |
| "epoch": 4.66, | |
| "learning_rate": 6.892109500805154e-06, | |
| "loss": 0.887, | |
| "step": 14560 | |
| }, | |
| { | |
| "epoch": 4.66, | |
| "learning_rate": 6.8276972624798724e-06, | |
| "loss": 0.9109, | |
| "step": 14570 | |
| }, | |
| { | |
| "epoch": 4.67, | |
| "learning_rate": 6.7632850241545894e-06, | |
| "loss": 0.8918, | |
| "step": 14580 | |
| }, | |
| { | |
| "epoch": 4.67, | |
| "learning_rate": 6.698872785829307e-06, | |
| "loss": 0.9071, | |
| "step": 14590 | |
| }, | |
| { | |
| "epoch": 4.67, | |
| "learning_rate": 6.634460547504026e-06, | |
| "loss": 0.9541, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 4.67, | |
| "eval_loss": 1.0271246433258057, | |
| "eval_runtime": 185.9235, | |
| "eval_samples_per_second": 10.757, | |
| "eval_steps_per_second": 1.345, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 4.67, | |
| "learning_rate": 6.570048309178745e-06, | |
| "loss": 0.9248, | |
| "step": 14610 | |
| }, | |
| { | |
| "epoch": 4.68, | |
| "learning_rate": 6.505636070853463e-06, | |
| "loss": 0.8734, | |
| "step": 14620 | |
| }, | |
| { | |
| "epoch": 4.68, | |
| "learning_rate": 6.44122383252818e-06, | |
| "loss": 0.9185, | |
| "step": 14630 | |
| }, | |
| { | |
| "epoch": 4.68, | |
| "learning_rate": 6.376811594202898e-06, | |
| "loss": 0.8952, | |
| "step": 14640 | |
| }, | |
| { | |
| "epoch": 4.69, | |
| "learning_rate": 6.312399355877617e-06, | |
| "loss": 0.9259, | |
| "step": 14650 | |
| }, | |
| { | |
| "epoch": 4.69, | |
| "learning_rate": 6.247987117552336e-06, | |
| "loss": 0.9393, | |
| "step": 14660 | |
| }, | |
| { | |
| "epoch": 4.69, | |
| "learning_rate": 6.1835748792270535e-06, | |
| "loss": 0.9438, | |
| "step": 14670 | |
| }, | |
| { | |
| "epoch": 4.7, | |
| "learning_rate": 6.119162640901771e-06, | |
| "loss": 0.9362, | |
| "step": 14680 | |
| }, | |
| { | |
| "epoch": 4.7, | |
| "learning_rate": 6.054750402576489e-06, | |
| "loss": 0.9455, | |
| "step": 14690 | |
| }, | |
| { | |
| "epoch": 4.7, | |
| "learning_rate": 5.990338164251208e-06, | |
| "loss": 0.8851, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 4.71, | |
| "learning_rate": 5.925925925925927e-06, | |
| "loss": 0.9641, | |
| "step": 14710 | |
| }, | |
| { | |
| "epoch": 4.71, | |
| "learning_rate": 5.8615136876006445e-06, | |
| "loss": 0.933, | |
| "step": 14720 | |
| }, | |
| { | |
| "epoch": 4.71, | |
| "learning_rate": 5.797101449275362e-06, | |
| "loss": 0.9546, | |
| "step": 14730 | |
| }, | |
| { | |
| "epoch": 4.72, | |
| "learning_rate": 5.732689210950081e-06, | |
| "loss": 0.923, | |
| "step": 14740 | |
| }, | |
| { | |
| "epoch": 4.72, | |
| "learning_rate": 5.668276972624799e-06, | |
| "loss": 0.9494, | |
| "step": 14750 | |
| }, | |
| { | |
| "epoch": 4.72, | |
| "learning_rate": 5.603864734299518e-06, | |
| "loss": 0.9442, | |
| "step": 14760 | |
| }, | |
| { | |
| "epoch": 4.73, | |
| "learning_rate": 5.5394524959742355e-06, | |
| "loss": 0.9571, | |
| "step": 14770 | |
| }, | |
| { | |
| "epoch": 4.73, | |
| "learning_rate": 5.475040257648953e-06, | |
| "loss": 0.9171, | |
| "step": 14780 | |
| }, | |
| { | |
| "epoch": 4.73, | |
| "learning_rate": 5.410628019323672e-06, | |
| "loss": 0.9183, | |
| "step": 14790 | |
| }, | |
| { | |
| "epoch": 4.74, | |
| "learning_rate": 5.34621578099839e-06, | |
| "loss": 0.9395, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 4.74, | |
| "eval_loss": 1.026661992073059, | |
| "eval_runtime": 185.6848, | |
| "eval_samples_per_second": 10.771, | |
| "eval_steps_per_second": 1.346, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 4.74, | |
| "learning_rate": 5.281803542673109e-06, | |
| "loss": 0.9001, | |
| "step": 14810 | |
| }, | |
| { | |
| "epoch": 4.74, | |
| "learning_rate": 5.217391304347826e-06, | |
| "loss": 0.9535, | |
| "step": 14820 | |
| }, | |
| { | |
| "epoch": 4.75, | |
| "learning_rate": 5.152979066022544e-06, | |
| "loss": 0.9204, | |
| "step": 14830 | |
| }, | |
| { | |
| "epoch": 4.75, | |
| "learning_rate": 5.088566827697263e-06, | |
| "loss": 0.9311, | |
| "step": 14840 | |
| }, | |
| { | |
| "epoch": 4.75, | |
| "learning_rate": 5.024154589371981e-06, | |
| "loss": 0.9117, | |
| "step": 14850 | |
| }, | |
| { | |
| "epoch": 4.75, | |
| "learning_rate": 4.9597423510467e-06, | |
| "loss": 0.9534, | |
| "step": 14860 | |
| }, | |
| { | |
| "epoch": 4.76, | |
| "learning_rate": 4.895330112721417e-06, | |
| "loss": 0.9449, | |
| "step": 14870 | |
| }, | |
| { | |
| "epoch": 4.76, | |
| "learning_rate": 4.830917874396135e-06, | |
| "loss": 0.9474, | |
| "step": 14880 | |
| }, | |
| { | |
| "epoch": 4.76, | |
| "learning_rate": 4.766505636070854e-06, | |
| "loss": 0.9429, | |
| "step": 14890 | |
| }, | |
| { | |
| "epoch": 4.77, | |
| "learning_rate": 4.702093397745572e-06, | |
| "loss": 1.0071, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 4.77, | |
| "learning_rate": 4.63768115942029e-06, | |
| "loss": 0.9075, | |
| "step": 14910 | |
| }, | |
| { | |
| "epoch": 4.77, | |
| "learning_rate": 4.5732689210950084e-06, | |
| "loss": 0.9184, | |
| "step": 14920 | |
| }, | |
| { | |
| "epoch": 4.78, | |
| "learning_rate": 4.508856682769726e-06, | |
| "loss": 0.928, | |
| "step": 14930 | |
| }, | |
| { | |
| "epoch": 4.78, | |
| "learning_rate": 4.444444444444445e-06, | |
| "loss": 0.9877, | |
| "step": 14940 | |
| }, | |
| { | |
| "epoch": 4.78, | |
| "learning_rate": 4.380032206119163e-06, | |
| "loss": 0.9049, | |
| "step": 14950 | |
| }, | |
| { | |
| "epoch": 4.79, | |
| "learning_rate": 4.315619967793881e-06, | |
| "loss": 0.9278, | |
| "step": 14960 | |
| }, | |
| { | |
| "epoch": 4.79, | |
| "learning_rate": 4.251207729468599e-06, | |
| "loss": 0.9215, | |
| "step": 14970 | |
| }, | |
| { | |
| "epoch": 4.79, | |
| "learning_rate": 4.186795491143317e-06, | |
| "loss": 0.9413, | |
| "step": 14980 | |
| }, | |
| { | |
| "epoch": 4.8, | |
| "learning_rate": 4.122383252818036e-06, | |
| "loss": 0.9088, | |
| "step": 14990 | |
| }, | |
| { | |
| "epoch": 4.8, | |
| "learning_rate": 4.057971014492754e-06, | |
| "loss": 0.9334, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 4.8, | |
| "eval_loss": 1.0265828371047974, | |
| "eval_runtime": 185.8809, | |
| "eval_samples_per_second": 10.76, | |
| "eval_steps_per_second": 1.345, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 4.8, | |
| "learning_rate": 3.993558776167472e-06, | |
| "loss": 0.9048, | |
| "step": 15010 | |
| }, | |
| { | |
| "epoch": 4.81, | |
| "learning_rate": 3.92914653784219e-06, | |
| "loss": 0.9295, | |
| "step": 15020 | |
| }, | |
| { | |
| "epoch": 4.81, | |
| "learning_rate": 3.864734299516908e-06, | |
| "loss": 0.9104, | |
| "step": 15030 | |
| }, | |
| { | |
| "epoch": 4.81, | |
| "learning_rate": 3.8003220611916265e-06, | |
| "loss": 0.9633, | |
| "step": 15040 | |
| }, | |
| { | |
| "epoch": 4.82, | |
| "learning_rate": 3.7359098228663444e-06, | |
| "loss": 0.9382, | |
| "step": 15050 | |
| }, | |
| { | |
| "epoch": 4.82, | |
| "learning_rate": 3.671497584541063e-06, | |
| "loss": 0.9009, | |
| "step": 15060 | |
| }, | |
| { | |
| "epoch": 4.82, | |
| "learning_rate": 3.6070853462157814e-06, | |
| "loss": 0.9157, | |
| "step": 15070 | |
| }, | |
| { | |
| "epoch": 4.83, | |
| "learning_rate": 3.5426731078904992e-06, | |
| "loss": 0.9522, | |
| "step": 15080 | |
| }, | |
| { | |
| "epoch": 4.83, | |
| "learning_rate": 3.4782608695652175e-06, | |
| "loss": 0.8791, | |
| "step": 15090 | |
| }, | |
| { | |
| "epoch": 4.83, | |
| "learning_rate": 3.4138486312399362e-06, | |
| "loss": 0.9306, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 4.83, | |
| "learning_rate": 3.3494363929146536e-06, | |
| "loss": 0.9234, | |
| "step": 15110 | |
| }, | |
| { | |
| "epoch": 4.84, | |
| "learning_rate": 3.2850241545893724e-06, | |
| "loss": 0.9262, | |
| "step": 15120 | |
| }, | |
| { | |
| "epoch": 4.84, | |
| "learning_rate": 3.22061191626409e-06, | |
| "loss": 0.878, | |
| "step": 15130 | |
| }, | |
| { | |
| "epoch": 4.84, | |
| "learning_rate": 3.1561996779388085e-06, | |
| "loss": 0.8833, | |
| "step": 15140 | |
| }, | |
| { | |
| "epoch": 4.85, | |
| "learning_rate": 3.0917874396135268e-06, | |
| "loss": 0.9158, | |
| "step": 15150 | |
| }, | |
| { | |
| "epoch": 4.85, | |
| "learning_rate": 3.0273752012882446e-06, | |
| "loss": 0.9675, | |
| "step": 15160 | |
| }, | |
| { | |
| "epoch": 4.85, | |
| "learning_rate": 2.9629629629629633e-06, | |
| "loss": 0.9967, | |
| "step": 15170 | |
| }, | |
| { | |
| "epoch": 4.86, | |
| "learning_rate": 2.898550724637681e-06, | |
| "loss": 0.918, | |
| "step": 15180 | |
| }, | |
| { | |
| "epoch": 4.86, | |
| "learning_rate": 2.8341384863123995e-06, | |
| "loss": 0.8854, | |
| "step": 15190 | |
| }, | |
| { | |
| "epoch": 4.86, | |
| "learning_rate": 2.7697262479871177e-06, | |
| "loss": 0.9589, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 4.86, | |
| "eval_loss": 1.0268317461013794, | |
| "eval_runtime": 185.8882, | |
| "eval_samples_per_second": 10.759, | |
| "eval_steps_per_second": 1.345, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 4.87, | |
| "learning_rate": 2.705314009661836e-06, | |
| "loss": 0.9624, | |
| "step": 15210 | |
| }, | |
| { | |
| "epoch": 4.87, | |
| "learning_rate": 2.6409017713365543e-06, | |
| "loss": 0.9221, | |
| "step": 15220 | |
| }, | |
| { | |
| "epoch": 4.87, | |
| "learning_rate": 2.576489533011272e-06, | |
| "loss": 0.9157, | |
| "step": 15230 | |
| }, | |
| { | |
| "epoch": 4.88, | |
| "learning_rate": 2.5120772946859904e-06, | |
| "loss": 0.9513, | |
| "step": 15240 | |
| }, | |
| { | |
| "epoch": 4.88, | |
| "learning_rate": 2.4476650563607083e-06, | |
| "loss": 0.8803, | |
| "step": 15250 | |
| }, | |
| { | |
| "epoch": 4.88, | |
| "learning_rate": 2.383252818035427e-06, | |
| "loss": 0.8862, | |
| "step": 15260 | |
| }, | |
| { | |
| "epoch": 4.89, | |
| "learning_rate": 2.318840579710145e-06, | |
| "loss": 0.8634, | |
| "step": 15270 | |
| }, | |
| { | |
| "epoch": 4.89, | |
| "learning_rate": 2.254428341384863e-06, | |
| "loss": 0.9381, | |
| "step": 15280 | |
| }, | |
| { | |
| "epoch": 4.89, | |
| "learning_rate": 2.1900161030595814e-06, | |
| "loss": 0.9557, | |
| "step": 15290 | |
| }, | |
| { | |
| "epoch": 4.9, | |
| "learning_rate": 2.1256038647342997e-06, | |
| "loss": 0.8925, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 4.9, | |
| "learning_rate": 2.061191626409018e-06, | |
| "loss": 0.9255, | |
| "step": 15310 | |
| }, | |
| { | |
| "epoch": 4.9, | |
| "learning_rate": 1.996779388083736e-06, | |
| "loss": 0.9385, | |
| "step": 15320 | |
| }, | |
| { | |
| "epoch": 4.91, | |
| "learning_rate": 1.932367149758454e-06, | |
| "loss": 0.9084, | |
| "step": 15330 | |
| }, | |
| { | |
| "epoch": 4.91, | |
| "learning_rate": 1.8679549114331722e-06, | |
| "loss": 0.9499, | |
| "step": 15340 | |
| }, | |
| { | |
| "epoch": 4.91, | |
| "learning_rate": 1.8035426731078907e-06, | |
| "loss": 0.8961, | |
| "step": 15350 | |
| }, | |
| { | |
| "epoch": 4.91, | |
| "learning_rate": 1.7391304347826088e-06, | |
| "loss": 0.8908, | |
| "step": 15360 | |
| }, | |
| { | |
| "epoch": 4.92, | |
| "learning_rate": 1.6747181964573268e-06, | |
| "loss": 0.9399, | |
| "step": 15370 | |
| }, | |
| { | |
| "epoch": 4.92, | |
| "learning_rate": 1.610305958132045e-06, | |
| "loss": 0.9366, | |
| "step": 15380 | |
| }, | |
| { | |
| "epoch": 4.92, | |
| "learning_rate": 1.5458937198067634e-06, | |
| "loss": 0.9091, | |
| "step": 15390 | |
| }, | |
| { | |
| "epoch": 4.93, | |
| "learning_rate": 1.4814814814814817e-06, | |
| "loss": 0.8723, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 4.93, | |
| "eval_loss": 1.026999831199646, | |
| "eval_runtime": 185.7802, | |
| "eval_samples_per_second": 10.765, | |
| "eval_steps_per_second": 1.346, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 4.93, | |
| "learning_rate": 1.4170692431561997e-06, | |
| "loss": 0.9343, | |
| "step": 15410 | |
| }, | |
| { | |
| "epoch": 4.93, | |
| "learning_rate": 1.352657004830918e-06, | |
| "loss": 0.9296, | |
| "step": 15420 | |
| }, | |
| { | |
| "epoch": 4.94, | |
| "learning_rate": 1.288244766505636e-06, | |
| "loss": 0.9574, | |
| "step": 15430 | |
| }, | |
| { | |
| "epoch": 4.94, | |
| "learning_rate": 1.2238325281803542e-06, | |
| "loss": 0.9422, | |
| "step": 15440 | |
| }, | |
| { | |
| "epoch": 4.94, | |
| "learning_rate": 1.1594202898550724e-06, | |
| "loss": 0.9331, | |
| "step": 15450 | |
| }, | |
| { | |
| "epoch": 4.95, | |
| "learning_rate": 1.0950080515297907e-06, | |
| "loss": 0.8342, | |
| "step": 15460 | |
| }, | |
| { | |
| "epoch": 4.95, | |
| "learning_rate": 1.030595813204509e-06, | |
| "loss": 0.9069, | |
| "step": 15470 | |
| }, | |
| { | |
| "epoch": 4.95, | |
| "learning_rate": 9.66183574879227e-07, | |
| "loss": 0.931, | |
| "step": 15480 | |
| }, | |
| { | |
| "epoch": 4.96, | |
| "learning_rate": 9.017713365539453e-07, | |
| "loss": 0.8903, | |
| "step": 15490 | |
| }, | |
| { | |
| "epoch": 4.96, | |
| "learning_rate": 8.373590982286634e-07, | |
| "loss": 0.902, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 4.96, | |
| "learning_rate": 7.729468599033817e-07, | |
| "loss": 0.9406, | |
| "step": 15510 | |
| }, | |
| { | |
| "epoch": 4.97, | |
| "learning_rate": 7.085346215780999e-07, | |
| "loss": 0.9086, | |
| "step": 15520 | |
| }, | |
| { | |
| "epoch": 4.97, | |
| "learning_rate": 6.44122383252818e-07, | |
| "loss": 0.9097, | |
| "step": 15530 | |
| }, | |
| { | |
| "epoch": 4.97, | |
| "learning_rate": 5.797101449275362e-07, | |
| "loss": 0.9068, | |
| "step": 15540 | |
| }, | |
| { | |
| "epoch": 4.98, | |
| "learning_rate": 5.152979066022545e-07, | |
| "loss": 0.9534, | |
| "step": 15550 | |
| }, | |
| { | |
| "epoch": 4.98, | |
| "learning_rate": 4.5088566827697267e-07, | |
| "loss": 0.9242, | |
| "step": 15560 | |
| }, | |
| { | |
| "epoch": 4.98, | |
| "learning_rate": 3.8647342995169085e-07, | |
| "loss": 0.9215, | |
| "step": 15570 | |
| }, | |
| { | |
| "epoch": 4.99, | |
| "learning_rate": 3.22061191626409e-07, | |
| "loss": 0.9169, | |
| "step": 15580 | |
| }, | |
| { | |
| "epoch": 4.99, | |
| "learning_rate": 2.5764895330112725e-07, | |
| "loss": 0.9331, | |
| "step": 15590 | |
| }, | |
| { | |
| "epoch": 4.99, | |
| "learning_rate": 1.9323671497584542e-07, | |
| "loss": 0.8916, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 4.99, | |
| "eval_loss": 1.0266457796096802, | |
| "eval_runtime": 186.0125, | |
| "eval_samples_per_second": 10.752, | |
| "eval_steps_per_second": 1.344, | |
| "step": 15600 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 15625, | |
| "num_train_epochs": 5, | |
| "save_steps": 200, | |
| "total_flos": 3.348372562869289e+18, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |