| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 32.78688524590164, | |
| "global_step": 2000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 2.4999999999999998e-05, | |
| "loss": 4.5271, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.2525749891599525e-05, | |
| "loss": 3.9027, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 3.6928031367991554e-05, | |
| "loss": 3.4367, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.005149978319905e-05, | |
| "loss": 3.3343, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.247425010840046e-05, | |
| "loss": 3.2186, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 4.445378125959108e-05, | |
| "loss": 3.1259, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 4.612745100035642e-05, | |
| "loss": 2.9406, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 4.757724967479858e-05, | |
| "loss": 2.8567, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 4.885606273598312e-05, | |
| "loss": 2.8893, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 4.9999999999999996e-05, | |
| "loss": 2.9067, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 5.1034817128955624e-05, | |
| "loss": 2.813, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 5.197953115119061e-05, | |
| "loss": 2.8364, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 5.2848583807670914e-05, | |
| "loss": 2.7439, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 5.365320089195593e-05, | |
| "loss": 2.7048, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 5.4402281476392025e-05, | |
| "loss": 2.6631, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 5.5102999566398106e-05, | |
| "loss": 2.6804, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 5.5761223034456847e-05, | |
| "loss": 2.6094, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 5.6381812627582644e-05, | |
| "loss": 2.6055, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 3.11, | |
| "learning_rate": 5.696884002382071e-05, | |
| "loss": 2.4945, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 3.28, | |
| "learning_rate": 5.752574989159952e-05, | |
| "loss": 2.5523, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 5.805548236834797e-05, | |
| "loss": 2.5406, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 3.61, | |
| "learning_rate": 5.856056702055516e-05, | |
| "loss": 2.5444, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 3.77, | |
| "learning_rate": 5.9043195900439815e-05, | |
| "loss": 2.568, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 3.93, | |
| "learning_rate": 5.950528104279014e-05, | |
| "loss": 2.5305, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 4.1, | |
| "learning_rate": 5.9948500216800926e-05, | |
| "loss": 2.4653, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 4.26, | |
| "learning_rate": 6.037433369927045e-05, | |
| "loss": 2.4644, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 4.43, | |
| "learning_rate": 6.078409410397467e-05, | |
| "loss": 2.4302, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 4.59, | |
| "learning_rate": 6.117895078355547e-05, | |
| "loss": 2.4025, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 4.75, | |
| "learning_rate": 6.15599499474739e-05, | |
| "loss": 2.4185, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 4.92, | |
| "learning_rate": 6.192803136799156e-05, | |
| "loss": 2.4651, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 5.08, | |
| "learning_rate": 6.22840423458568e-05, | |
| "loss": 2.4095, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 5.25, | |
| "learning_rate": 6.262874945799764e-05, | |
| "loss": 2.3613, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 5.41, | |
| "learning_rate": 6.296284849694718e-05, | |
| "loss": 2.381, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 5.57, | |
| "learning_rate": 6.328697292605637e-05, | |
| "loss": 2.3387, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 5.74, | |
| "learning_rate": 6.360170110875688e-05, | |
| "loss": 2.3331, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 5.9, | |
| "learning_rate": 6.390756251918218e-05, | |
| "loss": 2.3831, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 6.07, | |
| "learning_rate": 6.420504310167487e-05, | |
| "loss": 2.2977, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 6.23, | |
| "learning_rate": 6.449458991542025e-05, | |
| "loss": 2.2816, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 6.39, | |
| "learning_rate": 6.477661517566246e-05, | |
| "loss": 2.2826, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 6.56, | |
| "learning_rate": 6.505149978319905e-05, | |
| "loss": 2.3021, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 6.72, | |
| "learning_rate": 6.531959641799339e-05, | |
| "loss": 2.3166, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 6.89, | |
| "learning_rate": 6.55812322599475e-05, | |
| "loss": 2.3421, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 7.05, | |
| "learning_rate": 6.583671138948966e-05, | |
| "loss": 2.3094, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 7.21, | |
| "learning_rate": 6.608631691215467e-05, | |
| "loss": 2.2268, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 7.38, | |
| "learning_rate": 6.633031284438359e-05, | |
| "loss": 2.2722, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 7.54, | |
| "learning_rate": 6.656894579203935e-05, | |
| "loss": 2.1933, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 7.7, | |
| "learning_rate": 6.680244644839293e-05, | |
| "loss": 2.1991, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 7.87, | |
| "learning_rate": 6.703103093438967e-05, | |
| "loss": 2.2637, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 8.03, | |
| "learning_rate": 6.725490200071283e-05, | |
| "loss": 2.2277, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 8.2, | |
| "learning_rate": 6.747425010840046e-05, | |
| "loss": 2.1819, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 8.2, | |
| "eval_loss": 2.786494255065918, | |
| "eval_runtime": 13.6759, | |
| "eval_samples_per_second": 47.602, | |
| "eval_steps_per_second": 0.585, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 8.36, | |
| "learning_rate": 6.76892544024484e-05, | |
| "loss": 2.1427, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 8.52, | |
| "learning_rate": 6.790008359086997e-05, | |
| "loss": 2.173, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 8.69, | |
| "learning_rate": 6.810689674001973e-05, | |
| "loss": 2.1895, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 8.85, | |
| "learning_rate": 6.830984399557421e-05, | |
| "loss": 2.2101, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 9.02, | |
| "learning_rate": 6.850906723735608e-05, | |
| "loss": 2.1926, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 9.18, | |
| "learning_rate": 6.870470067515499e-05, | |
| "loss": 2.0861, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 9.34, | |
| "learning_rate": 6.889687139181228e-05, | |
| "loss": 2.1092, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 9.51, | |
| "learning_rate": 6.908569983907343e-05, | |
| "loss": 2.129, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 9.67, | |
| "learning_rate": 6.92713002910536e-05, | |
| "loss": 2.113, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 9.84, | |
| "learning_rate": 6.945378125959108e-05, | |
| "loss": 2.1234, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "learning_rate": 6.963324587526918e-05, | |
| "loss": 2.1028, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 10.16, | |
| "learning_rate": 6.980979223745634e-05, | |
| "loss": 2.0476, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 10.33, | |
| "learning_rate": 6.998351373633953e-05, | |
| "loss": 2.0879, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 10.49, | |
| "learning_rate": 7.015449934959717e-05, | |
| "loss": 2.0547, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 10.66, | |
| "learning_rate": 7.032283391607138e-05, | |
| "loss": 2.0791, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 10.82, | |
| "learning_rate": 7.048859838854671e-05, | |
| "loss": 2.1454, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 10.98, | |
| "learning_rate": 7.065187006752065e-05, | |
| "loss": 2.0957, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 11.15, | |
| "learning_rate": 7.08127228176559e-05, | |
| "loss": 2.0561, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 11.31, | |
| "learning_rate": 7.097122726843138e-05, | |
| "loss": 2.0563, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 11.48, | |
| "learning_rate": 7.112745100035642e-05, | |
| "loss": 2.027, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 11.64, | |
| "learning_rate": 7.128145871797688e-05, | |
| "loss": 2.0495, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 11.8, | |
| "learning_rate": 7.143331241078171e-05, | |
| "loss": 2.019, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 11.97, | |
| "learning_rate": 7.158307150301139e-05, | |
| "loss": 2.0242, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 12.13, | |
| "learning_rate": 7.17307929932744e-05, | |
| "loss": 1.9962, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 12.3, | |
| "learning_rate": 7.187653158479249e-05, | |
| "loss": 1.9971, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 12.46, | |
| "learning_rate": 7.202033980701978e-05, | |
| "loss": 2.0236, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 12.62, | |
| "learning_rate": 7.216226812931204e-05, | |
| "loss": 1.9923, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 12.79, | |
| "learning_rate": 7.2302365067262e-05, | |
| "loss": 2.0244, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 12.95, | |
| "learning_rate": 7.244067728226103e-05, | |
| "loss": 1.9846, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 13.11, | |
| "learning_rate": 7.257724967479857e-05, | |
| "loss": 1.9811, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 13.28, | |
| "learning_rate": 7.271212547196624e-05, | |
| "loss": 1.9709, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 13.44, | |
| "learning_rate": 7.284534630959291e-05, | |
| "loss": 1.9652, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 13.61, | |
| "learning_rate": 7.297695230940184e-05, | |
| "loss": 1.9605, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 13.77, | |
| "learning_rate": 7.310698215154704e-05, | |
| "loss": 1.9692, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 13.93, | |
| "learning_rate": 7.323547314285732e-05, | |
| "loss": 1.9945, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 14.1, | |
| "learning_rate": 7.336246128108918e-05, | |
| "loss": 1.9222, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 14.26, | |
| "learning_rate": 7.348798131546546e-05, | |
| "loss": 1.9283, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 14.43, | |
| "learning_rate": 7.36120668037542e-05, | |
| "loss": 1.9376, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 14.59, | |
| "learning_rate": 7.37347501661228e-05, | |
| "loss": 1.9247, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 14.75, | |
| "learning_rate": 7.385606273598311e-05, | |
| "loss": 1.9218, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 14.92, | |
| "learning_rate": 7.397603480802732e-05, | |
| "loss": 1.9492, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 15.08, | |
| "learning_rate": 7.409469568363888e-05, | |
| "loss": 1.9235, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 15.25, | |
| "learning_rate": 7.421207371384837e-05, | |
| "loss": 1.8671, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 15.41, | |
| "learning_rate": 7.432819633999247e-05, | |
| "loss": 1.909, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 15.57, | |
| "learning_rate": 7.444309013222118e-05, | |
| "loss": 1.8568, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 15.74, | |
| "learning_rate": 7.45567808259892e-05, | |
| "loss": 1.9199, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 15.9, | |
| "learning_rate": 7.46692933566561e-05, | |
| "loss": 1.9247, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 16.07, | |
| "learning_rate": 7.478065189231236e-05, | |
| "loss": 1.895, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 16.23, | |
| "learning_rate": 7.489087986493874e-05, | |
| "loss": 1.8821, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 16.39, | |
| "learning_rate": 7.5e-05, | |
| "loss": 1.8423, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 16.39, | |
| "eval_loss": 2.925347328186035, | |
| "eval_runtime": 13.9926, | |
| "eval_samples_per_second": 46.524, | |
| "eval_steps_per_second": 0.572, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 16.56, | |
| "learning_rate": 7.510803434456605e-05, | |
| "loss": 1.8519, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 16.72, | |
| "learning_rate": 7.521500429404794e-05, | |
| "loss": 1.8578, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 16.89, | |
| "learning_rate": 7.532093061762931e-05, | |
| "loss": 1.8676, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 17.05, | |
| "learning_rate": 7.54258334824695e-05, | |
| "loss": 1.8492, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 17.21, | |
| "learning_rate": 7.552973247674843e-05, | |
| "loss": 1.8542, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 17.38, | |
| "learning_rate": 7.563264663161926e-05, | |
| "loss": 1.8312, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 17.54, | |
| "learning_rate": 7.573459444213023e-05, | |
| "loss": 1.8554, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 17.7, | |
| "learning_rate": 7.583559388717374e-05, | |
| "loss": 1.8484, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 17.87, | |
| "learning_rate": 7.593566244851558e-05, | |
| "loss": 1.8485, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 18.03, | |
| "learning_rate": 7.603481712895562e-05, | |
| "loss": 1.8505, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 18.2, | |
| "learning_rate": 7.613307446966643e-05, | |
| "loss": 1.8163, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 18.36, | |
| "learning_rate": 7.623045056675453e-05, | |
| "loss": 1.8382, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 18.52, | |
| "learning_rate": 7.632696108708549e-05, | |
| "loss": 1.8251, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 18.69, | |
| "learning_rate": 7.642262128341181e-05, | |
| "loss": 1.8252, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 18.85, | |
| "learning_rate": 7.651744600884029e-05, | |
| "loss": 1.849, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 19.02, | |
| "learning_rate": 7.661144973067295e-05, | |
| "loss": 1.8202, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 19.18, | |
| "learning_rate": 7.670464654365404e-05, | |
| "loss": 1.8013, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 19.34, | |
| "learning_rate": 7.679705018265312e-05, | |
| "loss": 1.8149, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 19.51, | |
| "learning_rate": 7.688867403481326e-05, | |
| "loss": 1.7919, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 19.67, | |
| "learning_rate": 7.697953115119061e-05, | |
| "loss": 1.801, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 19.84, | |
| "learning_rate": 7.706963425791124e-05, | |
| "loss": 1.8286, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "learning_rate": 7.71589957668687e-05, | |
| "loss": 1.7945, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 20.16, | |
| "learning_rate": 7.724762778598493e-05, | |
| "loss": 1.7619, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 20.33, | |
| "learning_rate": 7.733554212905587e-05, | |
| "loss": 1.7693, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 20.49, | |
| "learning_rate": 7.74227503252014e-05, | |
| "loss": 1.7689, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 20.66, | |
| "learning_rate": 7.750926362793907e-05, | |
| "loss": 1.77, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 20.82, | |
| "learning_rate": 7.759509302389892e-05, | |
| "loss": 1.7765, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 20.98, | |
| "learning_rate": 7.768024924119671e-05, | |
| "loss": 1.7791, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 21.15, | |
| "learning_rate": 7.776474275748121e-05, | |
| "loss": 1.7514, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 21.31, | |
| "learning_rate": 7.784858380767091e-05, | |
| "loss": 1.7564, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 21.48, | |
| "learning_rate": 7.793178239139409e-05, | |
| "loss": 1.7541, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 21.64, | |
| "learning_rate": 7.801434828014625e-05, | |
| "loss": 1.7519, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 21.8, | |
| "learning_rate": 7.809629102417713e-05, | |
| "loss": 1.7862, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 21.97, | |
| "learning_rate": 7.817761995912018e-05, | |
| "loss": 1.7724, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 22.13, | |
| "learning_rate": 7.825834421237515e-05, | |
| "loss": 1.7565, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 22.3, | |
| "learning_rate": 7.833847270925543e-05, | |
| "loss": 1.7346, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 22.46, | |
| "learning_rate": 7.841801417891016e-05, | |
| "loss": 1.7238, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 22.62, | |
| "learning_rate": 7.84969771600309e-05, | |
| "loss": 1.738, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 22.79, | |
| "learning_rate": 7.857537000635237e-05, | |
| "loss": 1.7446, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 22.95, | |
| "learning_rate": 7.865320089195594e-05, | |
| "loss": 1.7395, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 23.11, | |
| "learning_rate": 7.87304778163845e-05, | |
| "loss": 1.7533, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 23.28, | |
| "learning_rate": 7.880720860957641e-05, | |
| "loss": 1.7101, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 23.44, | |
| "learning_rate": 7.888340093662653e-05, | |
| "loss": 1.7145, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 23.61, | |
| "learning_rate": 7.895906230238123e-05, | |
| "loss": 1.7496, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 23.77, | |
| "learning_rate": 7.903420005587436e-05, | |
| "loss": 1.7416, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 23.93, | |
| "learning_rate": 7.910882139461093e-05, | |
| "loss": 1.7315, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 24.1, | |
| "learning_rate": 7.918293336870439e-05, | |
| "loss": 1.7224, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 24.26, | |
| "learning_rate": 7.925654288487392e-05, | |
| "loss": 1.716, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 24.43, | |
| "learning_rate": 7.932965671030685e-05, | |
| "loss": 1.704, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 24.59, | |
| "learning_rate": 7.940228147639202e-05, | |
| "loss": 1.6873, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 24.59, | |
| "eval_loss": 3.106438159942627, | |
| "eval_runtime": 13.9459, | |
| "eval_samples_per_second": 46.681, | |
| "eval_steps_per_second": 0.574, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 24.75, | |
| "learning_rate": 7.947442368232923e-05, | |
| "loss": 1.7098, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 24.92, | |
| "learning_rate": 7.954608969861931e-05, | |
| "loss": 1.7058, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 25.08, | |
| "learning_rate": 7.961728577043997e-05, | |
| "loss": 1.7189, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 25.25, | |
| "learning_rate": 7.968801802091157e-05, | |
| "loss": 1.6689, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 25.41, | |
| "learning_rate": 7.975829245425728e-05, | |
| "loss": 1.709, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 25.57, | |
| "learning_rate": 7.982811495886153e-05, | |
| "loss": 1.6881, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 25.74, | |
| "learning_rate": 7.989749131023083e-05, | |
| "loss": 1.7032, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 25.9, | |
| "learning_rate": 7.996642717386056e-05, | |
| "loss": 1.6887, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 26.07, | |
| "learning_rate": 8.003492810801127e-05, | |
| "loss": 1.6961, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 26.23, | |
| "learning_rate": 8.01029995663981e-05, | |
| "loss": 1.6701, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 26.39, | |
| "learning_rate": 8.017064690079624e-05, | |
| "loss": 1.69, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 26.56, | |
| "learning_rate": 8.023787536356576e-05, | |
| "loss": 1.7125, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 26.72, | |
| "learning_rate": 8.030469011009893e-05, | |
| "loss": 1.6606, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 26.89, | |
| "learning_rate": 8.037109620119243e-05, | |
| "loss": 1.6649, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 27.05, | |
| "learning_rate": 8.043709860534764e-05, | |
| "loss": 1.6699, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 27.21, | |
| "learning_rate": 8.050270220100136e-05, | |
| "loss": 1.645, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 27.38, | |
| "learning_rate": 8.056791177868957e-05, | |
| "loss": 1.65, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 27.54, | |
| "learning_rate": 8.063273204314657e-05, | |
| "loss": 1.6552, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 27.7, | |
| "learning_rate": 8.069716761534183e-05, | |
| "loss": 1.6772, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 27.87, | |
| "learning_rate": 8.076122303445684e-05, | |
| "loss": 1.6664, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 28.03, | |
| "learning_rate": 8.082490275980384e-05, | |
| "loss": 1.6539, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 28.2, | |
| "learning_rate": 8.088821117268871e-05, | |
| "loss": 1.6616, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 28.36, | |
| "learning_rate": 8.095115257821987e-05, | |
| "loss": 1.6379, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 28.52, | |
| "learning_rate": 8.1013731207065e-05, | |
| "loss": 1.6536, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 28.69, | |
| "learning_rate": 8.107595121715735e-05, | |
| "loss": 1.6506, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 28.85, | |
| "learning_rate": 8.113781669535373e-05, | |
| "loss": 1.66, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 29.02, | |
| "learning_rate": 8.119933165904515e-05, | |
| "loss": 1.6548, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 29.18, | |
| "learning_rate": 8.126050005772234e-05, | |
| "loss": 1.6408, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 29.34, | |
| "learning_rate": 8.132132577449732e-05, | |
| "loss": 1.6533, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 29.51, | |
| "learning_rate": 8.138181262758264e-05, | |
| "loss": 1.6508, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 29.67, | |
| "learning_rate": 8.144196437172959e-05, | |
| "loss": 1.6302, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 29.84, | |
| "learning_rate": 8.150178469962686e-05, | |
| "loss": 1.6319, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "learning_rate": 8.156127724326073e-05, | |
| "loss": 1.623, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 30.16, | |
| "learning_rate": 8.16204455752384e-05, | |
| "loss": 1.6299, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 30.33, | |
| "learning_rate": 8.167929321007533e-05, | |
| "loss": 1.6187, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 30.49, | |
| "learning_rate": 8.17378236054479e-05, | |
| "loss": 1.6138, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 30.66, | |
| "learning_rate": 8.179604016341247e-05, | |
| "loss": 1.6418, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 30.82, | |
| "learning_rate": 8.1853946231592e-05, | |
| "loss": 1.6433, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 30.98, | |
| "learning_rate": 8.19115451043311e-05, | |
| "loss": 1.6416, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 31.15, | |
| "learning_rate": 8.196884002382071e-05, | |
| "loss": 1.6244, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 31.31, | |
| "learning_rate": 8.202583418119318e-05, | |
| "loss": 1.6141, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 31.48, | |
| "learning_rate": 8.208253071758874e-05, | |
| "loss": 1.6033, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 31.64, | |
| "learning_rate": 8.213893272519434e-05, | |
| "loss": 1.6287, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 31.8, | |
| "learning_rate": 8.219504324825564e-05, | |
| "loss": 1.6268, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 31.97, | |
| "learning_rate": 8.225086528406294e-05, | |
| "loss": 1.6297, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 32.13, | |
| "learning_rate": 8.23064017839119e-05, | |
| "loss": 1.6012, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 32.3, | |
| "learning_rate": 8.236165565403982e-05, | |
| "loss": 1.6203, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 32.46, | |
| "learning_rate": 8.241662975653826e-05, | |
| "loss": 1.6107, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 32.62, | |
| "learning_rate": 8.247132691024267e-05, | |
| "loss": 1.6107, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 32.79, | |
| "learning_rate": 8.252574989159953e-05, | |
| "loss": 1.5886, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 32.79, | |
| "eval_loss": 3.1871249675750732, | |
| "eval_runtime": 13.1185, | |
| "eval_samples_per_second": 49.624, | |
| "eval_steps_per_second": 0.61, | |
| "step": 2000 | |
| } | |
| ], | |
| "max_steps": 100000, | |
| "num_train_epochs": 1640, | |
| "total_flos": 394015948800.0, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |