| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.807554874936192, | |
| "global_step": 5500, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.05, | |
| "eval_gen_len": 11.1391, | |
| "eval_loss": 1.4340903759002686, | |
| "eval_rouge1": 23.6512, | |
| "eval_rouge2": 14.437, | |
| "eval_rougeL": 21.5776, | |
| "eval_rougeLsum": 21.9318, | |
| "eval_runtime": 142.6631, | |
| "eval_samples_per_second": 35.286, | |
| "eval_steps_per_second": 1.472, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "eval_gen_len": 9.006, | |
| "eval_loss": 1.498289704322815, | |
| "eval_rouge1": 15.3475, | |
| "eval_rouge2": 9.3183, | |
| "eval_rougeL": 14.1923, | |
| "eval_rougeLsum": 14.233, | |
| "eval_runtime": 142.2039, | |
| "eval_samples_per_second": 35.4, | |
| "eval_steps_per_second": 1.477, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "eval_gen_len": 5.8882, | |
| "eval_loss": 1.4406640529632568, | |
| "eval_rouge1": 4.4607, | |
| "eval_rouge2": 2.5402, | |
| "eval_rougeL": 4.0061, | |
| "eval_rougeLsum": 4.0532, | |
| "eval_runtime": 139.4058, | |
| "eval_samples_per_second": 36.11, | |
| "eval_steps_per_second": 1.506, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "eval_gen_len": 8.2084, | |
| "eval_loss": 1.4275349378585815, | |
| "eval_rouge1": 11.9263, | |
| "eval_rouge2": 6.9699, | |
| "eval_rougeL": 11.0646, | |
| "eval_rougeLsum": 11.1028, | |
| "eval_runtime": 141.9092, | |
| "eval_samples_per_second": 35.473, | |
| "eval_steps_per_second": 1.48, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.830866088140208e-05, | |
| "loss": 1.3664, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "eval_gen_len": 8.6985, | |
| "eval_loss": 1.3808070421218872, | |
| "eval_rouge1": 10.8243, | |
| "eval_rouge2": 6.646, | |
| "eval_rougeL": 10.1357, | |
| "eval_rougeLsum": 10.1831, | |
| "eval_runtime": 141.9295, | |
| "eval_samples_per_second": 35.468, | |
| "eval_steps_per_second": 1.48, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.824059894503999e-05, | |
| "loss": 1.2606, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.8172537008677897e-05, | |
| "loss": 1.2438, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.8104475072315808e-05, | |
| "loss": 1.2952, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.803641313595372e-05, | |
| "loss": 1.2639, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.796835119959163e-05, | |
| "loss": 1.2335, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.790028926322954e-05, | |
| "loss": 1.258, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.7832227326867452e-05, | |
| "loss": 1.2501, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.7764165390505363e-05, | |
| "loss": 1.2657, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.769610345414327e-05, | |
| "loss": 1.2659, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.762804151778118e-05, | |
| "loss": 1.2469, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.7559979581419092e-05, | |
| "loss": 1.2932, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.7491917645057003e-05, | |
| "loss": 1.2521, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7423855708694914e-05, | |
| "loss": 1.2671, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.7355793772332825e-05, | |
| "loss": 1.2515, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.7287731835970736e-05, | |
| "loss": 1.2208, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.7219669899608644e-05, | |
| "loss": 1.2493, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.7151607963246555e-05, | |
| "loss": 1.2419, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.7083546026884466e-05, | |
| "loss": 1.2693, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.7015484090522377e-05, | |
| "loss": 1.2141, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.6947422154160288e-05, | |
| "loss": 1.2465, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.68793602177982e-05, | |
| "loss": 1.2586, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.681129828143611e-05, | |
| "loss": 1.2394, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.6743236345074017e-05, | |
| "loss": 1.2733, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.6675174408711928e-05, | |
| "loss": 1.2734, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.660711247234984e-05, | |
| "loss": 1.2333, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 1.653905053598775e-05, | |
| "loss": 1.198, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 1.647098859962566e-05, | |
| "loss": 1.2297, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 1.6402926663263572e-05, | |
| "loss": 1.2702, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 1.6334864726901483e-05, | |
| "loss": 1.2799, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 1.6266802790539394e-05, | |
| "loss": 1.2275, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 1.61987408541773e-05, | |
| "loss": 1.2433, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 1.6130678917815213e-05, | |
| "loss": 1.2376, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 1.6062616981453123e-05, | |
| "loss": 1.2032, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.5994555045091034e-05, | |
| "loss": 1.2623, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.5926493108728945e-05, | |
| "loss": 1.2367, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.5858431172366856e-05, | |
| "loss": 1.2015, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.5790369236004764e-05, | |
| "loss": 1.1896, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.5722307299642675e-05, | |
| "loss": 1.2868, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.5654245363280586e-05, | |
| "loss": 1.1849, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.5586183426918497e-05, | |
| "loss": 1.2463, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.5518121490556408e-05, | |
| "loss": 1.2507, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.545005955419432e-05, | |
| "loss": 1.2211, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.538199761783223e-05, | |
| "loss": 1.2082, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.531393568147014e-05, | |
| "loss": 1.2116, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.5245873745108048e-05, | |
| "loss": 1.2223, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.517781180874596e-05, | |
| "loss": 1.2209, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.510974987238387e-05, | |
| "loss": 1.2513, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.5041687936021781e-05, | |
| "loss": 1.1825, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.4977029096477795e-05, | |
| "loss": 1.2369, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.4908967160115706e-05, | |
| "loss": 1.2623, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "eval_gen_len": 10.2668, | |
| "eval_loss": 1.3323159217834473, | |
| "eval_rouge1": 5.5, | |
| "eval_rouge2": 3.2122, | |
| "eval_rougeL": 5.0222, | |
| "eval_rougeLsum": 5.0534, | |
| "eval_runtime": 141.5634, | |
| "eval_samples_per_second": 35.56, | |
| "eval_steps_per_second": 1.483, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.4840905223753617e-05, | |
| "loss": 1.2388, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.4772843287391528e-05, | |
| "loss": 1.2583, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.4704781351029439e-05, | |
| "loss": 1.1844, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.4636719414667348e-05, | |
| "loss": 1.2002, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.4568657478305259e-05, | |
| "loss": 1.2349, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.450059554194317e-05, | |
| "loss": 1.2212, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.443253360558108e-05, | |
| "loss": 1.2169, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.4364471669218992e-05, | |
| "loss": 1.177, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.4296409732856901e-05, | |
| "loss": 1.1969, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.4228347796494812e-05, | |
| "loss": 1.2036, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 1.4160285860132721e-05, | |
| "loss": 1.1921, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 1.4092223923770632e-05, | |
| "loss": 1.2075, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 1.4024161987408542e-05, | |
| "loss": 1.2248, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 1.3956100051046453e-05, | |
| "loss": 1.1921, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 1.3888038114684363e-05, | |
| "loss": 1.2332, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.3819976178322274e-05, | |
| "loss": 1.2762, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.3751914241960185e-05, | |
| "loss": 1.24, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.3683852305598095e-05, | |
| "loss": 1.2117, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.3615790369236006e-05, | |
| "loss": 1.2361, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.3547728432873917e-05, | |
| "loss": 1.2175, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.3479666496511828e-05, | |
| "loss": 1.1877, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.3411604560149739e-05, | |
| "loss": 1.2637, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.3343542623787648e-05, | |
| "loss": 1.1752, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 1.3275480687425559e-05, | |
| "loss": 1.0205, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 1.320741875106347e-05, | |
| "loss": 1.0218, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 1.3139356814701379e-05, | |
| "loss": 1.0558, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 1.3071294878339288e-05, | |
| "loss": 1.0654, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 1.30032329419772e-05, | |
| "loss": 1.0294, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 1.293517100561511e-05, | |
| "loss": 1.0307, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.2867109069253021e-05, | |
| "loss": 1.0403, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 1.2799047132890932e-05, | |
| "loss": 1.0457, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 1.2730985196528841e-05, | |
| "loss": 0.9853, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 1.2662923260166752e-05, | |
| "loss": 1.0287, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.2594861323804663e-05, | |
| "loss": 1.0173, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 1.2526799387442574e-05, | |
| "loss": 1.0501, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.2458737451080485e-05, | |
| "loss": 1.0665, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 1.2390675514718396e-05, | |
| "loss": 1.0629, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 1.2322613578356306e-05, | |
| "loss": 1.0737, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 1.2254551641994217e-05, | |
| "loss": 1.0557, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 1.2186489705632126e-05, | |
| "loss": 1.0174, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.2118427769270035e-05, | |
| "loss": 1.0301, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.2050365832907946e-05, | |
| "loss": 1.0604, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 1.1982303896545857e-05, | |
| "loss": 1.0277, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 1.1914241960183768e-05, | |
| "loss": 1.0805, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 1.1846180023821679e-05, | |
| "loss": 1.0641, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 1.1778118087459588e-05, | |
| "loss": 1.0864, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 1.17100561510975e-05, | |
| "loss": 1.0462, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 1.164199421473541e-05, | |
| "loss": 1.0208, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 1.1573932278373321e-05, | |
| "loss": 1.0617, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 1.1505870342011232e-05, | |
| "loss": 1.039, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 1.1437808405649143e-05, | |
| "loss": 1.046, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 1.1369746469287052e-05, | |
| "loss": 1.0336, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 1.1301684532924963e-05, | |
| "loss": 1.0181, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 1.1233622596562874e-05, | |
| "loss": 1.0551, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 1.1165560660200782e-05, | |
| "loss": 1.0149, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 1.1097498723838693e-05, | |
| "loss": 1.0448, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 1.1029436787476604e-05, | |
| "loss": 1.0477, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 1.0961374851114515e-05, | |
| "loss": 1.0601, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 1.0893312914752426e-05, | |
| "loss": 1.0558, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 1.0825250978390337e-05, | |
| "loss": 1.0861, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 1.0757189042028246e-05, | |
| "loss": 1.0411, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 1.0689127105666157e-05, | |
| "loss": 1.0597, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 1.0621065169304068e-05, | |
| "loss": 1.0235, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 1.0553003232941979e-05, | |
| "loss": 1.063, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.048494129657989e-05, | |
| "loss": 1.0622, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.0416879360217799e-05, | |
| "loss": 1.0778, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.034881742385571e-05, | |
| "loss": 1.0465, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.0280755487493621e-05, | |
| "loss": 1.0644, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.021269355113153e-05, | |
| "loss": 1.0282, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.014463161476944e-05, | |
| "loss": 1.0326, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.007656967840735e-05, | |
| "loss": 1.0591, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.0008507742045262e-05, | |
| "loss": 1.0269, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.940445805683172e-06, | |
| "loss": 1.0326, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 9.872383869321083e-06, | |
| "loss": 1.0935, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.804321932958993e-06, | |
| "loss": 1.0214, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "eval_gen_len": 12.1136, | |
| "eval_loss": 1.3700170516967773, | |
| "eval_rouge1": 6.4308, | |
| "eval_rouge2": 3.9131, | |
| "eval_rougeL": 6.0216, | |
| "eval_rougeLsum": 6.0417, | |
| "eval_runtime": 143.5844, | |
| "eval_samples_per_second": 35.06, | |
| "eval_steps_per_second": 1.463, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 9.736259996596904e-06, | |
| "loss": 1.0338, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.668198060234815e-06, | |
| "loss": 1.046, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 9.600136123872726e-06, | |
| "loss": 1.0208, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.532074187510637e-06, | |
| "loss": 1.0093, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.464012251148546e-06, | |
| "loss": 1.0358, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 9.395950314786457e-06, | |
| "loss": 1.0709, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.327888378424366e-06, | |
| "loss": 1.0602, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.259826442062277e-06, | |
| "loss": 1.0062, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.191764505700188e-06, | |
| "loss": 1.0823, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.123702569338099e-06, | |
| "loss": 1.0439, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.05564063297601e-06, | |
| "loss": 1.0454, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 8.98757869661392e-06, | |
| "loss": 1.0806, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 8.91951676025183e-06, | |
| "loss": 1.0176, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 8.85145482388974e-06, | |
| "loss": 1.0799, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 8.78339288752765e-06, | |
| "loss": 1.0421, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 8.715330951165561e-06, | |
| "loss": 1.0136, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 8.647269014803472e-06, | |
| "loss": 0.9866, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 8.579207078441383e-06, | |
| "loss": 1.0355, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 8.511145142079293e-06, | |
| "loss": 1.0791, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 8.443083205717204e-06, | |
| "loss": 1.0524, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 8.375021269355113e-06, | |
| "loss": 1.0669, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 8.306959332993024e-06, | |
| "loss": 1.007, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 8.238897396630935e-06, | |
| "loss": 1.0121, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 8.170835460268846e-06, | |
| "loss": 1.0281, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 8.102773523906757e-06, | |
| "loss": 1.062, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 8.034711587544666e-06, | |
| "loss": 1.0041, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 7.966649651182577e-06, | |
| "loss": 1.043, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 7.898587714820486e-06, | |
| "loss": 1.0586, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 7.830525778458397e-06, | |
| "loss": 1.0733, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 7.762463842096308e-06, | |
| "loss": 1.0441, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 7.69440190573422e-06, | |
| "loss": 1.0267, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 7.626339969372129e-06, | |
| "loss": 1.0521, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 7.55827803301004e-06, | |
| "loss": 1.0168, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 7.4902160966479495e-06, | |
| "loss": 1.046, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 7.4221541602858605e-06, | |
| "loss": 1.0259, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 7.3540922239237715e-06, | |
| "loss": 1.0173, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 7.286030287561682e-06, | |
| "loss": 1.0507, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 7.2179683511995926e-06, | |
| "loss": 1.0145, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 7.149906414837503e-06, | |
| "loss": 1.0193, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 7.081844478475414e-06, | |
| "loss": 1.0452, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 7.013782542113323e-06, | |
| "loss": 1.0494, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 6.945720605751234e-06, | |
| "loss": 1.0334, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 6.877658669389145e-06, | |
| "loss": 0.9954, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 6.809596733027055e-06, | |
| "loss": 1.0286, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 6.741534796664966e-06, | |
| "loss": 1.0547, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 6.673472860302876e-06, | |
| "loss": 1.0648, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 6.605410923940787e-06, | |
| "loss": 0.9251, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 6.537348987578696e-06, | |
| "loss": 0.9152, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 6.469287051216607e-06, | |
| "loss": 0.9066, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 6.401225114854518e-06, | |
| "loss": 0.8861, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 6.333163178492428e-06, | |
| "loss": 0.8827, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 6.265101242130339e-06, | |
| "loss": 0.9346, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 6.1970393057682494e-06, | |
| "loss": 0.9073, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 6.12897736940616e-06, | |
| "loss": 0.9338, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 6.060915433044071e-06, | |
| "loss": 0.9163, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 5.992853496681981e-06, | |
| "loss": 0.9198, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 5.924791560319892e-06, | |
| "loss": 0.9399, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 5.856729623957802e-06, | |
| "loss": 0.8875, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 5.788667687595713e-06, | |
| "loss": 0.926, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 5.720605751233624e-06, | |
| "loss": 0.9618, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 5.652543814871534e-06, | |
| "loss": 0.9256, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 5.587884975327549e-06, | |
| "loss": 0.9011, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 5.519823038965459e-06, | |
| "loss": 0.9066, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 5.451761102603369e-06, | |
| "loss": 0.8843, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 5.383699166241279e-06, | |
| "loss": 0.9288, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 5.31563722987919e-06, | |
| "loss": 0.9352, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 5.247575293517101e-06, | |
| "loss": 0.937, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 5.1795133571550115e-06, | |
| "loss": 0.9036, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 5.1114514207929224e-06, | |
| "loss": 0.8833, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 5.0433894844308325e-06, | |
| "loss": 0.9567, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 4.975327548068743e-06, | |
| "loss": 0.9222, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 4.907265611706654e-06, | |
| "loss": 0.9528, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 4.839203675344564e-06, | |
| "loss": 0.9287, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 4.771141738982475e-06, | |
| "loss": 0.9269, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 4.703079802620385e-06, | |
| "loss": 0.8977, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "eval_gen_len": 12.9261, | |
| "eval_loss": 1.3723982572555542, | |
| "eval_rouge1": 4.2774, | |
| "eval_rouge2": 2.4929, | |
| "eval_rougeL": 3.9376, | |
| "eval_rougeLsum": 3.9429, | |
| "eval_runtime": 142.026, | |
| "eval_samples_per_second": 35.444, | |
| "eval_steps_per_second": 1.479, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 4.635017866258296e-06, | |
| "loss": 0.913, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 4.566955929896206e-06, | |
| "loss": 0.9012, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 4.498893993534116e-06, | |
| "loss": 0.9077, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 4.430832057172027e-06, | |
| "loss": 0.9181, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 4.362770120809938e-06, | |
| "loss": 0.9234, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 4.294708184447848e-06, | |
| "loss": 0.9014, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 4.226646248085758e-06, | |
| "loss": 0.8963, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 4.158584311723669e-06, | |
| "loss": 0.9097, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 4.090522375361579e-06, | |
| "loss": 0.903, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 4.0224604389994894e-06, | |
| "loss": 0.8938, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 3.9543985026374e-06, | |
| "loss": 0.9029, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 3.886336566275311e-06, | |
| "loss": 0.9338, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 3.8182746299132215e-06, | |
| "loss": 0.9295, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 3.7502126935511316e-06, | |
| "loss": 0.8946, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 3.6821507571890426e-06, | |
| "loss": 0.891, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 3.6140888208269527e-06, | |
| "loss": 0.9123, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 3.5460268844648632e-06, | |
| "loss": 0.9518, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 3.477964948102774e-06, | |
| "loss": 0.9378, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 3.4099030117406843e-06, | |
| "loss": 0.9089, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 3.3418410753785945e-06, | |
| "loss": 0.9241, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 3.2737791390165054e-06, | |
| "loss": 0.9281, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 3.205717202654416e-06, | |
| "loss": 0.9082, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 3.1376552662923265e-06, | |
| "loss": 0.915, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 3.0695933299302366e-06, | |
| "loss": 0.882, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 3.001531393568147e-06, | |
| "loss": 0.9182, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 2.9334694572060577e-06, | |
| "loss": 0.9369, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 2.865407520843968e-06, | |
| "loss": 0.9324, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 2.797345584481879e-06, | |
| "loss": 0.9069, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 2.7292836481197893e-06, | |
| "loss": 0.9079, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 2.6612217117577e-06, | |
| "loss": 0.8965, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 2.59315977539561e-06, | |
| "loss": 0.9312, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 2.5250978390335206e-06, | |
| "loss": 0.9454, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 2.4570359026714315e-06, | |
| "loss": 0.9115, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 2.3889739663093416e-06, | |
| "loss": 0.9265, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 2.320912029947252e-06, | |
| "loss": 0.9381, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 2.2528500935851627e-06, | |
| "loss": 0.9334, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 2.1847881572230733e-06, | |
| "loss": 0.914, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 2.116726220860984e-06, | |
| "loss": 0.8805, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 2.048664284498894e-06, | |
| "loss": 0.9417, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 1.980602348136805e-06, | |
| "loss": 0.8974, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 1.912540411774715e-06, | |
| "loss": 0.9722, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 1.8444784754126258e-06, | |
| "loss": 0.9155, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 1.7764165390505361e-06, | |
| "loss": 0.9088, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 1.7083546026884464e-06, | |
| "loss": 0.894, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 1.6402926663263572e-06, | |
| "loss": 0.936, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 1.5722307299642675e-06, | |
| "loss": 0.8867, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 1.504168793602178e-06, | |
| "loss": 0.9142, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 1.4361068572400886e-06, | |
| "loss": 0.9394, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 1.3680449208779992e-06, | |
| "loss": 0.9313, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 1.2999829845159095e-06, | |
| "loss": 0.9197, | |
| "step": 5500 | |
| } | |
| ], | |
| "max_steps": 5877, | |
| "num_train_epochs": 3, | |
| "total_flos": 1.5087907085549568e+16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |