| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.384191899331498, | |
| "eval_steps": 500, | |
| "global_step": 5280, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 8.680555555555556e-06, | |
| "loss": 1.4509, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.736111111111111e-05, | |
| "loss": 0.7733, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 2.604166666666667e-05, | |
| "loss": 0.4247, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 3.472222222222222e-05, | |
| "loss": 0.3633, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.340277777777778e-05, | |
| "loss": 0.3354, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 5.208333333333334e-05, | |
| "loss": 0.3234, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 6.076388888888889e-05, | |
| "loss": 0.3164, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 6.944444444444444e-05, | |
| "loss": 0.2978, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 7.8125e-05, | |
| "loss": 0.2887, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 8.680555555555556e-05, | |
| "loss": 0.2838, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.548611111111112e-05, | |
| "loss": 0.2853, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 9.999471159635539e-05, | |
| "loss": 0.2798, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 9.994973097955704e-05, | |
| "loss": 0.2753, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 9.98588929347481e-05, | |
| "loss": 0.279, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 9.972228085775511e-05, | |
| "loss": 0.2685, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 9.954002016824227e-05, | |
| "loss": 0.2633, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 9.931227819456718e-05, | |
| "loss": 0.2627, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.903926402016153e-05, | |
| "loss": 0.2549, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 9.872122829157737e-05, | |
| "loss": 0.2526, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.835846298837584e-05, | |
| "loss": 0.2518, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 9.795130115506887e-05, | |
| "loss": 0.2532, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 9.750011659536058e-05, | |
| "loss": 0.256, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 9.700532352896867e-05, | |
| "loss": 0.2451, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 9.646737621134112e-05, | |
| "loss": 0.2453, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 9.588676851661718e-05, | |
| "loss": 0.2481, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 9.526403348421544e-05, | |
| "loss": 0.2436, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 9.459974282946571e-05, | |
| "loss": 0.2386, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 9.389450641873323e-05, | |
| "loss": 0.2385, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.314897170951779e-05, | |
| "loss": 0.2319, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.23638231560414e-05, | |
| "loss": 0.2361, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 9.153978158087033e-05, | |
| "loss": 0.2429, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.067760351314838e-05, | |
| "loss": 0.2373, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 8.9778080494049e-05, | |
| "loss": 0.2311, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 8.884203835008382e-05, | |
| "loss": 0.2327, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 8.787033643493483e-05, | |
| "loss": 0.2321, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 8.68638668405062e-05, | |
| "loss": 0.2302, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 8.582355357792003e-05, | |
| "loss": 0.2179, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 8.475035172920804e-05, | |
| "loss": 0.2305, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 8.364524657047789e-05, | |
| "loss": 0.2241, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 8.250925266735918e-05, | |
| "loss": 0.2243, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 8.134341294355963e-05, | |
| "loss": 0.2219, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 8.014879772338649e-05, | |
| "loss": 0.2256, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 7.89265037491121e-05, | |
| "loss": 0.2187, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 7.767765317408613e-05, | |
| "loss": 0.2187, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 7.64033925325184e-05, | |
| "loss": 0.2207, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 7.510489168687851e-05, | |
| "loss": 0.2189, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 7.378334275387821e-05, | |
| "loss": 0.2159, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 7.243995901002312e-05, | |
| "loss": 0.2132, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 7.10759737777379e-05, | |
| "loss": 0.2133, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 6.969263929308812e-05, | |
| "loss": 0.2162, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 6.829122555613785e-05, | |
| "loss": 0.2139, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 6.687301916499871e-05, | |
| "loss": 0.211, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 6.543932213464079e-05, | |
| "loss": 0.2088, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 6.399145070154961e-05, | |
| "loss": 0.2061, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 6.253073411532693e-05, | |
| "loss": 0.2053, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 6.105851341834439e-05, | |
| "loss": 0.2058, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 5.957614021457072e-05, | |
| "loss": 0.2025, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 5.808497542870246e-05, | |
| "loss": 0.2021, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 5.658638805673777e-05, | |
| "loss": 0.205, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 5.5081753909140096e-05, | |
| "loss": 0.2031, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 5.357245434774567e-05, | |
| "loss": 0.2031, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 5.205987501757452e-05, | |
| "loss": 0.1999, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 5.054540457470912e-05, | |
| "loss": 0.1975, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.903043341140879e-05, | |
| "loss": 0.1978, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.751635237962999e-05, | |
| "loss": 0.2042, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.600455151412482e-05, | |
| "loss": 0.1877, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.4496418756289426e-05, | |
| "loss": 0.2019, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.299333867993469e-05, | |
| "loss": 0.1939, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.149669122014823e-05, | |
| "loss": 0.1965, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.0007850406415234e-05, | |
| "loss": 0.1937, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.8528183101161126e-05, | |
| "loss": 0.1908, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.705904774487396e-05, | |
| "loss": 0.1914, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 3.560179310895886e-05, | |
| "loss": 0.1897, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.415775705746934e-05, | |
| "loss": 0.188, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.272826531885229e-05, | |
| "loss": 0.1884, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.131463026883449e-05, | |
| "loss": 0.1878, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 2.9918149725567645e-05, | |
| "loss": 0.177, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 2.854010575813856e-05, | |
| "loss": 0.1666, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 2.7181763509537816e-05, | |
| "loss": 0.1679, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 2.5844370035168073e-05, | |
| "loss": 0.1694, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 2.4529153157957913e-05, | |
| "loss": 0.1684, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 2.3237320341132497e-05, | |
| "loss": 0.167, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 2.197005757967593e-05, | |
| "loss": 0.1669, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 2.0728528311502976e-05, | |
| "loss": 0.1618, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.9513872349339796e-05, | |
| "loss": 0.1641, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.832720483429426e-05, | |
| "loss": 0.1635, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 1.7169615212076613e-05, | |
| "loss": 0.162, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 1.6042166232810347e-05, | |
| "loss": 0.16, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 1.49458929753515e-05, | |
| "loss": 0.1626, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.3881801897012225e-05, | |
| "loss": 0.161, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.2850869909560904e-05, | |
| "loss": 0.1587, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 1.1854043482347243e-05, | |
| "loss": 0.1611, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 1.089223777337568e-05, | |
| "loss": 0.1621, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 9.966335789124809e-06, | |
| "loss": 0.1594, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 9.077187573884244e-06, | |
| "loss": 0.1604, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 8.225609429353187e-06, | |
| "loss": 0.1551, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 7.4123831652168965e-06, | |
| "loss": 0.1604, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 6.6382553813896075e-06, | |
| "loss": 0.1573, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 5.903936782582253e-06, | |
| "loss": 0.1525, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 5.210101525824746e-06, | |
| "loss": 0.1545, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 4.55738660154153e-06, | |
| "loss": 0.1552, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.946391248748821e-06, | |
| "loss": 0.1564, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.3776764049102684e-06, | |
| "loss": 0.1564, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 2.8517641909562077e-06, | |
| "loss": 0.1541, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.3691374319393166e-06, | |
| "loss": 0.156, | |
| "step": 5250 | |
| } | |
| ], | |
| "logging_steps": 50, | |
| "max_steps": 5760, | |
| "num_train_epochs": 2, | |
| "save_steps": 480, | |
| "total_flos": 6.995902140215624e+17, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |