| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 4.985976939856653, |
| "global_step": 4000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.06, |
| "learning_rate": 4.2e-05, |
| "loss": 0.8909, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 9.1e-05, |
| "loss": 0.8023, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.12, |
| "eval_loss": 0.7678152918815613, |
| "eval_runtime": 2324.1166, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 9.895140664961637e-05, |
| "loss": 0.7815, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 9.767263427109975e-05, |
| "loss": 0.8184, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.25, |
| "eval_loss": 0.7796664834022522, |
| "eval_runtime": 2323.4202, |
| "eval_samples_per_second": 0.633, |
| "eval_steps_per_second": 0.633, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 9.639386189258313e-05, |
| "loss": 0.7501, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 9.51150895140665e-05, |
| "loss": 0.7786, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.37, |
| "eval_loss": 0.7461986541748047, |
| "eval_runtime": 2323.156, |
| "eval_samples_per_second": 0.633, |
| "eval_steps_per_second": 0.633, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 9.383631713554988e-05, |
| "loss": 0.7247, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 9.255754475703325e-05, |
| "loss": 0.699, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.5, |
| "eval_loss": 0.7154579162597656, |
| "eval_runtime": 2326.716, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 9.127877237851663e-05, |
| "loss": 0.7498, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 9e-05, |
| "loss": 0.6729, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.62, |
| "eval_loss": 0.6918142437934875, |
| "eval_runtime": 2326.9582, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 8.872122762148338e-05, |
| "loss": 0.6821, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 8.744245524296676e-05, |
| "loss": 0.7392, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.75, |
| "eval_loss": 0.6949509978294373, |
| "eval_runtime": 2328.8704, |
| "eval_samples_per_second": 0.631, |
| "eval_steps_per_second": 0.631, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 8.616368286445013e-05, |
| "loss": 0.6848, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 8.488491048593351e-05, |
| "loss": 0.7601, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.87, |
| "eval_loss": 0.6541839241981506, |
| "eval_runtime": 2322.3854, |
| "eval_samples_per_second": 0.633, |
| "eval_steps_per_second": 0.633, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 8.360613810741689e-05, |
| "loss": 0.7163, |
| "step": 750 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 8.232736572890026e-05, |
| "loss": 0.7121, |
| "step": 800 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_loss": 0.7020866274833679, |
| "eval_runtime": 2329.0246, |
| "eval_samples_per_second": 0.631, |
| "eval_steps_per_second": 0.631, |
| "step": 800 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 8.104859335038364e-05, |
| "loss": 0.6303, |
| "step": 850 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 7.992327365728901e-05, |
| "loss": 0.834, |
| "step": 900 |
| }, |
| { |
| "epoch": 1.12, |
| "eval_loss": 0.8280826210975647, |
| "eval_runtime": 2322.7118, |
| "eval_samples_per_second": 0.633, |
| "eval_steps_per_second": 0.633, |
| "step": 900 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 7.867007672634272e-05, |
| "loss": 0.7102, |
| "step": 950 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 7.73913043478261e-05, |
| "loss": 0.6845, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1.25, |
| "eval_loss": 0.684939444065094, |
| "eval_runtime": 2322.9147, |
| "eval_samples_per_second": 0.633, |
| "eval_steps_per_second": 0.633, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 7.611253196930947e-05, |
| "loss": 0.6918, |
| "step": 1050 |
| }, |
| { |
| "epoch": 1.37, |
| "learning_rate": 7.483375959079285e-05, |
| "loss": 0.7139, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.37, |
| "eval_loss": 0.682366132736206, |
| "eval_runtime": 2323.8642, |
| "eval_samples_per_second": 0.633, |
| "eval_steps_per_second": 0.633, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 7.355498721227622e-05, |
| "loss": 0.6513, |
| "step": 1150 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 7.22762148337596e-05, |
| "loss": 0.6515, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1.5, |
| "eval_loss": 0.6614900827407837, |
| "eval_runtime": 2324.2619, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 7.099744245524297e-05, |
| "loss": 0.6325, |
| "step": 1250 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 6.971867007672635e-05, |
| "loss": 0.6658, |
| "step": 1300 |
| }, |
| { |
| "epoch": 1.62, |
| "eval_loss": 0.6691812872886658, |
| "eval_runtime": 2322.8992, |
| "eval_samples_per_second": 0.633, |
| "eval_steps_per_second": 0.633, |
| "step": 1300 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 6.843989769820973e-05, |
| "loss": 0.6709, |
| "step": 1350 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 6.71611253196931e-05, |
| "loss": 0.6571, |
| "step": 1400 |
| }, |
| { |
| "epoch": 1.75, |
| "eval_loss": 0.7406654357910156, |
| "eval_runtime": 2323.316, |
| "eval_samples_per_second": 0.633, |
| "eval_steps_per_second": 0.633, |
| "step": 1400 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 6.588235294117648e-05, |
| "loss": 0.6527, |
| "step": 1450 |
| }, |
| { |
| "epoch": 1.87, |
| "learning_rate": 6.460358056265985e-05, |
| "loss": 0.6432, |
| "step": 1500 |
| }, |
| { |
| "epoch": 1.87, |
| "eval_loss": 0.6645843982696533, |
| "eval_runtime": 2325.3057, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 1500 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 6.332480818414323e-05, |
| "loss": 0.6998, |
| "step": 1550 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 6.20460358056266e-05, |
| "loss": 0.6777, |
| "step": 1600 |
| }, |
| { |
| "epoch": 1.99, |
| "eval_loss": 0.7001596093177795, |
| "eval_runtime": 2324.698, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 1600 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 6.0767263427109976e-05, |
| "loss": 0.63, |
| "step": 1650 |
| }, |
| { |
| "epoch": 2.12, |
| "learning_rate": 5.948849104859335e-05, |
| "loss": 0.6027, |
| "step": 1700 |
| }, |
| { |
| "epoch": 2.12, |
| "eval_loss": 0.6740280389785767, |
| "eval_runtime": 2325.0824, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 1700 |
| }, |
| { |
| "epoch": 2.18, |
| "learning_rate": 5.820971867007673e-05, |
| "loss": 0.6171, |
| "step": 1750 |
| }, |
| { |
| "epoch": 2.24, |
| "learning_rate": 5.6930946291560104e-05, |
| "loss": 0.5711, |
| "step": 1800 |
| }, |
| { |
| "epoch": 2.24, |
| "eval_loss": 0.6770710349082947, |
| "eval_runtime": 2324.7006, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 1800 |
| }, |
| { |
| "epoch": 2.31, |
| "learning_rate": 5.565217391304348e-05, |
| "loss": 0.6193, |
| "step": 1850 |
| }, |
| { |
| "epoch": 2.37, |
| "learning_rate": 5.4373401534526856e-05, |
| "loss": 0.6544, |
| "step": 1900 |
| }, |
| { |
| "epoch": 2.37, |
| "eval_loss": 0.7468971014022827, |
| "eval_runtime": 2324.7323, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 1900 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 5.309462915601023e-05, |
| "loss": 0.6559, |
| "step": 1950 |
| }, |
| { |
| "epoch": 2.49, |
| "learning_rate": 5.181585677749361e-05, |
| "loss": 0.6981, |
| "step": 2000 |
| }, |
| { |
| "epoch": 2.49, |
| "eval_loss": 0.6552309393882751, |
| "eval_runtime": 2325.8405, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 2000 |
| }, |
| { |
| "epoch": 2.56, |
| "learning_rate": 5.0537084398976984e-05, |
| "loss": 0.6161, |
| "step": 2050 |
| }, |
| { |
| "epoch": 2.62, |
| "learning_rate": 4.925831202046036e-05, |
| "loss": 0.6425, |
| "step": 2100 |
| }, |
| { |
| "epoch": 2.62, |
| "eval_loss": 0.6608405113220215, |
| "eval_runtime": 2324.8602, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 2100 |
| }, |
| { |
| "epoch": 2.68, |
| "learning_rate": 4.7979539641943737e-05, |
| "loss": 0.5966, |
| "step": 2150 |
| }, |
| { |
| "epoch": 2.74, |
| "learning_rate": 4.670076726342711e-05, |
| "loss": 0.6435, |
| "step": 2200 |
| }, |
| { |
| "epoch": 2.74, |
| "eval_loss": 0.6722382307052612, |
| "eval_runtime": 2325.4748, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 2200 |
| }, |
| { |
| "epoch": 2.8, |
| "learning_rate": 4.542199488491049e-05, |
| "loss": 0.6493, |
| "step": 2250 |
| }, |
| { |
| "epoch": 2.87, |
| "learning_rate": 4.4143222506393865e-05, |
| "loss": 0.6281, |
| "step": 2300 |
| }, |
| { |
| "epoch": 2.87, |
| "eval_loss": 0.6553158760070801, |
| "eval_runtime": 2324.9442, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 2300 |
| }, |
| { |
| "epoch": 2.93, |
| "learning_rate": 4.286445012787724e-05, |
| "loss": 0.5495, |
| "step": 2350 |
| }, |
| { |
| "epoch": 2.99, |
| "learning_rate": 4.158567774936061e-05, |
| "loss": 0.6129, |
| "step": 2400 |
| }, |
| { |
| "epoch": 2.99, |
| "eval_loss": 0.6536511182785034, |
| "eval_runtime": 2325.6669, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 2400 |
| }, |
| { |
| "epoch": 3.05, |
| "learning_rate": 4.0306905370843986e-05, |
| "loss": 0.5085, |
| "step": 2450 |
| }, |
| { |
| "epoch": 3.12, |
| "learning_rate": 3.902813299232736e-05, |
| "loss": 0.5443, |
| "step": 2500 |
| }, |
| { |
| "epoch": 3.12, |
| "eval_loss": 0.659809947013855, |
| "eval_runtime": 2326.2914, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 2500 |
| }, |
| { |
| "epoch": 3.18, |
| "learning_rate": 3.774936061381074e-05, |
| "loss": 0.5447, |
| "step": 2550 |
| }, |
| { |
| "epoch": 3.24, |
| "learning_rate": 3.6470588235294114e-05, |
| "loss": 0.535, |
| "step": 2600 |
| }, |
| { |
| "epoch": 3.24, |
| "eval_loss": 0.6724178194999695, |
| "eval_runtime": 2327.4156, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 2600 |
| }, |
| { |
| "epoch": 3.3, |
| "learning_rate": 3.519181585677749e-05, |
| "loss": 0.6094, |
| "step": 2650 |
| }, |
| { |
| "epoch": 3.37, |
| "learning_rate": 3.3913043478260867e-05, |
| "loss": 0.5347, |
| "step": 2700 |
| }, |
| { |
| "epoch": 3.37, |
| "eval_loss": 0.6599487662315369, |
| "eval_runtime": 2327.143, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 2700 |
| }, |
| { |
| "epoch": 3.43, |
| "learning_rate": 3.263427109974424e-05, |
| "loss": 0.572, |
| "step": 2750 |
| }, |
| { |
| "epoch": 3.49, |
| "learning_rate": 3.135549872122762e-05, |
| "loss": 0.5493, |
| "step": 2800 |
| }, |
| { |
| "epoch": 3.49, |
| "eval_loss": 0.6231494545936584, |
| "eval_runtime": 2326.2678, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 2800 |
| }, |
| { |
| "epoch": 3.55, |
| "learning_rate": 3.0076726342710998e-05, |
| "loss": 0.5566, |
| "step": 2850 |
| }, |
| { |
| "epoch": 3.61, |
| "learning_rate": 2.8797953964194374e-05, |
| "loss": 0.5564, |
| "step": 2900 |
| }, |
| { |
| "epoch": 3.61, |
| "eval_loss": 0.6593871116638184, |
| "eval_runtime": 2327.2395, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 2900 |
| }, |
| { |
| "epoch": 3.68, |
| "learning_rate": 2.751918158567775e-05, |
| "loss": 0.5787, |
| "step": 2950 |
| }, |
| { |
| "epoch": 3.74, |
| "learning_rate": 2.6240409207161126e-05, |
| "loss": 0.509, |
| "step": 3000 |
| }, |
| { |
| "epoch": 3.74, |
| "eval_loss": 0.6348086595535278, |
| "eval_runtime": 2326.4619, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 3000 |
| }, |
| { |
| "epoch": 3.8, |
| "learning_rate": 2.4961636828644502e-05, |
| "loss": 0.5356, |
| "step": 3050 |
| }, |
| { |
| "epoch": 3.86, |
| "learning_rate": 2.368286445012788e-05, |
| "loss": 0.52, |
| "step": 3100 |
| }, |
| { |
| "epoch": 3.86, |
| "eval_loss": 0.6384520530700684, |
| "eval_runtime": 2327.6186, |
| "eval_samples_per_second": 0.632, |
| "eval_steps_per_second": 0.632, |
| "step": 3100 |
| }, |
| { |
| "epoch": 3.93, |
| "learning_rate": 2.2404092071611255e-05, |
| "loss": 0.5418, |
| "step": 3150 |
| }, |
| { |
| "epoch": 3.99, |
| "learning_rate": 2.112531969309463e-05, |
| "loss": 0.5569, |
| "step": 3200 |
| }, |
| { |
| "epoch": 3.99, |
| "eval_loss": 0.6191104650497437, |
| "eval_runtime": 2329.026, |
| "eval_samples_per_second": 0.631, |
| "eval_steps_per_second": 0.631, |
| "step": 3200 |
| }, |
| { |
| "epoch": 4.05, |
| "learning_rate": 1.9846547314578007e-05, |
| "loss": 0.4716, |
| "step": 3250 |
| }, |
| { |
| "epoch": 4.11, |
| "learning_rate": 1.8567774936061383e-05, |
| "loss": 0.501, |
| "step": 3300 |
| }, |
| { |
| "epoch": 4.11, |
| "eval_loss": 0.6244591474533081, |
| "eval_runtime": 2328.4154, |
| "eval_samples_per_second": 0.631, |
| "eval_steps_per_second": 0.631, |
| "step": 3300 |
| }, |
| { |
| "epoch": 4.18, |
| "learning_rate": 1.728900255754476e-05, |
| "loss": 0.4853, |
| "step": 3350 |
| }, |
| { |
| "epoch": 4.24, |
| "learning_rate": 1.6010230179028135e-05, |
| "loss": 0.5105, |
| "step": 3400 |
| }, |
| { |
| "epoch": 4.24, |
| "eval_loss": 0.6227459907531738, |
| "eval_runtime": 2328.6203, |
| "eval_samples_per_second": 0.631, |
| "eval_steps_per_second": 0.631, |
| "step": 3400 |
| }, |
| { |
| "epoch": 4.3, |
| "learning_rate": 1.4731457800511508e-05, |
| "loss": 0.4757, |
| "step": 3450 |
| }, |
| { |
| "epoch": 4.36, |
| "learning_rate": 1.3452685421994884e-05, |
| "loss": 0.5064, |
| "step": 3500 |
| }, |
| { |
| "epoch": 4.36, |
| "eval_loss": 0.628924548625946, |
| "eval_runtime": 2328.7385, |
| "eval_samples_per_second": 0.631, |
| "eval_steps_per_second": 0.631, |
| "step": 3500 |
| }, |
| { |
| "epoch": 4.43, |
| "learning_rate": 1.2173913043478261e-05, |
| "loss": 0.5107, |
| "step": 3550 |
| }, |
| { |
| "epoch": 4.49, |
| "learning_rate": 1.0895140664961638e-05, |
| "loss": 0.5088, |
| "step": 3600 |
| }, |
| { |
| "epoch": 4.49, |
| "eval_loss": 0.620337724685669, |
| "eval_runtime": 2329.0799, |
| "eval_samples_per_second": 0.631, |
| "eval_steps_per_second": 0.631, |
| "step": 3600 |
| }, |
| { |
| "epoch": 4.55, |
| "learning_rate": 9.616368286445012e-06, |
| "loss": 0.5184, |
| "step": 3650 |
| }, |
| { |
| "epoch": 4.61, |
| "learning_rate": 8.337595907928388e-06, |
| "loss": 0.5246, |
| "step": 3700 |
| }, |
| { |
| "epoch": 4.61, |
| "eval_loss": 0.618162989616394, |
| "eval_runtime": 2327.9739, |
| "eval_samples_per_second": 0.631, |
| "eval_steps_per_second": 0.631, |
| "step": 3700 |
| }, |
| { |
| "epoch": 4.67, |
| "learning_rate": 7.058823529411765e-06, |
| "loss": 0.5027, |
| "step": 3750 |
| }, |
| { |
| "epoch": 4.74, |
| "learning_rate": 5.780051150895141e-06, |
| "loss": 0.4927, |
| "step": 3800 |
| }, |
| { |
| "epoch": 4.74, |
| "eval_loss": 0.6198447942733765, |
| "eval_runtime": 2328.8672, |
| "eval_samples_per_second": 0.631, |
| "eval_steps_per_second": 0.631, |
| "step": 3800 |
| }, |
| { |
| "epoch": 4.8, |
| "learning_rate": 4.501278772378517e-06, |
| "loss": 0.5568, |
| "step": 3850 |
| }, |
| { |
| "epoch": 4.86, |
| "learning_rate": 3.2225063938618927e-06, |
| "loss": 0.5038, |
| "step": 3900 |
| }, |
| { |
| "epoch": 4.86, |
| "eval_loss": 0.6127236485481262, |
| "eval_runtime": 2329.1619, |
| "eval_samples_per_second": 0.631, |
| "eval_steps_per_second": 0.631, |
| "step": 3900 |
| }, |
| { |
| "epoch": 4.92, |
| "learning_rate": 1.943734015345269e-06, |
| "loss": 0.5114, |
| "step": 3950 |
| }, |
| { |
| "epoch": 4.99, |
| "learning_rate": 6.649616368286445e-07, |
| "loss": 0.5494, |
| "step": 4000 |
| }, |
| { |
| "epoch": 4.99, |
| "eval_loss": 0.614413321018219, |
| "eval_runtime": 2328.9991, |
| "eval_samples_per_second": 0.631, |
| "eval_steps_per_second": 0.631, |
| "step": 4000 |
| } |
| ], |
| "max_steps": 4010, |
| "num_train_epochs": 5, |
| "total_flos": 3.2509882552308204e+18, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|