| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.0, |
| "eval_steps": 500, |
| "global_step": 120516, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.01, |
| "learning_rate": 1.9944682310509254e-05, |
| "loss": 1.5408, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 1.988936462101851e-05, |
| "loss": 1.2376, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 1.9834046931527766e-05, |
| "loss": 1.1307, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 1.977872924203702e-05, |
| "loss": 1.103, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 1.9723411552546275e-05, |
| "loss": 1.0708, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.9668093863055527e-05, |
| "loss": 1.0288, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.9612776173564783e-05, |
| "loss": 1.0089, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.955745848407404e-05, |
| "loss": 1.0096, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.9502140794583292e-05, |
| "loss": 1.0213, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.9446823105092548e-05, |
| "loss": 0.9962, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.9391505415601804e-05, |
| "loss": 0.9709, |
| "step": 5500 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.9336187726111056e-05, |
| "loss": 0.9898, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.9280870036620312e-05, |
| "loss": 0.9536, |
| "step": 6500 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9225552347129565e-05, |
| "loss": 0.9591, |
| "step": 7000 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.917023465763882e-05, |
| "loss": 0.9851, |
| "step": 7500 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.9114916968148077e-05, |
| "loss": 0.9548, |
| "step": 8000 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.905959927865733e-05, |
| "loss": 1.0179, |
| "step": 8500 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.9004281589166585e-05, |
| "loss": 0.9737, |
| "step": 9000 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.894896389967584e-05, |
| "loss": 0.9592, |
| "step": 9500 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.8893646210185094e-05, |
| "loss": 0.9488, |
| "step": 10000 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.883832852069435e-05, |
| "loss": 0.9598, |
| "step": 10500 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.8783010831203603e-05, |
| "loss": 0.9572, |
| "step": 11000 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.872769314171286e-05, |
| "loss": 0.9789, |
| "step": 11500 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.8672375452222115e-05, |
| "loss": 0.9503, |
| "step": 12000 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.8617057762731367e-05, |
| "loss": 0.9296, |
| "step": 12500 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.8561740073240623e-05, |
| "loss": 0.933, |
| "step": 13000 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.850642238374988e-05, |
| "loss": 0.9438, |
| "step": 13500 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.845110469425913e-05, |
| "loss": 0.919, |
| "step": 14000 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.8395787004768388e-05, |
| "loss": 0.9704, |
| "step": 14500 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.834046931527764e-05, |
| "loss": 0.9099, |
| "step": 15000 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.8285151625786896e-05, |
| "loss": 0.9517, |
| "step": 15500 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.8229833936296152e-05, |
| "loss": 0.9463, |
| "step": 16000 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.8174516246805405e-05, |
| "loss": 0.9535, |
| "step": 16500 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.8119198557314657e-05, |
| "loss": 0.9372, |
| "step": 17000 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.8063880867823917e-05, |
| "loss": 0.9515, |
| "step": 17500 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.800856317833317e-05, |
| "loss": 0.9675, |
| "step": 18000 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.7953245488842422e-05, |
| "loss": 0.9579, |
| "step": 18500 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.7897927799351678e-05, |
| "loss": 0.961, |
| "step": 19000 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.7842610109860934e-05, |
| "loss": 0.9397, |
| "step": 19500 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.7787292420370186e-05, |
| "loss": 0.9518, |
| "step": 20000 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.7731974730879442e-05, |
| "loss": 0.9745, |
| "step": 20500 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.7676657041388695e-05, |
| "loss": 0.8979, |
| "step": 21000 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.762133935189795e-05, |
| "loss": 0.9224, |
| "step": 21500 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.7566021662407207e-05, |
| "loss": 0.9556, |
| "step": 22000 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.751070397291646e-05, |
| "loss": 0.978, |
| "step": 22500 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.7455386283425715e-05, |
| "loss": 0.9102, |
| "step": 23000 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.7400068593934968e-05, |
| "loss": 0.9124, |
| "step": 23500 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.7344750904444224e-05, |
| "loss": 0.9089, |
| "step": 24000 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.728943321495348e-05, |
| "loss": 0.9296, |
| "step": 24500 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.7234115525462733e-05, |
| "loss": 0.9406, |
| "step": 25000 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.717879783597199e-05, |
| "loss": 0.9625, |
| "step": 25500 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.7123480146481245e-05, |
| "loss": 0.9196, |
| "step": 26000 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.7068162456990497e-05, |
| "loss": 0.9402, |
| "step": 26500 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.7012844767499753e-05, |
| "loss": 0.977, |
| "step": 27000 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.6957527078009006e-05, |
| "loss": 0.9082, |
| "step": 27500 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.690220938851826e-05, |
| "loss": 0.937, |
| "step": 28000 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.6846891699027518e-05, |
| "loss": 0.8986, |
| "step": 28500 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.679157400953677e-05, |
| "loss": 0.9493, |
| "step": 29000 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.6736256320046026e-05, |
| "loss": 0.9245, |
| "step": 29500 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.6680938630555282e-05, |
| "loss": 0.8917, |
| "step": 30000 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.6625620941064535e-05, |
| "loss": 0.9436, |
| "step": 30500 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.657030325157379e-05, |
| "loss": 0.9, |
| "step": 31000 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 1.6514985562083043e-05, |
| "loss": 0.9481, |
| "step": 31500 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 1.64596678725923e-05, |
| "loss": 0.9146, |
| "step": 32000 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 1.6404350183101555e-05, |
| "loss": 0.9303, |
| "step": 32500 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 1.6349032493610808e-05, |
| "loss": 0.9305, |
| "step": 33000 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 1.629371480412006e-05, |
| "loss": 0.9222, |
| "step": 33500 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 1.623839711462932e-05, |
| "loss": 0.9495, |
| "step": 34000 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 1.6183079425138572e-05, |
| "loss": 0.8726, |
| "step": 34500 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 1.6127761735647825e-05, |
| "loss": 0.9479, |
| "step": 35000 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 1.607244404615708e-05, |
| "loss": 0.9395, |
| "step": 35500 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 1.6017126356666337e-05, |
| "loss": 0.9261, |
| "step": 36000 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 1.596180866717559e-05, |
| "loss": 0.8737, |
| "step": 36500 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 1.5906490977684845e-05, |
| "loss": 0.91, |
| "step": 37000 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 1.5851173288194098e-05, |
| "loss": 0.9309, |
| "step": 37500 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 1.5795855598703354e-05, |
| "loss": 0.9278, |
| "step": 38000 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 1.574053790921261e-05, |
| "loss": 0.9506, |
| "step": 38500 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 1.5685220219721863e-05, |
| "loss": 0.8787, |
| "step": 39000 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 1.562990253023112e-05, |
| "loss": 0.9097, |
| "step": 39500 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 1.557458484074037e-05, |
| "loss": 0.8812, |
| "step": 40000 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 1.5519267151249627e-05, |
| "loss": 0.8798, |
| "step": 40500 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 1.5463949461758883e-05, |
| "loss": 0.9398, |
| "step": 41000 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 1.5408631772268136e-05, |
| "loss": 0.9148, |
| "step": 41500 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 1.535331408277739e-05, |
| "loss": 0.9398, |
| "step": 42000 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 1.5297996393286648e-05, |
| "loss": 0.8984, |
| "step": 42500 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 1.5242678703795902e-05, |
| "loss": 0.8947, |
| "step": 43000 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 1.5187361014305155e-05, |
| "loss": 0.8767, |
| "step": 43500 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 1.5132043324814409e-05, |
| "loss": 0.906, |
| "step": 44000 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 1.5076725635323666e-05, |
| "loss": 0.9188, |
| "step": 44500 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 1.5021407945832919e-05, |
| "loss": 0.9534, |
| "step": 45000 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 1.4966090256342173e-05, |
| "loss": 0.913, |
| "step": 45500 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 1.4910772566851428e-05, |
| "loss": 0.8824, |
| "step": 46000 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 1.4855454877360684e-05, |
| "loss": 0.9155, |
| "step": 46500 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 1.4800137187869938e-05, |
| "loss": 0.9142, |
| "step": 47000 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 1.4744819498379192e-05, |
| "loss": 0.9195, |
| "step": 47500 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 1.4689501808888446e-05, |
| "loss": 0.9207, |
| "step": 48000 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 1.4634184119397702e-05, |
| "loss": 0.9057, |
| "step": 48500 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.4578866429906957e-05, |
| "loss": 0.9273, |
| "step": 49000 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.4523548740416211e-05, |
| "loss": 0.9333, |
| "step": 49500 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.4468231050925465e-05, |
| "loss": 0.8996, |
| "step": 50000 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.4412913361434721e-05, |
| "loss": 0.9298, |
| "step": 50500 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.4357595671943975e-05, |
| "loss": 0.933, |
| "step": 51000 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.430227798245323e-05, |
| "loss": 0.8895, |
| "step": 51500 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.4246960292962484e-05, |
| "loss": 0.9297, |
| "step": 52000 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 1.419164260347174e-05, |
| "loss": 0.9256, |
| "step": 52500 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 1.4136324913980994e-05, |
| "loss": 0.9088, |
| "step": 53000 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 1.4081007224490249e-05, |
| "loss": 0.9162, |
| "step": 53500 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 1.4025689534999503e-05, |
| "loss": 0.9249, |
| "step": 54000 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 1.3970371845508759e-05, |
| "loss": 0.9561, |
| "step": 54500 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 1.3915054156018013e-05, |
| "loss": 0.9376, |
| "step": 55000 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 1.3859736466527267e-05, |
| "loss": 0.9404, |
| "step": 55500 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 1.3804418777036522e-05, |
| "loss": 0.9654, |
| "step": 56000 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.3749101087545778e-05, |
| "loss": 0.8642, |
| "step": 56500 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.3693783398055032e-05, |
| "loss": 0.9651, |
| "step": 57000 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.3638465708564286e-05, |
| "loss": 0.928, |
| "step": 57500 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 1.358314801907354e-05, |
| "loss": 0.9215, |
| "step": 58000 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 1.3527830329582796e-05, |
| "loss": 0.8917, |
| "step": 58500 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.347251264009205e-05, |
| "loss": 0.8971, |
| "step": 59000 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 1.3417194950601305e-05, |
| "loss": 0.9151, |
| "step": 59500 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 1.336187726111056e-05, |
| "loss": 0.8999, |
| "step": 60000 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 1.3306559571619812e-05, |
| "loss": 0.9411, |
| "step": 60500 |
| }, |
| { |
| "epoch": 1.01, |
| "learning_rate": 1.325124188212907e-05, |
| "loss": 0.9059, |
| "step": 61000 |
| }, |
| { |
| "epoch": 1.02, |
| "learning_rate": 1.3195924192638324e-05, |
| "loss": 0.8741, |
| "step": 61500 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 1.3140606503147576e-05, |
| "loss": 0.93, |
| "step": 62000 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 1.308528881365683e-05, |
| "loss": 0.9157, |
| "step": 62500 |
| }, |
| { |
| "epoch": 1.05, |
| "learning_rate": 1.3029971124166088e-05, |
| "loss": 0.8834, |
| "step": 63000 |
| }, |
| { |
| "epoch": 1.05, |
| "learning_rate": 1.2974653434675341e-05, |
| "loss": 0.9047, |
| "step": 63500 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 1.2919335745184595e-05, |
| "loss": 0.9396, |
| "step": 64000 |
| }, |
| { |
| "epoch": 1.07, |
| "learning_rate": 1.286401805569385e-05, |
| "loss": 0.9448, |
| "step": 64500 |
| }, |
| { |
| "epoch": 1.08, |
| "learning_rate": 1.2808700366203106e-05, |
| "loss": 0.9332, |
| "step": 65000 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 1.275338267671236e-05, |
| "loss": 0.9394, |
| "step": 65500 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 1.2698064987221614e-05, |
| "loss": 0.9217, |
| "step": 66000 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 1.2642747297730868e-05, |
| "loss": 0.893, |
| "step": 66500 |
| }, |
| { |
| "epoch": 1.11, |
| "learning_rate": 1.2587429608240124e-05, |
| "loss": 0.8987, |
| "step": 67000 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 1.2532111918749379e-05, |
| "loss": 0.8831, |
| "step": 67500 |
| }, |
| { |
| "epoch": 1.13, |
| "learning_rate": 1.2476794229258633e-05, |
| "loss": 0.9192, |
| "step": 68000 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 1.2421476539767887e-05, |
| "loss": 0.9175, |
| "step": 68500 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 1.2366158850277143e-05, |
| "loss": 0.9207, |
| "step": 69000 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 1.2310841160786397e-05, |
| "loss": 0.8792, |
| "step": 69500 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 1.2255523471295652e-05, |
| "loss": 0.9052, |
| "step": 70000 |
| }, |
| { |
| "epoch": 1.17, |
| "learning_rate": 1.2200205781804906e-05, |
| "loss": 0.9362, |
| "step": 70500 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 1.2144888092314162e-05, |
| "loss": 0.8689, |
| "step": 71000 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 1.2089570402823416e-05, |
| "loss": 0.8964, |
| "step": 71500 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 1.203425271333267e-05, |
| "loss": 0.9057, |
| "step": 72000 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 1.1978935023841925e-05, |
| "loss": 0.9527, |
| "step": 72500 |
| }, |
| { |
| "epoch": 1.21, |
| "learning_rate": 1.192361733435118e-05, |
| "loss": 0.9061, |
| "step": 73000 |
| }, |
| { |
| "epoch": 1.22, |
| "learning_rate": 1.1868299644860435e-05, |
| "loss": 0.9662, |
| "step": 73500 |
| }, |
| { |
| "epoch": 1.23, |
| "learning_rate": 1.181298195536969e-05, |
| "loss": 0.9286, |
| "step": 74000 |
| }, |
| { |
| "epoch": 1.24, |
| "learning_rate": 1.1757664265878944e-05, |
| "loss": 0.953, |
| "step": 74500 |
| }, |
| { |
| "epoch": 1.24, |
| "learning_rate": 1.17023465763882e-05, |
| "loss": 0.9109, |
| "step": 75000 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 1.1647028886897454e-05, |
| "loss": 0.9167, |
| "step": 75500 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 1.1591711197406708e-05, |
| "loss": 0.9382, |
| "step": 76000 |
| }, |
| { |
| "epoch": 1.27, |
| "learning_rate": 1.1536393507915962e-05, |
| "loss": 0.9144, |
| "step": 76500 |
| }, |
| { |
| "epoch": 1.28, |
| "learning_rate": 1.1481075818425218e-05, |
| "loss": 0.8935, |
| "step": 77000 |
| }, |
| { |
| "epoch": 1.29, |
| "learning_rate": 1.1425758128934473e-05, |
| "loss": 0.9181, |
| "step": 77500 |
| }, |
| { |
| "epoch": 1.29, |
| "learning_rate": 1.1370440439443727e-05, |
| "loss": 0.9167, |
| "step": 78000 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 1.131512274995298e-05, |
| "loss": 0.8912, |
| "step": 78500 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 1.1259805060462234e-05, |
| "loss": 0.9078, |
| "step": 79000 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 1.1204487370971491e-05, |
| "loss": 0.9164, |
| "step": 79500 |
| }, |
| { |
| "epoch": 1.33, |
| "learning_rate": 1.1149169681480744e-05, |
| "loss": 0.9178, |
| "step": 80000 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 1.1093851991989998e-05, |
| "loss": 0.8802, |
| "step": 80500 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 1.1038534302499253e-05, |
| "loss": 0.8861, |
| "step": 81000 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 1.0983216613008509e-05, |
| "loss": 0.9514, |
| "step": 81500 |
| }, |
| { |
| "epoch": 1.36, |
| "learning_rate": 1.0927898923517763e-05, |
| "loss": 0.9349, |
| "step": 82000 |
| }, |
| { |
| "epoch": 1.37, |
| "learning_rate": 1.0872581234027017e-05, |
| "loss": 0.8864, |
| "step": 82500 |
| }, |
| { |
| "epoch": 1.38, |
| "learning_rate": 1.0817263544536271e-05, |
| "loss": 0.9031, |
| "step": 83000 |
| }, |
| { |
| "epoch": 1.39, |
| "learning_rate": 1.0761945855045527e-05, |
| "loss": 0.9202, |
| "step": 83500 |
| }, |
| { |
| "epoch": 1.39, |
| "learning_rate": 1.0706628165554782e-05, |
| "loss": 0.964, |
| "step": 84000 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 1.0651310476064036e-05, |
| "loss": 0.8934, |
| "step": 84500 |
| }, |
| { |
| "epoch": 1.41, |
| "learning_rate": 1.059599278657329e-05, |
| "loss": 0.8834, |
| "step": 85000 |
| }, |
| { |
| "epoch": 1.42, |
| "learning_rate": 1.0540675097082546e-05, |
| "loss": 0.9234, |
| "step": 85500 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 1.04853574075918e-05, |
| "loss": 0.9243, |
| "step": 86000 |
| }, |
| { |
| "epoch": 1.44, |
| "learning_rate": 1.0430039718101055e-05, |
| "loss": 0.9032, |
| "step": 86500 |
| }, |
| { |
| "epoch": 1.44, |
| "learning_rate": 1.0374722028610309e-05, |
| "loss": 0.8856, |
| "step": 87000 |
| }, |
| { |
| "epoch": 1.45, |
| "learning_rate": 1.0319404339119565e-05, |
| "loss": 0.918, |
| "step": 87500 |
| }, |
| { |
| "epoch": 1.46, |
| "learning_rate": 1.026408664962882e-05, |
| "loss": 0.9737, |
| "step": 88000 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 1.0208768960138074e-05, |
| "loss": 0.8813, |
| "step": 88500 |
| }, |
| { |
| "epoch": 1.48, |
| "learning_rate": 1.0153451270647328e-05, |
| "loss": 0.8957, |
| "step": 89000 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 1.0098133581156584e-05, |
| "loss": 0.9095, |
| "step": 89500 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 1.0042815891665838e-05, |
| "loss": 0.9336, |
| "step": 90000 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 9.987498202175092e-06, |
| "loss": 0.8921, |
| "step": 90500 |
| }, |
| { |
| "epoch": 1.51, |
| "learning_rate": 9.932180512684347e-06, |
| "loss": 0.8593, |
| "step": 91000 |
| }, |
| { |
| "epoch": 1.52, |
| "learning_rate": 9.876862823193601e-06, |
| "loss": 0.9258, |
| "step": 91500 |
| }, |
| { |
| "epoch": 1.53, |
| "learning_rate": 9.821545133702857e-06, |
| "loss": 0.927, |
| "step": 92000 |
| }, |
| { |
| "epoch": 1.54, |
| "learning_rate": 9.766227444212111e-06, |
| "loss": 0.9453, |
| "step": 92500 |
| }, |
| { |
| "epoch": 1.54, |
| "learning_rate": 9.710909754721366e-06, |
| "loss": 0.9271, |
| "step": 93000 |
| }, |
| { |
| "epoch": 1.55, |
| "learning_rate": 9.65559206523062e-06, |
| "loss": 0.9393, |
| "step": 93500 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 9.600274375739876e-06, |
| "loss": 0.9376, |
| "step": 94000 |
| }, |
| { |
| "epoch": 1.57, |
| "learning_rate": 9.54495668624913e-06, |
| "loss": 0.929, |
| "step": 94500 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 9.489638996758384e-06, |
| "loss": 0.9438, |
| "step": 95000 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 9.434321307267639e-06, |
| "loss": 0.9369, |
| "step": 95500 |
| }, |
| { |
| "epoch": 1.59, |
| "learning_rate": 9.379003617776895e-06, |
| "loss": 0.9364, |
| "step": 96000 |
| }, |
| { |
| "epoch": 1.6, |
| "learning_rate": 9.323685928286149e-06, |
| "loss": 0.9375, |
| "step": 96500 |
| }, |
| { |
| "epoch": 1.61, |
| "learning_rate": 9.268368238795403e-06, |
| "loss": 0.8816, |
| "step": 97000 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 9.213050549304657e-06, |
| "loss": 0.916, |
| "step": 97500 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 9.157732859813913e-06, |
| "loss": 0.994, |
| "step": 98000 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 9.102415170323166e-06, |
| "loss": 0.9202, |
| "step": 98500 |
| }, |
| { |
| "epoch": 1.64, |
| "learning_rate": 9.04709748083242e-06, |
| "loss": 0.9044, |
| "step": 99000 |
| }, |
| { |
| "epoch": 1.65, |
| "learning_rate": 8.991779791341676e-06, |
| "loss": 0.8644, |
| "step": 99500 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 8.93646210185093e-06, |
| "loss": 0.8926, |
| "step": 100000 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 8.881144412360185e-06, |
| "loss": 0.9647, |
| "step": 100500 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 8.825826722869439e-06, |
| "loss": 0.9277, |
| "step": 101000 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 8.770509033378695e-06, |
| "loss": 0.8776, |
| "step": 101500 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 8.71519134388795e-06, |
| "loss": 0.8863, |
| "step": 102000 |
| }, |
| { |
| "epoch": 1.7, |
| "learning_rate": 8.659873654397204e-06, |
| "loss": 0.9286, |
| "step": 102500 |
| }, |
| { |
| "epoch": 1.71, |
| "learning_rate": 8.604555964906458e-06, |
| "loss": 0.9329, |
| "step": 103000 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 8.549238275415714e-06, |
| "loss": 0.9091, |
| "step": 103500 |
| }, |
| { |
| "epoch": 1.73, |
| "learning_rate": 8.493920585924968e-06, |
| "loss": 0.9215, |
| "step": 104000 |
| }, |
| { |
| "epoch": 1.73, |
| "learning_rate": 8.438602896434222e-06, |
| "loss": 0.9013, |
| "step": 104500 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 8.383285206943477e-06, |
| "loss": 0.9064, |
| "step": 105000 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 8.327967517452733e-06, |
| "loss": 0.8798, |
| "step": 105500 |
| }, |
| { |
| "epoch": 1.76, |
| "learning_rate": 8.272649827961985e-06, |
| "loss": 0.9142, |
| "step": 106000 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 8.217332138471241e-06, |
| "loss": 0.9415, |
| "step": 106500 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 8.162014448980496e-06, |
| "loss": 0.8729, |
| "step": 107000 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 8.10669675948975e-06, |
| "loss": 0.9351, |
| "step": 107500 |
| }, |
| { |
| "epoch": 1.79, |
| "learning_rate": 8.051379069999004e-06, |
| "loss": 0.9186, |
| "step": 108000 |
| }, |
| { |
| "epoch": 1.8, |
| "learning_rate": 7.99606138050826e-06, |
| "loss": 0.9713, |
| "step": 108500 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 7.940743691017514e-06, |
| "loss": 0.9179, |
| "step": 109000 |
| }, |
| { |
| "epoch": 1.82, |
| "learning_rate": 7.885426001526769e-06, |
| "loss": 0.9631, |
| "step": 109500 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 7.830108312036023e-06, |
| "loss": 0.891, |
| "step": 110000 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 7.774790622545279e-06, |
| "loss": 0.9189, |
| "step": 110500 |
| }, |
| { |
| "epoch": 1.84, |
| "learning_rate": 7.719472933054533e-06, |
| "loss": 0.8631, |
| "step": 111000 |
| }, |
| { |
| "epoch": 1.85, |
| "learning_rate": 7.664155243563787e-06, |
| "loss": 0.9599, |
| "step": 111500 |
| }, |
| { |
| "epoch": 1.86, |
| "learning_rate": 7.608837554073042e-06, |
| "loss": 0.9375, |
| "step": 112000 |
| }, |
| { |
| "epoch": 1.87, |
| "learning_rate": 7.553519864582297e-06, |
| "loss": 0.921, |
| "step": 112500 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 7.498202175091551e-06, |
| "loss": 0.8958, |
| "step": 113000 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 7.442884485600806e-06, |
| "loss": 0.9505, |
| "step": 113500 |
| }, |
| { |
| "epoch": 1.89, |
| "learning_rate": 7.3875667961100605e-06, |
| "loss": 0.9405, |
| "step": 114000 |
| }, |
| { |
| "epoch": 1.9, |
| "learning_rate": 7.332249106619316e-06, |
| "loss": 0.8996, |
| "step": 114500 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 7.27693141712857e-06, |
| "loss": 0.8697, |
| "step": 115000 |
| }, |
| { |
| "epoch": 1.92, |
| "learning_rate": 7.221613727637825e-06, |
| "loss": 0.9324, |
| "step": 115500 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 7.166296038147079e-06, |
| "loss": 0.8865, |
| "step": 116000 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 7.1109783486563345e-06, |
| "loss": 0.9143, |
| "step": 116500 |
| }, |
| { |
| "epoch": 1.94, |
| "learning_rate": 7.055660659165589e-06, |
| "loss": 0.8974, |
| "step": 117000 |
| }, |
| { |
| "epoch": 1.95, |
| "learning_rate": 7.000342969674844e-06, |
| "loss": 0.926, |
| "step": 117500 |
| }, |
| { |
| "epoch": 1.96, |
| "learning_rate": 6.945025280184098e-06, |
| "loss": 0.9321, |
| "step": 118000 |
| }, |
| { |
| "epoch": 1.97, |
| "learning_rate": 6.889707590693352e-06, |
| "loss": 0.9175, |
| "step": 118500 |
| }, |
| { |
| "epoch": 1.97, |
| "learning_rate": 6.834389901202607e-06, |
| "loss": 0.9066, |
| "step": 119000 |
| }, |
| { |
| "epoch": 1.98, |
| "learning_rate": 6.779072211711861e-06, |
| "loss": 0.9149, |
| "step": 119500 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 6.723754522221116e-06, |
| "loss": 0.9249, |
| "step": 120000 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 6.66843683273037e-06, |
| "loss": 0.8616, |
| "step": 120500 |
| } |
| ], |
| "logging_steps": 500, |
| "max_steps": 180774, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 500, |
| "total_flos": 1.9790822990016e+17, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|