| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 29864, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.01674256630056255, | |
| "grad_norm": 0.26318585872650146, | |
| "learning_rate": 1.6538332775359894e-06, | |
| "loss": 0.1468, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.0334851326011251, | |
| "grad_norm": 40.848941802978516, | |
| "learning_rate": 3.3244057582859056e-06, | |
| "loss": 0.1348, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.05022769890168765, | |
| "grad_norm": 0.3210512399673462, | |
| "learning_rate": 4.9983260796786076e-06, | |
| "loss": 0.1292, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.0669702652022502, | |
| "grad_norm": 51.398277282714844, | |
| "learning_rate": 6.6722464010713095e-06, | |
| "loss": 0.1246, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.08371283150281275, | |
| "grad_norm": 1.8654736280441284, | |
| "learning_rate": 8.346166722464011e-06, | |
| "loss": 0.073, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.1004553978033753, | |
| "grad_norm": 2.204322338104248, | |
| "learning_rate": 9.997767607991965e-06, | |
| "loss": 0.0861, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.11719796410393785, | |
| "grad_norm": 0.029873741790652275, | |
| "learning_rate": 9.81173494065558e-06, | |
| "loss": 0.0626, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.1339405304045004, | |
| "grad_norm": 2.218162775039673, | |
| "learning_rate": 9.626074338653869e-06, | |
| "loss": 0.0731, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.15068309670506294, | |
| "grad_norm": 6.862820625305176, | |
| "learning_rate": 9.440041671317484e-06, | |
| "loss": 0.0756, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.1674256630056255, | |
| "grad_norm": 1.0643534660339355, | |
| "learning_rate": 9.2540090039811e-06, | |
| "loss": 0.0702, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.18416822930618806, | |
| "grad_norm": 0.14988207817077637, | |
| "learning_rate": 9.067976336644717e-06, | |
| "loss": 0.0675, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.2009107956067506, | |
| "grad_norm": 0.25116392970085144, | |
| "learning_rate": 8.882315734643004e-06, | |
| "loss": 0.0604, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.21765336190731316, | |
| "grad_norm": 0.3021365702152252, | |
| "learning_rate": 8.69628306730662e-06, | |
| "loss": 0.0515, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.2343959282078757, | |
| "grad_norm": 22.51474952697754, | |
| "learning_rate": 8.510250399970235e-06, | |
| "loss": 0.0643, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.2511384945084382, | |
| "grad_norm": 2.37544584274292, | |
| "learning_rate": 8.324217732633852e-06, | |
| "loss": 0.07, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.2678810608090008, | |
| "grad_norm": 57.23020553588867, | |
| "learning_rate": 8.138185065297467e-06, | |
| "loss": 0.0557, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.28462362710956335, | |
| "grad_norm": 0.0032292637042701244, | |
| "learning_rate": 7.952152397961083e-06, | |
| "loss": 0.0614, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.3013661934101259, | |
| "grad_norm": 5.582527160644531, | |
| "learning_rate": 7.766119730624698e-06, | |
| "loss": 0.0577, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.31810875971068847, | |
| "grad_norm": 0.04729452729225159, | |
| "learning_rate": 7.580087063288314e-06, | |
| "loss": 0.0582, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.334851326011251, | |
| "grad_norm": 2.683903217315674, | |
| "learning_rate": 7.394426461286602e-06, | |
| "loss": 0.0549, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.35159389231181354, | |
| "grad_norm": 0.9779844880104065, | |
| "learning_rate": 7.208393793950218e-06, | |
| "loss": 0.0615, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.3683364586123761, | |
| "grad_norm": 79.76528930664062, | |
| "learning_rate": 7.0223611266138335e-06, | |
| "loss": 0.0519, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.38507902491293866, | |
| "grad_norm": 0.3208529055118561, | |
| "learning_rate": 6.836328459277449e-06, | |
| "loss": 0.0591, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.4018215912135012, | |
| "grad_norm": 23.068653106689453, | |
| "learning_rate": 6.650295791941065e-06, | |
| "loss": 0.0539, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.41856415751406373, | |
| "grad_norm": 0.6496044993400574, | |
| "learning_rate": 6.464635189939353e-06, | |
| "loss": 0.0549, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.4353067238146263, | |
| "grad_norm": 38.81090545654297, | |
| "learning_rate": 6.2786025226029705e-06, | |
| "loss": 0.0543, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.45204929011518885, | |
| "grad_norm": 2.4993317127227783, | |
| "learning_rate": 6.092569855266585e-06, | |
| "loss": 0.0614, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.4687918564157514, | |
| "grad_norm": 0.5560814738273621, | |
| "learning_rate": 5.906537187930201e-06, | |
| "loss": 0.048, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.485534422716314, | |
| "grad_norm": 1.5382946729660034, | |
| "learning_rate": 5.720504520593817e-06, | |
| "loss": 0.0558, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.5022769890168765, | |
| "grad_norm": 0.026120496913790703, | |
| "learning_rate": 5.534471853257432e-06, | |
| "loss": 0.0492, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.519019555317439, | |
| "grad_norm": 1.6323872804641724, | |
| "learning_rate": 5.348439185921048e-06, | |
| "loss": 0.0633, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.5357621216180016, | |
| "grad_norm": 0.012650725431740284, | |
| "learning_rate": 5.162406518584664e-06, | |
| "loss": 0.0447, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.5525046879185641, | |
| "grad_norm": 8.42776107788086, | |
| "learning_rate": 4.976745916582952e-06, | |
| "loss": 0.043, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.5692472542191267, | |
| "grad_norm": 2.717674732208252, | |
| "learning_rate": 4.790713249246568e-06, | |
| "loss": 0.0517, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.5859898205196893, | |
| "grad_norm": 0.01320638321340084, | |
| "learning_rate": 4.604680581910183e-06, | |
| "loss": 0.0768, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.6027323868202518, | |
| "grad_norm": 68.84772491455078, | |
| "learning_rate": 4.418647914573799e-06, | |
| "loss": 0.0578, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.6194749531208144, | |
| "grad_norm": 0.05758577585220337, | |
| "learning_rate": 4.232987312572088e-06, | |
| "loss": 0.0575, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.6362175194213769, | |
| "grad_norm": 0.02908864989876747, | |
| "learning_rate": 4.046954645235704e-06, | |
| "loss": 0.0527, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.6529600857219394, | |
| "grad_norm": 5.406867027282715, | |
| "learning_rate": 3.86092197789932e-06, | |
| "loss": 0.0556, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.669702652022502, | |
| "grad_norm": 0.1083500012755394, | |
| "learning_rate": 3.6748893105629353e-06, | |
| "loss": 0.0607, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.6864452183230646, | |
| "grad_norm": 13.136815071105957, | |
| "learning_rate": 3.488856643226551e-06, | |
| "loss": 0.0608, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.7031877846236271, | |
| "grad_norm": 0.7505860328674316, | |
| "learning_rate": 3.3028239758901666e-06, | |
| "loss": 0.0691, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.7199303509241897, | |
| "grad_norm": 56.34339141845703, | |
| "learning_rate": 3.1167913085537823e-06, | |
| "loss": 0.0563, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 0.7366729172247523, | |
| "grad_norm": 0.267652302980423, | |
| "learning_rate": 2.9307586412173984e-06, | |
| "loss": 0.0688, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.7534154835253147, | |
| "grad_norm": 4.030105113983154, | |
| "learning_rate": 2.7450980392156867e-06, | |
| "loss": 0.0637, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 0.7701580498258773, | |
| "grad_norm": 1.6611367464065552, | |
| "learning_rate": 2.5590653718793024e-06, | |
| "loss": 0.0724, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.7869006161264399, | |
| "grad_norm": 65.19082641601562, | |
| "learning_rate": 2.3734047698775908e-06, | |
| "loss": 0.0759, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 0.8036431824270024, | |
| "grad_norm": 2.824739694595337, | |
| "learning_rate": 2.1873721025412064e-06, | |
| "loss": 0.0675, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.820385748727565, | |
| "grad_norm": 1.6871856451034546, | |
| "learning_rate": 2.001339435204822e-06, | |
| "loss": 0.1016, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 0.8371283150281275, | |
| "grad_norm": 73.87975311279297, | |
| "learning_rate": 1.8156788332031105e-06, | |
| "loss": 0.0995, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 0.85387088132869, | |
| "grad_norm": 1.2010084390640259, | |
| "learning_rate": 1.6296461658667261e-06, | |
| "loss": 0.0871, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 0.8706134476292526, | |
| "grad_norm": 8.91370964050293, | |
| "learning_rate": 1.443613498530342e-06, | |
| "loss": 0.0936, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 0.8873560139298151, | |
| "grad_norm": 8.569842338562012, | |
| "learning_rate": 1.2575808311939577e-06, | |
| "loss": 0.0927, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 0.9040985802303777, | |
| "grad_norm": 0.14943744242191315, | |
| "learning_rate": 1.0715481638575736e-06, | |
| "loss": 0.1108, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 0.9208411465309403, | |
| "grad_norm": 48.06157684326172, | |
| "learning_rate": 8.855154965211891e-07, | |
| "loss": 0.1141, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 0.9375837128315028, | |
| "grad_norm": 46.07848358154297, | |
| "learning_rate": 6.994828291848049e-07, | |
| "loss": 0.118, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 0.9543262791320654, | |
| "grad_norm": 1.0916130542755127, | |
| "learning_rate": 5.134501618484206e-07, | |
| "loss": 0.138, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 0.971068845432628, | |
| "grad_norm": 0.2534860074520111, | |
| "learning_rate": 3.277895598467091e-07, | |
| "loss": 0.1425, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 0.9878114117331904, | |
| "grad_norm": 0.08512990921735764, | |
| "learning_rate": 1.421289578449976e-07, | |
| "loss": 0.1432, | |
| "step": 29500 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 29864, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |