| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.0, |
| "eval_steps": 500, |
| "global_step": 750, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.013349463935588836, |
| "grad_norm": 3.8443384275181525, |
| "learning_rate": 6.000000000000001e-07, |
| "loss": 0.6301, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.026698927871177672, |
| "grad_norm": 1.1304019443903548, |
| "learning_rate": 1.2666666666666669e-06, |
| "loss": 0.6325, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.04004839180676651, |
| "grad_norm": 0.7924252148112436, |
| "learning_rate": 1.9333333333333336e-06, |
| "loss": 0.6061, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.053397855742355345, |
| "grad_norm": 0.7615554902176207, |
| "learning_rate": 2.6e-06, |
| "loss": 0.6039, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.06674731967794419, |
| "grad_norm": 0.6884584126571622, |
| "learning_rate": 3.266666666666667e-06, |
| "loss": 0.6035, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.08009678361353302, |
| "grad_norm": 0.6895907181910402, |
| "learning_rate": 3.9333333333333335e-06, |
| "loss": 0.6177, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.09344624754912186, |
| "grad_norm": 0.6789285267200114, |
| "learning_rate": 4.600000000000001e-06, |
| "loss": 0.603, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.10679571148471069, |
| "grad_norm": 0.6467096461947729, |
| "learning_rate": 4.99956677884892e-06, |
| "loss": 0.6118, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.12014517542029952, |
| "grad_norm": 0.6168635819855449, |
| "learning_rate": 4.994694765050121e-06, |
| "loss": 0.6116, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.13349463935588837, |
| "grad_norm": 0.5441117690789925, |
| "learning_rate": 4.984419797901491e-06, |
| "loss": 0.6046, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.1468441032914772, |
| "grad_norm": 0.5172864204707323, |
| "learning_rate": 4.9687641306938766e-06, |
| "loss": 0.5961, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.16019356722706604, |
| "grad_norm": 0.5120924898252496, |
| "learning_rate": 4.94776167011629e-06, |
| "loss": 0.5891, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.17354303116265488, |
| "grad_norm": 0.4936858545975663, |
| "learning_rate": 4.921457902821578e-06, |
| "loss": 0.5965, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.1868924950982437, |
| "grad_norm": 0.5096637378296062, |
| "learning_rate": 4.889909796912454e-06, |
| "loss": 0.5876, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.20024195903383255, |
| "grad_norm": 0.4893103328332192, |
| "learning_rate": 4.85318567856128e-06, |
| "loss": 0.5866, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.21359142296942138, |
| "grad_norm": 0.48634183325283653, |
| "learning_rate": 4.811365084030784e-06, |
| "loss": 0.5917, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.22694088690501021, |
| "grad_norm": 0.5077880866108633, |
| "learning_rate": 4.764538587416233e-06, |
| "loss": 0.5972, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.24029035084059905, |
| "grad_norm": 0.5531086916036718, |
| "learning_rate": 4.712807604482108e-06, |
| "loss": 0.5692, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.2536398147761879, |
| "grad_norm": 0.4815973346852316, |
| "learning_rate": 4.656284173018144e-06, |
| "loss": 0.5878, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.26698927871177675, |
| "grad_norm": 0.4477210102882603, |
| "learning_rate": 4.595090710190419e-06, |
| "loss": 0.606, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.2803387426473656, |
| "grad_norm": 0.4442300924031833, |
| "learning_rate": 4.529359747413038e-06, |
| "loss": 0.5921, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.2936882065829544, |
| "grad_norm": 0.4318914186907017, |
| "learning_rate": 4.4592336433146e-06, |
| "loss": 0.5954, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.30703767051854325, |
| "grad_norm": 0.4822511296612312, |
| "learning_rate": 4.384864275421109e-06, |
| "loss": 0.6013, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.3203871344541321, |
| "grad_norm": 0.44291005932494015, |
| "learning_rate": 4.30641271122307e-06, |
| "loss": 0.6083, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.3337365983897209, |
| "grad_norm": 0.48203103136482495, |
| "learning_rate": 4.224048859339175e-06, |
| "loss": 0.6216, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.34708606232530975, |
| "grad_norm": 0.45150578106097233, |
| "learning_rate": 4.1379511015320625e-06, |
| "loss": 0.6042, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.3604355262608986, |
| "grad_norm": 0.4551136191346761, |
| "learning_rate": 4.048305906373151e-06, |
| "loss": 0.6163, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.3737849901964874, |
| "grad_norm": 0.46692000639310954, |
| "learning_rate": 3.955307425393224e-06, |
| "loss": 0.576, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.38713445413207626, |
| "grad_norm": 0.43289733012383014, |
| "learning_rate": 3.859157072593459e-06, |
| "loss": 0.5905, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.4004839180676651, |
| "grad_norm": 0.47300144920672366, |
| "learning_rate": 3.760063088227542e-06, |
| "loss": 0.6046, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.4138333820032539, |
| "grad_norm": 0.43427902038611416, |
| "learning_rate": 3.658240087799655e-06, |
| "loss": 0.5795, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.42718284593884276, |
| "grad_norm": 0.4321634460614613, |
| "learning_rate": 3.5539085972550786e-06, |
| "loss": 0.6001, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.4405323098744316, |
| "grad_norm": 0.4751769823117227, |
| "learning_rate": 3.4472945753701038e-06, |
| "loss": 0.5969, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.45388177381002043, |
| "grad_norm": 0.4523747463622145, |
| "learning_rate": 3.338628924375638e-06, |
| "loss": 0.5809, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.46723123774560926, |
| "grad_norm": 0.4571186296564021, |
| "learning_rate": 3.228146989874389e-06, |
| "loss": 0.5884, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.4805807016811981, |
| "grad_norm": 0.48390017792164, |
| "learning_rate": 3.116088051134695e-06, |
| "loss": 0.5854, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.49393016561678693, |
| "grad_norm": 0.4656758402293932, |
| "learning_rate": 3.002694802864912e-06, |
| "loss": 0.6115, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.5072796295523758, |
| "grad_norm": 0.4624954532839225, |
| "learning_rate": 2.888212829590719e-06, |
| "loss": 0.6079, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.5206290934879646, |
| "grad_norm": 0.436497560779719, |
| "learning_rate": 2.77289007377372e-06, |
| "loss": 0.5787, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.5339785574235535, |
| "grad_norm": 0.46334201651691187, |
| "learning_rate": 2.6569762988232838e-06, |
| "loss": 0.6069, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.5473280213591423, |
| "grad_norm": 0.4806566661003276, |
| "learning_rate": 2.5407225481646146e-06, |
| "loss": 0.575, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.5606774852947312, |
| "grad_norm": 0.460214567270956, |
| "learning_rate": 2.4243806015345988e-06, |
| "loss": 0.5858, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.5740269492303199, |
| "grad_norm": 0.42739545970036125, |
| "learning_rate": 2.3082024296829538e-06, |
| "loss": 0.5699, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.5873764131659088, |
| "grad_norm": 0.4503645884785317, |
| "learning_rate": 2.192439648659699e-06, |
| "loss": 0.5823, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.6007258771014976, |
| "grad_norm": 0.4776520102412501, |
| "learning_rate": 2.0773429748708153e-06, |
| "loss": 0.5974, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.6140753410370865, |
| "grad_norm": 0.4600206541871604, |
| "learning_rate": 1.963161682082342e-06, |
| "loss": 0.5905, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.6274248049726753, |
| "grad_norm": 0.42642177736501347, |
| "learning_rate": 1.850143061548907e-06, |
| "loss": 0.5841, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.6407742689082642, |
| "grad_norm": 0.4374872861905679, |
| "learning_rate": 1.7385318864359304e-06, |
| "loss": 0.5883, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.654123732843853, |
| "grad_norm": 0.4539731186521712, |
| "learning_rate": 1.6285698816954626e-06, |
| "loss": 0.5981, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.6674731967794418, |
| "grad_norm": 0.49549560937788556, |
| "learning_rate": 1.520495200543754e-06, |
| "loss": 0.594, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.6808226607150306, |
| "grad_norm": 0.442220991522498, |
| "learning_rate": 1.41454190867443e-06, |
| "loss": 0.5978, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.6941721246506195, |
| "grad_norm": 0.44323894790772966, |
| "learning_rate": 1.3109394773243117e-06, |
| "loss": 0.5867, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.7075215885862083, |
| "grad_norm": 0.4233660170980954, |
| "learning_rate": 1.2099122862898214e-06, |
| "loss": 0.6018, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.7208710525217972, |
| "grad_norm": 0.43218789543451824, |
| "learning_rate": 1.1116791379703032e-06, |
| "loss": 0.6013, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.7342205164573861, |
| "grad_norm": 0.44159730183002366, |
| "learning_rate": 1.0164527834907468e-06, |
| "loss": 0.591, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.7475699803929748, |
| "grad_norm": 0.43870818069759426, |
| "learning_rate": 9.244394619302338e-07, |
| "loss": 0.5691, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.7609194443285637, |
| "grad_norm": 0.4477385893330783, |
| "learning_rate": 8.35838453654009e-07, |
| "loss": 0.6041, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.7742689082641525, |
| "grad_norm": 0.43072214351782917, |
| "learning_rate": 7.508416487165862e-07, |
| "loss": 0.5886, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.7876183721997414, |
| "grad_norm": 0.4279974728470172, |
| "learning_rate": 6.696331312706245e-07, |
| "loss": 0.5754, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.8009678361353302, |
| "grad_norm": 0.3917096350694494, |
| "learning_rate": 5.923887808816373e-07, |
| "loss": 0.5896, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.8143173000709191, |
| "grad_norm": 0.4434938485219406, |
| "learning_rate": 5.192758916120236e-07, |
| "loss": 0.5831, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.8276667640065078, |
| "grad_norm": 0.41679775408713765, |
| "learning_rate": 4.5045280969937847e-07, |
| "loss": 0.5937, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.8410162279420967, |
| "grad_norm": 0.4268405005566493, |
| "learning_rate": 3.86068590613804e-07, |
| "loss": 0.5933, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.8543656918776855, |
| "grad_norm": 0.4166184793700984, |
| "learning_rate": 3.262626762369525e-07, |
| "loss": 0.5993, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.8677151558132744, |
| "grad_norm": 0.41797554779045376, |
| "learning_rate": 2.7116459286195887e-07, |
| "loss": 0.6012, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.8810646197488632, |
| "grad_norm": 0.4255310425370243, |
| "learning_rate": 2.208936706683351e-07, |
| "loss": 0.5856, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.8944140836844521, |
| "grad_norm": 0.42943616856150774, |
| "learning_rate": 1.7555878527937164e-07, |
| "loss": 0.5929, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.9077635476200409, |
| "grad_norm": 0.41759560647733296, |
| "learning_rate": 1.352581219617824e-07, |
| "loss": 0.578, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.9211130115556297, |
| "grad_norm": 0.41217774020393527, |
| "learning_rate": 1.0007896297828113e-07, |
| "loss": 0.5815, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.9344624754912185, |
| "grad_norm": 0.4305528132965425, |
| "learning_rate": 7.009749855363457e-08, |
| "loss": 0.5984, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.9478119394268074, |
| "grad_norm": 0.4237537569089471, |
| "learning_rate": 4.537866186360207e-08, |
| "loss": 0.5797, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.9611614033623962, |
| "grad_norm": 0.4354601270782283, |
| "learning_rate": 2.59759884041369e-08, |
| "loss": 0.5822, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.9745108672979851, |
| "grad_norm": 0.4236057022989506, |
| "learning_rate": 1.193150004542204e-08, |
| "loss": 0.6017, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.9878603312335739, |
| "grad_norm": 0.4253092913652658, |
| "learning_rate": 3.275614021857609e-09, |
| "loss": 0.5792, |
| "step": 740 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.4153479478058976, |
| "learning_rate": 2.7077055103075233e-11, |
| "loss": 0.5653, |
| "step": 750 |
| }, |
| { |
| "epoch": 1.0, |
| "step": 750, |
| "total_flos": 9.514760021422899e+17, |
| "train_loss": 0.5942036323547363, |
| "train_runtime": 26089.1141, |
| "train_samples_per_second": 14.701, |
| "train_steps_per_second": 0.029 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 750, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 250, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 9.514760021422899e+17, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|