| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 11, | |
| "global_step": 101, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.009900990099009901, | |
| "grad_norm": 1.9865820407867432, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8807, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.019801980198019802, | |
| "grad_norm": 1.6995460987091064, | |
| "learning_rate": 1e-05, | |
| "loss": 0.7298, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.0297029702970297, | |
| "grad_norm": 1.870273232460022, | |
| "learning_rate": 9.997482711915926e-06, | |
| "loss": 0.8256, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.039603960396039604, | |
| "grad_norm": 2.035698175430298, | |
| "learning_rate": 9.989933382359423e-06, | |
| "loss": 0.8423, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.04950495049504951, | |
| "grad_norm": 1.5802401304244995, | |
| "learning_rate": 9.977359612865424e-06, | |
| "loss": 1.0321, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0594059405940594, | |
| "grad_norm": 1.3238128423690796, | |
| "learning_rate": 9.959774064153977e-06, | |
| "loss": 0.7894, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.06930693069306931, | |
| "grad_norm": 1.3449586629867554, | |
| "learning_rate": 9.937194443381972e-06, | |
| "loss": 0.69, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.07920792079207921, | |
| "grad_norm": 1.215697169303894, | |
| "learning_rate": 9.909643486313533e-06, | |
| "loss": 0.8263, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.0891089108910891, | |
| "grad_norm": 1.4947011470794678, | |
| "learning_rate": 9.877148934427037e-06, | |
| "loss": 0.7706, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.09900990099009901, | |
| "grad_norm": 0.9939355850219727, | |
| "learning_rate": 9.839743506981783e-06, | |
| "loss": 0.6854, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.10891089108910891, | |
| "grad_norm": 1.3050940036773682, | |
| "learning_rate": 9.797464868072489e-06, | |
| "loss": 0.9061, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.10891089108910891, | |
| "eval_loss": 0.7099791765213013, | |
| "eval_runtime": 1.6876, | |
| "eval_samples_per_second": 5.333, | |
| "eval_steps_per_second": 1.185, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.1188118811881188, | |
| "grad_norm": 1.0027287006378174, | |
| "learning_rate": 9.750355588704728e-06, | |
| "loss": 0.7043, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.12871287128712872, | |
| "grad_norm": 0.9371753334999084, | |
| "learning_rate": 9.698463103929542e-06, | |
| "loss": 0.6727, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.13861386138613863, | |
| "grad_norm": 0.9471062421798706, | |
| "learning_rate": 9.641839665080363e-06, | |
| "loss": 0.6772, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.1485148514851485, | |
| "grad_norm": 1.1542134284973145, | |
| "learning_rate": 9.580542287160348e-06, | |
| "loss": 0.7956, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.15841584158415842, | |
| "grad_norm": 1.1275238990783691, | |
| "learning_rate": 9.514632691433108e-06, | |
| "loss": 0.7486, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.16831683168316833, | |
| "grad_norm": 1.0453617572784424, | |
| "learning_rate": 9.444177243274619e-06, | |
| "loss": 0.659, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.1782178217821782, | |
| "grad_norm": 1.0331149101257324, | |
| "learning_rate": 9.369246885348926e-06, | |
| "loss": 0.7471, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.18811881188118812, | |
| "grad_norm": 0.8706634044647217, | |
| "learning_rate": 9.289917066174887e-06, | |
| "loss": 0.5547, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.19801980198019803, | |
| "grad_norm": 1.1781532764434814, | |
| "learning_rate": 9.206267664155906e-06, | |
| "loss": 0.6239, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.2079207920792079, | |
| "grad_norm": 1.2011443376541138, | |
| "learning_rate": 9.118382907149164e-06, | |
| "loss": 0.7847, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.21782178217821782, | |
| "grad_norm": 1.028536319732666, | |
| "learning_rate": 9.026351287655294e-06, | |
| "loss": 0.6565, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.21782178217821782, | |
| "eval_loss": 0.6684643030166626, | |
| "eval_runtime": 1.6848, | |
| "eval_samples_per_second": 5.342, | |
| "eval_steps_per_second": 1.187, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.22772277227722773, | |
| "grad_norm": 0.9103761315345764, | |
| "learning_rate": 8.930265473713939e-06, | |
| "loss": 0.6495, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.2376237623762376, | |
| "grad_norm": 1.0075730085372925, | |
| "learning_rate": 8.83022221559489e-06, | |
| "loss": 0.7109, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.24752475247524752, | |
| "grad_norm": 0.9617104530334473, | |
| "learning_rate": 8.726322248378775e-06, | |
| "loss": 0.6993, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.25742574257425743, | |
| "grad_norm": 1.072245717048645, | |
| "learning_rate": 8.61867019052535e-06, | |
| "loss": 0.8897, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.26732673267326734, | |
| "grad_norm": 1.0244859457015991, | |
| "learning_rate": 8.507374438531606e-06, | |
| "loss": 0.8158, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.27722772277227725, | |
| "grad_norm": 1.0862797498703003, | |
| "learning_rate": 8.392547057785662e-06, | |
| "loss": 0.7539, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.2871287128712871, | |
| "grad_norm": 0.9954370260238647, | |
| "learning_rate": 8.274303669726427e-06, | |
| "loss": 0.718, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.297029702970297, | |
| "grad_norm": 1.0018794536590576, | |
| "learning_rate": 8.152763335422612e-06, | |
| "loss": 0.7518, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.3069306930693069, | |
| "grad_norm": 0.9397566318511963, | |
| "learning_rate": 8.028048435688333e-06, | |
| "loss": 0.663, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.31683168316831684, | |
| "grad_norm": 0.8733705878257751, | |
| "learning_rate": 7.900284547855992e-06, | |
| "loss": 0.6565, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.32673267326732675, | |
| "grad_norm": 0.938222348690033, | |
| "learning_rate": 7.769600319330553e-06, | |
| "loss": 0.601, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.32673267326732675, | |
| "eval_loss": 0.648015022277832, | |
| "eval_runtime": 1.6839, | |
| "eval_samples_per_second": 5.345, | |
| "eval_steps_per_second": 1.188, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.33663366336633666, | |
| "grad_norm": 0.9610555768013, | |
| "learning_rate": 7.636127338052513e-06, | |
| "loss": 0.7488, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.3465346534653465, | |
| "grad_norm": 1.0051265954971313, | |
| "learning_rate": 7.500000000000001e-06, | |
| "loss": 0.7929, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.3564356435643564, | |
| "grad_norm": 0.8689800500869751, | |
| "learning_rate": 7.361355373863415e-06, | |
| "loss": 0.57, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.36633663366336633, | |
| "grad_norm": 1.0085370540618896, | |
| "learning_rate": 7.2203330630288714e-06, | |
| "loss": 0.672, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.37623762376237624, | |
| "grad_norm": 0.8325203061103821, | |
| "learning_rate": 7.0770750650094335e-06, | |
| "loss": 0.6143, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.38613861386138615, | |
| "grad_norm": 0.9245986342430115, | |
| "learning_rate": 6.931725628465643e-06, | |
| "loss": 0.7288, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.39603960396039606, | |
| "grad_norm": 1.1846799850463867, | |
| "learning_rate": 6.78443110795936e-06, | |
| "loss": 0.8207, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.40594059405940597, | |
| "grad_norm": 0.8949479460716248, | |
| "learning_rate": 6.635339816587109e-06, | |
| "loss": 0.6476, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.4158415841584158, | |
| "grad_norm": 0.9853402376174927, | |
| "learning_rate": 6.484601876641375e-06, | |
| "loss": 0.7146, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.42574257425742573, | |
| "grad_norm": 0.9564022421836853, | |
| "learning_rate": 6.332369068450175e-06, | |
| "loss": 0.7382, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.43564356435643564, | |
| "grad_norm": 1.0044441223144531, | |
| "learning_rate": 6.178794677547138e-06, | |
| "loss": 0.7668, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.43564356435643564, | |
| "eval_loss": 0.6342881917953491, | |
| "eval_runtime": 1.6847, | |
| "eval_samples_per_second": 5.342, | |
| "eval_steps_per_second": 1.187, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.44554455445544555, | |
| "grad_norm": 0.8848042488098145, | |
| "learning_rate": 6.024033340325954e-06, | |
| "loss": 0.6532, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.45544554455445546, | |
| "grad_norm": 0.8762900233268738, | |
| "learning_rate": 5.8682408883346535e-06, | |
| "loss": 0.674, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.46534653465346537, | |
| "grad_norm": 0.8674903512001038, | |
| "learning_rate": 5.711574191366427e-06, | |
| "loss": 0.6138, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.4752475247524752, | |
| "grad_norm": 0.9267687201499939, | |
| "learning_rate": 5.5541909995050554e-06, | |
| "loss": 0.6838, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.48514851485148514, | |
| "grad_norm": 0.8423399329185486, | |
| "learning_rate": 5.396249784283943e-06, | |
| "loss": 0.6081, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.49504950495049505, | |
| "grad_norm": 0.9809019565582275, | |
| "learning_rate": 5.237909579118713e-06, | |
| "loss": 0.7202, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.504950495049505, | |
| "grad_norm": 0.8688491582870483, | |
| "learning_rate": 5.07932981917404e-06, | |
| "loss": 0.6236, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.5148514851485149, | |
| "grad_norm": 0.9908064007759094, | |
| "learning_rate": 4.9206701808259605e-06, | |
| "loss": 0.7449, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.5247524752475248, | |
| "grad_norm": 0.935342013835907, | |
| "learning_rate": 4.762090420881289e-06, | |
| "loss": 0.6582, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.5346534653465347, | |
| "grad_norm": 1.0338435173034668, | |
| "learning_rate": 4.603750215716057e-06, | |
| "loss": 0.7808, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.5445544554455446, | |
| "grad_norm": 1.0253307819366455, | |
| "learning_rate": 4.445809000494945e-06, | |
| "loss": 0.8058, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.5445544554455446, | |
| "eval_loss": 0.6248189210891724, | |
| "eval_runtime": 1.6835, | |
| "eval_samples_per_second": 5.346, | |
| "eval_steps_per_second": 1.188, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.5544554455445545, | |
| "grad_norm": 0.9448875188827515, | |
| "learning_rate": 4.2884258086335755e-06, | |
| "loss": 0.7117, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.5643564356435643, | |
| "grad_norm": 0.8737557530403137, | |
| "learning_rate": 4.131759111665349e-06, | |
| "loss": 0.5988, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.5742574257425742, | |
| "grad_norm": 0.764798104763031, | |
| "learning_rate": 3.975966659674048e-06, | |
| "loss": 0.501, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.5841584158415841, | |
| "grad_norm": 0.9195338487625122, | |
| "learning_rate": 3.821205322452863e-06, | |
| "loss": 0.677, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.594059405940594, | |
| "grad_norm": 0.9102662801742554, | |
| "learning_rate": 3.667630931549826e-06, | |
| "loss": 0.6363, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.6039603960396039, | |
| "grad_norm": 1.1512144804000854, | |
| "learning_rate": 3.5153981233586277e-06, | |
| "loss": 0.7082, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.6138613861386139, | |
| "grad_norm": 1.0006070137023926, | |
| "learning_rate": 3.3646601834128924e-06, | |
| "loss": 0.6685, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.6237623762376238, | |
| "grad_norm": 0.9273307919502258, | |
| "learning_rate": 3.2155688920406415e-06, | |
| "loss": 0.6652, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.6336633663366337, | |
| "grad_norm": 0.9092214703559875, | |
| "learning_rate": 3.0682743715343565e-06, | |
| "loss": 0.6497, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.6435643564356436, | |
| "grad_norm": 0.8190207481384277, | |
| "learning_rate": 2.9229249349905686e-06, | |
| "loss": 0.5194, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.6534653465346535, | |
| "grad_norm": 0.9039291739463806, | |
| "learning_rate": 2.7796669369711294e-06, | |
| "loss": 0.675, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.6534653465346535, | |
| "eval_loss": 0.6190232038497925, | |
| "eval_runtime": 1.6846, | |
| "eval_samples_per_second": 5.342, | |
| "eval_steps_per_second": 1.187, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.6633663366336634, | |
| "grad_norm": 1.0311757326126099, | |
| "learning_rate": 2.6386446261365874e-06, | |
| "loss": 0.8244, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.6732673267326733, | |
| "grad_norm": 0.8113132119178772, | |
| "learning_rate": 2.5000000000000015e-06, | |
| "loss": 0.5488, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.6831683168316832, | |
| "grad_norm": 0.9003159403800964, | |
| "learning_rate": 2.363872661947488e-06, | |
| "loss": 0.7291, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.693069306930693, | |
| "grad_norm": 1.057133674621582, | |
| "learning_rate": 2.230399680669449e-06, | |
| "loss": 0.8625, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.7029702970297029, | |
| "grad_norm": 0.9384030699729919, | |
| "learning_rate": 2.09971545214401e-06, | |
| "loss": 0.6813, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.7128712871287128, | |
| "grad_norm": 0.8832238912582397, | |
| "learning_rate": 1.971951564311668e-06, | |
| "loss": 0.6099, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.7227722772277227, | |
| "grad_norm": 0.8297242522239685, | |
| "learning_rate": 1.8472366645773892e-06, | |
| "loss": 0.5465, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.7326732673267327, | |
| "grad_norm": 0.9183589816093445, | |
| "learning_rate": 1.7256963302735752e-06, | |
| "loss": 0.6971, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.7425742574257426, | |
| "grad_norm": 0.8353100419044495, | |
| "learning_rate": 1.6074529422143398e-06, | |
| "loss": 0.5426, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.7524752475247525, | |
| "grad_norm": 1.051756739616394, | |
| "learning_rate": 1.4926255614683931e-06, | |
| "loss": 0.7567, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.7623762376237624, | |
| "grad_norm": 0.8063908815383911, | |
| "learning_rate": 1.3813298094746491e-06, | |
| "loss": 0.5383, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.7623762376237624, | |
| "eval_loss": 0.6155756711959839, | |
| "eval_runtime": 1.6854, | |
| "eval_samples_per_second": 5.34, | |
| "eval_steps_per_second": 1.187, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.7722772277227723, | |
| "grad_norm": 0.8472511768341064, | |
| "learning_rate": 1.2736777516212267e-06, | |
| "loss": 0.5792, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.7821782178217822, | |
| "grad_norm": 0.8465410470962524, | |
| "learning_rate": 1.1697777844051105e-06, | |
| "loss": 0.6283, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.7920792079207921, | |
| "grad_norm": 0.7360741496086121, | |
| "learning_rate": 1.0697345262860638e-06, | |
| "loss": 0.4699, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.801980198019802, | |
| "grad_norm": 0.9224135875701904, | |
| "learning_rate": 9.73648712344707e-07, | |
| "loss": 0.6806, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.8118811881188119, | |
| "grad_norm": 0.871505081653595, | |
| "learning_rate": 8.816170928508367e-07, | |
| "loss": 0.6824, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.8217821782178217, | |
| "grad_norm": 0.8890396952629089, | |
| "learning_rate": 7.937323358440935e-07, | |
| "loss": 0.6427, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.8316831683168316, | |
| "grad_norm": 1.091757893562317, | |
| "learning_rate": 7.100829338251147e-07, | |
| "loss": 0.7535, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.8415841584158416, | |
| "grad_norm": 0.809833824634552, | |
| "learning_rate": 6.307531146510754e-07, | |
| "loss": 0.5965, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.8514851485148515, | |
| "grad_norm": 0.861588716506958, | |
| "learning_rate": 5.558227567253832e-07, | |
| "loss": 0.6629, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.8613861386138614, | |
| "grad_norm": 0.7921182513237, | |
| "learning_rate": 4.853673085668947e-07, | |
| "loss": 0.482, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.8712871287128713, | |
| "grad_norm": 0.7692471742630005, | |
| "learning_rate": 4.194577128396521e-07, | |
| "loss": 0.5044, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.8712871287128713, | |
| "eval_loss": 0.6139124035835266, | |
| "eval_runtime": 1.684, | |
| "eval_samples_per_second": 5.344, | |
| "eval_steps_per_second": 1.188, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.8811881188118812, | |
| "grad_norm": 0.8603846430778503, | |
| "learning_rate": 3.581603349196372e-07, | |
| "loss": 0.6439, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.8910891089108911, | |
| "grad_norm": 0.8953405022621155, | |
| "learning_rate": 3.015368960704584e-07, | |
| "loss": 0.6881, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.900990099009901, | |
| "grad_norm": 0.9108988046646118, | |
| "learning_rate": 2.4964441129527337e-07, | |
| "loss": 0.7118, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.9108910891089109, | |
| "grad_norm": 0.886647641658783, | |
| "learning_rate": 2.0253513192751374e-07, | |
| "loss": 0.7544, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.9207920792079208, | |
| "grad_norm": 0.943878173828125, | |
| "learning_rate": 1.6025649301821877e-07, | |
| "loss": 0.7527, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.9306930693069307, | |
| "grad_norm": 0.8354966044425964, | |
| "learning_rate": 1.2285106557296479e-07, | |
| "loss": 0.6079, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.9405940594059405, | |
| "grad_norm": 0.8121664524078369, | |
| "learning_rate": 9.035651368646647e-08, | |
| "loss": 0.5891, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.9504950495049505, | |
| "grad_norm": 0.9194057583808899, | |
| "learning_rate": 6.280555661802857e-08, | |
| "loss": 0.6953, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.9603960396039604, | |
| "grad_norm": 0.9447068572044373, | |
| "learning_rate": 4.02259358460233e-08, | |
| "loss": 0.7401, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.9702970297029703, | |
| "grad_norm": 0.8672422766685486, | |
| "learning_rate": 2.264038713457706e-08, | |
| "loss": 0.6349, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.9801980198019802, | |
| "grad_norm": 0.904771625995636, | |
| "learning_rate": 1.006661764057837e-08, | |
| "loss": 0.6322, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.9801980198019802, | |
| "eval_loss": 0.6136024594306946, | |
| "eval_runtime": 1.6864, | |
| "eval_samples_per_second": 5.337, | |
| "eval_steps_per_second": 1.186, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.9900990099009901, | |
| "grad_norm": 0.8689496517181396, | |
| "learning_rate": 2.5172880840745873e-09, | |
| "loss": 0.6191, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.8673611283302307, | |
| "learning_rate": 0.0, | |
| "loss": 0.5897, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 101, | |
| "total_flos": 1.5438974547499418e+17, | |
| "train_loss": 0.6889385657735391, | |
| "train_runtime": 730.4137, | |
| "train_samples_per_second": 1.101, | |
| "train_steps_per_second": 0.138 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 101, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 101, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.5438974547499418e+17, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |