| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 337, | |
| "global_step": 337, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.002967359050445104, | |
| "grad_norm": 0.94921875, | |
| "learning_rate": 1e-05, | |
| "loss": 2.0677, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.005934718100890208, | |
| "grad_norm": 0.76953125, | |
| "learning_rate": 9.97032640949555e-06, | |
| "loss": 2.0863, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.008902077151335312, | |
| "grad_norm": 0.6640625, | |
| "learning_rate": 9.940652818991098e-06, | |
| "loss": 1.9814, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.011869436201780416, | |
| "grad_norm": 0.61328125, | |
| "learning_rate": 9.910979228486648e-06, | |
| "loss": 2.0, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.01483679525222552, | |
| "grad_norm": 0.59765625, | |
| "learning_rate": 9.881305637982197e-06, | |
| "loss": 1.9165, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.017804154302670624, | |
| "grad_norm": 0.54296875, | |
| "learning_rate": 9.851632047477747e-06, | |
| "loss": 1.8981, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.020771513353115726, | |
| "grad_norm": 0.55078125, | |
| "learning_rate": 9.821958456973294e-06, | |
| "loss": 1.8915, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.02373887240356083, | |
| "grad_norm": 0.5234375, | |
| "learning_rate": 9.792284866468842e-06, | |
| "loss": 1.8848, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.026706231454005934, | |
| "grad_norm": 0.486328125, | |
| "learning_rate": 9.762611275964392e-06, | |
| "loss": 1.7873, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.02967359050445104, | |
| "grad_norm": 0.486328125, | |
| "learning_rate": 9.732937685459941e-06, | |
| "loss": 1.8174, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.032640949554896145, | |
| "grad_norm": 0.478515625, | |
| "learning_rate": 9.70326409495549e-06, | |
| "loss": 1.8172, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.03560830860534125, | |
| "grad_norm": 0.41796875, | |
| "learning_rate": 9.673590504451039e-06, | |
| "loss": 1.7265, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.03857566765578635, | |
| "grad_norm": 0.416015625, | |
| "learning_rate": 9.643916913946588e-06, | |
| "loss": 1.6981, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.04154302670623145, | |
| "grad_norm": 0.38671875, | |
| "learning_rate": 9.614243323442138e-06, | |
| "loss": 1.7138, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.04451038575667656, | |
| "grad_norm": 0.359375, | |
| "learning_rate": 9.584569732937687e-06, | |
| "loss": 1.7191, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.04747774480712166, | |
| "grad_norm": 0.345703125, | |
| "learning_rate": 9.554896142433235e-06, | |
| "loss": 1.6637, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.050445103857566766, | |
| "grad_norm": 0.34375, | |
| "learning_rate": 9.525222551928784e-06, | |
| "loss": 1.6311, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.05341246290801187, | |
| "grad_norm": 0.330078125, | |
| "learning_rate": 9.495548961424334e-06, | |
| "loss": 1.6814, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.05637982195845697, | |
| "grad_norm": 0.361328125, | |
| "learning_rate": 9.465875370919882e-06, | |
| "loss": 1.7252, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.05934718100890208, | |
| "grad_norm": 0.337890625, | |
| "learning_rate": 9.436201780415431e-06, | |
| "loss": 1.677, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.06231454005934718, | |
| "grad_norm": 0.302734375, | |
| "learning_rate": 9.406528189910979e-06, | |
| "loss": 1.6096, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.06528189910979229, | |
| "grad_norm": 0.294921875, | |
| "learning_rate": 9.376854599406528e-06, | |
| "loss": 1.6227, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.06824925816023739, | |
| "grad_norm": 0.26953125, | |
| "learning_rate": 9.347181008902078e-06, | |
| "loss": 1.5964, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.0712166172106825, | |
| "grad_norm": 0.296875, | |
| "learning_rate": 9.317507418397626e-06, | |
| "loss": 1.5578, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.07418397626112759, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 9.287833827893175e-06, | |
| "loss": 1.5712, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.0771513353115727, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 9.258160237388725e-06, | |
| "loss": 1.5594, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.08011869436201781, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 9.228486646884274e-06, | |
| "loss": 1.6137, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.0830860534124629, | |
| "grad_norm": 0.275390625, | |
| "learning_rate": 9.198813056379822e-06, | |
| "loss": 1.5647, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.08605341246290801, | |
| "grad_norm": 0.267578125, | |
| "learning_rate": 9.169139465875372e-06, | |
| "loss": 1.5569, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.08902077151335312, | |
| "grad_norm": 0.236328125, | |
| "learning_rate": 9.139465875370921e-06, | |
| "loss": 1.5277, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.09198813056379822, | |
| "grad_norm": 0.25, | |
| "learning_rate": 9.10979228486647e-06, | |
| "loss": 1.5045, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.09495548961424333, | |
| "grad_norm": 0.275390625, | |
| "learning_rate": 9.080118694362018e-06, | |
| "loss": 1.5668, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.09792284866468842, | |
| "grad_norm": 0.3359375, | |
| "learning_rate": 9.050445103857568e-06, | |
| "loss": 1.541, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.10089020771513353, | |
| "grad_norm": 0.26953125, | |
| "learning_rate": 9.020771513353116e-06, | |
| "loss": 1.5232, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.10385756676557864, | |
| "grad_norm": 0.22265625, | |
| "learning_rate": 8.991097922848665e-06, | |
| "loss": 1.5146, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.10682492581602374, | |
| "grad_norm": 0.220703125, | |
| "learning_rate": 8.961424332344215e-06, | |
| "loss": 1.5174, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.10979228486646884, | |
| "grad_norm": 0.2080078125, | |
| "learning_rate": 8.931750741839763e-06, | |
| "loss": 1.4951, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.11275964391691394, | |
| "grad_norm": 0.2109375, | |
| "learning_rate": 8.902077151335312e-06, | |
| "loss": 1.4719, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.11572700296735905, | |
| "grad_norm": 0.212890625, | |
| "learning_rate": 8.872403560830862e-06, | |
| "loss": 1.4433, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.11869436201780416, | |
| "grad_norm": 0.30859375, | |
| "learning_rate": 8.842729970326411e-06, | |
| "loss": 1.4952, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.12166172106824925, | |
| "grad_norm": 0.20703125, | |
| "learning_rate": 8.813056379821959e-06, | |
| "loss": 1.5169, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.12462908011869436, | |
| "grad_norm": 0.2412109375, | |
| "learning_rate": 8.783382789317508e-06, | |
| "loss": 1.554, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.12759643916913946, | |
| "grad_norm": 0.216796875, | |
| "learning_rate": 8.753709198813058e-06, | |
| "loss": 1.4951, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.13056379821958458, | |
| "grad_norm": 0.21875, | |
| "learning_rate": 8.724035608308606e-06, | |
| "loss": 1.5065, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.13353115727002968, | |
| "grad_norm": 0.19921875, | |
| "learning_rate": 8.694362017804155e-06, | |
| "loss": 1.4919, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.13649851632047477, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 8.664688427299705e-06, | |
| "loss": 1.5049, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.1394658753709199, | |
| "grad_norm": 0.1943359375, | |
| "learning_rate": 8.635014836795252e-06, | |
| "loss": 1.4194, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.142433234421365, | |
| "grad_norm": 0.2177734375, | |
| "learning_rate": 8.605341246290802e-06, | |
| "loss": 1.3983, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.14540059347181009, | |
| "grad_norm": 0.2099609375, | |
| "learning_rate": 8.57566765578635e-06, | |
| "loss": 1.4324, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.14836795252225518, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 8.5459940652819e-06, | |
| "loss": 1.4409, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.1513353115727003, | |
| "grad_norm": 0.1953125, | |
| "learning_rate": 8.516320474777449e-06, | |
| "loss": 1.4665, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.1543026706231454, | |
| "grad_norm": 0.1796875, | |
| "learning_rate": 8.486646884272998e-06, | |
| "loss": 1.4116, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.1572700296735905, | |
| "grad_norm": 0.1943359375, | |
| "learning_rate": 8.456973293768546e-06, | |
| "loss": 1.4634, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.16023738872403562, | |
| "grad_norm": 0.1884765625, | |
| "learning_rate": 8.427299703264096e-06, | |
| "loss": 1.4233, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.1632047477744807, | |
| "grad_norm": 0.1884765625, | |
| "learning_rate": 8.397626112759645e-06, | |
| "loss": 1.4375, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.1661721068249258, | |
| "grad_norm": 0.185546875, | |
| "learning_rate": 8.367952522255195e-06, | |
| "loss": 1.4266, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.16913946587537093, | |
| "grad_norm": 0.201171875, | |
| "learning_rate": 8.338278931750742e-06, | |
| "loss": 1.4455, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.17210682492581603, | |
| "grad_norm": 0.1806640625, | |
| "learning_rate": 8.308605341246292e-06, | |
| "loss": 1.3704, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.17507418397626112, | |
| "grad_norm": 0.173828125, | |
| "learning_rate": 8.278931750741841e-06, | |
| "loss": 1.3802, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.17804154302670624, | |
| "grad_norm": 0.18359375, | |
| "learning_rate": 8.24925816023739e-06, | |
| "loss": 1.4001, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.18100890207715134, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 8.219584569732939e-06, | |
| "loss": 1.4321, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.18397626112759644, | |
| "grad_norm": 0.232421875, | |
| "learning_rate": 8.189910979228487e-06, | |
| "loss": 1.4167, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.18694362017804153, | |
| "grad_norm": 0.1865234375, | |
| "learning_rate": 8.160237388724036e-06, | |
| "loss": 1.3671, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.18991097922848665, | |
| "grad_norm": 0.26171875, | |
| "learning_rate": 8.130563798219586e-06, | |
| "loss": 1.4385, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.19287833827893175, | |
| "grad_norm": 0.1875, | |
| "learning_rate": 8.100890207715133e-06, | |
| "loss": 1.4385, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.19584569732937684, | |
| "grad_norm": 0.1845703125, | |
| "learning_rate": 8.071216617210683e-06, | |
| "loss": 1.4083, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.19881305637982197, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 8.041543026706232e-06, | |
| "loss": 1.3803, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.20178041543026706, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 8.011869436201782e-06, | |
| "loss": 1.3954, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.20474777448071216, | |
| "grad_norm": 0.2138671875, | |
| "learning_rate": 7.98219584569733e-06, | |
| "loss": 1.3395, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.20771513353115728, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 7.95252225519288e-06, | |
| "loss": 1.3295, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.21068249258160238, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 7.922848664688429e-06, | |
| "loss": 1.3906, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.21364985163204747, | |
| "grad_norm": 0.208984375, | |
| "learning_rate": 7.893175074183978e-06, | |
| "loss": 1.3739, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.2166172106824926, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 7.863501483679526e-06, | |
| "loss": 1.4093, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.2195845697329377, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 7.833827893175074e-06, | |
| "loss": 1.3961, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.22255192878338279, | |
| "grad_norm": 0.2158203125, | |
| "learning_rate": 7.804154302670623e-06, | |
| "loss": 1.3947, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.22551928783382788, | |
| "grad_norm": 0.1806640625, | |
| "learning_rate": 7.774480712166173e-06, | |
| "loss": 1.3505, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.228486646884273, | |
| "grad_norm": 0.220703125, | |
| "learning_rate": 7.744807121661722e-06, | |
| "loss": 1.4362, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.2314540059347181, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 7.71513353115727e-06, | |
| "loss": 1.3679, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.2344213649851632, | |
| "grad_norm": 0.2216796875, | |
| "learning_rate": 7.68545994065282e-06, | |
| "loss": 1.3595, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.23738872403560832, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 7.655786350148369e-06, | |
| "loss": 1.3918, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.2403560830860534, | |
| "grad_norm": 0.1826171875, | |
| "learning_rate": 7.626112759643918e-06, | |
| "loss": 1.4121, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.2433234421364985, | |
| "grad_norm": 0.1884765625, | |
| "learning_rate": 7.5964391691394664e-06, | |
| "loss": 1.3981, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.24629080118694363, | |
| "grad_norm": 0.2080078125, | |
| "learning_rate": 7.566765578635016e-06, | |
| "loss": 1.3622, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.24925816023738873, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 7.537091988130565e-06, | |
| "loss": 1.374, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.2522255192878338, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 7.507418397626114e-06, | |
| "loss": 1.3944, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.2551928783382789, | |
| "grad_norm": 0.20703125, | |
| "learning_rate": 7.477744807121662e-06, | |
| "loss": 1.361, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.258160237388724, | |
| "grad_norm": 0.232421875, | |
| "learning_rate": 7.4480712166172105e-06, | |
| "loss": 1.333, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.26112759643916916, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 7.41839762611276e-06, | |
| "loss": 1.3303, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.26409495548961426, | |
| "grad_norm": 0.1865234375, | |
| "learning_rate": 7.388724035608309e-06, | |
| "loss": 1.3375, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.26706231454005935, | |
| "grad_norm": 0.1884765625, | |
| "learning_rate": 7.359050445103858e-06, | |
| "loss": 1.3534, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.27002967359050445, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 7.329376854599407e-06, | |
| "loss": 1.3671, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.27299703264094954, | |
| "grad_norm": 0.181640625, | |
| "learning_rate": 7.299703264094956e-06, | |
| "loss": 1.3794, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.27596439169139464, | |
| "grad_norm": 0.1865234375, | |
| "learning_rate": 7.270029673590505e-06, | |
| "loss": 1.3515, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.2789317507418398, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 7.2403560830860545e-06, | |
| "loss": 1.3377, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.2818991097922849, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 7.210682492581603e-06, | |
| "loss": 1.3657, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.28486646884273, | |
| "grad_norm": 0.2119140625, | |
| "learning_rate": 7.181008902077153e-06, | |
| "loss": 1.3964, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.2878338278931751, | |
| "grad_norm": 0.2197265625, | |
| "learning_rate": 7.151335311572701e-06, | |
| "loss": 1.2998, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.29080118694362017, | |
| "grad_norm": 0.2138671875, | |
| "learning_rate": 7.12166172106825e-06, | |
| "loss": 1.3535, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.29376854599406527, | |
| "grad_norm": 0.19140625, | |
| "learning_rate": 7.091988130563799e-06, | |
| "loss": 1.3267, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.29673590504451036, | |
| "grad_norm": 0.201171875, | |
| "learning_rate": 7.062314540059347e-06, | |
| "loss": 1.3391, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.2997032640949555, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 7.032640949554897e-06, | |
| "loss": 1.3941, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.3026706231454006, | |
| "grad_norm": 0.2490234375, | |
| "learning_rate": 7.0029673590504455e-06, | |
| "loss": 1.3128, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.3056379821958457, | |
| "grad_norm": 0.181640625, | |
| "learning_rate": 6.973293768545994e-06, | |
| "loss": 1.3351, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.3086053412462908, | |
| "grad_norm": 0.17578125, | |
| "learning_rate": 6.943620178041544e-06, | |
| "loss": 1.3271, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.3115727002967359, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 6.913946587537092e-06, | |
| "loss": 1.3283, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.314540059347181, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 6.884272997032642e-06, | |
| "loss": 1.294, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.31750741839762614, | |
| "grad_norm": 0.205078125, | |
| "learning_rate": 6.85459940652819e-06, | |
| "loss": 1.3383, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.32047477744807124, | |
| "grad_norm": 0.2119140625, | |
| "learning_rate": 6.82492581602374e-06, | |
| "loss": 1.3375, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.32344213649851633, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 6.795252225519289e-06, | |
| "loss": 1.3399, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.3264094955489614, | |
| "grad_norm": 0.1884765625, | |
| "learning_rate": 6.765578635014838e-06, | |
| "loss": 1.345, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.3293768545994065, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 6.735905044510387e-06, | |
| "loss": 1.2866, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.3323442136498516, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 6.7062314540059345e-06, | |
| "loss": 1.3115, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.3353115727002967, | |
| "grad_norm": 0.1845703125, | |
| "learning_rate": 6.676557863501484e-06, | |
| "loss": 1.3283, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.33827893175074186, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 6.646884272997033e-06, | |
| "loss": 1.3087, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.34124629080118696, | |
| "grad_norm": 0.19140625, | |
| "learning_rate": 6.617210682492582e-06, | |
| "loss": 1.3121, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.34421364985163205, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 6.587537091988131e-06, | |
| "loss": 1.3209, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.34718100890207715, | |
| "grad_norm": 0.185546875, | |
| "learning_rate": 6.55786350148368e-06, | |
| "loss": 1.3096, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.35014836795252224, | |
| "grad_norm": 0.1875, | |
| "learning_rate": 6.528189910979229e-06, | |
| "loss": 1.294, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.35311572700296734, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 6.4985163204747785e-06, | |
| "loss": 1.3233, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.3560830860534125, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 6.468842729970327e-06, | |
| "loss": 1.3177, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.3590504451038576, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 6.439169139465876e-06, | |
| "loss": 1.3206, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.3620178041543027, | |
| "grad_norm": 0.236328125, | |
| "learning_rate": 6.409495548961425e-06, | |
| "loss": 1.3653, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.3649851632047478, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 6.379821958456974e-06, | |
| "loss": 1.2995, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.36795252225519287, | |
| "grad_norm": 0.2109375, | |
| "learning_rate": 6.3501483679525235e-06, | |
| "loss": 1.3219, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.37091988130563797, | |
| "grad_norm": 0.287109375, | |
| "learning_rate": 6.320474777448071e-06, | |
| "loss": 1.3444, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.37388724035608306, | |
| "grad_norm": 0.251953125, | |
| "learning_rate": 6.29080118694362e-06, | |
| "loss": 1.2872, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.3768545994065282, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 6.2611275964391694e-06, | |
| "loss": 1.3334, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.3798219584569733, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 6.231454005934718e-06, | |
| "loss": 1.3167, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.3827893175074184, | |
| "grad_norm": 0.2041015625, | |
| "learning_rate": 6.201780415430268e-06, | |
| "loss": 1.3093, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.3857566765578635, | |
| "grad_norm": 0.2099609375, | |
| "learning_rate": 6.172106824925816e-06, | |
| "loss": 1.2717, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.3887240356083086, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 6.142433234421366e-06, | |
| "loss": 1.3157, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.3916913946587537, | |
| "grad_norm": 0.19921875, | |
| "learning_rate": 6.112759643916914e-06, | |
| "loss": 1.3099, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.39465875370919884, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 6.083086053412464e-06, | |
| "loss": 1.3178, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.39762611275964393, | |
| "grad_norm": 0.1806640625, | |
| "learning_rate": 6.0534124629080126e-06, | |
| "loss": 1.2858, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.40059347181008903, | |
| "grad_norm": 0.310546875, | |
| "learning_rate": 6.023738872403562e-06, | |
| "loss": 1.3109, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.4035608308605341, | |
| "grad_norm": 0.2138671875, | |
| "learning_rate": 5.994065281899111e-06, | |
| "loss": 1.3189, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.4065281899109792, | |
| "grad_norm": 0.1943359375, | |
| "learning_rate": 5.964391691394659e-06, | |
| "loss": 1.2719, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.4094955489614243, | |
| "grad_norm": 0.1884765625, | |
| "learning_rate": 5.934718100890208e-06, | |
| "loss": 1.2975, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.4124629080118694, | |
| "grad_norm": 0.20703125, | |
| "learning_rate": 5.905044510385757e-06, | |
| "loss": 1.2303, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.41543026706231456, | |
| "grad_norm": 0.2099609375, | |
| "learning_rate": 5.875370919881306e-06, | |
| "loss": 1.3048, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.41839762611275966, | |
| "grad_norm": 0.18359375, | |
| "learning_rate": 5.845697329376855e-06, | |
| "loss": 1.2735, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.42136498516320475, | |
| "grad_norm": 0.1806640625, | |
| "learning_rate": 5.8160237388724035e-06, | |
| "loss": 1.327, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.42433234421364985, | |
| "grad_norm": 0.2041015625, | |
| "learning_rate": 5.786350148367953e-06, | |
| "loss": 1.2812, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.42729970326409494, | |
| "grad_norm": 0.234375, | |
| "learning_rate": 5.756676557863502e-06, | |
| "loss": 1.3042, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.43026706231454004, | |
| "grad_norm": 0.19140625, | |
| "learning_rate": 5.727002967359051e-06, | |
| "loss": 1.2647, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.4332344213649852, | |
| "grad_norm": 0.205078125, | |
| "learning_rate": 5.6973293768546e-06, | |
| "loss": 1.3202, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.4362017804154303, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 5.667655786350149e-06, | |
| "loss": 1.3324, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.4391691394658754, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 5.637982195845698e-06, | |
| "loss": 1.2715, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.4421364985163205, | |
| "grad_norm": 0.185546875, | |
| "learning_rate": 5.6083086053412475e-06, | |
| "loss": 1.2783, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.44510385756676557, | |
| "grad_norm": 0.267578125, | |
| "learning_rate": 5.578635014836796e-06, | |
| "loss": 1.265, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.44807121661721067, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 5.548961424332344e-06, | |
| "loss": 1.2528, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.45103857566765576, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 5.5192878338278934e-06, | |
| "loss": 1.3136, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.4540059347181009, | |
| "grad_norm": 0.2099609375, | |
| "learning_rate": 5.489614243323442e-06, | |
| "loss": 1.298, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.456973293768546, | |
| "grad_norm": 0.19140625, | |
| "learning_rate": 5.459940652818992e-06, | |
| "loss": 1.2735, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.4599406528189911, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 5.43026706231454e-06, | |
| "loss": 1.3049, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.4629080118694362, | |
| "grad_norm": 0.201171875, | |
| "learning_rate": 5.40059347181009e-06, | |
| "loss": 1.2733, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.4658753709198813, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 5.370919881305638e-06, | |
| "loss": 1.288, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.4688427299703264, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 5.341246290801188e-06, | |
| "loss": 1.3038, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.47181008902077154, | |
| "grad_norm": 0.1884765625, | |
| "learning_rate": 5.3115727002967366e-06, | |
| "loss": 1.2611, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.47477744807121663, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 5.281899109792285e-06, | |
| "loss": 1.3067, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.47774480712166173, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 5.252225519287835e-06, | |
| "loss": 1.2886, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.4807121661721068, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 5.222551928783383e-06, | |
| "loss": 1.2595, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.4836795252225519, | |
| "grad_norm": 0.1875, | |
| "learning_rate": 5.192878338278933e-06, | |
| "loss": 1.2751, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.486646884272997, | |
| "grad_norm": 0.1875, | |
| "learning_rate": 5.163204747774481e-06, | |
| "loss": 1.2961, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.4896142433234421, | |
| "grad_norm": 0.240234375, | |
| "learning_rate": 5.133531157270029e-06, | |
| "loss": 1.3248, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.49258160237388726, | |
| "grad_norm": 0.185546875, | |
| "learning_rate": 5.103857566765579e-06, | |
| "loss": 1.3029, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.49554896142433236, | |
| "grad_norm": 0.201171875, | |
| "learning_rate": 5.0741839762611275e-06, | |
| "loss": 1.2941, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.49851632047477745, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 5.044510385756677e-06, | |
| "loss": 1.2685, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.5014836795252225, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 5.014836795252226e-06, | |
| "loss": 1.2844, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.5044510385756676, | |
| "grad_norm": 0.2041015625, | |
| "learning_rate": 4.985163204747775e-06, | |
| "loss": 1.2717, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.5074183976261127, | |
| "grad_norm": 0.2060546875, | |
| "learning_rate": 4.955489614243324e-06, | |
| "loss": 1.2669, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.5103857566765578, | |
| "grad_norm": 0.203125, | |
| "learning_rate": 4.925816023738873e-06, | |
| "loss": 1.2924, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.5133531157270029, | |
| "grad_norm": 0.2138671875, | |
| "learning_rate": 4.896142433234421e-06, | |
| "loss": 1.2299, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.516320474777448, | |
| "grad_norm": 0.1943359375, | |
| "learning_rate": 4.866468842729971e-06, | |
| "loss": 1.287, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.5192878338278932, | |
| "grad_norm": 0.20703125, | |
| "learning_rate": 4.836795252225519e-06, | |
| "loss": 1.3162, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.5222551928783383, | |
| "grad_norm": 0.2041015625, | |
| "learning_rate": 4.807121661721069e-06, | |
| "loss": 1.2701, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.5252225519287834, | |
| "grad_norm": 0.2119140625, | |
| "learning_rate": 4.7774480712166174e-06, | |
| "loss": 1.2998, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.5281899109792285, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 4.747774480712167e-06, | |
| "loss": 1.2894, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.5311572700296736, | |
| "grad_norm": 0.2099609375, | |
| "learning_rate": 4.718100890207716e-06, | |
| "loss": 1.2528, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.5341246290801187, | |
| "grad_norm": 0.259765625, | |
| "learning_rate": 4.688427299703264e-06, | |
| "loss": 1.2502, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.5370919881305638, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 4.658753709198813e-06, | |
| "loss": 1.2745, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.5400593471810089, | |
| "grad_norm": 0.1865234375, | |
| "learning_rate": 4.629080118694362e-06, | |
| "loss": 1.2996, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.543026706231454, | |
| "grad_norm": 0.1875, | |
| "learning_rate": 4.599406528189911e-06, | |
| "loss": 1.2463, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.5459940652818991, | |
| "grad_norm": 0.1875, | |
| "learning_rate": 4.5697329376854606e-06, | |
| "loss": 1.2661, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.5489614243323442, | |
| "grad_norm": 0.2080078125, | |
| "learning_rate": 4.540059347181009e-06, | |
| "loss": 1.2275, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.5519287833827893, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 4.510385756676558e-06, | |
| "loss": 1.2486, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.5548961424332344, | |
| "grad_norm": 0.2099609375, | |
| "learning_rate": 4.480712166172107e-06, | |
| "loss": 1.2555, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.5578635014836796, | |
| "grad_norm": 0.21875, | |
| "learning_rate": 4.451038575667656e-06, | |
| "loss": 1.3215, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.5608308605341247, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 4.4213649851632055e-06, | |
| "loss": 1.2401, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.5637982195845698, | |
| "grad_norm": 0.2099609375, | |
| "learning_rate": 4.391691394658754e-06, | |
| "loss": 1.2612, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.5667655786350149, | |
| "grad_norm": 0.203125, | |
| "learning_rate": 4.362017804154303e-06, | |
| "loss": 1.2806, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.56973293768546, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 4.332344213649852e-06, | |
| "loss": 1.2672, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.5727002967359051, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 4.302670623145401e-06, | |
| "loss": 1.237, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.5756676557863502, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 4.27299703264095e-06, | |
| "loss": 1.2676, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.5786350148367952, | |
| "grad_norm": 0.2158203125, | |
| "learning_rate": 4.243323442136499e-06, | |
| "loss": 1.2977, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.5816023738872403, | |
| "grad_norm": 0.205078125, | |
| "learning_rate": 4.213649851632048e-06, | |
| "loss": 1.2774, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.5845697329376854, | |
| "grad_norm": 0.20703125, | |
| "learning_rate": 4.183976261127597e-06, | |
| "loss": 1.2262, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.5875370919881305, | |
| "grad_norm": 0.2080078125, | |
| "learning_rate": 4.154302670623146e-06, | |
| "loss": 1.2861, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.5905044510385756, | |
| "grad_norm": 0.1865234375, | |
| "learning_rate": 4.124629080118695e-06, | |
| "loss": 1.2761, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.5934718100890207, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 4.094955489614243e-06, | |
| "loss": 1.2924, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.5964391691394659, | |
| "grad_norm": 0.21875, | |
| "learning_rate": 4.065281899109793e-06, | |
| "loss": 1.2484, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.599406528189911, | |
| "grad_norm": 0.2080078125, | |
| "learning_rate": 4.0356083086053414e-06, | |
| "loss": 1.2271, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.6023738872403561, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 4.005934718100891e-06, | |
| "loss": 1.284, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.6053412462908012, | |
| "grad_norm": 0.205078125, | |
| "learning_rate": 3.97626112759644e-06, | |
| "loss": 1.2713, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.6083086053412463, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 3.946587537091989e-06, | |
| "loss": 1.2357, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.6112759643916914, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 3.916913946587537e-06, | |
| "loss": 1.3009, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.6142433234421365, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 3.887240356083086e-06, | |
| "loss": 1.2574, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.6172106824925816, | |
| "grad_norm": 0.208984375, | |
| "learning_rate": 3.857566765578635e-06, | |
| "loss": 1.276, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.6201780415430267, | |
| "grad_norm": 0.212890625, | |
| "learning_rate": 3.8278931750741846e-06, | |
| "loss": 1.3123, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.6231454005934718, | |
| "grad_norm": 0.19921875, | |
| "learning_rate": 3.7982195845697332e-06, | |
| "loss": 1.315, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.6261127596439169, | |
| "grad_norm": 0.205078125, | |
| "learning_rate": 3.7685459940652823e-06, | |
| "loss": 1.2536, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.629080118694362, | |
| "grad_norm": 0.201171875, | |
| "learning_rate": 3.738872403560831e-06, | |
| "loss": 1.2606, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.6320474777448071, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 3.70919881305638e-06, | |
| "loss": 1.25, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.6350148367952523, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 3.679525222551929e-06, | |
| "loss": 1.2494, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.6379821958456974, | |
| "grad_norm": 0.1943359375, | |
| "learning_rate": 3.649851632047478e-06, | |
| "loss": 1.2372, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.6409495548961425, | |
| "grad_norm": 0.310546875, | |
| "learning_rate": 3.6201780415430273e-06, | |
| "loss": 1.3252, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.6439169139465876, | |
| "grad_norm": 0.203125, | |
| "learning_rate": 3.5905044510385763e-06, | |
| "loss": 1.2716, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.6468842729970327, | |
| "grad_norm": 0.22265625, | |
| "learning_rate": 3.560830860534125e-06, | |
| "loss": 1.306, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.6498516320474778, | |
| "grad_norm": 0.201171875, | |
| "learning_rate": 3.5311572700296736e-06, | |
| "loss": 1.2718, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.6528189910979229, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 3.5014836795252227e-06, | |
| "loss": 1.2342, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.655786350148368, | |
| "grad_norm": 0.19140625, | |
| "learning_rate": 3.471810089020772e-06, | |
| "loss": 1.2443, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.658753709198813, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 3.442136498516321e-06, | |
| "loss": 1.2524, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.6617210682492581, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 3.41246290801187e-06, | |
| "loss": 1.292, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.6646884272997032, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 3.382789317507419e-06, | |
| "loss": 1.2716, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.6676557863501483, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 3.3531157270029673e-06, | |
| "loss": 1.2327, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.6706231454005934, | |
| "grad_norm": 0.203125, | |
| "learning_rate": 3.3234421364985163e-06, | |
| "loss": 1.2138, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.6735905044510386, | |
| "grad_norm": 0.203125, | |
| "learning_rate": 3.2937685459940654e-06, | |
| "loss": 1.232, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.6765578635014837, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 3.2640949554896145e-06, | |
| "loss": 1.2347, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.6795252225519288, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 3.2344213649851636e-06, | |
| "loss": 1.2906, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.6824925816023739, | |
| "grad_norm": 0.1943359375, | |
| "learning_rate": 3.2047477744807127e-06, | |
| "loss": 1.2621, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.685459940652819, | |
| "grad_norm": 0.1826171875, | |
| "learning_rate": 3.1750741839762617e-06, | |
| "loss": 1.2336, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.6884272997032641, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 3.14540059347181e-06, | |
| "loss": 1.2533, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.6913946587537092, | |
| "grad_norm": 0.21484375, | |
| "learning_rate": 3.115727002967359e-06, | |
| "loss": 1.2594, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.6943620178041543, | |
| "grad_norm": 0.2255859375, | |
| "learning_rate": 3.086053412462908e-06, | |
| "loss": 1.3164, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.6973293768545994, | |
| "grad_norm": 0.205078125, | |
| "learning_rate": 3.056379821958457e-06, | |
| "loss": 1.2366, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.7002967359050445, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 3.0267062314540063e-06, | |
| "loss": 1.2423, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.7032640949554896, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 2.9970326409495554e-06, | |
| "loss": 1.2781, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.7062314540059347, | |
| "grad_norm": 0.185546875, | |
| "learning_rate": 2.967359050445104e-06, | |
| "loss": 1.2367, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.7091988130563798, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 2.937685459940653e-06, | |
| "loss": 1.2378, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.712166172106825, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 2.9080118694362018e-06, | |
| "loss": 1.2291, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.7151335311572701, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 2.878338278931751e-06, | |
| "loss": 1.273, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.7181008902077152, | |
| "grad_norm": 0.2099609375, | |
| "learning_rate": 2.8486646884273e-06, | |
| "loss": 1.241, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.7210682492581603, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 2.818991097922849e-06, | |
| "loss": 1.2248, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.7240356083086054, | |
| "grad_norm": 0.2109375, | |
| "learning_rate": 2.789317507418398e-06, | |
| "loss": 1.2397, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.7270029673590505, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 2.7596439169139467e-06, | |
| "loss": 1.2308, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.7299703264094956, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 2.729970326409496e-06, | |
| "loss": 1.244, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.7329376854599406, | |
| "grad_norm": 0.201171875, | |
| "learning_rate": 2.700296735905045e-06, | |
| "loss": 1.2553, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.7359050445103857, | |
| "grad_norm": 0.2197265625, | |
| "learning_rate": 2.670623145400594e-06, | |
| "loss": 1.2875, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.7388724035608308, | |
| "grad_norm": 0.205078125, | |
| "learning_rate": 2.6409495548961426e-06, | |
| "loss": 1.2506, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.7418397626112759, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 2.6112759643916917e-06, | |
| "loss": 1.2325, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.744807121661721, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 2.5816023738872403e-06, | |
| "loss": 1.2208, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.7477744807121661, | |
| "grad_norm": 0.1865234375, | |
| "learning_rate": 2.5519287833827894e-06, | |
| "loss": 1.2257, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.7507418397626113, | |
| "grad_norm": 0.1943359375, | |
| "learning_rate": 2.5222551928783385e-06, | |
| "loss": 1.232, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.7537091988130564, | |
| "grad_norm": 0.2099609375, | |
| "learning_rate": 2.4925816023738876e-06, | |
| "loss": 1.27, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.7566765578635015, | |
| "grad_norm": 0.18359375, | |
| "learning_rate": 2.4629080118694367e-06, | |
| "loss": 1.2424, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.7596439169139466, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 2.4332344213649853e-06, | |
| "loss": 1.2693, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.7626112759643917, | |
| "grad_norm": 0.19921875, | |
| "learning_rate": 2.4035608308605344e-06, | |
| "loss": 1.2664, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.7655786350148368, | |
| "grad_norm": 0.1943359375, | |
| "learning_rate": 2.3738872403560835e-06, | |
| "loss": 1.2501, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.7685459940652819, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 2.344213649851632e-06, | |
| "loss": 1.2318, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.771513353115727, | |
| "grad_norm": 0.2109375, | |
| "learning_rate": 2.314540059347181e-06, | |
| "loss": 1.3029, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.7744807121661721, | |
| "grad_norm": 0.2080078125, | |
| "learning_rate": 2.2848664688427303e-06, | |
| "loss": 1.2724, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.7774480712166172, | |
| "grad_norm": 0.201171875, | |
| "learning_rate": 2.255192878338279e-06, | |
| "loss": 1.2878, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.7804154302670623, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 2.225519287833828e-06, | |
| "loss": 1.2552, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.7833827893175074, | |
| "grad_norm": 0.201171875, | |
| "learning_rate": 2.195845697329377e-06, | |
| "loss": 1.2643, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.7863501483679525, | |
| "grad_norm": 0.2119140625, | |
| "learning_rate": 2.166172106824926e-06, | |
| "loss": 1.233, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.7893175074183977, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 2.136498516320475e-06, | |
| "loss": 1.2309, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.7922848664688428, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 2.106824925816024e-06, | |
| "loss": 1.2229, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.7952522255192879, | |
| "grad_norm": 0.1953125, | |
| "learning_rate": 2.077151335311573e-06, | |
| "loss": 1.2481, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.798219584569733, | |
| "grad_norm": 0.19140625, | |
| "learning_rate": 2.0474777448071216e-06, | |
| "loss": 1.2295, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.8011869436201781, | |
| "grad_norm": 0.1884765625, | |
| "learning_rate": 2.0178041543026707e-06, | |
| "loss": 1.2315, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.8041543026706232, | |
| "grad_norm": 0.19140625, | |
| "learning_rate": 1.98813056379822e-06, | |
| "loss": 1.2581, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.8071216617210683, | |
| "grad_norm": 0.1953125, | |
| "learning_rate": 1.9584569732937684e-06, | |
| "loss": 1.1988, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.8100890207715133, | |
| "grad_norm": 0.19140625, | |
| "learning_rate": 1.9287833827893175e-06, | |
| "loss": 1.2688, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.8130563798219584, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 1.8991097922848666e-06, | |
| "loss": 1.2592, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.8160237388724035, | |
| "grad_norm": 0.2578125, | |
| "learning_rate": 1.8694362017804155e-06, | |
| "loss": 1.262, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.8189910979228486, | |
| "grad_norm": 0.224609375, | |
| "learning_rate": 1.8397626112759646e-06, | |
| "loss": 1.2743, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.8219584569732937, | |
| "grad_norm": 0.2197265625, | |
| "learning_rate": 1.8100890207715136e-06, | |
| "loss": 1.2792, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.8249258160237388, | |
| "grad_norm": 0.22265625, | |
| "learning_rate": 1.7804154302670625e-06, | |
| "loss": 1.2854, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.827893175074184, | |
| "grad_norm": 0.208984375, | |
| "learning_rate": 1.7507418397626114e-06, | |
| "loss": 1.3017, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.8308605341246291, | |
| "grad_norm": 0.201171875, | |
| "learning_rate": 1.7210682492581604e-06, | |
| "loss": 1.2655, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.8338278931750742, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 1.6913946587537095e-06, | |
| "loss": 1.2525, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.8367952522255193, | |
| "grad_norm": 0.1875, | |
| "learning_rate": 1.6617210682492582e-06, | |
| "loss": 1.247, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.8397626112759644, | |
| "grad_norm": 0.185546875, | |
| "learning_rate": 1.6320474777448073e-06, | |
| "loss": 1.2561, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.8427299703264095, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 1.6023738872403563e-06, | |
| "loss": 1.2832, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.8456973293768546, | |
| "grad_norm": 0.203125, | |
| "learning_rate": 1.572700296735905e-06, | |
| "loss": 1.2847, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.8486646884272997, | |
| "grad_norm": 0.1953125, | |
| "learning_rate": 1.543026706231454e-06, | |
| "loss": 1.2585, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.8516320474777448, | |
| "grad_norm": 0.2099609375, | |
| "learning_rate": 1.5133531157270031e-06, | |
| "loss": 1.293, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.8545994065281899, | |
| "grad_norm": 0.2041015625, | |
| "learning_rate": 1.483679525222552e-06, | |
| "loss": 1.2895, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.857566765578635, | |
| "grad_norm": 0.1875, | |
| "learning_rate": 1.4540059347181009e-06, | |
| "loss": 1.2713, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.8605341246290801, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 1.42433234421365e-06, | |
| "loss": 1.2637, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.8635014836795252, | |
| "grad_norm": 0.21484375, | |
| "learning_rate": 1.394658753709199e-06, | |
| "loss": 1.2238, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.8664688427299704, | |
| "grad_norm": 0.2158203125, | |
| "learning_rate": 1.364985163204748e-06, | |
| "loss": 1.3296, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.8694362017804155, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 1.335311572700297e-06, | |
| "loss": 1.2612, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.8724035608308606, | |
| "grad_norm": 0.19921875, | |
| "learning_rate": 1.3056379821958458e-06, | |
| "loss": 1.2456, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.8753709198813057, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 1.2759643916913947e-06, | |
| "loss": 1.2478, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.8783382789317508, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 1.2462908011869438e-06, | |
| "loss": 1.2667, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.8813056379821959, | |
| "grad_norm": 0.2041015625, | |
| "learning_rate": 1.2166172106824927e-06, | |
| "loss": 1.2313, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.884272997032641, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 1.1869436201780417e-06, | |
| "loss": 1.2604, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.887240356083086, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 1.1572700296735906e-06, | |
| "loss": 1.2672, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.8902077151335311, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 1.1275964391691395e-06, | |
| "loss": 1.253, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.8931750741839762, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 1.0979228486646885e-06, | |
| "loss": 1.2247, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.8961424332344213, | |
| "grad_norm": 0.2119140625, | |
| "learning_rate": 1.0682492581602374e-06, | |
| "loss": 1.291, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.8991097922848664, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 1.0385756676557865e-06, | |
| "loss": 1.2788, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.9020771513353115, | |
| "grad_norm": 0.1884765625, | |
| "learning_rate": 1.0089020771513354e-06, | |
| "loss": 1.2542, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.9050445103857567, | |
| "grad_norm": 0.1884765625, | |
| "learning_rate": 9.792284866468842e-07, | |
| "loss": 1.2522, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.9080118694362018, | |
| "grad_norm": 0.1953125, | |
| "learning_rate": 9.495548961424333e-07, | |
| "loss": 1.2059, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.9109792284866469, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 9.198813056379823e-07, | |
| "loss": 1.2572, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.913946587537092, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 8.902077151335312e-07, | |
| "loss": 1.2355, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.9169139465875371, | |
| "grad_norm": 0.212890625, | |
| "learning_rate": 8.605341246290802e-07, | |
| "loss": 1.2088, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.9198813056379822, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 8.308605341246291e-07, | |
| "loss": 1.2711, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.9228486646884273, | |
| "grad_norm": 0.1845703125, | |
| "learning_rate": 8.011869436201782e-07, | |
| "loss": 1.2234, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.9258160237388724, | |
| "grad_norm": 0.205078125, | |
| "learning_rate": 7.71513353115727e-07, | |
| "loss": 1.3062, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.9287833827893175, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 7.41839762611276e-07, | |
| "loss": 1.2521, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.9317507418397626, | |
| "grad_norm": 0.2099609375, | |
| "learning_rate": 7.12166172106825e-07, | |
| "loss": 1.2534, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.9347181008902077, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 6.82492581602374e-07, | |
| "loss": 1.2616, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.9376854599406528, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 6.528189910979229e-07, | |
| "loss": 1.3005, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.9406528189910979, | |
| "grad_norm": 0.19921875, | |
| "learning_rate": 6.231454005934719e-07, | |
| "loss": 1.3083, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.9436201780415431, | |
| "grad_norm": 0.1845703125, | |
| "learning_rate": 5.934718100890209e-07, | |
| "loss": 1.2245, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.9465875370919882, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 5.637982195845697e-07, | |
| "loss": 1.2227, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.9495548961424333, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 5.341246290801187e-07, | |
| "loss": 1.2399, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.9525222551928784, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 5.044510385756677e-07, | |
| "loss": 1.2081, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.9554896142433235, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 4.7477744807121665e-07, | |
| "loss": 1.2437, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.9584569732937686, | |
| "grad_norm": 0.22265625, | |
| "learning_rate": 4.451038575667656e-07, | |
| "loss": 1.1966, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.9614243323442137, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 4.1543026706231454e-07, | |
| "loss": 1.2334, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.9643916913946587, | |
| "grad_norm": 0.201171875, | |
| "learning_rate": 3.857566765578635e-07, | |
| "loss": 1.2657, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.9673590504451038, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 3.560830860534125e-07, | |
| "loss": 1.2559, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.9703264094955489, | |
| "grad_norm": 0.18359375, | |
| "learning_rate": 3.2640949554896146e-07, | |
| "loss": 1.2565, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.973293768545994, | |
| "grad_norm": 0.1845703125, | |
| "learning_rate": 2.9673590504451043e-07, | |
| "loss": 1.2533, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.9762611275964391, | |
| "grad_norm": 0.1875, | |
| "learning_rate": 2.6706231454005935e-07, | |
| "loss": 1.2565, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.9792284866468842, | |
| "grad_norm": 0.2490234375, | |
| "learning_rate": 2.3738872403560833e-07, | |
| "loss": 1.331, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.9821958456973294, | |
| "grad_norm": 0.2138671875, | |
| "learning_rate": 2.0771513353115727e-07, | |
| "loss": 1.2968, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.9851632047477745, | |
| "grad_norm": 0.1865234375, | |
| "learning_rate": 1.7804154302670624e-07, | |
| "loss": 1.2477, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.9881305637982196, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 1.4836795252225522e-07, | |
| "loss": 1.2617, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.9910979228486647, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 1.1869436201780416e-07, | |
| "loss": 1.2204, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.9940652818991098, | |
| "grad_norm": 0.2265625, | |
| "learning_rate": 8.902077151335312e-08, | |
| "loss": 1.253, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.9970326409495549, | |
| "grad_norm": 0.208984375, | |
| "learning_rate": 5.934718100890208e-08, | |
| "loss": 1.2939, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.76171875, | |
| "learning_rate": 2.967359050445104e-08, | |
| "loss": 1.2321, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 1.2519956827163696, | |
| "eval_runtime": 12.2277, | |
| "eval_samples_per_second": 2.862, | |
| "eval_steps_per_second": 0.409, | |
| "step": 337 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 337, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 0, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.154519532014928e+18, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |