| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 32.142857142857146, | |
| "eval_steps": 50, | |
| "global_step": 450, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.07142857142857142, | |
| "grad_norm": 5.383044719696045, | |
| "learning_rate": 0.0, | |
| "loss": 1.5793, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.14285714285714285, | |
| "grad_norm": 3.96494197845459, | |
| "learning_rate": 1.2500000000000002e-07, | |
| "loss": 1.62, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.21428571428571427, | |
| "grad_norm": 4.022143363952637, | |
| "learning_rate": 2.5000000000000004e-07, | |
| "loss": 1.6035, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.2857142857142857, | |
| "grad_norm": 6.808481216430664, | |
| "learning_rate": 3.75e-07, | |
| "loss": 1.6309, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.35714285714285715, | |
| "grad_norm": 4.0089311599731445, | |
| "learning_rate": 5.000000000000001e-07, | |
| "loss": 1.596, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.42857142857142855, | |
| "grad_norm": 6.727263450622559, | |
| "learning_rate": 6.25e-07, | |
| "loss": 1.6312, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 7.648512840270996, | |
| "learning_rate": 7.5e-07, | |
| "loss": 1.6206, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.5714285714285714, | |
| "grad_norm": 11.115941047668457, | |
| "learning_rate": 8.750000000000001e-07, | |
| "loss": 1.6133, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.6428571428571429, | |
| "grad_norm": 8.82610034942627, | |
| "learning_rate": 1.0000000000000002e-06, | |
| "loss": 1.6419, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.7142857142857143, | |
| "grad_norm": 3.3373453617095947, | |
| "learning_rate": 1.125e-06, | |
| "loss": 1.5894, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.7857142857142857, | |
| "grad_norm": 7.521600723266602, | |
| "learning_rate": 1.25e-06, | |
| "loss": 1.6188, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.8571428571428571, | |
| "grad_norm": 5.497554302215576, | |
| "learning_rate": 1.3750000000000002e-06, | |
| "loss": 1.6221, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.9285714285714286, | |
| "grad_norm": 3.5530192852020264, | |
| "learning_rate": 1.5e-06, | |
| "loss": 1.6112, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 4.452667713165283, | |
| "learning_rate": 1.6250000000000001e-06, | |
| "loss": 1.5999, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 1.0714285714285714, | |
| "grad_norm": 2.4583044052124023, | |
| "learning_rate": 1.7500000000000002e-06, | |
| "loss": 1.5917, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 1.1428571428571428, | |
| "grad_norm": 9.272625923156738, | |
| "learning_rate": 1.875e-06, | |
| "loss": 1.6092, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 1.2142857142857142, | |
| "grad_norm": 4.451343059539795, | |
| "learning_rate": 2.0000000000000003e-06, | |
| "loss": 1.5686, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 1.2857142857142856, | |
| "grad_norm": 5.756788730621338, | |
| "learning_rate": 2.1250000000000004e-06, | |
| "loss": 1.6265, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 1.3571428571428572, | |
| "grad_norm": 4.12752103805542, | |
| "learning_rate": 2.25e-06, | |
| "loss": 1.6073, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 1.4285714285714286, | |
| "grad_norm": 2.2131590843200684, | |
| "learning_rate": 2.375e-06, | |
| "loss": 1.5929, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "grad_norm": 7.726499080657959, | |
| "learning_rate": 2.5e-06, | |
| "loss": 1.6232, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 1.5714285714285714, | |
| "grad_norm": 5.906811714172363, | |
| "learning_rate": 2.625e-06, | |
| "loss": 1.5777, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 1.6428571428571428, | |
| "grad_norm": 4.695427894592285, | |
| "learning_rate": 2.7500000000000004e-06, | |
| "loss": 1.591, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 1.7142857142857144, | |
| "grad_norm": 1.5481104850769043, | |
| "learning_rate": 2.8750000000000004e-06, | |
| "loss": 1.5743, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 1.7857142857142856, | |
| "grad_norm": 4.624354839324951, | |
| "learning_rate": 3e-06, | |
| "loss": 1.6065, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 1.8571428571428572, | |
| "grad_norm": 5.646456718444824, | |
| "learning_rate": 3.125e-06, | |
| "loss": 1.5975, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 1.9285714285714286, | |
| "grad_norm": 8.583253860473633, | |
| "learning_rate": 3.2500000000000002e-06, | |
| "loss": 1.5809, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 1.7897872924804688, | |
| "learning_rate": 3.3750000000000003e-06, | |
| "loss": 1.5839, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 2.0714285714285716, | |
| "grad_norm": 2.509343385696411, | |
| "learning_rate": 3.5000000000000004e-06, | |
| "loss": 1.5725, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 2.142857142857143, | |
| "grad_norm": 3.86081600189209, | |
| "learning_rate": 3.625e-06, | |
| "loss": 1.5766, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 2.2142857142857144, | |
| "grad_norm": 4.29940128326416, | |
| "learning_rate": 3.75e-06, | |
| "loss": 1.5727, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 2.2857142857142856, | |
| "grad_norm": 3.209973096847534, | |
| "learning_rate": 3.875e-06, | |
| "loss": 1.5783, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 2.357142857142857, | |
| "grad_norm": 2.813647985458374, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 1.5687, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 2.4285714285714284, | |
| "grad_norm": 2.750006914138794, | |
| "learning_rate": 4.125e-06, | |
| "loss": 1.5739, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "grad_norm": 4.352213382720947, | |
| "learning_rate": 4.250000000000001e-06, | |
| "loss": 1.5781, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 2.571428571428571, | |
| "grad_norm": 2.743788957595825, | |
| "learning_rate": 4.375e-06, | |
| "loss": 1.5731, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 2.642857142857143, | |
| "grad_norm": 5.008453845977783, | |
| "learning_rate": 4.5e-06, | |
| "loss": 1.5828, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 2.7142857142857144, | |
| "grad_norm": 3.1128811836242676, | |
| "learning_rate": 4.625e-06, | |
| "loss": 1.5706, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 2.7857142857142856, | |
| "grad_norm": 5.385458946228027, | |
| "learning_rate": 4.75e-06, | |
| "loss": 1.5834, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 2.857142857142857, | |
| "grad_norm": 1.6079157590866089, | |
| "learning_rate": 4.875000000000001e-06, | |
| "loss": 1.5721, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 2.928571428571429, | |
| "grad_norm": 3.1016640663146973, | |
| "learning_rate": 5e-06, | |
| "loss": 1.5763, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "grad_norm": 4.650889873504639, | |
| "learning_rate": 5.125e-06, | |
| "loss": 1.5426, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 3.0714285714285716, | |
| "grad_norm": 1.5727084875106812, | |
| "learning_rate": 5.25e-06, | |
| "loss": 1.5432, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 3.142857142857143, | |
| "grad_norm": 2.637890577316284, | |
| "learning_rate": 5.375e-06, | |
| "loss": 1.5472, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 3.2142857142857144, | |
| "grad_norm": 1.8277631998062134, | |
| "learning_rate": 5.500000000000001e-06, | |
| "loss": 1.5363, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 3.2857142857142856, | |
| "grad_norm": 1.6959493160247803, | |
| "learning_rate": 5.625e-06, | |
| "loss": 1.5397, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 3.357142857142857, | |
| "grad_norm": 7.91141414642334, | |
| "learning_rate": 5.750000000000001e-06, | |
| "loss": 1.5392, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 3.4285714285714284, | |
| "grad_norm": 8.705608367919922, | |
| "learning_rate": 5.875e-06, | |
| "loss": 1.5499, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 3.5, | |
| "grad_norm": 1.4096094369888306, | |
| "learning_rate": 6e-06, | |
| "loss": 1.5554, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 3.571428571428571, | |
| "grad_norm": 1.2479711771011353, | |
| "learning_rate": 6.125e-06, | |
| "loss": 1.5537, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 3.642857142857143, | |
| "grad_norm": 1.7531940937042236, | |
| "learning_rate": 6.25e-06, | |
| "loss": 1.533, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 3.7142857142857144, | |
| "grad_norm": 3.1394143104553223, | |
| "learning_rate": 6.375000000000001e-06, | |
| "loss": 1.5223, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 3.7857142857142856, | |
| "grad_norm": 5.379184246063232, | |
| "learning_rate": 6.5000000000000004e-06, | |
| "loss": 1.5204, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 3.857142857142857, | |
| "grad_norm": 1.6642723083496094, | |
| "learning_rate": 6.625000000000001e-06, | |
| "loss": 1.5195, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 3.928571428571429, | |
| "grad_norm": 1.8299416303634644, | |
| "learning_rate": 6.750000000000001e-06, | |
| "loss": 1.5111, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "grad_norm": 3.942272901535034, | |
| "learning_rate": 6.875000000000001e-06, | |
| "loss": 1.5134, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 4.071428571428571, | |
| "grad_norm": 1.4287347793579102, | |
| "learning_rate": 7.000000000000001e-06, | |
| "loss": 1.5119, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 4.142857142857143, | |
| "grad_norm": 2.0702247619628906, | |
| "learning_rate": 7.1249999999999995e-06, | |
| "loss": 1.4962, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 4.214285714285714, | |
| "grad_norm": 2.863966703414917, | |
| "learning_rate": 7.25e-06, | |
| "loss": 1.5204, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 4.285714285714286, | |
| "grad_norm": 2.398355722427368, | |
| "learning_rate": 7.375e-06, | |
| "loss": 1.4955, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 4.357142857142857, | |
| "grad_norm": 1.4015003442764282, | |
| "learning_rate": 7.5e-06, | |
| "loss": 1.4942, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 4.428571428571429, | |
| "grad_norm": 1.6098459959030151, | |
| "learning_rate": 7.625e-06, | |
| "loss": 1.4712, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 4.5, | |
| "grad_norm": 4.722957611083984, | |
| "learning_rate": 7.75e-06, | |
| "loss": 1.5288, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 4.571428571428571, | |
| "grad_norm": 3.6340856552124023, | |
| "learning_rate": 7.875e-06, | |
| "loss": 1.524, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 4.642857142857143, | |
| "grad_norm": 2.6402580738067627, | |
| "learning_rate": 8.000000000000001e-06, | |
| "loss": 1.5417, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 4.714285714285714, | |
| "grad_norm": 2.7053163051605225, | |
| "learning_rate": 8.125000000000001e-06, | |
| "loss": 1.5101, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 4.785714285714286, | |
| "grad_norm": 4.306228160858154, | |
| "learning_rate": 8.25e-06, | |
| "loss": 1.5474, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 4.857142857142857, | |
| "grad_norm": 3.2443113327026367, | |
| "learning_rate": 8.375e-06, | |
| "loss": 1.5324, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 4.928571428571429, | |
| "grad_norm": 2.3825535774230957, | |
| "learning_rate": 8.500000000000002e-06, | |
| "loss": 1.522, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 3.2580692768096924, | |
| "learning_rate": 8.625e-06, | |
| "loss": 1.5179, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 5.071428571428571, | |
| "grad_norm": 1.4018006324768066, | |
| "learning_rate": 8.75e-06, | |
| "loss": 1.4971, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 5.142857142857143, | |
| "grad_norm": 5.274407386779785, | |
| "learning_rate": 8.875e-06, | |
| "loss": 1.513, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 5.214285714285714, | |
| "grad_norm": 1.3473955392837524, | |
| "learning_rate": 9e-06, | |
| "loss": 1.4898, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 5.285714285714286, | |
| "grad_norm": 5.1161346435546875, | |
| "learning_rate": 9.125e-06, | |
| "loss": 1.5032, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 5.357142857142857, | |
| "grad_norm": 10.268465042114258, | |
| "learning_rate": 9.25e-06, | |
| "loss": 1.4926, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 5.428571428571429, | |
| "grad_norm": 1.676137089729309, | |
| "learning_rate": 9.375000000000001e-06, | |
| "loss": 1.4833, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 5.5, | |
| "grad_norm": 1.441231369972229, | |
| "learning_rate": 9.5e-06, | |
| "loss": 1.4934, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 5.571428571428571, | |
| "grad_norm": 3.917909860610962, | |
| "learning_rate": 9.625e-06, | |
| "loss": 1.5473, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 5.642857142857143, | |
| "grad_norm": 1.9755581617355347, | |
| "learning_rate": 9.750000000000002e-06, | |
| "loss": 1.5192, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 5.714285714285714, | |
| "grad_norm": 2.587038516998291, | |
| "learning_rate": 9.875000000000001e-06, | |
| "loss": 1.5212, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 5.785714285714286, | |
| "grad_norm": 2.155751943588257, | |
| "learning_rate": 1e-05, | |
| "loss": 1.5337, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 5.857142857142857, | |
| "grad_norm": 2.8446056842803955, | |
| "learning_rate": 1.0125e-05, | |
| "loss": 1.5174, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 5.928571428571429, | |
| "grad_norm": 2.9765541553497314, | |
| "learning_rate": 1.025e-05, | |
| "loss": 1.5267, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "grad_norm": 2.4112436771392822, | |
| "learning_rate": 1.0375e-05, | |
| "loss": 1.4897, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 6.071428571428571, | |
| "grad_norm": 3.6984405517578125, | |
| "learning_rate": 1.05e-05, | |
| "loss": 1.5304, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 6.142857142857143, | |
| "grad_norm": 3.8474667072296143, | |
| "learning_rate": 1.0625e-05, | |
| "loss": 1.5232, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 6.214285714285714, | |
| "grad_norm": 2.087263584136963, | |
| "learning_rate": 1.075e-05, | |
| "loss": 1.5285, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 6.285714285714286, | |
| "grad_norm": 2.1326711177825928, | |
| "learning_rate": 1.0875e-05, | |
| "loss": 1.5387, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 6.357142857142857, | |
| "grad_norm": 4.559895992279053, | |
| "learning_rate": 1.1000000000000001e-05, | |
| "loss": 1.524, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 6.428571428571429, | |
| "grad_norm": 1.8320558071136475, | |
| "learning_rate": 1.1125000000000001e-05, | |
| "loss": 1.5143, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 6.5, | |
| "grad_norm": 1.8108539581298828, | |
| "learning_rate": 1.125e-05, | |
| "loss": 1.4987, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 6.571428571428571, | |
| "grad_norm": 1.3095402717590332, | |
| "learning_rate": 1.1375e-05, | |
| "loss": 1.5025, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 6.642857142857143, | |
| "grad_norm": 4.299781799316406, | |
| "learning_rate": 1.1500000000000002e-05, | |
| "loss": 1.5251, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 6.714285714285714, | |
| "grad_norm": 1.1237058639526367, | |
| "learning_rate": 1.1625000000000001e-05, | |
| "loss": 1.5185, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 6.785714285714286, | |
| "grad_norm": 6.838643550872803, | |
| "learning_rate": 1.175e-05, | |
| "loss": 1.5032, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 6.857142857142857, | |
| "grad_norm": 1.886619210243225, | |
| "learning_rate": 1.1875e-05, | |
| "loss": 1.5394, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 6.928571428571429, | |
| "grad_norm": 2.4293034076690674, | |
| "learning_rate": 1.2e-05, | |
| "loss": 1.4958, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "grad_norm": 1.3641681671142578, | |
| "learning_rate": 1.2125e-05, | |
| "loss": 1.5182, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 7.071428571428571, | |
| "grad_norm": 1.6436591148376465, | |
| "learning_rate": 1.225e-05, | |
| "loss": 1.5007, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 7.142857142857143, | |
| "grad_norm": 1.4990618228912354, | |
| "learning_rate": 1.2375000000000001e-05, | |
| "loss": 1.4881, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 7.214285714285714, | |
| "grad_norm": 1.2271665334701538, | |
| "learning_rate": 1.25e-05, | |
| "loss": 1.5017, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 7.285714285714286, | |
| "grad_norm": 1.2291427850723267, | |
| "learning_rate": 1.2625e-05, | |
| "loss": 1.4979, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 7.357142857142857, | |
| "grad_norm": 5.997069358825684, | |
| "learning_rate": 1.2750000000000002e-05, | |
| "loss": 1.5238, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 7.428571428571429, | |
| "grad_norm": 1.4330472946166992, | |
| "learning_rate": 1.2875000000000001e-05, | |
| "loss": 1.5133, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 7.5, | |
| "grad_norm": 1.3329960107803345, | |
| "learning_rate": 1.3000000000000001e-05, | |
| "loss": 1.5044, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 7.571428571428571, | |
| "grad_norm": 2.7081637382507324, | |
| "learning_rate": 1.3125e-05, | |
| "loss": 1.4825, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 7.642857142857143, | |
| "grad_norm": 2.8797085285186768, | |
| "learning_rate": 1.3250000000000002e-05, | |
| "loss": 1.4711, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 7.714285714285714, | |
| "grad_norm": 4.054922580718994, | |
| "learning_rate": 1.3375000000000002e-05, | |
| "loss": 1.4982, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 7.785714285714286, | |
| "grad_norm": 7.4362311363220215, | |
| "learning_rate": 1.3500000000000001e-05, | |
| "loss": 1.4786, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 7.857142857142857, | |
| "grad_norm": 2.5698859691619873, | |
| "learning_rate": 1.3625e-05, | |
| "loss": 1.4572, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 7.928571428571429, | |
| "grad_norm": 2.9194092750549316, | |
| "learning_rate": 1.3750000000000002e-05, | |
| "loss": 1.4916, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "grad_norm": 2.044605016708374, | |
| "learning_rate": 1.3875000000000002e-05, | |
| "loss": 1.4879, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 8.071428571428571, | |
| "grad_norm": 1.3414772748947144, | |
| "learning_rate": 1.4000000000000001e-05, | |
| "loss": 1.4887, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 8.142857142857142, | |
| "grad_norm": 2.0897555351257324, | |
| "learning_rate": 1.4125e-05, | |
| "loss": 1.4908, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 8.214285714285714, | |
| "grad_norm": 3.721792697906494, | |
| "learning_rate": 1.4249999999999999e-05, | |
| "loss": 1.4978, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 8.285714285714286, | |
| "grad_norm": 1.3067519664764404, | |
| "learning_rate": 1.4374999999999999e-05, | |
| "loss": 1.4903, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 8.357142857142858, | |
| "grad_norm": 2.3188929557800293, | |
| "learning_rate": 1.45e-05, | |
| "loss": 1.475, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 8.428571428571429, | |
| "grad_norm": 2.131817102432251, | |
| "learning_rate": 1.4625e-05, | |
| "loss": 1.4877, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 8.5, | |
| "grad_norm": 5.011297225952148, | |
| "learning_rate": 1.475e-05, | |
| "loss": 1.4651, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 8.571428571428571, | |
| "grad_norm": 2.497302770614624, | |
| "learning_rate": 1.4875e-05, | |
| "loss": 1.4488, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 8.642857142857142, | |
| "grad_norm": 1.3582658767700195, | |
| "learning_rate": 1.5e-05, | |
| "loss": 1.4446, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 8.714285714285714, | |
| "grad_norm": 5.030256748199463, | |
| "learning_rate": 1.5125e-05, | |
| "loss": 1.4445, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 8.785714285714286, | |
| "grad_norm": 6.480685234069824, | |
| "learning_rate": 1.525e-05, | |
| "loss": 1.4853, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 8.857142857142858, | |
| "grad_norm": 9.358418464660645, | |
| "learning_rate": 1.5375e-05, | |
| "loss": 1.5026, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 8.928571428571429, | |
| "grad_norm": 8.761038780212402, | |
| "learning_rate": 1.55e-05, | |
| "loss": 1.4884, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "grad_norm": 1.6179229021072388, | |
| "learning_rate": 1.5625e-05, | |
| "loss": 1.4705, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 9.071428571428571, | |
| "grad_norm": 2.0552783012390137, | |
| "learning_rate": 1.575e-05, | |
| "loss": 1.4869, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 9.142857142857142, | |
| "grad_norm": 6.032142162322998, | |
| "learning_rate": 1.5875e-05, | |
| "loss": 1.4461, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 9.214285714285714, | |
| "grad_norm": 2.0145914554595947, | |
| "learning_rate": 1.6000000000000003e-05, | |
| "loss": 1.4625, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 9.285714285714286, | |
| "grad_norm": 5.505954742431641, | |
| "learning_rate": 1.6125000000000002e-05, | |
| "loss": 1.4764, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 9.357142857142858, | |
| "grad_norm": 4.161317348480225, | |
| "learning_rate": 1.6250000000000002e-05, | |
| "loss": 1.4558, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 9.428571428571429, | |
| "grad_norm": 6.453343391418457, | |
| "learning_rate": 1.6375e-05, | |
| "loss": 1.4914, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 9.5, | |
| "grad_norm": 2.7060706615448, | |
| "learning_rate": 1.65e-05, | |
| "loss": 1.4479, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 9.571428571428571, | |
| "grad_norm": 6.37178897857666, | |
| "learning_rate": 1.6625e-05, | |
| "loss": 1.4951, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 9.642857142857142, | |
| "grad_norm": 1.644932746887207, | |
| "learning_rate": 1.675e-05, | |
| "loss": 1.4492, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 9.714285714285714, | |
| "grad_norm": 4.173709392547607, | |
| "learning_rate": 1.6875000000000004e-05, | |
| "loss": 1.4419, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 9.785714285714286, | |
| "grad_norm": 2.661614179611206, | |
| "learning_rate": 1.7000000000000003e-05, | |
| "loss": 1.4419, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 9.857142857142858, | |
| "grad_norm": 2.468867778778076, | |
| "learning_rate": 1.7125000000000003e-05, | |
| "loss": 1.416, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 9.928571428571429, | |
| "grad_norm": 1.0798838138580322, | |
| "learning_rate": 1.725e-05, | |
| "loss": 1.4473, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "grad_norm": 1.5816832780838013, | |
| "learning_rate": 1.7375e-05, | |
| "loss": 1.4399, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 10.071428571428571, | |
| "grad_norm": 2.063046932220459, | |
| "learning_rate": 1.75e-05, | |
| "loss": 1.4415, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 10.142857142857142, | |
| "grad_norm": 3.9278879165649414, | |
| "learning_rate": 1.7625e-05, | |
| "loss": 1.4463, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 10.214285714285714, | |
| "grad_norm": 1.3775101900100708, | |
| "learning_rate": 1.775e-05, | |
| "loss": 1.4019, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 10.285714285714286, | |
| "grad_norm": 6.007996082305908, | |
| "learning_rate": 1.7875e-05, | |
| "loss": 1.4282, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 10.357142857142858, | |
| "grad_norm": 4.488779067993164, | |
| "learning_rate": 1.8e-05, | |
| "loss": 1.451, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 10.428571428571429, | |
| "grad_norm": 5.015976905822754, | |
| "learning_rate": 1.8125e-05, | |
| "loss": 1.4089, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 10.5, | |
| "grad_norm": 1.4387730360031128, | |
| "learning_rate": 1.825e-05, | |
| "loss": 1.4249, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 10.571428571428571, | |
| "grad_norm": 2.4665675163269043, | |
| "learning_rate": 1.8375e-05, | |
| "loss": 1.4327, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 10.642857142857142, | |
| "grad_norm": 1.8020267486572266, | |
| "learning_rate": 1.85e-05, | |
| "loss": 1.3995, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 10.714285714285714, | |
| "grad_norm": 1.6808390617370605, | |
| "learning_rate": 1.8625000000000002e-05, | |
| "loss": 1.389, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 10.785714285714286, | |
| "grad_norm": 2.3376588821411133, | |
| "learning_rate": 1.8750000000000002e-05, | |
| "loss": 1.4748, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 10.857142857142858, | |
| "grad_norm": 2.5257272720336914, | |
| "learning_rate": 1.8875e-05, | |
| "loss": 1.4221, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 10.928571428571429, | |
| "grad_norm": 1.6355801820755005, | |
| "learning_rate": 1.9e-05, | |
| "loss": 1.3544, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "grad_norm": 2.1227149963378906, | |
| "learning_rate": 1.9125e-05, | |
| "loss": 1.3493, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 11.071428571428571, | |
| "grad_norm": 4.7650370597839355, | |
| "learning_rate": 1.925e-05, | |
| "loss": 1.4708, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 11.142857142857142, | |
| "grad_norm": 2.83673095703125, | |
| "learning_rate": 1.9375e-05, | |
| "loss": 1.4064, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 11.214285714285714, | |
| "grad_norm": 2.4201812744140625, | |
| "learning_rate": 1.9500000000000003e-05, | |
| "loss": 1.3221, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 11.285714285714286, | |
| "grad_norm": 2.0276103019714355, | |
| "learning_rate": 1.9625000000000003e-05, | |
| "loss": 1.2859, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 11.357142857142858, | |
| "grad_norm": 7.189219951629639, | |
| "learning_rate": 1.9750000000000002e-05, | |
| "loss": 1.3526, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 11.428571428571429, | |
| "grad_norm": 2.9911937713623047, | |
| "learning_rate": 1.9875000000000002e-05, | |
| "loss": 1.319, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 11.5, | |
| "grad_norm": 7.41688346862793, | |
| "learning_rate": 2e-05, | |
| "loss": 1.2701, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 11.571428571428571, | |
| "grad_norm": 2.5221216678619385, | |
| "learning_rate": 2.0125e-05, | |
| "loss": 1.2342, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 11.642857142857142, | |
| "grad_norm": 3.1596360206604004, | |
| "learning_rate": 2.025e-05, | |
| "loss": 1.2818, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 11.714285714285714, | |
| "grad_norm": 4.858260631561279, | |
| "learning_rate": 2.0375e-05, | |
| "loss": 1.263, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 11.785714285714286, | |
| "grad_norm": 5.346636772155762, | |
| "learning_rate": 2.05e-05, | |
| "loss": 1.2936, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 11.857142857142858, | |
| "grad_norm": 4.936079502105713, | |
| "learning_rate": 2.0625e-05, | |
| "loss": 1.2727, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 11.928571428571429, | |
| "grad_norm": 10.273465156555176, | |
| "learning_rate": 2.075e-05, | |
| "loss": 1.2605, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "grad_norm": 4.918086051940918, | |
| "learning_rate": 2.0875e-05, | |
| "loss": 1.2611, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 12.071428571428571, | |
| "grad_norm": 12.318052291870117, | |
| "learning_rate": 2.1e-05, | |
| "loss": 1.2175, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 12.142857142857142, | |
| "grad_norm": 4.972585201263428, | |
| "learning_rate": 2.1125000000000002e-05, | |
| "loss": 1.1154, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 12.214285714285714, | |
| "grad_norm": 4.9560441970825195, | |
| "learning_rate": 2.125e-05, | |
| "loss": 1.1845, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 12.285714285714286, | |
| "grad_norm": 5.026245594024658, | |
| "learning_rate": 2.1375e-05, | |
| "loss": 1.1234, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 12.357142857142858, | |
| "grad_norm": 7.455307483673096, | |
| "learning_rate": 2.15e-05, | |
| "loss": 1.1865, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 12.428571428571429, | |
| "grad_norm": 5.8239946365356445, | |
| "learning_rate": 2.1625e-05, | |
| "loss": 1.2061, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 12.5, | |
| "grad_norm": 7.225159168243408, | |
| "learning_rate": 2.175e-05, | |
| "loss": 1.2647, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 12.571428571428571, | |
| "grad_norm": 3.666829824447632, | |
| "learning_rate": 2.1875e-05, | |
| "loss": 1.1374, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 12.642857142857142, | |
| "grad_norm": 6.110956192016602, | |
| "learning_rate": 2.2000000000000003e-05, | |
| "loss": 1.1125, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 12.714285714285714, | |
| "grad_norm": 10.124720573425293, | |
| "learning_rate": 2.2125000000000002e-05, | |
| "loss": 1.1762, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 12.785714285714286, | |
| "grad_norm": 6.521277904510498, | |
| "learning_rate": 2.2250000000000002e-05, | |
| "loss": 1.2196, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 12.857142857142858, | |
| "grad_norm": 3.0751357078552246, | |
| "learning_rate": 2.2375000000000002e-05, | |
| "loss": 1.1462, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 12.928571428571429, | |
| "grad_norm": 4.803239345550537, | |
| "learning_rate": 2.25e-05, | |
| "loss": 1.0391, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "grad_norm": 3.005739450454712, | |
| "learning_rate": 2.2625e-05, | |
| "loss": 1.0753, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 13.071428571428571, | |
| "grad_norm": 4.087983131408691, | |
| "learning_rate": 2.275e-05, | |
| "loss": 1.2163, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 13.142857142857142, | |
| "grad_norm": 3.475637674331665, | |
| "learning_rate": 2.2875e-05, | |
| "loss": 1.197, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 13.214285714285714, | |
| "grad_norm": 6.145781517028809, | |
| "learning_rate": 2.3000000000000003e-05, | |
| "loss": 1.1744, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 13.285714285714286, | |
| "grad_norm": 4.0928544998168945, | |
| "learning_rate": 2.3125000000000003e-05, | |
| "loss": 1.1189, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 13.357142857142858, | |
| "grad_norm": 9.976082801818848, | |
| "learning_rate": 2.3250000000000003e-05, | |
| "loss": 1.2488, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 13.428571428571429, | |
| "grad_norm": 6.045898914337158, | |
| "learning_rate": 2.3375000000000002e-05, | |
| "loss": 1.139, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 13.5, | |
| "grad_norm": 4.3854451179504395, | |
| "learning_rate": 2.35e-05, | |
| "loss": 1.0929, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 13.571428571428571, | |
| "grad_norm": 3.349039316177368, | |
| "learning_rate": 2.3624999999999998e-05, | |
| "loss": 1.0427, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 13.642857142857142, | |
| "grad_norm": 3.4710395336151123, | |
| "learning_rate": 2.375e-05, | |
| "loss": 1.1249, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 13.714285714285714, | |
| "grad_norm": 9.711404800415039, | |
| "learning_rate": 2.3875e-05, | |
| "loss": 1.2626, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 13.785714285714286, | |
| "grad_norm": 3.2990989685058594, | |
| "learning_rate": 2.4e-05, | |
| "loss": 1.1687, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 13.857142857142858, | |
| "grad_norm": 2.510340929031372, | |
| "learning_rate": 2.4125e-05, | |
| "loss": 1.0953, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 13.928571428571429, | |
| "grad_norm": 3.561298131942749, | |
| "learning_rate": 2.425e-05, | |
| "loss": 1.0411, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "grad_norm": 4.719476699829102, | |
| "learning_rate": 2.4375e-05, | |
| "loss": 1.0803, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 14.071428571428571, | |
| "grad_norm": 3.3607053756713867, | |
| "learning_rate": 2.45e-05, | |
| "loss": 1.0709, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 14.142857142857142, | |
| "grad_norm": 4.803561687469482, | |
| "learning_rate": 2.4625000000000002e-05, | |
| "loss": 1.0335, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 14.214285714285714, | |
| "grad_norm": 3.8269684314727783, | |
| "learning_rate": 2.4750000000000002e-05, | |
| "loss": 1.0742, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 14.285714285714286, | |
| "grad_norm": 2.7520132064819336, | |
| "learning_rate": 2.4875e-05, | |
| "loss": 1.17, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 14.357142857142858, | |
| "grad_norm": 5.653459548950195, | |
| "learning_rate": 2.5e-05, | |
| "loss": 1.0918, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 14.428571428571429, | |
| "grad_norm": 2.774630546569824, | |
| "learning_rate": 2.4999980961416097e-05, | |
| "loss": 1.058, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 14.5, | |
| "grad_norm": 3.926753520965576, | |
| "learning_rate": 2.499992384572238e-05, | |
| "loss": 1.0942, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 14.571428571428571, | |
| "grad_norm": 2.9734742641448975, | |
| "learning_rate": 2.4999828653092835e-05, | |
| "loss": 1.007, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 14.642857142857142, | |
| "grad_norm": 2.9620471000671387, | |
| "learning_rate": 2.4999695383817435e-05, | |
| "loss": 1.0881, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 14.714285714285714, | |
| "grad_norm": 5.124941349029541, | |
| "learning_rate": 2.499952403830214e-05, | |
| "loss": 1.0806, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 14.785714285714286, | |
| "grad_norm": 3.0979349613189697, | |
| "learning_rate": 2.4999314617068904e-05, | |
| "loss": 1.0281, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 14.857142857142858, | |
| "grad_norm": 3.4293224811553955, | |
| "learning_rate": 2.4999067120755652e-05, | |
| "loss": 1.051, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 14.928571428571429, | |
| "grad_norm": 6.487819194793701, | |
| "learning_rate": 2.4998781550116305e-05, | |
| "loss": 1.0308, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "grad_norm": 3.3049116134643555, | |
| "learning_rate": 2.499845790602076e-05, | |
| "loss": 0.9572, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 15.071428571428571, | |
| "grad_norm": 4.562008380889893, | |
| "learning_rate": 2.4998096189454893e-05, | |
| "loss": 1.0758, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 15.142857142857142, | |
| "grad_norm": 4.130517482757568, | |
| "learning_rate": 2.4997696401520555e-05, | |
| "loss": 1.0565, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 15.214285714285714, | |
| "grad_norm": 7.463335037231445, | |
| "learning_rate": 2.499725854343557e-05, | |
| "loss": 0.9738, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 15.285714285714286, | |
| "grad_norm": 5.149586200714111, | |
| "learning_rate": 2.4996782616533732e-05, | |
| "loss": 1.0234, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 15.357142857142858, | |
| "grad_norm": 5.69240140914917, | |
| "learning_rate": 2.499626862226479e-05, | |
| "loss": 1.1353, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 15.428571428571429, | |
| "grad_norm": 4.097994327545166, | |
| "learning_rate": 2.4995716562194465e-05, | |
| "loss": 1.0006, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 15.5, | |
| "grad_norm": 3.709420919418335, | |
| "learning_rate": 2.499512643800443e-05, | |
| "loss": 0.9626, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 15.571428571428571, | |
| "grad_norm": 6.93643856048584, | |
| "learning_rate": 2.4994498251492302e-05, | |
| "loss": 0.9727, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 15.642857142857142, | |
| "grad_norm": 2.359558582305908, | |
| "learning_rate": 2.4993832004571646e-05, | |
| "loss": 1.0169, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 15.714285714285714, | |
| "grad_norm": 3.7421205043792725, | |
| "learning_rate": 2.4993127699271966e-05, | |
| "loss": 1.0996, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 15.785714285714286, | |
| "grad_norm": 3.670799493789673, | |
| "learning_rate": 2.49923853377387e-05, | |
| "loss": 1.0078, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 15.857142857142858, | |
| "grad_norm": 3.7862744331359863, | |
| "learning_rate": 2.4991604922233204e-05, | |
| "loss": 0.9785, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 15.928571428571429, | |
| "grad_norm": 2.4504923820495605, | |
| "learning_rate": 2.4990786455132764e-05, | |
| "loss": 0.9815, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "grad_norm": 4.063878536224365, | |
| "learning_rate": 2.4989929938930576e-05, | |
| "loss": 0.9463, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 16.071428571428573, | |
| "grad_norm": 2.9702260494232178, | |
| "learning_rate": 2.498903537623573e-05, | |
| "loss": 0.9618, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 16.142857142857142, | |
| "grad_norm": 2.672618865966797, | |
| "learning_rate": 2.4988102769773227e-05, | |
| "loss": 0.9338, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 16.214285714285715, | |
| "grad_norm": 6.381713390350342, | |
| "learning_rate": 2.4987132122383936e-05, | |
| "loss": 1.0038, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 16.285714285714285, | |
| "grad_norm": 2.369471549987793, | |
| "learning_rate": 2.4986123437024627e-05, | |
| "loss": 1.0239, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 16.357142857142858, | |
| "grad_norm": 2.5277535915374756, | |
| "learning_rate": 2.4985076716767927e-05, | |
| "loss": 0.9014, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 16.428571428571427, | |
| "grad_norm": 3.924433946609497, | |
| "learning_rate": 2.4983991964802327e-05, | |
| "loss": 0.966, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 16.5, | |
| "grad_norm": 3.6045427322387695, | |
| "learning_rate": 2.4982869184432174e-05, | |
| "loss": 0.9695, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 16.571428571428573, | |
| "grad_norm": 6.360043048858643, | |
| "learning_rate": 2.498170837907765e-05, | |
| "loss": 1.0349, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 16.642857142857142, | |
| "grad_norm": 4.736248970031738, | |
| "learning_rate": 2.4980509552274765e-05, | |
| "loss": 0.992, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 16.714285714285715, | |
| "grad_norm": 2.659874677658081, | |
| "learning_rate": 2.4979272707675356e-05, | |
| "loss": 0.9548, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 16.785714285714285, | |
| "grad_norm": 2.803654909133911, | |
| "learning_rate": 2.497799784904707e-05, | |
| "loss": 0.8971, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 16.857142857142858, | |
| "grad_norm": 4.661594867706299, | |
| "learning_rate": 2.4976684980273338e-05, | |
| "loss": 0.9161, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 16.928571428571427, | |
| "grad_norm": 3.7907001972198486, | |
| "learning_rate": 2.4975334105353396e-05, | |
| "loss": 0.8663, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "grad_norm": 6.289381504058838, | |
| "learning_rate": 2.497394522840224e-05, | |
| "loss": 0.9157, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 17.071428571428573, | |
| "grad_norm": 3.7386388778686523, | |
| "learning_rate": 2.4972518353650626e-05, | |
| "loss": 0.9065, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 17.142857142857142, | |
| "grad_norm": 2.243206739425659, | |
| "learning_rate": 2.497105348544507e-05, | |
| "loss": 0.8066, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 17.214285714285715, | |
| "grad_norm": 3.9942452907562256, | |
| "learning_rate": 2.4969550628247805e-05, | |
| "loss": 0.9173, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 17.285714285714285, | |
| "grad_norm": 4.815189361572266, | |
| "learning_rate": 2.49680097866368e-05, | |
| "loss": 0.9629, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 17.357142857142858, | |
| "grad_norm": 5.594153881072998, | |
| "learning_rate": 2.4966430965305727e-05, | |
| "loss": 0.9673, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 17.428571428571427, | |
| "grad_norm": 4.274055004119873, | |
| "learning_rate": 2.4964814169063948e-05, | |
| "loss": 0.9222, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 17.5, | |
| "grad_norm": 3.3747873306274414, | |
| "learning_rate": 2.4963159402836506e-05, | |
| "loss": 0.8992, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 17.571428571428573, | |
| "grad_norm": 4.040309429168701, | |
| "learning_rate": 2.49614666716641e-05, | |
| "loss": 0.8257, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 17.642857142857142, | |
| "grad_norm": 3.115257740020752, | |
| "learning_rate": 2.495973598070309e-05, | |
| "loss": 0.9713, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 17.714285714285715, | |
| "grad_norm": 2.5480332374572754, | |
| "learning_rate": 2.4957967335225456e-05, | |
| "loss": 0.8687, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 17.785714285714285, | |
| "grad_norm": 3.8874218463897705, | |
| "learning_rate": 2.4956160740618806e-05, | |
| "loss": 0.9432, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 17.857142857142858, | |
| "grad_norm": 3.870405673980713, | |
| "learning_rate": 2.495431620238633e-05, | |
| "loss": 0.8584, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 17.928571428571427, | |
| "grad_norm": 2.534801483154297, | |
| "learning_rate": 2.495243372614682e-05, | |
| "loss": 0.9255, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "grad_norm": 2.785574197769165, | |
| "learning_rate": 2.495051331763462e-05, | |
| "loss": 0.8727, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 18.071428571428573, | |
| "grad_norm": 4.180792331695557, | |
| "learning_rate": 2.494855498269963e-05, | |
| "loss": 0.9683, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 18.142857142857142, | |
| "grad_norm": 5.634334564208984, | |
| "learning_rate": 2.4946558727307277e-05, | |
| "loss": 0.9485, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 18.214285714285715, | |
| "grad_norm": 4.938521862030029, | |
| "learning_rate": 2.4944524557538503e-05, | |
| "loss": 0.878, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 18.285714285714285, | |
| "grad_norm": 3.9465174674987793, | |
| "learning_rate": 2.4942452479589735e-05, | |
| "loss": 0.9731, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 18.357142857142858, | |
| "grad_norm": 3.601600170135498, | |
| "learning_rate": 2.494034249977289e-05, | |
| "loss": 0.9057, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 18.428571428571427, | |
| "grad_norm": 3.65264892578125, | |
| "learning_rate": 2.4938194624515333e-05, | |
| "loss": 0.9126, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 18.5, | |
| "grad_norm": 4.7837677001953125, | |
| "learning_rate": 2.4936008860359854e-05, | |
| "loss": 0.9847, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 18.571428571428573, | |
| "grad_norm": 4.8172831535339355, | |
| "learning_rate": 2.4933785213964677e-05, | |
| "loss": 0.9202, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 18.642857142857142, | |
| "grad_norm": 4.068408012390137, | |
| "learning_rate": 2.4931523692103418e-05, | |
| "loss": 0.9072, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 18.714285714285715, | |
| "grad_norm": 4.106720924377441, | |
| "learning_rate": 2.492922430166506e-05, | |
| "loss": 0.8837, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 18.785714285714285, | |
| "grad_norm": 3.7029213905334473, | |
| "learning_rate": 2.4926887049653943e-05, | |
| "loss": 0.9464, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 18.857142857142858, | |
| "grad_norm": 3.3589251041412354, | |
| "learning_rate": 2.492451194318975e-05, | |
| "loss": 0.8956, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 18.928571428571427, | |
| "grad_norm": 3.4119558334350586, | |
| "learning_rate": 2.4922098989507454e-05, | |
| "loss": 0.9177, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "grad_norm": 2.3078062534332275, | |
| "learning_rate": 2.4919648195957344e-05, | |
| "loss": 0.8086, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 19.071428571428573, | |
| "grad_norm": 2.213747978210449, | |
| "learning_rate": 2.4917159570004954e-05, | |
| "loss": 0.8417, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 19.142857142857142, | |
| "grad_norm": 2.629509687423706, | |
| "learning_rate": 2.491463311923108e-05, | |
| "loss": 0.9628, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 19.214285714285715, | |
| "grad_norm": 1.877100944519043, | |
| "learning_rate": 2.491206885133171e-05, | |
| "loss": 0.8174, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 19.285714285714285, | |
| "grad_norm": 2.653271436691284, | |
| "learning_rate": 2.490946677411807e-05, | |
| "loss": 0.8823, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 19.357142857142858, | |
| "grad_norm": 2.488171339035034, | |
| "learning_rate": 2.4906826895516528e-05, | |
| "loss": 0.888, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 19.428571428571427, | |
| "grad_norm": 2.5691208839416504, | |
| "learning_rate": 2.490414922356861e-05, | |
| "loss": 0.87, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 19.5, | |
| "grad_norm": 3.2644972801208496, | |
| "learning_rate": 2.4901433766430975e-05, | |
| "loss": 0.9013, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 19.571428571428573, | |
| "grad_norm": 5.025469779968262, | |
| "learning_rate": 2.4898680532375374e-05, | |
| "loss": 0.9013, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 19.642857142857142, | |
| "grad_norm": 3.607896566390991, | |
| "learning_rate": 2.489588952978863e-05, | |
| "loss": 0.8409, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 19.714285714285715, | |
| "grad_norm": 2.6993868350982666, | |
| "learning_rate": 2.4893060767172632e-05, | |
| "loss": 0.7845, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 19.785714285714285, | |
| "grad_norm": 2.4249427318573, | |
| "learning_rate": 2.489019425314427e-05, | |
| "loss": 0.8113, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 19.857142857142858, | |
| "grad_norm": 4.231166362762451, | |
| "learning_rate": 2.4887289996435452e-05, | |
| "loss": 0.8879, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 19.928571428571427, | |
| "grad_norm": 7.234214782714844, | |
| "learning_rate": 2.4884348005893045e-05, | |
| "loss": 0.8847, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "grad_norm": 4.228268146514893, | |
| "learning_rate": 2.488136829047886e-05, | |
| "loss": 0.8504, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 20.071428571428573, | |
| "grad_norm": 2.944539785385132, | |
| "learning_rate": 2.487835085926963e-05, | |
| "loss": 0.845, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 20.142857142857142, | |
| "grad_norm": 2.7270450592041016, | |
| "learning_rate": 2.487529572145697e-05, | |
| "loss": 0.8335, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 20.214285714285715, | |
| "grad_norm": 2.7035233974456787, | |
| "learning_rate": 2.4872202886347362e-05, | |
| "loss": 0.815, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 20.285714285714285, | |
| "grad_norm": 3.4698843955993652, | |
| "learning_rate": 2.486907236336212e-05, | |
| "loss": 0.875, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 20.357142857142858, | |
| "grad_norm": 4.334729194641113, | |
| "learning_rate": 2.4865904162037358e-05, | |
| "loss": 0.8447, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 20.428571428571427, | |
| "grad_norm": 3.5744895935058594, | |
| "learning_rate": 2.4862698292023963e-05, | |
| "loss": 0.792, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 20.5, | |
| "grad_norm": 3.5507006645202637, | |
| "learning_rate": 2.4859454763087577e-05, | |
| "loss": 0.7948, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 20.571428571428573, | |
| "grad_norm": 2.2967231273651123, | |
| "learning_rate": 2.4856173585108544e-05, | |
| "loss": 0.7975, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 20.642857142857142, | |
| "grad_norm": 2.5421581268310547, | |
| "learning_rate": 2.4852854768081912e-05, | |
| "loss": 0.8074, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 20.714285714285715, | |
| "grad_norm": 2.4942188262939453, | |
| "learning_rate": 2.4849498322117364e-05, | |
| "loss": 0.7962, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 20.785714285714285, | |
| "grad_norm": 4.169875621795654, | |
| "learning_rate": 2.4846104257439222e-05, | |
| "loss": 0.8181, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 20.857142857142858, | |
| "grad_norm": 6.13093376159668, | |
| "learning_rate": 2.4842672584386396e-05, | |
| "loss": 0.8434, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 20.928571428571427, | |
| "grad_norm": 3.3157453536987305, | |
| "learning_rate": 2.483920331341235e-05, | |
| "loss": 0.8108, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 21.0, | |
| "grad_norm": 3.0505266189575195, | |
| "learning_rate": 2.4835696455085093e-05, | |
| "loss": 0.7412, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 21.071428571428573, | |
| "grad_norm": 5.5504350662231445, | |
| "learning_rate": 2.483215202008712e-05, | |
| "loss": 0.8437, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 21.142857142857142, | |
| "grad_norm": 6.492306709289551, | |
| "learning_rate": 2.4828570019215396e-05, | |
| "loss": 0.8533, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 21.214285714285715, | |
| "grad_norm": 3.0854108333587646, | |
| "learning_rate": 2.4824950463381314e-05, | |
| "loss": 0.7881, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 21.285714285714285, | |
| "grad_norm": 3.006676435470581, | |
| "learning_rate": 2.482129336361067e-05, | |
| "loss": 0.8156, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 21.357142857142858, | |
| "grad_norm": 3.312669515609741, | |
| "learning_rate": 2.481759873104363e-05, | |
| "loss": 0.8011, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 21.428571428571427, | |
| "grad_norm": 5.346591949462891, | |
| "learning_rate": 2.4813866576934676e-05, | |
| "loss": 0.7895, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 21.5, | |
| "grad_norm": 5.8296074867248535, | |
| "learning_rate": 2.4810096912652604e-05, | |
| "loss": 0.893, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 21.571428571428573, | |
| "grad_norm": 3.022014617919922, | |
| "learning_rate": 2.480628974968046e-05, | |
| "loss": 0.7875, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 21.642857142857142, | |
| "grad_norm": 3.372377395629883, | |
| "learning_rate": 2.4802445099615525e-05, | |
| "loss": 0.8107, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 21.714285714285715, | |
| "grad_norm": 4.238394260406494, | |
| "learning_rate": 2.479856297416927e-05, | |
| "loss": 0.7879, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 21.785714285714285, | |
| "grad_norm": 4.552757263183594, | |
| "learning_rate": 2.4794643385167327e-05, | |
| "loss": 0.8352, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 21.857142857142858, | |
| "grad_norm": 4.247888088226318, | |
| "learning_rate": 2.4790686344549436e-05, | |
| "loss": 0.7733, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 21.928571428571427, | |
| "grad_norm": 2.4774329662323, | |
| "learning_rate": 2.478669186436943e-05, | |
| "loss": 0.7338, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 22.0, | |
| "grad_norm": 2.973707675933838, | |
| "learning_rate": 2.478265995679519e-05, | |
| "loss": 0.7532, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 22.071428571428573, | |
| "grad_norm": 5.437268257141113, | |
| "learning_rate": 2.4778590634108613e-05, | |
| "loss": 0.8951, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 22.142857142857142, | |
| "grad_norm": 5.616093635559082, | |
| "learning_rate": 2.4774483908705546e-05, | |
| "loss": 0.8117, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 22.214285714285715, | |
| "grad_norm": 3.7091429233551025, | |
| "learning_rate": 2.4770339793095802e-05, | |
| "loss": 0.8166, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 22.285714285714285, | |
| "grad_norm": 3.873770236968994, | |
| "learning_rate": 2.4766158299903062e-05, | |
| "loss": 0.7786, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 22.357142857142858, | |
| "grad_norm": 3.692065477371216, | |
| "learning_rate": 2.4761939441864895e-05, | |
| "loss": 0.8148, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 22.428571428571427, | |
| "grad_norm": 4.50544548034668, | |
| "learning_rate": 2.4757683231832662e-05, | |
| "loss": 0.9163, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 22.5, | |
| "grad_norm": 4.464144229888916, | |
| "learning_rate": 2.4753389682771523e-05, | |
| "loss": 0.8223, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 22.571428571428573, | |
| "grad_norm": 3.729602575302124, | |
| "learning_rate": 2.474905880776037e-05, | |
| "loss": 0.8479, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 22.642857142857142, | |
| "grad_norm": 3.674757957458496, | |
| "learning_rate": 2.47446906199918e-05, | |
| "loss": 0.7319, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 22.714285714285715, | |
| "grad_norm": 4.361830711364746, | |
| "learning_rate": 2.4740285132772072e-05, | |
| "loss": 0.8054, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 22.785714285714285, | |
| "grad_norm": 4.365119934082031, | |
| "learning_rate": 2.4735842359521064e-05, | |
| "loss": 0.801, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 22.857142857142858, | |
| "grad_norm": 4.092256546020508, | |
| "learning_rate": 2.4731362313772233e-05, | |
| "loss": 0.8389, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 22.928571428571427, | |
| "grad_norm": 3.239090919494629, | |
| "learning_rate": 2.4726845009172572e-05, | |
| "loss": 0.7814, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 23.0, | |
| "grad_norm": 2.6484477519989014, | |
| "learning_rate": 2.4722290459482578e-05, | |
| "loss": 0.7468, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 23.071428571428573, | |
| "grad_norm": 3.275247573852539, | |
| "learning_rate": 2.47176986785762e-05, | |
| "loss": 0.827, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 23.142857142857142, | |
| "grad_norm": 3.040330410003662, | |
| "learning_rate": 2.47130696804408e-05, | |
| "loss": 0.7849, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 23.214285714285715, | |
| "grad_norm": 4.0800395011901855, | |
| "learning_rate": 2.47084034791771e-05, | |
| "loss": 0.856, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 23.285714285714285, | |
| "grad_norm": 3.5290443897247314, | |
| "learning_rate": 2.4703700088999167e-05, | |
| "loss": 0.825, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 23.357142857142858, | |
| "grad_norm": 3.670090436935425, | |
| "learning_rate": 2.4698959524234346e-05, | |
| "loss": 0.8061, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 23.428571428571427, | |
| "grad_norm": 1.9602779150009155, | |
| "learning_rate": 2.4694181799323206e-05, | |
| "loss": 0.7803, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 23.5, | |
| "grad_norm": 3.8044676780700684, | |
| "learning_rate": 2.468936692881954e-05, | |
| "loss": 0.767, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 23.571428571428573, | |
| "grad_norm": 3.5850207805633545, | |
| "learning_rate": 2.4684514927390274e-05, | |
| "loss": 0.7555, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 23.642857142857142, | |
| "grad_norm": 2.4394097328186035, | |
| "learning_rate": 2.4679625809815443e-05, | |
| "loss": 0.7911, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 23.714285714285715, | |
| "grad_norm": 3.6944406032562256, | |
| "learning_rate": 2.467469959098815e-05, | |
| "loss": 0.7708, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 23.785714285714285, | |
| "grad_norm": 3.9482243061065674, | |
| "learning_rate": 2.4669736285914505e-05, | |
| "loss": 0.8035, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 23.857142857142858, | |
| "grad_norm": 4.452454566955566, | |
| "learning_rate": 2.4664735909713606e-05, | |
| "loss": 0.7837, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 23.928571428571427, | |
| "grad_norm": 3.8159029483795166, | |
| "learning_rate": 2.465969847761746e-05, | |
| "loss": 0.8188, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "grad_norm": 2.743255138397217, | |
| "learning_rate": 2.4654624004970957e-05, | |
| "loss": 0.7004, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 24.071428571428573, | |
| "grad_norm": 3.636960506439209, | |
| "learning_rate": 2.464951250723183e-05, | |
| "loss": 0.7809, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 24.142857142857142, | |
| "grad_norm": 3.9498450756073, | |
| "learning_rate": 2.4644363999970576e-05, | |
| "loss": 0.7903, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 24.214285714285715, | |
| "grad_norm": 3.0506386756896973, | |
| "learning_rate": 2.4639178498870452e-05, | |
| "loss": 0.8114, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 24.285714285714285, | |
| "grad_norm": 3.2507994174957275, | |
| "learning_rate": 2.4633956019727385e-05, | |
| "loss": 0.7556, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 24.357142857142858, | |
| "grad_norm": 5.005919933319092, | |
| "learning_rate": 2.4628696578449956e-05, | |
| "loss": 0.8456, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 24.428571428571427, | |
| "grad_norm": 3.5123672485351562, | |
| "learning_rate": 2.4623400191059335e-05, | |
| "loss": 0.7921, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 24.5, | |
| "grad_norm": 3.3725931644439697, | |
| "learning_rate": 2.4618066873689238e-05, | |
| "loss": 0.7986, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 24.571428571428573, | |
| "grad_norm": 2.6177029609680176, | |
| "learning_rate": 2.461269664258587e-05, | |
| "loss": 0.7449, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 24.642857142857142, | |
| "grad_norm": 5.391937732696533, | |
| "learning_rate": 2.4607289514107888e-05, | |
| "loss": 0.7433, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 24.714285714285715, | |
| "grad_norm": 2.888105630874634, | |
| "learning_rate": 2.460184550472635e-05, | |
| "loss": 0.7079, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 24.785714285714285, | |
| "grad_norm": 2.3010149002075195, | |
| "learning_rate": 2.4596364631024643e-05, | |
| "loss": 0.7376, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 24.857142857142858, | |
| "grad_norm": 3.590585470199585, | |
| "learning_rate": 2.459084690969846e-05, | |
| "loss": 0.7532, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 24.928571428571427, | |
| "grad_norm": 6.115037441253662, | |
| "learning_rate": 2.4585292357555746e-05, | |
| "loss": 0.7568, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "grad_norm": 2.7985963821411133, | |
| "learning_rate": 2.457970099151662e-05, | |
| "loss": 0.7001, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 25.071428571428573, | |
| "grad_norm": 5.373832702636719, | |
| "learning_rate": 2.4574072828613354e-05, | |
| "loss": 0.8254, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 25.142857142857142, | |
| "grad_norm": 3.988442897796631, | |
| "learning_rate": 2.4568407885990313e-05, | |
| "loss": 0.7305, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 25.214285714285715, | |
| "grad_norm": 3.5456085205078125, | |
| "learning_rate": 2.4562706180903894e-05, | |
| "loss": 0.7546, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 25.285714285714285, | |
| "grad_norm": 3.8878087997436523, | |
| "learning_rate": 2.4556967730722478e-05, | |
| "loss": 0.794, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 25.357142857142858, | |
| "grad_norm": 3.997696876525879, | |
| "learning_rate": 2.455119255292638e-05, | |
| "loss": 0.7983, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 25.428571428571427, | |
| "grad_norm": 5.822238445281982, | |
| "learning_rate": 2.4545380665107786e-05, | |
| "loss": 0.899, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 25.5, | |
| "grad_norm": 3.2456319332122803, | |
| "learning_rate": 2.453953208497073e-05, | |
| "loss": 0.7284, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 25.571428571428573, | |
| "grad_norm": 3.3598856925964355, | |
| "learning_rate": 2.4533646830330986e-05, | |
| "loss": 0.8056, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 25.642857142857142, | |
| "grad_norm": 3.6620326042175293, | |
| "learning_rate": 2.452772491911607e-05, | |
| "loss": 0.7994, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 25.714285714285715, | |
| "grad_norm": 4.005545139312744, | |
| "learning_rate": 2.4521766369365142e-05, | |
| "loss": 0.7773, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 25.785714285714285, | |
| "grad_norm": 4.023702144622803, | |
| "learning_rate": 2.4515771199228987e-05, | |
| "loss": 0.7717, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 25.857142857142858, | |
| "grad_norm": 1.9900436401367188, | |
| "learning_rate": 2.450973942696993e-05, | |
| "loss": 0.7331, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 25.928571428571427, | |
| "grad_norm": 5.718785285949707, | |
| "learning_rate": 2.450367107096179e-05, | |
| "loss": 0.7818, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 26.0, | |
| "grad_norm": 3.5384066104888916, | |
| "learning_rate": 2.449756614968984e-05, | |
| "loss": 0.8165, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 26.071428571428573, | |
| "grad_norm": 2.2805933952331543, | |
| "learning_rate": 2.449142468175072e-05, | |
| "loss": 0.68, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 26.142857142857142, | |
| "grad_norm": 2.818986654281616, | |
| "learning_rate": 2.4485246685852413e-05, | |
| "loss": 0.6765, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 26.214285714285715, | |
| "grad_norm": 3.089205741882324, | |
| "learning_rate": 2.4479032180814166e-05, | |
| "loss": 0.6901, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 26.285714285714285, | |
| "grad_norm": 2.990636110305786, | |
| "learning_rate": 2.447278118556644e-05, | |
| "loss": 0.6839, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 26.357142857142858, | |
| "grad_norm": 2.0169079303741455, | |
| "learning_rate": 2.446649371915084e-05, | |
| "loss": 0.6499, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 26.428571428571427, | |
| "grad_norm": 2.073349714279175, | |
| "learning_rate": 2.4460169800720095e-05, | |
| "loss": 0.6312, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 26.5, | |
| "grad_norm": 5.544507026672363, | |
| "learning_rate": 2.4453809449537947e-05, | |
| "loss": 0.7886, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 26.571428571428573, | |
| "grad_norm": 3.550513505935669, | |
| "learning_rate": 2.4447412684979127e-05, | |
| "loss": 0.7682, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 26.642857142857142, | |
| "grad_norm": 4.850046634674072, | |
| "learning_rate": 2.4440979526529295e-05, | |
| "loss": 0.7891, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 26.714285714285715, | |
| "grad_norm": 5.702778339385986, | |
| "learning_rate": 2.4434509993784972e-05, | |
| "loss": 0.7478, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 26.785714285714285, | |
| "grad_norm": 2.3614633083343506, | |
| "learning_rate": 2.4428004106453462e-05, | |
| "loss": 0.6622, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 26.857142857142858, | |
| "grad_norm": 4.546297073364258, | |
| "learning_rate": 2.4421461884352836e-05, | |
| "loss": 0.7229, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 26.928571428571427, | |
| "grad_norm": 4.9091386795043945, | |
| "learning_rate": 2.4414883347411836e-05, | |
| "loss": 0.7186, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 27.0, | |
| "grad_norm": 3.949092388153076, | |
| "learning_rate": 2.440826851566983e-05, | |
| "loss": 0.7257, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 27.071428571428573, | |
| "grad_norm": 5.0857439041137695, | |
| "learning_rate": 2.4401617409276735e-05, | |
| "loss": 0.6953, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 27.142857142857142, | |
| "grad_norm": 4.325730323791504, | |
| "learning_rate": 2.439493004849298e-05, | |
| "loss": 0.6696, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 27.214285714285715, | |
| "grad_norm": 3.15690541267395, | |
| "learning_rate": 2.438820645368942e-05, | |
| "loss": 0.7532, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 27.285714285714285, | |
| "grad_norm": 4.08896017074585, | |
| "learning_rate": 2.4381446645347297e-05, | |
| "loss": 0.7039, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 27.357142857142858, | |
| "grad_norm": 3.996779441833496, | |
| "learning_rate": 2.4374650644058156e-05, | |
| "loss": 0.7489, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 27.428571428571427, | |
| "grad_norm": 4.629286766052246, | |
| "learning_rate": 2.43678184705238e-05, | |
| "loss": 0.7634, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 27.5, | |
| "grad_norm": 4.668010234832764, | |
| "learning_rate": 2.4360950145556208e-05, | |
| "loss": 0.7321, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 27.571428571428573, | |
| "grad_norm": 2.8770227432250977, | |
| "learning_rate": 2.4354045690077492e-05, | |
| "loss": 0.6417, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 27.642857142857142, | |
| "grad_norm": 4.506302356719971, | |
| "learning_rate": 2.4347105125119824e-05, | |
| "loss": 0.6698, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 27.714285714285715, | |
| "grad_norm": 2.6643428802490234, | |
| "learning_rate": 2.4340128471825362e-05, | |
| "loss": 0.6938, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 27.785714285714285, | |
| "grad_norm": 2.8056280612945557, | |
| "learning_rate": 2.4333115751446208e-05, | |
| "loss": 0.6743, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 27.857142857142858, | |
| "grad_norm": 2.7733471393585205, | |
| "learning_rate": 2.4326066985344318e-05, | |
| "loss": 0.6573, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 27.928571428571427, | |
| "grad_norm": 5.0971174240112305, | |
| "learning_rate": 2.4318982194991463e-05, | |
| "loss": 0.6754, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "grad_norm": 6.083596706390381, | |
| "learning_rate": 2.4311861401969138e-05, | |
| "loss": 0.7646, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 28.071428571428573, | |
| "grad_norm": 4.150708198547363, | |
| "learning_rate": 2.4304704627968515e-05, | |
| "loss": 0.6958, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 28.142857142857142, | |
| "grad_norm": 3.6580939292907715, | |
| "learning_rate": 2.429751189479037e-05, | |
| "loss": 0.6848, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 28.214285714285715, | |
| "grad_norm": 2.1017942428588867, | |
| "learning_rate": 2.429028322434501e-05, | |
| "loss": 0.6576, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 28.285714285714285, | |
| "grad_norm": 1.8062525987625122, | |
| "learning_rate": 2.4283018638652234e-05, | |
| "loss": 0.6963, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 28.357142857142858, | |
| "grad_norm": 3.1975321769714355, | |
| "learning_rate": 2.427571815984121e-05, | |
| "loss": 0.6599, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 28.428571428571427, | |
| "grad_norm": 3.0259714126586914, | |
| "learning_rate": 2.4268381810150474e-05, | |
| "loss": 0.6959, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 28.5, | |
| "grad_norm": 5.031489849090576, | |
| "learning_rate": 2.426100961192782e-05, | |
| "loss": 0.7208, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 28.571428571428573, | |
| "grad_norm": 3.5637011528015137, | |
| "learning_rate": 2.4253601587630236e-05, | |
| "loss": 0.6896, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 28.642857142857142, | |
| "grad_norm": 3.333253860473633, | |
| "learning_rate": 2.4246157759823855e-05, | |
| "loss": 0.6815, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 28.714285714285715, | |
| "grad_norm": 3.422375440597534, | |
| "learning_rate": 2.4238678151183863e-05, | |
| "loss": 0.6606, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 28.785714285714285, | |
| "grad_norm": 2.458256721496582, | |
| "learning_rate": 2.423116278449445e-05, | |
| "loss": 0.6484, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 28.857142857142858, | |
| "grad_norm": 4.026810169219971, | |
| "learning_rate": 2.4223611682648724e-05, | |
| "loss": 0.7167, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 28.928571428571427, | |
| "grad_norm": 3.2683417797088623, | |
| "learning_rate": 2.4216024868648644e-05, | |
| "loss": 0.6899, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 29.0, | |
| "grad_norm": 3.54608416557312, | |
| "learning_rate": 2.4208402365604972e-05, | |
| "loss": 0.6843, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 29.071428571428573, | |
| "grad_norm": 4.817044734954834, | |
| "learning_rate": 2.420074419673717e-05, | |
| "loss": 0.6544, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 29.142857142857142, | |
| "grad_norm": 3.0443451404571533, | |
| "learning_rate": 2.4193050385373344e-05, | |
| "loss": 0.6134, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 29.214285714285715, | |
| "grad_norm": 6.059633731842041, | |
| "learning_rate": 2.418532095495018e-05, | |
| "loss": 0.6718, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 29.285714285714285, | |
| "grad_norm": 2.560316324234009, | |
| "learning_rate": 2.417755592901287e-05, | |
| "loss": 0.6134, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 29.357142857142858, | |
| "grad_norm": 3.06530499458313, | |
| "learning_rate": 2.4169755331215023e-05, | |
| "loss": 0.6158, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 29.428571428571427, | |
| "grad_norm": 2.439854621887207, | |
| "learning_rate": 2.4161919185318617e-05, | |
| "loss": 0.6428, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 29.5, | |
| "grad_norm": 4.74956750869751, | |
| "learning_rate": 2.4154047515193904e-05, | |
| "loss": 0.6847, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 29.571428571428573, | |
| "grad_norm": 3.4892289638519287, | |
| "learning_rate": 2.4146140344819363e-05, | |
| "loss": 0.6293, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 29.642857142857142, | |
| "grad_norm": 2.6727051734924316, | |
| "learning_rate": 2.4138197698281606e-05, | |
| "loss": 0.6429, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 29.714285714285715, | |
| "grad_norm": 4.598572254180908, | |
| "learning_rate": 2.413021959977531e-05, | |
| "loss": 0.6504, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 29.785714285714285, | |
| "grad_norm": 5.181699752807617, | |
| "learning_rate": 2.4122206073603142e-05, | |
| "loss": 0.6994, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 29.857142857142858, | |
| "grad_norm": 4.484119415283203, | |
| "learning_rate": 2.4114157144175703e-05, | |
| "loss": 0.7659, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 29.928571428571427, | |
| "grad_norm": 4.818487644195557, | |
| "learning_rate": 2.4106072836011422e-05, | |
| "loss": 0.7583, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "grad_norm": 3.408137083053589, | |
| "learning_rate": 2.40979531737365e-05, | |
| "loss": 0.6579, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 30.071428571428573, | |
| "grad_norm": 4.11182975769043, | |
| "learning_rate": 2.4089798182084845e-05, | |
| "loss": 0.69, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 30.142857142857142, | |
| "grad_norm": 4.751646518707275, | |
| "learning_rate": 2.4081607885897966e-05, | |
| "loss": 0.7584, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 30.214285714285715, | |
| "grad_norm": 5.190206050872803, | |
| "learning_rate": 2.407338231012494e-05, | |
| "loss": 0.7468, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 30.285714285714285, | |
| "grad_norm": 3.750694513320923, | |
| "learning_rate": 2.406512147982228e-05, | |
| "loss": 0.743, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 30.357142857142858, | |
| "grad_norm": 3.4604289531707764, | |
| "learning_rate": 2.4056825420153917e-05, | |
| "loss": 0.7051, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 30.428571428571427, | |
| "grad_norm": 2.6612353324890137, | |
| "learning_rate": 2.4048494156391087e-05, | |
| "loss": 0.6852, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 30.5, | |
| "grad_norm": 4.154074192047119, | |
| "learning_rate": 2.4040127713912264e-05, | |
| "loss": 0.6733, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 30.571428571428573, | |
| "grad_norm": 3.900482654571533, | |
| "learning_rate": 2.403172611820308e-05, | |
| "loss": 0.6134, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 30.642857142857142, | |
| "grad_norm": 5.209915637969971, | |
| "learning_rate": 2.4023289394856257e-05, | |
| "loss": 0.7324, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 30.714285714285715, | |
| "grad_norm": 2.668140411376953, | |
| "learning_rate": 2.401481756957152e-05, | |
| "loss": 0.6655, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 30.785714285714285, | |
| "grad_norm": 3.480163812637329, | |
| "learning_rate": 2.4006310668155508e-05, | |
| "loss": 0.7295, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 30.857142857142858, | |
| "grad_norm": 2.9695725440979004, | |
| "learning_rate": 2.3997768716521723e-05, | |
| "loss": 0.6949, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 30.928571428571427, | |
| "grad_norm": 2.298661470413208, | |
| "learning_rate": 2.398919174069043e-05, | |
| "loss": 0.6687, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 31.0, | |
| "grad_norm": 3.2847728729248047, | |
| "learning_rate": 2.398057976678859e-05, | |
| "loss": 0.6779, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 31.071428571428573, | |
| "grad_norm": 4.464836120605469, | |
| "learning_rate": 2.3971932821049765e-05, | |
| "loss": 0.7223, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 31.142857142857142, | |
| "grad_norm": 2.811166763305664, | |
| "learning_rate": 2.396325092981405e-05, | |
| "loss": 0.6677, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 31.214285714285715, | |
| "grad_norm": 3.1231977939605713, | |
| "learning_rate": 2.3954534119527996e-05, | |
| "loss": 0.6745, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 31.285714285714285, | |
| "grad_norm": 1.7231147289276123, | |
| "learning_rate": 2.3945782416744517e-05, | |
| "loss": 0.5596, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 31.357142857142858, | |
| "grad_norm": 2.8085505962371826, | |
| "learning_rate": 2.3936995848122812e-05, | |
| "loss": 0.5903, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 31.428571428571427, | |
| "grad_norm": 5.913967132568359, | |
| "learning_rate": 2.3928174440428297e-05, | |
| "loss": 0.6269, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 31.5, | |
| "grad_norm": 3.01706862449646, | |
| "learning_rate": 2.391931822053251e-05, | |
| "loss": 0.6192, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 31.571428571428573, | |
| "grad_norm": 7.216449737548828, | |
| "learning_rate": 2.3910427215413036e-05, | |
| "loss": 0.6928, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 31.642857142857142, | |
| "grad_norm": 3.636003255844116, | |
| "learning_rate": 2.390150145215341e-05, | |
| "loss": 0.6932, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 31.714285714285715, | |
| "grad_norm": 2.978515625, | |
| "learning_rate": 2.3892540957943067e-05, | |
| "loss": 0.6355, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 31.785714285714285, | |
| "grad_norm": 2.5752205848693848, | |
| "learning_rate": 2.3883545760077215e-05, | |
| "loss": 0.6208, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 31.857142857142858, | |
| "grad_norm": 2.671752452850342, | |
| "learning_rate": 2.3874515885956792e-05, | |
| "loss": 0.6078, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 31.928571428571427, | |
| "grad_norm": 3.273820400238037, | |
| "learning_rate": 2.386545136308836e-05, | |
| "loss": 0.6566, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "grad_norm": 3.485921859741211, | |
| "learning_rate": 2.3856352219084024e-05, | |
| "loss": 0.606, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 32.07142857142857, | |
| "grad_norm": 2.268240451812744, | |
| "learning_rate": 2.384721848166136e-05, | |
| "loss": 0.5724, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 32.142857142857146, | |
| "grad_norm": 4.0276689529418945, | |
| "learning_rate": 2.3838050178643312e-05, | |
| "loss": 0.5929, | |
| "step": 450 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 2000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 143, | |
| "save_steps": 200, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": false, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 64, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |