| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.0, | |
| "eval_steps": 500, | |
| "global_step": 1088, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.001838235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.0000000000000002e-07, | |
| "loss": 1.9627, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.003676470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.0000000000000003e-07, | |
| "loss": 1.6036, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.0055147058823529415, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.000000000000001e-07, | |
| "loss": 1.7332, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.007352941176470588, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.000000000000001e-07, | |
| "loss": 1.8943, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.009191176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0000000000000002e-06, | |
| "loss": 1.9555, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.011029411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2000000000000002e-06, | |
| "loss": 1.972, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.012867647058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4000000000000001e-06, | |
| "loss": 1.7141, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.014705882352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6000000000000001e-06, | |
| "loss": 1.8038, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.016544117647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.8000000000000001e-06, | |
| "loss": 1.9202, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.01838235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.0000000000000003e-06, | |
| "loss": 1.6748, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.02022058823529412, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.2e-06, | |
| "loss": 1.8197, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.022058823529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.4000000000000003e-06, | |
| "loss": 1.7396, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.02389705882352941, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.6e-06, | |
| "loss": 1.654, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.025735294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.8000000000000003e-06, | |
| "loss": 1.7685, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.027573529411764705, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3e-06, | |
| "loss": 1.5536, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.029411764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.2000000000000003e-06, | |
| "loss": 1.7035, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.03125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.4000000000000005e-06, | |
| "loss": 1.9268, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.03308823529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.6000000000000003e-06, | |
| "loss": 2.0697, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.034926470588235295, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.8000000000000005e-06, | |
| "loss": 1.9225, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.03676470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 1.8257, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.03860294117647059, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.2000000000000004e-06, | |
| "loss": 1.8509, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.04044117647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.4e-06, | |
| "loss": 1.8072, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.042279411764705885, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.600000000000001e-06, | |
| "loss": 1.7604, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.04411764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.800000000000001e-06, | |
| "loss": 1.7735, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.04595588235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5e-06, | |
| "loss": 1.8244, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.04779411764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.999989082004443e-06, | |
| "loss": 1.9756, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.04963235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.999956328113134e-06, | |
| "loss": 1.6347, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.051470588235294115, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.999901738612159e-06, | |
| "loss": 1.817, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.05330882352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.999825313978322e-06, | |
| "loss": 1.6143, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.05514705882352941, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.999727054879149e-06, | |
| "loss": 1.7087, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.05698529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.999606962172872e-06, | |
| "loss": 1.9148, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.058823529411764705, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.999465036908429e-06, | |
| "loss": 1.7264, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.06066176470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.999301280325452e-06, | |
| "loss": 1.5771, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.0625, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.999115693854255e-06, | |
| "loss": 1.5012, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.06433823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.998908279115825e-06, | |
| "loss": 1.8459, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.0661764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.998679037921803e-06, | |
| "loss": 1.7001, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.06801470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.998427972274473e-06, | |
| "loss": 1.719, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.06985294117647059, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.998155084366744e-06, | |
| "loss": 1.9945, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.07169117647058823, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.997860376582123e-06, | |
| "loss": 1.8024, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.07352941176470588, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.997543851494709e-06, | |
| "loss": 1.7099, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.07536764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.9972055118691545e-06, | |
| "loss": 1.5121, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.07720588235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.996845360660652e-06, | |
| "loss": 1.7949, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.07904411764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.996463401014908e-06, | |
| "loss": 1.6773, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.08088235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.9960596362681054e-06, | |
| "loss": 1.7022, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.08272058823529412, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.9956340699468896e-06, | |
| "loss": 1.7003, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.08455882352941177, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.995186705768322e-06, | |
| "loss": 1.5486, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.08639705882352941, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.9947175476398606e-06, | |
| "loss": 1.7186, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.08823529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.994226599659319e-06, | |
| "loss": 1.6925, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.0900735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.993713866114829e-06, | |
| "loss": 1.9749, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.09191176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.993179351484811e-06, | |
| "loss": 2.0403, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.09375, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.9926230604379275e-06, | |
| "loss": 1.8667, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.09558823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.992044997833044e-06, | |
| "loss": 1.8817, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.0974264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.991445168719189e-06, | |
| "loss": 1.909, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.09926470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.9908235783355075e-06, | |
| "loss": 1.6816, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.10110294117647059, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.990180232111217e-06, | |
| "loss": 1.627, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.10294117647058823, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.989515135665558e-06, | |
| "loss": 1.7085, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.10477941176470588, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.988828294807746e-06, | |
| "loss": 1.884, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.10661764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.988119715536922e-06, | |
| "loss": 1.7246, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.10845588235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.987389404042098e-06, | |
| "loss": 1.7751, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.11029411764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.986637366702105e-06, | |
| "loss": 1.7463, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.11213235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.985863610085534e-06, | |
| "loss": 1.695, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.11397058823529412, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.985068140950683e-06, | |
| "loss": 1.8484, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.11580882352941177, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.984250966245495e-06, | |
| "loss": 1.9041, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.11764705882352941, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.983412093107496e-06, | |
| "loss": 1.9226, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.11948529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.982551528863738e-06, | |
| "loss": 1.8654, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.1213235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.981669281030731e-06, | |
| "loss": 1.7118, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.12316176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.980765357314376e-06, | |
| "loss": 1.7896, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.9798397656099005e-06, | |
| "loss": 1.5778, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.12683823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.978892514001792e-06, | |
| "loss": 1.9112, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.12867647058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.977923610763719e-06, | |
| "loss": 1.8526, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.13051470588235295, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.976933064358467e-06, | |
| "loss": 1.6893, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.1323529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.975920883437862e-06, | |
| "loss": 1.8083, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.13419117647058823, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.974887076842694e-06, | |
| "loss": 1.9307, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.13602941176470587, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.973831653602637e-06, | |
| "loss": 1.828, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.13786764705882354, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.972754622936178e-06, | |
| "loss": 1.7038, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.13970588235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.971655994250529e-06, | |
| "loss": 1.6064, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.14154411764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.97053577714155e-06, | |
| "loss": 1.7252, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.14338235294117646, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.96939398139366e-06, | |
| "loss": 1.979, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.14522058823529413, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.968230616979755e-06, | |
| "loss": 1.9525, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.14705882352941177, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.967045694061122e-06, | |
| "loss": 1.912, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.1488970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.965839222987348e-06, | |
| "loss": 1.6298, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.15073529411764705, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.9646112142962295e-06, | |
| "loss": 1.6537, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.15257352941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.96336167871368e-06, | |
| "loss": 1.9089, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.15441176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.96209062715364e-06, | |
| "loss": 1.8703, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.15625, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.960798070717977e-06, | |
| "loss": 1.9559, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.15808823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.959484020696392e-06, | |
| "loss": 1.8444, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.15992647058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.9581484885663175e-06, | |
| "loss": 1.6396, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.16176470588235295, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.956791485992823e-06, | |
| "loss": 1.9488, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.1636029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.955413024828504e-06, | |
| "loss": 1.8038, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.16544117647058823, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.9540131171133884e-06, | |
| "loss": 1.7477, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.16727941176470587, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.952591775074825e-06, | |
| "loss": 1.7757, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.16911764705882354, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.951149011127379e-06, | |
| "loss": 1.7452, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.17095588235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.949684837872723e-06, | |
| "loss": 1.6137, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.17279411764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.948199268099525e-06, | |
| "loss": 1.8074, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.17463235294117646, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.946692314783342e-06, | |
| "loss": 1.7006, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.17647058823529413, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.9451639910865016e-06, | |
| "loss": 1.6746, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.17830882352941177, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.943614310357987e-06, | |
| "loss": 1.5338, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.1801470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.942043286133326e-06, | |
| "loss": 1.639, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.18198529411764705, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.940450932134467e-06, | |
| "loss": 1.8445, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.18382352941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.9388372622696605e-06, | |
| "loss": 1.6577, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.18566176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.937202290633337e-06, | |
| "loss": 1.9322, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.1875, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.935546031505991e-06, | |
| "loss": 1.8592, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.18933823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.933868499354043e-06, | |
| "loss": 1.8238, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.19117647058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.932169708829725e-06, | |
| "loss": 1.8892, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.19301470588235295, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.930449674770947e-06, | |
| "loss": 1.6668, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.1948529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.928708412201169e-06, | |
| "loss": 2.0112, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.19669117647058823, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.926945936329266e-06, | |
| "loss": 1.8705, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.19852941176470587, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.925162262549405e-06, | |
| "loss": 1.8025, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.20036764705882354, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.923357406440896e-06, | |
| "loss": 1.9824, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.20220588235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.921531383768071e-06, | |
| "loss": 1.6375, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.20404411764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.919684210480134e-06, | |
| "loss": 1.9491, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.20588235294117646, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.917815902711029e-06, | |
| "loss": 2.0238, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.20772058823529413, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.915926476779297e-06, | |
| "loss": 1.8125, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.20955882352941177, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.914015949187934e-06, | |
| "loss": 1.7428, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.2113970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.912084336624243e-06, | |
| "loss": 1.6849, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.21323529411764705, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.910131655959697e-06, | |
| "loss": 1.8793, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.21507352941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.908157924249781e-06, | |
| "loss": 1.9526, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.21691176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.906163158733851e-06, | |
| "loss": 1.7916, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.21875, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.904147376834979e-06, | |
| "loss": 1.8987, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.22058823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.9021105961598046e-06, | |
| "loss": 1.877, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.22242647058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.900052834498377e-06, | |
| "loss": 1.7896, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.22426470588235295, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.897974109824002e-06, | |
| "loss": 1.8914, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.2261029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.895874440293085e-06, | |
| "loss": 1.7302, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.22794117647058823, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.8937538442449724e-06, | |
| "loss": 1.807, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.22977941176470587, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.891612340201791e-06, | |
| "loss": 1.6097, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.23161764705882354, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.8894499468682865e-06, | |
| "loss": 1.9383, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.23345588235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.887266683131659e-06, | |
| "loss": 1.6959, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.23529411764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.885062568061399e-06, | |
| "loss": 1.9403, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.23713235294117646, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.882837620909121e-06, | |
| "loss": 1.6888, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.23897058823529413, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.880591861108397e-06, | |
| "loss": 1.7798, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.24080882352941177, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.878325308274583e-06, | |
| "loss": 1.796, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.2426470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.876037982204649e-06, | |
| "loss": 1.6234, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.24448529411764705, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.873729902877009e-06, | |
| "loss": 1.8065, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.24632352941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.871401090451342e-06, | |
| "loss": 1.9266, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.24816176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.869051565268419e-06, | |
| "loss": 1.772, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.866681347849925e-06, | |
| "loss": 1.5869, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.25183823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.8642904588982785e-06, | |
| "loss": 1.9012, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.2536764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.861878919296451e-06, | |
| "loss": 1.9242, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.2555147058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.859446750107786e-06, | |
| "loss": 2.0885, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.25735294117647056, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.856993972575813e-06, | |
| "loss": 1.5305, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.25919117647058826, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.854520608124063e-06, | |
| "loss": 1.8923, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.2610294117647059, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.8520266783558825e-06, | |
| "loss": 1.8581, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.26286764705882354, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.849512205054242e-06, | |
| "loss": 1.6467, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.2647058823529412, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.846977210181549e-06, | |
| "loss": 1.8146, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.2665441176470588, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.844421715879453e-06, | |
| "loss": 1.555, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.26838235294117646, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.841845744468655e-06, | |
| "loss": 1.7029, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.2702205882352941, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.83924931844871e-06, | |
| "loss": 1.7241, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.27205882352941174, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.836632460497832e-06, | |
| "loss": 1.667, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.27389705882352944, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.833995193472697e-06, | |
| "loss": 1.5294, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.2757352941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.831337540408239e-06, | |
| "loss": 1.7341, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.2775735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.828659524517455e-06, | |
| "loss": 1.7731, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.27941176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.825961169191196e-06, | |
| "loss": 1.891, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.28125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.8232424979979684e-06, | |
| "loss": 1.5459, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.28308823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.820503534683725e-06, | |
| "loss": 1.7663, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.2849264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.8177443031716545e-06, | |
| "loss": 1.9843, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.2867647058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.814964827561981e-06, | |
| "loss": 1.9345, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.28860294117647056, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.812165132131746e-06, | |
| "loss": 1.651, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.29044117647058826, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.809345241334598e-06, | |
| "loss": 1.7562, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.2922794117647059, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.806505179800583e-06, | |
| "loss": 1.7144, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.29411764705882354, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.803644972335925e-06, | |
| "loss": 1.8868, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.2959558823529412, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.800764643922806e-06, | |
| "loss": 1.7201, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.2977941176470588, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.797864219719161e-06, | |
| "loss": 2.0389, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.29963235294117646, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.794943725058441e-06, | |
| "loss": 1.7262, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.3014705882352941, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.792003185449406e-06, | |
| "loss": 1.8069, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.30330882352941174, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.789042626575895e-06, | |
| "loss": 1.8573, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.30514705882352944, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.786062074296602e-06, | |
| "loss": 1.7839, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.3069852941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.783061554644853e-06, | |
| "loss": 1.8461, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.3088235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.780041093828376e-06, | |
| "loss": 1.7538, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.31066176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.777000718229072e-06, | |
| "loss": 1.6497, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.3125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.773940454402789e-06, | |
| "loss": 1.6723, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.31433823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.770860329079083e-06, | |
| "loss": 1.8927, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.3161764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.7677603691609905e-06, | |
| "loss": 1.7236, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.3180147058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.7646406017247895e-06, | |
| "loss": 1.971, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.31985294117647056, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.761501054019766e-06, | |
| "loss": 1.8082, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.32169117647058826, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.758341753467975e-06, | |
| "loss": 1.9078, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.3235294117647059, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.755162727663998e-06, | |
| "loss": 1.6387, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.32536764705882354, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.751964004374709e-06, | |
| "loss": 2.0215, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.3272058823529412, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.748745611539024e-06, | |
| "loss": 1.8042, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.3290441176470588, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.745507577267663e-06, | |
| "loss": 2.1742, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.33088235294117646, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.7422499298429e-06, | |
| "loss": 1.8744, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.3327205882352941, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.738972697718319e-06, | |
| "loss": 1.7443, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.33455882352941174, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.735675909518565e-06, | |
| "loss": 1.9355, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.33639705882352944, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.732359594039094e-06, | |
| "loss": 1.5572, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.3382352941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.729023780245919e-06, | |
| "loss": 1.6816, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.3400735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.725668497275361e-06, | |
| "loss": 1.8776, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.34191176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.72229377443379e-06, | |
| "loss": 1.7296, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.34375, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.718899641197375e-06, | |
| "loss": 1.8298, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.34558823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.71548612721182e-06, | |
| "loss": 1.7406, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.3474264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.712053262292111e-06, | |
| "loss": 1.953, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.3492647058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.70860107642225e-06, | |
| "loss": 1.8692, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.35110294117647056, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.7051295997549964e-06, | |
| "loss": 1.8754, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.35294117647058826, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.701638862611605e-06, | |
| "loss": 1.8684, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.3547794117647059, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.698128895481557e-06, | |
| "loss": 1.5358, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.35661764705882354, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.694599729022297e-06, | |
| "loss": 1.56, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.3584558823529412, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.691051394058965e-06, | |
| "loss": 1.7223, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.3602941176470588, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.687483921584124e-06, | |
| "loss": 1.6848, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.36213235294117646, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.683897342757493e-06, | |
| "loss": 1.654, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.3639705882352941, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.680291688905674e-06, | |
| "loss": 1.6973, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.36580882352941174, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.676666991521876e-06, | |
| "loss": 1.6474, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.36764705882352944, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.673023282265645e-06, | |
| "loss": 1.5936, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.3694852941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.669360592962581e-06, | |
| "loss": 1.6647, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.3713235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.665678955604064e-06, | |
| "loss": 1.9738, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.37316176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.661978402346974e-06, | |
| "loss": 1.7933, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.375, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.658258965513412e-06, | |
| "loss": 1.9133, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.37683823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.654520677590412e-06, | |
| "loss": 1.8377, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.3786764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.650763571229664e-06, | |
| "loss": 1.79, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.3805147058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.646987679247223e-06, | |
| "loss": 1.5877, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.38235294117647056, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.643193034623229e-06, | |
| "loss": 1.7125, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.38419117647058826, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.6393796705016105e-06, | |
| "loss": 1.7207, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.3860294117647059, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.635547620189802e-06, | |
| "loss": 1.6849, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.38786764705882354, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.631696917158449e-06, | |
| "loss": 1.716, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.3897058823529412, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.62782759504112e-06, | |
| "loss": 1.7206, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.3915441176470588, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.623939687634009e-06, | |
| "loss": 1.4938, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.39338235294117646, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.620033228895639e-06, | |
| "loss": 1.9391, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.3952205882352941, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.616108252946568e-06, | |
| "loss": 1.688, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.39705882352941174, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.612164794069096e-06, | |
| "loss": 1.9585, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.39889705882352944, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.608202886706953e-06, | |
| "loss": 1.6469, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.4007352941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.6042225654650096e-06, | |
| "loss": 1.8181, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.4025735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.60022386510897e-06, | |
| "loss": 1.8259, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.40441176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.5962068205650674e-06, | |
| "loss": 1.8962, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.40625, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.592171466919762e-06, | |
| "loss": 1.868, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.40808823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.588117839419432e-06, | |
| "loss": 1.7946, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.4099264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.584045973470067e-06, | |
| "loss": 1.6068, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.4117647058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.579955904636959e-06, | |
| "loss": 1.8194, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.41360294117647056, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.5758476686443905e-06, | |
| "loss": 1.8958, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.41544117647058826, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.571721301375323e-06, | |
| "loss": 1.5318, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.4172794117647059, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.5675768388710855e-06, | |
| "loss": 1.6046, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.41911764705882354, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.563414317331053e-06, | |
| "loss": 1.6724, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.4209558823529412, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.559233773112343e-06, | |
| "loss": 1.8096, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.4227941176470588, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.5550352427294836e-06, | |
| "loss": 1.4821, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.42463235294117646, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.550818762854105e-06, | |
| "loss": 1.6695, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.4264705882352941, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.546584370314613e-06, | |
| "loss": 1.6973, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.42830882352941174, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.542332102095871e-06, | |
| "loss": 1.8328, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.43014705882352944, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.538061995338875e-06, | |
| "loss": 1.6589, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.4319852941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.533774087340431e-06, | |
| "loss": 1.7145, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.4338235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.529468415552829e-06, | |
| "loss": 1.4717, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.43566176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.52514501758351e-06, | |
| "loss": 1.7362, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.4375, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.520803931194747e-06, | |
| "loss": 1.8571, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.43933823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.5164451943033105e-06, | |
| "loss": 1.9605, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.4411764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.512068844980136e-06, | |
| "loss": 1.9368, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.4430147058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.507674921449994e-06, | |
| "loss": 1.45, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.44485294117647056, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.503263462091153e-06, | |
| "loss": 1.6417, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.44669117647058826, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.49883450543505e-06, | |
| "loss": 1.7531, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.4485294117647059, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.494388090165947e-06, | |
| "loss": 1.826, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.45036764705882354, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.489924255120597e-06, | |
| "loss": 1.5047, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.4522058823529412, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.485443039287907e-06, | |
| "loss": 1.7405, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.4540441176470588, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.48094448180859e-06, | |
| "loss": 1.7201, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.45588235294117646, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.476428621974833e-06, | |
| "loss": 1.9913, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.4577205882352941, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.471895499229946e-06, | |
| "loss": 1.5852, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.45955882352941174, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.467345153168018e-06, | |
| "loss": 1.5358, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.46139705882352944, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.462777623533577e-06, | |
| "loss": 1.7271, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.4632352941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.458192950221237e-06, | |
| "loss": 1.5025, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.4650735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.4535911732753535e-06, | |
| "loss": 1.9186, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.46691176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.448972332889669e-06, | |
| "loss": 1.8936, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.46875, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.444336469406968e-06, | |
| "loss": 1.699, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.47058823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.4396836233187195e-06, | |
| "loss": 1.9617, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.4724264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.435013835264725e-06, | |
| "loss": 1.9323, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.4742647058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.4303271460327655e-06, | |
| "loss": 1.6515, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.47610294117647056, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.425623596558243e-06, | |
| "loss": 1.6436, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.47794117647058826, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.420903227923823e-06, | |
| "loss": 1.9221, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.4797794117647059, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.416166081359077e-06, | |
| "loss": 1.9025, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.48161764705882354, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.411412198240119e-06, | |
| "loss": 1.866, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.4834558823529412, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.406641620089252e-06, | |
| "loss": 1.6989, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.4852941176470588, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.401854388574595e-06, | |
| "loss": 1.7039, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.48713235294117646, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.397050545509726e-06, | |
| "loss": 1.9074, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.4889705882352941, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.392230132853316e-06, | |
| "loss": 1.5768, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.49080882352941174, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.387393192708758e-06, | |
| "loss": 1.6722, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.49264705882352944, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.382539767323805e-06, | |
| "loss": 1.8256, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.4944852941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.377669899090202e-06, | |
| "loss": 1.6908, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.4963235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.372783630543305e-06, | |
| "loss": 1.6795, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.49816176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.3678810043617215e-06, | |
| "loss": 1.7813, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.362962063366933e-06, | |
| "loss": 1.6982, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.5018382352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.358026850522919e-06, | |
| "loss": 2.0328, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.5036764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.353075408935787e-06, | |
| "loss": 1.7353, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.5055147058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.348107781853389e-06, | |
| "loss": 1.7333, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.5073529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.34312401266495e-06, | |
| "loss": 1.914, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.5091911764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.338124144900685e-06, | |
| "loss": 1.6887, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.5110294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.333108222231423e-06, | |
| "loss": 1.6988, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.5128676470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.32807628846822e-06, | |
| "loss": 1.5326, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.5147058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.3230283875619815e-06, | |
| "loss": 1.9346, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.5165441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.317964563603073e-06, | |
| "loss": 1.6371, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.5183823529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.312884860820942e-06, | |
| "loss": 2.1047, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.5202205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.307789323583727e-06, | |
| "loss": 1.5355, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.5220588235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.302677996397868e-06, | |
| "loss": 1.5629, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.5238970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.297550923907726e-06, | |
| "loss": 1.9965, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.5257352941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.2924081508951824e-06, | |
| "loss": 1.8088, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.5275735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.287249722279257e-06, | |
| "loss": 1.8258, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.5294117647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.28207568311571e-06, | |
| "loss": 1.7162, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.53125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.27688607859665e-06, | |
| "loss": 1.6649, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.5330882352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.27168095405014e-06, | |
| "loss": 1.8048, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.5349264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.266460354939803e-06, | |
| "loss": 1.7108, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.5367647058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.26122432686442e-06, | |
| "loss": 1.9229, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.5386029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.255972915557537e-06, | |
| "loss": 1.9534, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.5404411764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.250706166887061e-06, | |
| "loss": 1.9423, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.5422794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.245424126854864e-06, | |
| "loss": 1.7946, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.5441176470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.240126841596377e-06, | |
| "loss": 1.6288, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.5459558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.234814357380189e-06, | |
| "loss": 1.9226, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.5477941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.229486720607645e-06, | |
| "loss": 1.5787, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.5496323529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.224143977812435e-06, | |
| "loss": 1.6782, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.5514705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.218786175660194e-06, | |
| "loss": 1.4888, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.5533088235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.213413360948089e-06, | |
| "loss": 1.7777, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.5551470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.208025580604413e-06, | |
| "loss": 1.7884, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.5569852941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.202622881688178e-06, | |
| "loss": 1.6578, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.5588235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.197205311388698e-06, | |
| "loss": 1.991, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.5606617647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.1917729170251765e-06, | |
| "loss": 1.7002, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.5625, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.186325746046302e-06, | |
| "loss": 1.9256, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.5643382352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.180863846029825e-06, | |
| "loss": 1.8477, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.5661764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.175387264682146e-06, | |
| "loss": 1.9161, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.5680147058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.169896049837899e-06, | |
| "loss": 1.6926, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.5698529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.164390249459526e-06, | |
| "loss": 1.8339, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.5716911764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.158869911636876e-06, | |
| "loss": 1.8295, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.5735294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.153335084586766e-06, | |
| "loss": 1.8681, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.5753676470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.147785816652569e-06, | |
| "loss": 1.6268, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.5772058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.142222156303792e-06, | |
| "loss": 1.773, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.5790441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.13664415213565e-06, | |
| "loss": 2.0425, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.5808823529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.131051852868643e-06, | |
| "loss": 1.8064, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.5827205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.125445307348129e-06, | |
| "loss": 1.5052, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.5845588235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.119824564543901e-06, | |
| "loss": 1.7783, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.5863970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.114189673549752e-06, | |
| "loss": 1.6945, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.5882352941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.108540683583057e-06, | |
| "loss": 1.7935, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.5900735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.102877643984332e-06, | |
| "loss": 2.0515, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.5919117647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.097200604216811e-06, | |
| "loss": 1.7803, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.59375, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.09150961386601e-06, | |
| "loss": 1.7374, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.5955882352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.085804722639293e-06, | |
| "loss": 1.6242, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.5974264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.0800859803654436e-06, | |
| "loss": 1.8858, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.5992647058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.074353436994223e-06, | |
| "loss": 1.8843, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.6011029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.068607142595939e-06, | |
| "loss": 1.4963, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.6029411764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.062847147361003e-06, | |
| "loss": 1.6638, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.6047794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.0570735015994986e-06, | |
| "loss": 1.9207, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.6066176470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.0512862557407365e-06, | |
| "loss": 1.5746, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.6084558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.045485460332815e-06, | |
| "loss": 1.8553, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.6102941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.0396711660421825e-06, | |
| "loss": 1.8915, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.6121323529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.03384342365319e-06, | |
| "loss": 1.8034, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.6139705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.02800228406765e-06, | |
| "loss": 1.9337, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.6158088235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.02214779830439e-06, | |
| "loss": 1.9172, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.6176470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.016280017498812e-06, | |
| "loss": 1.5344, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.6194852941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.010398992902437e-06, | |
| "loss": 1.6145, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.6213235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.004504775882467e-06, | |
| "loss": 1.6857, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.6231617647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.998597417921331e-06, | |
| "loss": 1.6453, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.625, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.992676970616233e-06, | |
| "loss": 1.9115, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.6268382352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.98674348567871e-06, | |
| "loss": 1.7092, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.6286764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.980797014934169e-06, | |
| "loss": 1.7614, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.6305147058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.974837610321445e-06, | |
| "loss": 1.6805, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.6323529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.968865323892339e-06, | |
| "loss": 1.9099, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.6341911764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.962880207811168e-06, | |
| "loss": 1.5895, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.6360294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.95688231435431e-06, | |
| "loss": 1.6966, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.6378676470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.950871695909744e-06, | |
| "loss": 1.6833, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.6397058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.944848404976593e-06, | |
| "loss": 1.8615, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.6415441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.93881249416467e-06, | |
| "loss": 1.93, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.6433823529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.932764016194013e-06, | |
| "loss": 1.8436, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.6452205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.926703023894424e-06, | |
| "loss": 1.4891, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.6470588235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.920629570205014e-06, | |
| "loss": 2.0484, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.6488970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.914543708173735e-06, | |
| "loss": 1.7981, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.6507352941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.90844549095692e-06, | |
| "loss": 1.8744, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.6525735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.9023349718188155e-06, | |
| "loss": 1.9961, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.6544117647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.8962122041311155e-06, | |
| "loss": 1.8839, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.65625, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.890077241372503e-06, | |
| "loss": 1.7395, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.6580882352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.883930137128175e-06, | |
| "loss": 1.6163, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.6599264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.877770945089377e-06, | |
| "loss": 1.7127, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.6617647058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.871599719052931e-06, | |
| "loss": 1.6822, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.6636029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.865416512920776e-06, | |
| "loss": 2.0061, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.6654411764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.859221380699482e-06, | |
| "loss": 1.4916, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.6672794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.853014376499792e-06, | |
| "loss": 1.5192, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.6691176470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.846795554536141e-06, | |
| "loss": 1.8608, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.6709558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.840564969126186e-06, | |
| "loss": 1.7084, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.6727941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.834322674690329e-06, | |
| "loss": 1.6686, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.6746323529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.828068725751245e-06, | |
| "loss": 1.7066, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.6764705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.8218031769334024e-06, | |
| "loss": 1.7413, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.6783088235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.81552608296259e-06, | |
| "loss": 1.9639, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.6801470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.809237498665434e-06, | |
| "loss": 1.6569, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.6819852941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.8029374789689234e-06, | |
| "loss": 1.7029, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.6838235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7966260788999278e-06, | |
| "loss": 1.4536, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.6856617647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7903033535847167e-06, | |
| "loss": 1.7632, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.6875, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7839693582484806e-06, | |
| "loss": 1.698, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.6893382352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7776241482148452e-06, | |
| "loss": 1.6655, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.6911764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.771267778905391e-06, | |
| "loss": 1.8655, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.6930147058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7649003058391664e-06, | |
| "loss": 1.7093, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.6948529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7585217846322075e-06, | |
| "loss": 1.6746, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.6966911764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7521322709970454e-06, | |
| "loss": 1.9697, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.6985294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.745731820742227e-06, | |
| "loss": 2.0496, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.7003676470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7393204897718194e-06, | |
| "loss": 1.8899, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.7022058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7328983340849324e-06, | |
| "loss": 1.9481, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.7040441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7264654097752173e-06, | |
| "loss": 1.8767, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.7058823529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7200217730303865e-06, | |
| "loss": 1.7622, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.7077205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.713567480131718e-06, | |
| "loss": 1.8596, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.7095588235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7071025874535643e-06, | |
| "loss": 1.7558, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.7113970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7006271514628617e-06, | |
| "loss": 2.0891, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.7132352941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.694141228718634e-06, | |
| "loss": 1.5486, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.7150735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.6876448758715028e-06, | |
| "loss": 1.6308, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.7169117647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.681138149663189e-06, | |
| "loss": 1.9086, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.71875, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.6746211069260197e-06, | |
| "loss": 1.9397, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.7205882352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.6680938045824284e-06, | |
| "loss": 1.6673, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.7224264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.661556299644462e-06, | |
| "loss": 1.7074, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.7242647058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.6550086492132804e-06, | |
| "loss": 1.298, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.7261029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.6484509104786582e-06, | |
| "loss": 1.711, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.7279411764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.6418831407184856e-06, | |
| "loss": 1.5623, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.7297794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.6353053972982676e-06, | |
| "loss": 1.7687, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.7316176470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.628717737670623e-06, | |
| "loss": 1.6687, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.7334558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.6221202193747818e-06, | |
| "loss": 1.6686, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.7352941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.6155129000360846e-06, | |
| "loss": 1.4463, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.7371323529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.6088958373654794e-06, | |
| "loss": 1.8492, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.7389705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.602269089159013e-06, | |
| "loss": 1.7782, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.7408088235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.5956327132973313e-06, | |
| "loss": 1.7213, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.7426470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.588986767745174e-06, | |
| "loss": 1.6712, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.7444852941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.5823313105508626e-06, | |
| "loss": 1.679, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.7463235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.575666399845799e-06, | |
| "loss": 1.8439, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.7481617647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.568992093843956e-06, | |
| "loss": 2.0623, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.5623084508413685e-06, | |
| "loss": 1.6133, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.7518382352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.555615529215623e-06, | |
| "loss": 1.8713, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.7536764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.5489133874253516e-06, | |
| "loss": 1.8986, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.7555147058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.5422020840097173e-06, | |
| "loss": 1.6793, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.7573529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.535481677587904e-06, | |
| "loss": 1.6225, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.7591911764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.5287522268586074e-06, | |
| "loss": 1.7254, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.7610294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.5220137905995165e-06, | |
| "loss": 1.7691, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.7628676470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.515266427666806e-06, | |
| "loss": 1.7055, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.7647058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.508510196994618e-06, | |
| "loss": 1.7593, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.7665441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.50174515759455e-06, | |
| "loss": 1.7012, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.7683823529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.4949713685551377e-06, | |
| "loss": 1.887, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.7702205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.488188889041341e-06, | |
| "loss": 1.7187, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.7720588235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.4813977782940234e-06, | |
| "loss": 1.8475, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.7738970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.4745980956294396e-06, | |
| "loss": 1.6388, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.7757352941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.4677899004387134e-06, | |
| "loss": 1.9088, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.7775735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.460973252187321e-06, | |
| "loss": 1.6816, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.7794117647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.4541482104145695e-06, | |
| "loss": 1.6671, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.78125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.447314834733081e-06, | |
| "loss": 1.6012, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.7830882352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.440473184828266e-06, | |
| "loss": 1.6968, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.7849264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.433623320457809e-06, | |
| "loss": 1.5466, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.7867647058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.4267653014511405e-06, | |
| "loss": 1.8788, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.7886029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.419899187708917e-06, | |
| "loss": 1.8398, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.7904411764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.4130250392024973e-06, | |
| "loss": 1.7668, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.7922794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.4061429159734207e-06, | |
| "loss": 1.7916, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.7941176470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.3992528781328793e-06, | |
| "loss": 1.628, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.7959558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.3923549858611958e-06, | |
| "loss": 1.6921, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.7977941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.385449299407296e-06, | |
| "loss": 1.7011, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.7996323529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.378535879088182e-06, | |
| "loss": 1.6465, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.8014705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.3716147852884073e-06, | |
| "loss": 1.7443, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.8033088235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.3646860784595512e-06, | |
| "loss": 1.7987, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.8051470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.357749819119685e-06, | |
| "loss": 1.9102, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.8069852941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.3508060678528464e-06, | |
| "loss": 1.6303, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.8088235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.3438548853085135e-06, | |
| "loss": 1.5065, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.8106617647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.3368963322010695e-06, | |
| "loss": 1.7563, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.8125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.329930469309276e-06, | |
| "loss": 1.6226, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.8143382352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.322957357475741e-06, | |
| "loss": 1.8419, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.8161764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.315977057606388e-06, | |
| "loss": 1.7456, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.8180147058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.3089896306699233e-06, | |
| "loss": 1.9603, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.8198529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.301995137697304e-06, | |
| "loss": 1.6238, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.8216911764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.2949936397812055e-06, | |
| "loss": 1.6546, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.8235294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.287985198075484e-06, | |
| "loss": 1.5644, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.8253676470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.2809698737946494e-06, | |
| "loss": 1.9652, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.8272058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.2739477282133253e-06, | |
| "loss": 1.6981, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.8290441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.266918822665715e-06, | |
| "loss": 1.8254, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.8308823529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.259883218545065e-06, | |
| "loss": 1.5648, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.8327205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.2528409773031322e-06, | |
| "loss": 1.649, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.8345588235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.2457921604496435e-06, | |
| "loss": 1.9707, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.8363970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.2387368295517586e-06, | |
| "loss": 1.4134, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.8382352941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.231675046233536e-06, | |
| "loss": 1.834, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.8400735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.22460687217539e-06, | |
| "loss": 1.7787, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 0.8419117647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.217532369113555e-06, | |
| "loss": 1.7776, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.84375, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.2104515988395456e-06, | |
| "loss": 1.8862, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.8455882352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.2033646231996167e-06, | |
| "loss": 1.5536, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.8474264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.196271504094223e-06, | |
| "loss": 1.6952, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 0.8492647058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.189172303477478e-06, | |
| "loss": 1.8626, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.8511029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.182067083356616e-06, | |
| "loss": 1.7898, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 0.8529411764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.174955905791444e-06, | |
| "loss": 1.7797, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.8547794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.1678388328938093e-06, | |
| "loss": 1.6529, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.8566176470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.1607159268270447e-06, | |
| "loss": 1.6602, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 0.8584558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.153587249805438e-06, | |
| "loss": 1.6258, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 0.8602941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.1464528640936797e-06, | |
| "loss": 1.7756, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.8621323529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.139312832006323e-06, | |
| "loss": 1.771, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 0.8639705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.132167215907238e-06, | |
| "loss": 1.9377, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.8658088235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.12501607820907e-06, | |
| "loss": 1.7851, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 0.8676470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.11785948137269e-06, | |
| "loss": 1.9384, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 0.8694852941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.1106974879066514e-06, | |
| "loss": 1.4842, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 0.8713235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.1035301603666456e-06, | |
| "loss": 1.7289, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 0.8731617647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.0963575613549523e-06, | |
| "loss": 1.8963, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.875, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.089179753519894e-06, | |
| "loss": 1.8238, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 0.8768382352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.0819967995552913e-06, | |
| "loss": 2.1243, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 0.8786764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.074808762199911e-06, | |
| "loss": 1.7607, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 0.8805147058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.0676157042369213e-06, | |
| "loss": 1.7313, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 0.8823529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.0604176884933422e-06, | |
| "loss": 1.6726, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.8841911764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.053214777839496e-06, | |
| "loss": 1.7602, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 0.8860294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.0460070351884614e-06, | |
| "loss": 1.7777, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 0.8878676470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.0387945234955187e-06, | |
| "loss": 1.759, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 0.8897058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.031577305757605e-06, | |
| "loss": 2.0917, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 0.8915441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.024355445012761e-06, | |
| "loss": 1.7402, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.8933823529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.0171290043395823e-06, | |
| "loss": 1.9261, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 0.8952205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.0098980468566663e-06, | |
| "loss": 1.6524, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 0.8970588235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.0026626357220623e-06, | |
| "loss": 1.8296, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 0.8988970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.9954228341327192e-06, | |
| "loss": 1.8665, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 0.9007352941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.988178705323934e-06, | |
| "loss": 1.8146, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.9025735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.9809303125688004e-06, | |
| "loss": 1.7391, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 0.9044117647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.9736777191776543e-06, | |
| "loss": 1.6417, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 0.90625, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.966420988497522e-06, | |
| "loss": 1.8464, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 0.9080882352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.959160183911565e-06, | |
| "loss": 1.8636, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 0.9099264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.9518953688385298e-06, | |
| "loss": 1.6568, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.9117647058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.9446266067321904e-06, | |
| "loss": 1.9179, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 0.9136029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.9373539610807983e-06, | |
| "loss": 1.9894, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 0.9154411764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.930077495406523e-06, | |
| "loss": 1.8537, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 0.9172794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.9227972732649e-06, | |
| "loss": 1.8176, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 0.9191176470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.915513358244276e-06, | |
| "loss": 1.7762, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.9209558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.9082258139652536e-06, | |
| "loss": 1.7569, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 0.9227941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.900934704080133e-06, | |
| "loss": 1.7657, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 0.9246323529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.893640092272357e-06, | |
| "loss": 1.7845, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 0.9264705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.8863420422559577e-06, | |
| "loss": 1.4962, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 0.9283088235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.8790406177749985e-06, | |
| "loss": 1.6051, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.9301470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.8717358826030158e-06, | |
| "loss": 1.8549, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 0.9319852941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.86442790054246e-06, | |
| "loss": 1.7021, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 0.9338235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.8571167354241445e-06, | |
| "loss": 1.6309, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 0.9356617647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.849802451106685e-06, | |
| "loss": 1.77, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 0.9375, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.84248511147594e-06, | |
| "loss": 1.7613, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.9393382352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.835164780444455e-06, | |
| "loss": 1.9886, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 0.9411764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.8278415219509025e-06, | |
| "loss": 1.6941, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 0.9430147058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.8205153999595253e-06, | |
| "loss": 1.641, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 0.9448529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.8131864784595788e-06, | |
| "loss": 1.8998, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 0.9466911764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.8058548214647674e-06, | |
| "loss": 1.7034, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.9485294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.798520493012691e-06, | |
| "loss": 1.7346, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 0.9503676470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.7911835571642816e-06, | |
| "loss": 2.0461, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 0.9522058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.783844078003245e-06, | |
| "loss": 1.7676, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 0.9540441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.7765021196355023e-06, | |
| "loss": 1.592, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 0.9558823529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.76915774618863e-06, | |
| "loss": 1.5685, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.9577205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.761811021811295e-06, | |
| "loss": 1.7379, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 0.9595588235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.754462010672701e-06, | |
| "loss": 1.9914, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 0.9613970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.7471107769620258e-06, | |
| "loss": 1.8213, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 0.9632352941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.739757384887859e-06, | |
| "loss": 1.6564, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 0.9650735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.732401898677642e-06, | |
| "loss": 1.8431, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.9669117647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.725044382577107e-06, | |
| "loss": 1.8911, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 0.96875, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.7176849008497165e-06, | |
| "loss": 1.6735, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 0.9705882352941176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.7103235177761018e-06, | |
| "loss": 1.6467, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 0.9724264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.702960297653501e-06, | |
| "loss": 1.7016, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 0.9742647058823529, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.695595304795197e-06, | |
| "loss": 1.8497, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.9761029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.688228603529959e-06, | |
| "loss": 1.9022, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 0.9779411764705882, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.680860258201475e-06, | |
| "loss": 1.6943, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 0.9797794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.6734903331677946e-06, | |
| "loss": 1.886, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 0.9816176470588235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.666118892800765e-06, | |
| "loss": 1.715, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 0.9834558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.658746001485469e-06, | |
| "loss": 1.7098, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.9852941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.651371723619661e-06, | |
| "loss": 1.6282, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 0.9871323529411765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.6439961236132083e-06, | |
| "loss": 1.8106, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 0.9889705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.6366192658875256e-06, | |
| "loss": 1.95, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 0.9908088235294118, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.629241214875013e-06, | |
| "loss": 1.5364, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 0.9926470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.621862035018492e-06, | |
| "loss": 1.8866, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.9944852941176471, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.6144817907706453e-06, | |
| "loss": 1.6631, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 0.9963235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.607100546593453e-06, | |
| "loss": 1.7325, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 0.9981617647058824, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.5997183669576264e-06, | |
| "loss": 1.6731, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.59233531634205e-06, | |
| "loss": 1.481, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 1.0018382352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.584951459233215e-06, | |
| "loss": 1.864, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 1.0036764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.5775668601246555e-06, | |
| "loss": 1.8555, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 1.005514705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.5701815835163896e-06, | |
| "loss": 1.8919, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 1.0073529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.5627956939143507e-06, | |
| "loss": 1.9853, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 1.0091911764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.555409255829825e-06, | |
| "loss": 2.0538, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 1.0110294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.548022333778892e-06, | |
| "loss": 1.632, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.0128676470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.540634992281858e-06, | |
| "loss": 1.5268, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 1.0147058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.5332472958626923e-06, | |
| "loss": 1.6835, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 1.0165441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.525859309048463e-06, | |
| "loss": 1.7067, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 1.0183823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.518471096368777e-06, | |
| "loss": 1.8755, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 1.0202205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.511082722355212e-06, | |
| "loss": 1.8055, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 1.0220588235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.503694251540757e-06, | |
| "loss": 1.5165, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 1.0238970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.496305748459244e-06, | |
| "loss": 1.4758, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 1.025735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.4889172776447885e-06, | |
| "loss": 1.811, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 1.0275735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.4815289036312236e-06, | |
| "loss": 1.8183, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 1.0294117647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.474140690951538e-06, | |
| "loss": 1.9046, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.03125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.4667527041373085e-06, | |
| "loss": 1.5942, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 1.0330882352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.459365007718143e-06, | |
| "loss": 1.5341, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 1.0349264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.4519776662211083e-06, | |
| "loss": 1.8902, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 1.036764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.444590744170176e-06, | |
| "loss": 1.9901, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 1.0386029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.437204306085651e-06, | |
| "loss": 1.421, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 1.0404411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.4298184164836104e-06, | |
| "loss": 1.8848, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 1.0422794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.422433139875345e-06, | |
| "loss": 1.5938, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 1.0441176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.4150485407667855e-06, | |
| "loss": 1.9612, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 1.0459558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.4076646836579508e-06, | |
| "loss": 1.7775, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 1.0477941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.4002816330423744e-06, | |
| "loss": 1.6265, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.0496323529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.392899453406548e-06, | |
| "loss": 1.9559, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 1.0514705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.385518209229355e-06, | |
| "loss": 1.6096, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 1.0533088235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.3781379649815094e-06, | |
| "loss": 2.0393, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 1.0551470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.3707587851249875e-06, | |
| "loss": 1.7581, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 1.056985294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.3633807341124753e-06, | |
| "loss": 1.857, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 1.0588235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.3560038763867913e-06, | |
| "loss": 1.8291, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 1.0606617647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.3486282763803397e-06, | |
| "loss": 1.6881, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 1.0625, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.3412539985145324e-06, | |
| "loss": 1.5025, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 1.0643382352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.3338811071992353e-06, | |
| "loss": 1.7593, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 1.0661764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.3265096668322063e-06, | |
| "loss": 1.8755, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.068014705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.319139741798525e-06, | |
| "loss": 1.7058, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 1.0698529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.3117713964700415e-06, | |
| "loss": 2.0831, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 1.0716911764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.304404695204804e-06, | |
| "loss": 1.8701, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 1.0735294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.2970397023465e-06, | |
| "loss": 1.8427, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 1.0753676470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.289676482223899e-06, | |
| "loss": 1.4616, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 1.0772058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.2823150991502844e-06, | |
| "loss": 1.6877, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 1.0790441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.274955617422894e-06, | |
| "loss": 1.7195, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 1.0808823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.267598101322359e-06, | |
| "loss": 1.7431, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 1.0827205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.2602426151121413e-06, | |
| "loss": 1.2884, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 1.0845588235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.2528892230379746e-06, | |
| "loss": 1.7814, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.0863970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.245537989327299e-06, | |
| "loss": 1.6408, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 1.088235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.2381889781887063e-06, | |
| "loss": 1.708, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 1.0900735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.2308422538113718e-06, | |
| "loss": 1.4733, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 1.0919117647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.2234978803644972e-06, | |
| "loss": 1.7233, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 1.09375, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.2161559219967555e-06, | |
| "loss": 1.6656, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 1.0955882352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.2088164428357196e-06, | |
| "loss": 1.7206, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 1.0974264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.2014795069873092e-06, | |
| "loss": 1.915, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 1.099264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.1941451785352334e-06, | |
| "loss": 1.7666, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 1.1011029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.1868135215404216e-06, | |
| "loss": 1.5042, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 1.1029411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.1794846000404756e-06, | |
| "loss": 1.9193, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.1047794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.1721584780490988e-06, | |
| "loss": 1.9347, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 1.1066176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.1648352195555456e-06, | |
| "loss": 1.8926, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 1.1084558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.157514888524061e-06, | |
| "loss": 1.73, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 1.1102941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.150197548893315e-06, | |
| "loss": 1.5678, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 1.1121323529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.142883264575856e-06, | |
| "loss": 1.837, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 1.1139705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.1355720994575414e-06, | |
| "loss": 1.6828, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 1.1158088235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.1282641173969855e-06, | |
| "loss": 1.6306, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 1.1176470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.120959382225002e-06, | |
| "loss": 1.8238, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 1.119485294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.113657957744042e-06, | |
| "loss": 1.9431, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 1.1213235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.106359907727644e-06, | |
| "loss": 1.754, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.1231617647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.099065295919869e-06, | |
| "loss": 1.8101, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 1.125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.0917741860347463e-06, | |
| "loss": 1.8708, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 1.1268382352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.0844866417557243e-06, | |
| "loss": 1.8487, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 1.1286764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.0772027267351e-06, | |
| "loss": 1.9463, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 1.130514705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.069922504593478e-06, | |
| "loss": 1.5628, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 1.1323529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.062646038919202e-06, | |
| "loss": 1.6557, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 1.1341911764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.0553733932678096e-06, | |
| "loss": 1.5397, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 1.1360294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.048104631161471e-06, | |
| "loss": 1.6991, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 1.1378676470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.040839816088436e-06, | |
| "loss": 1.8088, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 1.1397058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.0335790115024787e-06, | |
| "loss": 1.7756, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.1415441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.026322280822346e-06, | |
| "loss": 2.0245, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 1.1433823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.019069687431199e-06, | |
| "loss": 1.7583, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 1.1452205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.0118212946760664e-06, | |
| "loss": 1.7974, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 1.1470588235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.004577165867282e-06, | |
| "loss": 1.4694, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 1.1488970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.9973373642779385e-06, | |
| "loss": 1.7528, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 1.150735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.9901019531433345e-06, | |
| "loss": 1.5792, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 1.1525735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.9828709956604176e-06, | |
| "loss": 1.5808, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 1.1544117647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.9756445549872396e-06, | |
| "loss": 1.8014, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 1.15625, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.968422694242396e-06, | |
| "loss": 1.7312, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 1.1580882352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.961205476504482e-06, | |
| "loss": 1.5888, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.1599264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.9539929648115395e-06, | |
| "loss": 1.7464, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 1.161764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.9467852221605044e-06, | |
| "loss": 1.7115, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 1.1636029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.939582311506658e-06, | |
| "loss": 1.7716, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 1.1654411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.9323842957630795e-06, | |
| "loss": 1.7229, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 1.1672794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.925191237800089e-06, | |
| "loss": 1.7373, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 1.1691176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.9180032004447095e-06, | |
| "loss": 1.5658, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 1.1709558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.910820246480106e-06, | |
| "loss": 1.8198, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 1.1727941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.9036424386450487e-06, | |
| "loss": 1.6393, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 1.1746323529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.8964698396333555e-06, | |
| "loss": 1.872, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 1.1764705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.8893025120933488e-06, | |
| "loss": 1.5913, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.1783088235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.882140518627311e-06, | |
| "loss": 1.7481, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 1.1801470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.874983921790931e-06, | |
| "loss": 1.7131, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 1.181985294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.8678327840927627e-06, | |
| "loss": 1.5675, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 1.1838235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.860687167993678e-06, | |
| "loss": 1.889, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 1.1856617647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.8535471359063211e-06, | |
| "loss": 1.8815, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 1.1875, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.8464127501945625e-06, | |
| "loss": 1.5414, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 1.1893382352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.8392840731729561e-06, | |
| "loss": 1.8284, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 1.1911764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.8321611671061915e-06, | |
| "loss": 1.712, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 1.193014705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.8250440942085562e-06, | |
| "loss": 1.6119, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 1.1948529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.8179329166433846e-06, | |
| "loss": 1.6625, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.1966911764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.8108276965225225e-06, | |
| "loss": 1.9299, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 1.1985294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.8037284959057782e-06, | |
| "loss": 1.7874, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 1.2003676470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.7966353768003838e-06, | |
| "loss": 1.6945, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 1.2022058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.7895484011604553e-06, | |
| "loss": 1.7856, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 1.2040441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.7824676308864452e-06, | |
| "loss": 1.516, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 1.2058823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.7753931278246109e-06, | |
| "loss": 1.5837, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 1.2077205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.768324953766465e-06, | |
| "loss": 1.8389, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 1.2095588235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.761263170448242e-06, | |
| "loss": 1.5752, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 1.2113970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.7542078395503574e-06, | |
| "loss": 1.5576, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 1.213235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.7471590226968682e-06, | |
| "loss": 1.7369, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.2150735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.7401167814549353e-06, | |
| "loss": 1.7101, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 1.2169117647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.7330811773342864e-06, | |
| "loss": 1.7831, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 1.21875, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.7260522717866751e-06, | |
| "loss": 1.8079, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 1.2205882352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.719030126205351e-06, | |
| "loss": 1.9275, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 1.2224264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.7120148019245173e-06, | |
| "loss": 1.7212, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 1.224264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.7050063602187956e-06, | |
| "loss": 1.6668, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 1.2261029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6980048623026967e-06, | |
| "loss": 1.756, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 1.2279411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6910103693300767e-06, | |
| "loss": 2.0557, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 1.2297794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6840229423936127e-06, | |
| "loss": 1.5587, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 1.2316176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6770426425242603e-06, | |
| "loss": 1.9321, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.2334558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.670069530690725e-06, | |
| "loss": 1.8981, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 1.2352941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6631036677989315e-06, | |
| "loss": 1.7275, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 1.2371323529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6561451146914873e-06, | |
| "loss": 1.6767, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 1.2389705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.649193932147154e-06, | |
| "loss": 1.7657, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 1.2408088235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6422501808803165e-06, | |
| "loss": 1.9796, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 1.2426470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.635313921540449e-06, | |
| "loss": 1.7294, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 1.244485294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6283852147115931e-06, | |
| "loss": 1.6732, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 1.2463235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6214641209118186e-06, | |
| "loss": 1.8132, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 1.2481617647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6145507005927052e-06, | |
| "loss": 1.7224, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.607645014138805e-06, | |
| "loss": 1.9829, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.2518382352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6007471218671209e-06, | |
| "loss": 1.7143, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 1.2536764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.59385708402658e-06, | |
| "loss": 1.7415, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 1.2555147058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.586974960797503e-06, | |
| "loss": 1.7169, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 1.2573529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.5801008122910838e-06, | |
| "loss": 1.865, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 1.2591911764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.5732346985488605e-06, | |
| "loss": 1.9046, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 1.2610294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.5663766795421912e-06, | |
| "loss": 1.8258, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 1.2628676470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.5595268151717347e-06, | |
| "loss": 1.666, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 1.2647058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.55268516526692e-06, | |
| "loss": 1.9431, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 1.2665441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.5458517895854309e-06, | |
| "loss": 1.7691, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 1.2683823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.5390267478126802e-06, | |
| "loss": 1.9737, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.2702205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.5322100995612868e-06, | |
| "loss": 1.8369, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 1.2720588235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.525401904370561e-06, | |
| "loss": 1.8232, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 1.2738970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.5186022217059777e-06, | |
| "loss": 1.5475, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 1.2757352941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.5118111109586598e-06, | |
| "loss": 1.9996, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 1.2775735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.5050286314448632e-06, | |
| "loss": 1.7864, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 1.2794117647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4982548424054505e-06, | |
| "loss": 1.7143, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 1.28125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4914898030053827e-06, | |
| "loss": 1.7523, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 1.2830882352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4847335723331952e-06, | |
| "loss": 1.8499, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 1.2849264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4779862094004837e-06, | |
| "loss": 1.8997, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 1.2867647058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4712477731413935e-06, | |
| "loss": 1.825, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.2886029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4645183224120956e-06, | |
| "loss": 1.9566, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 1.2904411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.457797915990283e-06, | |
| "loss": 1.6981, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 1.2922794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4510866125746497e-06, | |
| "loss": 1.7447, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 1.2941176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4443844707843767e-06, | |
| "loss": 1.8724, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 1.2959558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.437691549158633e-06, | |
| "loss": 1.967, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 1.2977941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4310079061560438e-06, | |
| "loss": 1.7536, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 1.2996323529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.424333600154202e-06, | |
| "loss": 1.9413, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 1.3014705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4176686894491386e-06, | |
| "loss": 2.1014, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 1.3033088235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.411013232254827e-06, | |
| "loss": 1.7457, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 1.3051470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4043672867026693e-06, | |
| "loss": 1.7354, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.3069852941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3977309108409882e-06, | |
| "loss": 1.8861, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 1.3088235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3911041626345215e-06, | |
| "loss": 1.4903, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 1.3106617647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3844870999639154e-06, | |
| "loss": 1.788, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 1.3125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3778797806252186e-06, | |
| "loss": 1.7226, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 1.3143382352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3712822623293776e-06, | |
| "loss": 1.7147, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 1.3161764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3646946027017333e-06, | |
| "loss": 1.7169, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 1.3180147058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3581168592815142e-06, | |
| "loss": 1.7809, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 1.3198529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3515490895213428e-06, | |
| "loss": 1.6852, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 1.3216911764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3449913507867202e-06, | |
| "loss": 1.8936, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 1.3235294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3384437003555394e-06, | |
| "loss": 1.8982, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.3253676470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3319061954175733e-06, | |
| "loss": 1.5213, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 1.3272058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3253788930739816e-06, | |
| "loss": 1.5643, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 1.3290441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3188618503368117e-06, | |
| "loss": 1.6885, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 1.3308823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3123551241284976e-06, | |
| "loss": 1.8524, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 1.3327205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3058587712813665e-06, | |
| "loss": 1.9121, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 1.3345588235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.299372848537139e-06, | |
| "loss": 1.8104, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 1.3363970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.292897412546436e-06, | |
| "loss": 1.7105, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 1.3382352941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.286432519868282e-06, | |
| "loss": 1.7432, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 1.3400735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2799782269696137e-06, | |
| "loss": 1.8036, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 1.3419117647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2735345902247831e-06, | |
| "loss": 1.8589, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.34375, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2671016659150693e-06, | |
| "loss": 1.5688, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 1.3455882352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2606795102281806e-06, | |
| "loss": 1.6113, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 1.3474264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2542681792577749e-06, | |
| "loss": 1.799, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 1.3492647058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2478677290029554e-06, | |
| "loss": 1.6972, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 1.3511029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2414782153677935e-06, | |
| "loss": 1.586, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 1.3529411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.235099694160834e-06, | |
| "loss": 1.88, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 1.3547794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2287322210946098e-06, | |
| "loss": 1.6245, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 1.3566176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2223758517851556e-06, | |
| "loss": 1.8224, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 1.3584558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.21603064175152e-06, | |
| "loss": 1.9156, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 1.3602941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2096966464152837e-06, | |
| "loss": 1.8021, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.3621323529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2033739211000729e-06, | |
| "loss": 1.9186, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 1.3639705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1970625210310768e-06, | |
| "loss": 1.7245, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 1.3658088235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.190762501334566e-06, | |
| "loss": 1.6118, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 1.3676470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.184473917037411e-06, | |
| "loss": 1.8043, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 1.3694852941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1781968230665978e-06, | |
| "loss": 1.7861, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 1.3713235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1719312742487565e-06, | |
| "loss": 1.6989, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 1.3731617647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1656773253096716e-06, | |
| "loss": 1.7684, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 1.375, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1594350308738154e-06, | |
| "loss": 1.7307, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 1.3768382352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1532044454638598e-06, | |
| "loss": 1.733, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 1.3786764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1469856235002088e-06, | |
| "loss": 1.7285, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.3805147058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.140778619300519e-06, | |
| "loss": 1.8647, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 1.3823529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1345834870792256e-06, | |
| "loss": 1.6412, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 1.3841911764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1284002809470695e-06, | |
| "loss": 1.6432, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 1.3860294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1222290549106245e-06, | |
| "loss": 1.779, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 1.3878676470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1160698628718253e-06, | |
| "loss": 1.8016, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 1.3897058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1099227586274965e-06, | |
| "loss": 1.708, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 1.3915441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1037877958688856e-06, | |
| "loss": 1.5641, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 1.3933823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0976650281811855e-06, | |
| "loss": 1.6455, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 1.3952205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.091554509043081e-06, | |
| "loss": 1.6056, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 1.3970588235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0854562918262645e-06, | |
| "loss": 1.669, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.3988970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0793704297949872e-06, | |
| "loss": 1.8561, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 1.4007352941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0732969761055774e-06, | |
| "loss": 1.608, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 1.4025735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0672359838059884e-06, | |
| "loss": 1.681, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 1.4044117647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0611875058353301e-06, | |
| "loss": 1.8089, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 1.40625, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.055151595023407e-06, | |
| "loss": 1.8458, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 1.4080882352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0491283040902567e-06, | |
| "loss": 2.0752, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 1.4099264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0431176856456904e-06, | |
| "loss": 1.7032, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 1.4117647058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0371197921888324e-06, | |
| "loss": 1.9582, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 1.4136029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0311346761076619e-06, | |
| "loss": 1.7644, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 1.4154411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0251623896785553e-06, | |
| "loss": 1.8227, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.4172794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0192029850658306e-06, | |
| "loss": 1.9582, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 1.4191176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0132565143212907e-06, | |
| "loss": 1.9154, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 1.4209558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.007323029383766e-06, | |
| "loss": 1.769, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 1.4227941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.00140258207867e-06, | |
| "loss": 1.7419, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 1.4246323529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.954952241175325e-07, | |
| "loss": 1.7485, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 1.4264705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.896010070975638e-07, | |
| "loss": 1.8709, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 1.4283088235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.837199825011895e-07, | |
| "loss": 1.7689, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 1.4301470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.778522016956105e-07, | |
| "loss": 1.7562, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 1.4319852941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.71997715932351e-07, | |
| "loss": 1.6869, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 1.4338235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.661565763468106e-07, | |
| "loss": 1.8674, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.4356617647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.603288339578177e-07, | |
| "loss": 1.6821, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 1.4375, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.54514539667185e-07, | |
| "loss": 1.7868, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 1.4393382352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.487137442592642e-07, | |
| "loss": 1.8007, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 1.4411764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.429264984005015e-07, | |
| "loss": 1.8453, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 1.4430147058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.371528526389978e-07, | |
| "loss": 1.9375, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 1.4448529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.313928574040615e-07, | |
| "loss": 1.8563, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 1.4466911764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.256465630057778e-07, | |
| "loss": 1.7986, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 1.4485294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.199140196345569e-07, | |
| "loss": 1.7898, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 1.4503676470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.141952773607082e-07, | |
| "loss": 1.8142, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 1.4522058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.084903861339916e-07, | |
| "loss": 1.7731, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.4540441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.027993957831898e-07, | |
| "loss": 1.5631, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 1.4558823529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.971223560156686e-07, | |
| "loss": 1.7644, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 1.4577205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.914593164169438e-07, | |
| "loss": 1.7418, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 1.4595588235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.858103264502482e-07, | |
| "loss": 1.7843, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 1.4613970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.801754354561001e-07, | |
| "loss": 1.6739, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 1.4632352941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.745546926518714e-07, | |
| "loss": 1.8804, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 1.4650735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.689481471313574e-07, | |
| "loss": 1.8066, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 1.4669117647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.633558478643503e-07, | |
| "loss": 1.6212, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 1.46875, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.577778436962081e-07, | |
| "loss": 1.7998, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 1.4705882352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.522141833474318e-07, | |
| "loss": 1.693, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.4724264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.466649154132342e-07, | |
| "loss": 1.729, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 1.4742647058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.411300883631246e-07, | |
| "loss": 1.722, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 1.4761029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.356097505404742e-07, | |
| "loss": 1.6582, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 1.4779411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.301039501621028e-07, | |
| "loss": 1.6609, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 1.4797794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.24612735317854e-07, | |
| "loss": 1.8046, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 1.4816176470588236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.191361539701748e-07, | |
| "loss": 1.7152, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 1.4834558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.136742539536979e-07, | |
| "loss": 1.7417, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 1.4852941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.082270829748245e-07, | |
| "loss": 1.6673, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 1.4871323529411764, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.02794688611303e-07, | |
| "loss": 1.9246, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 1.4889705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.973771183118222e-07, | |
| "loss": 1.5252, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.4908088235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.919744193955864e-07, | |
| "loss": 1.6546, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 1.4926470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.865866390519114e-07, | |
| "loss": 1.7019, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 1.4944852941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.812138243398071e-07, | |
| "loss": 1.8009, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 1.4963235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.75856022187565e-07, | |
| "loss": 1.8821, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 1.4981617647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.705132793923559e-07, | |
| "loss": 1.4378, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.651856426198106e-07, | |
| "loss": 1.6485, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 1.5018382352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.59873158403624e-07, | |
| "loss": 1.8485, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 1.5036764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.54575873145137e-07, | |
| "loss": 1.5154, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 1.5055147058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.492938331129393e-07, | |
| "loss": 1.6477, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 1.5073529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.440270844424635e-07, | |
| "loss": 1.6652, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.5091911764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.3877567313558e-07, | |
| "loss": 1.8875, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 1.5110294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.335396450601972e-07, | |
| "loss": 1.7319, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 1.5128676470588234, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.283190459498607e-07, | |
| "loss": 1.6376, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 1.5147058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.231139214033505e-07, | |
| "loss": 1.6051, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 1.5165441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.179243168842911e-07, | |
| "loss": 1.8559, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 1.5183823529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.127502777207437e-07, | |
| "loss": 1.6349, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 1.5202205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.075918491048172e-07, | |
| "loss": 1.4663, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 1.5220588235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.024490760922748e-07, | |
| "loss": 1.6839, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 1.5238970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.973220036021313e-07, | |
| "loss": 1.5657, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 1.5257352941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.922106764162742e-07, | |
| "loss": 1.9338, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.5275735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.871151391790584e-07, | |
| "loss": 1.9015, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 1.5294117647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.820354363969276e-07, | |
| "loss": 1.4708, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 1.53125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.769716124380193e-07, | |
| "loss": 1.9289, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 1.5330882352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.7192371153178e-07, | |
| "loss": 1.6729, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 1.5349264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.668917777685771e-07, | |
| "loss": 1.828, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 1.5367647058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.618758550993157e-07, | |
| "loss": 1.8761, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 1.5386029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.568759873350505e-07, | |
| "loss": 1.7344, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 1.5404411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.518922181466123e-07, | |
| "loss": 1.7223, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 1.5422794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.469245910642135e-07, | |
| "loss": 1.5391, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 1.5441176470588234, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.419731494770814e-07, | |
| "loss": 1.7717, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.5459558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.370379366330682e-07, | |
| "loss": 1.7384, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 1.5477941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.321189956382795e-07, | |
| "loss": 1.7945, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 1.5496323529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.27216369456696e-07, | |
| "loss": 1.8308, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 1.5514705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.223301009097982e-07, | |
| "loss": 1.8692, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 1.5533088235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.174602326761947e-07, | |
| "loss": 1.8698, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 1.5551470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.126068072912431e-07, | |
| "loss": 2.0008, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 1.5569852941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.077698671466852e-07, | |
| "loss": 1.7407, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 1.5588235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.029494544902742e-07, | |
| "loss": 1.79, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 1.5606617647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.981456114254061e-07, | |
| "loss": 1.719, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 1.5625, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.933583799107482e-07, | |
| "loss": 1.7807, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.5643382352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.885878017598815e-07, | |
| "loss": 1.9422, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 1.5661764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.838339186409236e-07, | |
| "loss": 1.9177, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 1.5680147058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.790967720761778e-07, | |
| "loss": 1.5679, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 1.5698529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.743764034417579e-07, | |
| "loss": 1.8958, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 1.5716911764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.696728539672353e-07, | |
| "loss": 1.978, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 1.5735294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.649861647352758e-07, | |
| "loss": 1.5037, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 1.5753676470588234, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.603163766812817e-07, | |
| "loss": 1.7701, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 1.5772058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.556635305930327e-07, | |
| "loss": 1.4404, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 1.5790441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.510276671103315e-07, | |
| "loss": 1.8623, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 1.5808823529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.464088267246473e-07, | |
| "loss": 1.7948, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.5827205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.418070497787634e-07, | |
| "loss": 1.8761, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 1.5845588235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.372223764664236e-07, | |
| "loss": 1.8479, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 1.5863970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.326548468319825e-07, | |
| "loss": 1.7426, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 1.5882352941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.281045007700555e-07, | |
| "loss": 1.821, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 1.5900735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.235713780251669e-07, | |
| "loss": 1.8196, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 1.5919117647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.190555181914106e-07, | |
| "loss": 1.6626, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 1.59375, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.14556960712094e-07, | |
| "loss": 1.7291, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 1.5955882352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.100757448794039e-07, | |
| "loss": 1.5247, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 1.5974264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.056119098340545e-07, | |
| "loss": 1.5109, | |
| "step": 869 | |
| }, | |
| { | |
| "epoch": 1.5992647058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.01165494564951e-07, | |
| "loss": 1.9381, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.6011029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.967365379088473e-07, | |
| "loss": 1.7729, | |
| "step": 871 | |
| }, | |
| { | |
| "epoch": 1.6029411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.923250785500069e-07, | |
| "loss": 1.7253, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 1.6047794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.879311550198643e-07, | |
| "loss": 1.7911, | |
| "step": 873 | |
| }, | |
| { | |
| "epoch": 1.6066176470588234, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.835548056966899e-07, | |
| "loss": 1.8662, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 1.6084558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.791960688052533e-07, | |
| "loss": 1.1291, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 1.6102941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.748549824164908e-07, | |
| "loss": 1.6652, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 1.6121323529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.705315844471728e-07, | |
| "loss": 1.8267, | |
| "step": 877 | |
| }, | |
| { | |
| "epoch": 1.6139705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.6622591265956863e-07, | |
| "loss": 1.9036, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 1.6158088235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.6193800466112576e-07, | |
| "loss": 1.884, | |
| "step": 879 | |
| }, | |
| { | |
| "epoch": 1.6176470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.5766789790412963e-07, | |
| "loss": 1.7381, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.6194852941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.534156296853884e-07, | |
| "loss": 1.8417, | |
| "step": 881 | |
| }, | |
| { | |
| "epoch": 1.6213235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.491812371458962e-07, | |
| "loss": 1.5894, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 1.6231617647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.449647572705171e-07, | |
| "loss": 1.8028, | |
| "step": 883 | |
| }, | |
| { | |
| "epoch": 1.625, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.407662268876578e-07, | |
| "loss": 1.9584, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 1.6268382352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.3658568266894716e-07, | |
| "loss": 1.7494, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 1.6286764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.324231611289159e-07, | |
| "loss": 1.7979, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 1.6305147058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.282786986246773e-07, | |
| "loss": 1.7872, | |
| "step": 887 | |
| }, | |
| { | |
| "epoch": 1.6323529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.2415233135560977e-07, | |
| "loss": 2.0073, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 1.6341911764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.200440953630411e-07, | |
| "loss": 1.7227, | |
| "step": 889 | |
| }, | |
| { | |
| "epoch": 1.6360294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.1595402652993304e-07, | |
| "loss": 1.6261, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 1.6378676470588234, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.118821605805684e-07, | |
| "loss": 1.9185, | |
| "step": 891 | |
| }, | |
| { | |
| "epoch": 1.6397058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.0782853308023895e-07, | |
| "loss": 1.6901, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 1.6415441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.0379317943493287e-07, | |
| "loss": 2.045, | |
| "step": 893 | |
| }, | |
| { | |
| "epoch": 1.6433823529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.997761348910309e-07, | |
| "loss": 1.8319, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 1.6452205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.95777434534991e-07, | |
| "loss": 1.5535, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 1.6470588235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.917971132930476e-07, | |
| "loss": 1.8543, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 1.6488970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.8783520593090486e-07, | |
| "loss": 1.8723, | |
| "step": 897 | |
| }, | |
| { | |
| "epoch": 1.6507352941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.83891747053432e-07, | |
| "loss": 1.8325, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 1.6525735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7996677110436226e-07, | |
| "loss": 1.841, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 1.6544117647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.76060312365992e-07, | |
| "loss": 1.7471, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.65625, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.721724049588801e-07, | |
| "loss": 1.592, | |
| "step": 901 | |
| }, | |
| { | |
| "epoch": 1.6580882352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.6830308284155126e-07, | |
| "loss": 1.7401, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 1.6599264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.64452379810199e-07, | |
| "loss": 1.6918, | |
| "step": 903 | |
| }, | |
| { | |
| "epoch": 1.6617647058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.606203294983901e-07, | |
| "loss": 1.9804, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 1.6636029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.5680696537677193e-07, | |
| "loss": 1.7059, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 1.6654411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.5301232075277635e-07, | |
| "loss": 1.7507, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 1.6672794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.4923642877033674e-07, | |
| "loss": 1.5373, | |
| "step": 907 | |
| }, | |
| { | |
| "epoch": 1.6691176470588234, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.454793224095879e-07, | |
| "loss": 1.5712, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 1.6709558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.4174103448658875e-07, | |
| "loss": 1.8673, | |
| "step": 909 | |
| }, | |
| { | |
| "epoch": 1.6727941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.3802159765302627e-07, | |
| "loss": 1.6926, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 1.6746323529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.343210443959369e-07, | |
| "loss": 1.9163, | |
| "step": 911 | |
| }, | |
| { | |
| "epoch": 1.6764705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.3063940703742e-07, | |
| "loss": 1.7858, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 1.6783088235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.269767177343555e-07, | |
| "loss": 1.6493, | |
| "step": 913 | |
| }, | |
| { | |
| "epoch": 1.6801470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.2333300847812387e-07, | |
| "loss": 1.9067, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 1.6819852941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.1970831109432666e-07, | |
| "loss": 1.7049, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 1.6838235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.161026572425074e-07, | |
| "loss": 2.0241, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 1.6856617647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.125160784158765e-07, | |
| "loss": 1.4931, | |
| "step": 917 | |
| }, | |
| { | |
| "epoch": 1.6875, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.0894860594103597e-07, | |
| "loss": 1.8086, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 1.6893382352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.054002709777029e-07, | |
| "loss": 1.9141, | |
| "step": 919 | |
| }, | |
| { | |
| "epoch": 1.6911764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.018711045184436e-07, | |
| "loss": 1.5003, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.6930147058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.9836113738839544e-07, | |
| "loss": 2.0187, | |
| "step": 921 | |
| }, | |
| { | |
| "epoch": 1.6948529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.9487040024500424e-07, | |
| "loss": 1.723, | |
| "step": 922 | |
| }, | |
| { | |
| "epoch": 1.6966911764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.9139892357775115e-07, | |
| "loss": 1.881, | |
| "step": 923 | |
| }, | |
| { | |
| "epoch": 1.6985294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.879467377078896e-07, | |
| "loss": 1.6959, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 1.7003676470588234, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.845138727881802e-07, | |
| "loss": 1.5516, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 1.7022058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.8110035880262563e-07, | |
| "loss": 1.5723, | |
| "step": 926 | |
| }, | |
| { | |
| "epoch": 1.7040441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.7770622556621054e-07, | |
| "loss": 1.7377, | |
| "step": 927 | |
| }, | |
| { | |
| "epoch": 1.7058823529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.7433150272463985e-07, | |
| "loss": 1.5755, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 1.7077205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.7097621975408136e-07, | |
| "loss": 1.5451, | |
| "step": 929 | |
| }, | |
| { | |
| "epoch": 1.7095588235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.676404059609061e-07, | |
| "loss": 1.7694, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 1.7113970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.6432409048143483e-07, | |
| "loss": 2.0309, | |
| "step": 931 | |
| }, | |
| { | |
| "epoch": 1.7132352941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.610273022816812e-07, | |
| "loss": 1.6251, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 1.7150735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.577500701571009e-07, | |
| "loss": 1.8236, | |
| "step": 933 | |
| }, | |
| { | |
| "epoch": 1.7169117647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.5449242273233753e-07, | |
| "loss": 1.6628, | |
| "step": 934 | |
| }, | |
| { | |
| "epoch": 1.71875, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.5125438846097636e-07, | |
| "loss": 1.9572, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 1.7205882352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.480359956252912e-07, | |
| "loss": 1.7628, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 1.7224264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.4483727233600237e-07, | |
| "loss": 1.5424, | |
| "step": 937 | |
| }, | |
| { | |
| "epoch": 1.7242647058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.4165824653202605e-07, | |
| "loss": 1.772, | |
| "step": 938 | |
| }, | |
| { | |
| "epoch": 1.7261029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.384989459802345e-07, | |
| "loss": 1.806, | |
| "step": 939 | |
| }, | |
| { | |
| "epoch": 1.7279411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.3535939827521114e-07, | |
| "loss": 1.9861, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 1.7297794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.3223963083901014e-07, | |
| "loss": 1.7898, | |
| "step": 941 | |
| }, | |
| { | |
| "epoch": 1.7316176470588234, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.2913967092091706e-07, | |
| "loss": 1.9501, | |
| "step": 942 | |
| }, | |
| { | |
| "epoch": 1.7334558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.2605954559721116e-07, | |
| "loss": 1.4893, | |
| "step": 943 | |
| }, | |
| { | |
| "epoch": 1.7352941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.2299928177092784e-07, | |
| "loss": 1.9125, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 1.7371323529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.199589061716245e-07, | |
| "loss": 1.9624, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 1.7389705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.1693844535514763e-07, | |
| "loss": 1.8543, | |
| "step": 946 | |
| }, | |
| { | |
| "epoch": 1.7408088235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.13937925703398e-07, | |
| "loss": 1.5834, | |
| "step": 947 | |
| }, | |
| { | |
| "epoch": 1.7426470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.1095737342410577e-07, | |
| "loss": 1.5379, | |
| "step": 948 | |
| }, | |
| { | |
| "epoch": 1.7444852941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.0799681455059423e-07, | |
| "loss": 1.8171, | |
| "step": 949 | |
| }, | |
| { | |
| "epoch": 1.7463235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.0505627494155994e-07, | |
| "loss": 1.9249, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 1.7481617647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.0213578028084018e-07, | |
| "loss": 1.8258, | |
| "step": 951 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.9923535607719423e-07, | |
| "loss": 1.5903, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 1.7518382352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.9635502766407632e-07, | |
| "loss": 1.7791, | |
| "step": 953 | |
| }, | |
| { | |
| "epoch": 1.7536764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.934948201994169e-07, | |
| "loss": 1.687, | |
| "step": 954 | |
| }, | |
| { | |
| "epoch": 1.7555147058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.9065475866540172e-07, | |
| "loss": 1.8509, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 1.7573529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.8783486786825455e-07, | |
| "loss": 1.8326, | |
| "step": 956 | |
| }, | |
| { | |
| "epoch": 1.7591911764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.8503517243801922e-07, | |
| "loss": 1.8198, | |
| "step": 957 | |
| }, | |
| { | |
| "epoch": 1.7610294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.8225569682834565e-07, | |
| "loss": 1.7581, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 1.7628676470588234, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.794964653162759e-07, | |
| "loss": 1.9452, | |
| "step": 959 | |
| }, | |
| { | |
| "epoch": 1.7647058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.7675750200203152e-07, | |
| "loss": 1.7043, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 1.7665441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.7403883080880424e-07, | |
| "loss": 1.6634, | |
| "step": 961 | |
| }, | |
| { | |
| "epoch": 1.7683823529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.713404754825454e-07, | |
| "loss": 1.8329, | |
| "step": 962 | |
| }, | |
| { | |
| "epoch": 1.7702205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6866245959176157e-07, | |
| "loss": 1.3605, | |
| "step": 963 | |
| }, | |
| { | |
| "epoch": 1.7720588235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.66004806527304e-07, | |
| "loss": 1.8235, | |
| "step": 964 | |
| }, | |
| { | |
| "epoch": 1.7738970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6336753950216833e-07, | |
| "loss": 1.6423, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 1.7757352941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6075068155129075e-07, | |
| "loss": 1.5276, | |
| "step": 966 | |
| }, | |
| { | |
| "epoch": 1.7775735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.5815425553134562e-07, | |
| "loss": 1.6792, | |
| "step": 967 | |
| }, | |
| { | |
| "epoch": 1.7794117647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.555782841205475e-07, | |
| "loss": 1.7641, | |
| "step": 968 | |
| }, | |
| { | |
| "epoch": 1.78125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.5302278981845136e-07, | |
| "loss": 1.6337, | |
| "step": 969 | |
| }, | |
| { | |
| "epoch": 1.7830882352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.5048779494575838e-07, | |
| "loss": 1.6087, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 1.7849264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4797332164411816e-07, | |
| "loss": 1.8162, | |
| "step": 971 | |
| }, | |
| { | |
| "epoch": 1.7867647058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.454793918759373e-07, | |
| "loss": 1.7674, | |
| "step": 972 | |
| }, | |
| { | |
| "epoch": 1.7886029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.430060274241876e-07, | |
| "loss": 1.8653, | |
| "step": 973 | |
| }, | |
| { | |
| "epoch": 1.7904411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4055324989221464e-07, | |
| "loss": 1.7376, | |
| "step": 974 | |
| }, | |
| { | |
| "epoch": 1.7922794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3812108070354908e-07, | |
| "loss": 1.7221, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 1.7941176470588234, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3570954110172203e-07, | |
| "loss": 1.9376, | |
| "step": 976 | |
| }, | |
| { | |
| "epoch": 1.7959558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3331865215007482e-07, | |
| "loss": 1.7548, | |
| "step": 977 | |
| }, | |
| { | |
| "epoch": 1.7977941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.309484347315812e-07, | |
| "loss": 1.8472, | |
| "step": 978 | |
| }, | |
| { | |
| "epoch": 1.7996323529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2859890954865867e-07, | |
| "loss": 1.6442, | |
| "step": 979 | |
| }, | |
| { | |
| "epoch": 1.8014705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.262700971229916e-07, | |
| "loss": 1.6644, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 1.8033088235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2396201779535154e-07, | |
| "loss": 1.6133, | |
| "step": 981 | |
| }, | |
| { | |
| "epoch": 1.8051470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2167469172541773e-07, | |
| "loss": 1.7667, | |
| "step": 982 | |
| }, | |
| { | |
| "epoch": 1.8069852941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.194081388916035e-07, | |
| "loss": 1.8097, | |
| "step": 983 | |
| }, | |
| { | |
| "epoch": 1.8088235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1716237909087991e-07, | |
| "loss": 1.9465, | |
| "step": 984 | |
| }, | |
| { | |
| "epoch": 1.8106617647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1493743193860207e-07, | |
| "loss": 1.8992, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 1.8125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.12733316868342e-07, | |
| "loss": 1.8624, | |
| "step": 986 | |
| }, | |
| { | |
| "epoch": 1.8143382352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1055005313171413e-07, | |
| "loss": 1.7256, | |
| "step": 987 | |
| }, | |
| { | |
| "epoch": 1.8161764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0838765979820892e-07, | |
| "loss": 1.7664, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 1.8180147058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0624615575502789e-07, | |
| "loss": 1.6467, | |
| "step": 989 | |
| }, | |
| { | |
| "epoch": 1.8198529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0412555970691519e-07, | |
| "loss": 1.7614, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 1.8216911764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0202589017599878e-07, | |
| "loss": 1.8051, | |
| "step": 991 | |
| }, | |
| { | |
| "epoch": 1.8235294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.994716550162376e-08, | |
| "loss": 2.0208, | |
| "step": 992 | |
| }, | |
| { | |
| "epoch": 1.8253676470588234, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.788940384019591e-08, | |
| "loss": 1.8248, | |
| "step": 993 | |
| }, | |
| { | |
| "epoch": 1.8272058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.585262316502114e-08, | |
| "loss": 1.7762, | |
| "step": 994 | |
| }, | |
| { | |
| "epoch": 1.8290441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.383684126614945e-08, | |
| "loss": 1.6026, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 1.8308823529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.184207575021947e-08, | |
| "loss": 1.5231, | |
| "step": 996 | |
| }, | |
| { | |
| "epoch": 1.8327205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.986834404030364e-08, | |
| "loss": 1.6845, | |
| "step": 997 | |
| }, | |
| { | |
| "epoch": 1.8345588235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.79156633757569e-08, | |
| "loss": 1.6766, | |
| "step": 998 | |
| }, | |
| { | |
| "epoch": 1.8363970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.598405081206712e-08, | |
| "loss": 1.7158, | |
| "step": 999 | |
| }, | |
| { | |
| "epoch": 1.8382352941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.407352322070323e-08, | |
| "loss": 1.9128, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.8400735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.218409728897148e-08, | |
| "loss": 1.892, | |
| "step": 1001 | |
| }, | |
| { | |
| "epoch": 1.8419117647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.031578951986697e-08, | |
| "loss": 1.699, | |
| "step": 1002 | |
| }, | |
| { | |
| "epoch": 1.84375, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.846861623192953e-08, | |
| "loss": 1.5973, | |
| "step": 1003 | |
| }, | |
| { | |
| "epoch": 1.8455882352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.66425935591042e-08, | |
| "loss": 1.8389, | |
| "step": 1004 | |
| }, | |
| { | |
| "epoch": 1.8474264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.483773745059569e-08, | |
| "loss": 1.7224, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 1.8492647058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.305406367073387e-08, | |
| "loss": 1.7114, | |
| "step": 1006 | |
| }, | |
| { | |
| "epoch": 1.8511029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.129158779883211e-08, | |
| "loss": 1.8995, | |
| "step": 1007 | |
| }, | |
| { | |
| "epoch": 1.8529411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.955032522905331e-08, | |
| "loss": 1.8587, | |
| "step": 1008 | |
| }, | |
| { | |
| "epoch": 1.8547794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.78302911702755e-08, | |
| "loss": 1.934, | |
| "step": 1009 | |
| }, | |
| { | |
| "epoch": 1.8566176470588234, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.613150064595786e-08, | |
| "loss": 1.9598, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 1.8584558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.445396849400987e-08, | |
| "loss": 1.5031, | |
| "step": 1011 | |
| }, | |
| { | |
| "epoch": 1.8602941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.27977093666629e-08, | |
| "loss": 1.8191, | |
| "step": 1012 | |
| }, | |
| { | |
| "epoch": 1.8621323529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.11627377303406e-08, | |
| "loss": 1.856, | |
| "step": 1013 | |
| }, | |
| { | |
| "epoch": 1.8639705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.954906786553361e-08, | |
| "loss": 1.4916, | |
| "step": 1014 | |
| }, | |
| { | |
| "epoch": 1.8658088235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.795671386667423e-08, | |
| "loss": 1.7956, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 1.8676470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.638568964201313e-08, | |
| "loss": 1.8146, | |
| "step": 1016 | |
| }, | |
| { | |
| "epoch": 1.8694852941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.4836008913499164e-08, | |
| "loss": 1.6043, | |
| "step": 1017 | |
| }, | |
| { | |
| "epoch": 1.8713235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.33076852166578e-08, | |
| "loss": 1.8612, | |
| "step": 1018 | |
| }, | |
| { | |
| "epoch": 1.8731617647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.1800731900474855e-08, | |
| "loss": 1.6387, | |
| "step": 1019 | |
| }, | |
| { | |
| "epoch": 1.875, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.0315162127277384e-08, | |
| "loss": 1.7442, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.8768382352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.885098887262074e-08, | |
| "loss": 1.9966, | |
| "step": 1021 | |
| }, | |
| { | |
| "epoch": 1.8786764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.7408224925174776e-08, | |
| "loss": 1.7452, | |
| "step": 1022 | |
| }, | |
| { | |
| "epoch": 1.8805147058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.598688288661196e-08, | |
| "loss": 1.6622, | |
| "step": 1023 | |
| }, | |
| { | |
| "epoch": 1.8823529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.4586975171496686e-08, | |
| "loss": 1.6588, | |
| "step": 1024 | |
| }, | |
| { | |
| "epoch": 1.8841911764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.320851400717835e-08, | |
| "loss": 1.6102, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 1.8860294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.18515114336826e-08, | |
| "loss": 1.6396, | |
| "step": 1026 | |
| }, | |
| { | |
| "epoch": 1.8878676470588234, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.05159793036089e-08, | |
| "loss": 1.7462, | |
| "step": 1027 | |
| }, | |
| { | |
| "epoch": 1.8897058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.9201929282023375e-08, | |
| "loss": 1.7489, | |
| "step": 1028 | |
| }, | |
| { | |
| "epoch": 1.8915441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7909372846360593e-08, | |
| "loss": 1.8584, | |
| "step": 1029 | |
| }, | |
| { | |
| "epoch": 1.8933823529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.663832128632028e-08, | |
| "loss": 1.9724, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 1.8952205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.53887857037713e-08, | |
| "loss": 1.9278, | |
| "step": 1031 | |
| }, | |
| { | |
| "epoch": 1.8970588235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.416077701265203e-08, | |
| "loss": 1.8119, | |
| "step": 1032 | |
| }, | |
| { | |
| "epoch": 1.8988970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.295430593887788e-08, | |
| "loss": 1.7219, | |
| "step": 1033 | |
| }, | |
| { | |
| "epoch": 1.9007352941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.176938302024535e-08, | |
| "loss": 1.7584, | |
| "step": 1034 | |
| }, | |
| { | |
| "epoch": 1.9025735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.060601860634088e-08, | |
| "loss": 1.6304, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 1.9044117647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.9464222858450465e-08, | |
| "loss": 1.8839, | |
| "step": 1036 | |
| }, | |
| { | |
| "epoch": 1.90625, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.834400574947077e-08, | |
| "loss": 1.8399, | |
| "step": 1037 | |
| }, | |
| { | |
| "epoch": 1.9080882352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.72453770638223e-08, | |
| "loss": 1.656, | |
| "step": 1038 | |
| }, | |
| { | |
| "epoch": 1.9099264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.616834639736332e-08, | |
| "loss": 1.7035, | |
| "step": 1039 | |
| }, | |
| { | |
| "epoch": 1.9117647058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.5112923157306902e-08, | |
| "loss": 1.9063, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.9136029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.4079116562137906e-08, | |
| "loss": 1.4835, | |
| "step": 1041 | |
| }, | |
| { | |
| "epoch": 1.9154411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.3066935641533063e-08, | |
| "loss": 1.7912, | |
| "step": 1042 | |
| }, | |
| { | |
| "epoch": 1.9172794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.207638923628158e-08, | |
| "loss": 1.7931, | |
| "step": 1043 | |
| }, | |
| { | |
| "epoch": 1.9191176470588234, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.110748599820883e-08, | |
| "loss": 1.7911, | |
| "step": 1044 | |
| }, | |
| { | |
| "epoch": 1.9209558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.0160234390099454e-08, | |
| "loss": 1.8429, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 1.9227941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.9234642685624915e-08, | |
| "loss": 1.4734, | |
| "step": 1046 | |
| }, | |
| { | |
| "epoch": 1.9246323529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.833071896926969e-08, | |
| "loss": 1.7346, | |
| "step": 1047 | |
| }, | |
| { | |
| "epoch": 1.9264705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.744847113626186e-08, | |
| "loss": 1.6973, | |
| "step": 1048 | |
| }, | |
| { | |
| "epoch": 1.9283088235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.658790689250428e-08, | |
| "loss": 1.9301, | |
| "step": 1049 | |
| }, | |
| { | |
| "epoch": 1.9301470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.574903375450576e-08, | |
| "loss": 1.7079, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.9319852941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.493185904931721e-08, | |
| "loss": 1.4348, | |
| "step": 1051 | |
| }, | |
| { | |
| "epoch": 1.9338235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4136389914466142e-08, | |
| "loss": 1.9293, | |
| "step": 1052 | |
| }, | |
| { | |
| "epoch": 1.9356617647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3362633297895334e-08, | |
| "loss": 1.5467, | |
| "step": 1053 | |
| }, | |
| { | |
| "epoch": 1.9375, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2610595957902039e-08, | |
| "loss": 1.8591, | |
| "step": 1054 | |
| }, | |
| { | |
| "epoch": 1.9393382352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1880284463078596e-08, | |
| "loss": 1.871, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 1.9411764705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1171705192254689e-08, | |
| "loss": 1.966, | |
| "step": 1056 | |
| }, | |
| { | |
| "epoch": 1.9430147058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0484864334442956e-08, | |
| "loss": 1.5263, | |
| "step": 1057 | |
| }, | |
| { | |
| "epoch": 1.9448529411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.819767888783749e-09, | |
| "loss": 1.6543, | |
| "step": 1058 | |
| }, | |
| { | |
| "epoch": 1.9466911764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.176421664492952e-09, | |
| "loss": 1.7169, | |
| "step": 1059 | |
| }, | |
| { | |
| "epoch": 1.9485294117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.554831280811748e-09, | |
| "loss": 1.7802, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.9503676470588234, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.95500216695666e-09, | |
| "loss": 1.6599, | |
| "step": 1061 | |
| }, | |
| { | |
| "epoch": 1.9522058823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.3769395620731934e-09, | |
| "loss": 1.7335, | |
| "step": 1062 | |
| }, | |
| { | |
| "epoch": 1.9540441176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.82064851518921e-09, | |
| "loss": 1.8277, | |
| "step": 1063 | |
| }, | |
| { | |
| "epoch": 1.9558823529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.286133885171075e-09, | |
| "loss": 1.938, | |
| "step": 1064 | |
| }, | |
| { | |
| "epoch": 1.9577205882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.773400340682023e-09, | |
| "loss": 1.8225, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 1.9595588235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.282452360139689e-09, | |
| "loss": 1.7666, | |
| "step": 1066 | |
| }, | |
| { | |
| "epoch": 1.9613970588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.813294231678367e-09, | |
| "loss": 1.5706, | |
| "step": 1067 | |
| }, | |
| { | |
| "epoch": 1.9632352941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.3659300531112555e-09, | |
| "loss": 1.9548, | |
| "step": 1068 | |
| }, | |
| { | |
| "epoch": 1.9650735294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.9403637318943814e-09, | |
| "loss": 1.8425, | |
| "step": 1069 | |
| }, | |
| { | |
| "epoch": 1.9669117647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.536598985092732e-09, | |
| "loss": 1.8042, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.96875, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.154639339347787e-09, | |
| "loss": 1.6248, | |
| "step": 1071 | |
| }, | |
| { | |
| "epoch": 1.9705882352941178, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.7944881308464288e-09, | |
| "loss": 1.7392, | |
| "step": 1072 | |
| }, | |
| { | |
| "epoch": 1.9724264705882353, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.4561485052920774e-09, | |
| "loss": 1.6152, | |
| "step": 1073 | |
| }, | |
| { | |
| "epoch": 1.9742647058823528, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.1396234178769347e-09, | |
| "loss": 1.8047, | |
| "step": 1074 | |
| }, | |
| { | |
| "epoch": 1.9761029411764706, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.844915633257005e-09, | |
| "loss": 1.8002, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 1.9779411764705883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.5720277255268368e-09, | |
| "loss": 1.874, | |
| "step": 1076 | |
| }, | |
| { | |
| "epoch": 1.9797794117647058, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3209620781973188e-09, | |
| "loss": 1.8465, | |
| "step": 1077 | |
| }, | |
| { | |
| "epoch": 1.9816176470588234, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0917208841756954e-09, | |
| "loss": 1.7619, | |
| "step": 1078 | |
| }, | |
| { | |
| "epoch": 1.9834558823529411, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.843061457450286e-10, | |
| "loss": 1.8742, | |
| "step": 1079 | |
| }, | |
| { | |
| "epoch": 1.9852941176470589, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.98719674548376e-10, | |
| "loss": 1.6781, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.9871323529411766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.349630915710279e-10, | |
| "loss": 1.6849, | |
| "step": 1081 | |
| }, | |
| { | |
| "epoch": 1.9889705882352942, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.930378271282953e-10, | |
| "loss": 1.6883, | |
| "step": 1082 | |
| }, | |
| { | |
| "epoch": 1.9908088235294117, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.7294512085163093e-10, | |
| "loss": 1.8295, | |
| "step": 1083 | |
| }, | |
| { | |
| "epoch": 1.9926470588235294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.746860216783608e-10, | |
| "loss": 1.7199, | |
| "step": 1084 | |
| }, | |
| { | |
| "epoch": 1.9944852941176472, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.82613878422467e-11, | |
| "loss": 1.8275, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 1.9963235294117647, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.3671886866269995e-11, | |
| "loss": 1.9304, | |
| "step": 1086 | |
| }, | |
| { | |
| "epoch": 1.9981617647058822, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0917995557080397e-11, | |
| "loss": 1.8221, | |
| "step": 1087 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 1.4433, | |
| "step": 1088 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 1088, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 272, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 9.797617676955156e+18, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |