| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 3782, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.00026441036488630354, | |
| "grad_norm": 3.609375, | |
| "learning_rate": 5.277044854881267e-07, | |
| "loss": 1.8784, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0013220518244315177, | |
| "grad_norm": 3.609375, | |
| "learning_rate": 2.6385224274406333e-06, | |
| "loss": 1.8748, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0026441036488630354, | |
| "grad_norm": 3.578125, | |
| "learning_rate": 5.2770448548812665e-06, | |
| "loss": 1.8755, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.003966155473294554, | |
| "grad_norm": 3.5, | |
| "learning_rate": 7.915567282321901e-06, | |
| "loss": 1.8741, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.005288207297726071, | |
| "grad_norm": 3.359375, | |
| "learning_rate": 1.0554089709762533e-05, | |
| "loss": 1.8644, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.006610259122157589, | |
| "grad_norm": 2.9375, | |
| "learning_rate": 1.3192612137203167e-05, | |
| "loss": 1.8583, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.007932310946589107, | |
| "grad_norm": 2.5625, | |
| "learning_rate": 1.5831134564643802e-05, | |
| "loss": 1.8416, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.009254362771020624, | |
| "grad_norm": 1.8671875, | |
| "learning_rate": 1.8469656992084432e-05, | |
| "loss": 1.8236, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.010576414595452142, | |
| "grad_norm": 1.4921875, | |
| "learning_rate": 2.1108179419525066e-05, | |
| "loss": 1.8066, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.011898466419883659, | |
| "grad_norm": 1.2265625, | |
| "learning_rate": 2.37467018469657e-05, | |
| "loss": 1.7951, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.013220518244315178, | |
| "grad_norm": 1.1171875, | |
| "learning_rate": 2.6385224274406334e-05, | |
| "loss": 1.7783, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.014542570068746695, | |
| "grad_norm": 0.92578125, | |
| "learning_rate": 2.9023746701846964e-05, | |
| "loss": 1.7577, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.015864621893178214, | |
| "grad_norm": 0.8125, | |
| "learning_rate": 3.1662269129287604e-05, | |
| "loss": 1.7413, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.01718667371760973, | |
| "grad_norm": 0.6953125, | |
| "learning_rate": 3.430079155672823e-05, | |
| "loss": 1.7217, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.01850872554204125, | |
| "grad_norm": 0.69921875, | |
| "learning_rate": 3.6939313984168865e-05, | |
| "loss": 1.7064, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.019830777366472766, | |
| "grad_norm": 0.57421875, | |
| "learning_rate": 3.95778364116095e-05, | |
| "loss": 1.6914, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.021152829190904283, | |
| "grad_norm": 0.57421875, | |
| "learning_rate": 4.221635883905013e-05, | |
| "loss": 1.679, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.0224748810153358, | |
| "grad_norm": 0.55859375, | |
| "learning_rate": 4.4854881266490766e-05, | |
| "loss": 1.6632, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.023796932839767318, | |
| "grad_norm": 0.58203125, | |
| "learning_rate": 4.74934036939314e-05, | |
| "loss": 1.6472, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.025118984664198835, | |
| "grad_norm": 0.546875, | |
| "learning_rate": 5.0131926121372033e-05, | |
| "loss": 1.6343, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.026441036488630356, | |
| "grad_norm": 0.380859375, | |
| "learning_rate": 5.277044854881267e-05, | |
| "loss": 1.6313, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.027763088313061873, | |
| "grad_norm": 0.484375, | |
| "learning_rate": 5.540897097625331e-05, | |
| "loss": 1.6094, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.02908514013749339, | |
| "grad_norm": 0.5234375, | |
| "learning_rate": 5.804749340369393e-05, | |
| "loss": 1.5984, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.030407191961924908, | |
| "grad_norm": 0.51171875, | |
| "learning_rate": 6.068601583113457e-05, | |
| "loss": 1.5843, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.03172924378635643, | |
| "grad_norm": 0.322265625, | |
| "learning_rate": 6.332453825857521e-05, | |
| "loss": 1.5689, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.03305129561078794, | |
| "grad_norm": 0.44921875, | |
| "learning_rate": 6.596306068601583e-05, | |
| "loss": 1.5575, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.03437334743521946, | |
| "grad_norm": 0.3203125, | |
| "learning_rate": 6.860158311345646e-05, | |
| "loss": 1.5451, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.03569539925965098, | |
| "grad_norm": 0.4140625, | |
| "learning_rate": 7.124010554089711e-05, | |
| "loss": 1.5333, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.0370174510840825, | |
| "grad_norm": 0.48828125, | |
| "learning_rate": 7.387862796833773e-05, | |
| "loss": 1.5237, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.03833950290851401, | |
| "grad_norm": 0.5, | |
| "learning_rate": 7.651715039577836e-05, | |
| "loss": 1.5105, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.03966155473294553, | |
| "grad_norm": 0.3984375, | |
| "learning_rate": 7.9155672823219e-05, | |
| "loss": 1.4946, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.040983606557377046, | |
| "grad_norm": 0.39453125, | |
| "learning_rate": 8.179419525065963e-05, | |
| "loss": 1.4843, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.04230565838180857, | |
| "grad_norm": 0.3515625, | |
| "learning_rate": 8.443271767810026e-05, | |
| "loss": 1.4732, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.04362771020624009, | |
| "grad_norm": 0.421875, | |
| "learning_rate": 8.70712401055409e-05, | |
| "loss": 1.4611, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.0449497620306716, | |
| "grad_norm": 0.31640625, | |
| "learning_rate": 8.970976253298153e-05, | |
| "loss": 1.4485, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.04627181385510312, | |
| "grad_norm": 0.341796875, | |
| "learning_rate": 9.234828496042217e-05, | |
| "loss": 1.4406, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.047593865679534636, | |
| "grad_norm": 0.359375, | |
| "learning_rate": 9.49868073878628e-05, | |
| "loss": 1.4236, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.04891591750396616, | |
| "grad_norm": 0.32421875, | |
| "learning_rate": 9.762532981530343e-05, | |
| "loss": 1.4107, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.05023796932839767, | |
| "grad_norm": 0.435546875, | |
| "learning_rate": 0.00010026385224274407, | |
| "loss": 1.4052, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.05156002115282919, | |
| "grad_norm": 0.458984375, | |
| "learning_rate": 0.00010290237467018471, | |
| "loss": 1.3859, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.05288207297726071, | |
| "grad_norm": 0.51953125, | |
| "learning_rate": 0.00010554089709762533, | |
| "loss": 1.3742, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.054204124801692226, | |
| "grad_norm": 0.478515625, | |
| "learning_rate": 0.00010817941952506597, | |
| "loss": 1.365, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.05552617662612375, | |
| "grad_norm": 0.57421875, | |
| "learning_rate": 0.00011081794195250662, | |
| "loss": 1.3482, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.05684822845055526, | |
| "grad_norm": 0.412109375, | |
| "learning_rate": 0.00011345646437994724, | |
| "loss": 1.3375, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.05817028027498678, | |
| "grad_norm": 0.384765625, | |
| "learning_rate": 0.00011609498680738786, | |
| "loss": 1.333, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.059492332099418295, | |
| "grad_norm": 0.63671875, | |
| "learning_rate": 0.00011873350923482852, | |
| "loss": 1.3238, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.060814383923849816, | |
| "grad_norm": 0.55859375, | |
| "learning_rate": 0.00012137203166226914, | |
| "loss": 1.315, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.06213643574828133, | |
| "grad_norm": 0.671875, | |
| "learning_rate": 0.00012401055408970977, | |
| "loss": 1.3028, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.06345848757271286, | |
| "grad_norm": 0.470703125, | |
| "learning_rate": 0.00012664907651715042, | |
| "loss": 1.2869, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.06478053939714437, | |
| "grad_norm": 0.435546875, | |
| "learning_rate": 0.00012928759894459104, | |
| "loss": 1.281, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.06610259122157588, | |
| "grad_norm": 0.671875, | |
| "learning_rate": 0.00013192612137203166, | |
| "loss": 1.2747, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.0674246430460074, | |
| "grad_norm": 0.625, | |
| "learning_rate": 0.0001345646437994723, | |
| "loss": 1.2697, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.06874669487043893, | |
| "grad_norm": 1.0859375, | |
| "learning_rate": 0.00013720316622691292, | |
| "loss": 1.2549, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.07006874669487044, | |
| "grad_norm": 0.75, | |
| "learning_rate": 0.00013984168865435357, | |
| "loss": 1.246, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.07139079851930195, | |
| "grad_norm": 0.451171875, | |
| "learning_rate": 0.00014248021108179422, | |
| "loss": 1.2412, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.07271285034373347, | |
| "grad_norm": 0.515625, | |
| "learning_rate": 0.00014511873350923484, | |
| "loss": 1.2343, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.074034902168165, | |
| "grad_norm": 0.5390625, | |
| "learning_rate": 0.00014775725593667546, | |
| "loss": 1.2205, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.07535695399259651, | |
| "grad_norm": 0.8984375, | |
| "learning_rate": 0.00015039577836411608, | |
| "loss": 1.2203, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.07667900581702802, | |
| "grad_norm": 0.6875, | |
| "learning_rate": 0.00015303430079155673, | |
| "loss": 1.2122, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.07800105764145955, | |
| "grad_norm": 1.1015625, | |
| "learning_rate": 0.00015567282321899737, | |
| "loss": 1.2035, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.07932310946589106, | |
| "grad_norm": 0.3515625, | |
| "learning_rate": 0.000158311345646438, | |
| "loss": 1.1933, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.08064516129032258, | |
| "grad_norm": 0.4453125, | |
| "learning_rate": 0.00016094986807387864, | |
| "loss": 1.1926, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.08196721311475409, | |
| "grad_norm": 0.404296875, | |
| "learning_rate": 0.00016358839050131926, | |
| "loss": 1.1833, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.08328926493918562, | |
| "grad_norm": 0.5703125, | |
| "learning_rate": 0.00016622691292875988, | |
| "loss": 1.1753, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.08461131676361713, | |
| "grad_norm": 0.6484375, | |
| "learning_rate": 0.00016886543535620053, | |
| "loss": 1.1717, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.08593336858804865, | |
| "grad_norm": 0.66015625, | |
| "learning_rate": 0.00017150395778364118, | |
| "loss": 1.1684, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.08725542041248018, | |
| "grad_norm": 0.58984375, | |
| "learning_rate": 0.0001741424802110818, | |
| "loss": 1.1608, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.08857747223691169, | |
| "grad_norm": 0.5625, | |
| "learning_rate": 0.00017678100263852244, | |
| "loss": 1.1563, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.0898995240613432, | |
| "grad_norm": 0.47265625, | |
| "learning_rate": 0.00017941952506596306, | |
| "loss": 1.1501, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.09122157588577472, | |
| "grad_norm": 0.58203125, | |
| "learning_rate": 0.00018205804749340368, | |
| "loss": 1.1451, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.09254362771020624, | |
| "grad_norm": 0.59765625, | |
| "learning_rate": 0.00018469656992084433, | |
| "loss": 1.1338, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.09386567953463776, | |
| "grad_norm": 0.484375, | |
| "learning_rate": 0.00018733509234828498, | |
| "loss": 1.1326, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.09518773135906927, | |
| "grad_norm": 0.55859375, | |
| "learning_rate": 0.0001899736147757256, | |
| "loss": 1.1286, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.0965097831835008, | |
| "grad_norm": 0.62109375, | |
| "learning_rate": 0.00019261213720316625, | |
| "loss": 1.1219, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.09783183500793231, | |
| "grad_norm": 0.69140625, | |
| "learning_rate": 0.00019525065963060687, | |
| "loss": 1.1238, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.09915388683236383, | |
| "grad_norm": 0.57421875, | |
| "learning_rate": 0.00019788918205804749, | |
| "loss": 1.1165, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.10047593865679534, | |
| "grad_norm": 0.490234375, | |
| "learning_rate": 0.00019999995738663592, | |
| "loss": 1.1111, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.10179799048122687, | |
| "grad_norm": 0.6484375, | |
| "learning_rate": 0.00019999846592270624, | |
| "loss": 1.0969, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.10312004230565838, | |
| "grad_norm": 0.8046875, | |
| "learning_rate": 0.00019999484382688995, | |
| "loss": 1.1008, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.1044420941300899, | |
| "grad_norm": 0.65625, | |
| "learning_rate": 0.00019998909117636182, | |
| "loss": 1.0969, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.10576414595452142, | |
| "grad_norm": 0.56640625, | |
| "learning_rate": 0.00019998120809369154, | |
| "loss": 1.0888, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.10708619777895294, | |
| "grad_norm": 1.1875, | |
| "learning_rate": 0.0001999711947468411, | |
| "loss": 1.0926, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.10840824960338445, | |
| "grad_norm": 1.3515625, | |
| "learning_rate": 0.00019995905134916143, | |
| "loss": 1.08, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.10973030142781597, | |
| "grad_norm": 0.58984375, | |
| "learning_rate": 0.0001999447781593875, | |
| "loss": 1.0778, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.1110523532522475, | |
| "grad_norm": 0.453125, | |
| "learning_rate": 0.00019992837548163316, | |
| "loss": 1.0771, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.11237440507667901, | |
| "grad_norm": 0.70703125, | |
| "learning_rate": 0.00019990984366538442, | |
| "loss": 1.0722, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.11369645690111052, | |
| "grad_norm": 0.49609375, | |
| "learning_rate": 0.00019988918310549222, | |
| "loss": 1.0628, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.11501850872554203, | |
| "grad_norm": 0.45703125, | |
| "learning_rate": 0.0001998663942421637, | |
| "loss": 1.0611, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.11634056054997356, | |
| "grad_norm": 0.54296875, | |
| "learning_rate": 0.00019984147756095308, | |
| "loss": 1.0627, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.11766261237440508, | |
| "grad_norm": 0.78515625, | |
| "learning_rate": 0.0001998144335927513, | |
| "loss": 1.0562, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.11898466419883659, | |
| "grad_norm": 0.439453125, | |
| "learning_rate": 0.00019978526291377464, | |
| "loss": 1.0543, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.12030671602326812, | |
| "grad_norm": 0.392578125, | |
| "learning_rate": 0.0001997539661455524, | |
| "loss": 1.0532, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.12162876784769963, | |
| "grad_norm": 0.439453125, | |
| "learning_rate": 0.00019972054395491368, | |
| "loss": 1.0477, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.12295081967213115, | |
| "grad_norm": 0.59765625, | |
| "learning_rate": 0.00019968499705397331, | |
| "loss": 1.044, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.12427287149656266, | |
| "grad_norm": 0.451171875, | |
| "learning_rate": 0.00019964732620011651, | |
| "loss": 1.0451, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.12559492332099417, | |
| "grad_norm": 0.365234375, | |
| "learning_rate": 0.00019960753219598281, | |
| "loss": 1.0385, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.12691697514542571, | |
| "grad_norm": 0.4375, | |
| "learning_rate": 0.00019956561588944897, | |
| "loss": 1.0376, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.12823902696985723, | |
| "grad_norm": 0.39453125, | |
| "learning_rate": 0.00019952157817361088, | |
| "loss": 1.0396, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.12956107879428874, | |
| "grad_norm": 0.484375, | |
| "learning_rate": 0.0001994754199867645, | |
| "loss": 1.0342, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.13088313061872026, | |
| "grad_norm": 0.640625, | |
| "learning_rate": 0.00019942714231238604, | |
| "loss": 1.036, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.13220518244315177, | |
| "grad_norm": 0.451171875, | |
| "learning_rate": 0.00019937674617911077, | |
| "loss": 1.0231, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.13352723426758328, | |
| "grad_norm": 0.400390625, | |
| "learning_rate": 0.00019932423266071122, | |
| "loss": 1.0286, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.1348492860920148, | |
| "grad_norm": 0.322265625, | |
| "learning_rate": 0.00019926960287607436, | |
| "loss": 1.0238, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.1361713379164463, | |
| "grad_norm": 0.39453125, | |
| "learning_rate": 0.0001992128579891777, | |
| "loss": 1.0216, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.13749338974087785, | |
| "grad_norm": 0.287109375, | |
| "learning_rate": 0.00019915399920906432, | |
| "loss": 1.0237, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.13881544156530937, | |
| "grad_norm": 0.443359375, | |
| "learning_rate": 0.0001990930277898175, | |
| "loss": 1.0137, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.14013749338974088, | |
| "grad_norm": 0.5, | |
| "learning_rate": 0.00019902994503053363, | |
| "loss": 1.0206, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.1414595452141724, | |
| "grad_norm": 0.470703125, | |
| "learning_rate": 0.0001989647522752947, | |
| "loss": 1.0189, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.1427815970386039, | |
| "grad_norm": 0.486328125, | |
| "learning_rate": 0.0001988974509131397, | |
| "loss": 1.0109, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.14410364886303542, | |
| "grad_norm": 0.337890625, | |
| "learning_rate": 0.00019882804237803488, | |
| "loss": 1.0069, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.14542570068746694, | |
| "grad_norm": 0.224609375, | |
| "learning_rate": 0.00019875652814884326, | |
| "loss": 1.0097, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.14674775251189848, | |
| "grad_norm": 0.48828125, | |
| "learning_rate": 0.00019868290974929328, | |
| "loss": 1.0033, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.14806980433633, | |
| "grad_norm": 0.8671875, | |
| "learning_rate": 0.00019860718874794602, | |
| "loss": 1.0034, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.1493918561607615, | |
| "grad_norm": 0.69921875, | |
| "learning_rate": 0.00019852936675816209, | |
| "loss": 1.0019, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.15071390798519302, | |
| "grad_norm": 0.62890625, | |
| "learning_rate": 0.000198449445438067, | |
| "loss": 1.0004, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.15203595980962453, | |
| "grad_norm": 0.5078125, | |
| "learning_rate": 0.00019836742649051602, | |
| "loss": 0.998, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.15335801163405605, | |
| "grad_norm": 0.69140625, | |
| "learning_rate": 0.00019828331166305785, | |
| "loss": 0.9944, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.15468006345848756, | |
| "grad_norm": 0.427734375, | |
| "learning_rate": 0.00019819710274789727, | |
| "loss": 0.9951, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.1560021152829191, | |
| "grad_norm": 0.5, | |
| "learning_rate": 0.0001981088015818571, | |
| "loss": 0.9962, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.15732416710735062, | |
| "grad_norm": 0.7265625, | |
| "learning_rate": 0.00019801841004633906, | |
| "loss": 0.9884, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.15864621893178213, | |
| "grad_norm": 0.5546875, | |
| "learning_rate": 0.00019792593006728347, | |
| "loss": 0.9918, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.15996827075621364, | |
| "grad_norm": 0.796875, | |
| "learning_rate": 0.00019783136361512858, | |
| "loss": 0.993, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.16129032258064516, | |
| "grad_norm": 0.6875, | |
| "learning_rate": 0.00019773471270476822, | |
| "loss": 0.9902, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.16261237440507667, | |
| "grad_norm": 0.5390625, | |
| "learning_rate": 0.0001976359793955091, | |
| "loss": 0.9843, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.16393442622950818, | |
| "grad_norm": 0.306640625, | |
| "learning_rate": 0.0001975351657910269, | |
| "loss": 0.988, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.16525647805393973, | |
| "grad_norm": 0.54296875, | |
| "learning_rate": 0.00019743227403932134, | |
| "loss": 0.9874, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.16657852987837124, | |
| "grad_norm": 0.54296875, | |
| "learning_rate": 0.00019732730633267056, | |
| "loss": 0.9782, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.16790058170280275, | |
| "grad_norm": 0.6171875, | |
| "learning_rate": 0.00019722026490758424, | |
| "loss": 0.9813, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.16922263352723427, | |
| "grad_norm": 0.412109375, | |
| "learning_rate": 0.00019711115204475616, | |
| "loss": 0.9793, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.17054468535166578, | |
| "grad_norm": 0.36328125, | |
| "learning_rate": 0.0001969999700690154, | |
| "loss": 0.9749, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.1718667371760973, | |
| "grad_norm": 0.255859375, | |
| "learning_rate": 0.00019688672134927693, | |
| "loss": 0.9768, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.1731887890005288, | |
| "grad_norm": 0.38671875, | |
| "learning_rate": 0.00019677140829849112, | |
| "loss": 0.973, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.17451084082496035, | |
| "grad_norm": 0.4453125, | |
| "learning_rate": 0.00019665403337359227, | |
| "loss": 0.972, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.17583289264939186, | |
| "grad_norm": 0.291015625, | |
| "learning_rate": 0.00019653459907544634, | |
| "loss": 0.9759, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.17715494447382338, | |
| "grad_norm": 0.3671875, | |
| "learning_rate": 0.0001964131079487976, | |
| "loss": 0.9691, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.1784769962982549, | |
| "grad_norm": 0.330078125, | |
| "learning_rate": 0.0001962895625822144, | |
| "loss": 0.9681, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.1797990481226864, | |
| "grad_norm": 0.48828125, | |
| "learning_rate": 0.0001961639656080342, | |
| "loss": 0.9695, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.18112109994711792, | |
| "grad_norm": 0.435546875, | |
| "learning_rate": 0.00019603631970230713, | |
| "loss": 0.9695, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.18244315177154943, | |
| "grad_norm": 0.369140625, | |
| "learning_rate": 0.00019590662758473934, | |
| "loss": 0.9685, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.18376520359598097, | |
| "grad_norm": 0.37109375, | |
| "learning_rate": 0.0001957748920186348, | |
| "loss": 0.9681, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.1850872554204125, | |
| "grad_norm": 0.3984375, | |
| "learning_rate": 0.00019564111581083657, | |
| "loss": 0.9632, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.186409307244844, | |
| "grad_norm": 0.330078125, | |
| "learning_rate": 0.00019550530181166692, | |
| "loss": 0.9683, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.18773135906927552, | |
| "grad_norm": 0.4453125, | |
| "learning_rate": 0.0001953674529148666, | |
| "loss": 0.9628, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.18905341089370703, | |
| "grad_norm": 0.259765625, | |
| "learning_rate": 0.00019522757205753323, | |
| "loss": 0.9625, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.19037546271813854, | |
| "grad_norm": 0.294921875, | |
| "learning_rate": 0.00019508566222005866, | |
| "loss": 0.9583, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.19169751454257006, | |
| "grad_norm": 0.330078125, | |
| "learning_rate": 0.00019494172642606553, | |
| "loss": 0.9634, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.1930195663670016, | |
| "grad_norm": 0.4140625, | |
| "learning_rate": 0.00019479576774234284, | |
| "loss": 0.9602, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.1943416181914331, | |
| "grad_norm": 0.345703125, | |
| "learning_rate": 0.00019464778927878048, | |
| "loss": 0.967, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.19566367001586463, | |
| "grad_norm": 0.44140625, | |
| "learning_rate": 0.00019449779418830322, | |
| "loss": 0.9549, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.19698572184029614, | |
| "grad_norm": 0.279296875, | |
| "learning_rate": 0.0001943457856668033, | |
| "loss": 0.9652, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.19830777366472765, | |
| "grad_norm": 0.400390625, | |
| "learning_rate": 0.00019419176695307245, | |
| "loss": 0.9541, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.19962982548915917, | |
| "grad_norm": 0.419921875, | |
| "learning_rate": 0.00019403574132873276, | |
| "loss": 0.9506, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.20095187731359068, | |
| "grad_norm": 0.421875, | |
| "learning_rate": 0.00019387771211816705, | |
| "loss": 0.9572, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.20227392913802222, | |
| "grad_norm": 0.2265625, | |
| "learning_rate": 0.00019371768268844762, | |
| "loss": 0.9525, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.20359598096245374, | |
| "grad_norm": 0.3203125, | |
| "learning_rate": 0.0001935556564492649, | |
| "loss": 0.9542, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.20491803278688525, | |
| "grad_norm": 0.388671875, | |
| "learning_rate": 0.0001933916368528545, | |
| "loss": 0.9532, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.20624008461131677, | |
| "grad_norm": 0.318359375, | |
| "learning_rate": 0.00019322562739392394, | |
| "loss": 0.9511, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.20756213643574828, | |
| "grad_norm": 0.4296875, | |
| "learning_rate": 0.00019305763160957788, | |
| "loss": 0.9536, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.2088841882601798, | |
| "grad_norm": 0.322265625, | |
| "learning_rate": 0.00019288765307924299, | |
| "loss": 0.9511, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.2102062400846113, | |
| "grad_norm": 0.3125, | |
| "learning_rate": 0.00019271569542459165, | |
| "loss": 0.9497, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.21152829190904285, | |
| "grad_norm": 0.259765625, | |
| "learning_rate": 0.00019254176230946462, | |
| "loss": 0.9484, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.21285034373347436, | |
| "grad_norm": 0.466796875, | |
| "learning_rate": 0.00019236585743979322, | |
| "loss": 0.9442, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.21417239555790588, | |
| "grad_norm": 0.3671875, | |
| "learning_rate": 0.00019218798456352012, | |
| "loss": 0.9426, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.2154944473823374, | |
| "grad_norm": 0.40234375, | |
| "learning_rate": 0.00019200814747051976, | |
| "loss": 0.9455, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.2168164992067689, | |
| "grad_norm": 0.349609375, | |
| "learning_rate": 0.00019182634999251728, | |
| "loss": 0.9419, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.21813855103120042, | |
| "grad_norm": 0.29296875, | |
| "learning_rate": 0.00019164259600300723, | |
| "loss": 0.947, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.21946060285563193, | |
| "grad_norm": 0.3359375, | |
| "learning_rate": 0.00019145688941717075, | |
| "loss": 0.9478, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.22078265468006344, | |
| "grad_norm": 0.396484375, | |
| "learning_rate": 0.0001912692341917922, | |
| "loss": 0.9433, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.222104706504495, | |
| "grad_norm": 0.310546875, | |
| "learning_rate": 0.00019107963432517505, | |
| "loss": 0.9461, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.2234267583289265, | |
| "grad_norm": 0.416015625, | |
| "learning_rate": 0.00019088809385705646, | |
| "loss": 0.9421, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.22474881015335801, | |
| "grad_norm": 0.439453125, | |
| "learning_rate": 0.00019069461686852134, | |
| "loss": 0.9436, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.22607086197778953, | |
| "grad_norm": 0.26171875, | |
| "learning_rate": 0.0001904992074819153, | |
| "loss": 0.9381, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.22739291380222104, | |
| "grad_norm": 0.2470703125, | |
| "learning_rate": 0.00019030186986075703, | |
| "loss": 0.9408, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.22871496562665256, | |
| "grad_norm": 0.3515625, | |
| "learning_rate": 0.0001901026082096492, | |
| "loss": 0.9371, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.23003701745108407, | |
| "grad_norm": 0.404296875, | |
| "learning_rate": 0.00018990142677418923, | |
| "loss": 0.9387, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.2313590692755156, | |
| "grad_norm": 0.455078125, | |
| "learning_rate": 0.00018969832984087873, | |
| "loss": 0.9372, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.23268112109994712, | |
| "grad_norm": 0.265625, | |
| "learning_rate": 0.00018949332173703206, | |
| "loss": 0.936, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.23400317292437864, | |
| "grad_norm": 0.298828125, | |
| "learning_rate": 0.0001892864068306843, | |
| "loss": 0.9405, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.23532522474881015, | |
| "grad_norm": 0.373046875, | |
| "learning_rate": 0.00018907758953049805, | |
| "loss": 0.9333, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.23664727657324167, | |
| "grad_norm": 0.3984375, | |
| "learning_rate": 0.00018886687428566954, | |
| "loss": 0.9356, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.23796932839767318, | |
| "grad_norm": 0.50390625, | |
| "learning_rate": 0.00018865426558583383, | |
| "loss": 0.9344, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.2392913802221047, | |
| "grad_norm": 0.291015625, | |
| "learning_rate": 0.00018843976796096917, | |
| "loss": 0.9355, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.24061343204653624, | |
| "grad_norm": 0.353515625, | |
| "learning_rate": 0.00018822338598130047, | |
| "loss": 0.9336, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.24193548387096775, | |
| "grad_norm": 0.2490234375, | |
| "learning_rate": 0.0001880051242572019, | |
| "loss": 0.9304, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.24325753569539926, | |
| "grad_norm": 0.5703125, | |
| "learning_rate": 0.00018778498743909873, | |
| "loss": 0.931, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.24457958751983078, | |
| "grad_norm": 0.259765625, | |
| "learning_rate": 0.00018756298021736808, | |
| "loss": 0.934, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.2459016393442623, | |
| "grad_norm": 0.2890625, | |
| "learning_rate": 0.00018733910732223925, | |
| "loss": 0.9273, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.2472236911686938, | |
| "grad_norm": 0.279296875, | |
| "learning_rate": 0.00018711337352369264, | |
| "loss": 0.9315, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.24854574299312532, | |
| "grad_norm": 0.341796875, | |
| "learning_rate": 0.0001868857836313584, | |
| "loss": 0.9316, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.24986779481755686, | |
| "grad_norm": 0.2578125, | |
| "learning_rate": 0.00018665634249441366, | |
| "loss": 0.9319, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.25118984664198835, | |
| "grad_norm": 0.337890625, | |
| "learning_rate": 0.0001864250550014795, | |
| "loss": 0.9297, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.25251189846641986, | |
| "grad_norm": 0.43359375, | |
| "learning_rate": 0.00018619192608051655, | |
| "loss": 0.9237, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.25383395029085143, | |
| "grad_norm": 0.255859375, | |
| "learning_rate": 0.00018595696069872013, | |
| "loss": 0.9293, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.25515600211528294, | |
| "grad_norm": 0.279296875, | |
| "learning_rate": 0.00018572016386241442, | |
| "loss": 0.9308, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.25647805393971446, | |
| "grad_norm": 0.546875, | |
| "learning_rate": 0.00018548154061694572, | |
| "loss": 0.9207, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.25780010576414597, | |
| "grad_norm": 0.369140625, | |
| "learning_rate": 0.00018524109604657496, | |
| "loss": 0.9285, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.2591221575885775, | |
| "grad_norm": 0.3359375, | |
| "learning_rate": 0.00018499883527436947, | |
| "loss": 0.9268, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.260444209413009, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 0.0001847547634620936, | |
| "loss": 0.9284, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.2617662612374405, | |
| "grad_norm": 0.4296875, | |
| "learning_rate": 0.00018450888581009908, | |
| "loss": 0.9264, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.263088313061872, | |
| "grad_norm": 0.359375, | |
| "learning_rate": 0.00018426120755721386, | |
| "loss": 0.926, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.26441036488630354, | |
| "grad_norm": 0.345703125, | |
| "learning_rate": 0.0001840117339806308, | |
| "loss": 0.9218, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.26573241671073505, | |
| "grad_norm": 0.326171875, | |
| "learning_rate": 0.00018376047039579495, | |
| "loss": 0.9245, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.26705446853516657, | |
| "grad_norm": 0.2578125, | |
| "learning_rate": 0.00018350742215629055, | |
| "loss": 0.9221, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.2683765203595981, | |
| "grad_norm": 0.42578125, | |
| "learning_rate": 0.00018325259465372677, | |
| "loss": 0.926, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 0.2696985721840296, | |
| "grad_norm": 0.4140625, | |
| "learning_rate": 0.0001829959933176229, | |
| "loss": 0.9235, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.2710206240084611, | |
| "grad_norm": 0.4375, | |
| "learning_rate": 0.00018273762361529274, | |
| "loss": 0.9237, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.2723426758328926, | |
| "grad_norm": 0.2314453125, | |
| "learning_rate": 0.000182477491051728, | |
| "loss": 0.9203, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.2736647276573242, | |
| "grad_norm": 0.396484375, | |
| "learning_rate": 0.00018221560116948103, | |
| "loss": 0.9206, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.2749867794817557, | |
| "grad_norm": 0.345703125, | |
| "learning_rate": 0.00018195195954854676, | |
| "loss": 0.924, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.2763088313061872, | |
| "grad_norm": 0.53125, | |
| "learning_rate": 0.00018168657180624384, | |
| "loss": 0.918, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 0.27763088313061873, | |
| "grad_norm": 0.337890625, | |
| "learning_rate": 0.0001814194435970949, | |
| "loss": 0.9298, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.27895293495505025, | |
| "grad_norm": 0.7421875, | |
| "learning_rate": 0.00018115058061270598, | |
| "loss": 0.9195, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 0.28027498677948176, | |
| "grad_norm": 0.57421875, | |
| "learning_rate": 0.0001808799885816455, | |
| "loss": 0.9196, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.2815970386039133, | |
| "grad_norm": 0.39453125, | |
| "learning_rate": 0.00018060767326932194, | |
| "loss": 0.9154, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.2829190904283448, | |
| "grad_norm": 0.38671875, | |
| "learning_rate": 0.00018033364047786128, | |
| "loss": 0.9186, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.2842411422527763, | |
| "grad_norm": 0.400390625, | |
| "learning_rate": 0.00018005789604598303, | |
| "loss": 0.9173, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.2855631940772078, | |
| "grad_norm": 0.384765625, | |
| "learning_rate": 0.0001797804458488762, | |
| "loss": 0.9199, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.28688524590163933, | |
| "grad_norm": 0.71875, | |
| "learning_rate": 0.00017950129579807374, | |
| "loss": 0.9202, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 0.28820729772607084, | |
| "grad_norm": 0.55859375, | |
| "learning_rate": 0.00017922045184132698, | |
| "loss": 0.922, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.28952934955050236, | |
| "grad_norm": 0.640625, | |
| "learning_rate": 0.00017893791996247856, | |
| "loss": 0.9157, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.29085140137493387, | |
| "grad_norm": 0.263671875, | |
| "learning_rate": 0.00017865370618133511, | |
| "loss": 0.9145, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.29217345319936544, | |
| "grad_norm": 0.306640625, | |
| "learning_rate": 0.00017836781655353905, | |
| "loss": 0.9114, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 0.29349550502379695, | |
| "grad_norm": 0.314453125, | |
| "learning_rate": 0.00017808025717043938, | |
| "loss": 0.9128, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.29481755684822847, | |
| "grad_norm": 0.2392578125, | |
| "learning_rate": 0.00017779103415896193, | |
| "loss": 0.9182, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 0.29613960867266, | |
| "grad_norm": 0.421875, | |
| "learning_rate": 0.000177500153681479, | |
| "loss": 0.9127, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.2974616604970915, | |
| "grad_norm": 0.59765625, | |
| "learning_rate": 0.00017720762193567787, | |
| "loss": 0.9122, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.298783712321523, | |
| "grad_norm": 0.255859375, | |
| "learning_rate": 0.00017691344515442879, | |
| "loss": 0.9117, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.3001057641459545, | |
| "grad_norm": 0.2333984375, | |
| "learning_rate": 0.00017661762960565223, | |
| "loss": 0.9097, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 0.30142781597038604, | |
| "grad_norm": 0.376953125, | |
| "learning_rate": 0.0001763201815921853, | |
| "loss": 0.9085, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.30274986779481755, | |
| "grad_norm": 0.2412109375, | |
| "learning_rate": 0.0001760211074516474, | |
| "loss": 0.915, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 0.30407191961924906, | |
| "grad_norm": 0.291015625, | |
| "learning_rate": 0.00017572041355630536, | |
| "loss": 0.9145, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.3053939714436806, | |
| "grad_norm": 0.2294921875, | |
| "learning_rate": 0.00017541810631293742, | |
| "loss": 0.9144, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 0.3067160232681121, | |
| "grad_norm": 0.25, | |
| "learning_rate": 0.00017511419216269695, | |
| "loss": 0.9129, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.3080380750925436, | |
| "grad_norm": 0.2490234375, | |
| "learning_rate": 0.00017480867758097506, | |
| "loss": 0.9126, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 0.3093601269169751, | |
| "grad_norm": 0.2431640625, | |
| "learning_rate": 0.00017450156907726272, | |
| "loss": 0.9139, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.3106821787414067, | |
| "grad_norm": 0.2421875, | |
| "learning_rate": 0.00017419287319501197, | |
| "loss": 0.9064, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 0.3120042305658382, | |
| "grad_norm": 0.2578125, | |
| "learning_rate": 0.00017388259651149673, | |
| "loss": 0.9165, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.3133262823902697, | |
| "grad_norm": 0.263671875, | |
| "learning_rate": 0.00017357074563767225, | |
| "loss": 0.9102, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 0.31464833421470123, | |
| "grad_norm": 0.2412109375, | |
| "learning_rate": 0.00017325732721803466, | |
| "loss": 0.9102, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.31597038603913274, | |
| "grad_norm": 0.302734375, | |
| "learning_rate": 0.0001729423479304792, | |
| "loss": 0.9092, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 0.31729243786356426, | |
| "grad_norm": 0.30078125, | |
| "learning_rate": 0.00017262581448615794, | |
| "loss": 0.9076, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.31861448968799577, | |
| "grad_norm": 0.265625, | |
| "learning_rate": 0.00017230773362933687, | |
| "loss": 0.9073, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 0.3199365415124273, | |
| "grad_norm": 0.294921875, | |
| "learning_rate": 0.0001719881121372521, | |
| "loss": 0.9074, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.3212585933368588, | |
| "grad_norm": 0.2197265625, | |
| "learning_rate": 0.00017166695681996555, | |
| "loss": 0.912, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 0.3225806451612903, | |
| "grad_norm": 0.361328125, | |
| "learning_rate": 0.00017134427452021974, | |
| "loss": 0.9061, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.3239026969857218, | |
| "grad_norm": 0.2890625, | |
| "learning_rate": 0.00017102007211329214, | |
| "loss": 0.9084, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 0.32522474881015334, | |
| "grad_norm": 0.251953125, | |
| "learning_rate": 0.0001706943565068486, | |
| "loss": 0.9091, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.32654680063458486, | |
| "grad_norm": 0.2080078125, | |
| "learning_rate": 0.00017036713464079612, | |
| "loss": 0.9092, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 0.32786885245901637, | |
| "grad_norm": 0.326171875, | |
| "learning_rate": 0.0001700384134871351, | |
| "loss": 0.9069, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.32919090428344794, | |
| "grad_norm": 0.310546875, | |
| "learning_rate": 0.00016970820004981067, | |
| "loss": 0.9082, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 0.33051295610787945, | |
| "grad_norm": 0.33203125, | |
| "learning_rate": 0.00016937650136456358, | |
| "loss": 0.9086, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.33183500793231097, | |
| "grad_norm": 0.1875, | |
| "learning_rate": 0.0001690433244987802, | |
| "loss": 0.9041, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 0.3331570597567425, | |
| "grad_norm": 0.2578125, | |
| "learning_rate": 0.000168708676551342, | |
| "loss": 0.9098, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.334479111581174, | |
| "grad_norm": 0.30859375, | |
| "learning_rate": 0.00016837256465247418, | |
| "loss": 0.9024, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 0.3358011634056055, | |
| "grad_norm": 0.2412109375, | |
| "learning_rate": 0.00016803499596359392, | |
| "loss": 0.9061, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.337123215230037, | |
| "grad_norm": 0.232421875, | |
| "learning_rate": 0.0001676959776771577, | |
| "loss": 0.9094, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 0.33844526705446853, | |
| "grad_norm": 0.234375, | |
| "learning_rate": 0.00016735551701650803, | |
| "loss": 0.9035, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.33976731887890005, | |
| "grad_norm": 0.294921875, | |
| "learning_rate": 0.00016701362123571959, | |
| "loss": 0.9071, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 0.34108937070333156, | |
| "grad_norm": 0.421875, | |
| "learning_rate": 0.00016667029761944466, | |
| "loss": 0.9064, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.3424114225277631, | |
| "grad_norm": 0.2373046875, | |
| "learning_rate": 0.00016632555348275788, | |
| "loss": 0.9043, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 0.3437334743521946, | |
| "grad_norm": 0.32421875, | |
| "learning_rate": 0.00016597939617100046, | |
| "loss": 0.9039, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.3450555261766261, | |
| "grad_norm": 0.482421875, | |
| "learning_rate": 0.00016563183305962363, | |
| "loss": 0.9042, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 0.3463775780010576, | |
| "grad_norm": 0.310546875, | |
| "learning_rate": 0.0001652828715540314, | |
| "loss": 0.9026, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.34769962982548913, | |
| "grad_norm": 0.29296875, | |
| "learning_rate": 0.00016493251908942302, | |
| "loss": 0.9043, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 0.3490216816499207, | |
| "grad_norm": 0.392578125, | |
| "learning_rate": 0.0001645807831306343, | |
| "loss": 0.9035, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.3503437334743522, | |
| "grad_norm": 0.384765625, | |
| "learning_rate": 0.00016422767117197867, | |
| "loss": 0.9016, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 0.35166578529878373, | |
| "grad_norm": 0.3125, | |
| "learning_rate": 0.0001638731907370876, | |
| "loss": 0.8989, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.35298783712321524, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 0.00016351734937875007, | |
| "loss": 0.8979, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 0.35430988894764676, | |
| "grad_norm": 0.3125, | |
| "learning_rate": 0.00016316015467875188, | |
| "loss": 0.904, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.35563194077207827, | |
| "grad_norm": 0.326171875, | |
| "learning_rate": 0.00016280161424771396, | |
| "loss": 0.8991, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 0.3569539925965098, | |
| "grad_norm": 0.27734375, | |
| "learning_rate": 0.0001624417357249302, | |
| "loss": 0.901, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.3582760444209413, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 0.00016208052677820484, | |
| "loss": 0.904, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 0.3595980962453728, | |
| "grad_norm": 0.2470703125, | |
| "learning_rate": 0.00016171799510368895, | |
| "loss": 0.8977, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.3609201480698043, | |
| "grad_norm": 0.296875, | |
| "learning_rate": 0.00016135414842571643, | |
| "loss": 0.9002, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 0.36224219989423584, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 0.0001609889944966396, | |
| "loss": 0.8996, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.36356425171866735, | |
| "grad_norm": 0.27734375, | |
| "learning_rate": 0.0001606225410966638, | |
| "loss": 0.9047, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 0.36488630354309887, | |
| "grad_norm": 0.33984375, | |
| "learning_rate": 0.0001602547960336819, | |
| "loss": 0.9004, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.3662083553675304, | |
| "grad_norm": 0.3125, | |
| "learning_rate": 0.00015988576714310755, | |
| "loss": 0.902, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 0.36753040719196195, | |
| "grad_norm": 0.25, | |
| "learning_rate": 0.00015951546228770868, | |
| "loss": 0.8999, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.36885245901639346, | |
| "grad_norm": 0.248046875, | |
| "learning_rate": 0.0001591438893574396, | |
| "loss": 0.9047, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 0.370174510840825, | |
| "grad_norm": 0.2451171875, | |
| "learning_rate": 0.0001587710562692731, | |
| "loss": 0.8982, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.3714965626652565, | |
| "grad_norm": 0.26171875, | |
| "learning_rate": 0.00015839697096703176, | |
| "loss": 0.9009, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 0.372818614489688, | |
| "grad_norm": 0.359375, | |
| "learning_rate": 0.00015802164142121854, | |
| "loss": 0.9023, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.3741406663141195, | |
| "grad_norm": 0.392578125, | |
| "learning_rate": 0.0001576450756288471, | |
| "loss": 0.9018, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 0.37546271813855103, | |
| "grad_norm": 0.341796875, | |
| "learning_rate": 0.0001572672816132714, | |
| "loss": 0.9006, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.37678476996298255, | |
| "grad_norm": 0.212890625, | |
| "learning_rate": 0.00015688826742401465, | |
| "loss": 0.8996, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 0.37810682178741406, | |
| "grad_norm": 0.2314453125, | |
| "learning_rate": 0.00015650804113659793, | |
| "loss": 0.8957, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.3794288736118456, | |
| "grad_norm": 0.369140625, | |
| "learning_rate": 0.00015612661085236807, | |
| "loss": 0.901, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 0.3807509254362771, | |
| "grad_norm": 0.2265625, | |
| "learning_rate": 0.00015574398469832493, | |
| "loss": 0.8984, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.3820729772607086, | |
| "grad_norm": 0.287109375, | |
| "learning_rate": 0.00015536017082694846, | |
| "loss": 0.8983, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 0.3833950290851401, | |
| "grad_norm": 0.2392578125, | |
| "learning_rate": 0.00015497517741602486, | |
| "loss": 0.8992, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.38471708090957163, | |
| "grad_norm": 0.2158203125, | |
| "learning_rate": 0.00015458901266847232, | |
| "loss": 0.8988, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 0.3860391327340032, | |
| "grad_norm": 0.369140625, | |
| "learning_rate": 0.0001542016848121663, | |
| "loss": 0.8982, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.3873611845584347, | |
| "grad_norm": 0.314453125, | |
| "learning_rate": 0.00015381320209976426, | |
| "loss": 0.9009, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 0.3886832363828662, | |
| "grad_norm": 0.3203125, | |
| "learning_rate": 0.0001534235728085297, | |
| "loss": 0.9006, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.39000528820729774, | |
| "grad_norm": 0.2216796875, | |
| "learning_rate": 0.00015303280524015585, | |
| "loss": 0.9001, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 0.39132734003172925, | |
| "grad_norm": 0.228515625, | |
| "learning_rate": 0.0001526409077205889, | |
| "loss": 0.894, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.39264939185616077, | |
| "grad_norm": 0.244140625, | |
| "learning_rate": 0.00015224788859985043, | |
| "loss": 0.8955, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 0.3939714436805923, | |
| "grad_norm": 0.322265625, | |
| "learning_rate": 0.00015185375625185964, | |
| "loss": 0.902, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.3952934955050238, | |
| "grad_norm": 0.322265625, | |
| "learning_rate": 0.00015145851907425484, | |
| "loss": 0.8995, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 0.3966155473294553, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 0.0001510621854882145, | |
| "loss": 0.8997, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.3979375991538868, | |
| "grad_norm": 0.2158203125, | |
| "learning_rate": 0.00015066476393827799, | |
| "loss": 0.8915, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 0.39925965097831834, | |
| "grad_norm": 0.2421875, | |
| "learning_rate": 0.00015026626289216542, | |
| "loss": 0.8916, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.40058170280274985, | |
| "grad_norm": 0.3125, | |
| "learning_rate": 0.0001498666908405975, | |
| "loss": 0.8961, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 0.40190375462718136, | |
| "grad_norm": 0.3359375, | |
| "learning_rate": 0.00014946605629711425, | |
| "loss": 0.8991, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.4032258064516129, | |
| "grad_norm": 0.306640625, | |
| "learning_rate": 0.00014906436779789402, | |
| "loss": 0.8932, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 0.40454785827604445, | |
| "grad_norm": 0.267578125, | |
| "learning_rate": 0.00014866163390157136, | |
| "loss": 0.8952, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.40586991010047596, | |
| "grad_norm": 0.51953125, | |
| "learning_rate": 0.00014825786318905469, | |
| "loss": 0.893, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 0.4071919619249075, | |
| "grad_norm": 0.43359375, | |
| "learning_rate": 0.00014785306426334354, | |
| "loss": 0.8961, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.408514013749339, | |
| "grad_norm": 0.244140625, | |
| "learning_rate": 0.0001474472457493452, | |
| "loss": 0.8916, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 0.4098360655737705, | |
| "grad_norm": 0.263671875, | |
| "learning_rate": 0.00014704041629369106, | |
| "loss": 0.9021, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.411158117398202, | |
| "grad_norm": 0.265625, | |
| "learning_rate": 0.00014663258456455212, | |
| "loss": 0.8992, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 0.41248016922263353, | |
| "grad_norm": 0.251953125, | |
| "learning_rate": 0.00014622375925145455, | |
| "loss": 0.8926, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.41380222104706504, | |
| "grad_norm": 0.25, | |
| "learning_rate": 0.0001458139490650945, | |
| "loss": 0.894, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 0.41512427287149656, | |
| "grad_norm": 0.33984375, | |
| "learning_rate": 0.0001454031627371524, | |
| "loss": 0.8948, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.41644632469592807, | |
| "grad_norm": 0.26953125, | |
| "learning_rate": 0.000144991409020107, | |
| "loss": 0.8933, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 0.4177683765203596, | |
| "grad_norm": 0.287109375, | |
| "learning_rate": 0.00014457869668704895, | |
| "loss": 0.9002, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.4190904283447911, | |
| "grad_norm": 0.201171875, | |
| "learning_rate": 0.0001441650345314936, | |
| "loss": 0.8961, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 0.4204124801692226, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 0.00014375043136719402, | |
| "loss": 0.8939, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.4217345319936541, | |
| "grad_norm": 0.2275390625, | |
| "learning_rate": 0.00014333489602795293, | |
| "loss": 0.8915, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 0.4230565838180857, | |
| "grad_norm": 0.265625, | |
| "learning_rate": 0.0001429184373674346, | |
| "loss": 0.8952, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.4243786356425172, | |
| "grad_norm": 0.2470703125, | |
| "learning_rate": 0.0001425010642589762, | |
| "loss": 0.8931, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 0.4257006874669487, | |
| "grad_norm": 0.283203125, | |
| "learning_rate": 0.0001420827855953986, | |
| "loss": 0.891, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.42702273929138024, | |
| "grad_norm": 0.267578125, | |
| "learning_rate": 0.00014166361028881716, | |
| "loss": 0.891, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 0.42834479111581175, | |
| "grad_norm": 0.2333984375, | |
| "learning_rate": 0.00014124354727045163, | |
| "loss": 0.8917, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.42966684294024327, | |
| "grad_norm": 0.2177734375, | |
| "learning_rate": 0.00014082260549043592, | |
| "loss": 0.8897, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 0.4309888947646748, | |
| "grad_norm": 0.34375, | |
| "learning_rate": 0.0001404007939176274, | |
| "loss": 0.895, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.4323109465891063, | |
| "grad_norm": 0.2060546875, | |
| "learning_rate": 0.00013997812153941584, | |
| "loss": 0.8965, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 0.4336329984135378, | |
| "grad_norm": 0.2314453125, | |
| "learning_rate": 0.00013955459736153183, | |
| "loss": 0.8954, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.4349550502379693, | |
| "grad_norm": 0.234375, | |
| "learning_rate": 0.00013913023040785502, | |
| "loss": 0.8982, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 0.43627710206240083, | |
| "grad_norm": 0.30859375, | |
| "learning_rate": 0.00013870502972022173, | |
| "loss": 0.8906, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.43759915388683235, | |
| "grad_norm": 0.29296875, | |
| "learning_rate": 0.00013827900435823236, | |
| "loss": 0.8945, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 0.43892120571126386, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 0.00013785216339905838, | |
| "loss": 0.8904, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.4402432575356954, | |
| "grad_norm": 0.41796875, | |
| "learning_rate": 0.0001374245159372489, | |
| "loss": 0.8924, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 0.4415653093601269, | |
| "grad_norm": 0.35546875, | |
| "learning_rate": 0.00013699607108453684, | |
| "loss": 0.8855, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.44288736118455846, | |
| "grad_norm": 0.26171875, | |
| "learning_rate": 0.00013656683796964495, | |
| "loss": 0.8914, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 0.44420941300899, | |
| "grad_norm": 0.2451171875, | |
| "learning_rate": 0.00013613682573809113, | |
| "loss": 0.8931, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.4455314648334215, | |
| "grad_norm": 0.359375, | |
| "learning_rate": 0.0001357060435519936, | |
| "loss": 0.8926, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 0.446853516657853, | |
| "grad_norm": 0.25390625, | |
| "learning_rate": 0.00013527450058987585, | |
| "loss": 0.8915, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.4481755684822845, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 0.0001348422060464709, | |
| "loss": 0.8878, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 0.44949762030671603, | |
| "grad_norm": 0.23046875, | |
| "learning_rate": 0.00013440916913252536, | |
| "loss": 0.89, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.45081967213114754, | |
| "grad_norm": 0.1767578125, | |
| "learning_rate": 0.0001339753990746034, | |
| "loss": 0.8881, | |
| "step": 1705 | |
| }, | |
| { | |
| "epoch": 0.45214172395557906, | |
| "grad_norm": 0.287109375, | |
| "learning_rate": 0.00013354090511488994, | |
| "loss": 0.8911, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.45346377578001057, | |
| "grad_norm": 0.283203125, | |
| "learning_rate": 0.00013310569651099388, | |
| "loss": 0.8904, | |
| "step": 1715 | |
| }, | |
| { | |
| "epoch": 0.4547858276044421, | |
| "grad_norm": 0.2314453125, | |
| "learning_rate": 0.00013266978253575074, | |
| "loss": 0.8902, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.4561078794288736, | |
| "grad_norm": 0.29296875, | |
| "learning_rate": 0.00013223317247702517, | |
| "loss": 0.8923, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 0.4574299312533051, | |
| "grad_norm": 0.244140625, | |
| "learning_rate": 0.000131795875637513, | |
| "loss": 0.8918, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.4587519830777366, | |
| "grad_norm": 0.34375, | |
| "learning_rate": 0.00013135790133454305, | |
| "loss": 0.8906, | |
| "step": 1735 | |
| }, | |
| { | |
| "epoch": 0.46007403490216814, | |
| "grad_norm": 0.31640625, | |
| "learning_rate": 0.00013091925889987865, | |
| "loss": 0.8894, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.4613960867265997, | |
| "grad_norm": 0.23046875, | |
| "learning_rate": 0.00013047995767951883, | |
| "loss": 0.8945, | |
| "step": 1745 | |
| }, | |
| { | |
| "epoch": 0.4627181385510312, | |
| "grad_norm": 0.306640625, | |
| "learning_rate": 0.00013004000703349891, | |
| "loss": 0.8918, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.46404019037546274, | |
| "grad_norm": 0.29296875, | |
| "learning_rate": 0.00012959941633569153, | |
| "loss": 0.8951, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 0.46536224219989425, | |
| "grad_norm": 0.1943359375, | |
| "learning_rate": 0.00012915819497360658, | |
| "loss": 0.8887, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.46668429402432576, | |
| "grad_norm": 0.251953125, | |
| "learning_rate": 0.00012871635234819136, | |
| "loss": 0.8897, | |
| "step": 1765 | |
| }, | |
| { | |
| "epoch": 0.4680063458487573, | |
| "grad_norm": 0.25390625, | |
| "learning_rate": 0.00012827389787363006, | |
| "loss": 0.892, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.4693283976731888, | |
| "grad_norm": 0.2119140625, | |
| "learning_rate": 0.00012783084097714347, | |
| "loss": 0.8879, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 0.4706504494976203, | |
| "grad_norm": 0.240234375, | |
| "learning_rate": 0.00012738719109878794, | |
| "loss": 0.891, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.4719725013220518, | |
| "grad_norm": 0.265625, | |
| "learning_rate": 0.00012694295769125424, | |
| "loss": 0.8906, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 0.47329455314648333, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 0.0001264981502196662, | |
| "loss": 0.8911, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.47461660497091485, | |
| "grad_norm": 0.26171875, | |
| "learning_rate": 0.000126052778161379, | |
| "loss": 0.895, | |
| "step": 1795 | |
| }, | |
| { | |
| "epoch": 0.47593865679534636, | |
| "grad_norm": 0.1806640625, | |
| "learning_rate": 0.00012560685100577738, | |
| "loss": 0.8892, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.4772607086197779, | |
| "grad_norm": 0.275390625, | |
| "learning_rate": 0.0001251603782540732, | |
| "loss": 0.8869, | |
| "step": 1805 | |
| }, | |
| { | |
| "epoch": 0.4785827604442094, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 0.0001247133694191032, | |
| "loss": 0.8881, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.47990481226864096, | |
| "grad_norm": 0.2216796875, | |
| "learning_rate": 0.0001242658340251263, | |
| "loss": 0.8821, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 0.48122686409307247, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 0.00012381778160762058, | |
| "loss": 0.8936, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.482548915917504, | |
| "grad_norm": 0.208984375, | |
| "learning_rate": 0.00012336922171308018, | |
| "loss": 0.8916, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 0.4838709677419355, | |
| "grad_norm": 0.23828125, | |
| "learning_rate": 0.00012292016389881184, | |
| "loss": 0.8879, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.485193019566367, | |
| "grad_norm": 0.240234375, | |
| "learning_rate": 0.00012247061773273135, | |
| "loss": 0.8883, | |
| "step": 1835 | |
| }, | |
| { | |
| "epoch": 0.4865150713907985, | |
| "grad_norm": 0.3046875, | |
| "learning_rate": 0.00012202059279315954, | |
| "loss": 0.8888, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.48783712321523004, | |
| "grad_norm": 0.365234375, | |
| "learning_rate": 0.00012157009866861842, | |
| "loss": 0.8884, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 0.48915917503966155, | |
| "grad_norm": 0.2314453125, | |
| "learning_rate": 0.00012111914495762669, | |
| "loss": 0.8899, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.49048122686409307, | |
| "grad_norm": 0.427734375, | |
| "learning_rate": 0.00012066774126849529, | |
| "loss": 0.8879, | |
| "step": 1855 | |
| }, | |
| { | |
| "epoch": 0.4918032786885246, | |
| "grad_norm": 0.263671875, | |
| "learning_rate": 0.00012021589721912277, | |
| "loss": 0.8925, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.4931253305129561, | |
| "grad_norm": 0.275390625, | |
| "learning_rate": 0.00011976362243679014, | |
| "loss": 0.8885, | |
| "step": 1865 | |
| }, | |
| { | |
| "epoch": 0.4944473823373876, | |
| "grad_norm": 0.22265625, | |
| "learning_rate": 0.00011931092655795597, | |
| "loss": 0.8869, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.4957694341618191, | |
| "grad_norm": 0.44140625, | |
| "learning_rate": 0.00011885781922805101, | |
| "loss": 0.8892, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 0.49709148598625064, | |
| "grad_norm": 0.228515625, | |
| "learning_rate": 0.0001184043101012726, | |
| "loss": 0.8913, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.4984135378106822, | |
| "grad_norm": 0.232421875, | |
| "learning_rate": 0.000117950408840379, | |
| "loss": 0.8877, | |
| "step": 1885 | |
| }, | |
| { | |
| "epoch": 0.4997355896351137, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 0.00011749612511648362, | |
| "loss": 0.8914, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.5010576414595452, | |
| "grad_norm": 0.30078125, | |
| "learning_rate": 0.0001170414686088488, | |
| "loss": 0.8906, | |
| "step": 1895 | |
| }, | |
| { | |
| "epoch": 0.5023796932839767, | |
| "grad_norm": 0.2451171875, | |
| "learning_rate": 0.00011658644900467965, | |
| "loss": 0.8923, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.5037017451084083, | |
| "grad_norm": 0.314453125, | |
| "learning_rate": 0.00011613107599891775, | |
| "loss": 0.8924, | |
| "step": 1905 | |
| }, | |
| { | |
| "epoch": 0.5050237969328397, | |
| "grad_norm": 0.259765625, | |
| "learning_rate": 0.00011567535929403435, | |
| "loss": 0.8869, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.5063458487572713, | |
| "grad_norm": 0.240234375, | |
| "learning_rate": 0.00011521930859982382, | |
| "loss": 0.8896, | |
| "step": 1915 | |
| }, | |
| { | |
| "epoch": 0.5076679005817029, | |
| "grad_norm": 0.2109375, | |
| "learning_rate": 0.00011476293363319675, | |
| "loss": 0.8869, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.5089899524061343, | |
| "grad_norm": 0.2197265625, | |
| "learning_rate": 0.0001143062441179729, | |
| "loss": 0.8869, | |
| "step": 1925 | |
| }, | |
| { | |
| "epoch": 0.5103120042305659, | |
| "grad_norm": 0.22265625, | |
| "learning_rate": 0.00011384924978467397, | |
| "loss": 0.8854, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.5116340560549973, | |
| "grad_norm": 0.24609375, | |
| "learning_rate": 0.00011339196037031627, | |
| "loss": 0.8905, | |
| "step": 1935 | |
| }, | |
| { | |
| "epoch": 0.5129561078794289, | |
| "grad_norm": 0.23828125, | |
| "learning_rate": 0.00011293438561820343, | |
| "loss": 0.8885, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.5142781597038604, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 0.00011247653527771847, | |
| "loss": 0.8869, | |
| "step": 1945 | |
| }, | |
| { | |
| "epoch": 0.5156002115282919, | |
| "grad_norm": 0.2138671875, | |
| "learning_rate": 0.00011201841910411652, | |
| "loss": 0.8873, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.5169222633527234, | |
| "grad_norm": 0.2177734375, | |
| "learning_rate": 0.00011156004685831648, | |
| "loss": 0.893, | |
| "step": 1955 | |
| }, | |
| { | |
| "epoch": 0.518244315177155, | |
| "grad_norm": 0.298828125, | |
| "learning_rate": 0.0001111014283066935, | |
| "loss": 0.8841, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.5195663670015864, | |
| "grad_norm": 0.291015625, | |
| "learning_rate": 0.0001106425732208705, | |
| "loss": 0.8796, | |
| "step": 1965 | |
| }, | |
| { | |
| "epoch": 0.520888418826018, | |
| "grad_norm": 0.2890625, | |
| "learning_rate": 0.0001101834913775103, | |
| "loss": 0.8917, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.5222104706504495, | |
| "grad_norm": 0.35546875, | |
| "learning_rate": 0.00010972419255810704, | |
| "loss": 0.8918, | |
| "step": 1975 | |
| }, | |
| { | |
| "epoch": 0.523532522474881, | |
| "grad_norm": 0.259765625, | |
| "learning_rate": 0.000109264686548778, | |
| "loss": 0.8866, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.5248545742993125, | |
| "grad_norm": 0.203125, | |
| "learning_rate": 0.00010880498314005488, | |
| "loss": 0.8911, | |
| "step": 1985 | |
| }, | |
| { | |
| "epoch": 0.526176626123744, | |
| "grad_norm": 0.265625, | |
| "learning_rate": 0.00010834509212667537, | |
| "loss": 0.8856, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.5274986779481756, | |
| "grad_norm": 0.2890625, | |
| "learning_rate": 0.00010788502330737438, | |
| "loss": 0.8899, | |
| "step": 1995 | |
| }, | |
| { | |
| "epoch": 0.5288207297726071, | |
| "grad_norm": 0.248046875, | |
| "learning_rate": 0.00010742478648467522, | |
| "loss": 0.8883, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.5301427815970386, | |
| "grad_norm": 0.2138671875, | |
| "learning_rate": 0.00010696439146468085, | |
| "loss": 0.8899, | |
| "step": 2005 | |
| }, | |
| { | |
| "epoch": 0.5314648334214701, | |
| "grad_norm": 0.21875, | |
| "learning_rate": 0.00010650384805686482, | |
| "loss": 0.8894, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.5327868852459017, | |
| "grad_norm": 0.220703125, | |
| "learning_rate": 0.0001060431660738624, | |
| "loss": 0.8882, | |
| "step": 2015 | |
| }, | |
| { | |
| "epoch": 0.5341089370703331, | |
| "grad_norm": 0.236328125, | |
| "learning_rate": 0.00010558235533126142, | |
| "loss": 0.8855, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.5354309888947647, | |
| "grad_norm": 0.236328125, | |
| "learning_rate": 0.00010512142564739302, | |
| "loss": 0.8841, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 0.5367530407191962, | |
| "grad_norm": 0.1826171875, | |
| "learning_rate": 0.00010466038684312278, | |
| "loss": 0.8912, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.5380750925436277, | |
| "grad_norm": 0.2314453125, | |
| "learning_rate": 0.0001041992487416411, | |
| "loss": 0.8883, | |
| "step": 2035 | |
| }, | |
| { | |
| "epoch": 0.5393971443680592, | |
| "grad_norm": 0.24609375, | |
| "learning_rate": 0.00010373802116825418, | |
| "loss": 0.8863, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.5407191961924908, | |
| "grad_norm": 0.275390625, | |
| "learning_rate": 0.00010327671395017451, | |
| "loss": 0.8831, | |
| "step": 2045 | |
| }, | |
| { | |
| "epoch": 0.5420412480169222, | |
| "grad_norm": 0.283203125, | |
| "learning_rate": 0.00010281533691631155, | |
| "loss": 0.8831, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.5433632998413538, | |
| "grad_norm": 0.2333984375, | |
| "learning_rate": 0.00010235389989706232, | |
| "loss": 0.8888, | |
| "step": 2055 | |
| }, | |
| { | |
| "epoch": 0.5446853516657852, | |
| "grad_norm": 0.2431640625, | |
| "learning_rate": 0.0001018924127241019, | |
| "loss": 0.893, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.5460074034902168, | |
| "grad_norm": 0.2109375, | |
| "learning_rate": 0.000101430885230174, | |
| "loss": 0.8872, | |
| "step": 2065 | |
| }, | |
| { | |
| "epoch": 0.5473294553146484, | |
| "grad_norm": 0.294921875, | |
| "learning_rate": 0.00010096932724888146, | |
| "loss": 0.888, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.5486515071390798, | |
| "grad_norm": 0.2119140625, | |
| "learning_rate": 0.00010050774861447662, | |
| "loss": 0.8897, | |
| "step": 2075 | |
| }, | |
| { | |
| "epoch": 0.5499735589635114, | |
| "grad_norm": 0.1806640625, | |
| "learning_rate": 0.00010004615916165191, | |
| "loss": 0.8865, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.5512956107879429, | |
| "grad_norm": 0.21875, | |
| "learning_rate": 9.958456872533031e-05, | |
| "loss": 0.8856, | |
| "step": 2085 | |
| }, | |
| { | |
| "epoch": 0.5526176626123744, | |
| "grad_norm": 0.181640625, | |
| "learning_rate": 9.91229871404557e-05, | |
| "loss": 0.8828, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.5539397144368059, | |
| "grad_norm": 0.158203125, | |
| "learning_rate": 9.866142424178341e-05, | |
| "loss": 0.8896, | |
| "step": 2095 | |
| }, | |
| { | |
| "epoch": 0.5552617662612375, | |
| "grad_norm": 0.2060546875, | |
| "learning_rate": 9.819988986367053e-05, | |
| "loss": 0.8845, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.5565838180856689, | |
| "grad_norm": 0.208984375, | |
| "learning_rate": 9.773839383986664e-05, | |
| "loss": 0.8877, | |
| "step": 2105 | |
| }, | |
| { | |
| "epoch": 0.5579058699101005, | |
| "grad_norm": 0.2119140625, | |
| "learning_rate": 9.727694600330395e-05, | |
| "loss": 0.8835, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.559227921734532, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 9.681555618588808e-05, | |
| "loss": 0.8855, | |
| "step": 2115 | |
| }, | |
| { | |
| "epoch": 0.5605499735589635, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 9.635423421828835e-05, | |
| "loss": 0.8838, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.561872025383395, | |
| "grad_norm": 0.23046875, | |
| "learning_rate": 9.589298992972857e-05, | |
| "loss": 0.8901, | |
| "step": 2125 | |
| }, | |
| { | |
| "epoch": 0.5631940772078265, | |
| "grad_norm": 0.306640625, | |
| "learning_rate": 9.543183314777732e-05, | |
| "loss": 0.888, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.5645161290322581, | |
| "grad_norm": 0.244140625, | |
| "learning_rate": 9.497077369813885e-05, | |
| "loss": 0.8865, | |
| "step": 2135 | |
| }, | |
| { | |
| "epoch": 0.5658381808566896, | |
| "grad_norm": 0.29296875, | |
| "learning_rate": 9.450982140444344e-05, | |
| "loss": 0.8916, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.5671602326811211, | |
| "grad_norm": 0.21875, | |
| "learning_rate": 9.404898608803836e-05, | |
| "loss": 0.8869, | |
| "step": 2145 | |
| }, | |
| { | |
| "epoch": 0.5684822845055526, | |
| "grad_norm": 0.2412109375, | |
| "learning_rate": 9.358827756777837e-05, | |
| "loss": 0.8844, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.5698043363299842, | |
| "grad_norm": 0.23046875, | |
| "learning_rate": 9.312770565981677e-05, | |
| "loss": 0.8861, | |
| "step": 2155 | |
| }, | |
| { | |
| "epoch": 0.5711263881544156, | |
| "grad_norm": 0.2412109375, | |
| "learning_rate": 9.266728017739601e-05, | |
| "loss": 0.888, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.5724484399788472, | |
| "grad_norm": 0.1953125, | |
| "learning_rate": 9.220701093063868e-05, | |
| "loss": 0.8857, | |
| "step": 2165 | |
| }, | |
| { | |
| "epoch": 0.5737704918032787, | |
| "grad_norm": 0.2421875, | |
| "learning_rate": 9.174690772633865e-05, | |
| "loss": 0.8896, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 0.5750925436277102, | |
| "grad_norm": 0.326171875, | |
| "learning_rate": 9.128698036775179e-05, | |
| "loss": 0.884, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 0.5764145954521417, | |
| "grad_norm": 0.208984375, | |
| "learning_rate": 9.082723865438747e-05, | |
| "loss": 0.8903, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.5777366472765733, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 9.036769238179947e-05, | |
| "loss": 0.8927, | |
| "step": 2185 | |
| }, | |
| { | |
| "epoch": 0.5790586991010047, | |
| "grad_norm": 0.21484375, | |
| "learning_rate": 8.990835134137744e-05, | |
| "loss": 0.8848, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.5803807509254363, | |
| "grad_norm": 0.166015625, | |
| "learning_rate": 8.944922532013811e-05, | |
| "loss": 0.884, | |
| "step": 2195 | |
| }, | |
| { | |
| "epoch": 0.5817028027498677, | |
| "grad_norm": 0.2578125, | |
| "learning_rate": 8.899032410051708e-05, | |
| "loss": 0.8832, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.5830248545742993, | |
| "grad_norm": 0.2119140625, | |
| "learning_rate": 8.853165746015997e-05, | |
| "loss": 0.8851, | |
| "step": 2205 | |
| }, | |
| { | |
| "epoch": 0.5843469063987309, | |
| "grad_norm": 0.21484375, | |
| "learning_rate": 8.807323517171444e-05, | |
| "loss": 0.8899, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 0.5856689582231623, | |
| "grad_norm": 0.181640625, | |
| "learning_rate": 8.761506700262172e-05, | |
| "loss": 0.8875, | |
| "step": 2215 | |
| }, | |
| { | |
| "epoch": 0.5869910100475939, | |
| "grad_norm": 0.19921875, | |
| "learning_rate": 8.715716271490877e-05, | |
| "loss": 0.8899, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.5883130618720254, | |
| "grad_norm": 0.185546875, | |
| "learning_rate": 8.669953206497994e-05, | |
| "loss": 0.8819, | |
| "step": 2225 | |
| }, | |
| { | |
| "epoch": 0.5896351136964569, | |
| "grad_norm": 0.25390625, | |
| "learning_rate": 8.624218480340942e-05, | |
| "loss": 0.8879, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 0.5909571655208884, | |
| "grad_norm": 0.296875, | |
| "learning_rate": 8.578513067473322e-05, | |
| "loss": 0.8877, | |
| "step": 2235 | |
| }, | |
| { | |
| "epoch": 0.59227921734532, | |
| "grad_norm": 0.1875, | |
| "learning_rate": 8.532837941724181e-05, | |
| "loss": 0.8849, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.5936012691697514, | |
| "grad_norm": 0.2109375, | |
| "learning_rate": 8.487194076277234e-05, | |
| "loss": 0.8826, | |
| "step": 2245 | |
| }, | |
| { | |
| "epoch": 0.594923320994183, | |
| "grad_norm": 0.259765625, | |
| "learning_rate": 8.441582443650161e-05, | |
| "loss": 0.8847, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.5962453728186144, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 8.396004015673853e-05, | |
| "loss": 0.8825, | |
| "step": 2255 | |
| }, | |
| { | |
| "epoch": 0.597567424643046, | |
| "grad_norm": 0.2158203125, | |
| "learning_rate": 8.35045976347173e-05, | |
| "loss": 0.888, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.5988894764674775, | |
| "grad_norm": 0.1865234375, | |
| "learning_rate": 8.304950657439033e-05, | |
| "loss": 0.885, | |
| "step": 2265 | |
| }, | |
| { | |
| "epoch": 0.600211528291909, | |
| "grad_norm": 0.2353515625, | |
| "learning_rate": 8.259477667222172e-05, | |
| "loss": 0.8857, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 0.6015335801163406, | |
| "grad_norm": 0.296875, | |
| "learning_rate": 8.214041761698031e-05, | |
| "loss": 0.8876, | |
| "step": 2275 | |
| }, | |
| { | |
| "epoch": 0.6028556319407721, | |
| "grad_norm": 0.232421875, | |
| "learning_rate": 8.168643908953359e-05, | |
| "loss": 0.8846, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.6041776837652036, | |
| "grad_norm": 0.255859375, | |
| "learning_rate": 8.12328507626411e-05, | |
| "loss": 0.8863, | |
| "step": 2285 | |
| }, | |
| { | |
| "epoch": 0.6054997355896351, | |
| "grad_norm": 0.3359375, | |
| "learning_rate": 8.077966230074872e-05, | |
| "loss": 0.8889, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 0.6068217874140667, | |
| "grad_norm": 0.2197265625, | |
| "learning_rate": 8.032688335978247e-05, | |
| "loss": 0.8844, | |
| "step": 2295 | |
| }, | |
| { | |
| "epoch": 0.6081438392384981, | |
| "grad_norm": 0.2470703125, | |
| "learning_rate": 7.987452358694278e-05, | |
| "loss": 0.8894, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.6094658910629297, | |
| "grad_norm": 0.2060546875, | |
| "learning_rate": 7.942259262049911e-05, | |
| "loss": 0.8825, | |
| "step": 2305 | |
| }, | |
| { | |
| "epoch": 0.6107879428873612, | |
| "grad_norm": 0.236328125, | |
| "learning_rate": 7.897110008958441e-05, | |
| "loss": 0.8851, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.6121099947117927, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 7.852005561399014e-05, | |
| "loss": 0.8903, | |
| "step": 2315 | |
| }, | |
| { | |
| "epoch": 0.6134320465362242, | |
| "grad_norm": 0.27734375, | |
| "learning_rate": 7.806946880396109e-05, | |
| "loss": 0.8873, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.6147540983606558, | |
| "grad_norm": 0.27734375, | |
| "learning_rate": 7.761934925999086e-05, | |
| "loss": 0.8862, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 0.6160761501850872, | |
| "grad_norm": 0.23046875, | |
| "learning_rate": 7.716970657261702e-05, | |
| "loss": 0.8846, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 0.6173982020095188, | |
| "grad_norm": 0.2255859375, | |
| "learning_rate": 7.672055032221709e-05, | |
| "loss": 0.8897, | |
| "step": 2335 | |
| }, | |
| { | |
| "epoch": 0.6187202538339502, | |
| "grad_norm": 0.2197265625, | |
| "learning_rate": 7.627189007880413e-05, | |
| "loss": 0.8873, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.6200423056583818, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 7.582373540182304e-05, | |
| "loss": 0.8904, | |
| "step": 2345 | |
| }, | |
| { | |
| "epoch": 0.6213643574828134, | |
| "grad_norm": 0.208984375, | |
| "learning_rate": 7.537609583994667e-05, | |
| "loss": 0.8834, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.6226864093072448, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 7.492898093087269e-05, | |
| "loss": 0.8825, | |
| "step": 2355 | |
| }, | |
| { | |
| "epoch": 0.6240084611316764, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 7.448240020111993e-05, | |
| "loss": 0.8837, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.6253305129561079, | |
| "grad_norm": 0.1669921875, | |
| "learning_rate": 7.403636316582594e-05, | |
| "loss": 0.8885, | |
| "step": 2365 | |
| }, | |
| { | |
| "epoch": 0.6266525647805394, | |
| "grad_norm": 0.2158203125, | |
| "learning_rate": 7.359087932854371e-05, | |
| "loss": 0.8944, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.6279746166049709, | |
| "grad_norm": 0.369140625, | |
| "learning_rate": 7.314595818103959e-05, | |
| "loss": 0.889, | |
| "step": 2375 | |
| }, | |
| { | |
| "epoch": 0.6292966684294025, | |
| "grad_norm": 0.232421875, | |
| "learning_rate": 7.270160920309077e-05, | |
| "loss": 0.8869, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.6306187202538339, | |
| "grad_norm": 0.208984375, | |
| "learning_rate": 7.225784186228359e-05, | |
| "loss": 0.8829, | |
| "step": 2385 | |
| }, | |
| { | |
| "epoch": 0.6319407720782655, | |
| "grad_norm": 0.25390625, | |
| "learning_rate": 7.181466561381152e-05, | |
| "loss": 0.8889, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 0.633262823902697, | |
| "grad_norm": 0.287109375, | |
| "learning_rate": 7.13720899002739e-05, | |
| "loss": 0.8823, | |
| "step": 2395 | |
| }, | |
| { | |
| "epoch": 0.6345848757271285, | |
| "grad_norm": 0.21484375, | |
| "learning_rate": 7.093012415147462e-05, | |
| "loss": 0.884, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.63590692755156, | |
| "grad_norm": 0.1591796875, | |
| "learning_rate": 7.048877778422138e-05, | |
| "loss": 0.8848, | |
| "step": 2405 | |
| }, | |
| { | |
| "epoch": 0.6372289793759915, | |
| "grad_norm": 0.259765625, | |
| "learning_rate": 7.00480602021248e-05, | |
| "loss": 0.8821, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 0.638551031200423, | |
| "grad_norm": 0.291015625, | |
| "learning_rate": 6.960798079539835e-05, | |
| "loss": 0.8879, | |
| "step": 2415 | |
| }, | |
| { | |
| "epoch": 0.6398730830248546, | |
| "grad_norm": 0.2080078125, | |
| "learning_rate": 6.916854894065796e-05, | |
| "loss": 0.886, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.6411951348492861, | |
| "grad_norm": 0.251953125, | |
| "learning_rate": 6.87297740007225e-05, | |
| "loss": 0.8895, | |
| "step": 2425 | |
| }, | |
| { | |
| "epoch": 0.6425171866737176, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 6.829166532441421e-05, | |
| "loss": 0.8839, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.6438392384981492, | |
| "grad_norm": 0.203125, | |
| "learning_rate": 6.785423224635941e-05, | |
| "loss": 0.8834, | |
| "step": 2435 | |
| }, | |
| { | |
| "epoch": 0.6451612903225806, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 6.741748408678975e-05, | |
| "loss": 0.8849, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.6464833421470122, | |
| "grad_norm": 0.1787109375, | |
| "learning_rate": 6.698143015134349e-05, | |
| "loss": 0.8804, | |
| "step": 2445 | |
| }, | |
| { | |
| "epoch": 0.6478053939714437, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 6.65460797308674e-05, | |
| "loss": 0.8845, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.6491274457958752, | |
| "grad_norm": 0.1611328125, | |
| "learning_rate": 6.611144210121861e-05, | |
| "loss": 0.8897, | |
| "step": 2455 | |
| }, | |
| { | |
| "epoch": 0.6504494976203067, | |
| "grad_norm": 0.1630859375, | |
| "learning_rate": 6.567752652306717e-05, | |
| "loss": 0.8913, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.6517715494447383, | |
| "grad_norm": 0.2470703125, | |
| "learning_rate": 6.52443422416985e-05, | |
| "loss": 0.8884, | |
| "step": 2465 | |
| }, | |
| { | |
| "epoch": 0.6530936012691697, | |
| "grad_norm": 0.17578125, | |
| "learning_rate": 6.48118984868167e-05, | |
| "loss": 0.8894, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 0.6544156530936013, | |
| "grad_norm": 0.2265625, | |
| "learning_rate": 6.438020447234759e-05, | |
| "loss": 0.8837, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 0.6557377049180327, | |
| "grad_norm": 0.1748046875, | |
| "learning_rate": 6.394926939624267e-05, | |
| "loss": 0.8864, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.6570597567424643, | |
| "grad_norm": 0.1796875, | |
| "learning_rate": 6.351910244028285e-05, | |
| "loss": 0.8859, | |
| "step": 2485 | |
| }, | |
| { | |
| "epoch": 0.6583818085668959, | |
| "grad_norm": 0.2314453125, | |
| "learning_rate": 6.308971276988313e-05, | |
| "loss": 0.885, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.6597038603913273, | |
| "grad_norm": 0.263671875, | |
| "learning_rate": 6.266110953389704e-05, | |
| "loss": 0.8835, | |
| "step": 2495 | |
| }, | |
| { | |
| "epoch": 0.6610259122157589, | |
| "grad_norm": 0.1806640625, | |
| "learning_rate": 6.223330186442194e-05, | |
| "loss": 0.8834, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.6623479640401904, | |
| "grad_norm": 0.201171875, | |
| "learning_rate": 6.180629887660425e-05, | |
| "loss": 0.8839, | |
| "step": 2505 | |
| }, | |
| { | |
| "epoch": 0.6636700158646219, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 6.138010966844538e-05, | |
| "loss": 0.8868, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 0.6649920676890534, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 6.095474332060776e-05, | |
| "loss": 0.8857, | |
| "step": 2515 | |
| }, | |
| { | |
| "epoch": 0.666314119513485, | |
| "grad_norm": 0.205078125, | |
| "learning_rate": 6.0530208896221527e-05, | |
| "loss": 0.8823, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.6676361713379164, | |
| "grad_norm": 0.28125, | |
| "learning_rate": 6.010651544069119e-05, | |
| "loss": 0.8835, | |
| "step": 2525 | |
| }, | |
| { | |
| "epoch": 0.668958223162348, | |
| "grad_norm": 0.259765625, | |
| "learning_rate": 5.968367198150316e-05, | |
| "loss": 0.8826, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 0.6702802749867794, | |
| "grad_norm": 0.1650390625, | |
| "learning_rate": 5.92616875280332e-05, | |
| "loss": 0.8825, | |
| "step": 2535 | |
| }, | |
| { | |
| "epoch": 0.671602326811211, | |
| "grad_norm": 0.2294921875, | |
| "learning_rate": 5.8840571071354565e-05, | |
| "loss": 0.8874, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.6729243786356425, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 5.842033158404636e-05, | |
| "loss": 0.8914, | |
| "step": 2545 | |
| }, | |
| { | |
| "epoch": 0.674246430460074, | |
| "grad_norm": 0.181640625, | |
| "learning_rate": 5.800097802000257e-05, | |
| "loss": 0.8844, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.6755684822845055, | |
| "grad_norm": 0.240234375, | |
| "learning_rate": 5.758251931424089e-05, | |
| "loss": 0.8809, | |
| "step": 2555 | |
| }, | |
| { | |
| "epoch": 0.6768905341089371, | |
| "grad_norm": 0.2421875, | |
| "learning_rate": 5.716496438271277e-05, | |
| "loss": 0.8828, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.6782125859333686, | |
| "grad_norm": 0.185546875, | |
| "learning_rate": 5.6748322122113264e-05, | |
| "loss": 0.8832, | |
| "step": 2565 | |
| }, | |
| { | |
| "epoch": 0.6795346377578001, | |
| "grad_norm": 0.23046875, | |
| "learning_rate": 5.633260140969132e-05, | |
| "loss": 0.885, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 0.6808566895822317, | |
| "grad_norm": 0.162109375, | |
| "learning_rate": 5.591781110306096e-05, | |
| "loss": 0.8837, | |
| "step": 2575 | |
| }, | |
| { | |
| "epoch": 0.6821787414066631, | |
| "grad_norm": 0.173828125, | |
| "learning_rate": 5.55039600400122e-05, | |
| "loss": 0.8817, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.6835007932310947, | |
| "grad_norm": 0.20703125, | |
| "learning_rate": 5.509105703832313e-05, | |
| "loss": 0.8858, | |
| "step": 2585 | |
| }, | |
| { | |
| "epoch": 0.6848228450555262, | |
| "grad_norm": 0.2451171875, | |
| "learning_rate": 5.467911089557169e-05, | |
| "loss": 0.8863, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 0.6861448968799577, | |
| "grad_norm": 0.208984375, | |
| "learning_rate": 5.426813038894839e-05, | |
| "loss": 0.8844, | |
| "step": 2595 | |
| }, | |
| { | |
| "epoch": 0.6874669487043892, | |
| "grad_norm": 0.203125, | |
| "learning_rate": 5.3858124275069244e-05, | |
| "loss": 0.8836, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.6887890005288207, | |
| "grad_norm": 0.25390625, | |
| "learning_rate": 5.344910128978934e-05, | |
| "loss": 0.8837, | |
| "step": 2605 | |
| }, | |
| { | |
| "epoch": 0.6901110523532522, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 5.304107014801645e-05, | |
| "loss": 0.8845, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.6914331041776838, | |
| "grad_norm": 0.21875, | |
| "learning_rate": 5.263403954352564e-05, | |
| "loss": 0.8878, | |
| "step": 2615 | |
| }, | |
| { | |
| "epoch": 0.6927551560021152, | |
| "grad_norm": 0.1767578125, | |
| "learning_rate": 5.222801814877369e-05, | |
| "loss": 0.8879, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.6940772078265468, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 5.182301461471474e-05, | |
| "loss": 0.8905, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 0.6953992596509783, | |
| "grad_norm": 0.201171875, | |
| "learning_rate": 5.141903757061555e-05, | |
| "loss": 0.8889, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 0.6967213114754098, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 5.1016095623871865e-05, | |
| "loss": 0.8805, | |
| "step": 2635 | |
| }, | |
| { | |
| "epoch": 0.6980433632998414, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 5.0614197359824955e-05, | |
| "loss": 0.8848, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.6993654151242729, | |
| "grad_norm": 0.2294921875, | |
| "learning_rate": 5.021335134157885e-05, | |
| "loss": 0.883, | |
| "step": 2645 | |
| }, | |
| { | |
| "epoch": 0.7006874669487044, | |
| "grad_norm": 0.216796875, | |
| "learning_rate": 4.981356610981756e-05, | |
| "loss": 0.8817, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.7020095187731359, | |
| "grad_norm": 0.1591796875, | |
| "learning_rate": 4.9414850182623465e-05, | |
| "loss": 0.8831, | |
| "step": 2655 | |
| }, | |
| { | |
| "epoch": 0.7033315705975675, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 4.901721205529548e-05, | |
| "loss": 0.8845, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.7046536224219989, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 4.8620660200168425e-05, | |
| "loss": 0.8821, | |
| "step": 2665 | |
| }, | |
| { | |
| "epoch": 0.7059756742464305, | |
| "grad_norm": 0.19140625, | |
| "learning_rate": 4.822520306643207e-05, | |
| "loss": 0.8835, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.7072977260708619, | |
| "grad_norm": 0.255859375, | |
| "learning_rate": 4.783084907995156e-05, | |
| "loss": 0.8831, | |
| "step": 2675 | |
| }, | |
| { | |
| "epoch": 0.7086197778952935, | |
| "grad_norm": 0.158203125, | |
| "learning_rate": 4.743760664308753e-05, | |
| "loss": 0.8841, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.709941829719725, | |
| "grad_norm": 0.298828125, | |
| "learning_rate": 4.704548413451728e-05, | |
| "loss": 0.8843, | |
| "step": 2685 | |
| }, | |
| { | |
| "epoch": 0.7112638815441565, | |
| "grad_norm": 0.220703125, | |
| "learning_rate": 4.665448990905613e-05, | |
| "loss": 0.8843, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 0.712585933368588, | |
| "grad_norm": 0.265625, | |
| "learning_rate": 4.626463229747958e-05, | |
| "loss": 0.8832, | |
| "step": 2695 | |
| }, | |
| { | |
| "epoch": 0.7139079851930196, | |
| "grad_norm": 0.181640625, | |
| "learning_rate": 4.587591960634565e-05, | |
| "loss": 0.8852, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.7152300370174511, | |
| "grad_norm": 0.2421875, | |
| "learning_rate": 4.54883601178179e-05, | |
| "loss": 0.8844, | |
| "step": 2705 | |
| }, | |
| { | |
| "epoch": 0.7165520888418826, | |
| "grad_norm": 0.1943359375, | |
| "learning_rate": 4.510196208948909e-05, | |
| "loss": 0.8873, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 0.7178741406663142, | |
| "grad_norm": 0.17578125, | |
| "learning_rate": 4.4716733754205064e-05, | |
| "loss": 0.8872, | |
| "step": 2715 | |
| }, | |
| { | |
| "epoch": 0.7191961924907456, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 4.433268331988956e-05, | |
| "loss": 0.8891, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 0.7205182443151772, | |
| "grad_norm": 0.2392578125, | |
| "learning_rate": 4.394981896936908e-05, | |
| "loss": 0.8861, | |
| "step": 2725 | |
| }, | |
| { | |
| "epoch": 0.7218402961396087, | |
| "grad_norm": 0.17578125, | |
| "learning_rate": 4.356814886019871e-05, | |
| "loss": 0.8884, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 0.7231623479640402, | |
| "grad_norm": 0.162109375, | |
| "learning_rate": 4.3187681124488196e-05, | |
| "loss": 0.8865, | |
| "step": 2735 | |
| }, | |
| { | |
| "epoch": 0.7244843997884717, | |
| "grad_norm": 0.19921875, | |
| "learning_rate": 4.280842386872889e-05, | |
| "loss": 0.8817, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 0.7258064516129032, | |
| "grad_norm": 0.1865234375, | |
| "learning_rate": 4.243038517362068e-05, | |
| "loss": 0.8832, | |
| "step": 2745 | |
| }, | |
| { | |
| "epoch": 0.7271285034373347, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 4.205357309390021e-05, | |
| "loss": 0.8856, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.7284505552617663, | |
| "grad_norm": 0.14453125, | |
| "learning_rate": 4.167799565816889e-05, | |
| "loss": 0.8862, | |
| "step": 2755 | |
| }, | |
| { | |
| "epoch": 0.7297726070861977, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 4.130366086872215e-05, | |
| "loss": 0.8826, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.7310946589106293, | |
| "grad_norm": 0.22265625, | |
| "learning_rate": 4.093057670137865e-05, | |
| "loss": 0.8871, | |
| "step": 2765 | |
| }, | |
| { | |
| "epoch": 0.7324167107350608, | |
| "grad_norm": 0.2099609375, | |
| "learning_rate": 4.055875110531068e-05, | |
| "loss": 0.8862, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 0.7337387625594923, | |
| "grad_norm": 0.20703125, | |
| "learning_rate": 4.0188192002874404e-05, | |
| "loss": 0.8849, | |
| "step": 2775 | |
| }, | |
| { | |
| "epoch": 0.7350608143839239, | |
| "grad_norm": 0.2431640625, | |
| "learning_rate": 3.9818907289441375e-05, | |
| "loss": 0.8831, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 0.7363828662083554, | |
| "grad_norm": 0.205078125, | |
| "learning_rate": 3.9450904833230094e-05, | |
| "loss": 0.8879, | |
| "step": 2785 | |
| }, | |
| { | |
| "epoch": 0.7377049180327869, | |
| "grad_norm": 0.162109375, | |
| "learning_rate": 3.908419247513862e-05, | |
| "loss": 0.884, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 0.7390269698572184, | |
| "grad_norm": 0.173828125, | |
| "learning_rate": 3.871877802857714e-05, | |
| "loss": 0.8931, | |
| "step": 2795 | |
| }, | |
| { | |
| "epoch": 0.74034902168165, | |
| "grad_norm": 0.2578125, | |
| "learning_rate": 3.8354669279301914e-05, | |
| "loss": 0.8881, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.7416710735060814, | |
| "grad_norm": 0.294921875, | |
| "learning_rate": 3.7991873985248996e-05, | |
| "loss": 0.8866, | |
| "step": 2805 | |
| }, | |
| { | |
| "epoch": 0.742993125330513, | |
| "grad_norm": 0.2294921875, | |
| "learning_rate": 3.763039987636924e-05, | |
| "loss": 0.8826, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 0.7443151771549444, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 3.7270254654463356e-05, | |
| "loss": 0.8841, | |
| "step": 2815 | |
| }, | |
| { | |
| "epoch": 0.745637228979376, | |
| "grad_norm": 0.2041015625, | |
| "learning_rate": 3.6911445993018076e-05, | |
| "loss": 0.8903, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.7469592808038075, | |
| "grad_norm": 0.2177734375, | |
| "learning_rate": 3.6553981537042406e-05, | |
| "loss": 0.8823, | |
| "step": 2825 | |
| }, | |
| { | |
| "epoch": 0.748281332628239, | |
| "grad_norm": 0.259765625, | |
| "learning_rate": 3.619786890290482e-05, | |
| "loss": 0.8846, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 0.7496033844526705, | |
| "grad_norm": 0.205078125, | |
| "learning_rate": 3.5843115678171125e-05, | |
| "loss": 0.881, | |
| "step": 2835 | |
| }, | |
| { | |
| "epoch": 0.7509254362771021, | |
| "grad_norm": 0.255859375, | |
| "learning_rate": 3.5489729421442516e-05, | |
| "loss": 0.8845, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 0.7522474881015336, | |
| "grad_norm": 0.2412109375, | |
| "learning_rate": 3.513771766219487e-05, | |
| "loss": 0.8857, | |
| "step": 2845 | |
| }, | |
| { | |
| "epoch": 0.7535695399259651, | |
| "grad_norm": 0.220703125, | |
| "learning_rate": 3.4787087900617955e-05, | |
| "loss": 0.8882, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.7548915917503967, | |
| "grad_norm": 0.2421875, | |
| "learning_rate": 3.443784760745593e-05, | |
| "loss": 0.8853, | |
| "step": 2855 | |
| }, | |
| { | |
| "epoch": 0.7562136435748281, | |
| "grad_norm": 0.216796875, | |
| "learning_rate": 3.409000422384797e-05, | |
| "loss": 0.8892, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 0.7575356953992597, | |
| "grad_norm": 0.1845703125, | |
| "learning_rate": 3.3743565161169835e-05, | |
| "loss": 0.8842, | |
| "step": 2865 | |
| }, | |
| { | |
| "epoch": 0.7588577472236911, | |
| "grad_norm": 0.1953125, | |
| "learning_rate": 3.339853780087584e-05, | |
| "loss": 0.8863, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 0.7601797990481227, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 3.3054929494341805e-05, | |
| "loss": 0.884, | |
| "step": 2875 | |
| }, | |
| { | |
| "epoch": 0.7615018508725542, | |
| "grad_norm": 0.2119140625, | |
| "learning_rate": 3.2712747562708115e-05, | |
| "loss": 0.886, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.7628239026969857, | |
| "grad_norm": 0.216796875, | |
| "learning_rate": 3.237199929672401e-05, | |
| "loss": 0.8833, | |
| "step": 2885 | |
| }, | |
| { | |
| "epoch": 0.7641459545214172, | |
| "grad_norm": 0.2431640625, | |
| "learning_rate": 3.203269195659202e-05, | |
| "loss": 0.8824, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 0.7654680063458488, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 3.169483277181351e-05, | |
| "loss": 0.8864, | |
| "step": 2895 | |
| }, | |
| { | |
| "epoch": 0.7667900581702802, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 3.135842894103436e-05, | |
| "loss": 0.8867, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.7681121099947118, | |
| "grad_norm": 0.1806640625, | |
| "learning_rate": 3.102348763189188e-05, | |
| "loss": 0.891, | |
| "step": 2905 | |
| }, | |
| { | |
| "epoch": 0.7694341618191433, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 3.069001598086184e-05, | |
| "loss": 0.8902, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 0.7707562136435748, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 3.0358021093106594e-05, | |
| "loss": 0.885, | |
| "step": 2915 | |
| }, | |
| { | |
| "epoch": 0.7720782654680064, | |
| "grad_norm": 0.23046875, | |
| "learning_rate": 3.0027510042323537e-05, | |
| "loss": 0.8866, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 0.7734003172924379, | |
| "grad_norm": 0.2275390625, | |
| "learning_rate": 2.9698489870594616e-05, | |
| "loss": 0.8831, | |
| "step": 2925 | |
| }, | |
| { | |
| "epoch": 0.7747223691168694, | |
| "grad_norm": 0.244140625, | |
| "learning_rate": 2.937096758823601e-05, | |
| "loss": 0.8869, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 0.7760444209413009, | |
| "grad_norm": 0.1748046875, | |
| "learning_rate": 2.9044950173649e-05, | |
| "loss": 0.8816, | |
| "step": 2935 | |
| }, | |
| { | |
| "epoch": 0.7773664727657325, | |
| "grad_norm": 0.16796875, | |
| "learning_rate": 2.872044457317109e-05, | |
| "loss": 0.8875, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.7786885245901639, | |
| "grad_norm": 0.171875, | |
| "learning_rate": 2.839745770092821e-05, | |
| "loss": 0.8828, | |
| "step": 2945 | |
| }, | |
| { | |
| "epoch": 0.7800105764145955, | |
| "grad_norm": 0.1884765625, | |
| "learning_rate": 2.8075996438687148e-05, | |
| "loss": 0.8818, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.7813326282390269, | |
| "grad_norm": 0.212890625, | |
| "learning_rate": 2.7756067635709204e-05, | |
| "loss": 0.8911, | |
| "step": 2955 | |
| }, | |
| { | |
| "epoch": 0.7826546800634585, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 2.7437678108604003e-05, | |
| "loss": 0.8859, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 0.78397673188789, | |
| "grad_norm": 0.1796875, | |
| "learning_rate": 2.712083464118441e-05, | |
| "loss": 0.8892, | |
| "step": 2965 | |
| }, | |
| { | |
| "epoch": 0.7852987837123215, | |
| "grad_norm": 0.205078125, | |
| "learning_rate": 2.68055439843219e-05, | |
| "loss": 0.8826, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 0.786620835536753, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 2.6491812855802834e-05, | |
| "loss": 0.8877, | |
| "step": 2975 | |
| }, | |
| { | |
| "epoch": 0.7879428873611846, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 2.6179647940185238e-05, | |
| "loss": 0.8867, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 0.789264939185616, | |
| "grad_norm": 0.232421875, | |
| "learning_rate": 2.5869055888656336e-05, | |
| "loss": 0.8858, | |
| "step": 2985 | |
| }, | |
| { | |
| "epoch": 0.7905869910100476, | |
| "grad_norm": 0.234375, | |
| "learning_rate": 2.556004331889098e-05, | |
| "loss": 0.8859, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 0.7919090428344792, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 2.525261681491047e-05, | |
| "loss": 0.8849, | |
| "step": 2995 | |
| }, | |
| { | |
| "epoch": 0.7932310946589106, | |
| "grad_norm": 0.248046875, | |
| "learning_rate": 2.494678292694247e-05, | |
| "loss": 0.8823, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.7945531464833422, | |
| "grad_norm": 0.162109375, | |
| "learning_rate": 2.4642548171281244e-05, | |
| "loss": 0.8886, | |
| "step": 3005 | |
| }, | |
| { | |
| "epoch": 0.7958751983077736, | |
| "grad_norm": 0.2041015625, | |
| "learning_rate": 2.433991903014896e-05, | |
| "loss": 0.8889, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 0.7971972501322052, | |
| "grad_norm": 0.2255859375, | |
| "learning_rate": 2.403890195155748e-05, | |
| "loss": 0.8858, | |
| "step": 3015 | |
| }, | |
| { | |
| "epoch": 0.7985193019566367, | |
| "grad_norm": 0.1728515625, | |
| "learning_rate": 2.373950334917111e-05, | |
| "loss": 0.8818, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 0.7998413537810682, | |
| "grad_norm": 0.263671875, | |
| "learning_rate": 2.3441729602169748e-05, | |
| "loss": 0.8859, | |
| "step": 3025 | |
| }, | |
| { | |
| "epoch": 0.8011634056054997, | |
| "grad_norm": 0.2197265625, | |
| "learning_rate": 2.3145587055113183e-05, | |
| "loss": 0.8876, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 0.8024854574299313, | |
| "grad_norm": 0.25, | |
| "learning_rate": 2.2851082017805703e-05, | |
| "loss": 0.8843, | |
| "step": 3035 | |
| }, | |
| { | |
| "epoch": 0.8038075092543627, | |
| "grad_norm": 0.2412109375, | |
| "learning_rate": 2.255822076516184e-05, | |
| "loss": 0.8846, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 0.8051295610787943, | |
| "grad_norm": 0.2392578125, | |
| "learning_rate": 2.2267009537072536e-05, | |
| "loss": 0.8877, | |
| "step": 3045 | |
| }, | |
| { | |
| "epoch": 0.8064516129032258, | |
| "grad_norm": 0.263671875, | |
| "learning_rate": 2.197745453827226e-05, | |
| "loss": 0.8836, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.8077736647276573, | |
| "grad_norm": 0.171875, | |
| "learning_rate": 2.168956193820676e-05, | |
| "loss": 0.885, | |
| "step": 3055 | |
| }, | |
| { | |
| "epoch": 0.8090957165520889, | |
| "grad_norm": 0.1669921875, | |
| "learning_rate": 2.140333787090172e-05, | |
| "loss": 0.8896, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.8104177683765204, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 2.1118788434831894e-05, | |
| "loss": 0.8819, | |
| "step": 3065 | |
| }, | |
| { | |
| "epoch": 0.8117398202009519, | |
| "grad_norm": 0.19921875, | |
| "learning_rate": 2.083591969279136e-05, | |
| "loss": 0.8838, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 0.8130618720253834, | |
| "grad_norm": 0.1884765625, | |
| "learning_rate": 2.0554737671764134e-05, | |
| "loss": 0.8825, | |
| "step": 3075 | |
| }, | |
| { | |
| "epoch": 0.814383923849815, | |
| "grad_norm": 0.181640625, | |
| "learning_rate": 2.0275248362795995e-05, | |
| "loss": 0.8857, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.8157059756742464, | |
| "grad_norm": 0.30078125, | |
| "learning_rate": 1.999745772086655e-05, | |
| "loss": 0.8852, | |
| "step": 3085 | |
| }, | |
| { | |
| "epoch": 0.817028027498678, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 1.9721371664762666e-05, | |
| "loss": 0.8864, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 0.8183500793231094, | |
| "grad_norm": 0.2451171875, | |
| "learning_rate": 1.9446996076952094e-05, | |
| "loss": 0.88, | |
| "step": 3095 | |
| }, | |
| { | |
| "epoch": 0.819672131147541, | |
| "grad_norm": 0.20703125, | |
| "learning_rate": 1.917433680345829e-05, | |
| "loss": 0.8872, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.8209941829719725, | |
| "grad_norm": 0.2119140625, | |
| "learning_rate": 1.8903399653735766e-05, | |
| "loss": 0.888, | |
| "step": 3105 | |
| }, | |
| { | |
| "epoch": 0.822316234796404, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 1.8634190400546415e-05, | |
| "loss": 0.8882, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 0.8236382866208355, | |
| "grad_norm": 0.1845703125, | |
| "learning_rate": 1.8366714779836434e-05, | |
| "loss": 0.8887, | |
| "step": 3115 | |
| }, | |
| { | |
| "epoch": 0.8249603384452671, | |
| "grad_norm": 0.1728515625, | |
| "learning_rate": 1.8100978490614085e-05, | |
| "loss": 0.8856, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.8262823902696985, | |
| "grad_norm": 0.1796875, | |
| "learning_rate": 1.783698719482836e-05, | |
| "loss": 0.886, | |
| "step": 3125 | |
| }, | |
| { | |
| "epoch": 0.8276044420941301, | |
| "grad_norm": 0.181640625, | |
| "learning_rate": 1.7574746517248254e-05, | |
| "loss": 0.8807, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 0.8289264939185617, | |
| "grad_norm": 0.2431640625, | |
| "learning_rate": 1.7314262045343e-05, | |
| "loss": 0.8861, | |
| "step": 3135 | |
| }, | |
| { | |
| "epoch": 0.8302485457429931, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 1.7055539329162963e-05, | |
| "loss": 0.8825, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 0.8315705975674247, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 1.6798583881221374e-05, | |
| "loss": 0.8803, | |
| "step": 3145 | |
| }, | |
| { | |
| "epoch": 0.8328926493918561, | |
| "grad_norm": 0.23046875, | |
| "learning_rate": 1.6543401176376916e-05, | |
| "loss": 0.8877, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.8342147012162877, | |
| "grad_norm": 0.2041015625, | |
| "learning_rate": 1.628999665171712e-05, | |
| "loss": 0.8832, | |
| "step": 3155 | |
| }, | |
| { | |
| "epoch": 0.8355367530407192, | |
| "grad_norm": 0.1796875, | |
| "learning_rate": 1.603837570644238e-05, | |
| "loss": 0.8854, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 0.8368588048651507, | |
| "grad_norm": 0.2578125, | |
| "learning_rate": 1.578854370175107e-05, | |
| "loss": 0.883, | |
| "step": 3165 | |
| }, | |
| { | |
| "epoch": 0.8381808566895822, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 1.554050596072516e-05, | |
| "loss": 0.8855, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 0.8395029085140138, | |
| "grad_norm": 0.1611328125, | |
| "learning_rate": 1.5294267768216975e-05, | |
| "loss": 0.8873, | |
| "step": 3175 | |
| }, | |
| { | |
| "epoch": 0.8408249603384452, | |
| "grad_norm": 0.212890625, | |
| "learning_rate": 1.5049834370736405e-05, | |
| "loss": 0.8869, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 0.8421470121628768, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 1.4807210976339291e-05, | |
| "loss": 0.8853, | |
| "step": 3185 | |
| }, | |
| { | |
| "epoch": 0.8434690639873083, | |
| "grad_norm": 0.212890625, | |
| "learning_rate": 1.4566402754516328e-05, | |
| "loss": 0.8862, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 0.8447911158117398, | |
| "grad_norm": 0.2255859375, | |
| "learning_rate": 1.4327414836082976e-05, | |
| "loss": 0.884, | |
| "step": 3195 | |
| }, | |
| { | |
| "epoch": 0.8461131676361714, | |
| "grad_norm": 0.18359375, | |
| "learning_rate": 1.409025231307013e-05, | |
| "loss": 0.8904, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.8474352194606029, | |
| "grad_norm": 0.20703125, | |
| "learning_rate": 1.3854920238615688e-05, | |
| "loss": 0.8857, | |
| "step": 3205 | |
| }, | |
| { | |
| "epoch": 0.8487572712850344, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 1.3621423626856756e-05, | |
| "loss": 0.8863, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 0.8500793231094659, | |
| "grad_norm": 0.25390625, | |
| "learning_rate": 1.3389767452822943e-05, | |
| "loss": 0.8851, | |
| "step": 3215 | |
| }, | |
| { | |
| "epoch": 0.8514013749338974, | |
| "grad_norm": 0.2177734375, | |
| "learning_rate": 1.3159956652330251e-05, | |
| "loss": 0.886, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 0.8527234267583289, | |
| "grad_norm": 0.21484375, | |
| "learning_rate": 1.2931996121876033e-05, | |
| "loss": 0.8822, | |
| "step": 3225 | |
| }, | |
| { | |
| "epoch": 0.8540454785827605, | |
| "grad_norm": 0.1640625, | |
| "learning_rate": 1.2705890718534508e-05, | |
| "loss": 0.8833, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 0.8553675304071919, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 1.2481645259853436e-05, | |
| "loss": 0.8854, | |
| "step": 3235 | |
| }, | |
| { | |
| "epoch": 0.8566895822316235, | |
| "grad_norm": 0.220703125, | |
| "learning_rate": 1.2259264523751335e-05, | |
| "loss": 0.8882, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 0.858011634056055, | |
| "grad_norm": 0.18359375, | |
| "learning_rate": 1.2038753248415768e-05, | |
| "loss": 0.8862, | |
| "step": 3245 | |
| }, | |
| { | |
| "epoch": 0.8593336858804865, | |
| "grad_norm": 0.1572265625, | |
| "learning_rate": 1.1820116132202374e-05, | |
| "loss": 0.8882, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.860655737704918, | |
| "grad_norm": 0.2431640625, | |
| "learning_rate": 1.16033578335347e-05, | |
| "loss": 0.888, | |
| "step": 3255 | |
| }, | |
| { | |
| "epoch": 0.8619777895293496, | |
| "grad_norm": 0.212890625, | |
| "learning_rate": 1.138848297080507e-05, | |
| "loss": 0.887, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 0.863299841353781, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 1.1175496122276008e-05, | |
| "loss": 0.8841, | |
| "step": 3265 | |
| }, | |
| { | |
| "epoch": 0.8646218931782126, | |
| "grad_norm": 0.1728515625, | |
| "learning_rate": 1.0964401825982895e-05, | |
| "loss": 0.8828, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 0.8659439450026442, | |
| "grad_norm": 0.2294921875, | |
| "learning_rate": 1.075520457963708e-05, | |
| "loss": 0.8828, | |
| "step": 3275 | |
| }, | |
| { | |
| "epoch": 0.8672659968270756, | |
| "grad_norm": 0.2412109375, | |
| "learning_rate": 1.0547908840530197e-05, | |
| "loss": 0.8892, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 0.8685880486515072, | |
| "grad_norm": 0.1513671875, | |
| "learning_rate": 1.034251902543908e-05, | |
| "loss": 0.8842, | |
| "step": 3285 | |
| }, | |
| { | |
| "epoch": 0.8699101004759386, | |
| "grad_norm": 0.1669921875, | |
| "learning_rate": 1.01390395105318e-05, | |
| "loss": 0.8858, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 0.8712321523003702, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 9.937474631274257e-06, | |
| "loss": 0.8896, | |
| "step": 3295 | |
| }, | |
| { | |
| "epoch": 0.8725542041248017, | |
| "grad_norm": 0.220703125, | |
| "learning_rate": 9.737828682337969e-06, | |
| "loss": 0.8844, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.8738762559492332, | |
| "grad_norm": 0.1806640625, | |
| "learning_rate": 9.540105917508391e-06, | |
| "loss": 0.885, | |
| "step": 3305 | |
| }, | |
| { | |
| "epoch": 0.8751983077736647, | |
| "grad_norm": 0.1826171875, | |
| "learning_rate": 9.34431054959447e-06, | |
| "loss": 0.8878, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 0.8765203595980963, | |
| "grad_norm": 0.1748046875, | |
| "learning_rate": 9.150446750338714e-06, | |
| "loss": 0.8867, | |
| "step": 3315 | |
| }, | |
| { | |
| "epoch": 0.8778424114225277, | |
| "grad_norm": 0.255859375, | |
| "learning_rate": 8.958518650328429e-06, | |
| "loss": 0.8846, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 0.8791644632469593, | |
| "grad_norm": 0.169921875, | |
| "learning_rate": 8.768530338907655e-06, | |
| "loss": 0.8856, | |
| "step": 3325 | |
| }, | |
| { | |
| "epoch": 0.8804865150713908, | |
| "grad_norm": 0.2099609375, | |
| "learning_rate": 8.580485864090015e-06, | |
| "loss": 0.8829, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 0.8818085668958223, | |
| "grad_norm": 0.1748046875, | |
| "learning_rate": 8.394389232472499e-06, | |
| "loss": 0.8895, | |
| "step": 3335 | |
| }, | |
| { | |
| "epoch": 0.8831306187202538, | |
| "grad_norm": 0.2470703125, | |
| "learning_rate": 8.210244409150158e-06, | |
| "loss": 0.8858, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 0.8844526705446853, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 8.028055317631455e-06, | |
| "loss": 0.8871, | |
| "step": 3345 | |
| }, | |
| { | |
| "epoch": 0.8857747223691169, | |
| "grad_norm": 0.2236328125, | |
| "learning_rate": 7.847825839754852e-06, | |
| "loss": 0.8837, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.8870967741935484, | |
| "grad_norm": 0.216796875, | |
| "learning_rate": 7.669559815605965e-06, | |
| "loss": 0.8839, | |
| "step": 3355 | |
| }, | |
| { | |
| "epoch": 0.88841882601798, | |
| "grad_norm": 0.1533203125, | |
| "learning_rate": 7.493261043435828e-06, | |
| "loss": 0.8802, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 0.8897408778424114, | |
| "grad_norm": 0.169921875, | |
| "learning_rate": 7.318933279579909e-06, | |
| "loss": 0.8831, | |
| "step": 3365 | |
| }, | |
| { | |
| "epoch": 0.891062929666843, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 7.146580238378131e-06, | |
| "loss": 0.8866, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 0.8923849814912744, | |
| "grad_norm": 0.1689453125, | |
| "learning_rate": 6.976205592095675e-06, | |
| "loss": 0.886, | |
| "step": 3375 | |
| }, | |
| { | |
| "epoch": 0.893707033315706, | |
| "grad_norm": 0.2119140625, | |
| "learning_rate": 6.8078129708447516e-06, | |
| "loss": 0.8897, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 0.8950290851401375, | |
| "grad_norm": 0.18359375, | |
| "learning_rate": 6.641405962507307e-06, | |
| "loss": 0.886, | |
| "step": 3385 | |
| }, | |
| { | |
| "epoch": 0.896351136964569, | |
| "grad_norm": 0.18359375, | |
| "learning_rate": 6.4769881126584996e-06, | |
| "loss": 0.8848, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 0.8976731887890005, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 6.314562924491229e-06, | |
| "loss": 0.8863, | |
| "step": 3395 | |
| }, | |
| { | |
| "epoch": 0.8989952406134321, | |
| "grad_norm": 0.208984375, | |
| "learning_rate": 6.154133858741407e-06, | |
| "loss": 0.8861, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.9003172924378635, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 5.995704333614327e-06, | |
| "loss": 0.8867, | |
| "step": 3405 | |
| }, | |
| { | |
| "epoch": 0.9016393442622951, | |
| "grad_norm": 0.28125, | |
| "learning_rate": 5.839277724711733e-06, | |
| "loss": 0.8874, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 0.9029613960867267, | |
| "grad_norm": 0.2373046875, | |
| "learning_rate": 5.6848573649599834e-06, | |
| "loss": 0.8782, | |
| "step": 3415 | |
| }, | |
| { | |
| "epoch": 0.9042834479111581, | |
| "grad_norm": 0.185546875, | |
| "learning_rate": 5.53244654453896e-06, | |
| "loss": 0.8851, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 0.9056054997355897, | |
| "grad_norm": 0.1953125, | |
| "learning_rate": 5.382048510812032e-06, | |
| "loss": 0.8814, | |
| "step": 3425 | |
| }, | |
| { | |
| "epoch": 0.9069275515600211, | |
| "grad_norm": 0.2294921875, | |
| "learning_rate": 5.233666468256804e-06, | |
| "loss": 0.888, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 0.9082496033844527, | |
| "grad_norm": 0.2041015625, | |
| "learning_rate": 5.0873035783969066e-06, | |
| "loss": 0.8862, | |
| "step": 3435 | |
| }, | |
| { | |
| "epoch": 0.9095716552088842, | |
| "grad_norm": 0.16796875, | |
| "learning_rate": 4.942962959734565e-06, | |
| "loss": 0.8835, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 0.9108937070333157, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 4.800647687684223e-06, | |
| "loss": 0.8893, | |
| "step": 3445 | |
| }, | |
| { | |
| "epoch": 0.9122157588577472, | |
| "grad_norm": 0.2197265625, | |
| "learning_rate": 4.660360794506946e-06, | |
| "loss": 0.8826, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.9135378106821788, | |
| "grad_norm": 0.166015625, | |
| "learning_rate": 4.52210526924588e-06, | |
| "loss": 0.8845, | |
| "step": 3455 | |
| }, | |
| { | |
| "epoch": 0.9148598625066102, | |
| "grad_norm": 0.1943359375, | |
| "learning_rate": 4.385884057662515e-06, | |
| "loss": 0.8862, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 0.9161819143310418, | |
| "grad_norm": 0.2216796875, | |
| "learning_rate": 4.251700062173947e-06, | |
| "loss": 0.886, | |
| "step": 3465 | |
| }, | |
| { | |
| "epoch": 0.9175039661554732, | |
| "grad_norm": 0.220703125, | |
| "learning_rate": 4.119556141791048e-06, | |
| "loss": 0.888, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 0.9188260179799048, | |
| "grad_norm": 0.1826171875, | |
| "learning_rate": 3.989455112057483e-06, | |
| "loss": 0.8857, | |
| "step": 3475 | |
| }, | |
| { | |
| "epoch": 0.9201480698043363, | |
| "grad_norm": 0.2451171875, | |
| "learning_rate": 3.861399744989813e-06, | |
| "loss": 0.8871, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 0.9214701216287678, | |
| "grad_norm": 0.203125, | |
| "learning_rate": 3.7353927690183867e-06, | |
| "loss": 0.8852, | |
| "step": 3485 | |
| }, | |
| { | |
| "epoch": 0.9227921734531994, | |
| "grad_norm": 0.1884765625, | |
| "learning_rate": 3.6114368689291856e-06, | |
| "loss": 0.8834, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 0.9241142252776309, | |
| "grad_norm": 0.267578125, | |
| "learning_rate": 3.4895346858066724e-06, | |
| "loss": 0.8847, | |
| "step": 3495 | |
| }, | |
| { | |
| "epoch": 0.9254362771020624, | |
| "grad_norm": 0.2138671875, | |
| "learning_rate": 3.3696888169774677e-06, | |
| "loss": 0.8825, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.9267583289264939, | |
| "grad_norm": 0.1806640625, | |
| "learning_rate": 3.2519018159550496e-06, | |
| "loss": 0.8884, | |
| "step": 3505 | |
| }, | |
| { | |
| "epoch": 0.9280803807509255, | |
| "grad_norm": 0.3671875, | |
| "learning_rate": 3.1361761923853095e-06, | |
| "loss": 0.8816, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 0.9294024325753569, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 3.0225144119931158e-06, | |
| "loss": 0.8828, | |
| "step": 3515 | |
| }, | |
| { | |
| "epoch": 0.9307244843997885, | |
| "grad_norm": 0.2734375, | |
| "learning_rate": 2.910918896529757e-06, | |
| "loss": 0.8851, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 0.93204653622422, | |
| "grad_norm": 0.236328125, | |
| "learning_rate": 2.8013920237213277e-06, | |
| "loss": 0.8851, | |
| "step": 3525 | |
| }, | |
| { | |
| "epoch": 0.9333685880486515, | |
| "grad_norm": 0.2451171875, | |
| "learning_rate": 2.693936127218133e-06, | |
| "loss": 0.8877, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 0.934690639873083, | |
| "grad_norm": 0.2294921875, | |
| "learning_rate": 2.5885534965448768e-06, | |
| "loss": 0.8848, | |
| "step": 3535 | |
| }, | |
| { | |
| "epoch": 0.9360126916975146, | |
| "grad_norm": 0.1513671875, | |
| "learning_rate": 2.4852463770519417e-06, | |
| "loss": 0.8823, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 0.937334743521946, | |
| "grad_norm": 0.2197265625, | |
| "learning_rate": 2.3840169698675087e-06, | |
| "loss": 0.8865, | |
| "step": 3545 | |
| }, | |
| { | |
| "epoch": 0.9386567953463776, | |
| "grad_norm": 0.2431640625, | |
| "learning_rate": 2.2848674318507125e-06, | |
| "loss": 0.8829, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.9399788471708092, | |
| "grad_norm": 0.2216796875, | |
| "learning_rate": 2.1877998755456266e-06, | |
| "loss": 0.8851, | |
| "step": 3555 | |
| }, | |
| { | |
| "epoch": 0.9413008989952406, | |
| "grad_norm": 0.1884765625, | |
| "learning_rate": 2.0928163691362856e-06, | |
| "loss": 0.8876, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 0.9426229508196722, | |
| "grad_norm": 0.169921875, | |
| "learning_rate": 1.9999189364025894e-06, | |
| "loss": 0.8861, | |
| "step": 3565 | |
| }, | |
| { | |
| "epoch": 0.9439450026441036, | |
| "grad_norm": 0.2275390625, | |
| "learning_rate": 1.909109556677269e-06, | |
| "loss": 0.8874, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 0.9452670544685352, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 1.8203901648035648e-06, | |
| "loss": 0.8865, | |
| "step": 3575 | |
| }, | |
| { | |
| "epoch": 0.9465891062929667, | |
| "grad_norm": 0.181640625, | |
| "learning_rate": 1.7337626510941619e-06, | |
| "loss": 0.8876, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 0.9479111581173982, | |
| "grad_norm": 0.3046875, | |
| "learning_rate": 1.6492288612907748e-06, | |
| "loss": 0.8889, | |
| "step": 3585 | |
| }, | |
| { | |
| "epoch": 0.9492332099418297, | |
| "grad_norm": 0.208984375, | |
| "learning_rate": 1.5667905965249363e-06, | |
| "loss": 0.8842, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 0.9505552617662613, | |
| "grad_norm": 0.1630859375, | |
| "learning_rate": 1.486449613279539e-06, | |
| "loss": 0.8857, | |
| "step": 3595 | |
| }, | |
| { | |
| "epoch": 0.9518773135906927, | |
| "grad_norm": 0.208984375, | |
| "learning_rate": 1.408207623351454e-06, | |
| "loss": 0.8847, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.9531993654151243, | |
| "grad_norm": 0.2353515625, | |
| "learning_rate": 1.3320662938150485e-06, | |
| "loss": 0.8828, | |
| "step": 3605 | |
| }, | |
| { | |
| "epoch": 0.9545214172395557, | |
| "grad_norm": 0.1669921875, | |
| "learning_rate": 1.2580272469866483e-06, | |
| "loss": 0.8787, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 0.9558434690639873, | |
| "grad_norm": 0.2080078125, | |
| "learning_rate": 1.186092060389976e-06, | |
| "loss": 0.886, | |
| "step": 3615 | |
| }, | |
| { | |
| "epoch": 0.9571655208884188, | |
| "grad_norm": 0.146484375, | |
| "learning_rate": 1.1162622667226008e-06, | |
| "loss": 0.8837, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 0.9584875727128503, | |
| "grad_norm": 0.1806640625, | |
| "learning_rate": 1.0485393538231747e-06, | |
| "loss": 0.8846, | |
| "step": 3625 | |
| }, | |
| { | |
| "epoch": 0.9598096245372819, | |
| "grad_norm": 0.234375, | |
| "learning_rate": 9.829247646398144e-07, | |
| "loss": 0.8867, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 0.9611316763617134, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 9.194198971993362e-07, | |
| "loss": 0.8869, | |
| "step": 3635 | |
| }, | |
| { | |
| "epoch": 0.9624537281861449, | |
| "grad_norm": 0.265625, | |
| "learning_rate": 8.580261045774474e-07, | |
| "loss": 0.8824, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 0.9637757800105764, | |
| "grad_norm": 0.16796875, | |
| "learning_rate": 7.98744694869924e-07, | |
| "loss": 0.8875, | |
| "step": 3645 | |
| }, | |
| { | |
| "epoch": 0.965097831835008, | |
| "grad_norm": 0.291015625, | |
| "learning_rate": 7.41576931164778e-07, | |
| "loss": 0.8856, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.9664198836594394, | |
| "grad_norm": 0.2236328125, | |
| "learning_rate": 6.86524031515301e-07, | |
| "loss": 0.8861, | |
| "step": 3655 | |
| }, | |
| { | |
| "epoch": 0.967741935483871, | |
| "grad_norm": 0.1806640625, | |
| "learning_rate": 6.335871689140959e-07, | |
| "loss": 0.8868, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 0.9690639873083025, | |
| "grad_norm": 0.1650390625, | |
| "learning_rate": 5.827674712681752e-07, | |
| "loss": 0.8873, | |
| "step": 3665 | |
| }, | |
| { | |
| "epoch": 0.970386039132734, | |
| "grad_norm": 0.2138671875, | |
| "learning_rate": 5.340660213748017e-07, | |
| "loss": 0.8833, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 0.9717080909571655, | |
| "grad_norm": 0.2041015625, | |
| "learning_rate": 4.874838568985296e-07, | |
| "loss": 0.8803, | |
| "step": 3675 | |
| }, | |
| { | |
| "epoch": 0.973030142781597, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 4.430219703490335e-07, | |
| "loss": 0.8846, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 0.9743521946060285, | |
| "grad_norm": 0.28515625, | |
| "learning_rate": 4.0068130905996924e-07, | |
| "loss": 0.8873, | |
| "step": 3685 | |
| }, | |
| { | |
| "epoch": 0.9756742464304601, | |
| "grad_norm": 0.2392578125, | |
| "learning_rate": 3.604627751688239e-07, | |
| "loss": 0.8817, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 0.9769962982548915, | |
| "grad_norm": 0.1845703125, | |
| "learning_rate": 3.2236722559764176e-07, | |
| "loss": 0.8877, | |
| "step": 3695 | |
| }, | |
| { | |
| "epoch": 0.9783183500793231, | |
| "grad_norm": 0.19140625, | |
| "learning_rate": 2.86395472034795e-07, | |
| "loss": 0.8879, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.9796404019037547, | |
| "grad_norm": 0.2578125, | |
| "learning_rate": 2.525482809177082e-07, | |
| "loss": 0.8908, | |
| "step": 3705 | |
| }, | |
| { | |
| "epoch": 0.9809624537281861, | |
| "grad_norm": 0.173828125, | |
| "learning_rate": 2.2082637341647173e-07, | |
| "loss": 0.881, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 0.9822845055526177, | |
| "grad_norm": 0.2080078125, | |
| "learning_rate": 1.9123042541854265e-07, | |
| "loss": 0.8863, | |
| "step": 3715 | |
| }, | |
| { | |
| "epoch": 0.9836065573770492, | |
| "grad_norm": 0.16796875, | |
| "learning_rate": 1.6376106751430087e-07, | |
| "loss": 0.8909, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 0.9849286092014807, | |
| "grad_norm": 0.21484375, | |
| "learning_rate": 1.3841888498361544e-07, | |
| "loss": 0.8815, | |
| "step": 3725 | |
| }, | |
| { | |
| "epoch": 0.9862506610259122, | |
| "grad_norm": 0.23828125, | |
| "learning_rate": 1.1520441778339885e-07, | |
| "loss": 0.8851, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 0.9875727128503438, | |
| "grad_norm": 0.259765625, | |
| "learning_rate": 9.411816053608302e-08, | |
| "loss": 0.8855, | |
| "step": 3735 | |
| }, | |
| { | |
| "epoch": 0.9888947646747752, | |
| "grad_norm": 0.197265625, | |
| "learning_rate": 7.51605625190721e-08, | |
| "loss": 0.8903, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 0.9902168164992068, | |
| "grad_norm": 0.1845703125, | |
| "learning_rate": 5.833202765519463e-08, | |
| "loss": 0.8867, | |
| "step": 3745 | |
| }, | |
| { | |
| "epoch": 0.9915388683236382, | |
| "grad_norm": 0.1943359375, | |
| "learning_rate": 4.3632914504077026e-08, | |
| "loss": 0.8839, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.9928609201480698, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 3.106353625451641e-08, | |
| "loss": 0.8824, | |
| "step": 3755 | |
| }, | |
| { | |
| "epoch": 0.9941829719725013, | |
| "grad_norm": 0.1708984375, | |
| "learning_rate": 2.062416071780815e-08, | |
| "loss": 0.8881, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 0.9955050237969328, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 1.2315010322028198e-08, | |
| "loss": 0.8869, | |
| "step": 3765 | |
| }, | |
| { | |
| "epoch": 0.9968270756213644, | |
| "grad_norm": 0.19921875, | |
| "learning_rate": 6.136262107292456e-09, | |
| "loss": 0.8892, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 0.9981491274457959, | |
| "grad_norm": 0.23828125, | |
| "learning_rate": 2.0880477220042163e-09, | |
| "loss": 0.8853, | |
| "step": 3775 | |
| }, | |
| { | |
| "epoch": 0.9994711792702274, | |
| "grad_norm": 0.234375, | |
| "learning_rate": 1.7045342003418895e-10, | |
| "loss": 0.8854, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 3782, | |
| "total_flos": 1.011923420184576e+18, | |
| "train_loss": 0.9601045426144794, | |
| "train_runtime": 898.0661, | |
| "train_samples_per_second": 2156.094, | |
| "train_steps_per_second": 4.211 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 3782, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.011923420184576e+18, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |