diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,6727 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.980443285528032, + "eval_steps": 500, + "global_step": 955, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.005215123859191656, + "grad_norm": 8.640251838986332, + "learning_rate": 8.333333333333333e-07, + "loss": 1.3109, + "step": 1 + }, + { + "epoch": 0.010430247718383311, + "grad_norm": 8.50165895125568, + "learning_rate": 1.6666666666666667e-06, + "loss": 1.2827, + "step": 2 + }, + { + "epoch": 0.01564537157757497, + "grad_norm": 8.535323811460232, + "learning_rate": 2.5e-06, + "loss": 1.2971, + "step": 3 + }, + { + "epoch": 0.020860495436766623, + "grad_norm": 8.005773755772996, + "learning_rate": 3.3333333333333333e-06, + "loss": 1.2835, + "step": 4 + }, + { + "epoch": 0.02607561929595828, + "grad_norm": 6.542532290299446, + "learning_rate": 4.166666666666667e-06, + "loss": 1.2424, + "step": 5 + }, + { + "epoch": 0.03129074315514994, + "grad_norm": 3.513863098870053, + "learning_rate": 5e-06, + "loss": 1.1855, + "step": 6 + }, + { + "epoch": 0.03650586701434159, + "grad_norm": 2.7215053265085425, + "learning_rate": 5.833333333333334e-06, + "loss": 1.1771, + "step": 7 + }, + { + "epoch": 0.041720990873533245, + "grad_norm": 6.45835816206034, + "learning_rate": 6.666666666666667e-06, + "loss": 1.1818, + "step": 8 + }, + { + "epoch": 0.0469361147327249, + "grad_norm": 6.896643054762279, + "learning_rate": 7.500000000000001e-06, + "loss": 1.1927, + "step": 9 + }, + { + "epoch": 0.05215123859191656, + "grad_norm": 7.106049193713323, + "learning_rate": 8.333333333333334e-06, + "loss": 1.192, + "step": 10 + }, + { + "epoch": 0.05736636245110821, + "grad_norm": 5.8848429339749755, + "learning_rate": 9.166666666666666e-06, + "loss": 1.1358, + "step": 11 + }, + { + "epoch": 0.06258148631029987, + "grad_norm": 5.16279034442262, + "learning_rate": 1e-05, + "loss": 1.1187, + "step": 12 + }, + { + "epoch": 0.06779661016949153, + "grad_norm": 3.4057013372590914, + "learning_rate": 1.0833333333333334e-05, + "loss": 1.1044, + "step": 13 + }, + { + "epoch": 0.07301173402868318, + "grad_norm": 2.2087972161452516, + "learning_rate": 1.1666666666666668e-05, + "loss": 1.0906, + "step": 14 + }, + { + "epoch": 0.07822685788787484, + "grad_norm": 2.1528478641437006, + "learning_rate": 1.25e-05, + "loss": 1.0749, + "step": 15 + }, + { + "epoch": 0.08344198174706649, + "grad_norm": 2.0837468590704984, + "learning_rate": 1.3333333333333333e-05, + "loss": 1.041, + "step": 16 + }, + { + "epoch": 0.08865710560625815, + "grad_norm": 2.3435856245272064, + "learning_rate": 1.416666666666667e-05, + "loss": 1.0308, + "step": 17 + }, + { + "epoch": 0.0938722294654498, + "grad_norm": 1.7734881445436932, + "learning_rate": 1.5000000000000002e-05, + "loss": 1.0104, + "step": 18 + }, + { + "epoch": 0.09908735332464146, + "grad_norm": 1.2852782680220982, + "learning_rate": 1.5833333333333333e-05, + "loss": 1.0159, + "step": 19 + }, + { + "epoch": 0.10430247718383312, + "grad_norm": 1.4664459559013807, + "learning_rate": 1.6666666666666667e-05, + "loss": 0.9997, + "step": 20 + }, + { + "epoch": 0.10951760104302477, + "grad_norm": 1.2747674832880032, + "learning_rate": 1.7500000000000002e-05, + "loss": 0.9999, + "step": 21 + }, + { + "epoch": 0.11473272490221642, + "grad_norm": 1.1325966855930794, + "learning_rate": 1.8333333333333333e-05, + "loss": 0.9845, + "step": 22 + }, + { + "epoch": 0.11994784876140809, + "grad_norm": 1.178213874446251, + "learning_rate": 1.916666666666667e-05, + "loss": 0.9754, + "step": 23 + }, + { + "epoch": 0.12516297262059975, + "grad_norm": 1.0070300787437625, + "learning_rate": 2e-05, + "loss": 0.9916, + "step": 24 + }, + { + "epoch": 0.1303780964797914, + "grad_norm": 1.1301814564159125, + "learning_rate": 2.0833333333333336e-05, + "loss": 0.9535, + "step": 25 + }, + { + "epoch": 0.13559322033898305, + "grad_norm": 1.1582102527170561, + "learning_rate": 2.1666666666666667e-05, + "loss": 0.9559, + "step": 26 + }, + { + "epoch": 0.1408083441981747, + "grad_norm": 1.28096007382199, + "learning_rate": 2.25e-05, + "loss": 0.953, + "step": 27 + }, + { + "epoch": 0.14602346805736635, + "grad_norm": 1.3963901299642703, + "learning_rate": 2.3333333333333336e-05, + "loss": 0.9467, + "step": 28 + }, + { + "epoch": 0.15123859191655803, + "grad_norm": 1.2008437237669338, + "learning_rate": 2.4166666666666667e-05, + "loss": 0.9319, + "step": 29 + }, + { + "epoch": 0.15645371577574968, + "grad_norm": 1.4579402450445522, + "learning_rate": 2.5e-05, + "loss": 0.9583, + "step": 30 + }, + { + "epoch": 0.16166883963494133, + "grad_norm": 1.332654985022459, + "learning_rate": 2.5833333333333336e-05, + "loss": 0.9222, + "step": 31 + }, + { + "epoch": 0.16688396349413298, + "grad_norm": 1.0118249538528512, + "learning_rate": 2.6666666666666667e-05, + "loss": 0.919, + "step": 32 + }, + { + "epoch": 0.17209908735332463, + "grad_norm": 1.9244594897055562, + "learning_rate": 2.75e-05, + "loss": 0.9432, + "step": 33 + }, + { + "epoch": 0.1773142112125163, + "grad_norm": 1.1988103983333642, + "learning_rate": 2.833333333333334e-05, + "loss": 0.9309, + "step": 34 + }, + { + "epoch": 0.18252933507170796, + "grad_norm": 1.375838275777245, + "learning_rate": 2.9166666666666666e-05, + "loss": 0.9164, + "step": 35 + }, + { + "epoch": 0.1877444589308996, + "grad_norm": 1.662174214128309, + "learning_rate": 3.0000000000000004e-05, + "loss": 0.9116, + "step": 36 + }, + { + "epoch": 0.19295958279009126, + "grad_norm": 19.880826748141878, + "learning_rate": 3.0833333333333335e-05, + "loss": 0.9136, + "step": 37 + }, + { + "epoch": 0.1981747066492829, + "grad_norm": 2.1089494461311897, + "learning_rate": 3.1666666666666666e-05, + "loss": 0.9288, + "step": 38 + }, + { + "epoch": 0.2033898305084746, + "grad_norm": 1.0754107081344952, + "learning_rate": 3.2500000000000004e-05, + "loss": 0.9173, + "step": 39 + }, + { + "epoch": 0.20860495436766624, + "grad_norm": 3.9611232218120964, + "learning_rate": 3.3333333333333335e-05, + "loss": 0.9227, + "step": 40 + }, + { + "epoch": 0.2138200782268579, + "grad_norm": 3.0041123063486306, + "learning_rate": 3.4166666666666666e-05, + "loss": 0.9319, + "step": 41 + }, + { + "epoch": 0.21903520208604954, + "grad_norm": 2.7847797531368066, + "learning_rate": 3.5000000000000004e-05, + "loss": 0.9233, + "step": 42 + }, + { + "epoch": 0.2242503259452412, + "grad_norm": 2.2845773931740725, + "learning_rate": 3.5833333333333335e-05, + "loss": 0.9286, + "step": 43 + }, + { + "epoch": 0.22946544980443284, + "grad_norm": 2.9550586599454363, + "learning_rate": 3.6666666666666666e-05, + "loss": 0.9194, + "step": 44 + }, + { + "epoch": 0.23468057366362452, + "grad_norm": 1.713180532466099, + "learning_rate": 3.7500000000000003e-05, + "loss": 0.8909, + "step": 45 + }, + { + "epoch": 0.23989569752281617, + "grad_norm": 3.6245500815283984, + "learning_rate": 3.833333333333334e-05, + "loss": 0.9061, + "step": 46 + }, + { + "epoch": 0.24511082138200782, + "grad_norm": 2.7780460507158034, + "learning_rate": 3.9166666666666665e-05, + "loss": 0.9165, + "step": 47 + }, + { + "epoch": 0.2503259452411995, + "grad_norm": 3.3854708521313563, + "learning_rate": 4e-05, + "loss": 0.9076, + "step": 48 + }, + { + "epoch": 0.25554106910039115, + "grad_norm": 2.788766938646062, + "learning_rate": 4.0833333333333334e-05, + "loss": 0.9086, + "step": 49 + }, + { + "epoch": 0.2607561929595828, + "grad_norm": 3.769154968802035, + "learning_rate": 4.166666666666667e-05, + "loss": 0.8985, + "step": 50 + }, + { + "epoch": 0.26597131681877445, + "grad_norm": 3.6676032273552783, + "learning_rate": 4.25e-05, + "loss": 0.9186, + "step": 51 + }, + { + "epoch": 0.2711864406779661, + "grad_norm": 2.300855067051272, + "learning_rate": 4.3333333333333334e-05, + "loss": 0.8785, + "step": 52 + }, + { + "epoch": 0.27640156453715775, + "grad_norm": 2.0855212942964156, + "learning_rate": 4.416666666666667e-05, + "loss": 0.9044, + "step": 53 + }, + { + "epoch": 0.2816166883963494, + "grad_norm": 3.1541260770808144, + "learning_rate": 4.5e-05, + "loss": 0.9134, + "step": 54 + }, + { + "epoch": 0.28683181225554105, + "grad_norm": 2.3439141139233635, + "learning_rate": 4.5833333333333334e-05, + "loss": 0.8849, + "step": 55 + }, + { + "epoch": 0.2920469361147327, + "grad_norm": 3.622853136869387, + "learning_rate": 4.666666666666667e-05, + "loss": 0.8968, + "step": 56 + }, + { + "epoch": 0.29726205997392435, + "grad_norm": 3.3444739650293793, + "learning_rate": 4.75e-05, + "loss": 0.9119, + "step": 57 + }, + { + "epoch": 0.30247718383311606, + "grad_norm": 2.4056743976018007, + "learning_rate": 4.8333333333333334e-05, + "loss": 0.8978, + "step": 58 + }, + { + "epoch": 0.3076923076923077, + "grad_norm": 2.0165924530855692, + "learning_rate": 4.916666666666667e-05, + "loss": 0.8842, + "step": 59 + }, + { + "epoch": 0.31290743155149936, + "grad_norm": 3.199274379857031, + "learning_rate": 5e-05, + "loss": 0.8926, + "step": 60 + }, + { + "epoch": 0.318122555410691, + "grad_norm": 2.145560466910737, + "learning_rate": 5.0833333333333333e-05, + "loss": 0.8875, + "step": 61 + }, + { + "epoch": 0.32333767926988266, + "grad_norm": 3.3364603622456923, + "learning_rate": 5.166666666666667e-05, + "loss": 0.8891, + "step": 62 + }, + { + "epoch": 0.3285528031290743, + "grad_norm": 3.088241155388676, + "learning_rate": 5.25e-05, + "loss": 0.8989, + "step": 63 + }, + { + "epoch": 0.33376792698826596, + "grad_norm": 2.2703016974586836, + "learning_rate": 5.333333333333333e-05, + "loss": 0.8848, + "step": 64 + }, + { + "epoch": 0.3389830508474576, + "grad_norm": 2.1056335353722875, + "learning_rate": 5.416666666666667e-05, + "loss": 0.8895, + "step": 65 + }, + { + "epoch": 0.34419817470664926, + "grad_norm": 2.195474617684906, + "learning_rate": 5.5e-05, + "loss": 0.8855, + "step": 66 + }, + { + "epoch": 0.3494132985658409, + "grad_norm": 1.8388111784516368, + "learning_rate": 5.583333333333333e-05, + "loss": 0.88, + "step": 67 + }, + { + "epoch": 0.3546284224250326, + "grad_norm": 2.205537218893776, + "learning_rate": 5.666666666666668e-05, + "loss": 0.8923, + "step": 68 + }, + { + "epoch": 0.35984354628422427, + "grad_norm": 1.43087259231813, + "learning_rate": 5.75e-05, + "loss": 0.8783, + "step": 69 + }, + { + "epoch": 0.3650586701434159, + "grad_norm": 3.170207545618331, + "learning_rate": 5.833333333333333e-05, + "loss": 0.8979, + "step": 70 + }, + { + "epoch": 0.37027379400260757, + "grad_norm": 2.4965183591219686, + "learning_rate": 5.916666666666668e-05, + "loss": 0.887, + "step": 71 + }, + { + "epoch": 0.3754889178617992, + "grad_norm": 2.18025601493941, + "learning_rate": 6.000000000000001e-05, + "loss": 0.8885, + "step": 72 + }, + { + "epoch": 0.38070404172099087, + "grad_norm": 2.320595439694094, + "learning_rate": 6.083333333333333e-05, + "loss": 0.8815, + "step": 73 + }, + { + "epoch": 0.3859191655801825, + "grad_norm": 2.0185818109580467, + "learning_rate": 6.166666666666667e-05, + "loss": 0.8831, + "step": 74 + }, + { + "epoch": 0.39113428943937417, + "grad_norm": 2.2105832623786745, + "learning_rate": 6.25e-05, + "loss": 0.8666, + "step": 75 + }, + { + "epoch": 0.3963494132985658, + "grad_norm": 1.9340187658703214, + "learning_rate": 6.333333333333333e-05, + "loss": 0.888, + "step": 76 + }, + { + "epoch": 0.4015645371577575, + "grad_norm": 1.811786813233452, + "learning_rate": 6.416666666666668e-05, + "loss": 0.889, + "step": 77 + }, + { + "epoch": 0.4067796610169492, + "grad_norm": 2.0985830559238563, + "learning_rate": 6.500000000000001e-05, + "loss": 0.8945, + "step": 78 + }, + { + "epoch": 0.41199478487614083, + "grad_norm": 2.7706210707741885, + "learning_rate": 6.583333333333334e-05, + "loss": 0.903, + "step": 79 + }, + { + "epoch": 0.4172099087353325, + "grad_norm": 2.328367128642788, + "learning_rate": 6.666666666666667e-05, + "loss": 0.886, + "step": 80 + }, + { + "epoch": 0.42242503259452413, + "grad_norm": 2.7310201710261857, + "learning_rate": 6.75e-05, + "loss": 0.8728, + "step": 81 + }, + { + "epoch": 0.4276401564537158, + "grad_norm": 2.012060559904938, + "learning_rate": 6.833333333333333e-05, + "loss": 0.8807, + "step": 82 + }, + { + "epoch": 0.43285528031290743, + "grad_norm": 1.8442023724268992, + "learning_rate": 6.916666666666668e-05, + "loss": 0.8832, + "step": 83 + }, + { + "epoch": 0.4380704041720991, + "grad_norm": 3.3879560119725225, + "learning_rate": 7.000000000000001e-05, + "loss": 0.8848, + "step": 84 + }, + { + "epoch": 0.44328552803129073, + "grad_norm": 2.0302673957424457, + "learning_rate": 7.083333333333334e-05, + "loss": 0.8716, + "step": 85 + }, + { + "epoch": 0.4485006518904824, + "grad_norm": 3.238439147300149, + "learning_rate": 7.166666666666667e-05, + "loss": 0.8896, + "step": 86 + }, + { + "epoch": 0.45371577574967403, + "grad_norm": 2.718682181237891, + "learning_rate": 7.25e-05, + "loss": 0.902, + "step": 87 + }, + { + "epoch": 0.4589308996088657, + "grad_norm": 2.4102500849494035, + "learning_rate": 7.333333333333333e-05, + "loss": 0.8865, + "step": 88 + }, + { + "epoch": 0.4641460234680574, + "grad_norm": 2.057159117257536, + "learning_rate": 7.416666666666668e-05, + "loss": 0.8714, + "step": 89 + }, + { + "epoch": 0.46936114732724904, + "grad_norm": 2.303058894217337, + "learning_rate": 7.500000000000001e-05, + "loss": 0.875, + "step": 90 + }, + { + "epoch": 0.4745762711864407, + "grad_norm": 1.853131768077206, + "learning_rate": 7.583333333333334e-05, + "loss": 0.8697, + "step": 91 + }, + { + "epoch": 0.47979139504563234, + "grad_norm": 1.5089115884426068, + "learning_rate": 7.666666666666668e-05, + "loss": 0.8672, + "step": 92 + }, + { + "epoch": 0.485006518904824, + "grad_norm": 3.288728672540363, + "learning_rate": 7.75e-05, + "loss": 0.8787, + "step": 93 + }, + { + "epoch": 0.49022164276401564, + "grad_norm": 2.8218607718403197, + "learning_rate": 7.833333333333333e-05, + "loss": 0.8886, + "step": 94 + }, + { + "epoch": 0.4954367666232073, + "grad_norm": 1.966995373674031, + "learning_rate": 7.916666666666668e-05, + "loss": 0.8738, + "step": 95 + }, + { + "epoch": 0.500651890482399, + "grad_norm": 4.193763943813693, + "learning_rate": 8e-05, + "loss": 0.8813, + "step": 96 + }, + { + "epoch": 0.5058670143415906, + "grad_norm": 2.2253144313432514, + "learning_rate": 7.99997324882088e-05, + "loss": 0.8778, + "step": 97 + }, + { + "epoch": 0.5110821382007823, + "grad_norm": 4.059040973559077, + "learning_rate": 7.999892995641334e-05, + "loss": 0.8844, + "step": 98 + }, + { + "epoch": 0.516297262059974, + "grad_norm": 2.814215374527903, + "learning_rate": 7.999759241534794e-05, + "loss": 0.8899, + "step": 99 + }, + { + "epoch": 0.5215123859191656, + "grad_norm": 3.1446586623159383, + "learning_rate": 7.9995719882903e-05, + "loss": 0.8933, + "step": 100 + }, + { + "epoch": 0.5267275097783573, + "grad_norm": 2.5831777088882335, + "learning_rate": 7.999331238412474e-05, + "loss": 0.885, + "step": 101 + }, + { + "epoch": 0.5319426336375489, + "grad_norm": 2.571515738465628, + "learning_rate": 7.99903699512149e-05, + "loss": 0.8773, + "step": 102 + }, + { + "epoch": 0.5371577574967406, + "grad_norm": 2.028582558881839, + "learning_rate": 7.998689262353024e-05, + "loss": 0.875, + "step": 103 + }, + { + "epoch": 0.5423728813559322, + "grad_norm": 1.9833279742128762, + "learning_rate": 7.998288044758206e-05, + "loss": 0.8864, + "step": 104 + }, + { + "epoch": 0.5475880052151239, + "grad_norm": 1.595344120443328, + "learning_rate": 7.99783334770356e-05, + "loss": 0.8772, + "step": 105 + }, + { + "epoch": 0.5528031290743155, + "grad_norm": 2.3698581665478025, + "learning_rate": 7.997325177270926e-05, + "loss": 0.8587, + "step": 106 + }, + { + "epoch": 0.5580182529335072, + "grad_norm": 1.8069354168584124, + "learning_rate": 7.996763540257382e-05, + "loss": 0.8786, + "step": 107 + }, + { + "epoch": 0.5632333767926988, + "grad_norm": 2.8802931859293968, + "learning_rate": 7.996148444175155e-05, + "loss": 0.8651, + "step": 108 + }, + { + "epoch": 0.5684485006518905, + "grad_norm": 2.163396645739927, + "learning_rate": 7.99547989725152e-05, + "loss": 0.8779, + "step": 109 + }, + { + "epoch": 0.5736636245110821, + "grad_norm": 1.7103752366008835, + "learning_rate": 7.994757908428683e-05, + "loss": 0.8627, + "step": 110 + }, + { + "epoch": 0.5788787483702738, + "grad_norm": 2.6236911355349366, + "learning_rate": 7.99398248736367e-05, + "loss": 0.8775, + "step": 111 + }, + { + "epoch": 0.5840938722294654, + "grad_norm": 1.7695200626233123, + "learning_rate": 7.993153644428198e-05, + "loss": 0.8675, + "step": 112 + }, + { + "epoch": 0.5893089960886571, + "grad_norm": 3.1189819236023597, + "learning_rate": 7.992271390708529e-05, + "loss": 0.864, + "step": 113 + }, + { + "epoch": 0.5945241199478487, + "grad_norm": 2.519671466442974, + "learning_rate": 7.991335738005325e-05, + "loss": 0.8753, + "step": 114 + }, + { + "epoch": 0.5997392438070405, + "grad_norm": 2.445885184903824, + "learning_rate": 7.990346698833493e-05, + "loss": 0.8791, + "step": 115 + }, + { + "epoch": 0.6049543676662321, + "grad_norm": 2.6107239711114634, + "learning_rate": 7.989304286422016e-05, + "loss": 0.8688, + "step": 116 + }, + { + "epoch": 0.6101694915254238, + "grad_norm": 2.3280917526561926, + "learning_rate": 7.988208514713773e-05, + "loss": 0.8558, + "step": 117 + }, + { + "epoch": 0.6153846153846154, + "grad_norm": 2.01951998260186, + "learning_rate": 7.987059398365358e-05, + "loss": 0.8611, + "step": 118 + }, + { + "epoch": 0.6205997392438071, + "grad_norm": 2.67003693356188, + "learning_rate": 7.98585695274688e-05, + "loss": 0.8599, + "step": 119 + }, + { + "epoch": 0.6258148631029987, + "grad_norm": 1.3989184556669023, + "learning_rate": 7.984601193941757e-05, + "loss": 0.8619, + "step": 120 + }, + { + "epoch": 0.6310299869621904, + "grad_norm": 2.9657726548979007, + "learning_rate": 7.983292138746504e-05, + "loss": 0.8533, + "step": 121 + }, + { + "epoch": 0.636245110821382, + "grad_norm": 2.2111378936066717, + "learning_rate": 7.981929804670505e-05, + "loss": 0.8647, + "step": 122 + }, + { + "epoch": 0.6414602346805737, + "grad_norm": 1.7933021378995146, + "learning_rate": 7.980514209935783e-05, + "loss": 0.8572, + "step": 123 + }, + { + "epoch": 0.6466753585397653, + "grad_norm": 2.6146019104563036, + "learning_rate": 7.97904537347675e-05, + "loss": 0.8629, + "step": 124 + }, + { + "epoch": 0.651890482398957, + "grad_norm": 1.7226667254129786, + "learning_rate": 7.977523314939961e-05, + "loss": 0.8728, + "step": 125 + }, + { + "epoch": 0.6571056062581486, + "grad_norm": 2.4681433772246804, + "learning_rate": 7.975948054683847e-05, + "loss": 0.8772, + "step": 126 + }, + { + "epoch": 0.6623207301173403, + "grad_norm": 1.8531960262958524, + "learning_rate": 7.974319613778441e-05, + "loss": 0.8568, + "step": 127 + }, + { + "epoch": 0.6675358539765319, + "grad_norm": 3.3273564870681067, + "learning_rate": 7.972638014005102e-05, + "loss": 0.8674, + "step": 128 + }, + { + "epoch": 0.6727509778357236, + "grad_norm": 1.6565232049947463, + "learning_rate": 7.970903277856216e-05, + "loss": 0.8593, + "step": 129 + }, + { + "epoch": 0.6779661016949152, + "grad_norm": 3.2772231491550157, + "learning_rate": 7.969115428534904e-05, + "loss": 0.8689, + "step": 130 + }, + { + "epoch": 0.6831812255541069, + "grad_norm": 2.9366197326510175, + "learning_rate": 7.967274489954703e-05, + "loss": 0.8562, + "step": 131 + }, + { + "epoch": 0.6883963494132985, + "grad_norm": 1.579970961599737, + "learning_rate": 7.965380486739253e-05, + "loss": 0.8422, + "step": 132 + }, + { + "epoch": 0.6936114732724902, + "grad_norm": 3.1594495489774967, + "learning_rate": 7.963433444221964e-05, + "loss": 0.875, + "step": 133 + }, + { + "epoch": 0.6988265971316818, + "grad_norm": 2.3403703396896423, + "learning_rate": 7.961433388445676e-05, + "loss": 0.8591, + "step": 134 + }, + { + "epoch": 0.7040417209908736, + "grad_norm": 1.8764160357986022, + "learning_rate": 7.959380346162314e-05, + "loss": 0.8585, + "step": 135 + }, + { + "epoch": 0.7092568448500652, + "grad_norm": 2.954668141615744, + "learning_rate": 7.957274344832533e-05, + "loss": 0.8558, + "step": 136 + }, + { + "epoch": 0.7144719687092569, + "grad_norm": 2.1132431617150647, + "learning_rate": 7.955115412625337e-05, + "loss": 0.8597, + "step": 137 + }, + { + "epoch": 0.7196870925684485, + "grad_norm": 1.9269531529548758, + "learning_rate": 7.952903578417719e-05, + "loss": 0.8553, + "step": 138 + }, + { + "epoch": 0.7249022164276402, + "grad_norm": 1.9600506890917169, + "learning_rate": 7.950638871794268e-05, + "loss": 0.8498, + "step": 139 + }, + { + "epoch": 0.7301173402868318, + "grad_norm": 1.295131341774357, + "learning_rate": 7.948321323046766e-05, + "loss": 0.862, + "step": 140 + }, + { + "epoch": 0.7353324641460235, + "grad_norm": 2.598262766807541, + "learning_rate": 7.945950963173797e-05, + "loss": 0.8743, + "step": 141 + }, + { + "epoch": 0.7405475880052151, + "grad_norm": 1.7173227811522496, + "learning_rate": 7.943527823880321e-05, + "loss": 0.857, + "step": 142 + }, + { + "epoch": 0.7457627118644068, + "grad_norm": 2.778103098366881, + "learning_rate": 7.941051937577255e-05, + "loss": 0.8644, + "step": 143 + }, + { + "epoch": 0.7509778357235984, + "grad_norm": 1.9721804979121416, + "learning_rate": 7.938523337381036e-05, + "loss": 0.8641, + "step": 144 + }, + { + "epoch": 0.7561929595827901, + "grad_norm": 2.534467842230536, + "learning_rate": 7.935942057113185e-05, + "loss": 0.8607, + "step": 145 + }, + { + "epoch": 0.7614080834419817, + "grad_norm": 2.420990126278799, + "learning_rate": 7.933308131299846e-05, + "loss": 0.8532, + "step": 146 + }, + { + "epoch": 0.7666232073011734, + "grad_norm": 2.0250401094262775, + "learning_rate": 7.93062159517133e-05, + "loss": 0.8463, + "step": 147 + }, + { + "epoch": 0.771838331160365, + "grad_norm": 1.949074222131184, + "learning_rate": 7.92788248466164e-05, + "loss": 0.8505, + "step": 148 + }, + { + "epoch": 0.7770534550195567, + "grad_norm": 1.9136794648235018, + "learning_rate": 7.925090836407997e-05, + "loss": 0.8539, + "step": 149 + }, + { + "epoch": 0.7822685788787483, + "grad_norm": 1.3862902439859852, + "learning_rate": 7.922246687750341e-05, + "loss": 0.8511, + "step": 150 + }, + { + "epoch": 0.78748370273794, + "grad_norm": 2.238512517144782, + "learning_rate": 7.919350076730836e-05, + "loss": 0.846, + "step": 151 + }, + { + "epoch": 0.7926988265971316, + "grad_norm": 2.0835619443377134, + "learning_rate": 7.916401042093361e-05, + "loss": 0.8461, + "step": 152 + }, + { + "epoch": 0.7979139504563233, + "grad_norm": 1.8060011410999197, + "learning_rate": 7.913399623282997e-05, + "loss": 0.833, + "step": 153 + }, + { + "epoch": 0.803129074315515, + "grad_norm": 2.0703223704187357, + "learning_rate": 7.910345860445487e-05, + "loss": 0.8447, + "step": 154 + }, + { + "epoch": 0.8083441981747066, + "grad_norm": 2.299135311289539, + "learning_rate": 7.90723979442671e-05, + "loss": 0.8533, + "step": 155 + }, + { + "epoch": 0.8135593220338984, + "grad_norm": 1.8730518820522357, + "learning_rate": 7.90408146677213e-05, + "loss": 0.8372, + "step": 156 + }, + { + "epoch": 0.81877444589309, + "grad_norm": 1.5402936247797026, + "learning_rate": 7.900870919726244e-05, + "loss": 0.8408, + "step": 157 + }, + { + "epoch": 0.8239895697522817, + "grad_norm": 2.378224013864058, + "learning_rate": 7.897608196232007e-05, + "loss": 0.8492, + "step": 158 + }, + { + "epoch": 0.8292046936114733, + "grad_norm": 1.8241375846314414, + "learning_rate": 7.894293339930272e-05, + "loss": 0.8338, + "step": 159 + }, + { + "epoch": 0.834419817470665, + "grad_norm": 2.2711975231391826, + "learning_rate": 7.890926395159197e-05, + "loss": 0.8385, + "step": 160 + }, + { + "epoch": 0.8396349413298566, + "grad_norm": 1.6195285561704085, + "learning_rate": 7.887507406953651e-05, + "loss": 0.8489, + "step": 161 + }, + { + "epoch": 0.8448500651890483, + "grad_norm": 2.2453846392163768, + "learning_rate": 7.884036421044618e-05, + "loss": 0.8487, + "step": 162 + }, + { + "epoch": 0.8500651890482399, + "grad_norm": 1.940408392169657, + "learning_rate": 7.880513483858583e-05, + "loss": 0.8398, + "step": 163 + }, + { + "epoch": 0.8552803129074316, + "grad_norm": 2.2159286823222675, + "learning_rate": 7.876938642516905e-05, + "loss": 0.8492, + "step": 164 + }, + { + "epoch": 0.8604954367666232, + "grad_norm": 1.4842300453687591, + "learning_rate": 7.873311944835195e-05, + "loss": 0.8376, + "step": 165 + }, + { + "epoch": 0.8657105606258149, + "grad_norm": 2.491755340230041, + "learning_rate": 7.869633439322674e-05, + "loss": 0.8386, + "step": 166 + }, + { + "epoch": 0.8709256844850065, + "grad_norm": 1.8127014745278216, + "learning_rate": 7.865903175181521e-05, + "loss": 0.8318, + "step": 167 + }, + { + "epoch": 0.8761408083441982, + "grad_norm": 1.5105514574695558, + "learning_rate": 7.862121202306217e-05, + "loss": 0.8317, + "step": 168 + }, + { + "epoch": 0.8813559322033898, + "grad_norm": 2.485520131584514, + "learning_rate": 7.858287571282882e-05, + "loss": 0.8496, + "step": 169 + }, + { + "epoch": 0.8865710560625815, + "grad_norm": 1.7516963763143856, + "learning_rate": 7.854402333388587e-05, + "loss": 0.8433, + "step": 170 + }, + { + "epoch": 0.8917861799217731, + "grad_norm": 2.462296826145561, + "learning_rate": 7.850465540590684e-05, + "loss": 0.8537, + "step": 171 + }, + { + "epoch": 0.8970013037809648, + "grad_norm": 1.1958019189200906, + "learning_rate": 7.846477245546094e-05, + "loss": 0.8323, + "step": 172 + }, + { + "epoch": 0.9022164276401564, + "grad_norm": 2.6026749193200693, + "learning_rate": 7.842437501600616e-05, + "loss": 0.8516, + "step": 173 + }, + { + "epoch": 0.9074315514993481, + "grad_norm": 1.7619131106705077, + "learning_rate": 7.838346362788206e-05, + "loss": 0.838, + "step": 174 + }, + { + "epoch": 0.9126466753585397, + "grad_norm": 2.4386182828768312, + "learning_rate": 7.834203883830259e-05, + "loss": 0.8444, + "step": 175 + }, + { + "epoch": 0.9178617992177314, + "grad_norm": 2.0969277412774274, + "learning_rate": 7.830010120134873e-05, + "loss": 0.8622, + "step": 176 + }, + { + "epoch": 0.9230769230769231, + "grad_norm": 1.4692953467652599, + "learning_rate": 7.825765127796108e-05, + "loss": 0.8338, + "step": 177 + }, + { + "epoch": 0.9282920469361148, + "grad_norm": 1.5923417027165594, + "learning_rate": 7.821468963593242e-05, + "loss": 0.8391, + "step": 178 + }, + { + "epoch": 0.9335071707953064, + "grad_norm": 2.2009607061094942, + "learning_rate": 7.817121684990004e-05, + "loss": 0.8589, + "step": 179 + }, + { + "epoch": 0.9387222946544981, + "grad_norm": 1.5116758680331852, + "learning_rate": 7.812723350133805e-05, + "loss": 0.8341, + "step": 180 + }, + { + "epoch": 0.9439374185136897, + "grad_norm": 2.038874732993399, + "learning_rate": 7.80827401785497e-05, + "loss": 0.8547, + "step": 181 + }, + { + "epoch": 0.9491525423728814, + "grad_norm": 2.2985530062759514, + "learning_rate": 7.80377374766594e-05, + "loss": 0.8391, + "step": 182 + }, + { + "epoch": 0.954367666232073, + "grad_norm": 1.257017121878146, + "learning_rate": 7.799222599760481e-05, + "loss": 0.8325, + "step": 183 + }, + { + "epoch": 0.9595827900912647, + "grad_norm": 2.873446007386103, + "learning_rate": 7.794620635012883e-05, + "loss": 0.853, + "step": 184 + }, + { + "epoch": 0.9647979139504563, + "grad_norm": 2.0836694911738207, + "learning_rate": 7.789967914977134e-05, + "loss": 0.8554, + "step": 185 + }, + { + "epoch": 0.970013037809648, + "grad_norm": 2.095712201456811, + "learning_rate": 7.785264501886108e-05, + "loss": 0.8439, + "step": 186 + }, + { + "epoch": 0.9752281616688396, + "grad_norm": 1.668972201159559, + "learning_rate": 7.78051045865073e-05, + "loss": 0.8442, + "step": 187 + }, + { + "epoch": 0.9804432855280313, + "grad_norm": 2.6548529934832787, + "learning_rate": 7.77570584885913e-05, + "loss": 0.8509, + "step": 188 + }, + { + "epoch": 0.9856584093872229, + "grad_norm": 1.6440679506158635, + "learning_rate": 7.770850736775796e-05, + "loss": 0.8367, + "step": 189 + }, + { + "epoch": 0.9908735332464146, + "grad_norm": 2.895095407046929, + "learning_rate": 7.765945187340715e-05, + "loss": 0.8476, + "step": 190 + }, + { + "epoch": 0.9960886571056062, + "grad_norm": 2.1880375444766655, + "learning_rate": 7.760989266168503e-05, + "loss": 0.8466, + "step": 191 + }, + { + "epoch": 1.001303780964798, + "grad_norm": 2.491373495658068, + "learning_rate": 7.755983039547528e-05, + "loss": 1.0483, + "step": 192 + }, + { + "epoch": 1.0065189048239895, + "grad_norm": 2.1065510338703195, + "learning_rate": 7.750926574439019e-05, + "loss": 0.8317, + "step": 193 + }, + { + "epoch": 1.0117340286831813, + "grad_norm": 1.6048527323347803, + "learning_rate": 7.745819938476184e-05, + "loss": 0.8309, + "step": 194 + }, + { + "epoch": 1.0169491525423728, + "grad_norm": 1.3686572978259008, + "learning_rate": 7.740663199963284e-05, + "loss": 0.8207, + "step": 195 + }, + { + "epoch": 1.0221642764015646, + "grad_norm": 1.6395534429827354, + "learning_rate": 7.73545642787474e-05, + "loss": 0.824, + "step": 196 + }, + { + "epoch": 1.0273794002607561, + "grad_norm": 1.829519017394267, + "learning_rate": 7.730199691854198e-05, + "loss": 0.8346, + "step": 197 + }, + { + "epoch": 1.032594524119948, + "grad_norm": 2.086222113236519, + "learning_rate": 7.724893062213602e-05, + "loss": 0.8204, + "step": 198 + }, + { + "epoch": 1.0378096479791394, + "grad_norm": 1.6543925285010792, + "learning_rate": 7.71953660993225e-05, + "loss": 0.8238, + "step": 199 + }, + { + "epoch": 1.0430247718383312, + "grad_norm": 1.6861891172176104, + "learning_rate": 7.71413040665585e-05, + "loss": 0.8215, + "step": 200 + }, + { + "epoch": 1.0482398956975227, + "grad_norm": 2.313274305017442, + "learning_rate": 7.708674524695559e-05, + "loss": 0.8082, + "step": 201 + }, + { + "epoch": 1.0534550195567145, + "grad_norm": 1.8334252249650502, + "learning_rate": 7.703169037027014e-05, + "loss": 0.824, + "step": 202 + }, + { + "epoch": 1.058670143415906, + "grad_norm": 1.7126012022750563, + "learning_rate": 7.697614017289357e-05, + "loss": 0.8105, + "step": 203 + }, + { + "epoch": 1.0638852672750978, + "grad_norm": 1.3173869406989056, + "learning_rate": 7.692009539784255e-05, + "loss": 0.8303, + "step": 204 + }, + { + "epoch": 1.0691003911342893, + "grad_norm": 2.0262932335464012, + "learning_rate": 7.686355679474898e-05, + "loss": 0.8197, + "step": 205 + }, + { + "epoch": 1.074315514993481, + "grad_norm": 2.2714312463405397, + "learning_rate": 7.680652511985e-05, + "loss": 0.8196, + "step": 206 + }, + { + "epoch": 1.0795306388526726, + "grad_norm": 1.5002234528402432, + "learning_rate": 7.674900113597787e-05, + "loss": 0.8141, + "step": 207 + }, + { + "epoch": 1.0847457627118644, + "grad_norm": 2.059909369219652, + "learning_rate": 7.669098561254983e-05, + "loss": 0.826, + "step": 208 + }, + { + "epoch": 1.0899608865710562, + "grad_norm": 1.6415840594311673, + "learning_rate": 7.663247932555767e-05, + "loss": 0.8314, + "step": 209 + }, + { + "epoch": 1.0951760104302477, + "grad_norm": 2.094580841761285, + "learning_rate": 7.65734830575575e-05, + "loss": 0.8405, + "step": 210 + }, + { + "epoch": 1.1003911342894395, + "grad_norm": 1.4276266650819935, + "learning_rate": 7.651399759765915e-05, + "loss": 0.8172, + "step": 211 + }, + { + "epoch": 1.105606258148631, + "grad_norm": 2.406763290618582, + "learning_rate": 7.645402374151575e-05, + "loss": 0.8166, + "step": 212 + }, + { + "epoch": 1.1108213820078228, + "grad_norm": 1.4071963692309266, + "learning_rate": 7.639356229131298e-05, + "loss": 0.8252, + "step": 213 + }, + { + "epoch": 1.1160365058670143, + "grad_norm": 1.434300709398784, + "learning_rate": 7.633261405575838e-05, + "loss": 0.8342, + "step": 214 + }, + { + "epoch": 1.121251629726206, + "grad_norm": 2.6026163076276223, + "learning_rate": 7.627117985007052e-05, + "loss": 0.8277, + "step": 215 + }, + { + "epoch": 1.1264667535853976, + "grad_norm": 1.545445917096449, + "learning_rate": 7.620926049596814e-05, + "loss": 0.8224, + "step": 216 + }, + { + "epoch": 1.1316818774445894, + "grad_norm": 2.349483330688553, + "learning_rate": 7.61468568216591e-05, + "loss": 0.8317, + "step": 217 + }, + { + "epoch": 1.136897001303781, + "grad_norm": 1.6493926282653744, + "learning_rate": 7.60839696618293e-05, + "loss": 0.8235, + "step": 218 + }, + { + "epoch": 1.1421121251629727, + "grad_norm": 2.523447089439544, + "learning_rate": 7.602059985763165e-05, + "loss": 0.8259, + "step": 219 + }, + { + "epoch": 1.1473272490221642, + "grad_norm": 1.910781874766756, + "learning_rate": 7.595674825667457e-05, + "loss": 0.826, + "step": 220 + }, + { + "epoch": 1.152542372881356, + "grad_norm": 2.2270989973834032, + "learning_rate": 7.589241571301091e-05, + "loss": 0.8309, + "step": 221 + }, + { + "epoch": 1.1577574967405475, + "grad_norm": 1.7934683365685906, + "learning_rate": 7.582760308712634e-05, + "loss": 0.8302, + "step": 222 + }, + { + "epoch": 1.1629726205997393, + "grad_norm": 2.0662257630292062, + "learning_rate": 7.5762311245928e-05, + "loss": 0.8277, + "step": 223 + }, + { + "epoch": 1.1681877444589308, + "grad_norm": 1.6790104250183842, + "learning_rate": 7.569654106273268e-05, + "loss": 0.8254, + "step": 224 + }, + { + "epoch": 1.1734028683181226, + "grad_norm": 2.11893755909352, + "learning_rate": 7.563029341725541e-05, + "loss": 0.8183, + "step": 225 + }, + { + "epoch": 1.1786179921773141, + "grad_norm": 1.5877658958807297, + "learning_rate": 7.55635691955975e-05, + "loss": 0.8232, + "step": 226 + }, + { + "epoch": 1.1838331160365059, + "grad_norm": 2.147915736111546, + "learning_rate": 7.549636929023471e-05, + "loss": 0.8185, + "step": 227 + }, + { + "epoch": 1.1890482398956976, + "grad_norm": 1.7400398545071791, + "learning_rate": 7.542869460000544e-05, + "loss": 0.8219, + "step": 228 + }, + { + "epoch": 1.1942633637548892, + "grad_norm": 1.8017174411766266, + "learning_rate": 7.536054603009856e-05, + "loss": 0.8099, + "step": 229 + }, + { + "epoch": 1.1994784876140807, + "grad_norm": 1.7295101992953383, + "learning_rate": 7.529192449204137e-05, + "loss": 0.8199, + "step": 230 + }, + { + "epoch": 1.2046936114732725, + "grad_norm": 1.81480851347367, + "learning_rate": 7.522283090368739e-05, + "loss": 0.8163, + "step": 231 + }, + { + "epoch": 1.2099087353324642, + "grad_norm": 1.4532183826025098, + "learning_rate": 7.515326618920409e-05, + "loss": 0.8243, + "step": 232 + }, + { + "epoch": 1.2151238591916558, + "grad_norm": 2.025221417951849, + "learning_rate": 7.508323127906055e-05, + "loss": 0.8104, + "step": 233 + }, + { + "epoch": 1.2203389830508475, + "grad_norm": 2.839580655118928, + "learning_rate": 7.5012727110015e-05, + "loss": 0.8156, + "step": 234 + }, + { + "epoch": 1.225554106910039, + "grad_norm": 0.9612813608575648, + "learning_rate": 7.494175462510225e-05, + "loss": 0.8136, + "step": 235 + }, + { + "epoch": 1.2307692307692308, + "grad_norm": 4.580627208479269, + "learning_rate": 7.487031477362112e-05, + "loss": 0.8363, + "step": 236 + }, + { + "epoch": 1.2359843546284224, + "grad_norm": 3.3062212088133585, + "learning_rate": 7.479840851112175e-05, + "loss": 0.8457, + "step": 237 + }, + { + "epoch": 1.2411994784876141, + "grad_norm": 3.9085448168352577, + "learning_rate": 7.47260367993928e-05, + "loss": 0.8369, + "step": 238 + }, + { + "epoch": 1.2464146023468057, + "grad_norm": 3.2990078411550416, + "learning_rate": 7.465320060644857e-05, + "loss": 0.83, + "step": 239 + }, + { + "epoch": 1.2516297262059974, + "grad_norm": 3.171914889017843, + "learning_rate": 7.45799009065161e-05, + "loss": 0.8245, + "step": 240 + }, + { + "epoch": 1.256844850065189, + "grad_norm": 2.6816905524834436, + "learning_rate": 7.450613868002208e-05, + "loss": 0.8363, + "step": 241 + }, + { + "epoch": 1.2620599739243807, + "grad_norm": 3.0890905030641256, + "learning_rate": 7.443191491357976e-05, + "loss": 0.823, + "step": 242 + }, + { + "epoch": 1.2672750977835723, + "grad_norm": 2.3490265350478956, + "learning_rate": 7.435723059997581e-05, + "loss": 0.8276, + "step": 243 + }, + { + "epoch": 1.272490221642764, + "grad_norm": 3.5496378180168957, + "learning_rate": 7.428208673815693e-05, + "loss": 0.8247, + "step": 244 + }, + { + "epoch": 1.2777053455019556, + "grad_norm": 3.053170271456407, + "learning_rate": 7.420648433321659e-05, + "loss": 0.8306, + "step": 245 + }, + { + "epoch": 1.2829204693611473, + "grad_norm": 2.387063335590222, + "learning_rate": 7.41304243963815e-05, + "loss": 0.831, + "step": 246 + }, + { + "epoch": 1.288135593220339, + "grad_norm": 1.8497576227055157, + "learning_rate": 7.405390794499819e-05, + "loss": 0.8066, + "step": 247 + }, + { + "epoch": 1.2933507170795306, + "grad_norm": 2.9897041114580727, + "learning_rate": 7.397693600251929e-05, + "loss": 0.8134, + "step": 248 + }, + { + "epoch": 1.2985658409387222, + "grad_norm": 1.980452439525295, + "learning_rate": 7.389950959848992e-05, + "loss": 0.8252, + "step": 249 + }, + { + "epoch": 1.303780964797914, + "grad_norm": 3.4714237886205748, + "learning_rate": 7.382162976853387e-05, + "loss": 0.8294, + "step": 250 + }, + { + "epoch": 1.3089960886571057, + "grad_norm": 3.2963021601322535, + "learning_rate": 7.37432975543398e-05, + "loss": 0.8232, + "step": 251 + }, + { + "epoch": 1.3142112125162972, + "grad_norm": 2.1372765339871806, + "learning_rate": 7.366451400364723e-05, + "loss": 0.8228, + "step": 252 + }, + { + "epoch": 1.3194263363754888, + "grad_norm": 1.61329278641058, + "learning_rate": 7.358528017023262e-05, + "loss": 0.8149, + "step": 253 + }, + { + "epoch": 1.3246414602346805, + "grad_norm": 2.149446564090082, + "learning_rate": 7.350559711389518e-05, + "loss": 0.803, + "step": 254 + }, + { + "epoch": 1.3298565840938723, + "grad_norm": 1.2080706850732208, + "learning_rate": 7.342546590044279e-05, + "loss": 0.8309, + "step": 255 + }, + { + "epoch": 1.3350717079530638, + "grad_norm": 2.060776426075165, + "learning_rate": 7.334488760167768e-05, + "loss": 0.8218, + "step": 256 + }, + { + "epoch": 1.3402868318122556, + "grad_norm": 1.4270579486920896, + "learning_rate": 7.326386329538207e-05, + "loss": 0.8482, + "step": 257 + }, + { + "epoch": 1.3455019556714471, + "grad_norm": 2.320654966360905, + "learning_rate": 7.318239406530386e-05, + "loss": 0.8284, + "step": 258 + }, + { + "epoch": 1.350717079530639, + "grad_norm": 2.040787679253322, + "learning_rate": 7.3100481001142e-05, + "loss": 0.8436, + "step": 259 + }, + { + "epoch": 1.3559322033898304, + "grad_norm": 1.450733455499567, + "learning_rate": 7.301812519853203e-05, + "loss": 0.8067, + "step": 260 + }, + { + "epoch": 1.3611473272490222, + "grad_norm": 1.4473957425879223, + "learning_rate": 7.293532775903137e-05, + "loss": 0.8172, + "step": 261 + }, + { + "epoch": 1.3663624511082137, + "grad_norm": 1.4567295321246068, + "learning_rate": 7.285208979010458e-05, + "loss": 0.833, + "step": 262 + }, + { + "epoch": 1.3715775749674055, + "grad_norm": 2.2303402593920234, + "learning_rate": 7.276841240510858e-05, + "loss": 0.8241, + "step": 263 + }, + { + "epoch": 1.376792698826597, + "grad_norm": 1.3757747352916987, + "learning_rate": 7.26842967232777e-05, + "loss": 0.8106, + "step": 264 + }, + { + "epoch": 1.3820078226857888, + "grad_norm": 1.960651210566042, + "learning_rate": 7.25997438697088e-05, + "loss": 0.8258, + "step": 265 + }, + { + "epoch": 1.3872229465449806, + "grad_norm": 1.6973151493940888, + "learning_rate": 7.251475497534615e-05, + "loss": 0.8421, + "step": 266 + }, + { + "epoch": 1.3924380704041721, + "grad_norm": 1.4278961151807503, + "learning_rate": 7.242933117696628e-05, + "loss": 0.8107, + "step": 267 + }, + { + "epoch": 1.3976531942633637, + "grad_norm": 1.6765492543318952, + "learning_rate": 7.234347361716291e-05, + "loss": 0.8126, + "step": 268 + }, + { + "epoch": 1.4028683181225554, + "grad_norm": 1.7901534079828416, + "learning_rate": 7.225718344433149e-05, + "loss": 0.8209, + "step": 269 + }, + { + "epoch": 1.4080834419817472, + "grad_norm": 1.1879344799911462, + "learning_rate": 7.217046181265394e-05, + "loss": 0.8241, + "step": 270 + }, + { + "epoch": 1.4132985658409387, + "grad_norm": 2.316592389288362, + "learning_rate": 7.208330988208324e-05, + "loss": 0.8233, + "step": 271 + }, + { + "epoch": 1.4185136897001303, + "grad_norm": 1.3009721963229537, + "learning_rate": 7.199572881832784e-05, + "loss": 0.8094, + "step": 272 + }, + { + "epoch": 1.423728813559322, + "grad_norm": 2.068602624965321, + "learning_rate": 7.190771979283608e-05, + "loss": 0.8221, + "step": 273 + }, + { + "epoch": 1.4289439374185138, + "grad_norm": 1.9940233989402643, + "learning_rate": 7.181928398278058e-05, + "loss": 0.8225, + "step": 274 + }, + { + "epoch": 1.4341590612777053, + "grad_norm": 1.7328843255622977, + "learning_rate": 7.173042257104243e-05, + "loss": 0.8142, + "step": 275 + }, + { + "epoch": 1.439374185136897, + "grad_norm": 1.5984687241828228, + "learning_rate": 7.164113674619542e-05, + "loss": 0.8062, + "step": 276 + }, + { + "epoch": 1.4445893089960886, + "grad_norm": 1.9941450850887774, + "learning_rate": 7.155142770249008e-05, + "loss": 0.8156, + "step": 277 + }, + { + "epoch": 1.4498044328552804, + "grad_norm": 1.2629214490257985, + "learning_rate": 7.146129663983775e-05, + "loss": 0.8029, + "step": 278 + }, + { + "epoch": 1.455019556714472, + "grad_norm": 1.5341743253202347, + "learning_rate": 7.137074476379454e-05, + "loss": 0.8184, + "step": 279 + }, + { + "epoch": 1.4602346805736637, + "grad_norm": 1.205052243506182, + "learning_rate": 7.127977328554518e-05, + "loss": 0.8297, + "step": 280 + }, + { + "epoch": 1.4654498044328552, + "grad_norm": 1.7584530348059961, + "learning_rate": 7.118838342188683e-05, + "loss": 0.8183, + "step": 281 + }, + { + "epoch": 1.470664928292047, + "grad_norm": 1.8223656535526023, + "learning_rate": 7.10965763952128e-05, + "loss": 0.8154, + "step": 282 + }, + { + "epoch": 1.4758800521512385, + "grad_norm": 1.3665387213451412, + "learning_rate": 7.100435343349617e-05, + "loss": 0.8114, + "step": 283 + }, + { + "epoch": 1.4810951760104303, + "grad_norm": 1.499553247270366, + "learning_rate": 7.091171577027344e-05, + "loss": 0.813, + "step": 284 + }, + { + "epoch": 1.4863102998696218, + "grad_norm": 2.1007419745906866, + "learning_rate": 7.081866464462798e-05, + "loss": 0.8216, + "step": 285 + }, + { + "epoch": 1.4915254237288136, + "grad_norm": 1.512999754469945, + "learning_rate": 7.072520130117344e-05, + "loss": 0.8182, + "step": 286 + }, + { + "epoch": 1.4967405475880051, + "grad_norm": 1.6975647047303697, + "learning_rate": 7.063132699003716e-05, + "loss": 0.8244, + "step": 287 + }, + { + "epoch": 1.5019556714471969, + "grad_norm": 1.7957380356125672, + "learning_rate": 7.053704296684337e-05, + "loss": 0.8162, + "step": 288 + }, + { + "epoch": 1.5071707953063886, + "grad_norm": 1.2813149409377043, + "learning_rate": 7.044235049269649e-05, + "loss": 0.8095, + "step": 289 + }, + { + "epoch": 1.5123859191655802, + "grad_norm": 1.56534754361993, + "learning_rate": 7.034725083416419e-05, + "loss": 0.8258, + "step": 290 + }, + { + "epoch": 1.5176010430247717, + "grad_norm": 1.7975943131313281, + "learning_rate": 7.025174526326045e-05, + "loss": 0.8025, + "step": 291 + }, + { + "epoch": 1.5228161668839635, + "grad_norm": 1.1853911402078072, + "learning_rate": 7.015583505742857e-05, + "loss": 0.8249, + "step": 292 + }, + { + "epoch": 1.5280312907431552, + "grad_norm": 1.8494955443278849, + "learning_rate": 7.005952149952416e-05, + "loss": 0.8378, + "step": 293 + }, + { + "epoch": 1.5332464146023468, + "grad_norm": 1.3622849536120247, + "learning_rate": 6.996280587779778e-05, + "loss": 0.8354, + "step": 294 + }, + { + "epoch": 1.5384615384615383, + "grad_norm": 2.0112831346634708, + "learning_rate": 6.986568948587792e-05, + "loss": 0.83, + "step": 295 + }, + { + "epoch": 1.54367666232073, + "grad_norm": 1.5420042484474124, + "learning_rate": 6.976817362275357e-05, + "loss": 0.8109, + "step": 296 + }, + { + "epoch": 1.5488917861799218, + "grad_norm": 2.0929864944342804, + "learning_rate": 6.96702595927569e-05, + "loss": 0.8334, + "step": 297 + }, + { + "epoch": 1.5541069100391134, + "grad_norm": 1.8922406476791735, + "learning_rate": 6.957194870554578e-05, + "loss": 0.806, + "step": 298 + }, + { + "epoch": 1.559322033898305, + "grad_norm": 1.6880868554685944, + "learning_rate": 6.947324227608628e-05, + "loss": 0.8212, + "step": 299 + }, + { + "epoch": 1.5645371577574967, + "grad_norm": 1.6346451668072235, + "learning_rate": 6.937414162463509e-05, + "loss": 0.8014, + "step": 300 + }, + { + "epoch": 1.5697522816166884, + "grad_norm": 1.577939054921778, + "learning_rate": 6.927464807672186e-05, + "loss": 0.8187, + "step": 301 + }, + { + "epoch": 1.57496740547588, + "grad_norm": 1.4316029714279948, + "learning_rate": 6.917476296313145e-05, + "loss": 0.8046, + "step": 302 + }, + { + "epoch": 1.5801825293350718, + "grad_norm": 1.5381955264209675, + "learning_rate": 6.907448761988612e-05, + "loss": 0.8077, + "step": 303 + }, + { + "epoch": 1.5853976531942635, + "grad_norm": 1.0983288794814963, + "learning_rate": 6.897382338822772e-05, + "loss": 0.804, + "step": 304 + }, + { + "epoch": 1.590612777053455, + "grad_norm": 1.2952124244327796, + "learning_rate": 6.88727716145997e-05, + "loss": 0.8014, + "step": 305 + }, + { + "epoch": 1.5958279009126466, + "grad_norm": 1.1248306945768556, + "learning_rate": 6.877133365062911e-05, + "loss": 0.8086, + "step": 306 + }, + { + "epoch": 1.6010430247718384, + "grad_norm": 1.7722044591752057, + "learning_rate": 6.86695108531085e-05, + "loss": 0.8061, + "step": 307 + }, + { + "epoch": 1.6062581486310301, + "grad_norm": 1.8332828396584284, + "learning_rate": 6.856730458397787e-05, + "loss": 0.8177, + "step": 308 + }, + { + "epoch": 1.6114732724902217, + "grad_norm": 1.3222660990882034, + "learning_rate": 6.846471621030626e-05, + "loss": 0.8012, + "step": 309 + }, + { + "epoch": 1.6166883963494132, + "grad_norm": 1.3548943859194524, + "learning_rate": 6.836174710427369e-05, + "loss": 0.8171, + "step": 310 + }, + { + "epoch": 1.621903520208605, + "grad_norm": 1.2347347259073755, + "learning_rate": 6.825839864315264e-05, + "loss": 0.7839, + "step": 311 + }, + { + "epoch": 1.6271186440677967, + "grad_norm": 1.4912678731047493, + "learning_rate": 6.815467220928972e-05, + "loss": 0.8004, + "step": 312 + }, + { + "epoch": 1.6323337679269883, + "grad_norm": 1.0089833155389005, + "learning_rate": 6.805056919008714e-05, + "loss": 0.806, + "step": 313 + }, + { + "epoch": 1.6375488917861798, + "grad_norm": 1.8923057463186084, + "learning_rate": 6.794609097798414e-05, + "loss": 0.8149, + "step": 314 + }, + { + "epoch": 1.6427640156453716, + "grad_norm": 1.3911432602736478, + "learning_rate": 6.784123897043841e-05, + "loss": 0.8261, + "step": 315 + }, + { + "epoch": 1.6479791395045633, + "grad_norm": 1.2299562475949994, + "learning_rate": 6.773601456990739e-05, + "loss": 0.8025, + "step": 316 + }, + { + "epoch": 1.6531942633637549, + "grad_norm": 1.749738991866774, + "learning_rate": 6.763041918382945e-05, + "loss": 0.8087, + "step": 317 + }, + { + "epoch": 1.6584093872229464, + "grad_norm": 1.2129819476495074, + "learning_rate": 6.752445422460513e-05, + "loss": 0.8058, + "step": 318 + }, + { + "epoch": 1.6636245110821382, + "grad_norm": 1.9218816457335506, + "learning_rate": 6.741812110957823e-05, + "loss": 0.8199, + "step": 319 + }, + { + "epoch": 1.66883963494133, + "grad_norm": 1.5263095695339857, + "learning_rate": 6.731142126101688e-05, + "loss": 0.8098, + "step": 320 + }, + { + "epoch": 1.6740547588005215, + "grad_norm": 1.7388018103606915, + "learning_rate": 6.720435610609443e-05, + "loss": 0.7964, + "step": 321 + }, + { + "epoch": 1.6792698826597132, + "grad_norm": 1.1143428753117783, + "learning_rate": 6.709692707687047e-05, + "loss": 0.8026, + "step": 322 + }, + { + "epoch": 1.684485006518905, + "grad_norm": 1.4518152074151873, + "learning_rate": 6.69891356102716e-05, + "loss": 0.8199, + "step": 323 + }, + { + "epoch": 1.6897001303780965, + "grad_norm": 1.2302866722544838, + "learning_rate": 6.688098314807221e-05, + "loss": 0.8116, + "step": 324 + }, + { + "epoch": 1.694915254237288, + "grad_norm": 1.355915097493558, + "learning_rate": 6.677247113687527e-05, + "loss": 0.8184, + "step": 325 + }, + { + "epoch": 1.7001303780964798, + "grad_norm": 1.0973851180923702, + "learning_rate": 6.666360102809289e-05, + "loss": 0.8066, + "step": 326 + }, + { + "epoch": 1.7053455019556716, + "grad_norm": 2.034989183783714, + "learning_rate": 6.655437427792698e-05, + "loss": 0.8068, + "step": 327 + }, + { + "epoch": 1.7105606258148631, + "grad_norm": 1.352963702330486, + "learning_rate": 6.644479234734971e-05, + "loss": 0.8337, + "step": 328 + }, + { + "epoch": 1.7157757496740547, + "grad_norm": 1.486986240165971, + "learning_rate": 6.6334856702084e-05, + "loss": 0.8142, + "step": 329 + }, + { + "epoch": 1.7209908735332464, + "grad_norm": 1.2755542839313279, + "learning_rate": 6.622456881258392e-05, + "loss": 0.8224, + "step": 330 + }, + { + "epoch": 1.7262059973924382, + "grad_norm": 1.4619956893213686, + "learning_rate": 6.6113930154015e-05, + "loss": 0.8077, + "step": 331 + }, + { + "epoch": 1.7314211212516297, + "grad_norm": 1.7136751212219052, + "learning_rate": 6.600294220623457e-05, + "loss": 0.8089, + "step": 332 + }, + { + "epoch": 1.7366362451108213, + "grad_norm": 1.1691508723533794, + "learning_rate": 6.589160645377181e-05, + "loss": 0.8192, + "step": 333 + }, + { + "epoch": 1.741851368970013, + "grad_norm": 2.2934106958814544, + "learning_rate": 6.57799243858081e-05, + "loss": 0.8123, + "step": 334 + }, + { + "epoch": 1.7470664928292048, + "grad_norm": 1.3479467101781495, + "learning_rate": 6.566789749615691e-05, + "loss": 0.8016, + "step": 335 + }, + { + "epoch": 1.7522816166883963, + "grad_norm": 1.4715014206564454, + "learning_rate": 6.555552728324394e-05, + "loss": 0.8135, + "step": 336 + }, + { + "epoch": 1.7574967405475879, + "grad_norm": 2.4679525667464155, + "learning_rate": 6.544281525008703e-05, + "loss": 0.8047, + "step": 337 + }, + { + "epoch": 1.7627118644067796, + "grad_norm": 1.2701148601075805, + "learning_rate": 6.532976290427611e-05, + "loss": 0.8155, + "step": 338 + }, + { + "epoch": 1.7679269882659714, + "grad_norm": 2.7926000172133967, + "learning_rate": 6.521637175795292e-05, + "loss": 0.8114, + "step": 339 + }, + { + "epoch": 1.773142112125163, + "grad_norm": 1.8634719780362607, + "learning_rate": 6.51026433277909e-05, + "loss": 0.8279, + "step": 340 + }, + { + "epoch": 1.7783572359843545, + "grad_norm": 2.527061321023331, + "learning_rate": 6.498857913497485e-05, + "loss": 0.8308, + "step": 341 + }, + { + "epoch": 1.7835723598435462, + "grad_norm": 1.7861432214936908, + "learning_rate": 6.487418070518063e-05, + "loss": 0.8237, + "step": 342 + }, + { + "epoch": 1.788787483702738, + "grad_norm": 2.4741757997509812, + "learning_rate": 6.475944956855463e-05, + "loss": 0.8098, + "step": 343 + }, + { + "epoch": 1.7940026075619295, + "grad_norm": 2.5470632776776583, + "learning_rate": 6.464438725969348e-05, + "loss": 0.8153, + "step": 344 + }, + { + "epoch": 1.7992177314211213, + "grad_norm": 1.4497598146752289, + "learning_rate": 6.452899531762338e-05, + "loss": 0.809, + "step": 345 + }, + { + "epoch": 1.804432855280313, + "grad_norm": 1.6671814059233867, + "learning_rate": 6.44132752857796e-05, + "loss": 0.8206, + "step": 346 + }, + { + "epoch": 1.8096479791395046, + "grad_norm": 1.4862285893517284, + "learning_rate": 6.429722871198579e-05, + "loss": 0.8125, + "step": 347 + }, + { + "epoch": 1.8148631029986961, + "grad_norm": 1.4717475694489375, + "learning_rate": 6.418085714843328e-05, + "loss": 0.8028, + "step": 348 + }, + { + "epoch": 1.820078226857888, + "grad_norm": 1.2961734030931862, + "learning_rate": 6.406416215166035e-05, + "loss": 0.814, + "step": 349 + }, + { + "epoch": 1.8252933507170797, + "grad_norm": 1.375416590864907, + "learning_rate": 6.394714528253137e-05, + "loss": 0.8084, + "step": 350 + }, + { + "epoch": 1.8305084745762712, + "grad_norm": 1.1939033449159728, + "learning_rate": 6.382980810621595e-05, + "loss": 0.8054, + "step": 351 + }, + { + "epoch": 1.8357235984354627, + "grad_norm": 1.905725265353936, + "learning_rate": 6.371215219216801e-05, + "loss": 0.799, + "step": 352 + }, + { + "epoch": 1.8409387222946545, + "grad_norm": 1.5504247264478312, + "learning_rate": 6.359417911410477e-05, + "loss": 0.8184, + "step": 353 + }, + { + "epoch": 1.8461538461538463, + "grad_norm": 1.034848478726046, + "learning_rate": 6.347589044998568e-05, + "loss": 0.8058, + "step": 354 + }, + { + "epoch": 1.8513689700130378, + "grad_norm": 2.3826958255372217, + "learning_rate": 6.335728778199139e-05, + "loss": 0.8066, + "step": 355 + }, + { + "epoch": 1.8565840938722293, + "grad_norm": 1.3959972285328377, + "learning_rate": 6.323837269650249e-05, + "loss": 0.8217, + "step": 356 + }, + { + "epoch": 1.861799217731421, + "grad_norm": 2.506085020659937, + "learning_rate": 6.311914678407837e-05, + "loss": 0.8216, + "step": 357 + }, + { + "epoch": 1.8670143415906129, + "grad_norm": 1.8279565139453056, + "learning_rate": 6.299961163943587e-05, + "loss": 0.8213, + "step": 358 + }, + { + "epoch": 1.8722294654498044, + "grad_norm": 2.0962030329324204, + "learning_rate": 6.287976886142806e-05, + "loss": 0.8206, + "step": 359 + }, + { + "epoch": 1.877444589308996, + "grad_norm": 1.70186031477118, + "learning_rate": 6.275962005302273e-05, + "loss": 0.8301, + "step": 360 + }, + { + "epoch": 1.8826597131681877, + "grad_norm": 2.1788285828805365, + "learning_rate": 6.263916682128104e-05, + "loss": 0.817, + "step": 361 + }, + { + "epoch": 1.8878748370273795, + "grad_norm": 1.805067718405396, + "learning_rate": 6.251841077733595e-05, + "loss": 0.8158, + "step": 362 + }, + { + "epoch": 1.893089960886571, + "grad_norm": 2.1305468630587066, + "learning_rate": 6.239735353637076e-05, + "loss": 0.8086, + "step": 363 + }, + { + "epoch": 1.8983050847457628, + "grad_norm": 1.8535876445861497, + "learning_rate": 6.227599671759745e-05, + "loss": 0.8088, + "step": 364 + }, + { + "epoch": 1.9035202086049545, + "grad_norm": 2.0851098014276914, + "learning_rate": 6.215434194423499e-05, + "loss": 0.8053, + "step": 365 + }, + { + "epoch": 1.908735332464146, + "grad_norm": 1.6617436320103747, + "learning_rate": 6.203239084348772e-05, + "loss": 0.8272, + "step": 366 + }, + { + "epoch": 1.9139504563233376, + "grad_norm": 1.9519893634609065, + "learning_rate": 6.191014504652352e-05, + "loss": 0.8052, + "step": 367 + }, + { + "epoch": 1.9191655801825294, + "grad_norm": 1.8496089676435563, + "learning_rate": 6.178760618845194e-05, + "loss": 0.8152, + "step": 368 + }, + { + "epoch": 1.9243807040417211, + "grad_norm": 1.5391906529413488, + "learning_rate": 6.166477590830252e-05, + "loss": 0.8129, + "step": 369 + }, + { + "epoch": 1.9295958279009127, + "grad_norm": 1.5353016752039255, + "learning_rate": 6.154165584900263e-05, + "loss": 0.7994, + "step": 370 + }, + { + "epoch": 1.9348109517601042, + "grad_norm": 1.4284130690464092, + "learning_rate": 6.141824765735567e-05, + "loss": 0.8169, + "step": 371 + }, + { + "epoch": 1.940026075619296, + "grad_norm": 1.3150434684077639, + "learning_rate": 6.129455298401894e-05, + "loss": 0.7936, + "step": 372 + }, + { + "epoch": 1.9452411994784877, + "grad_norm": 1.3747947345260623, + "learning_rate": 6.117057348348164e-05, + "loss": 0.8007, + "step": 373 + }, + { + "epoch": 1.9504563233376793, + "grad_norm": 1.0176646734905497, + "learning_rate": 6.104631081404269e-05, + "loss": 0.8085, + "step": 374 + }, + { + "epoch": 1.9556714471968708, + "grad_norm": 1.773293381624066, + "learning_rate": 6.092176663778851e-05, + "loss": 0.8075, + "step": 375 + }, + { + "epoch": 1.9608865710560626, + "grad_norm": 1.4739056885963973, + "learning_rate": 6.079694262057094e-05, + "loss": 0.8125, + "step": 376 + }, + { + "epoch": 1.9661016949152543, + "grad_norm": 1.6264911494594487, + "learning_rate": 6.067184043198476e-05, + "loss": 0.8093, + "step": 377 + }, + { + "epoch": 1.9713168187744459, + "grad_norm": 1.5269696160976443, + "learning_rate": 6.054646174534552e-05, + "loss": 0.8073, + "step": 378 + }, + { + "epoch": 1.9765319426336374, + "grad_norm": 1.4672155013368013, + "learning_rate": 6.0420808237667055e-05, + "loss": 0.7873, + "step": 379 + }, + { + "epoch": 1.9817470664928292, + "grad_norm": 1.6071190706840761, + "learning_rate": 6.029488158963912e-05, + "loss": 0.8185, + "step": 380 + }, + { + "epoch": 1.986962190352021, + "grad_norm": 1.1200471714942668, + "learning_rate": 6.016868348560488e-05, + "loss": 0.7945, + "step": 381 + }, + { + "epoch": 1.9921773142112125, + "grad_norm": 1.2021517615584545, + "learning_rate": 6.004221561353838e-05, + "loss": 0.8019, + "step": 382 + }, + { + "epoch": 1.997392438070404, + "grad_norm": 1.1953523539998745, + "learning_rate": 5.991547966502195e-05, + "loss": 0.8847, + "step": 383 + }, + { + "epoch": 2.002607561929596, + "grad_norm": 1.2139914038103075, + "learning_rate": 5.978847733522363e-05, + "loss": 0.9192, + "step": 384 + }, + { + "epoch": 2.0078226857887875, + "grad_norm": 1.2663014964519022, + "learning_rate": 5.9661210322874456e-05, + "loss": 0.7934, + "step": 385 + }, + { + "epoch": 2.013037809647979, + "grad_norm": 1.7536553812874152, + "learning_rate": 5.953368033024576e-05, + "loss": 0.7716, + "step": 386 + }, + { + "epoch": 2.0182529335071706, + "grad_norm": 1.3505679779762105, + "learning_rate": 5.940588906312636e-05, + "loss": 0.79, + "step": 387 + }, + { + "epoch": 2.0234680573663626, + "grad_norm": 1.100812193500458, + "learning_rate": 5.9277838230799816e-05, + "loss": 0.7831, + "step": 388 + }, + { + "epoch": 2.028683181225554, + "grad_norm": 1.312836314335182, + "learning_rate": 5.9149529546021486e-05, + "loss": 0.7833, + "step": 389 + }, + { + "epoch": 2.0338983050847457, + "grad_norm": 1.9463142850045887, + "learning_rate": 5.902096472499569e-05, + "loss": 0.7829, + "step": 390 + }, + { + "epoch": 2.039113428943937, + "grad_norm": 0.7033565907846064, + "learning_rate": 5.889214548735269e-05, + "loss": 0.7842, + "step": 391 + }, + { + "epoch": 2.044328552803129, + "grad_norm": 1.743510351487419, + "learning_rate": 5.876307355612575e-05, + "loss": 0.7679, + "step": 392 + }, + { + "epoch": 2.0495436766623207, + "grad_norm": 1.2465470038883548, + "learning_rate": 5.8633750657728033e-05, + "loss": 0.7712, + "step": 393 + }, + { + "epoch": 2.0547588005215123, + "grad_norm": 2.068655430465104, + "learning_rate": 5.850417852192956e-05, + "loss": 0.7853, + "step": 394 + }, + { + "epoch": 2.0599739243807043, + "grad_norm": 0.9560005505258112, + "learning_rate": 5.837435888183403e-05, + "loss": 0.785, + "step": 395 + }, + { + "epoch": 2.065189048239896, + "grad_norm": 1.7062918904685223, + "learning_rate": 5.8244293473855664e-05, + "loss": 0.7849, + "step": 396 + }, + { + "epoch": 2.0704041720990873, + "grad_norm": 1.623721715677207, + "learning_rate": 5.8113984037695984e-05, + "loss": 0.7857, + "step": 397 + }, + { + "epoch": 2.075619295958279, + "grad_norm": 1.484064136635307, + "learning_rate": 5.798343231632053e-05, + "loss": 0.7779, + "step": 398 + }, + { + "epoch": 2.080834419817471, + "grad_norm": 1.34676834959028, + "learning_rate": 5.785264005593553e-05, + "loss": 0.7934, + "step": 399 + }, + { + "epoch": 2.0860495436766624, + "grad_norm": 1.6199775628522408, + "learning_rate": 5.772160900596456e-05, + "loss": 0.7735, + "step": 400 + }, + { + "epoch": 2.091264667535854, + "grad_norm": 1.5660844096047937, + "learning_rate": 5.7590340919025204e-05, + "loss": 0.7833, + "step": 401 + }, + { + "epoch": 2.0964797913950455, + "grad_norm": 1.0421277103166564, + "learning_rate": 5.7458837550905486e-05, + "loss": 0.7837, + "step": 402 + }, + { + "epoch": 2.1016949152542375, + "grad_norm": 1.7458400592594279, + "learning_rate": 5.73271006605405e-05, + "loss": 0.7855, + "step": 403 + }, + { + "epoch": 2.106910039113429, + "grad_norm": 1.5186591541100996, + "learning_rate": 5.7195132009988814e-05, + "loss": 0.7853, + "step": 404 + }, + { + "epoch": 2.1121251629726205, + "grad_norm": 0.8719695266664192, + "learning_rate": 5.706293336440894e-05, + "loss": 0.7896, + "step": 405 + }, + { + "epoch": 2.117340286831812, + "grad_norm": 1.0314918237610309, + "learning_rate": 5.693050649203568e-05, + "loss": 0.7807, + "step": 406 + }, + { + "epoch": 2.122555410691004, + "grad_norm": 0.7603723547082539, + "learning_rate": 5.679785316415654e-05, + "loss": 0.7763, + "step": 407 + }, + { + "epoch": 2.1277705345501956, + "grad_norm": 0.9173514911262751, + "learning_rate": 5.6664975155088004e-05, + "loss": 0.7919, + "step": 408 + }, + { + "epoch": 2.132985658409387, + "grad_norm": 1.2215202346945515, + "learning_rate": 5.6531874242151746e-05, + "loss": 0.772, + "step": 409 + }, + { + "epoch": 2.1382007822685787, + "grad_norm": 2.368630198683683, + "learning_rate": 5.639855220565098e-05, + "loss": 0.801, + "step": 410 + }, + { + "epoch": 2.1434159061277707, + "grad_norm": 1.2419888378957813, + "learning_rate": 5.626501082884654e-05, + "loss": 0.7948, + "step": 411 + }, + { + "epoch": 2.148631029986962, + "grad_norm": 3.0547619405266446, + "learning_rate": 5.613125189793305e-05, + "loss": 0.7761, + "step": 412 + }, + { + "epoch": 2.1538461538461537, + "grad_norm": 2.1394143129005987, + "learning_rate": 5.5997277202015085e-05, + "loss": 0.7815, + "step": 413 + }, + { + "epoch": 2.1590612777053453, + "grad_norm": 3.041921943681318, + "learning_rate": 5.586308853308319e-05, + "loss": 0.7889, + "step": 414 + }, + { + "epoch": 2.1642764015645373, + "grad_norm": 2.4712686274458853, + "learning_rate": 5.572868768598993e-05, + "loss": 0.7764, + "step": 415 + }, + { + "epoch": 2.169491525423729, + "grad_norm": 2.5501423533072605, + "learning_rate": 5.5594076458425864e-05, + "loss": 0.7995, + "step": 416 + }, + { + "epoch": 2.1747066492829203, + "grad_norm": 2.2877612282195416, + "learning_rate": 5.545925665089552e-05, + "loss": 0.7882, + "step": 417 + }, + { + "epoch": 2.1799217731421123, + "grad_norm": 2.1483606998937144, + "learning_rate": 5.532423006669332e-05, + "loss": 0.7873, + "step": 418 + }, + { + "epoch": 2.185136897001304, + "grad_norm": 1.7462471775177655, + "learning_rate": 5.518899851187942e-05, + "loss": 0.7964, + "step": 419 + }, + { + "epoch": 2.1903520208604954, + "grad_norm": 2.756935578457596, + "learning_rate": 5.505356379525559e-05, + "loss": 0.7891, + "step": 420 + }, + { + "epoch": 2.195567144719687, + "grad_norm": 2.4106523309041314, + "learning_rate": 5.491792772834103e-05, + "loss": 0.7804, + "step": 421 + }, + { + "epoch": 2.200782268578879, + "grad_norm": 1.8981144967918866, + "learning_rate": 5.478209212534809e-05, + "loss": 0.7796, + "step": 422 + }, + { + "epoch": 2.2059973924380705, + "grad_norm": 1.487529019552184, + "learning_rate": 5.464605880315803e-05, + "loss": 0.7773, + "step": 423 + }, + { + "epoch": 2.211212516297262, + "grad_norm": 2.44109555313169, + "learning_rate": 5.4509829581296774e-05, + "loss": 0.777, + "step": 424 + }, + { + "epoch": 2.2164276401564535, + "grad_norm": 1.9157281312835157, + "learning_rate": 5.4373406281910434e-05, + "loss": 0.7835, + "step": 425 + }, + { + "epoch": 2.2216427640156455, + "grad_norm": 2.6110440355014477, + "learning_rate": 5.423679072974109e-05, + "loss": 0.7792, + "step": 426 + }, + { + "epoch": 2.226857887874837, + "grad_norm": 2.3658010170065853, + "learning_rate": 5.4099984752102295e-05, + "loss": 0.7914, + "step": 427 + }, + { + "epoch": 2.2320730117340286, + "grad_norm": 1.768060696859615, + "learning_rate": 5.396299017885465e-05, + "loss": 0.7789, + "step": 428 + }, + { + "epoch": 2.23728813559322, + "grad_norm": 1.661248122612779, + "learning_rate": 5.3825808842381345e-05, + "loss": 0.7735, + "step": 429 + }, + { + "epoch": 2.242503259452412, + "grad_norm": 2.362109109009534, + "learning_rate": 5.3688442577563646e-05, + "loss": 0.7886, + "step": 430 + }, + { + "epoch": 2.2477183833116037, + "grad_norm": 2.0017993267041905, + "learning_rate": 5.355089322175629e-05, + "loss": 0.7735, + "step": 431 + }, + { + "epoch": 2.252933507170795, + "grad_norm": 1.985452281861886, + "learning_rate": 5.3413162614763043e-05, + "loss": 0.7904, + "step": 432 + }, + { + "epoch": 2.2581486310299868, + "grad_norm": 1.8249568105371123, + "learning_rate": 5.327525259881196e-05, + "loss": 0.7792, + "step": 433 + }, + { + "epoch": 2.2633637548891787, + "grad_norm": 1.987099316108934, + "learning_rate": 5.3137165018530805e-05, + "loss": 0.7718, + "step": 434 + }, + { + "epoch": 2.2685788787483703, + "grad_norm": 1.7309972017145459, + "learning_rate": 5.299890172092238e-05, + "loss": 0.7737, + "step": 435 + }, + { + "epoch": 2.273794002607562, + "grad_norm": 2.083040780695706, + "learning_rate": 5.286046455533981e-05, + "loss": 0.7861, + "step": 436 + }, + { + "epoch": 2.279009126466754, + "grad_norm": 1.7833638732253791, + "learning_rate": 5.27218553734618e-05, + "loss": 0.7688, + "step": 437 + }, + { + "epoch": 2.2842242503259453, + "grad_norm": 2.2540602964270997, + "learning_rate": 5.2583076029267864e-05, + "loss": 0.7868, + "step": 438 + }, + { + "epoch": 2.289439374185137, + "grad_norm": 2.1042743671125046, + "learning_rate": 5.2444128379013564e-05, + "loss": 0.7806, + "step": 439 + }, + { + "epoch": 2.2946544980443284, + "grad_norm": 1.638491731440727, + "learning_rate": 5.2305014281205634e-05, + "loss": 0.7778, + "step": 440 + }, + { + "epoch": 2.2998696219035204, + "grad_norm": 1.4490203056402047, + "learning_rate": 5.2165735596577146e-05, + "loss": 0.7776, + "step": 441 + }, + { + "epoch": 2.305084745762712, + "grad_norm": 2.285181885567823, + "learning_rate": 5.202629418806263e-05, + "loss": 0.7767, + "step": 442 + }, + { + "epoch": 2.3102998696219035, + "grad_norm": 1.9464604198183288, + "learning_rate": 5.1886691920773116e-05, + "loss": 0.7651, + "step": 443 + }, + { + "epoch": 2.315514993481095, + "grad_norm": 1.7631237894703213, + "learning_rate": 5.174693066197125e-05, + "loss": 0.7726, + "step": 444 + }, + { + "epoch": 2.320730117340287, + "grad_norm": 1.56838244742276, + "learning_rate": 5.160701228104626e-05, + "loss": 0.7819, + "step": 445 + }, + { + "epoch": 2.3259452411994785, + "grad_norm": 2.202901692990395, + "learning_rate": 5.146693864948898e-05, + "loss": 0.7856, + "step": 446 + }, + { + "epoch": 2.33116036505867, + "grad_norm": 1.92108099189494, + "learning_rate": 5.132671164086682e-05, + "loss": 0.775, + "step": 447 + }, + { + "epoch": 2.3363754889178616, + "grad_norm": 1.760952455824346, + "learning_rate": 5.118633313079869e-05, + "loss": 0.7807, + "step": 448 + }, + { + "epoch": 2.3415906127770536, + "grad_norm": 1.5869940859426193, + "learning_rate": 5.104580499692992e-05, + "loss": 0.7845, + "step": 449 + }, + { + "epoch": 2.346805736636245, + "grad_norm": 2.0260440621389364, + "learning_rate": 5.090512911890715e-05, + "loss": 0.7857, + "step": 450 + }, + { + "epoch": 2.3520208604954367, + "grad_norm": 1.7167308373113537, + "learning_rate": 5.076430737835318e-05, + "loss": 0.7815, + "step": 451 + }, + { + "epoch": 2.3572359843546282, + "grad_norm": 1.881887746447444, + "learning_rate": 5.062334165884182e-05, + "loss": 0.7811, + "step": 452 + }, + { + "epoch": 2.36245110821382, + "grad_norm": 1.6824543517537593, + "learning_rate": 5.0482233845872674e-05, + "loss": 0.776, + "step": 453 + }, + { + "epoch": 2.3676662320730117, + "grad_norm": 1.8755854127797766, + "learning_rate": 5.034098582684595e-05, + "loss": 0.7803, + "step": 454 + }, + { + "epoch": 2.3728813559322033, + "grad_norm": 1.6777197590484048, + "learning_rate": 5.019959949103715e-05, + "loss": 0.7848, + "step": 455 + }, + { + "epoch": 2.3780964797913953, + "grad_norm": 1.81809003546367, + "learning_rate": 5.005807672957188e-05, + "loss": 0.7858, + "step": 456 + }, + { + "epoch": 2.383311603650587, + "grad_norm": 1.6133734346606823, + "learning_rate": 4.9916419435400516e-05, + "loss": 0.7808, + "step": 457 + }, + { + "epoch": 2.3885267275097783, + "grad_norm": 1.8363482997949565, + "learning_rate": 4.9774629503272874e-05, + "loss": 0.7796, + "step": 458 + }, + { + "epoch": 2.39374185136897, + "grad_norm": 1.5541908015103185, + "learning_rate": 4.96327088297129e-05, + "loss": 0.7771, + "step": 459 + }, + { + "epoch": 2.3989569752281614, + "grad_norm": 1.754540259013617, + "learning_rate": 4.949065931299328e-05, + "loss": 0.7858, + "step": 460 + }, + { + "epoch": 2.4041720990873534, + "grad_norm": 1.539083403172259, + "learning_rate": 4.934848285311002e-05, + "loss": 0.778, + "step": 461 + }, + { + "epoch": 2.409387222946545, + "grad_norm": 1.8702666142803812, + "learning_rate": 4.920618135175712e-05, + "loss": 0.7815, + "step": 462 + }, + { + "epoch": 2.4146023468057365, + "grad_norm": 1.644430924102753, + "learning_rate": 4.9063756712301036e-05, + "loss": 0.773, + "step": 463 + }, + { + "epoch": 2.4198174706649285, + "grad_norm": 1.6756469842658643, + "learning_rate": 4.8921210839755304e-05, + "loss": 0.7737, + "step": 464 + }, + { + "epoch": 2.42503259452412, + "grad_norm": 1.4353920458126117, + "learning_rate": 4.877854564075499e-05, + "loss": 0.7861, + "step": 465 + }, + { + "epoch": 2.4302477183833116, + "grad_norm": 1.8834730757505287, + "learning_rate": 4.863576302353125e-05, + "loss": 0.7638, + "step": 466 + }, + { + "epoch": 2.435462842242503, + "grad_norm": 1.6285082262536497, + "learning_rate": 4.849286489788579e-05, + "loss": 0.7719, + "step": 467 + }, + { + "epoch": 2.440677966101695, + "grad_norm": 1.6841805396754381, + "learning_rate": 4.834985317516525e-05, + "loss": 0.7833, + "step": 468 + }, + { + "epoch": 2.4458930899608866, + "grad_norm": 1.5254872325101743, + "learning_rate": 4.8206729768235756e-05, + "loss": 0.7766, + "step": 469 + }, + { + "epoch": 2.451108213820078, + "grad_norm": 1.5946022371035584, + "learning_rate": 4.8063496591457256e-05, + "loss": 0.7688, + "step": 470 + }, + { + "epoch": 2.4563233376792697, + "grad_norm": 1.3044832158013293, + "learning_rate": 4.792015556065793e-05, + "loss": 0.7745, + "step": 471 + }, + { + "epoch": 2.4615384615384617, + "grad_norm": 1.822367193607616, + "learning_rate": 4.777670859310857e-05, + "loss": 0.7839, + "step": 472 + }, + { + "epoch": 2.466753585397653, + "grad_norm": 1.5933250932152585, + "learning_rate": 4.763315760749695e-05, + "loss": 0.7886, + "step": 473 + }, + { + "epoch": 2.4719687092568448, + "grad_norm": 1.588436894654484, + "learning_rate": 4.748950452390212e-05, + "loss": 0.7742, + "step": 474 + }, + { + "epoch": 2.4771838331160367, + "grad_norm": 1.4522634410247943, + "learning_rate": 4.734575126376876e-05, + "loss": 0.7823, + "step": 475 + }, + { + "epoch": 2.4823989569752283, + "grad_norm": 1.586237253805528, + "learning_rate": 4.7201899749881504e-05, + "loss": 0.7926, + "step": 476 + }, + { + "epoch": 2.48761408083442, + "grad_norm": 1.3312635236777999, + "learning_rate": 4.705795190633915e-05, + "loss": 0.786, + "step": 477 + }, + { + "epoch": 2.4928292046936114, + "grad_norm": 1.7378087684763845, + "learning_rate": 4.691390965852893e-05, + "loss": 0.7693, + "step": 478 + }, + { + "epoch": 2.498044328552803, + "grad_norm": 1.554126764803909, + "learning_rate": 4.676977493310088e-05, + "loss": 0.7911, + "step": 479 + }, + { + "epoch": 2.503259452411995, + "grad_norm": 1.4897288575191927, + "learning_rate": 4.662554965794192e-05, + "loss": 0.7809, + "step": 480 + }, + { + "epoch": 2.5084745762711864, + "grad_norm": 1.2742943025314304, + "learning_rate": 4.648123576215011e-05, + "loss": 0.7811, + "step": 481 + }, + { + "epoch": 2.513689700130378, + "grad_norm": 1.714780336353614, + "learning_rate": 4.633683517600891e-05, + "loss": 0.7707, + "step": 482 + }, + { + "epoch": 2.51890482398957, + "grad_norm": 1.514593536382248, + "learning_rate": 4.61923498309613e-05, + "loss": 0.7817, + "step": 483 + }, + { + "epoch": 2.5241199478487615, + "grad_norm": 1.4058404125363677, + "learning_rate": 4.604778165958392e-05, + "loss": 0.7691, + "step": 484 + }, + { + "epoch": 2.529335071707953, + "grad_norm": 1.1465871672288677, + "learning_rate": 4.590313259556132e-05, + "loss": 0.7635, + "step": 485 + }, + { + "epoch": 2.5345501955671446, + "grad_norm": 1.6402163570623103, + "learning_rate": 4.575840457366001e-05, + "loss": 0.7876, + "step": 486 + }, + { + "epoch": 2.539765319426336, + "grad_norm": 1.3178513882789409, + "learning_rate": 4.561359952970259e-05, + "loss": 0.7664, + "step": 487 + }, + { + "epoch": 2.544980443285528, + "grad_norm": 1.6022661486448093, + "learning_rate": 4.546871940054191e-05, + "loss": 0.7905, + "step": 488 + }, + { + "epoch": 2.5501955671447196, + "grad_norm": 1.4090335919146642, + "learning_rate": 4.5323766124035115e-05, + "loss": 0.7819, + "step": 489 + }, + { + "epoch": 2.555410691003911, + "grad_norm": 1.324862817997592, + "learning_rate": 4.5178741639017736e-05, + "loss": 0.7787, + "step": 490 + }, + { + "epoch": 2.560625814863103, + "grad_norm": 1.1347550661741326, + "learning_rate": 4.503364788527774e-05, + "loss": 0.7831, + "step": 491 + }, + { + "epoch": 2.5658409387222947, + "grad_norm": 1.5760247018602986, + "learning_rate": 4.488848680352965e-05, + "loss": 0.7796, + "step": 492 + }, + { + "epoch": 2.5710560625814862, + "grad_norm": 1.3090981394784642, + "learning_rate": 4.4743260335388516e-05, + "loss": 0.7724, + "step": 493 + }, + { + "epoch": 2.576271186440678, + "grad_norm": 1.3548998817619167, + "learning_rate": 4.459797042334395e-05, + "loss": 0.781, + "step": 494 + }, + { + "epoch": 2.5814863102998697, + "grad_norm": 1.2234336308095453, + "learning_rate": 4.445261901073419e-05, + "loss": 0.7743, + "step": 495 + }, + { + "epoch": 2.5867014341590613, + "grad_norm": 1.2594794059995964, + "learning_rate": 4.430720804172008e-05, + "loss": 0.7878, + "step": 496 + }, + { + "epoch": 2.591916558018253, + "grad_norm": 1.0475475763925866, + "learning_rate": 4.416173946125906e-05, + "loss": 0.7851, + "step": 497 + }, + { + "epoch": 2.5971316818774444, + "grad_norm": 1.3034377416715606, + "learning_rate": 4.401621521507914e-05, + "loss": 0.7661, + "step": 498 + }, + { + "epoch": 2.6023468057366363, + "grad_norm": 1.0178177147229814, + "learning_rate": 4.387063724965292e-05, + "loss": 0.768, + "step": 499 + }, + { + "epoch": 2.607561929595828, + "grad_norm": 1.4358534072422087, + "learning_rate": 4.372500751217153e-05, + "loss": 0.7886, + "step": 500 + }, + { + "epoch": 2.6127770534550194, + "grad_norm": 1.150853401699166, + "learning_rate": 4.357932795051852e-05, + "loss": 0.7934, + "step": 501 + }, + { + "epoch": 2.6179921773142114, + "grad_norm": 1.1810592463532203, + "learning_rate": 4.3433600513243965e-05, + "loss": 0.7563, + "step": 502 + }, + { + "epoch": 2.623207301173403, + "grad_norm": 1.0034045366911863, + "learning_rate": 4.328782714953823e-05, + "loss": 0.79, + "step": 503 + }, + { + "epoch": 2.6284224250325945, + "grad_norm": 1.0980643139376323, + "learning_rate": 4.3142009809205986e-05, + "loss": 0.7861, + "step": 504 + }, + { + "epoch": 2.633637548891786, + "grad_norm": 0.8881990742171157, + "learning_rate": 4.2996150442640163e-05, + "loss": 0.7758, + "step": 505 + }, + { + "epoch": 2.6388526727509776, + "grad_norm": 0.8675797674117819, + "learning_rate": 4.2850251000795735e-05, + "loss": 0.7801, + "step": 506 + }, + { + "epoch": 2.6440677966101696, + "grad_norm": 0.7883090264802906, + "learning_rate": 4.270431343516379e-05, + "loss": 0.7745, + "step": 507 + }, + { + "epoch": 2.649282920469361, + "grad_norm": 0.7583525803187019, + "learning_rate": 4.2558339697745297e-05, + "loss": 0.7827, + "step": 508 + }, + { + "epoch": 2.6544980443285526, + "grad_norm": 0.7762877342404452, + "learning_rate": 4.2412331741025045e-05, + "loss": 0.7967, + "step": 509 + }, + { + "epoch": 2.6597131681877446, + "grad_norm": 0.7014761796032362, + "learning_rate": 4.226629151794555e-05, + "loss": 0.7858, + "step": 510 + }, + { + "epoch": 2.664928292046936, + "grad_norm": 0.7396414732409832, + "learning_rate": 4.2120220981880875e-05, + "loss": 0.7794, + "step": 511 + }, + { + "epoch": 2.6701434159061277, + "grad_norm": 0.6998054830397707, + "learning_rate": 4.197412208661058e-05, + "loss": 0.7777, + "step": 512 + }, + { + "epoch": 2.6753585397653197, + "grad_norm": 0.6386491344974894, + "learning_rate": 4.182799678629351e-05, + "loss": 0.7823, + "step": 513 + }, + { + "epoch": 2.680573663624511, + "grad_norm": 0.5877328765456284, + "learning_rate": 4.168184703544171e-05, + "loss": 0.7744, + "step": 514 + }, + { + "epoch": 2.6857887874837028, + "grad_norm": 0.5920983354896419, + "learning_rate": 4.153567478889426e-05, + "loss": 0.7769, + "step": 515 + }, + { + "epoch": 2.6910039113428943, + "grad_norm": 0.4867319202645289, + "learning_rate": 4.138948200179115e-05, + "loss": 0.7807, + "step": 516 + }, + { + "epoch": 2.696219035202086, + "grad_norm": 0.5746181168372155, + "learning_rate": 4.124327062954707e-05, + "loss": 0.7708, + "step": 517 + }, + { + "epoch": 2.701434159061278, + "grad_norm": 0.35394623638423245, + "learning_rate": 4.1097042627825325e-05, + "loss": 0.7823, + "step": 518 + }, + { + "epoch": 2.7066492829204694, + "grad_norm": 0.6245548216599335, + "learning_rate": 4.095079995251168e-05, + "loss": 0.784, + "step": 519 + }, + { + "epoch": 2.711864406779661, + "grad_norm": 0.3951096697035697, + "learning_rate": 4.080454455968812e-05, + "loss": 0.7711, + "step": 520 + }, + { + "epoch": 2.717079530638853, + "grad_norm": 0.5759028178578062, + "learning_rate": 4.065827840560674e-05, + "loss": 0.7773, + "step": 521 + }, + { + "epoch": 2.7222946544980444, + "grad_norm": 0.39680693057913397, + "learning_rate": 4.0512003446663576e-05, + "loss": 0.7736, + "step": 522 + }, + { + "epoch": 2.727509778357236, + "grad_norm": 0.49858076462377626, + "learning_rate": 4.0365721639372476e-05, + "loss": 0.7884, + "step": 523 + }, + { + "epoch": 2.7327249022164275, + "grad_norm": 0.4117350165519155, + "learning_rate": 4.021943494033882e-05, + "loss": 0.7707, + "step": 524 + }, + { + "epoch": 2.737940026075619, + "grad_norm": 0.4515042047115184, + "learning_rate": 4.007314530623348e-05, + "loss": 0.7778, + "step": 525 + }, + { + "epoch": 2.743155149934811, + "grad_norm": 0.42547404001436967, + "learning_rate": 3.9926854693766536e-05, + "loss": 0.7688, + "step": 526 + }, + { + "epoch": 2.7483702737940026, + "grad_norm": 0.35069455268010413, + "learning_rate": 3.978056505966119e-05, + "loss": 0.7744, + "step": 527 + }, + { + "epoch": 2.753585397653194, + "grad_norm": 0.4444789741021684, + "learning_rate": 3.963427836062753e-05, + "loss": 0.775, + "step": 528 + }, + { + "epoch": 2.758800521512386, + "grad_norm": 0.43329608911430956, + "learning_rate": 3.948799655333644e-05, + "loss": 0.7748, + "step": 529 + }, + { + "epoch": 2.7640156453715776, + "grad_norm": 0.4209440665741905, + "learning_rate": 3.9341721594393276e-05, + "loss": 0.7762, + "step": 530 + }, + { + "epoch": 2.769230769230769, + "grad_norm": 0.4322531684473173, + "learning_rate": 3.919545544031189e-05, + "loss": 0.7878, + "step": 531 + }, + { + "epoch": 2.774445893089961, + "grad_norm": 0.4720335982176997, + "learning_rate": 3.904920004748833e-05, + "loss": 0.777, + "step": 532 + }, + { + "epoch": 2.7796610169491527, + "grad_norm": 0.37675833736688985, + "learning_rate": 3.8902957372174675e-05, + "loss": 0.7728, + "step": 533 + }, + { + "epoch": 2.7848761408083442, + "grad_norm": 0.3961055004093581, + "learning_rate": 3.8756729370452936e-05, + "loss": 0.7745, + "step": 534 + }, + { + "epoch": 2.7900912646675358, + "grad_norm": 0.39787870628829797, + "learning_rate": 3.8610517998208866e-05, + "loss": 0.779, + "step": 535 + }, + { + "epoch": 2.7953063885267273, + "grad_norm": 0.366576904221954, + "learning_rate": 3.846432521110574e-05, + "loss": 0.7823, + "step": 536 + }, + { + "epoch": 2.8005215123859193, + "grad_norm": 0.39669396154921643, + "learning_rate": 3.831815296455829e-05, + "loss": 0.7776, + "step": 537 + }, + { + "epoch": 2.805736636245111, + "grad_norm": 0.5485537263803792, + "learning_rate": 3.817200321370649e-05, + "loss": 0.7821, + "step": 538 + }, + { + "epoch": 2.8109517601043024, + "grad_norm": 0.5046596554915849, + "learning_rate": 3.802587791338943e-05, + "loss": 0.757, + "step": 539 + }, + { + "epoch": 2.8161668839634943, + "grad_norm": 0.4297088800078496, + "learning_rate": 3.787977901811914e-05, + "loss": 0.7819, + "step": 540 + }, + { + "epoch": 2.821382007822686, + "grad_norm": 0.2792196466948258, + "learning_rate": 3.7733708482054476e-05, + "loss": 0.7908, + "step": 541 + }, + { + "epoch": 2.8265971316818774, + "grad_norm": 0.3389535808547031, + "learning_rate": 3.758766825897497e-05, + "loss": 0.7789, + "step": 542 + }, + { + "epoch": 2.831812255541069, + "grad_norm": 0.33504887483040935, + "learning_rate": 3.7441660302254724e-05, + "loss": 0.7641, + "step": 543 + }, + { + "epoch": 2.8370273794002605, + "grad_norm": 0.35811016749224023, + "learning_rate": 3.729568656483623e-05, + "loss": 0.7711, + "step": 544 + }, + { + "epoch": 2.8422425032594525, + "grad_norm": 0.3304033008238814, + "learning_rate": 3.714974899920428e-05, + "loss": 0.781, + "step": 545 + }, + { + "epoch": 2.847457627118644, + "grad_norm": 0.42274779776394655, + "learning_rate": 3.7003849557359863e-05, + "loss": 0.7738, + "step": 546 + }, + { + "epoch": 2.8526727509778356, + "grad_norm": 0.33608657379975826, + "learning_rate": 3.685799019079402e-05, + "loss": 0.7858, + "step": 547 + }, + { + "epoch": 2.8578878748370276, + "grad_norm": 0.4402897321734762, + "learning_rate": 3.6712172850461785e-05, + "loss": 0.7679, + "step": 548 + }, + { + "epoch": 2.863102998696219, + "grad_norm": 0.29934112633294296, + "learning_rate": 3.6566399486756055e-05, + "loss": 0.7889, + "step": 549 + }, + { + "epoch": 2.8683181225554106, + "grad_norm": 0.38015963993433893, + "learning_rate": 3.642067204948149e-05, + "loss": 0.7695, + "step": 550 + }, + { + "epoch": 2.8735332464146026, + "grad_norm": 0.38644710923105946, + "learning_rate": 3.627499248782849e-05, + "loss": 0.7745, + "step": 551 + }, + { + "epoch": 2.878748370273794, + "grad_norm": 0.44283035266374077, + "learning_rate": 3.612936275034709e-05, + "loss": 0.7829, + "step": 552 + }, + { + "epoch": 2.8839634941329857, + "grad_norm": 0.426691177545099, + "learning_rate": 3.598378478492087e-05, + "loss": 0.7775, + "step": 553 + }, + { + "epoch": 2.8891786179921772, + "grad_norm": 0.41202751066509746, + "learning_rate": 3.583826053874096e-05, + "loss": 0.7792, + "step": 554 + }, + { + "epoch": 2.8943937418513688, + "grad_norm": 0.40337608056255875, + "learning_rate": 3.5692791958279924e-05, + "loss": 0.7737, + "step": 555 + }, + { + "epoch": 2.8996088657105608, + "grad_norm": 0.42109772590116307, + "learning_rate": 3.554738098926582e-05, + "loss": 0.7757, + "step": 556 + }, + { + "epoch": 2.9048239895697523, + "grad_norm": 0.6458006850893471, + "learning_rate": 3.540202957665606e-05, + "loss": 0.7775, + "step": 557 + }, + { + "epoch": 2.910039113428944, + "grad_norm": 0.5129008593035667, + "learning_rate": 3.525673966461149e-05, + "loss": 0.7793, + "step": 558 + }, + { + "epoch": 2.915254237288136, + "grad_norm": 0.37986325064356524, + "learning_rate": 3.5111513196470356e-05, + "loss": 0.7706, + "step": 559 + }, + { + "epoch": 2.9204693611473274, + "grad_norm": 0.3127213742675476, + "learning_rate": 3.4966352114722264e-05, + "loss": 0.779, + "step": 560 + }, + { + "epoch": 2.925684485006519, + "grad_norm": 0.44150964872249004, + "learning_rate": 3.482125836098227e-05, + "loss": 0.7726, + "step": 561 + }, + { + "epoch": 2.9308996088657104, + "grad_norm": 0.37918325106659184, + "learning_rate": 3.46762338759649e-05, + "loss": 0.7763, + "step": 562 + }, + { + "epoch": 2.936114732724902, + "grad_norm": 0.34801331615309594, + "learning_rate": 3.4531280599458096e-05, + "loss": 0.7792, + "step": 563 + }, + { + "epoch": 2.941329856584094, + "grad_norm": 0.3417071688239163, + "learning_rate": 3.438640047029741e-05, + "loss": 0.783, + "step": 564 + }, + { + "epoch": 2.9465449804432855, + "grad_norm": 0.4027724323607999, + "learning_rate": 3.424159542634001e-05, + "loss": 0.7799, + "step": 565 + }, + { + "epoch": 2.951760104302477, + "grad_norm": 0.359846745274617, + "learning_rate": 3.409686740443869e-05, + "loss": 0.7874, + "step": 566 + }, + { + "epoch": 2.956975228161669, + "grad_norm": 0.39877444581949706, + "learning_rate": 3.395221834041608e-05, + "loss": 0.7796, + "step": 567 + }, + { + "epoch": 2.9621903520208606, + "grad_norm": 0.31627637641393286, + "learning_rate": 3.3807650169038714e-05, + "loss": 0.7828, + "step": 568 + }, + { + "epoch": 2.967405475880052, + "grad_norm": 0.2915329688447508, + "learning_rate": 3.366316482399111e-05, + "loss": 0.7611, + "step": 569 + }, + { + "epoch": 2.9726205997392436, + "grad_norm": 0.3003806981610255, + "learning_rate": 3.351876423784991e-05, + "loss": 0.7753, + "step": 570 + }, + { + "epoch": 2.9778357235984356, + "grad_norm": 0.32825206856441885, + "learning_rate": 3.33744503420581e-05, + "loss": 0.7805, + "step": 571 + }, + { + "epoch": 2.983050847457627, + "grad_norm": 0.28811623986201, + "learning_rate": 3.323022506689913e-05, + "loss": 0.7787, + "step": 572 + }, + { + "epoch": 2.9882659713168187, + "grad_norm": 0.3007386340259681, + "learning_rate": 3.308609034147109e-05, + "loss": 0.7693, + "step": 573 + }, + { + "epoch": 2.9934810951760102, + "grad_norm": 0.28601621207061495, + "learning_rate": 3.294204809366088e-05, + "loss": 0.7811, + "step": 574 + }, + { + "epoch": 2.9986962190352022, + "grad_norm": 0.38527843914278115, + "learning_rate": 3.27981002501185e-05, + "loss": 0.8962, + "step": 575 + }, + { + "epoch": 3.0039113428943938, + "grad_norm": 0.3891353205481252, + "learning_rate": 3.265424873623125e-05, + "loss": 0.8279, + "step": 576 + }, + { + "epoch": 3.0091264667535853, + "grad_norm": 0.34677276173333527, + "learning_rate": 3.251049547609789e-05, + "loss": 0.7617, + "step": 577 + }, + { + "epoch": 3.014341590612777, + "grad_norm": 0.3887463573130963, + "learning_rate": 3.2366842392503065e-05, + "loss": 0.7404, + "step": 578 + }, + { + "epoch": 3.019556714471969, + "grad_norm": 0.45850333580212727, + "learning_rate": 3.222329140689144e-05, + "loss": 0.7509, + "step": 579 + }, + { + "epoch": 3.0247718383311604, + "grad_norm": 0.42139454484056155, + "learning_rate": 3.207984443934208e-05, + "loss": 0.7474, + "step": 580 + }, + { + "epoch": 3.029986962190352, + "grad_norm": 0.32213156347235194, + "learning_rate": 3.193650340854275e-05, + "loss": 0.7449, + "step": 581 + }, + { + "epoch": 3.035202086049544, + "grad_norm": 0.2882489783740716, + "learning_rate": 3.179327023176426e-05, + "loss": 0.7565, + "step": 582 + }, + { + "epoch": 3.0404172099087354, + "grad_norm": 0.31398533972182546, + "learning_rate": 3.1650146824834765e-05, + "loss": 0.7663, + "step": 583 + }, + { + "epoch": 3.045632333767927, + "grad_norm": 0.42483377951795925, + "learning_rate": 3.1507135102114224e-05, + "loss": 0.7524, + "step": 584 + }, + { + "epoch": 3.0508474576271185, + "grad_norm": 0.4203137878607459, + "learning_rate": 3.1364236976468755e-05, + "loss": 0.7476, + "step": 585 + }, + { + "epoch": 3.0560625814863105, + "grad_norm": 0.3536224112892052, + "learning_rate": 3.122145435924502e-05, + "loss": 0.7479, + "step": 586 + }, + { + "epoch": 3.061277705345502, + "grad_norm": 0.3355710189906583, + "learning_rate": 3.107878916024471e-05, + "loss": 0.7475, + "step": 587 + }, + { + "epoch": 3.0664928292046936, + "grad_norm": 0.3578880245402906, + "learning_rate": 3.093624328769898e-05, + "loss": 0.7542, + "step": 588 + }, + { + "epoch": 3.071707953063885, + "grad_norm": 0.3807896298844658, + "learning_rate": 3.079381864824289e-05, + "loss": 0.7617, + "step": 589 + }, + { + "epoch": 3.076923076923077, + "grad_norm": 0.37153630367778145, + "learning_rate": 3.065151714688998e-05, + "loss": 0.7581, + "step": 590 + }, + { + "epoch": 3.0821382007822686, + "grad_norm": 0.41123425605078595, + "learning_rate": 3.0509340687006735e-05, + "loss": 0.7538, + "step": 591 + }, + { + "epoch": 3.08735332464146, + "grad_norm": 0.33981481049719536, + "learning_rate": 3.0367291170287102e-05, + "loss": 0.7524, + "step": 592 + }, + { + "epoch": 3.0925684485006517, + "grad_norm": 0.3651990332335353, + "learning_rate": 3.0225370496727125e-05, + "loss": 0.7584, + "step": 593 + }, + { + "epoch": 3.0977835723598437, + "grad_norm": 0.2918824973106232, + "learning_rate": 3.0083580564599484e-05, + "loss": 0.7563, + "step": 594 + }, + { + "epoch": 3.1029986962190352, + "grad_norm": 0.3423081259716029, + "learning_rate": 2.9941923270428125e-05, + "loss": 0.7471, + "step": 595 + }, + { + "epoch": 3.1082138200782268, + "grad_norm": 0.3604784090698713, + "learning_rate": 2.9800400508962854e-05, + "loss": 0.7458, + "step": 596 + }, + { + "epoch": 3.1134289439374183, + "grad_norm": 0.44107810877163156, + "learning_rate": 2.9659014173154053e-05, + "loss": 0.761, + "step": 597 + }, + { + "epoch": 3.1186440677966103, + "grad_norm": 0.36068201923449394, + "learning_rate": 2.9517766154127332e-05, + "loss": 0.7495, + "step": 598 + }, + { + "epoch": 3.123859191655802, + "grad_norm": 0.32775307660721575, + "learning_rate": 2.93766583411582e-05, + "loss": 0.7463, + "step": 599 + }, + { + "epoch": 3.1290743155149934, + "grad_norm": 0.4150847775843986, + "learning_rate": 2.923569262164684e-05, + "loss": 0.7459, + "step": 600 + }, + { + "epoch": 3.134289439374185, + "grad_norm": 0.3453247625813097, + "learning_rate": 2.909487088109287e-05, + "loss": 0.756, + "step": 601 + }, + { + "epoch": 3.139504563233377, + "grad_norm": 0.3884807540864722, + "learning_rate": 2.8954195003070106e-05, + "loss": 0.7469, + "step": 602 + }, + { + "epoch": 3.1447196870925684, + "grad_norm": 0.32406875922905615, + "learning_rate": 2.8813666869201323e-05, + "loss": 0.7492, + "step": 603 + }, + { + "epoch": 3.14993481095176, + "grad_norm": 0.3124660936977556, + "learning_rate": 2.867328835913319e-05, + "loss": 0.7412, + "step": 604 + }, + { + "epoch": 3.155149934810952, + "grad_norm": 0.402564278458608, + "learning_rate": 2.853306135051103e-05, + "loss": 0.7628, + "step": 605 + }, + { + "epoch": 3.1603650586701435, + "grad_norm": 0.3914753370029631, + "learning_rate": 2.8392987718953748e-05, + "loss": 0.7467, + "step": 606 + }, + { + "epoch": 3.165580182529335, + "grad_norm": 0.36094618715771837, + "learning_rate": 2.8253069338028756e-05, + "loss": 0.7528, + "step": 607 + }, + { + "epoch": 3.1707953063885266, + "grad_norm": 0.4160569360480524, + "learning_rate": 2.8113308079226897e-05, + "loss": 0.753, + "step": 608 + }, + { + "epoch": 3.1760104302477186, + "grad_norm": 0.36846901981179736, + "learning_rate": 2.7973705811937387e-05, + "loss": 0.762, + "step": 609 + }, + { + "epoch": 3.18122555410691, + "grad_norm": 0.3181951199229672, + "learning_rate": 2.7834264403422864e-05, + "loss": 0.7581, + "step": 610 + }, + { + "epoch": 3.1864406779661016, + "grad_norm": 0.28604043043751265, + "learning_rate": 2.769498571879438e-05, + "loss": 0.7591, + "step": 611 + }, + { + "epoch": 3.191655801825293, + "grad_norm": 0.2634604899316907, + "learning_rate": 2.7555871620986453e-05, + "loss": 0.7554, + "step": 612 + }, + { + "epoch": 3.196870925684485, + "grad_norm": 0.3039718480846403, + "learning_rate": 2.7416923970732146e-05, + "loss": 0.7588, + "step": 613 + }, + { + "epoch": 3.2020860495436767, + "grad_norm": 0.288103576580876, + "learning_rate": 2.7278144626538212e-05, + "loss": 0.7487, + "step": 614 + }, + { + "epoch": 3.2073011734028682, + "grad_norm": 0.32239905156935256, + "learning_rate": 2.71395354446602e-05, + "loss": 0.7459, + "step": 615 + }, + { + "epoch": 3.21251629726206, + "grad_norm": 0.29806440969996567, + "learning_rate": 2.7001098279077623e-05, + "loss": 0.7548, + "step": 616 + }, + { + "epoch": 3.2177314211212518, + "grad_norm": 0.25998375637060805, + "learning_rate": 2.6862834981469198e-05, + "loss": 0.7486, + "step": 617 + }, + { + "epoch": 3.2229465449804433, + "grad_norm": 0.39237992094856566, + "learning_rate": 2.6724747401188055e-05, + "loss": 0.7532, + "step": 618 + }, + { + "epoch": 3.228161668839635, + "grad_norm": 0.2548892235121485, + "learning_rate": 2.6586837385236966e-05, + "loss": 0.7484, + "step": 619 + }, + { + "epoch": 3.2333767926988264, + "grad_norm": 0.3225804950825967, + "learning_rate": 2.6449106778243708e-05, + "loss": 0.7529, + "step": 620 + }, + { + "epoch": 3.2385919165580184, + "grad_norm": 0.45928173129349054, + "learning_rate": 2.631155742243637e-05, + "loss": 0.7549, + "step": 621 + }, + { + "epoch": 3.24380704041721, + "grad_norm": 0.3308854647358001, + "learning_rate": 2.6174191157618652e-05, + "loss": 0.7554, + "step": 622 + }, + { + "epoch": 3.2490221642764014, + "grad_norm": 0.2795489427507208, + "learning_rate": 2.6037009821145346e-05, + "loss": 0.7593, + "step": 623 + }, + { + "epoch": 3.2542372881355934, + "grad_norm": 0.4430769575631773, + "learning_rate": 2.5900015247897714e-05, + "loss": 0.7545, + "step": 624 + }, + { + "epoch": 3.259452411994785, + "grad_norm": 0.33341725027604363, + "learning_rate": 2.576320927025892e-05, + "loss": 0.753, + "step": 625 + }, + { + "epoch": 3.2646675358539765, + "grad_norm": 0.33122719602795536, + "learning_rate": 2.562659371808959e-05, + "loss": 0.7436, + "step": 626 + }, + { + "epoch": 3.269882659713168, + "grad_norm": 0.3963689954079041, + "learning_rate": 2.549017041870325e-05, + "loss": 0.7676, + "step": 627 + }, + { + "epoch": 3.27509778357236, + "grad_norm": 0.3538575496536462, + "learning_rate": 2.535394119684197e-05, + "loss": 0.7548, + "step": 628 + }, + { + "epoch": 3.2803129074315516, + "grad_norm": 0.24662943382751912, + "learning_rate": 2.521790787465193e-05, + "loss": 0.7517, + "step": 629 + }, + { + "epoch": 3.285528031290743, + "grad_norm": 0.39858795017503024, + "learning_rate": 2.5082072271658982e-05, + "loss": 0.7482, + "step": 630 + }, + { + "epoch": 3.2907431551499347, + "grad_norm": 0.24293852218374687, + "learning_rate": 2.494643620474442e-05, + "loss": 0.7445, + "step": 631 + }, + { + "epoch": 3.2959582790091266, + "grad_norm": 0.3499814647894016, + "learning_rate": 2.4811001488120598e-05, + "loss": 0.7555, + "step": 632 + }, + { + "epoch": 3.301173402868318, + "grad_norm": 0.27161500960796325, + "learning_rate": 2.4675769933306696e-05, + "loss": 0.7568, + "step": 633 + }, + { + "epoch": 3.3063885267275097, + "grad_norm": 0.2881787773315299, + "learning_rate": 2.4540743349104485e-05, + "loss": 0.7442, + "step": 634 + }, + { + "epoch": 3.3116036505867013, + "grad_norm": 0.2740026111414901, + "learning_rate": 2.440592354157415e-05, + "loss": 0.7579, + "step": 635 + }, + { + "epoch": 3.3168187744458932, + "grad_norm": 0.2605377452426752, + "learning_rate": 2.427131231401008e-05, + "loss": 0.7518, + "step": 636 + }, + { + "epoch": 3.3220338983050848, + "grad_norm": 0.2482525542273476, + "learning_rate": 2.413691146691681e-05, + "loss": 0.7615, + "step": 637 + }, + { + "epoch": 3.3272490221642763, + "grad_norm": 0.33206747294953354, + "learning_rate": 2.400272279798492e-05, + "loss": 0.7498, + "step": 638 + }, + { + "epoch": 3.332464146023468, + "grad_norm": 0.29080523675344855, + "learning_rate": 2.3868748102066964e-05, + "loss": 0.7457, + "step": 639 + }, + { + "epoch": 3.33767926988266, + "grad_norm": 0.31636040314912883, + "learning_rate": 2.3734989171153475e-05, + "loss": 0.7605, + "step": 640 + }, + { + "epoch": 3.3428943937418514, + "grad_norm": 0.3319918116811582, + "learning_rate": 2.3601447794349035e-05, + "loss": 0.7601, + "step": 641 + }, + { + "epoch": 3.348109517601043, + "grad_norm": 0.24931361371615676, + "learning_rate": 2.346812575784826e-05, + "loss": 0.752, + "step": 642 + }, + { + "epoch": 3.353324641460235, + "grad_norm": 0.30718347134905344, + "learning_rate": 2.333502484491202e-05, + "loss": 0.7616, + "step": 643 + }, + { + "epoch": 3.3585397653194264, + "grad_norm": 0.282987877244827, + "learning_rate": 2.3202146835843458e-05, + "loss": 0.7462, + "step": 644 + }, + { + "epoch": 3.363754889178618, + "grad_norm": 0.24823892221787283, + "learning_rate": 2.3069493507964328e-05, + "loss": 0.7508, + "step": 645 + }, + { + "epoch": 3.3689700130378095, + "grad_norm": 0.2747764846839421, + "learning_rate": 2.2937066635591082e-05, + "loss": 0.757, + "step": 646 + }, + { + "epoch": 3.374185136897001, + "grad_norm": 0.2496756674179984, + "learning_rate": 2.2804867990011186e-05, + "loss": 0.7578, + "step": 647 + }, + { + "epoch": 3.379400260756193, + "grad_norm": 0.2395621072609135, + "learning_rate": 2.26728993394595e-05, + "loss": 0.765, + "step": 648 + }, + { + "epoch": 3.3846153846153846, + "grad_norm": 0.25006905141016617, + "learning_rate": 2.2541162449094517e-05, + "loss": 0.7558, + "step": 649 + }, + { + "epoch": 3.389830508474576, + "grad_norm": 0.22593196210127933, + "learning_rate": 2.2409659080974792e-05, + "loss": 0.7441, + "step": 650 + }, + { + "epoch": 3.395045632333768, + "grad_norm": 0.25905615560125267, + "learning_rate": 2.2278390994035437e-05, + "loss": 0.759, + "step": 651 + }, + { + "epoch": 3.4002607561929596, + "grad_norm": 0.21460137257291464, + "learning_rate": 2.214735994406449e-05, + "loss": 0.7356, + "step": 652 + }, + { + "epoch": 3.405475880052151, + "grad_norm": 0.2844041034563339, + "learning_rate": 2.2016567683679474e-05, + "loss": 0.7611, + "step": 653 + }, + { + "epoch": 3.4106910039113427, + "grad_norm": 0.2193094602636308, + "learning_rate": 2.188601596230402e-05, + "loss": 0.7628, + "step": 654 + }, + { + "epoch": 3.4159061277705347, + "grad_norm": 0.2688886990907944, + "learning_rate": 2.1755706526144346e-05, + "loss": 0.7555, + "step": 655 + }, + { + "epoch": 3.4211212516297262, + "grad_norm": 0.22979387277456514, + "learning_rate": 2.1625641118165993e-05, + "loss": 0.7506, + "step": 656 + }, + { + "epoch": 3.426336375488918, + "grad_norm": 0.2575836540395883, + "learning_rate": 2.1495821478070463e-05, + "loss": 0.7454, + "step": 657 + }, + { + "epoch": 3.4315514993481093, + "grad_norm": 0.22113251166352366, + "learning_rate": 2.1366249342271983e-05, + "loss": 0.7607, + "step": 658 + }, + { + "epoch": 3.4367666232073013, + "grad_norm": 0.2541994114755118, + "learning_rate": 2.123692644387427e-05, + "loss": 0.7529, + "step": 659 + }, + { + "epoch": 3.441981747066493, + "grad_norm": 0.17984867947865277, + "learning_rate": 2.110785451264733e-05, + "loss": 0.7584, + "step": 660 + }, + { + "epoch": 3.4471968709256844, + "grad_norm": 0.2623173957981238, + "learning_rate": 2.0979035275004326e-05, + "loss": 0.7589, + "step": 661 + }, + { + "epoch": 3.4524119947848764, + "grad_norm": 0.17573507602846267, + "learning_rate": 2.0850470453978527e-05, + "loss": 0.7575, + "step": 662 + }, + { + "epoch": 3.457627118644068, + "grad_norm": 0.23281049598480352, + "learning_rate": 2.07221617692002e-05, + "loss": 0.7495, + "step": 663 + }, + { + "epoch": 3.4628422425032594, + "grad_norm": 0.19792669337262203, + "learning_rate": 2.0594110936873646e-05, + "loss": 0.7554, + "step": 664 + }, + { + "epoch": 3.468057366362451, + "grad_norm": 0.2348813479171187, + "learning_rate": 2.0466319669754257e-05, + "loss": 0.7459, + "step": 665 + }, + { + "epoch": 3.4732724902216425, + "grad_norm": 0.17805434883842863, + "learning_rate": 2.033878967712556e-05, + "loss": 0.7482, + "step": 666 + }, + { + "epoch": 3.4784876140808345, + "grad_norm": 0.2324152396008409, + "learning_rate": 2.0211522664776378e-05, + "loss": 0.7627, + "step": 667 + }, + { + "epoch": 3.483702737940026, + "grad_norm": 0.21569790923016002, + "learning_rate": 2.0084520334978066e-05, + "loss": 0.7475, + "step": 668 + }, + { + "epoch": 3.4889178617992176, + "grad_norm": 0.22787366870500253, + "learning_rate": 1.9957784386461642e-05, + "loss": 0.7539, + "step": 669 + }, + { + "epoch": 3.4941329856584096, + "grad_norm": 0.19346838428327817, + "learning_rate": 1.9831316514395125e-05, + "loss": 0.7482, + "step": 670 + }, + { + "epoch": 3.499348109517601, + "grad_norm": 0.22386565166114242, + "learning_rate": 1.9705118410360888e-05, + "loss": 0.7516, + "step": 671 + }, + { + "epoch": 3.5045632333767927, + "grad_norm": 0.19045550096072955, + "learning_rate": 1.9579191762332962e-05, + "loss": 0.7524, + "step": 672 + }, + { + "epoch": 3.509778357235984, + "grad_norm": 0.2568792722114023, + "learning_rate": 1.9453538254654492e-05, + "loss": 0.7608, + "step": 673 + }, + { + "epoch": 3.514993481095176, + "grad_norm": 0.2309953251505408, + "learning_rate": 1.9328159568015253e-05, + "loss": 0.761, + "step": 674 + }, + { + "epoch": 3.5202086049543677, + "grad_norm": 0.20090449798296708, + "learning_rate": 1.920305737942908e-05, + "loss": 0.7498, + "step": 675 + }, + { + "epoch": 3.5254237288135593, + "grad_norm": 0.1995433496289568, + "learning_rate": 1.9078233362211488e-05, + "loss": 0.7635, + "step": 676 + }, + { + "epoch": 3.530638852672751, + "grad_norm": 0.20023193260506214, + "learning_rate": 1.895368918595733e-05, + "loss": 0.7497, + "step": 677 + }, + { + "epoch": 3.5358539765319428, + "grad_norm": 0.19621170906491942, + "learning_rate": 1.882942651651835e-05, + "loss": 0.7434, + "step": 678 + }, + { + "epoch": 3.5410691003911343, + "grad_norm": 0.20574042364789344, + "learning_rate": 1.8705447015981056e-05, + "loss": 0.7436, + "step": 679 + }, + { + "epoch": 3.546284224250326, + "grad_norm": 0.18782121755639544, + "learning_rate": 1.858175234264434e-05, + "loss": 0.7556, + "step": 680 + }, + { + "epoch": 3.551499348109518, + "grad_norm": 0.1936150072481402, + "learning_rate": 1.845834415099736e-05, + "loss": 0.757, + "step": 681 + }, + { + "epoch": 3.5567144719687094, + "grad_norm": 0.1926949621185055, + "learning_rate": 1.833522409169748e-05, + "loss": 0.756, + "step": 682 + }, + { + "epoch": 3.561929595827901, + "grad_norm": 0.20476384841537432, + "learning_rate": 1.8212393811548074e-05, + "loss": 0.7341, + "step": 683 + }, + { + "epoch": 3.5671447196870925, + "grad_norm": 0.1696138534253153, + "learning_rate": 1.8089854953476503e-05, + "loss": 0.7521, + "step": 684 + }, + { + "epoch": 3.572359843546284, + "grad_norm": 0.22686575356319796, + "learning_rate": 1.796760915651229e-05, + "loss": 0.7621, + "step": 685 + }, + { + "epoch": 3.577574967405476, + "grad_norm": 0.18860960425351292, + "learning_rate": 1.784565805576503e-05, + "loss": 0.7537, + "step": 686 + }, + { + "epoch": 3.5827900912646675, + "grad_norm": 0.19640065748239693, + "learning_rate": 1.7724003282402567e-05, + "loss": 0.7592, + "step": 687 + }, + { + "epoch": 3.588005215123859, + "grad_norm": 0.18973579313916158, + "learning_rate": 1.760264646362926e-05, + "loss": 0.7477, + "step": 688 + }, + { + "epoch": 3.593220338983051, + "grad_norm": 0.20298759538425024, + "learning_rate": 1.7481589222664076e-05, + "loss": 0.7486, + "step": 689 + }, + { + "epoch": 3.5984354628422426, + "grad_norm": 0.17790150008707342, + "learning_rate": 1.736083317871898e-05, + "loss": 0.7458, + "step": 690 + }, + { + "epoch": 3.603650586701434, + "grad_norm": 0.20575599062083585, + "learning_rate": 1.7240379946977283e-05, + "loss": 0.7544, + "step": 691 + }, + { + "epoch": 3.6088657105606257, + "grad_norm": 0.17855398739046716, + "learning_rate": 1.712023113857195e-05, + "loss": 0.756, + "step": 692 + }, + { + "epoch": 3.614080834419817, + "grad_norm": 0.1947379717329557, + "learning_rate": 1.700038836056413e-05, + "loss": 0.7547, + "step": 693 + }, + { + "epoch": 3.619295958279009, + "grad_norm": 0.17740265903389008, + "learning_rate": 1.6880853215921642e-05, + "loss": 0.7532, + "step": 694 + }, + { + "epoch": 3.6245110821382007, + "grad_norm": 0.18082866727173452, + "learning_rate": 1.676162730349751e-05, + "loss": 0.7466, + "step": 695 + }, + { + "epoch": 3.6297262059973923, + "grad_norm": 0.18963190902387492, + "learning_rate": 1.6642712218008613e-05, + "loss": 0.742, + "step": 696 + }, + { + "epoch": 3.6349413298565842, + "grad_norm": 0.18716923070682592, + "learning_rate": 1.6524109550014323e-05, + "loss": 0.7515, + "step": 697 + }, + { + "epoch": 3.640156453715776, + "grad_norm": 0.20784228257641144, + "learning_rate": 1.6405820885895242e-05, + "loss": 0.7578, + "step": 698 + }, + { + "epoch": 3.6453715775749673, + "grad_norm": 0.1920361900700138, + "learning_rate": 1.6287847807832e-05, + "loss": 0.7538, + "step": 699 + }, + { + "epoch": 3.6505867014341593, + "grad_norm": 0.1652320731622616, + "learning_rate": 1.617019189378407e-05, + "loss": 0.7511, + "step": 700 + }, + { + "epoch": 3.655801825293351, + "grad_norm": 0.19431451990268614, + "learning_rate": 1.6052854717468644e-05, + "loss": 0.7515, + "step": 701 + }, + { + "epoch": 3.6610169491525424, + "grad_norm": 0.19087250562535932, + "learning_rate": 1.5935837848339667e-05, + "loss": 0.7568, + "step": 702 + }, + { + "epoch": 3.666232073011734, + "grad_norm": 0.20029152301458664, + "learning_rate": 1.581914285156673e-05, + "loss": 0.7465, + "step": 703 + }, + { + "epoch": 3.6714471968709255, + "grad_norm": 0.18770482135472025, + "learning_rate": 1.5702771288014206e-05, + "loss": 0.7516, + "step": 704 + }, + { + "epoch": 3.6766623207301175, + "grad_norm": 0.17463919163255975, + "learning_rate": 1.5586724714220397e-05, + "loss": 0.7498, + "step": 705 + }, + { + "epoch": 3.681877444589309, + "grad_norm": 0.1942555556221587, + "learning_rate": 1.5471004682376626e-05, + "loss": 0.743, + "step": 706 + }, + { + "epoch": 3.6870925684485005, + "grad_norm": 0.1946562516538543, + "learning_rate": 1.535561274030652e-05, + "loss": 0.7569, + "step": 707 + }, + { + "epoch": 3.6923076923076925, + "grad_norm": 0.17582331062620302, + "learning_rate": 1.5240550431445376e-05, + "loss": 0.7519, + "step": 708 + }, + { + "epoch": 3.697522816166884, + "grad_norm": 0.1861426528732765, + "learning_rate": 1.5125819294819391e-05, + "loss": 0.7482, + "step": 709 + }, + { + "epoch": 3.7027379400260756, + "grad_norm": 0.18269562874012577, + "learning_rate": 1.5011420865025143e-05, + "loss": 0.7523, + "step": 710 + }, + { + "epoch": 3.707953063885267, + "grad_norm": 0.18506228544951506, + "learning_rate": 1.4897356672209112e-05, + "loss": 0.7547, + "step": 711 + }, + { + "epoch": 3.7131681877444587, + "grad_norm": 0.20416348855073826, + "learning_rate": 1.4783628242047097e-05, + "loss": 0.7502, + "step": 712 + }, + { + "epoch": 3.7183833116036507, + "grad_norm": 0.16997764410348623, + "learning_rate": 1.467023709572391e-05, + "loss": 0.7512, + "step": 713 + }, + { + "epoch": 3.723598435462842, + "grad_norm": 0.18897244039596858, + "learning_rate": 1.4557184749912981e-05, + "loss": 0.7473, + "step": 714 + }, + { + "epoch": 3.7288135593220337, + "grad_norm": 0.19172065409291622, + "learning_rate": 1.4444472716756074e-05, + "loss": 0.7497, + "step": 715 + }, + { + "epoch": 3.7340286831812257, + "grad_norm": 0.171487618996453, + "learning_rate": 1.433210250384311e-05, + "loss": 0.7526, + "step": 716 + }, + { + "epoch": 3.7392438070404173, + "grad_norm": 0.17442752265085087, + "learning_rate": 1.4220075614191924e-05, + "loss": 0.7515, + "step": 717 + }, + { + "epoch": 3.744458930899609, + "grad_norm": 0.17508720890471086, + "learning_rate": 1.4108393546228194e-05, + "loss": 0.7578, + "step": 718 + }, + { + "epoch": 3.749674054758801, + "grad_norm": 0.19876602903764187, + "learning_rate": 1.3997057793765452e-05, + "loss": 0.7483, + "step": 719 + }, + { + "epoch": 3.7548891786179923, + "grad_norm": 0.17214365448338131, + "learning_rate": 1.3886069845985013e-05, + "loss": 0.7568, + "step": 720 + }, + { + "epoch": 3.760104302477184, + "grad_norm": 0.1920697005442559, + "learning_rate": 1.3775431187416097e-05, + "loss": 0.7504, + "step": 721 + }, + { + "epoch": 3.7653194263363754, + "grad_norm": 0.18503507230287503, + "learning_rate": 1.3665143297916018e-05, + "loss": 0.7636, + "step": 722 + }, + { + "epoch": 3.770534550195567, + "grad_norm": 0.19252969937704967, + "learning_rate": 1.3555207652650308e-05, + "loss": 0.7526, + "step": 723 + }, + { + "epoch": 3.775749674054759, + "grad_norm": 0.1554286223570067, + "learning_rate": 1.3445625722073023e-05, + "loss": 0.7512, + "step": 724 + }, + { + "epoch": 3.7809647979139505, + "grad_norm": 0.16978176727895614, + "learning_rate": 1.3336398971907114e-05, + "loss": 0.7556, + "step": 725 + }, + { + "epoch": 3.786179921773142, + "grad_norm": 0.15542348945118697, + "learning_rate": 1.3227528863124745e-05, + "loss": 0.7512, + "step": 726 + }, + { + "epoch": 3.791395045632334, + "grad_norm": 0.16399605746487972, + "learning_rate": 1.3119016851927798e-05, + "loss": 0.756, + "step": 727 + }, + { + "epoch": 3.7966101694915255, + "grad_norm": 0.16707283374852486, + "learning_rate": 1.301086438972842e-05, + "loss": 0.7475, + "step": 728 + }, + { + "epoch": 3.801825293350717, + "grad_norm": 0.15983694259778197, + "learning_rate": 1.2903072923129547e-05, + "loss": 0.7474, + "step": 729 + }, + { + "epoch": 3.8070404172099086, + "grad_norm": 0.14876964520461528, + "learning_rate": 1.2795643893905575e-05, + "loss": 0.7471, + "step": 730 + }, + { + "epoch": 3.8122555410691, + "grad_norm": 0.19924804420654435, + "learning_rate": 1.2688578738983131e-05, + "loss": 0.7464, + "step": 731 + }, + { + "epoch": 3.817470664928292, + "grad_norm": 0.19191423801767207, + "learning_rate": 1.2581878890421777e-05, + "loss": 0.7562, + "step": 732 + }, + { + "epoch": 3.8226857887874837, + "grad_norm": 0.20765129848679256, + "learning_rate": 1.2475545775394879e-05, + "loss": 0.7565, + "step": 733 + }, + { + "epoch": 3.827900912646675, + "grad_norm": 0.1697957267489159, + "learning_rate": 1.2369580816170563e-05, + "loss": 0.7586, + "step": 734 + }, + { + "epoch": 3.833116036505867, + "grad_norm": 0.1716215315213691, + "learning_rate": 1.2263985430092618e-05, + "loss": 0.7524, + "step": 735 + }, + { + "epoch": 3.8383311603650587, + "grad_norm": 0.18944814607450408, + "learning_rate": 1.2158761029561582e-05, + "loss": 0.7547, + "step": 736 + }, + { + "epoch": 3.8435462842242503, + "grad_norm": 0.14989091598529683, + "learning_rate": 1.2053909022015863e-05, + "loss": 0.7446, + "step": 737 + }, + { + "epoch": 3.8487614080834422, + "grad_norm": 0.1858141017872855, + "learning_rate": 1.194943080991287e-05, + "loss": 0.7503, + "step": 738 + }, + { + "epoch": 3.853976531942634, + "grad_norm": 0.14994883200639647, + "learning_rate": 1.1845327790710276e-05, + "loss": 0.747, + "step": 739 + }, + { + "epoch": 3.8591916558018253, + "grad_norm": 0.14941004189650142, + "learning_rate": 1.1741601356847365e-05, + "loss": 0.7541, + "step": 740 + }, + { + "epoch": 3.864406779661017, + "grad_norm": 0.16071992101640972, + "learning_rate": 1.1638252895726327e-05, + "loss": 0.7496, + "step": 741 + }, + { + "epoch": 3.8696219035202084, + "grad_norm": 0.1374205892222837, + "learning_rate": 1.1535283789693756e-05, + "loss": 0.753, + "step": 742 + }, + { + "epoch": 3.8748370273794004, + "grad_norm": 0.1394433360371365, + "learning_rate": 1.1432695416022158e-05, + "loss": 0.7535, + "step": 743 + }, + { + "epoch": 3.880052151238592, + "grad_norm": 0.16866985633081483, + "learning_rate": 1.1330489146891494e-05, + "loss": 0.7517, + "step": 744 + }, + { + "epoch": 3.8852672750977835, + "grad_norm": 0.15730043046261316, + "learning_rate": 1.1228666349370897e-05, + "loss": 0.7474, + "step": 745 + }, + { + "epoch": 3.8904823989569755, + "grad_norm": 0.1848086766193644, + "learning_rate": 1.112722838540031e-05, + "loss": 0.7615, + "step": 746 + }, + { + "epoch": 3.895697522816167, + "grad_norm": 0.16625357168861077, + "learning_rate": 1.1026176611772286e-05, + "loss": 0.7488, + "step": 747 + }, + { + "epoch": 3.9009126466753585, + "grad_norm": 0.1629255000824365, + "learning_rate": 1.0925512380113892e-05, + "loss": 0.7577, + "step": 748 + }, + { + "epoch": 3.90612777053455, + "grad_norm": 0.18675770122521296, + "learning_rate": 1.0825237036868575e-05, + "loss": 0.7531, + "step": 749 + }, + { + "epoch": 3.9113428943937416, + "grad_norm": 0.15755029689979122, + "learning_rate": 1.0725351923278144e-05, + "loss": 0.7548, + "step": 750 + }, + { + "epoch": 3.9165580182529336, + "grad_norm": 0.18913533326276255, + "learning_rate": 1.0625858375364917e-05, + "loss": 0.754, + "step": 751 + }, + { + "epoch": 3.921773142112125, + "grad_norm": 0.18170716935082054, + "learning_rate": 1.0526757723913735e-05, + "loss": 0.7441, + "step": 752 + }, + { + "epoch": 3.9269882659713167, + "grad_norm": 0.13608488064261914, + "learning_rate": 1.042805129445423e-05, + "loss": 0.7508, + "step": 753 + }, + { + "epoch": 3.9322033898305087, + "grad_norm": 0.17632467806971025, + "learning_rate": 1.0329740407243105e-05, + "loss": 0.7747, + "step": 754 + }, + { + "epoch": 3.9374185136897, + "grad_norm": 0.17072961754308208, + "learning_rate": 1.0231826377246437e-05, + "loss": 0.7462, + "step": 755 + }, + { + "epoch": 3.9426336375488917, + "grad_norm": 0.16478645139827036, + "learning_rate": 1.0134310514122082e-05, + "loss": 0.7563, + "step": 756 + }, + { + "epoch": 3.9478487614080837, + "grad_norm": 0.17542376854887573, + "learning_rate": 1.0037194122202227e-05, + "loss": 0.7567, + "step": 757 + }, + { + "epoch": 3.9530638852672753, + "grad_norm": 0.17159268136488617, + "learning_rate": 9.940478500475858e-06, + "loss": 0.7503, + "step": 758 + }, + { + "epoch": 3.958279009126467, + "grad_norm": 0.15300026372913983, + "learning_rate": 9.844164942571424e-06, + "loss": 0.7523, + "step": 759 + }, + { + "epoch": 3.9634941329856583, + "grad_norm": 0.159680621058249, + "learning_rate": 9.748254736739571e-06, + "loss": 0.757, + "step": 760 + }, + { + "epoch": 3.96870925684485, + "grad_norm": 0.15768593487660648, + "learning_rate": 9.652749165835828e-06, + "loss": 0.7458, + "step": 761 + }, + { + "epoch": 3.973924380704042, + "grad_norm": 0.16943127341342135, + "learning_rate": 9.557649507303508e-06, + "loss": 0.7456, + "step": 762 + }, + { + "epoch": 3.9791395045632334, + "grad_norm": 0.14842747623637131, + "learning_rate": 9.462957033156632e-06, + "loss": 0.7497, + "step": 763 + }, + { + "epoch": 3.984354628422425, + "grad_norm": 0.1501263870664962, + "learning_rate": 9.368673009962852e-06, + "loss": 0.7514, + "step": 764 + }, + { + "epoch": 3.989569752281617, + "grad_norm": 0.17717065342340108, + "learning_rate": 9.274798698826557e-06, + "loss": 0.7531, + "step": 765 + }, + { + "epoch": 3.9947848761408085, + "grad_norm": 0.15866607398462376, + "learning_rate": 9.181335355372028e-06, + "loss": 0.7512, + "step": 766 + }, + { + "epoch": 4.0, + "grad_norm": 0.17952277081095838, + "learning_rate": 9.088284229726572e-06, + "loss": 0.9412, + "step": 767 + }, + { + "epoch": 4.005215123859192, + "grad_norm": 0.23274451897753248, + "learning_rate": 8.995646566503838e-06, + "loss": 0.7375, + "step": 768 + }, + { + "epoch": 4.010430247718383, + "grad_norm": 0.17288112228894242, + "learning_rate": 8.90342360478723e-06, + "loss": 0.7373, + "step": 769 + }, + { + "epoch": 4.015645371577575, + "grad_norm": 0.20174270563988336, + "learning_rate": 8.81161657811318e-06, + "loss": 0.7453, + "step": 770 + }, + { + "epoch": 4.020860495436767, + "grad_norm": 0.1726715461295719, + "learning_rate": 8.720226714454831e-06, + "loss": 0.7421, + "step": 771 + }, + { + "epoch": 4.026075619295958, + "grad_norm": 0.20664612564398913, + "learning_rate": 8.629255236205475e-06, + "loss": 0.7393, + "step": 772 + }, + { + "epoch": 4.03129074315515, + "grad_norm": 0.183032821290067, + "learning_rate": 8.538703360162267e-06, + "loss": 0.7262, + "step": 773 + }, + { + "epoch": 4.036505867014341, + "grad_norm": 0.19162204178734604, + "learning_rate": 8.44857229750994e-06, + "loss": 0.7229, + "step": 774 + }, + { + "epoch": 4.041720990873533, + "grad_norm": 0.19714648419474126, + "learning_rate": 8.3588632538046e-06, + "loss": 0.7406, + "step": 775 + }, + { + "epoch": 4.046936114732725, + "grad_norm": 0.18562646027448648, + "learning_rate": 8.269577428957571e-06, + "loss": 0.7375, + "step": 776 + }, + { + "epoch": 4.052151238591916, + "grad_norm": 0.19319081667416735, + "learning_rate": 8.180716017219433e-06, + "loss": 0.7315, + "step": 777 + }, + { + "epoch": 4.057366362451108, + "grad_norm": 0.17933639021135425, + "learning_rate": 8.092280207163941e-06, + "loss": 0.7439, + "step": 778 + }, + { + "epoch": 4.0625814863103, + "grad_norm": 0.16796265908006924, + "learning_rate": 8.004271181672178e-06, + "loss": 0.7219, + "step": 779 + }, + { + "epoch": 4.067796610169491, + "grad_norm": 0.16506690334266957, + "learning_rate": 7.916690117916772e-06, + "loss": 0.7409, + "step": 780 + }, + { + "epoch": 4.073011734028683, + "grad_norm": 0.16443977378096947, + "learning_rate": 7.829538187346077e-06, + "loss": 0.7377, + "step": 781 + }, + { + "epoch": 4.078226857887874, + "grad_norm": 0.1997012117535286, + "learning_rate": 7.742816555668532e-06, + "loss": 0.7399, + "step": 782 + }, + { + "epoch": 4.083441981747066, + "grad_norm": 0.14695927923208477, + "learning_rate": 7.656526382837106e-06, + "loss": 0.7327, + "step": 783 + }, + { + "epoch": 4.088657105606258, + "grad_norm": 0.1799749838620325, + "learning_rate": 7.570668823033727e-06, + "loss": 0.7309, + "step": 784 + }, + { + "epoch": 4.0938722294654495, + "grad_norm": 0.17989297489241943, + "learning_rate": 7.4852450246538685e-06, + "loss": 0.7343, + "step": 785 + }, + { + "epoch": 4.0990873533246415, + "grad_norm": 0.15506025585209593, + "learning_rate": 7.40025613029121e-06, + "loss": 0.736, + "step": 786 + }, + { + "epoch": 4.1043024771838335, + "grad_norm": 0.18105126244534, + "learning_rate": 7.315703276722317e-06, + "loss": 0.7342, + "step": 787 + }, + { + "epoch": 4.1095176010430245, + "grad_norm": 0.19050717050961422, + "learning_rate": 7.231587594891438e-06, + "loss": 0.7458, + "step": 788 + }, + { + "epoch": 4.1147327249022165, + "grad_norm": 0.14216335950493533, + "learning_rate": 7.147910209895435e-06, + "loss": 0.7318, + "step": 789 + }, + { + "epoch": 4.1199478487614085, + "grad_norm": 0.1655844728014387, + "learning_rate": 7.064672240968638e-06, + "loss": 0.7442, + "step": 790 + }, + { + "epoch": 4.1251629726206, + "grad_norm": 0.14883381171151777, + "learning_rate": 6.9818748014679785e-06, + "loss": 0.7403, + "step": 791 + }, + { + "epoch": 4.130378096479792, + "grad_norm": 0.1602685057342838, + "learning_rate": 6.899518998858017e-06, + "loss": 0.7405, + "step": 792 + }, + { + "epoch": 4.135593220338983, + "grad_norm": 0.1649659601993552, + "learning_rate": 6.81760593469615e-06, + "loss": 0.7346, + "step": 793 + }, + { + "epoch": 4.140808344198175, + "grad_norm": 0.12921893028933829, + "learning_rate": 6.7361367046179325e-06, + "loss": 0.7303, + "step": 794 + }, + { + "epoch": 4.146023468057367, + "grad_norm": 0.12489877860585255, + "learning_rate": 6.655112398322332e-06, + "loss": 0.731, + "step": 795 + }, + { + "epoch": 4.151238591916558, + "grad_norm": 0.17428621687246298, + "learning_rate": 6.574534099557204e-06, + "loss": 0.7307, + "step": 796 + }, + { + "epoch": 4.15645371577575, + "grad_norm": 0.1251987161714777, + "learning_rate": 6.49440288610482e-06, + "loss": 0.7275, + "step": 797 + }, + { + "epoch": 4.161668839634942, + "grad_norm": 0.15533418379259104, + "learning_rate": 6.414719829767406e-06, + "loss": 0.7339, + "step": 798 + }, + { + "epoch": 4.166883963494133, + "grad_norm": 0.16236363314856642, + "learning_rate": 6.335485996352782e-06, + "loss": 0.7418, + "step": 799 + }, + { + "epoch": 4.172099087353325, + "grad_norm": 0.13801981121095086, + "learning_rate": 6.256702445660221e-06, + "loss": 0.7304, + "step": 800 + }, + { + "epoch": 4.177314211212516, + "grad_norm": 0.14142425026968236, + "learning_rate": 6.178370231466142e-06, + "loss": 0.743, + "step": 801 + }, + { + "epoch": 4.182529335071708, + "grad_norm": 0.16216801700272557, + "learning_rate": 6.100490401510089e-06, + "loss": 0.7437, + "step": 802 + }, + { + "epoch": 4.1877444589309, + "grad_norm": 0.14030637701048476, + "learning_rate": 6.0230639974807206e-06, + "loss": 0.7383, + "step": 803 + }, + { + "epoch": 4.192959582790091, + "grad_norm": 0.13754640300037912, + "learning_rate": 5.946092055001824e-06, + "loss": 0.7329, + "step": 804 + }, + { + "epoch": 4.198174706649283, + "grad_norm": 0.16139415971113086, + "learning_rate": 5.869575603618507e-06, + "loss": 0.7417, + "step": 805 + }, + { + "epoch": 4.203389830508475, + "grad_norm": 0.15626356140455866, + "learning_rate": 5.793515666783429e-06, + "loss": 0.7337, + "step": 806 + }, + { + "epoch": 4.208604954367666, + "grad_norm": 0.12177414306535114, + "learning_rate": 5.717913261843078e-06, + "loss": 0.733, + "step": 807 + }, + { + "epoch": 4.213820078226858, + "grad_norm": 0.12926025156980447, + "learning_rate": 5.642769400024199e-06, + "loss": 0.7422, + "step": 808 + }, + { + "epoch": 4.219035202086049, + "grad_norm": 0.13704261747895924, + "learning_rate": 5.568085086420247e-06, + "loss": 0.7404, + "step": 809 + }, + { + "epoch": 4.224250325945241, + "grad_norm": 0.13842188890961235, + "learning_rate": 5.4938613199779334e-06, + "loss": 0.7334, + "step": 810 + }, + { + "epoch": 4.229465449804433, + "grad_norm": 0.13979962035636995, + "learning_rate": 5.420099093483911e-06, + "loss": 0.7222, + "step": 811 + }, + { + "epoch": 4.234680573663624, + "grad_norm": 0.12565163857223183, + "learning_rate": 5.346799393551436e-06, + "loss": 0.7307, + "step": 812 + }, + { + "epoch": 4.239895697522816, + "grad_norm": 0.12679243834545414, + "learning_rate": 5.273963200607206e-06, + "loss": 0.7387, + "step": 813 + }, + { + "epoch": 4.245110821382008, + "grad_norm": 0.13813903814909753, + "learning_rate": 5.201591488878257e-06, + "loss": 0.7251, + "step": 814 + }, + { + "epoch": 4.250325945241199, + "grad_norm": 0.1285263356419264, + "learning_rate": 5.129685226378893e-06, + "loss": 0.7234, + "step": 815 + }, + { + "epoch": 4.255541069100391, + "grad_norm": 0.12820199717697753, + "learning_rate": 5.058245374897763e-06, + "loss": 0.7271, + "step": 816 + }, + { + "epoch": 4.260756192959583, + "grad_norm": 0.12495602397125814, + "learning_rate": 4.987272889985009e-06, + "loss": 0.7393, + "step": 817 + }, + { + "epoch": 4.265971316818774, + "grad_norm": 0.15151715317554978, + "learning_rate": 4.916768720939451e-06, + "loss": 0.7332, + "step": 818 + }, + { + "epoch": 4.271186440677966, + "grad_norm": 0.14747133461014475, + "learning_rate": 4.846733810795914e-06, + "loss": 0.7297, + "step": 819 + }, + { + "epoch": 4.276401564537157, + "grad_norm": 0.14100514701144612, + "learning_rate": 4.777169096312624e-06, + "loss": 0.7276, + "step": 820 + }, + { + "epoch": 4.281616688396349, + "grad_norm": 0.1646113420856527, + "learning_rate": 4.708075507958642e-06, + "loss": 0.7281, + "step": 821 + }, + { + "epoch": 4.286831812255541, + "grad_norm": 0.13983624281016735, + "learning_rate": 4.639453969901442e-06, + "loss": 0.7276, + "step": 822 + }, + { + "epoch": 4.292046936114732, + "grad_norm": 0.13237794200876413, + "learning_rate": 4.571305399994561e-06, + "loss": 0.7409, + "step": 823 + }, + { + "epoch": 4.297262059973924, + "grad_norm": 0.180777223722747, + "learning_rate": 4.503630709765294e-06, + "loss": 0.741, + "step": 824 + }, + { + "epoch": 4.302477183833116, + "grad_norm": 0.1367694718240071, + "learning_rate": 4.436430804402521e-06, + "loss": 0.7331, + "step": 825 + }, + { + "epoch": 4.3076923076923075, + "grad_norm": 0.15547588436169807, + "learning_rate": 4.3697065827446e-06, + "loss": 0.7376, + "step": 826 + }, + { + "epoch": 4.3129074315514995, + "grad_norm": 0.17342743882348932, + "learning_rate": 4.303458937267326e-06, + "loss": 0.7267, + "step": 827 + }, + { + "epoch": 4.318122555410691, + "grad_norm": 0.16682655867364918, + "learning_rate": 4.237688754072022e-06, + "loss": 0.7309, + "step": 828 + }, + { + "epoch": 4.3233376792698825, + "grad_norm": 0.16450623559503522, + "learning_rate": 4.172396912873659e-06, + "loss": 0.7342, + "step": 829 + }, + { + "epoch": 4.3285528031290745, + "grad_norm": 0.19380979384775532, + "learning_rate": 4.107584286989097e-06, + "loss": 0.7392, + "step": 830 + }, + { + "epoch": 4.333767926988266, + "grad_norm": 0.18721173933947008, + "learning_rate": 4.043251743325436e-06, + "loss": 0.7303, + "step": 831 + }, + { + "epoch": 4.338983050847458, + "grad_norm": 0.15835333882238062, + "learning_rate": 3.979400142368368e-06, + "loss": 0.7301, + "step": 832 + }, + { + "epoch": 4.34419817470665, + "grad_norm": 0.18385561342339599, + "learning_rate": 3.916030338170696e-06, + "loss": 0.7374, + "step": 833 + }, + { + "epoch": 4.349413298565841, + "grad_norm": 0.2022205872843893, + "learning_rate": 3.8531431783409165e-06, + "loss": 0.7387, + "step": 834 + }, + { + "epoch": 4.354628422425033, + "grad_norm": 0.15316197738525242, + "learning_rate": 3.7907395040318685e-06, + "loss": 0.7385, + "step": 835 + }, + { + "epoch": 4.359843546284225, + "grad_norm": 0.1500484412205804, + "learning_rate": 3.7288201499294797e-06, + "loss": 0.7268, + "step": 836 + }, + { + "epoch": 4.365058670143416, + "grad_norm": 0.14189535258541974, + "learning_rate": 3.667385944241626e-06, + "loss": 0.7453, + "step": 837 + }, + { + "epoch": 4.370273794002608, + "grad_norm": 0.15984630251802126, + "learning_rate": 3.606437708687023e-06, + "loss": 0.7395, + "step": 838 + }, + { + "epoch": 4.375488917861799, + "grad_norm": 0.1642822923840156, + "learning_rate": 3.545976258484247e-06, + "loss": 0.7354, + "step": 839 + }, + { + "epoch": 4.380704041720991, + "grad_norm": 0.13221533894356902, + "learning_rate": 3.4860024023408577e-06, + "loss": 0.7459, + "step": 840 + }, + { + "epoch": 4.385919165580183, + "grad_norm": 0.18318603173559903, + "learning_rate": 3.426516942442524e-06, + "loss": 0.7509, + "step": 841 + }, + { + "epoch": 4.391134289439374, + "grad_norm": 0.17040203525878148, + "learning_rate": 3.3675206744423394e-06, + "loss": 0.7375, + "step": 842 + }, + { + "epoch": 4.396349413298566, + "grad_norm": 0.157205271311182, + "learning_rate": 3.309014387450189e-06, + "loss": 0.7302, + "step": 843 + }, + { + "epoch": 4.401564537157758, + "grad_norm": 0.15260420463119376, + "learning_rate": 3.2509988640221414e-06, + "loss": 0.7407, + "step": 844 + }, + { + "epoch": 4.406779661016949, + "grad_norm": 0.16154440296062286, + "learning_rate": 3.1934748801500184e-06, + "loss": 0.7255, + "step": 845 + }, + { + "epoch": 4.411994784876141, + "grad_norm": 0.14388929124722757, + "learning_rate": 3.136443205251034e-06, + "loss": 0.7369, + "step": 846 + }, + { + "epoch": 4.417209908735332, + "grad_norm": 0.14793139417234433, + "learning_rate": 3.0799046021574574e-06, + "loss": 0.7337, + "step": 847 + }, + { + "epoch": 4.422425032594524, + "grad_norm": 0.14906886650426293, + "learning_rate": 3.0238598271064323e-06, + "loss": 0.7319, + "step": 848 + }, + { + "epoch": 4.427640156453716, + "grad_norm": 0.18885017044608088, + "learning_rate": 2.9683096297298753e-06, + "loss": 0.7226, + "step": 849 + }, + { + "epoch": 4.432855280312907, + "grad_norm": 0.16698890562230234, + "learning_rate": 2.9132547530444254e-06, + "loss": 0.74, + "step": 850 + }, + { + "epoch": 4.438070404172099, + "grad_norm": 0.12976275259498832, + "learning_rate": 2.8586959334415064e-06, + "loss": 0.7325, + "step": 851 + }, + { + "epoch": 4.443285528031291, + "grad_norm": 0.14807620276113978, + "learning_rate": 2.8046339006775115e-06, + "loss": 0.7197, + "step": 852 + }, + { + "epoch": 4.448500651890482, + "grad_norm": 0.15235231363332793, + "learning_rate": 2.751069377863993e-06, + "loss": 0.7302, + "step": 853 + }, + { + "epoch": 4.453715775749674, + "grad_norm": 0.15202060711926385, + "learning_rate": 2.698003081458018e-06, + "loss": 0.7289, + "step": 854 + }, + { + "epoch": 4.458930899608866, + "grad_norm": 0.12630616368200856, + "learning_rate": 2.6454357212526026e-06, + "loss": 0.7279, + "step": 855 + }, + { + "epoch": 4.464146023468057, + "grad_norm": 0.11739628579126643, + "learning_rate": 2.5933680003671625e-06, + "loss": 0.7335, + "step": 856 + }, + { + "epoch": 4.469361147327249, + "grad_norm": 0.1434952804888338, + "learning_rate": 2.541800615238175e-06, + "loss": 0.7266, + "step": 857 + }, + { + "epoch": 4.47457627118644, + "grad_norm": 0.176277045965935, + "learning_rate": 2.4907342556098125e-06, + "loss": 0.7269, + "step": 858 + }, + { + "epoch": 4.479791395045632, + "grad_norm": 0.18491464302255675, + "learning_rate": 2.440169604524738e-06, + "loss": 0.7453, + "step": 859 + }, + { + "epoch": 4.485006518904824, + "grad_norm": 0.12415001369259007, + "learning_rate": 2.390107338314982e-06, + "loss": 0.7319, + "step": 860 + }, + { + "epoch": 4.490221642764015, + "grad_norm": 0.10478588253479926, + "learning_rate": 2.340548126592861e-06, + "loss": 0.737, + "step": 861 + }, + { + "epoch": 4.495436766623207, + "grad_norm": 0.14165809120057415, + "learning_rate": 2.2914926322420515e-06, + "loss": 0.7343, + "step": 862 + }, + { + "epoch": 4.500651890482399, + "grad_norm": 0.11590106390711405, + "learning_rate": 2.242941511408714e-06, + "loss": 0.737, + "step": 863 + }, + { + "epoch": 4.50586701434159, + "grad_norm": 0.12634034846215603, + "learning_rate": 2.1948954134927103e-06, + "loss": 0.7377, + "step": 864 + }, + { + "epoch": 4.511082138200782, + "grad_norm": 0.11268712032715957, + "learning_rate": 2.147354981138925e-06, + "loss": 0.7257, + "step": 865 + }, + { + "epoch": 4.5162972620599735, + "grad_norm": 0.11916219994594991, + "learning_rate": 2.1003208502286744e-06, + "loss": 0.729, + "step": 866 + }, + { + "epoch": 4.5215123859191655, + "grad_norm": 0.13119722935564196, + "learning_rate": 2.0537936498711898e-06, + "loss": 0.7316, + "step": 867 + }, + { + "epoch": 4.5267275097783575, + "grad_norm": 0.12883622307786546, + "learning_rate": 2.007774002395193e-06, + "loss": 0.7337, + "step": 868 + }, + { + "epoch": 4.531942633637549, + "grad_norm": 0.10574304590110693, + "learning_rate": 1.9622625233406143e-06, + "loss": 0.7268, + "step": 869 + }, + { + "epoch": 4.5371577574967406, + "grad_norm": 0.11189460238334994, + "learning_rate": 1.9172598214503146e-06, + "loss": 0.736, + "step": 870 + }, + { + "epoch": 4.5423728813559325, + "grad_norm": 0.12169459450452168, + "learning_rate": 1.8727664986619576e-06, + "loss": 0.7359, + "step": 871 + }, + { + "epoch": 4.547588005215124, + "grad_norm": 0.11602787776789947, + "learning_rate": 1.8287831500999775e-06, + "loss": 0.733, + "step": 872 + }, + { + "epoch": 4.552803129074316, + "grad_norm": 0.11542784974890338, + "learning_rate": 1.7853103640675852e-06, + "loss": 0.7274, + "step": 873 + }, + { + "epoch": 4.558018252933508, + "grad_norm": 0.10042034550407807, + "learning_rate": 1.74234872203892e-06, + "loss": 0.7346, + "step": 874 + }, + { + "epoch": 4.563233376792699, + "grad_norm": 0.11267524138272358, + "learning_rate": 1.699898798651285e-06, + "loss": 0.7355, + "step": 875 + }, + { + "epoch": 4.568448500651891, + "grad_norm": 0.14020374405591798, + "learning_rate": 1.6579611616974212e-06, + "loss": 0.7323, + "step": 876 + }, + { + "epoch": 4.573663624511082, + "grad_norm": 0.1329010216092144, + "learning_rate": 1.6165363721179471e-06, + "loss": 0.7333, + "step": 877 + }, + { + "epoch": 4.578878748370274, + "grad_norm": 0.1111745646565014, + "learning_rate": 1.575624983993853e-06, + "loss": 0.7198, + "step": 878 + }, + { + "epoch": 4.584093872229466, + "grad_norm": 0.12252637821649354, + "learning_rate": 1.5352275445390752e-06, + "loss": 0.7265, + "step": 879 + }, + { + "epoch": 4.589308996088657, + "grad_norm": 0.10588650578235918, + "learning_rate": 1.4953445940931688e-06, + "loss": 0.7328, + "step": 880 + }, + { + "epoch": 4.594524119947849, + "grad_norm": 0.10242904609486453, + "learning_rate": 1.4559766661141317e-06, + "loss": 0.7305, + "step": 881 + }, + { + "epoch": 4.599739243807041, + "grad_norm": 0.11484087780107081, + "learning_rate": 1.4171242871711922e-06, + "loss": 0.7432, + "step": 882 + }, + { + "epoch": 4.604954367666232, + "grad_norm": 0.10614866491620119, + "learning_rate": 1.3787879769378277e-06, + "loss": 0.7317, + "step": 883 + }, + { + "epoch": 4.610169491525424, + "grad_norm": 0.11216621248162341, + "learning_rate": 1.3409682481848063e-06, + "loss": 0.7379, + "step": 884 + }, + { + "epoch": 4.615384615384615, + "grad_norm": 0.09993770692863328, + "learning_rate": 1.3036656067732679e-06, + "loss": 0.739, + "step": 885 + }, + { + "epoch": 4.620599739243807, + "grad_norm": 0.09930632699157292, + "learning_rate": 1.2668805516480577e-06, + "loss": 0.743, + "step": 886 + }, + { + "epoch": 4.625814863102999, + "grad_norm": 0.1008961163196706, + "learning_rate": 1.2306135748309633e-06, + "loss": 0.746, + "step": 887 + }, + { + "epoch": 4.63102998696219, + "grad_norm": 0.1255524490616452, + "learning_rate": 1.1948651614141783e-06, + "loss": 0.7317, + "step": 888 + }, + { + "epoch": 4.636245110821382, + "grad_norm": 0.10155914604030952, + "learning_rate": 1.1596357895538213e-06, + "loss": 0.7213, + "step": 889 + }, + { + "epoch": 4.641460234680574, + "grad_norm": 0.10501816031109035, + "learning_rate": 1.1249259304634986e-06, + "loss": 0.7398, + "step": 890 + }, + { + "epoch": 4.646675358539765, + "grad_norm": 0.13819236460373796, + "learning_rate": 1.0907360484080409e-06, + "loss": 0.7368, + "step": 891 + }, + { + "epoch": 4.651890482398957, + "grad_norm": 0.10830515113482954, + "learning_rate": 1.0570666006972875e-06, + "loss": 0.7543, + "step": 892 + }, + { + "epoch": 4.657105606258149, + "grad_norm": 0.10654993027327486, + "learning_rate": 1.023918037679943e-06, + "loss": 0.7243, + "step": 893 + }, + { + "epoch": 4.66232073011734, + "grad_norm": 0.10012139135715376, + "learning_rate": 9.912908027375745e-07, + "loss": 0.7373, + "step": 894 + }, + { + "epoch": 4.667535853976532, + "grad_norm": 0.0950421289561811, + "learning_rate": 9.591853322787003e-07, + "loss": 0.7412, + "step": 895 + }, + { + "epoch": 4.672750977835723, + "grad_norm": 0.09904664343265647, + "learning_rate": 9.276020557329101e-07, + "loss": 0.729, + "step": 896 + }, + { + "epoch": 4.677966101694915, + "grad_norm": 0.1051253462879674, + "learning_rate": 8.965413955451363e-07, + "loss": 0.7329, + "step": 897 + }, + { + "epoch": 4.683181225554107, + "grad_norm": 0.10272032718312339, + "learning_rate": 8.660037671700405e-07, + "loss": 0.7387, + "step": 898 + }, + { + "epoch": 4.688396349413298, + "grad_norm": 0.09813662944001211, + "learning_rate": 8.359895790663963e-07, + "loss": 0.7377, + "step": 899 + }, + { + "epoch": 4.69361147327249, + "grad_norm": 0.10536763446890161, + "learning_rate": 8.064992326916577e-07, + "loss": 0.7353, + "step": 900 + }, + { + "epoch": 4.698826597131681, + "grad_norm": 0.10236023154880665, + "learning_rate": 7.775331224966076e-07, + "loss": 0.7272, + "step": 901 + }, + { + "epoch": 4.704041720990873, + "grad_norm": 0.09414675460502064, + "learning_rate": 7.49091635920034e-07, + "loss": 0.7292, + "step": 902 + }, + { + "epoch": 4.709256844850065, + "grad_norm": 0.0943501127919075, + "learning_rate": 7.211751533835998e-07, + "loss": 0.7416, + "step": 903 + }, + { + "epoch": 4.7144719687092564, + "grad_norm": 0.1021023935601688, + "learning_rate": 6.937840482867142e-07, + "loss": 0.7337, + "step": 904 + }, + { + "epoch": 4.719687092568448, + "grad_norm": 0.09881423844437276, + "learning_rate": 6.669186870015454e-07, + "loss": 0.7324, + "step": 905 + }, + { + "epoch": 4.72490221642764, + "grad_norm": 0.09987028627197758, + "learning_rate": 6.405794288681577e-07, + "loss": 0.7374, + "step": 906 + }, + { + "epoch": 4.7301173402868315, + "grad_norm": 0.09574731953052935, + "learning_rate": 6.147666261896445e-07, + "loss": 0.7326, + "step": 907 + }, + { + "epoch": 4.7353324641460235, + "grad_norm": 0.09726177695264816, + "learning_rate": 5.894806242274565e-07, + "loss": 0.7375, + "step": 908 + }, + { + "epoch": 4.7405475880052155, + "grad_norm": 0.09663506260906218, + "learning_rate": 5.647217611967914e-07, + "loss": 0.7366, + "step": 909 + }, + { + "epoch": 4.745762711864407, + "grad_norm": 0.09898203274690626, + "learning_rate": 5.404903682620299e-07, + "loss": 0.7282, + "step": 910 + }, + { + "epoch": 4.7509778357235986, + "grad_norm": 0.09888749932985194, + "learning_rate": 5.167867695323426e-07, + "loss": 0.7234, + "step": 911 + }, + { + "epoch": 4.7561929595827905, + "grad_norm": 0.09677300838881348, + "learning_rate": 4.936112820573335e-07, + "loss": 0.734, + "step": 912 + }, + { + "epoch": 4.761408083441982, + "grad_norm": 0.09338358111347181, + "learning_rate": 4.709642158228134e-07, + "loss": 0.7319, + "step": 913 + }, + { + "epoch": 4.766623207301174, + "grad_norm": 0.09475677222529419, + "learning_rate": 4.4884587374663727e-07, + "loss": 0.7345, + "step": 914 + }, + { + "epoch": 4.771838331160365, + "grad_norm": 0.09660842094296682, + "learning_rate": 4.2725655167468626e-07, + "loss": 0.7328, + "step": 915 + }, + { + "epoch": 4.777053455019557, + "grad_norm": 0.09522844691697419, + "learning_rate": 4.061965383768529e-07, + "loss": 0.7395, + "step": 916 + }, + { + "epoch": 4.782268578878749, + "grad_norm": 0.09658310921334537, + "learning_rate": 3.856661155432484e-07, + "loss": 0.7282, + "step": 917 + }, + { + "epoch": 4.78748370273794, + "grad_norm": 0.09209868347782292, + "learning_rate": 3.656655577803703e-07, + "loss": 0.7321, + "step": 918 + }, + { + "epoch": 4.792698826597132, + "grad_norm": 0.09809162403374437, + "learning_rate": 3.4619513260746973e-07, + "loss": 0.721, + "step": 919 + }, + { + "epoch": 4.797913950456323, + "grad_norm": 0.09728296962426303, + "learning_rate": 3.2725510045297225e-07, + "loss": 0.7383, + "step": 920 + }, + { + "epoch": 4.803129074315515, + "grad_norm": 0.0982190410036559, + "learning_rate": 3.0884571465096936e-07, + "loss": 0.7356, + "step": 921 + }, + { + "epoch": 4.808344198174707, + "grad_norm": 0.09138351849890962, + "learning_rate": 2.909672214378434e-07, + "loss": 0.7375, + "step": 922 + }, + { + "epoch": 4.813559322033898, + "grad_norm": 0.08931700971308501, + "learning_rate": 2.736198599489903e-07, + "loss": 0.7399, + "step": 923 + }, + { + "epoch": 4.81877444589309, + "grad_norm": 0.09559406138911163, + "learning_rate": 2.568038622155955e-07, + "loss": 0.7509, + "step": 924 + }, + { + "epoch": 4.823989569752282, + "grad_norm": 0.09150057561208538, + "learning_rate": 2.4051945316153846e-07, + "loss": 0.7284, + "step": 925 + }, + { + "epoch": 4.829204693611473, + "grad_norm": 0.09349065890924081, + "learning_rate": 2.2476685060039527e-07, + "loss": 0.7299, + "step": 926 + }, + { + "epoch": 4.834419817470665, + "grad_norm": 0.09556536713909304, + "learning_rate": 2.0954626523251197e-07, + "loss": 0.7276, + "step": 927 + }, + { + "epoch": 4.839634941329857, + "grad_norm": 0.09471713249250732, + "learning_rate": 1.948579006421847e-07, + "loss": 0.7428, + "step": 928 + }, + { + "epoch": 4.844850065189048, + "grad_norm": 0.09109503005212934, + "learning_rate": 1.8070195329495944e-07, + "loss": 0.7401, + "step": 929 + }, + { + "epoch": 4.85006518904824, + "grad_norm": 0.09359860047319951, + "learning_rate": 1.6707861253497214e-07, + "loss": 0.7409, + "step": 930 + }, + { + "epoch": 4.855280312907432, + "grad_norm": 0.09252775154456866, + "learning_rate": 1.539880605824351e-07, + "loss": 0.7362, + "step": 931 + }, + { + "epoch": 4.860495436766623, + "grad_norm": 0.09910390005741515, + "learning_rate": 1.4143047253120322e-07, + "loss": 0.7332, + "step": 932 + }, + { + "epoch": 4.865710560625815, + "grad_norm": 0.0900655265782037, + "learning_rate": 1.2940601634642059e-07, + "loss": 0.7233, + "step": 933 + }, + { + "epoch": 4.870925684485006, + "grad_norm": 0.09409378931281938, + "learning_rate": 1.1791485286227311e-07, + "loss": 0.7395, + "step": 934 + }, + { + "epoch": 4.876140808344198, + "grad_norm": 0.10015850940263663, + "learning_rate": 1.0695713577984824e-07, + "loss": 0.7231, + "step": 935 + }, + { + "epoch": 4.88135593220339, + "grad_norm": 0.0903358412925811, + "learning_rate": 9.653301166507422e-08, + "loss": 0.7351, + "step": 936 + }, + { + "epoch": 4.886571056062581, + "grad_norm": 0.0902382557022594, + "learning_rate": 8.664261994675738e-08, + "loss": 0.7333, + "step": 937 + }, + { + "epoch": 4.891786179921773, + "grad_norm": 0.09428682569360029, + "learning_rate": 7.728609291471678e-08, + "loss": 0.7353, + "step": 938 + }, + { + "epoch": 4.897001303780964, + "grad_norm": 0.09406554806150552, + "learning_rate": 6.846355571801688e-08, + "loss": 0.7387, + "step": 939 + }, + { + "epoch": 4.902216427640156, + "grad_norm": 0.10257884244588195, + "learning_rate": 6.017512636329325e-08, + "loss": 0.7327, + "step": 940 + }, + { + "epoch": 4.907431551499348, + "grad_norm": 0.08993907897520984, + "learning_rate": 5.242091571318053e-08, + "loss": 0.7323, + "step": 941 + }, + { + "epoch": 4.912646675358539, + "grad_norm": 0.09752580757548222, + "learning_rate": 4.5201027484811365e-08, + "loss": 0.7369, + "step": 942 + }, + { + "epoch": 4.917861799217731, + "grad_norm": 0.09295306846482325, + "learning_rate": 3.851555824844866e-08, + "loss": 0.7269, + "step": 943 + }, + { + "epoch": 4.923076923076923, + "grad_norm": 0.09090791000944588, + "learning_rate": 3.23645974261888e-08, + "loss": 0.72, + "step": 944 + }, + { + "epoch": 4.9282920469361144, + "grad_norm": 0.0941176803792802, + "learning_rate": 2.674822729075377e-08, + "loss": 0.7302, + "step": 945 + }, + { + "epoch": 4.933507170795306, + "grad_norm": 0.09476591232745427, + "learning_rate": 2.166652296440752e-08, + "loss": 0.7389, + "step": 946 + }, + { + "epoch": 4.938722294654498, + "grad_norm": 0.09487039421752298, + "learning_rate": 1.7119552417943496e-08, + "loss": 0.7384, + "step": 947 + }, + { + "epoch": 4.9439374185136895, + "grad_norm": 0.09350946727423737, + "learning_rate": 1.3107376469769783e-08, + "loss": 0.7352, + "step": 948 + }, + { + "epoch": 4.9491525423728815, + "grad_norm": 0.09008974287657726, + "learning_rate": 9.630048785105318e-09, + "loss": 0.7202, + "step": 949 + }, + { + "epoch": 4.9543676662320735, + "grad_norm": 0.09424907250830968, + "learning_rate": 6.687615875264897e-09, + "loss": 0.7352, + "step": 950 + }, + { + "epoch": 4.959582790091265, + "grad_norm": 0.08996524126321254, + "learning_rate": 4.280117097015257e-09, + "loss": 0.7342, + "step": 951 + }, + { + "epoch": 4.9647979139504566, + "grad_norm": 0.09830845644976817, + "learning_rate": 2.407584652073247e-09, + "loss": 0.7357, + "step": 952 + }, + { + "epoch": 4.970013037809648, + "grad_norm": 0.0969409044778751, + "learning_rate": 1.0700435866706216e-09, + "loss": 0.7212, + "step": 953 + }, + { + "epoch": 4.97522816166884, + "grad_norm": 0.09285007352436725, + "learning_rate": 2.6751179119877124e-10, + "loss": 0.7353, + "step": 954 + }, + { + "epoch": 4.980443285528032, + "grad_norm": 0.09324700428805709, + "learning_rate": 0.0, + "loss": 0.7386, + "step": 955 + }, + { + "epoch": 4.980443285528032, + "step": 955, + "total_flos": 1.9663195796108476e+19, + "train_loss": 0.7990808349005215, + "train_runtime": 49935.3477, + "train_samples_per_second": 9.826, + "train_steps_per_second": 0.019 + } + ], + "logging_steps": 1.0, + "max_steps": 955, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.9663195796108476e+19, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}