{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 312, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.003205128205128205, "grad_norm": 13.91187858581543, "learning_rate": 2.2222222222222223e-05, "loss": 4.8108, "step": 1 }, { "epoch": 0.00641025641025641, "grad_norm": 14.33206558227539, "learning_rate": 4.4444444444444447e-05, "loss": 4.8801, "step": 2 }, { "epoch": 0.009615384615384616, "grad_norm": 14.014120101928711, "learning_rate": 6.666666666666667e-05, "loss": 4.8164, "step": 3 }, { "epoch": 0.01282051282051282, "grad_norm": 13.723206520080566, "learning_rate": 8.888888888888889e-05, "loss": 4.7503, "step": 4 }, { "epoch": 0.016025641025641024, "grad_norm": 14.325286865234375, "learning_rate": 0.00011111111111111112, "loss": 4.3225, "step": 5 }, { "epoch": 0.019230769230769232, "grad_norm": 13.791594505310059, "learning_rate": 0.00013333333333333334, "loss": 3.3933, "step": 6 }, { "epoch": 0.022435897435897436, "grad_norm": 10.824278831481934, "learning_rate": 0.00015555555555555556, "loss": 2.8485, "step": 7 }, { "epoch": 0.02564102564102564, "grad_norm": 9.639212608337402, "learning_rate": 0.00017777777777777779, "loss": 2.2245, "step": 8 }, { "epoch": 0.028846153846153848, "grad_norm": 7.029407501220703, "learning_rate": 0.0002, "loss": 1.6899, "step": 9 }, { "epoch": 0.03205128205128205, "grad_norm": 5.85064697265625, "learning_rate": 0.00019999462497359466, "loss": 1.4971, "step": 10 }, { "epoch": 0.035256410256410256, "grad_norm": 3.8360719680786133, "learning_rate": 0.0001999785004721968, "loss": 1.159, "step": 11 }, { "epoch": 0.038461538461538464, "grad_norm": 1.9359312057495117, "learning_rate": 0.00019995162822919883, "loss": 0.9306, "step": 12 }, { "epoch": 0.041666666666666664, "grad_norm": 2.1098594665527344, "learning_rate": 0.00019991401113338104, "loss": 0.9273, "step": 13 }, { "epoch": 0.04487179487179487, "grad_norm": 3.5646820068359375, "learning_rate": 0.00019986565322860115, "loss": 0.9156, "step": 14 }, { "epoch": 0.04807692307692308, "grad_norm": 2.836466073989868, "learning_rate": 0.00019980655971335945, "loss": 0.8981, "step": 15 }, { "epoch": 0.05128205128205128, "grad_norm": 2.151491641998291, "learning_rate": 0.00019973673694024, "loss": 0.8187, "step": 16 }, { "epoch": 0.05448717948717949, "grad_norm": 2.594736337661743, "learning_rate": 0.0001996561924152278, "loss": 0.8162, "step": 17 }, { "epoch": 0.057692307692307696, "grad_norm": 4.105694770812988, "learning_rate": 0.0001995649347969019, "loss": 0.8538, "step": 18 }, { "epoch": 0.060897435897435896, "grad_norm": 2.181260585784912, "learning_rate": 0.00019946297389550433, "loss": 0.8083, "step": 19 }, { "epoch": 0.0641025641025641, "grad_norm": 1.9404771327972412, "learning_rate": 0.0001993503206718859, "loss": 0.8117, "step": 20 }, { "epoch": 0.0673076923076923, "grad_norm": 1.6177146434783936, "learning_rate": 0.00019922698723632767, "loss": 0.668, "step": 21 }, { "epoch": 0.07051282051282051, "grad_norm": 2.5855164527893066, "learning_rate": 0.00019909298684723904, "loss": 0.9246, "step": 22 }, { "epoch": 0.07371794871794872, "grad_norm": 0.7814646363258362, "learning_rate": 0.00019894833390973266, "loss": 0.7877, "step": 23 }, { "epoch": 0.07692307692307693, "grad_norm": 1.592220664024353, "learning_rate": 0.0001987930439740757, "loss": 0.8776, "step": 24 }, { "epoch": 0.08012820512820513, "grad_norm": 3.1788816452026367, "learning_rate": 0.0001986271337340182, "loss": 0.943, "step": 25 }, { "epoch": 0.08333333333333333, "grad_norm": 0.8750837445259094, "learning_rate": 0.0001984506210249986, "loss": 0.9156, "step": 26 }, { "epoch": 0.08653846153846154, "grad_norm": 1.8515149354934692, "learning_rate": 0.00019826352482222638, "loss": 0.8454, "step": 27 }, { "epoch": 0.08974358974358974, "grad_norm": 1.592309594154358, "learning_rate": 0.0001980658652386421, "loss": 0.939, "step": 28 }, { "epoch": 0.09294871794871795, "grad_norm": 1.8106122016906738, "learning_rate": 0.00019785766352275542, "loss": 0.9646, "step": 29 }, { "epoch": 0.09615384615384616, "grad_norm": 1.220379114151001, "learning_rate": 0.00019763894205636072, "loss": 0.8782, "step": 30 }, { "epoch": 0.09935897435897435, "grad_norm": 1.5412297248840332, "learning_rate": 0.00019740972435213115, "loss": 0.7935, "step": 31 }, { "epoch": 0.10256410256410256, "grad_norm": 1.016067385673523, "learning_rate": 0.00019717003505109095, "loss": 0.905, "step": 32 }, { "epoch": 0.10576923076923077, "grad_norm": 0.7788788080215454, "learning_rate": 0.00019691989991996663, "loss": 0.7594, "step": 33 }, { "epoch": 0.10897435897435898, "grad_norm": 1.0361382961273193, "learning_rate": 0.00019665934584841682, "loss": 0.8002, "step": 34 }, { "epoch": 0.11217948717948718, "grad_norm": 0.8088793158531189, "learning_rate": 0.00019638840084614182, "loss": 0.8022, "step": 35 }, { "epoch": 0.11538461538461539, "grad_norm": 1.6871492862701416, "learning_rate": 0.00019610709403987246, "loss": 0.9974, "step": 36 }, { "epoch": 0.11858974358974358, "grad_norm": 0.8181275725364685, "learning_rate": 0.000195815455670239, "loss": 0.8345, "step": 37 }, { "epoch": 0.12179487179487179, "grad_norm": 1.2367571592330933, "learning_rate": 0.0001955135170885202, "loss": 0.9184, "step": 38 }, { "epoch": 0.125, "grad_norm": 0.8114968538284302, "learning_rate": 0.00019520131075327298, "loss": 0.7724, "step": 39 }, { "epoch": 0.1282051282051282, "grad_norm": 0.9863741993904114, "learning_rate": 0.00019487887022684336, "loss": 0.7895, "step": 40 }, { "epoch": 0.13141025641025642, "grad_norm": 0.8909745216369629, "learning_rate": 0.00019454623017175812, "loss": 0.8602, "step": 41 }, { "epoch": 0.1346153846153846, "grad_norm": 1.0723499059677124, "learning_rate": 0.0001942034263469989, "loss": 0.8249, "step": 42 }, { "epoch": 0.13782051282051283, "grad_norm": 2.3453705310821533, "learning_rate": 0.00019385049560415794, "loss": 0.8398, "step": 43 }, { "epoch": 0.14102564102564102, "grad_norm": 2.1112656593322754, "learning_rate": 0.00019348747588347637, "loss": 0.8648, "step": 44 }, { "epoch": 0.14423076923076922, "grad_norm": 0.7072998285293579, "learning_rate": 0.00019311440620976597, "loss": 0.7766, "step": 45 }, { "epoch": 0.14743589743589744, "grad_norm": 1.122727632522583, "learning_rate": 0.00019273132668821364, "loss": 0.8375, "step": 46 }, { "epoch": 0.15064102564102563, "grad_norm": 0.43819618225097656, "learning_rate": 0.00019233827850007027, "loss": 0.7336, "step": 47 }, { "epoch": 0.15384615384615385, "grad_norm": 0.6066083908081055, "learning_rate": 0.00019193530389822363, "loss": 0.7079, "step": 48 }, { "epoch": 0.15705128205128205, "grad_norm": 0.9152003526687622, "learning_rate": 0.0001915224462026563, "loss": 0.8533, "step": 49 }, { "epoch": 0.16025641025641027, "grad_norm": 1.2656763792037964, "learning_rate": 0.0001910997497957885, "loss": 0.8655, "step": 50 }, { "epoch": 0.16346153846153846, "grad_norm": 0.5455206036567688, "learning_rate": 0.00019066726011770726, "loss": 0.7714, "step": 51 }, { "epoch": 0.16666666666666666, "grad_norm": 0.77585369348526, "learning_rate": 0.00019022502366128135, "loss": 0.8717, "step": 52 }, { "epoch": 0.16987179487179488, "grad_norm": 0.5681566596031189, "learning_rate": 0.0001897730879671634, "loss": 0.8441, "step": 53 }, { "epoch": 0.17307692307692307, "grad_norm": 1.26576566696167, "learning_rate": 0.00018931150161867916, "loss": 0.9617, "step": 54 }, { "epoch": 0.1762820512820513, "grad_norm": 0.49213504791259766, "learning_rate": 0.0001888403142366049, "loss": 0.8874, "step": 55 }, { "epoch": 0.1794871794871795, "grad_norm": 0.5622470378875732, "learning_rate": 0.00018835957647383303, "loss": 0.7956, "step": 56 }, { "epoch": 0.18269230769230768, "grad_norm": 0.945149302482605, "learning_rate": 0.00018786934000992688, "loss": 0.8357, "step": 57 }, { "epoch": 0.1858974358974359, "grad_norm": 0.5550402998924255, "learning_rate": 0.00018736965754556528, "loss": 0.8787, "step": 58 }, { "epoch": 0.1891025641025641, "grad_norm": 0.841452419757843, "learning_rate": 0.00018686058279687698, "loss": 0.8429, "step": 59 }, { "epoch": 0.19230769230769232, "grad_norm": 0.45196735858917236, "learning_rate": 0.00018634217048966637, "loss": 0.8261, "step": 60 }, { "epoch": 0.1955128205128205, "grad_norm": 0.9931226372718811, "learning_rate": 0.0001858144763535302, "loss": 0.8594, "step": 61 }, { "epoch": 0.1987179487179487, "grad_norm": 0.7350529432296753, "learning_rate": 0.00018527755711586678, "loss": 0.8443, "step": 62 }, { "epoch": 0.20192307692307693, "grad_norm": 0.5463083386421204, "learning_rate": 0.00018473147049577774, "loss": 0.7869, "step": 63 }, { "epoch": 0.20512820512820512, "grad_norm": 1.1669566631317139, "learning_rate": 0.00018417627519786315, "loss": 0.9341, "step": 64 }, { "epoch": 0.20833333333333334, "grad_norm": 0.588135302066803, "learning_rate": 0.00018361203090591071, "loss": 0.7736, "step": 65 }, { "epoch": 0.21153846153846154, "grad_norm": 0.5148831605911255, "learning_rate": 0.00018303879827647975, "loss": 0.8559, "step": 66 }, { "epoch": 0.21474358974358973, "grad_norm": 0.6593618392944336, "learning_rate": 0.00018245663893238075, "loss": 0.7974, "step": 67 }, { "epoch": 0.21794871794871795, "grad_norm": 0.6366052627563477, "learning_rate": 0.00018186561545605054, "loss": 0.7543, "step": 68 }, { "epoch": 0.22115384615384615, "grad_norm": 0.5705471634864807, "learning_rate": 0.00018126579138282503, "loss": 0.8161, "step": 69 }, { "epoch": 0.22435897435897437, "grad_norm": 0.5039474368095398, "learning_rate": 0.00018065723119410884, "loss": 0.8071, "step": 70 }, { "epoch": 0.22756410256410256, "grad_norm": 0.7973775267601013, "learning_rate": 0.0001800400003104436, "loss": 0.7785, "step": 71 }, { "epoch": 0.23076923076923078, "grad_norm": 0.9587669372558594, "learning_rate": 0.00017941416508447536, "loss": 0.8423, "step": 72 }, { "epoch": 0.23397435897435898, "grad_norm": 0.4499201476573944, "learning_rate": 0.00017877979279382135, "loss": 0.8052, "step": 73 }, { "epoch": 0.23717948717948717, "grad_norm": 0.9492788910865784, "learning_rate": 0.0001781369516338378, "loss": 0.9093, "step": 74 }, { "epoch": 0.2403846153846154, "grad_norm": 0.6145200133323669, "learning_rate": 0.000177485710710289, "loss": 0.8359, "step": 75 }, { "epoch": 0.24358974358974358, "grad_norm": 0.5011272430419922, "learning_rate": 0.00017682614003191807, "loss": 0.8054, "step": 76 }, { "epoch": 0.2467948717948718, "grad_norm": 1.112430453300476, "learning_rate": 0.0001761583105029213, "loss": 0.8851, "step": 77 }, { "epoch": 0.25, "grad_norm": 0.8186496496200562, "learning_rate": 0.00017548229391532572, "loss": 0.757, "step": 78 }, { "epoch": 0.2532051282051282, "grad_norm": 1.00787353515625, "learning_rate": 0.00017479816294127152, "loss": 0.8089, "step": 79 }, { "epoch": 0.2564102564102564, "grad_norm": 0.9003360271453857, "learning_rate": 0.0001741059911251997, "loss": 0.9457, "step": 80 }, { "epoch": 0.25961538461538464, "grad_norm": 1.0145341157913208, "learning_rate": 0.00017340585287594604, "loss": 0.796, "step": 81 }, { "epoch": 0.26282051282051283, "grad_norm": 1.2034144401550293, "learning_rate": 0.00017269782345874203, "loss": 0.8504, "step": 82 }, { "epoch": 0.266025641025641, "grad_norm": 0.5833753347396851, "learning_rate": 0.00017198197898712404, "loss": 0.819, "step": 83 }, { "epoch": 0.2692307692307692, "grad_norm": 0.47029227018356323, "learning_rate": 0.00017125839641475072, "loss": 0.7965, "step": 84 }, { "epoch": 0.2724358974358974, "grad_norm": 0.6297673583030701, "learning_rate": 0.00017052715352713075, "loss": 0.7564, "step": 85 }, { "epoch": 0.27564102564102566, "grad_norm": 0.7019922137260437, "learning_rate": 0.00016978832893326074, "loss": 0.8576, "step": 86 }, { "epoch": 0.27884615384615385, "grad_norm": 0.7785760760307312, "learning_rate": 0.0001690420020571747, "loss": 0.8872, "step": 87 }, { "epoch": 0.28205128205128205, "grad_norm": 0.47651761770248413, "learning_rate": 0.00016828825312940592, "loss": 0.8333, "step": 88 }, { "epoch": 0.28525641025641024, "grad_norm": 0.5962091684341431, "learning_rate": 0.00016752716317836229, "loss": 0.9013, "step": 89 }, { "epoch": 0.28846153846153844, "grad_norm": 0.4600299596786499, "learning_rate": 0.00016675881402161536, "loss": 0.8245, "step": 90 }, { "epoch": 0.2916666666666667, "grad_norm": 0.5112613439559937, "learning_rate": 0.00016598328825710533, "loss": 0.8138, "step": 91 }, { "epoch": 0.2948717948717949, "grad_norm": 0.7249051332473755, "learning_rate": 0.00016520066925426144, "loss": 0.8401, "step": 92 }, { "epoch": 0.2980769230769231, "grad_norm": 0.6219087839126587, "learning_rate": 0.0001644110411450398, "loss": 0.7938, "step": 93 }, { "epoch": 0.30128205128205127, "grad_norm": 0.4909549951553345, "learning_rate": 0.00016361448881487914, "loss": 0.8425, "step": 94 }, { "epoch": 0.30448717948717946, "grad_norm": 0.45718565583229065, "learning_rate": 0.0001628110978935756, "loss": 0.8709, "step": 95 }, { "epoch": 0.3076923076923077, "grad_norm": 0.7159481048583984, "learning_rate": 0.00016200095474607753, "loss": 0.8469, "step": 96 }, { "epoch": 0.3108974358974359, "grad_norm": 0.6644757986068726, "learning_rate": 0.0001611841464632011, "loss": 0.8352, "step": 97 }, { "epoch": 0.3141025641025641, "grad_norm": 0.40486231446266174, "learning_rate": 0.00016036076085226814, "loss": 0.7915, "step": 98 }, { "epoch": 0.3173076923076923, "grad_norm": 0.588192880153656, "learning_rate": 0.0001595308864276666, "loss": 0.8802, "step": 99 }, { "epoch": 0.32051282051282054, "grad_norm": 0.2998511493206024, "learning_rate": 0.0001586946124013354, "loss": 0.7983, "step": 100 }, { "epoch": 0.32371794871794873, "grad_norm": 0.5495243072509766, "learning_rate": 0.00015785202867317407, "loss": 0.806, "step": 101 }, { "epoch": 0.3269230769230769, "grad_norm": 0.3516915738582611, "learning_rate": 0.00015700322582137827, "loss": 0.8455, "step": 102 }, { "epoch": 0.3301282051282051, "grad_norm": 0.6122769117355347, "learning_rate": 0.0001561482950927029, "loss": 0.8538, "step": 103 }, { "epoch": 0.3333333333333333, "grad_norm": 0.5758071541786194, "learning_rate": 0.00015528732839265272, "loss": 0.7786, "step": 104 }, { "epoch": 0.33653846153846156, "grad_norm": 0.606889009475708, "learning_rate": 0.00015442041827560274, "loss": 0.8535, "step": 105 }, { "epoch": 0.33974358974358976, "grad_norm": 0.5758721232414246, "learning_rate": 0.00015354765793484834, "loss": 0.8368, "step": 106 }, { "epoch": 0.34294871794871795, "grad_norm": 0.41782525181770325, "learning_rate": 0.000152669141192587, "loss": 0.8372, "step": 107 }, { "epoch": 0.34615384615384615, "grad_norm": 0.44819149374961853, "learning_rate": 0.00015178496248983254, "loss": 0.7807, "step": 108 }, { "epoch": 0.34935897435897434, "grad_norm": 0.42602694034576416, "learning_rate": 0.00015089521687626243, "loss": 0.7363, "step": 109 }, { "epoch": 0.3525641025641026, "grad_norm": 0.5617648363113403, "learning_rate": 0.00015000000000000001, "loss": 0.8311, "step": 110 }, { "epoch": 0.3557692307692308, "grad_norm": 0.38159874081611633, "learning_rate": 0.00014909940809733222, "loss": 0.7536, "step": 111 }, { "epoch": 0.358974358974359, "grad_norm": 0.3874197006225586, "learning_rate": 0.00014819353798236427, "loss": 0.8124, "step": 112 }, { "epoch": 0.36217948717948717, "grad_norm": 0.3920519948005676, "learning_rate": 0.00014728248703661182, "loss": 0.7821, "step": 113 }, { "epoch": 0.36538461538461536, "grad_norm": 0.31610092520713806, "learning_rate": 0.00014636635319853275, "loss": 0.8025, "step": 114 }, { "epoch": 0.3685897435897436, "grad_norm": 0.9610485434532166, "learning_rate": 0.00014544523495299842, "loss": 0.8597, "step": 115 }, { "epoch": 0.3717948717948718, "grad_norm": 0.6983340978622437, "learning_rate": 0.0001445192313207067, "loss": 0.8763, "step": 116 }, { "epoch": 0.375, "grad_norm": 0.37733662128448486, "learning_rate": 0.00014358844184753712, "loss": 0.787, "step": 117 }, { "epoch": 0.3782051282051282, "grad_norm": 0.6264870166778564, "learning_rate": 0.00014265296659384956, "loss": 0.6865, "step": 118 }, { "epoch": 0.3814102564102564, "grad_norm": 0.6618545651435852, "learning_rate": 0.0001417129061237278, "loss": 0.8416, "step": 119 }, { "epoch": 0.38461538461538464, "grad_norm": 0.5489368438720703, "learning_rate": 0.00014076836149416887, "loss": 0.8113, "step": 120 }, { "epoch": 0.38782051282051283, "grad_norm": 0.5143547654151917, "learning_rate": 0.00013981943424421932, "loss": 0.7212, "step": 121 }, { "epoch": 0.391025641025641, "grad_norm": 0.3641842007637024, "learning_rate": 0.00013886622638405952, "loss": 0.7081, "step": 122 }, { "epoch": 0.3942307692307692, "grad_norm": 0.5204703211784363, "learning_rate": 0.00013790884038403795, "loss": 0.8671, "step": 123 }, { "epoch": 0.3974358974358974, "grad_norm": 0.6415024995803833, "learning_rate": 0.00013694737916365517, "loss": 0.7905, "step": 124 }, { "epoch": 0.40064102564102566, "grad_norm": 0.5021610260009766, "learning_rate": 0.0001359819460805001, "loss": 0.8396, "step": 125 }, { "epoch": 0.40384615384615385, "grad_norm": 0.6839991211891174, "learning_rate": 0.00013501264491913906, "loss": 0.8807, "step": 126 }, { "epoch": 0.40705128205128205, "grad_norm": 0.3714783787727356, "learning_rate": 0.00013403957987995882, "loss": 0.7946, "step": 127 }, { "epoch": 0.41025641025641024, "grad_norm": 0.5053361654281616, "learning_rate": 0.00013306285556796495, "loss": 0.8481, "step": 128 }, { "epoch": 0.41346153846153844, "grad_norm": 0.5429531335830688, "learning_rate": 0.00013208257698153677, "loss": 0.7554, "step": 129 }, { "epoch": 0.4166666666666667, "grad_norm": 0.3953702449798584, "learning_rate": 0.00013109884950114007, "loss": 0.8, "step": 130 }, { "epoch": 0.4198717948717949, "grad_norm": 0.6481508016586304, "learning_rate": 0.00013011177887799845, "loss": 0.8627, "step": 131 }, { "epoch": 0.4230769230769231, "grad_norm": 0.7805072069168091, "learning_rate": 0.00012912147122272523, "loss": 0.8365, "step": 132 }, { "epoch": 0.42628205128205127, "grad_norm": 0.6625213027000427, "learning_rate": 0.00012812803299391628, "loss": 0.8348, "step": 133 }, { "epoch": 0.42948717948717946, "grad_norm": 0.5890776515007019, "learning_rate": 0.0001271315709867059, "loss": 0.8482, "step": 134 }, { "epoch": 0.4326923076923077, "grad_norm": 0.8653396964073181, "learning_rate": 0.00012613219232128608, "loss": 0.7627, "step": 135 }, { "epoch": 0.4358974358974359, "grad_norm": 0.4106805920600891, "learning_rate": 0.00012513000443139112, "loss": 0.8412, "step": 136 }, { "epoch": 0.4391025641025641, "grad_norm": 0.9956967234611511, "learning_rate": 0.00012412511505274844, "loss": 0.9127, "step": 137 }, { "epoch": 0.4423076923076923, "grad_norm": 0.46844878792762756, "learning_rate": 0.000123117632211497, "loss": 0.774, "step": 138 }, { "epoch": 0.44551282051282054, "grad_norm": 0.966278612613678, "learning_rate": 0.0001221076642125742, "loss": 0.8659, "step": 139 }, { "epoch": 0.44871794871794873, "grad_norm": 0.5847793221473694, "learning_rate": 0.00012109531962807332, "loss": 0.7507, "step": 140 }, { "epoch": 0.4519230769230769, "grad_norm": 0.5927683115005493, "learning_rate": 0.00012008070728557186, "loss": 0.8908, "step": 141 }, { "epoch": 0.4551282051282051, "grad_norm": 0.42794185876846313, "learning_rate": 0.00011906393625643244, "loss": 0.7478, "step": 142 }, { "epoch": 0.4583333333333333, "grad_norm": 0.6426613926887512, "learning_rate": 0.00011804511584407763, "loss": 0.7498, "step": 143 }, { "epoch": 0.46153846153846156, "grad_norm": 0.40398654341697693, "learning_rate": 0.00011702435557223987, "loss": 0.7539, "step": 144 }, { "epoch": 0.46474358974358976, "grad_norm": 0.4018569886684418, "learning_rate": 0.00011600176517318741, "loss": 0.8294, "step": 145 }, { "epoch": 0.46794871794871795, "grad_norm": 0.33486562967300415, "learning_rate": 0.00011497745457592816, "loss": 0.8322, "step": 146 }, { "epoch": 0.47115384615384615, "grad_norm": 0.539478063583374, "learning_rate": 0.00011395153389439233, "loss": 0.8697, "step": 147 }, { "epoch": 0.47435897435897434, "grad_norm": 0.4381621479988098, "learning_rate": 0.0001129241134155949, "loss": 0.8216, "step": 148 }, { "epoch": 0.4775641025641026, "grad_norm": 1.0025720596313477, "learning_rate": 0.00011189530358778005, "loss": 0.9283, "step": 149 }, { "epoch": 0.4807692307692308, "grad_norm": 0.6928549408912659, "learning_rate": 0.00011086521500854745, "loss": 0.8163, "step": 150 }, { "epoch": 0.483974358974359, "grad_norm": 0.3628673851490021, "learning_rate": 0.00010983395841296348, "loss": 0.7866, "step": 151 }, { "epoch": 0.48717948717948717, "grad_norm": 0.6659094095230103, "learning_rate": 0.00010880164466165674, "loss": 0.767, "step": 152 }, { "epoch": 0.49038461538461536, "grad_norm": 0.4747048318386078, "learning_rate": 0.00010776838472890065, "loss": 0.7455, "step": 153 }, { "epoch": 0.4935897435897436, "grad_norm": 0.5610392093658447, "learning_rate": 0.00010673428969068364, "loss": 0.7943, "step": 154 }, { "epoch": 0.4967948717948718, "grad_norm": 0.8494670391082764, "learning_rate": 0.00010569947071276847, "loss": 0.8778, "step": 155 }, { "epoch": 0.5, "grad_norm": 0.4916403293609619, "learning_rate": 0.00010466403903874176, "loss": 0.7698, "step": 156 }, { "epoch": 0.5032051282051282, "grad_norm": 1.1345371007919312, "learning_rate": 0.00010362810597805526, "loss": 0.9339, "step": 157 }, { "epoch": 0.5064102564102564, "grad_norm": 0.8065443634986877, "learning_rate": 0.00010259178289406011, "loss": 0.7978, "step": 158 }, { "epoch": 0.5096153846153846, "grad_norm": 1.3246110677719116, "learning_rate": 0.0001015551811920351, "loss": 0.9917, "step": 159 }, { "epoch": 0.5128205128205128, "grad_norm": 0.5924827456474304, "learning_rate": 0.00010051841230721065, "loss": 0.8334, "step": 160 }, { "epoch": 0.5160256410256411, "grad_norm": 0.5810033679008484, "learning_rate": 9.948158769278939e-05, "loss": 0.8431, "step": 161 }, { "epoch": 0.5192307692307693, "grad_norm": 0.517737627029419, "learning_rate": 9.844481880796491e-05, "loss": 0.8149, "step": 162 }, { "epoch": 0.5224358974358975, "grad_norm": 0.6809769868850708, "learning_rate": 9.740821710593989e-05, "loss": 0.9241, "step": 163 }, { "epoch": 0.5256410256410257, "grad_norm": 0.5955147743225098, "learning_rate": 9.637189402194476e-05, "loss": 0.761, "step": 164 }, { "epoch": 0.5288461538461539, "grad_norm": 0.6794393062591553, "learning_rate": 9.533596096125825e-05, "loss": 0.8291, "step": 165 }, { "epoch": 0.532051282051282, "grad_norm": 0.6929296255111694, "learning_rate": 9.430052928723153e-05, "loss": 0.7753, "step": 166 }, { "epoch": 0.5352564102564102, "grad_norm": 0.6990981698036194, "learning_rate": 9.326571030931637e-05, "loss": 0.8932, "step": 167 }, { "epoch": 0.5384615384615384, "grad_norm": 0.591814398765564, "learning_rate": 9.223161527109937e-05, "loss": 0.8114, "step": 168 }, { "epoch": 0.5416666666666666, "grad_norm": 0.6408451199531555, "learning_rate": 9.119835533834331e-05, "loss": 0.7942, "step": 169 }, { "epoch": 0.5448717948717948, "grad_norm": 0.6977259516716003, "learning_rate": 9.016604158703654e-05, "loss": 0.912, "step": 170 }, { "epoch": 0.5480769230769231, "grad_norm": 0.604136049747467, "learning_rate": 8.913478499145254e-05, "loss": 0.789, "step": 171 }, { "epoch": 0.5512820512820513, "grad_norm": 0.5467025637626648, "learning_rate": 8.810469641222001e-05, "loss": 0.7825, "step": 172 }, { "epoch": 0.5544871794871795, "grad_norm": 0.7560824155807495, "learning_rate": 8.707588658440511e-05, "loss": 0.7609, "step": 173 }, { "epoch": 0.5576923076923077, "grad_norm": 0.9269303679466248, "learning_rate": 8.604846610560771e-05, "loss": 0.8706, "step": 174 }, { "epoch": 0.5608974358974359, "grad_norm": 0.7465280294418335, "learning_rate": 8.502254542407186e-05, "loss": 0.7986, "step": 175 }, { "epoch": 0.5641025641025641, "grad_norm": 0.8691240549087524, "learning_rate": 8.399823482681262e-05, "loss": 0.7843, "step": 176 }, { "epoch": 0.5673076923076923, "grad_norm": 0.59829181432724, "learning_rate": 8.297564442776014e-05, "loss": 0.7441, "step": 177 }, { "epoch": 0.5705128205128205, "grad_norm": 0.9093548655509949, "learning_rate": 8.195488415592238e-05, "loss": 0.8501, "step": 178 }, { "epoch": 0.5737179487179487, "grad_norm": 0.8965813517570496, "learning_rate": 8.093606374356759e-05, "loss": 0.8309, "step": 179 }, { "epoch": 0.5769230769230769, "grad_norm": 0.7008676528930664, "learning_rate": 7.991929271442817e-05, "loss": 0.6563, "step": 180 }, { "epoch": 0.5801282051282052, "grad_norm": 0.8001984357833862, "learning_rate": 7.89046803719267e-05, "loss": 0.8305, "step": 181 }, { "epoch": 0.5833333333333334, "grad_norm": 0.8418125510215759, "learning_rate": 7.789233578742582e-05, "loss": 0.8296, "step": 182 }, { "epoch": 0.5865384615384616, "grad_norm": 0.8208316564559937, "learning_rate": 7.688236778850306e-05, "loss": 0.7777, "step": 183 }, { "epoch": 0.5897435897435898, "grad_norm": 0.8942621946334839, "learning_rate": 7.587488494725157e-05, "loss": 0.9101, "step": 184 }, { "epoch": 0.592948717948718, "grad_norm": 0.6328911781311035, "learning_rate": 7.48699955686089e-05, "loss": 0.7004, "step": 185 }, { "epoch": 0.5961538461538461, "grad_norm": 0.5853061079978943, "learning_rate": 7.386780767871397e-05, "loss": 0.8373, "step": 186 }, { "epoch": 0.5993589743589743, "grad_norm": 0.9095894694328308, "learning_rate": 7.286842901329412e-05, "loss": 0.833, "step": 187 }, { "epoch": 0.6025641025641025, "grad_norm": 0.562564492225647, "learning_rate": 7.187196700608373e-05, "loss": 0.7822, "step": 188 }, { "epoch": 0.6057692307692307, "grad_norm": 0.812235951423645, "learning_rate": 7.087852877727481e-05, "loss": 0.823, "step": 189 }, { "epoch": 0.6089743589743589, "grad_norm": 0.9851651191711426, "learning_rate": 6.988822112200156e-05, "loss": 0.8251, "step": 190 }, { "epoch": 0.6121794871794872, "grad_norm": 0.9343296885490417, "learning_rate": 6.890115049885994e-05, "loss": 0.7613, "step": 191 }, { "epoch": 0.6153846153846154, "grad_norm": 0.6352812647819519, "learning_rate": 6.791742301846326e-05, "loss": 0.7761, "step": 192 }, { "epoch": 0.6185897435897436, "grad_norm": 1.490954041481018, "learning_rate": 6.693714443203507e-05, "loss": 1.0911, "step": 193 }, { "epoch": 0.6217948717948718, "grad_norm": 0.7327432036399841, "learning_rate": 6.59604201200412e-05, "loss": 0.7859, "step": 194 }, { "epoch": 0.625, "grad_norm": 0.7584162950515747, "learning_rate": 6.498735508086093e-05, "loss": 0.6795, "step": 195 }, { "epoch": 0.6282051282051282, "grad_norm": 0.9393740892410278, "learning_rate": 6.40180539194999e-05, "loss": 0.7382, "step": 196 }, { "epoch": 0.6314102564102564, "grad_norm": 1.4272030591964722, "learning_rate": 6.305262083634488e-05, "loss": 0.8325, "step": 197 }, { "epoch": 0.6346153846153846, "grad_norm": 1.7151983976364136, "learning_rate": 6.209115961596208e-05, "loss": 0.9443, "step": 198 }, { "epoch": 0.6378205128205128, "grad_norm": 0.856202244758606, "learning_rate": 6.113377361594049e-05, "loss": 0.7448, "step": 199 }, { "epoch": 0.6410256410256411, "grad_norm": 1.3806068897247314, "learning_rate": 6.018056575578075e-05, "loss": 0.9231, "step": 200 }, { "epoch": 0.6442307692307693, "grad_norm": 0.8063072562217712, "learning_rate": 5.923163850583113e-05, "loss": 0.7305, "step": 201 }, { "epoch": 0.6474358974358975, "grad_norm": 1.1578689813613892, "learning_rate": 5.828709387627218e-05, "loss": 0.7892, "step": 202 }, { "epoch": 0.6506410256410257, "grad_norm": 0.881730854511261, "learning_rate": 5.73470334061505e-05, "loss": 0.7162, "step": 203 }, { "epoch": 0.6538461538461539, "grad_norm": 0.9713548421859741, "learning_rate": 5.6411558152462894e-05, "loss": 0.7252, "step": 204 }, { "epoch": 0.657051282051282, "grad_norm": 1.12835693359375, "learning_rate": 5.54807686792933e-05, "loss": 0.7446, "step": 205 }, { "epoch": 0.6602564102564102, "grad_norm": 1.3624720573425293, "learning_rate": 5.4554765047001613e-05, "loss": 0.7664, "step": 206 }, { "epoch": 0.6634615384615384, "grad_norm": 1.169715166091919, "learning_rate": 5.363364680146725e-05, "loss": 0.7936, "step": 207 }, { "epoch": 0.6666666666666666, "grad_norm": 1.94355309009552, "learning_rate": 5.271751296338823e-05, "loss": 0.8667, "step": 208 }, { "epoch": 0.6698717948717948, "grad_norm": 1.5784542560577393, "learning_rate": 5.180646201763577e-05, "loss": 0.8658, "step": 209 }, { "epoch": 0.6730769230769231, "grad_norm": 0.9312598705291748, "learning_rate": 5.090059190266779e-05, "loss": 0.7996, "step": 210 }, { "epoch": 0.6762820512820513, "grad_norm": 0.9065409302711487, "learning_rate": 5.000000000000002e-05, "loss": 0.7893, "step": 211 }, { "epoch": 0.6794871794871795, "grad_norm": 1.3561393022537231, "learning_rate": 4.9104783123737566e-05, "loss": 0.8413, "step": 212 }, { "epoch": 0.6826923076923077, "grad_norm": 0.8747413158416748, "learning_rate": 4.821503751016746e-05, "loss": 0.799, "step": 213 }, { "epoch": 0.6858974358974359, "grad_norm": 0.8541598916053772, "learning_rate": 4.733085880741301e-05, "loss": 0.7965, "step": 214 }, { "epoch": 0.6891025641025641, "grad_norm": 1.1472944021224976, "learning_rate": 4.645234206515171e-05, "loss": 0.7915, "step": 215 }, { "epoch": 0.6923076923076923, "grad_norm": 1.0317872762680054, "learning_rate": 4.5579581724397255e-05, "loss": 0.8, "step": 216 }, { "epoch": 0.6955128205128205, "grad_norm": 1.4401482343673706, "learning_rate": 4.471267160734731e-05, "loss": 0.7188, "step": 217 }, { "epoch": 0.6987179487179487, "grad_norm": 0.7811070680618286, "learning_rate": 4.385170490729712e-05, "loss": 0.8123, "step": 218 }, { "epoch": 0.7019230769230769, "grad_norm": 1.1124131679534912, "learning_rate": 4.2996774178621736e-05, "loss": 0.9242, "step": 219 }, { "epoch": 0.7051282051282052, "grad_norm": 0.9388600587844849, "learning_rate": 4.2147971326825966e-05, "loss": 0.8065, "step": 220 }, { "epoch": 0.7083333333333334, "grad_norm": 0.8694519400596619, "learning_rate": 4.130538759866457e-05, "loss": 0.8012, "step": 221 }, { "epoch": 0.7115384615384616, "grad_norm": 0.756574809551239, "learning_rate": 4.046911357233343e-05, "loss": 0.6749, "step": 222 }, { "epoch": 0.7147435897435898, "grad_norm": 0.8523420095443726, "learning_rate": 3.963923914773187e-05, "loss": 0.7184, "step": 223 }, { "epoch": 0.717948717948718, "grad_norm": 0.8101953864097595, "learning_rate": 3.8815853536798904e-05, "loss": 0.8064, "step": 224 }, { "epoch": 0.7211538461538461, "grad_norm": 0.9404911994934082, "learning_rate": 3.79990452539225e-05, "loss": 0.7452, "step": 225 }, { "epoch": 0.7243589743589743, "grad_norm": 0.8205745220184326, "learning_rate": 3.7188902106424416e-05, "loss": 0.7859, "step": 226 }, { "epoch": 0.7275641025641025, "grad_norm": 0.9722088575363159, "learning_rate": 3.638551118512089e-05, "loss": 0.7248, "step": 227 }, { "epoch": 0.7307692307692307, "grad_norm": 0.7296696901321411, "learning_rate": 3.558895885496023e-05, "loss": 0.7329, "step": 228 }, { "epoch": 0.7339743589743589, "grad_norm": 1.253202199935913, "learning_rate": 3.479933074573858e-05, "loss": 0.8777, "step": 229 }, { "epoch": 0.7371794871794872, "grad_norm": 1.1534240245819092, "learning_rate": 3.401671174289469e-05, "loss": 0.7485, "step": 230 }, { "epoch": 0.7403846153846154, "grad_norm": 1.1224077939987183, "learning_rate": 3.324118597838464e-05, "loss": 0.7728, "step": 231 }, { "epoch": 0.7435897435897436, "grad_norm": 0.8976945281028748, "learning_rate": 3.2472836821637744e-05, "loss": 0.7541, "step": 232 }, { "epoch": 0.7467948717948718, "grad_norm": 0.840302050113678, "learning_rate": 3.1711746870594086e-05, "loss": 0.6973, "step": 233 }, { "epoch": 0.75, "grad_norm": 0.988998293876648, "learning_rate": 3.0957997942825336e-05, "loss": 0.6981, "step": 234 }, { "epoch": 0.7532051282051282, "grad_norm": 1.0511542558670044, "learning_rate": 3.021167106673928e-05, "loss": 0.6618, "step": 235 }, { "epoch": 0.7564102564102564, "grad_norm": 1.1155120134353638, "learning_rate": 2.9472846472869298e-05, "loss": 0.7775, "step": 236 }, { "epoch": 0.7596153846153846, "grad_norm": 1.064674735069275, "learning_rate": 2.874160358524931e-05, "loss": 0.747, "step": 237 }, { "epoch": 0.7628205128205128, "grad_norm": 1.1070621013641357, "learning_rate": 2.8018021012875994e-05, "loss": 0.8083, "step": 238 }, { "epoch": 0.7660256410256411, "grad_norm": 0.8089228868484497, "learning_rate": 2.7302176541257986e-05, "loss": 0.7037, "step": 239 }, { "epoch": 0.7692307692307693, "grad_norm": 0.7307572960853577, "learning_rate": 2.659414712405398e-05, "loss": 0.7206, "step": 240 }, { "epoch": 0.7724358974358975, "grad_norm": 1.0370497703552246, "learning_rate": 2.5894008874800325e-05, "loss": 0.7379, "step": 241 }, { "epoch": 0.7756410256410257, "grad_norm": 1.3511935472488403, "learning_rate": 2.5201837058728505e-05, "loss": 0.7198, "step": 242 }, { "epoch": 0.7788461538461539, "grad_norm": 0.8478284478187561, "learning_rate": 2.451770608467432e-05, "loss": 0.8123, "step": 243 }, { "epoch": 0.782051282051282, "grad_norm": 1.0344876050949097, "learning_rate": 2.3841689497078746e-05, "loss": 0.7769, "step": 244 }, { "epoch": 0.7852564102564102, "grad_norm": 1.0734611749649048, "learning_rate": 2.3173859968081944e-05, "loss": 0.7204, "step": 245 }, { "epoch": 0.7884615384615384, "grad_norm": 1.7516003847122192, "learning_rate": 2.251428928971102e-05, "loss": 0.7088, "step": 246 }, { "epoch": 0.7916666666666666, "grad_norm": 1.0216751098632812, "learning_rate": 2.1863048366162208e-05, "loss": 0.7828, "step": 247 }, { "epoch": 0.7948717948717948, "grad_norm": 1.4974614381790161, "learning_rate": 2.1220207206178688e-05, "loss": 0.8549, "step": 248 }, { "epoch": 0.7980769230769231, "grad_norm": 1.2746527194976807, "learning_rate": 2.058583491552465e-05, "loss": 0.7863, "step": 249 }, { "epoch": 0.8012820512820513, "grad_norm": 1.0694308280944824, "learning_rate": 1.995999968955641e-05, "loss": 0.7612, "step": 250 }, { "epoch": 0.8044871794871795, "grad_norm": 1.4098316431045532, "learning_rate": 1.9342768805891178e-05, "loss": 0.7792, "step": 251 }, { "epoch": 0.8076923076923077, "grad_norm": 0.9499948024749756, "learning_rate": 1.8734208617174988e-05, "loss": 0.8431, "step": 252 }, { "epoch": 0.8108974358974359, "grad_norm": 1.0735160112380981, "learning_rate": 1.8134384543949478e-05, "loss": 0.6933, "step": 253 }, { "epoch": 0.8141025641025641, "grad_norm": 1.3675204515457153, "learning_rate": 1.754336106761927e-05, "loss": 0.7038, "step": 254 }, { "epoch": 0.8173076923076923, "grad_norm": 1.0891550779342651, "learning_rate": 1.696120172352025e-05, "loss": 0.6681, "step": 255 }, { "epoch": 0.8205128205128205, "grad_norm": 1.0907151699066162, "learning_rate": 1.6387969094089316e-05, "loss": 0.8043, "step": 256 }, { "epoch": 0.8237179487179487, "grad_norm": 0.935387134552002, "learning_rate": 1.5823724802136865e-05, "loss": 0.8365, "step": 257 }, { "epoch": 0.8269230769230769, "grad_norm": 0.8595330715179443, "learning_rate": 1.526852950422226e-05, "loss": 0.7014, "step": 258 }, { "epoch": 0.8301282051282052, "grad_norm": 1.1038745641708374, "learning_rate": 1.4722442884133214e-05, "loss": 0.7596, "step": 259 }, { "epoch": 0.8333333333333334, "grad_norm": 1.1582711935043335, "learning_rate": 1.4185523646469822e-05, "loss": 0.6689, "step": 260 }, { "epoch": 0.8365384615384616, "grad_norm": 1.1454054117202759, "learning_rate": 1.3657829510333654e-05, "loss": 0.8129, "step": 261 }, { "epoch": 0.8397435897435898, "grad_norm": 1.1534847021102905, "learning_rate": 1.3139417203123027e-05, "loss": 0.6611, "step": 262 }, { "epoch": 0.842948717948718, "grad_norm": 1.1782026290893555, "learning_rate": 1.263034245443473e-05, "loss": 0.7166, "step": 263 }, { "epoch": 0.8461538461538461, "grad_norm": 1.3515312671661377, "learning_rate": 1.2130659990073146e-05, "loss": 0.789, "step": 264 }, { "epoch": 0.8493589743589743, "grad_norm": 1.0562694072723389, "learning_rate": 1.1640423526166988e-05, "loss": 0.7457, "step": 265 }, { "epoch": 0.8525641025641025, "grad_norm": 1.0518815517425537, "learning_rate": 1.1159685763395111e-05, "loss": 0.7977, "step": 266 }, { "epoch": 0.8557692307692307, "grad_norm": 1.1951806545257568, "learning_rate": 1.0688498381320855e-05, "loss": 0.7813, "step": 267 }, { "epoch": 0.8589743589743589, "grad_norm": 1.2347551584243774, "learning_rate": 1.0226912032836611e-05, "loss": 0.6702, "step": 268 }, { "epoch": 0.8621794871794872, "grad_norm": 2.0265257358551025, "learning_rate": 9.774976338718677e-06, "loss": 0.9113, "step": 269 }, { "epoch": 0.8653846153846154, "grad_norm": 1.1159110069274902, "learning_rate": 9.332739882292752e-06, "loss": 0.7241, "step": 270 }, { "epoch": 0.8685897435897436, "grad_norm": 1.0101889371871948, "learning_rate": 8.900250204211514e-06, "loss": 0.7407, "step": 271 }, { "epoch": 0.8717948717948718, "grad_norm": 1.1704083681106567, "learning_rate": 8.47755379734373e-06, "loss": 0.6598, "step": 272 }, { "epoch": 0.875, "grad_norm": 0.9426171779632568, "learning_rate": 8.064696101776358e-06, "loss": 0.7204, "step": 273 }, { "epoch": 0.8782051282051282, "grad_norm": 1.2830696105957031, "learning_rate": 7.661721499929753e-06, "loss": 0.7574, "step": 274 }, { "epoch": 0.8814102564102564, "grad_norm": 1.5801949501037598, "learning_rate": 7.2686733117863784e-06, "loss": 0.7968, "step": 275 }, { "epoch": 0.8846153846153846, "grad_norm": 1.4336384534835815, "learning_rate": 6.8855937902340576e-06, "loss": 0.6974, "step": 276 }, { "epoch": 0.8878205128205128, "grad_norm": 0.9879603385925293, "learning_rate": 6.512524116523633e-06, "loss": 0.634, "step": 277 }, { "epoch": 0.8910256410256411, "grad_norm": 1.2858949899673462, "learning_rate": 6.149504395842087e-06, "loss": 0.7762, "step": 278 }, { "epoch": 0.8942307692307693, "grad_norm": 1.6432108879089355, "learning_rate": 5.7965736530010916e-06, "loss": 0.821, "step": 279 }, { "epoch": 0.8974358974358975, "grad_norm": 1.3427728414535522, "learning_rate": 5.453769828241872e-06, "loss": 0.7391, "step": 280 }, { "epoch": 0.9006410256410257, "grad_norm": 1.405120849609375, "learning_rate": 5.121129773156663e-06, "loss": 0.7824, "step": 281 }, { "epoch": 0.9038461538461539, "grad_norm": 1.232499122619629, "learning_rate": 4.798689246727006e-06, "loss": 0.7732, "step": 282 }, { "epoch": 0.907051282051282, "grad_norm": 1.4786527156829834, "learning_rate": 4.486482911479839e-06, "loss": 0.7186, "step": 283 }, { "epoch": 0.9102564102564102, "grad_norm": 1.3841599225997925, "learning_rate": 4.184544329761009e-06, "loss": 0.7316, "step": 284 }, { "epoch": 0.9134615384615384, "grad_norm": 1.110377550125122, "learning_rate": 3.892905960127546e-06, "loss": 0.7166, "step": 285 }, { "epoch": 0.9166666666666666, "grad_norm": 1.2394565343856812, "learning_rate": 3.611599153858214e-06, "loss": 0.6894, "step": 286 }, { "epoch": 0.9198717948717948, "grad_norm": 2.2001726627349854, "learning_rate": 3.3406541515832003e-06, "loss": 0.8162, "step": 287 }, { "epoch": 0.9230769230769231, "grad_norm": 1.696381688117981, "learning_rate": 3.0801000800333877e-06, "loss": 0.6764, "step": 288 }, { "epoch": 0.9262820512820513, "grad_norm": 1.1135708093643188, "learning_rate": 2.8299649489090475e-06, "loss": 0.7059, "step": 289 }, { "epoch": 0.9294871794871795, "grad_norm": 1.3587653636932373, "learning_rate": 2.590275647868867e-06, "loss": 0.7251, "step": 290 }, { "epoch": 0.9326923076923077, "grad_norm": 1.686203956604004, "learning_rate": 2.3610579436393e-06, "loss": 0.6625, "step": 291 }, { "epoch": 0.9358974358974359, "grad_norm": 1.2855104207992554, "learning_rate": 2.1423364772445887e-06, "loss": 0.7562, "step": 292 }, { "epoch": 0.9391025641025641, "grad_norm": 1.6745336055755615, "learning_rate": 1.9341347613579087e-06, "loss": 0.7244, "step": 293 }, { "epoch": 0.9423076923076923, "grad_norm": 1.2314362525939941, "learning_rate": 1.7364751777736332e-06, "loss": 0.6981, "step": 294 }, { "epoch": 0.9455128205128205, "grad_norm": 0.9058342576026917, "learning_rate": 1.5493789750014031e-06, "loss": 0.7398, "step": 295 }, { "epoch": 0.9487179487179487, "grad_norm": 1.4387075901031494, "learning_rate": 1.3728662659818204e-06, "loss": 0.6769, "step": 296 }, { "epoch": 0.9519230769230769, "grad_norm": 1.3964595794677734, "learning_rate": 1.2069560259243328e-06, "loss": 0.685, "step": 297 }, { "epoch": 0.9551282051282052, "grad_norm": 1.1776217222213745, "learning_rate": 1.0516660902673448e-06, "loss": 0.7934, "step": 298 }, { "epoch": 0.9583333333333334, "grad_norm": 1.2884339094161987, "learning_rate": 9.070131527609604e-07, "loss": 0.6712, "step": 299 }, { "epoch": 0.9615384615384616, "grad_norm": 1.7066468000411987, "learning_rate": 7.730127636723539e-07, "loss": 0.6243, "step": 300 }, { "epoch": 0.9647435897435898, "grad_norm": 0.9889491200447083, "learning_rate": 6.496793281141056e-07, "loss": 0.6915, "step": 301 }, { "epoch": 0.967948717948718, "grad_norm": 0.9342219233512878, "learning_rate": 5.370261044956971e-07, "loss": 0.7163, "step": 302 }, { "epoch": 0.9711538461538461, "grad_norm": 1.2161084413528442, "learning_rate": 4.3506520309813947e-07, "loss": 0.6957, "step": 303 }, { "epoch": 0.9743589743589743, "grad_norm": 1.4092503786087036, "learning_rate": 3.4380758477219333e-07, "loss": 0.7295, "step": 304 }, { "epoch": 0.9775641025641025, "grad_norm": 1.5113911628723145, "learning_rate": 2.6326305976001055e-07, "loss": 0.6846, "step": 305 }, { "epoch": 0.9807692307692307, "grad_norm": 1.2269924879074097, "learning_rate": 1.9344028664056713e-07, "loss": 0.7242, "step": 306 }, { "epoch": 0.9839743589743589, "grad_norm": 1.3103538751602173, "learning_rate": 1.3434677139885222e-07, "loss": 0.6923, "step": 307 }, { "epoch": 0.9871794871794872, "grad_norm": 1.182924747467041, "learning_rate": 8.598886661895788e-08, "loss": 0.693, "step": 308 }, { "epoch": 0.9903846153846154, "grad_norm": 1.258489966392517, "learning_rate": 4.837177080119215e-08, "loss": 0.7017, "step": 309 }, { "epoch": 0.9935897435897436, "grad_norm": 1.2968997955322266, "learning_rate": 2.1499527803214846e-08, "loss": 0.7148, "step": 310 }, { "epoch": 0.9967948717948718, "grad_norm": 1.3392126560211182, "learning_rate": 5.375026405352035e-09, "loss": 0.7221, "step": 311 }, { "epoch": 1.0, "grad_norm": 1.7223619222640991, "learning_rate": 0.0, "loss": 0.5389, "step": 312 } ], "logging_steps": 1, "max_steps": 312, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3978364611723264.0, "train_batch_size": 16, "trial_name": null, "trial_params": null }