{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.3910355486862442, "eval_steps": 500, "global_step": 18000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0077279752704791345, "grad_norm": 2.382732391357422, "learning_rate": 4.987764039155075e-05, "loss": 2.2319, "step": 100 }, { "epoch": 0.015455950540958269, "grad_norm": 1.1049453020095825, "learning_rate": 4.974884080370943e-05, "loss": 1.4633, "step": 200 }, { "epoch": 0.023183925811437404, "grad_norm": 2.012437582015991, "learning_rate": 4.962004121586811e-05, "loss": 1.4898, "step": 300 }, { "epoch": 0.030911901081916538, "grad_norm": 2.1583738327026367, "learning_rate": 4.949124162802679e-05, "loss": 1.2682, "step": 400 }, { "epoch": 0.03863987635239567, "grad_norm": 1.3249982595443726, "learning_rate": 4.9362442040185474e-05, "loss": 1.3189, "step": 500 }, { "epoch": 0.04636785162287481, "grad_norm": 1.0867700576782227, "learning_rate": 4.923364245234415e-05, "loss": 1.2482, "step": 600 }, { "epoch": 0.05409582689335394, "grad_norm": 0.9986234307289124, "learning_rate": 4.910484286450284e-05, "loss": 1.1299, "step": 700 }, { "epoch": 0.061823802163833076, "grad_norm": 0.8884647488594055, "learning_rate": 4.897604327666152e-05, "loss": 1.2575, "step": 800 }, { "epoch": 0.0695517774343122, "grad_norm": 1.6838749647140503, "learning_rate": 4.8847243688820196e-05, "loss": 1.1865, "step": 900 }, { "epoch": 0.07727975270479134, "grad_norm": 2.3805487155914307, "learning_rate": 4.8718444100978875e-05, "loss": 1.1084, "step": 1000 }, { "epoch": 0.08500772797527048, "grad_norm": 1.653010606765747, "learning_rate": 4.858964451313756e-05, "loss": 1.1401, "step": 1100 }, { "epoch": 0.09273570324574962, "grad_norm": 1.6932406425476074, "learning_rate": 4.846084492529624e-05, "loss": 1.2395, "step": 1200 }, { "epoch": 0.10046367851622875, "grad_norm": 1.2242937088012695, "learning_rate": 4.8332045337454925e-05, "loss": 1.1328, "step": 1300 }, { "epoch": 0.10819165378670788, "grad_norm": 2.8659744262695312, "learning_rate": 4.8203245749613604e-05, "loss": 1.1238, "step": 1400 }, { "epoch": 0.11591962905718702, "grad_norm": 2.079432249069214, "learning_rate": 4.807444616177228e-05, "loss": 1.0579, "step": 1500 }, { "epoch": 0.12364760432766615, "grad_norm": 0.5808582901954651, "learning_rate": 4.794564657393096e-05, "loss": 0.9646, "step": 1600 }, { "epoch": 0.13137557959814528, "grad_norm": 1.380146861076355, "learning_rate": 4.781684698608965e-05, "loss": 1.1086, "step": 1700 }, { "epoch": 0.1391035548686244, "grad_norm": 1.4197566509246826, "learning_rate": 4.7688047398248325e-05, "loss": 1.1619, "step": 1800 }, { "epoch": 0.14683153013910355, "grad_norm": 1.1287027597427368, "learning_rate": 4.755924781040701e-05, "loss": 1.0074, "step": 1900 }, { "epoch": 0.1545595054095827, "grad_norm": 1.1685930490493774, "learning_rate": 4.743044822256569e-05, "loss": 1.1573, "step": 2000 }, { "epoch": 0.16228748068006182, "grad_norm": 1.008876919746399, "learning_rate": 4.730164863472437e-05, "loss": 1.0858, "step": 2100 }, { "epoch": 0.17001545595054096, "grad_norm": 0.9628033638000488, "learning_rate": 4.717284904688305e-05, "loss": 1.1267, "step": 2200 }, { "epoch": 0.1777434312210201, "grad_norm": 1.298677682876587, "learning_rate": 4.704404945904173e-05, "loss": 1.0804, "step": 2300 }, { "epoch": 0.18547140649149924, "grad_norm": 1.6227502822875977, "learning_rate": 4.691524987120041e-05, "loss": 1.1371, "step": 2400 }, { "epoch": 0.19319938176197837, "grad_norm": 1.6179583072662354, "learning_rate": 4.67864502833591e-05, "loss": 1.0538, "step": 2500 }, { "epoch": 0.2009273570324575, "grad_norm": 1.242937445640564, "learning_rate": 4.6657650695517776e-05, "loss": 0.9714, "step": 2600 }, { "epoch": 0.20865533230293662, "grad_norm": 1.4579142332077026, "learning_rate": 4.6528851107676455e-05, "loss": 1.1136, "step": 2700 }, { "epoch": 0.21638330757341576, "grad_norm": 1.5873304605484009, "learning_rate": 4.6400051519835134e-05, "loss": 1.0059, "step": 2800 }, { "epoch": 0.2241112828438949, "grad_norm": 1.6348575353622437, "learning_rate": 4.627125193199382e-05, "loss": 1.009, "step": 2900 }, { "epoch": 0.23183925811437403, "grad_norm": 1.373192548751831, "learning_rate": 4.6142452344152505e-05, "loss": 1.0069, "step": 3000 }, { "epoch": 0.23956723338485317, "grad_norm": 0.9459308385848999, "learning_rate": 4.6013652756311184e-05, "loss": 0.9666, "step": 3100 }, { "epoch": 0.2472952086553323, "grad_norm": 1.5119107961654663, "learning_rate": 4.588485316846986e-05, "loss": 1.0447, "step": 3200 }, { "epoch": 0.2550231839258114, "grad_norm": 1.3911887407302856, "learning_rate": 4.575605358062854e-05, "loss": 1.0636, "step": 3300 }, { "epoch": 0.26275115919629055, "grad_norm": 1.434175729751587, "learning_rate": 4.562725399278723e-05, "loss": 1.0425, "step": 3400 }, { "epoch": 0.2704791344667697, "grad_norm": 1.7139261960983276, "learning_rate": 4.5498454404945905e-05, "loss": 1.0795, "step": 3500 }, { "epoch": 0.2782071097372488, "grad_norm": 0.8860535025596619, "learning_rate": 4.536965481710459e-05, "loss": 0.9276, "step": 3600 }, { "epoch": 0.28593508500772796, "grad_norm": 1.6701558828353882, "learning_rate": 4.524085522926327e-05, "loss": 0.9323, "step": 3700 }, { "epoch": 0.2936630602782071, "grad_norm": 1.3985803127288818, "learning_rate": 4.511205564142195e-05, "loss": 0.9605, "step": 3800 }, { "epoch": 0.30139103554868624, "grad_norm": 0.7279998064041138, "learning_rate": 4.498325605358063e-05, "loss": 0.9523, "step": 3900 }, { "epoch": 0.3091190108191654, "grad_norm": 0.7744489312171936, "learning_rate": 4.485445646573931e-05, "loss": 1.03, "step": 4000 }, { "epoch": 0.3168469860896445, "grad_norm": 1.330718755722046, "learning_rate": 4.472565687789799e-05, "loss": 1.0305, "step": 4100 }, { "epoch": 0.32457496136012365, "grad_norm": 1.212430477142334, "learning_rate": 4.459685729005668e-05, "loss": 1.055, "step": 4200 }, { "epoch": 0.3323029366306028, "grad_norm": 1.0525803565979004, "learning_rate": 4.4468057702215356e-05, "loss": 1.0114, "step": 4300 }, { "epoch": 0.3400309119010819, "grad_norm": 1.147479772567749, "learning_rate": 4.433925811437404e-05, "loss": 0.9928, "step": 4400 }, { "epoch": 0.34775888717156106, "grad_norm": 1.313454508781433, "learning_rate": 4.4210458526532714e-05, "loss": 0.974, "step": 4500 }, { "epoch": 0.3554868624420402, "grad_norm": 1.4218354225158691, "learning_rate": 4.40816589386914e-05, "loss": 0.9413, "step": 4600 }, { "epoch": 0.36321483771251933, "grad_norm": 1.2902942895889282, "learning_rate": 4.395285935085008e-05, "loss": 1.002, "step": 4700 }, { "epoch": 0.37094281298299847, "grad_norm": 0.638493001461029, "learning_rate": 4.3824059763008763e-05, "loss": 0.9175, "step": 4800 }, { "epoch": 0.3786707882534776, "grad_norm": 1.4392635822296143, "learning_rate": 4.369526017516744e-05, "loss": 0.89, "step": 4900 }, { "epoch": 0.38639876352395675, "grad_norm": 2.024583339691162, "learning_rate": 4.356646058732613e-05, "loss": 0.9731, "step": 5000 }, { "epoch": 0.3941267387944359, "grad_norm": 1.8554579019546509, "learning_rate": 4.34376609994848e-05, "loss": 0.9251, "step": 5100 }, { "epoch": 0.401854714064915, "grad_norm": 0.7937321662902832, "learning_rate": 4.3308861411643485e-05, "loss": 0.9758, "step": 5200 }, { "epoch": 0.4095826893353941, "grad_norm": 1.3737086057662964, "learning_rate": 4.3180061823802164e-05, "loss": 1.0431, "step": 5300 }, { "epoch": 0.41731066460587324, "grad_norm": 1.1854668855667114, "learning_rate": 4.305126223596085e-05, "loss": 0.9832, "step": 5400 }, { "epoch": 0.4250386398763524, "grad_norm": 1.4984533786773682, "learning_rate": 4.292246264811953e-05, "loss": 0.9949, "step": 5500 }, { "epoch": 0.4327666151468315, "grad_norm": 0.8790501356124878, "learning_rate": 4.2793663060278214e-05, "loss": 1.0039, "step": 5600 }, { "epoch": 0.44049459041731065, "grad_norm": 0.7149144411087036, "learning_rate": 4.2664863472436886e-05, "loss": 0.9687, "step": 5700 }, { "epoch": 0.4482225656877898, "grad_norm": 1.146522045135498, "learning_rate": 4.253606388459557e-05, "loss": 0.9247, "step": 5800 }, { "epoch": 0.4559505409582689, "grad_norm": 1.8279156684875488, "learning_rate": 4.240726429675425e-05, "loss": 0.998, "step": 5900 }, { "epoch": 0.46367851622874806, "grad_norm": 0.974434494972229, "learning_rate": 4.2278464708912936e-05, "loss": 1.0235, "step": 6000 }, { "epoch": 0.4714064914992272, "grad_norm": 1.2241114377975464, "learning_rate": 4.2149665121071615e-05, "loss": 0.854, "step": 6100 }, { "epoch": 0.47913446676970634, "grad_norm": 1.2750343084335327, "learning_rate": 4.20208655332303e-05, "loss": 1.005, "step": 6200 }, { "epoch": 0.4868624420401855, "grad_norm": 1.21708345413208, "learning_rate": 4.189206594538897e-05, "loss": 0.9381, "step": 6300 }, { "epoch": 0.4945904173106646, "grad_norm": 1.3617980480194092, "learning_rate": 4.176326635754766e-05, "loss": 0.8568, "step": 6400 }, { "epoch": 0.5023183925811437, "grad_norm": 1.8272175788879395, "learning_rate": 4.163446676970634e-05, "loss": 0.8578, "step": 6500 }, { "epoch": 0.5100463678516228, "grad_norm": 0.9534468650817871, "learning_rate": 4.150566718186502e-05, "loss": 0.9316, "step": 6600 }, { "epoch": 0.517774343122102, "grad_norm": 0.8899122476577759, "learning_rate": 4.13768675940237e-05, "loss": 0.8945, "step": 6700 }, { "epoch": 0.5255023183925811, "grad_norm": 0.9442152380943298, "learning_rate": 4.1248068006182387e-05, "loss": 0.8847, "step": 6800 }, { "epoch": 0.5332302936630603, "grad_norm": 1.1126829385757446, "learning_rate": 4.111926841834106e-05, "loss": 0.9536, "step": 6900 }, { "epoch": 0.5409582689335394, "grad_norm": 1.2574111223220825, "learning_rate": 4.0990468830499744e-05, "loss": 0.9258, "step": 7000 }, { "epoch": 0.5486862442040186, "grad_norm": 1.5313997268676758, "learning_rate": 4.086166924265842e-05, "loss": 0.914, "step": 7100 }, { "epoch": 0.5564142194744977, "grad_norm": 1.064965009689331, "learning_rate": 4.073286965481711e-05, "loss": 0.9365, "step": 7200 }, { "epoch": 0.5641421947449768, "grad_norm": 1.4868042469024658, "learning_rate": 4.060407006697579e-05, "loss": 0.9697, "step": 7300 }, { "epoch": 0.5718701700154559, "grad_norm": 0.8288936018943787, "learning_rate": 4.047527047913447e-05, "loss": 0.8882, "step": 7400 }, { "epoch": 0.5795981452859351, "grad_norm": 0.5883311033248901, "learning_rate": 4.0346470891293145e-05, "loss": 0.9866, "step": 7500 }, { "epoch": 0.5873261205564142, "grad_norm": 0.6529534459114075, "learning_rate": 4.021767130345183e-05, "loss": 0.9215, "step": 7600 }, { "epoch": 0.5950540958268934, "grad_norm": 0.7607959508895874, "learning_rate": 4.008887171561051e-05, "loss": 0.9043, "step": 7700 }, { "epoch": 0.6027820710973725, "grad_norm": 1.3604129552841187, "learning_rate": 3.9960072127769195e-05, "loss": 0.9164, "step": 7800 }, { "epoch": 0.6105100463678517, "grad_norm": 1.1792867183685303, "learning_rate": 3.9831272539927873e-05, "loss": 0.9893, "step": 7900 }, { "epoch": 0.6182380216383307, "grad_norm": 0.9397594332695007, "learning_rate": 3.970247295208656e-05, "loss": 0.9113, "step": 8000 }, { "epoch": 0.6259659969088099, "grad_norm": 0.5270503759384155, "learning_rate": 3.957367336424523e-05, "loss": 0.9062, "step": 8100 }, { "epoch": 0.633693972179289, "grad_norm": 0.9222223162651062, "learning_rate": 3.944487377640392e-05, "loss": 0.9085, "step": 8200 }, { "epoch": 0.6414219474497682, "grad_norm": 1.1142535209655762, "learning_rate": 3.9316074188562595e-05, "loss": 0.9342, "step": 8300 }, { "epoch": 0.6491499227202473, "grad_norm": 0.9540684223175049, "learning_rate": 3.918727460072128e-05, "loss": 0.9318, "step": 8400 }, { "epoch": 0.6568778979907264, "grad_norm": 1.5218793153762817, "learning_rate": 3.905847501287996e-05, "loss": 0.952, "step": 8500 }, { "epoch": 0.6646058732612056, "grad_norm": 0.6363744735717773, "learning_rate": 3.8929675425038645e-05, "loss": 0.8996, "step": 8600 }, { "epoch": 0.6723338485316847, "grad_norm": 1.2216317653656006, "learning_rate": 3.880087583719732e-05, "loss": 0.9585, "step": 8700 }, { "epoch": 0.6800618238021638, "grad_norm": 0.5407654643058777, "learning_rate": 3.8672076249356e-05, "loss": 0.9532, "step": 8800 }, { "epoch": 0.6877897990726429, "grad_norm": 0.936690092086792, "learning_rate": 3.854327666151468e-05, "loss": 0.8751, "step": 8900 }, { "epoch": 0.6955177743431221, "grad_norm": 1.137471079826355, "learning_rate": 3.841447707367337e-05, "loss": 0.8881, "step": 9000 }, { "epoch": 0.7032457496136012, "grad_norm": 0.28785666823387146, "learning_rate": 3.828567748583205e-05, "loss": 0.9673, "step": 9100 }, { "epoch": 0.7109737248840804, "grad_norm": 0.8596464991569519, "learning_rate": 3.815687789799073e-05, "loss": 0.9919, "step": 9200 }, { "epoch": 0.7187017001545595, "grad_norm": 1.2839220762252808, "learning_rate": 3.802807831014941e-05, "loss": 0.9603, "step": 9300 }, { "epoch": 0.7264296754250387, "grad_norm": 1.0451067686080933, "learning_rate": 3.789927872230809e-05, "loss": 0.9291, "step": 9400 }, { "epoch": 0.7341576506955177, "grad_norm": 0.7771898508071899, "learning_rate": 3.7770479134466775e-05, "loss": 0.9414, "step": 9500 }, { "epoch": 0.7418856259659969, "grad_norm": 0.885811984539032, "learning_rate": 3.7641679546625453e-05, "loss": 0.8987, "step": 9600 }, { "epoch": 0.749613601236476, "grad_norm": 1.3268210887908936, "learning_rate": 3.751287995878414e-05, "loss": 0.9359, "step": 9700 }, { "epoch": 0.7573415765069552, "grad_norm": 0.6823452711105347, "learning_rate": 3.738408037094282e-05, "loss": 0.8704, "step": 9800 }, { "epoch": 0.7650695517774343, "grad_norm": 0.8156200647354126, "learning_rate": 3.7255280783101497e-05, "loss": 0.8954, "step": 9900 }, { "epoch": 0.7727975270479135, "grad_norm": 0.8661110997200012, "learning_rate": 3.7126481195260175e-05, "loss": 0.8889, "step": 10000 }, { "epoch": 0.7805255023183926, "grad_norm": 0.5882070064544678, "learning_rate": 3.699768160741886e-05, "loss": 0.916, "step": 10100 }, { "epoch": 0.7882534775888718, "grad_norm": 0.9691454768180847, "learning_rate": 3.686888201957754e-05, "loss": 0.834, "step": 10200 }, { "epoch": 0.7959814528593508, "grad_norm": 1.3009424209594727, "learning_rate": 3.6740082431736225e-05, "loss": 0.9496, "step": 10300 }, { "epoch": 0.80370942812983, "grad_norm": 1.1349143981933594, "learning_rate": 3.6611282843894904e-05, "loss": 0.8839, "step": 10400 }, { "epoch": 0.8114374034003091, "grad_norm": 0.924774706363678, "learning_rate": 3.648248325605358e-05, "loss": 0.9256, "step": 10500 }, { "epoch": 0.8191653786707882, "grad_norm": 0.7730056047439575, "learning_rate": 3.635368366821226e-05, "loss": 0.9233, "step": 10600 }, { "epoch": 0.8268933539412674, "grad_norm": 1.2240443229675293, "learning_rate": 3.622488408037095e-05, "loss": 0.9311, "step": 10700 }, { "epoch": 0.8346213292117465, "grad_norm": 1.022357702255249, "learning_rate": 3.6096084492529626e-05, "loss": 0.8764, "step": 10800 }, { "epoch": 0.8423493044822257, "grad_norm": 0.9591399431228638, "learning_rate": 3.596728490468831e-05, "loss": 0.8918, "step": 10900 }, { "epoch": 0.8500772797527048, "grad_norm": 0.5928884744644165, "learning_rate": 3.583848531684699e-05, "loss": 0.9446, "step": 11000 }, { "epoch": 0.8578052550231839, "grad_norm": 1.6488679647445679, "learning_rate": 3.5710973724884085e-05, "loss": 0.8841, "step": 11100 }, { "epoch": 0.865533230293663, "grad_norm": 1.309877872467041, "learning_rate": 3.5582174137042763e-05, "loss": 0.849, "step": 11200 }, { "epoch": 0.8732612055641422, "grad_norm": 1.1240078210830688, "learning_rate": 3.545337454920145e-05, "loss": 0.9454, "step": 11300 }, { "epoch": 0.8809891808346213, "grad_norm": 0.9316561222076416, "learning_rate": 3.532457496136012e-05, "loss": 0.9605, "step": 11400 }, { "epoch": 0.8887171561051005, "grad_norm": 0.7135329842567444, "learning_rate": 3.5195775373518807e-05, "loss": 0.8883, "step": 11500 }, { "epoch": 0.8964451313755796, "grad_norm": 0.9502096772193909, "learning_rate": 3.5066975785677485e-05, "loss": 0.8585, "step": 11600 }, { "epoch": 0.9041731066460588, "grad_norm": 1.4689842462539673, "learning_rate": 3.493817619783617e-05, "loss": 0.9306, "step": 11700 }, { "epoch": 0.9119010819165378, "grad_norm": 0.41735073924064636, "learning_rate": 3.480937660999485e-05, "loss": 0.8678, "step": 11800 }, { "epoch": 0.919629057187017, "grad_norm": 0.6292606592178345, "learning_rate": 3.4680577022153535e-05, "loss": 0.8864, "step": 11900 }, { "epoch": 0.9273570324574961, "grad_norm": 0.7318680882453918, "learning_rate": 3.455177743431221e-05, "loss": 0.8658, "step": 12000 }, { "epoch": 0.9350850077279753, "grad_norm": 1.8119789361953735, "learning_rate": 3.442297784647089e-05, "loss": 0.8236, "step": 12100 }, { "epoch": 0.9428129829984544, "grad_norm": 2.469245195388794, "learning_rate": 3.429417825862957e-05, "loss": 0.9092, "step": 12200 }, { "epoch": 0.9505409582689336, "grad_norm": 0.9728156924247742, "learning_rate": 3.416537867078826e-05, "loss": 0.9168, "step": 12300 }, { "epoch": 0.9582689335394127, "grad_norm": 1.0292718410491943, "learning_rate": 3.4036579082946936e-05, "loss": 0.8535, "step": 12400 }, { "epoch": 0.9659969088098919, "grad_norm": 1.157630443572998, "learning_rate": 3.390777949510562e-05, "loss": 0.8482, "step": 12500 }, { "epoch": 0.973724884080371, "grad_norm": 0.944819986820221, "learning_rate": 3.3778979907264293e-05, "loss": 0.9321, "step": 12600 }, { "epoch": 0.98145285935085, "grad_norm": 0.8043965697288513, "learning_rate": 3.365018031942298e-05, "loss": 0.905, "step": 12700 }, { "epoch": 0.9891808346213292, "grad_norm": 0.81541508436203, "learning_rate": 3.352138073158166e-05, "loss": 0.8518, "step": 12800 }, { "epoch": 0.9969088098918083, "grad_norm": 0.9929108619689941, "learning_rate": 3.339258114374034e-05, "loss": 0.8439, "step": 12900 }, { "epoch": 1.0046367851622875, "grad_norm": 1.1995909214019775, "learning_rate": 3.326378155589902e-05, "loss": 0.8566, "step": 13000 }, { "epoch": 1.0123647604327666, "grad_norm": 0.7053945064544678, "learning_rate": 3.313498196805771e-05, "loss": 0.8943, "step": 13100 }, { "epoch": 1.0200927357032457, "grad_norm": 0.7920809984207153, "learning_rate": 3.300618238021638e-05, "loss": 0.8741, "step": 13200 }, { "epoch": 1.027820710973725, "grad_norm": 1.0155510902404785, "learning_rate": 3.2877382792375065e-05, "loss": 0.8022, "step": 13300 }, { "epoch": 1.035548686244204, "grad_norm": 1.0440579652786255, "learning_rate": 3.2748583204533744e-05, "loss": 0.8103, "step": 13400 }, { "epoch": 1.0432766615146831, "grad_norm": 0.9690439105033875, "learning_rate": 3.261978361669243e-05, "loss": 0.8003, "step": 13500 }, { "epoch": 1.0510046367851622, "grad_norm": 0.9963751435279846, "learning_rate": 3.249098402885111e-05, "loss": 0.8653, "step": 13600 }, { "epoch": 1.0587326120556415, "grad_norm": 0.49158361554145813, "learning_rate": 3.2362184441009794e-05, "loss": 0.8642, "step": 13700 }, { "epoch": 1.0664605873261206, "grad_norm": 0.9656476378440857, "learning_rate": 3.2233384853168466e-05, "loss": 0.8053, "step": 13800 }, { "epoch": 1.0741885625965997, "grad_norm": 1.2530566453933716, "learning_rate": 3.210458526532715e-05, "loss": 0.8381, "step": 13900 }, { "epoch": 1.0819165378670788, "grad_norm": 0.8346812725067139, "learning_rate": 3.197578567748583e-05, "loss": 0.8065, "step": 14000 }, { "epoch": 1.089644513137558, "grad_norm": 1.1836985349655151, "learning_rate": 3.1846986089644516e-05, "loss": 0.8483, "step": 14100 }, { "epoch": 1.0973724884080371, "grad_norm": 0.9917717576026917, "learning_rate": 3.1718186501803195e-05, "loss": 0.8459, "step": 14200 }, { "epoch": 1.1051004636785162, "grad_norm": 1.0433543920516968, "learning_rate": 3.158938691396188e-05, "loss": 0.8446, "step": 14300 }, { "epoch": 1.1128284389489953, "grad_norm": 1.1880654096603394, "learning_rate": 3.146058732612055e-05, "loss": 0.8122, "step": 14400 }, { "epoch": 1.1205564142194744, "grad_norm": 0.46447160840034485, "learning_rate": 3.133178773827924e-05, "loss": 0.7701, "step": 14500 }, { "epoch": 1.1282843894899537, "grad_norm": 1.4256715774536133, "learning_rate": 3.1202988150437917e-05, "loss": 0.8635, "step": 14600 }, { "epoch": 1.1360123647604328, "grad_norm": 0.9805368781089783, "learning_rate": 3.10741885625966e-05, "loss": 0.8218, "step": 14700 }, { "epoch": 1.1437403400309119, "grad_norm": 1.1528363227844238, "learning_rate": 3.094538897475529e-05, "loss": 0.8366, "step": 14800 }, { "epoch": 1.1514683153013912, "grad_norm": 0.9413203001022339, "learning_rate": 3.0816589386913966e-05, "loss": 0.8323, "step": 14900 }, { "epoch": 1.1591962905718702, "grad_norm": 0.9010465145111084, "learning_rate": 3.0687789799072645e-05, "loss": 0.8863, "step": 15000 }, { "epoch": 1.1669242658423493, "grad_norm": 0.8075920343399048, "learning_rate": 3.0558990211231324e-05, "loss": 0.8015, "step": 15100 }, { "epoch": 1.1746522411128284, "grad_norm": 0.7727937698364258, "learning_rate": 3.043019062339001e-05, "loss": 0.822, "step": 15200 }, { "epoch": 1.1823802163833075, "grad_norm": 0.8842815160751343, "learning_rate": 3.030139103554869e-05, "loss": 0.8069, "step": 15300 }, { "epoch": 1.1901081916537868, "grad_norm": 1.687024474143982, "learning_rate": 3.017259144770737e-05, "loss": 0.8723, "step": 15400 }, { "epoch": 1.1978361669242659, "grad_norm": 0.9511739015579224, "learning_rate": 3.004379185986605e-05, "loss": 0.7482, "step": 15500 }, { "epoch": 1.205564142194745, "grad_norm": 1.404523253440857, "learning_rate": 2.9914992272024735e-05, "loss": 0.8435, "step": 15600 }, { "epoch": 1.213292117465224, "grad_norm": 1.1667975187301636, "learning_rate": 2.978619268418341e-05, "loss": 0.8717, "step": 15700 }, { "epoch": 1.2210200927357033, "grad_norm": 1.286392331123352, "learning_rate": 2.9657393096342096e-05, "loss": 0.8148, "step": 15800 }, { "epoch": 1.2287480680061824, "grad_norm": 0.915986955165863, "learning_rate": 2.9528593508500775e-05, "loss": 0.8285, "step": 15900 }, { "epoch": 1.2364760432766615, "grad_norm": 0.9131502509117126, "learning_rate": 2.9399793920659457e-05, "loss": 0.8686, "step": 16000 }, { "epoch": 1.2442040185471406, "grad_norm": 1.6465792655944824, "learning_rate": 2.9270994332818136e-05, "loss": 0.7989, "step": 16100 }, { "epoch": 1.2519319938176197, "grad_norm": 0.6225204467773438, "learning_rate": 2.914219474497682e-05, "loss": 0.8871, "step": 16200 }, { "epoch": 1.259659969088099, "grad_norm": 1.4327455759048462, "learning_rate": 2.9013395157135496e-05, "loss": 0.8531, "step": 16300 }, { "epoch": 1.267387944358578, "grad_norm": 1.1303315162658691, "learning_rate": 2.8884595569294182e-05, "loss": 0.8464, "step": 16400 }, { "epoch": 1.2751159196290571, "grad_norm": 0.8437920212745667, "learning_rate": 2.875579598145286e-05, "loss": 0.8394, "step": 16500 }, { "epoch": 1.2828438948995364, "grad_norm": 1.1533674001693726, "learning_rate": 2.8626996393611543e-05, "loss": 0.8548, "step": 16600 }, { "epoch": 1.2905718701700155, "grad_norm": 0.858174204826355, "learning_rate": 2.8498196805770222e-05, "loss": 0.8281, "step": 16700 }, { "epoch": 1.2982998454404946, "grad_norm": 0.8084927797317505, "learning_rate": 2.8369397217928907e-05, "loss": 0.8215, "step": 16800 }, { "epoch": 1.3060278207109737, "grad_norm": 1.1206783056259155, "learning_rate": 2.8240597630087583e-05, "loss": 0.8648, "step": 16900 }, { "epoch": 1.3137557959814528, "grad_norm": 1.336126685142517, "learning_rate": 2.8111798042246268e-05, "loss": 0.8386, "step": 17000 }, { "epoch": 1.321483771251932, "grad_norm": 1.350467324256897, "learning_rate": 2.798428645028336e-05, "loss": 0.7863, "step": 17100 }, { "epoch": 1.3292117465224111, "grad_norm": 1.1663061380386353, "learning_rate": 2.785548686244204e-05, "loss": 0.8944, "step": 17200 }, { "epoch": 1.3369397217928902, "grad_norm": 1.4521268606185913, "learning_rate": 2.7727975270479132e-05, "loss": 0.8633, "step": 17300 }, { "epoch": 1.3446676970633695, "grad_norm": 1.0562511682510376, "learning_rate": 2.7599175682637818e-05, "loss": 0.8684, "step": 17400 }, { "epoch": 1.3523956723338486, "grad_norm": 0.7794145941734314, "learning_rate": 2.7470376094796497e-05, "loss": 0.8334, "step": 17500 }, { "epoch": 1.3601236476043277, "grad_norm": 1.1900535821914673, "learning_rate": 2.734157650695518e-05, "loss": 0.8278, "step": 17600 }, { "epoch": 1.3678516228748068, "grad_norm": 1.2511202096939087, "learning_rate": 2.7212776919113858e-05, "loss": 0.8056, "step": 17700 }, { "epoch": 1.3755795981452859, "grad_norm": 0.5958196520805359, "learning_rate": 2.7083977331272543e-05, "loss": 0.7903, "step": 17800 }, { "epoch": 1.383307573415765, "grad_norm": 0.9310637712478638, "learning_rate": 2.695517774343122e-05, "loss": 0.806, "step": 17900 }, { "epoch": 1.3910355486862442, "grad_norm": 0.8104686737060547, "learning_rate": 2.6826378155589904e-05, "loss": 0.8067, "step": 18000 } ], "logging_steps": 100, "max_steps": 38820, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.8813026304e+16, "train_batch_size": 4, "trial_name": null, "trial_params": null }