| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.49992924461551524, | |
| "eval_steps": 276, | |
| "global_step": 1104, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.00045283446070245947, | |
| "grad_norm": 4.009160995483398, | |
| "learning_rate": 1.5e-06, | |
| "loss": 1.777, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.00045283446070245947, | |
| "eval_loss": 1.9080619812011719, | |
| "eval_runtime": 19.3772, | |
| "eval_samples_per_second": 5.883, | |
| "eval_steps_per_second": 0.774, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0009056689214049189, | |
| "grad_norm": 3.954909086227417, | |
| "learning_rate": 3e-06, | |
| "loss": 1.7658, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.0013585033821073783, | |
| "grad_norm": 3.939889907836914, | |
| "learning_rate": 4.5e-06, | |
| "loss": 1.7708, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.0018113378428098379, | |
| "grad_norm": 2.411039352416992, | |
| "learning_rate": 6e-06, | |
| "loss": 1.7529, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.0022641723035122974, | |
| "grad_norm": 1.929020643234253, | |
| "learning_rate": 7.5e-06, | |
| "loss": 1.6849, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0027170067642147567, | |
| "grad_norm": 2.1016135215759277, | |
| "learning_rate": 9e-06, | |
| "loss": 1.7142, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.003169841224917216, | |
| "grad_norm": 2.308230400085449, | |
| "learning_rate": 1.05e-05, | |
| "loss": 1.7518, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.0036226756856196757, | |
| "grad_norm": 1.8725852966308594, | |
| "learning_rate": 1.2e-05, | |
| "loss": 1.7033, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.004075510146322135, | |
| "grad_norm": 2.1712286472320557, | |
| "learning_rate": 1.3500000000000001e-05, | |
| "loss": 1.6949, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.004528344607024595, | |
| "grad_norm": 1.3823119401931763, | |
| "learning_rate": 1.5e-05, | |
| "loss": 1.647, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.004981179067727054, | |
| "grad_norm": 1.3454489707946777, | |
| "learning_rate": 1.4999999153937153e-05, | |
| "loss": 1.6661, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.005434013528429513, | |
| "grad_norm": 1.0993516445159912, | |
| "learning_rate": 1.4999996615748808e-05, | |
| "loss": 1.5813, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.005886847989131973, | |
| "grad_norm": 1.3689382076263428, | |
| "learning_rate": 1.4999992385435533e-05, | |
| "loss": 1.5947, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.006339682449834432, | |
| "grad_norm": 1.1131012439727783, | |
| "learning_rate": 1.4999986462998284e-05, | |
| "loss": 1.6002, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.006792516910536892, | |
| "grad_norm": 1.140008807182312, | |
| "learning_rate": 1.4999978848438399e-05, | |
| "loss": 1.5594, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.0072453513712393515, | |
| "grad_norm": 0.921599268913269, | |
| "learning_rate": 1.4999969541757593e-05, | |
| "loss": 1.5514, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.007698185831941811, | |
| "grad_norm": 0.9127187132835388, | |
| "learning_rate": 1.4999958542957966e-05, | |
| "loss": 1.5352, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.00815102029264427, | |
| "grad_norm": 0.7846677303314209, | |
| "learning_rate": 1.4999945852042004e-05, | |
| "loss": 1.5188, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.00860385475334673, | |
| "grad_norm": 0.843215823173523, | |
| "learning_rate": 1.4999931469012563e-05, | |
| "loss": 1.5185, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.00905668921404919, | |
| "grad_norm": 0.8540645241737366, | |
| "learning_rate": 1.4999915393872895e-05, | |
| "loss": 1.5068, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.00950952367475165, | |
| "grad_norm": 0.7134804725646973, | |
| "learning_rate": 1.4999897626626621e-05, | |
| "loss": 1.515, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.009962358135454107, | |
| "grad_norm": 0.870681643486023, | |
| "learning_rate": 1.4999878167277757e-05, | |
| "loss": 1.4962, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.010415192596156567, | |
| "grad_norm": 0.7976385951042175, | |
| "learning_rate": 1.4999857015830684e-05, | |
| "loss": 1.4902, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.010868027056859027, | |
| "grad_norm": 0.6572742462158203, | |
| "learning_rate": 1.4999834172290182e-05, | |
| "loss": 1.4845, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.011320861517561486, | |
| "grad_norm": 0.7284879684448242, | |
| "learning_rate": 1.49998096366614e-05, | |
| "loss": 1.4778, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.011773695978263946, | |
| "grad_norm": 0.7442657351493835, | |
| "learning_rate": 1.4999783408949875e-05, | |
| "loss": 1.4279, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.012226530438966406, | |
| "grad_norm": 0.7158825993537903, | |
| "learning_rate": 1.4999755489161527e-05, | |
| "loss": 1.4394, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.012679364899668864, | |
| "grad_norm": 0.6678286790847778, | |
| "learning_rate": 1.4999725877302652e-05, | |
| "loss": 1.4384, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.013132199360371324, | |
| "grad_norm": 0.6753094792366028, | |
| "learning_rate": 1.4999694573379931e-05, | |
| "loss": 1.4938, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.013585033821073783, | |
| "grad_norm": 0.8098469972610474, | |
| "learning_rate": 1.499966157740043e-05, | |
| "loss": 1.413, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.014037868281776243, | |
| "grad_norm": 0.6817219257354736, | |
| "learning_rate": 1.4999626889371588e-05, | |
| "loss": 1.4387, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.014490702742478703, | |
| "grad_norm": 0.6587890982627869, | |
| "learning_rate": 1.4999590509301237e-05, | |
| "loss": 1.3972, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.014943537203181163, | |
| "grad_norm": 0.6853736042976379, | |
| "learning_rate": 1.4999552437197582e-05, | |
| "loss": 1.4371, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.015396371663883622, | |
| "grad_norm": 0.6654878854751587, | |
| "learning_rate": 1.4999512673069211e-05, | |
| "loss": 1.3717, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.015849206124586082, | |
| "grad_norm": 0.6802534461021423, | |
| "learning_rate": 1.4999471216925097e-05, | |
| "loss": 1.3965, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.01630204058528854, | |
| "grad_norm": 0.638325035572052, | |
| "learning_rate": 1.4999428068774597e-05, | |
| "loss": 1.3658, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.016754875045991, | |
| "grad_norm": 0.6029091477394104, | |
| "learning_rate": 1.499938322862744e-05, | |
| "loss": 1.3593, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.01720770950669346, | |
| "grad_norm": 0.6502748727798462, | |
| "learning_rate": 1.4999336696493746e-05, | |
| "loss": 1.3406, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.017660543967395918, | |
| "grad_norm": 0.5874539613723755, | |
| "learning_rate": 1.4999288472384011e-05, | |
| "loss": 1.3959, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.01811337842809838, | |
| "grad_norm": 0.5745269060134888, | |
| "learning_rate": 1.499923855630912e-05, | |
| "loss": 1.3568, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.018566212888800837, | |
| "grad_norm": 0.6025118827819824, | |
| "learning_rate": 1.4999186948280329e-05, | |
| "loss": 1.3679, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.0190190473495033, | |
| "grad_norm": 0.6154868006706238, | |
| "learning_rate": 1.4999133648309286e-05, | |
| "loss": 1.362, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.019471881810205757, | |
| "grad_norm": 0.5797818303108215, | |
| "learning_rate": 1.4999078656408013e-05, | |
| "loss": 1.3357, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.019924716270908215, | |
| "grad_norm": 0.6571888327598572, | |
| "learning_rate": 1.499902197258892e-05, | |
| "loss": 1.3554, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.020377550731610676, | |
| "grad_norm": 0.6282890439033508, | |
| "learning_rate": 1.4998963596864794e-05, | |
| "loss": 1.3436, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.020830385192313134, | |
| "grad_norm": 0.541675865650177, | |
| "learning_rate": 1.4998903529248806e-05, | |
| "loss": 1.3465, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.021283219653015596, | |
| "grad_norm": 0.629550576210022, | |
| "learning_rate": 1.499884176975451e-05, | |
| "loss": 1.3405, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.021736054113718054, | |
| "grad_norm": 0.6273293495178223, | |
| "learning_rate": 1.4998778318395837e-05, | |
| "loss": 1.3425, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.022188888574420515, | |
| "grad_norm": 0.5797371864318848, | |
| "learning_rate": 1.4998713175187105e-05, | |
| "loss": 1.3066, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.022641723035122973, | |
| "grad_norm": 0.6254112720489502, | |
| "learning_rate": 1.499864634014301e-05, | |
| "loss": 1.287, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.02309455749582543, | |
| "grad_norm": 0.5634319186210632, | |
| "learning_rate": 1.4998577813278631e-05, | |
| "loss": 1.3163, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.023547391956527892, | |
| "grad_norm": 0.634515106678009, | |
| "learning_rate": 1.4998507594609432e-05, | |
| "loss": 1.293, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.02400022641723035, | |
| "grad_norm": 0.6572280526161194, | |
| "learning_rate": 1.499843568415125e-05, | |
| "loss": 1.3049, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.024453060877932812, | |
| "grad_norm": 0.627619206905365, | |
| "learning_rate": 1.4998362081920313e-05, | |
| "loss": 1.293, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.02490589533863527, | |
| "grad_norm": 0.5970916152000427, | |
| "learning_rate": 1.4998286787933226e-05, | |
| "loss": 1.3552, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.025358729799337728, | |
| "grad_norm": 0.5540199875831604, | |
| "learning_rate": 1.4998209802206977e-05, | |
| "loss": 1.3124, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.02581156426004019, | |
| "grad_norm": 0.5490045547485352, | |
| "learning_rate": 1.4998131124758935e-05, | |
| "loss": 1.2801, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.026264398720742647, | |
| "grad_norm": 0.6576207280158997, | |
| "learning_rate": 1.4998050755606851e-05, | |
| "loss": 1.2807, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.02671723318144511, | |
| "grad_norm": 0.617120623588562, | |
| "learning_rate": 1.4997968694768856e-05, | |
| "loss": 1.2434, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.027170067642147567, | |
| "grad_norm": 0.6149354577064514, | |
| "learning_rate": 1.4997884942263468e-05, | |
| "loss": 1.2947, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.02762290210285003, | |
| "grad_norm": 0.6856628060340881, | |
| "learning_rate": 1.499779949810958e-05, | |
| "loss": 1.304, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.028075736563552486, | |
| "grad_norm": 0.5651207566261292, | |
| "learning_rate": 1.4997712362326468e-05, | |
| "loss": 1.2328, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.028528571024254944, | |
| "grad_norm": 0.6854087114334106, | |
| "learning_rate": 1.4997623534933796e-05, | |
| "loss": 1.2766, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.028981405484957406, | |
| "grad_norm": 0.609596848487854, | |
| "learning_rate": 1.4997533015951603e-05, | |
| "loss": 1.2682, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.029434239945659864, | |
| "grad_norm": 0.575474202632904, | |
| "learning_rate": 1.499744080540031e-05, | |
| "loss": 1.1962, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.029887074406362325, | |
| "grad_norm": 0.6126519441604614, | |
| "learning_rate": 1.4997346903300723e-05, | |
| "loss": 1.2239, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.030339908867064783, | |
| "grad_norm": 0.5966687798500061, | |
| "learning_rate": 1.4997251309674028e-05, | |
| "loss": 1.2515, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.030792743327767245, | |
| "grad_norm": 0.7234688997268677, | |
| "learning_rate": 1.499715402454179e-05, | |
| "loss": 1.2173, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.031245577788469703, | |
| "grad_norm": 0.5534050464630127, | |
| "learning_rate": 1.4997055047925962e-05, | |
| "loss": 1.2269, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.031698412249172164, | |
| "grad_norm": 0.7033390402793884, | |
| "learning_rate": 1.499695437984887e-05, | |
| "loss": 1.2371, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.03215124670987462, | |
| "grad_norm": 0.7099838852882385, | |
| "learning_rate": 1.4996852020333232e-05, | |
| "loss": 1.2639, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.03260408117057708, | |
| "grad_norm": 0.5796670913696289, | |
| "learning_rate": 1.4996747969402139e-05, | |
| "loss": 1.2291, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.03305691563127954, | |
| "grad_norm": 0.6576344966888428, | |
| "learning_rate": 1.4996642227079065e-05, | |
| "loss": 1.1869, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.033509750091982, | |
| "grad_norm": 0.6396910548210144, | |
| "learning_rate": 1.4996534793387871e-05, | |
| "loss": 1.2227, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.03396258455268446, | |
| "grad_norm": 0.6648237109184265, | |
| "learning_rate": 1.4996425668352793e-05, | |
| "loss": 1.2715, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.03441541901338692, | |
| "grad_norm": 0.6780776381492615, | |
| "learning_rate": 1.4996314851998453e-05, | |
| "loss": 1.226, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.03486825347408938, | |
| "grad_norm": 0.5589417815208435, | |
| "learning_rate": 1.4996202344349853e-05, | |
| "loss": 1.2053, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.035321087934791835, | |
| "grad_norm": 0.5351216793060303, | |
| "learning_rate": 1.4996088145432375e-05, | |
| "loss": 1.2103, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.0357739223954943, | |
| "grad_norm": 0.4985811114311218, | |
| "learning_rate": 1.4995972255271787e-05, | |
| "loss": 1.2327, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.03622675685619676, | |
| "grad_norm": 0.5302553176879883, | |
| "learning_rate": 1.499585467389423e-05, | |
| "loss": 1.1468, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.03667959131689921, | |
| "grad_norm": 0.6337281465530396, | |
| "learning_rate": 1.499573540132624e-05, | |
| "loss": 1.2117, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.037132425777601674, | |
| "grad_norm": 0.8564434051513672, | |
| "learning_rate": 1.4995614437594721e-05, | |
| "loss": 1.1857, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.037585260238304136, | |
| "grad_norm": 0.7630729675292969, | |
| "learning_rate": 1.4995491782726968e-05, | |
| "loss": 1.2389, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.0380380946990066, | |
| "grad_norm": 0.7095062732696533, | |
| "learning_rate": 1.499536743675065e-05, | |
| "loss": 1.162, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.03849092915970905, | |
| "grad_norm": 0.7915695309638977, | |
| "learning_rate": 1.4995241399693827e-05, | |
| "loss": 1.2049, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.03894376362041151, | |
| "grad_norm": 0.6757071018218994, | |
| "learning_rate": 1.4995113671584933e-05, | |
| "loss": 1.2344, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.039396598081113975, | |
| "grad_norm": 0.8018056750297546, | |
| "learning_rate": 1.4994984252452782e-05, | |
| "loss": 1.1507, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.03984943254181643, | |
| "grad_norm": 0.8417953848838806, | |
| "learning_rate": 1.4994853142326578e-05, | |
| "loss": 1.1799, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.04030226700251889, | |
| "grad_norm": 0.7193191051483154, | |
| "learning_rate": 1.4994720341235898e-05, | |
| "loss": 1.1755, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.04075510146322135, | |
| "grad_norm": 1.2904716730117798, | |
| "learning_rate": 1.4994585849210707e-05, | |
| "loss": 1.1643, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.041207935923923814, | |
| "grad_norm": 0.7058910727500916, | |
| "learning_rate": 1.4994449666281348e-05, | |
| "loss": 1.1737, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.04166077038462627, | |
| "grad_norm": 0.9418469667434692, | |
| "learning_rate": 1.4994311792478543e-05, | |
| "loss": 1.1885, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.04211360484532873, | |
| "grad_norm": 0.8901066780090332, | |
| "learning_rate": 1.4994172227833402e-05, | |
| "loss": 1.1702, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.04256643930603119, | |
| "grad_norm": 0.5510401129722595, | |
| "learning_rate": 1.4994030972377414e-05, | |
| "loss": 1.1792, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.043019273766733646, | |
| "grad_norm": 0.7584707140922546, | |
| "learning_rate": 1.4993888026142448e-05, | |
| "loss": 1.1627, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.04347210822743611, | |
| "grad_norm": 0.7134795188903809, | |
| "learning_rate": 1.499374338916075e-05, | |
| "loss": 1.236, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.04392494268813857, | |
| "grad_norm": 0.6071385145187378, | |
| "learning_rate": 1.499359706146496e-05, | |
| "loss": 1.1651, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.04437777714884103, | |
| "grad_norm": 0.6142598390579224, | |
| "learning_rate": 1.4993449043088088e-05, | |
| "loss": 1.129, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.044830611609543485, | |
| "grad_norm": 0.713774561882019, | |
| "learning_rate": 1.4993299334063528e-05, | |
| "loss": 1.1959, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.045283446070245946, | |
| "grad_norm": 0.6220033764839172, | |
| "learning_rate": 1.499314793442506e-05, | |
| "loss": 1.1464, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.04573628053094841, | |
| "grad_norm": 0.9813457727432251, | |
| "learning_rate": 1.4992994844206843e-05, | |
| "loss": 1.1228, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.04618911499165086, | |
| "grad_norm": 0.9431173205375671, | |
| "learning_rate": 1.499284006344341e-05, | |
| "loss": 1.1478, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.04664194945235332, | |
| "grad_norm": 0.5895239114761353, | |
| "learning_rate": 1.4992683592169691e-05, | |
| "loss": 1.1035, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.047094783913055785, | |
| "grad_norm": 1.2074604034423828, | |
| "learning_rate": 1.4992525430420984e-05, | |
| "loss": 1.1472, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.047547618373758246, | |
| "grad_norm": 0.8212599158287048, | |
| "learning_rate": 1.499236557823297e-05, | |
| "loss": 1.1222, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.0480004528344607, | |
| "grad_norm": 0.7460811138153076, | |
| "learning_rate": 1.4992204035641721e-05, | |
| "loss": 1.1551, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.04845328729516316, | |
| "grad_norm": 0.9674639105796814, | |
| "learning_rate": 1.499204080268368e-05, | |
| "loss": 1.15, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.048906121755865624, | |
| "grad_norm": 0.7379394173622131, | |
| "learning_rate": 1.4991875879395677e-05, | |
| "loss": 1.096, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.04935895621656808, | |
| "grad_norm": 0.7826714515686035, | |
| "learning_rate": 1.4991709265814918e-05, | |
| "loss": 1.1086, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.04981179067727054, | |
| "grad_norm": 0.7406508326530457, | |
| "learning_rate": 1.4991540961978997e-05, | |
| "loss": 1.1203, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.050264625137973, | |
| "grad_norm": 0.6145086884498596, | |
| "learning_rate": 1.4991370967925882e-05, | |
| "loss": 1.1688, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.050717459598675456, | |
| "grad_norm": 0.7194088697433472, | |
| "learning_rate": 1.4991199283693933e-05, | |
| "loss": 1.0994, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.05117029405937792, | |
| "grad_norm": 0.6828657984733582, | |
| "learning_rate": 1.499102590932188e-05, | |
| "loss": 1.1362, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.05162312852008038, | |
| "grad_norm": 0.6402639150619507, | |
| "learning_rate": 1.4990850844848841e-05, | |
| "loss": 1.1146, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.05207596298078284, | |
| "grad_norm": 0.7166230082511902, | |
| "learning_rate": 1.4990674090314313e-05, | |
| "loss": 1.085, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.052528797441485295, | |
| "grad_norm": 0.6384372711181641, | |
| "learning_rate": 1.4990495645758174e-05, | |
| "loss": 1.0982, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.052981631902187756, | |
| "grad_norm": 0.6809221506118774, | |
| "learning_rate": 1.4990315511220685e-05, | |
| "loss": 1.1414, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.05343446636289022, | |
| "grad_norm": 0.6124834418296814, | |
| "learning_rate": 1.4990133686742488e-05, | |
| "loss": 1.0886, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.05388730082359267, | |
| "grad_norm": 0.7249141335487366, | |
| "learning_rate": 1.4989950172364603e-05, | |
| "loss": 1.1568, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.054340135284295134, | |
| "grad_norm": 0.6189077496528625, | |
| "learning_rate": 1.498976496812844e-05, | |
| "loss": 1.0731, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.054792969744997595, | |
| "grad_norm": 0.8025009036064148, | |
| "learning_rate": 1.4989578074075777e-05, | |
| "loss": 1.1252, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.05524580420570006, | |
| "grad_norm": 0.6181627511978149, | |
| "learning_rate": 1.4989389490248783e-05, | |
| "loss": 1.0713, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.05569863866640251, | |
| "grad_norm": 0.8665769696235657, | |
| "learning_rate": 1.4989199216690006e-05, | |
| "loss": 1.1481, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.05615147312710497, | |
| "grad_norm": 0.7832885384559631, | |
| "learning_rate": 1.4989007253442377e-05, | |
| "loss": 1.0775, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.056604307587807434, | |
| "grad_norm": 0.7431057095527649, | |
| "learning_rate": 1.4988813600549202e-05, | |
| "loss": 1.1013, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.05705714204850989, | |
| "grad_norm": 1.1883816719055176, | |
| "learning_rate": 1.4988618258054176e-05, | |
| "loss": 1.1003, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.05750997650921235, | |
| "grad_norm": 0.6886972188949585, | |
| "learning_rate": 1.498842122600137e-05, | |
| "loss": 1.0754, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.05796281096991481, | |
| "grad_norm": 0.8941016793251038, | |
| "learning_rate": 1.4988222504435235e-05, | |
| "loss": 1.0577, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.05841564543061727, | |
| "grad_norm": 0.9873454570770264, | |
| "learning_rate": 1.498802209340061e-05, | |
| "loss": 1.089, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.05886847989131973, | |
| "grad_norm": 0.6695548892021179, | |
| "learning_rate": 1.4987819992942712e-05, | |
| "loss": 1.0799, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.05932131435202219, | |
| "grad_norm": 0.9906266331672668, | |
| "learning_rate": 1.4987616203107134e-05, | |
| "loss": 1.0753, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.05977414881272465, | |
| "grad_norm": 0.6169360876083374, | |
| "learning_rate": 1.4987410723939857e-05, | |
| "loss": 1.0863, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.060226983273427105, | |
| "grad_norm": 1.0210751295089722, | |
| "learning_rate": 1.4987203555487242e-05, | |
| "loss": 1.0493, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.06067981773412957, | |
| "grad_norm": 0.675287663936615, | |
| "learning_rate": 1.4986994697796022e-05, | |
| "loss": 1.0669, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.06113265219483203, | |
| "grad_norm": 0.8825522661209106, | |
| "learning_rate": 1.4986784150913329e-05, | |
| "loss": 1.0482, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.06158548665553449, | |
| "grad_norm": 0.6397138237953186, | |
| "learning_rate": 1.4986571914886662e-05, | |
| "loss": 1.0544, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.062038321116236944, | |
| "grad_norm": 1.0741486549377441, | |
| "learning_rate": 1.4986357989763901e-05, | |
| "loss": 1.037, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.062491155576939406, | |
| "grad_norm": 0.6419923901557922, | |
| "learning_rate": 1.4986142375593315e-05, | |
| "loss": 1.0751, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.06294399003764187, | |
| "grad_norm": 0.8368469476699829, | |
| "learning_rate": 1.498592507242355e-05, | |
| "loss": 1.0357, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.06339682449834433, | |
| "grad_norm": 0.7393881678581238, | |
| "learning_rate": 1.4985706080303634e-05, | |
| "loss": 1.0733, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.06384965895904679, | |
| "grad_norm": 0.8628896474838257, | |
| "learning_rate": 1.4985485399282973e-05, | |
| "loss": 1.0808, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.06430249341974924, | |
| "grad_norm": 0.6672838926315308, | |
| "learning_rate": 1.4985263029411356e-05, | |
| "loss": 1.0266, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.0647553278804517, | |
| "grad_norm": 0.7643623352050781, | |
| "learning_rate": 1.4985038970738959e-05, | |
| "loss": 1.0044, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.06520816234115416, | |
| "grad_norm": 0.8692808747291565, | |
| "learning_rate": 1.4984813223316326e-05, | |
| "loss": 1.0128, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.06566099680185662, | |
| "grad_norm": 0.7556530833244324, | |
| "learning_rate": 1.4984585787194392e-05, | |
| "loss": 1.0772, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.06611383126255908, | |
| "grad_norm": 0.8598385453224182, | |
| "learning_rate": 1.4984356662424473e-05, | |
| "loss": 1.0511, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.06656666572326154, | |
| "grad_norm": 0.6620029807090759, | |
| "learning_rate": 1.498412584905826e-05, | |
| "loss": 1.0401, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.067019500183964, | |
| "grad_norm": 0.8549764752388, | |
| "learning_rate": 1.498389334714783e-05, | |
| "loss": 1.0172, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.06747233464466645, | |
| "grad_norm": 0.6814466714859009, | |
| "learning_rate": 1.498365915674564e-05, | |
| "loss": 1.0638, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.06792516910536892, | |
| "grad_norm": 0.7864624261856079, | |
| "learning_rate": 1.4983423277904526e-05, | |
| "loss": 1.0069, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.06837800356607138, | |
| "grad_norm": 0.7862728238105774, | |
| "learning_rate": 1.4983185710677705e-05, | |
| "loss": 1.0092, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.06883083802677384, | |
| "grad_norm": 0.656066358089447, | |
| "learning_rate": 1.498294645511878e-05, | |
| "loss": 0.9948, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.0692836724874763, | |
| "grad_norm": 0.8510375022888184, | |
| "learning_rate": 1.4982705511281728e-05, | |
| "loss": 1.0411, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.06973650694817876, | |
| "grad_norm": 0.6606243848800659, | |
| "learning_rate": 1.4982462879220911e-05, | |
| "loss": 1.0054, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.07018934140888121, | |
| "grad_norm": 0.7300797700881958, | |
| "learning_rate": 1.498221855899107e-05, | |
| "loss": 0.9801, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.07064217586958367, | |
| "grad_norm": 0.8046183586120605, | |
| "learning_rate": 1.4981972550647328e-05, | |
| "loss": 0.9794, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.07109501033028613, | |
| "grad_norm": 0.6511421203613281, | |
| "learning_rate": 1.4981724854245189e-05, | |
| "loss": 1.065, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.0715478447909886, | |
| "grad_norm": 0.8642223477363586, | |
| "learning_rate": 1.4981475469840538e-05, | |
| "loss": 0.9808, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.07200067925169105, | |
| "grad_norm": 0.7534111738204956, | |
| "learning_rate": 1.498122439748964e-05, | |
| "loss": 1.0371, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.07245351371239352, | |
| "grad_norm": 0.7241458892822266, | |
| "learning_rate": 1.4980971637249141e-05, | |
| "loss": 0.9568, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.07290634817309598, | |
| "grad_norm": 0.6701076030731201, | |
| "learning_rate": 1.4980717189176066e-05, | |
| "loss": 1.0158, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.07335918263379843, | |
| "grad_norm": 0.6040867567062378, | |
| "learning_rate": 1.4980461053327829e-05, | |
| "loss": 0.9966, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.07381201709450089, | |
| "grad_norm": 0.7389962077140808, | |
| "learning_rate": 1.4980203229762208e-05, | |
| "loss": 0.962, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.07426485155520335, | |
| "grad_norm": 0.6864858269691467, | |
| "learning_rate": 1.4979943718537383e-05, | |
| "loss": 1.0563, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.07471768601590581, | |
| "grad_norm": 0.6471516489982605, | |
| "learning_rate": 1.4979682519711897e-05, | |
| "loss": 1.0448, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.07517052047660827, | |
| "grad_norm": 0.7508234977722168, | |
| "learning_rate": 1.4979419633344686e-05, | |
| "loss": 1.0353, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.07562335493731073, | |
| "grad_norm": 0.6425490379333496, | |
| "learning_rate": 1.4979155059495056e-05, | |
| "loss": 0.9819, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.0760761893980132, | |
| "grad_norm": 0.6431488990783691, | |
| "learning_rate": 1.4978888798222703e-05, | |
| "loss": 1.0394, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.07652902385871564, | |
| "grad_norm": 0.5777195692062378, | |
| "learning_rate": 1.49786208495877e-05, | |
| "loss": 0.9881, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.0769818583194181, | |
| "grad_norm": 0.7131640315055847, | |
| "learning_rate": 1.4978351213650498e-05, | |
| "loss": 1.0054, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.07743469278012056, | |
| "grad_norm": 0.7139365673065186, | |
| "learning_rate": 1.4978079890471935e-05, | |
| "loss": 1.0341, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.07788752724082303, | |
| "grad_norm": 0.6020310521125793, | |
| "learning_rate": 1.4977806880113223e-05, | |
| "loss": 0.9785, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.07834036170152549, | |
| "grad_norm": 0.7464612126350403, | |
| "learning_rate": 1.4977532182635963e-05, | |
| "loss": 0.9976, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.07879319616222795, | |
| "grad_norm": 0.8518685698509216, | |
| "learning_rate": 1.4977255798102122e-05, | |
| "loss": 0.9976, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.07924603062293041, | |
| "grad_norm": 0.6868304014205933, | |
| "learning_rate": 1.4976977726574065e-05, | |
| "loss": 0.9989, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.07969886508363286, | |
| "grad_norm": 0.7775158882141113, | |
| "learning_rate": 1.4976697968114529e-05, | |
| "loss": 0.9655, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.08015169954433532, | |
| "grad_norm": 0.9228640794754028, | |
| "learning_rate": 1.4976416522786626e-05, | |
| "loss": 0.9497, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.08060453400503778, | |
| "grad_norm": 0.6497952342033386, | |
| "learning_rate": 1.4976133390653861e-05, | |
| "loss": 0.94, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.08105736846574024, | |
| "grad_norm": 0.853291928768158, | |
| "learning_rate": 1.497584857178011e-05, | |
| "loss": 0.9267, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.0815102029264427, | |
| "grad_norm": 1.0284061431884766, | |
| "learning_rate": 1.4975562066229635e-05, | |
| "loss": 0.9772, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.08196303738714517, | |
| "grad_norm": 0.7046072483062744, | |
| "learning_rate": 1.4975273874067078e-05, | |
| "loss": 1.0039, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.08241587184784763, | |
| "grad_norm": 0.8511521220207214, | |
| "learning_rate": 1.4974983995357456e-05, | |
| "loss": 0.935, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.08286870630855007, | |
| "grad_norm": 0.9605690240859985, | |
| "learning_rate": 1.497469243016617e-05, | |
| "loss": 0.9885, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.08332154076925254, | |
| "grad_norm": 0.7198604345321655, | |
| "learning_rate": 1.4974399178559007e-05, | |
| "loss": 0.9757, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.083774375229955, | |
| "grad_norm": 0.8301750421524048, | |
| "learning_rate": 1.4974104240602127e-05, | |
| "loss": 0.9232, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.08422720969065746, | |
| "grad_norm": 0.8821402788162231, | |
| "learning_rate": 1.497380761636207e-05, | |
| "loss": 0.9657, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.08468004415135992, | |
| "grad_norm": 1.1182292699813843, | |
| "learning_rate": 1.4973509305905762e-05, | |
| "loss": 1.005, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.08513287861206238, | |
| "grad_norm": 0.7202256917953491, | |
| "learning_rate": 1.4973209309300508e-05, | |
| "loss": 0.916, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.08558571307276484, | |
| "grad_norm": 1.0314316749572754, | |
| "learning_rate": 1.4972907626613993e-05, | |
| "loss": 0.9722, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.08603854753346729, | |
| "grad_norm": 0.8789144158363342, | |
| "learning_rate": 1.4972604257914277e-05, | |
| "loss": 0.9835, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.08649138199416975, | |
| "grad_norm": 0.8964241743087769, | |
| "learning_rate": 1.497229920326981e-05, | |
| "loss": 0.9562, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.08694421645487221, | |
| "grad_norm": 0.9407457709312439, | |
| "learning_rate": 1.4971992462749413e-05, | |
| "loss": 0.935, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.08739705091557468, | |
| "grad_norm": 0.7482190132141113, | |
| "learning_rate": 1.4971684036422295e-05, | |
| "loss": 0.9561, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.08784988537627714, | |
| "grad_norm": 0.7358856201171875, | |
| "learning_rate": 1.4971373924358041e-05, | |
| "loss": 0.9106, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.0883027198369796, | |
| "grad_norm": 0.7911369204521179, | |
| "learning_rate": 1.4971062126626617e-05, | |
| "loss": 0.9625, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.08875555429768206, | |
| "grad_norm": 0.6739444136619568, | |
| "learning_rate": 1.497074864329837e-05, | |
| "loss": 0.936, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.08920838875838451, | |
| "grad_norm": 0.7796293497085571, | |
| "learning_rate": 1.497043347444403e-05, | |
| "loss": 0.9645, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.08966122321908697, | |
| "grad_norm": 0.7281574010848999, | |
| "learning_rate": 1.4970116620134701e-05, | |
| "loss": 0.985, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.09011405767978943, | |
| "grad_norm": 0.7205932140350342, | |
| "learning_rate": 1.4969798080441872e-05, | |
| "loss": 0.9253, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.09056689214049189, | |
| "grad_norm": 0.8215734958648682, | |
| "learning_rate": 1.496947785543741e-05, | |
| "loss": 0.9391, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.09101972660119435, | |
| "grad_norm": 0.6455587148666382, | |
| "learning_rate": 1.4969155945193562e-05, | |
| "loss": 0.8646, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.09147256106189681, | |
| "grad_norm": 0.6836207509040833, | |
| "learning_rate": 1.496883234978296e-05, | |
| "loss": 0.9728, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.09192539552259928, | |
| "grad_norm": 0.6182245016098022, | |
| "learning_rate": 1.496850706927861e-05, | |
| "loss": 0.8912, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.09237822998330172, | |
| "grad_norm": 1.1609197854995728, | |
| "learning_rate": 1.4968180103753901e-05, | |
| "loss": 0.9035, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.09283106444400419, | |
| "grad_norm": 0.7769509553909302, | |
| "learning_rate": 1.4967851453282601e-05, | |
| "loss": 0.8651, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.09328389890470665, | |
| "grad_norm": 0.8045849204063416, | |
| "learning_rate": 1.4967521117938861e-05, | |
| "loss": 0.9432, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.09373673336540911, | |
| "grad_norm": 0.747963547706604, | |
| "learning_rate": 1.496718909779721e-05, | |
| "loss": 0.9142, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.09418956782611157, | |
| "grad_norm": 0.8653061389923096, | |
| "learning_rate": 1.4966855392932558e-05, | |
| "loss": 0.9138, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.09464240228681403, | |
| "grad_norm": 0.7377287745475769, | |
| "learning_rate": 1.4966520003420191e-05, | |
| "loss": 0.9411, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.09509523674751649, | |
| "grad_norm": 0.8704388737678528, | |
| "learning_rate": 1.496618292933578e-05, | |
| "loss": 0.9223, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.09554807120821894, | |
| "grad_norm": 0.8722823262214661, | |
| "learning_rate": 1.4965844170755376e-05, | |
| "loss": 0.9261, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.0960009056689214, | |
| "grad_norm": 0.6991630792617798, | |
| "learning_rate": 1.4965503727755408e-05, | |
| "loss": 0.9053, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.09645374012962386, | |
| "grad_norm": 0.762243926525116, | |
| "learning_rate": 1.4965161600412686e-05, | |
| "loss": 0.8989, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.09690657459032632, | |
| "grad_norm": 0.6228227019309998, | |
| "learning_rate": 1.4964817788804398e-05, | |
| "loss": 0.897, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.09735940905102879, | |
| "grad_norm": 0.871157705783844, | |
| "learning_rate": 1.4964472293008114e-05, | |
| "loss": 0.9551, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.09781224351173125, | |
| "grad_norm": 0.7100203037261963, | |
| "learning_rate": 1.4964125113101787e-05, | |
| "loss": 0.9334, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.0982650779724337, | |
| "grad_norm": 0.6398739218711853, | |
| "learning_rate": 1.4963776249163742e-05, | |
| "loss": 0.9256, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.09871791243313616, | |
| "grad_norm": 0.7361502051353455, | |
| "learning_rate": 1.496342570127269e-05, | |
| "loss": 0.9207, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.09917074689383862, | |
| "grad_norm": 0.6598243713378906, | |
| "learning_rate": 1.4963073469507722e-05, | |
| "loss": 0.9029, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.09962358135454108, | |
| "grad_norm": 0.7405319213867188, | |
| "learning_rate": 1.4962719553948306e-05, | |
| "loss": 0.9661, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.10007641581524354, | |
| "grad_norm": 0.7439802289009094, | |
| "learning_rate": 1.4962363954674294e-05, | |
| "loss": 0.9133, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.100529250275946, | |
| "grad_norm": 0.7841500639915466, | |
| "learning_rate": 1.4962006671765911e-05, | |
| "loss": 0.8958, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.10098208473664846, | |
| "grad_norm": 0.8495052456855774, | |
| "learning_rate": 1.4961647705303765e-05, | |
| "loss": 0.9755, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.10143491919735091, | |
| "grad_norm": 0.7986276149749756, | |
| "learning_rate": 1.4961287055368853e-05, | |
| "loss": 0.994, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.10188775365805337, | |
| "grad_norm": 0.8611103892326355, | |
| "learning_rate": 1.4960924722042536e-05, | |
| "loss": 0.8454, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.10234058811875583, | |
| "grad_norm": 1.167007565498352, | |
| "learning_rate": 1.4960560705406563e-05, | |
| "loss": 0.9154, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.1027934225794583, | |
| "grad_norm": 0.7711225748062134, | |
| "learning_rate": 1.4960195005543066e-05, | |
| "loss": 0.8714, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.10324625704016076, | |
| "grad_norm": 0.7879083752632141, | |
| "learning_rate": 1.4959827622534551e-05, | |
| "loss": 0.8909, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.10369909150086322, | |
| "grad_norm": 0.964492678642273, | |
| "learning_rate": 1.4959458556463905e-05, | |
| "loss": 0.8881, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.10415192596156568, | |
| "grad_norm": 1.0750067234039307, | |
| "learning_rate": 1.4959087807414397e-05, | |
| "loss": 0.8857, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.10460476042226813, | |
| "grad_norm": 0.7446492910385132, | |
| "learning_rate": 1.4958715375469674e-05, | |
| "loss": 0.9002, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.10505759488297059, | |
| "grad_norm": 1.0399891138076782, | |
| "learning_rate": 1.4958341260713762e-05, | |
| "loss": 0.9695, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.10551042934367305, | |
| "grad_norm": 0.7288771867752075, | |
| "learning_rate": 1.495796546323107e-05, | |
| "loss": 0.9033, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.10596326380437551, | |
| "grad_norm": 1.0252315998077393, | |
| "learning_rate": 1.495758798310638e-05, | |
| "loss": 0.8462, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.10641609826507797, | |
| "grad_norm": 0.7485429644584656, | |
| "learning_rate": 1.4957208820424859e-05, | |
| "loss": 0.8822, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.10686893272578044, | |
| "grad_norm": 0.9481981992721558, | |
| "learning_rate": 1.4956827975272054e-05, | |
| "loss": 0.8892, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.1073217671864829, | |
| "grad_norm": 0.8089492917060852, | |
| "learning_rate": 1.495644544773389e-05, | |
| "loss": 0.8245, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.10777460164718534, | |
| "grad_norm": 0.9913555383682251, | |
| "learning_rate": 1.4956061237896671e-05, | |
| "loss": 0.8442, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.1082274361078878, | |
| "grad_norm": 0.8869263529777527, | |
| "learning_rate": 1.4955675345847084e-05, | |
| "loss": 0.9323, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.10868027056859027, | |
| "grad_norm": 0.8241413235664368, | |
| "learning_rate": 1.4955287771672187e-05, | |
| "loss": 0.8985, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.10913310502929273, | |
| "grad_norm": 0.7058445811271667, | |
| "learning_rate": 1.4954898515459428e-05, | |
| "loss": 0.817, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.10958593948999519, | |
| "grad_norm": 0.7939592599868774, | |
| "learning_rate": 1.4954507577296628e-05, | |
| "loss": 0.7496, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.11003877395069765, | |
| "grad_norm": 0.8440166115760803, | |
| "learning_rate": 1.4954114957271988e-05, | |
| "loss": 0.9368, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.11049160841140011, | |
| "grad_norm": 0.8787747621536255, | |
| "learning_rate": 1.4953720655474092e-05, | |
| "loss": 0.8562, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.11094444287210256, | |
| "grad_norm": 0.7784985899925232, | |
| "learning_rate": 1.4953324671991897e-05, | |
| "loss": 0.8718, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.11139727733280502, | |
| "grad_norm": 0.9408149719238281, | |
| "learning_rate": 1.495292700691475e-05, | |
| "loss": 0.9013, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.11185011179350748, | |
| "grad_norm": 0.8225075602531433, | |
| "learning_rate": 1.4952527660332368e-05, | |
| "loss": 0.8769, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.11230294625420995, | |
| "grad_norm": 0.7701689004898071, | |
| "learning_rate": 1.4952126632334847e-05, | |
| "loss": 0.8824, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.11275578071491241, | |
| "grad_norm": 0.7905251979827881, | |
| "learning_rate": 1.4951723923012667e-05, | |
| "loss": 0.8753, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.11320861517561487, | |
| "grad_norm": 0.7921510338783264, | |
| "learning_rate": 1.495131953245669e-05, | |
| "loss": 0.8127, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.11366144963631733, | |
| "grad_norm": 0.8098307251930237, | |
| "learning_rate": 1.495091346075815e-05, | |
| "loss": 0.9178, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.11411428409701978, | |
| "grad_norm": 0.789347767829895, | |
| "learning_rate": 1.4950505708008665e-05, | |
| "loss": 0.8268, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.11456711855772224, | |
| "grad_norm": 0.8333251476287842, | |
| "learning_rate": 1.495009627430023e-05, | |
| "loss": 0.8422, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.1150199530184247, | |
| "grad_norm": 0.9456163048744202, | |
| "learning_rate": 1.4949685159725219e-05, | |
| "loss": 0.8445, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.11547278747912716, | |
| "grad_norm": 0.7319315075874329, | |
| "learning_rate": 1.4949272364376389e-05, | |
| "loss": 0.8195, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.11592562193982962, | |
| "grad_norm": 0.9746232628822327, | |
| "learning_rate": 1.494885788834687e-05, | |
| "loss": 0.8909, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.11637845640053208, | |
| "grad_norm": 0.9130688905715942, | |
| "learning_rate": 1.4948441731730177e-05, | |
| "loss": 0.8685, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.11683129086123455, | |
| "grad_norm": 0.7087278962135315, | |
| "learning_rate": 1.4948023894620201e-05, | |
| "loss": 0.9413, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.117284125321937, | |
| "grad_norm": 0.9176363348960876, | |
| "learning_rate": 1.4947604377111216e-05, | |
| "loss": 0.8186, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.11773695978263946, | |
| "grad_norm": 0.8463951945304871, | |
| "learning_rate": 1.494718317929787e-05, | |
| "loss": 0.8986, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.11818979424334192, | |
| "grad_norm": 0.8302780389785767, | |
| "learning_rate": 1.494676030127519e-05, | |
| "loss": 0.8204, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.11864262870404438, | |
| "grad_norm": 0.7446224093437195, | |
| "learning_rate": 1.4946335743138587e-05, | |
| "loss": 0.8383, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.11909546316474684, | |
| "grad_norm": 0.7819026112556458, | |
| "learning_rate": 1.4945909504983848e-05, | |
| "loss": 0.8248, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.1195482976254493, | |
| "grad_norm": 0.7390825152397156, | |
| "learning_rate": 1.4945481586907141e-05, | |
| "loss": 0.8604, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.12000113208615176, | |
| "grad_norm": 0.7803400158882141, | |
| "learning_rate": 1.494505198900501e-05, | |
| "loss": 0.8302, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.12045396654685421, | |
| "grad_norm": 0.9145985841751099, | |
| "learning_rate": 1.4944620711374377e-05, | |
| "loss": 0.8732, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.12090680100755667, | |
| "grad_norm": 0.7339340448379517, | |
| "learning_rate": 1.494418775411255e-05, | |
| "loss": 0.8042, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.12135963546825913, | |
| "grad_norm": 0.7937312722206116, | |
| "learning_rate": 1.4943753117317208e-05, | |
| "loss": 0.8347, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.1218124699289616, | |
| "grad_norm": 0.7886615991592407, | |
| "learning_rate": 1.4943316801086414e-05, | |
| "loss": 0.8511, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.12226530438966406, | |
| "grad_norm": 0.8435689210891724, | |
| "learning_rate": 1.4942878805518608e-05, | |
| "loss": 0.8187, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.12271813885036652, | |
| "grad_norm": 0.6933366060256958, | |
| "learning_rate": 1.494243913071261e-05, | |
| "loss": 0.762, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.12317097331106898, | |
| "grad_norm": 0.743684709072113, | |
| "learning_rate": 1.4941997776767616e-05, | |
| "loss": 0.8686, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.12362380777177143, | |
| "grad_norm": 0.772577702999115, | |
| "learning_rate": 1.4941554743783204e-05, | |
| "loss": 0.8271, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.12407664223247389, | |
| "grad_norm": 0.7139607667922974, | |
| "learning_rate": 1.4941110031859327e-05, | |
| "loss": 0.7724, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.12452947669317635, | |
| "grad_norm": 0.6617469191551208, | |
| "learning_rate": 1.4940663641096325e-05, | |
| "loss": 0.8092, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.12498231115387881, | |
| "grad_norm": 0.841748833656311, | |
| "learning_rate": 1.4940215571594908e-05, | |
| "loss": 0.9361, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.12498231115387881, | |
| "eval_loss": 0.8086027503013611, | |
| "eval_runtime": 19.4718, | |
| "eval_samples_per_second": 5.855, | |
| "eval_steps_per_second": 0.77, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.12543514561458127, | |
| "grad_norm": 0.6689454913139343, | |
| "learning_rate": 1.4939765823456168e-05, | |
| "loss": 0.8363, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.12588798007528373, | |
| "grad_norm": 0.7700071334838867, | |
| "learning_rate": 1.4939314396781575e-05, | |
| "loss": 0.8393, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.1263408145359862, | |
| "grad_norm": 0.7219974994659424, | |
| "learning_rate": 1.4938861291672983e-05, | |
| "loss": 0.8039, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.12679364899668866, | |
| "grad_norm": 0.6459200382232666, | |
| "learning_rate": 1.4938406508232613e-05, | |
| "loss": 0.819, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.12724648345739112, | |
| "grad_norm": 0.7447069883346558, | |
| "learning_rate": 1.4937950046563075e-05, | |
| "loss": 0.8567, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.12769931791809358, | |
| "grad_norm": 0.8317793011665344, | |
| "learning_rate": 1.4937491906767357e-05, | |
| "loss": 0.8555, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.128152152378796, | |
| "grad_norm": 0.6332393884658813, | |
| "learning_rate": 1.4937032088948819e-05, | |
| "loss": 0.851, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.12860498683949848, | |
| "grad_norm": 0.6920958757400513, | |
| "learning_rate": 1.4936570593211203e-05, | |
| "loss": 0.7737, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.12905782130020094, | |
| "grad_norm": 0.7432802319526672, | |
| "learning_rate": 1.4936107419658635e-05, | |
| "loss": 0.8177, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.1295106557609034, | |
| "grad_norm": 0.7308900952339172, | |
| "learning_rate": 1.4935642568395613e-05, | |
| "loss": 0.751, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.12996349022160586, | |
| "grad_norm": 0.9042837619781494, | |
| "learning_rate": 1.4935176039527014e-05, | |
| "loss": 0.8192, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.13041632468230832, | |
| "grad_norm": 0.7064964175224304, | |
| "learning_rate": 1.4934707833158094e-05, | |
| "loss": 0.7918, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.13086915914301078, | |
| "grad_norm": 1.04722261428833, | |
| "learning_rate": 1.4934237949394492e-05, | |
| "loss": 0.8003, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.13132199360371324, | |
| "grad_norm": 0.6335585117340088, | |
| "learning_rate": 1.4933766388342215e-05, | |
| "loss": 0.7773, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.1317748280644157, | |
| "grad_norm": 0.8007393479347229, | |
| "learning_rate": 1.4933293150107663e-05, | |
| "loss": 0.8016, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.13222766252511817, | |
| "grad_norm": 0.7477773427963257, | |
| "learning_rate": 1.49328182347976e-05, | |
| "loss": 0.7772, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.13268049698582063, | |
| "grad_norm": 0.8117051124572754, | |
| "learning_rate": 1.4932341642519178e-05, | |
| "loss": 0.8021, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.1331333314465231, | |
| "grad_norm": 1.086212158203125, | |
| "learning_rate": 1.4931863373379923e-05, | |
| "loss": 0.8271, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.13358616590722555, | |
| "grad_norm": 0.6811758279800415, | |
| "learning_rate": 1.4931383427487741e-05, | |
| "loss": 0.7714, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.134039000367928, | |
| "grad_norm": 0.810197114944458, | |
| "learning_rate": 1.4930901804950918e-05, | |
| "loss": 0.7556, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.13449183482863045, | |
| "grad_norm": 0.656467080116272, | |
| "learning_rate": 1.4930418505878113e-05, | |
| "loss": 0.7308, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.1349446692893329, | |
| "grad_norm": 0.7282403111457825, | |
| "learning_rate": 1.4929933530378367e-05, | |
| "loss": 0.7529, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.13539750375003537, | |
| "grad_norm": 0.7933338284492493, | |
| "learning_rate": 1.4929446878561098e-05, | |
| "loss": 0.8193, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.13585033821073783, | |
| "grad_norm": 0.7075956463813782, | |
| "learning_rate": 1.4928958550536107e-05, | |
| "loss": 0.7736, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.1363031726714403, | |
| "grad_norm": 0.6675400137901306, | |
| "learning_rate": 1.4928468546413563e-05, | |
| "loss": 0.7477, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.13675600713214275, | |
| "grad_norm": 0.8855708241462708, | |
| "learning_rate": 1.4927976866304024e-05, | |
| "loss": 0.7988, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.13720884159284522, | |
| "grad_norm": 0.7367263436317444, | |
| "learning_rate": 1.4927483510318417e-05, | |
| "loss": 0.8018, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.13766167605354768, | |
| "grad_norm": 0.7198050022125244, | |
| "learning_rate": 1.4926988478568055e-05, | |
| "loss": 0.8311, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.13811451051425014, | |
| "grad_norm": 0.6773768067359924, | |
| "learning_rate": 1.4926491771164623e-05, | |
| "loss": 0.8388, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.1385673449749526, | |
| "grad_norm": 0.7650995850563049, | |
| "learning_rate": 1.4925993388220189e-05, | |
| "loss": 0.7779, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.13902017943565506, | |
| "grad_norm": 0.7893260717391968, | |
| "learning_rate": 1.4925493329847195e-05, | |
| "loss": 0.7559, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.13947301389635752, | |
| "grad_norm": 0.7186692357063293, | |
| "learning_rate": 1.4924991596158462e-05, | |
| "loss": 0.8107, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.13992584835705998, | |
| "grad_norm": 0.7286010980606079, | |
| "learning_rate": 1.492448818726719e-05, | |
| "loss": 0.6999, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.14037868281776242, | |
| "grad_norm": 0.7160384058952332, | |
| "learning_rate": 1.4923983103286957e-05, | |
| "loss": 0.7403, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.14083151727846488, | |
| "grad_norm": 0.858772337436676, | |
| "learning_rate": 1.492347634433172e-05, | |
| "loss": 0.7424, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.14128435173916734, | |
| "grad_norm": 0.7329276204109192, | |
| "learning_rate": 1.4922967910515809e-05, | |
| "loss": 0.7232, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.1417371861998698, | |
| "grad_norm": 0.8894422054290771, | |
| "learning_rate": 1.4922457801953934e-05, | |
| "loss": 0.7737, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.14219002066057226, | |
| "grad_norm": 0.7823007106781006, | |
| "learning_rate": 1.492194601876119e-05, | |
| "loss": 0.7835, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.14264285512127473, | |
| "grad_norm": 0.8172221779823303, | |
| "learning_rate": 1.492143256105304e-05, | |
| "loss": 0.7721, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.1430956895819772, | |
| "grad_norm": 0.9521918892860413, | |
| "learning_rate": 1.492091742894533e-05, | |
| "loss": 0.7202, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.14354852404267965, | |
| "grad_norm": 0.6912105679512024, | |
| "learning_rate": 1.492040062255428e-05, | |
| "loss": 0.7555, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.1440013585033821, | |
| "grad_norm": 0.713433563709259, | |
| "learning_rate": 1.4919882141996493e-05, | |
| "loss": 0.7422, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.14445419296408457, | |
| "grad_norm": 0.6809953451156616, | |
| "learning_rate": 1.4919361987388942e-05, | |
| "loss": 0.7298, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.14490702742478703, | |
| "grad_norm": 0.8265045285224915, | |
| "learning_rate": 1.491884015884899e-05, | |
| "loss": 0.8056, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.1453598618854895, | |
| "grad_norm": 0.7027236819267273, | |
| "learning_rate": 1.491831665649437e-05, | |
| "loss": 0.7647, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.14581269634619196, | |
| "grad_norm": 0.6985670328140259, | |
| "learning_rate": 1.4917791480443183e-05, | |
| "loss": 0.7859, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.14626553080689442, | |
| "grad_norm": 0.6507831811904907, | |
| "learning_rate": 1.4917264630813925e-05, | |
| "loss": 0.7143, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.14671836526759685, | |
| "grad_norm": 0.8120051026344299, | |
| "learning_rate": 1.4916736107725463e-05, | |
| "loss": 0.7825, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.1471711997282993, | |
| "grad_norm": 0.7564501166343689, | |
| "learning_rate": 1.4916205911297039e-05, | |
| "loss": 0.8156, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.14762403418900177, | |
| "grad_norm": 0.7937654852867126, | |
| "learning_rate": 1.4915674041648274e-05, | |
| "loss": 0.7779, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.14807686864970424, | |
| "grad_norm": 0.6390685439109802, | |
| "learning_rate": 1.4915140498899165e-05, | |
| "loss": 0.7297, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.1485297031104067, | |
| "grad_norm": 0.8413376212120056, | |
| "learning_rate": 1.4914605283170092e-05, | |
| "loss": 0.7459, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.14898253757110916, | |
| "grad_norm": 0.8402100801467896, | |
| "learning_rate": 1.4914068394581805e-05, | |
| "loss": 0.7285, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.14943537203181162, | |
| "grad_norm": 0.8007932901382446, | |
| "learning_rate": 1.4913529833255436e-05, | |
| "loss": 0.8058, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.14988820649251408, | |
| "grad_norm": 0.913092851638794, | |
| "learning_rate": 1.4912989599312496e-05, | |
| "loss": 0.7506, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.15034104095321654, | |
| "grad_norm": 0.9280517101287842, | |
| "learning_rate": 1.4912447692874865e-05, | |
| "loss": 0.8402, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.150793875413919, | |
| "grad_norm": 0.8286890387535095, | |
| "learning_rate": 1.4911904114064816e-05, | |
| "loss": 0.6831, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.15124670987462147, | |
| "grad_norm": 1.3162356615066528, | |
| "learning_rate": 1.491135886300498e-05, | |
| "loss": 0.7654, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.15169954433532393, | |
| "grad_norm": 0.6534262895584106, | |
| "learning_rate": 1.4910811939818376e-05, | |
| "loss": 0.7818, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.1521523787960264, | |
| "grad_norm": 1.4352598190307617, | |
| "learning_rate": 1.4910263344628405e-05, | |
| "loss": 0.7497, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.15260521325672885, | |
| "grad_norm": 0.7645979523658752, | |
| "learning_rate": 1.4909713077558834e-05, | |
| "loss": 0.7298, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.15305804771743128, | |
| "grad_norm": 1.1644870042800903, | |
| "learning_rate": 1.4909161138733815e-05, | |
| "loss": 0.7725, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.15351088217813375, | |
| "grad_norm": 0.7654008865356445, | |
| "learning_rate": 1.4908607528277873e-05, | |
| "loss": 0.6841, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.1539637166388362, | |
| "grad_norm": 0.9176437258720398, | |
| "learning_rate": 1.4908052246315915e-05, | |
| "loss": 0.7965, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.15441655109953867, | |
| "grad_norm": 0.8297829031944275, | |
| "learning_rate": 1.4907495292973215e-05, | |
| "loss": 0.7624, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.15486938556024113, | |
| "grad_norm": 0.7733472585678101, | |
| "learning_rate": 1.490693666837544e-05, | |
| "loss": 0.7679, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.1553222200209436, | |
| "grad_norm": 0.8300292491912842, | |
| "learning_rate": 1.4906376372648618e-05, | |
| "loss": 0.7582, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.15577505448164605, | |
| "grad_norm": 0.7907745838165283, | |
| "learning_rate": 1.4905814405919163e-05, | |
| "loss": 0.6966, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.15622788894234851, | |
| "grad_norm": 1.0779638290405273, | |
| "learning_rate": 1.4905250768313865e-05, | |
| "loss": 0.698, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.15668072340305098, | |
| "grad_norm": 0.7394607663154602, | |
| "learning_rate": 1.490468545995989e-05, | |
| "loss": 0.7973, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.15713355786375344, | |
| "grad_norm": 0.9171550869941711, | |
| "learning_rate": 1.4904118480984782e-05, | |
| "loss": 0.7185, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.1575863923244559, | |
| "grad_norm": 0.7766843438148499, | |
| "learning_rate": 1.4903549831516459e-05, | |
| "loss": 0.692, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.15803922678515836, | |
| "grad_norm": 0.8801367282867432, | |
| "learning_rate": 1.490297951168322e-05, | |
| "loss": 0.7797, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.15849206124586082, | |
| "grad_norm": 0.8276110887527466, | |
| "learning_rate": 1.4902407521613735e-05, | |
| "loss": 0.7405, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.15894489570656328, | |
| "grad_norm": 0.8306746482849121, | |
| "learning_rate": 1.4901833861437058e-05, | |
| "loss": 0.7268, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.15939773016726572, | |
| "grad_norm": 0.9609989523887634, | |
| "learning_rate": 1.4901258531282616e-05, | |
| "loss": 0.7233, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.15985056462796818, | |
| "grad_norm": 0.8433880805969238, | |
| "learning_rate": 1.4900681531280212e-05, | |
| "loss": 0.7076, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.16030339908867064, | |
| "grad_norm": 0.7875558137893677, | |
| "learning_rate": 1.4900102861560027e-05, | |
| "loss": 0.6939, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.1607562335493731, | |
| "grad_norm": 0.7184025645256042, | |
| "learning_rate": 1.4899522522252618e-05, | |
| "loss": 0.7195, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.16120906801007556, | |
| "grad_norm": 0.7774180173873901, | |
| "learning_rate": 1.4898940513488921e-05, | |
| "loss": 0.713, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.16166190247077802, | |
| "grad_norm": 0.8321492075920105, | |
| "learning_rate": 1.4898356835400247e-05, | |
| "loss": 0.7299, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.16211473693148049, | |
| "grad_norm": 0.8344357013702393, | |
| "learning_rate": 1.4897771488118281e-05, | |
| "loss": 0.723, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.16256757139218295, | |
| "grad_norm": 0.8810961842536926, | |
| "learning_rate": 1.489718447177509e-05, | |
| "loss": 0.723, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.1630204058528854, | |
| "grad_norm": 0.6815357804298401, | |
| "learning_rate": 1.4896595786503114e-05, | |
| "loss": 0.7178, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.16347324031358787, | |
| "grad_norm": 0.7081362009048462, | |
| "learning_rate": 1.4896005432435168e-05, | |
| "loss": 0.6913, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.16392607477429033, | |
| "grad_norm": 0.712780237197876, | |
| "learning_rate": 1.4895413409704448e-05, | |
| "loss": 0.7063, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.1643789092349928, | |
| "grad_norm": 0.7857349514961243, | |
| "learning_rate": 1.4894819718444525e-05, | |
| "loss": 0.6999, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.16483174369569525, | |
| "grad_norm": 0.7282727360725403, | |
| "learning_rate": 1.4894224358789344e-05, | |
| "loss": 0.7216, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.16528457815639772, | |
| "grad_norm": 0.7465353012084961, | |
| "learning_rate": 1.4893627330873227e-05, | |
| "loss": 0.7342, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.16573741261710015, | |
| "grad_norm": 0.8547371029853821, | |
| "learning_rate": 1.4893028634830877e-05, | |
| "loss": 0.6788, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.1661902470778026, | |
| "grad_norm": 0.7159935832023621, | |
| "learning_rate": 1.4892428270797368e-05, | |
| "loss": 0.6689, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.16664308153850507, | |
| "grad_norm": 0.8785684108734131, | |
| "learning_rate": 1.489182623890815e-05, | |
| "loss": 0.7402, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.16709591599920753, | |
| "grad_norm": 0.8778204917907715, | |
| "learning_rate": 1.4891222539299058e-05, | |
| "loss": 0.6354, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.16754875045991, | |
| "grad_norm": 0.8214927315711975, | |
| "learning_rate": 1.4890617172106286e-05, | |
| "loss": 0.7109, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.16800158492061246, | |
| "grad_norm": 0.7724870443344116, | |
| "learning_rate": 1.4890010137466428e-05, | |
| "loss": 0.7082, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.16845441938131492, | |
| "grad_norm": 0.7708100080490112, | |
| "learning_rate": 1.4889401435516431e-05, | |
| "loss": 0.6824, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.16890725384201738, | |
| "grad_norm": 0.6783519387245178, | |
| "learning_rate": 1.4888791066393632e-05, | |
| "loss": 0.733, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.16936008830271984, | |
| "grad_norm": 0.7518247961997986, | |
| "learning_rate": 1.488817903023574e-05, | |
| "loss": 0.7549, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.1698129227634223, | |
| "grad_norm": 0.7920773029327393, | |
| "learning_rate": 1.4887565327180842e-05, | |
| "loss": 0.7505, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.17026575722412476, | |
| "grad_norm": 0.7498000860214233, | |
| "learning_rate": 1.4886949957367398e-05, | |
| "loss": 0.738, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.17071859168482723, | |
| "grad_norm": 0.682931661605835, | |
| "learning_rate": 1.4886332920934247e-05, | |
| "loss": 0.6858, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.1711714261455297, | |
| "grad_norm": 0.705777108669281, | |
| "learning_rate": 1.4885714218020604e-05, | |
| "loss": 0.7319, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.17162426060623212, | |
| "grad_norm": 0.7821267247200012, | |
| "learning_rate": 1.4885093848766055e-05, | |
| "loss": 0.6815, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.17207709506693458, | |
| "grad_norm": 0.7366465330123901, | |
| "learning_rate": 1.4884471813310567e-05, | |
| "loss": 0.7158, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.17252992952763704, | |
| "grad_norm": 0.7276479601860046, | |
| "learning_rate": 1.4883848111794484e-05, | |
| "loss": 0.6813, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.1729827639883395, | |
| "grad_norm": 0.6883667707443237, | |
| "learning_rate": 1.488322274435852e-05, | |
| "loss": 0.6472, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.17343559844904197, | |
| "grad_norm": 0.6681039333343506, | |
| "learning_rate": 1.4882595711143772e-05, | |
| "loss": 0.6808, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.17388843290974443, | |
| "grad_norm": 0.9255726933479309, | |
| "learning_rate": 1.4881967012291707e-05, | |
| "loss": 0.6653, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.1743412673704469, | |
| "grad_norm": 0.7656347155570984, | |
| "learning_rate": 1.488133664794417e-05, | |
| "loss": 0.6502, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.17479410183114935, | |
| "grad_norm": 0.6879505515098572, | |
| "learning_rate": 1.4880704618243382e-05, | |
| "loss": 0.6785, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.1752469362918518, | |
| "grad_norm": 0.7322476506233215, | |
| "learning_rate": 1.4880070923331942e-05, | |
| "loss": 0.6697, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.17569977075255427, | |
| "grad_norm": 0.7081751823425293, | |
| "learning_rate": 1.4879435563352815e-05, | |
| "loss": 0.6706, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.17615260521325674, | |
| "grad_norm": 0.9385679364204407, | |
| "learning_rate": 1.4878798538449358e-05, | |
| "loss": 0.6689, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.1766054396739592, | |
| "grad_norm": 1.0547739267349243, | |
| "learning_rate": 1.4878159848765288e-05, | |
| "loss": 0.7344, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.17705827413466166, | |
| "grad_norm": 0.772883951663971, | |
| "learning_rate": 1.4877519494444707e-05, | |
| "loss": 0.7109, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.17751110859536412, | |
| "grad_norm": 0.9401255249977112, | |
| "learning_rate": 1.4876877475632089e-05, | |
| "loss": 0.6846, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.17796394305606655, | |
| "grad_norm": 0.6944053769111633, | |
| "learning_rate": 1.4876233792472284e-05, | |
| "loss": 0.6735, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.17841677751676902, | |
| "grad_norm": 1.1043004989624023, | |
| "learning_rate": 1.4875588445110517e-05, | |
| "loss": 0.6569, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.17886961197747148, | |
| "grad_norm": 0.6451114416122437, | |
| "learning_rate": 1.4874941433692393e-05, | |
| "loss": 0.6748, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.17932244643817394, | |
| "grad_norm": 0.8441924452781677, | |
| "learning_rate": 1.4874292758363882e-05, | |
| "loss": 0.729, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.1797752808988764, | |
| "grad_norm": 0.7901274561882019, | |
| "learning_rate": 1.4873642419271342e-05, | |
| "loss": 0.7124, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.18022811535957886, | |
| "grad_norm": 1.0363836288452148, | |
| "learning_rate": 1.4872990416561499e-05, | |
| "loss": 0.6816, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.18068094982028132, | |
| "grad_norm": 0.7908300161361694, | |
| "learning_rate": 1.4872336750381452e-05, | |
| "loss": 0.6721, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.18113378428098378, | |
| "grad_norm": 0.8819090127944946, | |
| "learning_rate": 1.4871681420878683e-05, | |
| "loss": 0.6521, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.18158661874168625, | |
| "grad_norm": 1.1245052814483643, | |
| "learning_rate": 1.4871024428201043e-05, | |
| "loss": 0.7597, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.1820394532023887, | |
| "grad_norm": 0.962951123714447, | |
| "learning_rate": 1.4870365772496764e-05, | |
| "loss": 0.7093, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.18249228766309117, | |
| "grad_norm": 1.236630916595459, | |
| "learning_rate": 1.4869705453914446e-05, | |
| "loss": 0.7013, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.18294512212379363, | |
| "grad_norm": 0.8700730204582214, | |
| "learning_rate": 1.486904347260307e-05, | |
| "loss": 0.6539, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.1833979565844961, | |
| "grad_norm": 1.0793607234954834, | |
| "learning_rate": 1.4868379828711991e-05, | |
| "loss": 0.6784, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.18385079104519855, | |
| "grad_norm": 0.9725819826126099, | |
| "learning_rate": 1.4867714522390934e-05, | |
| "loss": 0.7212, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.184303625505901, | |
| "grad_norm": 0.9909386038780212, | |
| "learning_rate": 1.4867047553790007e-05, | |
| "loss": 0.67, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.18475645996660345, | |
| "grad_norm": 0.993817925453186, | |
| "learning_rate": 1.4866378923059687e-05, | |
| "loss": 0.6186, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.1852092944273059, | |
| "grad_norm": 1.0599784851074219, | |
| "learning_rate": 1.4865708630350834e-05, | |
| "loss": 0.674, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.18566212888800837, | |
| "grad_norm": 1.0941828489303589, | |
| "learning_rate": 1.4865036675814669e-05, | |
| "loss": 0.6587, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.18611496334871083, | |
| "grad_norm": 3.841005802154541, | |
| "learning_rate": 1.48643630596028e-05, | |
| "loss": 0.729, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.1865677978094133, | |
| "grad_norm": 1.5443971157073975, | |
| "learning_rate": 1.4863687781867209e-05, | |
| "loss": 0.6768, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.18702063227011576, | |
| "grad_norm": 0.9371129870414734, | |
| "learning_rate": 1.4863010842760246e-05, | |
| "loss": 0.6169, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.18747346673081822, | |
| "grad_norm": 1.437848448753357, | |
| "learning_rate": 1.4862332242434639e-05, | |
| "loss": 0.7218, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.18792630119152068, | |
| "grad_norm": 1.4850507974624634, | |
| "learning_rate": 1.4861651981043495e-05, | |
| "loss": 0.6684, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.18837913565222314, | |
| "grad_norm": 0.7159014940261841, | |
| "learning_rate": 1.486097005874029e-05, | |
| "loss": 0.6782, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.1888319701129256, | |
| "grad_norm": 1.2282770872116089, | |
| "learning_rate": 1.4860286475678877e-05, | |
| "loss": 0.7224, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.18928480457362806, | |
| "grad_norm": 1.0234930515289307, | |
| "learning_rate": 1.4859601232013488e-05, | |
| "loss": 0.693, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.18973763903433052, | |
| "grad_norm": 0.8315566778182983, | |
| "learning_rate": 1.485891432789872e-05, | |
| "loss": 0.6348, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.19019047349503299, | |
| "grad_norm": 1.353355884552002, | |
| "learning_rate": 1.4858225763489552e-05, | |
| "loss": 0.7412, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.19064330795573542, | |
| "grad_norm": 0.8341365456581116, | |
| "learning_rate": 1.4857535538941339e-05, | |
| "loss": 0.6768, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.19109614241643788, | |
| "grad_norm": 0.8955153226852417, | |
| "learning_rate": 1.4856843654409802e-05, | |
| "loss": 0.65, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.19154897687714034, | |
| "grad_norm": 0.8571435809135437, | |
| "learning_rate": 1.4856150110051044e-05, | |
| "loss": 0.6437, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.1920018113378428, | |
| "grad_norm": 0.76616370677948, | |
| "learning_rate": 1.485545490602154e-05, | |
| "loss": 0.6599, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.19245464579854527, | |
| "grad_norm": 0.7934353947639465, | |
| "learning_rate": 1.485475804247814e-05, | |
| "loss": 0.6271, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.19290748025924773, | |
| "grad_norm": 0.8939810395240784, | |
| "learning_rate": 1.4854059519578069e-05, | |
| "loss": 0.659, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.1933603147199502, | |
| "grad_norm": 0.6890648007392883, | |
| "learning_rate": 1.4853359337478923e-05, | |
| "loss": 0.7456, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.19381314918065265, | |
| "grad_norm": 0.8249607682228088, | |
| "learning_rate": 1.4852657496338678e-05, | |
| "loss": 0.6701, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.1942659836413551, | |
| "grad_norm": 0.8182385563850403, | |
| "learning_rate": 1.4851953996315678e-05, | |
| "loss": 0.6081, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.19471881810205757, | |
| "grad_norm": 0.7631829977035522, | |
| "learning_rate": 1.4851248837568646e-05, | |
| "loss": 0.5855, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.19517165256276003, | |
| "grad_norm": 0.9246631860733032, | |
| "learning_rate": 1.4850542020256677e-05, | |
| "loss": 0.6469, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.1956244870234625, | |
| "grad_norm": 0.7482291460037231, | |
| "learning_rate": 1.4849833544539242e-05, | |
| "loss": 0.6801, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.19607732148416496, | |
| "grad_norm": 0.8039886951446533, | |
| "learning_rate": 1.4849123410576183e-05, | |
| "loss": 0.71, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.1965301559448674, | |
| "grad_norm": 0.7532901763916016, | |
| "learning_rate": 1.484841161852772e-05, | |
| "loss": 0.6619, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.19698299040556985, | |
| "grad_norm": 0.7386736273765564, | |
| "learning_rate": 1.4847698168554447e-05, | |
| "loss": 0.69, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.1974358248662723, | |
| "grad_norm": 0.6776562929153442, | |
| "learning_rate": 1.4846983060817324e-05, | |
| "loss": 0.6575, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.19788865932697478, | |
| "grad_norm": 0.7260106205940247, | |
| "learning_rate": 1.4846266295477698e-05, | |
| "loss": 0.6311, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.19834149378767724, | |
| "grad_norm": 0.7524548172950745, | |
| "learning_rate": 1.4845547872697279e-05, | |
| "loss": 0.6287, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.1987943282483797, | |
| "grad_norm": 0.8071796298027039, | |
| "learning_rate": 1.4844827792638158e-05, | |
| "loss": 0.6258, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.19924716270908216, | |
| "grad_norm": 0.8051711320877075, | |
| "learning_rate": 1.4844106055462793e-05, | |
| "loss": 0.6966, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.19969999716978462, | |
| "grad_norm": 0.8135205507278442, | |
| "learning_rate": 1.4843382661334025e-05, | |
| "loss": 0.6637, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.20015283163048708, | |
| "grad_norm": 0.8501531481742859, | |
| "learning_rate": 1.4842657610415061e-05, | |
| "loss": 0.6969, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.20060566609118954, | |
| "grad_norm": 0.7646431922912598, | |
| "learning_rate": 1.4841930902869486e-05, | |
| "loss": 0.6786, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.201058500551892, | |
| "grad_norm": 0.7486881613731384, | |
| "learning_rate": 1.4841202538861255e-05, | |
| "loss": 0.6009, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.20151133501259447, | |
| "grad_norm": 0.7672154307365417, | |
| "learning_rate": 1.4840472518554702e-05, | |
| "loss": 0.6538, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.20196416947329693, | |
| "grad_norm": 0.7490094900131226, | |
| "learning_rate": 1.4839740842114529e-05, | |
| "loss": 0.624, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.2024170039339994, | |
| "grad_norm": 0.6690830588340759, | |
| "learning_rate": 1.4839007509705819e-05, | |
| "loss": 0.6335, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.20286983839470182, | |
| "grad_norm": 0.7209163308143616, | |
| "learning_rate": 1.483827252149402e-05, | |
| "loss": 0.596, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.20332267285540429, | |
| "grad_norm": 0.7429115772247314, | |
| "learning_rate": 1.4837535877644957e-05, | |
| "loss": 0.6344, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.20377550731610675, | |
| "grad_norm": 0.7298805713653564, | |
| "learning_rate": 1.4836797578324833e-05, | |
| "loss": 0.6197, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.2042283417768092, | |
| "grad_norm": 0.750446081161499, | |
| "learning_rate": 1.4836057623700218e-05, | |
| "loss": 0.5782, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.20468117623751167, | |
| "grad_norm": 0.811336100101471, | |
| "learning_rate": 1.483531601393806e-05, | |
| "loss": 0.6411, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.20513401069821413, | |
| "grad_norm": 0.7158386707305908, | |
| "learning_rate": 1.4834572749205675e-05, | |
| "loss": 0.6706, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.2055868451589166, | |
| "grad_norm": 0.6980316042900085, | |
| "learning_rate": 1.483382782967076e-05, | |
| "loss": 0.607, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.20603967961961905, | |
| "grad_norm": 0.7448428273200989, | |
| "learning_rate": 1.483308125550138e-05, | |
| "loss": 0.666, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.20649251408032152, | |
| "grad_norm": 0.6907296776771545, | |
| "learning_rate": 1.4832333026865973e-05, | |
| "loss": 0.6145, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.20694534854102398, | |
| "grad_norm": 0.7163165211677551, | |
| "learning_rate": 1.4831583143933353e-05, | |
| "loss": 0.6306, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 0.20739818300172644, | |
| "grad_norm": 0.6344444155693054, | |
| "learning_rate": 1.483083160687271e-05, | |
| "loss": 0.619, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.2078510174624289, | |
| "grad_norm": 0.6727322936058044, | |
| "learning_rate": 1.4830078415853596e-05, | |
| "loss": 0.6618, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.20830385192313136, | |
| "grad_norm": 0.7381990551948547, | |
| "learning_rate": 1.482932357104595e-05, | |
| "loss": 0.5994, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.20875668638383382, | |
| "grad_norm": 0.7138816118240356, | |
| "learning_rate": 1.4828567072620074e-05, | |
| "loss": 0.6549, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 0.20920952084453626, | |
| "grad_norm": 0.708865761756897, | |
| "learning_rate": 1.4827808920746646e-05, | |
| "loss": 0.6603, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.20966235530523872, | |
| "grad_norm": 0.7122054696083069, | |
| "learning_rate": 1.4827049115596722e-05, | |
| "loss": 0.6423, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 0.21011518976594118, | |
| "grad_norm": 0.7929750680923462, | |
| "learning_rate": 1.4826287657341723e-05, | |
| "loss": 0.6772, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.21056802422664364, | |
| "grad_norm": 0.7480528354644775, | |
| "learning_rate": 1.4825524546153449e-05, | |
| "loss": 0.6118, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.2110208586873461, | |
| "grad_norm": 0.8834244608879089, | |
| "learning_rate": 1.4824759782204068e-05, | |
| "loss": 0.622, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 0.21147369314804856, | |
| "grad_norm": 0.6908849477767944, | |
| "learning_rate": 1.4823993365666127e-05, | |
| "loss": 0.6625, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 0.21192652760875103, | |
| "grad_norm": 0.7661241888999939, | |
| "learning_rate": 1.4823225296712539e-05, | |
| "loss": 0.6409, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.2123793620694535, | |
| "grad_norm": 0.6954145431518555, | |
| "learning_rate": 1.4822455575516595e-05, | |
| "loss": 0.5925, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 0.21283219653015595, | |
| "grad_norm": 0.6936039924621582, | |
| "learning_rate": 1.4821684202251958e-05, | |
| "loss": 0.6555, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.2132850309908584, | |
| "grad_norm": 0.8155099749565125, | |
| "learning_rate": 1.482091117709266e-05, | |
| "loss": 0.6868, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 0.21373786545156087, | |
| "grad_norm": 0.6670776605606079, | |
| "learning_rate": 1.4820136500213112e-05, | |
| "loss": 0.602, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 0.21419069991226333, | |
| "grad_norm": 0.6749069690704346, | |
| "learning_rate": 1.4819360171788092e-05, | |
| "loss": 0.5757, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 0.2146435343729658, | |
| "grad_norm": 0.8901804685592651, | |
| "learning_rate": 1.4818582191992752e-05, | |
| "loss": 0.6199, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 0.21509636883366826, | |
| "grad_norm": 1.0478390455245972, | |
| "learning_rate": 1.4817802561002619e-05, | |
| "loss": 0.5611, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.2155492032943707, | |
| "grad_norm": 0.679456353187561, | |
| "learning_rate": 1.4817021278993588e-05, | |
| "loss": 0.5717, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 0.21600203775507315, | |
| "grad_norm": 1.1234318017959595, | |
| "learning_rate": 1.4816238346141934e-05, | |
| "loss": 0.6477, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 0.2164548722157756, | |
| "grad_norm": 0.7910907864570618, | |
| "learning_rate": 1.4815453762624295e-05, | |
| "loss": 0.5973, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 0.21690770667647807, | |
| "grad_norm": 0.8617794513702393, | |
| "learning_rate": 1.481466752861769e-05, | |
| "loss": 0.5653, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 0.21736054113718054, | |
| "grad_norm": 0.7190772294998169, | |
| "learning_rate": 1.4813879644299502e-05, | |
| "loss": 0.6063, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.217813375597883, | |
| "grad_norm": 0.9855300188064575, | |
| "learning_rate": 1.4813090109847495e-05, | |
| "loss": 0.6402, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 0.21826621005858546, | |
| "grad_norm": 0.7089081406593323, | |
| "learning_rate": 1.4812298925439799e-05, | |
| "loss": 0.6129, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 0.21871904451928792, | |
| "grad_norm": 0.8527977466583252, | |
| "learning_rate": 1.4811506091254922e-05, | |
| "loss": 0.6116, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 0.21917187897999038, | |
| "grad_norm": 0.8156578540802002, | |
| "learning_rate": 1.4810711607471737e-05, | |
| "loss": 0.6608, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 0.21962471344069284, | |
| "grad_norm": 0.7411708235740662, | |
| "learning_rate": 1.4809915474269493e-05, | |
| "loss": 0.5877, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.2200775479013953, | |
| "grad_norm": 0.9931669235229492, | |
| "learning_rate": 1.4809117691827812e-05, | |
| "loss": 0.6118, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 0.22053038236209777, | |
| "grad_norm": 0.79165118932724, | |
| "learning_rate": 1.4808318260326687e-05, | |
| "loss": 0.5565, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 0.22098321682280023, | |
| "grad_norm": 0.8213533163070679, | |
| "learning_rate": 1.4807517179946483e-05, | |
| "loss": 0.6048, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 0.2214360512835027, | |
| "grad_norm": 0.7628992795944214, | |
| "learning_rate": 1.4806714450867937e-05, | |
| "loss": 0.5837, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 0.22188888574420512, | |
| "grad_norm": 0.7759009003639221, | |
| "learning_rate": 1.4805910073272159e-05, | |
| "loss": 0.6224, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.22234172020490758, | |
| "grad_norm": 0.7115432024002075, | |
| "learning_rate": 1.4805104047340628e-05, | |
| "loss": 0.6427, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 0.22279455466561005, | |
| "grad_norm": 0.756812334060669, | |
| "learning_rate": 1.4804296373255201e-05, | |
| "loss": 0.6536, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 0.2232473891263125, | |
| "grad_norm": 0.7481681108474731, | |
| "learning_rate": 1.48034870511981e-05, | |
| "loss": 0.6257, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 0.22370022358701497, | |
| "grad_norm": 0.7567417621612549, | |
| "learning_rate": 1.4802676081351922e-05, | |
| "loss": 0.5999, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 0.22415305804771743, | |
| "grad_norm": 0.8136939406394958, | |
| "learning_rate": 1.4801863463899634e-05, | |
| "loss": 0.6432, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.2246058925084199, | |
| "grad_norm": 0.7277612686157227, | |
| "learning_rate": 1.4801049199024577e-05, | |
| "loss": 0.6153, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 0.22505872696912235, | |
| "grad_norm": 0.8077271580696106, | |
| "learning_rate": 1.4800233286910462e-05, | |
| "loss": 0.5886, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 0.22551156142982481, | |
| "grad_norm": 0.8224574327468872, | |
| "learning_rate": 1.4799415727741376e-05, | |
| "loss": 0.5642, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 0.22596439589052728, | |
| "grad_norm": 0.7653172612190247, | |
| "learning_rate": 1.479859652170177e-05, | |
| "loss": 0.605, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 0.22641723035122974, | |
| "grad_norm": 0.74990314245224, | |
| "learning_rate": 1.4797775668976473e-05, | |
| "loss": 0.619, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.2268700648119322, | |
| "grad_norm": 0.8035469055175781, | |
| "learning_rate": 1.4796953169750684e-05, | |
| "loss": 0.5863, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 0.22732289927263466, | |
| "grad_norm": 0.7987021207809448, | |
| "learning_rate": 1.4796129024209968e-05, | |
| "loss": 0.5786, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 0.2277757337333371, | |
| "grad_norm": 0.7279565334320068, | |
| "learning_rate": 1.4795303232540272e-05, | |
| "loss": 0.646, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 0.22822856819403955, | |
| "grad_norm": 0.753925621509552, | |
| "learning_rate": 1.4794475794927902e-05, | |
| "loss": 0.6106, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 0.22868140265474202, | |
| "grad_norm": 0.8243687152862549, | |
| "learning_rate": 1.4793646711559549e-05, | |
| "loss": 0.5846, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.22913423711544448, | |
| "grad_norm": 0.7034870982170105, | |
| "learning_rate": 1.4792815982622264e-05, | |
| "loss": 0.6537, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 0.22958707157614694, | |
| "grad_norm": 0.7104423642158508, | |
| "learning_rate": 1.4791983608303472e-05, | |
| "loss": 0.6009, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 0.2300399060368494, | |
| "grad_norm": 0.8007641434669495, | |
| "learning_rate": 1.4791149588790973e-05, | |
| "loss": 0.5634, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 0.23049274049755186, | |
| "grad_norm": 1.0584112405776978, | |
| "learning_rate": 1.4790313924272935e-05, | |
| "loss": 0.5732, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 0.23094557495825432, | |
| "grad_norm": 0.7870271801948547, | |
| "learning_rate": 1.4789476614937899e-05, | |
| "loss": 0.6032, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.23139840941895679, | |
| "grad_norm": 0.8935226202011108, | |
| "learning_rate": 1.4788637660974773e-05, | |
| "loss": 0.5845, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 0.23185124387965925, | |
| "grad_norm": 0.7717947363853455, | |
| "learning_rate": 1.4787797062572842e-05, | |
| "loss": 0.6126, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 0.2323040783403617, | |
| "grad_norm": 0.8335773944854736, | |
| "learning_rate": 1.4786954819921759e-05, | |
| "loss": 0.6342, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 0.23275691280106417, | |
| "grad_norm": 0.7727064490318298, | |
| "learning_rate": 1.4786110933211548e-05, | |
| "loss": 0.5998, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 0.23320974726176663, | |
| "grad_norm": 0.8422814011573792, | |
| "learning_rate": 1.4785265402632602e-05, | |
| "loss": 0.6381, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.2336625817224691, | |
| "grad_norm": 0.8143975138664246, | |
| "learning_rate": 1.4784418228375688e-05, | |
| "loss": 0.5694, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 0.23411541618317153, | |
| "grad_norm": 0.764120876789093, | |
| "learning_rate": 1.4783569410631942e-05, | |
| "loss": 0.566, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 0.234568250643874, | |
| "grad_norm": 0.8712199926376343, | |
| "learning_rate": 1.4782718949592873e-05, | |
| "loss": 0.6068, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 0.23502108510457645, | |
| "grad_norm": 0.8048343062400818, | |
| "learning_rate": 1.4781866845450361e-05, | |
| "loss": 0.5919, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 0.2354739195652789, | |
| "grad_norm": 0.8512641191482544, | |
| "learning_rate": 1.478101309839665e-05, | |
| "loss": 0.6131, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.23592675402598137, | |
| "grad_norm": 0.7001386880874634, | |
| "learning_rate": 1.4780157708624364e-05, | |
| "loss": 0.5971, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 0.23637958848668383, | |
| "grad_norm": 0.7913312911987305, | |
| "learning_rate": 1.477930067632649e-05, | |
| "loss": 0.6156, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 0.2368324229473863, | |
| "grad_norm": 0.7557013034820557, | |
| "learning_rate": 1.4778442001696392e-05, | |
| "loss": 0.5684, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 0.23728525740808876, | |
| "grad_norm": 0.6847025156021118, | |
| "learning_rate": 1.4777581684927797e-05, | |
| "loss": 0.6304, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 0.23773809186879122, | |
| "grad_norm": 0.7465041875839233, | |
| "learning_rate": 1.4776719726214811e-05, | |
| "loss": 0.5996, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.23819092632949368, | |
| "grad_norm": 0.776152491569519, | |
| "learning_rate": 1.4775856125751907e-05, | |
| "loss": 0.6457, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 0.23864376079019614, | |
| "grad_norm": 0.7176225185394287, | |
| "learning_rate": 1.4774990883733924e-05, | |
| "loss": 0.6206, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 0.2390965952508986, | |
| "grad_norm": 0.6723232865333557, | |
| "learning_rate": 1.4774124000356074e-05, | |
| "loss": 0.5649, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 0.23954942971160106, | |
| "grad_norm": 0.7886945605278015, | |
| "learning_rate": 1.4773255475813948e-05, | |
| "loss": 0.6057, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 0.24000226417230353, | |
| "grad_norm": 0.8538393974304199, | |
| "learning_rate": 1.4772385310303492e-05, | |
| "loss": 0.5672, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.24045509863300596, | |
| "grad_norm": 0.8311111927032471, | |
| "learning_rate": 1.4771513504021034e-05, | |
| "loss": 0.6067, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 0.24090793309370842, | |
| "grad_norm": 0.6744326949119568, | |
| "learning_rate": 1.4770640057163265e-05, | |
| "loss": 0.5392, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 0.24136076755441088, | |
| "grad_norm": 0.8035153150558472, | |
| "learning_rate": 1.4769764969927255e-05, | |
| "loss": 0.5968, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 0.24181360201511334, | |
| "grad_norm": 0.9007952213287354, | |
| "learning_rate": 1.4768888242510433e-05, | |
| "loss": 0.5998, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 0.2422664364758158, | |
| "grad_norm": 1.0340895652770996, | |
| "learning_rate": 1.4768009875110603e-05, | |
| "loss": 0.6269, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.24271927093651827, | |
| "grad_norm": 0.7285747528076172, | |
| "learning_rate": 1.4767129867925942e-05, | |
| "loss": 0.6182, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 0.24317210539722073, | |
| "grad_norm": 0.7769334316253662, | |
| "learning_rate": 1.4766248221154996e-05, | |
| "loss": 0.6133, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 0.2436249398579232, | |
| "grad_norm": 0.8429601788520813, | |
| "learning_rate": 1.4765364934996674e-05, | |
| "loss": 0.5653, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 0.24407777431862565, | |
| "grad_norm": 0.7434231638908386, | |
| "learning_rate": 1.4764480009650264e-05, | |
| "loss": 0.5946, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 0.2445306087793281, | |
| "grad_norm": 0.8999215960502625, | |
| "learning_rate": 1.4763593445315422e-05, | |
| "loss": 0.5909, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.24498344324003057, | |
| "grad_norm": 0.749334454536438, | |
| "learning_rate": 1.4762705242192165e-05, | |
| "loss": 0.6052, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 0.24543627770073304, | |
| "grad_norm": 0.8988975882530212, | |
| "learning_rate": 1.4761815400480892e-05, | |
| "loss": 0.5615, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 0.2458891121614355, | |
| "grad_norm": 0.7588968276977539, | |
| "learning_rate": 1.4760923920382364e-05, | |
| "loss": 0.5623, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 0.24634194662213796, | |
| "grad_norm": 0.8823333978652954, | |
| "learning_rate": 1.4760030802097715e-05, | |
| "loss": 0.5761, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 0.2467947810828404, | |
| "grad_norm": 0.6606689095497131, | |
| "learning_rate": 1.4759136045828447e-05, | |
| "loss": 0.5995, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.24724761554354285, | |
| "grad_norm": 0.9664527773857117, | |
| "learning_rate": 1.4758239651776434e-05, | |
| "loss": 0.6049, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 0.24770045000424532, | |
| "grad_norm": 0.6853055357933044, | |
| "learning_rate": 1.4757341620143913e-05, | |
| "loss": 0.6096, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 0.24815328446494778, | |
| "grad_norm": 0.845271110534668, | |
| "learning_rate": 1.4756441951133497e-05, | |
| "loss": 0.607, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 0.24860611892565024, | |
| "grad_norm": 0.7086828947067261, | |
| "learning_rate": 1.4755540644948168e-05, | |
| "loss": 0.5752, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 0.2490589533863527, | |
| "grad_norm": 0.7899136543273926, | |
| "learning_rate": 1.4754637701791273e-05, | |
| "loss": 0.5464, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.24951178784705516, | |
| "grad_norm": 0.7464292645454407, | |
| "learning_rate": 1.4753733121866534e-05, | |
| "loss": 0.5789, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 0.24996462230775762, | |
| "grad_norm": 0.8088657259941101, | |
| "learning_rate": 1.4752826905378039e-05, | |
| "loss": 0.6007, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 0.24996462230775762, | |
| "eval_loss": 0.5881260633468628, | |
| "eval_runtime": 19.4408, | |
| "eval_samples_per_second": 5.864, | |
| "eval_steps_per_second": 0.772, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 0.25041745676846006, | |
| "grad_norm": 0.8311989307403564, | |
| "learning_rate": 1.4751919052530245e-05, | |
| "loss": 0.5806, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 0.25087029122916255, | |
| "grad_norm": 0.9002801179885864, | |
| "learning_rate": 1.4751009563527977e-05, | |
| "loss": 0.5398, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 0.251323125689865, | |
| "grad_norm": 0.8495317697525024, | |
| "learning_rate": 1.4750098438576434e-05, | |
| "loss": 0.6073, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.25177596015056747, | |
| "grad_norm": 0.7978039979934692, | |
| "learning_rate": 1.474918567788118e-05, | |
| "loss": 0.6079, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 0.2522287946112699, | |
| "grad_norm": 0.7676923274993896, | |
| "learning_rate": 1.4748271281648145e-05, | |
| "loss": 0.633, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 0.2526816290719724, | |
| "grad_norm": 0.7194515466690063, | |
| "learning_rate": 1.4747355250083639e-05, | |
| "loss": 0.5659, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 0.2531344635326748, | |
| "grad_norm": 0.8001179099082947, | |
| "learning_rate": 1.474643758339433e-05, | |
| "loss": 0.56, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 0.2535872979933773, | |
| "grad_norm": 0.7714060544967651, | |
| "learning_rate": 1.4745518281787259e-05, | |
| "loss": 0.553, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.25404013245407975, | |
| "grad_norm": 0.8208261132240295, | |
| "learning_rate": 1.474459734546984e-05, | |
| "loss": 0.614, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 0.25449296691478224, | |
| "grad_norm": 0.6942229866981506, | |
| "learning_rate": 1.4743674774649845e-05, | |
| "loss": 0.5483, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 0.25494580137548467, | |
| "grad_norm": 0.7569373250007629, | |
| "learning_rate": 1.4742750569535425e-05, | |
| "loss": 0.615, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 0.25539863583618716, | |
| "grad_norm": 0.7039315700531006, | |
| "learning_rate": 1.4741824730335098e-05, | |
| "loss": 0.5252, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 0.2558514702968896, | |
| "grad_norm": 0.8170211315155029, | |
| "learning_rate": 1.4740897257257745e-05, | |
| "loss": 0.6154, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.256304304757592, | |
| "grad_norm": 0.688449501991272, | |
| "learning_rate": 1.4739968150512622e-05, | |
| "loss": 0.5537, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 0.2567571392182945, | |
| "grad_norm": 0.74798983335495, | |
| "learning_rate": 1.473903741030935e-05, | |
| "loss": 0.5854, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 0.25720997367899695, | |
| "grad_norm": 0.7894384264945984, | |
| "learning_rate": 1.4738105036857921e-05, | |
| "loss": 0.5017, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 0.25766280813969944, | |
| "grad_norm": 0.8515793085098267, | |
| "learning_rate": 1.4737171030368693e-05, | |
| "loss": 0.5644, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 0.2581156426004019, | |
| "grad_norm": 0.9285520315170288, | |
| "learning_rate": 1.4736235391052393e-05, | |
| "loss": 0.5714, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.25856847706110436, | |
| "grad_norm": 0.7274629473686218, | |
| "learning_rate": 1.4735298119120116e-05, | |
| "loss": 0.4978, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 0.2590213115218068, | |
| "grad_norm": 1.1050032377243042, | |
| "learning_rate": 1.473435921478333e-05, | |
| "loss": 0.5982, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 0.2594741459825093, | |
| "grad_norm": 0.8430832028388977, | |
| "learning_rate": 1.4733418678253865e-05, | |
| "loss": 0.5126, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 0.2599269804432117, | |
| "grad_norm": 0.7927145957946777, | |
| "learning_rate": 1.4732476509743919e-05, | |
| "loss": 0.5066, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 0.2603798149039142, | |
| "grad_norm": 0.8087596893310547, | |
| "learning_rate": 1.473153270946607e-05, | |
| "loss": 0.6461, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.26083264936461664, | |
| "grad_norm": 1.0339022874832153, | |
| "learning_rate": 1.4730587277633245e-05, | |
| "loss": 0.6035, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 0.26128548382531913, | |
| "grad_norm": 0.8909124732017517, | |
| "learning_rate": 1.4729640214458754e-05, | |
| "loss": 0.573, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 0.26173831828602157, | |
| "grad_norm": 1.0848407745361328, | |
| "learning_rate": 1.4728691520156269e-05, | |
| "loss": 0.5968, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 0.262191152746724, | |
| "grad_norm": 0.8249455094337463, | |
| "learning_rate": 1.4727741194939832e-05, | |
| "loss": 0.5785, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 0.2626439872074265, | |
| "grad_norm": 1.190679669380188, | |
| "learning_rate": 1.4726789239023856e-05, | |
| "loss": 0.597, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.2630968216681289, | |
| "grad_norm": 0.784512460231781, | |
| "learning_rate": 1.4725835652623112e-05, | |
| "loss": 0.534, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 0.2635496561288314, | |
| "grad_norm": 0.8939950466156006, | |
| "learning_rate": 1.4724880435952747e-05, | |
| "loss": 0.5271, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 0.26400249058953384, | |
| "grad_norm": 0.7243357300758362, | |
| "learning_rate": 1.4723923589228276e-05, | |
| "loss": 0.5108, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 0.26445532505023633, | |
| "grad_norm": 0.7685579657554626, | |
| "learning_rate": 1.4722965112665577e-05, | |
| "loss": 0.5129, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 0.26490815951093877, | |
| "grad_norm": 0.8200727105140686, | |
| "learning_rate": 1.47220050064809e-05, | |
| "loss": 0.5488, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.26536099397164126, | |
| "grad_norm": 0.8088552355766296, | |
| "learning_rate": 1.4721043270890862e-05, | |
| "loss": 0.503, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 0.2658138284323437, | |
| "grad_norm": 0.7941896915435791, | |
| "learning_rate": 1.4720079906112444e-05, | |
| "loss": 0.5661, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 0.2662666628930462, | |
| "grad_norm": 0.7164261937141418, | |
| "learning_rate": 1.4719114912363e-05, | |
| "loss": 0.5767, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 0.2667194973537486, | |
| "grad_norm": 0.92717444896698, | |
| "learning_rate": 1.4718148289860247e-05, | |
| "loss": 0.5572, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 0.2671723318144511, | |
| "grad_norm": 0.7904344797134399, | |
| "learning_rate": 1.471718003882227e-05, | |
| "loss": 0.5009, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.26762516627515354, | |
| "grad_norm": 0.8072219491004944, | |
| "learning_rate": 1.4716210159467527e-05, | |
| "loss": 0.5609, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 0.268078000735856, | |
| "grad_norm": 0.7373447418212891, | |
| "learning_rate": 1.4715238652014834e-05, | |
| "loss": 0.5378, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 0.26853083519655846, | |
| "grad_norm": 0.8908963203430176, | |
| "learning_rate": 1.4714265516683383e-05, | |
| "loss": 0.5076, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 0.2689836696572609, | |
| "grad_norm": 0.7174452543258667, | |
| "learning_rate": 1.4713290753692728e-05, | |
| "loss": 0.5344, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 0.2694365041179634, | |
| "grad_norm": 0.7640471458435059, | |
| "learning_rate": 1.471231436326279e-05, | |
| "loss": 0.5525, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.2698893385786658, | |
| "grad_norm": 0.8122929334640503, | |
| "learning_rate": 1.4711336345613864e-05, | |
| "loss": 0.5774, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 0.2703421730393683, | |
| "grad_norm": 0.740322470664978, | |
| "learning_rate": 1.4710356700966603e-05, | |
| "loss": 0.5363, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 0.27079500750007074, | |
| "grad_norm": 0.7563768625259399, | |
| "learning_rate": 1.4709375429542034e-05, | |
| "loss": 0.5507, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 0.27124784196077323, | |
| "grad_norm": 0.6652012467384338, | |
| "learning_rate": 1.4708392531561544e-05, | |
| "loss": 0.538, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 0.27170067642147566, | |
| "grad_norm": 0.785305917263031, | |
| "learning_rate": 1.4707408007246898e-05, | |
| "loss": 0.6097, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.27215351088217815, | |
| "grad_norm": 0.7112635970115662, | |
| "learning_rate": 1.4706421856820214e-05, | |
| "loss": 0.5993, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 0.2726063453428806, | |
| "grad_norm": 0.7882818579673767, | |
| "learning_rate": 1.4705434080503988e-05, | |
| "loss": 0.5176, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 0.2730591798035831, | |
| "grad_norm": 1.0064949989318848, | |
| "learning_rate": 1.470444467852108e-05, | |
| "loss": 0.5876, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 0.2735120142642855, | |
| "grad_norm": 0.8223081827163696, | |
| "learning_rate": 1.4703453651094713e-05, | |
| "loss": 0.5901, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 0.273964848724988, | |
| "grad_norm": 0.9124444723129272, | |
| "learning_rate": 1.4702460998448482e-05, | |
| "loss": 0.5698, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.27441768318569043, | |
| "grad_norm": 0.8722511529922485, | |
| "learning_rate": 1.4701466720806342e-05, | |
| "loss": 0.6134, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 0.27487051764639286, | |
| "grad_norm": 0.6496700048446655, | |
| "learning_rate": 1.4700470818392621e-05, | |
| "loss": 0.5107, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 0.27532335210709535, | |
| "grad_norm": 0.696812629699707, | |
| "learning_rate": 1.4699473291432012e-05, | |
| "loss": 0.5341, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 0.2757761865677978, | |
| "grad_norm": 0.7689089775085449, | |
| "learning_rate": 1.4698474140149576e-05, | |
| "loss": 0.5457, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 0.2762290210285003, | |
| "grad_norm": 0.7031781673431396, | |
| "learning_rate": 1.4697473364770731e-05, | |
| "loss": 0.5142, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.2766818554892027, | |
| "grad_norm": 0.8002460598945618, | |
| "learning_rate": 1.4696470965521276e-05, | |
| "loss": 0.585, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 0.2771346899499052, | |
| "grad_norm": 0.8532388210296631, | |
| "learning_rate": 1.4695466942627365e-05, | |
| "loss": 0.5656, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 0.27758752441060763, | |
| "grad_norm": 0.7644308805465698, | |
| "learning_rate": 1.4694461296315525e-05, | |
| "loss": 0.5082, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 0.2780403588713101, | |
| "grad_norm": 0.8224950432777405, | |
| "learning_rate": 1.4693454026812643e-05, | |
| "loss": 0.5058, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 0.27849319333201256, | |
| "grad_norm": 0.6774716973304749, | |
| "learning_rate": 1.4692445134345981e-05, | |
| "loss": 0.4961, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.27894602779271505, | |
| "grad_norm": 0.8179674744606018, | |
| "learning_rate": 1.4691434619143159e-05, | |
| "loss": 0.4869, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 0.2793988622534175, | |
| "grad_norm": 0.6929678320884705, | |
| "learning_rate": 1.4690422481432164e-05, | |
| "loss": 0.5274, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 0.27985169671411997, | |
| "grad_norm": 0.7661529779434204, | |
| "learning_rate": 1.4689408721441357e-05, | |
| "loss": 0.4941, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 0.2803045311748224, | |
| "grad_norm": 0.7640644907951355, | |
| "learning_rate": 1.4688393339399453e-05, | |
| "loss": 0.5352, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 0.28075736563552484, | |
| "grad_norm": 1.1076126098632812, | |
| "learning_rate": 1.4687376335535543e-05, | |
| "loss": 0.5875, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.2812102000962273, | |
| "grad_norm": 0.7371631860733032, | |
| "learning_rate": 1.468635771007908e-05, | |
| "loss": 0.524, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 0.28166303455692976, | |
| "grad_norm": 0.7571913003921509, | |
| "learning_rate": 1.468533746325988e-05, | |
| "loss": 0.5778, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 0.28211586901763225, | |
| "grad_norm": 0.7033771276473999, | |
| "learning_rate": 1.4684315595308132e-05, | |
| "loss": 0.5334, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 0.2825687034783347, | |
| "grad_norm": 1.0408718585968018, | |
| "learning_rate": 1.4683292106454383e-05, | |
| "loss": 0.5654, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 0.28302153793903717, | |
| "grad_norm": 0.7843034267425537, | |
| "learning_rate": 1.4682266996929551e-05, | |
| "loss": 0.5262, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.2834743723997396, | |
| "grad_norm": 0.8728121519088745, | |
| "learning_rate": 1.4681240266964919e-05, | |
| "loss": 0.51, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 0.2839272068604421, | |
| "grad_norm": 0.640317976474762, | |
| "learning_rate": 1.4680211916792133e-05, | |
| "loss": 0.5431, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 0.2843800413211445, | |
| "grad_norm": 0.9380479454994202, | |
| "learning_rate": 1.4679181946643205e-05, | |
| "loss": 0.539, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 0.284832875781847, | |
| "grad_norm": 0.7024577856063843, | |
| "learning_rate": 1.4678150356750514e-05, | |
| "loss": 0.4875, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 0.28528571024254945, | |
| "grad_norm": 0.7149611711502075, | |
| "learning_rate": 1.4677117147346806e-05, | |
| "loss": 0.4813, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.28573854470325194, | |
| "grad_norm": 0.6758112907409668, | |
| "learning_rate": 1.4676082318665188e-05, | |
| "loss": 0.5813, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 0.2861913791639544, | |
| "grad_norm": 0.733073890209198, | |
| "learning_rate": 1.4675045870939137e-05, | |
| "loss": 0.5057, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 0.28664421362465686, | |
| "grad_norm": 0.9430336356163025, | |
| "learning_rate": 1.467400780440249e-05, | |
| "loss": 0.5309, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 0.2870970480853593, | |
| "grad_norm": 0.779950737953186, | |
| "learning_rate": 1.4672968119289455e-05, | |
| "loss": 0.5352, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 0.28754988254606173, | |
| "grad_norm": 0.804973840713501, | |
| "learning_rate": 1.4671926815834602e-05, | |
| "loss": 0.5636, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.2880027170067642, | |
| "grad_norm": 0.6549612283706665, | |
| "learning_rate": 1.4670883894272865e-05, | |
| "loss": 0.5356, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 0.28845555146746665, | |
| "grad_norm": 0.7596065998077393, | |
| "learning_rate": 1.4669839354839545e-05, | |
| "loss": 0.5719, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 0.28890838592816914, | |
| "grad_norm": 0.8659469485282898, | |
| "learning_rate": 1.4668793197770308e-05, | |
| "loss": 0.5028, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 0.2893612203888716, | |
| "grad_norm": 0.8582921028137207, | |
| "learning_rate": 1.4667745423301185e-05, | |
| "loss": 0.509, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 0.28981405484957407, | |
| "grad_norm": 0.751609206199646, | |
| "learning_rate": 1.466669603166857e-05, | |
| "loss": 0.5419, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.2902668893102765, | |
| "grad_norm": 1.0048867464065552, | |
| "learning_rate": 1.4665645023109226e-05, | |
| "loss": 0.5745, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 0.290719723770979, | |
| "grad_norm": 0.7976627945899963, | |
| "learning_rate": 1.4664592397860274e-05, | |
| "loss": 0.4956, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 0.2911725582316814, | |
| "grad_norm": 0.7978916168212891, | |
| "learning_rate": 1.4663538156159208e-05, | |
| "loss": 0.5252, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 0.2916253926923839, | |
| "grad_norm": 0.824474573135376, | |
| "learning_rate": 1.466248229824388e-05, | |
| "loss": 0.5338, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 0.29207822715308634, | |
| "grad_norm": 0.7580894827842712, | |
| "learning_rate": 1.4661424824352512e-05, | |
| "loss": 0.5074, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.29253106161378883, | |
| "grad_norm": 0.7472822070121765, | |
| "learning_rate": 1.4660365734723684e-05, | |
| "loss": 0.4755, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 0.29298389607449127, | |
| "grad_norm": 0.7068562507629395, | |
| "learning_rate": 1.465930502959635e-05, | |
| "loss": 0.5129, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 0.2934367305351937, | |
| "grad_norm": 1.2026171684265137, | |
| "learning_rate": 1.4658242709209815e-05, | |
| "loss": 0.538, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 0.2938895649958962, | |
| "grad_norm": 0.7542515397071838, | |
| "learning_rate": 1.4657178773803761e-05, | |
| "loss": 0.5344, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 0.2943423994565986, | |
| "grad_norm": 1.083905816078186, | |
| "learning_rate": 1.4656113223618232e-05, | |
| "loss": 0.5295, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.2947952339173011, | |
| "grad_norm": 0.7265732884407043, | |
| "learning_rate": 1.4655046058893629e-05, | |
| "loss": 0.5152, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 0.29524806837800355, | |
| "grad_norm": 0.8517547249794006, | |
| "learning_rate": 1.4653977279870725e-05, | |
| "loss": 0.5316, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 0.29570090283870604, | |
| "grad_norm": 0.7927708625793457, | |
| "learning_rate": 1.4652906886790654e-05, | |
| "loss": 0.5193, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 0.29615373729940847, | |
| "grad_norm": 0.661551296710968, | |
| "learning_rate": 1.4651834879894914e-05, | |
| "loss": 0.5148, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 0.29660657176011096, | |
| "grad_norm": 0.7670228481292725, | |
| "learning_rate": 1.4650761259425369e-05, | |
| "loss": 0.5274, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.2970594062208134, | |
| "grad_norm": 0.7227180004119873, | |
| "learning_rate": 1.4649686025624244e-05, | |
| "loss": 0.5116, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 0.2975122406815159, | |
| "grad_norm": 0.8273769617080688, | |
| "learning_rate": 1.4648609178734132e-05, | |
| "loss": 0.4723, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 0.2979650751422183, | |
| "grad_norm": 0.6958010196685791, | |
| "learning_rate": 1.4647530718997987e-05, | |
| "loss": 0.5077, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 0.2984179096029208, | |
| "grad_norm": 0.8402749300003052, | |
| "learning_rate": 1.4646450646659126e-05, | |
| "loss": 0.5319, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 0.29887074406362324, | |
| "grad_norm": 0.7076095938682556, | |
| "learning_rate": 1.4645368961961232e-05, | |
| "loss": 0.5306, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.29932357852432573, | |
| "grad_norm": 0.8241704702377319, | |
| "learning_rate": 1.4644285665148352e-05, | |
| "loss": 0.5652, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 0.29977641298502816, | |
| "grad_norm": 0.6605017781257629, | |
| "learning_rate": 1.4643200756464895e-05, | |
| "loss": 0.4739, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 0.3002292474457306, | |
| "grad_norm": 0.7162265181541443, | |
| "learning_rate": 1.4642114236155635e-05, | |
| "loss": 0.5339, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 0.3006820819064331, | |
| "grad_norm": 0.920584499835968, | |
| "learning_rate": 1.464102610446571e-05, | |
| "loss": 0.477, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 0.3011349163671355, | |
| "grad_norm": 1.0123887062072754, | |
| "learning_rate": 1.463993636164062e-05, | |
| "loss": 0.5005, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.301587750827838, | |
| "grad_norm": 0.8752802610397339, | |
| "learning_rate": 1.4638845007926232e-05, | |
| "loss": 0.5035, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 0.30204058528854044, | |
| "grad_norm": 0.9289608001708984, | |
| "learning_rate": 1.4637752043568767e-05, | |
| "loss": 0.4911, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 0.30249341974924293, | |
| "grad_norm": 0.8423551321029663, | |
| "learning_rate": 1.463665746881482e-05, | |
| "loss": 0.5333, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 0.30294625420994536, | |
| "grad_norm": 1.0525884628295898, | |
| "learning_rate": 1.463556128391135e-05, | |
| "loss": 0.5097, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 0.30339908867064785, | |
| "grad_norm": 0.909200131893158, | |
| "learning_rate": 1.4634463489105666e-05, | |
| "loss": 0.5584, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.3038519231313503, | |
| "grad_norm": 1.0725759267807007, | |
| "learning_rate": 1.4633364084645457e-05, | |
| "loss": 0.5722, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 0.3043047575920528, | |
| "grad_norm": 0.7647678256034851, | |
| "learning_rate": 1.4632263070778758e-05, | |
| "loss": 0.5064, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 0.3047575920527552, | |
| "grad_norm": 1.1567573547363281, | |
| "learning_rate": 1.4631160447753985e-05, | |
| "loss": 0.5108, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 0.3052104265134577, | |
| "grad_norm": 1.012374758720398, | |
| "learning_rate": 1.4630056215819904e-05, | |
| "loss": 0.5589, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 0.30566326097416013, | |
| "grad_norm": 1.4147967100143433, | |
| "learning_rate": 1.4628950375225648e-05, | |
| "loss": 0.5355, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.30611609543486257, | |
| "grad_norm": 0.9129096269607544, | |
| "learning_rate": 1.4627842926220715e-05, | |
| "loss": 0.5529, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 0.30656892989556506, | |
| "grad_norm": 0.9252450466156006, | |
| "learning_rate": 1.4626733869054964e-05, | |
| "loss": 0.4604, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 0.3070217643562675, | |
| "grad_norm": 1.167197585105896, | |
| "learning_rate": 1.4625623203978614e-05, | |
| "loss": 0.5238, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 0.30747459881697, | |
| "grad_norm": 0.672999918460846, | |
| "learning_rate": 1.4624510931242252e-05, | |
| "loss": 0.4724, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 0.3079274332776724, | |
| "grad_norm": 0.8699397444725037, | |
| "learning_rate": 1.4623397051096825e-05, | |
| "loss": 0.5158, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.3083802677383749, | |
| "grad_norm": 0.7273291945457458, | |
| "learning_rate": 1.4622281563793643e-05, | |
| "loss": 0.5246, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 0.30883310219907734, | |
| "grad_norm": 0.9489529728889465, | |
| "learning_rate": 1.4621164469584378e-05, | |
| "loss": 0.5135, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 0.3092859366597798, | |
| "grad_norm": 0.786284863948822, | |
| "learning_rate": 1.4620045768721065e-05, | |
| "loss": 0.4938, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 0.30973877112048226, | |
| "grad_norm": 0.894683837890625, | |
| "learning_rate": 1.4618925461456106e-05, | |
| "loss": 0.5306, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 0.31019160558118475, | |
| "grad_norm": 0.9071069955825806, | |
| "learning_rate": 1.4617803548042254e-05, | |
| "loss": 0.5439, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.3106444400418872, | |
| "grad_norm": 0.8423410654067993, | |
| "learning_rate": 1.4616680028732637e-05, | |
| "loss": 0.4799, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 0.31109727450258967, | |
| "grad_norm": 0.7595556378364563, | |
| "learning_rate": 1.4615554903780738e-05, | |
| "loss": 0.4923, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 0.3115501089632921, | |
| "grad_norm": 0.7766554355621338, | |
| "learning_rate": 1.4614428173440403e-05, | |
| "loss": 0.5725, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 0.31200294342399454, | |
| "grad_norm": 0.8758528828620911, | |
| "learning_rate": 1.4613299837965843e-05, | |
| "loss": 0.5016, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 0.31245577788469703, | |
| "grad_norm": 0.8516387343406677, | |
| "learning_rate": 1.4612169897611626e-05, | |
| "loss": 0.511, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.31290861234539946, | |
| "grad_norm": 0.7681635022163391, | |
| "learning_rate": 1.461103835263269e-05, | |
| "loss": 0.4992, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 0.31336144680610195, | |
| "grad_norm": 0.7421332597732544, | |
| "learning_rate": 1.4609905203284329e-05, | |
| "loss": 0.4717, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 0.3138142812668044, | |
| "grad_norm": 0.7615931630134583, | |
| "learning_rate": 1.46087704498222e-05, | |
| "loss": 0.505, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 0.3142671157275069, | |
| "grad_norm": 0.7673407793045044, | |
| "learning_rate": 1.460763409250232e-05, | |
| "loss": 0.4748, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 0.3147199501882093, | |
| "grad_norm": 0.8670549988746643, | |
| "learning_rate": 1.4606496131581075e-05, | |
| "loss": 0.496, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.3151727846489118, | |
| "grad_norm": 0.8095021843910217, | |
| "learning_rate": 1.4605356567315204e-05, | |
| "loss": 0.4385, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 0.31562561910961423, | |
| "grad_norm": 0.7151927947998047, | |
| "learning_rate": 1.4604215399961815e-05, | |
| "loss": 0.5262, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 0.3160784535703167, | |
| "grad_norm": 1.0873347520828247, | |
| "learning_rate": 1.4603072629778372e-05, | |
| "loss": 0.444, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 0.31653128803101915, | |
| "grad_norm": 0.8482450246810913, | |
| "learning_rate": 1.4601928257022705e-05, | |
| "loss": 0.5739, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 0.31698412249172164, | |
| "grad_norm": 0.7151737213134766, | |
| "learning_rate": 1.4600782281953e-05, | |
| "loss": 0.4712, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.3174369569524241, | |
| "grad_norm": 0.685892641544342, | |
| "learning_rate": 1.4599634704827813e-05, | |
| "loss": 0.4681, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 0.31788979141312657, | |
| "grad_norm": 0.8122037649154663, | |
| "learning_rate": 1.4598485525906055e-05, | |
| "loss": 0.4768, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 0.318342625873829, | |
| "grad_norm": 0.6169959902763367, | |
| "learning_rate": 1.4597334745447e-05, | |
| "loss": 0.4869, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 0.31879546033453143, | |
| "grad_norm": 0.7096338272094727, | |
| "learning_rate": 1.4596182363710281e-05, | |
| "loss": 0.4805, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 0.3192482947952339, | |
| "grad_norm": 0.6837482452392578, | |
| "learning_rate": 1.4595028380955897e-05, | |
| "loss": 0.5615, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.31970112925593636, | |
| "grad_norm": 0.7366883158683777, | |
| "learning_rate": 1.4593872797444206e-05, | |
| "loss": 0.4706, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 0.32015396371663885, | |
| "grad_norm": 0.8051954507827759, | |
| "learning_rate": 1.4592715613435927e-05, | |
| "loss": 0.4847, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 0.3206067981773413, | |
| "grad_norm": 0.8133533596992493, | |
| "learning_rate": 1.4591556829192137e-05, | |
| "loss": 0.4954, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 0.32105963263804377, | |
| "grad_norm": 0.7073555588722229, | |
| "learning_rate": 1.4590396444974284e-05, | |
| "loss": 0.4308, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 0.3215124670987462, | |
| "grad_norm": 0.8296173810958862, | |
| "learning_rate": 1.4589234461044162e-05, | |
| "loss": 0.5252, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.3219653015594487, | |
| "grad_norm": 0.7163283824920654, | |
| "learning_rate": 1.458807087766394e-05, | |
| "loss": 0.5206, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 0.3224181360201511, | |
| "grad_norm": 0.8291743993759155, | |
| "learning_rate": 1.4586905695096137e-05, | |
| "loss": 0.4864, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 0.3228709704808536, | |
| "grad_norm": 0.7736324667930603, | |
| "learning_rate": 1.4585738913603644e-05, | |
| "loss": 0.5039, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 0.32332380494155605, | |
| "grad_norm": 0.7839064002037048, | |
| "learning_rate": 1.45845705334497e-05, | |
| "loss": 0.4585, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 0.32377663940225854, | |
| "grad_norm": 0.782141923904419, | |
| "learning_rate": 1.4583400554897916e-05, | |
| "loss": 0.5271, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.32422947386296097, | |
| "grad_norm": 0.7646268010139465, | |
| "learning_rate": 1.4582228978212256e-05, | |
| "loss": 0.5123, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 0.3246823083236634, | |
| "grad_norm": 0.8933156728744507, | |
| "learning_rate": 1.458105580365705e-05, | |
| "loss": 0.4291, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 0.3251351427843659, | |
| "grad_norm": 0.796796441078186, | |
| "learning_rate": 1.4579881031496982e-05, | |
| "loss": 0.5057, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 0.3255879772450683, | |
| "grad_norm": 0.8032305836677551, | |
| "learning_rate": 1.45787046619971e-05, | |
| "loss": 0.5003, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 0.3260408117057708, | |
| "grad_norm": 0.788154125213623, | |
| "learning_rate": 1.457752669542282e-05, | |
| "loss": 0.4792, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.32649364616647325, | |
| "grad_norm": 0.75339674949646, | |
| "learning_rate": 1.4576347132039902e-05, | |
| "loss": 0.4287, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 0.32694648062717574, | |
| "grad_norm": 1.0328612327575684, | |
| "learning_rate": 1.457516597211448e-05, | |
| "loss": 0.5249, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 0.3273993150878782, | |
| "grad_norm": 1.01662278175354, | |
| "learning_rate": 1.4573983215913043e-05, | |
| "loss": 0.5471, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 0.32785214954858066, | |
| "grad_norm": 0.8054636120796204, | |
| "learning_rate": 1.4572798863702442e-05, | |
| "loss": 0.4786, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 0.3283049840092831, | |
| "grad_norm": 0.8108761310577393, | |
| "learning_rate": 1.4571612915749885e-05, | |
| "loss": 0.4995, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.3287578184699856, | |
| "grad_norm": 1.0297499895095825, | |
| "learning_rate": 1.457042537232294e-05, | |
| "loss": 0.5023, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 0.329210652930688, | |
| "grad_norm": 0.8287637829780579, | |
| "learning_rate": 1.4569236233689537e-05, | |
| "loss": 0.4844, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 0.3296634873913905, | |
| "grad_norm": 0.8962579965591431, | |
| "learning_rate": 1.456804550011797e-05, | |
| "loss": 0.4787, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 0.33011632185209294, | |
| "grad_norm": 0.7159314751625061, | |
| "learning_rate": 1.4566853171876886e-05, | |
| "loss": 0.4616, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 0.33056915631279543, | |
| "grad_norm": 0.9780300855636597, | |
| "learning_rate": 1.456565924923529e-05, | |
| "loss": 0.4748, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.33102199077349787, | |
| "grad_norm": 0.8780568242073059, | |
| "learning_rate": 1.4564463732462559e-05, | |
| "loss": 0.5119, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 0.3314748252342003, | |
| "grad_norm": 0.7616591453552246, | |
| "learning_rate": 1.4563266621828413e-05, | |
| "loss": 0.4763, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 0.3319276596949028, | |
| "grad_norm": 0.913154661655426, | |
| "learning_rate": 1.4562067917602947e-05, | |
| "loss": 0.5038, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 0.3323804941556052, | |
| "grad_norm": 0.7419286370277405, | |
| "learning_rate": 1.4560867620056608e-05, | |
| "loss": 0.5218, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 0.3328333286163077, | |
| "grad_norm": 0.759284496307373, | |
| "learning_rate": 1.4559665729460198e-05, | |
| "loss": 0.4869, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.33328616307701014, | |
| "grad_norm": 0.885164201259613, | |
| "learning_rate": 1.455846224608489e-05, | |
| "loss": 0.5, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 0.33373899753771263, | |
| "grad_norm": 0.7694923281669617, | |
| "learning_rate": 1.4557257170202207e-05, | |
| "loss": 0.4784, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 0.33419183199841507, | |
| "grad_norm": 0.7753260731697083, | |
| "learning_rate": 1.4556050502084034e-05, | |
| "loss": 0.4699, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 0.33464466645911756, | |
| "grad_norm": 0.6947426795959473, | |
| "learning_rate": 1.4554842242002614e-05, | |
| "loss": 0.4374, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 0.33509750091982, | |
| "grad_norm": 0.796204686164856, | |
| "learning_rate": 1.4553632390230554e-05, | |
| "loss": 0.5163, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.3355503353805225, | |
| "grad_norm": 0.7171505093574524, | |
| "learning_rate": 1.4552420947040817e-05, | |
| "loss": 0.484, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 0.3360031698412249, | |
| "grad_norm": 0.7220250964164734, | |
| "learning_rate": 1.4551207912706725e-05, | |
| "loss": 0.4531, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 0.3364560043019274, | |
| "grad_norm": 0.7847347855567932, | |
| "learning_rate": 1.4549993287501955e-05, | |
| "loss": 0.5114, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 0.33690883876262984, | |
| "grad_norm": 0.7722388505935669, | |
| "learning_rate": 1.4548777071700549e-05, | |
| "loss": 0.4938, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 0.33736167322333227, | |
| "grad_norm": 0.7462444305419922, | |
| "learning_rate": 1.4547559265576907e-05, | |
| "loss": 0.4641, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.33781450768403476, | |
| "grad_norm": 0.8149163722991943, | |
| "learning_rate": 1.4546339869405783e-05, | |
| "loss": 0.4787, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 0.3382673421447372, | |
| "grad_norm": 0.8563286066055298, | |
| "learning_rate": 1.4545118883462298e-05, | |
| "loss": 0.4962, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 0.3387201766054397, | |
| "grad_norm": 0.7843456268310547, | |
| "learning_rate": 1.4543896308021924e-05, | |
| "loss": 0.4664, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 0.3391730110661421, | |
| "grad_norm": 0.7675468325614929, | |
| "learning_rate": 1.4542672143360493e-05, | |
| "loss": 0.4841, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 0.3396258455268446, | |
| "grad_norm": 0.8849632740020752, | |
| "learning_rate": 1.4541446389754201e-05, | |
| "loss": 0.5167, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.34007867998754704, | |
| "grad_norm": 0.778249979019165, | |
| "learning_rate": 1.4540219047479597e-05, | |
| "loss": 0.4565, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 0.34053151444824953, | |
| "grad_norm": 0.7831906080245972, | |
| "learning_rate": 1.4538990116813589e-05, | |
| "loss": 0.4558, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 0.34098434890895196, | |
| "grad_norm": 0.777973473072052, | |
| "learning_rate": 1.4537759598033446e-05, | |
| "loss": 0.4822, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 0.34143718336965445, | |
| "grad_norm": 0.6689452528953552, | |
| "learning_rate": 1.453652749141679e-05, | |
| "loss": 0.4402, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 0.3418900178303569, | |
| "grad_norm": 0.6991743445396423, | |
| "learning_rate": 1.453529379724161e-05, | |
| "loss": 0.4195, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.3423428522910594, | |
| "grad_norm": 0.8085342049598694, | |
| "learning_rate": 1.4534058515786245e-05, | |
| "loss": 0.4789, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 0.3427956867517618, | |
| "grad_norm": 0.7133055329322815, | |
| "learning_rate": 1.4532821647329397e-05, | |
| "loss": 0.4706, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 0.34324852121246424, | |
| "grad_norm": 0.7966130375862122, | |
| "learning_rate": 1.4531583192150119e-05, | |
| "loss": 0.4782, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 0.34370135567316673, | |
| "grad_norm": 0.7512375712394714, | |
| "learning_rate": 1.4530343150527836e-05, | |
| "loss": 0.4674, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 0.34415419013386916, | |
| "grad_norm": 0.9407236576080322, | |
| "learning_rate": 1.4529101522742311e-05, | |
| "loss": 0.4598, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.34460702459457165, | |
| "grad_norm": 0.8138443231582642, | |
| "learning_rate": 1.4527858309073688e-05, | |
| "loss": 0.4619, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 0.3450598590552741, | |
| "grad_norm": 0.6962406039237976, | |
| "learning_rate": 1.452661350980245e-05, | |
| "loss": 0.4936, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 0.3455126935159766, | |
| "grad_norm": 0.7135865092277527, | |
| "learning_rate": 1.4525367125209445e-05, | |
| "loss": 0.4329, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 0.345965527976679, | |
| "grad_norm": 0.8204096555709839, | |
| "learning_rate": 1.452411915557588e-05, | |
| "loss": 0.4502, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 0.3464183624373815, | |
| "grad_norm": 0.7781338691711426, | |
| "learning_rate": 1.4522869601183318e-05, | |
| "loss": 0.4148, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.34687119689808393, | |
| "grad_norm": 0.7633736729621887, | |
| "learning_rate": 1.4521618462313679e-05, | |
| "loss": 0.4297, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 0.3473240313587864, | |
| "grad_norm": 0.904326856136322, | |
| "learning_rate": 1.4520365739249236e-05, | |
| "loss": 0.4972, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 0.34777686581948886, | |
| "grad_norm": 0.7865485548973083, | |
| "learning_rate": 1.4519111432272632e-05, | |
| "loss": 0.4718, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 0.34822970028019135, | |
| "grad_norm": 0.9248954057693481, | |
| "learning_rate": 1.4517855541666855e-05, | |
| "loss": 0.4736, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 0.3486825347408938, | |
| "grad_norm": 0.8533254861831665, | |
| "learning_rate": 1.4516598067715257e-05, | |
| "loss": 0.4452, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.34913536920159627, | |
| "grad_norm": 0.9142417311668396, | |
| "learning_rate": 1.4515339010701544e-05, | |
| "loss": 0.3997, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 0.3495882036622987, | |
| "grad_norm": 0.9811474680900574, | |
| "learning_rate": 1.4514078370909782e-05, | |
| "loss": 0.4839, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 0.35004103812300114, | |
| "grad_norm": 0.8523197770118713, | |
| "learning_rate": 1.451281614862439e-05, | |
| "loss": 0.5039, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 0.3504938725837036, | |
| "grad_norm": 0.9271774888038635, | |
| "learning_rate": 1.4511552344130147e-05, | |
| "loss": 0.526, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 0.35094670704440606, | |
| "grad_norm": 0.7626346349716187, | |
| "learning_rate": 1.4510286957712192e-05, | |
| "loss": 0.4031, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.35139954150510855, | |
| "grad_norm": 0.9117299318313599, | |
| "learning_rate": 1.4509019989656013e-05, | |
| "loss": 0.4756, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 0.351852375965811, | |
| "grad_norm": 0.7210586071014404, | |
| "learning_rate": 1.4507751440247461e-05, | |
| "loss": 0.4667, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 0.35230521042651347, | |
| "grad_norm": 0.7991892695426941, | |
| "learning_rate": 1.4506481309772743e-05, | |
| "loss": 0.4789, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 0.3527580448872159, | |
| "grad_norm": 0.7282071113586426, | |
| "learning_rate": 1.4505209598518418e-05, | |
| "loss": 0.4288, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 0.3532108793479184, | |
| "grad_norm": 0.8729153275489807, | |
| "learning_rate": 1.450393630677141e-05, | |
| "loss": 0.4637, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.3536637138086208, | |
| "grad_norm": 0.8181951642036438, | |
| "learning_rate": 1.4502661434818992e-05, | |
| "loss": 0.4957, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 0.3541165482693233, | |
| "grad_norm": 0.8652138710021973, | |
| "learning_rate": 1.4501384982948799e-05, | |
| "loss": 0.4715, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 0.35456938273002575, | |
| "grad_norm": 0.9875885248184204, | |
| "learning_rate": 1.4500106951448816e-05, | |
| "loss": 0.4996, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 0.35502221719072824, | |
| "grad_norm": 0.7315845489501953, | |
| "learning_rate": 1.4498827340607393e-05, | |
| "loss": 0.46, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 0.3554750516514307, | |
| "grad_norm": 0.93685382604599, | |
| "learning_rate": 1.4497546150713227e-05, | |
| "loss": 0.4265, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.3559278861121331, | |
| "grad_norm": 0.7243474125862122, | |
| "learning_rate": 1.449626338205538e-05, | |
| "loss": 0.4844, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 0.3563807205728356, | |
| "grad_norm": 0.8693944215774536, | |
| "learning_rate": 1.4494979034923265e-05, | |
| "loss": 0.5018, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 0.35683355503353803, | |
| "grad_norm": 0.7501546144485474, | |
| "learning_rate": 1.4493693109606649e-05, | |
| "loss": 0.4865, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 0.3572863894942405, | |
| "grad_norm": 0.8437265157699585, | |
| "learning_rate": 1.4492405606395663e-05, | |
| "loss": 0.495, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 0.35773922395494295, | |
| "grad_norm": 0.782289981842041, | |
| "learning_rate": 1.4491116525580786e-05, | |
| "loss": 0.4567, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.35819205841564544, | |
| "grad_norm": 0.755082368850708, | |
| "learning_rate": 1.4489825867452858e-05, | |
| "loss": 0.4891, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 0.3586448928763479, | |
| "grad_norm": 0.7294925451278687, | |
| "learning_rate": 1.4488533632303072e-05, | |
| "loss": 0.4613, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 0.35909772733705037, | |
| "grad_norm": 0.8621608018875122, | |
| "learning_rate": 1.448723982042298e-05, | |
| "loss": 0.4479, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 0.3595505617977528, | |
| "grad_norm": 0.7976484298706055, | |
| "learning_rate": 1.4485944432104483e-05, | |
| "loss": 0.4942, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 0.3600033962584553, | |
| "grad_norm": 0.6837937235832214, | |
| "learning_rate": 1.4484647467639848e-05, | |
| "loss": 0.4596, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.3604562307191577, | |
| "grad_norm": 0.7584710717201233, | |
| "learning_rate": 1.4483348927321688e-05, | |
| "loss": 0.4767, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 0.3609090651798602, | |
| "grad_norm": 0.8236560225486755, | |
| "learning_rate": 1.4482048811442978e-05, | |
| "loss": 0.4923, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 0.36136189964056264, | |
| "grad_norm": 0.7171486020088196, | |
| "learning_rate": 1.4480747120297043e-05, | |
| "loss": 0.4381, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 0.3618147341012651, | |
| "grad_norm": 0.7314976453781128, | |
| "learning_rate": 1.447944385417757e-05, | |
| "loss": 0.4573, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 0.36226756856196757, | |
| "grad_norm": 0.7224847674369812, | |
| "learning_rate": 1.4478139013378597e-05, | |
| "loss": 0.5023, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.36272040302267, | |
| "grad_norm": 0.7627956867218018, | |
| "learning_rate": 1.4476832598194514e-05, | |
| "loss": 0.4326, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 0.3631732374833725, | |
| "grad_norm": 0.8306934237480164, | |
| "learning_rate": 1.4475524608920071e-05, | |
| "loss": 0.5251, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 0.3636260719440749, | |
| "grad_norm": 0.8524587750434875, | |
| "learning_rate": 1.4474215045850377e-05, | |
| "loss": 0.4546, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 0.3640789064047774, | |
| "grad_norm": 0.6962935924530029, | |
| "learning_rate": 1.4472903909280888e-05, | |
| "loss": 0.4495, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 0.36453174086547985, | |
| "grad_norm": 0.847118616104126, | |
| "learning_rate": 1.4471591199507417e-05, | |
| "loss": 0.4508, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.36498457532618234, | |
| "grad_norm": 0.7857109308242798, | |
| "learning_rate": 1.4470276916826138e-05, | |
| "loss": 0.456, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 0.36543740978688477, | |
| "grad_norm": 0.9544152021408081, | |
| "learning_rate": 1.446896106153357e-05, | |
| "loss": 0.4344, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 0.36589024424758726, | |
| "grad_norm": 0.6988970041275024, | |
| "learning_rate": 1.4467643633926598e-05, | |
| "loss": 0.4755, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 0.3663430787082897, | |
| "grad_norm": 0.9396758675575256, | |
| "learning_rate": 1.4466324634302448e-05, | |
| "loss": 0.4444, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 0.3667959131689922, | |
| "grad_norm": 0.813913106918335, | |
| "learning_rate": 1.4465004062958716e-05, | |
| "loss": 0.4326, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.3672487476296946, | |
| "grad_norm": 0.8034855723381042, | |
| "learning_rate": 1.4463681920193341e-05, | |
| "loss": 0.4905, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 0.3677015820903971, | |
| "grad_norm": 0.7723264694213867, | |
| "learning_rate": 1.446235820630462e-05, | |
| "loss": 0.376, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 0.36815441655109954, | |
| "grad_norm": 0.8094017505645752, | |
| "learning_rate": 1.4461032921591207e-05, | |
| "loss": 0.5052, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 0.368607251011802, | |
| "grad_norm": 0.7081817984580994, | |
| "learning_rate": 1.4459706066352105e-05, | |
| "loss": 0.4714, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 0.36906008547250446, | |
| "grad_norm": 0.7871371507644653, | |
| "learning_rate": 1.4458377640886682e-05, | |
| "loss": 0.4692, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.3695129199332069, | |
| "grad_norm": 0.8632165789604187, | |
| "learning_rate": 1.4457047645494645e-05, | |
| "loss": 0.4475, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 0.3699657543939094, | |
| "grad_norm": 0.875834047794342, | |
| "learning_rate": 1.4455716080476068e-05, | |
| "loss": 0.4508, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 0.3704185888546118, | |
| "grad_norm": 0.7545289397239685, | |
| "learning_rate": 1.4454382946131373e-05, | |
| "loss": 0.4841, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 0.3708714233153143, | |
| "grad_norm": 0.8614468574523926, | |
| "learning_rate": 1.4453048242761337e-05, | |
| "loss": 0.477, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 0.37132425777601674, | |
| "grad_norm": 0.9642696380615234, | |
| "learning_rate": 1.4451711970667094e-05, | |
| "loss": 0.4879, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.37177709223671923, | |
| "grad_norm": 0.7030223608016968, | |
| "learning_rate": 1.4450374130150124e-05, | |
| "loss": 0.4212, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 0.37222992669742166, | |
| "grad_norm": 0.9804565906524658, | |
| "learning_rate": 1.4449034721512271e-05, | |
| "loss": 0.4439, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 0.37268276115812415, | |
| "grad_norm": 0.732269287109375, | |
| "learning_rate": 1.4447693745055728e-05, | |
| "loss": 0.4869, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 0.3731355956188266, | |
| "grad_norm": 0.914057195186615, | |
| "learning_rate": 1.444635120108304e-05, | |
| "loss": 0.4113, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 0.3735884300795291, | |
| "grad_norm": 0.8915912508964539, | |
| "learning_rate": 1.4445007089897108e-05, | |
| "loss": 0.4566, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.3740412645402315, | |
| "grad_norm": 0.803848147392273, | |
| "learning_rate": 1.4443661411801183e-05, | |
| "loss": 0.4411, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 0.37449409900093394, | |
| "grad_norm": 0.8054463267326355, | |
| "learning_rate": 1.4442314167098878e-05, | |
| "loss": 0.4521, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 0.37494693346163643, | |
| "grad_norm": 0.6809094548225403, | |
| "learning_rate": 1.444096535609415e-05, | |
| "loss": 0.4083, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 0.37494693346163643, | |
| "eval_loss": 0.4509795606136322, | |
| "eval_runtime": 19.4619, | |
| "eval_samples_per_second": 5.858, | |
| "eval_steps_per_second": 0.771, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 0.37539976792233887, | |
| "grad_norm": 0.759226381778717, | |
| "learning_rate": 1.4439614979091316e-05, | |
| "loss": 0.3904, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 0.37585260238304136, | |
| "grad_norm": 0.7871491312980652, | |
| "learning_rate": 1.4438263036395045e-05, | |
| "loss": 0.4172, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.3763054368437438, | |
| "grad_norm": 0.6610317230224609, | |
| "learning_rate": 1.4436909528310352e-05, | |
| "loss": 0.4328, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 0.3767582713044463, | |
| "grad_norm": 0.7470182776451111, | |
| "learning_rate": 1.4435554455142614e-05, | |
| "loss": 0.4792, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 0.3772111057651487, | |
| "grad_norm": 0.7115527391433716, | |
| "learning_rate": 1.4434197817197558e-05, | |
| "loss": 0.4253, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 0.3776639402258512, | |
| "grad_norm": 0.8027327656745911, | |
| "learning_rate": 1.4432839614781268e-05, | |
| "loss": 0.4349, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 0.37811677468655364, | |
| "grad_norm": 0.6858651041984558, | |
| "learning_rate": 1.4431479848200174e-05, | |
| "loss": 0.464, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.3785696091472561, | |
| "grad_norm": 0.780019998550415, | |
| "learning_rate": 1.443011851776106e-05, | |
| "loss": 0.4301, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 0.37902244360795856, | |
| "grad_norm": 0.8875818252563477, | |
| "learning_rate": 1.4428755623771068e-05, | |
| "loss": 0.4638, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 0.37947527806866105, | |
| "grad_norm": 0.7595096230506897, | |
| "learning_rate": 1.442739116653769e-05, | |
| "loss": 0.4761, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 0.3799281125293635, | |
| "grad_norm": 0.7235725522041321, | |
| "learning_rate": 1.4426025146368768e-05, | |
| "loss": 0.4596, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 0.38038094699006597, | |
| "grad_norm": 0.8074544072151184, | |
| "learning_rate": 1.4424657563572499e-05, | |
| "loss": 0.5049, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.3808337814507684, | |
| "grad_norm": 0.784801185131073, | |
| "learning_rate": 1.4423288418457437e-05, | |
| "loss": 0.4387, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 0.38128661591147084, | |
| "grad_norm": 0.6974912285804749, | |
| "learning_rate": 1.442191771133248e-05, | |
| "loss": 0.4004, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 0.38173945037217333, | |
| "grad_norm": 0.6632540822029114, | |
| "learning_rate": 1.4420545442506885e-05, | |
| "loss": 0.4515, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 0.38219228483287576, | |
| "grad_norm": 0.7197034358978271, | |
| "learning_rate": 1.4419171612290256e-05, | |
| "loss": 0.4033, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 0.38264511929357825, | |
| "grad_norm": 0.6470990180969238, | |
| "learning_rate": 1.4417796220992553e-05, | |
| "loss": 0.4453, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.3830979537542807, | |
| "grad_norm": 0.7469446659088135, | |
| "learning_rate": 1.441641926892409e-05, | |
| "loss": 0.4657, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 0.3835507882149832, | |
| "grad_norm": 0.7464520335197449, | |
| "learning_rate": 1.4415040756395527e-05, | |
| "loss": 0.4376, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 0.3840036226756856, | |
| "grad_norm": 0.800579309463501, | |
| "learning_rate": 1.441366068371788e-05, | |
| "loss": 0.4434, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 0.3844564571363881, | |
| "grad_norm": 0.6948139071464539, | |
| "learning_rate": 1.4412279051202518e-05, | |
| "loss": 0.3879, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 0.38490929159709053, | |
| "grad_norm": 0.6828701496124268, | |
| "learning_rate": 1.441089585916116e-05, | |
| "loss": 0.4162, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.385362126057793, | |
| "grad_norm": 0.9038822054862976, | |
| "learning_rate": 1.4409511107905878e-05, | |
| "loss": 0.457, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 0.38581496051849545, | |
| "grad_norm": 0.7000515460968018, | |
| "learning_rate": 1.4408124797749093e-05, | |
| "loss": 0.4054, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 0.38626779497919794, | |
| "grad_norm": 0.8235939145088196, | |
| "learning_rate": 1.4406736929003582e-05, | |
| "loss": 0.4283, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 0.3867206294399004, | |
| "grad_norm": 0.824226975440979, | |
| "learning_rate": 1.4405347501982473e-05, | |
| "loss": 0.4499, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 0.3871734639006028, | |
| "grad_norm": 0.7194985151290894, | |
| "learning_rate": 1.4403956516999235e-05, | |
| "loss": 0.4278, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.3876262983613053, | |
| "grad_norm": 0.7407351732254028, | |
| "learning_rate": 1.4402563974367709e-05, | |
| "loss": 0.4495, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 0.38807913282200773, | |
| "grad_norm": 0.7273187041282654, | |
| "learning_rate": 1.4401169874402071e-05, | |
| "loss": 0.389, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 0.3885319672827102, | |
| "grad_norm": 0.6498844623565674, | |
| "learning_rate": 1.4399774217416852e-05, | |
| "loss": 0.3919, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 0.38898480174341266, | |
| "grad_norm": 0.7037122845649719, | |
| "learning_rate": 1.4398377003726936e-05, | |
| "loss": 0.4601, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 0.38943763620411515, | |
| "grad_norm": 0.7716974020004272, | |
| "learning_rate": 1.4396978233647561e-05, | |
| "loss": 0.4147, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.3898904706648176, | |
| "grad_norm": 0.7838717699050903, | |
| "learning_rate": 1.439557790749431e-05, | |
| "loss": 0.4185, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 0.39034330512552007, | |
| "grad_norm": 0.7486699819564819, | |
| "learning_rate": 1.439417602558312e-05, | |
| "loss": 0.4397, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 0.3907961395862225, | |
| "grad_norm": 0.7644836902618408, | |
| "learning_rate": 1.4392772588230281e-05, | |
| "loss": 0.4814, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 0.391248974046925, | |
| "grad_norm": 0.8007204532623291, | |
| "learning_rate": 1.439136759575243e-05, | |
| "loss": 0.47, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 0.3917018085076274, | |
| "grad_norm": 0.7783535122871399, | |
| "learning_rate": 1.4389961048466557e-05, | |
| "loss": 0.3774, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.3921546429683299, | |
| "grad_norm": 0.7233867645263672, | |
| "learning_rate": 1.4388552946690004e-05, | |
| "loss": 0.4344, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 0.39260747742903235, | |
| "grad_norm": 0.6840167045593262, | |
| "learning_rate": 1.4387143290740462e-05, | |
| "loss": 0.4511, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 0.3930603118897348, | |
| "grad_norm": 0.8994132280349731, | |
| "learning_rate": 1.4385732080935974e-05, | |
| "loss": 0.4716, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 0.39351314635043727, | |
| "grad_norm": 0.8595420122146606, | |
| "learning_rate": 1.4384319317594926e-05, | |
| "loss": 0.4388, | |
| "step": 869 | |
| }, | |
| { | |
| "epoch": 0.3939659808111397, | |
| "grad_norm": 0.675515353679657, | |
| "learning_rate": 1.4382905001036073e-05, | |
| "loss": 0.4134, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.3944188152718422, | |
| "grad_norm": 1.7675050497055054, | |
| "learning_rate": 1.4381489131578499e-05, | |
| "loss": 0.4454, | |
| "step": 871 | |
| }, | |
| { | |
| "epoch": 0.3948716497325446, | |
| "grad_norm": 1.0099380016326904, | |
| "learning_rate": 1.438007170954165e-05, | |
| "loss": 0.4738, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 0.3953244841932471, | |
| "grad_norm": 0.7462349534034729, | |
| "learning_rate": 1.437865273524532e-05, | |
| "loss": 0.3707, | |
| "step": 873 | |
| }, | |
| { | |
| "epoch": 0.39577731865394955, | |
| "grad_norm": 0.7054492235183716, | |
| "learning_rate": 1.4377232209009657e-05, | |
| "loss": 0.4068, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 0.39623015311465204, | |
| "grad_norm": 0.7035594582557678, | |
| "learning_rate": 1.4375810131155152e-05, | |
| "loss": 0.4456, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.3966829875753545, | |
| "grad_norm": 0.691320538520813, | |
| "learning_rate": 1.4374386502002649e-05, | |
| "loss": 0.4188, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 0.39713582203605696, | |
| "grad_norm": 0.7369733452796936, | |
| "learning_rate": 1.4372961321873344e-05, | |
| "loss": 0.4572, | |
| "step": 877 | |
| }, | |
| { | |
| "epoch": 0.3975886564967594, | |
| "grad_norm": 0.687791109085083, | |
| "learning_rate": 1.4371534591088785e-05, | |
| "loss": 0.4136, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 0.3980414909574619, | |
| "grad_norm": 0.7917645573616028, | |
| "learning_rate": 1.437010630997086e-05, | |
| "loss": 0.4474, | |
| "step": 879 | |
| }, | |
| { | |
| "epoch": 0.3984943254181643, | |
| "grad_norm": 0.7191473245620728, | |
| "learning_rate": 1.4368676478841818e-05, | |
| "loss": 0.4608, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.3989471598788668, | |
| "grad_norm": 0.7102738618850708, | |
| "learning_rate": 1.4367245098024248e-05, | |
| "loss": 0.4014, | |
| "step": 881 | |
| }, | |
| { | |
| "epoch": 0.39939999433956924, | |
| "grad_norm": 0.8336265683174133, | |
| "learning_rate": 1.4365812167841096e-05, | |
| "loss": 0.4597, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 0.3998528288002717, | |
| "grad_norm": 0.6486582159996033, | |
| "learning_rate": 1.436437768861566e-05, | |
| "loss": 0.4273, | |
| "step": 883 | |
| }, | |
| { | |
| "epoch": 0.40030566326097416, | |
| "grad_norm": 0.779009222984314, | |
| "learning_rate": 1.4362941660671574e-05, | |
| "loss": 0.4285, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 0.4007584977216766, | |
| "grad_norm": 0.8129340410232544, | |
| "learning_rate": 1.4361504084332835e-05, | |
| "loss": 0.4729, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.4012113321823791, | |
| "grad_norm": 0.765630841255188, | |
| "learning_rate": 1.4360064959923784e-05, | |
| "loss": 0.4146, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 0.4016641666430815, | |
| "grad_norm": 0.7774229049682617, | |
| "learning_rate": 1.435862428776911e-05, | |
| "loss": 0.442, | |
| "step": 887 | |
| }, | |
| { | |
| "epoch": 0.402117001103784, | |
| "grad_norm": 0.7265660166740417, | |
| "learning_rate": 1.4357182068193856e-05, | |
| "loss": 0.4161, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 0.40256983556448644, | |
| "grad_norm": 0.6984989643096924, | |
| "learning_rate": 1.4355738301523405e-05, | |
| "loss": 0.4508, | |
| "step": 889 | |
| }, | |
| { | |
| "epoch": 0.40302267002518893, | |
| "grad_norm": 0.8146981596946716, | |
| "learning_rate": 1.4354292988083503e-05, | |
| "loss": 0.3812, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.40347550448589137, | |
| "grad_norm": 0.7826554775238037, | |
| "learning_rate": 1.435284612820023e-05, | |
| "loss": 0.4889, | |
| "step": 891 | |
| }, | |
| { | |
| "epoch": 0.40392833894659386, | |
| "grad_norm": 0.7819204926490784, | |
| "learning_rate": 1.4351397722200023e-05, | |
| "loss": 0.4392, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 0.4043811734072963, | |
| "grad_norm": 0.851521909236908, | |
| "learning_rate": 1.4349947770409669e-05, | |
| "loss": 0.457, | |
| "step": 893 | |
| }, | |
| { | |
| "epoch": 0.4048340078679988, | |
| "grad_norm": 0.751393735408783, | |
| "learning_rate": 1.4348496273156302e-05, | |
| "loss": 0.4016, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 0.4052868423287012, | |
| "grad_norm": 0.7438313961029053, | |
| "learning_rate": 1.4347043230767402e-05, | |
| "loss": 0.4529, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.40573967678940365, | |
| "grad_norm": 0.8485840559005737, | |
| "learning_rate": 1.43455886435708e-05, | |
| "loss": 0.4588, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 0.40619251125010614, | |
| "grad_norm": 0.6871113777160645, | |
| "learning_rate": 1.4344132511894673e-05, | |
| "loss": 0.4313, | |
| "step": 897 | |
| }, | |
| { | |
| "epoch": 0.40664534571080857, | |
| "grad_norm": 0.7333969473838806, | |
| "learning_rate": 1.4342674836067553e-05, | |
| "loss": 0.4039, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 0.40709818017151106, | |
| "grad_norm": 0.7326341867446899, | |
| "learning_rate": 1.4341215616418316e-05, | |
| "loss": 0.4024, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 0.4075510146322135, | |
| "grad_norm": 0.754438042640686, | |
| "learning_rate": 1.4339754853276182e-05, | |
| "loss": 0.4381, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.408003849092916, | |
| "grad_norm": 0.7537654638290405, | |
| "learning_rate": 1.4338292546970729e-05, | |
| "loss": 0.3864, | |
| "step": 901 | |
| }, | |
| { | |
| "epoch": 0.4084566835536184, | |
| "grad_norm": 0.7983335852622986, | |
| "learning_rate": 1.4336828697831873e-05, | |
| "loss": 0.465, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 0.4089095180143209, | |
| "grad_norm": 0.6547751426696777, | |
| "learning_rate": 1.4335363306189887e-05, | |
| "loss": 0.4017, | |
| "step": 903 | |
| }, | |
| { | |
| "epoch": 0.40936235247502334, | |
| "grad_norm": 0.8735558390617371, | |
| "learning_rate": 1.4333896372375383e-05, | |
| "loss": 0.4062, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 0.40981518693572583, | |
| "grad_norm": 0.6709791421890259, | |
| "learning_rate": 1.4332427896719331e-05, | |
| "loss": 0.4209, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.41026802139642826, | |
| "grad_norm": 1.0297253131866455, | |
| "learning_rate": 1.433095787955304e-05, | |
| "loss": 0.4074, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 0.41072085585713075, | |
| "grad_norm": 0.6717653274536133, | |
| "learning_rate": 1.4329486321208176e-05, | |
| "loss": 0.4291, | |
| "step": 907 | |
| }, | |
| { | |
| "epoch": 0.4111736903178332, | |
| "grad_norm": 0.789406418800354, | |
| "learning_rate": 1.4328013222016739e-05, | |
| "loss": 0.4289, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 0.4116265247785357, | |
| "grad_norm": 0.7695425152778625, | |
| "learning_rate": 1.432653858231109e-05, | |
| "loss": 0.3966, | |
| "step": 909 | |
| }, | |
| { | |
| "epoch": 0.4120793592392381, | |
| "grad_norm": 0.8856887817382812, | |
| "learning_rate": 1.4325062402423932e-05, | |
| "loss": 0.3951, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.41253219369994054, | |
| "grad_norm": 0.8140108585357666, | |
| "learning_rate": 1.4323584682688317e-05, | |
| "loss": 0.4092, | |
| "step": 911 | |
| }, | |
| { | |
| "epoch": 0.41298502816064303, | |
| "grad_norm": 0.8315860033035278, | |
| "learning_rate": 1.4322105423437639e-05, | |
| "loss": 0.4223, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 0.41343786262134546, | |
| "grad_norm": 0.7078157067298889, | |
| "learning_rate": 1.4320624625005648e-05, | |
| "loss": 0.4425, | |
| "step": 913 | |
| }, | |
| { | |
| "epoch": 0.41389069708204795, | |
| "grad_norm": 0.740716814994812, | |
| "learning_rate": 1.4319142287726436e-05, | |
| "loss": 0.417, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 0.4143435315427504, | |
| "grad_norm": 0.7005321979522705, | |
| "learning_rate": 1.4317658411934442e-05, | |
| "loss": 0.3627, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.4147963660034529, | |
| "grad_norm": 0.6697372794151306, | |
| "learning_rate": 1.4316172997964452e-05, | |
| "loss": 0.4012, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 0.4152492004641553, | |
| "grad_norm": 0.9033950567245483, | |
| "learning_rate": 1.4314686046151605e-05, | |
| "loss": 0.4569, | |
| "step": 917 | |
| }, | |
| { | |
| "epoch": 0.4157020349248578, | |
| "grad_norm": 0.7096682786941528, | |
| "learning_rate": 1.4313197556831378e-05, | |
| "loss": 0.4487, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 0.41615486938556023, | |
| "grad_norm": 0.7417932152748108, | |
| "learning_rate": 1.4311707530339601e-05, | |
| "loss": 0.4104, | |
| "step": 919 | |
| }, | |
| { | |
| "epoch": 0.4166077038462627, | |
| "grad_norm": 0.6767085790634155, | |
| "learning_rate": 1.4310215967012449e-05, | |
| "loss": 0.4512, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.41706053830696516, | |
| "grad_norm": 0.9227702617645264, | |
| "learning_rate": 1.4308722867186442e-05, | |
| "loss": 0.3907, | |
| "step": 921 | |
| }, | |
| { | |
| "epoch": 0.41751337276766765, | |
| "grad_norm": 0.7554175853729248, | |
| "learning_rate": 1.430722823119845e-05, | |
| "loss": 0.4507, | |
| "step": 922 | |
| }, | |
| { | |
| "epoch": 0.4179662072283701, | |
| "grad_norm": 1.1104964017868042, | |
| "learning_rate": 1.4305732059385687e-05, | |
| "loss": 0.3809, | |
| "step": 923 | |
| }, | |
| { | |
| "epoch": 0.4184190416890725, | |
| "grad_norm": 0.6951049566268921, | |
| "learning_rate": 1.4304234352085715e-05, | |
| "loss": 0.4229, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 0.418871876149775, | |
| "grad_norm": 0.9023833870887756, | |
| "learning_rate": 1.4302735109636442e-05, | |
| "loss": 0.4166, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.41932471061047744, | |
| "grad_norm": 0.6952568888664246, | |
| "learning_rate": 1.430123433237612e-05, | |
| "loss": 0.3956, | |
| "step": 926 | |
| }, | |
| { | |
| "epoch": 0.4197775450711799, | |
| "grad_norm": 0.8569507002830505, | |
| "learning_rate": 1.4299732020643354e-05, | |
| "loss": 0.4372, | |
| "step": 927 | |
| }, | |
| { | |
| "epoch": 0.42023037953188236, | |
| "grad_norm": 0.8564557433128357, | |
| "learning_rate": 1.4298228174777086e-05, | |
| "loss": 0.4311, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 0.42068321399258485, | |
| "grad_norm": 0.9847443699836731, | |
| "learning_rate": 1.4296722795116614e-05, | |
| "loss": 0.4198, | |
| "step": 929 | |
| }, | |
| { | |
| "epoch": 0.4211360484532873, | |
| "grad_norm": 0.9145928621292114, | |
| "learning_rate": 1.4295215882001572e-05, | |
| "loss": 0.4092, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.42158888291398977, | |
| "grad_norm": 1.0060977935791016, | |
| "learning_rate": 1.4293707435771947e-05, | |
| "loss": 0.38, | |
| "step": 931 | |
| }, | |
| { | |
| "epoch": 0.4220417173746922, | |
| "grad_norm": 0.8660867810249329, | |
| "learning_rate": 1.4292197456768069e-05, | |
| "loss": 0.4302, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 0.4224945518353947, | |
| "grad_norm": 0.7605267763137817, | |
| "learning_rate": 1.4290685945330615e-05, | |
| "loss": 0.4346, | |
| "step": 933 | |
| }, | |
| { | |
| "epoch": 0.4229473862960971, | |
| "grad_norm": 0.9517015814781189, | |
| "learning_rate": 1.4289172901800606e-05, | |
| "loss": 0.4111, | |
| "step": 934 | |
| }, | |
| { | |
| "epoch": 0.4234002207567996, | |
| "grad_norm": 0.7545219659805298, | |
| "learning_rate": 1.4287658326519414e-05, | |
| "loss": 0.4231, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.42385305521750205, | |
| "grad_norm": 0.7649204134941101, | |
| "learning_rate": 1.4286142219828748e-05, | |
| "loss": 0.4532, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 0.4243058896782045, | |
| "grad_norm": 0.8143602013587952, | |
| "learning_rate": 1.428462458207067e-05, | |
| "loss": 0.4212, | |
| "step": 937 | |
| }, | |
| { | |
| "epoch": 0.424758724138907, | |
| "grad_norm": 0.9045494794845581, | |
| "learning_rate": 1.4283105413587581e-05, | |
| "loss": 0.4313, | |
| "step": 938 | |
| }, | |
| { | |
| "epoch": 0.4252115585996094, | |
| "grad_norm": 0.9623934030532837, | |
| "learning_rate": 1.4281584714722237e-05, | |
| "loss": 0.4485, | |
| "step": 939 | |
| }, | |
| { | |
| "epoch": 0.4256643930603119, | |
| "grad_norm": 0.7864290475845337, | |
| "learning_rate": 1.4280062485817727e-05, | |
| "loss": 0.4093, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.42611722752101433, | |
| "grad_norm": 0.6734517812728882, | |
| "learning_rate": 1.4278538727217495e-05, | |
| "loss": 0.4376, | |
| "step": 941 | |
| }, | |
| { | |
| "epoch": 0.4265700619817168, | |
| "grad_norm": 0.7887140512466431, | |
| "learning_rate": 1.4277013439265325e-05, | |
| "loss": 0.3778, | |
| "step": 942 | |
| }, | |
| { | |
| "epoch": 0.42702289644241925, | |
| "grad_norm": 0.8578718900680542, | |
| "learning_rate": 1.4275486622305348e-05, | |
| "loss": 0.4567, | |
| "step": 943 | |
| }, | |
| { | |
| "epoch": 0.42747573090312174, | |
| "grad_norm": 0.775020956993103, | |
| "learning_rate": 1.427395827668204e-05, | |
| "loss": 0.4177, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 0.4279285653638242, | |
| "grad_norm": 0.7445123791694641, | |
| "learning_rate": 1.427242840274022e-05, | |
| "loss": 0.4109, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.42838139982452667, | |
| "grad_norm": 0.660199761390686, | |
| "learning_rate": 1.4270897000825052e-05, | |
| "loss": 0.3661, | |
| "step": 946 | |
| }, | |
| { | |
| "epoch": 0.4288342342852291, | |
| "grad_norm": 0.960015594959259, | |
| "learning_rate": 1.426936407128205e-05, | |
| "loss": 0.4006, | |
| "step": 947 | |
| }, | |
| { | |
| "epoch": 0.4292870687459316, | |
| "grad_norm": 0.7209600806236267, | |
| "learning_rate": 1.4267829614457067e-05, | |
| "loss": 0.4041, | |
| "step": 948 | |
| }, | |
| { | |
| "epoch": 0.429739903206634, | |
| "grad_norm": 0.7578160762786865, | |
| "learning_rate": 1.42662936306963e-05, | |
| "loss": 0.3883, | |
| "step": 949 | |
| }, | |
| { | |
| "epoch": 0.4301927376673365, | |
| "grad_norm": 0.7112842798233032, | |
| "learning_rate": 1.4264756120346294e-05, | |
| "loss": 0.4147, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.43064557212803894, | |
| "grad_norm": 0.8537875413894653, | |
| "learning_rate": 1.4263217083753938e-05, | |
| "loss": 0.4498, | |
| "step": 951 | |
| }, | |
| { | |
| "epoch": 0.4310984065887414, | |
| "grad_norm": 0.6418316960334778, | |
| "learning_rate": 1.4261676521266466e-05, | |
| "loss": 0.3858, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 0.43155124104944387, | |
| "grad_norm": 0.8194690346717834, | |
| "learning_rate": 1.4260134433231449e-05, | |
| "loss": 0.3749, | |
| "step": 953 | |
| }, | |
| { | |
| "epoch": 0.4320040755101463, | |
| "grad_norm": 1.0442348718643188, | |
| "learning_rate": 1.4258590819996813e-05, | |
| "loss": 0.3945, | |
| "step": 954 | |
| }, | |
| { | |
| "epoch": 0.4324569099708488, | |
| "grad_norm": 0.778914213180542, | |
| "learning_rate": 1.4257045681910822e-05, | |
| "loss": 0.4199, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.4329097444315512, | |
| "grad_norm": 0.7745048999786377, | |
| "learning_rate": 1.4255499019322084e-05, | |
| "loss": 0.4514, | |
| "step": 956 | |
| }, | |
| { | |
| "epoch": 0.4333625788922537, | |
| "grad_norm": 0.7798154354095459, | |
| "learning_rate": 1.4253950832579552e-05, | |
| "loss": 0.4377, | |
| "step": 957 | |
| }, | |
| { | |
| "epoch": 0.43381541335295615, | |
| "grad_norm": 0.7294392585754395, | |
| "learning_rate": 1.4252401122032525e-05, | |
| "loss": 0.3638, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 0.43426824781365864, | |
| "grad_norm": 0.795789361000061, | |
| "learning_rate": 1.4250849888030641e-05, | |
| "loss": 0.4311, | |
| "step": 959 | |
| }, | |
| { | |
| "epoch": 0.43472108227436107, | |
| "grad_norm": 1.5630682706832886, | |
| "learning_rate": 1.4249297130923885e-05, | |
| "loss": 0.396, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.43517391673506356, | |
| "grad_norm": 0.7337138056755066, | |
| "learning_rate": 1.4247742851062587e-05, | |
| "loss": 0.4561, | |
| "step": 961 | |
| }, | |
| { | |
| "epoch": 0.435626751195766, | |
| "grad_norm": 0.7088420987129211, | |
| "learning_rate": 1.4246187048797414e-05, | |
| "loss": 0.4247, | |
| "step": 962 | |
| }, | |
| { | |
| "epoch": 0.4360795856564685, | |
| "grad_norm": 0.9262889623641968, | |
| "learning_rate": 1.4244629724479387e-05, | |
| "loss": 0.4006, | |
| "step": 963 | |
| }, | |
| { | |
| "epoch": 0.4365324201171709, | |
| "grad_norm": 0.8591921329498291, | |
| "learning_rate": 1.4243070878459862e-05, | |
| "loss": 0.4003, | |
| "step": 964 | |
| }, | |
| { | |
| "epoch": 0.43698525457787335, | |
| "grad_norm": 1.0036776065826416, | |
| "learning_rate": 1.4241510511090539e-05, | |
| "loss": 0.4098, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.43743808903857584, | |
| "grad_norm": 0.6864471435546875, | |
| "learning_rate": 1.4239948622723465e-05, | |
| "loss": 0.4326, | |
| "step": 966 | |
| }, | |
| { | |
| "epoch": 0.4378909234992783, | |
| "grad_norm": 1.0244603157043457, | |
| "learning_rate": 1.4238385213711027e-05, | |
| "loss": 0.4419, | |
| "step": 967 | |
| }, | |
| { | |
| "epoch": 0.43834375795998076, | |
| "grad_norm": 0.7093055844306946, | |
| "learning_rate": 1.4236820284405956e-05, | |
| "loss": 0.3951, | |
| "step": 968 | |
| }, | |
| { | |
| "epoch": 0.4387965924206832, | |
| "grad_norm": 0.8618746995925903, | |
| "learning_rate": 1.4235253835161329e-05, | |
| "loss": 0.4756, | |
| "step": 969 | |
| }, | |
| { | |
| "epoch": 0.4392494268813857, | |
| "grad_norm": 0.7317725419998169, | |
| "learning_rate": 1.4233685866330561e-05, | |
| "loss": 0.3789, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.4397022613420881, | |
| "grad_norm": 0.8535686731338501, | |
| "learning_rate": 1.4232116378267412e-05, | |
| "loss": 0.4096, | |
| "step": 971 | |
| }, | |
| { | |
| "epoch": 0.4401550958027906, | |
| "grad_norm": 0.7266016006469727, | |
| "learning_rate": 1.4230545371325985e-05, | |
| "loss": 0.3706, | |
| "step": 972 | |
| }, | |
| { | |
| "epoch": 0.44060793026349304, | |
| "grad_norm": 0.7952896356582642, | |
| "learning_rate": 1.4228972845860725e-05, | |
| "loss": 0.3392, | |
| "step": 973 | |
| }, | |
| { | |
| "epoch": 0.44106076472419553, | |
| "grad_norm": 0.8086087107658386, | |
| "learning_rate": 1.422739880222642e-05, | |
| "loss": 0.3972, | |
| "step": 974 | |
| }, | |
| { | |
| "epoch": 0.44151359918489796, | |
| "grad_norm": 0.8605599403381348, | |
| "learning_rate": 1.4225823240778207e-05, | |
| "loss": 0.422, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.44196643364560045, | |
| "grad_norm": 0.752622663974762, | |
| "learning_rate": 1.4224246161871548e-05, | |
| "loss": 0.3965, | |
| "step": 976 | |
| }, | |
| { | |
| "epoch": 0.4424192681063029, | |
| "grad_norm": 0.8178388476371765, | |
| "learning_rate": 1.4222667565862264e-05, | |
| "loss": 0.4438, | |
| "step": 977 | |
| }, | |
| { | |
| "epoch": 0.4428721025670054, | |
| "grad_norm": 0.7991887927055359, | |
| "learning_rate": 1.4221087453106515e-05, | |
| "loss": 0.3748, | |
| "step": 978 | |
| }, | |
| { | |
| "epoch": 0.4433249370277078, | |
| "grad_norm": 0.7863441705703735, | |
| "learning_rate": 1.4219505823960798e-05, | |
| "loss": 0.3352, | |
| "step": 979 | |
| }, | |
| { | |
| "epoch": 0.44377777148841024, | |
| "grad_norm": 0.7784522175788879, | |
| "learning_rate": 1.4217922678781955e-05, | |
| "loss": 0.4319, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.44423060594911273, | |
| "grad_norm": 0.8807750344276428, | |
| "learning_rate": 1.4216338017927171e-05, | |
| "loss": 0.4165, | |
| "step": 981 | |
| }, | |
| { | |
| "epoch": 0.44468344040981517, | |
| "grad_norm": 0.6990790367126465, | |
| "learning_rate": 1.4214751841753971e-05, | |
| "loss": 0.4141, | |
| "step": 982 | |
| }, | |
| { | |
| "epoch": 0.44513627487051766, | |
| "grad_norm": 0.9423395991325378, | |
| "learning_rate": 1.4213164150620227e-05, | |
| "loss": 0.3669, | |
| "step": 983 | |
| }, | |
| { | |
| "epoch": 0.4455891093312201, | |
| "grad_norm": 0.7699115872383118, | |
| "learning_rate": 1.4211574944884143e-05, | |
| "loss": 0.4115, | |
| "step": 984 | |
| }, | |
| { | |
| "epoch": 0.4460419437919226, | |
| "grad_norm": 0.8870624899864197, | |
| "learning_rate": 1.4209984224904272e-05, | |
| "loss": 0.3953, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.446494778252625, | |
| "grad_norm": 0.8935746550559998, | |
| "learning_rate": 1.4208391991039509e-05, | |
| "loss": 0.4049, | |
| "step": 986 | |
| }, | |
| { | |
| "epoch": 0.4469476127133275, | |
| "grad_norm": 0.647849440574646, | |
| "learning_rate": 1.4206798243649085e-05, | |
| "loss": 0.3873, | |
| "step": 987 | |
| }, | |
| { | |
| "epoch": 0.44740044717402994, | |
| "grad_norm": 0.7980324625968933, | |
| "learning_rate": 1.4205202983092583e-05, | |
| "loss": 0.4253, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 0.4478532816347324, | |
| "grad_norm": 0.6993192434310913, | |
| "learning_rate": 1.4203606209729912e-05, | |
| "loss": 0.409, | |
| "step": 989 | |
| }, | |
| { | |
| "epoch": 0.44830611609543486, | |
| "grad_norm": 0.7574604153633118, | |
| "learning_rate": 1.4202007923921338e-05, | |
| "loss": 0.345, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.44875895055613735, | |
| "grad_norm": 0.8472474217414856, | |
| "learning_rate": 1.4200408126027458e-05, | |
| "loss": 0.3847, | |
| "step": 991 | |
| }, | |
| { | |
| "epoch": 0.4492117850168398, | |
| "grad_norm": 0.7414373159408569, | |
| "learning_rate": 1.4198806816409211e-05, | |
| "loss": 0.3923, | |
| "step": 992 | |
| }, | |
| { | |
| "epoch": 0.4496646194775422, | |
| "grad_norm": 0.6760424971580505, | |
| "learning_rate": 1.4197203995427883e-05, | |
| "loss": 0.3471, | |
| "step": 993 | |
| }, | |
| { | |
| "epoch": 0.4501174539382447, | |
| "grad_norm": 0.6407291889190674, | |
| "learning_rate": 1.4195599663445096e-05, | |
| "loss": 0.4312, | |
| "step": 994 | |
| }, | |
| { | |
| "epoch": 0.45057028839894714, | |
| "grad_norm": 0.6934922337532043, | |
| "learning_rate": 1.4193993820822815e-05, | |
| "loss": 0.4067, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.45102312285964963, | |
| "grad_norm": 0.7909060716629028, | |
| "learning_rate": 1.4192386467923343e-05, | |
| "loss": 0.4303, | |
| "step": 996 | |
| }, | |
| { | |
| "epoch": 0.45147595732035206, | |
| "grad_norm": 0.7516681551933289, | |
| "learning_rate": 1.4190777605109326e-05, | |
| "loss": 0.4274, | |
| "step": 997 | |
| }, | |
| { | |
| "epoch": 0.45192879178105455, | |
| "grad_norm": 0.7599420547485352, | |
| "learning_rate": 1.4189167232743754e-05, | |
| "loss": 0.3949, | |
| "step": 998 | |
| }, | |
| { | |
| "epoch": 0.452381626241757, | |
| "grad_norm": 0.7314419150352478, | |
| "learning_rate": 1.4187555351189948e-05, | |
| "loss": 0.4176, | |
| "step": 999 | |
| }, | |
| { | |
| "epoch": 0.4528344607024595, | |
| "grad_norm": 0.6400272846221924, | |
| "learning_rate": 1.418594196081158e-05, | |
| "loss": 0.3949, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.4532872951631619, | |
| "grad_norm": 0.8566839098930359, | |
| "learning_rate": 1.4184327061972657e-05, | |
| "loss": 0.4034, | |
| "step": 1001 | |
| }, | |
| { | |
| "epoch": 0.4537401296238644, | |
| "grad_norm": 0.7478398084640503, | |
| "learning_rate": 1.4182710655037527e-05, | |
| "loss": 0.4593, | |
| "step": 1002 | |
| }, | |
| { | |
| "epoch": 0.45419296408456683, | |
| "grad_norm": 0.8677337169647217, | |
| "learning_rate": 1.4181092740370878e-05, | |
| "loss": 0.432, | |
| "step": 1003 | |
| }, | |
| { | |
| "epoch": 0.4546457985452693, | |
| "grad_norm": 0.7963004112243652, | |
| "learning_rate": 1.4179473318337738e-05, | |
| "loss": 0.3899, | |
| "step": 1004 | |
| }, | |
| { | |
| "epoch": 0.45509863300597175, | |
| "grad_norm": 0.7104138135910034, | |
| "learning_rate": 1.4177852389303479e-05, | |
| "loss": 0.3823, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.4555514674666742, | |
| "grad_norm": 0.6873833537101746, | |
| "learning_rate": 1.4176229953633806e-05, | |
| "loss": 0.3799, | |
| "step": 1006 | |
| }, | |
| { | |
| "epoch": 0.4560043019273767, | |
| "grad_norm": 0.8924839496612549, | |
| "learning_rate": 1.4174606011694767e-05, | |
| "loss": 0.4167, | |
| "step": 1007 | |
| }, | |
| { | |
| "epoch": 0.4564571363880791, | |
| "grad_norm": 0.8046056032180786, | |
| "learning_rate": 1.4172980563852757e-05, | |
| "loss": 0.3823, | |
| "step": 1008 | |
| }, | |
| { | |
| "epoch": 0.4569099708487816, | |
| "grad_norm": 0.7857614159584045, | |
| "learning_rate": 1.4171353610474496e-05, | |
| "loss": 0.4105, | |
| "step": 1009 | |
| }, | |
| { | |
| "epoch": 0.45736280530948403, | |
| "grad_norm": 0.715185284614563, | |
| "learning_rate": 1.416972515192706e-05, | |
| "loss": 0.3557, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.4578156397701865, | |
| "grad_norm": 0.7968368530273438, | |
| "learning_rate": 1.416809518857785e-05, | |
| "loss": 0.3438, | |
| "step": 1011 | |
| }, | |
| { | |
| "epoch": 0.45826847423088896, | |
| "grad_norm": 0.9046311974525452, | |
| "learning_rate": 1.4166463720794615e-05, | |
| "loss": 0.4533, | |
| "step": 1012 | |
| }, | |
| { | |
| "epoch": 0.45872130869159145, | |
| "grad_norm": 0.6228207349777222, | |
| "learning_rate": 1.4164830748945445e-05, | |
| "loss": 0.3698, | |
| "step": 1013 | |
| }, | |
| { | |
| "epoch": 0.4591741431522939, | |
| "grad_norm": 0.8371472954750061, | |
| "learning_rate": 1.416319627339876e-05, | |
| "loss": 0.4385, | |
| "step": 1014 | |
| }, | |
| { | |
| "epoch": 0.45962697761299637, | |
| "grad_norm": 0.6872038841247559, | |
| "learning_rate": 1.416156029452333e-05, | |
| "loss": 0.3892, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 0.4600798120736988, | |
| "grad_norm": 0.7857869863510132, | |
| "learning_rate": 1.4159922812688255e-05, | |
| "loss": 0.4179, | |
| "step": 1016 | |
| }, | |
| { | |
| "epoch": 0.4605326465344013, | |
| "grad_norm": 0.8071677684783936, | |
| "learning_rate": 1.4158283828262982e-05, | |
| "loss": 0.4028, | |
| "step": 1017 | |
| }, | |
| { | |
| "epoch": 0.4609854809951037, | |
| "grad_norm": 0.7820652723312378, | |
| "learning_rate": 1.415664334161729e-05, | |
| "loss": 0.379, | |
| "step": 1018 | |
| }, | |
| { | |
| "epoch": 0.4614383154558062, | |
| "grad_norm": 0.7160912752151489, | |
| "learning_rate": 1.4155001353121304e-05, | |
| "loss": 0.393, | |
| "step": 1019 | |
| }, | |
| { | |
| "epoch": 0.46189114991650865, | |
| "grad_norm": 0.7849195003509521, | |
| "learning_rate": 1.415335786314548e-05, | |
| "loss": 0.4273, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.4623439843772111, | |
| "grad_norm": 0.7774779200553894, | |
| "learning_rate": 1.4151712872060623e-05, | |
| "loss": 0.356, | |
| "step": 1021 | |
| }, | |
| { | |
| "epoch": 0.46279681883791357, | |
| "grad_norm": 0.7844209671020508, | |
| "learning_rate": 1.4150066380237864e-05, | |
| "loss": 0.4261, | |
| "step": 1022 | |
| }, | |
| { | |
| "epoch": 0.463249653298616, | |
| "grad_norm": 0.840189516544342, | |
| "learning_rate": 1.4148418388048683e-05, | |
| "loss": 0.3801, | |
| "step": 1023 | |
| }, | |
| { | |
| "epoch": 0.4637024877593185, | |
| "grad_norm": 0.8453295230865479, | |
| "learning_rate": 1.4146768895864892e-05, | |
| "loss": 0.4049, | |
| "step": 1024 | |
| }, | |
| { | |
| "epoch": 0.4641553222200209, | |
| "grad_norm": 0.7747737169265747, | |
| "learning_rate": 1.4145117904058645e-05, | |
| "loss": 0.4002, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.4646081566807234, | |
| "grad_norm": 0.7298173904418945, | |
| "learning_rate": 1.4143465413002435e-05, | |
| "loss": 0.3816, | |
| "step": 1026 | |
| }, | |
| { | |
| "epoch": 0.46506099114142585, | |
| "grad_norm": 0.8087685704231262, | |
| "learning_rate": 1.4141811423069092e-05, | |
| "loss": 0.3819, | |
| "step": 1027 | |
| }, | |
| { | |
| "epoch": 0.46551382560212834, | |
| "grad_norm": 0.6976639032363892, | |
| "learning_rate": 1.4140155934631783e-05, | |
| "loss": 0.3915, | |
| "step": 1028 | |
| }, | |
| { | |
| "epoch": 0.4659666600628308, | |
| "grad_norm": 0.6582549214363098, | |
| "learning_rate": 1.413849894806401e-05, | |
| "loss": 0.3571, | |
| "step": 1029 | |
| }, | |
| { | |
| "epoch": 0.46641949452353326, | |
| "grad_norm": 0.7097831964492798, | |
| "learning_rate": 1.4136840463739623e-05, | |
| "loss": 0.4004, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.4668723289842357, | |
| "grad_norm": 0.8044737577438354, | |
| "learning_rate": 1.4135180482032801e-05, | |
| "loss": 0.4087, | |
| "step": 1031 | |
| }, | |
| { | |
| "epoch": 0.4673251634449382, | |
| "grad_norm": 0.7350354194641113, | |
| "learning_rate": 1.4133519003318063e-05, | |
| "loss": 0.4101, | |
| "step": 1032 | |
| }, | |
| { | |
| "epoch": 0.4677779979056406, | |
| "grad_norm": 0.7550399899482727, | |
| "learning_rate": 1.4131856027970269e-05, | |
| "loss": 0.3657, | |
| "step": 1033 | |
| }, | |
| { | |
| "epoch": 0.46823083236634305, | |
| "grad_norm": 0.65655118227005, | |
| "learning_rate": 1.413019155636461e-05, | |
| "loss": 0.373, | |
| "step": 1034 | |
| }, | |
| { | |
| "epoch": 0.46868366682704554, | |
| "grad_norm": 0.7206447124481201, | |
| "learning_rate": 1.4128525588876624e-05, | |
| "loss": 0.353, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.469136501287748, | |
| "grad_norm": 0.7007904052734375, | |
| "learning_rate": 1.4126858125882176e-05, | |
| "loss": 0.3668, | |
| "step": 1036 | |
| }, | |
| { | |
| "epoch": 0.46958933574845046, | |
| "grad_norm": 0.9158049821853638, | |
| "learning_rate": 1.4125189167757477e-05, | |
| "loss": 0.3746, | |
| "step": 1037 | |
| }, | |
| { | |
| "epoch": 0.4700421702091529, | |
| "grad_norm": 0.8141727447509766, | |
| "learning_rate": 1.412351871487907e-05, | |
| "loss": 0.3728, | |
| "step": 1038 | |
| }, | |
| { | |
| "epoch": 0.4704950046698554, | |
| "grad_norm": 0.696324348449707, | |
| "learning_rate": 1.4121846767623839e-05, | |
| "loss": 0.3415, | |
| "step": 1039 | |
| }, | |
| { | |
| "epoch": 0.4709478391305578, | |
| "grad_norm": 0.8677844405174255, | |
| "learning_rate": 1.4120173326369003e-05, | |
| "loss": 0.3584, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.4714006735912603, | |
| "grad_norm": 0.8021166324615479, | |
| "learning_rate": 1.4118498391492118e-05, | |
| "loss": 0.3832, | |
| "step": 1041 | |
| }, | |
| { | |
| "epoch": 0.47185350805196274, | |
| "grad_norm": 0.9213194847106934, | |
| "learning_rate": 1.4116821963371075e-05, | |
| "loss": 0.4097, | |
| "step": 1042 | |
| }, | |
| { | |
| "epoch": 0.47230634251266523, | |
| "grad_norm": 0.8504592776298523, | |
| "learning_rate": 1.4115144042384107e-05, | |
| "loss": 0.4031, | |
| "step": 1043 | |
| }, | |
| { | |
| "epoch": 0.47275917697336767, | |
| "grad_norm": 0.746096134185791, | |
| "learning_rate": 1.411346462890978e-05, | |
| "loss": 0.4064, | |
| "step": 1044 | |
| }, | |
| { | |
| "epoch": 0.47321201143407016, | |
| "grad_norm": 0.9829161167144775, | |
| "learning_rate": 1.4111783723327e-05, | |
| "loss": 0.409, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 0.4736648458947726, | |
| "grad_norm": 0.7933005690574646, | |
| "learning_rate": 1.4110101326015007e-05, | |
| "loss": 0.4209, | |
| "step": 1046 | |
| }, | |
| { | |
| "epoch": 0.474117680355475, | |
| "grad_norm": 0.8225368857383728, | |
| "learning_rate": 1.4108417437353374e-05, | |
| "loss": 0.3688, | |
| "step": 1047 | |
| }, | |
| { | |
| "epoch": 0.4745705148161775, | |
| "grad_norm": 0.6937693357467651, | |
| "learning_rate": 1.4106732057722018e-05, | |
| "loss": 0.4077, | |
| "step": 1048 | |
| }, | |
| { | |
| "epoch": 0.47502334927687995, | |
| "grad_norm": 0.7931678891181946, | |
| "learning_rate": 1.410504518750119e-05, | |
| "loss": 0.3273, | |
| "step": 1049 | |
| }, | |
| { | |
| "epoch": 0.47547618373758244, | |
| "grad_norm": 0.7295666933059692, | |
| "learning_rate": 1.4103356827071473e-05, | |
| "loss": 0.3798, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.47592901819828487, | |
| "grad_norm": 0.7447585463523865, | |
| "learning_rate": 1.410166697681379e-05, | |
| "loss": 0.3923, | |
| "step": 1051 | |
| }, | |
| { | |
| "epoch": 0.47638185265898736, | |
| "grad_norm": 0.732587993144989, | |
| "learning_rate": 1.4099975637109401e-05, | |
| "loss": 0.3618, | |
| "step": 1052 | |
| }, | |
| { | |
| "epoch": 0.4768346871196898, | |
| "grad_norm": 0.6432980895042419, | |
| "learning_rate": 1.4098282808339899e-05, | |
| "loss": 0.3782, | |
| "step": 1053 | |
| }, | |
| { | |
| "epoch": 0.4772875215803923, | |
| "grad_norm": 0.9398018717765808, | |
| "learning_rate": 1.4096588490887218e-05, | |
| "loss": 0.4002, | |
| "step": 1054 | |
| }, | |
| { | |
| "epoch": 0.4777403560410947, | |
| "grad_norm": 0.845302164554596, | |
| "learning_rate": 1.409489268513362e-05, | |
| "loss": 0.3981, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 0.4781931905017972, | |
| "grad_norm": 0.7976647615432739, | |
| "learning_rate": 1.409319539146171e-05, | |
| "loss": 0.3828, | |
| "step": 1056 | |
| }, | |
| { | |
| "epoch": 0.47864602496249964, | |
| "grad_norm": 0.7224624156951904, | |
| "learning_rate": 1.4091496610254426e-05, | |
| "loss": 0.4466, | |
| "step": 1057 | |
| }, | |
| { | |
| "epoch": 0.47909885942320213, | |
| "grad_norm": 0.6982882618904114, | |
| "learning_rate": 1.4089796341895036e-05, | |
| "loss": 0.3732, | |
| "step": 1058 | |
| }, | |
| { | |
| "epoch": 0.47955169388390456, | |
| "grad_norm": 0.6954671144485474, | |
| "learning_rate": 1.408809458676716e-05, | |
| "loss": 0.3646, | |
| "step": 1059 | |
| }, | |
| { | |
| "epoch": 0.48000452834460705, | |
| "grad_norm": 0.6358652710914612, | |
| "learning_rate": 1.4086391345254732e-05, | |
| "loss": 0.382, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.4804573628053095, | |
| "grad_norm": 0.7169029116630554, | |
| "learning_rate": 1.4084686617742037e-05, | |
| "loss": 0.384, | |
| "step": 1061 | |
| }, | |
| { | |
| "epoch": 0.4809101972660119, | |
| "grad_norm": 0.7383198738098145, | |
| "learning_rate": 1.4082980404613687e-05, | |
| "loss": 0.3487, | |
| "step": 1062 | |
| }, | |
| { | |
| "epoch": 0.4813630317267144, | |
| "grad_norm": 0.6351798176765442, | |
| "learning_rate": 1.4081272706254637e-05, | |
| "loss": 0.3988, | |
| "step": 1063 | |
| }, | |
| { | |
| "epoch": 0.48181586618741684, | |
| "grad_norm": 0.7320924401283264, | |
| "learning_rate": 1.4079563523050169e-05, | |
| "loss": 0.3706, | |
| "step": 1064 | |
| }, | |
| { | |
| "epoch": 0.48226870064811933, | |
| "grad_norm": 0.7490411996841431, | |
| "learning_rate": 1.4077852855385902e-05, | |
| "loss": 0.3777, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.48272153510882176, | |
| "grad_norm": 0.8109067678451538, | |
| "learning_rate": 1.4076140703647794e-05, | |
| "loss": 0.391, | |
| "step": 1066 | |
| }, | |
| { | |
| "epoch": 0.48317436956952425, | |
| "grad_norm": 0.7815183401107788, | |
| "learning_rate": 1.4074427068222135e-05, | |
| "loss": 0.3854, | |
| "step": 1067 | |
| }, | |
| { | |
| "epoch": 0.4836272040302267, | |
| "grad_norm": 0.8142592906951904, | |
| "learning_rate": 1.407271194949555e-05, | |
| "loss": 0.3907, | |
| "step": 1068 | |
| }, | |
| { | |
| "epoch": 0.4840800384909292, | |
| "grad_norm": 0.7322883009910583, | |
| "learning_rate": 1.4070995347854997e-05, | |
| "loss": 0.3879, | |
| "step": 1069 | |
| }, | |
| { | |
| "epoch": 0.4845328729516316, | |
| "grad_norm": 0.7850584983825684, | |
| "learning_rate": 1.4069277263687772e-05, | |
| "loss": 0.3516, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.4849857074123341, | |
| "grad_norm": 0.6609899401664734, | |
| "learning_rate": 1.40675576973815e-05, | |
| "loss": 0.3677, | |
| "step": 1071 | |
| }, | |
| { | |
| "epoch": 0.48543854187303653, | |
| "grad_norm": 0.6708543300628662, | |
| "learning_rate": 1.4065836649324148e-05, | |
| "loss": 0.3708, | |
| "step": 1072 | |
| }, | |
| { | |
| "epoch": 0.485891376333739, | |
| "grad_norm": 1.1243339776992798, | |
| "learning_rate": 1.4064114119904012e-05, | |
| "loss": 0.4169, | |
| "step": 1073 | |
| }, | |
| { | |
| "epoch": 0.48634421079444146, | |
| "grad_norm": 0.7623947858810425, | |
| "learning_rate": 1.4062390109509722e-05, | |
| "loss": 0.3518, | |
| "step": 1074 | |
| }, | |
| { | |
| "epoch": 0.4867970452551439, | |
| "grad_norm": 0.7470114827156067, | |
| "learning_rate": 1.4060664618530247e-05, | |
| "loss": 0.3892, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.4872498797158464, | |
| "grad_norm": 0.8398137092590332, | |
| "learning_rate": 1.4058937647354882e-05, | |
| "loss": 0.4376, | |
| "step": 1076 | |
| }, | |
| { | |
| "epoch": 0.4877027141765488, | |
| "grad_norm": 0.7564401626586914, | |
| "learning_rate": 1.4057209196373265e-05, | |
| "loss": 0.3588, | |
| "step": 1077 | |
| }, | |
| { | |
| "epoch": 0.4881555486372513, | |
| "grad_norm": 0.7910377383232117, | |
| "learning_rate": 1.4055479265975361e-05, | |
| "loss": 0.3916, | |
| "step": 1078 | |
| }, | |
| { | |
| "epoch": 0.48860838309795374, | |
| "grad_norm": 0.720329999923706, | |
| "learning_rate": 1.405374785655147e-05, | |
| "loss": 0.3732, | |
| "step": 1079 | |
| }, | |
| { | |
| "epoch": 0.4890612175586562, | |
| "grad_norm": 0.7369146347045898, | |
| "learning_rate": 1.4052014968492232e-05, | |
| "loss": 0.397, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.48951405201935866, | |
| "grad_norm": 0.6554039120674133, | |
| "learning_rate": 1.4050280602188613e-05, | |
| "loss": 0.3219, | |
| "step": 1081 | |
| }, | |
| { | |
| "epoch": 0.48996688648006115, | |
| "grad_norm": 0.7499859929084778, | |
| "learning_rate": 1.4048544758031912e-05, | |
| "loss": 0.3816, | |
| "step": 1082 | |
| }, | |
| { | |
| "epoch": 0.4904197209407636, | |
| "grad_norm": 0.855495810508728, | |
| "learning_rate": 1.4046807436413768e-05, | |
| "loss": 0.3978, | |
| "step": 1083 | |
| }, | |
| { | |
| "epoch": 0.49087255540146607, | |
| "grad_norm": 0.6773454546928406, | |
| "learning_rate": 1.404506863772615e-05, | |
| "loss": 0.389, | |
| "step": 1084 | |
| }, | |
| { | |
| "epoch": 0.4913253898621685, | |
| "grad_norm": 0.7844047546386719, | |
| "learning_rate": 1.4043328362361359e-05, | |
| "loss": 0.3594, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 0.491778224322871, | |
| "grad_norm": 0.7174739241600037, | |
| "learning_rate": 1.404158661071203e-05, | |
| "loss": 0.3667, | |
| "step": 1086 | |
| }, | |
| { | |
| "epoch": 0.4922310587835734, | |
| "grad_norm": 0.6895620226860046, | |
| "learning_rate": 1.4039843383171132e-05, | |
| "loss": 0.4006, | |
| "step": 1087 | |
| }, | |
| { | |
| "epoch": 0.4926838932442759, | |
| "grad_norm": 0.6506296992301941, | |
| "learning_rate": 1.4038098680131966e-05, | |
| "loss": 0.3277, | |
| "step": 1088 | |
| }, | |
| { | |
| "epoch": 0.49313672770497835, | |
| "grad_norm": 0.7601318359375, | |
| "learning_rate": 1.4036352501988168e-05, | |
| "loss": 0.3736, | |
| "step": 1089 | |
| }, | |
| { | |
| "epoch": 0.4935895621656808, | |
| "grad_norm": 0.6839517951011658, | |
| "learning_rate": 1.4034604849133703e-05, | |
| "loss": 0.3365, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.4940423966263833, | |
| "grad_norm": 0.6496476531028748, | |
| "learning_rate": 1.4032855721962869e-05, | |
| "loss": 0.3697, | |
| "step": 1091 | |
| }, | |
| { | |
| "epoch": 0.4944952310870857, | |
| "grad_norm": 0.7376840710639954, | |
| "learning_rate": 1.4031105120870303e-05, | |
| "loss": 0.4017, | |
| "step": 1092 | |
| }, | |
| { | |
| "epoch": 0.4949480655477882, | |
| "grad_norm": 0.7958086133003235, | |
| "learning_rate": 1.4029353046250965e-05, | |
| "loss": 0.3657, | |
| "step": 1093 | |
| }, | |
| { | |
| "epoch": 0.49540090000849063, | |
| "grad_norm": 0.6121701002120972, | |
| "learning_rate": 1.4027599498500157e-05, | |
| "loss": 0.3189, | |
| "step": 1094 | |
| }, | |
| { | |
| "epoch": 0.4958537344691931, | |
| "grad_norm": 0.7224218249320984, | |
| "learning_rate": 1.4025844478013505e-05, | |
| "loss": 0.3443, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.49630656892989555, | |
| "grad_norm": 0.8465177416801453, | |
| "learning_rate": 1.4024087985186975e-05, | |
| "loss": 0.3789, | |
| "step": 1096 | |
| }, | |
| { | |
| "epoch": 0.49675940339059804, | |
| "grad_norm": 0.6270837783813477, | |
| "learning_rate": 1.4022330020416858e-05, | |
| "loss": 0.3388, | |
| "step": 1097 | |
| }, | |
| { | |
| "epoch": 0.4972122378513005, | |
| "grad_norm": 0.6935149431228638, | |
| "learning_rate": 1.4020570584099782e-05, | |
| "loss": 0.336, | |
| "step": 1098 | |
| }, | |
| { | |
| "epoch": 0.49766507231200297, | |
| "grad_norm": 0.7307567596435547, | |
| "learning_rate": 1.4018809676632703e-05, | |
| "loss": 0.4503, | |
| "step": 1099 | |
| }, | |
| { | |
| "epoch": 0.4981179067727054, | |
| "grad_norm": 0.7322489023208618, | |
| "learning_rate": 1.4017047298412914e-05, | |
| "loss": 0.3521, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.4985707412334079, | |
| "grad_norm": 0.7978783249855042, | |
| "learning_rate": 1.4015283449838037e-05, | |
| "loss": 0.3992, | |
| "step": 1101 | |
| }, | |
| { | |
| "epoch": 0.4990235756941103, | |
| "grad_norm": 0.6273020505905151, | |
| "learning_rate": 1.4013518131306024e-05, | |
| "loss": 0.3647, | |
| "step": 1102 | |
| }, | |
| { | |
| "epoch": 0.49947641015481276, | |
| "grad_norm": 0.6884061098098755, | |
| "learning_rate": 1.401175134321516e-05, | |
| "loss": 0.3903, | |
| "step": 1103 | |
| }, | |
| { | |
| "epoch": 0.49992924461551524, | |
| "grad_norm": 0.7667953968048096, | |
| "learning_rate": 1.4009983085964064e-05, | |
| "loss": 0.3562, | |
| "step": 1104 | |
| }, | |
| { | |
| "epoch": 0.49992924461551524, | |
| "eval_loss": 0.3898637592792511, | |
| "eval_runtime": 19.4358, | |
| "eval_samples_per_second": 5.865, | |
| "eval_steps_per_second": 0.772, | |
| "step": 1104 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 6624, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 552, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 9.82217497538577e+19, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |