| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 12.0, | |
| "eval_steps": 500, | |
| "global_step": 5892, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.05091649694501019, | |
| "grad_norm": 2.806674003601074, | |
| "learning_rate": 0.0002, | |
| "loss": 1.8959, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.10183299389002037, | |
| "grad_norm": 3.7694294452667236, | |
| "learning_rate": 0.0002, | |
| "loss": 1.5457, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.15274949083503056, | |
| "grad_norm": 2.5580432415008545, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2675, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.20366598778004075, | |
| "grad_norm": 3.3819141387939453, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1332, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.2545824847250509, | |
| "grad_norm": 2.542970895767212, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1962, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.3054989816700611, | |
| "grad_norm": 3.139821767807007, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0642, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.3564154786150713, | |
| "grad_norm": 2.8020615577697754, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1643, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.4073319755600815, | |
| "grad_norm": 2.742553234100342, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0901, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.45824847250509165, | |
| "grad_norm": 2.8044979572296143, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1531, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.5091649694501018, | |
| "grad_norm": 2.295903444290161, | |
| "learning_rate": 0.0002, | |
| "loss": 0.981, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.560081466395112, | |
| "grad_norm": 2.5899577140808105, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1367, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.6109979633401222, | |
| "grad_norm": 3.2374327182769775, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0535, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.6619144602851323, | |
| "grad_norm": 2.467207193374634, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1175, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.7128309572301426, | |
| "grad_norm": 2.5034849643707275, | |
| "learning_rate": 0.0002, | |
| "loss": 1.078, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.7637474541751528, | |
| "grad_norm": 2.7688446044921875, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0662, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.814663951120163, | |
| "grad_norm": 2.9710495471954346, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0392, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.8655804480651731, | |
| "grad_norm": 2.105386734008789, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0551, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.9164969450101833, | |
| "grad_norm": 2.467886209487915, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0188, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.9674134419551935, | |
| "grad_norm": 2.3255183696746826, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0985, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 1.0183299389002036, | |
| "grad_norm": 1.9878566265106201, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9582, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.0692464358452138, | |
| "grad_norm": 2.2030067443847656, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8562, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 1.120162932790224, | |
| "grad_norm": 2.0550765991210938, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8564, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.1710794297352343, | |
| "grad_norm": 2.8419840335845947, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8446, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 1.2219959266802445, | |
| "grad_norm": 1.9293944835662842, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8239, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.2729124236252547, | |
| "grad_norm": 2.2307465076446533, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7875, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 1.3238289205702647, | |
| "grad_norm": 2.4380366802215576, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8686, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.374745417515275, | |
| "grad_norm": 2.5356967449188232, | |
| "learning_rate": 0.0002, | |
| "loss": 0.891, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 1.4256619144602851, | |
| "grad_norm": 1.957305669784546, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8923, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.4765784114052953, | |
| "grad_norm": 3.196012020111084, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8906, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 1.5274949083503055, | |
| "grad_norm": 2.050201654434204, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9081, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.5784114052953155, | |
| "grad_norm": 2.0173072814941406, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9078, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 1.629327902240326, | |
| "grad_norm": 2.157409906387329, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8954, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.680244399185336, | |
| "grad_norm": 3.089580535888672, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8553, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 1.7311608961303462, | |
| "grad_norm": 2.2709248065948486, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9085, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.7820773930753564, | |
| "grad_norm": 2.866403102874756, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8565, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 1.8329938900203666, | |
| "grad_norm": 2.305607795715332, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8674, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.8839103869653768, | |
| "grad_norm": 2.287306070327759, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9006, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 1.9348268839103868, | |
| "grad_norm": 2.0112550258636475, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8583, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 1.9857433808553973, | |
| "grad_norm": 2.2255215644836426, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8373, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 2.0366598778004072, | |
| "grad_norm": 2.1194534301757812, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7602, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 2.0875763747454177, | |
| "grad_norm": 1.874557614326477, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6402, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 2.1384928716904277, | |
| "grad_norm": 2.008828639984131, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6897, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 2.189409368635438, | |
| "grad_norm": 1.8525766134262085, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6611, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 2.240325865580448, | |
| "grad_norm": 2.184070348739624, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7043, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 2.291242362525458, | |
| "grad_norm": 1.9643105268478394, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6444, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 2.3421588594704685, | |
| "grad_norm": 2.2633492946624756, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6915, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 2.3930753564154785, | |
| "grad_norm": 2.3236186504364014, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6539, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 2.443991853360489, | |
| "grad_norm": 2.5817580223083496, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6922, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 2.494908350305499, | |
| "grad_norm": 2.4054062366485596, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6485, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 2.5458248472505094, | |
| "grad_norm": 2.730226516723633, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7163, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 2.5967413441955194, | |
| "grad_norm": 2.436521530151367, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6899, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 2.6476578411405294, | |
| "grad_norm": 2.4914846420288086, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7301, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 2.69857433808554, | |
| "grad_norm": 2.550816774368286, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6853, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 2.74949083503055, | |
| "grad_norm": 2.2780368328094482, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7165, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 2.8004073319755602, | |
| "grad_norm": 2.6065120697021484, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6841, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 2.8513238289205702, | |
| "grad_norm": 2.3689310550689697, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7647, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 2.9022403258655807, | |
| "grad_norm": 3.0321147441864014, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6813, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 2.9531568228105907, | |
| "grad_norm": 2.069146156311035, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7123, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 3.0040733197556007, | |
| "grad_norm": 2.1185224056243896, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7189, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 3.054989816700611, | |
| "grad_norm": 2.558671236038208, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5302, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 3.105906313645621, | |
| "grad_norm": 2.6581151485443115, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4756, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 3.1568228105906315, | |
| "grad_norm": 2.602369785308838, | |
| "learning_rate": 0.0002, | |
| "loss": 0.527, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 3.2077393075356415, | |
| "grad_norm": 2.8149938583374023, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5173, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 3.258655804480652, | |
| "grad_norm": 2.065443515777588, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5756, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 3.309572301425662, | |
| "grad_norm": 2.308039903640747, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5096, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 3.360488798370672, | |
| "grad_norm": 2.4268381595611572, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5538, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 3.4114052953156824, | |
| "grad_norm": 2.6498641967773438, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5189, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 3.4623217922606924, | |
| "grad_norm": 2.2553043365478516, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5786, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 3.513238289205703, | |
| "grad_norm": 2.8816471099853516, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5232, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 3.564154786150713, | |
| "grad_norm": 1.7681572437286377, | |
| "learning_rate": 0.0002, | |
| "loss": 0.569, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 3.6150712830957232, | |
| "grad_norm": 2.772834062576294, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5357, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 3.6659877800407332, | |
| "grad_norm": 2.4505577087402344, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5848, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 3.716904276985743, | |
| "grad_norm": 3.0437214374542236, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5571, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 3.7678207739307537, | |
| "grad_norm": 3.2281458377838135, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5956, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 3.8187372708757636, | |
| "grad_norm": 2.437544345855713, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5542, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 3.869653767820774, | |
| "grad_norm": 2.759650230407715, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5823, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 3.920570264765784, | |
| "grad_norm": 2.3260252475738525, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5724, | |
| "step": 1925 | |
| }, | |
| { | |
| "epoch": 3.9714867617107945, | |
| "grad_norm": 2.2468202114105225, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5884, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 4.022403258655804, | |
| "grad_norm": 2.221639394760132, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5266, | |
| "step": 1975 | |
| }, | |
| { | |
| "epoch": 4.0733197556008145, | |
| "grad_norm": 1.9102641344070435, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3866, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 4.124236252545825, | |
| "grad_norm": 2.9819774627685547, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4212, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 4.175152749490835, | |
| "grad_norm": 2.39497709274292, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4073, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 4.226069246435845, | |
| "grad_norm": 2.4051284790039062, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4328, | |
| "step": 2075 | |
| }, | |
| { | |
| "epoch": 4.276985743380855, | |
| "grad_norm": 2.677963972091675, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4294, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 4.327902240325866, | |
| "grad_norm": 2.484499216079712, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4386, | |
| "step": 2125 | |
| }, | |
| { | |
| "epoch": 4.378818737270876, | |
| "grad_norm": 3.8193187713623047, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4317, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 4.429735234215886, | |
| "grad_norm": 2.5229299068450928, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4597, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 4.480651731160896, | |
| "grad_norm": 2.6942062377929688, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4462, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 4.531568228105907, | |
| "grad_norm": 2.4558463096618652, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4645, | |
| "step": 2225 | |
| }, | |
| { | |
| "epoch": 4.582484725050916, | |
| "grad_norm": 2.276397466659546, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4721, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 4.633401221995927, | |
| "grad_norm": 2.844794750213623, | |
| "learning_rate": 0.0002, | |
| "loss": 0.477, | |
| "step": 2275 | |
| }, | |
| { | |
| "epoch": 4.684317718940937, | |
| "grad_norm": 2.6256089210510254, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4553, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 4.7352342158859475, | |
| "grad_norm": 2.599666118621826, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4753, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 4.786150712830957, | |
| "grad_norm": 2.470028877258301, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4433, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 4.8370672097759675, | |
| "grad_norm": 2.4316930770874023, | |
| "learning_rate": 0.0002, | |
| "loss": 0.493, | |
| "step": 2375 | |
| }, | |
| { | |
| "epoch": 4.887983706720978, | |
| "grad_norm": 2.4588210582733154, | |
| "learning_rate": 0.0002, | |
| "loss": 0.465, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 4.9389002036659875, | |
| "grad_norm": 2.5438883304595947, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4802, | |
| "step": 2425 | |
| }, | |
| { | |
| "epoch": 4.989816700610998, | |
| "grad_norm": 2.764341354370117, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4788, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 5.040733197556008, | |
| "grad_norm": 2.4877161979675293, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3577, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 5.091649694501018, | |
| "grad_norm": 2.5782573223114014, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3434, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 5.142566191446028, | |
| "grad_norm": 2.363449811935425, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3595, | |
| "step": 2525 | |
| }, | |
| { | |
| "epoch": 5.193482688391039, | |
| "grad_norm": 2.2839596271514893, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3566, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 5.244399185336049, | |
| "grad_norm": 2.3375349044799805, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3704, | |
| "step": 2575 | |
| }, | |
| { | |
| "epoch": 5.295315682281059, | |
| "grad_norm": 2.2514779567718506, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3586, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 5.346232179226069, | |
| "grad_norm": 2.687700033187866, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3797, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 5.39714867617108, | |
| "grad_norm": 2.498506784439087, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3761, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 5.44806517311609, | |
| "grad_norm": 2.632052183151245, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4052, | |
| "step": 2675 | |
| }, | |
| { | |
| "epoch": 5.4989816700611, | |
| "grad_norm": 2.7227554321289062, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3939, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 5.54989816700611, | |
| "grad_norm": 2.041354179382324, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3995, | |
| "step": 2725 | |
| }, | |
| { | |
| "epoch": 5.6008146639511205, | |
| "grad_norm": 4.209399223327637, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3882, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 5.65173116089613, | |
| "grad_norm": 1.9786779880523682, | |
| "learning_rate": 0.0002, | |
| "loss": 0.401, | |
| "step": 2775 | |
| }, | |
| { | |
| "epoch": 5.7026476578411405, | |
| "grad_norm": 3.278057098388672, | |
| "learning_rate": 0.0002, | |
| "loss": 0.391, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 5.753564154786151, | |
| "grad_norm": 2.4366769790649414, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4022, | |
| "step": 2825 | |
| }, | |
| { | |
| "epoch": 5.804480651731161, | |
| "grad_norm": 2.606545925140381, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3864, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 5.855397148676171, | |
| "grad_norm": 2.517683982849121, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4029, | |
| "step": 2875 | |
| }, | |
| { | |
| "epoch": 5.906313645621181, | |
| "grad_norm": 2.00010085105896, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3909, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 5.957230142566192, | |
| "grad_norm": 2.561739683151245, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3977, | |
| "step": 2925 | |
| }, | |
| { | |
| "epoch": 6.008146639511201, | |
| "grad_norm": 2.0152885913848877, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3885, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 6.059063136456212, | |
| "grad_norm": 2.1359171867370605, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3042, | |
| "step": 2975 | |
| }, | |
| { | |
| "epoch": 6.109979633401222, | |
| "grad_norm": 1.833449363708496, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2968, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 6.160896130346233, | |
| "grad_norm": 1.7801095247268677, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3064, | |
| "step": 3025 | |
| }, | |
| { | |
| "epoch": 6.211812627291242, | |
| "grad_norm": 2.25486159324646, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3364, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 6.262729124236253, | |
| "grad_norm": 2.518338680267334, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3269, | |
| "step": 3075 | |
| }, | |
| { | |
| "epoch": 6.313645621181263, | |
| "grad_norm": 1.4665368795394897, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3359, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 6.364562118126273, | |
| "grad_norm": 2.3365631103515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3233, | |
| "step": 3125 | |
| }, | |
| { | |
| "epoch": 6.415478615071283, | |
| "grad_norm": 2.406109094619751, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3409, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 6.4663951120162935, | |
| "grad_norm": 3.327075958251953, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3484, | |
| "step": 3175 | |
| }, | |
| { | |
| "epoch": 6.517311608961304, | |
| "grad_norm": 2.125439167022705, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3423, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 6.5682281059063135, | |
| "grad_norm": 2.0253591537475586, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3325, | |
| "step": 3225 | |
| }, | |
| { | |
| "epoch": 6.619144602851324, | |
| "grad_norm": 2.2506959438323975, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3592, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 6.670061099796334, | |
| "grad_norm": 3.3501861095428467, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3404, | |
| "step": 3275 | |
| }, | |
| { | |
| "epoch": 6.720977596741344, | |
| "grad_norm": 2.8377532958984375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3683, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 6.771894093686354, | |
| "grad_norm": 2.3918981552124023, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3506, | |
| "step": 3325 | |
| }, | |
| { | |
| "epoch": 6.822810590631365, | |
| "grad_norm": 2.026571273803711, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3714, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 6.873727087576375, | |
| "grad_norm": 1.9378567934036255, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3529, | |
| "step": 3375 | |
| }, | |
| { | |
| "epoch": 6.924643584521385, | |
| "grad_norm": 2.5375149250030518, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3614, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 6.975560081466395, | |
| "grad_norm": 2.6719155311584473, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3543, | |
| "step": 3425 | |
| }, | |
| { | |
| "epoch": 7.026476578411406, | |
| "grad_norm": 1.9606503248214722, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3204, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 7.077393075356415, | |
| "grad_norm": 2.2155184745788574, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2657, | |
| "step": 3475 | |
| }, | |
| { | |
| "epoch": 7.128309572301426, | |
| "grad_norm": 2.116616725921631, | |
| "learning_rate": 0.0002, | |
| "loss": 0.29, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 7.179226069246436, | |
| "grad_norm": 1.7466342449188232, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2942, | |
| "step": 3525 | |
| }, | |
| { | |
| "epoch": 7.2301425661914465, | |
| "grad_norm": 2.126159906387329, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2951, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 7.281059063136456, | |
| "grad_norm": 1.983241081237793, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3043, | |
| "step": 3575 | |
| }, | |
| { | |
| "epoch": 7.3319755600814664, | |
| "grad_norm": 2.3558452129364014, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3011, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 7.382892057026477, | |
| "grad_norm": 2.587038516998291, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3116, | |
| "step": 3625 | |
| }, | |
| { | |
| "epoch": 7.433808553971486, | |
| "grad_norm": 2.0329151153564453, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3139, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 7.484725050916497, | |
| "grad_norm": 1.955492377281189, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2996, | |
| "step": 3675 | |
| }, | |
| { | |
| "epoch": 7.535641547861507, | |
| "grad_norm": 1.8513798713684082, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3288, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 7.586558044806518, | |
| "grad_norm": 1.7869365215301514, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2988, | |
| "step": 3725 | |
| }, | |
| { | |
| "epoch": 7.637474541751527, | |
| "grad_norm": 2.4369406700134277, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3168, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 7.688391038696538, | |
| "grad_norm": 2.287158727645874, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3253, | |
| "step": 3775 | |
| }, | |
| { | |
| "epoch": 7.739307535641548, | |
| "grad_norm": 2.4751217365264893, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3143, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 7.790224032586558, | |
| "grad_norm": 2.351529836654663, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3135, | |
| "step": 3825 | |
| }, | |
| { | |
| "epoch": 7.841140529531568, | |
| "grad_norm": 1.7692896127700806, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3226, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 7.892057026476579, | |
| "grad_norm": 1.483668327331543, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3172, | |
| "step": 3875 | |
| }, | |
| { | |
| "epoch": 7.942973523421589, | |
| "grad_norm": 2.4964892864227295, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3395, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 7.993890020366599, | |
| "grad_norm": 1.4458813667297363, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3376, | |
| "step": 3925 | |
| }, | |
| { | |
| "epoch": 8.044806517311608, | |
| "grad_norm": 1.859740972518921, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2504, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 8.095723014256619, | |
| "grad_norm": 1.3244179487228394, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2643, | |
| "step": 3975 | |
| }, | |
| { | |
| "epoch": 8.146639511201629, | |
| "grad_norm": 1.5618747472763062, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2579, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 8.19755600814664, | |
| "grad_norm": 1.806612491607666, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2722, | |
| "step": 4025 | |
| }, | |
| { | |
| "epoch": 8.24847250509165, | |
| "grad_norm": 3.7666194438934326, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2762, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 8.29938900203666, | |
| "grad_norm": 2.5826241970062256, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2749, | |
| "step": 4075 | |
| }, | |
| { | |
| "epoch": 8.35030549898167, | |
| "grad_norm": 1.5352802276611328, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2827, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 8.40122199592668, | |
| "grad_norm": 1.3692405223846436, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2844, | |
| "step": 4125 | |
| }, | |
| { | |
| "epoch": 8.45213849287169, | |
| "grad_norm": 2.2680575847625732, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2921, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 8.5030549898167, | |
| "grad_norm": 1.5913770198822021, | |
| "learning_rate": 0.0002, | |
| "loss": 0.294, | |
| "step": 4175 | |
| }, | |
| { | |
| "epoch": 8.55397148676171, | |
| "grad_norm": 2.9854307174682617, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2971, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 8.604887983706721, | |
| "grad_norm": 2.6485278606414795, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3029, | |
| "step": 4225 | |
| }, | |
| { | |
| "epoch": 8.655804480651732, | |
| "grad_norm": 1.9538838863372803, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2878, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 8.706720977596742, | |
| "grad_norm": 1.6330325603485107, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3144, | |
| "step": 4275 | |
| }, | |
| { | |
| "epoch": 8.757637474541752, | |
| "grad_norm": 1.4803214073181152, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2976, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 8.808553971486761, | |
| "grad_norm": 3.5660393238067627, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3142, | |
| "step": 4325 | |
| }, | |
| { | |
| "epoch": 8.859470468431772, | |
| "grad_norm": 2.3036141395568848, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3012, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 8.910386965376782, | |
| "grad_norm": 2.528514862060547, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3145, | |
| "step": 4375 | |
| }, | |
| { | |
| "epoch": 8.961303462321792, | |
| "grad_norm": 1.4351972341537476, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3136, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 9.012219959266803, | |
| "grad_norm": 1.339784860610962, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2982, | |
| "step": 4425 | |
| }, | |
| { | |
| "epoch": 9.063136456211813, | |
| "grad_norm": 1.6082321405410767, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2447, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 9.114052953156822, | |
| "grad_norm": 2.039848566055298, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2687, | |
| "step": 4475 | |
| }, | |
| { | |
| "epoch": 9.164969450101832, | |
| "grad_norm": 1.8532267808914185, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2656, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 9.215885947046843, | |
| "grad_norm": 2.1240642070770264, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2685, | |
| "step": 4525 | |
| }, | |
| { | |
| "epoch": 9.266802443991853, | |
| "grad_norm": 1.2580517530441284, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2563, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 9.317718940936864, | |
| "grad_norm": 1.3006818294525146, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2726, | |
| "step": 4575 | |
| }, | |
| { | |
| "epoch": 9.368635437881874, | |
| "grad_norm": 1.2992304563522339, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2611, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 9.419551934826885, | |
| "grad_norm": 2.6281898021698, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2844, | |
| "step": 4625 | |
| }, | |
| { | |
| "epoch": 9.470468431771893, | |
| "grad_norm": 1.8424818515777588, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2744, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 9.521384928716904, | |
| "grad_norm": 2.217327356338501, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2943, | |
| "step": 4675 | |
| }, | |
| { | |
| "epoch": 9.572301425661914, | |
| "grad_norm": 1.463914394378662, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2651, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 9.623217922606925, | |
| "grad_norm": 1.7850229740142822, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2853, | |
| "step": 4725 | |
| }, | |
| { | |
| "epoch": 9.674134419551935, | |
| "grad_norm": 1.7324199676513672, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2755, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 9.725050916496945, | |
| "grad_norm": 1.1688644886016846, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2895, | |
| "step": 4775 | |
| }, | |
| { | |
| "epoch": 9.775967413441956, | |
| "grad_norm": 1.8048006296157837, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2728, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 9.826883910386965, | |
| "grad_norm": 1.9764938354492188, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2993, | |
| "step": 4825 | |
| }, | |
| { | |
| "epoch": 9.877800407331975, | |
| "grad_norm": 2.7224347591400146, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2882, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 9.928716904276985, | |
| "grad_norm": 2.1471643447875977, | |
| "learning_rate": 0.0002, | |
| "loss": 0.292, | |
| "step": 4875 | |
| }, | |
| { | |
| "epoch": 9.979633401221996, | |
| "grad_norm": 2.122901678085327, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2771, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 10.030549898167006, | |
| "grad_norm": 1.5285823345184326, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2568, | |
| "step": 4925 | |
| }, | |
| { | |
| "epoch": 10.081466395112017, | |
| "grad_norm": 1.8790913820266724, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2382, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 10.132382892057027, | |
| "grad_norm": 1.7328804731369019, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2489, | |
| "step": 4975 | |
| }, | |
| { | |
| "epoch": 10.183299389002036, | |
| "grad_norm": 1.6446770429611206, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2591, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 10.234215885947046, | |
| "grad_norm": 2.407752275466919, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2636, | |
| "step": 5025 | |
| }, | |
| { | |
| "epoch": 10.285132382892057, | |
| "grad_norm": 1.7200428247451782, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2528, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 10.336048879837067, | |
| "grad_norm": 4.040430545806885, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2609, | |
| "step": 5075 | |
| }, | |
| { | |
| "epoch": 10.386965376782078, | |
| "grad_norm": 1.499848484992981, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2699, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 10.437881873727088, | |
| "grad_norm": 1.3807271718978882, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2698, | |
| "step": 5125 | |
| }, | |
| { | |
| "epoch": 10.488798370672098, | |
| "grad_norm": 2.1875016689300537, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2718, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 10.539714867617107, | |
| "grad_norm": 1.8573893308639526, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2658, | |
| "step": 5175 | |
| }, | |
| { | |
| "epoch": 10.590631364562118, | |
| "grad_norm": 1.7177698612213135, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2666, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 10.641547861507128, | |
| "grad_norm": 2.5458731651306152, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2699, | |
| "step": 5225 | |
| }, | |
| { | |
| "epoch": 10.692464358452138, | |
| "grad_norm": 1.9686267375946045, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2765, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 10.743380855397149, | |
| "grad_norm": 2.519333600997925, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2746, | |
| "step": 5275 | |
| }, | |
| { | |
| "epoch": 10.79429735234216, | |
| "grad_norm": 1.456548810005188, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2746, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 10.84521384928717, | |
| "grad_norm": 2.32214617729187, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2921, | |
| "step": 5325 | |
| }, | |
| { | |
| "epoch": 10.89613034623218, | |
| "grad_norm": 1.5902856588363647, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2849, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 10.947046843177189, | |
| "grad_norm": 2.1129226684570312, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2725, | |
| "step": 5375 | |
| }, | |
| { | |
| "epoch": 10.9979633401222, | |
| "grad_norm": 2.040208578109741, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2917, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 11.04887983706721, | |
| "grad_norm": 2.841794967651367, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2304, | |
| "step": 5425 | |
| }, | |
| { | |
| "epoch": 11.09979633401222, | |
| "grad_norm": 1.1914277076721191, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2458, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 11.15071283095723, | |
| "grad_norm": 1.49674391746521, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2369, | |
| "step": 5475 | |
| }, | |
| { | |
| "epoch": 11.201629327902241, | |
| "grad_norm": 0.9921414256095886, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2586, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 11.25254582484725, | |
| "grad_norm": 2.710859537124634, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2436, | |
| "step": 5525 | |
| }, | |
| { | |
| "epoch": 11.30346232179226, | |
| "grad_norm": 1.3944896459579468, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2654, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 11.35437881873727, | |
| "grad_norm": 1.8961577415466309, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2411, | |
| "step": 5575 | |
| }, | |
| { | |
| "epoch": 11.405295315682281, | |
| "grad_norm": 3.050171136856079, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2615, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 11.456211812627291, | |
| "grad_norm": 2.27951979637146, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2534, | |
| "step": 5625 | |
| }, | |
| { | |
| "epoch": 11.507128309572302, | |
| "grad_norm": 1.4935775995254517, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2687, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 11.558044806517312, | |
| "grad_norm": 2.3738980293273926, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2485, | |
| "step": 5675 | |
| }, | |
| { | |
| "epoch": 11.608961303462323, | |
| "grad_norm": 2.942415237426758, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2735, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 11.659877800407331, | |
| "grad_norm": 2.6978747844696045, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2488, | |
| "step": 5725 | |
| }, | |
| { | |
| "epoch": 11.710794297352342, | |
| "grad_norm": 1.4899920225143433, | |
| "learning_rate": 0.0002, | |
| "loss": 0.267, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 11.761710794297352, | |
| "grad_norm": 2.248117208480835, | |
| "learning_rate": 0.0002, | |
| "loss": 0.253, | |
| "step": 5775 | |
| }, | |
| { | |
| "epoch": 11.812627291242363, | |
| "grad_norm": 1.3980762958526611, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2742, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 11.863543788187373, | |
| "grad_norm": 3.564595937728882, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2679, | |
| "step": 5825 | |
| }, | |
| { | |
| "epoch": 11.914460285132384, | |
| "grad_norm": 1.835825800895691, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2799, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 11.965376782077392, | |
| "grad_norm": 1.2737057209014893, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2761, | |
| "step": 5875 | |
| } | |
| ], | |
| "logging_steps": 25, | |
| "max_steps": 5892, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 12, | |
| "save_steps": 25, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 5.5000869805056e+16, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |