| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 4.982343499197432, | |
| "eval_steps": 500, | |
| "global_step": 385, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.012841091492776886, | |
| "grad_norm": 0.40566644072532654, | |
| "learning_rate": 1.6666666666666667e-05, | |
| "loss": 0.2239, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.025682182985553772, | |
| "grad_norm": 0.5796589851379395, | |
| "learning_rate": 3.3333333333333335e-05, | |
| "loss": 0.2932, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.038523274478330656, | |
| "grad_norm": 0.620160698890686, | |
| "learning_rate": 5e-05, | |
| "loss": 0.3059, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.051364365971107544, | |
| "grad_norm": 0.4742761254310608, | |
| "learning_rate": 6.666666666666667e-05, | |
| "loss": 0.2751, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.06420545746388442, | |
| "grad_norm": 0.39537593722343445, | |
| "learning_rate": 8.333333333333334e-05, | |
| "loss": 0.2436, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.07704654895666131, | |
| "grad_norm": 0.33797481656074524, | |
| "learning_rate": 0.0001, | |
| "loss": 0.2592, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.0898876404494382, | |
| "grad_norm": 0.3221365511417389, | |
| "learning_rate": 0.00011666666666666668, | |
| "loss": 0.2416, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.10272873194221509, | |
| "grad_norm": 0.28052330017089844, | |
| "learning_rate": 0.00013333333333333334, | |
| "loss": 0.2092, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.11556982343499198, | |
| "grad_norm": 0.29556986689567566, | |
| "learning_rate": 0.00015000000000000001, | |
| "loss": 0.2056, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.12841091492776885, | |
| "grad_norm": 0.30732303857803345, | |
| "learning_rate": 0.0001666666666666667, | |
| "loss": 0.1879, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.14125200642054575, | |
| "grad_norm": 0.30247950553894043, | |
| "learning_rate": 0.00018333333333333334, | |
| "loss": 0.1839, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.15409309791332262, | |
| "grad_norm": 0.3433546721935272, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1908, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.16693418940609953, | |
| "grad_norm": 0.19485895335674286, | |
| "learning_rate": 0.00019999645309530845, | |
| "loss": 0.1743, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.1797752808988764, | |
| "grad_norm": 0.18382272124290466, | |
| "learning_rate": 0.00019998581263284444, | |
| "loss": 0.1843, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.1926163723916533, | |
| "grad_norm": 0.1688140630722046, | |
| "learning_rate": 0.00019996807936742208, | |
| "loss": 0.1765, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.20545746388443017, | |
| "grad_norm": 0.1830463707447052, | |
| "learning_rate": 0.00019994325455700542, | |
| "loss": 0.1922, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.21829855537720708, | |
| "grad_norm": 0.22317859530448914, | |
| "learning_rate": 0.00019991133996261922, | |
| "loss": 0.2321, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.23113964686998395, | |
| "grad_norm": 0.17708703875541687, | |
| "learning_rate": 0.00019987233784822395, | |
| "loss": 0.1965, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.24398073836276082, | |
| "grad_norm": 0.18069303035736084, | |
| "learning_rate": 0.00019982625098055525, | |
| "loss": 0.1589, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.2568218298555377, | |
| "grad_norm": 0.2211710810661316, | |
| "learning_rate": 0.0001997730826289277, | |
| "loss": 0.1806, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.2696629213483146, | |
| "grad_norm": 0.2257445603609085, | |
| "learning_rate": 0.00019971283656500276, | |
| "loss": 0.1807, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.2825040128410915, | |
| "grad_norm": 0.22210289537906647, | |
| "learning_rate": 0.00019964551706252144, | |
| "loss": 0.1869, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.2953451043338684, | |
| "grad_norm": 0.19516178965568542, | |
| "learning_rate": 0.00019957112889700085, | |
| "loss": 0.1646, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.30818619582664525, | |
| "grad_norm": 0.22516900300979614, | |
| "learning_rate": 0.00019948967734539571, | |
| "loss": 0.164, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.32102728731942215, | |
| "grad_norm": 0.2650158107280731, | |
| "learning_rate": 0.0001994011681857238, | |
| "loss": 0.1733, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.33386837881219905, | |
| "grad_norm": 0.16395625472068787, | |
| "learning_rate": 0.00019930560769665617, | |
| "loss": 0.1448, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.3467094703049759, | |
| "grad_norm": 0.21180006861686707, | |
| "learning_rate": 0.00019920300265707184, | |
| "loss": 0.1933, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.3595505617977528, | |
| "grad_norm": 0.17538852989673615, | |
| "learning_rate": 0.0001990933603455767, | |
| "loss": 0.1767, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.3723916532905297, | |
| "grad_norm": 0.19017687439918518, | |
| "learning_rate": 0.00019897668853998725, | |
| "loss": 0.2061, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.3852327447833066, | |
| "grad_norm": 0.1627688705921173, | |
| "learning_rate": 0.0001988529955167791, | |
| "loss": 0.165, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.39807383627608345, | |
| "grad_norm": 0.23488560318946838, | |
| "learning_rate": 0.00019872229005049946, | |
| "loss": 0.1807, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.41091492776886035, | |
| "grad_norm": 0.16365641355514526, | |
| "learning_rate": 0.00019858458141314503, | |
| "loss": 0.1606, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.42375601926163725, | |
| "grad_norm": 0.1977553814649582, | |
| "learning_rate": 0.00019843987937350396, | |
| "loss": 0.1885, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.43659711075441415, | |
| "grad_norm": 0.24794891476631165, | |
| "learning_rate": 0.00019828819419646316, | |
| "loss": 0.1695, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.449438202247191, | |
| "grad_norm": 0.19939230382442474, | |
| "learning_rate": 0.00019812953664228, | |
| "loss": 0.1516, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.4622792937399679, | |
| "grad_norm": 0.2248964011669159, | |
| "learning_rate": 0.00019796391796581887, | |
| "loss": 0.1509, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.4751203852327448, | |
| "grad_norm": 0.2671429514884949, | |
| "learning_rate": 0.000197791349915753, | |
| "loss": 0.1936, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.48796147672552165, | |
| "grad_norm": 0.1820429414510727, | |
| "learning_rate": 0.00019761184473373095, | |
| "loss": 0.1297, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.5008025682182986, | |
| "grad_norm": 0.15517914295196533, | |
| "learning_rate": 0.00019742541515350813, | |
| "loss": 0.1664, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.5136436597110754, | |
| "grad_norm": 0.1475019007921219, | |
| "learning_rate": 0.00019723207440004362, | |
| "loss": 0.1606, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.5264847512038523, | |
| "grad_norm": 0.1478549689054489, | |
| "learning_rate": 0.0001970318361885619, | |
| "loss": 0.1629, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.5393258426966292, | |
| "grad_norm": 0.1522587686777115, | |
| "learning_rate": 0.00019682471472358003, | |
| "loss": 0.1854, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.5521669341894061, | |
| "grad_norm": 0.16658446192741394, | |
| "learning_rate": 0.00019661072469789992, | |
| "loss": 0.16, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.565008025682183, | |
| "grad_norm": 0.1563337743282318, | |
| "learning_rate": 0.0001963898812915661, | |
| "loss": 0.1646, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.5778491171749599, | |
| "grad_norm": 0.18028098344802856, | |
| "learning_rate": 0.00019616220017078882, | |
| "loss": 0.1679, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.5906902086677368, | |
| "grad_norm": 0.17881852388381958, | |
| "learning_rate": 0.00019592769748683287, | |
| "loss": 0.1567, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.6035313001605136, | |
| "grad_norm": 0.18338599801063538, | |
| "learning_rate": 0.00019568638987487155, | |
| "loss": 0.1475, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.6163723916532905, | |
| "grad_norm": 0.19950391352176666, | |
| "learning_rate": 0.0001954382944528069, | |
| "loss": 0.1604, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.6292134831460674, | |
| "grad_norm": 0.2285371720790863, | |
| "learning_rate": 0.00019518342882005532, | |
| "loss": 0.1463, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.6420545746388443, | |
| "grad_norm": 0.26193344593048096, | |
| "learning_rate": 0.00019492181105629886, | |
| "loss": 0.1649, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.6548956661316212, | |
| "grad_norm": 0.14862482249736786, | |
| "learning_rate": 0.00019465345972020313, | |
| "loss": 0.1303, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.6677367576243981, | |
| "grad_norm": 0.14517804980278015, | |
| "learning_rate": 0.00019437839384810028, | |
| "loss": 0.161, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.680577849117175, | |
| "grad_norm": 0.14767995476722717, | |
| "learning_rate": 0.00019409663295263902, | |
| "loss": 0.162, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.6934189406099518, | |
| "grad_norm": 0.15824687480926514, | |
| "learning_rate": 0.00019380819702140016, | |
| "loss": 0.1555, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.7062600321027287, | |
| "grad_norm": 0.1738673746585846, | |
| "learning_rate": 0.00019351310651547884, | |
| "loss": 0.1807, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.7191011235955056, | |
| "grad_norm": 0.1739288568496704, | |
| "learning_rate": 0.00019321138236803311, | |
| "loss": 0.1824, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.7319422150882825, | |
| "grad_norm": 0.1934700608253479, | |
| "learning_rate": 0.0001929030459827988, | |
| "loss": 0.1986, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.7447833065810594, | |
| "grad_norm": 0.18333880603313446, | |
| "learning_rate": 0.00019258811923257137, | |
| "loss": 0.1798, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.7576243980738363, | |
| "grad_norm": 0.17110952734947205, | |
| "learning_rate": 0.00019226662445765417, | |
| "loss": 0.1279, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.7704654895666132, | |
| "grad_norm": 0.18288837373256683, | |
| "learning_rate": 0.0001919385844642737, | |
| "loss": 0.1372, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.78330658105939, | |
| "grad_norm": 0.21783368289470673, | |
| "learning_rate": 0.0001916040225229618, | |
| "loss": 0.1669, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.7961476725521669, | |
| "grad_norm": 0.24256236851215363, | |
| "learning_rate": 0.00019126296236690485, | |
| "loss": 0.1504, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.8089887640449438, | |
| "grad_norm": 0.16231147944927216, | |
| "learning_rate": 0.00019091542819026024, | |
| "loss": 0.1243, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.8218298555377207, | |
| "grad_norm": 0.14093080163002014, | |
| "learning_rate": 0.0001905614446464399, | |
| "loss": 0.1547, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.8346709470304976, | |
| "grad_norm": 0.1700276881456375, | |
| "learning_rate": 0.00019020103684636177, | |
| "loss": 0.1762, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.8475120385232745, | |
| "grad_norm": 0.15964765846729279, | |
| "learning_rate": 0.00018983423035666817, | |
| "loss": 0.1593, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.8603531300160514, | |
| "grad_norm": 0.14215022325515747, | |
| "learning_rate": 0.0001894610511979123, | |
| "loss": 0.1454, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.8731942215088283, | |
| "grad_norm": 0.1703556478023529, | |
| "learning_rate": 0.00018908152584271227, | |
| "loss": 0.1681, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.8860353130016051, | |
| "grad_norm": 0.15497173368930817, | |
| "learning_rate": 0.00018869568121387343, | |
| "loss": 0.1503, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.898876404494382, | |
| "grad_norm": 0.15470866858959198, | |
| "learning_rate": 0.00018830354468247817, | |
| "loss": 0.1417, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.9117174959871589, | |
| "grad_norm": 0.17999379336833954, | |
| "learning_rate": 0.00018790514406594465, | |
| "loss": 0.1601, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.9245585874799358, | |
| "grad_norm": 0.15672680735588074, | |
| "learning_rate": 0.00018750050762605312, | |
| "loss": 0.1282, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.9373996789727127, | |
| "grad_norm": 0.21854081749916077, | |
| "learning_rate": 0.0001870896640669413, | |
| "loss": 0.1523, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.9502407704654896, | |
| "grad_norm": 0.19732460379600525, | |
| "learning_rate": 0.00018667264253306823, | |
| "loss": 0.1554, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.9630818619582665, | |
| "grad_norm": 0.24896161258220673, | |
| "learning_rate": 0.00018624947260714652, | |
| "loss": 0.1747, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.9759229534510433, | |
| "grad_norm": 0.13258929550647736, | |
| "learning_rate": 0.0001858201843080441, | |
| "loss": 0.1484, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.9887640449438202, | |
| "grad_norm": 0.17648810148239136, | |
| "learning_rate": 0.00018538480808865464, | |
| "loss": 0.1701, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 1.0112359550561798, | |
| "grad_norm": 0.5716731548309326, | |
| "learning_rate": 0.00018494337483373726, | |
| "loss": 0.248, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 1.0240770465489566, | |
| "grad_norm": 0.21971853077411652, | |
| "learning_rate": 0.00018449591585772553, | |
| "loss": 0.1502, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 1.0369181380417336, | |
| "grad_norm": 0.20206806063652039, | |
| "learning_rate": 0.00018404246290250638, | |
| "loss": 0.1179, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.0497592295345104, | |
| "grad_norm": 0.4210855960845947, | |
| "learning_rate": 0.000183583048135168, | |
| "loss": 0.1414, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 1.0626003210272874, | |
| "grad_norm": 0.2643897831439972, | |
| "learning_rate": 0.00018311770414571835, | |
| "loss": 0.1228, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 1.0754414125200642, | |
| "grad_norm": 0.27934709191322327, | |
| "learning_rate": 0.000182646463944773, | |
| "loss": 0.1244, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 1.088282504012841, | |
| "grad_norm": 0.2761859893798828, | |
| "learning_rate": 0.00018216936096121348, | |
| "loss": 0.1373, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 1.101123595505618, | |
| "grad_norm": 0.3699813187122345, | |
| "learning_rate": 0.00018168642903981607, | |
| "loss": 0.123, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 1.1139646869983948, | |
| "grad_norm": 0.38410046696662903, | |
| "learning_rate": 0.00018119770243885065, | |
| "loss": 0.126, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 1.1268057784911718, | |
| "grad_norm": 0.36278554797172546, | |
| "learning_rate": 0.0001807032158276508, | |
| "loss": 0.1081, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 1.1396468699839486, | |
| "grad_norm": 0.4462052583694458, | |
| "learning_rate": 0.00018020300428415407, | |
| "loss": 0.1134, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 1.1524879614767256, | |
| "grad_norm": 0.4787779152393341, | |
| "learning_rate": 0.00017969710329241386, | |
| "loss": 0.1165, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 1.1653290529695024, | |
| "grad_norm": 0.34396493434906006, | |
| "learning_rate": 0.00017918554874008226, | |
| "loss": 0.1239, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 1.1781701444622792, | |
| "grad_norm": 0.2252732515335083, | |
| "learning_rate": 0.00017866837691586404, | |
| "loss": 0.1598, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 1.1910112359550562, | |
| "grad_norm": 0.28776365518569946, | |
| "learning_rate": 0.00017814562450694266, | |
| "loss": 0.1315, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 1.203852327447833, | |
| "grad_norm": 0.2533372938632965, | |
| "learning_rate": 0.00017761732859637746, | |
| "loss": 0.1319, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 1.21669341894061, | |
| "grad_norm": 0.258912593126297, | |
| "learning_rate": 0.0001770835266604734, | |
| "loss": 0.1492, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 1.2295345104333868, | |
| "grad_norm": 0.2596625089645386, | |
| "learning_rate": 0.0001765442565661222, | |
| "loss": 0.1362, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 1.2423756019261638, | |
| "grad_norm": 0.3291197717189789, | |
| "learning_rate": 0.00017599955656811653, | |
| "loss": 0.1325, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 1.2552166934189406, | |
| "grad_norm": 0.3218838572502136, | |
| "learning_rate": 0.0001754494653064359, | |
| "loss": 0.1297, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 1.2680577849117176, | |
| "grad_norm": 0.4448782801628113, | |
| "learning_rate": 0.00017489402180350582, | |
| "loss": 0.1291, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 1.2808988764044944, | |
| "grad_norm": 0.33252349495887756, | |
| "learning_rate": 0.00017433326546142968, | |
| "loss": 0.1123, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 1.2937399678972712, | |
| "grad_norm": 0.4413713216781616, | |
| "learning_rate": 0.00017376723605919345, | |
| "loss": 0.124, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 1.3065810593900482, | |
| "grad_norm": 0.407705157995224, | |
| "learning_rate": 0.00017319597374984395, | |
| "loss": 0.1194, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 1.319422150882825, | |
| "grad_norm": 0.4282243251800537, | |
| "learning_rate": 0.00017261951905764056, | |
| "loss": 0.1281, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 1.332263242375602, | |
| "grad_norm": 0.30847808718681335, | |
| "learning_rate": 0.00017203791287518028, | |
| "loss": 0.127, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 1.3451043338683788, | |
| "grad_norm": 0.3466252386569977, | |
| "learning_rate": 0.00017145119646049705, | |
| "loss": 0.1737, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 1.3579454253611556, | |
| "grad_norm": 0.409016489982605, | |
| "learning_rate": 0.00017085941143413496, | |
| "loss": 0.1345, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 1.3707865168539326, | |
| "grad_norm": 0.3255477547645569, | |
| "learning_rate": 0.0001702625997761957, | |
| "loss": 0.1621, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 1.3836276083467094, | |
| "grad_norm": 0.2995210587978363, | |
| "learning_rate": 0.00016966080382336074, | |
| "loss": 0.1752, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 1.3964686998394864, | |
| "grad_norm": 0.3129432797431946, | |
| "learning_rate": 0.00016905406626588777, | |
| "loss": 0.1763, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 1.4093097913322632, | |
| "grad_norm": 0.31171512603759766, | |
| "learning_rate": 0.00016844243014458262, | |
| "loss": 0.1328, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 1.4221508828250402, | |
| "grad_norm": 0.29522988200187683, | |
| "learning_rate": 0.00016782593884774586, | |
| "loss": 0.1394, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 1.434991974317817, | |
| "grad_norm": 0.31598663330078125, | |
| "learning_rate": 0.0001672046361080949, | |
| "loss": 0.1351, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 1.447833065810594, | |
| "grad_norm": 0.3824458718299866, | |
| "learning_rate": 0.00016657856599966182, | |
| "loss": 0.1334, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 1.4606741573033708, | |
| "grad_norm": 0.39797738194465637, | |
| "learning_rate": 0.0001659477729346667, | |
| "loss": 0.1208, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 1.4735152487961476, | |
| "grad_norm": 0.4944222867488861, | |
| "learning_rate": 0.0001653123016603672, | |
| "loss": 0.1392, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 1.4863563402889246, | |
| "grad_norm": 0.30534660816192627, | |
| "learning_rate": 0.0001646721972558842, | |
| "loss": 0.1243, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 1.4991974317817014, | |
| "grad_norm": 0.33155277371406555, | |
| "learning_rate": 0.00016402750512900397, | |
| "loss": 0.1633, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 1.5120385232744784, | |
| "grad_norm": 0.343932569026947, | |
| "learning_rate": 0.0001633782710129571, | |
| "loss": 0.1701, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 1.5248796147672552, | |
| "grad_norm": 0.27937018871307373, | |
| "learning_rate": 0.00016272454096317432, | |
| "loss": 0.1549, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 1.537720706260032, | |
| "grad_norm": 0.35195839405059814, | |
| "learning_rate": 0.00016206636135401913, | |
| "loss": 0.1805, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 1.550561797752809, | |
| "grad_norm": 0.3309479057788849, | |
| "learning_rate": 0.00016140377887549843, | |
| "loss": 0.138, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 1.563402889245586, | |
| "grad_norm": 0.2956474721431732, | |
| "learning_rate": 0.0001607368405299503, | |
| "loss": 0.1484, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 1.5762439807383628, | |
| "grad_norm": 0.300729364156723, | |
| "learning_rate": 0.00016006559362870964, | |
| "loss": 0.1577, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 1.5890850722311396, | |
| "grad_norm": 0.35056933760643005, | |
| "learning_rate": 0.00015939008578875214, | |
| "loss": 0.1341, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 1.6019261637239164, | |
| "grad_norm": 0.3127368688583374, | |
| "learning_rate": 0.00015871036492931632, | |
| "loss": 0.1245, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 1.6147672552166934, | |
| "grad_norm": 0.31962379813194275, | |
| "learning_rate": 0.00015802647926850424, | |
| "loss": 0.1269, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 1.6276083467094704, | |
| "grad_norm": 0.30289211869239807, | |
| "learning_rate": 0.00015733847731986113, | |
| "loss": 0.1189, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 1.6404494382022472, | |
| "grad_norm": 0.42945238947868347, | |
| "learning_rate": 0.00015664640788893376, | |
| "loss": 0.1386, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 1.653290529695024, | |
| "grad_norm": 0.2242669314146042, | |
| "learning_rate": 0.0001559503200698084, | |
| "loss": 0.1136, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 1.666131621187801, | |
| "grad_norm": 0.382952481508255, | |
| "learning_rate": 0.00015525026324162805, | |
| "loss": 0.1661, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 1.6789727126805778, | |
| "grad_norm": 0.31802457571029663, | |
| "learning_rate": 0.00015454628706508962, | |
| "loss": 0.1715, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 1.6918138041733548, | |
| "grad_norm": 0.31699663400650024, | |
| "learning_rate": 0.00015383844147892126, | |
| "loss": 0.1454, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 1.7046548956661316, | |
| "grad_norm": 0.3550918400287628, | |
| "learning_rate": 0.00015312677669633952, | |
| "loss": 0.1454, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 1.7174959871589084, | |
| "grad_norm": 0.41878822445869446, | |
| "learning_rate": 0.00015241134320148752, | |
| "loss": 0.1565, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 1.7303370786516854, | |
| "grad_norm": 0.2940831780433655, | |
| "learning_rate": 0.00015169219174585372, | |
| "loss": 0.1405, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 1.7431781701444624, | |
| "grad_norm": 0.40646418929100037, | |
| "learning_rate": 0.00015096937334467151, | |
| "loss": 0.148, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 1.7560192616372392, | |
| "grad_norm": 0.296655535697937, | |
| "learning_rate": 0.00015024293927330047, | |
| "loss": 0.1325, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 1.768860353130016, | |
| "grad_norm": 0.2861349582672119, | |
| "learning_rate": 0.00014951294106358887, | |
| "loss": 0.1137, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 1.7817014446227928, | |
| "grad_norm": 0.35022491216659546, | |
| "learning_rate": 0.00014877943050021827, | |
| "loss": 0.1437, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 1.7945425361155698, | |
| "grad_norm": 0.39922255277633667, | |
| "learning_rate": 0.0001480424596170298, | |
| "loss": 0.1271, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 1.8073836276083468, | |
| "grad_norm": 0.21837952733039856, | |
| "learning_rate": 0.00014730208069333313, | |
| "loss": 0.1044, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 1.8202247191011236, | |
| "grad_norm": 0.28346794843673706, | |
| "learning_rate": 0.00014655834625019787, | |
| "loss": 0.1535, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 1.8330658105939004, | |
| "grad_norm": 0.3031107783317566, | |
| "learning_rate": 0.00014581130904672764, | |
| "loss": 0.1509, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 1.8459069020866774, | |
| "grad_norm": 0.288959801197052, | |
| "learning_rate": 0.00014506102207631773, | |
| "loss": 0.1565, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 1.8587479935794544, | |
| "grad_norm": 0.26415252685546875, | |
| "learning_rate": 0.00014430753856289565, | |
| "loss": 0.1542, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 1.8715890850722312, | |
| "grad_norm": 0.29712775349617004, | |
| "learning_rate": 0.0001435509119571456, | |
| "loss": 0.1442, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 1.884430176565008, | |
| "grad_norm": 0.3248273730278015, | |
| "learning_rate": 0.00014279119593271666, | |
| "loss": 0.142, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 1.8972712680577848, | |
| "grad_norm": 0.34400874376296997, | |
| "learning_rate": 0.00014202844438241546, | |
| "loss": 0.1495, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 1.9101123595505618, | |
| "grad_norm": 0.6406733989715576, | |
| "learning_rate": 0.00014126271141438315, | |
| "loss": 0.1178, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 1.9229534510433388, | |
| "grad_norm": 0.26794880628585815, | |
| "learning_rate": 0.00014049405134825677, | |
| "loss": 0.1219, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 1.9357945425361156, | |
| "grad_norm": 0.35011592507362366, | |
| "learning_rate": 0.00013972251871131625, | |
| "loss": 0.1417, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 1.9486356340288924, | |
| "grad_norm": 0.4136745035648346, | |
| "learning_rate": 0.0001389481682346162, | |
| "loss": 0.1319, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 1.9614767255216692, | |
| "grad_norm": 0.4271896779537201, | |
| "learning_rate": 0.00013817105484910334, | |
| "loss": 0.1226, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 1.9743178170144462, | |
| "grad_norm": 0.26580625772476196, | |
| "learning_rate": 0.00013739123368171994, | |
| "loss": 0.1477, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 1.9871589085072232, | |
| "grad_norm": 0.36320167779922485, | |
| "learning_rate": 0.00013660876005149318, | |
| "loss": 0.1446, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 2.009630818619583, | |
| "grad_norm": 0.556969404220581, | |
| "learning_rate": 0.00013582368946561083, | |
| "loss": 0.2174, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 2.0224719101123596, | |
| "grad_norm": 0.29500341415405273, | |
| "learning_rate": 0.00013503607761548384, | |
| "loss": 0.1428, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 2.0353130016051364, | |
| "grad_norm": 0.3006628155708313, | |
| "learning_rate": 0.00013424598037279544, | |
| "loss": 0.1346, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 2.048154093097913, | |
| "grad_norm": 0.3429463803768158, | |
| "learning_rate": 0.00013345345378553805, | |
| "loss": 0.1418, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 2.0609951845906904, | |
| "grad_norm": 0.31199735403060913, | |
| "learning_rate": 0.000132658554074037, | |
| "loss": 0.1335, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 2.073836276083467, | |
| "grad_norm": 0.35404714941978455, | |
| "learning_rate": 0.00013186133762696267, | |
| "loss": 0.1295, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 2.086677367576244, | |
| "grad_norm": 0.2951603829860687, | |
| "learning_rate": 0.00013106186099733018, | |
| "loss": 0.1309, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 2.099518459069021, | |
| "grad_norm": 0.32650235295295715, | |
| "learning_rate": 0.0001302601808984877, | |
| "loss": 0.1414, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 2.1123595505617976, | |
| "grad_norm": 0.5657181143760681, | |
| "learning_rate": 0.0001294563542000933, | |
| "loss": 0.1141, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 2.125200642054575, | |
| "grad_norm": 0.31900152564048767, | |
| "learning_rate": 0.0001286504379240807, | |
| "loss": 0.1105, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 2.1380417335473516, | |
| "grad_norm": 0.3072754442691803, | |
| "learning_rate": 0.0001278424892406143, | |
| "loss": 0.118, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 2.1508828250401284, | |
| "grad_norm": 0.36620378494262695, | |
| "learning_rate": 0.00012703256546403374, | |
| "loss": 0.1147, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 2.163723916532905, | |
| "grad_norm": 0.35902220010757446, | |
| "learning_rate": 0.00012622072404878774, | |
| "loss": 0.0722, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 2.176565008025682, | |
| "grad_norm": 0.3162679970264435, | |
| "learning_rate": 0.0001254070225853589, | |
| "loss": 0.1503, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 2.189406099518459, | |
| "grad_norm": 0.2725870907306671, | |
| "learning_rate": 0.00012459151879617785, | |
| "loss": 0.1488, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 2.202247191011236, | |
| "grad_norm": 0.28934144973754883, | |
| "learning_rate": 0.00012377427053152903, | |
| "loss": 0.1199, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 2.215088282504013, | |
| "grad_norm": 0.23648247122764587, | |
| "learning_rate": 0.00012295533576544648, | |
| "loss": 0.1137, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 2.2279293739967896, | |
| "grad_norm": 0.3106015622615814, | |
| "learning_rate": 0.00012213477259160146, | |
| "loss": 0.1433, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 2.240770465489567, | |
| "grad_norm": 0.25510647892951965, | |
| "learning_rate": 0.00012131263921918143, | |
| "loss": 0.1107, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 2.2536115569823436, | |
| "grad_norm": 0.33392271399497986, | |
| "learning_rate": 0.00012048899396876065, | |
| "loss": 0.1245, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 2.2664526484751204, | |
| "grad_norm": 0.2912399172782898, | |
| "learning_rate": 0.00011966389526816322, | |
| "loss": 0.1124, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 2.279293739967897, | |
| "grad_norm": 0.30920150876045227, | |
| "learning_rate": 0.00011883740164831818, | |
| "loss": 0.0881, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 2.292134831460674, | |
| "grad_norm": 0.3264816701412201, | |
| "learning_rate": 0.00011800957173910748, | |
| "loss": 0.1108, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 2.304975922953451, | |
| "grad_norm": 0.33401617407798767, | |
| "learning_rate": 0.00011718046426520689, | |
| "loss": 0.1088, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 2.317817014446228, | |
| "grad_norm": 0.3508036434650421, | |
| "learning_rate": 0.00011635013804192015, | |
| "loss": 0.1157, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 2.330658105939005, | |
| "grad_norm": 0.31395018100738525, | |
| "learning_rate": 0.00011551865197100686, | |
| "loss": 0.1019, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 2.3434991974317816, | |
| "grad_norm": 0.27874499559402466, | |
| "learning_rate": 0.00011468606503650394, | |
| "loss": 0.1243, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 2.3563402889245584, | |
| "grad_norm": 0.24277067184448242, | |
| "learning_rate": 0.00011385243630054144, | |
| "loss": 0.1447, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 2.3691813804173356, | |
| "grad_norm": 0.2734696567058563, | |
| "learning_rate": 0.00011301782489915287, | |
| "loss": 0.1438, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 2.3820224719101124, | |
| "grad_norm": 0.2731485068798065, | |
| "learning_rate": 0.00011218229003808012, | |
| "loss": 0.1431, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 2.394863563402889, | |
| "grad_norm": 0.47293299436569214, | |
| "learning_rate": 0.00011134589098857356, | |
| "loss": 0.1287, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 2.407704654895666, | |
| "grad_norm": 0.3162679076194763, | |
| "learning_rate": 0.00011050868708318747, | |
| "loss": 0.1217, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 2.420545746388443, | |
| "grad_norm": 0.39970093965530396, | |
| "learning_rate": 0.00010967073771157099, | |
| "loss": 0.1352, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 2.43338683788122, | |
| "grad_norm": 0.2743336856365204, | |
| "learning_rate": 0.00010883210231625534, | |
| "loss": 0.0977, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 2.446227929373997, | |
| "grad_norm": 0.37246617674827576, | |
| "learning_rate": 0.00010799284038843686, | |
| "loss": 0.1047, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 2.4590690208667736, | |
| "grad_norm": 0.3695964813232422, | |
| "learning_rate": 0.00010715301146375694, | |
| "loss": 0.1127, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 2.4719101123595504, | |
| "grad_norm": 0.35048359632492065, | |
| "learning_rate": 0.00010631267511807861, | |
| "loss": 0.1153, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 2.4847512038523276, | |
| "grad_norm": 0.20369504392147064, | |
| "learning_rate": 0.0001054718909632604, | |
| "loss": 0.0828, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 2.4975922953451044, | |
| "grad_norm": 0.3476305902004242, | |
| "learning_rate": 0.00010463071864292764, | |
| "loss": 0.1384, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 2.510433386837881, | |
| "grad_norm": 0.2885417640209198, | |
| "learning_rate": 0.00010378921782824128, | |
| "loss": 0.1496, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 2.523274478330658, | |
| "grad_norm": 0.2671290338039398, | |
| "learning_rate": 0.00010294744821366504, | |
| "loss": 0.14, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 2.5361155698234352, | |
| "grad_norm": 0.30776816606521606, | |
| "learning_rate": 0.0001021054695127309, | |
| "loss": 0.1442, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 2.548956661316212, | |
| "grad_norm": 0.2914685904979706, | |
| "learning_rate": 0.00010126334145380288, | |
| "loss": 0.1321, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 2.561797752808989, | |
| "grad_norm": 0.23030787706375122, | |
| "learning_rate": 0.00010042112377584028, | |
| "loss": 0.1177, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 2.5746388443017656, | |
| "grad_norm": 0.2737199068069458, | |
| "learning_rate": 9.957887622415975e-05, | |
| "loss": 0.1198, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 2.5874799357945424, | |
| "grad_norm": 0.26640573143959045, | |
| "learning_rate": 9.873665854619715e-05, | |
| "loss": 0.1084, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 2.600321027287319, | |
| "grad_norm": 0.3201727569103241, | |
| "learning_rate": 9.789453048726912e-05, | |
| "loss": 0.1022, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 2.6131621187800964, | |
| "grad_norm": 0.262690007686615, | |
| "learning_rate": 9.705255178633497e-05, | |
| "loss": 0.094, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 2.626003210272873, | |
| "grad_norm": 0.43916091322898865, | |
| "learning_rate": 9.621078217175876e-05, | |
| "loss": 0.1059, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 2.63884430176565, | |
| "grad_norm": 0.3432372212409973, | |
| "learning_rate": 9.53692813570724e-05, | |
| "loss": 0.1142, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 2.6516853932584272, | |
| "grad_norm": 0.23425506055355072, | |
| "learning_rate": 9.452810903673963e-05, | |
| "loss": 0.1118, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 2.664526484751204, | |
| "grad_norm": 0.2430526316165924, | |
| "learning_rate": 9.368732488192143e-05, | |
| "loss": 0.0987, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 2.677367576243981, | |
| "grad_norm": 0.28097087144851685, | |
| "learning_rate": 9.28469885362431e-05, | |
| "loss": 0.1198, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 2.6902086677367576, | |
| "grad_norm": 0.2345268875360489, | |
| "learning_rate": 9.200715961156317e-05, | |
| "loss": 0.1115, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 2.7030497592295344, | |
| "grad_norm": 0.4230293035507202, | |
| "learning_rate": 9.116789768374467e-05, | |
| "loss": 0.1497, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 2.715890850722311, | |
| "grad_norm": 0.2744080424308777, | |
| "learning_rate": 9.032926228842902e-05, | |
| "loss": 0.1296, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 2.7287319422150884, | |
| "grad_norm": 0.3323458135128021, | |
| "learning_rate": 8.949131291681257e-05, | |
| "loss": 0.143, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 2.741573033707865, | |
| "grad_norm": 0.2832281291484833, | |
| "learning_rate": 8.865410901142645e-05, | |
| "loss": 0.1154, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 2.754414125200642, | |
| "grad_norm": 0.3109094798564911, | |
| "learning_rate": 8.781770996191992e-05, | |
| "loss": 0.1146, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 2.767255216693419, | |
| "grad_norm": 0.27764928340911865, | |
| "learning_rate": 8.698217510084717e-05, | |
| "loss": 0.0992, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 2.780096308186196, | |
| "grad_norm": 0.3414210379123688, | |
| "learning_rate": 8.614756369945856e-05, | |
| "loss": 0.0996, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 2.792937399678973, | |
| "grad_norm": 0.276713490486145, | |
| "learning_rate": 8.531393496349606e-05, | |
| "loss": 0.0948, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 2.8057784911717496, | |
| "grad_norm": 0.24623610079288483, | |
| "learning_rate": 8.448134802899314e-05, | |
| "loss": 0.0819, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 2.8186195826645264, | |
| "grad_norm": 0.21359184384346008, | |
| "learning_rate": 8.364986195807986e-05, | |
| "loss": 0.1337, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 2.831460674157303, | |
| "grad_norm": 0.28975728154182434, | |
| "learning_rate": 8.281953573479315e-05, | |
| "loss": 0.1421, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 2.8443017656500804, | |
| "grad_norm": 0.25927087664604187, | |
| "learning_rate": 8.199042826089252e-05, | |
| "loss": 0.1162, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 2.857142857142857, | |
| "grad_norm": 0.2766461968421936, | |
| "learning_rate": 8.116259835168183e-05, | |
| "loss": 0.1, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 2.869983948635634, | |
| "grad_norm": 0.28379619121551514, | |
| "learning_rate": 8.033610473183678e-05, | |
| "loss": 0.1001, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 2.882825040128411, | |
| "grad_norm": 0.2919027805328369, | |
| "learning_rate": 7.951100603123936e-05, | |
| "loss": 0.1081, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 2.895666131621188, | |
| "grad_norm": 0.26831895112991333, | |
| "learning_rate": 7.868736078081859e-05, | |
| "loss": 0.1092, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 2.908507223113965, | |
| "grad_norm": 0.2470114827156067, | |
| "learning_rate": 7.786522740839852e-05, | |
| "loss": 0.1069, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 2.9213483146067416, | |
| "grad_norm": 0.27561888098716736, | |
| "learning_rate": 7.704466423455353e-05, | |
| "loss": 0.0981, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 2.9341894060995184, | |
| "grad_norm": 0.26214268803596497, | |
| "learning_rate": 7.622572946847098e-05, | |
| "loss": 0.0933, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 2.947030497592295, | |
| "grad_norm": 0.3250983655452728, | |
| "learning_rate": 7.540848120382214e-05, | |
| "loss": 0.1005, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 2.959871589085072, | |
| "grad_norm": 0.3818821609020233, | |
| "learning_rate": 7.459297741464115e-05, | |
| "loss": 0.0977, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 2.972712680577849, | |
| "grad_norm": 0.1819436401128769, | |
| "learning_rate": 7.37792759512123e-05, | |
| "loss": 0.1218, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 2.985553772070626, | |
| "grad_norm": 0.28734883666038513, | |
| "learning_rate": 7.296743453596631e-05, | |
| "loss": 0.1154, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 3.0080256821829856, | |
| "grad_norm": 0.5932813882827759, | |
| "learning_rate": 7.215751075938571e-05, | |
| "loss": 0.2066, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 3.0208667736757624, | |
| "grad_norm": 0.19985750317573547, | |
| "learning_rate": 7.134956207591936e-05, | |
| "loss": 0.1257, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 3.033707865168539, | |
| "grad_norm": 0.2326143980026245, | |
| "learning_rate": 7.054364579990677e-05, | |
| "loss": 0.1019, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 3.0465489566613164, | |
| "grad_norm": 0.24029302597045898, | |
| "learning_rate": 6.973981910151234e-05, | |
| "loss": 0.1013, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 3.059390048154093, | |
| "grad_norm": 0.19977405667304993, | |
| "learning_rate": 6.893813900266987e-05, | |
| "loss": 0.087, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 3.07223113964687, | |
| "grad_norm": 0.2730342447757721, | |
| "learning_rate": 6.813866237303735e-05, | |
| "loss": 0.0991, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 3.085072231139647, | |
| "grad_norm": 0.2400568425655365, | |
| "learning_rate": 6.734144592596303e-05, | |
| "loss": 0.1135, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 3.0979133226324236, | |
| "grad_norm": 0.22795532643795013, | |
| "learning_rate": 6.654654621446202e-05, | |
| "loss": 0.1163, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 3.110754414125201, | |
| "grad_norm": 0.2850149869918823, | |
| "learning_rate": 6.57540196272046e-05, | |
| "loss": 0.1007, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 3.1235955056179776, | |
| "grad_norm": 0.2573602497577667, | |
| "learning_rate": 6.496392238451621e-05, | |
| "loss": 0.0706, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 3.1364365971107544, | |
| "grad_norm": 0.27073922753334045, | |
| "learning_rate": 6.417631053438917e-05, | |
| "loss": 0.0842, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 3.149277688603531, | |
| "grad_norm": 0.24778737127780914, | |
| "learning_rate": 6.339123994850684e-05, | |
| "loss": 0.0822, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 3.162118780096308, | |
| "grad_norm": 0.24561458826065063, | |
| "learning_rate": 6.26087663182801e-05, | |
| "loss": 0.0441, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 3.174959871589085, | |
| "grad_norm": 0.27637338638305664, | |
| "learning_rate": 6.182894515089672e-05, | |
| "loss": 0.1313, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 3.187800963081862, | |
| "grad_norm": 0.2668153941631317, | |
| "learning_rate": 6.105183176538385e-05, | |
| "loss": 0.1261, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 3.200642054574639, | |
| "grad_norm": 0.2336243838071823, | |
| "learning_rate": 6.0277481288683765e-05, | |
| "loss": 0.0968, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 3.2134831460674156, | |
| "grad_norm": 0.2731676995754242, | |
| "learning_rate": 5.950594865174325e-05, | |
| "loss": 0.1, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 3.226324237560193, | |
| "grad_norm": 0.19752222299575806, | |
| "learning_rate": 5.8737288585616867e-05, | |
| "loss": 0.1018, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 3.2391653290529696, | |
| "grad_norm": 0.274467796087265, | |
| "learning_rate": 5.797155561758454e-05, | |
| "loss": 0.1193, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 3.2520064205457464, | |
| "grad_norm": 0.2718445062637329, | |
| "learning_rate": 5.7208804067283396e-05, | |
| "loss": 0.0881, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 3.264847512038523, | |
| "grad_norm": 0.2557855248451233, | |
| "learning_rate": 5.644908804285447e-05, | |
| "loss": 0.0833, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 3.2776886035313, | |
| "grad_norm": 0.2935195863246918, | |
| "learning_rate": 5.5692461437104335e-05, | |
| "loss": 0.0881, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 3.290529695024077, | |
| "grad_norm": 0.2287595272064209, | |
| "learning_rate": 5.493897792368228e-05, | |
| "loss": 0.0688, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 3.303370786516854, | |
| "grad_norm": 0.2637540102005005, | |
| "learning_rate": 5.418869095327237e-05, | |
| "loss": 0.0665, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 3.316211878009631, | |
| "grad_norm": 0.3496418595314026, | |
| "learning_rate": 5.344165374980217e-05, | |
| "loss": 0.0901, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 3.3290529695024076, | |
| "grad_norm": 0.1895870864391327, | |
| "learning_rate": 5.269791930666688e-05, | |
| "loss": 0.0885, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 3.341894060995185, | |
| "grad_norm": 0.2253326177597046, | |
| "learning_rate": 5.195754038297023e-05, | |
| "loss": 0.1087, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 3.3547351524879616, | |
| "grad_norm": 0.18178820610046387, | |
| "learning_rate": 5.122056949978176e-05, | |
| "loss": 0.1029, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 3.3675762439807384, | |
| "grad_norm": 0.26113295555114746, | |
| "learning_rate": 5.0487058936411144e-05, | |
| "loss": 0.1091, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 3.380417335473515, | |
| "grad_norm": 0.2430870234966278, | |
| "learning_rate": 4.975706072669958e-05, | |
| "loss": 0.1085, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 3.393258426966292, | |
| "grad_norm": 0.2758089005947113, | |
| "learning_rate": 4.9030626655328516e-05, | |
| "loss": 0.0981, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 3.4060995184590688, | |
| "grad_norm": 0.26935628056526184, | |
| "learning_rate": 4.8307808254146266e-05, | |
| "loss": 0.0873, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 3.418940609951846, | |
| "grad_norm": 0.2657538652420044, | |
| "learning_rate": 4.758865679851247e-05, | |
| "loss": 0.0843, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 3.431781701444623, | |
| "grad_norm": 0.24748681485652924, | |
| "learning_rate": 4.68732233036605e-05, | |
| "loss": 0.0798, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 3.4446227929373996, | |
| "grad_norm": 0.456748366355896, | |
| "learning_rate": 4.616155852107877e-05, | |
| "loss": 0.0851, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 3.4574638844301764, | |
| "grad_norm": 0.23406369984149933, | |
| "learning_rate": 4.54537129349104e-05, | |
| "loss": 0.0831, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 3.4703049759229536, | |
| "grad_norm": 0.24098779261112213, | |
| "learning_rate": 4.4749736758372e-05, | |
| "loss": 0.081, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 3.4831460674157304, | |
| "grad_norm": 0.13690468668937683, | |
| "learning_rate": 4.404967993019162e-05, | |
| "loss": 0.0439, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 3.495987158908507, | |
| "grad_norm": 0.1986464560031891, | |
| "learning_rate": 4.335359211106624e-05, | |
| "loss": 0.1084, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 3.508828250401284, | |
| "grad_norm": 0.2006080001592636, | |
| "learning_rate": 4.26615226801389e-05, | |
| "loss": 0.1235, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 3.521669341894061, | |
| "grad_norm": 0.2104363888502121, | |
| "learning_rate": 4.1973520731495765e-05, | |
| "loss": 0.1134, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 3.534510433386838, | |
| "grad_norm": 0.24415647983551025, | |
| "learning_rate": 4.128963507068371e-05, | |
| "loss": 0.1146, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 3.547351524879615, | |
| "grad_norm": 0.2873344421386719, | |
| "learning_rate": 4.0609914211247866e-05, | |
| "loss": 0.0897, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 3.5601926163723916, | |
| "grad_norm": 0.19503550231456757, | |
| "learning_rate": 3.993440637129036e-05, | |
| "loss": 0.0896, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 3.5730337078651684, | |
| "grad_norm": 0.27025559544563293, | |
| "learning_rate": 3.926315947004971e-05, | |
| "loss": 0.0979, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 3.5858747993579456, | |
| "grad_norm": 0.3529897630214691, | |
| "learning_rate": 3.859622112450158e-05, | |
| "loss": 0.0784, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 3.5987158908507224, | |
| "grad_norm": 0.31557080149650574, | |
| "learning_rate": 3.7933638645980905e-05, | |
| "loss": 0.0921, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 3.611556982343499, | |
| "grad_norm": 0.25294047594070435, | |
| "learning_rate": 3.727545903682572e-05, | |
| "loss": 0.086, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 3.624398073836276, | |
| "grad_norm": 0.20190520584583282, | |
| "learning_rate": 3.6621728987042905e-05, | |
| "loss": 0.0605, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 3.637239165329053, | |
| "grad_norm": 0.2746667265892029, | |
| "learning_rate": 3.5972494870996063e-05, | |
| "loss": 0.077, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 3.65008025682183, | |
| "grad_norm": 0.23425573110580444, | |
| "learning_rate": 3.532780274411581e-05, | |
| "loss": 0.0991, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 3.662921348314607, | |
| "grad_norm": 0.15118825435638428, | |
| "learning_rate": 3.468769833963279e-05, | |
| "loss": 0.1051, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 3.6757624398073836, | |
| "grad_norm": 0.23835481703281403, | |
| "learning_rate": 3.40522270653333e-05, | |
| "loss": 0.0939, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 3.6886035313001604, | |
| "grad_norm": 0.17009741067886353, | |
| "learning_rate": 3.3421434000338194e-05, | |
| "loss": 0.085, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 3.7014446227929376, | |
| "grad_norm": 0.23825782537460327, | |
| "learning_rate": 3.279536389190512e-05, | |
| "loss": 0.0982, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 3.7142857142857144, | |
| "grad_norm": 0.21432097256183624, | |
| "learning_rate": 3.2174061152254166e-05, | |
| "loss": 0.079, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 3.727126805778491, | |
| "grad_norm": 0.21313104033470154, | |
| "learning_rate": 3.1557569855417394e-05, | |
| "loss": 0.0852, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 3.739967897271268, | |
| "grad_norm": 0.28463664650917053, | |
| "learning_rate": 3.094593373411224e-05, | |
| "loss": 0.0883, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 3.752808988764045, | |
| "grad_norm": 0.24580109119415283, | |
| "learning_rate": 3.0339196176639283e-05, | |
| "loss": 0.0771, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 3.7656500802568216, | |
| "grad_norm": 0.24096901714801788, | |
| "learning_rate": 2.973740022380428e-05, | |
| "loss": 0.0749, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 3.778491171749599, | |
| "grad_norm": 0.25022003054618835, | |
| "learning_rate": 2.9140588565865077e-05, | |
| "loss": 0.0748, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 3.7913322632423756, | |
| "grad_norm": 0.2789011597633362, | |
| "learning_rate": 2.8548803539502987e-05, | |
| "loss": 0.0717, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 3.8041733547351524, | |
| "grad_norm": 0.1361372172832489, | |
| "learning_rate": 2.7962087124819757e-05, | |
| "loss": 0.0431, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 3.8170144462279296, | |
| "grad_norm": 0.22512777149677277, | |
| "learning_rate": 2.7380480942359477e-05, | |
| "loss": 0.1221, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 3.8298555377207064, | |
| "grad_norm": 0.2254047840833664, | |
| "learning_rate": 2.6804026250156077e-05, | |
| "loss": 0.1062, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 3.842696629213483, | |
| "grad_norm": 0.27256065607070923, | |
| "learning_rate": 2.6232763940806606e-05, | |
| "loss": 0.1023, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 3.85553772070626, | |
| "grad_norm": 0.2014126181602478, | |
| "learning_rate": 2.5666734538570337e-05, | |
| "loss": 0.0999, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 3.868378812199037, | |
| "grad_norm": 0.27657970786094666, | |
| "learning_rate": 2.5105978196494184e-05, | |
| "loss": 0.0859, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 3.8812199036918136, | |
| "grad_norm": 0.287130743265152, | |
| "learning_rate": 2.455053469356413e-05, | |
| "loss": 0.0946, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 3.894060995184591, | |
| "grad_norm": 0.28707414865493774, | |
| "learning_rate": 2.400044343188349e-05, | |
| "loss": 0.0848, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 3.9069020866773676, | |
| "grad_norm": 0.20366519689559937, | |
| "learning_rate": 2.3455743433877808e-05, | |
| "loss": 0.0774, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 3.9197431781701444, | |
| "grad_norm": 0.2152596265077591, | |
| "learning_rate": 2.2916473339526644e-05, | |
| "loss": 0.0715, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 3.932584269662921, | |
| "grad_norm": 0.26020580530166626, | |
| "learning_rate": 2.238267140362257e-05, | |
| "loss": 0.0721, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 3.9454253611556984, | |
| "grad_norm": 0.3326077461242676, | |
| "learning_rate": 2.1854375493057388e-05, | |
| "loss": 0.0757, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 3.958266452648475, | |
| "grad_norm": 0.26648905873298645, | |
| "learning_rate": 2.1331623084135976e-05, | |
| "loss": 0.0832, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 3.971107544141252, | |
| "grad_norm": 0.21082034707069397, | |
| "learning_rate": 2.0814451259917767e-05, | |
| "loss": 0.0825, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 3.983948635634029, | |
| "grad_norm": 0.19309554994106293, | |
| "learning_rate": 2.0302896707586127e-05, | |
| "loss": 0.0906, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 4.006420545746389, | |
| "grad_norm": 0.35690435767173767, | |
| "learning_rate": 1.9796995715845955e-05, | |
| "loss": 0.1062, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 4.019261637239166, | |
| "grad_norm": 0.1724165827035904, | |
| "learning_rate": 1.9296784172349223e-05, | |
| "loss": 0.1117, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 4.032102728731942, | |
| "grad_norm": 0.17502835392951965, | |
| "learning_rate": 1.8802297561149363e-05, | |
| "loss": 0.1141, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 4.044943820224719, | |
| "grad_norm": 0.21144062280654907, | |
| "learning_rate": 1.8313570960183977e-05, | |
| "loss": 0.0925, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 4.057784911717496, | |
| "grad_norm": 0.25978124141693115, | |
| "learning_rate": 1.7830639038786555e-05, | |
| "loss": 0.0865, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 4.070626003210273, | |
| "grad_norm": 0.20121395587921143, | |
| "learning_rate": 1.7353536055227047e-05, | |
| "loss": 0.0658, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 4.08346709470305, | |
| "grad_norm": 0.3032616376876831, | |
| "learning_rate": 1.688229585428167e-05, | |
| "loss": 0.0741, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 4.096308186195826, | |
| "grad_norm": 0.19196993112564087, | |
| "learning_rate": 1.6416951864832008e-05, | |
| "loss": 0.069, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 4.109149277688603, | |
| "grad_norm": 0.2227143496274948, | |
| "learning_rate": 1.5957537097493637e-05, | |
| "loss": 0.064, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 4.121990369181381, | |
| "grad_norm": 0.2821972072124481, | |
| "learning_rate": 1.5504084142274477e-05, | |
| "loss": 0.0563, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 4.134831460674158, | |
| "grad_norm": 0.23713919520378113, | |
| "learning_rate": 1.5056625166262772e-05, | |
| "loss": 0.0657, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 4.147672552166934, | |
| "grad_norm": 0.18978498876094818, | |
| "learning_rate": 1.4615191911345371e-05, | |
| "loss": 0.0609, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 4.160513643659711, | |
| "grad_norm": 0.12965190410614014, | |
| "learning_rate": 1.4179815691955923e-05, | |
| "loss": 0.0341, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 4.173354735152488, | |
| "grad_norm": 0.1908600926399231, | |
| "learning_rate": 1.3750527392853519e-05, | |
| "loss": 0.0838, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 4.186195826645265, | |
| "grad_norm": 0.22569157183170319, | |
| "learning_rate": 1.3327357466931811e-05, | |
| "loss": 0.0833, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 4.199036918138042, | |
| "grad_norm": 0.20586968958377838, | |
| "learning_rate": 1.2910335933058714e-05, | |
| "loss": 0.0921, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 4.211878009630818, | |
| "grad_norm": 0.22846491634845734, | |
| "learning_rate": 1.2499492373946919e-05, | |
| "loss": 0.0935, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 4.224719101123595, | |
| "grad_norm": 0.2915958762168884, | |
| "learning_rate": 1.2094855934055361e-05, | |
| "loss": 0.0931, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 4.237560192616373, | |
| "grad_norm": 0.22078849375247955, | |
| "learning_rate": 1.1696455317521826e-05, | |
| "loss": 0.0873, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 4.25040128410915, | |
| "grad_norm": 0.2387520968914032, | |
| "learning_rate": 1.1304318786126588e-05, | |
| "loss": 0.0819, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 4.263242375601926, | |
| "grad_norm": 0.23125675320625305, | |
| "learning_rate": 1.0918474157287728e-05, | |
| "loss": 0.0713, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 4.276083467094703, | |
| "grad_norm": 0.25590938329696655, | |
| "learning_rate": 1.0538948802087722e-05, | |
| "loss": 0.0734, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 4.28892455858748, | |
| "grad_norm": 0.23935070633888245, | |
| "learning_rate": 1.0165769643331836e-05, | |
| "loss": 0.0677, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 4.301765650080257, | |
| "grad_norm": 0.18608956038951874, | |
| "learning_rate": 9.79896315363823e-06, | |
| "loss": 0.0541, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 4.314606741573034, | |
| "grad_norm": 0.3140433132648468, | |
| "learning_rate": 9.438555353560107e-06, | |
| "loss": 0.0586, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 4.32744783306581, | |
| "grad_norm": 0.31043872237205505, | |
| "learning_rate": 9.084571809739806e-06, | |
| "loss": 0.0726, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 4.340288924558587, | |
| "grad_norm": 0.19324089586734772, | |
| "learning_rate": 8.737037633095168e-06, | |
| "loss": 0.082, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 4.353130016051364, | |
| "grad_norm": 0.2725972831249237, | |
| "learning_rate": 8.395977477038198e-06, | |
| "loss": 0.1139, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 4.365971107544142, | |
| "grad_norm": 0.16910400986671448, | |
| "learning_rate": 8.061415535726303e-06, | |
| "loss": 0.0728, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 4.378812199036918, | |
| "grad_norm": 0.22665587067604065, | |
| "learning_rate": 7.733375542345832e-06, | |
| "loss": 0.096, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 4.391653290529695, | |
| "grad_norm": 0.26520606875419617, | |
| "learning_rate": 7.411880767428636e-06, | |
| "loss": 0.0902, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 4.404494382022472, | |
| "grad_norm": 0.19831718504428864, | |
| "learning_rate": 7.096954017201207e-06, | |
| "loss": 0.0816, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 4.417335473515249, | |
| "grad_norm": 0.20937596261501312, | |
| "learning_rate": 6.788617631966909e-06, | |
| "loss": 0.065, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 4.430176565008026, | |
| "grad_norm": 0.284106582403183, | |
| "learning_rate": 6.486893484521162e-06, | |
| "loss": 0.0653, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 4.443017656500802, | |
| "grad_norm": 0.23955251276493073, | |
| "learning_rate": 6.19180297859987e-06, | |
| "loss": 0.0483, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 4.455858747993579, | |
| "grad_norm": 0.2952927052974701, | |
| "learning_rate": 5.9033670473610085e-06, | |
| "loss": 0.0719, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 4.468699839486356, | |
| "grad_norm": 0.2724378705024719, | |
| "learning_rate": 5.621606151899716e-06, | |
| "loss": 0.0544, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 4.481540930979134, | |
| "grad_norm": 0.17572782933712006, | |
| "learning_rate": 5.346540279796886e-06, | |
| "loss": 0.0446, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 4.49438202247191, | |
| "grad_norm": 0.22200757265090942, | |
| "learning_rate": 5.078188943701123e-06, | |
| "loss": 0.1018, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 4.507223113964687, | |
| "grad_norm": 0.2331586480140686, | |
| "learning_rate": 4.816571179944706e-06, | |
| "loss": 0.0991, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 4.520064205457464, | |
| "grad_norm": 0.18275317549705505, | |
| "learning_rate": 4.561705547193096e-06, | |
| "loss": 0.0768, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 4.532905296950241, | |
| "grad_norm": 0.1845731884241104, | |
| "learning_rate": 4.313610125128464e-06, | |
| "loss": 0.0922, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 4.545746388443018, | |
| "grad_norm": 0.2531459331512451, | |
| "learning_rate": 4.072302513167148e-06, | |
| "loss": 0.0731, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 4.558587479935794, | |
| "grad_norm": 0.17735856771469116, | |
| "learning_rate": 3.837799829211164e-06, | |
| "loss": 0.0664, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 4.571428571428571, | |
| "grad_norm": 0.1611146628856659, | |
| "learning_rate": 3.610118708433907e-06, | |
| "loss": 0.0681, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 4.584269662921348, | |
| "grad_norm": 0.2285057157278061, | |
| "learning_rate": 3.389275302100081e-06, | |
| "loss": 0.0685, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 4.597110754414125, | |
| "grad_norm": 0.21249911189079285, | |
| "learning_rate": 3.1752852764199812e-06, | |
| "loss": 0.0642, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 4.609951845906902, | |
| "grad_norm": 0.27474352717399597, | |
| "learning_rate": 2.9681638114381184e-06, | |
| "loss": 0.0629, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 4.622792937399679, | |
| "grad_norm": 0.23277732729911804, | |
| "learning_rate": 2.767925599956411e-06, | |
| "loss": 0.0533, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 4.635634028892456, | |
| "grad_norm": 0.34809577465057373, | |
| "learning_rate": 2.57458484649189e-06, | |
| "loss": 0.0653, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 4.648475120385233, | |
| "grad_norm": 0.19076129794120789, | |
| "learning_rate": 2.3881552662690765e-06, | |
| "loss": 0.0871, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 4.66131621187801, | |
| "grad_norm": 0.15909138321876526, | |
| "learning_rate": 2.2086500842470214e-06, | |
| "loss": 0.0811, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 4.674157303370786, | |
| "grad_norm": 0.1952279508113861, | |
| "learning_rate": 2.0360820341811635e-06, | |
| "loss": 0.0826, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 4.686998394863563, | |
| "grad_norm": 0.14884842932224274, | |
| "learning_rate": 1.8704633577200204e-06, | |
| "loss": 0.0582, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 4.69983948635634, | |
| "grad_norm": 0.2566874027252197, | |
| "learning_rate": 1.7118058035368368e-06, | |
| "loss": 0.0891, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 4.712680577849117, | |
| "grad_norm": 0.23755532503128052, | |
| "learning_rate": 1.5601206264960644e-06, | |
| "loss": 0.0904, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 4.725521669341894, | |
| "grad_norm": 0.2856099009513855, | |
| "learning_rate": 1.4154185868550106e-06, | |
| "loss": 0.0754, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 4.738362760834671, | |
| "grad_norm": 0.24481110274791718, | |
| "learning_rate": 1.2777099495005495e-06, | |
| "loss": 0.0777, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 4.751203852327448, | |
| "grad_norm": 0.2835271656513214, | |
| "learning_rate": 1.147004483220926e-06, | |
| "loss": 0.0688, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 4.764044943820225, | |
| "grad_norm": 0.26856350898742676, | |
| "learning_rate": 1.0233114600127547e-06, | |
| "loss": 0.0639, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 4.776886035313002, | |
| "grad_norm": 0.2596482038497925, | |
| "learning_rate": 9.066396544233468e-07, | |
| "loss": 0.0557, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 4.789727126805778, | |
| "grad_norm": 0.25499358773231506, | |
| "learning_rate": 7.969973429281741e-07, | |
| "loss": 0.0536, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 4.802568218298555, | |
| "grad_norm": 0.1461191475391388, | |
| "learning_rate": 6.94392303343816e-07, | |
| "loss": 0.0412, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 4.815409309791332, | |
| "grad_norm": 0.32968461513519287, | |
| "learning_rate": 5.988318142762128e-07, | |
| "loss": 0.1382, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 4.828250401284109, | |
| "grad_norm": 0.19888003170490265, | |
| "learning_rate": 5.103226546043005e-07, | |
| "loss": 0.0932, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 4.841091492776886, | |
| "grad_norm": 0.22171039879322052, | |
| "learning_rate": 4.28871102999151e-07, | |
| "loss": 0.0748, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 4.853932584269663, | |
| "grad_norm": 0.18117375671863556, | |
| "learning_rate": 3.54482937478573e-07, | |
| "loss": 0.0805, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 4.86677367576244, | |
| "grad_norm": 0.18704640865325928, | |
| "learning_rate": 2.871634349972285e-07, | |
| "loss": 0.0959, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 4.879614767255217, | |
| "grad_norm": 0.2359827756881714, | |
| "learning_rate": 2.2691737107231048e-07, | |
| "loss": 0.0779, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 4.892455858747994, | |
| "grad_norm": 0.19910109043121338, | |
| "learning_rate": 1.7374901944474663e-07, | |
| "loss": 0.0737, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 4.90529695024077, | |
| "grad_norm": 0.2202347368001938, | |
| "learning_rate": 1.2766215177605345e-07, | |
| "loss": 0.0775, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 4.918138041733547, | |
| "grad_norm": 0.2741416096687317, | |
| "learning_rate": 8.866003738078332e-08, | |
| "loss": 0.0816, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 4.930979133226324, | |
| "grad_norm": 0.19433164596557617, | |
| "learning_rate": 5.674544299457685e-08, | |
| "loss": 0.0703, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 4.943820224719101, | |
| "grad_norm": 0.24093866348266602, | |
| "learning_rate": 3.192063257793088e-08, | |
| "loss": 0.0588, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 4.956661316211878, | |
| "grad_norm": 0.3715647757053375, | |
| "learning_rate": 1.4187367155582555e-08, | |
| "loss": 0.0546, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 4.969502407704655, | |
| "grad_norm": 0.26011672616004944, | |
| "learning_rate": 3.546904691564823e-09, | |
| "loss": 0.0893, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 4.982343499197432, | |
| "grad_norm": 0.19426091015338898, | |
| "learning_rate": 0.0, | |
| "loss": 0.0736, | |
| "step": 385 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 385, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.2073448843153244e+18, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |