| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.964487905301081, | |
| "eval_steps": 200, | |
| "global_step": 180, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.016469377251672673, | |
| "grad_norm": 25.988508224487305, | |
| "learning_rate": 1.111111111111111e-06, | |
| "loss": 0.5961, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.032938754503345345, | |
| "grad_norm": 26.565134048461914, | |
| "learning_rate": 2.222222222222222e-06, | |
| "loss": 0.595, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.049408131755018014, | |
| "grad_norm": 22.700319290161133, | |
| "learning_rate": 3.3333333333333333e-06, | |
| "loss": 0.5211, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.06587750900669069, | |
| "grad_norm": 17.019222259521484, | |
| "learning_rate": 4.444444444444444e-06, | |
| "loss": 0.4043, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.08234688625836335, | |
| "grad_norm": 15.759174346923828, | |
| "learning_rate": 5.555555555555557e-06, | |
| "loss": 0.3571, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.09881626351003603, | |
| "grad_norm": 12.310270309448242, | |
| "learning_rate": 6.666666666666667e-06, | |
| "loss": 0.351, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.1152856407617087, | |
| "grad_norm": 6.430344104766846, | |
| "learning_rate": 7.77777777777778e-06, | |
| "loss": 0.3183, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.13175501801338138, | |
| "grad_norm": 29.118850708007812, | |
| "learning_rate": 8.888888888888888e-06, | |
| "loss": 0.2986, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.14822439526505404, | |
| "grad_norm": 5.387965202331543, | |
| "learning_rate": 1e-05, | |
| "loss": 0.3031, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.1646937725167267, | |
| "grad_norm": 4.416501998901367, | |
| "learning_rate": 1.1111111111111113e-05, | |
| "loss": 0.2769, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.1811631497683994, | |
| "grad_norm": 3.6543190479278564, | |
| "learning_rate": 1.2222222222222224e-05, | |
| "loss": 0.2596, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.19763252702007206, | |
| "grad_norm": 3.7170989513397217, | |
| "learning_rate": 1.3333333333333333e-05, | |
| "loss": 0.2532, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.21410190427174472, | |
| "grad_norm": 2.963078498840332, | |
| "learning_rate": 1.4444444444444446e-05, | |
| "loss": 0.241, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.2305712815234174, | |
| "grad_norm": 3.8106493949890137, | |
| "learning_rate": 1.555555555555556e-05, | |
| "loss": 0.2355, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.24704065877509007, | |
| "grad_norm": 3.3100316524505615, | |
| "learning_rate": 1.6666666666666667e-05, | |
| "loss": 0.227, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.26351003602676276, | |
| "grad_norm": 3.8181209564208984, | |
| "learning_rate": 1.7777777777777777e-05, | |
| "loss": 0.222, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.2799794132784354, | |
| "grad_norm": 3.1601641178131104, | |
| "learning_rate": 1.888888888888889e-05, | |
| "loss": 0.216, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.2964487905301081, | |
| "grad_norm": 3.8540680408477783, | |
| "learning_rate": 2e-05, | |
| "loss": 0.2219, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.3129181677817808, | |
| "grad_norm": 2.4376087188720703, | |
| "learning_rate": 1.9998119704485016e-05, | |
| "loss": 0.2124, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.3293875450334534, | |
| "grad_norm": 2.0443239212036133, | |
| "learning_rate": 1.9992479525042305e-05, | |
| "loss": 0.2072, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.3458569222851261, | |
| "grad_norm": 2.7866272926330566, | |
| "learning_rate": 1.9983081582712684e-05, | |
| "loss": 0.2023, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.3623262995367988, | |
| "grad_norm": 2.3178961277008057, | |
| "learning_rate": 1.996992941167792e-05, | |
| "loss": 0.2013, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.3787956767884714, | |
| "grad_norm": 2.5984015464782715, | |
| "learning_rate": 1.9953027957931658e-05, | |
| "loss": 0.1933, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.3952650540401441, | |
| "grad_norm": 1.8709274530410767, | |
| "learning_rate": 1.9932383577419432e-05, | |
| "loss": 0.1945, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.4117344312918168, | |
| "grad_norm": 2.1134493350982666, | |
| "learning_rate": 1.9908004033648452e-05, | |
| "loss": 0.1916, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.42820380854348944, | |
| "grad_norm": 2.076616048812866, | |
| "learning_rate": 1.9879898494768093e-05, | |
| "loss": 0.1902, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.44467318579516213, | |
| "grad_norm": 2.05837082862854, | |
| "learning_rate": 1.9848077530122083e-05, | |
| "loss": 0.1849, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.4611425630468348, | |
| "grad_norm": 1.7282419204711914, | |
| "learning_rate": 1.9812553106273848e-05, | |
| "loss": 0.1843, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.47761194029850745, | |
| "grad_norm": 2.2622146606445312, | |
| "learning_rate": 1.9773338582506357e-05, | |
| "loss": 0.1872, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.49408131755018014, | |
| "grad_norm": 1.5245953798294067, | |
| "learning_rate": 1.973044870579824e-05, | |
| "loss": 0.182, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.5105506948018528, | |
| "grad_norm": 1.590587854385376, | |
| "learning_rate": 1.9683899605278062e-05, | |
| "loss": 0.1767, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.5270200720535255, | |
| "grad_norm": 1.6935125589370728, | |
| "learning_rate": 1.9633708786158803e-05, | |
| "loss": 0.1739, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.5434894493051982, | |
| "grad_norm": 1.6206973791122437, | |
| "learning_rate": 1.957989512315489e-05, | |
| "loss": 0.1771, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.5599588265568708, | |
| "grad_norm": 1.561724066734314, | |
| "learning_rate": 1.9522478853384154e-05, | |
| "loss": 0.1706, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.5764282038085435, | |
| "grad_norm": 1.4770928621292114, | |
| "learning_rate": 1.946148156875751e-05, | |
| "loss": 0.1711, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.5928975810602162, | |
| "grad_norm": 1.5764074325561523, | |
| "learning_rate": 1.9396926207859085e-05, | |
| "loss": 0.1756, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.6093669583118888, | |
| "grad_norm": 2.4715301990509033, | |
| "learning_rate": 1.932883704732001e-05, | |
| "loss": 0.1757, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.6258363355635616, | |
| "grad_norm": 1.3919658660888672, | |
| "learning_rate": 1.9257239692688907e-05, | |
| "loss": 0.1658, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.6423057128152342, | |
| "grad_norm": 2.575564384460449, | |
| "learning_rate": 1.9182161068802742e-05, | |
| "loss": 0.1768, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.6587750900669068, | |
| "grad_norm": 1.6217772960662842, | |
| "learning_rate": 1.9103629409661468e-05, | |
| "loss": 0.1695, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.6752444673185796, | |
| "grad_norm": 2.0268895626068115, | |
| "learning_rate": 1.902167424781038e-05, | |
| "loss": 0.1749, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.6917138445702522, | |
| "grad_norm": 1.6548528671264648, | |
| "learning_rate": 1.8936326403234125e-05, | |
| "loss": 0.1723, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.7081832218219248, | |
| "grad_norm": 1.6716668605804443, | |
| "learning_rate": 1.8847617971766577e-05, | |
| "loss": 0.1724, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.7246525990735976, | |
| "grad_norm": 1.6575418710708618, | |
| "learning_rate": 1.8755582313020912e-05, | |
| "loss": 0.1701, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.7411219763252702, | |
| "grad_norm": 1.429405689239502, | |
| "learning_rate": 1.866025403784439e-05, | |
| "loss": 0.1665, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.7575913535769428, | |
| "grad_norm": 1.5980753898620605, | |
| "learning_rate": 1.8561668995302668e-05, | |
| "loss": 0.1677, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.7740607308286156, | |
| "grad_norm": 1.3052290678024292, | |
| "learning_rate": 1.845986425919841e-05, | |
| "loss": 0.1674, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.7905301080802882, | |
| "grad_norm": 1.3834303617477417, | |
| "learning_rate": 1.8354878114129368e-05, | |
| "loss": 0.1621, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.8069994853319609, | |
| "grad_norm": 1.314771294593811, | |
| "learning_rate": 1.824675004109107e-05, | |
| "loss": 0.1603, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.8234688625836336, | |
| "grad_norm": 1.255486249923706, | |
| "learning_rate": 1.8135520702629677e-05, | |
| "loss": 0.1602, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.8399382398353062, | |
| "grad_norm": 1.1701654195785522, | |
| "learning_rate": 1.802123192755044e-05, | |
| "loss": 0.1595, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.8564076170869789, | |
| "grad_norm": 1.1197280883789062, | |
| "learning_rate": 1.7903926695187595e-05, | |
| "loss": 0.1589, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.8728769943386516, | |
| "grad_norm": 1.26057767868042, | |
| "learning_rate": 1.7783649119241603e-05, | |
| "loss": 0.1627, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.8893463715903243, | |
| "grad_norm": 1.1180353164672852, | |
| "learning_rate": 1.766044443118978e-05, | |
| "loss": 0.1577, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.9058157488419969, | |
| "grad_norm": 1.1337931156158447, | |
| "learning_rate": 1.7534358963276606e-05, | |
| "loss": 0.1563, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.9222851260936696, | |
| "grad_norm": 1.1284104585647583, | |
| "learning_rate": 1.740544013109005e-05, | |
| "loss": 0.1552, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.9387545033453423, | |
| "grad_norm": 1.1995372772216797, | |
| "learning_rate": 1.7273736415730488e-05, | |
| "loss": 0.1571, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.9552238805970149, | |
| "grad_norm": 1.189568042755127, | |
| "learning_rate": 1.7139297345578992e-05, | |
| "loss": 0.1543, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.9716932578486875, | |
| "grad_norm": 1.1965142488479614, | |
| "learning_rate": 1.7002173477671685e-05, | |
| "loss": 0.1601, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.9881626351003603, | |
| "grad_norm": 1.1370137929916382, | |
| "learning_rate": 1.686241637868734e-05, | |
| "loss": 0.1554, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.004632012352033, | |
| "grad_norm": 1.208176612854004, | |
| "learning_rate": 1.6720078605555227e-05, | |
| "loss": 0.1458, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 1.0211013896037056, | |
| "grad_norm": 1.1520111560821533, | |
| "learning_rate": 1.657521368569064e-05, | |
| "loss": 0.1315, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 1.0375707668553782, | |
| "grad_norm": 1.1699644327163696, | |
| "learning_rate": 1.6427876096865394e-05, | |
| "loss": 0.1289, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 1.054040144107051, | |
| "grad_norm": 1.2921233177185059, | |
| "learning_rate": 1.627812124672099e-05, | |
| "loss": 0.1292, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 1.0705095213587237, | |
| "grad_norm": 1.1183018684387207, | |
| "learning_rate": 1.6126005451932028e-05, | |
| "loss": 0.1301, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 1.0869788986103963, | |
| "grad_norm": 1.132805347442627, | |
| "learning_rate": 1.5971585917027864e-05, | |
| "loss": 0.1291, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 1.103448275862069, | |
| "grad_norm": 1.1662507057189941, | |
| "learning_rate": 1.5814920712880267e-05, | |
| "loss": 0.1296, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 1.1199176531137416, | |
| "grad_norm": 1.230918288230896, | |
| "learning_rate": 1.5656068754865388e-05, | |
| "loss": 0.1283, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 1.1363870303654142, | |
| "grad_norm": 1.111428141593933, | |
| "learning_rate": 1.5495089780708062e-05, | |
| "loss": 0.1264, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 1.152856407617087, | |
| "grad_norm": 1.118282437324524, | |
| "learning_rate": 1.5332044328016916e-05, | |
| "loss": 0.1326, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 1.1693257848687597, | |
| "grad_norm": 1.0411953926086426, | |
| "learning_rate": 1.5166993711518631e-05, | |
| "loss": 0.1273, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 1.1857951621204323, | |
| "grad_norm": 1.0174416303634644, | |
| "learning_rate": 1.5000000000000002e-05, | |
| "loss": 0.1236, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 1.202264539372105, | |
| "grad_norm": 1.1264220476150513, | |
| "learning_rate": 1.4831125992966386e-05, | |
| "loss": 0.1262, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 1.2187339166237776, | |
| "grad_norm": 1.1279516220092773, | |
| "learning_rate": 1.4660435197025391e-05, | |
| "loss": 0.1259, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 1.2352032938754502, | |
| "grad_norm": 0.9918281435966492, | |
| "learning_rate": 1.4487991802004625e-05, | |
| "loss": 0.1263, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 1.2516726711271229, | |
| "grad_norm": 1.0115736722946167, | |
| "learning_rate": 1.4313860656812537e-05, | |
| "loss": 0.1281, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 1.2681420483787957, | |
| "grad_norm": 0.993411123752594, | |
| "learning_rate": 1.4138107245051394e-05, | |
| "loss": 0.1287, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 1.2846114256304684, | |
| "grad_norm": 1.0473469495773315, | |
| "learning_rate": 1.396079766039157e-05, | |
| "loss": 0.1272, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 1.301080802882141, | |
| "grad_norm": 0.9457677006721497, | |
| "learning_rate": 1.3781998581716427e-05, | |
| "loss": 0.125, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 1.3175501801338136, | |
| "grad_norm": 0.9976021647453308, | |
| "learning_rate": 1.3601777248047105e-05, | |
| "loss": 0.1256, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.3340195573854863, | |
| "grad_norm": 1.0927587747573853, | |
| "learning_rate": 1.342020143325669e-05, | |
| "loss": 0.1265, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 1.3504889346371591, | |
| "grad_norm": 1.0087074041366577, | |
| "learning_rate": 1.3237339420583213e-05, | |
| "loss": 0.1252, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 1.3669583118888318, | |
| "grad_norm": 1.046044945716858, | |
| "learning_rate": 1.3053259976951134e-05, | |
| "loss": 0.1247, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 1.3834276891405044, | |
| "grad_norm": 1.0403574705123901, | |
| "learning_rate": 1.2868032327110904e-05, | |
| "loss": 0.1254, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 1.399897066392177, | |
| "grad_norm": 0.9709586501121521, | |
| "learning_rate": 1.2681726127606374e-05, | |
| "loss": 0.1295, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 1.4163664436438497, | |
| "grad_norm": 0.9234594106674194, | |
| "learning_rate": 1.2494411440579814e-05, | |
| "loss": 0.1226, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 1.4328358208955223, | |
| "grad_norm": 1.01310133934021, | |
| "learning_rate": 1.2306158707424402e-05, | |
| "loss": 0.1243, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 1.449305198147195, | |
| "grad_norm": 0.977340579032898, | |
| "learning_rate": 1.211703872229411e-05, | |
| "loss": 0.1252, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 1.4657745753988678, | |
| "grad_norm": 0.9476578235626221, | |
| "learning_rate": 1.1927122605480899e-05, | |
| "loss": 0.1235, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 1.4822439526505404, | |
| "grad_norm": 0.9919003248214722, | |
| "learning_rate": 1.1736481776669307e-05, | |
| "loss": 0.1238, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 1.498713329902213, | |
| "grad_norm": 0.9731943011283875, | |
| "learning_rate": 1.1545187928078407e-05, | |
| "loss": 0.1269, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 1.5151827071538857, | |
| "grad_norm": 0.9613791108131409, | |
| "learning_rate": 1.1353312997501313e-05, | |
| "loss": 0.1278, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 1.5316520844055583, | |
| "grad_norm": 0.9685577750205994, | |
| "learning_rate": 1.1160929141252303e-05, | |
| "loss": 0.1255, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 1.5481214616572312, | |
| "grad_norm": 0.9065141081809998, | |
| "learning_rate": 1.0968108707031792e-05, | |
| "loss": 0.1229, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 1.5645908389089036, | |
| "grad_norm": 0.9448639750480652, | |
| "learning_rate": 1.077492420671931e-05, | |
| "loss": 0.1228, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 1.5810602161605765, | |
| "grad_norm": 0.91968834400177, | |
| "learning_rate": 1.0581448289104759e-05, | |
| "loss": 0.1238, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 1.597529593412249, | |
| "grad_norm": 0.9806727170944214, | |
| "learning_rate": 1.038775371256817e-05, | |
| "loss": 0.1256, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 1.6139989706639217, | |
| "grad_norm": 0.9167839884757996, | |
| "learning_rate": 1.0193913317718245e-05, | |
| "loss": 0.1249, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 1.6304683479155946, | |
| "grad_norm": 0.9122770428657532, | |
| "learning_rate": 1e-05, | |
| "loss": 0.1254, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 1.646937725167267, | |
| "grad_norm": 0.9491678476333618, | |
| "learning_rate": 9.806086682281759e-06, | |
| "loss": 0.1216, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 1.6634071024189399, | |
| "grad_norm": 0.8831928968429565, | |
| "learning_rate": 9.612246287431832e-06, | |
| "loss": 0.1212, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 1.6798764796706125, | |
| "grad_norm": 0.8818361759185791, | |
| "learning_rate": 9.418551710895243e-06, | |
| "loss": 0.1206, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 1.6963458569222851, | |
| "grad_norm": 0.9016006588935852, | |
| "learning_rate": 9.225075793280693e-06, | |
| "loss": 0.1218, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 1.7128152341739578, | |
| "grad_norm": 0.9079029560089111, | |
| "learning_rate": 9.03189129296821e-06, | |
| "loss": 0.1198, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 1.7292846114256304, | |
| "grad_norm": 0.9520952105522156, | |
| "learning_rate": 8.839070858747697e-06, | |
| "loss": 0.123, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 1.7457539886773032, | |
| "grad_norm": 0.9180439114570618, | |
| "learning_rate": 8.646687002498692e-06, | |
| "loss": 0.1213, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 1.7622233659289757, | |
| "grad_norm": 0.8695424199104309, | |
| "learning_rate": 8.454812071921597e-06, | |
| "loss": 0.1205, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 1.7786927431806485, | |
| "grad_norm": 0.8801153302192688, | |
| "learning_rate": 8.263518223330698e-06, | |
| "loss": 0.1169, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 1.7951621204323212, | |
| "grad_norm": 0.8642145395278931, | |
| "learning_rate": 8.072877394519103e-06, | |
| "loss": 0.1176, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 1.8116314976839938, | |
| "grad_norm": 0.8979960680007935, | |
| "learning_rate": 7.882961277705897e-06, | |
| "loss": 0.1204, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 1.8281008749356666, | |
| "grad_norm": 0.820036768913269, | |
| "learning_rate": 7.6938412925756e-06, | |
| "loss": 0.1213, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 1.844570252187339, | |
| "grad_norm": 0.8929469585418701, | |
| "learning_rate": 7.505588559420188e-06, | |
| "loss": 0.1186, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 1.861039629439012, | |
| "grad_norm": 0.8560171127319336, | |
| "learning_rate": 7.3182738723936255e-06, | |
| "loss": 0.1207, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 1.8775090066906845, | |
| "grad_norm": 0.8577896952629089, | |
| "learning_rate": 7.131967672889101e-06, | |
| "loss": 0.119, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 1.8939783839423572, | |
| "grad_norm": 0.8649168610572815, | |
| "learning_rate": 6.94674002304887e-06, | |
| "loss": 0.1171, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 1.9104477611940298, | |
| "grad_norm": 0.8647825121879578, | |
| "learning_rate": 6.762660579416791e-06, | |
| "loss": 0.119, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 1.9269171384457024, | |
| "grad_norm": 0.8913580775260925, | |
| "learning_rate": 6.579798566743314e-06, | |
| "loss": 0.1178, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 1.9433865156973753, | |
| "grad_norm": 0.8711913824081421, | |
| "learning_rate": 6.3982227519528986e-06, | |
| "loss": 0.1177, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 1.9598558929490477, | |
| "grad_norm": 0.887152910232544, | |
| "learning_rate": 6.218001418283577e-06, | |
| "loss": 0.1208, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 1.9763252702007206, | |
| "grad_norm": 0.9313440918922424, | |
| "learning_rate": 6.039202339608432e-06, | |
| "loss": 0.1174, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 1.9927946474523932, | |
| "grad_norm": 0.971422553062439, | |
| "learning_rate": 5.8618927549486095e-06, | |
| "loss": 0.1205, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 2.009264024704066, | |
| "grad_norm": 0.9346736669540405, | |
| "learning_rate": 5.686139343187468e-06, | |
| "loss": 0.1045, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 2.0257334019557387, | |
| "grad_norm": 0.9971660375595093, | |
| "learning_rate": 5.512008197995379e-06, | |
| "loss": 0.0927, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 2.042202779207411, | |
| "grad_norm": 0.8701250553131104, | |
| "learning_rate": 5.339564802974615e-06, | |
| "loss": 0.0917, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 2.058672156459084, | |
| "grad_norm": 0.9550381898880005, | |
| "learning_rate": 5.168874007033615e-06, | |
| "loss": 0.089, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 2.0751415337107564, | |
| "grad_norm": 0.985183835029602, | |
| "learning_rate": 5.000000000000003e-06, | |
| "loss": 0.0878, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 2.0916109109624292, | |
| "grad_norm": 1.021644949913025, | |
| "learning_rate": 4.8330062884813714e-06, | |
| "loss": 0.0877, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 2.108080288214102, | |
| "grad_norm": 0.967758297920227, | |
| "learning_rate": 4.66795567198309e-06, | |
| "loss": 0.0888, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 2.1245496654657745, | |
| "grad_norm": 0.9479052424430847, | |
| "learning_rate": 4.504910219291941e-06, | |
| "loss": 0.0893, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 2.1410190427174474, | |
| "grad_norm": 0.8988873362541199, | |
| "learning_rate": 4.343931245134616e-06, | |
| "loss": 0.0885, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 2.1574884199691198, | |
| "grad_norm": 0.8895907402038574, | |
| "learning_rate": 4.185079287119733e-06, | |
| "loss": 0.0875, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 2.1739577972207926, | |
| "grad_norm": 0.8918712139129639, | |
| "learning_rate": 4.028414082972141e-06, | |
| "loss": 0.0874, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 2.1904271744724655, | |
| "grad_norm": 0.8613003492355347, | |
| "learning_rate": 3.873994548067972e-06, | |
| "loss": 0.0864, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 2.206896551724138, | |
| "grad_norm": 0.8328737020492554, | |
| "learning_rate": 3.7218787532790167e-06, | |
| "loss": 0.0869, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 2.2233659289758108, | |
| "grad_norm": 0.8785966038703918, | |
| "learning_rate": 3.5721239031346067e-06, | |
| "loss": 0.0866, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 2.239835306227483, | |
| "grad_norm": 0.9229673743247986, | |
| "learning_rate": 3.424786314309365e-06, | |
| "loss": 0.0865, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 2.256304683479156, | |
| "grad_norm": 0.9559910893440247, | |
| "learning_rate": 3.279921394444776e-06, | |
| "loss": 0.0865, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 2.2727740607308284, | |
| "grad_norm": 0.9126923680305481, | |
| "learning_rate": 3.1375836213126653e-06, | |
| "loss": 0.0853, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 2.2892434379825013, | |
| "grad_norm": 0.8650088310241699, | |
| "learning_rate": 2.9978265223283152e-06, | |
| "loss": 0.0859, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 2.305712815234174, | |
| "grad_norm": 0.8619425296783447, | |
| "learning_rate": 2.8607026544210115e-06, | |
| "loss": 0.0854, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 2.3221821924858466, | |
| "grad_norm": 0.930046558380127, | |
| "learning_rate": 2.726263584269513e-06, | |
| "loss": 0.088, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 2.3386515697375194, | |
| "grad_norm": 0.8604719638824463, | |
| "learning_rate": 2.594559868909956e-06, | |
| "loss": 0.0853, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 2.355120946989192, | |
| "grad_norm": 0.8624026775360107, | |
| "learning_rate": 2.4656410367233928e-06, | |
| "loss": 0.0856, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 2.3715903242408647, | |
| "grad_norm": 0.8655623197555542, | |
| "learning_rate": 2.339555568810221e-06, | |
| "loss": 0.087, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 2.388059701492537, | |
| "grad_norm": 0.8891249895095825, | |
| "learning_rate": 2.2163508807584e-06, | |
| "loss": 0.0875, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 2.40452907874421, | |
| "grad_norm": 0.8799765706062317, | |
| "learning_rate": 2.0960733048124082e-06, | |
| "loss": 0.0861, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 2.420998455995883, | |
| "grad_norm": 0.8611019849777222, | |
| "learning_rate": 1.9787680724495617e-06, | |
| "loss": 0.0849, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 2.4374678332475552, | |
| "grad_norm": 0.8410822749137878, | |
| "learning_rate": 1.8644792973703252e-06, | |
| "loss": 0.085, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 2.453937210499228, | |
| "grad_norm": 0.9146021604537964, | |
| "learning_rate": 1.7532499589089324e-06, | |
| "loss": 0.0852, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 2.4704065877509005, | |
| "grad_norm": 0.875053346157074, | |
| "learning_rate": 1.6451218858706374e-06, | |
| "loss": 0.0852, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 2.4868759650025734, | |
| "grad_norm": 0.8440834283828735, | |
| "learning_rate": 1.5401357408015893e-06, | |
| "loss": 0.086, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 2.5033453422542458, | |
| "grad_norm": 0.8311520218849182, | |
| "learning_rate": 1.4383310046973365e-06, | |
| "loss": 0.0849, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 2.5198147195059186, | |
| "grad_norm": 0.8406962752342224, | |
| "learning_rate": 1.339745962155613e-06, | |
| "loss": 0.083, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 2.5362840967575915, | |
| "grad_norm": 0.8209373354911804, | |
| "learning_rate": 1.2444176869790925e-06, | |
| "loss": 0.0813, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 2.552753474009264, | |
| "grad_norm": 0.8779584169387817, | |
| "learning_rate": 1.152382028233422e-06, | |
| "loss": 0.0868, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 2.5692228512609367, | |
| "grad_norm": 0.8810158967971802, | |
| "learning_rate": 1.0636735967658785e-06, | |
| "loss": 0.0865, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 2.5856922285126096, | |
| "grad_norm": 0.865736722946167, | |
| "learning_rate": 9.783257521896228e-07, | |
| "loss": 0.0846, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 2.602161605764282, | |
| "grad_norm": 0.845588207244873, | |
| "learning_rate": 8.963705903385344e-07, | |
| "loss": 0.0847, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 2.618630983015955, | |
| "grad_norm": 0.815373420715332, | |
| "learning_rate": 8.178389311972612e-07, | |
| "loss": 0.0845, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 2.6351003602676273, | |
| "grad_norm": 0.7676054835319519, | |
| "learning_rate": 7.427603073110967e-07, | |
| "loss": 0.0833, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 2.6515697375193, | |
| "grad_norm": 0.7454907894134521, | |
| "learning_rate": 6.711629526799946e-07, | |
| "loss": 0.0819, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 2.6680391147709726, | |
| "grad_norm": 0.7519494891166687, | |
| "learning_rate": 6.030737921409169e-07, | |
| "loss": 0.0847, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 2.6845084920226454, | |
| "grad_norm": 0.7272472977638245, | |
| "learning_rate": 5.385184312424973e-07, | |
| "loss": 0.0823, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 2.7009778692743183, | |
| "grad_norm": 0.7485241293907166, | |
| "learning_rate": 4.775211466158469e-07, | |
| "loss": 0.0865, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 2.7174472465259907, | |
| "grad_norm": 0.7801215052604675, | |
| "learning_rate": 4.2010487684511105e-07, | |
| "loss": 0.0858, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 2.7339166237776635, | |
| "grad_norm": 0.7571792602539062, | |
| "learning_rate": 3.662912138411967e-07, | |
| "loss": 0.083, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 2.750386001029336, | |
| "grad_norm": 0.7191453576087952, | |
| "learning_rate": 3.161003947219421e-07, | |
| "loss": 0.0833, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 2.766855378281009, | |
| "grad_norm": 0.7716163992881775, | |
| "learning_rate": 2.6955129420176193e-07, | |
| "loss": 0.0841, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 2.783324755532681, | |
| "grad_norm": 0.7539470195770264, | |
| "learning_rate": 2.2666141749364434e-07, | |
| "loss": 0.0836, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 2.799794132784354, | |
| "grad_norm": 0.7242315411567688, | |
| "learning_rate": 1.874468937261531e-07, | |
| "loss": 0.0821, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 2.816263510036027, | |
| "grad_norm": 0.7200111746788025, | |
| "learning_rate": 1.519224698779198e-07, | |
| "loss": 0.0837, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 2.8327328872876993, | |
| "grad_norm": 0.7370088696479797, | |
| "learning_rate": 1.201015052319099e-07, | |
| "loss": 0.084, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 2.849202264539372, | |
| "grad_norm": 0.7719192504882812, | |
| "learning_rate": 9.199596635154684e-08, | |
| "loss": 0.0837, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 2.8656716417910446, | |
| "grad_norm": 0.7894994020462036, | |
| "learning_rate": 6.761642258056977e-08, | |
| "loss": 0.0835, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 2.8821410190427175, | |
| "grad_norm": 0.7965743541717529, | |
| "learning_rate": 4.6972042068341714e-08, | |
| "loss": 0.0823, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 2.89861039629439, | |
| "grad_norm": 0.7946169376373291, | |
| "learning_rate": 3.0070588322079765e-08, | |
| "loss": 0.0806, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 2.9150797735460627, | |
| "grad_norm": 0.8272018432617188, | |
| "learning_rate": 1.6918417287318245e-08, | |
| "loss": 0.0828, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 2.9315491507977356, | |
| "grad_norm": 0.8682279586791992, | |
| "learning_rate": 7.520474957699586e-09, | |
| "loss": 0.0849, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 2.948018528049408, | |
| "grad_norm": 0.834924578666687, | |
| "learning_rate": 1.8802955149865854e-09, | |
| "loss": 0.0837, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 2.964487905301081, | |
| "grad_norm": 0.8157399892807007, | |
| "learning_rate": 0.0, | |
| "loss": 0.0827, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 2.964487905301081, | |
| "step": 180, | |
| "total_flos": 8.05255292248916e+18, | |
| "train_loss": 0.14365306976768705, | |
| "train_runtime": 16152.8942, | |
| "train_samples_per_second": 5.774, | |
| "train_steps_per_second": 0.011 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 180, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 8.05255292248916e+18, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |