Invalid JSON: Unexpected token 'N', ..."on_loss": NaN,
"... is not valid JSON
| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 4.0, | |
| "eval_steps": 2000.0, | |
| "global_step": 412, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.009708737864077669, | |
| "grad_norm": 5.575473016079204, | |
| "kl": 0.0, | |
| "learning_rate": 5.000000000000001e-07, | |
| "loss": 0.4845, | |
| "step": 1, | |
| "step_loss": 0.490234375 | |
| }, | |
| { | |
| "epoch": 0.019417475728155338, | |
| "grad_norm": 9.7143780359321, | |
| "kl": 0.0020599365234375, | |
| "learning_rate": 2.438044511330269e-06, | |
| "loss": 0.4845, | |
| "step": 2, | |
| "step_loss": 0.48046875 | |
| }, | |
| { | |
| "epoch": 0.02912621359223301, | |
| "grad_norm": 7.3775865577265884, | |
| "kl": 0.002471923828125, | |
| "learning_rate": 3.5717278751869343e-06, | |
| "loss": 0.4945, | |
| "step": 3, | |
| "step_loss": 0.546875 | |
| }, | |
| { | |
| "epoch": 0.038834951456310676, | |
| "grad_norm": 15.596262146563706, | |
| "kl": 0.002349853515625, | |
| "learning_rate": 4.376089022660538e-06, | |
| "loss": 0.4586, | |
| "step": 4, | |
| "step_loss": 0.32421875 | |
| }, | |
| { | |
| "epoch": 0.04854368932038835, | |
| "grad_norm": 6.642139360876633, | |
| "kl": 0.021240234375, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4483, | |
| "step": 5, | |
| "step_loss": 0.71875 | |
| }, | |
| { | |
| "epoch": 0.05825242718446602, | |
| "grad_norm": 10.4028109346367, | |
| "kl": 0.037353515625, | |
| "learning_rate": 4.999957311534636e-06, | |
| "loss": 0.4224, | |
| "step": 6, | |
| "step_loss": 0.306640625 | |
| }, | |
| { | |
| "epoch": 0.06796116504854369, | |
| "grad_norm": 8.965364568204704, | |
| "kl": 0.043212890625, | |
| "learning_rate": 4.99982924775837e-06, | |
| "loss": 0.3771, | |
| "step": 7, | |
| "step_loss": 0.361328125 | |
| }, | |
| { | |
| "epoch": 0.07766990291262135, | |
| "grad_norm": 5.86329979229663, | |
| "kl": 0.06396484375, | |
| "learning_rate": 4.999615813530619e-06, | |
| "loss": 0.372, | |
| "step": 8, | |
| "step_loss": 0.345703125 | |
| }, | |
| { | |
| "epoch": 0.08737864077669903, | |
| "grad_norm": 7.680252752913332, | |
| "kl": 0.07763671875, | |
| "learning_rate": 4.999317016950212e-06, | |
| "loss": 0.4717, | |
| "step": 9, | |
| "step_loss": 0.482421875 | |
| }, | |
| { | |
| "epoch": 0.0970873786407767, | |
| "grad_norm": 7.28382768968269, | |
| "kl": 0.12060546875, | |
| "learning_rate": 4.998932869355074e-06, | |
| "loss": 0.3988, | |
| "step": 10, | |
| "step_loss": 0.259765625 | |
| }, | |
| { | |
| "epoch": 0.10679611650485436, | |
| "grad_norm": 3.742924436507147, | |
| "kl": 0.07958984375, | |
| "learning_rate": 4.998463385321802e-06, | |
| "loss": 0.3597, | |
| "step": 11, | |
| "step_loss": 0.42578125 | |
| }, | |
| { | |
| "epoch": 0.11650485436893204, | |
| "grad_norm": 2.077338507259391, | |
| "kl": 0.0986328125, | |
| "learning_rate": 4.997908582665111e-06, | |
| "loss": 0.3828, | |
| "step": 12, | |
| "step_loss": 0.2197265625 | |
| }, | |
| { | |
| "epoch": 0.1262135922330097, | |
| "grad_norm": 2.0936762099287187, | |
| "kl": 0.2578125, | |
| "learning_rate": 4.997268482437153e-06, | |
| "loss": 0.3686, | |
| "step": 13, | |
| "step_loss": 0.40234375 | |
| }, | |
| { | |
| "epoch": 0.13592233009708737, | |
| "grad_norm": 1.5539690974487896, | |
| "kl": 0.11572265625, | |
| "learning_rate": 4.9965431089267265e-06, | |
| "loss": 0.3481, | |
| "step": 14, | |
| "step_loss": 0.318359375 | |
| }, | |
| { | |
| "epoch": 0.14563106796116504, | |
| "grad_norm": 11.763990149324268, | |
| "kl": 0.19921875, | |
| "learning_rate": 4.9957324896583495e-06, | |
| "loss": 0.3518, | |
| "step": 15, | |
| "step_loss": 0.419921875 | |
| }, | |
| { | |
| "epoch": 0.1553398058252427, | |
| "grad_norm": 1.6021280918226926, | |
| "kl": 0.2080078125, | |
| "learning_rate": 4.9948366553912146e-06, | |
| "loss": 0.3468, | |
| "step": 16, | |
| "step_loss": 0.498046875 | |
| }, | |
| { | |
| "epoch": 0.1650485436893204, | |
| "grad_norm": 2.243950658979944, | |
| "kl": 0.1279296875, | |
| "learning_rate": 4.993855640118024e-06, | |
| "loss": 0.3336, | |
| "step": 17, | |
| "step_loss": 0.267578125 | |
| }, | |
| { | |
| "epoch": 0.17475728155339806, | |
| "grad_norm": 1.5344029876580454, | |
| "kl": 0.2216796875, | |
| "learning_rate": 4.992789481063699e-06, | |
| "loss": 0.3456, | |
| "step": 18, | |
| "step_loss": 0.3671875 | |
| }, | |
| { | |
| "epoch": 0.18446601941747573, | |
| "grad_norm": 1.7390327115176747, | |
| "kl": 0.1259765625, | |
| "learning_rate": 4.9916382186839665e-06, | |
| "loss": 0.3427, | |
| "step": 19, | |
| "step_loss": 0.298828125 | |
| }, | |
| { | |
| "epoch": 0.1941747572815534, | |
| "grad_norm": 5.2081774675570855, | |
| "kl": 0.1591796875, | |
| "learning_rate": 4.990401896663829e-06, | |
| "loss": 0.2956, | |
| "step": 20, | |
| "step_loss": 0.2353515625 | |
| }, | |
| { | |
| "epoch": 0.20388349514563106, | |
| "grad_norm": 2.087334097477202, | |
| "kl": 0.169921875, | |
| "learning_rate": 4.989080561915895e-06, | |
| "loss": 0.3336, | |
| "step": 21, | |
| "step_loss": 0.2109375 | |
| }, | |
| { | |
| "epoch": 0.21359223300970873, | |
| "grad_norm": 1.8894907173358744, | |
| "kl": 0.1845703125, | |
| "learning_rate": 4.987674264578615e-06, | |
| "loss": 0.2974, | |
| "step": 22, | |
| "step_loss": 0.2314453125 | |
| }, | |
| { | |
| "epoch": 0.22330097087378642, | |
| "grad_norm": 3.672752113284032, | |
| "kl": 0.1748046875, | |
| "learning_rate": 4.9861830580143665e-06, | |
| "loss": 0.3722, | |
| "step": 23, | |
| "step_loss": 0.255859375 | |
| }, | |
| { | |
| "epoch": 0.23300970873786409, | |
| "grad_norm": 1.0485056233907877, | |
| "kl": 0.1494140625, | |
| "learning_rate": 4.984606998807432e-06, | |
| "loss": 0.2997, | |
| "step": 24, | |
| "step_loss": 0.396484375 | |
| }, | |
| { | |
| "epoch": 0.24271844660194175, | |
| "grad_norm": 1.3694786710161069, | |
| "kl": 0.11474609375, | |
| "learning_rate": 4.982946146761856e-06, | |
| "loss": 0.2984, | |
| "step": 25, | |
| "step_loss": 0.1396484375 | |
| }, | |
| { | |
| "epoch": 0.2524271844660194, | |
| "grad_norm": 1.0149380047661418, | |
| "kl": 0.2314453125, | |
| "learning_rate": 4.981200564899172e-06, | |
| "loss": 0.2926, | |
| "step": 26, | |
| "step_loss": 0.34375 | |
| }, | |
| { | |
| "epoch": 0.2621359223300971, | |
| "grad_norm": 3.3346882609517627, | |
| "kl": 0.1630859375, | |
| "learning_rate": 4.979370319456011e-06, | |
| "loss": 0.325, | |
| "step": 27, | |
| "step_loss": 0.2197265625 | |
| }, | |
| { | |
| "epoch": 0.27184466019417475, | |
| "grad_norm": 1.1026809101068904, | |
| "kl": 0.1171875, | |
| "learning_rate": 4.977455479881591e-06, | |
| "loss": 0.3124, | |
| "step": 28, | |
| "step_loss": 0.42578125 | |
| }, | |
| { | |
| "epoch": 0.2815533980582524, | |
| "grad_norm": 1.4388631908391791, | |
| "kl": 0.2470703125, | |
| "learning_rate": 4.975456118835079e-06, | |
| "loss": 0.2766, | |
| "step": 29, | |
| "step_loss": 0.2275390625 | |
| }, | |
| { | |
| "epoch": 0.2912621359223301, | |
| "grad_norm": 1.654229929536635, | |
| "kl": 0.23828125, | |
| "learning_rate": 4.973372312182835e-06, | |
| "loss": 0.3192, | |
| "step": 30, | |
| "step_loss": 0.5546875 | |
| }, | |
| { | |
| "epoch": 0.30097087378640774, | |
| "grad_norm": 1.0784118934051907, | |
| "kl": 0.138671875, | |
| "learning_rate": 4.971204138995531e-06, | |
| "loss": 0.2877, | |
| "step": 31, | |
| "step_loss": 0.26171875 | |
| }, | |
| { | |
| "epoch": 0.3106796116504854, | |
| "grad_norm": 1.789773991131577, | |
| "kl": 0.2138671875, | |
| "learning_rate": 4.968951681545156e-06, | |
| "loss": 0.3597, | |
| "step": 32, | |
| "step_loss": 0.23828125 | |
| }, | |
| { | |
| "epoch": 0.32038834951456313, | |
| "grad_norm": 1.3349882623963718, | |
| "kl": 0.1240234375, | |
| "learning_rate": 4.96661502530189e-06, | |
| "loss": 0.2909, | |
| "step": 33, | |
| "step_loss": 0.2431640625 | |
| }, | |
| { | |
| "epoch": 0.3300970873786408, | |
| "grad_norm": 0.8694048190676366, | |
| "kl": 0.12109375, | |
| "learning_rate": 4.96419425893086e-06, | |
| "loss": 0.2964, | |
| "step": 34, | |
| "step_loss": 0.25 | |
| }, | |
| { | |
| "epoch": 0.33980582524271846, | |
| "grad_norm": 1.247182149510005, | |
| "kl": 0.1650390625, | |
| "learning_rate": 4.96168947428878e-06, | |
| "loss": 0.3136, | |
| "step": 35, | |
| "step_loss": 0.25 | |
| }, | |
| { | |
| "epoch": 0.34951456310679613, | |
| "grad_norm": 0.9312530287110918, | |
| "kl": 0.1630859375, | |
| "learning_rate": 4.959100766420458e-06, | |
| "loss": 0.3164, | |
| "step": 36, | |
| "step_loss": 0.306640625 | |
| }, | |
| { | |
| "epoch": 0.3592233009708738, | |
| "grad_norm": 0.9175318225986004, | |
| "kl": 0.22265625, | |
| "learning_rate": 4.9564282335552e-06, | |
| "loss": 0.2954, | |
| "step": 37, | |
| "step_loss": 0.259765625 | |
| }, | |
| { | |
| "epoch": 0.36893203883495146, | |
| "grad_norm": 0.993550543291835, | |
| "kl": 0.193359375, | |
| "learning_rate": 4.953671977103074e-06, | |
| "loss": 0.2802, | |
| "step": 38, | |
| "step_loss": 0.28125 | |
| }, | |
| { | |
| "epoch": 0.3786407766990291, | |
| "grad_norm": 1.0512485528634326, | |
| "kl": 0.7578125, | |
| "learning_rate": 4.950832101651063e-06, | |
| "loss": 0.2843, | |
| "step": 39, | |
| "step_loss": 0.58203125 | |
| }, | |
| { | |
| "epoch": 0.3883495145631068, | |
| "grad_norm": 0.6686286343760834, | |
| "kl": 0.11376953125, | |
| "learning_rate": 4.947908714959102e-06, | |
| "loss": 0.2512, | |
| "step": 40, | |
| "step_loss": 0.142578125 | |
| }, | |
| { | |
| "epoch": 0.39805825242718446, | |
| "grad_norm": 0.8469829082363365, | |
| "kl": 0.1650390625, | |
| "learning_rate": 4.944901927955983e-06, | |
| "loss": 0.3058, | |
| "step": 41, | |
| "step_loss": 0.5546875 | |
| }, | |
| { | |
| "epoch": 0.4077669902912621, | |
| "grad_norm": 0.7813325456880021, | |
| "kl": 0.2421875, | |
| "learning_rate": 4.941811854735148e-06, | |
| "loss": 0.2823, | |
| "step": 42, | |
| "step_loss": 0.193359375 | |
| }, | |
| { | |
| "epoch": 0.4174757281553398, | |
| "grad_norm": 0.9078321786346385, | |
| "kl": 0.11865234375, | |
| "learning_rate": 4.938638612550361e-06, | |
| "loss": 0.2902, | |
| "step": 43, | |
| "step_loss": 0.22265625 | |
| }, | |
| { | |
| "epoch": 0.42718446601941745, | |
| "grad_norm": 0.8942167570500859, | |
| "kl": 0.138671875, | |
| "learning_rate": 4.935382321811256e-06, | |
| "loss": 0.3012, | |
| "step": 44, | |
| "step_loss": 0.3828125 | |
| }, | |
| { | |
| "epoch": 0.4368932038834951, | |
| "grad_norm": 0.8140639878037927, | |
| "kl": 0.166015625, | |
| "learning_rate": 4.932043106078772e-06, | |
| "loss": 0.2861, | |
| "step": 45, | |
| "step_loss": 0.197265625 | |
| }, | |
| { | |
| "epoch": 0.44660194174757284, | |
| "grad_norm": 1.4604431784813239, | |
| "kl": 0.302734375, | |
| "learning_rate": 4.928621092060457e-06, | |
| "loss": 0.3029, | |
| "step": 46, | |
| "step_loss": 0.49609375 | |
| }, | |
| { | |
| "epoch": 0.4563106796116505, | |
| "grad_norm": 0.7529571567774104, | |
| "kl": 0.1318359375, | |
| "learning_rate": 4.925116409605672e-06, | |
| "loss": 0.2834, | |
| "step": 47, | |
| "step_loss": 0.275390625 | |
| }, | |
| { | |
| "epoch": 0.46601941747572817, | |
| "grad_norm": 0.7826578425297724, | |
| "kl": 0.267578125, | |
| "learning_rate": 4.92152919170065e-06, | |
| "loss": 0.3034, | |
| "step": 48, | |
| "step_loss": 0.5234375 | |
| }, | |
| { | |
| "epoch": 0.47572815533980584, | |
| "grad_norm": 1.066438164425578, | |
| "kl": 0.2109375, | |
| "learning_rate": 4.917859574463462e-06, | |
| "loss": 0.3071, | |
| "step": 49, | |
| "step_loss": 0.2236328125 | |
| }, | |
| { | |
| "epoch": 0.4854368932038835, | |
| "grad_norm": 1.7914744281397004, | |
| "kl": 0.1728515625, | |
| "learning_rate": 4.9141076971388435e-06, | |
| "loss": 0.289, | |
| "step": 50, | |
| "step_loss": 0.3046875 | |
| }, | |
| { | |
| "epoch": 0.49514563106796117, | |
| "grad_norm": 0.7820108904555855, | |
| "kl": 0.126953125, | |
| "learning_rate": 4.9102737020929135e-06, | |
| "loss": 0.2661, | |
| "step": 51, | |
| "step_loss": 0.21875 | |
| }, | |
| { | |
| "epoch": 0.5048543689320388, | |
| "grad_norm": 0.8477559536427011, | |
| "kl": 0.1875, | |
| "learning_rate": 4.906357734807776e-06, | |
| "loss": 0.3048, | |
| "step": 52, | |
| "step_loss": 0.2578125 | |
| }, | |
| { | |
| "epoch": 0.5145631067961165, | |
| "grad_norm": 0.781267699994807, | |
| "kl": 0.150390625, | |
| "learning_rate": 4.902359943875992e-06, | |
| "loss": 0.2837, | |
| "step": 53, | |
| "step_loss": 0.2734375 | |
| }, | |
| { | |
| "epoch": 0.5242718446601942, | |
| "grad_norm": 0.7373706510227717, | |
| "kl": 0.146484375, | |
| "learning_rate": 4.89828048099495e-06, | |
| "loss": 0.2632, | |
| "step": 54, | |
| "step_loss": 0.16015625 | |
| }, | |
| { | |
| "epoch": 0.5339805825242718, | |
| "grad_norm": 0.7396378743294025, | |
| "kl": 0.2421875, | |
| "learning_rate": 4.894119500961103e-06, | |
| "loss": 0.2848, | |
| "step": 55, | |
| "step_loss": 0.458984375 | |
| }, | |
| { | |
| "epoch": 0.5436893203883495, | |
| "grad_norm": 0.8384171545494072, | |
| "kl": 0.1728515625, | |
| "learning_rate": 4.889877161664096e-06, | |
| "loss": 0.311, | |
| "step": 56, | |
| "step_loss": 0.4609375 | |
| }, | |
| { | |
| "epoch": 0.5533980582524272, | |
| "grad_norm": 0.7847312422227414, | |
| "kl": 0.205078125, | |
| "learning_rate": 4.885553624080778e-06, | |
| "loss": 0.2971, | |
| "step": 57, | |
| "step_loss": 0.203125 | |
| }, | |
| { | |
| "epoch": 0.5631067961165048, | |
| "grad_norm": 0.791715637164775, | |
| "kl": 0.1494140625, | |
| "learning_rate": 4.881149052269091e-06, | |
| "loss": 0.2983, | |
| "step": 58, | |
| "step_loss": 0.197265625 | |
| }, | |
| { | |
| "epoch": 0.5728155339805825, | |
| "grad_norm": 0.6776733370951615, | |
| "kl": 0.158203125, | |
| "learning_rate": 4.876663613361844e-06, | |
| "loss": 0.2871, | |
| "step": 59, | |
| "step_loss": 0.2490234375 | |
| }, | |
| { | |
| "epoch": 0.5825242718446602, | |
| "grad_norm": 0.6974646233573695, | |
| "kl": 0.20703125, | |
| "learning_rate": 4.8720974775603745e-06, | |
| "loss": 0.2983, | |
| "step": 60, | |
| "step_loss": 0.203125 | |
| }, | |
| { | |
| "epoch": 0.5922330097087378, | |
| "grad_norm": 1.1133484937202898, | |
| "kl": 1.1171875, | |
| "learning_rate": 4.867450818128086e-06, | |
| "loss": 0.2969, | |
| "step": 61, | |
| "step_loss": 0.89453125 | |
| }, | |
| { | |
| "epoch": 0.6019417475728155, | |
| "grad_norm": 0.7147826288493739, | |
| "kl": 0.1279296875, | |
| "learning_rate": 4.862723811383878e-06, | |
| "loss": 0.2819, | |
| "step": 62, | |
| "step_loss": 0.341796875 | |
| }, | |
| { | |
| "epoch": 0.6116504854368932, | |
| "grad_norm": 0.762521447990066, | |
| "kl": 0.1591796875, | |
| "learning_rate": 4.857916636695449e-06, | |
| "loss": 0.2812, | |
| "step": 63, | |
| "step_loss": 0.1533203125 | |
| }, | |
| { | |
| "epoch": 0.6213592233009708, | |
| "grad_norm": 0.6945241251902946, | |
| "kl": 0.2294921875, | |
| "learning_rate": 4.853029476472499e-06, | |
| "loss": 0.2986, | |
| "step": 64, | |
| "step_loss": 0.275390625 | |
| }, | |
| { | |
| "epoch": 0.6310679611650486, | |
| "grad_norm": 0.7038147827740143, | |
| "kl": 0.1708984375, | |
| "learning_rate": 4.848062516159801e-06, | |
| "loss": 0.3015, | |
| "step": 65, | |
| "step_loss": 0.400390625 | |
| }, | |
| { | |
| "epoch": 0.6407766990291263, | |
| "grad_norm": 0.8016050985516795, | |
| "kl": 0.19140625, | |
| "learning_rate": 4.843015944230166e-06, | |
| "loss": 0.2873, | |
| "step": 66, | |
| "step_loss": 0.279296875 | |
| }, | |
| { | |
| "epoch": 0.6504854368932039, | |
| "grad_norm": 0.7769984336135177, | |
| "kl": 0.2451171875, | |
| "learning_rate": 4.837889952177294e-06, | |
| "loss": 0.2899, | |
| "step": 67, | |
| "step_loss": 0.56640625 | |
| }, | |
| { | |
| "epoch": 0.6601941747572816, | |
| "grad_norm": 0.6998891333931025, | |
| "kl": 0.138671875, | |
| "learning_rate": 4.832684734508502e-06, | |
| "loss": 0.2643, | |
| "step": 68, | |
| "step_loss": 0.330078125 | |
| }, | |
| { | |
| "epoch": 0.6699029126213593, | |
| "grad_norm": 1.0392663511926639, | |
| "kl": 0.1640625, | |
| "learning_rate": 4.827400488737351e-06, | |
| "loss": 0.2863, | |
| "step": 69, | |
| "step_loss": 0.287109375 | |
| }, | |
| { | |
| "epoch": 0.6796116504854369, | |
| "grad_norm": 0.8630658049176575, | |
| "kl": 0.40234375, | |
| "learning_rate": 4.822037415376147e-06, | |
| "loss": 0.2894, | |
| "step": 70, | |
| "step_loss": 0.435546875 | |
| }, | |
| { | |
| "epoch": 0.6893203883495146, | |
| "grad_norm": 0.7810137963853107, | |
| "kl": 0.271484375, | |
| "learning_rate": 4.816595717928327e-06, | |
| "loss": 0.2842, | |
| "step": 71, | |
| "step_loss": 0.2451171875 | |
| }, | |
| { | |
| "epoch": 0.6990291262135923, | |
| "grad_norm": 0.7611975818745967, | |
| "kl": 0.181640625, | |
| "learning_rate": 4.8110756028807506e-06, | |
| "loss": 0.2618, | |
| "step": 72, | |
| "step_loss": 0.3359375 | |
| }, | |
| { | |
| "epoch": 0.7087378640776699, | |
| "grad_norm": 0.9039894718510777, | |
| "kl": 0.17578125, | |
| "learning_rate": 4.805477279695852e-06, | |
| "loss": 0.304, | |
| "step": 73, | |
| "step_loss": 0.369140625 | |
| }, | |
| { | |
| "epoch": 0.7184466019417476, | |
| "grad_norm": 0.6848452637621099, | |
| "kl": 0.154296875, | |
| "learning_rate": 4.799800960803699e-06, | |
| "loss": 0.2583, | |
| "step": 74, | |
| "step_loss": 0.265625 | |
| }, | |
| { | |
| "epoch": 0.7281553398058253, | |
| "grad_norm": 0.7706941057886834, | |
| "kl": 0.11962890625, | |
| "learning_rate": 4.794046861593929e-06, | |
| "loss": 0.2913, | |
| "step": 75, | |
| "step_loss": 0.19921875 | |
| }, | |
| { | |
| "epoch": 0.7378640776699029, | |
| "grad_norm": 0.7340458477237756, | |
| "kl": 0.1953125, | |
| "learning_rate": 4.788215200407576e-06, | |
| "loss": 0.2725, | |
| "step": 76, | |
| "step_loss": 0.357421875 | |
| }, | |
| { | |
| "epoch": 0.7475728155339806, | |
| "grad_norm": 1.0806357960026565, | |
| "kl": 0.23046875, | |
| "learning_rate": 4.7823061985287906e-06, | |
| "loss": 0.2781, | |
| "step": 77, | |
| "step_loss": 0.193359375 | |
| }, | |
| { | |
| "epoch": 0.7572815533980582, | |
| "grad_norm": 0.7704916059097427, | |
| "kl": 0.34375, | |
| "learning_rate": 4.776320080176434e-06, | |
| "loss": 0.2742, | |
| "step": 78, | |
| "step_loss": 0.3203125 | |
| }, | |
| { | |
| "epoch": 0.7669902912621359, | |
| "grad_norm": 0.7492040543213823, | |
| "kl": 0.18359375, | |
| "learning_rate": 4.770257072495581e-06, | |
| "loss": 0.2776, | |
| "step": 79, | |
| "step_loss": 0.2109375 | |
| }, | |
| { | |
| "epoch": 0.7766990291262136, | |
| "grad_norm": 0.7467369782034328, | |
| "kl": 0.51171875, | |
| "learning_rate": 4.764117405548891e-06, | |
| "loss": 0.2377, | |
| "step": 80, | |
| "step_loss": 0.365234375 | |
| }, | |
| { | |
| "epoch": 0.7864077669902912, | |
| "grad_norm": 0.9406968288941178, | |
| "kl": 0.1513671875, | |
| "learning_rate": 4.757901312307882e-06, | |
| "loss": 0.2708, | |
| "step": 81, | |
| "step_loss": 0.162109375 | |
| }, | |
| { | |
| "epoch": 0.7961165048543689, | |
| "grad_norm": 0.7660157121486119, | |
| "kl": 0.2333984375, | |
| "learning_rate": 4.751609028644097e-06, | |
| "loss": 0.2877, | |
| "step": 82, | |
| "step_loss": 0.30078125 | |
| }, | |
| { | |
| "epoch": 0.8058252427184466, | |
| "grad_norm": 0.6957183682004972, | |
| "kl": 0.1806640625, | |
| "learning_rate": 4.7452407933201395e-06, | |
| "loss": 0.2403, | |
| "step": 83, | |
| "step_loss": 0.265625 | |
| }, | |
| { | |
| "epoch": 0.8155339805825242, | |
| "grad_norm": 0.7043740093288546, | |
| "kl": 0.1767578125, | |
| "learning_rate": 4.738796847980627e-06, | |
| "loss": 0.2735, | |
| "step": 84, | |
| "step_loss": 0.3359375 | |
| }, | |
| { | |
| "epoch": 0.8252427184466019, | |
| "grad_norm": 0.6782132306690141, | |
| "kl": 0.1572265625, | |
| "learning_rate": 4.732277437143015e-06, | |
| "loss": 0.2517, | |
| "step": 85, | |
| "step_loss": 0.2119140625 | |
| }, | |
| { | |
| "epoch": 0.8349514563106796, | |
| "grad_norm": 0.7479330827900228, | |
| "kl": 0.21484375, | |
| "learning_rate": 4.725682808188321e-06, | |
| "loss": 0.2641, | |
| "step": 86, | |
| "step_loss": 0.232421875 | |
| }, | |
| { | |
| "epoch": 0.8446601941747572, | |
| "grad_norm": 0.7620154952078262, | |
| "kl": 0.10791015625, | |
| "learning_rate": 4.719013211351734e-06, | |
| "loss": 0.2572, | |
| "step": 87, | |
| "step_loss": 0.146484375 | |
| }, | |
| { | |
| "epoch": 0.8543689320388349, | |
| "grad_norm": 0.6678822109983321, | |
| "kl": 0.2060546875, | |
| "learning_rate": 4.712268899713125e-06, | |
| "loss": 0.2878, | |
| "step": 88, | |
| "step_loss": 0.275390625 | |
| }, | |
| { | |
| "epoch": 0.8640776699029126, | |
| "grad_norm": 0.7456708178708048, | |
| "kl": 0.1298828125, | |
| "learning_rate": 4.705450129187439e-06, | |
| "loss": 0.2746, | |
| "step": 89, | |
| "step_loss": 0.2275390625 | |
| }, | |
| { | |
| "epoch": 0.8737864077669902, | |
| "grad_norm": 0.6177231600838415, | |
| "kl": 0.2236328125, | |
| "learning_rate": 4.698557158514988e-06, | |
| "loss": 0.2491, | |
| "step": 90, | |
| "step_loss": 0.412109375 | |
| }, | |
| { | |
| "epoch": 0.883495145631068, | |
| "grad_norm": 0.8797248927244058, | |
| "kl": 0.1513671875, | |
| "learning_rate": 4.691590249251627e-06, | |
| "loss": 0.2745, | |
| "step": 91, | |
| "step_loss": 0.12353515625 | |
| }, | |
| { | |
| "epoch": 0.8932038834951457, | |
| "grad_norm": 0.7015383302055267, | |
| "kl": 0.1162109375, | |
| "learning_rate": 4.684549665758839e-06, | |
| "loss": 0.2778, | |
| "step": 92, | |
| "step_loss": 0.189453125 | |
| }, | |
| { | |
| "epoch": 0.9029126213592233, | |
| "grad_norm": 0.7073373711435095, | |
| "kl": 0.2490234375, | |
| "learning_rate": 4.677435675193692e-06, | |
| "loss": 0.2862, | |
| "step": 93, | |
| "step_loss": 0.26171875 | |
| }, | |
| { | |
| "epoch": 0.912621359223301, | |
| "grad_norm": 0.6637944746473083, | |
| "kl": 0.14453125, | |
| "learning_rate": 4.670248547498712e-06, | |
| "loss": 0.2734, | |
| "step": 94, | |
| "step_loss": 0.1376953125 | |
| }, | |
| { | |
| "epoch": 0.9223300970873787, | |
| "grad_norm": 0.720119043217982, | |
| "kl": 0.1865234375, | |
| "learning_rate": 4.662988555391632e-06, | |
| "loss": 0.2899, | |
| "step": 95, | |
| "step_loss": 0.423828125 | |
| }, | |
| { | |
| "epoch": 0.9320388349514563, | |
| "grad_norm": 0.6938324080314517, | |
| "kl": 0.16796875, | |
| "learning_rate": 4.655655974355051e-06, | |
| "loss": 0.2714, | |
| "step": 96, | |
| "step_loss": 0.255859375 | |
| }, | |
| { | |
| "epoch": 0.941747572815534, | |
| "grad_norm": 0.6610518480589428, | |
| "kl": 0.1630859375, | |
| "learning_rate": 4.648251082625975e-06, | |
| "loss": 0.2837, | |
| "step": 97, | |
| "step_loss": 0.29296875 | |
| }, | |
| { | |
| "epoch": 0.9514563106796117, | |
| "grad_norm": 0.6877843962165877, | |
| "kl": 0.1748046875, | |
| "learning_rate": 4.640774161185259e-06, | |
| "loss": 0.3036, | |
| "step": 98, | |
| "step_loss": 0.515625 | |
| }, | |
| { | |
| "epoch": 0.9611650485436893, | |
| "grad_norm": 0.6332816751225963, | |
| "kl": 0.1533203125, | |
| "learning_rate": 4.633225493746951e-06, | |
| "loss": 0.2489, | |
| "step": 99, | |
| "step_loss": 0.1826171875 | |
| }, | |
| { | |
| "epoch": 0.970873786407767, | |
| "grad_norm": 0.7362793924005384, | |
| "kl": 0.123046875, | |
| "learning_rate": 4.625605366747519e-06, | |
| "loss": 0.2622, | |
| "step": 100, | |
| "step_loss": 0.12890625 | |
| }, | |
| { | |
| "epoch": 0.9805825242718447, | |
| "grad_norm": 0.6742913243561754, | |
| "kl": 0.2119140625, | |
| "learning_rate": 4.617914069334989e-06, | |
| "loss": 0.2544, | |
| "step": 101, | |
| "step_loss": 0.2578125 | |
| }, | |
| { | |
| "epoch": 0.9902912621359223, | |
| "grad_norm": 0.6047332480224725, | |
| "kl": 0.259765625, | |
| "learning_rate": 4.610151893357968e-06, | |
| "loss": 0.2546, | |
| "step": 102, | |
| "step_loss": 0.37890625 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.6546589091756503, | |
| "kl": 0.2138671875, | |
| "learning_rate": 4.602319133354571e-06, | |
| "loss": 0.2566, | |
| "step": 103, | |
| "step_loss": 0.2177734375 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_test_transformed.json_loss": NaN, | |
| "eval_test_transformed.json_runtime": 8.816, | |
| "eval_test_transformed.json_samples_per_second": 56.715, | |
| "eval_test_transformed.json_steps_per_second": 1.815, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 1.0097087378640777, | |
| "grad_norm": 0.639387377398209, | |
| "kl": NaN, | |
| "learning_rate": 4.594416086541248e-06, | |
| "loss": 0.2493, | |
| "step": 104, | |
| "step_loss": NaN | |
| }, | |
| { | |
| "epoch": 1.0194174757281553, | |
| "grad_norm": 0.7392491046589192, | |
| "kl": 0.1640625, | |
| "learning_rate": 4.5864430528014996e-06, | |
| "loss": 0.2413, | |
| "step": 105, | |
| "step_loss": 0.1396484375 | |
| }, | |
| { | |
| "epoch": 1.029126213592233, | |
| "grad_norm": 0.5649339416591039, | |
| "kl": 0.2001953125, | |
| "learning_rate": 4.578400334674503e-06, | |
| "loss": 0.2223, | |
| "step": 106, | |
| "step_loss": 0.318359375 | |
| }, | |
| { | |
| "epoch": 1.0388349514563107, | |
| "grad_norm": 0.6022576613202164, | |
| "kl": 0.1708984375, | |
| "learning_rate": 4.570288237343632e-06, | |
| "loss": 0.2368, | |
| "step": 107, | |
| "step_loss": 0.24609375 | |
| }, | |
| { | |
| "epoch": 1.0485436893203883, | |
| "grad_norm": 0.6973100963269709, | |
| "kl": 0.474609375, | |
| "learning_rate": 4.562107068624874e-06, | |
| "loss": 0.2363, | |
| "step": 108, | |
| "step_loss": 0.255859375 | |
| }, | |
| { | |
| "epoch": 1.058252427184466, | |
| "grad_norm": 0.6281000265214681, | |
| "kl": 0.16796875, | |
| "learning_rate": 4.55385713895515e-06, | |
| "loss": 0.2271, | |
| "step": 109, | |
| "step_loss": 0.2578125 | |
| }, | |
| { | |
| "epoch": 1.0679611650485437, | |
| "grad_norm": 0.6251998581820843, | |
| "kl": 0.19921875, | |
| "learning_rate": 4.545538761380539e-06, | |
| "loss": 0.2444, | |
| "step": 110, | |
| "step_loss": 0.1533203125 | |
| }, | |
| { | |
| "epoch": 1.0776699029126213, | |
| "grad_norm": 0.6499920789529428, | |
| "kl": 0.140625, | |
| "learning_rate": 4.537152251544394e-06, | |
| "loss": 0.2575, | |
| "step": 111, | |
| "step_loss": 0.21484375 | |
| }, | |
| { | |
| "epoch": 1.087378640776699, | |
| "grad_norm": 0.6196063483230071, | |
| "kl": 0.34765625, | |
| "learning_rate": 4.5286979276753675e-06, | |
| "loss": 0.2216, | |
| "step": 112, | |
| "step_loss": 0.24609375 | |
| }, | |
| { | |
| "epoch": 1.0970873786407767, | |
| "grad_norm": 0.6519876067572186, | |
| "kl": 0.2265625, | |
| "learning_rate": 4.520176110575338e-06, | |
| "loss": 0.2209, | |
| "step": 113, | |
| "step_loss": 0.1748046875 | |
| }, | |
| { | |
| "epoch": 1.1067961165048543, | |
| "grad_norm": 0.6464673181225589, | |
| "kl": 0.2041015625, | |
| "learning_rate": 4.511587123607232e-06, | |
| "loss": 0.2242, | |
| "step": 114, | |
| "step_loss": 0.271484375 | |
| }, | |
| { | |
| "epoch": 1.116504854368932, | |
| "grad_norm": 0.5620248695246799, | |
| "kl": 0.23046875, | |
| "learning_rate": 4.502931292682759e-06, | |
| "loss": 0.225, | |
| "step": 115, | |
| "step_loss": 0.255859375 | |
| }, | |
| { | |
| "epoch": 1.1262135922330097, | |
| "grad_norm": 0.6212910752562149, | |
| "kl": 0.3203125, | |
| "learning_rate": 4.494208946250042e-06, | |
| "loss": 0.2322, | |
| "step": 116, | |
| "step_loss": 0.166015625 | |
| }, | |
| { | |
| "epoch": 1.1359223300970873, | |
| "grad_norm": 0.6102956517360447, | |
| "kl": 0.134765625, | |
| "learning_rate": 4.485420415281157e-06, | |
| "loss": 0.246, | |
| "step": 117, | |
| "step_loss": 0.2099609375 | |
| }, | |
| { | |
| "epoch": 1.145631067961165, | |
| "grad_norm": 0.8959975572620882, | |
| "kl": 0.341796875, | |
| "learning_rate": 4.4765660332595686e-06, | |
| "loss": 0.2519, | |
| "step": 118, | |
| "step_loss": 0.154296875 | |
| }, | |
| { | |
| "epoch": 1.1553398058252426, | |
| "grad_norm": 0.5883698395469736, | |
| "kl": 0.30078125, | |
| "learning_rate": 4.467646136167482e-06, | |
| "loss": 0.2548, | |
| "step": 119, | |
| "step_loss": 0.22265625 | |
| }, | |
| { | |
| "epoch": 1.1650485436893203, | |
| "grad_norm": 0.5899290313710992, | |
| "kl": 0.1669921875, | |
| "learning_rate": 4.458661062473091e-06, | |
| "loss": 0.2327, | |
| "step": 120, | |
| "step_loss": 0.1533203125 | |
| }, | |
| { | |
| "epoch": 1.174757281553398, | |
| "grad_norm": 0.7434899396388887, | |
| "kl": 0.294921875, | |
| "learning_rate": 4.449611153117736e-06, | |
| "loss": 0.2415, | |
| "step": 121, | |
| "step_loss": 0.16796875 | |
| }, | |
| { | |
| "epoch": 1.1844660194174756, | |
| "grad_norm": 0.5645672375917218, | |
| "kl": 0.21875, | |
| "learning_rate": 4.4404967515029655e-06, | |
| "loss": 0.198, | |
| "step": 122, | |
| "step_loss": 0.27734375 | |
| }, | |
| { | |
| "epoch": 1.1941747572815533, | |
| "grad_norm": 0.61620433580154, | |
| "kl": 0.28125, | |
| "learning_rate": 4.431318203477505e-06, | |
| "loss": 0.214, | |
| "step": 123, | |
| "step_loss": 0.18359375 | |
| }, | |
| { | |
| "epoch": 1.203883495145631, | |
| "grad_norm": 0.5927124243417978, | |
| "kl": 0.29296875, | |
| "learning_rate": 4.422075857324138e-06, | |
| "loss": 0.2182, | |
| "step": 124, | |
| "step_loss": 0.25 | |
| }, | |
| { | |
| "epoch": 1.2135922330097086, | |
| "grad_norm": 0.6201091855952103, | |
| "kl": 0.18359375, | |
| "learning_rate": 4.412770063746483e-06, | |
| "loss": 0.2207, | |
| "step": 125, | |
| "step_loss": 0.1865234375 | |
| }, | |
| { | |
| "epoch": 1.2233009708737863, | |
| "grad_norm": 0.911604159360823, | |
| "kl": 0.162109375, | |
| "learning_rate": 4.403401175855695e-06, | |
| "loss": 0.2471, | |
| "step": 126, | |
| "step_loss": 0.1474609375 | |
| }, | |
| { | |
| "epoch": 1.233009708737864, | |
| "grad_norm": 0.6792470228209272, | |
| "kl": 0.0927734375, | |
| "learning_rate": 4.3939695491570596e-06, | |
| "loss": 0.2351, | |
| "step": 127, | |
| "step_loss": 0.146484375 | |
| }, | |
| { | |
| "epoch": 1.2427184466019416, | |
| "grad_norm": 0.6012469615978469, | |
| "kl": 0.212890625, | |
| "learning_rate": 4.384475541536505e-06, | |
| "loss": 0.2434, | |
| "step": 128, | |
| "step_loss": 0.2041015625 | |
| }, | |
| { | |
| "epoch": 1.2524271844660193, | |
| "grad_norm": 0.589209563012478, | |
| "kl": 0.2041015625, | |
| "learning_rate": 4.374919513247021e-06, | |
| "loss": 0.2129, | |
| "step": 129, | |
| "step_loss": 0.109375 | |
| }, | |
| { | |
| "epoch": 1.262135922330097, | |
| "grad_norm": 0.5987454630415056, | |
| "kl": 0.22265625, | |
| "learning_rate": 4.3653018268949945e-06, | |
| "loss": 0.2179, | |
| "step": 130, | |
| "step_loss": 0.166015625 | |
| }, | |
| { | |
| "epoch": 1.2718446601941746, | |
| "grad_norm": 0.5813161566217064, | |
| "kl": 0.2041015625, | |
| "learning_rate": 4.355622847426444e-06, | |
| "loss": 0.2183, | |
| "step": 131, | |
| "step_loss": 0.431640625 | |
| }, | |
| { | |
| "epoch": 1.2815533980582523, | |
| "grad_norm": 0.644992270460315, | |
| "kl": 0.25, | |
| "learning_rate": 4.345882942113171e-06, | |
| "loss": 0.2412, | |
| "step": 132, | |
| "step_loss": 0.1474609375 | |
| }, | |
| { | |
| "epoch": 1.29126213592233, | |
| "grad_norm": 0.5838255650379847, | |
| "kl": 0.4140625, | |
| "learning_rate": 4.336082480538832e-06, | |
| "loss": 0.2223, | |
| "step": 133, | |
| "step_loss": 0.171875 | |
| }, | |
| { | |
| "epoch": 1.3009708737864076, | |
| "grad_norm": 0.7975539886352869, | |
| "kl": 0.3125, | |
| "learning_rate": 4.326221834584905e-06, | |
| "loss": 0.2218, | |
| "step": 134, | |
| "step_loss": 0.47265625 | |
| }, | |
| { | |
| "epoch": 1.3106796116504853, | |
| "grad_norm": 0.5800964403953691, | |
| "kl": 0.181640625, | |
| "learning_rate": 4.316301378416585e-06, | |
| "loss": 0.239, | |
| "step": 135, | |
| "step_loss": 0.1552734375 | |
| }, | |
| { | |
| "epoch": 1.3203883495145632, | |
| "grad_norm": 0.5247261922814125, | |
| "kl": 0.1923828125, | |
| "learning_rate": 4.306321488468583e-06, | |
| "loss": 0.2337, | |
| "step": 136, | |
| "step_loss": 0.119140625 | |
| }, | |
| { | |
| "epoch": 1.3300970873786409, | |
| "grad_norm": 0.6297520228191021, | |
| "kl": 0.1435546875, | |
| "learning_rate": 4.296282543430841e-06, | |
| "loss": 0.24, | |
| "step": 137, | |
| "step_loss": 0.29296875 | |
| }, | |
| { | |
| "epoch": 1.3398058252427185, | |
| "grad_norm": 0.5734205082343037, | |
| "kl": 0.216796875, | |
| "learning_rate": 4.286184924234168e-06, | |
| "loss": 0.2218, | |
| "step": 138, | |
| "step_loss": 0.15234375 | |
| }, | |
| { | |
| "epoch": 1.3495145631067962, | |
| "grad_norm": 0.5442629148889834, | |
| "kl": 0.2255859375, | |
| "learning_rate": 4.276029014035777e-06, | |
| "loss": 0.2189, | |
| "step": 139, | |
| "step_loss": 0.2265625 | |
| }, | |
| { | |
| "epoch": 1.3592233009708738, | |
| "grad_norm": 0.5288732036595232, | |
| "kl": 0.12060546875, | |
| "learning_rate": 4.265815198204754e-06, | |
| "loss": 0.2428, | |
| "step": 140, | |
| "step_loss": 0.169921875 | |
| }, | |
| { | |
| "epoch": 1.3689320388349515, | |
| "grad_norm": 0.6546290678837918, | |
| "kl": 0.2451171875, | |
| "learning_rate": 4.2555438643074315e-06, | |
| "loss": 0.2422, | |
| "step": 141, | |
| "step_loss": 0.21875 | |
| }, | |
| { | |
| "epoch": 1.3786407766990292, | |
| "grad_norm": 0.585078334332048, | |
| "kl": 0.2041015625, | |
| "learning_rate": 4.245215402092681e-06, | |
| "loss": 0.2455, | |
| "step": 142, | |
| "step_loss": 0.205078125 | |
| }, | |
| { | |
| "epoch": 1.3883495145631068, | |
| "grad_norm": 0.556943790797957, | |
| "kl": 0.1748046875, | |
| "learning_rate": 4.234830203477126e-06, | |
| "loss": 0.2064, | |
| "step": 143, | |
| "step_loss": 0.201171875 | |
| }, | |
| { | |
| "epoch": 1.3980582524271845, | |
| "grad_norm": 0.5946911580488923, | |
| "kl": 0.14453125, | |
| "learning_rate": 4.224388662530271e-06, | |
| "loss": 0.2218, | |
| "step": 144, | |
| "step_loss": 0.2431640625 | |
| }, | |
| { | |
| "epoch": 1.4077669902912622, | |
| "grad_norm": 0.5810717668347283, | |
| "kl": 0.326171875, | |
| "learning_rate": 4.213891175459545e-06, | |
| "loss": 0.2442, | |
| "step": 145, | |
| "step_loss": 0.1796875 | |
| }, | |
| { | |
| "epoch": 1.4174757281553398, | |
| "grad_norm": 0.6115382263091327, | |
| "kl": 0.1865234375, | |
| "learning_rate": 4.203338140595272e-06, | |
| "loss": 0.2202, | |
| "step": 146, | |
| "step_loss": 0.26171875 | |
| }, | |
| { | |
| "epoch": 1.4271844660194175, | |
| "grad_norm": 0.5465729015270102, | |
| "kl": 0.248046875, | |
| "learning_rate": 4.192729958375552e-06, | |
| "loss": 0.2129, | |
| "step": 147, | |
| "step_loss": 0.28125 | |
| }, | |
| { | |
| "epoch": 1.4368932038834952, | |
| "grad_norm": 0.5834391929303593, | |
| "kl": 0.162109375, | |
| "learning_rate": 4.1820670313310684e-06, | |
| "loss": 0.2328, | |
| "step": 148, | |
| "step_loss": 0.1103515625 | |
| }, | |
| { | |
| "epoch": 1.4466019417475728, | |
| "grad_norm": 0.6239303256729286, | |
| "kl": 0.19921875, | |
| "learning_rate": 4.171349764069815e-06, | |
| "loss": 0.2468, | |
| "step": 149, | |
| "step_loss": 0.2275390625 | |
| }, | |
| { | |
| "epoch": 1.4563106796116505, | |
| "grad_norm": 0.5464914339167269, | |
| "kl": 0.154296875, | |
| "learning_rate": 4.16057856326174e-06, | |
| "loss": 0.2145, | |
| "step": 150, | |
| "step_loss": 0.21484375 | |
| }, | |
| { | |
| "epoch": 1.4660194174757282, | |
| "grad_norm": 0.5714192820756144, | |
| "kl": 0.34765625, | |
| "learning_rate": 4.149753837623317e-06, | |
| "loss": 0.2293, | |
| "step": 151, | |
| "step_loss": 0.1142578125 | |
| }, | |
| { | |
| "epoch": 1.4757281553398058, | |
| "grad_norm": 0.6222270965621777, | |
| "kl": 0.20703125, | |
| "learning_rate": 4.1388759979020346e-06, | |
| "loss": 0.2378, | |
| "step": 152, | |
| "step_loss": 0.2158203125 | |
| }, | |
| { | |
| "epoch": 1.4854368932038835, | |
| "grad_norm": 0.6459648372732245, | |
| "kl": 0.1728515625, | |
| "learning_rate": 4.127945456860813e-06, | |
| "loss": 0.2301, | |
| "step": 153, | |
| "step_loss": 0.154296875 | |
| }, | |
| { | |
| "epoch": 1.4951456310679612, | |
| "grad_norm": 0.613782544176392, | |
| "kl": 0.21875, | |
| "learning_rate": 4.116962629262339e-06, | |
| "loss": 0.2155, | |
| "step": 154, | |
| "step_loss": 0.2099609375 | |
| }, | |
| { | |
| "epoch": 1.5048543689320388, | |
| "grad_norm": 0.569655314094469, | |
| "kl": 0.220703125, | |
| "learning_rate": 4.105927931853327e-06, | |
| "loss": 0.231, | |
| "step": 155, | |
| "step_loss": 0.26953125 | |
| }, | |
| { | |
| "epoch": 1.5145631067961165, | |
| "grad_norm": 0.6309412616695066, | |
| "kl": 0.369140625, | |
| "learning_rate": 4.094841783348711e-06, | |
| "loss": 0.2435, | |
| "step": 156, | |
| "step_loss": 0.1689453125 | |
| }, | |
| { | |
| "epoch": 1.5242718446601942, | |
| "grad_norm": 0.6105234133068124, | |
| "kl": 0.2138671875, | |
| "learning_rate": 4.083704604415749e-06, | |
| "loss": 0.2052, | |
| "step": 157, | |
| "step_loss": 0.2412109375 | |
| }, | |
| { | |
| "epoch": 1.5339805825242718, | |
| "grad_norm": 0.5804176749635781, | |
| "kl": 0.2197265625, | |
| "learning_rate": 4.072516817658065e-06, | |
| "loss": 0.2508, | |
| "step": 158, | |
| "step_loss": 0.171875 | |
| }, | |
| { | |
| "epoch": 1.5436893203883495, | |
| "grad_norm": 0.55765374326049, | |
| "kl": 0.306640625, | |
| "learning_rate": 4.0612788475996125e-06, | |
| "loss": 0.2164, | |
| "step": 159, | |
| "step_loss": 0.2001953125 | |
| }, | |
| { | |
| "epoch": 1.5533980582524272, | |
| "grad_norm": 0.6074565129354751, | |
| "kl": 0.7265625, | |
| "learning_rate": 4.049991120668566e-06, | |
| "loss": 0.229, | |
| "step": 160, | |
| "step_loss": 0.40625 | |
| }, | |
| { | |
| "epoch": 1.5631067961165048, | |
| "grad_norm": 0.7748587294998153, | |
| "kl": 1.328125, | |
| "learning_rate": 4.038654065181137e-06, | |
| "loss": 0.281, | |
| "step": 161, | |
| "step_loss": 0.8984375 | |
| }, | |
| { | |
| "epoch": 1.5728155339805825, | |
| "grad_norm": 0.5644659021112801, | |
| "kl": 0.205078125, | |
| "learning_rate": 4.027268111325328e-06, | |
| "loss": 0.2397, | |
| "step": 162, | |
| "step_loss": 0.26953125 | |
| }, | |
| { | |
| "epoch": 1.5825242718446602, | |
| "grad_norm": 0.5648454404940179, | |
| "kl": 0.1533203125, | |
| "learning_rate": 4.015833691144603e-06, | |
| "loss": 0.2157, | |
| "step": 163, | |
| "step_loss": 0.1416015625 | |
| }, | |
| { | |
| "epoch": 1.5922330097087378, | |
| "grad_norm": 0.5903127460469126, | |
| "kl": 0.224609375, | |
| "learning_rate": 4.0043512385214924e-06, | |
| "loss": 0.2308, | |
| "step": 164, | |
| "step_loss": 0.1669921875 | |
| }, | |
| { | |
| "epoch": 1.6019417475728155, | |
| "grad_norm": 0.5951640957860757, | |
| "kl": 0.26953125, | |
| "learning_rate": 3.9928211891611385e-06, | |
| "loss": 0.2136, | |
| "step": 165, | |
| "step_loss": 0.1875 | |
| }, | |
| { | |
| "epoch": 1.6116504854368932, | |
| "grad_norm": 0.5261435237006341, | |
| "kl": 0.17578125, | |
| "learning_rate": 3.981243980574751e-06, | |
| "loss": 0.2382, | |
| "step": 166, | |
| "step_loss": 0.234375 | |
| }, | |
| { | |
| "epoch": 1.6213592233009708, | |
| "grad_norm": 0.5887020505829587, | |
| "kl": 0.2255859375, | |
| "learning_rate": 3.969620052063012e-06, | |
| "loss": 0.2352, | |
| "step": 167, | |
| "step_loss": 0.2236328125 | |
| }, | |
| { | |
| "epoch": 1.6310679611650487, | |
| "grad_norm": 0.6254868280303059, | |
| "kl": 0.2001953125, | |
| "learning_rate": 3.957949844699405e-06, | |
| "loss": 0.2235, | |
| "step": 168, | |
| "step_loss": 0.162109375 | |
| }, | |
| { | |
| "epoch": 1.6407766990291264, | |
| "grad_norm": 0.5478297482531744, | |
| "kl": 0.23046875, | |
| "learning_rate": 3.946233801313482e-06, | |
| "loss": 0.2201, | |
| "step": 169, | |
| "step_loss": 0.30859375 | |
| }, | |
| { | |
| "epoch": 1.650485436893204, | |
| "grad_norm": 0.5691412826879887, | |
| "kl": 0.2177734375, | |
| "learning_rate": 3.934472366474051e-06, | |
| "loss": 0.2014, | |
| "step": 170, | |
| "step_loss": 0.2001953125 | |
| }, | |
| { | |
| "epoch": 1.6601941747572817, | |
| "grad_norm": 0.5604014457233358, | |
| "kl": 0.16796875, | |
| "learning_rate": 3.922665986472316e-06, | |
| "loss": 0.2348, | |
| "step": 171, | |
| "step_loss": 0.27734375 | |
| }, | |
| { | |
| "epoch": 1.6699029126213594, | |
| "grad_norm": 0.5166321657054415, | |
| "kl": 0.1708984375, | |
| "learning_rate": 3.91081510930494e-06, | |
| "loss": 0.2125, | |
| "step": 172, | |
| "step_loss": 0.263671875 | |
| }, | |
| { | |
| "epoch": 1.679611650485437, | |
| "grad_norm": 0.567633227241871, | |
| "kl": 0.2197265625, | |
| "learning_rate": 3.8989201846570405e-06, | |
| "loss": 0.2257, | |
| "step": 173, | |
| "step_loss": 0.16796875 | |
| }, | |
| { | |
| "epoch": 1.6893203883495147, | |
| "grad_norm": 0.6483568118346154, | |
| "kl": 0.185546875, | |
| "learning_rate": 3.886981663885134e-06, | |
| "loss": 0.2426, | |
| "step": 174, | |
| "step_loss": 0.25390625 | |
| }, | |
| { | |
| "epoch": 1.6990291262135924, | |
| "grad_norm": 0.6784974565198278, | |
| "kl": 0.28125, | |
| "learning_rate": 3.875e-06, | |
| "loss": 0.2508, | |
| "step": 175, | |
| "step_loss": 0.453125 | |
| }, | |
| { | |
| "epoch": 1.70873786407767, | |
| "grad_norm": 0.5532390595417498, | |
| "kl": 0.20703125, | |
| "learning_rate": 3.862975647649503e-06, | |
| "loss": 0.2183, | |
| "step": 176, | |
| "step_loss": 0.12890625 | |
| }, | |
| { | |
| "epoch": 1.7184466019417477, | |
| "grad_norm": 0.5636800996705107, | |
| "kl": 0.228515625, | |
| "learning_rate": 3.850909063101328e-06, | |
| "loss": 0.2107, | |
| "step": 177, | |
| "step_loss": 0.1845703125 | |
| }, | |
| { | |
| "epoch": 1.7281553398058254, | |
| "grad_norm": 0.5703475366568808, | |
| "kl": 0.1337890625, | |
| "learning_rate": 3.838800704225679e-06, | |
| "loss": 0.2135, | |
| "step": 178, | |
| "step_loss": 0.1708984375 | |
| }, | |
| { | |
| "epoch": 1.737864077669903, | |
| "grad_norm": 0.6101234925448211, | |
| "kl": 0.1904296875, | |
| "learning_rate": 3.826651030477896e-06, | |
| "loss": 0.2385, | |
| "step": 179, | |
| "step_loss": 0.177734375 | |
| }, | |
| { | |
| "epoch": 1.7475728155339807, | |
| "grad_norm": 0.5393184109703284, | |
| "kl": 0.1708984375, | |
| "learning_rate": 3.8144605028810233e-06, | |
| "loss": 0.2215, | |
| "step": 180, | |
| "step_loss": 0.18359375 | |
| }, | |
| { | |
| "epoch": 1.7572815533980584, | |
| "grad_norm": 0.560735690651583, | |
| "kl": 0.322265625, | |
| "learning_rate": 3.802229584008321e-06, | |
| "loss": 0.2235, | |
| "step": 181, | |
| "step_loss": 0.2353515625 | |
| }, | |
| { | |
| "epoch": 1.766990291262136, | |
| "grad_norm": 0.5687197299908796, | |
| "kl": 0.318359375, | |
| "learning_rate": 3.789958737965705e-06, | |
| "loss": 0.2377, | |
| "step": 182, | |
| "step_loss": 0.42578125 | |
| }, | |
| { | |
| "epoch": 1.7766990291262137, | |
| "grad_norm": 0.517105808317797, | |
| "kl": 0.2265625, | |
| "learning_rate": 3.777648430374142e-06, | |
| "loss": 0.2247, | |
| "step": 183, | |
| "step_loss": 0.423828125 | |
| }, | |
| { | |
| "epoch": 1.7864077669902914, | |
| "grad_norm": 0.5846543567424347, | |
| "kl": 0.150390625, | |
| "learning_rate": 3.765299128351977e-06, | |
| "loss": 0.2233, | |
| "step": 184, | |
| "step_loss": 0.14453125 | |
| }, | |
| { | |
| "epoch": 1.796116504854369, | |
| "grad_norm": 0.550349343500017, | |
| "kl": 0.5078125, | |
| "learning_rate": 3.7529113004972117e-06, | |
| "loss": 0.2296, | |
| "step": 185, | |
| "step_loss": 0.32421875 | |
| }, | |
| { | |
| "epoch": 1.8058252427184467, | |
| "grad_norm": 0.6054819413255546, | |
| "kl": 0.2890625, | |
| "learning_rate": 3.740485416869722e-06, | |
| "loss": 0.2266, | |
| "step": 186, | |
| "step_loss": 0.2373046875 | |
| }, | |
| { | |
| "epoch": 1.8155339805825244, | |
| "grad_norm": 0.5882865611694021, | |
| "kl": 0.2001953125, | |
| "learning_rate": 3.7280219489734214e-06, | |
| "loss": 0.2265, | |
| "step": 187, | |
| "step_loss": 0.1259765625 | |
| }, | |
| { | |
| "epoch": 1.825242718446602, | |
| "grad_norm": 0.6392508865666573, | |
| "kl": 0.328125, | |
| "learning_rate": 3.7155213697383702e-06, | |
| "loss": 0.2463, | |
| "step": 188, | |
| "step_loss": 0.173828125 | |
| }, | |
| { | |
| "epoch": 1.8349514563106797, | |
| "grad_norm": 0.5196239704198046, | |
| "kl": 0.1923828125, | |
| "learning_rate": 3.70298415350283e-06, | |
| "loss": 0.2359, | |
| "step": 189, | |
| "step_loss": 0.123046875 | |
| }, | |
| { | |
| "epoch": 1.8446601941747574, | |
| "grad_norm": 0.6494083949250458, | |
| "kl": 1.0859375, | |
| "learning_rate": 3.690410775995263e-06, | |
| "loss": 0.2307, | |
| "step": 190, | |
| "step_loss": 0.44921875 | |
| }, | |
| { | |
| "epoch": 1.854368932038835, | |
| "grad_norm": 0.5435542784787667, | |
| "kl": 0.3359375, | |
| "learning_rate": 3.677801714316283e-06, | |
| "loss": 0.2252, | |
| "step": 191, | |
| "step_loss": 0.2265625 | |
| }, | |
| { | |
| "epoch": 1.8640776699029127, | |
| "grad_norm": 0.6511666190719619, | |
| "kl": 0.248046875, | |
| "learning_rate": 3.665157446920551e-06, | |
| "loss": 0.2525, | |
| "step": 192, | |
| "step_loss": 0.189453125 | |
| }, | |
| { | |
| "epoch": 1.8737864077669903, | |
| "grad_norm": 0.6128430106715034, | |
| "kl": 0.306640625, | |
| "learning_rate": 3.6524784535986175e-06, | |
| "loss": 0.2723, | |
| "step": 193, | |
| "step_loss": 0.298828125 | |
| }, | |
| { | |
| "epoch": 1.883495145631068, | |
| "grad_norm": 0.5849812397434826, | |
| "kl": 0.1875, | |
| "learning_rate": 3.639765215458721e-06, | |
| "loss": 0.2487, | |
| "step": 194, | |
| "step_loss": 0.33984375 | |
| }, | |
| { | |
| "epoch": 1.8932038834951457, | |
| "grad_norm": 0.5998143147199343, | |
| "kl": 0.2060546875, | |
| "learning_rate": 3.6270182149085288e-06, | |
| "loss": 0.2265, | |
| "step": 195, | |
| "step_loss": 0.1162109375 | |
| }, | |
| { | |
| "epoch": 1.9029126213592233, | |
| "grad_norm": 0.5559548956050891, | |
| "kl": 0.2041015625, | |
| "learning_rate": 3.6142379356368334e-06, | |
| "loss": 0.2266, | |
| "step": 196, | |
| "step_loss": 0.12109375 | |
| }, | |
| { | |
| "epoch": 1.912621359223301, | |
| "grad_norm": 0.6722163508708774, | |
| "kl": 0.2265625, | |
| "learning_rate": 3.6014248625951987e-06, | |
| "loss": 0.2349, | |
| "step": 197, | |
| "step_loss": 0.296875 | |
| }, | |
| { | |
| "epoch": 1.9223300970873787, | |
| "grad_norm": 0.5488054807070938, | |
| "kl": 0.0947265625, | |
| "learning_rate": 3.5885794819795566e-06, | |
| "loss": 0.216, | |
| "step": 198, | |
| "step_loss": 0.09326171875 | |
| }, | |
| { | |
| "epoch": 1.9320388349514563, | |
| "grad_norm": 0.5550608922872593, | |
| "kl": 0.423828125, | |
| "learning_rate": 3.5757022812117625e-06, | |
| "loss": 0.2304, | |
| "step": 199, | |
| "step_loss": 0.49609375 | |
| }, | |
| { | |
| "epoch": 1.941747572815534, | |
| "grad_norm": 0.5964742703384457, | |
| "kl": 0.341796875, | |
| "learning_rate": 3.562793748921095e-06, | |
| "loss": 0.2283, | |
| "step": 200, | |
| "step_loss": 0.32421875 | |
| }, | |
| { | |
| "epoch": 1.9514563106796117, | |
| "grad_norm": 0.5752160371372506, | |
| "kl": 0.263671875, | |
| "learning_rate": 3.5498543749257164e-06, | |
| "loss": 0.2166, | |
| "step": 201, | |
| "step_loss": 0.2041015625 | |
| }, | |
| { | |
| "epoch": 1.9611650485436893, | |
| "grad_norm": 0.547718813991668, | |
| "kl": 0.1904296875, | |
| "learning_rate": 3.536884650214088e-06, | |
| "loss": 0.2374, | |
| "step": 202, | |
| "step_loss": 0.193359375 | |
| }, | |
| { | |
| "epoch": 1.970873786407767, | |
| "grad_norm": 0.5834419512668307, | |
| "kl": 0.20703125, | |
| "learning_rate": 3.5238850669263386e-06, | |
| "loss": 0.2419, | |
| "step": 203, | |
| "step_loss": 0.404296875 | |
| }, | |
| { | |
| "epoch": 1.9805825242718447, | |
| "grad_norm": 0.5492481511414377, | |
| "kl": 0.171875, | |
| "learning_rate": 3.510856118335589e-06, | |
| "loss": 0.2304, | |
| "step": 204, | |
| "step_loss": 0.26953125 | |
| }, | |
| { | |
| "epoch": 1.9902912621359223, | |
| "grad_norm": 0.6611396642693328, | |
| "kl": 0.34765625, | |
| "learning_rate": 3.497798298829234e-06, | |
| "loss": 0.2235, | |
| "step": 205, | |
| "step_loss": 0.158203125 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 0.6014585529041143, | |
| "kl": 0.19140625, | |
| "learning_rate": 3.484712103890188e-06, | |
| "loss": 0.2067, | |
| "step": 206, | |
| "step_loss": 0.23828125 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_test_transformed.json_loss": NaN, | |
| "eval_test_transformed.json_runtime": 8.5733, | |
| "eval_test_transformed.json_samples_per_second": 58.321, | |
| "eval_test_transformed.json_steps_per_second": 1.866, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 2.0097087378640777, | |
| "grad_norm": 0.5791202947407729, | |
| "kl": NaN, | |
| "learning_rate": 3.471598030078074e-06, | |
| "loss": 0.1895, | |
| "step": 207, | |
| "step_loss": NaN | |
| }, | |
| { | |
| "epoch": 2.0194174757281553, | |
| "grad_norm": 0.5095592680348658, | |
| "kl": 0.2373046875, | |
| "learning_rate": 3.4584565750103932e-06, | |
| "loss": 0.1777, | |
| "step": 208, | |
| "step_loss": 0.25390625 | |
| }, | |
| { | |
| "epoch": 2.029126213592233, | |
| "grad_norm": 0.52090270216247, | |
| "kl": 0.2041015625, | |
| "learning_rate": 3.445288237343632e-06, | |
| "loss": 0.1885, | |
| "step": 209, | |
| "step_loss": 0.240234375 | |
| }, | |
| { | |
| "epoch": 2.0388349514563107, | |
| "grad_norm": 0.4976045327134996, | |
| "kl": 0.2431640625, | |
| "learning_rate": 3.432093516754348e-06, | |
| "loss": 0.1863, | |
| "step": 210, | |
| "step_loss": 0.1064453125 | |
| }, | |
| { | |
| "epoch": 2.0485436893203883, | |
| "grad_norm": 0.47846076806594573, | |
| "kl": 0.25390625, | |
| "learning_rate": 3.4188729139202063e-06, | |
| "loss": 0.1806, | |
| "step": 211, | |
| "step_loss": 0.138671875 | |
| }, | |
| { | |
| "epoch": 2.058252427184466, | |
| "grad_norm": 0.4911716210346588, | |
| "kl": 0.1787109375, | |
| "learning_rate": 3.4056269305009807e-06, | |
| "loss": 0.1963, | |
| "step": 212, | |
| "step_loss": 0.1650390625 | |
| }, | |
| { | |
| "epoch": 2.0679611650485437, | |
| "grad_norm": 0.5097950309369077, | |
| "kl": 0.1708984375, | |
| "learning_rate": 3.3923560691195194e-06, | |
| "loss": 0.1745, | |
| "step": 213, | |
| "step_loss": 0.232421875 | |
| }, | |
| { | |
| "epoch": 2.0776699029126213, | |
| "grad_norm": 0.49796056560568963, | |
| "kl": 0.390625, | |
| "learning_rate": 3.379060833342673e-06, | |
| "loss": 0.1956, | |
| "step": 214, | |
| "step_loss": 0.134765625 | |
| }, | |
| { | |
| "epoch": 2.087378640776699, | |
| "grad_norm": 0.5533404313296716, | |
| "kl": 0.1796875, | |
| "learning_rate": 3.3657417276621867e-06, | |
| "loss": 0.1977, | |
| "step": 215, | |
| "step_loss": 0.2890625 | |
| }, | |
| { | |
| "epoch": 2.0970873786407767, | |
| "grad_norm": 0.5296091938932913, | |
| "kl": 0.251953125, | |
| "learning_rate": 3.352399257475553e-06, | |
| "loss": 0.1799, | |
| "step": 216, | |
| "step_loss": 0.296875 | |
| }, | |
| { | |
| "epoch": 2.1067961165048543, | |
| "grad_norm": 0.6482221636505494, | |
| "kl": 0.435546875, | |
| "learning_rate": 3.339033929066841e-06, | |
| "loss": 0.1938, | |
| "step": 217, | |
| "step_loss": 0.134765625 | |
| }, | |
| { | |
| "epoch": 2.116504854368932, | |
| "grad_norm": 0.5671573525815887, | |
| "kl": 0.28515625, | |
| "learning_rate": 3.3256462495874804e-06, | |
| "loss": 0.2086, | |
| "step": 218, | |
| "step_loss": 0.259765625 | |
| }, | |
| { | |
| "epoch": 2.1262135922330097, | |
| "grad_norm": 0.5195068124172081, | |
| "kl": 0.1962890625, | |
| "learning_rate": 3.3122367270370193e-06, | |
| "loss": 0.1836, | |
| "step": 219, | |
| "step_loss": 0.33984375 | |
| }, | |
| { | |
| "epoch": 2.1359223300970873, | |
| "grad_norm": 0.5586205055924973, | |
| "kl": 0.2373046875, | |
| "learning_rate": 3.2988058702438493e-06, | |
| "loss": 0.2151, | |
| "step": 220, | |
| "step_loss": 0.1455078125 | |
| }, | |
| { | |
| "epoch": 2.145631067961165, | |
| "grad_norm": 0.4963165143500007, | |
| "kl": 0.3046875, | |
| "learning_rate": 3.285354188845892e-06, | |
| "loss": 0.1908, | |
| "step": 221, | |
| "step_loss": 0.140625 | |
| }, | |
| { | |
| "epoch": 2.1553398058252426, | |
| "grad_norm": 0.5444372065813423, | |
| "kl": 0.296875, | |
| "learning_rate": 3.271882193271271e-06, | |
| "loss": 0.2084, | |
| "step": 222, | |
| "step_loss": 0.18359375 | |
| }, | |
| { | |
| "epoch": 2.1650485436893203, | |
| "grad_norm": 0.5399029855207185, | |
| "kl": 1.0546875, | |
| "learning_rate": 3.258390394718933e-06, | |
| "loss": 0.2045, | |
| "step": 223, | |
| "step_loss": 0.3125 | |
| }, | |
| { | |
| "epoch": 2.174757281553398, | |
| "grad_norm": 0.5319895925471867, | |
| "kl": 0.345703125, | |
| "learning_rate": 3.2448793051392546e-06, | |
| "loss": 0.1971, | |
| "step": 224, | |
| "step_loss": 0.109375 | |
| }, | |
| { | |
| "epoch": 2.1844660194174756, | |
| "grad_norm": 0.6543149803275666, | |
| "kl": 0.162109375, | |
| "learning_rate": 3.231349437214619e-06, | |
| "loss": 0.1733, | |
| "step": 225, | |
| "step_loss": 0.267578125 | |
| }, | |
| { | |
| "epoch": 2.1941747572815533, | |
| "grad_norm": 0.5510963728874484, | |
| "kl": 0.2021484375, | |
| "learning_rate": 3.2178013043399588e-06, | |
| "loss": 0.172, | |
| "step": 226, | |
| "step_loss": 0.2109375 | |
| }, | |
| { | |
| "epoch": 2.203883495145631, | |
| "grad_norm": 0.5155354624371081, | |
| "kl": 0.435546875, | |
| "learning_rate": 3.2042354206032733e-06, | |
| "loss": 0.1785, | |
| "step": 227, | |
| "step_loss": 0.1484375 | |
| }, | |
| { | |
| "epoch": 2.2135922330097086, | |
| "grad_norm": 0.5053157167916185, | |
| "kl": 0.341796875, | |
| "learning_rate": 3.190652300766126e-06, | |
| "loss": 0.1685, | |
| "step": 228, | |
| "step_loss": 0.0595703125 | |
| }, | |
| { | |
| "epoch": 2.2233009708737863, | |
| "grad_norm": 0.5412285451182728, | |
| "kl": 0.294921875, | |
| "learning_rate": 3.1770524602441085e-06, | |
| "loss": 0.1945, | |
| "step": 229, | |
| "step_loss": 0.171875 | |
| }, | |
| { | |
| "epoch": 2.233009708737864, | |
| "grad_norm": 0.7645124019667211, | |
| "kl": 0.337890625, | |
| "learning_rate": 3.163436415087283e-06, | |
| "loss": 0.1786, | |
| "step": 230, | |
| "step_loss": 0.0947265625 | |
| }, | |
| { | |
| "epoch": 2.2427184466019416, | |
| "grad_norm": 0.5902838675490253, | |
| "kl": 0.177734375, | |
| "learning_rate": 3.1498046819606046e-06, | |
| "loss": 0.1995, | |
| "step": 231, | |
| "step_loss": 0.1884765625 | |
| }, | |
| { | |
| "epoch": 2.2524271844660193, | |
| "grad_norm": 0.6016854494954839, | |
| "kl": 0.416015625, | |
| "learning_rate": 3.13615777812431e-06, | |
| "loss": 0.2059, | |
| "step": 232, | |
| "step_loss": 0.189453125 | |
| }, | |
| { | |
| "epoch": 2.262135922330097, | |
| "grad_norm": 0.49283185288305054, | |
| "kl": 0.169921875, | |
| "learning_rate": 3.122496221414293e-06, | |
| "loss": 0.1833, | |
| "step": 233, | |
| "step_loss": 0.1328125 | |
| }, | |
| { | |
| "epoch": 2.2718446601941746, | |
| "grad_norm": 0.5176674531829142, | |
| "kl": 0.1953125, | |
| "learning_rate": 3.108820530222458e-06, | |
| "loss": 0.2065, | |
| "step": 234, | |
| "step_loss": 0.2734375 | |
| }, | |
| { | |
| "epoch": 2.2815533980582523, | |
| "grad_norm": 0.571466923601719, | |
| "kl": 0.234375, | |
| "learning_rate": 3.0951312234770427e-06, | |
| "loss": 0.1995, | |
| "step": 235, | |
| "step_loss": 0.1318359375 | |
| }, | |
| { | |
| "epoch": 2.29126213592233, | |
| "grad_norm": 0.46009947680771285, | |
| "kl": 0.283203125, | |
| "learning_rate": 3.081428820622935e-06, | |
| "loss": 0.1821, | |
| "step": 236, | |
| "step_loss": 0.1689453125 | |
| }, | |
| { | |
| "epoch": 2.3009708737864076, | |
| "grad_norm": 0.4656876349084823, | |
| "kl": 0.2158203125, | |
| "learning_rate": 3.067713841601956e-06, | |
| "loss": 0.1744, | |
| "step": 237, | |
| "step_loss": 0.1328125 | |
| }, | |
| { | |
| "epoch": 2.3106796116504853, | |
| "grad_norm": 0.5238872133477445, | |
| "kl": 0.1767578125, | |
| "learning_rate": 3.0539868068331345e-06, | |
| "loss": 0.21, | |
| "step": 238, | |
| "step_loss": 0.353515625 | |
| }, | |
| { | |
| "epoch": 2.320388349514563, | |
| "grad_norm": 0.5445354007250789, | |
| "kl": 0.240234375, | |
| "learning_rate": 3.040248237192958e-06, | |
| "loss": 0.1729, | |
| "step": 239, | |
| "step_loss": 0.12109375 | |
| }, | |
| { | |
| "epoch": 2.3300970873786406, | |
| "grad_norm": 0.5420652323121445, | |
| "kl": 0.193359375, | |
| "learning_rate": 3.026498653995607e-06, | |
| "loss": 0.2176, | |
| "step": 240, | |
| "step_loss": 0.330078125 | |
| }, | |
| { | |
| "epoch": 2.3398058252427183, | |
| "grad_norm": 0.5088097390569825, | |
| "kl": 0.361328125, | |
| "learning_rate": 3.0127385789731773e-06, | |
| "loss": 0.199, | |
| "step": 241, | |
| "step_loss": 0.107421875 | |
| }, | |
| { | |
| "epoch": 2.349514563106796, | |
| "grad_norm": 0.5272400930748196, | |
| "kl": 0.1708984375, | |
| "learning_rate": 2.9989685342558776e-06, | |
| "loss": 0.1906, | |
| "step": 242, | |
| "step_loss": 0.21484375 | |
| }, | |
| { | |
| "epoch": 2.3592233009708736, | |
| "grad_norm": 0.5369149150820804, | |
| "kl": 0.16015625, | |
| "learning_rate": 2.9851890423522214e-06, | |
| "loss": 0.1698, | |
| "step": 243, | |
| "step_loss": 0.2080078125 | |
| }, | |
| { | |
| "epoch": 2.3689320388349513, | |
| "grad_norm": 0.48739367878418577, | |
| "kl": 0.18359375, | |
| "learning_rate": 2.9714006261291967e-06, | |
| "loss": 0.1903, | |
| "step": 244, | |
| "step_loss": 0.123046875 | |
| }, | |
| { | |
| "epoch": 2.378640776699029, | |
| "grad_norm": 0.5086942793116336, | |
| "kl": 0.28515625, | |
| "learning_rate": 2.9576038087924304e-06, | |
| "loss": 0.1919, | |
| "step": 245, | |
| "step_loss": 0.1181640625 | |
| }, | |
| { | |
| "epoch": 2.3883495145631066, | |
| "grad_norm": 0.511491215329325, | |
| "kl": 0.287109375, | |
| "learning_rate": 2.943799113866329e-06, | |
| "loss": 0.17, | |
| "step": 246, | |
| "step_loss": 0.1376953125 | |
| }, | |
| { | |
| "epoch": 2.3980582524271843, | |
| "grad_norm": 0.4862420492331248, | |
| "kl": 0.2109375, | |
| "learning_rate": 2.929987065174219e-06, | |
| "loss": 0.1858, | |
| "step": 247, | |
| "step_loss": 0.0888671875 | |
| }, | |
| { | |
| "epoch": 2.407766990291262, | |
| "grad_norm": 0.5342355311910822, | |
| "kl": 0.408203125, | |
| "learning_rate": 2.9161681868184673e-06, | |
| "loss": 0.1747, | |
| "step": 248, | |
| "step_loss": 0.103515625 | |
| }, | |
| { | |
| "epoch": 2.4174757281553396, | |
| "grad_norm": 0.48746202027489255, | |
| "kl": 0.359375, | |
| "learning_rate": 2.9023430031605928e-06, | |
| "loss": 0.1628, | |
| "step": 249, | |
| "step_loss": 0.12890625 | |
| }, | |
| { | |
| "epoch": 2.4271844660194173, | |
| "grad_norm": 0.5273774877988189, | |
| "kl": 0.30078125, | |
| "learning_rate": 2.888512038801372e-06, | |
| "loss": 0.1826, | |
| "step": 250, | |
| "step_loss": 0.1435546875 | |
| }, | |
| { | |
| "epoch": 2.436893203883495, | |
| "grad_norm": 0.5621617942900121, | |
| "kl": 0.2373046875, | |
| "learning_rate": 2.874675818560933e-06, | |
| "loss": 0.2022, | |
| "step": 251, | |
| "step_loss": 0.1572265625 | |
| }, | |
| { | |
| "epoch": 2.4466019417475726, | |
| "grad_norm": 0.5789885643882243, | |
| "kl": 0.208984375, | |
| "learning_rate": 2.8608348674588383e-06, | |
| "loss": 0.1643, | |
| "step": 252, | |
| "step_loss": 0.1279296875 | |
| }, | |
| { | |
| "epoch": 2.4563106796116507, | |
| "grad_norm": 0.5470908147857678, | |
| "kl": 0.12060546875, | |
| "learning_rate": 2.8469897106941657e-06, | |
| "loss": 0.1879, | |
| "step": 253, | |
| "step_loss": 0.1455078125 | |
| }, | |
| { | |
| "epoch": 2.466019417475728, | |
| "grad_norm": 0.5350773885496009, | |
| "kl": 0.24609375, | |
| "learning_rate": 2.8331408736255766e-06, | |
| "loss": 0.1936, | |
| "step": 254, | |
| "step_loss": 0.2197265625 | |
| }, | |
| { | |
| "epoch": 2.475728155339806, | |
| "grad_norm": 0.5529985369495677, | |
| "kl": 0.251953125, | |
| "learning_rate": 2.8192888817513844e-06, | |
| "loss": 0.1665, | |
| "step": 255, | |
| "step_loss": 0.1728515625 | |
| }, | |
| { | |
| "epoch": 2.4854368932038833, | |
| "grad_norm": 0.5432455331659981, | |
| "kl": 0.10888671875, | |
| "learning_rate": 2.8054342606896102e-06, | |
| "loss": 0.1795, | |
| "step": 256, | |
| "step_loss": 0.09912109375 | |
| }, | |
| { | |
| "epoch": 2.4951456310679614, | |
| "grad_norm": 0.6472916647922633, | |
| "kl": 0.306640625, | |
| "learning_rate": 2.7915775361580427e-06, | |
| "loss": 0.178, | |
| "step": 257, | |
| "step_loss": 0.1982421875 | |
| }, | |
| { | |
| "epoch": 2.5048543689320386, | |
| "grad_norm": 0.5736017610369268, | |
| "kl": 0.349609375, | |
| "learning_rate": 2.7777192339542867e-06, | |
| "loss": 0.22, | |
| "step": 258, | |
| "step_loss": 0.248046875 | |
| }, | |
| { | |
| "epoch": 2.5145631067961167, | |
| "grad_norm": 0.5113226835633047, | |
| "kl": 0.19140625, | |
| "learning_rate": 2.7638598799358123e-06, | |
| "loss": 0.1816, | |
| "step": 259, | |
| "step_loss": 0.12109375 | |
| }, | |
| { | |
| "epoch": 2.524271844660194, | |
| "grad_norm": 0.4926253445141819, | |
| "kl": 0.333984375, | |
| "learning_rate": 2.7500000000000004e-06, | |
| "loss": 0.187, | |
| "step": 260, | |
| "step_loss": 0.1181640625 | |
| }, | |
| { | |
| "epoch": 2.533980582524272, | |
| "grad_norm": 0.5196079398321654, | |
| "kl": 0.15625, | |
| "learning_rate": 2.7361401200641884e-06, | |
| "loss": 0.2047, | |
| "step": 261, | |
| "step_loss": 0.16796875 | |
| }, | |
| { | |
| "epoch": 2.5436893203883493, | |
| "grad_norm": 0.5367574475148829, | |
| "kl": 0.134765625, | |
| "learning_rate": 2.722280766045714e-06, | |
| "loss": 0.1701, | |
| "step": 262, | |
| "step_loss": 0.2275390625 | |
| }, | |
| { | |
| "epoch": 2.5533980582524274, | |
| "grad_norm": 0.6249660005858519, | |
| "kl": 0.298828125, | |
| "learning_rate": 2.708422463841958e-06, | |
| "loss": 0.2044, | |
| "step": 263, | |
| "step_loss": 0.3203125 | |
| }, | |
| { | |
| "epoch": 2.5631067961165046, | |
| "grad_norm": 0.5255108947368616, | |
| "kl": 0.142578125, | |
| "learning_rate": 2.6945657393103913e-06, | |
| "loss": 0.2174, | |
| "step": 264, | |
| "step_loss": 0.267578125 | |
| }, | |
| { | |
| "epoch": 2.5728155339805827, | |
| "grad_norm": 0.5000574555764202, | |
| "kl": 0.1171875, | |
| "learning_rate": 2.680711118248617e-06, | |
| "loss": 0.1692, | |
| "step": 265, | |
| "step_loss": 0.10302734375 | |
| }, | |
| { | |
| "epoch": 2.58252427184466, | |
| "grad_norm": 0.5928812196080804, | |
| "kl": 0.1923828125, | |
| "learning_rate": 2.666859126374425e-06, | |
| "loss": 0.2014, | |
| "step": 266, | |
| "step_loss": 0.3203125 | |
| }, | |
| { | |
| "epoch": 2.592233009708738, | |
| "grad_norm": 0.520932366174803, | |
| "kl": 0.2490234375, | |
| "learning_rate": 2.653010289305835e-06, | |
| "loss": 0.1857, | |
| "step": 267, | |
| "step_loss": 0.134765625 | |
| }, | |
| { | |
| "epoch": 2.6019417475728153, | |
| "grad_norm": 0.5199064176512976, | |
| "kl": 0.255859375, | |
| "learning_rate": 2.639165132541162e-06, | |
| "loss": 0.1805, | |
| "step": 268, | |
| "step_loss": 0.234375 | |
| }, | |
| { | |
| "epoch": 2.6116504854368934, | |
| "grad_norm": 0.5189304175947423, | |
| "kl": 0.259765625, | |
| "learning_rate": 2.625324181439068e-06, | |
| "loss": 0.1857, | |
| "step": 269, | |
| "step_loss": 0.333984375 | |
| }, | |
| { | |
| "epoch": 2.6213592233009706, | |
| "grad_norm": 0.5477820911661071, | |
| "kl": 0.1962890625, | |
| "learning_rate": 2.611487961198629e-06, | |
| "loss": 0.1705, | |
| "step": 270, | |
| "step_loss": 0.267578125 | |
| }, | |
| { | |
| "epoch": 2.6310679611650487, | |
| "grad_norm": 0.5144960586834338, | |
| "kl": 0.3515625, | |
| "learning_rate": 2.5976569968394084e-06, | |
| "loss": 0.1824, | |
| "step": 271, | |
| "step_loss": 0.12890625 | |
| }, | |
| { | |
| "epoch": 2.6407766990291264, | |
| "grad_norm": 0.5129225163884132, | |
| "kl": 0.1611328125, | |
| "learning_rate": 2.583831813181534e-06, | |
| "loss": 0.1936, | |
| "step": 272, | |
| "step_loss": 0.240234375 | |
| }, | |
| { | |
| "epoch": 2.650485436893204, | |
| "grad_norm": 0.5747902549826789, | |
| "kl": 0.376953125, | |
| "learning_rate": 2.570012934825782e-06, | |
| "loss": 0.1803, | |
| "step": 273, | |
| "step_loss": 0.171875 | |
| }, | |
| { | |
| "epoch": 2.6601941747572817, | |
| "grad_norm": 0.5303667804765286, | |
| "kl": 0.2353515625, | |
| "learning_rate": 2.556200886133672e-06, | |
| "loss": 0.1638, | |
| "step": 274, | |
| "step_loss": 0.353515625 | |
| }, | |
| { | |
| "epoch": 2.6699029126213594, | |
| "grad_norm": 0.552737610069486, | |
| "kl": 0.166015625, | |
| "learning_rate": 2.542396191207571e-06, | |
| "loss": 0.1896, | |
| "step": 275, | |
| "step_loss": 0.212890625 | |
| }, | |
| { | |
| "epoch": 2.679611650485437, | |
| "grad_norm": 0.5138863525528505, | |
| "kl": 0.267578125, | |
| "learning_rate": 2.528599373870804e-06, | |
| "loss": 0.1647, | |
| "step": 276, | |
| "step_loss": 0.1650390625 | |
| }, | |
| { | |
| "epoch": 2.6893203883495147, | |
| "grad_norm": 0.5816837784109217, | |
| "kl": 0.3203125, | |
| "learning_rate": 2.5148109576477798e-06, | |
| "loss": 0.2022, | |
| "step": 277, | |
| "step_loss": 0.2734375 | |
| }, | |
| { | |
| "epoch": 2.6990291262135924, | |
| "grad_norm": 0.5946206686019718, | |
| "kl": 0.296875, | |
| "learning_rate": 2.501031465744123e-06, | |
| "loss": 0.1793, | |
| "step": 278, | |
| "step_loss": 0.1650390625 | |
| }, | |
| { | |
| "epoch": 2.70873786407767, | |
| "grad_norm": 0.5505786008910999, | |
| "kl": 0.26171875, | |
| "learning_rate": 2.487261421026823e-06, | |
| "loss": 0.1869, | |
| "step": 279, | |
| "step_loss": 0.189453125 | |
| }, | |
| { | |
| "epoch": 2.7184466019417477, | |
| "grad_norm": 0.5607618391578882, | |
| "kl": 0.396484375, | |
| "learning_rate": 2.4735013460043933e-06, | |
| "loss": 0.1782, | |
| "step": 280, | |
| "step_loss": 0.140625 | |
| }, | |
| { | |
| "epoch": 2.7281553398058254, | |
| "grad_norm": 0.532008359083315, | |
| "kl": 0.443359375, | |
| "learning_rate": 2.4597517628070427e-06, | |
| "loss": 0.2004, | |
| "step": 281, | |
| "step_loss": 0.1953125 | |
| }, | |
| { | |
| "epoch": 2.737864077669903, | |
| "grad_norm": 0.5409775868618988, | |
| "kl": 0.412109375, | |
| "learning_rate": 2.446013193166866e-06, | |
| "loss": 0.1968, | |
| "step": 282, | |
| "step_loss": 0.10302734375 | |
| }, | |
| { | |
| "epoch": 2.7475728155339807, | |
| "grad_norm": 0.5494599421845879, | |
| "kl": 0.291015625, | |
| "learning_rate": 2.4322861583980445e-06, | |
| "loss": 0.1882, | |
| "step": 283, | |
| "step_loss": 0.138671875 | |
| }, | |
| { | |
| "epoch": 2.7572815533980584, | |
| "grad_norm": 0.5270774451468452, | |
| "kl": 0.275390625, | |
| "learning_rate": 2.4185711793770662e-06, | |
| "loss": 0.1773, | |
| "step": 284, | |
| "step_loss": 0.1728515625 | |
| }, | |
| { | |
| "epoch": 2.766990291262136, | |
| "grad_norm": 0.556579747133581, | |
| "kl": 0.177734375, | |
| "learning_rate": 2.4048687765229584e-06, | |
| "loss": 0.1875, | |
| "step": 285, | |
| "step_loss": 0.12109375 | |
| }, | |
| { | |
| "epoch": 2.7766990291262137, | |
| "grad_norm": 0.503176078809862, | |
| "kl": 0.189453125, | |
| "learning_rate": 2.3911794697775437e-06, | |
| "loss": 0.1847, | |
| "step": 286, | |
| "step_loss": 0.328125 | |
| }, | |
| { | |
| "epoch": 2.7864077669902914, | |
| "grad_norm": 0.5914318608290587, | |
| "kl": 0.251953125, | |
| "learning_rate": 2.377503778585708e-06, | |
| "loss": 0.2007, | |
| "step": 287, | |
| "step_loss": 0.197265625 | |
| }, | |
| { | |
| "epoch": 2.796116504854369, | |
| "grad_norm": 0.47771283753849275, | |
| "kl": 0.365234375, | |
| "learning_rate": 2.3638422218756905e-06, | |
| "loss": 0.1968, | |
| "step": 288, | |
| "step_loss": 0.13671875 | |
| }, | |
| { | |
| "epoch": 2.8058252427184467, | |
| "grad_norm": 0.544279916739223, | |
| "kl": 0.1630859375, | |
| "learning_rate": 2.350195318039396e-06, | |
| "loss": 0.1844, | |
| "step": 289, | |
| "step_loss": 0.1279296875 | |
| }, | |
| { | |
| "epoch": 2.8155339805825244, | |
| "grad_norm": 0.7140740591904084, | |
| "kl": 0.2080078125, | |
| "learning_rate": 2.336563584912717e-06, | |
| "loss": 0.2153, | |
| "step": 290, | |
| "step_loss": 0.294921875 | |
| }, | |
| { | |
| "epoch": 2.825242718446602, | |
| "grad_norm": 0.5978474529315786, | |
| "kl": 0.1943359375, | |
| "learning_rate": 2.322947539755892e-06, | |
| "loss": 0.2018, | |
| "step": 291, | |
| "step_loss": 0.25390625 | |
| }, | |
| { | |
| "epoch": 2.8349514563106797, | |
| "grad_norm": 0.5338842771148811, | |
| "kl": 0.4296875, | |
| "learning_rate": 2.3093476992338747e-06, | |
| "loss": 0.2051, | |
| "step": 292, | |
| "step_loss": 0.142578125 | |
| }, | |
| { | |
| "epoch": 2.8446601941747574, | |
| "grad_norm": 0.5434580731788923, | |
| "kl": 0.2021484375, | |
| "learning_rate": 2.295764579396727e-06, | |
| "loss": 0.1833, | |
| "step": 293, | |
| "step_loss": 0.1259765625 | |
| }, | |
| { | |
| "epoch": 2.854368932038835, | |
| "grad_norm": 0.48934430178338256, | |
| "kl": 0.1484375, | |
| "learning_rate": 2.2821986956600415e-06, | |
| "loss": 0.1602, | |
| "step": 294, | |
| "step_loss": 0.10595703125 | |
| }, | |
| { | |
| "epoch": 2.8640776699029127, | |
| "grad_norm": 0.5408973977811805, | |
| "kl": 0.169921875, | |
| "learning_rate": 2.2686505627853812e-06, | |
| "loss": 0.1594, | |
| "step": 295, | |
| "step_loss": 0.1103515625 | |
| }, | |
| { | |
| "epoch": 2.8737864077669903, | |
| "grad_norm": 0.5283594028868295, | |
| "kl": 0.36328125, | |
| "learning_rate": 2.255120694860746e-06, | |
| "loss": 0.1916, | |
| "step": 296, | |
| "step_loss": 0.126953125 | |
| }, | |
| { | |
| "epoch": 2.883495145631068, | |
| "grad_norm": 0.5035651183385199, | |
| "kl": 0.2177734375, | |
| "learning_rate": 2.2416096052810683e-06, | |
| "loss": 0.1824, | |
| "step": 297, | |
| "step_loss": 0.220703125 | |
| }, | |
| { | |
| "epoch": 2.8932038834951457, | |
| "grad_norm": 0.5459862981254268, | |
| "kl": 0.2470703125, | |
| "learning_rate": 2.2281178067287294e-06, | |
| "loss": 0.2043, | |
| "step": 298, | |
| "step_loss": 0.205078125 | |
| }, | |
| { | |
| "epoch": 2.9029126213592233, | |
| "grad_norm": 0.5238184894583012, | |
| "kl": 0.30078125, | |
| "learning_rate": 2.214645811154108e-06, | |
| "loss": 0.1867, | |
| "step": 299, | |
| "step_loss": 0.2119140625 | |
| }, | |
| { | |
| "epoch": 2.912621359223301, | |
| "grad_norm": 0.5030938038128431, | |
| "kl": 0.251953125, | |
| "learning_rate": 2.201194129756152e-06, | |
| "loss": 0.1743, | |
| "step": 300, | |
| "step_loss": 0.1611328125 | |
| }, | |
| { | |
| "epoch": 2.9223300970873787, | |
| "grad_norm": 0.6344887467056015, | |
| "kl": 0.1796875, | |
| "learning_rate": 2.187763272962981e-06, | |
| "loss": 0.1986, | |
| "step": 301, | |
| "step_loss": 0.1328125 | |
| }, | |
| { | |
| "epoch": 2.9320388349514563, | |
| "grad_norm": 0.5310139131483351, | |
| "kl": 0.189453125, | |
| "learning_rate": 2.1743537504125208e-06, | |
| "loss": 0.1998, | |
| "step": 302, | |
| "step_loss": 0.287109375 | |
| }, | |
| { | |
| "epoch": 2.941747572815534, | |
| "grad_norm": 0.5172075780561182, | |
| "kl": 0.4375, | |
| "learning_rate": 2.1609660709331598e-06, | |
| "loss": 0.2207, | |
| "step": 303, | |
| "step_loss": 0.31640625 | |
| }, | |
| { | |
| "epoch": 2.9514563106796117, | |
| "grad_norm": 0.5624076449552908, | |
| "kl": 0.1748046875, | |
| "learning_rate": 2.1476007425244476e-06, | |
| "loss": 0.1955, | |
| "step": 304, | |
| "step_loss": 0.1015625 | |
| }, | |
| { | |
| "epoch": 2.9611650485436893, | |
| "grad_norm": 0.5232258269061456, | |
| "kl": 0.306640625, | |
| "learning_rate": 2.134258272337814e-06, | |
| "loss": 0.1884, | |
| "step": 305, | |
| "step_loss": 0.21484375 | |
| }, | |
| { | |
| "epoch": 2.970873786407767, | |
| "grad_norm": 0.5856686801966252, | |
| "kl": 0.291015625, | |
| "learning_rate": 2.120939166657327e-06, | |
| "loss": 0.202, | |
| "step": 306, | |
| "step_loss": 0.2021484375 | |
| }, | |
| { | |
| "epoch": 2.9805825242718447, | |
| "grad_norm": 0.5360709399599322, | |
| "kl": 0.29296875, | |
| "learning_rate": 2.1076439308804813e-06, | |
| "loss": 0.1946, | |
| "step": 307, | |
| "step_loss": 0.150390625 | |
| }, | |
| { | |
| "epoch": 2.9902912621359223, | |
| "grad_norm": 0.5186377564757945, | |
| "kl": 0.1923828125, | |
| "learning_rate": 2.0943730694990204e-06, | |
| "loss": 0.1776, | |
| "step": 308, | |
| "step_loss": 0.1015625 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "grad_norm": 0.5769108575358282, | |
| "kl": 0.48828125, | |
| "learning_rate": 2.081127086079795e-06, | |
| "loss": 0.1656, | |
| "step": 309, | |
| "step_loss": 0.240234375 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_test_transformed.json_loss": NaN, | |
| "eval_test_transformed.json_runtime": 8.3956, | |
| "eval_test_transformed.json_samples_per_second": 59.555, | |
| "eval_test_transformed.json_steps_per_second": 1.906, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 3.0097087378640777, | |
| "grad_norm": 0.5711934327513365, | |
| "kl": NaN, | |
| "learning_rate": 2.0679064832456523e-06, | |
| "loss": 0.1695, | |
| "step": 310, | |
| "step_loss": NaN | |
| }, | |
| { | |
| "epoch": 3.0194174757281553, | |
| "grad_norm": 0.5080776402736003, | |
| "kl": 0.2255859375, | |
| "learning_rate": 2.054711762656369e-06, | |
| "loss": 0.1599, | |
| "step": 311, | |
| "step_loss": 0.2373046875 | |
| }, | |
| { | |
| "epoch": 3.029126213592233, | |
| "grad_norm": 0.460829560524234, | |
| "kl": 0.154296875, | |
| "learning_rate": 2.0415434249896075e-06, | |
| "loss": 0.1554, | |
| "step": 312, | |
| "step_loss": 0.08984375 | |
| }, | |
| { | |
| "epoch": 3.0388349514563107, | |
| "grad_norm": 0.532111983258558, | |
| "kl": 0.259765625, | |
| "learning_rate": 2.0284019699219265e-06, | |
| "loss": 0.1561, | |
| "step": 313, | |
| "step_loss": 0.1689453125 | |
| }, | |
| { | |
| "epoch": 3.0485436893203883, | |
| "grad_norm": 0.5807336256828239, | |
| "kl": 0.369140625, | |
| "learning_rate": 2.0152878961098133e-06, | |
| "loss": 0.1513, | |
| "step": 314, | |
| "step_loss": 0.06201171875 | |
| }, | |
| { | |
| "epoch": 3.058252427184466, | |
| "grad_norm": 0.5177785233543187, | |
| "kl": 0.1982421875, | |
| "learning_rate": 2.0022017011707663e-06, | |
| "loss": 0.1816, | |
| "step": 315, | |
| "step_loss": 0.2236328125 | |
| }, | |
| { | |
| "epoch": 3.0679611650485437, | |
| "grad_norm": 0.5969479421413822, | |
| "kl": 0.333984375, | |
| "learning_rate": 1.989143881664412e-06, | |
| "loss": 0.1782, | |
| "step": 316, | |
| "step_loss": 0.240234375 | |
| }, | |
| { | |
| "epoch": 3.0776699029126213, | |
| "grad_norm": 0.489029791228728, | |
| "kl": 0.59765625, | |
| "learning_rate": 1.9761149330736625e-06, | |
| "loss": 0.1308, | |
| "step": 317, | |
| "step_loss": 0.064453125 | |
| }, | |
| { | |
| "epoch": 3.087378640776699, | |
| "grad_norm": 0.7131615528898937, | |
| "kl": 0.177734375, | |
| "learning_rate": 1.9631153497859127e-06, | |
| "loss": 0.179, | |
| "step": 318, | |
| "step_loss": 0.07861328125 | |
| }, | |
| { | |
| "epoch": 3.0970873786407767, | |
| "grad_norm": 0.4640669864219324, | |
| "kl": 0.390625, | |
| "learning_rate": 1.950145625074285e-06, | |
| "loss": 0.1331, | |
| "step": 319, | |
| "step_loss": 0.06494140625 | |
| }, | |
| { | |
| "epoch": 3.1067961165048543, | |
| "grad_norm": 0.4889269717181464, | |
| "kl": 0.2138671875, | |
| "learning_rate": 1.9372062510789064e-06, | |
| "loss": 0.1696, | |
| "step": 320, | |
| "step_loss": 0.07861328125 | |
| }, | |
| { | |
| "epoch": 3.116504854368932, | |
| "grad_norm": 0.4894140759210115, | |
| "kl": 0.294921875, | |
| "learning_rate": 1.924297718788238e-06, | |
| "loss": 0.1532, | |
| "step": 321, | |
| "step_loss": 0.1728515625 | |
| }, | |
| { | |
| "epoch": 3.1262135922330097, | |
| "grad_norm": 0.4851533726900198, | |
| "kl": 0.271484375, | |
| "learning_rate": 1.9114205180204437e-06, | |
| "loss": 0.1563, | |
| "step": 322, | |
| "step_loss": 0.08984375 | |
| }, | |
| { | |
| "epoch": 3.1359223300970873, | |
| "grad_norm": 0.4812749750914589, | |
| "kl": 0.3359375, | |
| "learning_rate": 1.8985751374048022e-06, | |
| "loss": 0.1597, | |
| "step": 323, | |
| "step_loss": 0.083984375 | |
| }, | |
| { | |
| "epoch": 3.145631067961165, | |
| "grad_norm": 0.4735366094358067, | |
| "kl": 0.57421875, | |
| "learning_rate": 1.8857620643631675e-06, | |
| "loss": 0.164, | |
| "step": 324, | |
| "step_loss": 0.19140625 | |
| }, | |
| { | |
| "epoch": 3.1553398058252426, | |
| "grad_norm": 0.4616311559837366, | |
| "kl": 0.3203125, | |
| "learning_rate": 1.8729817850914717e-06, | |
| "loss": 0.169, | |
| "step": 325, | |
| "step_loss": 0.2294921875 | |
| }, | |
| { | |
| "epoch": 3.1650485436893203, | |
| "grad_norm": 0.5563217588279268, | |
| "kl": 0.224609375, | |
| "learning_rate": 1.8602347845412799e-06, | |
| "loss": 0.1502, | |
| "step": 326, | |
| "step_loss": 0.28515625 | |
| }, | |
| { | |
| "epoch": 3.174757281553398, | |
| "grad_norm": 0.45499622442963633, | |
| "kl": 0.330078125, | |
| "learning_rate": 1.847521546401383e-06, | |
| "loss": 0.1547, | |
| "step": 327, | |
| "step_loss": 0.062255859375 | |
| }, | |
| { | |
| "epoch": 3.1844660194174756, | |
| "grad_norm": 0.5222912664403042, | |
| "kl": 0.33984375, | |
| "learning_rate": 1.83484255307945e-06, | |
| "loss": 0.1694, | |
| "step": 328, | |
| "step_loss": 0.1083984375 | |
| }, | |
| { | |
| "epoch": 3.1941747572815533, | |
| "grad_norm": 0.4698877989732488, | |
| "kl": 0.1611328125, | |
| "learning_rate": 1.8221982856837177e-06, | |
| "loss": 0.1616, | |
| "step": 329, | |
| "step_loss": 0.203125 | |
| }, | |
| { | |
| "epoch": 3.203883495145631, | |
| "grad_norm": 0.5161358585670528, | |
| "kl": 0.2578125, | |
| "learning_rate": 1.8095892240047375e-06, | |
| "loss": 0.1483, | |
| "step": 330, | |
| "step_loss": 0.193359375 | |
| }, | |
| { | |
| "epoch": 3.2135922330097086, | |
| "grad_norm": 0.5707034508549506, | |
| "kl": 0.384765625, | |
| "learning_rate": 1.7970158464971704e-06, | |
| "loss": 0.1764, | |
| "step": 331, | |
| "step_loss": 0.1435546875 | |
| }, | |
| { | |
| "epoch": 3.2233009708737863, | |
| "grad_norm": 0.6845693547436684, | |
| "kl": 0.30078125, | |
| "learning_rate": 1.78447863026163e-06, | |
| "loss": 0.1686, | |
| "step": 332, | |
| "step_loss": 0.146484375 | |
| }, | |
| { | |
| "epoch": 3.233009708737864, | |
| "grad_norm": 0.539242866306224, | |
| "kl": 0.279296875, | |
| "learning_rate": 1.77197805102658e-06, | |
| "loss": 0.1602, | |
| "step": 333, | |
| "step_loss": 0.30078125 | |
| }, | |
| { | |
| "epoch": 3.2427184466019416, | |
| "grad_norm": 0.5376628174473539, | |
| "kl": 0.318359375, | |
| "learning_rate": 1.759514583130279e-06, | |
| "loss": 0.1596, | |
| "step": 334, | |
| "step_loss": 0.0859375 | |
| }, | |
| { | |
| "epoch": 3.2524271844660193, | |
| "grad_norm": 0.488395082556356, | |
| "kl": 0.421875, | |
| "learning_rate": 1.7470886995027902e-06, | |
| "loss": 0.1437, | |
| "step": 335, | |
| "step_loss": 0.07080078125 | |
| }, | |
| { | |
| "epoch": 3.262135922330097, | |
| "grad_norm": 0.5055787180594609, | |
| "kl": 0.2099609375, | |
| "learning_rate": 1.734700871648024e-06, | |
| "loss": 0.1565, | |
| "step": 336, | |
| "step_loss": 0.1953125 | |
| }, | |
| { | |
| "epoch": 3.2718446601941746, | |
| "grad_norm": 0.4884350496305089, | |
| "kl": 0.2431640625, | |
| "learning_rate": 1.722351569625859e-06, | |
| "loss": 0.1445, | |
| "step": 337, | |
| "step_loss": 0.07666015625 | |
| }, | |
| { | |
| "epoch": 3.2815533980582523, | |
| "grad_norm": 0.5055878842267785, | |
| "kl": 0.1669921875, | |
| "learning_rate": 1.710041262034296e-06, | |
| "loss": 0.1662, | |
| "step": 338, | |
| "step_loss": 0.1298828125 | |
| }, | |
| { | |
| "epoch": 3.29126213592233, | |
| "grad_norm": 0.6568055753017068, | |
| "kl": 0.27734375, | |
| "learning_rate": 1.6977704159916801e-06, | |
| "loss": 0.1607, | |
| "step": 339, | |
| "step_loss": 0.13671875 | |
| }, | |
| { | |
| "epoch": 3.3009708737864076, | |
| "grad_norm": 0.46693217339820037, | |
| "kl": 0.326171875, | |
| "learning_rate": 1.6855394971189779e-06, | |
| "loss": 0.1519, | |
| "step": 340, | |
| "step_loss": 0.0751953125 | |
| }, | |
| { | |
| "epoch": 3.3106796116504853, | |
| "grad_norm": 0.46750716114679314, | |
| "kl": 0.224609375, | |
| "learning_rate": 1.6733489695221056e-06, | |
| "loss": 0.165, | |
| "step": 341, | |
| "step_loss": 0.25 | |
| }, | |
| { | |
| "epoch": 3.320388349514563, | |
| "grad_norm": 0.46663445875307946, | |
| "kl": 0.359375, | |
| "learning_rate": 1.6611992957743217e-06, | |
| "loss": 0.151, | |
| "step": 342, | |
| "step_loss": 0.1318359375 | |
| }, | |
| { | |
| "epoch": 3.3300970873786406, | |
| "grad_norm": 0.48545644382954584, | |
| "kl": 0.1982421875, | |
| "learning_rate": 1.6490909368986725e-06, | |
| "loss": 0.1624, | |
| "step": 343, | |
| "step_loss": 0.1748046875 | |
| }, | |
| { | |
| "epoch": 3.3398058252427183, | |
| "grad_norm": 0.4692569851993047, | |
| "kl": 0.232421875, | |
| "learning_rate": 1.637024352350498e-06, | |
| "loss": 0.1414, | |
| "step": 344, | |
| "step_loss": 0.271484375 | |
| }, | |
| { | |
| "epoch": 3.349514563106796, | |
| "grad_norm": 0.4758629432786183, | |
| "kl": 0.2490234375, | |
| "learning_rate": 1.6250000000000007e-06, | |
| "loss": 0.1427, | |
| "step": 345, | |
| "step_loss": 0.150390625 | |
| }, | |
| { | |
| "epoch": 3.3592233009708736, | |
| "grad_norm": 0.4871466615503954, | |
| "kl": 0.29296875, | |
| "learning_rate": 1.6130183361148675e-06, | |
| "loss": 0.1593, | |
| "step": 346, | |
| "step_loss": 0.1748046875 | |
| }, | |
| { | |
| "epoch": 3.3689320388349513, | |
| "grad_norm": 0.5474079532412777, | |
| "kl": 0.32421875, | |
| "learning_rate": 1.60107981534296e-06, | |
| "loss": 0.178, | |
| "step": 347, | |
| "step_loss": 0.267578125 | |
| }, | |
| { | |
| "epoch": 3.378640776699029, | |
| "grad_norm": 2.4760425876288865, | |
| "kl": 0.25390625, | |
| "learning_rate": 1.5891848906950602e-06, | |
| "loss": 0.1768, | |
| "step": 348, | |
| "step_loss": 0.3515625 | |
| }, | |
| { | |
| "epoch": 3.3883495145631066, | |
| "grad_norm": 0.499334607836154, | |
| "kl": 0.3359375, | |
| "learning_rate": 1.5773340135276843e-06, | |
| "loss": 0.1309, | |
| "step": 349, | |
| "step_loss": 0.07470703125 | |
| }, | |
| { | |
| "epoch": 3.3980582524271843, | |
| "grad_norm": 0.5121888774785807, | |
| "kl": 0.3046875, | |
| "learning_rate": 1.5655276335259495e-06, | |
| "loss": 0.146, | |
| "step": 350, | |
| "step_loss": 0.126953125 | |
| }, | |
| { | |
| "epoch": 3.407766990291262, | |
| "grad_norm": 0.5311085568794023, | |
| "kl": 0.427734375, | |
| "learning_rate": 1.5537661986865196e-06, | |
| "loss": 0.1675, | |
| "step": 351, | |
| "step_loss": 0.1103515625 | |
| }, | |
| { | |
| "epoch": 3.4174757281553396, | |
| "grad_norm": 0.5128192279567273, | |
| "kl": 0.271484375, | |
| "learning_rate": 1.542050155300595e-06, | |
| "loss": 0.1672, | |
| "step": 352, | |
| "step_loss": 0.2197265625 | |
| }, | |
| { | |
| "epoch": 3.4271844660194173, | |
| "grad_norm": 0.49745836081284944, | |
| "kl": 0.26171875, | |
| "learning_rate": 1.53037994793699e-06, | |
| "loss": 0.1568, | |
| "step": 353, | |
| "step_loss": 0.18359375 | |
| }, | |
| { | |
| "epoch": 3.436893203883495, | |
| "grad_norm": 0.48051427609991737, | |
| "kl": 0.263671875, | |
| "learning_rate": 1.51875601942525e-06, | |
| "loss": 0.1625, | |
| "step": 354, | |
| "step_loss": 0.2265625 | |
| }, | |
| { | |
| "epoch": 3.4466019417475726, | |
| "grad_norm": 0.48409810919704643, | |
| "kl": 0.29296875, | |
| "learning_rate": 1.507178810838863e-06, | |
| "loss": 0.1706, | |
| "step": 355, | |
| "step_loss": 0.1845703125 | |
| }, | |
| { | |
| "epoch": 3.4563106796116507, | |
| "grad_norm": 0.537826558649176, | |
| "kl": 0.349609375, | |
| "learning_rate": 1.4956487614785076e-06, | |
| "loss": 0.16, | |
| "step": 356, | |
| "step_loss": 0.15625 | |
| }, | |
| { | |
| "epoch": 3.466019417475728, | |
| "grad_norm": 0.507752429637045, | |
| "kl": 0.345703125, | |
| "learning_rate": 1.4841663088553992e-06, | |
| "loss": 0.1625, | |
| "step": 357, | |
| "step_loss": 0.12890625 | |
| }, | |
| { | |
| "epoch": 3.475728155339806, | |
| "grad_norm": 0.5130052193009929, | |
| "kl": 0.341796875, | |
| "learning_rate": 1.4727318886746725e-06, | |
| "loss": 0.1741, | |
| "step": 358, | |
| "step_loss": 0.11181640625 | |
| }, | |
| { | |
| "epoch": 3.4854368932038833, | |
| "grad_norm": 0.5556488303190134, | |
| "kl": 0.46484375, | |
| "learning_rate": 1.4613459348188635e-06, | |
| "loss": 0.1242, | |
| "step": 359, | |
| "step_loss": 0.0869140625 | |
| }, | |
| { | |
| "epoch": 3.4951456310679614, | |
| "grad_norm": 0.477620232983351, | |
| "kl": 0.39453125, | |
| "learning_rate": 1.4500088793314351e-06, | |
| "loss": 0.1439, | |
| "step": 360, | |
| "step_loss": 0.0703125 | |
| }, | |
| { | |
| "epoch": 3.5048543689320386, | |
| "grad_norm": 0.5136494794829761, | |
| "kl": 0.2734375, | |
| "learning_rate": 1.438721152400388e-06, | |
| "loss": 0.1511, | |
| "step": 361, | |
| "step_loss": 0.326171875 | |
| }, | |
| { | |
| "epoch": 3.5145631067961167, | |
| "grad_norm": 0.4992556694121463, | |
| "kl": 0.26171875, | |
| "learning_rate": 1.427483182341936e-06, | |
| "loss": 0.1641, | |
| "step": 362, | |
| "step_loss": 0.0810546875 | |
| }, | |
| { | |
| "epoch": 3.524271844660194, | |
| "grad_norm": 0.4965898781365568, | |
| "kl": 0.291015625, | |
| "learning_rate": 1.4162953955842518e-06, | |
| "loss": 0.1717, | |
| "step": 363, | |
| "step_loss": 0.06640625 | |
| }, | |
| { | |
| "epoch": 3.533980582524272, | |
| "grad_norm": 0.4885949309810252, | |
| "kl": 0.2490234375, | |
| "learning_rate": 1.4051582166512895e-06, | |
| "loss": 0.1377, | |
| "step": 364, | |
| "step_loss": 0.0830078125 | |
| }, | |
| { | |
| "epoch": 3.5436893203883493, | |
| "grad_norm": 0.7644545502958816, | |
| "kl": 0.3203125, | |
| "learning_rate": 1.3940720681466735e-06, | |
| "loss": 0.1581, | |
| "step": 365, | |
| "step_loss": 0.17578125 | |
| }, | |
| { | |
| "epoch": 3.5533980582524274, | |
| "grad_norm": 0.549140542595152, | |
| "kl": 0.26953125, | |
| "learning_rate": 1.3830373707376623e-06, | |
| "loss": 0.1455, | |
| "step": 366, | |
| "step_loss": 0.1533203125 | |
| }, | |
| { | |
| "epoch": 3.5631067961165046, | |
| "grad_norm": 0.7512378808438823, | |
| "kl": 0.333984375, | |
| "learning_rate": 1.372054543139188e-06, | |
| "loss": 0.15, | |
| "step": 367, | |
| "step_loss": 0.1103515625 | |
| }, | |
| { | |
| "epoch": 3.5728155339805827, | |
| "grad_norm": 0.5020029362406481, | |
| "kl": 0.19921875, | |
| "learning_rate": 1.3611240020979655e-06, | |
| "loss": 0.1427, | |
| "step": 368, | |
| "step_loss": 0.06494140625 | |
| }, | |
| { | |
| "epoch": 3.58252427184466, | |
| "grad_norm": 0.46369749478886507, | |
| "kl": 0.1650390625, | |
| "learning_rate": 1.3502461623766842e-06, | |
| "loss": 0.1668, | |
| "step": 369, | |
| "step_loss": 0.1728515625 | |
| }, | |
| { | |
| "epoch": 3.592233009708738, | |
| "grad_norm": 0.527596307252248, | |
| "kl": 0.404296875, | |
| "learning_rate": 1.3394214367382602e-06, | |
| "loss": 0.1543, | |
| "step": 370, | |
| "step_loss": 0.19921875 | |
| }, | |
| { | |
| "epoch": 3.6019417475728153, | |
| "grad_norm": 0.5072354485956617, | |
| "kl": 0.2578125, | |
| "learning_rate": 1.3286502359301863e-06, | |
| "loss": 0.1599, | |
| "step": 371, | |
| "step_loss": 0.1904296875 | |
| }, | |
| { | |
| "epoch": 3.6116504854368934, | |
| "grad_norm": 0.5160691631303242, | |
| "kl": 0.439453125, | |
| "learning_rate": 1.3179329686689318e-06, | |
| "loss": 0.1498, | |
| "step": 372, | |
| "step_loss": 0.169921875 | |
| }, | |
| { | |
| "epoch": 3.6213592233009706, | |
| "grad_norm": 0.6554485149692931, | |
| "kl": 0.2412109375, | |
| "learning_rate": 1.3072700416244494e-06, | |
| "loss": 0.1542, | |
| "step": 373, | |
| "step_loss": 0.228515625 | |
| }, | |
| { | |
| "epoch": 3.6310679611650487, | |
| "grad_norm": 0.4910381460677276, | |
| "kl": 0.240234375, | |
| "learning_rate": 1.2966618594047285e-06, | |
| "loss": 0.1407, | |
| "step": 374, | |
| "step_loss": 0.322265625 | |
| }, | |
| { | |
| "epoch": 3.6407766990291264, | |
| "grad_norm": 0.5460180925997682, | |
| "kl": 0.25, | |
| "learning_rate": 1.286108824540456e-06, | |
| "loss": 0.1558, | |
| "step": 375, | |
| "step_loss": 0.2236328125 | |
| }, | |
| { | |
| "epoch": 3.650485436893204, | |
| "grad_norm": 0.584914369592297, | |
| "kl": 0.53125, | |
| "learning_rate": 1.2756113374697294e-06, | |
| "loss": 0.1594, | |
| "step": 376, | |
| "step_loss": 0.1318359375 | |
| }, | |
| { | |
| "epoch": 3.6601941747572817, | |
| "grad_norm": 0.5256624617590849, | |
| "kl": 0.189453125, | |
| "learning_rate": 1.2651697965228748e-06, | |
| "loss": 0.1696, | |
| "step": 377, | |
| "step_loss": 0.1123046875 | |
| }, | |
| { | |
| "epoch": 3.6699029126213594, | |
| "grad_norm": 0.5033922457112555, | |
| "kl": 0.216796875, | |
| "learning_rate": 1.2547845979073194e-06, | |
| "loss": 0.1567, | |
| "step": 378, | |
| "step_loss": 0.2412109375 | |
| }, | |
| { | |
| "epoch": 3.679611650485437, | |
| "grad_norm": 0.5191896792873844, | |
| "kl": 0.1494140625, | |
| "learning_rate": 1.2444561356925692e-06, | |
| "loss": 0.1413, | |
| "step": 379, | |
| "step_loss": 0.10205078125 | |
| }, | |
| { | |
| "epoch": 3.6893203883495147, | |
| "grad_norm": 0.5166370978149198, | |
| "kl": 0.46484375, | |
| "learning_rate": 1.2341848017952464e-06, | |
| "loss": 0.1658, | |
| "step": 380, | |
| "step_loss": 0.1005859375 | |
| }, | |
| { | |
| "epoch": 3.6990291262135924, | |
| "grad_norm": 0.5574979370660633, | |
| "kl": 0.251953125, | |
| "learning_rate": 1.2239709859642237e-06, | |
| "loss": 0.1484, | |
| "step": 381, | |
| "step_loss": 0.2119140625 | |
| }, | |
| { | |
| "epoch": 3.70873786407767, | |
| "grad_norm": 0.5625584889838852, | |
| "kl": 0.5390625, | |
| "learning_rate": 1.2138150757658328e-06, | |
| "loss": 0.1802, | |
| "step": 382, | |
| "step_loss": 0.17578125 | |
| }, | |
| { | |
| "epoch": 3.7184466019417477, | |
| "grad_norm": 0.47927966720420184, | |
| "kl": 0.283203125, | |
| "learning_rate": 1.2037174565691591e-06, | |
| "loss": 0.1506, | |
| "step": 383, | |
| "step_loss": 0.146484375 | |
| }, | |
| { | |
| "epoch": 3.7281553398058254, | |
| "grad_norm": 0.5194167964748884, | |
| "kl": 0.4453125, | |
| "learning_rate": 1.1936785115314176e-06, | |
| "loss": 0.1395, | |
| "step": 384, | |
| "step_loss": 0.10498046875 | |
| }, | |
| { | |
| "epoch": 3.737864077669903, | |
| "grad_norm": 0.5468254611874791, | |
| "kl": 0.306640625, | |
| "learning_rate": 1.1836986215834153e-06, | |
| "loss": 0.1423, | |
| "step": 385, | |
| "step_loss": 0.12158203125 | |
| }, | |
| { | |
| "epoch": 3.7475728155339807, | |
| "grad_norm": 0.5071633518038373, | |
| "kl": 0.11865234375, | |
| "learning_rate": 1.1737781654150955e-06, | |
| "loss": 0.162, | |
| "step": 386, | |
| "step_loss": 0.12451171875 | |
| }, | |
| { | |
| "epoch": 3.7572815533980584, | |
| "grad_norm": 0.48157726901389747, | |
| "kl": 0.2353515625, | |
| "learning_rate": 1.1639175194611693e-06, | |
| "loss": 0.1547, | |
| "step": 387, | |
| "step_loss": 0.1328125 | |
| }, | |
| { | |
| "epoch": 3.766990291262136, | |
| "grad_norm": 0.541213622511581, | |
| "kl": 0.251953125, | |
| "learning_rate": 1.15411705788683e-06, | |
| "loss": 0.1655, | |
| "step": 388, | |
| "step_loss": 0.107421875 | |
| }, | |
| { | |
| "epoch": 3.7766990291262137, | |
| "grad_norm": 0.5615405005474776, | |
| "kl": 0.212890625, | |
| "learning_rate": 1.1443771525735577e-06, | |
| "loss": 0.1576, | |
| "step": 389, | |
| "step_loss": 0.1767578125 | |
| }, | |
| { | |
| "epoch": 3.7864077669902914, | |
| "grad_norm": 0.4950987080693749, | |
| "kl": 0.169921875, | |
| "learning_rate": 1.1346981731050051e-06, | |
| "loss": 0.1565, | |
| "step": 390, | |
| "step_loss": 0.2578125 | |
| }, | |
| { | |
| "epoch": 3.796116504854369, | |
| "grad_norm": 0.49370094007294774, | |
| "kl": 0.375, | |
| "learning_rate": 1.1250804867529794e-06, | |
| "loss": 0.17, | |
| "step": 391, | |
| "step_loss": 0.0888671875 | |
| }, | |
| { | |
| "epoch": 3.8058252427184467, | |
| "grad_norm": 0.5639925863204843, | |
| "kl": 0.251953125, | |
| "learning_rate": 1.1155244584634953e-06, | |
| "loss": 0.1395, | |
| "step": 392, | |
| "step_loss": 0.08984375 | |
| }, | |
| { | |
| "epoch": 3.8155339805825244, | |
| "grad_norm": 0.5187019670608374, | |
| "kl": 0.173828125, | |
| "learning_rate": 1.1060304508429407e-06, | |
| "loss": 0.1556, | |
| "step": 393, | |
| "step_loss": 0.154296875 | |
| }, | |
| { | |
| "epoch": 3.825242718446602, | |
| "grad_norm": 0.4683311180263721, | |
| "kl": 0.23046875, | |
| "learning_rate": 1.0965988241443043e-06, | |
| "loss": 0.1584, | |
| "step": 394, | |
| "step_loss": 0.1357421875 | |
| }, | |
| { | |
| "epoch": 3.8349514563106797, | |
| "grad_norm": 0.46054186247992845, | |
| "kl": 0.2412109375, | |
| "learning_rate": 1.0872299362535175e-06, | |
| "loss": 0.1232, | |
| "step": 395, | |
| "step_loss": 0.0966796875 | |
| }, | |
| { | |
| "epoch": 3.8446601941747574, | |
| "grad_norm": 0.514740300530401, | |
| "kl": 0.404296875, | |
| "learning_rate": 1.0779241426758628e-06, | |
| "loss": 0.1634, | |
| "step": 396, | |
| "step_loss": 0.09912109375 | |
| }, | |
| { | |
| "epoch": 3.854368932038835, | |
| "grad_norm": 0.5249887226800867, | |
| "kl": 0.2421875, | |
| "learning_rate": 1.068681796522496e-06, | |
| "loss": 0.1515, | |
| "step": 397, | |
| "step_loss": 0.078125 | |
| }, | |
| { | |
| "epoch": 3.8640776699029127, | |
| "grad_norm": 0.4882053926671313, | |
| "kl": 0.2734375, | |
| "learning_rate": 1.0595032484970354e-06, | |
| "loss": 0.1623, | |
| "step": 398, | |
| "step_loss": 0.06787109375 | |
| }, | |
| { | |
| "epoch": 3.8737864077669903, | |
| "grad_norm": 0.5046889857220114, | |
| "kl": 0.1982421875, | |
| "learning_rate": 1.0503888468822648e-06, | |
| "loss": 0.1485, | |
| "step": 399, | |
| "step_loss": 0.2275390625 | |
| }, | |
| { | |
| "epoch": 3.883495145631068, | |
| "grad_norm": 0.5742122623829672, | |
| "kl": 0.291015625, | |
| "learning_rate": 1.0413389375269098e-06, | |
| "loss": 0.1645, | |
| "step": 400, | |
| "step_loss": 0.109375 | |
| }, | |
| { | |
| "epoch": 3.8932038834951457, | |
| "grad_norm": 0.4869305965525205, | |
| "kl": 0.373046875, | |
| "learning_rate": 1.0323538638325185e-06, | |
| "loss": 0.1506, | |
| "step": 401, | |
| "step_loss": 0.06298828125 | |
| }, | |
| { | |
| "epoch": 3.9029126213592233, | |
| "grad_norm": 0.4895770160456859, | |
| "kl": 0.1748046875, | |
| "learning_rate": 1.0234339667404326e-06, | |
| "loss": 0.1535, | |
| "step": 402, | |
| "step_loss": 0.12255859375 | |
| }, | |
| { | |
| "epoch": 3.912621359223301, | |
| "grad_norm": 0.4718475674872005, | |
| "kl": 0.2138671875, | |
| "learning_rate": 1.0145795847188435e-06, | |
| "loss": 0.1563, | |
| "step": 403, | |
| "step_loss": 0.244140625 | |
| }, | |
| { | |
| "epoch": 3.9223300970873787, | |
| "grad_norm": 0.46310640287447863, | |
| "kl": 0.28125, | |
| "learning_rate": 1.0057910537499585e-06, | |
| "loss": 0.1502, | |
| "step": 404, | |
| "step_loss": 0.08642578125 | |
| }, | |
| { | |
| "epoch": 3.9320388349514563, | |
| "grad_norm": 0.5464878892001633, | |
| "kl": 0.2275390625, | |
| "learning_rate": 9.970687073172416e-07, | |
| "loss": 0.1648, | |
| "step": 405, | |
| "step_loss": 0.1259765625 | |
| }, | |
| { | |
| "epoch": 3.941747572815534, | |
| "grad_norm": 0.5101772415025155, | |
| "kl": 0.1318359375, | |
| "learning_rate": 9.884128763927692e-07, | |
| "loss": 0.1584, | |
| "step": 406, | |
| "step_loss": 0.0927734375 | |
| }, | |
| { | |
| "epoch": 3.9514563106796117, | |
| "grad_norm": 0.5262085741097766, | |
| "kl": 0.3359375, | |
| "learning_rate": 9.798238894246628e-07, | |
| "loss": 0.1498, | |
| "step": 407, | |
| "step_loss": 0.134765625 | |
| }, | |
| { | |
| "epoch": 3.9611650485436893, | |
| "grad_norm": 0.5133055914879928, | |
| "kl": 0.16015625, | |
| "learning_rate": 9.713020723246332e-07, | |
| "loss": 0.1707, | |
| "step": 408, | |
| "step_loss": 0.1513671875 | |
| }, | |
| { | |
| "epoch": 3.970873786407767, | |
| "grad_norm": 0.5499931811863684, | |
| "kl": 0.306640625, | |
| "learning_rate": 9.628477484556066e-07, | |
| "loss": 0.1547, | |
| "step": 409, | |
| "step_loss": 0.11474609375 | |
| }, | |
| { | |
| "epoch": 3.9805825242718447, | |
| "grad_norm": 0.5051010516067944, | |
| "kl": 0.162109375, | |
| "learning_rate": 9.54461238619462e-07, | |
| "loss": 0.1586, | |
| "step": 410, | |
| "step_loss": 0.109375 | |
| }, | |
| { | |
| "epoch": 3.9902912621359223, | |
| "grad_norm": 0.49187747334266807, | |
| "kl": 0.2333984375, | |
| "learning_rate": 9.461428610448503e-07, | |
| "loss": 0.1461, | |
| "step": 411, | |
| "step_loss": 0.1416015625 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "grad_norm": 0.44028016997121006, | |
| "kl": 0.2734375, | |
| "learning_rate": 9.378929313751267e-07, | |
| "loss": 0.1499, | |
| "step": 412, | |
| "step_loss": 0.08056640625 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_test_transformed.json_loss": NaN, | |
| "eval_test_transformed.json_runtime": 8.4799, | |
| "eval_test_transformed.json_samples_per_second": 58.963, | |
| "eval_test_transformed.json_steps_per_second": 1.887, | |
| "step": 412 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 515, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 50.0, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 43413797257216.0, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |