| { | |
| "best_metric": 0.8874096274375916, | |
| "best_model_checkpoint": "/workspace/previous_works/MedBLIP/output/MedBLIP-0007/checkpoint-10000", | |
| "epoch": 3.0, | |
| "eval_steps": 10000, | |
| "global_step": 14319, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0031426775612822125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 6.4738, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.006285355122564425, | |
| "grad_norm": 143.78610229492188, | |
| "learning_rate": 1.3966480446927375e-07, | |
| "loss": 6.4427, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.009428032683846637, | |
| "grad_norm": 125.2808609008789, | |
| "learning_rate": 3.3519553072625703e-07, | |
| "loss": 6.1729, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.01257071024512885, | |
| "grad_norm": 55.64238357543945, | |
| "learning_rate": 5.446927374301677e-07, | |
| "loss": 5.6704, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.01571338780641106, | |
| "grad_norm": 58.784568786621094, | |
| "learning_rate": 7.402234636871509e-07, | |
| "loss": 4.6903, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.018856065367693273, | |
| "grad_norm": 101.14881896972656, | |
| "learning_rate": 9.497206703910615e-07, | |
| "loss": 4.009, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.02199874292897549, | |
| "grad_norm": 40.7690315246582, | |
| "learning_rate": 1.159217877094972e-06, | |
| "loss": 3.3713, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.0251414204902577, | |
| "grad_norm": 22.23933219909668, | |
| "learning_rate": 1.3687150837988828e-06, | |
| "loss": 2.9785, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.028284098051539912, | |
| "grad_norm": 16.25871467590332, | |
| "learning_rate": 1.5782122905027933e-06, | |
| "loss": 2.8118, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.03142677561282212, | |
| "grad_norm": 23.991317749023438, | |
| "learning_rate": 1.787709497206704e-06, | |
| "loss": 2.6758, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.034569453174104335, | |
| "grad_norm": 27.5322265625, | |
| "learning_rate": 1.9972067039106146e-06, | |
| "loss": 2.5263, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.03771213073538655, | |
| "grad_norm": 18.619874954223633, | |
| "learning_rate": 2.2067039106145253e-06, | |
| "loss": 2.4554, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.04085480829666876, | |
| "grad_norm": 17.445724487304688, | |
| "learning_rate": 2.416201117318436e-06, | |
| "loss": 2.3279, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.04399748585795098, | |
| "grad_norm": 19.981834411621094, | |
| "learning_rate": 2.6256983240223464e-06, | |
| "loss": 2.2568, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.04714016341923319, | |
| "grad_norm": 15.571565628051758, | |
| "learning_rate": 2.8351955307262576e-06, | |
| "loss": 2.1021, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.0502828409805154, | |
| "grad_norm": 13.631041526794434, | |
| "learning_rate": 3.044692737430168e-06, | |
| "loss": 1.9724, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.05342551854179761, | |
| "grad_norm": 10.806873321533203, | |
| "learning_rate": 3.2541899441340786e-06, | |
| "loss": 1.8049, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.056568196103079824, | |
| "grad_norm": 14.863541603088379, | |
| "learning_rate": 3.4636871508379893e-06, | |
| "loss": 1.7358, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.059710873664362035, | |
| "grad_norm": 12.929265022277832, | |
| "learning_rate": 3.6731843575418996e-06, | |
| "loss": 1.6953, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.06285355122564425, | |
| "grad_norm": 12.591845512390137, | |
| "learning_rate": 3.88268156424581e-06, | |
| "loss": 1.6874, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.06599622878692646, | |
| "grad_norm": 13.552529335021973, | |
| "learning_rate": 4.0921787709497215e-06, | |
| "loss": 1.6485, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.06913890634820867, | |
| "grad_norm": 9.15485668182373, | |
| "learning_rate": 4.301675977653632e-06, | |
| "loss": 1.6044, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.07228158390949088, | |
| "grad_norm": 7.887852668762207, | |
| "learning_rate": 4.511173184357542e-06, | |
| "loss": 1.4664, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.0754242614707731, | |
| "grad_norm": 12.257076263427734, | |
| "learning_rate": 4.7206703910614525e-06, | |
| "loss": 1.5338, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.0785669390320553, | |
| "grad_norm": 8.181825637817383, | |
| "learning_rate": 4.930167597765364e-06, | |
| "loss": 1.4954, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.08170961659333752, | |
| "grad_norm": 9.838497161865234, | |
| "learning_rate": 5.139664804469274e-06, | |
| "loss": 1.4217, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.08485229415461974, | |
| "grad_norm": 15.192863464355469, | |
| "learning_rate": 5.349162011173184e-06, | |
| "loss": 1.4725, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.08799497171590195, | |
| "grad_norm": 7.013632297515869, | |
| "learning_rate": 5.558659217877096e-06, | |
| "loss": 1.3989, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.09113764927718417, | |
| "grad_norm": 8.598616600036621, | |
| "learning_rate": 5.768156424581007e-06, | |
| "loss": 1.4828, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.09428032683846638, | |
| "grad_norm": 5.625289440155029, | |
| "learning_rate": 5.977653631284917e-06, | |
| "loss": 1.4513, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.09742300439974859, | |
| "grad_norm": 7.280072212219238, | |
| "learning_rate": 6.187150837988828e-06, | |
| "loss": 1.4092, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.1005656819610308, | |
| "grad_norm": 7.608973979949951, | |
| "learning_rate": 6.396648044692738e-06, | |
| "loss": 1.4124, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.10370835952231301, | |
| "grad_norm": 8.430741310119629, | |
| "learning_rate": 6.606145251396649e-06, | |
| "loss": 1.4466, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.10685103708359522, | |
| "grad_norm": 8.642643928527832, | |
| "learning_rate": 6.815642458100559e-06, | |
| "loss": 1.3558, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.10999371464487744, | |
| "grad_norm": 6.632693290710449, | |
| "learning_rate": 7.02513966480447e-06, | |
| "loss": 1.4129, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.11313639220615965, | |
| "grad_norm": 12.695171356201172, | |
| "learning_rate": 7.2346368715083805e-06, | |
| "loss": 1.3444, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.11627906976744186, | |
| "grad_norm": 5.393228054046631, | |
| "learning_rate": 7.444134078212291e-06, | |
| "loss": 1.3485, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.11942174732872407, | |
| "grad_norm": 8.173951148986816, | |
| "learning_rate": 7.653631284916202e-06, | |
| "loss": 1.4175, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.12256442489000628, | |
| "grad_norm": 14.139352798461914, | |
| "learning_rate": 7.863128491620112e-06, | |
| "loss": 1.3797, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.1257071024512885, | |
| "grad_norm": 4.9999613761901855, | |
| "learning_rate": 8.072625698324023e-06, | |
| "loss": 1.3716, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.12884978001257072, | |
| "grad_norm": 4.344189643859863, | |
| "learning_rate": 8.282122905027935e-06, | |
| "loss": 1.3011, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.13199245757385292, | |
| "grad_norm": 11.033843040466309, | |
| "learning_rate": 8.491620111731845e-06, | |
| "loss": 1.3672, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.13513513513513514, | |
| "grad_norm": 4.3888068199157715, | |
| "learning_rate": 8.701117318435755e-06, | |
| "loss": 1.2773, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.13827781269641734, | |
| "grad_norm": 4.085406303405762, | |
| "learning_rate": 8.910614525139666e-06, | |
| "loss": 1.3057, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.14142049025769957, | |
| "grad_norm": 11.730964660644531, | |
| "learning_rate": 9.120111731843576e-06, | |
| "loss": 1.3752, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.14456316781898176, | |
| "grad_norm": 4.986327648162842, | |
| "learning_rate": 9.329608938547486e-06, | |
| "loss": 1.3055, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.147705845380264, | |
| "grad_norm": 5.643145561218262, | |
| "learning_rate": 9.539106145251398e-06, | |
| "loss": 1.3608, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.1508485229415462, | |
| "grad_norm": 12.661754608154297, | |
| "learning_rate": 9.748603351955308e-06, | |
| "loss": 1.4083, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.1539912005028284, | |
| "grad_norm": 6.018293857574463, | |
| "learning_rate": 9.958100558659219e-06, | |
| "loss": 1.2323, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.1571338780641106, | |
| "grad_norm": 4.150609493255615, | |
| "learning_rate": 1.0167597765363129e-05, | |
| "loss": 1.3068, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.16027655562539284, | |
| "grad_norm": 4.229065895080566, | |
| "learning_rate": 1.037709497206704e-05, | |
| "loss": 1.3257, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.16341923318667503, | |
| "grad_norm": 9.254191398620605, | |
| "learning_rate": 1.0586592178770951e-05, | |
| "loss": 1.2674, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.16656191074795726, | |
| "grad_norm": 3.52528977394104, | |
| "learning_rate": 1.0796089385474862e-05, | |
| "loss": 1.271, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.16970458830923948, | |
| "grad_norm": 6.313427448272705, | |
| "learning_rate": 1.1005586592178772e-05, | |
| "loss": 1.2948, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.17284726587052168, | |
| "grad_norm": 3.691848039627075, | |
| "learning_rate": 1.1215083798882682e-05, | |
| "loss": 1.292, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.1759899434318039, | |
| "grad_norm": 3.9371867179870605, | |
| "learning_rate": 1.1424581005586593e-05, | |
| "loss": 1.3135, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.1791326209930861, | |
| "grad_norm": 3.455486297607422, | |
| "learning_rate": 1.1634078212290503e-05, | |
| "loss": 1.2495, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.18227529855436833, | |
| "grad_norm": 10.44939136505127, | |
| "learning_rate": 1.1843575418994415e-05, | |
| "loss": 1.356, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.18541797611565053, | |
| "grad_norm": 3.6111254692077637, | |
| "learning_rate": 1.2053072625698325e-05, | |
| "loss": 1.3196, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.18856065367693275, | |
| "grad_norm": 3.531568765640259, | |
| "learning_rate": 1.2262569832402236e-05, | |
| "loss": 1.2307, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.19170333123821495, | |
| "grad_norm": 3.490081310272217, | |
| "learning_rate": 1.2472067039106146e-05, | |
| "loss": 1.2445, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.19484600879949718, | |
| "grad_norm": 5.011629581451416, | |
| "learning_rate": 1.2681564245810056e-05, | |
| "loss": 1.3192, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.19798868636077938, | |
| "grad_norm": 7.499322414398193, | |
| "learning_rate": 1.2891061452513967e-05, | |
| "loss": 1.2827, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.2011313639220616, | |
| "grad_norm": 3.792930841445923, | |
| "learning_rate": 1.3100558659217879e-05, | |
| "loss": 1.336, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.2042740414833438, | |
| "grad_norm": 4.43565559387207, | |
| "learning_rate": 1.3310055865921789e-05, | |
| "loss": 1.3371, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.20741671904462602, | |
| "grad_norm": 4.2933173179626465, | |
| "learning_rate": 1.3519553072625699e-05, | |
| "loss": 1.3118, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.21055939660590822, | |
| "grad_norm": 3.68387770652771, | |
| "learning_rate": 1.372905027932961e-05, | |
| "loss": 1.3975, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.21370207416719045, | |
| "grad_norm": 5.9917707443237305, | |
| "learning_rate": 1.393854748603352e-05, | |
| "loss": 1.2626, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.21684475172847265, | |
| "grad_norm": 3.1892921924591064, | |
| "learning_rate": 1.414804469273743e-05, | |
| "loss": 1.2781, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.21998742928975487, | |
| "grad_norm": 3.4965710639953613, | |
| "learning_rate": 1.4357541899441342e-05, | |
| "loss": 1.2557, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.2231301068510371, | |
| "grad_norm": 7.907847881317139, | |
| "learning_rate": 1.4567039106145252e-05, | |
| "loss": 1.2341, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.2262727844123193, | |
| "grad_norm": 3.6374635696411133, | |
| "learning_rate": 1.4776536312849163e-05, | |
| "loss": 1.3225, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.22941546197360152, | |
| "grad_norm": 6.52725887298584, | |
| "learning_rate": 1.4986033519553073e-05, | |
| "loss": 1.2137, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.23255813953488372, | |
| "grad_norm": 4.659281253814697, | |
| "learning_rate": 1.5195530726256983e-05, | |
| "loss": 1.3373, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.23570081709616594, | |
| "grad_norm": 3.1662540435791016, | |
| "learning_rate": 1.5405027932960895e-05, | |
| "loss": 1.3348, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.23884349465744814, | |
| "grad_norm": 3.8003756999969482, | |
| "learning_rate": 1.5614525139664806e-05, | |
| "loss": 1.2948, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.24198617221873037, | |
| "grad_norm": 3.0090460777282715, | |
| "learning_rate": 1.5824022346368716e-05, | |
| "loss": 1.2724, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 0.24512884978001256, | |
| "grad_norm": 3.1680350303649902, | |
| "learning_rate": 1.6033519553072626e-05, | |
| "loss": 1.2402, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.2482715273412948, | |
| "grad_norm": 3.944425106048584, | |
| "learning_rate": 1.6243016759776537e-05, | |
| "loss": 1.303, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 0.251414204902577, | |
| "grad_norm": 8.234729766845703, | |
| "learning_rate": 1.6452513966480447e-05, | |
| "loss": 1.2196, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.2545568824638592, | |
| "grad_norm": 6.725916862487793, | |
| "learning_rate": 1.6662011173184357e-05, | |
| "loss": 1.3213, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 0.25769956002514144, | |
| "grad_norm": 3.2696242332458496, | |
| "learning_rate": 1.687150837988827e-05, | |
| "loss": 1.2611, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.2608422375864236, | |
| "grad_norm": 3.0667693614959717, | |
| "learning_rate": 1.708100558659218e-05, | |
| "loss": 1.22, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 0.26398491514770583, | |
| "grad_norm": 3.5260438919067383, | |
| "learning_rate": 1.729050279329609e-05, | |
| "loss": 1.3142, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.26712759270898806, | |
| "grad_norm": 3.51269793510437, | |
| "learning_rate": 1.7500000000000002e-05, | |
| "loss": 1.2031, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 0.2702702702702703, | |
| "grad_norm": 2.9083192348480225, | |
| "learning_rate": 1.7709497206703912e-05, | |
| "loss": 1.1711, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.27341294783155246, | |
| "grad_norm": 2.915454626083374, | |
| "learning_rate": 1.7918994413407822e-05, | |
| "loss": 1.2452, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 0.2765556253928347, | |
| "grad_norm": 4.736166954040527, | |
| "learning_rate": 1.8128491620111733e-05, | |
| "loss": 1.2251, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.2796983029541169, | |
| "grad_norm": 2.881985664367676, | |
| "learning_rate": 1.8337988826815643e-05, | |
| "loss": 1.2541, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 0.28284098051539913, | |
| "grad_norm": 4.29525899887085, | |
| "learning_rate": 1.8547486033519553e-05, | |
| "loss": 1.2336, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.28598365807668136, | |
| "grad_norm": 3.046929359436035, | |
| "learning_rate": 1.8756983240223464e-05, | |
| "loss": 1.2576, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 0.2891263356379635, | |
| "grad_norm": 4.023289680480957, | |
| "learning_rate": 1.8966480446927374e-05, | |
| "loss": 1.2371, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.29226901319924575, | |
| "grad_norm": 6.221644878387451, | |
| "learning_rate": 1.9175977653631284e-05, | |
| "loss": 1.1985, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 0.295411690760528, | |
| "grad_norm": 6.021862030029297, | |
| "learning_rate": 1.9385474860335195e-05, | |
| "loss": 1.2741, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.2985543683218102, | |
| "grad_norm": 6.475122451782227, | |
| "learning_rate": 1.959497206703911e-05, | |
| "loss": 1.2405, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 0.3016970458830924, | |
| "grad_norm": 6.075967788696289, | |
| "learning_rate": 1.980446927374302e-05, | |
| "loss": 1.1969, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.3048397234443746, | |
| "grad_norm": 2.591923713684082, | |
| "learning_rate": 1.9999999702856602e-05, | |
| "loss": 1.1979, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 0.3079824010056568, | |
| "grad_norm": 6.351642608642578, | |
| "learning_rate": 1.9999923931385287e-05, | |
| "loss": 1.2887, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.31112507856693905, | |
| "grad_norm": 2.5136075019836426, | |
| "learning_rate": 1.9999714446548786e-05, | |
| "loss": 1.2974, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 0.3142677561282212, | |
| "grad_norm": 3.683798313140869, | |
| "learning_rate": 1.9999371251148214e-05, | |
| "loss": 1.2073, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.31741043368950345, | |
| "grad_norm": 6.616783142089844, | |
| "learning_rate": 1.9998894349772585e-05, | |
| "loss": 1.224, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 0.32055311125078567, | |
| "grad_norm": 3.386265277862549, | |
| "learning_rate": 1.9998283748798757e-05, | |
| "loss": 1.2237, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.3236957888120679, | |
| "grad_norm": 3.2038071155548096, | |
| "learning_rate": 1.9997539456391343e-05, | |
| "loss": 1.2805, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 0.32683846637335007, | |
| "grad_norm": 5.434456825256348, | |
| "learning_rate": 1.9996661482502602e-05, | |
| "loss": 1.295, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.3299811439346323, | |
| "grad_norm": 3.0071728229522705, | |
| "learning_rate": 1.999564983887231e-05, | |
| "loss": 1.2207, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 0.3331238214959145, | |
| "grad_norm": 4.010280609130859, | |
| "learning_rate": 1.9994504539027594e-05, | |
| "loss": 1.2005, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.33626649905719674, | |
| "grad_norm": 4.2148237228393555, | |
| "learning_rate": 1.9993225598282768e-05, | |
| "loss": 1.1608, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 0.33940917661847897, | |
| "grad_norm": 5.541611194610596, | |
| "learning_rate": 1.9991813033739097e-05, | |
| "loss": 1.1649, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.34255185417976114, | |
| "grad_norm": 5.37580680847168, | |
| "learning_rate": 1.9990266864284607e-05, | |
| "loss": 1.2087, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 0.34569453174104336, | |
| "grad_norm": 3.72402286529541, | |
| "learning_rate": 1.998858711059381e-05, | |
| "loss": 1.2606, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.3488372093023256, | |
| "grad_norm": 3.596006393432617, | |
| "learning_rate": 1.9986773795127425e-05, | |
| "loss": 1.2509, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 0.3519798868636078, | |
| "grad_norm": 2.3336808681488037, | |
| "learning_rate": 1.9984826942132087e-05, | |
| "loss": 1.2389, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.35512256442489, | |
| "grad_norm": 2.766104221343994, | |
| "learning_rate": 1.9982746577640024e-05, | |
| "loss": 1.2334, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 0.3582652419861722, | |
| "grad_norm": 2.6721198558807373, | |
| "learning_rate": 1.9980532729468698e-05, | |
| "loss": 1.2069, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.36140791954745444, | |
| "grad_norm": 2.4807119369506836, | |
| "learning_rate": 1.9978185427220447e-05, | |
| "loss": 1.264, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 0.36455059710873666, | |
| "grad_norm": 2.6071507930755615, | |
| "learning_rate": 1.9975704702282074e-05, | |
| "loss": 1.2745, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.36769327467001883, | |
| "grad_norm": 2.5577876567840576, | |
| "learning_rate": 1.9973090587824443e-05, | |
| "loss": 1.2244, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 0.37083595223130106, | |
| "grad_norm": 5.360003471374512, | |
| "learning_rate": 1.997034311880202e-05, | |
| "loss": 1.1723, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.3739786297925833, | |
| "grad_norm": 2.606682777404785, | |
| "learning_rate": 1.996746233195242e-05, | |
| "loss": 1.2349, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 0.3771213073538655, | |
| "grad_norm": 2.898453712463379, | |
| "learning_rate": 1.9964448265795903e-05, | |
| "loss": 1.2428, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.3802639849151477, | |
| "grad_norm": 2.8664214611053467, | |
| "learning_rate": 1.9961300960634876e-05, | |
| "loss": 1.2338, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 0.3834066624764299, | |
| "grad_norm": 3.170452117919922, | |
| "learning_rate": 1.9958020458553326e-05, | |
| "loss": 1.1729, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.38654934003771213, | |
| "grad_norm": 2.6928646564483643, | |
| "learning_rate": 1.995460680341629e-05, | |
| "loss": 1.2791, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 0.38969201759899436, | |
| "grad_norm": 5.1588287353515625, | |
| "learning_rate": 1.9951060040869245e-05, | |
| "loss": 1.1712, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.3928346951602766, | |
| "grad_norm": 4.63333797454834, | |
| "learning_rate": 1.994738021833751e-05, | |
| "loss": 1.285, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 0.39597737272155875, | |
| "grad_norm": 2.860305070877075, | |
| "learning_rate": 1.9943567385025605e-05, | |
| "loss": 1.1853, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.399120050282841, | |
| "grad_norm": 3.4676976203918457, | |
| "learning_rate": 1.9939621591916587e-05, | |
| "loss": 1.2164, | |
| "step": 1905 | |
| }, | |
| { | |
| "epoch": 0.4022627278441232, | |
| "grad_norm": 4.743997573852539, | |
| "learning_rate": 1.9935542891771392e-05, | |
| "loss": 1.2182, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.40540540540540543, | |
| "grad_norm": 4.906563758850098, | |
| "learning_rate": 1.9931331339128107e-05, | |
| "loss": 1.1844, | |
| "step": 1935 | |
| }, | |
| { | |
| "epoch": 0.4085480829666876, | |
| "grad_norm": 2.863513946533203, | |
| "learning_rate": 1.9926986990301244e-05, | |
| "loss": 1.2124, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.4116907605279698, | |
| "grad_norm": 2.1002469062805176, | |
| "learning_rate": 1.9922509903381e-05, | |
| "loss": 1.1589, | |
| "step": 1965 | |
| }, | |
| { | |
| "epoch": 0.41483343808925205, | |
| "grad_norm": 5.061679840087891, | |
| "learning_rate": 1.991790013823246e-05, | |
| "loss": 1.2208, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.4179761156505343, | |
| "grad_norm": 5.046812534332275, | |
| "learning_rate": 1.9913157756494822e-05, | |
| "loss": 1.2122, | |
| "step": 1995 | |
| }, | |
| { | |
| "epoch": 0.42111879321181644, | |
| "grad_norm": 2.990056037902832, | |
| "learning_rate": 1.9908282821580544e-05, | |
| "loss": 1.2296, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.42426147077309867, | |
| "grad_norm": 2.477032423019409, | |
| "learning_rate": 1.9903275398674523e-05, | |
| "loss": 1.2259, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 0.4274041483343809, | |
| "grad_norm": 2.7786192893981934, | |
| "learning_rate": 1.9898135554733198e-05, | |
| "loss": 1.2062, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.4305468258956631, | |
| "grad_norm": 2.325221061706543, | |
| "learning_rate": 1.989286335848368e-05, | |
| "loss": 1.2676, | |
| "step": 2055 | |
| }, | |
| { | |
| "epoch": 0.4336895034569453, | |
| "grad_norm": 2.26428484916687, | |
| "learning_rate": 1.988745888042281e-05, | |
| "loss": 1.1832, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.4368321810182275, | |
| "grad_norm": 2.361360788345337, | |
| "learning_rate": 1.988192219281624e-05, | |
| "loss": 1.1814, | |
| "step": 2085 | |
| }, | |
| { | |
| "epoch": 0.43997485857950974, | |
| "grad_norm": 2.2548787593841553, | |
| "learning_rate": 1.987625336969744e-05, | |
| "loss": 1.1371, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.44311753614079197, | |
| "grad_norm": 2.9371745586395264, | |
| "learning_rate": 1.9870452486866733e-05, | |
| "loss": 1.1185, | |
| "step": 2115 | |
| }, | |
| { | |
| "epoch": 0.4462602137020742, | |
| "grad_norm": 2.250372886657715, | |
| "learning_rate": 1.9864519621890263e-05, | |
| "loss": 1.1919, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.44940289126335636, | |
| "grad_norm": 2.822431802749634, | |
| "learning_rate": 1.9858454854098968e-05, | |
| "loss": 1.1479, | |
| "step": 2145 | |
| }, | |
| { | |
| "epoch": 0.4525455688246386, | |
| "grad_norm": 5.055064678192139, | |
| "learning_rate": 1.985225826458751e-05, | |
| "loss": 1.2237, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.4556882463859208, | |
| "grad_norm": 2.3341803550720215, | |
| "learning_rate": 1.9845929936213216e-05, | |
| "loss": 1.1988, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 0.45883092394720304, | |
| "grad_norm": 2.5791430473327637, | |
| "learning_rate": 1.9839469953594927e-05, | |
| "loss": 1.19, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.4619736015084852, | |
| "grad_norm": 3.167426824569702, | |
| "learning_rate": 1.9832878403111908e-05, | |
| "loss": 1.1852, | |
| "step": 2205 | |
| }, | |
| { | |
| "epoch": 0.46511627906976744, | |
| "grad_norm": 4.721558570861816, | |
| "learning_rate": 1.9826155372902667e-05, | |
| "loss": 1.198, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.46825895663104966, | |
| "grad_norm": 2.5482370853424072, | |
| "learning_rate": 1.9819300952863783e-05, | |
| "loss": 1.1782, | |
| "step": 2235 | |
| }, | |
| { | |
| "epoch": 0.4714016341923319, | |
| "grad_norm": 3.1147351264953613, | |
| "learning_rate": 1.9812315234648714e-05, | |
| "loss": 1.2138, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.47454431175361406, | |
| "grad_norm": 4.730167865753174, | |
| "learning_rate": 1.980519831166656e-05, | |
| "loss": 1.1622, | |
| "step": 2265 | |
| }, | |
| { | |
| "epoch": 0.4776869893148963, | |
| "grad_norm": 1.9187180995941162, | |
| "learning_rate": 1.979795027908082e-05, | |
| "loss": 1.1624, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.4808296668761785, | |
| "grad_norm": 2.371293067932129, | |
| "learning_rate": 1.9790571233808108e-05, | |
| "loss": 1.1823, | |
| "step": 2295 | |
| }, | |
| { | |
| "epoch": 0.48397234443746073, | |
| "grad_norm": 2.103717803955078, | |
| "learning_rate": 1.978306127451687e-05, | |
| "loss": 1.2298, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.4871150219987429, | |
| "grad_norm": 3.5898678302764893, | |
| "learning_rate": 1.9775420501626073e-05, | |
| "loss": 1.2612, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 0.49025769956002513, | |
| "grad_norm": 2.680027723312378, | |
| "learning_rate": 1.976764901730383e-05, | |
| "loss": 1.2161, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.49340037712130735, | |
| "grad_norm": 2.514291286468506, | |
| "learning_rate": 1.9759746925466073e-05, | |
| "loss": 1.1772, | |
| "step": 2355 | |
| }, | |
| { | |
| "epoch": 0.4965430546825896, | |
| "grad_norm": 2.453972816467285, | |
| "learning_rate": 1.9751714331775125e-05, | |
| "loss": 1.1982, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.4996857322438718, | |
| "grad_norm": 6.017053604125977, | |
| "learning_rate": 1.9743551343638324e-05, | |
| "loss": 1.181, | |
| "step": 2385 | |
| }, | |
| { | |
| "epoch": 0.502828409805154, | |
| "grad_norm": 2.46274733543396, | |
| "learning_rate": 1.973525807020656e-05, | |
| "loss": 1.1537, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.5059710873664363, | |
| "grad_norm": 2.6295454502105713, | |
| "learning_rate": 1.972683462237282e-05, | |
| "loss": 1.1961, | |
| "step": 2415 | |
| }, | |
| { | |
| "epoch": 0.5091137649277184, | |
| "grad_norm": 5.519942760467529, | |
| "learning_rate": 1.9718281112770723e-05, | |
| "loss": 1.1998, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.5122564424890006, | |
| "grad_norm": 2.5744876861572266, | |
| "learning_rate": 1.9709597655772993e-05, | |
| "loss": 1.1511, | |
| "step": 2445 | |
| }, | |
| { | |
| "epoch": 0.5153991200502829, | |
| "grad_norm": 2.621332883834839, | |
| "learning_rate": 1.970078436748994e-05, | |
| "loss": 1.1398, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.518541797611565, | |
| "grad_norm": 5.679701328277588, | |
| "learning_rate": 1.9691841365767904e-05, | |
| "loss": 1.1899, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 0.5216844751728472, | |
| "grad_norm": 2.662759780883789, | |
| "learning_rate": 1.9682768770187675e-05, | |
| "loss": 1.1442, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.5248271527341295, | |
| "grad_norm": 2.7816829681396484, | |
| "learning_rate": 1.967356670206291e-05, | |
| "loss": 1.2164, | |
| "step": 2505 | |
| }, | |
| { | |
| "epoch": 0.5279698302954117, | |
| "grad_norm": 5.280804634094238, | |
| "learning_rate": 1.9664235284438495e-05, | |
| "loss": 1.186, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.531112507856694, | |
| "grad_norm": 4.3773112297058105, | |
| "learning_rate": 1.96547746420889e-05, | |
| "loss": 1.2107, | |
| "step": 2535 | |
| }, | |
| { | |
| "epoch": 0.5342551854179761, | |
| "grad_norm": 2.773582696914673, | |
| "learning_rate": 1.9645184901516525e-05, | |
| "loss": 1.2216, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.5373978629792583, | |
| "grad_norm": 2.643582820892334, | |
| "learning_rate": 1.9635466190949987e-05, | |
| "loss": 1.1888, | |
| "step": 2565 | |
| }, | |
| { | |
| "epoch": 0.5405405405405406, | |
| "grad_norm": 2.277792453765869, | |
| "learning_rate": 1.962561864034243e-05, | |
| "loss": 1.225, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.5436832181018227, | |
| "grad_norm": 2.580367088317871, | |
| "learning_rate": 1.9615642381369765e-05, | |
| "loss": 1.2064, | |
| "step": 2595 | |
| }, | |
| { | |
| "epoch": 0.5468258956631049, | |
| "grad_norm": 4.304893493652344, | |
| "learning_rate": 1.9605537547428922e-05, | |
| "loss": 1.1652, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.5499685732243872, | |
| "grad_norm": 2.5215561389923096, | |
| "learning_rate": 1.959530427363607e-05, | |
| "loss": 1.1909, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 0.5531112507856694, | |
| "grad_norm": 3.4145307540893555, | |
| "learning_rate": 1.958494269682479e-05, | |
| "loss": 1.1211, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.5562539283469516, | |
| "grad_norm": 4.1698079109191895, | |
| "learning_rate": 1.9574452955544274e-05, | |
| "loss": 1.1235, | |
| "step": 2655 | |
| }, | |
| { | |
| "epoch": 0.5593966059082338, | |
| "grad_norm": 4.188419818878174, | |
| "learning_rate": 1.9563835190057442e-05, | |
| "loss": 1.1925, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.562539283469516, | |
| "grad_norm": 4.869725704193115, | |
| "learning_rate": 1.9553089542339093e-05, | |
| "loss": 1.172, | |
| "step": 2685 | |
| }, | |
| { | |
| "epoch": 0.5656819610307983, | |
| "grad_norm": 4.675424575805664, | |
| "learning_rate": 1.9542216156073994e-05, | |
| "loss": 1.218, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.5688246385920804, | |
| "grad_norm": 2.655273914337158, | |
| "learning_rate": 1.9531215176654953e-05, | |
| "loss": 1.128, | |
| "step": 2715 | |
| }, | |
| { | |
| "epoch": 0.5719673161533627, | |
| "grad_norm": 3.345499277114868, | |
| "learning_rate": 1.9520086751180894e-05, | |
| "loss": 1.2236, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 0.5751099937146449, | |
| "grad_norm": 2.5058796405792236, | |
| "learning_rate": 1.9508831028454862e-05, | |
| "loss": 1.1895, | |
| "step": 2745 | |
| }, | |
| { | |
| "epoch": 0.578252671275927, | |
| "grad_norm": 2.2655258178710938, | |
| "learning_rate": 1.9497448158982062e-05, | |
| "loss": 1.1486, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.5813953488372093, | |
| "grad_norm": 4.1772894859313965, | |
| "learning_rate": 1.948593829496783e-05, | |
| "loss": 1.1616, | |
| "step": 2775 | |
| }, | |
| { | |
| "epoch": 0.5845380263984915, | |
| "grad_norm": 2.290619373321533, | |
| "learning_rate": 1.9474301590315605e-05, | |
| "loss": 1.1508, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 0.5876807039597737, | |
| "grad_norm": 2.4885716438293457, | |
| "learning_rate": 1.9462538200624855e-05, | |
| "loss": 1.1727, | |
| "step": 2805 | |
| }, | |
| { | |
| "epoch": 0.590823381521056, | |
| "grad_norm": 2.564606189727783, | |
| "learning_rate": 1.9450648283189035e-05, | |
| "loss": 1.2178, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.5939660590823381, | |
| "grad_norm": 2.756286859512329, | |
| "learning_rate": 1.9438631996993424e-05, | |
| "loss": 1.2143, | |
| "step": 2835 | |
| }, | |
| { | |
| "epoch": 0.5971087366436204, | |
| "grad_norm": 4.876669883728027, | |
| "learning_rate": 1.9426489502713062e-05, | |
| "loss": 1.1908, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.6002514142049026, | |
| "grad_norm": 1.9889987707138062, | |
| "learning_rate": 1.9414220962710556e-05, | |
| "loss": 1.1404, | |
| "step": 2865 | |
| }, | |
| { | |
| "epoch": 0.6033940917661847, | |
| "grad_norm": 2.6921072006225586, | |
| "learning_rate": 1.9401826541033935e-05, | |
| "loss": 1.2159, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.606536769327467, | |
| "grad_norm": 1.9745949506759644, | |
| "learning_rate": 1.9389306403414438e-05, | |
| "loss": 1.1828, | |
| "step": 2895 | |
| }, | |
| { | |
| "epoch": 0.6096794468887492, | |
| "grad_norm": 2.012610912322998, | |
| "learning_rate": 1.9376660717264316e-05, | |
| "loss": 1.1413, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 0.6128221244500315, | |
| "grad_norm": 3.3427295684814453, | |
| "learning_rate": 1.9363889651674575e-05, | |
| "loss": 1.1888, | |
| "step": 2925 | |
| }, | |
| { | |
| "epoch": 0.6159648020113137, | |
| "grad_norm": 3.9471700191497803, | |
| "learning_rate": 1.935099337741274e-05, | |
| "loss": 1.1882, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.6191074795725958, | |
| "grad_norm": 2.1639480590820312, | |
| "learning_rate": 1.9337972066920533e-05, | |
| "loss": 1.1518, | |
| "step": 2955 | |
| }, | |
| { | |
| "epoch": 0.6222501571338781, | |
| "grad_norm": 2.18038272857666, | |
| "learning_rate": 1.9324825894311613e-05, | |
| "loss": 1.1424, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 0.6253928346951603, | |
| "grad_norm": 2.4745776653289795, | |
| "learning_rate": 1.9311555035369206e-05, | |
| "loss": 1.2065, | |
| "step": 2985 | |
| }, | |
| { | |
| "epoch": 0.6285355122564424, | |
| "grad_norm": 2.4123799800872803, | |
| "learning_rate": 1.9298159667543795e-05, | |
| "loss": 1.1677, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.6316781898177247, | |
| "grad_norm": 2.8787992000579834, | |
| "learning_rate": 1.92846399699507e-05, | |
| "loss": 1.2035, | |
| "step": 3015 | |
| }, | |
| { | |
| "epoch": 0.6348208673790069, | |
| "grad_norm": 2.3905246257781982, | |
| "learning_rate": 1.927099612336773e-05, | |
| "loss": 1.1933, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 0.6379635449402892, | |
| "grad_norm": 3.8858561515808105, | |
| "learning_rate": 1.9257228310232738e-05, | |
| "loss": 1.1913, | |
| "step": 3045 | |
| }, | |
| { | |
| "epoch": 0.6411062225015713, | |
| "grad_norm": 3.567950963973999, | |
| "learning_rate": 1.924333671464118e-05, | |
| "loss": 1.2417, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.6442489000628535, | |
| "grad_norm": 1.8170944452285767, | |
| "learning_rate": 1.9229321522343677e-05, | |
| "loss": 1.1476, | |
| "step": 3075 | |
| }, | |
| { | |
| "epoch": 0.6473915776241358, | |
| "grad_norm": 3.044593095779419, | |
| "learning_rate": 1.9215182920743502e-05, | |
| "loss": 1.1939, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 0.650534255185418, | |
| "grad_norm": 3.13097882270813, | |
| "learning_rate": 1.9200921098894094e-05, | |
| "loss": 1.2007, | |
| "step": 3105 | |
| }, | |
| { | |
| "epoch": 0.6536769327467001, | |
| "grad_norm": 2.1468851566314697, | |
| "learning_rate": 1.918653624749652e-05, | |
| "loss": 1.1587, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.6568196103079824, | |
| "grad_norm": 2.9996323585510254, | |
| "learning_rate": 1.9172028558896935e-05, | |
| "loss": 1.1762, | |
| "step": 3135 | |
| }, | |
| { | |
| "epoch": 0.6599622878692646, | |
| "grad_norm": 2.5897107124328613, | |
| "learning_rate": 1.9157398227084002e-05, | |
| "loss": 1.1334, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.6631049654305469, | |
| "grad_norm": 3.80074143409729, | |
| "learning_rate": 1.91426454476863e-05, | |
| "loss": 1.2642, | |
| "step": 3165 | |
| }, | |
| { | |
| "epoch": 0.666247642991829, | |
| "grad_norm": 1.9756152629852295, | |
| "learning_rate": 1.9127770417969703e-05, | |
| "loss": 1.1418, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 0.6693903205531112, | |
| "grad_norm": 1.7655247449874878, | |
| "learning_rate": 1.9112773336834754e-05, | |
| "loss": 1.1605, | |
| "step": 3195 | |
| }, | |
| { | |
| "epoch": 0.6725329981143935, | |
| "grad_norm": 2.2654428482055664, | |
| "learning_rate": 1.9097654404814e-05, | |
| "loss": 1.2366, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 0.6756756756756757, | |
| "grad_norm": 2.2166404724121094, | |
| "learning_rate": 1.9082413824069302e-05, | |
| "loss": 1.1469, | |
| "step": 3225 | |
| }, | |
| { | |
| "epoch": 0.6788183532369579, | |
| "grad_norm": 2.409104824066162, | |
| "learning_rate": 1.9067051798389148e-05, | |
| "loss": 1.1303, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 0.6819610307982401, | |
| "grad_norm": 2.2395966053009033, | |
| "learning_rate": 1.905156853318591e-05, | |
| "loss": 1.1507, | |
| "step": 3255 | |
| }, | |
| { | |
| "epoch": 0.6851037083595223, | |
| "grad_norm": 2.369009017944336, | |
| "learning_rate": 1.903596423549312e-05, | |
| "loss": 1.1468, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 0.6882463859208046, | |
| "grad_norm": 4.745908260345459, | |
| "learning_rate": 1.9020239113962677e-05, | |
| "loss": 1.1675, | |
| "step": 3285 | |
| }, | |
| { | |
| "epoch": 0.6913890634820867, | |
| "grad_norm": 2.9272751808166504, | |
| "learning_rate": 1.900439337886207e-05, | |
| "loss": 1.1604, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.6945317410433689, | |
| "grad_norm": 1.923449993133545, | |
| "learning_rate": 1.898842724207157e-05, | |
| "loss": 1.136, | |
| "step": 3315 | |
| }, | |
| { | |
| "epoch": 0.6976744186046512, | |
| "grad_norm": 2.657238006591797, | |
| "learning_rate": 1.897234091708139e-05, | |
| "loss": 1.212, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 0.7008170961659334, | |
| "grad_norm": 2.484586715698242, | |
| "learning_rate": 1.8956134618988828e-05, | |
| "loss": 1.1938, | |
| "step": 3345 | |
| }, | |
| { | |
| "epoch": 0.7039597737272156, | |
| "grad_norm": 2.127962589263916, | |
| "learning_rate": 1.89398085644954e-05, | |
| "loss": 1.138, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 0.7071024512884978, | |
| "grad_norm": 4.683898448944092, | |
| "learning_rate": 1.8923362971903932e-05, | |
| "loss": 1.2393, | |
| "step": 3375 | |
| }, | |
| { | |
| "epoch": 0.71024512884978, | |
| "grad_norm": 2.090588092803955, | |
| "learning_rate": 1.8906798061115654e-05, | |
| "loss": 1.1721, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 0.7133878064110623, | |
| "grad_norm": 2.328129768371582, | |
| "learning_rate": 1.8890114053627244e-05, | |
| "loss": 1.1351, | |
| "step": 3405 | |
| }, | |
| { | |
| "epoch": 0.7165304839723444, | |
| "grad_norm": 2.18684458732605, | |
| "learning_rate": 1.8873311172527883e-05, | |
| "loss": 1.1793, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 0.7196731615336267, | |
| "grad_norm": 3.865600109100342, | |
| "learning_rate": 1.885638964249625e-05, | |
| "loss": 1.1605, | |
| "step": 3435 | |
| }, | |
| { | |
| "epoch": 0.7228158390949089, | |
| "grad_norm": 1.9220081567764282, | |
| "learning_rate": 1.8839349689797542e-05, | |
| "loss": 1.1056, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.725958516656191, | |
| "grad_norm": 2.822507858276367, | |
| "learning_rate": 1.8822191542280433e-05, | |
| "loss": 1.2242, | |
| "step": 3465 | |
| }, | |
| { | |
| "epoch": 0.7291011942174733, | |
| "grad_norm": 2.1855289936065674, | |
| "learning_rate": 1.880491542937404e-05, | |
| "loss": 1.1582, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 0.7322438717787555, | |
| "grad_norm": 2.4138951301574707, | |
| "learning_rate": 1.8787521582084833e-05, | |
| "loss": 1.1889, | |
| "step": 3495 | |
| }, | |
| { | |
| "epoch": 0.7353865493400377, | |
| "grad_norm": 4.04121732711792, | |
| "learning_rate": 1.8770010232993562e-05, | |
| "loss": 1.1682, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 0.73852922690132, | |
| "grad_norm": 2.7555859088897705, | |
| "learning_rate": 1.875238161625216e-05, | |
| "loss": 1.1242, | |
| "step": 3525 | |
| }, | |
| { | |
| "epoch": 0.7416719044626021, | |
| "grad_norm": 2.1336309909820557, | |
| "learning_rate": 1.873463596758058e-05, | |
| "loss": 1.1823, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 0.7448145820238844, | |
| "grad_norm": 2.1928298473358154, | |
| "learning_rate": 1.871677352426367e-05, | |
| "loss": 1.1785, | |
| "step": 3555 | |
| }, | |
| { | |
| "epoch": 0.7479572595851666, | |
| "grad_norm": 1.8896750211715698, | |
| "learning_rate": 1.8698794525147988e-05, | |
| "loss": 1.1871, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 0.7510999371464487, | |
| "grad_norm": 2.354185104370117, | |
| "learning_rate": 1.8680699210638608e-05, | |
| "loss": 1.072, | |
| "step": 3585 | |
| }, | |
| { | |
| "epoch": 0.754242614707731, | |
| "grad_norm": 2.3728153705596924, | |
| "learning_rate": 1.8662487822695906e-05, | |
| "loss": 1.1607, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.7573852922690132, | |
| "grad_norm": 1.997661828994751, | |
| "learning_rate": 1.8644160604832344e-05, | |
| "loss": 1.1824, | |
| "step": 3615 | |
| }, | |
| { | |
| "epoch": 0.7605279698302954, | |
| "grad_norm": 2.4399378299713135, | |
| "learning_rate": 1.8625717802109173e-05, | |
| "loss": 1.1794, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 0.7636706473915776, | |
| "grad_norm": 2.1330130100250244, | |
| "learning_rate": 1.860715966113319e-05, | |
| "loss": 1.2458, | |
| "step": 3645 | |
| }, | |
| { | |
| "epoch": 0.7668133249528598, | |
| "grad_norm": 2.0650925636291504, | |
| "learning_rate": 1.8588486430053442e-05, | |
| "loss": 1.1785, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 0.7699560025141421, | |
| "grad_norm": 2.570436477661133, | |
| "learning_rate": 1.856969835855787e-05, | |
| "loss": 1.154, | |
| "step": 3675 | |
| }, | |
| { | |
| "epoch": 0.7730986800754243, | |
| "grad_norm": 4.064576625823975, | |
| "learning_rate": 1.855079569787002e-05, | |
| "loss": 1.1833, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 0.7762413576367064, | |
| "grad_norm": 2.199369430541992, | |
| "learning_rate": 1.8531778700745652e-05, | |
| "loss": 1.1681, | |
| "step": 3705 | |
| }, | |
| { | |
| "epoch": 0.7793840351979887, | |
| "grad_norm": 2.3457729816436768, | |
| "learning_rate": 1.851264762146937e-05, | |
| "loss": 1.1177, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 0.7825267127592709, | |
| "grad_norm": 1.790176272392273, | |
| "learning_rate": 1.849340271585121e-05, | |
| "loss": 1.1603, | |
| "step": 3735 | |
| }, | |
| { | |
| "epoch": 0.7856693903205532, | |
| "grad_norm": 1.7440158128738403, | |
| "learning_rate": 1.8474044241223247e-05, | |
| "loss": 1.1867, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.7888120678818353, | |
| "grad_norm": 2.208021640777588, | |
| "learning_rate": 1.8454572456436126e-05, | |
| "loss": 1.1381, | |
| "step": 3765 | |
| }, | |
| { | |
| "epoch": 0.7919547454431175, | |
| "grad_norm": 2.744593858718872, | |
| "learning_rate": 1.8434987621855615e-05, | |
| "loss": 1.1186, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 0.7950974230043998, | |
| "grad_norm": 2.6648318767547607, | |
| "learning_rate": 1.841528999935912e-05, | |
| "loss": 1.1345, | |
| "step": 3795 | |
| }, | |
| { | |
| "epoch": 0.798240100565682, | |
| "grad_norm": 2.1822752952575684, | |
| "learning_rate": 1.8395479852332185e-05, | |
| "loss": 1.2148, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 0.8013827781269641, | |
| "grad_norm": 2.2398853302001953, | |
| "learning_rate": 1.8375557445664962e-05, | |
| "loss": 1.1915, | |
| "step": 3825 | |
| }, | |
| { | |
| "epoch": 0.8045254556882464, | |
| "grad_norm": 3.89241886138916, | |
| "learning_rate": 1.835552304574869e-05, | |
| "loss": 1.2088, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 0.8076681332495286, | |
| "grad_norm": 2.0591068267822266, | |
| "learning_rate": 1.8335376920472098e-05, | |
| "loss": 1.2107, | |
| "step": 3855 | |
| }, | |
| { | |
| "epoch": 0.8108108108108109, | |
| "grad_norm": 2.421823501586914, | |
| "learning_rate": 1.831511933921787e-05, | |
| "loss": 1.1696, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 0.813953488372093, | |
| "grad_norm": 3.9787890911102295, | |
| "learning_rate": 1.829475057285899e-05, | |
| "loss": 1.1756, | |
| "step": 3885 | |
| }, | |
| { | |
| "epoch": 0.8170961659333752, | |
| "grad_norm": 2.6043264865875244, | |
| "learning_rate": 1.8274270893755166e-05, | |
| "loss": 1.1729, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.8202388434946575, | |
| "grad_norm": 2.131624221801758, | |
| "learning_rate": 1.825368057574916e-05, | |
| "loss": 1.1465, | |
| "step": 3915 | |
| }, | |
| { | |
| "epoch": 0.8233815210559396, | |
| "grad_norm": 2.2352840900421143, | |
| "learning_rate": 1.823297989416314e-05, | |
| "loss": 1.1126, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 0.8265241986172219, | |
| "grad_norm": 1.9127349853515625, | |
| "learning_rate": 1.821216912579499e-05, | |
| "loss": 1.1081, | |
| "step": 3945 | |
| }, | |
| { | |
| "epoch": 0.8296668761785041, | |
| "grad_norm": 1.8138941526412964, | |
| "learning_rate": 1.819124854891461e-05, | |
| "loss": 1.1369, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 0.8328095537397863, | |
| "grad_norm": 3.4601540565490723, | |
| "learning_rate": 1.817021844326021e-05, | |
| "loss": 1.1378, | |
| "step": 3975 | |
| }, | |
| { | |
| "epoch": 0.8359522313010685, | |
| "grad_norm": 2.171710968017578, | |
| "learning_rate": 1.814907909003454e-05, | |
| "loss": 1.146, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 0.8390949088623507, | |
| "grad_norm": 2.376568078994751, | |
| "learning_rate": 1.8127830771901157e-05, | |
| "loss": 1.1759, | |
| "step": 4005 | |
| }, | |
| { | |
| "epoch": 0.8422375864236329, | |
| "grad_norm": 4.197197437286377, | |
| "learning_rate": 1.810647377298064e-05, | |
| "loss": 1.1319, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 0.8453802639849152, | |
| "grad_norm": 1.8338217735290527, | |
| "learning_rate": 1.8085008378846777e-05, | |
| "loss": 1.126, | |
| "step": 4035 | |
| }, | |
| { | |
| "epoch": 0.8485229415461973, | |
| "grad_norm": 4.26235818862915, | |
| "learning_rate": 1.806343487652276e-05, | |
| "loss": 1.1889, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.8516656191074796, | |
| "grad_norm": 2.5566670894622803, | |
| "learning_rate": 1.8041753554477344e-05, | |
| "loss": 1.1377, | |
| "step": 4065 | |
| }, | |
| { | |
| "epoch": 0.8548082966687618, | |
| "grad_norm": 3.85239839553833, | |
| "learning_rate": 1.8019964702620987e-05, | |
| "loss": 1.1859, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 0.857950974230044, | |
| "grad_norm": 1.832269549369812, | |
| "learning_rate": 1.7998068612301976e-05, | |
| "loss": 1.1816, | |
| "step": 4095 | |
| }, | |
| { | |
| "epoch": 0.8610936517913262, | |
| "grad_norm": 2.670729160308838, | |
| "learning_rate": 1.7976065576302528e-05, | |
| "loss": 1.1686, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 0.8642363293526084, | |
| "grad_norm": 2.3271324634552, | |
| "learning_rate": 1.7953955888834885e-05, | |
| "loss": 1.1683, | |
| "step": 4125 | |
| }, | |
| { | |
| "epoch": 0.8673790069138906, | |
| "grad_norm": 2.2425904273986816, | |
| "learning_rate": 1.7931739845537365e-05, | |
| "loss": 1.182, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 0.8705216844751729, | |
| "grad_norm": 1.5735398530960083, | |
| "learning_rate": 1.7909417743470415e-05, | |
| "loss": 1.0888, | |
| "step": 4155 | |
| }, | |
| { | |
| "epoch": 0.873664362036455, | |
| "grad_norm": 2.4160172939300537, | |
| "learning_rate": 1.7886989881112647e-05, | |
| "loss": 1.1943, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 0.8768070395977373, | |
| "grad_norm": 2.265516519546509, | |
| "learning_rate": 1.786445655835684e-05, | |
| "loss": 1.1819, | |
| "step": 4185 | |
| }, | |
| { | |
| "epoch": 0.8799497171590195, | |
| "grad_norm": 3.3854482173919678, | |
| "learning_rate": 1.784181807650592e-05, | |
| "loss": 1.1299, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.8830923947203017, | |
| "grad_norm": 1.209488868713379, | |
| "learning_rate": 1.7819074738268956e-05, | |
| "loss": 1.0379, | |
| "step": 4215 | |
| }, | |
| { | |
| "epoch": 0.8862350722815839, | |
| "grad_norm": 2.617237091064453, | |
| "learning_rate": 1.7796226847757084e-05, | |
| "loss": 1.2145, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 0.8893777498428661, | |
| "grad_norm": 4.028698921203613, | |
| "learning_rate": 1.7773274710479473e-05, | |
| "loss": 1.117, | |
| "step": 4245 | |
| }, | |
| { | |
| "epoch": 0.8925204274041484, | |
| "grad_norm": 2.382817506790161, | |
| "learning_rate": 1.7750218633339194e-05, | |
| "loss": 1.1122, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 0.8956631049654306, | |
| "grad_norm": 2.063629150390625, | |
| "learning_rate": 1.7727058924629165e-05, | |
| "loss": 1.1553, | |
| "step": 4275 | |
| }, | |
| { | |
| "epoch": 0.8988057825267127, | |
| "grad_norm": 1.7411893606185913, | |
| "learning_rate": 1.7703795894028003e-05, | |
| "loss": 1.1544, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 0.901948460087995, | |
| "grad_norm": 1.9747669696807861, | |
| "learning_rate": 1.768042985259588e-05, | |
| "loss": 1.1875, | |
| "step": 4305 | |
| }, | |
| { | |
| "epoch": 0.9050911376492772, | |
| "grad_norm": 2.338653326034546, | |
| "learning_rate": 1.765696111277038e-05, | |
| "loss": 1.1609, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 0.9082338152105593, | |
| "grad_norm": 2.3144114017486572, | |
| "learning_rate": 1.7633389988362305e-05, | |
| "loss": 1.1843, | |
| "step": 4335 | |
| }, | |
| { | |
| "epoch": 0.9113764927718416, | |
| "grad_norm": 2.072577953338623, | |
| "learning_rate": 1.760971679455149e-05, | |
| "loss": 1.1109, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 0.9145191703331238, | |
| "grad_norm": 2.2952466011047363, | |
| "learning_rate": 1.758594184788258e-05, | |
| "loss": 1.1714, | |
| "step": 4365 | |
| }, | |
| { | |
| "epoch": 0.9176618478944061, | |
| "grad_norm": 2.254108428955078, | |
| "learning_rate": 1.7562065466260816e-05, | |
| "loss": 1.1627, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 0.9208045254556882, | |
| "grad_norm": 2.286694049835205, | |
| "learning_rate": 1.753808796894775e-05, | |
| "loss": 1.135, | |
| "step": 4395 | |
| }, | |
| { | |
| "epoch": 0.9239472030169704, | |
| "grad_norm": 3.5377025604248047, | |
| "learning_rate": 1.751400967655701e-05, | |
| "loss": 1.1172, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 0.9270898805782527, | |
| "grad_norm": 1.9915852546691895, | |
| "learning_rate": 1.7489830911049995e-05, | |
| "loss": 1.1235, | |
| "step": 4425 | |
| }, | |
| { | |
| "epoch": 0.9302325581395349, | |
| "grad_norm": 1.916908860206604, | |
| "learning_rate": 1.746555199573158e-05, | |
| "loss": 1.1604, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 0.933375235700817, | |
| "grad_norm": 2.122201442718506, | |
| "learning_rate": 1.7441173255245773e-05, | |
| "loss": 1.17, | |
| "step": 4455 | |
| }, | |
| { | |
| "epoch": 0.9365179132620993, | |
| "grad_norm": 2.4088730812072754, | |
| "learning_rate": 1.74166950155714e-05, | |
| "loss": 1.1222, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 0.9396605908233815, | |
| "grad_norm": 2.2933411598205566, | |
| "learning_rate": 1.7392117604017733e-05, | |
| "loss": 1.1322, | |
| "step": 4485 | |
| }, | |
| { | |
| "epoch": 0.9428032683846638, | |
| "grad_norm": 2.084245204925537, | |
| "learning_rate": 1.7367441349220108e-05, | |
| "loss": 1.139, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.9459459459459459, | |
| "grad_norm": 2.220381498336792, | |
| "learning_rate": 1.734266658113555e-05, | |
| "loss": 1.1533, | |
| "step": 4515 | |
| }, | |
| { | |
| "epoch": 0.9490886235072281, | |
| "grad_norm": 2.4368200302124023, | |
| "learning_rate": 1.731779363103833e-05, | |
| "loss": 1.1276, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 0.9522313010685104, | |
| "grad_norm": 3.9131979942321777, | |
| "learning_rate": 1.729282283151557e-05, | |
| "loss": 1.2134, | |
| "step": 4545 | |
| }, | |
| { | |
| "epoch": 0.9553739786297926, | |
| "grad_norm": 2.0956265926361084, | |
| "learning_rate": 1.7267754516462773e-05, | |
| "loss": 1.1336, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 0.9585166561910748, | |
| "grad_norm": 2.3662939071655273, | |
| "learning_rate": 1.724258902107936e-05, | |
| "loss": 1.1807, | |
| "step": 4575 | |
| }, | |
| { | |
| "epoch": 0.961659333752357, | |
| "grad_norm": 2.2511234283447266, | |
| "learning_rate": 1.721732668186421e-05, | |
| "loss": 1.1407, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 0.9648020113136392, | |
| "grad_norm": 2.2256953716278076, | |
| "learning_rate": 1.7191967836611115e-05, | |
| "loss": 1.1797, | |
| "step": 4605 | |
| }, | |
| { | |
| "epoch": 0.9679446888749215, | |
| "grad_norm": 1.6837149858474731, | |
| "learning_rate": 1.716651282440432e-05, | |
| "loss": 1.1586, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 0.9710873664362036, | |
| "grad_norm": 1.941137433052063, | |
| "learning_rate": 1.7140961985613952e-05, | |
| "loss": 1.1739, | |
| "step": 4635 | |
| }, | |
| { | |
| "epoch": 0.9742300439974858, | |
| "grad_norm": 3.3290293216705322, | |
| "learning_rate": 1.711531566189146e-05, | |
| "loss": 1.1511, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 0.9773727215587681, | |
| "grad_norm": 2.164433002471924, | |
| "learning_rate": 1.7089574196165087e-05, | |
| "loss": 1.1144, | |
| "step": 4665 | |
| }, | |
| { | |
| "epoch": 0.9805153991200503, | |
| "grad_norm": 2.117098569869995, | |
| "learning_rate": 1.706373793263525e-05, | |
| "loss": 1.0804, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 0.9836580766813325, | |
| "grad_norm": 2.2358627319335938, | |
| "learning_rate": 1.7037807216769952e-05, | |
| "loss": 1.158, | |
| "step": 4695 | |
| }, | |
| { | |
| "epoch": 0.9868007542426147, | |
| "grad_norm": 2.1659979820251465, | |
| "learning_rate": 1.7011782395300152e-05, | |
| "loss": 1.1633, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 0.9899434318038969, | |
| "grad_norm": 1.9542781114578247, | |
| "learning_rate": 1.6985663816215156e-05, | |
| "loss": 1.2018, | |
| "step": 4725 | |
| }, | |
| { | |
| "epoch": 0.9930861093651792, | |
| "grad_norm": 2.3751957416534424, | |
| "learning_rate": 1.6959451828757915e-05, | |
| "loss": 1.1924, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 0.9962287869264613, | |
| "grad_norm": 2.5773661136627197, | |
| "learning_rate": 1.6933146783420412e-05, | |
| "loss": 1.2159, | |
| "step": 4755 | |
| }, | |
| { | |
| "epoch": 0.9993714644877436, | |
| "grad_norm": 1.986519694328308, | |
| "learning_rate": 1.6906749031938932e-05, | |
| "loss": 1.1635, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 1.0025141420490258, | |
| "grad_norm": 2.238046646118164, | |
| "learning_rate": 1.688025892728938e-05, | |
| "loss": 1.1517, | |
| "step": 4785 | |
| }, | |
| { | |
| "epoch": 1.005656819610308, | |
| "grad_norm": 2.3779406547546387, | |
| "learning_rate": 1.685367682368255e-05, | |
| "loss": 1.1341, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 1.0087994971715901, | |
| "grad_norm": 2.1286470890045166, | |
| "learning_rate": 1.6827003076559403e-05, | |
| "loss": 1.1293, | |
| "step": 4815 | |
| }, | |
| { | |
| "epoch": 1.0119421747328725, | |
| "grad_norm": 3.604013442993164, | |
| "learning_rate": 1.68002380425863e-05, | |
| "loss": 1.1733, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 1.0150848522941547, | |
| "grad_norm": 1.7821085453033447, | |
| "learning_rate": 1.6773382079650237e-05, | |
| "loss": 1.1215, | |
| "step": 4845 | |
| }, | |
| { | |
| "epoch": 1.0182275298554369, | |
| "grad_norm": 2.2310407161712646, | |
| "learning_rate": 1.6746435546854068e-05, | |
| "loss": 1.1364, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 1.021370207416719, | |
| "grad_norm": 2.5142269134521484, | |
| "learning_rate": 1.6719398804511682e-05, | |
| "loss": 1.1585, | |
| "step": 4875 | |
| }, | |
| { | |
| "epoch": 1.0245128849780012, | |
| "grad_norm": 2.2651455402374268, | |
| "learning_rate": 1.6692272214143224e-05, | |
| "loss": 1.1408, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 1.0276555625392834, | |
| "grad_norm": 2.247772693634033, | |
| "learning_rate": 1.6665056138470212e-05, | |
| "loss": 1.1134, | |
| "step": 4905 | |
| }, | |
| { | |
| "epoch": 1.0307982401005658, | |
| "grad_norm": 2.3006253242492676, | |
| "learning_rate": 1.6637750941410725e-05, | |
| "loss": 1.1252, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 1.033940917661848, | |
| "grad_norm": 1.8263660669326782, | |
| "learning_rate": 1.6610356988074517e-05, | |
| "loss": 1.1427, | |
| "step": 4935 | |
| }, | |
| { | |
| "epoch": 1.03708359522313, | |
| "grad_norm": 2.199976921081543, | |
| "learning_rate": 1.658287464475814e-05, | |
| "loss": 1.1208, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 1.0402262727844123, | |
| "grad_norm": 2.1440887451171875, | |
| "learning_rate": 1.6555304278940053e-05, | |
| "loss": 1.1305, | |
| "step": 4965 | |
| }, | |
| { | |
| "epoch": 1.0433689503456944, | |
| "grad_norm": 2.2581751346588135, | |
| "learning_rate": 1.65276462592757e-05, | |
| "loss": 1.1683, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 1.0465116279069768, | |
| "grad_norm": 1.7286102771759033, | |
| "learning_rate": 1.6499900955592587e-05, | |
| "loss": 1.1857, | |
| "step": 4995 | |
| }, | |
| { | |
| "epoch": 1.049654305468259, | |
| "grad_norm": 1.8500277996063232, | |
| "learning_rate": 1.6472068738885322e-05, | |
| "loss": 1.0997, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 1.0527969830295412, | |
| "grad_norm": 2.4594027996063232, | |
| "learning_rate": 1.644414998131067e-05, | |
| "loss": 1.0588, | |
| "step": 5025 | |
| }, | |
| { | |
| "epoch": 1.0559396605908233, | |
| "grad_norm": 2.3058676719665527, | |
| "learning_rate": 1.6416145056182577e-05, | |
| "loss": 1.1304, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 1.0590823381521055, | |
| "grad_norm": 1.9614354372024536, | |
| "learning_rate": 1.6388054337967173e-05, | |
| "loss": 1.1542, | |
| "step": 5055 | |
| }, | |
| { | |
| "epoch": 1.062225015713388, | |
| "grad_norm": 2.4512290954589844, | |
| "learning_rate": 1.635987820227776e-05, | |
| "loss": 1.1455, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 1.06536769327467, | |
| "grad_norm": 3.3408472537994385, | |
| "learning_rate": 1.6331617025869787e-05, | |
| "loss": 1.1008, | |
| "step": 5085 | |
| }, | |
| { | |
| "epoch": 1.0685103708359522, | |
| "grad_norm": 3.3527655601501465, | |
| "learning_rate": 1.6303271186635837e-05, | |
| "loss": 1.1311, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 1.0716530483972344, | |
| "grad_norm": 2.1265969276428223, | |
| "learning_rate": 1.627484106360055e-05, | |
| "loss": 1.1328, | |
| "step": 5115 | |
| }, | |
| { | |
| "epoch": 1.0747957259585166, | |
| "grad_norm": 2.1224098205566406, | |
| "learning_rate": 1.624632703691556e-05, | |
| "loss": 1.1432, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 1.077938403519799, | |
| "grad_norm": 2.1932625770568848, | |
| "learning_rate": 1.6217729487854404e-05, | |
| "loss": 1.1336, | |
| "step": 5145 | |
| }, | |
| { | |
| "epoch": 1.0810810810810811, | |
| "grad_norm": 1.477196455001831, | |
| "learning_rate": 1.6189048798807453e-05, | |
| "loss": 1.0739, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 1.0842237586423633, | |
| "grad_norm": 2.2613472938537598, | |
| "learning_rate": 1.616028535327677e-05, | |
| "loss": 1.0342, | |
| "step": 5175 | |
| }, | |
| { | |
| "epoch": 1.0873664362036455, | |
| "grad_norm": 2.0612030029296875, | |
| "learning_rate": 1.6131439535870982e-05, | |
| "loss": 1.2227, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 1.0905091137649277, | |
| "grad_norm": 2.0832862854003906, | |
| "learning_rate": 1.6102511732300156e-05, | |
| "loss": 1.1417, | |
| "step": 5205 | |
| }, | |
| { | |
| "epoch": 1.0936517913262098, | |
| "grad_norm": 2.326160192489624, | |
| "learning_rate": 1.607350232937064e-05, | |
| "loss": 1.1104, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 1.0967944688874922, | |
| "grad_norm": 1.928688883781433, | |
| "learning_rate": 1.6044411714979866e-05, | |
| "loss": 1.1503, | |
| "step": 5235 | |
| }, | |
| { | |
| "epoch": 1.0999371464487744, | |
| "grad_norm": 2.209092617034912, | |
| "learning_rate": 1.6015240278111184e-05, | |
| "loss": 1.0927, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 1.1030798240100566, | |
| "grad_norm": 2.2068710327148438, | |
| "learning_rate": 1.5987941027952896e-05, | |
| "loss": 1.097, | |
| "step": 5265 | |
| }, | |
| { | |
| "epoch": 1.1062225015713387, | |
| "grad_norm": 2.1571450233459473, | |
| "learning_rate": 1.5958614441290784e-05, | |
| "loss": 1.0841, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 1.109365179132621, | |
| "grad_norm": 2.010093927383423, | |
| "learning_rate": 1.5929208179383714e-05, | |
| "loss": 1.1261, | |
| "step": 5295 | |
| }, | |
| { | |
| "epoch": 1.1125078566939033, | |
| "grad_norm": 2.4886395931243896, | |
| "learning_rate": 1.58997226354357e-05, | |
| "loss": 1.1357, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 1.1156505342551855, | |
| "grad_norm": 1.658639907836914, | |
| "learning_rate": 1.5870158203710876e-05, | |
| "loss": 1.0573, | |
| "step": 5325 | |
| }, | |
| { | |
| "epoch": 1.1187932118164676, | |
| "grad_norm": 2.1249542236328125, | |
| "learning_rate": 1.584051527952821e-05, | |
| "loss": 1.1161, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 1.1219358893777498, | |
| "grad_norm": 2.086566209793091, | |
| "learning_rate": 1.5810794259256238e-05, | |
| "loss": 1.1282, | |
| "step": 5355 | |
| }, | |
| { | |
| "epoch": 1.125078566939032, | |
| "grad_norm": 2.5150198936462402, | |
| "learning_rate": 1.5780995540307734e-05, | |
| "loss": 1.1408, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 1.1282212445003144, | |
| "grad_norm": 3.7854840755462646, | |
| "learning_rate": 1.575111952113444e-05, | |
| "loss": 1.1125, | |
| "step": 5385 | |
| }, | |
| { | |
| "epoch": 1.1313639220615965, | |
| "grad_norm": 2.2000720500946045, | |
| "learning_rate": 1.5721166601221697e-05, | |
| "loss": 1.099, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 1.1345065996228787, | |
| "grad_norm": 2.055802345275879, | |
| "learning_rate": 1.5691137181083126e-05, | |
| "loss": 1.1057, | |
| "step": 5415 | |
| }, | |
| { | |
| "epoch": 1.1376492771841609, | |
| "grad_norm": 1.9730473756790161, | |
| "learning_rate": 1.5661031662255254e-05, | |
| "loss": 1.0642, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 1.140791954745443, | |
| "grad_norm": 2.1444499492645264, | |
| "learning_rate": 1.563085044729218e-05, | |
| "loss": 1.1095, | |
| "step": 5445 | |
| }, | |
| { | |
| "epoch": 1.1439346323067254, | |
| "grad_norm": 3.2961201667785645, | |
| "learning_rate": 1.5600593939760146e-05, | |
| "loss": 1.1613, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 1.1470773098680076, | |
| "grad_norm": 2.405123710632324, | |
| "learning_rate": 1.557026254423218e-05, | |
| "loss": 1.1211, | |
| "step": 5475 | |
| }, | |
| { | |
| "epoch": 1.1502199874292898, | |
| "grad_norm": 3.308258533477783, | |
| "learning_rate": 1.5539856666282655e-05, | |
| "loss": 1.1381, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 1.153362664990572, | |
| "grad_norm": 2.2963814735412598, | |
| "learning_rate": 1.5509376712481898e-05, | |
| "loss": 1.1293, | |
| "step": 5505 | |
| }, | |
| { | |
| "epoch": 1.156505342551854, | |
| "grad_norm": 1.9797133207321167, | |
| "learning_rate": 1.5478823090390726e-05, | |
| "loss": 1.0968, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 1.1596480201131363, | |
| "grad_norm": 2.6032137870788574, | |
| "learning_rate": 1.5448196208555014e-05, | |
| "loss": 1.1462, | |
| "step": 5535 | |
| }, | |
| { | |
| "epoch": 1.1627906976744187, | |
| "grad_norm": 1.9197511672973633, | |
| "learning_rate": 1.5417496476500212e-05, | |
| "loss": 1.0224, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 1.1659333752357008, | |
| "grad_norm": 2.2676188945770264, | |
| "learning_rate": 1.538672430472589e-05, | |
| "loss": 1.0625, | |
| "step": 5565 | |
| }, | |
| { | |
| "epoch": 1.169076052796983, | |
| "grad_norm": 2.4030728340148926, | |
| "learning_rate": 1.535588010470026e-05, | |
| "loss": 1.1171, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 1.1722187303582652, | |
| "grad_norm": 1.387372374534607, | |
| "learning_rate": 1.5324964288854615e-05, | |
| "loss": 1.0836, | |
| "step": 5595 | |
| }, | |
| { | |
| "epoch": 1.1753614079195476, | |
| "grad_norm": 2.1212260723114014, | |
| "learning_rate": 1.5293977270577895e-05, | |
| "loss": 1.1069, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 1.1785040854808297, | |
| "grad_norm": 2.2747840881347656, | |
| "learning_rate": 1.526291946421109e-05, | |
| "loss": 1.134, | |
| "step": 5625 | |
| }, | |
| { | |
| "epoch": 1.181646763042112, | |
| "grad_norm": 1.7659993171691895, | |
| "learning_rate": 1.5231791285041741e-05, | |
| "loss": 1.1512, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 1.184789440603394, | |
| "grad_norm": 1.815944790840149, | |
| "learning_rate": 1.5200593149298375e-05, | |
| "loss": 1.0769, | |
| "step": 5655 | |
| }, | |
| { | |
| "epoch": 1.1879321181646763, | |
| "grad_norm": 2.154893636703491, | |
| "learning_rate": 1.5169325474144936e-05, | |
| "loss": 1.1198, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 1.1910747957259584, | |
| "grad_norm": 1.9412492513656616, | |
| "learning_rate": 1.5137988677675215e-05, | |
| "loss": 1.115, | |
| "step": 5685 | |
| }, | |
| { | |
| "epoch": 1.1942174732872408, | |
| "grad_norm": 1.8665941953659058, | |
| "learning_rate": 1.5106583178907244e-05, | |
| "loss": 1.1067, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 1.197360150848523, | |
| "grad_norm": 1.9800127744674683, | |
| "learning_rate": 1.5075109397777714e-05, | |
| "loss": 1.0744, | |
| "step": 5715 | |
| }, | |
| { | |
| "epoch": 1.2005028284098052, | |
| "grad_norm": 3.109827756881714, | |
| "learning_rate": 1.5043567755136346e-05, | |
| "loss": 1.0699, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 1.2036455059710873, | |
| "grad_norm": 1.9781428575515747, | |
| "learning_rate": 1.5011958672740261e-05, | |
| "loss": 1.1205, | |
| "step": 5745 | |
| }, | |
| { | |
| "epoch": 1.2067881835323695, | |
| "grad_norm": 2.4044976234436035, | |
| "learning_rate": 1.498028257324836e-05, | |
| "loss": 1.11, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 1.2099308610936519, | |
| "grad_norm": 2.0579380989074707, | |
| "learning_rate": 1.4948539880215642e-05, | |
| "loss": 1.1586, | |
| "step": 5775 | |
| }, | |
| { | |
| "epoch": 1.213073538654934, | |
| "grad_norm": 3.235210418701172, | |
| "learning_rate": 1.4916731018087576e-05, | |
| "loss": 1.1164, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 1.2162162162162162, | |
| "grad_norm": 2.258272647857666, | |
| "learning_rate": 1.4884856412194386e-05, | |
| "loss": 1.1057, | |
| "step": 5805 | |
| }, | |
| { | |
| "epoch": 1.2193588937774984, | |
| "grad_norm": 3.207092761993408, | |
| "learning_rate": 1.4852916488745409e-05, | |
| "loss": 1.1112, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 1.2225015713387806, | |
| "grad_norm": 2.802191972732544, | |
| "learning_rate": 1.4820911674823345e-05, | |
| "loss": 1.0854, | |
| "step": 5835 | |
| }, | |
| { | |
| "epoch": 1.2256442489000627, | |
| "grad_norm": 2.0806314945220947, | |
| "learning_rate": 1.4788842398378595e-05, | |
| "loss": 1.1596, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 1.2287869264613451, | |
| "grad_norm": 2.3855271339416504, | |
| "learning_rate": 1.4756709088223508e-05, | |
| "loss": 1.1108, | |
| "step": 5865 | |
| }, | |
| { | |
| "epoch": 1.2319296040226273, | |
| "grad_norm": 2.3936338424682617, | |
| "learning_rate": 1.4724512174026653e-05, | |
| "loss": 1.1988, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 1.2350722815839095, | |
| "grad_norm": 1.9991942644119263, | |
| "learning_rate": 1.4692252086307079e-05, | |
| "loss": 1.1588, | |
| "step": 5895 | |
| }, | |
| { | |
| "epoch": 1.2382149591451916, | |
| "grad_norm": 2.2678310871124268, | |
| "learning_rate": 1.465992925642856e-05, | |
| "loss": 1.084, | |
| "step": 5910 | |
| }, | |
| { | |
| "epoch": 1.241357636706474, | |
| "grad_norm": 2.767444610595703, | |
| "learning_rate": 1.4627544116593816e-05, | |
| "loss": 1.1242, | |
| "step": 5925 | |
| }, | |
| { | |
| "epoch": 1.2445003142677562, | |
| "grad_norm": 1.6328760385513306, | |
| "learning_rate": 1.4595097099838747e-05, | |
| "loss": 1.087, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 1.2476429918290384, | |
| "grad_norm": 1.8721544742584229, | |
| "learning_rate": 1.4562588640026629e-05, | |
| "loss": 1.1078, | |
| "step": 5955 | |
| }, | |
| { | |
| "epoch": 1.2507856693903205, | |
| "grad_norm": 2.0327460765838623, | |
| "learning_rate": 1.453001917184233e-05, | |
| "loss": 1.0601, | |
| "step": 5970 | |
| }, | |
| { | |
| "epoch": 1.2539283469516027, | |
| "grad_norm": 1.8650814294815063, | |
| "learning_rate": 1.4497389130786474e-05, | |
| "loss": 1.0553, | |
| "step": 5985 | |
| }, | |
| { | |
| "epoch": 1.2570710245128849, | |
| "grad_norm": 2.416186809539795, | |
| "learning_rate": 1.4464698953169649e-05, | |
| "loss": 1.1542, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 1.260213702074167, | |
| "grad_norm": 3.0162692070007324, | |
| "learning_rate": 1.443194907610654e-05, | |
| "loss": 1.1424, | |
| "step": 6015 | |
| }, | |
| { | |
| "epoch": 1.2633563796354494, | |
| "grad_norm": 2.833125352859497, | |
| "learning_rate": 1.4399139937510107e-05, | |
| "loss": 1.1691, | |
| "step": 6030 | |
| }, | |
| { | |
| "epoch": 1.2664990571967316, | |
| "grad_norm": 1.900730848312378, | |
| "learning_rate": 1.4366271976085719e-05, | |
| "loss": 1.0463, | |
| "step": 6045 | |
| }, | |
| { | |
| "epoch": 1.2696417347580138, | |
| "grad_norm": 2.243030071258545, | |
| "learning_rate": 1.4333345631325294e-05, | |
| "loss": 1.1671, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 1.2727844123192962, | |
| "grad_norm": 2.186521530151367, | |
| "learning_rate": 1.430036134350142e-05, | |
| "loss": 1.1653, | |
| "step": 6075 | |
| }, | |
| { | |
| "epoch": 1.2759270898805783, | |
| "grad_norm": 1.8847732543945312, | |
| "learning_rate": 1.4267319553661456e-05, | |
| "loss": 1.0842, | |
| "step": 6090 | |
| }, | |
| { | |
| "epoch": 1.2790697674418605, | |
| "grad_norm": 2.303593158721924, | |
| "learning_rate": 1.4234220703621658e-05, | |
| "loss": 1.0481, | |
| "step": 6105 | |
| }, | |
| { | |
| "epoch": 1.2822124450031427, | |
| "grad_norm": 3.213543653488159, | |
| "learning_rate": 1.4201065235961258e-05, | |
| "loss": 1.1446, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 1.2853551225644249, | |
| "grad_norm": 2.283177137374878, | |
| "learning_rate": 1.4167853594016539e-05, | |
| "loss": 1.114, | |
| "step": 6135 | |
| }, | |
| { | |
| "epoch": 1.288497800125707, | |
| "grad_norm": 3.1280677318573, | |
| "learning_rate": 1.4134586221874922e-05, | |
| "loss": 1.1568, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 1.2916404776869892, | |
| "grad_norm": 3.3405239582061768, | |
| "learning_rate": 1.410126356436902e-05, | |
| "loss": 1.0806, | |
| "step": 6165 | |
| }, | |
| { | |
| "epoch": 1.2947831552482716, | |
| "grad_norm": 2.9373440742492676, | |
| "learning_rate": 1.406788606707069e-05, | |
| "loss": 1.128, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 1.2979258328095538, | |
| "grad_norm": 2.1123921871185303, | |
| "learning_rate": 1.4034454176285073e-05, | |
| "loss": 1.017, | |
| "step": 6195 | |
| }, | |
| { | |
| "epoch": 1.301068510370836, | |
| "grad_norm": 2.298046827316284, | |
| "learning_rate": 1.4000968339044637e-05, | |
| "loss": 1.157, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 1.304211187932118, | |
| "grad_norm": 3.053687572479248, | |
| "learning_rate": 1.3967429003103175e-05, | |
| "loss": 1.1575, | |
| "step": 6225 | |
| }, | |
| { | |
| "epoch": 1.3073538654934005, | |
| "grad_norm": 2.9560487270355225, | |
| "learning_rate": 1.3933836616929856e-05, | |
| "loss": 1.1414, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 1.3104965430546827, | |
| "grad_norm": 2.401231050491333, | |
| "learning_rate": 1.3900191629703192e-05, | |
| "loss": 1.0565, | |
| "step": 6255 | |
| }, | |
| { | |
| "epoch": 1.3136392206159648, | |
| "grad_norm": 2.4497265815734863, | |
| "learning_rate": 1.3866494491305058e-05, | |
| "loss": 1.1296, | |
| "step": 6270 | |
| }, | |
| { | |
| "epoch": 1.316781898177247, | |
| "grad_norm": 2.0227253437042236, | |
| "learning_rate": 1.3832745652314652e-05, | |
| "loss": 1.1229, | |
| "step": 6285 | |
| }, | |
| { | |
| "epoch": 1.3199245757385292, | |
| "grad_norm": 1.9386115074157715, | |
| "learning_rate": 1.3798945564002493e-05, | |
| "loss": 1.0128, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 1.3230672532998113, | |
| "grad_norm": 1.967246413230896, | |
| "learning_rate": 1.376509467832437e-05, | |
| "loss": 1.1388, | |
| "step": 6315 | |
| }, | |
| { | |
| "epoch": 1.3262099308610937, | |
| "grad_norm": 3.249537706375122, | |
| "learning_rate": 1.3731193447915316e-05, | |
| "loss": 1.0584, | |
| "step": 6330 | |
| }, | |
| { | |
| "epoch": 1.329352608422376, | |
| "grad_norm": 2.7894890308380127, | |
| "learning_rate": 1.3697242326083536e-05, | |
| "loss": 1.1075, | |
| "step": 6345 | |
| }, | |
| { | |
| "epoch": 1.332495285983658, | |
| "grad_norm": 3.2315704822540283, | |
| "learning_rate": 1.3663241766804359e-05, | |
| "loss": 1.0972, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 1.3356379635449402, | |
| "grad_norm": 2.295436143875122, | |
| "learning_rate": 1.3629192224714163e-05, | |
| "loss": 1.1194, | |
| "step": 6375 | |
| }, | |
| { | |
| "epoch": 1.3387806411062226, | |
| "grad_norm": 1.2325416803359985, | |
| "learning_rate": 1.3595094155104297e-05, | |
| "loss": 1.099, | |
| "step": 6390 | |
| }, | |
| { | |
| "epoch": 1.3419233186675048, | |
| "grad_norm": 2.3862216472625732, | |
| "learning_rate": 1.3560948013914997e-05, | |
| "loss": 1.0684, | |
| "step": 6405 | |
| }, | |
| { | |
| "epoch": 1.345065996228787, | |
| "grad_norm": 2.1027228832244873, | |
| "learning_rate": 1.3526754257729273e-05, | |
| "loss": 1.0776, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 1.3482086737900691, | |
| "grad_norm": 2.421541213989258, | |
| "learning_rate": 1.3492513343766831e-05, | |
| "loss": 1.1078, | |
| "step": 6435 | |
| }, | |
| { | |
| "epoch": 1.3513513513513513, | |
| "grad_norm": 2.7943713665008545, | |
| "learning_rate": 1.3458225729877933e-05, | |
| "loss": 1.1076, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 1.3544940289126335, | |
| "grad_norm": 2.077890396118164, | |
| "learning_rate": 1.3423891874537289e-05, | |
| "loss": 1.1609, | |
| "step": 6465 | |
| }, | |
| { | |
| "epoch": 1.3576367064739157, | |
| "grad_norm": 2.0880138874053955, | |
| "learning_rate": 1.3389512236837931e-05, | |
| "loss": 1.0667, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 1.360779384035198, | |
| "grad_norm": 2.1045663356781006, | |
| "learning_rate": 1.3355087276485055e-05, | |
| "loss": 1.113, | |
| "step": 6495 | |
| }, | |
| { | |
| "epoch": 1.3639220615964802, | |
| "grad_norm": 2.125664472579956, | |
| "learning_rate": 1.3320617453789894e-05, | |
| "loss": 1.0619, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 1.3670647391577624, | |
| "grad_norm": 2.1109349727630615, | |
| "learning_rate": 1.3286103229663554e-05, | |
| "loss": 1.1235, | |
| "step": 6525 | |
| }, | |
| { | |
| "epoch": 1.3702074167190446, | |
| "grad_norm": 2.0411548614501953, | |
| "learning_rate": 1.3251545065610848e-05, | |
| "loss": 1.0872, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 1.373350094280327, | |
| "grad_norm": 3.083622694015503, | |
| "learning_rate": 1.3216943423724135e-05, | |
| "loss": 1.117, | |
| "step": 6555 | |
| }, | |
| { | |
| "epoch": 1.3764927718416091, | |
| "grad_norm": 1.8469839096069336, | |
| "learning_rate": 1.3182298766677136e-05, | |
| "loss": 1.1404, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 1.3796354494028913, | |
| "grad_norm": 2.121103525161743, | |
| "learning_rate": 1.3147611557718733e-05, | |
| "loss": 1.1007, | |
| "step": 6585 | |
| }, | |
| { | |
| "epoch": 1.3827781269641735, | |
| "grad_norm": 2.183732271194458, | |
| "learning_rate": 1.3112882260666805e-05, | |
| "loss": 1.1362, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 1.3859208045254556, | |
| "grad_norm": 2.0963168144226074, | |
| "learning_rate": 1.3078111339902004e-05, | |
| "loss": 1.1019, | |
| "step": 6615 | |
| }, | |
| { | |
| "epoch": 1.3890634820867378, | |
| "grad_norm": 3.269718885421753, | |
| "learning_rate": 1.3043299260361545e-05, | |
| "loss": 1.1073, | |
| "step": 6630 | |
| }, | |
| { | |
| "epoch": 1.3922061596480202, | |
| "grad_norm": 2.942408323287964, | |
| "learning_rate": 1.3008446487532998e-05, | |
| "loss": 1.0783, | |
| "step": 6645 | |
| }, | |
| { | |
| "epoch": 1.3953488372093024, | |
| "grad_norm": 2.6441168785095215, | |
| "learning_rate": 1.2973553487448068e-05, | |
| "loss": 1.1656, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 1.3984915147705845, | |
| "grad_norm": 2.102250099182129, | |
| "learning_rate": 1.2938620726676345e-05, | |
| "loss": 1.1408, | |
| "step": 6675 | |
| }, | |
| { | |
| "epoch": 1.4016341923318667, | |
| "grad_norm": 3.086590051651001, | |
| "learning_rate": 1.2903648672319084e-05, | |
| "loss": 1.1059, | |
| "step": 6690 | |
| }, | |
| { | |
| "epoch": 1.404776869893149, | |
| "grad_norm": 2.3277883529663086, | |
| "learning_rate": 1.2868637792002952e-05, | |
| "loss": 1.1665, | |
| "step": 6705 | |
| }, | |
| { | |
| "epoch": 1.4079195474544313, | |
| "grad_norm": 1.6698582172393799, | |
| "learning_rate": 1.2833588553873768e-05, | |
| "loss": 1.0649, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 1.4110622250157134, | |
| "grad_norm": 2.26218581199646, | |
| "learning_rate": 1.2798501426590262e-05, | |
| "loss": 1.1894, | |
| "step": 6735 | |
| }, | |
| { | |
| "epoch": 1.4142049025769956, | |
| "grad_norm": 1.8945990800857544, | |
| "learning_rate": 1.276337687931778e-05, | |
| "loss": 1.1507, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 1.4173475801382778, | |
| "grad_norm": 2.481565475463867, | |
| "learning_rate": 1.2728215381722044e-05, | |
| "loss": 1.0937, | |
| "step": 6765 | |
| }, | |
| { | |
| "epoch": 1.42049025769956, | |
| "grad_norm": 3.565307140350342, | |
| "learning_rate": 1.2693017403962842e-05, | |
| "loss": 1.197, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 1.4236329352608421, | |
| "grad_norm": 2.697265863418579, | |
| "learning_rate": 1.2657783416687763e-05, | |
| "loss": 1.0842, | |
| "step": 6795 | |
| }, | |
| { | |
| "epoch": 1.4267756128221245, | |
| "grad_norm": 2.244889736175537, | |
| "learning_rate": 1.2622513891025889e-05, | |
| "loss": 1.0942, | |
| "step": 6810 | |
| }, | |
| { | |
| "epoch": 1.4299182903834067, | |
| "grad_norm": 2.428379774093628, | |
| "learning_rate": 1.2587209298581506e-05, | |
| "loss": 1.131, | |
| "step": 6825 | |
| }, | |
| { | |
| "epoch": 1.4330609679446888, | |
| "grad_norm": 2.320680618286133, | |
| "learning_rate": 1.2551870111427791e-05, | |
| "loss": 1.1279, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 1.436203645505971, | |
| "grad_norm": 2.004258632659912, | |
| "learning_rate": 1.25164968021005e-05, | |
| "loss": 1.0718, | |
| "step": 6855 | |
| }, | |
| { | |
| "epoch": 1.4393463230672534, | |
| "grad_norm": 3.006464719772339, | |
| "learning_rate": 1.2481089843591655e-05, | |
| "loss": 1.1785, | |
| "step": 6870 | |
| }, | |
| { | |
| "epoch": 1.4424890006285356, | |
| "grad_norm": 2.481252670288086, | |
| "learning_rate": 1.2445649709343217e-05, | |
| "loss": 1.1382, | |
| "step": 6885 | |
| }, | |
| { | |
| "epoch": 1.4456316781898177, | |
| "grad_norm": 2.247004985809326, | |
| "learning_rate": 1.2410176873240751e-05, | |
| "loss": 1.1179, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 1.4487743557511, | |
| "grad_norm": 2.7122623920440674, | |
| "learning_rate": 1.237467180960709e-05, | |
| "loss": 1.0569, | |
| "step": 6915 | |
| }, | |
| { | |
| "epoch": 1.451917033312382, | |
| "grad_norm": 2.119673013687134, | |
| "learning_rate": 1.2339134993196007e-05, | |
| "loss": 1.0939, | |
| "step": 6930 | |
| }, | |
| { | |
| "epoch": 1.4550597108736643, | |
| "grad_norm": 2.4448978900909424, | |
| "learning_rate": 1.2303566899185842e-05, | |
| "loss": 1.0598, | |
| "step": 6945 | |
| }, | |
| { | |
| "epoch": 1.4582023884349467, | |
| "grad_norm": 2.211289405822754, | |
| "learning_rate": 1.2267968003173166e-05, | |
| "loss": 1.1679, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 1.4613450659962288, | |
| "grad_norm": 1.9416890144348145, | |
| "learning_rate": 1.2232338781166422e-05, | |
| "loss": 1.0983, | |
| "step": 6975 | |
| }, | |
| { | |
| "epoch": 1.464487743557511, | |
| "grad_norm": 2.0862271785736084, | |
| "learning_rate": 1.2196679709579546e-05, | |
| "loss": 1.0911, | |
| "step": 6990 | |
| }, | |
| { | |
| "epoch": 1.4676304211187932, | |
| "grad_norm": 2.4440407752990723, | |
| "learning_rate": 1.2160991265225621e-05, | |
| "loss": 1.1073, | |
| "step": 7005 | |
| }, | |
| { | |
| "epoch": 1.4707730986800756, | |
| "grad_norm": 2.1422340869903564, | |
| "learning_rate": 1.2125273925310465e-05, | |
| "loss": 1.0904, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 1.4739157762413577, | |
| "grad_norm": 2.2515573501586914, | |
| "learning_rate": 1.2089528167426294e-05, | |
| "loss": 1.1244, | |
| "step": 7035 | |
| }, | |
| { | |
| "epoch": 1.47705845380264, | |
| "grad_norm": 2.149007558822632, | |
| "learning_rate": 1.2053754469545291e-05, | |
| "loss": 1.0963, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 1.480201131363922, | |
| "grad_norm": 3.1129860877990723, | |
| "learning_rate": 1.2017953310013255e-05, | |
| "loss": 1.1655, | |
| "step": 7065 | |
| }, | |
| { | |
| "epoch": 1.4833438089252042, | |
| "grad_norm": 1.7863973379135132, | |
| "learning_rate": 1.1982125167543175e-05, | |
| "loss": 1.0844, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 1.4864864864864864, | |
| "grad_norm": 2.28974986076355, | |
| "learning_rate": 1.1946270521208845e-05, | |
| "loss": 1.0585, | |
| "step": 7095 | |
| }, | |
| { | |
| "epoch": 1.4896291640477686, | |
| "grad_norm": 2.3061442375183105, | |
| "learning_rate": 1.1910389850438457e-05, | |
| "loss": 1.1377, | |
| "step": 7110 | |
| }, | |
| { | |
| "epoch": 1.492771841609051, | |
| "grad_norm": 1.6387587785720825, | |
| "learning_rate": 1.1874483635008183e-05, | |
| "loss": 1.1109, | |
| "step": 7125 | |
| }, | |
| { | |
| "epoch": 1.4959145191703331, | |
| "grad_norm": 1.6554300785064697, | |
| "learning_rate": 1.1838552355035761e-05, | |
| "loss": 1.0686, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 1.4990571967316153, | |
| "grad_norm": 2.2356905937194824, | |
| "learning_rate": 1.1802596490974088e-05, | |
| "loss": 1.0877, | |
| "step": 7155 | |
| }, | |
| { | |
| "epoch": 1.5021998742928977, | |
| "grad_norm": 2.838385820388794, | |
| "learning_rate": 1.1766616523604774e-05, | |
| "loss": 1.1015, | |
| "step": 7170 | |
| }, | |
| { | |
| "epoch": 1.5053425518541799, | |
| "grad_norm": 1.8725031614303589, | |
| "learning_rate": 1.1730612934031737e-05, | |
| "loss": 1.0806, | |
| "step": 7185 | |
| }, | |
| { | |
| "epoch": 1.508485229415462, | |
| "grad_norm": 2.853379011154175, | |
| "learning_rate": 1.1694586203674742e-05, | |
| "loss": 1.0796, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 1.5116279069767442, | |
| "grad_norm": 2.5570309162139893, | |
| "learning_rate": 1.1658536814263e-05, | |
| "loss": 1.0773, | |
| "step": 7215 | |
| }, | |
| { | |
| "epoch": 1.5147705845380264, | |
| "grad_norm": 2.4062445163726807, | |
| "learning_rate": 1.1622465247828681e-05, | |
| "loss": 0.9971, | |
| "step": 7230 | |
| }, | |
| { | |
| "epoch": 1.5179132620993085, | |
| "grad_norm": 2.4074044227600098, | |
| "learning_rate": 1.1586371986700522e-05, | |
| "loss": 1.0774, | |
| "step": 7245 | |
| }, | |
| { | |
| "epoch": 1.5210559396605907, | |
| "grad_norm": 2.0330376625061035, | |
| "learning_rate": 1.155025751349732e-05, | |
| "loss": 1.1209, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 1.5241986172218729, | |
| "grad_norm": 2.229322671890259, | |
| "learning_rate": 1.1516531959834856e-05, | |
| "loss": 1.1011, | |
| "step": 7275 | |
| }, | |
| { | |
| "epoch": 1.5273412947831553, | |
| "grad_norm": 2.2459731101989746, | |
| "learning_rate": 1.1480377846161259e-05, | |
| "loss": 1.095, | |
| "step": 7290 | |
| }, | |
| { | |
| "epoch": 1.5304839723444374, | |
| "grad_norm": 1.9737571477890015, | |
| "learning_rate": 1.1444203937706724e-05, | |
| "loss": 1.0764, | |
| "step": 7305 | |
| }, | |
| { | |
| "epoch": 1.5336266499057196, | |
| "grad_norm": 2.0487782955169678, | |
| "learning_rate": 1.1408010718168447e-05, | |
| "loss": 1.1322, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 1.536769327467002, | |
| "grad_norm": 2.7282018661499023, | |
| "learning_rate": 1.1371798671501813e-05, | |
| "loss": 1.1322, | |
| "step": 7335 | |
| }, | |
| { | |
| "epoch": 1.5399120050282842, | |
| "grad_norm": 3.496860980987549, | |
| "learning_rate": 1.133556828191398e-05, | |
| "loss": 1.0829, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 1.5430546825895664, | |
| "grad_norm": 1.9811838865280151, | |
| "learning_rate": 1.1299320033857355e-05, | |
| "loss": 1.0813, | |
| "step": 7365 | |
| }, | |
| { | |
| "epoch": 1.5461973601508485, | |
| "grad_norm": 2.5440211296081543, | |
| "learning_rate": 1.1263054412023158e-05, | |
| "loss": 1.0347, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 1.5493400377121307, | |
| "grad_norm": 1.8516343832015991, | |
| "learning_rate": 1.1226771901334901e-05, | |
| "loss": 1.0719, | |
| "step": 7395 | |
| }, | |
| { | |
| "epoch": 1.5524827152734129, | |
| "grad_norm": 2.147573709487915, | |
| "learning_rate": 1.1190472986941939e-05, | |
| "loss": 1.068, | |
| "step": 7410 | |
| }, | |
| { | |
| "epoch": 1.555625392834695, | |
| "grad_norm": 1.999443531036377, | |
| "learning_rate": 1.1154158154212964e-05, | |
| "loss": 1.0674, | |
| "step": 7425 | |
| }, | |
| { | |
| "epoch": 1.5587680703959774, | |
| "grad_norm": 1.5025156736373901, | |
| "learning_rate": 1.111782788872952e-05, | |
| "loss": 1.1003, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 1.5619107479572596, | |
| "grad_norm": 2.162722587585449, | |
| "learning_rate": 1.1081482676279504e-05, | |
| "loss": 1.078, | |
| "step": 7455 | |
| }, | |
| { | |
| "epoch": 1.5650534255185418, | |
| "grad_norm": 2.049753427505493, | |
| "learning_rate": 1.1045123002850683e-05, | |
| "loss": 1.1825, | |
| "step": 7470 | |
| }, | |
| { | |
| "epoch": 1.5681961030798242, | |
| "grad_norm": 2.4640021324157715, | |
| "learning_rate": 1.1008749354624181e-05, | |
| "loss": 1.0942, | |
| "step": 7485 | |
| }, | |
| { | |
| "epoch": 1.5713387806411063, | |
| "grad_norm": 2.2149126529693604, | |
| "learning_rate": 1.0972362217967998e-05, | |
| "loss": 1.1415, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 1.5744814582023885, | |
| "grad_norm": 1.953383207321167, | |
| "learning_rate": 1.0935962079430468e-05, | |
| "loss": 1.0875, | |
| "step": 7515 | |
| }, | |
| { | |
| "epoch": 1.5776241357636707, | |
| "grad_norm": 2.371173620223999, | |
| "learning_rate": 1.0899549425733811e-05, | |
| "loss": 1.1068, | |
| "step": 7530 | |
| }, | |
| { | |
| "epoch": 1.5807668133249528, | |
| "grad_norm": 2.135591983795166, | |
| "learning_rate": 1.0863124743767566e-05, | |
| "loss": 1.1094, | |
| "step": 7545 | |
| }, | |
| { | |
| "epoch": 1.583909490886235, | |
| "grad_norm": 2.1090261936187744, | |
| "learning_rate": 1.0826688520582124e-05, | |
| "loss": 1.0459, | |
| "step": 7560 | |
| }, | |
| { | |
| "epoch": 1.5870521684475172, | |
| "grad_norm": 2.8963959217071533, | |
| "learning_rate": 1.0790241243382187e-05, | |
| "loss": 1.1019, | |
| "step": 7575 | |
| }, | |
| { | |
| "epoch": 1.5901948460087993, | |
| "grad_norm": 2.2832436561584473, | |
| "learning_rate": 1.0753783399520281e-05, | |
| "loss": 1.1125, | |
| "step": 7590 | |
| }, | |
| { | |
| "epoch": 1.5933375235700817, | |
| "grad_norm": 2.1328582763671875, | |
| "learning_rate": 1.0717315476490205e-05, | |
| "loss": 1.1088, | |
| "step": 7605 | |
| }, | |
| { | |
| "epoch": 1.596480201131364, | |
| "grad_norm": 1.7953342199325562, | |
| "learning_rate": 1.0680837961920538e-05, | |
| "loss": 1.0812, | |
| "step": 7620 | |
| }, | |
| { | |
| "epoch": 1.5996228786926463, | |
| "grad_norm": 2.0836551189422607, | |
| "learning_rate": 1.064435134356812e-05, | |
| "loss": 1.1015, | |
| "step": 7635 | |
| }, | |
| { | |
| "epoch": 1.6027655562539285, | |
| "grad_norm": 3.2132134437561035, | |
| "learning_rate": 1.0607856109311508e-05, | |
| "loss": 1.1377, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 1.6059082338152106, | |
| "grad_norm": 2.149661064147949, | |
| "learning_rate": 1.0571352747144477e-05, | |
| "loss": 1.0525, | |
| "step": 7665 | |
| }, | |
| { | |
| "epoch": 1.6090509113764928, | |
| "grad_norm": 1.8691046237945557, | |
| "learning_rate": 1.053484174516948e-05, | |
| "loss": 1.0954, | |
| "step": 7680 | |
| }, | |
| { | |
| "epoch": 1.612193588937775, | |
| "grad_norm": 1.9283403158187866, | |
| "learning_rate": 1.049832359159112e-05, | |
| "loss": 1.1199, | |
| "step": 7695 | |
| }, | |
| { | |
| "epoch": 1.6153362664990571, | |
| "grad_norm": 1.9671255350112915, | |
| "learning_rate": 1.0461798774709635e-05, | |
| "loss": 1.068, | |
| "step": 7710 | |
| }, | |
| { | |
| "epoch": 1.6184789440603393, | |
| "grad_norm": 2.721006393432617, | |
| "learning_rate": 1.0425267782914359e-05, | |
| "loss": 1.1274, | |
| "step": 7725 | |
| }, | |
| { | |
| "epoch": 1.6216216216216215, | |
| "grad_norm": 2.0059800148010254, | |
| "learning_rate": 1.0388731104677191e-05, | |
| "loss": 1.0832, | |
| "step": 7740 | |
| }, | |
| { | |
| "epoch": 1.6247642991829039, | |
| "grad_norm": 2.346212863922119, | |
| "learning_rate": 1.0352189228546068e-05, | |
| "loss": 1.1092, | |
| "step": 7755 | |
| }, | |
| { | |
| "epoch": 1.627906976744186, | |
| "grad_norm": 1.6566693782806396, | |
| "learning_rate": 1.031564264313843e-05, | |
| "loss": 1.033, | |
| "step": 7770 | |
| }, | |
| { | |
| "epoch": 1.6310496543054682, | |
| "grad_norm": 2.4089715480804443, | |
| "learning_rate": 1.0279091837134681e-05, | |
| "loss": 1.114, | |
| "step": 7785 | |
| }, | |
| { | |
| "epoch": 1.6341923318667506, | |
| "grad_norm": 2.9500551223754883, | |
| "learning_rate": 1.0242537299271675e-05, | |
| "loss": 1.1282, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 1.6373350094280328, | |
| "grad_norm": 1.8416279554367065, | |
| "learning_rate": 1.0205979518336148e-05, | |
| "loss": 1.0857, | |
| "step": 7815 | |
| }, | |
| { | |
| "epoch": 1.640477686989315, | |
| "grad_norm": 2.1354591846466064, | |
| "learning_rate": 1.0169418983158218e-05, | |
| "loss": 1.0807, | |
| "step": 7830 | |
| }, | |
| { | |
| "epoch": 1.6436203645505971, | |
| "grad_norm": 1.705767035484314, | |
| "learning_rate": 1.013285618260482e-05, | |
| "loss": 1.0273, | |
| "step": 7845 | |
| }, | |
| { | |
| "epoch": 1.6467630421118793, | |
| "grad_norm": 1.9959582090377808, | |
| "learning_rate": 1.009629160557318e-05, | |
| "loss": 1.0705, | |
| "step": 7860 | |
| }, | |
| { | |
| "epoch": 1.6499057196731615, | |
| "grad_norm": 2.004734754562378, | |
| "learning_rate": 1.0059725740984285e-05, | |
| "loss": 1.1216, | |
| "step": 7875 | |
| }, | |
| { | |
| "epoch": 1.6530483972344436, | |
| "grad_norm": 2.1259968280792236, | |
| "learning_rate": 1.0023159077776332e-05, | |
| "loss": 1.1618, | |
| "step": 7890 | |
| }, | |
| { | |
| "epoch": 1.6561910747957258, | |
| "grad_norm": 1.855248212814331, | |
| "learning_rate": 9.986592104898202e-06, | |
| "loss": 1.1007, | |
| "step": 7905 | |
| }, | |
| { | |
| "epoch": 1.6593337523570082, | |
| "grad_norm": 2.1948564052581787, | |
| "learning_rate": 9.950025311302914e-06, | |
| "loss": 1.1203, | |
| "step": 7920 | |
| }, | |
| { | |
| "epoch": 1.6624764299182904, | |
| "grad_norm": 2.140425205230713, | |
| "learning_rate": 9.913459185941085e-06, | |
| "loss": 1.0788, | |
| "step": 7935 | |
| }, | |
| { | |
| "epoch": 1.6656191074795728, | |
| "grad_norm": 3.158501148223877, | |
| "learning_rate": 9.87689421775441e-06, | |
| "loss": 1.0806, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 1.668761785040855, | |
| "grad_norm": 1.9126044511795044, | |
| "learning_rate": 9.840330895669096e-06, | |
| "loss": 1.086, | |
| "step": 7965 | |
| }, | |
| { | |
| "epoch": 1.671904462602137, | |
| "grad_norm": 2.586374044418335, | |
| "learning_rate": 9.803769708589352e-06, | |
| "loss": 1.0511, | |
| "step": 7980 | |
| }, | |
| { | |
| "epoch": 1.6750471401634193, | |
| "grad_norm": 2.192415475845337, | |
| "learning_rate": 9.767211145390827e-06, | |
| "loss": 1.1149, | |
| "step": 7995 | |
| }, | |
| { | |
| "epoch": 1.6781898177247014, | |
| "grad_norm": 2.12808895111084, | |
| "learning_rate": 9.73065569491409e-06, | |
| "loss": 1.1015, | |
| "step": 8010 | |
| }, | |
| { | |
| "epoch": 1.6813324952859836, | |
| "grad_norm": 2.3615734577178955, | |
| "learning_rate": 9.694103845958095e-06, | |
| "loss": 1.126, | |
| "step": 8025 | |
| }, | |
| { | |
| "epoch": 1.6844751728472658, | |
| "grad_norm": 1.6970425844192505, | |
| "learning_rate": 9.657556087273632e-06, | |
| "loss": 1.0458, | |
| "step": 8040 | |
| }, | |
| { | |
| "epoch": 1.687617850408548, | |
| "grad_norm": 2.4654476642608643, | |
| "learning_rate": 9.621012907556793e-06, | |
| "loss": 1.1316, | |
| "step": 8055 | |
| }, | |
| { | |
| "epoch": 1.6907605279698303, | |
| "grad_norm": 2.2715861797332764, | |
| "learning_rate": 9.584474795442454e-06, | |
| "loss": 1.0997, | |
| "step": 8070 | |
| }, | |
| { | |
| "epoch": 1.6939032055311125, | |
| "grad_norm": 2.117677688598633, | |
| "learning_rate": 9.547942239497716e-06, | |
| "loss": 1.0896, | |
| "step": 8085 | |
| }, | |
| { | |
| "epoch": 1.6970458830923947, | |
| "grad_norm": 1.9031733274459839, | |
| "learning_rate": 9.511415728215407e-06, | |
| "loss": 1.0468, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 1.700188560653677, | |
| "grad_norm": 3.147984504699707, | |
| "learning_rate": 9.474895750007508e-06, | |
| "loss": 1.0695, | |
| "step": 8115 | |
| }, | |
| { | |
| "epoch": 1.7033312382149592, | |
| "grad_norm": 2.6544809341430664, | |
| "learning_rate": 9.438382793198655e-06, | |
| "loss": 1.1083, | |
| "step": 8130 | |
| }, | |
| { | |
| "epoch": 1.7064739157762414, | |
| "grad_norm": 2.071148633956909, | |
| "learning_rate": 9.401877346019587e-06, | |
| "loss": 1.1407, | |
| "step": 8145 | |
| }, | |
| { | |
| "epoch": 1.7096165933375236, | |
| "grad_norm": 2.063387155532837, | |
| "learning_rate": 9.36537989660065e-06, | |
| "loss": 1.0749, | |
| "step": 8160 | |
| }, | |
| { | |
| "epoch": 1.7127592708988058, | |
| "grad_norm": 1.9773895740509033, | |
| "learning_rate": 9.328890932965225e-06, | |
| "loss": 1.0723, | |
| "step": 8175 | |
| }, | |
| { | |
| "epoch": 1.715901948460088, | |
| "grad_norm": 3.0571677684783936, | |
| "learning_rate": 9.29241094302324e-06, | |
| "loss": 1.1296, | |
| "step": 8190 | |
| }, | |
| { | |
| "epoch": 1.71904462602137, | |
| "grad_norm": 2.203171730041504, | |
| "learning_rate": 9.255940414564621e-06, | |
| "loss": 1.1401, | |
| "step": 8205 | |
| }, | |
| { | |
| "epoch": 1.7221873035826523, | |
| "grad_norm": 1.9974406957626343, | |
| "learning_rate": 9.21947983525279e-06, | |
| "loss": 1.1158, | |
| "step": 8220 | |
| }, | |
| { | |
| "epoch": 1.7253299811439347, | |
| "grad_norm": 2.110973834991455, | |
| "learning_rate": 9.183029692618134e-06, | |
| "loss": 1.0684, | |
| "step": 8235 | |
| }, | |
| { | |
| "epoch": 1.7284726587052168, | |
| "grad_norm": 1.8630388975143433, | |
| "learning_rate": 9.146590474051483e-06, | |
| "loss": 1.0376, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 1.7316153362664992, | |
| "grad_norm": 2.1341989040374756, | |
| "learning_rate": 9.110162666797595e-06, | |
| "loss": 1.0955, | |
| "step": 8265 | |
| }, | |
| { | |
| "epoch": 1.7347580138277814, | |
| "grad_norm": 2.5358388423919678, | |
| "learning_rate": 9.076174105269065e-06, | |
| "loss": 1.1651, | |
| "step": 8280 | |
| }, | |
| { | |
| "epoch": 1.7379006913890636, | |
| "grad_norm": 2.8575525283813477, | |
| "learning_rate": 9.039769740923183e-06, | |
| "loss": 1.1082, | |
| "step": 8295 | |
| }, | |
| { | |
| "epoch": 1.7410433689503457, | |
| "grad_norm": 2.287961959838867, | |
| "learning_rate": 9.003378216236902e-06, | |
| "loss": 1.1151, | |
| "step": 8310 | |
| }, | |
| { | |
| "epoch": 1.744186046511628, | |
| "grad_norm": 1.2058899402618408, | |
| "learning_rate": 8.967000017817247e-06, | |
| "loss": 1.0793, | |
| "step": 8325 | |
| }, | |
| { | |
| "epoch": 1.74732872407291, | |
| "grad_norm": 2.0018842220306396, | |
| "learning_rate": 8.930635632093046e-06, | |
| "loss": 1.0819, | |
| "step": 8340 | |
| }, | |
| { | |
| "epoch": 1.7504714016341922, | |
| "grad_norm": 1.9017311334609985, | |
| "learning_rate": 8.894285545308437e-06, | |
| "loss": 1.0658, | |
| "step": 8355 | |
| }, | |
| { | |
| "epoch": 1.7536140791954744, | |
| "grad_norm": 1.9596132040023804, | |
| "learning_rate": 8.857950243516364e-06, | |
| "loss": 1.1065, | |
| "step": 8370 | |
| }, | |
| { | |
| "epoch": 1.7567567567567568, | |
| "grad_norm": 1.979177713394165, | |
| "learning_rate": 8.821630212572074e-06, | |
| "loss": 1.0362, | |
| "step": 8385 | |
| }, | |
| { | |
| "epoch": 1.759899434318039, | |
| "grad_norm": 2.1770269870758057, | |
| "learning_rate": 8.785325938126608e-06, | |
| "loss": 1.0974, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 1.7630421118793211, | |
| "grad_norm": 2.125962734222412, | |
| "learning_rate": 8.749037905620334e-06, | |
| "loss": 1.1541, | |
| "step": 8415 | |
| }, | |
| { | |
| "epoch": 1.7661847894406035, | |
| "grad_norm": 2.0153768062591553, | |
| "learning_rate": 8.712766600276435e-06, | |
| "loss": 1.1233, | |
| "step": 8430 | |
| }, | |
| { | |
| "epoch": 1.7693274670018857, | |
| "grad_norm": 2.15311336517334, | |
| "learning_rate": 8.676512507094438e-06, | |
| "loss": 1.0341, | |
| "step": 8445 | |
| }, | |
| { | |
| "epoch": 1.7724701445631679, | |
| "grad_norm": 2.8032867908477783, | |
| "learning_rate": 8.640276110843702e-06, | |
| "loss": 1.0944, | |
| "step": 8460 | |
| }, | |
| { | |
| "epoch": 1.77561282212445, | |
| "grad_norm": 2.2355639934539795, | |
| "learning_rate": 8.604057896056965e-06, | |
| "loss": 1.1147, | |
| "step": 8475 | |
| }, | |
| { | |
| "epoch": 1.7787554996857322, | |
| "grad_norm": 2.222898006439209, | |
| "learning_rate": 8.567858347023843e-06, | |
| "loss": 1.1063, | |
| "step": 8490 | |
| }, | |
| { | |
| "epoch": 1.7818981772470144, | |
| "grad_norm": 2.043031692504883, | |
| "learning_rate": 8.531677947784383e-06, | |
| "loss": 1.1142, | |
| "step": 8505 | |
| }, | |
| { | |
| "epoch": 1.7850408548082966, | |
| "grad_norm": 2.3380000591278076, | |
| "learning_rate": 8.495517182122546e-06, | |
| "loss": 1.1379, | |
| "step": 8520 | |
| }, | |
| { | |
| "epoch": 1.7881835323695787, | |
| "grad_norm": 1.9827522039413452, | |
| "learning_rate": 8.459376533559784e-06, | |
| "loss": 1.1067, | |
| "step": 8535 | |
| }, | |
| { | |
| "epoch": 1.7913262099308611, | |
| "grad_norm": 2.2196788787841797, | |
| "learning_rate": 8.423256485348543e-06, | |
| "loss": 1.0563, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 1.7944688874921433, | |
| "grad_norm": 2.038891315460205, | |
| "learning_rate": 8.387157520465816e-06, | |
| "loss": 1.1151, | |
| "step": 8565 | |
| }, | |
| { | |
| "epoch": 1.7976115650534257, | |
| "grad_norm": 2.597698211669922, | |
| "learning_rate": 8.351080121606685e-06, | |
| "loss": 1.1025, | |
| "step": 8580 | |
| }, | |
| { | |
| "epoch": 1.8007542426147078, | |
| "grad_norm": 2.8073418140411377, | |
| "learning_rate": 8.315024771177862e-06, | |
| "loss": 1.136, | |
| "step": 8595 | |
| }, | |
| { | |
| "epoch": 1.80389692017599, | |
| "grad_norm": 2.143312692642212, | |
| "learning_rate": 8.278991951291234e-06, | |
| "loss": 1.0655, | |
| "step": 8610 | |
| }, | |
| { | |
| "epoch": 1.8070395977372722, | |
| "grad_norm": 1.5993940830230713, | |
| "learning_rate": 8.242982143757429e-06, | |
| "loss": 1.1023, | |
| "step": 8625 | |
| }, | |
| { | |
| "epoch": 1.8101822752985544, | |
| "grad_norm": 2.127690553665161, | |
| "learning_rate": 8.20699583007936e-06, | |
| "loss": 1.104, | |
| "step": 8640 | |
| }, | |
| { | |
| "epoch": 1.8133249528598365, | |
| "grad_norm": 2.57829213142395, | |
| "learning_rate": 8.171033491445806e-06, | |
| "loss": 1.0658, | |
| "step": 8655 | |
| }, | |
| { | |
| "epoch": 1.8164676304211187, | |
| "grad_norm": 2.4844679832458496, | |
| "learning_rate": 8.13509560872495e-06, | |
| "loss": 1.1207, | |
| "step": 8670 | |
| }, | |
| { | |
| "epoch": 1.8196103079824009, | |
| "grad_norm": 1.7117849588394165, | |
| "learning_rate": 8.099182662457972e-06, | |
| "loss": 1.1019, | |
| "step": 8685 | |
| }, | |
| { | |
| "epoch": 1.8227529855436833, | |
| "grad_norm": 1.6486339569091797, | |
| "learning_rate": 8.063295132852616e-06, | |
| "loss": 1.1566, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 1.8258956631049654, | |
| "grad_norm": 2.264791965484619, | |
| "learning_rate": 8.027433499776768e-06, | |
| "loss": 1.1269, | |
| "step": 8715 | |
| }, | |
| { | |
| "epoch": 1.8290383406662476, | |
| "grad_norm": 2.0826590061187744, | |
| "learning_rate": 7.99159824275204e-06, | |
| "loss": 1.1668, | |
| "step": 8730 | |
| }, | |
| { | |
| "epoch": 1.83218101822753, | |
| "grad_norm": 3.4813215732574463, | |
| "learning_rate": 7.955789840947357e-06, | |
| "loss": 1.0613, | |
| "step": 8745 | |
| }, | |
| { | |
| "epoch": 1.8353236957888122, | |
| "grad_norm": 1.950697898864746, | |
| "learning_rate": 7.920008773172549e-06, | |
| "loss": 1.1151, | |
| "step": 8760 | |
| }, | |
| { | |
| "epoch": 1.8384663733500943, | |
| "grad_norm": 2.761364459991455, | |
| "learning_rate": 7.884255517871955e-06, | |
| "loss": 1.0418, | |
| "step": 8775 | |
| }, | |
| { | |
| "epoch": 1.8416090509113765, | |
| "grad_norm": 2.035162925720215, | |
| "learning_rate": 7.848530553118024e-06, | |
| "loss": 1.1395, | |
| "step": 8790 | |
| }, | |
| { | |
| "epoch": 1.8447517284726587, | |
| "grad_norm": 2.061817169189453, | |
| "learning_rate": 7.812834356604913e-06, | |
| "loss": 1.1337, | |
| "step": 8805 | |
| }, | |
| { | |
| "epoch": 1.8478944060339408, | |
| "grad_norm": 2.214123010635376, | |
| "learning_rate": 7.777167405642109e-06, | |
| "loss": 1.09, | |
| "step": 8820 | |
| }, | |
| { | |
| "epoch": 1.851037083595223, | |
| "grad_norm": 2.1376426219940186, | |
| "learning_rate": 7.741530177148041e-06, | |
| "loss": 1.0908, | |
| "step": 8835 | |
| }, | |
| { | |
| "epoch": 1.8541797611565052, | |
| "grad_norm": 1.6116961240768433, | |
| "learning_rate": 7.70592314764371e-06, | |
| "loss": 1.1011, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 1.8573224387177876, | |
| "grad_norm": 2.253972291946411, | |
| "learning_rate": 7.670346793246317e-06, | |
| "loss": 1.0856, | |
| "step": 8865 | |
| }, | |
| { | |
| "epoch": 1.8604651162790697, | |
| "grad_norm": 1.9321953058242798, | |
| "learning_rate": 7.634801589662882e-06, | |
| "loss": 1.0851, | |
| "step": 8880 | |
| }, | |
| { | |
| "epoch": 1.8636077938403521, | |
| "grad_norm": 2.758145332336426, | |
| "learning_rate": 7.599288012183897e-06, | |
| "loss": 1.0952, | |
| "step": 8895 | |
| }, | |
| { | |
| "epoch": 1.8667504714016343, | |
| "grad_norm": 3.209368944168091, | |
| "learning_rate": 7.563806535676974e-06, | |
| "loss": 1.1081, | |
| "step": 8910 | |
| }, | |
| { | |
| "epoch": 1.8698931489629165, | |
| "grad_norm": 3.290693998336792, | |
| "learning_rate": 7.528357634580486e-06, | |
| "loss": 1.0403, | |
| "step": 8925 | |
| }, | |
| { | |
| "epoch": 1.8730358265241986, | |
| "grad_norm": 2.0125486850738525, | |
| "learning_rate": 7.4929417828972205e-06, | |
| "loss": 1.0801, | |
| "step": 8940 | |
| }, | |
| { | |
| "epoch": 1.8761785040854808, | |
| "grad_norm": 2.3519296646118164, | |
| "learning_rate": 7.457559454188052e-06, | |
| "loss": 1.1264, | |
| "step": 8955 | |
| }, | |
| { | |
| "epoch": 1.879321181646763, | |
| "grad_norm": 2.394660472869873, | |
| "learning_rate": 7.422211121565607e-06, | |
| "loss": 1.1271, | |
| "step": 8970 | |
| }, | |
| { | |
| "epoch": 1.8824638592080452, | |
| "grad_norm": 2.127760648727417, | |
| "learning_rate": 7.386897257687924e-06, | |
| "loss": 1.1308, | |
| "step": 8985 | |
| }, | |
| { | |
| "epoch": 1.8856065367693273, | |
| "grad_norm": 2.5212738513946533, | |
| "learning_rate": 7.3516183347521575e-06, | |
| "loss": 1.0758, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 1.8887492143306097, | |
| "grad_norm": 2.113893508911133, | |
| "learning_rate": 7.316374824488247e-06, | |
| "loss": 1.1463, | |
| "step": 9015 | |
| }, | |
| { | |
| "epoch": 1.8918918918918919, | |
| "grad_norm": 1.9831585884094238, | |
| "learning_rate": 7.281167198152615e-06, | |
| "loss": 1.1409, | |
| "step": 9030 | |
| }, | |
| { | |
| "epoch": 1.895034569453174, | |
| "grad_norm": 2.290170192718506, | |
| "learning_rate": 7.245995926521854e-06, | |
| "loss": 1.1156, | |
| "step": 9045 | |
| }, | |
| { | |
| "epoch": 1.8981772470144564, | |
| "grad_norm": 2.2289717197418213, | |
| "learning_rate": 7.2108614798864585e-06, | |
| "loss": 1.0908, | |
| "step": 9060 | |
| }, | |
| { | |
| "epoch": 1.9013199245757386, | |
| "grad_norm": 2.7847399711608887, | |
| "learning_rate": 7.17576432804451e-06, | |
| "loss": 1.1137, | |
| "step": 9075 | |
| }, | |
| { | |
| "epoch": 1.9044626021370208, | |
| "grad_norm": 2.2532174587249756, | |
| "learning_rate": 7.140704940295404e-06, | |
| "loss": 1.0993, | |
| "step": 9090 | |
| }, | |
| { | |
| "epoch": 1.907605279698303, | |
| "grad_norm": 3.097644329071045, | |
| "learning_rate": 7.1056837854335804e-06, | |
| "loss": 1.1322, | |
| "step": 9105 | |
| }, | |
| { | |
| "epoch": 1.9107479572595851, | |
| "grad_norm": 2.270334005355835, | |
| "learning_rate": 7.0707013317422446e-06, | |
| "loss": 1.0477, | |
| "step": 9120 | |
| }, | |
| { | |
| "epoch": 1.9138906348208673, | |
| "grad_norm": 2.698101282119751, | |
| "learning_rate": 7.035758046987122e-06, | |
| "loss": 1.0722, | |
| "step": 9135 | |
| }, | |
| { | |
| "epoch": 1.9170333123821495, | |
| "grad_norm": 2.2851641178131104, | |
| "learning_rate": 7.000854398410182e-06, | |
| "loss": 1.1588, | |
| "step": 9150 | |
| }, | |
| { | |
| "epoch": 1.9201759899434316, | |
| "grad_norm": 2.714940071105957, | |
| "learning_rate": 6.965990852723404e-06, | |
| "loss": 1.0664, | |
| "step": 9165 | |
| }, | |
| { | |
| "epoch": 1.923318667504714, | |
| "grad_norm": 2.079648971557617, | |
| "learning_rate": 6.9311678761025416e-06, | |
| "loss": 1.0701, | |
| "step": 9180 | |
| }, | |
| { | |
| "epoch": 1.9264613450659962, | |
| "grad_norm": 1.5704066753387451, | |
| "learning_rate": 6.896385934180867e-06, | |
| "loss": 1.085, | |
| "step": 9195 | |
| }, | |
| { | |
| "epoch": 1.9296040226272786, | |
| "grad_norm": 2.2583060264587402, | |
| "learning_rate": 6.861645492042977e-06, | |
| "loss": 1.1032, | |
| "step": 9210 | |
| }, | |
| { | |
| "epoch": 1.9327467001885608, | |
| "grad_norm": 2.0253028869628906, | |
| "learning_rate": 6.826947014218543e-06, | |
| "loss": 1.1232, | |
| "step": 9225 | |
| }, | |
| { | |
| "epoch": 1.935889377749843, | |
| "grad_norm": 3.0294480323791504, | |
| "learning_rate": 6.792290964676123e-06, | |
| "loss": 1.1583, | |
| "step": 9240 | |
| }, | |
| { | |
| "epoch": 1.939032055311125, | |
| "grad_norm": 2.480926036834717, | |
| "learning_rate": 6.7576778068169335e-06, | |
| "loss": 1.126, | |
| "step": 9255 | |
| }, | |
| { | |
| "epoch": 1.9421747328724073, | |
| "grad_norm": 2.393428087234497, | |
| "learning_rate": 6.725411298947324e-06, | |
| "loss": 1.0878, | |
| "step": 9270 | |
| }, | |
| { | |
| "epoch": 1.9453174104336894, | |
| "grad_norm": 2.0476040840148926, | |
| "learning_rate": 6.690882376871407e-06, | |
| "loss": 1.0794, | |
| "step": 9285 | |
| }, | |
| { | |
| "epoch": 1.9484600879949716, | |
| "grad_norm": 2.075193166732788, | |
| "learning_rate": 6.656397702457396e-06, | |
| "loss": 1.0825, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 1.9516027655562538, | |
| "grad_norm": 2.0196661949157715, | |
| "learning_rate": 6.621957736814983e-06, | |
| "loss": 1.0755, | |
| "step": 9315 | |
| }, | |
| { | |
| "epoch": 1.9547454431175362, | |
| "grad_norm": 2.4812753200531006, | |
| "learning_rate": 6.58756294045604e-06, | |
| "loss": 1.0883, | |
| "step": 9330 | |
| }, | |
| { | |
| "epoch": 1.9578881206788183, | |
| "grad_norm": 2.0828120708465576, | |
| "learning_rate": 6.553213773288465e-06, | |
| "loss": 1.1287, | |
| "step": 9345 | |
| }, | |
| { | |
| "epoch": 1.9610307982401005, | |
| "grad_norm": 2.273803949356079, | |
| "learning_rate": 6.518910694610023e-06, | |
| "loss": 1.0481, | |
| "step": 9360 | |
| }, | |
| { | |
| "epoch": 1.964173475801383, | |
| "grad_norm": 3.1020278930664062, | |
| "learning_rate": 6.4846541631022155e-06, | |
| "loss": 1.078, | |
| "step": 9375 | |
| }, | |
| { | |
| "epoch": 1.967316153362665, | |
| "grad_norm": 2.7231500148773193, | |
| "learning_rate": 6.4504446368241315e-06, | |
| "loss": 1.0389, | |
| "step": 9390 | |
| }, | |
| { | |
| "epoch": 1.9704588309239472, | |
| "grad_norm": 2.4498469829559326, | |
| "learning_rate": 6.416282573206341e-06, | |
| "loss": 1.1052, | |
| "step": 9405 | |
| }, | |
| { | |
| "epoch": 1.9736015084852294, | |
| "grad_norm": 2.266944408416748, | |
| "learning_rate": 6.382168429044769e-06, | |
| "loss": 1.1163, | |
| "step": 9420 | |
| }, | |
| { | |
| "epoch": 1.9767441860465116, | |
| "grad_norm": 2.877861976623535, | |
| "learning_rate": 6.34810266049459e-06, | |
| "loss": 1.1106, | |
| "step": 9435 | |
| }, | |
| { | |
| "epoch": 1.9798868636077938, | |
| "grad_norm": 2.256692886352539, | |
| "learning_rate": 6.314085723064118e-06, | |
| "loss": 1.0929, | |
| "step": 9450 | |
| }, | |
| { | |
| "epoch": 1.983029541169076, | |
| "grad_norm": 2.337369203567505, | |
| "learning_rate": 6.2801180716087315e-06, | |
| "loss": 1.1019, | |
| "step": 9465 | |
| }, | |
| { | |
| "epoch": 1.9861722187303583, | |
| "grad_norm": 3.803745746612549, | |
| "learning_rate": 6.246200160324789e-06, | |
| "loss": 1.0641, | |
| "step": 9480 | |
| }, | |
| { | |
| "epoch": 1.9893148962916405, | |
| "grad_norm": 2.354900598526001, | |
| "learning_rate": 6.21233244274354e-06, | |
| "loss": 1.1353, | |
| "step": 9495 | |
| }, | |
| { | |
| "epoch": 1.9924575738529227, | |
| "grad_norm": 2.0362517833709717, | |
| "learning_rate": 6.178515371725083e-06, | |
| "loss": 1.0638, | |
| "step": 9510 | |
| }, | |
| { | |
| "epoch": 1.995600251414205, | |
| "grad_norm": 1.7420244216918945, | |
| "learning_rate": 6.144749399452294e-06, | |
| "loss": 1.0539, | |
| "step": 9525 | |
| }, | |
| { | |
| "epoch": 1.9987429289754872, | |
| "grad_norm": 1.9186304807662964, | |
| "learning_rate": 6.11103497742478e-06, | |
| "loss": 1.0553, | |
| "step": 9540 | |
| }, | |
| { | |
| "epoch": 2.0018856065367694, | |
| "grad_norm": 1.9180148839950562, | |
| "learning_rate": 6.077372556452858e-06, | |
| "loss": 1.078, | |
| "step": 9555 | |
| }, | |
| { | |
| "epoch": 2.0050282840980516, | |
| "grad_norm": 1.8619567155838013, | |
| "learning_rate": 6.043762586651511e-06, | |
| "loss": 1.0238, | |
| "step": 9570 | |
| }, | |
| { | |
| "epoch": 2.0081709616593337, | |
| "grad_norm": 2.054584503173828, | |
| "learning_rate": 6.010205517434373e-06, | |
| "loss": 1.0361, | |
| "step": 9585 | |
| }, | |
| { | |
| "epoch": 2.011313639220616, | |
| "grad_norm": 1.6725273132324219, | |
| "learning_rate": 5.97670179750772e-06, | |
| "loss": 1.0048, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 2.014456316781898, | |
| "grad_norm": 2.8712847232818604, | |
| "learning_rate": 5.943251874864476e-06, | |
| "loss": 1.0199, | |
| "step": 9615 | |
| }, | |
| { | |
| "epoch": 2.0175989943431802, | |
| "grad_norm": 2.331249237060547, | |
| "learning_rate": 5.9098561967782165e-06, | |
| "loss": 1.0697, | |
| "step": 9630 | |
| }, | |
| { | |
| "epoch": 2.0207416719044624, | |
| "grad_norm": 2.359811782836914, | |
| "learning_rate": 5.876515209797189e-06, | |
| "loss": 1.0052, | |
| "step": 9645 | |
| }, | |
| { | |
| "epoch": 2.023884349465745, | |
| "grad_norm": 3.5132248401641846, | |
| "learning_rate": 5.843229359738336e-06, | |
| "loss": 1.0845, | |
| "step": 9660 | |
| }, | |
| { | |
| "epoch": 2.027027027027027, | |
| "grad_norm": 1.9716084003448486, | |
| "learning_rate": 5.809999091681349e-06, | |
| "loss": 1.1177, | |
| "step": 9675 | |
| }, | |
| { | |
| "epoch": 2.0301697045883094, | |
| "grad_norm": 2.3846986293792725, | |
| "learning_rate": 5.776824849962706e-06, | |
| "loss": 1.0848, | |
| "step": 9690 | |
| }, | |
| { | |
| "epoch": 2.0333123821495915, | |
| "grad_norm": 1.9413962364196777, | |
| "learning_rate": 5.743707078169731e-06, | |
| "loss": 1.0824, | |
| "step": 9705 | |
| }, | |
| { | |
| "epoch": 2.0364550597108737, | |
| "grad_norm": 1.9300832748413086, | |
| "learning_rate": 5.710646219134666e-06, | |
| "loss": 1.0474, | |
| "step": 9720 | |
| }, | |
| { | |
| "epoch": 2.039597737272156, | |
| "grad_norm": 2.1981208324432373, | |
| "learning_rate": 5.67764271492874e-06, | |
| "loss": 1.1053, | |
| "step": 9735 | |
| }, | |
| { | |
| "epoch": 2.042740414833438, | |
| "grad_norm": 2.165830612182617, | |
| "learning_rate": 5.644697006856268e-06, | |
| "loss": 1.0203, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 2.04588309239472, | |
| "grad_norm": 2.567518711090088, | |
| "learning_rate": 5.61180953544876e-06, | |
| "loss": 1.0509, | |
| "step": 9765 | |
| }, | |
| { | |
| "epoch": 2.0490257699560024, | |
| "grad_norm": 2.09709095954895, | |
| "learning_rate": 5.578980740459009e-06, | |
| "loss": 1.0906, | |
| "step": 9780 | |
| }, | |
| { | |
| "epoch": 2.0521684475172846, | |
| "grad_norm": 1.5784721374511719, | |
| "learning_rate": 5.546211060855211e-06, | |
| "loss": 1.0694, | |
| "step": 9795 | |
| }, | |
| { | |
| "epoch": 2.0553111250785667, | |
| "grad_norm": 3.4078545570373535, | |
| "learning_rate": 5.513500934815113e-06, | |
| "loss": 1.0352, | |
| "step": 9810 | |
| }, | |
| { | |
| "epoch": 2.0584538026398493, | |
| "grad_norm": 2.0774762630462646, | |
| "learning_rate": 5.4808507997201544e-06, | |
| "loss": 1.0513, | |
| "step": 9825 | |
| }, | |
| { | |
| "epoch": 2.0615964802011315, | |
| "grad_norm": 3.19948148727417, | |
| "learning_rate": 5.448261092149596e-06, | |
| "loss": 1.0432, | |
| "step": 9840 | |
| }, | |
| { | |
| "epoch": 2.0647391577624137, | |
| "grad_norm": 2.4168691635131836, | |
| "learning_rate": 5.415732247874696e-06, | |
| "loss": 1.0354, | |
| "step": 9855 | |
| }, | |
| { | |
| "epoch": 2.067881835323696, | |
| "grad_norm": 1.7410550117492676, | |
| "learning_rate": 5.383264701852886e-06, | |
| "loss": 1.0389, | |
| "step": 9870 | |
| }, | |
| { | |
| "epoch": 2.071024512884978, | |
| "grad_norm": 2.2156288623809814, | |
| "learning_rate": 5.350858888221957e-06, | |
| "loss": 1.0614, | |
| "step": 9885 | |
| }, | |
| { | |
| "epoch": 2.07416719044626, | |
| "grad_norm": 3.3456056118011475, | |
| "learning_rate": 5.318515240294241e-06, | |
| "loss": 1.099, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 2.0773098680075424, | |
| "grad_norm": 3.286379814147949, | |
| "learning_rate": 5.28623419055083e-06, | |
| "loss": 1.0667, | |
| "step": 9915 | |
| }, | |
| { | |
| "epoch": 2.0804525455688245, | |
| "grad_norm": 2.086221218109131, | |
| "learning_rate": 5.2540161706357855e-06, | |
| "loss": 1.0458, | |
| "step": 9930 | |
| }, | |
| { | |
| "epoch": 2.0835952231301067, | |
| "grad_norm": 3.3937273025512695, | |
| "learning_rate": 5.221861611350371e-06, | |
| "loss": 1.0696, | |
| "step": 9945 | |
| }, | |
| { | |
| "epoch": 2.086737900691389, | |
| "grad_norm": 2.162592887878418, | |
| "learning_rate": 5.189770942647291e-06, | |
| "loss": 1.034, | |
| "step": 9960 | |
| }, | |
| { | |
| "epoch": 2.0898805782526715, | |
| "grad_norm": 2.144193649291992, | |
| "learning_rate": 5.157744593624939e-06, | |
| "loss": 1.0997, | |
| "step": 9975 | |
| }, | |
| { | |
| "epoch": 2.0930232558139537, | |
| "grad_norm": 2.1846065521240234, | |
| "learning_rate": 5.125782992521664e-06, | |
| "loss": 1.034, | |
| "step": 9990 | |
| }, | |
| { | |
| "epoch": 2.0951183741881416, | |
| "eval_accuracy": 0.8874096274375916, | |
| "eval_loss": 1.1195236444473267, | |
| "eval_runtime": 3488.4257, | |
| "eval_samples_per_second": 1.372, | |
| "eval_steps_per_second": 0.343, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 2.096165933375236, | |
| "grad_norm": 1.9800641536712646, | |
| "learning_rate": 5.09388656671004e-06, | |
| "loss": 1.0096, | |
| "step": 10005 | |
| }, | |
| { | |
| "epoch": 2.099308610936518, | |
| "grad_norm": 2.919948101043701, | |
| "learning_rate": 5.062055742691154e-06, | |
| "loss": 1.0455, | |
| "step": 10020 | |
| }, | |
| { | |
| "epoch": 2.1024512884978, | |
| "grad_norm": 2.2051897048950195, | |
| "learning_rate": 5.0302909460889045e-06, | |
| "loss": 1.0675, | |
| "step": 10035 | |
| }, | |
| { | |
| "epoch": 2.1055939660590823, | |
| "grad_norm": 3.871448278427124, | |
| "learning_rate": 4.998592601644306e-06, | |
| "loss": 1.0813, | |
| "step": 10050 | |
| }, | |
| { | |
| "epoch": 2.1087366436203645, | |
| "grad_norm": 2.2782132625579834, | |
| "learning_rate": 4.9669611332098155e-06, | |
| "loss": 1.0833, | |
| "step": 10065 | |
| }, | |
| { | |
| "epoch": 2.1118793211816467, | |
| "grad_norm": 2.384378433227539, | |
| "learning_rate": 4.935396963743661e-06, | |
| "loss": 1.1207, | |
| "step": 10080 | |
| }, | |
| { | |
| "epoch": 2.115021998742929, | |
| "grad_norm": 2.561387300491333, | |
| "learning_rate": 4.9039005153041755e-06, | |
| "loss": 1.0639, | |
| "step": 10095 | |
| }, | |
| { | |
| "epoch": 2.118164676304211, | |
| "grad_norm": 3.0452916622161865, | |
| "learning_rate": 4.872472209044182e-06, | |
| "loss": 1.0579, | |
| "step": 10110 | |
| }, | |
| { | |
| "epoch": 2.121307353865493, | |
| "grad_norm": 2.2836525440216064, | |
| "learning_rate": 4.841112465205337e-06, | |
| "loss": 1.0984, | |
| "step": 10125 | |
| }, | |
| { | |
| "epoch": 2.124450031426776, | |
| "grad_norm": 3.4332149028778076, | |
| "learning_rate": 4.809821703112518e-06, | |
| "loss": 1.0762, | |
| "step": 10140 | |
| }, | |
| { | |
| "epoch": 2.127592708988058, | |
| "grad_norm": 2.4914424419403076, | |
| "learning_rate": 4.778600341168208e-06, | |
| "loss": 1.0371, | |
| "step": 10155 | |
| }, | |
| { | |
| "epoch": 2.13073538654934, | |
| "grad_norm": 3.004932165145874, | |
| "learning_rate": 4.747448796846932e-06, | |
| "loss": 1.084, | |
| "step": 10170 | |
| }, | |
| { | |
| "epoch": 2.1338780641106223, | |
| "grad_norm": 2.9469964504241943, | |
| "learning_rate": 4.716367486689636e-06, | |
| "loss": 1.0769, | |
| "step": 10185 | |
| }, | |
| { | |
| "epoch": 2.1370207416719045, | |
| "grad_norm": 1.8354699611663818, | |
| "learning_rate": 4.685356826298143e-06, | |
| "loss": 1.0308, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 2.1401634192331866, | |
| "grad_norm": 1.9137917757034302, | |
| "learning_rate": 4.654417230329575e-06, | |
| "loss": 1.0568, | |
| "step": 10215 | |
| }, | |
| { | |
| "epoch": 2.143306096794469, | |
| "grad_norm": 2.0806703567504883, | |
| "learning_rate": 4.6235491124908295e-06, | |
| "loss": 1.0535, | |
| "step": 10230 | |
| }, | |
| { | |
| "epoch": 2.146448774355751, | |
| "grad_norm": 3.0409634113311768, | |
| "learning_rate": 4.5927528855330425e-06, | |
| "loss": 1.0421, | |
| "step": 10245 | |
| }, | |
| { | |
| "epoch": 2.149591451917033, | |
| "grad_norm": 2.494769811630249, | |
| "learning_rate": 4.562028961246065e-06, | |
| "loss": 1.0494, | |
| "step": 10260 | |
| }, | |
| { | |
| "epoch": 2.1527341294783153, | |
| "grad_norm": 2.0933990478515625, | |
| "learning_rate": 4.5334188940857585e-06, | |
| "loss": 1.1007, | |
| "step": 10275 | |
| }, | |
| { | |
| "epoch": 2.155876807039598, | |
| "grad_norm": 2.1855649948120117, | |
| "learning_rate": 4.502835919016285e-06, | |
| "loss": 1.1062, | |
| "step": 10290 | |
| }, | |
| { | |
| "epoch": 2.15901948460088, | |
| "grad_norm": 1.9125105142593384, | |
| "learning_rate": 4.472326448936779e-06, | |
| "loss": 1.0911, | |
| "step": 10305 | |
| }, | |
| { | |
| "epoch": 2.1621621621621623, | |
| "grad_norm": 2.457319736480713, | |
| "learning_rate": 4.441890891802735e-06, | |
| "loss": 1.0553, | |
| "step": 10320 | |
| }, | |
| { | |
| "epoch": 2.1653048397234445, | |
| "grad_norm": 1.8059813976287842, | |
| "learning_rate": 4.411529654581325e-06, | |
| "loss": 1.0426, | |
| "step": 10335 | |
| }, | |
| { | |
| "epoch": 2.1684475172847266, | |
| "grad_norm": 1.840040922164917, | |
| "learning_rate": 4.381243143245957e-06, | |
| "loss": 1.0701, | |
| "step": 10350 | |
| }, | |
| { | |
| "epoch": 2.171590194846009, | |
| "grad_norm": 2.875662326812744, | |
| "learning_rate": 4.351031762770848e-06, | |
| "loss": 1.0217, | |
| "step": 10365 | |
| }, | |
| { | |
| "epoch": 2.174732872407291, | |
| "grad_norm": 2.7019827365875244, | |
| "learning_rate": 4.320895917125604e-06, | |
| "loss": 1.0767, | |
| "step": 10380 | |
| }, | |
| { | |
| "epoch": 2.177875549968573, | |
| "grad_norm": 2.331372022628784, | |
| "learning_rate": 4.290836009269824e-06, | |
| "loss": 1.0414, | |
| "step": 10395 | |
| }, | |
| { | |
| "epoch": 2.1810182275298553, | |
| "grad_norm": 1.7536449432373047, | |
| "learning_rate": 4.260852441147711e-06, | |
| "loss": 1.0328, | |
| "step": 10410 | |
| }, | |
| { | |
| "epoch": 2.1841609050911375, | |
| "grad_norm": 2.233278512954712, | |
| "learning_rate": 4.230945613682696e-06, | |
| "loss": 1.0464, | |
| "step": 10425 | |
| }, | |
| { | |
| "epoch": 2.1873035826524196, | |
| "grad_norm": 2.3253679275512695, | |
| "learning_rate": 4.2011159267720645e-06, | |
| "loss": 1.0734, | |
| "step": 10440 | |
| }, | |
| { | |
| "epoch": 2.1904462602137023, | |
| "grad_norm": 2.1976492404937744, | |
| "learning_rate": 4.171363779281644e-06, | |
| "loss": 1.0089, | |
| "step": 10455 | |
| }, | |
| { | |
| "epoch": 2.1935889377749844, | |
| "grad_norm": 1.6631523370742798, | |
| "learning_rate": 4.141689569040433e-06, | |
| "loss": 1.0598, | |
| "step": 10470 | |
| }, | |
| { | |
| "epoch": 2.1967316153362666, | |
| "grad_norm": 2.9074153900146484, | |
| "learning_rate": 4.112093692835306e-06, | |
| "loss": 1.0564, | |
| "step": 10485 | |
| }, | |
| { | |
| "epoch": 2.1998742928975488, | |
| "grad_norm": 2.0175225734710693, | |
| "learning_rate": 4.082576546405683e-06, | |
| "loss": 1.009, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 2.203016970458831, | |
| "grad_norm": 2.74937105178833, | |
| "learning_rate": 4.053138524438268e-06, | |
| "loss": 1.0614, | |
| "step": 10515 | |
| }, | |
| { | |
| "epoch": 2.206159648020113, | |
| "grad_norm": 1.9003492593765259, | |
| "learning_rate": 4.0237800205617585e-06, | |
| "loss": 1.0904, | |
| "step": 10530 | |
| }, | |
| { | |
| "epoch": 2.2093023255813953, | |
| "grad_norm": 2.6317954063415527, | |
| "learning_rate": 3.994501427341575e-06, | |
| "loss": 1.0716, | |
| "step": 10545 | |
| }, | |
| { | |
| "epoch": 2.2124450031426774, | |
| "grad_norm": 1.7894374132156372, | |
| "learning_rate": 3.965303136274614e-06, | |
| "loss": 1.1267, | |
| "step": 10560 | |
| }, | |
| { | |
| "epoch": 2.2155876807039596, | |
| "grad_norm": 2.2193076610565186, | |
| "learning_rate": 3.936185537784019e-06, | |
| "loss": 1.0509, | |
| "step": 10575 | |
| }, | |
| { | |
| "epoch": 2.218730358265242, | |
| "grad_norm": 2.435281753540039, | |
| "learning_rate": 3.9071490212139665e-06, | |
| "loss": 1.0699, | |
| "step": 10590 | |
| }, | |
| { | |
| "epoch": 2.2218730358265244, | |
| "grad_norm": 2.5646512508392334, | |
| "learning_rate": 3.878193974824444e-06, | |
| "loss": 1.0761, | |
| "step": 10605 | |
| }, | |
| { | |
| "epoch": 2.2250157133878066, | |
| "grad_norm": 2.3595800399780273, | |
| "learning_rate": 3.849320785786062e-06, | |
| "loss": 1.0842, | |
| "step": 10620 | |
| }, | |
| { | |
| "epoch": 2.2281583909490887, | |
| "grad_norm": 2.8445732593536377, | |
| "learning_rate": 3.820529840174885e-06, | |
| "loss": 1.0093, | |
| "step": 10635 | |
| }, | |
| { | |
| "epoch": 2.231301068510371, | |
| "grad_norm": 2.228774309158325, | |
| "learning_rate": 3.7918215229672606e-06, | |
| "loss": 1.1004, | |
| "step": 10650 | |
| }, | |
| { | |
| "epoch": 2.234443746071653, | |
| "grad_norm": 2.617530584335327, | |
| "learning_rate": 3.7631962180346905e-06, | |
| "loss": 1.0769, | |
| "step": 10665 | |
| }, | |
| { | |
| "epoch": 2.2375864236329353, | |
| "grad_norm": 2.4638450145721436, | |
| "learning_rate": 3.73465430813866e-06, | |
| "loss": 1.0703, | |
| "step": 10680 | |
| }, | |
| { | |
| "epoch": 2.2407291011942174, | |
| "grad_norm": 2.0117123126983643, | |
| "learning_rate": 3.706196174925558e-06, | |
| "loss": 1.0354, | |
| "step": 10695 | |
| }, | |
| { | |
| "epoch": 2.2438717787554996, | |
| "grad_norm": 2.4193451404571533, | |
| "learning_rate": 3.6778221989215537e-06, | |
| "loss": 1.0742, | |
| "step": 10710 | |
| }, | |
| { | |
| "epoch": 2.2470144563167818, | |
| "grad_norm": 2.2464051246643066, | |
| "learning_rate": 3.6495327595275223e-06, | |
| "loss": 1.0563, | |
| "step": 10725 | |
| }, | |
| { | |
| "epoch": 2.250157133878064, | |
| "grad_norm": 2.0514557361602783, | |
| "learning_rate": 3.6213282350139455e-06, | |
| "loss": 1.078, | |
| "step": 10740 | |
| }, | |
| { | |
| "epoch": 2.253299811439346, | |
| "grad_norm": 3.3520901203155518, | |
| "learning_rate": 3.593209002515885e-06, | |
| "loss": 1.0656, | |
| "step": 10755 | |
| }, | |
| { | |
| "epoch": 2.2564424890006287, | |
| "grad_norm": 1.7619084119796753, | |
| "learning_rate": 3.56517543802792e-06, | |
| "loss": 1.0591, | |
| "step": 10770 | |
| }, | |
| { | |
| "epoch": 2.259585166561911, | |
| "grad_norm": 2.095613718032837, | |
| "learning_rate": 3.5372279163991285e-06, | |
| "loss": 1.0567, | |
| "step": 10785 | |
| }, | |
| { | |
| "epoch": 2.262727844123193, | |
| "grad_norm": 2.4718666076660156, | |
| "learning_rate": 3.5093668113280688e-06, | |
| "loss": 1.0847, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 2.2658705216844752, | |
| "grad_norm": 3.0717573165893555, | |
| "learning_rate": 3.481592495357786e-06, | |
| "loss": 1.125, | |
| "step": 10815 | |
| }, | |
| { | |
| "epoch": 2.2690131992457574, | |
| "grad_norm": 2.8368752002716064, | |
| "learning_rate": 3.4539053398708313e-06, | |
| "loss": 1.0832, | |
| "step": 10830 | |
| }, | |
| { | |
| "epoch": 2.2721558768070396, | |
| "grad_norm": 2.2466132640838623, | |
| "learning_rate": 3.4263057150842927e-06, | |
| "loss": 1.0572, | |
| "step": 10845 | |
| }, | |
| { | |
| "epoch": 2.2752985543683217, | |
| "grad_norm": 2.3740978240966797, | |
| "learning_rate": 3.398793990044851e-06, | |
| "loss": 1.0886, | |
| "step": 10860 | |
| }, | |
| { | |
| "epoch": 2.278441231929604, | |
| "grad_norm": 2.0145745277404785, | |
| "learning_rate": 3.3713705326238354e-06, | |
| "loss": 0.9999, | |
| "step": 10875 | |
| }, | |
| { | |
| "epoch": 2.281583909490886, | |
| "grad_norm": 2.8063840866088867, | |
| "learning_rate": 3.3440357095123142e-06, | |
| "loss": 1.029, | |
| "step": 10890 | |
| }, | |
| { | |
| "epoch": 2.2847265870521687, | |
| "grad_norm": 2.4196486473083496, | |
| "learning_rate": 3.3167898862161896e-06, | |
| "loss": 1.0736, | |
| "step": 10905 | |
| }, | |
| { | |
| "epoch": 2.287869264613451, | |
| "grad_norm": 2.1821863651275635, | |
| "learning_rate": 3.289633427051295e-06, | |
| "loss": 1.0106, | |
| "step": 10920 | |
| }, | |
| { | |
| "epoch": 2.291011942174733, | |
| "grad_norm": 2.0182254314422607, | |
| "learning_rate": 3.262566695138557e-06, | |
| "loss": 1.0717, | |
| "step": 10935 | |
| }, | |
| { | |
| "epoch": 2.294154619736015, | |
| "grad_norm": 3.4760282039642334, | |
| "learning_rate": 3.2355900523991057e-06, | |
| "loss": 1.081, | |
| "step": 10950 | |
| }, | |
| { | |
| "epoch": 2.2972972972972974, | |
| "grad_norm": 1.5960103273391724, | |
| "learning_rate": 3.2087038595494567e-06, | |
| "loss": 0.9858, | |
| "step": 10965 | |
| }, | |
| { | |
| "epoch": 2.3004399748585795, | |
| "grad_norm": 2.3802037239074707, | |
| "learning_rate": 3.18190847609667e-06, | |
| "loss": 1.1179, | |
| "step": 10980 | |
| }, | |
| { | |
| "epoch": 2.3035826524198617, | |
| "grad_norm": 2.2108407020568848, | |
| "learning_rate": 3.155204260333561e-06, | |
| "loss": 1.0281, | |
| "step": 10995 | |
| }, | |
| { | |
| "epoch": 2.306725329981144, | |
| "grad_norm": 2.4052934646606445, | |
| "learning_rate": 3.1285915693339074e-06, | |
| "loss": 1.0728, | |
| "step": 11010 | |
| }, | |
| { | |
| "epoch": 2.309868007542426, | |
| "grad_norm": 2.564016342163086, | |
| "learning_rate": 3.1020707589476638e-06, | |
| "loss": 1.0539, | |
| "step": 11025 | |
| }, | |
| { | |
| "epoch": 2.313010685103708, | |
| "grad_norm": 1.9333407878875732, | |
| "learning_rate": 3.0756421837962013e-06, | |
| "loss": 1.0316, | |
| "step": 11040 | |
| }, | |
| { | |
| "epoch": 2.3161533626649904, | |
| "grad_norm": 2.317944049835205, | |
| "learning_rate": 3.0493061972675808e-06, | |
| "loss": 1.0568, | |
| "step": 11055 | |
| }, | |
| { | |
| "epoch": 2.3192960402262726, | |
| "grad_norm": 1.7732441425323486, | |
| "learning_rate": 3.0230631515118236e-06, | |
| "loss": 1.0107, | |
| "step": 11070 | |
| }, | |
| { | |
| "epoch": 2.322438717787555, | |
| "grad_norm": 2.109452724456787, | |
| "learning_rate": 2.9969133974361985e-06, | |
| "loss": 1.0543, | |
| "step": 11085 | |
| }, | |
| { | |
| "epoch": 2.3255813953488373, | |
| "grad_norm": 2.198519468307495, | |
| "learning_rate": 2.970857284700517e-06, | |
| "loss": 1.0958, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 2.3287240729101195, | |
| "grad_norm": 2.7568089962005615, | |
| "learning_rate": 2.9448951617124878e-06, | |
| "loss": 1.0141, | |
| "step": 11115 | |
| }, | |
| { | |
| "epoch": 2.3318667504714017, | |
| "grad_norm": 3.3935024738311768, | |
| "learning_rate": 2.919027375623028e-06, | |
| "loss": 1.0298, | |
| "step": 11130 | |
| }, | |
| { | |
| "epoch": 2.335009428032684, | |
| "grad_norm": 2.245539903640747, | |
| "learning_rate": 2.893254272321655e-06, | |
| "loss": 1.0677, | |
| "step": 11145 | |
| }, | |
| { | |
| "epoch": 2.338152105593966, | |
| "grad_norm": 3.133040428161621, | |
| "learning_rate": 2.8675761964318148e-06, | |
| "loss": 1.0253, | |
| "step": 11160 | |
| }, | |
| { | |
| "epoch": 2.341294783155248, | |
| "grad_norm": 2.1391761302948, | |
| "learning_rate": 2.8419934913063175e-06, | |
| "loss": 1.0297, | |
| "step": 11175 | |
| }, | |
| { | |
| "epoch": 2.3444374607165304, | |
| "grad_norm": 3.1094396114349365, | |
| "learning_rate": 2.8165064990227255e-06, | |
| "loss": 1.0369, | |
| "step": 11190 | |
| }, | |
| { | |
| "epoch": 2.3475801382778125, | |
| "grad_norm": 3.4849114418029785, | |
| "learning_rate": 2.791115560378781e-06, | |
| "loss": 1.0666, | |
| "step": 11205 | |
| }, | |
| { | |
| "epoch": 2.350722815839095, | |
| "grad_norm": 2.6266283988952637, | |
| "learning_rate": 2.7658210148878515e-06, | |
| "loss": 1.1017, | |
| "step": 11220 | |
| }, | |
| { | |
| "epoch": 2.3538654934003773, | |
| "grad_norm": 3.290055990219116, | |
| "learning_rate": 2.7406232007743892e-06, | |
| "loss": 1.017, | |
| "step": 11235 | |
| }, | |
| { | |
| "epoch": 2.3570081709616595, | |
| "grad_norm": 3.80057954788208, | |
| "learning_rate": 2.7155224549694114e-06, | |
| "loss": 1.0566, | |
| "step": 11250 | |
| }, | |
| { | |
| "epoch": 2.3601508485229417, | |
| "grad_norm": 2.253140926361084, | |
| "learning_rate": 2.690519113105988e-06, | |
| "loss": 1.1236, | |
| "step": 11265 | |
| }, | |
| { | |
| "epoch": 2.363293526084224, | |
| "grad_norm": 2.4676170349121094, | |
| "learning_rate": 2.667270835657341e-06, | |
| "loss": 1.0442, | |
| "step": 11280 | |
| }, | |
| { | |
| "epoch": 2.366436203645506, | |
| "grad_norm": 2.411430597305298, | |
| "learning_rate": 2.6424567549411838e-06, | |
| "loss": 1.0645, | |
| "step": 11295 | |
| }, | |
| { | |
| "epoch": 2.369578881206788, | |
| "grad_norm": 2.473489284515381, | |
| "learning_rate": 2.6177410551600802e-06, | |
| "loss": 1.1046, | |
| "step": 11310 | |
| }, | |
| { | |
| "epoch": 2.3727215587680703, | |
| "grad_norm": 2.323939800262451, | |
| "learning_rate": 2.593124066798477e-06, | |
| "loss": 1.0366, | |
| "step": 11325 | |
| }, | |
| { | |
| "epoch": 2.3758642363293525, | |
| "grad_norm": 2.2324061393737793, | |
| "learning_rate": 2.5686061190209134e-06, | |
| "loss": 1.036, | |
| "step": 11340 | |
| }, | |
| { | |
| "epoch": 2.3790069138906347, | |
| "grad_norm": 3.393950939178467, | |
| "learning_rate": 2.5441875396676186e-06, | |
| "loss": 1.0795, | |
| "step": 11355 | |
| }, | |
| { | |
| "epoch": 2.382149591451917, | |
| "grad_norm": 2.0376970767974854, | |
| "learning_rate": 2.5198686552501196e-06, | |
| "loss": 1.0682, | |
| "step": 11370 | |
| }, | |
| { | |
| "epoch": 2.385292269013199, | |
| "grad_norm": 2.5424187183380127, | |
| "learning_rate": 2.4956497909468713e-06, | |
| "loss": 1.0599, | |
| "step": 11385 | |
| }, | |
| { | |
| "epoch": 2.3884349465744816, | |
| "grad_norm": 2.3934738636016846, | |
| "learning_rate": 2.4715312705989236e-06, | |
| "loss": 1.0725, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 2.391577624135764, | |
| "grad_norm": 2.5349321365356445, | |
| "learning_rate": 2.447513416705579e-06, | |
| "loss": 0.977, | |
| "step": 11415 | |
| }, | |
| { | |
| "epoch": 2.394720301697046, | |
| "grad_norm": 2.4864916801452637, | |
| "learning_rate": 2.4235965504200974e-06, | |
| "loss": 1.0452, | |
| "step": 11430 | |
| }, | |
| { | |
| "epoch": 2.397862979258328, | |
| "grad_norm": 2.2827324867248535, | |
| "learning_rate": 2.3997809915453686e-06, | |
| "loss": 1.0459, | |
| "step": 11445 | |
| }, | |
| { | |
| "epoch": 2.4010056568196103, | |
| "grad_norm": 2.4698305130004883, | |
| "learning_rate": 2.376067058529672e-06, | |
| "loss": 1.0619, | |
| "step": 11460 | |
| }, | |
| { | |
| "epoch": 2.4041483343808925, | |
| "grad_norm": 2.9925546646118164, | |
| "learning_rate": 2.3524550684623927e-06, | |
| "loss": 1.0541, | |
| "step": 11475 | |
| }, | |
| { | |
| "epoch": 2.4072910119421747, | |
| "grad_norm": 2.6975173950195312, | |
| "learning_rate": 2.328945337069809e-06, | |
| "loss": 1.1299, | |
| "step": 11490 | |
| }, | |
| { | |
| "epoch": 2.410433689503457, | |
| "grad_norm": 2.239131212234497, | |
| "learning_rate": 2.305538178710831e-06, | |
| "loss": 1.0085, | |
| "step": 11505 | |
| }, | |
| { | |
| "epoch": 2.413576367064739, | |
| "grad_norm": 2.0249500274658203, | |
| "learning_rate": 2.282233906372835e-06, | |
| "loss": 1.0513, | |
| "step": 11520 | |
| }, | |
| { | |
| "epoch": 2.4167190446260216, | |
| "grad_norm": 2.393962860107422, | |
| "learning_rate": 2.2590328316674582e-06, | |
| "loss": 1.0077, | |
| "step": 11535 | |
| }, | |
| { | |
| "epoch": 2.4198617221873038, | |
| "grad_norm": 3.019960880279541, | |
| "learning_rate": 2.2359352648264387e-06, | |
| "loss": 1.0446, | |
| "step": 11550 | |
| }, | |
| { | |
| "epoch": 2.423004399748586, | |
| "grad_norm": 2.6998488903045654, | |
| "learning_rate": 2.2129415146974665e-06, | |
| "loss": 1.0773, | |
| "step": 11565 | |
| }, | |
| { | |
| "epoch": 2.426147077309868, | |
| "grad_norm": 3.3861067295074463, | |
| "learning_rate": 2.190051888740048e-06, | |
| "loss": 1.103, | |
| "step": 11580 | |
| }, | |
| { | |
| "epoch": 2.4292897548711503, | |
| "grad_norm": 2.977205514907837, | |
| "learning_rate": 2.1672666930214047e-06, | |
| "loss": 1.0307, | |
| "step": 11595 | |
| }, | |
| { | |
| "epoch": 2.4324324324324325, | |
| "grad_norm": 3.0046980381011963, | |
| "learning_rate": 2.1445862322123734e-06, | |
| "loss": 1.0402, | |
| "step": 11610 | |
| }, | |
| { | |
| "epoch": 2.4355751099937146, | |
| "grad_norm": 2.431718111038208, | |
| "learning_rate": 2.1220108095833345e-06, | |
| "loss": 1.0802, | |
| "step": 11625 | |
| }, | |
| { | |
| "epoch": 2.438717787554997, | |
| "grad_norm": 1.7797611951828003, | |
| "learning_rate": 2.099540727000159e-06, | |
| "loss": 1.0833, | |
| "step": 11640 | |
| }, | |
| { | |
| "epoch": 2.441860465116279, | |
| "grad_norm": 2.224626302719116, | |
| "learning_rate": 2.077176284920166e-06, | |
| "loss": 1.0358, | |
| "step": 11655 | |
| }, | |
| { | |
| "epoch": 2.445003142677561, | |
| "grad_norm": 2.7776806354522705, | |
| "learning_rate": 2.054917782388113e-06, | |
| "loss": 1.0251, | |
| "step": 11670 | |
| }, | |
| { | |
| "epoch": 2.4481458202388433, | |
| "grad_norm": 2.1015563011169434, | |
| "learning_rate": 2.0327655170321915e-06, | |
| "loss": 1.0326, | |
| "step": 11685 | |
| }, | |
| { | |
| "epoch": 2.4512884978001255, | |
| "grad_norm": 3.1175501346588135, | |
| "learning_rate": 2.010719785060048e-06, | |
| "loss": 1.0697, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 2.454431175361408, | |
| "grad_norm": 2.0512046813964844, | |
| "learning_rate": 1.9887808812548272e-06, | |
| "loss": 1.063, | |
| "step": 11715 | |
| }, | |
| { | |
| "epoch": 2.4575738529226903, | |
| "grad_norm": 2.4269354343414307, | |
| "learning_rate": 1.966949098971227e-06, | |
| "loss": 1.0623, | |
| "step": 11730 | |
| }, | |
| { | |
| "epoch": 2.4607165304839724, | |
| "grad_norm": 2.026348352432251, | |
| "learning_rate": 1.945224730131576e-06, | |
| "loss": 1.0249, | |
| "step": 11745 | |
| }, | |
| { | |
| "epoch": 2.4638592080452546, | |
| "grad_norm": 2.1796414852142334, | |
| "learning_rate": 1.9236080652219248e-06, | |
| "loss": 1.0627, | |
| "step": 11760 | |
| }, | |
| { | |
| "epoch": 2.4670018856065368, | |
| "grad_norm": 2.8909108638763428, | |
| "learning_rate": 1.9020993932881803e-06, | |
| "loss": 0.9954, | |
| "step": 11775 | |
| }, | |
| { | |
| "epoch": 2.470144563167819, | |
| "grad_norm": 1.939376950263977, | |
| "learning_rate": 1.8806990019322169e-06, | |
| "loss": 0.9819, | |
| "step": 11790 | |
| }, | |
| { | |
| "epoch": 2.473287240729101, | |
| "grad_norm": 2.434191942214966, | |
| "learning_rate": 1.8594071773080492e-06, | |
| "loss": 1.0858, | |
| "step": 11805 | |
| }, | |
| { | |
| "epoch": 2.4764299182903833, | |
| "grad_norm": 2.3099939823150635, | |
| "learning_rate": 1.8382242041179876e-06, | |
| "loss": 1.0571, | |
| "step": 11820 | |
| }, | |
| { | |
| "epoch": 2.4795725958516655, | |
| "grad_norm": 2.7791974544525146, | |
| "learning_rate": 1.8171503656088574e-06, | |
| "loss": 1.0598, | |
| "step": 11835 | |
| }, | |
| { | |
| "epoch": 2.482715273412948, | |
| "grad_norm": 2.9822537899017334, | |
| "learning_rate": 1.7961859435681882e-06, | |
| "loss": 1.029, | |
| "step": 11850 | |
| }, | |
| { | |
| "epoch": 2.4858579509742302, | |
| "grad_norm": 2.644841194152832, | |
| "learning_rate": 1.7753312183204486e-06, | |
| "loss": 1.12, | |
| "step": 11865 | |
| }, | |
| { | |
| "epoch": 2.4890006285355124, | |
| "grad_norm": 2.436312437057495, | |
| "learning_rate": 1.7545864687233127e-06, | |
| "loss": 1.0353, | |
| "step": 11880 | |
| }, | |
| { | |
| "epoch": 2.4921433060967946, | |
| "grad_norm": 2.3745834827423096, | |
| "learning_rate": 1.7339519721639164e-06, | |
| "loss": 1.0758, | |
| "step": 11895 | |
| }, | |
| { | |
| "epoch": 2.4952859836580767, | |
| "grad_norm": 2.9353575706481934, | |
| "learning_rate": 1.713428004555162e-06, | |
| "loss": 1.0455, | |
| "step": 11910 | |
| }, | |
| { | |
| "epoch": 2.498428661219359, | |
| "grad_norm": 2.2262260913848877, | |
| "learning_rate": 1.693014840332009e-06, | |
| "loss": 1.0345, | |
| "step": 11925 | |
| }, | |
| { | |
| "epoch": 2.501571338780641, | |
| "grad_norm": 2.1293249130249023, | |
| "learning_rate": 1.6727127524478227e-06, | |
| "loss": 1.0262, | |
| "step": 11940 | |
| }, | |
| { | |
| "epoch": 2.5047140163419233, | |
| "grad_norm": 2.169814348220825, | |
| "learning_rate": 1.6525220123707176e-06, | |
| "loss": 1.0121, | |
| "step": 11955 | |
| }, | |
| { | |
| "epoch": 2.5078566939032054, | |
| "grad_norm": 2.504615545272827, | |
| "learning_rate": 1.6324428900799273e-06, | |
| "loss": 1.0202, | |
| "step": 11970 | |
| }, | |
| { | |
| "epoch": 2.5109993714644876, | |
| "grad_norm": 2.8434998989105225, | |
| "learning_rate": 1.6124756540621966e-06, | |
| "loss": 1.0042, | |
| "step": 11985 | |
| }, | |
| { | |
| "epoch": 2.5141420490257698, | |
| "grad_norm": 1.939417839050293, | |
| "learning_rate": 1.5926205713081876e-06, | |
| "loss": 1.1108, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 2.517284726587052, | |
| "grad_norm": 2.569380283355713, | |
| "learning_rate": 1.572877907308913e-06, | |
| "loss": 1.0067, | |
| "step": 12015 | |
| }, | |
| { | |
| "epoch": 2.520427404148334, | |
| "grad_norm": 2.747483253479004, | |
| "learning_rate": 1.5532479260521849e-06, | |
| "loss": 1.0509, | |
| "step": 12030 | |
| }, | |
| { | |
| "epoch": 2.5235700817096167, | |
| "grad_norm": 1.7206883430480957, | |
| "learning_rate": 1.533730890019086e-06, | |
| "loss": 1.0901, | |
| "step": 12045 | |
| }, | |
| { | |
| "epoch": 2.526712759270899, | |
| "grad_norm": 2.222593069076538, | |
| "learning_rate": 1.5143270601804582e-06, | |
| "loss": 1.0182, | |
| "step": 12060 | |
| }, | |
| { | |
| "epoch": 2.529855436832181, | |
| "grad_norm": 2.6068830490112305, | |
| "learning_rate": 1.4950366959934116e-06, | |
| "loss": 1.0768, | |
| "step": 12075 | |
| }, | |
| { | |
| "epoch": 2.5329981143934632, | |
| "grad_norm": 2.674351692199707, | |
| "learning_rate": 1.475860055397862e-06, | |
| "loss": 1.0114, | |
| "step": 12090 | |
| }, | |
| { | |
| "epoch": 2.5361407919547454, | |
| "grad_norm": 3.2743771076202393, | |
| "learning_rate": 1.4567973948130731e-06, | |
| "loss": 1.0805, | |
| "step": 12105 | |
| }, | |
| { | |
| "epoch": 2.5392834695160276, | |
| "grad_norm": 3.0732719898223877, | |
| "learning_rate": 1.4378489691342345e-06, | |
| "loss": 1.0802, | |
| "step": 12120 | |
| }, | |
| { | |
| "epoch": 2.5424261470773097, | |
| "grad_norm": 2.502260446548462, | |
| "learning_rate": 1.4190150317290485e-06, | |
| "loss": 1.0701, | |
| "step": 12135 | |
| }, | |
| { | |
| "epoch": 2.5455688246385924, | |
| "grad_norm": 2.1941142082214355, | |
| "learning_rate": 1.4002958344343465e-06, | |
| "loss": 1.0671, | |
| "step": 12150 | |
| }, | |
| { | |
| "epoch": 2.5487115021998745, | |
| "grad_norm": 2.5131101608276367, | |
| "learning_rate": 1.3816916275527193e-06, | |
| "loss": 1.0483, | |
| "step": 12165 | |
| }, | |
| { | |
| "epoch": 2.5518541797611567, | |
| "grad_norm": 2.770385503768921, | |
| "learning_rate": 1.36320265984917e-06, | |
| "loss": 1.0708, | |
| "step": 12180 | |
| }, | |
| { | |
| "epoch": 2.554996857322439, | |
| "grad_norm": 2.3241212368011475, | |
| "learning_rate": 1.3448291785477874e-06, | |
| "loss": 1.0521, | |
| "step": 12195 | |
| }, | |
| { | |
| "epoch": 2.558139534883721, | |
| "grad_norm": 3.1057024002075195, | |
| "learning_rate": 1.3265714293284414e-06, | |
| "loss": 1.0705, | |
| "step": 12210 | |
| }, | |
| { | |
| "epoch": 2.561282212445003, | |
| "grad_norm": 1.8851613998413086, | |
| "learning_rate": 1.3084296563235033e-06, | |
| "loss": 1.037, | |
| "step": 12225 | |
| }, | |
| { | |
| "epoch": 2.5644248900062854, | |
| "grad_norm": 2.893693685531616, | |
| "learning_rate": 1.2904041021145597e-06, | |
| "loss": 1.0582, | |
| "step": 12240 | |
| }, | |
| { | |
| "epoch": 2.5675675675675675, | |
| "grad_norm": 2.2626419067382812, | |
| "learning_rate": 1.2724950077292042e-06, | |
| "loss": 1.0213, | |
| "step": 12255 | |
| }, | |
| { | |
| "epoch": 2.5707102451288497, | |
| "grad_norm": 2.1927521228790283, | |
| "learning_rate": 1.2547026126377849e-06, | |
| "loss": 1.0542, | |
| "step": 12270 | |
| }, | |
| { | |
| "epoch": 2.573852922690132, | |
| "grad_norm": 2.205655097961426, | |
| "learning_rate": 1.2382018758160751e-06, | |
| "loss": 1.0494, | |
| "step": 12285 | |
| }, | |
| { | |
| "epoch": 2.576995600251414, | |
| "grad_norm": 2.98274827003479, | |
| "learning_rate": 1.2206357725870321e-06, | |
| "loss": 1.0926, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 2.5801382778126962, | |
| "grad_norm": 2.7129874229431152, | |
| "learning_rate": 1.2031870620844844e-06, | |
| "loss": 1.0396, | |
| "step": 12315 | |
| }, | |
| { | |
| "epoch": 2.5832809553739784, | |
| "grad_norm": 2.8410415649414062, | |
| "learning_rate": 1.1858559776227863e-06, | |
| "loss": 1.0353, | |
| "step": 12330 | |
| }, | |
| { | |
| "epoch": 2.586423632935261, | |
| "grad_norm": 2.896530866622925, | |
| "learning_rate": 1.1686427509434617e-06, | |
| "loss": 1.1285, | |
| "step": 12345 | |
| }, | |
| { | |
| "epoch": 2.589566310496543, | |
| "grad_norm": 2.223740339279175, | |
| "learning_rate": 1.1515476122121094e-06, | |
| "loss": 1.0727, | |
| "step": 12360 | |
| }, | |
| { | |
| "epoch": 2.5927089880578253, | |
| "grad_norm": 2.6268301010131836, | |
| "learning_rate": 1.1345707900153203e-06, | |
| "loss": 1.016, | |
| "step": 12375 | |
| }, | |
| { | |
| "epoch": 2.5958516656191075, | |
| "grad_norm": 2.7400355339050293, | |
| "learning_rate": 1.1177125113576237e-06, | |
| "loss": 1.0317, | |
| "step": 12390 | |
| }, | |
| { | |
| "epoch": 2.5989943431803897, | |
| "grad_norm": 2.5901083946228027, | |
| "learning_rate": 1.100973001658453e-06, | |
| "loss": 1.0526, | |
| "step": 12405 | |
| }, | |
| { | |
| "epoch": 2.602137020741672, | |
| "grad_norm": 2.201157569885254, | |
| "learning_rate": 1.084352484749126e-06, | |
| "loss": 1.0881, | |
| "step": 12420 | |
| }, | |
| { | |
| "epoch": 2.605279698302954, | |
| "grad_norm": 3.0149240493774414, | |
| "learning_rate": 1.067851182869859e-06, | |
| "loss": 1.02, | |
| "step": 12435 | |
| }, | |
| { | |
| "epoch": 2.608422375864236, | |
| "grad_norm": 2.6754496097564697, | |
| "learning_rate": 1.051469316666791e-06, | |
| "loss": 1.0289, | |
| "step": 12450 | |
| }, | |
| { | |
| "epoch": 2.611565053425519, | |
| "grad_norm": 2.752760648727417, | |
| "learning_rate": 1.0352071051890333e-06, | |
| "loss": 1.0687, | |
| "step": 12465 | |
| }, | |
| { | |
| "epoch": 2.614707730986801, | |
| "grad_norm": 2.3588600158691406, | |
| "learning_rate": 1.0190647658857433e-06, | |
| "loss": 1.0514, | |
| "step": 12480 | |
| }, | |
| { | |
| "epoch": 2.617850408548083, | |
| "grad_norm": 2.367644786834717, | |
| "learning_rate": 1.0030425146032141e-06, | |
| "loss": 0.9977, | |
| "step": 12495 | |
| }, | |
| { | |
| "epoch": 2.6209930861093653, | |
| "grad_norm": 3.080481767654419, | |
| "learning_rate": 9.8714056558199e-07, | |
| "loss": 1.1113, | |
| "step": 12510 | |
| }, | |
| { | |
| "epoch": 2.6241357636706475, | |
| "grad_norm": 3.0548975467681885, | |
| "learning_rate": 9.713591314539938e-07, | |
| "loss": 1.0369, | |
| "step": 12525 | |
| }, | |
| { | |
| "epoch": 2.6272784412319297, | |
| "grad_norm": 2.131619930267334, | |
| "learning_rate": 9.556984232397037e-07, | |
| "loss": 1.0072, | |
| "step": 12540 | |
| }, | |
| { | |
| "epoch": 2.630421118793212, | |
| "grad_norm": 2.963510036468506, | |
| "learning_rate": 9.40158650345312e-07, | |
| "loss": 1.0475, | |
| "step": 12555 | |
| }, | |
| { | |
| "epoch": 2.633563796354494, | |
| "grad_norm": 2.7937393188476562, | |
| "learning_rate": 9.24740020559931e-07, | |
| "loss": 1.0605, | |
| "step": 12570 | |
| }, | |
| { | |
| "epoch": 2.636706473915776, | |
| "grad_norm": 2.5828826427459717, | |
| "learning_rate": 9.094427400528128e-07, | |
| "loss": 1.0657, | |
| "step": 12585 | |
| }, | |
| { | |
| "epoch": 2.6398491514770583, | |
| "grad_norm": 2.3063597679138184, | |
| "learning_rate": 8.942670133706044e-07, | |
| "loss": 1.015, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 2.6429918290383405, | |
| "grad_norm": 2.2541069984436035, | |
| "learning_rate": 8.792130434345947e-07, | |
| "loss": 1.0939, | |
| "step": 12615 | |
| }, | |
| { | |
| "epoch": 2.6461345065996227, | |
| "grad_norm": 2.0006487369537354, | |
| "learning_rate": 8.642810315380168e-07, | |
| "loss": 1.0214, | |
| "step": 12630 | |
| }, | |
| { | |
| "epoch": 2.649277184160905, | |
| "grad_norm": 2.5538716316223145, | |
| "learning_rate": 8.494711773433395e-07, | |
| "loss": 1.0569, | |
| "step": 12645 | |
| }, | |
| { | |
| "epoch": 2.6524198617221875, | |
| "grad_norm": 2.3980712890625, | |
| "learning_rate": 8.347836788796148e-07, | |
| "loss": 1.033, | |
| "step": 12660 | |
| }, | |
| { | |
| "epoch": 2.6555625392834696, | |
| "grad_norm": 2.408301830291748, | |
| "learning_rate": 8.202187325398248e-07, | |
| "loss": 1.0013, | |
| "step": 12675 | |
| }, | |
| { | |
| "epoch": 2.658705216844752, | |
| "grad_norm": 1.9503532648086548, | |
| "learning_rate": 8.057765330782486e-07, | |
| "loss": 0.9753, | |
| "step": 12690 | |
| }, | |
| { | |
| "epoch": 2.661847894406034, | |
| "grad_norm": 2.117751359939575, | |
| "learning_rate": 7.914572736078618e-07, | |
| "loss": 1.091, | |
| "step": 12705 | |
| }, | |
| { | |
| "epoch": 2.664990571967316, | |
| "grad_norm": 2.651381731033325, | |
| "learning_rate": 7.772611455977586e-07, | |
| "loss": 1.0366, | |
| "step": 12720 | |
| }, | |
| { | |
| "epoch": 2.6681332495285983, | |
| "grad_norm": 2.568176031112671, | |
| "learning_rate": 7.631883388705897e-07, | |
| "loss": 1.0775, | |
| "step": 12735 | |
| }, | |
| { | |
| "epoch": 2.6712759270898805, | |
| "grad_norm": 2.860661268234253, | |
| "learning_rate": 7.492390416000217e-07, | |
| "loss": 1.0576, | |
| "step": 12750 | |
| }, | |
| { | |
| "epoch": 2.6744186046511627, | |
| "grad_norm": 2.9709537029266357, | |
| "learning_rate": 7.354134403082191e-07, | |
| "loss": 1.0724, | |
| "step": 12765 | |
| }, | |
| { | |
| "epoch": 2.6775612822124453, | |
| "grad_norm": 1.8460935354232788, | |
| "learning_rate": 7.217117198633561e-07, | |
| "loss": 1.0227, | |
| "step": 12780 | |
| }, | |
| { | |
| "epoch": 2.6807039597737274, | |
| "grad_norm": 2.6005988121032715, | |
| "learning_rate": 7.081340634771394e-07, | |
| "loss": 1.0787, | |
| "step": 12795 | |
| }, | |
| { | |
| "epoch": 2.6838466373350096, | |
| "grad_norm": 2.098374366760254, | |
| "learning_rate": 6.946806527023686e-07, | |
| "loss": 1.0402, | |
| "step": 12810 | |
| }, | |
| { | |
| "epoch": 2.686989314896292, | |
| "grad_norm": 2.3875677585601807, | |
| "learning_rate": 6.813516674304888e-07, | |
| "loss": 1.0367, | |
| "step": 12825 | |
| }, | |
| { | |
| "epoch": 2.690131992457574, | |
| "grad_norm": 1.9632166624069214, | |
| "learning_rate": 6.681472858892068e-07, | |
| "loss": 1.0026, | |
| "step": 12840 | |
| }, | |
| { | |
| "epoch": 2.693274670018856, | |
| "grad_norm": 2.6119518280029297, | |
| "learning_rate": 6.550676846400927e-07, | |
| "loss": 1.0865, | |
| "step": 12855 | |
| }, | |
| { | |
| "epoch": 2.6964173475801383, | |
| "grad_norm": 2.205327272415161, | |
| "learning_rate": 6.421130385762275e-07, | |
| "loss": 1.0165, | |
| "step": 12870 | |
| }, | |
| { | |
| "epoch": 2.6995600251414205, | |
| "grad_norm": 3.523000478744507, | |
| "learning_rate": 6.292835209198622e-07, | |
| "loss": 1.0645, | |
| "step": 12885 | |
| }, | |
| { | |
| "epoch": 2.7027027027027026, | |
| "grad_norm": 2.800053119659424, | |
| "learning_rate": 6.165793032200984e-07, | |
| "loss": 1.028, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 2.705845380263985, | |
| "grad_norm": 2.6920459270477295, | |
| "learning_rate": 6.040005553505979e-07, | |
| "loss": 1.0499, | |
| "step": 12915 | |
| }, | |
| { | |
| "epoch": 2.708988057825267, | |
| "grad_norm": 2.7640254497528076, | |
| "learning_rate": 5.915474455073111e-07, | |
| "loss": 1.0495, | |
| "step": 12930 | |
| }, | |
| { | |
| "epoch": 2.712130735386549, | |
| "grad_norm": 2.8138866424560547, | |
| "learning_rate": 5.792201402062281e-07, | |
| "loss": 1.0698, | |
| "step": 12945 | |
| }, | |
| { | |
| "epoch": 2.7152734129478313, | |
| "grad_norm": 3.2253429889678955, | |
| "learning_rate": 5.670188042811498e-07, | |
| "loss": 1.0635, | |
| "step": 12960 | |
| }, | |
| { | |
| "epoch": 2.718416090509114, | |
| "grad_norm": 2.7197041511535645, | |
| "learning_rate": 5.549436008814868e-07, | |
| "loss": 1.0507, | |
| "step": 12975 | |
| }, | |
| { | |
| "epoch": 2.721558768070396, | |
| "grad_norm": 2.0629515647888184, | |
| "learning_rate": 5.429946914700745e-07, | |
| "loss": 1.0665, | |
| "step": 12990 | |
| }, | |
| { | |
| "epoch": 2.7247014456316783, | |
| "grad_norm": 1.5886244773864746, | |
| "learning_rate": 5.311722358210147e-07, | |
| "loss": 1.0672, | |
| "step": 13005 | |
| }, | |
| { | |
| "epoch": 2.7278441231929604, | |
| "grad_norm": 2.3268556594848633, | |
| "learning_rate": 5.194763920175461e-07, | |
| "loss": 1.0415, | |
| "step": 13020 | |
| }, | |
| { | |
| "epoch": 2.7309868007542426, | |
| "grad_norm": 2.352407693862915, | |
| "learning_rate": 5.079073164499216e-07, | |
| "loss": 1.0628, | |
| "step": 13035 | |
| }, | |
| { | |
| "epoch": 2.7341294783155248, | |
| "grad_norm": 2.2687618732452393, | |
| "learning_rate": 4.964651638133211e-07, | |
| "loss": 1.1221, | |
| "step": 13050 | |
| }, | |
| { | |
| "epoch": 2.737272155876807, | |
| "grad_norm": 2.1698989868164062, | |
| "learning_rate": 4.851500871057791e-07, | |
| "loss": 1.0301, | |
| "step": 13065 | |
| }, | |
| { | |
| "epoch": 2.740414833438089, | |
| "grad_norm": 2.6248748302459717, | |
| "learning_rate": 4.7396223762614545e-07, | |
| "loss": 1.1054, | |
| "step": 13080 | |
| }, | |
| { | |
| "epoch": 2.7435575109993717, | |
| "grad_norm": 2.520930767059326, | |
| "learning_rate": 4.629017649720602e-07, | |
| "loss": 1.0437, | |
| "step": 13095 | |
| }, | |
| { | |
| "epoch": 2.746700188560654, | |
| "grad_norm": 3.120213508605957, | |
| "learning_rate": 4.519688170379499e-07, | |
| "loss": 1.0116, | |
| "step": 13110 | |
| }, | |
| { | |
| "epoch": 2.749842866121936, | |
| "grad_norm": 2.3425211906433105, | |
| "learning_rate": 4.411635400130521e-07, | |
| "loss": 1.0608, | |
| "step": 13125 | |
| }, | |
| { | |
| "epoch": 2.7529855436832182, | |
| "grad_norm": 2.232712745666504, | |
| "learning_rate": 4.3048607837945954e-07, | |
| "loss": 1.0401, | |
| "step": 13140 | |
| }, | |
| { | |
| "epoch": 2.7561282212445004, | |
| "grad_norm": 2.2147791385650635, | |
| "learning_rate": 4.1993657491019355e-07, | |
| "loss": 1.0217, | |
| "step": 13155 | |
| }, | |
| { | |
| "epoch": 2.7592708988057826, | |
| "grad_norm": 2.526965618133545, | |
| "learning_rate": 4.095151706672884e-07, | |
| "loss": 1.0915, | |
| "step": 13170 | |
| }, | |
| { | |
| "epoch": 2.7624135763670647, | |
| "grad_norm": 1.9940340518951416, | |
| "learning_rate": 3.9922200499990404e-07, | |
| "loss": 1.0553, | |
| "step": 13185 | |
| }, | |
| { | |
| "epoch": 2.765556253928347, | |
| "grad_norm": 2.973738431930542, | |
| "learning_rate": 3.890572155424721e-07, | |
| "loss": 1.0615, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 2.768698931489629, | |
| "grad_norm": 2.3078935146331787, | |
| "learning_rate": 3.790209382128451e-07, | |
| "loss": 1.0285, | |
| "step": 13215 | |
| }, | |
| { | |
| "epoch": 2.7718416090509113, | |
| "grad_norm": 2.037052869796753, | |
| "learning_rate": 3.691133072104891e-07, | |
| "loss": 1.0545, | |
| "step": 13230 | |
| }, | |
| { | |
| "epoch": 2.7749842866121934, | |
| "grad_norm": 2.9037740230560303, | |
| "learning_rate": 3.59334455014676e-07, | |
| "loss": 1.0151, | |
| "step": 13245 | |
| }, | |
| { | |
| "epoch": 2.7781269641734756, | |
| "grad_norm": 2.245582103729248, | |
| "learning_rate": 3.4968451238272527e-07, | |
| "loss": 1.0715, | |
| "step": 13260 | |
| }, | |
| { | |
| "epoch": 2.7812696417347578, | |
| "grad_norm": 2.7545788288116455, | |
| "learning_rate": 3.4079431818805573e-07, | |
| "loss": 1.0315, | |
| "step": 13275 | |
| }, | |
| { | |
| "epoch": 2.7844123192960404, | |
| "grad_norm": 3.0721166133880615, | |
| "learning_rate": 3.313939650727771e-07, | |
| "loss": 1.0796, | |
| "step": 13290 | |
| }, | |
| { | |
| "epoch": 2.7875549968573226, | |
| "grad_norm": 2.1562039852142334, | |
| "learning_rate": 3.22122895125897e-07, | |
| "loss": 1.0105, | |
| "step": 13305 | |
| }, | |
| { | |
| "epoch": 2.7906976744186047, | |
| "grad_norm": 2.4602880477905273, | |
| "learning_rate": 3.129812323149528e-07, | |
| "loss": 1.0805, | |
| "step": 13320 | |
| }, | |
| { | |
| "epoch": 2.793840351979887, | |
| "grad_norm": 2.485700845718384, | |
| "learning_rate": 3.039690988771271e-07, | |
| "loss": 1.1113, | |
| "step": 13335 | |
| }, | |
| { | |
| "epoch": 2.796983029541169, | |
| "grad_norm": 2.2793211936950684, | |
| "learning_rate": 2.950866153176024e-07, | |
| "loss": 1.0269, | |
| "step": 13350 | |
| }, | |
| { | |
| "epoch": 2.8001257071024512, | |
| "grad_norm": 2.4900388717651367, | |
| "learning_rate": 2.8633390040796014e-07, | |
| "loss": 1.0586, | |
| "step": 13365 | |
| }, | |
| { | |
| "epoch": 2.8032683846637334, | |
| "grad_norm": 2.7282629013061523, | |
| "learning_rate": 2.7771107118458874e-07, | |
| "loss": 1.0752, | |
| "step": 13380 | |
| }, | |
| { | |
| "epoch": 2.8064110622250156, | |
| "grad_norm": 2.986842632293701, | |
| "learning_rate": 2.692182429471157e-07, | |
| "loss": 1.0826, | |
| "step": 13395 | |
| }, | |
| { | |
| "epoch": 2.809553739786298, | |
| "grad_norm": 1.311810851097107, | |
| "learning_rate": 2.608555292568671e-07, | |
| "loss": 1.0583, | |
| "step": 13410 | |
| }, | |
| { | |
| "epoch": 2.8126964173475804, | |
| "grad_norm": 2.533581256866455, | |
| "learning_rate": 2.526230419353537e-07, | |
| "loss": 1.0604, | |
| "step": 13425 | |
| }, | |
| { | |
| "epoch": 2.8158390949088625, | |
| "grad_norm": 2.3201465606689453, | |
| "learning_rate": 2.445208910627717e-07, | |
| "loss": 1.0868, | |
| "step": 13440 | |
| }, | |
| { | |
| "epoch": 2.8189817724701447, | |
| "grad_norm": 2.0772528648376465, | |
| "learning_rate": 2.3654918497653334e-07, | |
| "loss": 1.0854, | |
| "step": 13455 | |
| }, | |
| { | |
| "epoch": 2.822124450031427, | |
| "grad_norm": 2.761883020401001, | |
| "learning_rate": 2.287080302698097e-07, | |
| "loss": 1.0408, | |
| "step": 13470 | |
| }, | |
| { | |
| "epoch": 2.825267127592709, | |
| "grad_norm": 2.217202663421631, | |
| "learning_rate": 2.2099753179011807e-07, | |
| "loss": 1.0712, | |
| "step": 13485 | |
| }, | |
| { | |
| "epoch": 2.828409805153991, | |
| "grad_norm": 2.955160140991211, | |
| "learning_rate": 2.134177926379144e-07, | |
| "loss": 1.0678, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 2.8315524827152734, | |
| "grad_norm": 2.9105584621429443, | |
| "learning_rate": 2.0596891416521215e-07, | |
| "loss": 1.1113, | |
| "step": 13515 | |
| }, | |
| { | |
| "epoch": 2.8346951602765555, | |
| "grad_norm": 2.016284704208374, | |
| "learning_rate": 1.9865099597422887e-07, | |
| "loss": 1.0373, | |
| "step": 13530 | |
| }, | |
| { | |
| "epoch": 2.8378378378378377, | |
| "grad_norm": 2.6535935401916504, | |
| "learning_rate": 1.9146413591605516e-07, | |
| "loss": 1.0911, | |
| "step": 13545 | |
| }, | |
| { | |
| "epoch": 2.84098051539912, | |
| "grad_norm": 3.204658031463623, | |
| "learning_rate": 1.844084300893456e-07, | |
| "loss": 1.0636, | |
| "step": 13560 | |
| }, | |
| { | |
| "epoch": 2.844123192960402, | |
| "grad_norm": 2.872605085372925, | |
| "learning_rate": 1.7748397283903874e-07, | |
| "loss": 1.093, | |
| "step": 13575 | |
| }, | |
| { | |
| "epoch": 2.8472658705216842, | |
| "grad_norm": 2.8483595848083496, | |
| "learning_rate": 1.7069085675508356e-07, | |
| "loss": 1.0961, | |
| "step": 13590 | |
| }, | |
| { | |
| "epoch": 2.850408548082967, | |
| "grad_norm": 2.082428455352783, | |
| "learning_rate": 1.6402917267121398e-07, | |
| "loss": 1.0593, | |
| "step": 13605 | |
| }, | |
| { | |
| "epoch": 2.853551225644249, | |
| "grad_norm": 2.1791653633117676, | |
| "learning_rate": 1.5749900966372523e-07, | |
| "loss": 1.0552, | |
| "step": 13620 | |
| }, | |
| { | |
| "epoch": 2.856693903205531, | |
| "grad_norm": 1.7146304845809937, | |
| "learning_rate": 1.5110045505029036e-07, | |
| "loss": 1.0435, | |
| "step": 13635 | |
| }, | |
| { | |
| "epoch": 2.8598365807668134, | |
| "grad_norm": 2.64682674407959, | |
| "learning_rate": 1.4483359438878574e-07, | |
| "loss": 1.023, | |
| "step": 13650 | |
| }, | |
| { | |
| "epoch": 2.8629792583280955, | |
| "grad_norm": 2.390026569366455, | |
| "learning_rate": 1.386985114761541e-07, | |
| "loss": 1.0559, | |
| "step": 13665 | |
| }, | |
| { | |
| "epoch": 2.8661219358893777, | |
| "grad_norm": 2.797459602355957, | |
| "learning_rate": 1.326952883472743e-07, | |
| "loss": 1.1022, | |
| "step": 13680 | |
| }, | |
| { | |
| "epoch": 2.86926461345066, | |
| "grad_norm": 1.6154651641845703, | |
| "learning_rate": 1.2682400527387674e-07, | |
| "loss": 1.0415, | |
| "step": 13695 | |
| }, | |
| { | |
| "epoch": 2.872407291011942, | |
| "grad_norm": 2.7609522342681885, | |
| "learning_rate": 1.2108474076345855e-07, | |
| "loss": 1.0901, | |
| "step": 13710 | |
| }, | |
| { | |
| "epoch": 2.8755499685732246, | |
| "grad_norm": 2.618767499923706, | |
| "learning_rate": 1.1547757155824124e-07, | |
| "loss": 1.0118, | |
| "step": 13725 | |
| }, | |
| { | |
| "epoch": 2.878692646134507, | |
| "grad_norm": 2.90389347076416, | |
| "learning_rate": 1.1000257263414139e-07, | |
| "loss": 1.0729, | |
| "step": 13740 | |
| }, | |
| { | |
| "epoch": 2.881835323695789, | |
| "grad_norm": 2.361311435699463, | |
| "learning_rate": 1.0465981719976814e-07, | |
| "loss": 1.0597, | |
| "step": 13755 | |
| }, | |
| { | |
| "epoch": 2.884978001257071, | |
| "grad_norm": 1.77293062210083, | |
| "learning_rate": 9.944937669544741e-08, | |
| "loss": 1.0155, | |
| "step": 13770 | |
| }, | |
| { | |
| "epoch": 2.8881206788183533, | |
| "grad_norm": 2.5535075664520264, | |
| "learning_rate": 9.437132079226142e-08, | |
| "loss": 1.0966, | |
| "step": 13785 | |
| }, | |
| { | |
| "epoch": 2.8912633563796355, | |
| "grad_norm": 2.4783518314361572, | |
| "learning_rate": 8.942571739112171e-08, | |
| "loss": 1.0554, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 2.8944060339409177, | |
| "grad_norm": 1.9884682893753052, | |
| "learning_rate": 8.46126326218577e-08, | |
| "loss": 0.9958, | |
| "step": 13815 | |
| }, | |
| { | |
| "epoch": 2.8975487115022, | |
| "grad_norm": 1.811047911643982, | |
| "learning_rate": 7.99321308423362e-08, | |
| "loss": 1.0913, | |
| "step": 13830 | |
| }, | |
| { | |
| "epoch": 2.900691389063482, | |
| "grad_norm": 2.5902936458587646, | |
| "learning_rate": 7.538427463759768e-08, | |
| "loss": 1.0423, | |
| "step": 13845 | |
| }, | |
| { | |
| "epoch": 2.903834066624764, | |
| "grad_norm": 2.312286853790283, | |
| "learning_rate": 7.096912481901919e-08, | |
| "loss": 1.0178, | |
| "step": 13860 | |
| }, | |
| { | |
| "epoch": 2.9069767441860463, | |
| "grad_norm": 1.918196201324463, | |
| "learning_rate": 6.668674042350498e-08, | |
| "loss": 1.053, | |
| "step": 13875 | |
| }, | |
| { | |
| "epoch": 2.9101194217473285, | |
| "grad_norm": 1.8703136444091797, | |
| "learning_rate": 6.253717871269271e-08, | |
| "loss": 1.0538, | |
| "step": 13890 | |
| }, | |
| { | |
| "epoch": 2.9132620993086107, | |
| "grad_norm": 2.3557560443878174, | |
| "learning_rate": 5.8520495172190714e-08, | |
| "loss": 1.0793, | |
| "step": 13905 | |
| }, | |
| { | |
| "epoch": 2.9164047768698933, | |
| "grad_norm": 2.194120407104492, | |
| "learning_rate": 5.4636743510835256e-08, | |
| "loss": 1.0484, | |
| "step": 13920 | |
| }, | |
| { | |
| "epoch": 2.9195474544311755, | |
| "grad_norm": 1.7967839241027832, | |
| "learning_rate": 5.088597565997111e-08, | |
| "loss": 1.0467, | |
| "step": 13935 | |
| }, | |
| { | |
| "epoch": 2.9226901319924576, | |
| "grad_norm": 1.9517879486083984, | |
| "learning_rate": 4.72682417727599e-08, | |
| "loss": 1.0425, | |
| "step": 13950 | |
| }, | |
| { | |
| "epoch": 2.92583280955374, | |
| "grad_norm": 2.5437710285186768, | |
| "learning_rate": 4.37835902235062e-08, | |
| "loss": 1.0693, | |
| "step": 13965 | |
| }, | |
| { | |
| "epoch": 2.928975487115022, | |
| "grad_norm": 2.4388108253479004, | |
| "learning_rate": 4.043206760701246e-08, | |
| "loss": 1.1279, | |
| "step": 13980 | |
| }, | |
| { | |
| "epoch": 2.932118164676304, | |
| "grad_norm": 2.955836057662964, | |
| "learning_rate": 3.7213718737958425e-08, | |
| "loss": 1.0869, | |
| "step": 13995 | |
| }, | |
| { | |
| "epoch": 2.9352608422375863, | |
| "grad_norm": 2.5022730827331543, | |
| "learning_rate": 3.412858665029495e-08, | |
| "loss": 1.0562, | |
| "step": 14010 | |
| }, | |
| { | |
| "epoch": 2.9384035197988685, | |
| "grad_norm": 2.614485502243042, | |
| "learning_rate": 3.117671259667554e-08, | |
| "loss": 1.0316, | |
| "step": 14025 | |
| }, | |
| { | |
| "epoch": 2.941546197360151, | |
| "grad_norm": 1.9243937730789185, | |
| "learning_rate": 2.8358136047902385e-08, | |
| "loss": 1.0355, | |
| "step": 14040 | |
| }, | |
| { | |
| "epoch": 2.9446888749214333, | |
| "grad_norm": 2.475581407546997, | |
| "learning_rate": 2.5672894692397865e-08, | |
| "loss": 1.0007, | |
| "step": 14055 | |
| }, | |
| { | |
| "epoch": 2.9478315524827154, | |
| "grad_norm": 2.3144288063049316, | |
| "learning_rate": 2.312102443569941e-08, | |
| "loss": 1.0889, | |
| "step": 14070 | |
| }, | |
| { | |
| "epoch": 2.9509742300439976, | |
| "grad_norm": 1.9073032140731812, | |
| "learning_rate": 2.0702559399985444e-08, | |
| "loss": 1.0502, | |
| "step": 14085 | |
| }, | |
| { | |
| "epoch": 2.95411690760528, | |
| "grad_norm": 2.616192579269409, | |
| "learning_rate": 1.8417531923609066e-08, | |
| "loss": 1.0767, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 2.957259585166562, | |
| "grad_norm": 2.3188884258270264, | |
| "learning_rate": 1.6265972560675082e-08, | |
| "loss": 1.0091, | |
| "step": 14115 | |
| }, | |
| { | |
| "epoch": 2.960402262727844, | |
| "grad_norm": 2.5977001190185547, | |
| "learning_rate": 1.424791008062587e-08, | |
| "loss": 1.0617, | |
| "step": 14130 | |
| }, | |
| { | |
| "epoch": 2.9635449402891263, | |
| "grad_norm": 2.5720694065093994, | |
| "learning_rate": 1.2363371467861706e-08, | |
| "loss": 1.0382, | |
| "step": 14145 | |
| }, | |
| { | |
| "epoch": 2.9666876178504085, | |
| "grad_norm": 2.201608657836914, | |
| "learning_rate": 1.061238192137104e-08, | |
| "loss": 1.0358, | |
| "step": 14160 | |
| }, | |
| { | |
| "epoch": 2.9698302954116906, | |
| "grad_norm": 2.547421455383301, | |
| "learning_rate": 8.994964854404098e-09, | |
| "loss": 1.0699, | |
| "step": 14175 | |
| }, | |
| { | |
| "epoch": 2.972972972972973, | |
| "grad_norm": 2.6077804565429688, | |
| "learning_rate": 7.51114189415314e-09, | |
| "loss": 1.0301, | |
| "step": 14190 | |
| }, | |
| { | |
| "epoch": 2.976115650534255, | |
| "grad_norm": 2.2802724838256836, | |
| "learning_rate": 6.16093288146491e-09, | |
| "loss": 1.0323, | |
| "step": 14205 | |
| }, | |
| { | |
| "epoch": 2.979258328095537, | |
| "grad_norm": 3.556347608566284, | |
| "learning_rate": 4.9443558705764e-09, | |
| "loss": 1.057, | |
| "step": 14220 | |
| }, | |
| { | |
| "epoch": 2.9824010056568198, | |
| "grad_norm": 2.6890947818756104, | |
| "learning_rate": 3.86142712887172e-09, | |
| "loss": 1.0276, | |
| "step": 14235 | |
| }, | |
| { | |
| "epoch": 2.985543683218102, | |
| "grad_norm": 1.9261900186538696, | |
| "learning_rate": 2.912161136665592e-09, | |
| "loss": 0.9678, | |
| "step": 14250 | |
| }, | |
| { | |
| "epoch": 2.988686360779384, | |
| "grad_norm": 1.9330463409423828, | |
| "learning_rate": 2.096570587010183e-09, | |
| "loss": 1.07, | |
| "step": 14265 | |
| }, | |
| { | |
| "epoch": 2.9918290383406663, | |
| "grad_norm": 2.8647425174713135, | |
| "learning_rate": 1.4559673424208165e-09, | |
| "loss": 1.0322, | |
| "step": 14280 | |
| }, | |
| { | |
| "epoch": 2.9949717159019484, | |
| "grad_norm": 2.028707504272461, | |
| "learning_rate": 8.98845328231035e-10, | |
| "loss": 1.0043, | |
| "step": 14295 | |
| }, | |
| { | |
| "epoch": 2.9981143934632306, | |
| "grad_norm": 2.955282688140869, | |
| "learning_rate": 4.754256775185706e-10, | |
| "loss": 1.0309, | |
| "step": 14310 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 14319, | |
| "total_flos": 3.0631036932776264e+18, | |
| "train_loss": 1.1692348350084263, | |
| "train_runtime": 44795.2526, | |
| "train_samples_per_second": 1.279, | |
| "train_steps_per_second": 0.32 | |
| } | |
| ], | |
| "logging_steps": 15, | |
| "max_steps": 14319, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 10000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 3.0631036932776264e+18, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |