| { | |
| "best_metric": 0.8874096274375916, | |
| "best_model_checkpoint": "/workspace/previous_works/MedBLIP/output/MedBLIP-0007/checkpoint-10000", | |
| "epoch": 2.0951183741881416, | |
| "eval_steps": 10000, | |
| "global_step": 10000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0031426775612822125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0.0, | |
| "loss": 6.4738, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.006285355122564425, | |
| "grad_norm": 143.78610229492188, | |
| "learning_rate": 1.3966480446927375e-07, | |
| "loss": 6.4427, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.009428032683846637, | |
| "grad_norm": 125.2808609008789, | |
| "learning_rate": 3.3519553072625703e-07, | |
| "loss": 6.1729, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.01257071024512885, | |
| "grad_norm": 55.64238357543945, | |
| "learning_rate": 5.446927374301677e-07, | |
| "loss": 5.6704, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.01571338780641106, | |
| "grad_norm": 58.784568786621094, | |
| "learning_rate": 7.402234636871509e-07, | |
| "loss": 4.6903, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.018856065367693273, | |
| "grad_norm": 101.14881896972656, | |
| "learning_rate": 9.497206703910615e-07, | |
| "loss": 4.009, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.02199874292897549, | |
| "grad_norm": 40.7690315246582, | |
| "learning_rate": 1.159217877094972e-06, | |
| "loss": 3.3713, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.0251414204902577, | |
| "grad_norm": 22.23933219909668, | |
| "learning_rate": 1.3687150837988828e-06, | |
| "loss": 2.9785, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.028284098051539912, | |
| "grad_norm": 16.25871467590332, | |
| "learning_rate": 1.5782122905027933e-06, | |
| "loss": 2.8118, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.03142677561282212, | |
| "grad_norm": 23.991317749023438, | |
| "learning_rate": 1.787709497206704e-06, | |
| "loss": 2.6758, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.034569453174104335, | |
| "grad_norm": 27.5322265625, | |
| "learning_rate": 1.9972067039106146e-06, | |
| "loss": 2.5263, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.03771213073538655, | |
| "grad_norm": 18.619874954223633, | |
| "learning_rate": 2.2067039106145253e-06, | |
| "loss": 2.4554, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.04085480829666876, | |
| "grad_norm": 17.445724487304688, | |
| "learning_rate": 2.416201117318436e-06, | |
| "loss": 2.3279, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.04399748585795098, | |
| "grad_norm": 19.981834411621094, | |
| "learning_rate": 2.6256983240223464e-06, | |
| "loss": 2.2568, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.04714016341923319, | |
| "grad_norm": 15.571565628051758, | |
| "learning_rate": 2.8351955307262576e-06, | |
| "loss": 2.1021, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.0502828409805154, | |
| "grad_norm": 13.631041526794434, | |
| "learning_rate": 3.044692737430168e-06, | |
| "loss": 1.9724, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.05342551854179761, | |
| "grad_norm": 10.806873321533203, | |
| "learning_rate": 3.2541899441340786e-06, | |
| "loss": 1.8049, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.056568196103079824, | |
| "grad_norm": 14.863541603088379, | |
| "learning_rate": 3.4636871508379893e-06, | |
| "loss": 1.7358, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.059710873664362035, | |
| "grad_norm": 12.929265022277832, | |
| "learning_rate": 3.6731843575418996e-06, | |
| "loss": 1.6953, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.06285355122564425, | |
| "grad_norm": 12.591845512390137, | |
| "learning_rate": 3.88268156424581e-06, | |
| "loss": 1.6874, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.06599622878692646, | |
| "grad_norm": 13.552529335021973, | |
| "learning_rate": 4.0921787709497215e-06, | |
| "loss": 1.6485, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.06913890634820867, | |
| "grad_norm": 9.15485668182373, | |
| "learning_rate": 4.301675977653632e-06, | |
| "loss": 1.6044, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.07228158390949088, | |
| "grad_norm": 7.887852668762207, | |
| "learning_rate": 4.511173184357542e-06, | |
| "loss": 1.4664, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.0754242614707731, | |
| "grad_norm": 12.257076263427734, | |
| "learning_rate": 4.7206703910614525e-06, | |
| "loss": 1.5338, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.0785669390320553, | |
| "grad_norm": 8.181825637817383, | |
| "learning_rate": 4.930167597765364e-06, | |
| "loss": 1.4954, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.08170961659333752, | |
| "grad_norm": 9.838497161865234, | |
| "learning_rate": 5.139664804469274e-06, | |
| "loss": 1.4217, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.08485229415461974, | |
| "grad_norm": 15.192863464355469, | |
| "learning_rate": 5.349162011173184e-06, | |
| "loss": 1.4725, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.08799497171590195, | |
| "grad_norm": 7.013632297515869, | |
| "learning_rate": 5.558659217877096e-06, | |
| "loss": 1.3989, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.09113764927718417, | |
| "grad_norm": 8.598616600036621, | |
| "learning_rate": 5.768156424581007e-06, | |
| "loss": 1.4828, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.09428032683846638, | |
| "grad_norm": 5.625289440155029, | |
| "learning_rate": 5.977653631284917e-06, | |
| "loss": 1.4513, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.09742300439974859, | |
| "grad_norm": 7.280072212219238, | |
| "learning_rate": 6.187150837988828e-06, | |
| "loss": 1.4092, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.1005656819610308, | |
| "grad_norm": 7.608973979949951, | |
| "learning_rate": 6.396648044692738e-06, | |
| "loss": 1.4124, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.10370835952231301, | |
| "grad_norm": 8.430741310119629, | |
| "learning_rate": 6.606145251396649e-06, | |
| "loss": 1.4466, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.10685103708359522, | |
| "grad_norm": 8.642643928527832, | |
| "learning_rate": 6.815642458100559e-06, | |
| "loss": 1.3558, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.10999371464487744, | |
| "grad_norm": 6.632693290710449, | |
| "learning_rate": 7.02513966480447e-06, | |
| "loss": 1.4129, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.11313639220615965, | |
| "grad_norm": 12.695171356201172, | |
| "learning_rate": 7.2346368715083805e-06, | |
| "loss": 1.3444, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.11627906976744186, | |
| "grad_norm": 5.393228054046631, | |
| "learning_rate": 7.444134078212291e-06, | |
| "loss": 1.3485, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.11942174732872407, | |
| "grad_norm": 8.173951148986816, | |
| "learning_rate": 7.653631284916202e-06, | |
| "loss": 1.4175, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.12256442489000628, | |
| "grad_norm": 14.139352798461914, | |
| "learning_rate": 7.863128491620112e-06, | |
| "loss": 1.3797, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.1257071024512885, | |
| "grad_norm": 4.9999613761901855, | |
| "learning_rate": 8.072625698324023e-06, | |
| "loss": 1.3716, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.12884978001257072, | |
| "grad_norm": 4.344189643859863, | |
| "learning_rate": 8.282122905027935e-06, | |
| "loss": 1.3011, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.13199245757385292, | |
| "grad_norm": 11.033843040466309, | |
| "learning_rate": 8.491620111731845e-06, | |
| "loss": 1.3672, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.13513513513513514, | |
| "grad_norm": 4.3888068199157715, | |
| "learning_rate": 8.701117318435755e-06, | |
| "loss": 1.2773, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.13827781269641734, | |
| "grad_norm": 4.085406303405762, | |
| "learning_rate": 8.910614525139666e-06, | |
| "loss": 1.3057, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.14142049025769957, | |
| "grad_norm": 11.730964660644531, | |
| "learning_rate": 9.120111731843576e-06, | |
| "loss": 1.3752, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.14456316781898176, | |
| "grad_norm": 4.986327648162842, | |
| "learning_rate": 9.329608938547486e-06, | |
| "loss": 1.3055, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.147705845380264, | |
| "grad_norm": 5.643145561218262, | |
| "learning_rate": 9.539106145251398e-06, | |
| "loss": 1.3608, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.1508485229415462, | |
| "grad_norm": 12.661754608154297, | |
| "learning_rate": 9.748603351955308e-06, | |
| "loss": 1.4083, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.1539912005028284, | |
| "grad_norm": 6.018293857574463, | |
| "learning_rate": 9.958100558659219e-06, | |
| "loss": 1.2323, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.1571338780641106, | |
| "grad_norm": 4.150609493255615, | |
| "learning_rate": 1.0167597765363129e-05, | |
| "loss": 1.3068, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.16027655562539284, | |
| "grad_norm": 4.229065895080566, | |
| "learning_rate": 1.037709497206704e-05, | |
| "loss": 1.3257, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.16341923318667503, | |
| "grad_norm": 9.254191398620605, | |
| "learning_rate": 1.0586592178770951e-05, | |
| "loss": 1.2674, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.16656191074795726, | |
| "grad_norm": 3.52528977394104, | |
| "learning_rate": 1.0796089385474862e-05, | |
| "loss": 1.271, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.16970458830923948, | |
| "grad_norm": 6.313427448272705, | |
| "learning_rate": 1.1005586592178772e-05, | |
| "loss": 1.2948, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.17284726587052168, | |
| "grad_norm": 3.691848039627075, | |
| "learning_rate": 1.1215083798882682e-05, | |
| "loss": 1.292, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.1759899434318039, | |
| "grad_norm": 3.9371867179870605, | |
| "learning_rate": 1.1424581005586593e-05, | |
| "loss": 1.3135, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.1791326209930861, | |
| "grad_norm": 3.455486297607422, | |
| "learning_rate": 1.1634078212290503e-05, | |
| "loss": 1.2495, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.18227529855436833, | |
| "grad_norm": 10.44939136505127, | |
| "learning_rate": 1.1843575418994415e-05, | |
| "loss": 1.356, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.18541797611565053, | |
| "grad_norm": 3.6111254692077637, | |
| "learning_rate": 1.2053072625698325e-05, | |
| "loss": 1.3196, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.18856065367693275, | |
| "grad_norm": 3.531568765640259, | |
| "learning_rate": 1.2262569832402236e-05, | |
| "loss": 1.2307, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.19170333123821495, | |
| "grad_norm": 3.490081310272217, | |
| "learning_rate": 1.2472067039106146e-05, | |
| "loss": 1.2445, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.19484600879949718, | |
| "grad_norm": 5.011629581451416, | |
| "learning_rate": 1.2681564245810056e-05, | |
| "loss": 1.3192, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.19798868636077938, | |
| "grad_norm": 7.499322414398193, | |
| "learning_rate": 1.2891061452513967e-05, | |
| "loss": 1.2827, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.2011313639220616, | |
| "grad_norm": 3.792930841445923, | |
| "learning_rate": 1.3100558659217879e-05, | |
| "loss": 1.336, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.2042740414833438, | |
| "grad_norm": 4.43565559387207, | |
| "learning_rate": 1.3310055865921789e-05, | |
| "loss": 1.3371, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.20741671904462602, | |
| "grad_norm": 4.2933173179626465, | |
| "learning_rate": 1.3519553072625699e-05, | |
| "loss": 1.3118, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.21055939660590822, | |
| "grad_norm": 3.68387770652771, | |
| "learning_rate": 1.372905027932961e-05, | |
| "loss": 1.3975, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.21370207416719045, | |
| "grad_norm": 5.9917707443237305, | |
| "learning_rate": 1.393854748603352e-05, | |
| "loss": 1.2626, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.21684475172847265, | |
| "grad_norm": 3.1892921924591064, | |
| "learning_rate": 1.414804469273743e-05, | |
| "loss": 1.2781, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.21998742928975487, | |
| "grad_norm": 3.4965710639953613, | |
| "learning_rate": 1.4357541899441342e-05, | |
| "loss": 1.2557, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.2231301068510371, | |
| "grad_norm": 7.907847881317139, | |
| "learning_rate": 1.4567039106145252e-05, | |
| "loss": 1.2341, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.2262727844123193, | |
| "grad_norm": 3.6374635696411133, | |
| "learning_rate": 1.4776536312849163e-05, | |
| "loss": 1.3225, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.22941546197360152, | |
| "grad_norm": 6.52725887298584, | |
| "learning_rate": 1.4986033519553073e-05, | |
| "loss": 1.2137, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.23255813953488372, | |
| "grad_norm": 4.659281253814697, | |
| "learning_rate": 1.5195530726256983e-05, | |
| "loss": 1.3373, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.23570081709616594, | |
| "grad_norm": 3.1662540435791016, | |
| "learning_rate": 1.5405027932960895e-05, | |
| "loss": 1.3348, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.23884349465744814, | |
| "grad_norm": 3.8003756999969482, | |
| "learning_rate": 1.5614525139664806e-05, | |
| "loss": 1.2948, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.24198617221873037, | |
| "grad_norm": 3.0090460777282715, | |
| "learning_rate": 1.5824022346368716e-05, | |
| "loss": 1.2724, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 0.24512884978001256, | |
| "grad_norm": 3.1680350303649902, | |
| "learning_rate": 1.6033519553072626e-05, | |
| "loss": 1.2402, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.2482715273412948, | |
| "grad_norm": 3.944425106048584, | |
| "learning_rate": 1.6243016759776537e-05, | |
| "loss": 1.303, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 0.251414204902577, | |
| "grad_norm": 8.234729766845703, | |
| "learning_rate": 1.6452513966480447e-05, | |
| "loss": 1.2196, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.2545568824638592, | |
| "grad_norm": 6.725916862487793, | |
| "learning_rate": 1.6662011173184357e-05, | |
| "loss": 1.3213, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 0.25769956002514144, | |
| "grad_norm": 3.2696242332458496, | |
| "learning_rate": 1.687150837988827e-05, | |
| "loss": 1.2611, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.2608422375864236, | |
| "grad_norm": 3.0667693614959717, | |
| "learning_rate": 1.708100558659218e-05, | |
| "loss": 1.22, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 0.26398491514770583, | |
| "grad_norm": 3.5260438919067383, | |
| "learning_rate": 1.729050279329609e-05, | |
| "loss": 1.3142, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.26712759270898806, | |
| "grad_norm": 3.51269793510437, | |
| "learning_rate": 1.7500000000000002e-05, | |
| "loss": 1.2031, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 0.2702702702702703, | |
| "grad_norm": 2.9083192348480225, | |
| "learning_rate": 1.7709497206703912e-05, | |
| "loss": 1.1711, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.27341294783155246, | |
| "grad_norm": 2.915454626083374, | |
| "learning_rate": 1.7918994413407822e-05, | |
| "loss": 1.2452, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 0.2765556253928347, | |
| "grad_norm": 4.736166954040527, | |
| "learning_rate": 1.8128491620111733e-05, | |
| "loss": 1.2251, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.2796983029541169, | |
| "grad_norm": 2.881985664367676, | |
| "learning_rate": 1.8337988826815643e-05, | |
| "loss": 1.2541, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 0.28284098051539913, | |
| "grad_norm": 4.29525899887085, | |
| "learning_rate": 1.8547486033519553e-05, | |
| "loss": 1.2336, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.28598365807668136, | |
| "grad_norm": 3.046929359436035, | |
| "learning_rate": 1.8756983240223464e-05, | |
| "loss": 1.2576, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 0.2891263356379635, | |
| "grad_norm": 4.023289680480957, | |
| "learning_rate": 1.8966480446927374e-05, | |
| "loss": 1.2371, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.29226901319924575, | |
| "grad_norm": 6.221644878387451, | |
| "learning_rate": 1.9175977653631284e-05, | |
| "loss": 1.1985, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 0.295411690760528, | |
| "grad_norm": 6.021862030029297, | |
| "learning_rate": 1.9385474860335195e-05, | |
| "loss": 1.2741, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.2985543683218102, | |
| "grad_norm": 6.475122451782227, | |
| "learning_rate": 1.959497206703911e-05, | |
| "loss": 1.2405, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 0.3016970458830924, | |
| "grad_norm": 6.075967788696289, | |
| "learning_rate": 1.980446927374302e-05, | |
| "loss": 1.1969, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.3048397234443746, | |
| "grad_norm": 2.591923713684082, | |
| "learning_rate": 1.9999999702856602e-05, | |
| "loss": 1.1979, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 0.3079824010056568, | |
| "grad_norm": 6.351642608642578, | |
| "learning_rate": 1.9999923931385287e-05, | |
| "loss": 1.2887, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.31112507856693905, | |
| "grad_norm": 2.5136075019836426, | |
| "learning_rate": 1.9999714446548786e-05, | |
| "loss": 1.2974, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 0.3142677561282212, | |
| "grad_norm": 3.683798313140869, | |
| "learning_rate": 1.9999371251148214e-05, | |
| "loss": 1.2073, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.31741043368950345, | |
| "grad_norm": 6.616783142089844, | |
| "learning_rate": 1.9998894349772585e-05, | |
| "loss": 1.224, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 0.32055311125078567, | |
| "grad_norm": 3.386265277862549, | |
| "learning_rate": 1.9998283748798757e-05, | |
| "loss": 1.2237, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.3236957888120679, | |
| "grad_norm": 3.2038071155548096, | |
| "learning_rate": 1.9997539456391343e-05, | |
| "loss": 1.2805, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 0.32683846637335007, | |
| "grad_norm": 5.434456825256348, | |
| "learning_rate": 1.9996661482502602e-05, | |
| "loss": 1.295, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.3299811439346323, | |
| "grad_norm": 3.0071728229522705, | |
| "learning_rate": 1.999564983887231e-05, | |
| "loss": 1.2207, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 0.3331238214959145, | |
| "grad_norm": 4.010280609130859, | |
| "learning_rate": 1.9994504539027594e-05, | |
| "loss": 1.2005, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.33626649905719674, | |
| "grad_norm": 4.2148237228393555, | |
| "learning_rate": 1.9993225598282768e-05, | |
| "loss": 1.1608, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 0.33940917661847897, | |
| "grad_norm": 5.541611194610596, | |
| "learning_rate": 1.9991813033739097e-05, | |
| "loss": 1.1649, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.34255185417976114, | |
| "grad_norm": 5.37580680847168, | |
| "learning_rate": 1.9990266864284607e-05, | |
| "loss": 1.2087, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 0.34569453174104336, | |
| "grad_norm": 3.72402286529541, | |
| "learning_rate": 1.998858711059381e-05, | |
| "loss": 1.2606, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.3488372093023256, | |
| "grad_norm": 3.596006393432617, | |
| "learning_rate": 1.9986773795127425e-05, | |
| "loss": 1.2509, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 0.3519798868636078, | |
| "grad_norm": 2.3336808681488037, | |
| "learning_rate": 1.9984826942132087e-05, | |
| "loss": 1.2389, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.35512256442489, | |
| "grad_norm": 2.766104221343994, | |
| "learning_rate": 1.9982746577640024e-05, | |
| "loss": 1.2334, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 0.3582652419861722, | |
| "grad_norm": 2.6721198558807373, | |
| "learning_rate": 1.9980532729468698e-05, | |
| "loss": 1.2069, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.36140791954745444, | |
| "grad_norm": 2.4807119369506836, | |
| "learning_rate": 1.9978185427220447e-05, | |
| "loss": 1.264, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 0.36455059710873666, | |
| "grad_norm": 2.6071507930755615, | |
| "learning_rate": 1.9975704702282074e-05, | |
| "loss": 1.2745, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.36769327467001883, | |
| "grad_norm": 2.5577876567840576, | |
| "learning_rate": 1.9973090587824443e-05, | |
| "loss": 1.2244, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 0.37083595223130106, | |
| "grad_norm": 5.360003471374512, | |
| "learning_rate": 1.997034311880202e-05, | |
| "loss": 1.1723, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.3739786297925833, | |
| "grad_norm": 2.606682777404785, | |
| "learning_rate": 1.996746233195242e-05, | |
| "loss": 1.2349, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 0.3771213073538655, | |
| "grad_norm": 2.898453712463379, | |
| "learning_rate": 1.9964448265795903e-05, | |
| "loss": 1.2428, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.3802639849151477, | |
| "grad_norm": 2.8664214611053467, | |
| "learning_rate": 1.9961300960634876e-05, | |
| "loss": 1.2338, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 0.3834066624764299, | |
| "grad_norm": 3.170452117919922, | |
| "learning_rate": 1.9958020458553326e-05, | |
| "loss": 1.1729, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.38654934003771213, | |
| "grad_norm": 2.6928646564483643, | |
| "learning_rate": 1.995460680341629e-05, | |
| "loss": 1.2791, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 0.38969201759899436, | |
| "grad_norm": 5.1588287353515625, | |
| "learning_rate": 1.9951060040869245e-05, | |
| "loss": 1.1712, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.3928346951602766, | |
| "grad_norm": 4.63333797454834, | |
| "learning_rate": 1.994738021833751e-05, | |
| "loss": 1.285, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 0.39597737272155875, | |
| "grad_norm": 2.860305070877075, | |
| "learning_rate": 1.9943567385025605e-05, | |
| "loss": 1.1853, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.399120050282841, | |
| "grad_norm": 3.4676976203918457, | |
| "learning_rate": 1.9939621591916587e-05, | |
| "loss": 1.2164, | |
| "step": 1905 | |
| }, | |
| { | |
| "epoch": 0.4022627278441232, | |
| "grad_norm": 4.743997573852539, | |
| "learning_rate": 1.9935542891771392e-05, | |
| "loss": 1.2182, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.40540540540540543, | |
| "grad_norm": 4.906563758850098, | |
| "learning_rate": 1.9931331339128107e-05, | |
| "loss": 1.1844, | |
| "step": 1935 | |
| }, | |
| { | |
| "epoch": 0.4085480829666876, | |
| "grad_norm": 2.863513946533203, | |
| "learning_rate": 1.9926986990301244e-05, | |
| "loss": 1.2124, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.4116907605279698, | |
| "grad_norm": 2.1002469062805176, | |
| "learning_rate": 1.9922509903381e-05, | |
| "loss": 1.1589, | |
| "step": 1965 | |
| }, | |
| { | |
| "epoch": 0.41483343808925205, | |
| "grad_norm": 5.061679840087891, | |
| "learning_rate": 1.991790013823246e-05, | |
| "loss": 1.2208, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.4179761156505343, | |
| "grad_norm": 5.046812534332275, | |
| "learning_rate": 1.9913157756494822e-05, | |
| "loss": 1.2122, | |
| "step": 1995 | |
| }, | |
| { | |
| "epoch": 0.42111879321181644, | |
| "grad_norm": 2.990056037902832, | |
| "learning_rate": 1.9908282821580544e-05, | |
| "loss": 1.2296, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.42426147077309867, | |
| "grad_norm": 2.477032423019409, | |
| "learning_rate": 1.9903275398674523e-05, | |
| "loss": 1.2259, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 0.4274041483343809, | |
| "grad_norm": 2.7786192893981934, | |
| "learning_rate": 1.9898135554733198e-05, | |
| "loss": 1.2062, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.4305468258956631, | |
| "grad_norm": 2.325221061706543, | |
| "learning_rate": 1.989286335848368e-05, | |
| "loss": 1.2676, | |
| "step": 2055 | |
| }, | |
| { | |
| "epoch": 0.4336895034569453, | |
| "grad_norm": 2.26428484916687, | |
| "learning_rate": 1.988745888042281e-05, | |
| "loss": 1.1832, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.4368321810182275, | |
| "grad_norm": 2.361360788345337, | |
| "learning_rate": 1.988192219281624e-05, | |
| "loss": 1.1814, | |
| "step": 2085 | |
| }, | |
| { | |
| "epoch": 0.43997485857950974, | |
| "grad_norm": 2.2548787593841553, | |
| "learning_rate": 1.987625336969744e-05, | |
| "loss": 1.1371, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.44311753614079197, | |
| "grad_norm": 2.9371745586395264, | |
| "learning_rate": 1.9870452486866733e-05, | |
| "loss": 1.1185, | |
| "step": 2115 | |
| }, | |
| { | |
| "epoch": 0.4462602137020742, | |
| "grad_norm": 2.250372886657715, | |
| "learning_rate": 1.9864519621890263e-05, | |
| "loss": 1.1919, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.44940289126335636, | |
| "grad_norm": 2.822431802749634, | |
| "learning_rate": 1.9858454854098968e-05, | |
| "loss": 1.1479, | |
| "step": 2145 | |
| }, | |
| { | |
| "epoch": 0.4525455688246386, | |
| "grad_norm": 5.055064678192139, | |
| "learning_rate": 1.985225826458751e-05, | |
| "loss": 1.2237, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.4556882463859208, | |
| "grad_norm": 2.3341803550720215, | |
| "learning_rate": 1.9845929936213216e-05, | |
| "loss": 1.1988, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 0.45883092394720304, | |
| "grad_norm": 2.5791430473327637, | |
| "learning_rate": 1.9839469953594927e-05, | |
| "loss": 1.19, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.4619736015084852, | |
| "grad_norm": 3.167426824569702, | |
| "learning_rate": 1.9832878403111908e-05, | |
| "loss": 1.1852, | |
| "step": 2205 | |
| }, | |
| { | |
| "epoch": 0.46511627906976744, | |
| "grad_norm": 4.721558570861816, | |
| "learning_rate": 1.9826155372902667e-05, | |
| "loss": 1.198, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.46825895663104966, | |
| "grad_norm": 2.5482370853424072, | |
| "learning_rate": 1.9819300952863783e-05, | |
| "loss": 1.1782, | |
| "step": 2235 | |
| }, | |
| { | |
| "epoch": 0.4714016341923319, | |
| "grad_norm": 3.1147351264953613, | |
| "learning_rate": 1.9812315234648714e-05, | |
| "loss": 1.2138, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.47454431175361406, | |
| "grad_norm": 4.730167865753174, | |
| "learning_rate": 1.980519831166656e-05, | |
| "loss": 1.1622, | |
| "step": 2265 | |
| }, | |
| { | |
| "epoch": 0.4776869893148963, | |
| "grad_norm": 1.9187180995941162, | |
| "learning_rate": 1.979795027908082e-05, | |
| "loss": 1.1624, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.4808296668761785, | |
| "grad_norm": 2.371293067932129, | |
| "learning_rate": 1.9790571233808108e-05, | |
| "loss": 1.1823, | |
| "step": 2295 | |
| }, | |
| { | |
| "epoch": 0.48397234443746073, | |
| "grad_norm": 2.103717803955078, | |
| "learning_rate": 1.978306127451687e-05, | |
| "loss": 1.2298, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.4871150219987429, | |
| "grad_norm": 3.5898678302764893, | |
| "learning_rate": 1.9775420501626073e-05, | |
| "loss": 1.2612, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 0.49025769956002513, | |
| "grad_norm": 2.680027723312378, | |
| "learning_rate": 1.976764901730383e-05, | |
| "loss": 1.2161, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.49340037712130735, | |
| "grad_norm": 2.514291286468506, | |
| "learning_rate": 1.9759746925466073e-05, | |
| "loss": 1.1772, | |
| "step": 2355 | |
| }, | |
| { | |
| "epoch": 0.4965430546825896, | |
| "grad_norm": 2.453972816467285, | |
| "learning_rate": 1.9751714331775125e-05, | |
| "loss": 1.1982, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.4996857322438718, | |
| "grad_norm": 6.017053604125977, | |
| "learning_rate": 1.9743551343638324e-05, | |
| "loss": 1.181, | |
| "step": 2385 | |
| }, | |
| { | |
| "epoch": 0.502828409805154, | |
| "grad_norm": 2.46274733543396, | |
| "learning_rate": 1.973525807020656e-05, | |
| "loss": 1.1537, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.5059710873664363, | |
| "grad_norm": 2.6295454502105713, | |
| "learning_rate": 1.972683462237282e-05, | |
| "loss": 1.1961, | |
| "step": 2415 | |
| }, | |
| { | |
| "epoch": 0.5091137649277184, | |
| "grad_norm": 5.519942760467529, | |
| "learning_rate": 1.9718281112770723e-05, | |
| "loss": 1.1998, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.5122564424890006, | |
| "grad_norm": 2.5744876861572266, | |
| "learning_rate": 1.9709597655772993e-05, | |
| "loss": 1.1511, | |
| "step": 2445 | |
| }, | |
| { | |
| "epoch": 0.5153991200502829, | |
| "grad_norm": 2.621332883834839, | |
| "learning_rate": 1.970078436748994e-05, | |
| "loss": 1.1398, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.518541797611565, | |
| "grad_norm": 5.679701328277588, | |
| "learning_rate": 1.9691841365767904e-05, | |
| "loss": 1.1899, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 0.5216844751728472, | |
| "grad_norm": 2.662759780883789, | |
| "learning_rate": 1.9682768770187675e-05, | |
| "loss": 1.1442, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.5248271527341295, | |
| "grad_norm": 2.7816829681396484, | |
| "learning_rate": 1.967356670206291e-05, | |
| "loss": 1.2164, | |
| "step": 2505 | |
| }, | |
| { | |
| "epoch": 0.5279698302954117, | |
| "grad_norm": 5.280804634094238, | |
| "learning_rate": 1.9664235284438495e-05, | |
| "loss": 1.186, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.531112507856694, | |
| "grad_norm": 4.3773112297058105, | |
| "learning_rate": 1.96547746420889e-05, | |
| "loss": 1.2107, | |
| "step": 2535 | |
| }, | |
| { | |
| "epoch": 0.5342551854179761, | |
| "grad_norm": 2.773582696914673, | |
| "learning_rate": 1.9645184901516525e-05, | |
| "loss": 1.2216, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.5373978629792583, | |
| "grad_norm": 2.643582820892334, | |
| "learning_rate": 1.9635466190949987e-05, | |
| "loss": 1.1888, | |
| "step": 2565 | |
| }, | |
| { | |
| "epoch": 0.5405405405405406, | |
| "grad_norm": 2.277792453765869, | |
| "learning_rate": 1.962561864034243e-05, | |
| "loss": 1.225, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.5436832181018227, | |
| "grad_norm": 2.580367088317871, | |
| "learning_rate": 1.9615642381369765e-05, | |
| "loss": 1.2064, | |
| "step": 2595 | |
| }, | |
| { | |
| "epoch": 0.5468258956631049, | |
| "grad_norm": 4.304893493652344, | |
| "learning_rate": 1.9605537547428922e-05, | |
| "loss": 1.1652, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.5499685732243872, | |
| "grad_norm": 2.5215561389923096, | |
| "learning_rate": 1.959530427363607e-05, | |
| "loss": 1.1909, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 0.5531112507856694, | |
| "grad_norm": 3.4145307540893555, | |
| "learning_rate": 1.958494269682479e-05, | |
| "loss": 1.1211, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.5562539283469516, | |
| "grad_norm": 4.1698079109191895, | |
| "learning_rate": 1.9574452955544274e-05, | |
| "loss": 1.1235, | |
| "step": 2655 | |
| }, | |
| { | |
| "epoch": 0.5593966059082338, | |
| "grad_norm": 4.188419818878174, | |
| "learning_rate": 1.9563835190057442e-05, | |
| "loss": 1.1925, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.562539283469516, | |
| "grad_norm": 4.869725704193115, | |
| "learning_rate": 1.9553089542339093e-05, | |
| "loss": 1.172, | |
| "step": 2685 | |
| }, | |
| { | |
| "epoch": 0.5656819610307983, | |
| "grad_norm": 4.675424575805664, | |
| "learning_rate": 1.9542216156073994e-05, | |
| "loss": 1.218, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.5688246385920804, | |
| "grad_norm": 2.655273914337158, | |
| "learning_rate": 1.9531215176654953e-05, | |
| "loss": 1.128, | |
| "step": 2715 | |
| }, | |
| { | |
| "epoch": 0.5719673161533627, | |
| "grad_norm": 3.345499277114868, | |
| "learning_rate": 1.9520086751180894e-05, | |
| "loss": 1.2236, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 0.5751099937146449, | |
| "grad_norm": 2.5058796405792236, | |
| "learning_rate": 1.9508831028454862e-05, | |
| "loss": 1.1895, | |
| "step": 2745 | |
| }, | |
| { | |
| "epoch": 0.578252671275927, | |
| "grad_norm": 2.2655258178710938, | |
| "learning_rate": 1.9497448158982062e-05, | |
| "loss": 1.1486, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.5813953488372093, | |
| "grad_norm": 4.1772894859313965, | |
| "learning_rate": 1.948593829496783e-05, | |
| "loss": 1.1616, | |
| "step": 2775 | |
| }, | |
| { | |
| "epoch": 0.5845380263984915, | |
| "grad_norm": 2.290619373321533, | |
| "learning_rate": 1.9474301590315605e-05, | |
| "loss": 1.1508, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 0.5876807039597737, | |
| "grad_norm": 2.4885716438293457, | |
| "learning_rate": 1.9462538200624855e-05, | |
| "loss": 1.1727, | |
| "step": 2805 | |
| }, | |
| { | |
| "epoch": 0.590823381521056, | |
| "grad_norm": 2.564606189727783, | |
| "learning_rate": 1.9450648283189035e-05, | |
| "loss": 1.2178, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.5939660590823381, | |
| "grad_norm": 2.756286859512329, | |
| "learning_rate": 1.9438631996993424e-05, | |
| "loss": 1.2143, | |
| "step": 2835 | |
| }, | |
| { | |
| "epoch": 0.5971087366436204, | |
| "grad_norm": 4.876669883728027, | |
| "learning_rate": 1.9426489502713062e-05, | |
| "loss": 1.1908, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.6002514142049026, | |
| "grad_norm": 1.9889987707138062, | |
| "learning_rate": 1.9414220962710556e-05, | |
| "loss": 1.1404, | |
| "step": 2865 | |
| }, | |
| { | |
| "epoch": 0.6033940917661847, | |
| "grad_norm": 2.6921072006225586, | |
| "learning_rate": 1.9401826541033935e-05, | |
| "loss": 1.2159, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.606536769327467, | |
| "grad_norm": 1.9745949506759644, | |
| "learning_rate": 1.9389306403414438e-05, | |
| "loss": 1.1828, | |
| "step": 2895 | |
| }, | |
| { | |
| "epoch": 0.6096794468887492, | |
| "grad_norm": 2.012610912322998, | |
| "learning_rate": 1.9376660717264316e-05, | |
| "loss": 1.1413, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 0.6128221244500315, | |
| "grad_norm": 3.3427295684814453, | |
| "learning_rate": 1.9363889651674575e-05, | |
| "loss": 1.1888, | |
| "step": 2925 | |
| }, | |
| { | |
| "epoch": 0.6159648020113137, | |
| "grad_norm": 3.9471700191497803, | |
| "learning_rate": 1.935099337741274e-05, | |
| "loss": 1.1882, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.6191074795725958, | |
| "grad_norm": 2.1639480590820312, | |
| "learning_rate": 1.9337972066920533e-05, | |
| "loss": 1.1518, | |
| "step": 2955 | |
| }, | |
| { | |
| "epoch": 0.6222501571338781, | |
| "grad_norm": 2.18038272857666, | |
| "learning_rate": 1.9324825894311613e-05, | |
| "loss": 1.1424, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 0.6253928346951603, | |
| "grad_norm": 2.4745776653289795, | |
| "learning_rate": 1.9311555035369206e-05, | |
| "loss": 1.2065, | |
| "step": 2985 | |
| }, | |
| { | |
| "epoch": 0.6285355122564424, | |
| "grad_norm": 2.4123799800872803, | |
| "learning_rate": 1.9298159667543795e-05, | |
| "loss": 1.1677, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.6316781898177247, | |
| "grad_norm": 2.8787992000579834, | |
| "learning_rate": 1.92846399699507e-05, | |
| "loss": 1.2035, | |
| "step": 3015 | |
| }, | |
| { | |
| "epoch": 0.6348208673790069, | |
| "grad_norm": 2.3905246257781982, | |
| "learning_rate": 1.927099612336773e-05, | |
| "loss": 1.1933, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 0.6379635449402892, | |
| "grad_norm": 3.8858561515808105, | |
| "learning_rate": 1.9257228310232738e-05, | |
| "loss": 1.1913, | |
| "step": 3045 | |
| }, | |
| { | |
| "epoch": 0.6411062225015713, | |
| "grad_norm": 3.567950963973999, | |
| "learning_rate": 1.924333671464118e-05, | |
| "loss": 1.2417, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.6442489000628535, | |
| "grad_norm": 1.8170944452285767, | |
| "learning_rate": 1.9229321522343677e-05, | |
| "loss": 1.1476, | |
| "step": 3075 | |
| }, | |
| { | |
| "epoch": 0.6473915776241358, | |
| "grad_norm": 3.044593095779419, | |
| "learning_rate": 1.9215182920743502e-05, | |
| "loss": 1.1939, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 0.650534255185418, | |
| "grad_norm": 3.13097882270813, | |
| "learning_rate": 1.9200921098894094e-05, | |
| "loss": 1.2007, | |
| "step": 3105 | |
| }, | |
| { | |
| "epoch": 0.6536769327467001, | |
| "grad_norm": 2.1468851566314697, | |
| "learning_rate": 1.918653624749652e-05, | |
| "loss": 1.1587, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.6568196103079824, | |
| "grad_norm": 2.9996323585510254, | |
| "learning_rate": 1.9172028558896935e-05, | |
| "loss": 1.1762, | |
| "step": 3135 | |
| }, | |
| { | |
| "epoch": 0.6599622878692646, | |
| "grad_norm": 2.5897107124328613, | |
| "learning_rate": 1.9157398227084002e-05, | |
| "loss": 1.1334, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.6631049654305469, | |
| "grad_norm": 3.80074143409729, | |
| "learning_rate": 1.91426454476863e-05, | |
| "loss": 1.2642, | |
| "step": 3165 | |
| }, | |
| { | |
| "epoch": 0.666247642991829, | |
| "grad_norm": 1.9756152629852295, | |
| "learning_rate": 1.9127770417969703e-05, | |
| "loss": 1.1418, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 0.6693903205531112, | |
| "grad_norm": 1.7655247449874878, | |
| "learning_rate": 1.9112773336834754e-05, | |
| "loss": 1.1605, | |
| "step": 3195 | |
| }, | |
| { | |
| "epoch": 0.6725329981143935, | |
| "grad_norm": 2.2654428482055664, | |
| "learning_rate": 1.9097654404814e-05, | |
| "loss": 1.2366, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 0.6756756756756757, | |
| "grad_norm": 2.2166404724121094, | |
| "learning_rate": 1.9082413824069302e-05, | |
| "loss": 1.1469, | |
| "step": 3225 | |
| }, | |
| { | |
| "epoch": 0.6788183532369579, | |
| "grad_norm": 2.409104824066162, | |
| "learning_rate": 1.9067051798389148e-05, | |
| "loss": 1.1303, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 0.6819610307982401, | |
| "grad_norm": 2.2395966053009033, | |
| "learning_rate": 1.905156853318591e-05, | |
| "loss": 1.1507, | |
| "step": 3255 | |
| }, | |
| { | |
| "epoch": 0.6851037083595223, | |
| "grad_norm": 2.369009017944336, | |
| "learning_rate": 1.903596423549312e-05, | |
| "loss": 1.1468, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 0.6882463859208046, | |
| "grad_norm": 4.745908260345459, | |
| "learning_rate": 1.9020239113962677e-05, | |
| "loss": 1.1675, | |
| "step": 3285 | |
| }, | |
| { | |
| "epoch": 0.6913890634820867, | |
| "grad_norm": 2.9272751808166504, | |
| "learning_rate": 1.900439337886207e-05, | |
| "loss": 1.1604, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.6945317410433689, | |
| "grad_norm": 1.923449993133545, | |
| "learning_rate": 1.898842724207157e-05, | |
| "loss": 1.136, | |
| "step": 3315 | |
| }, | |
| { | |
| "epoch": 0.6976744186046512, | |
| "grad_norm": 2.657238006591797, | |
| "learning_rate": 1.897234091708139e-05, | |
| "loss": 1.212, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 0.7008170961659334, | |
| "grad_norm": 2.484586715698242, | |
| "learning_rate": 1.8956134618988828e-05, | |
| "loss": 1.1938, | |
| "step": 3345 | |
| }, | |
| { | |
| "epoch": 0.7039597737272156, | |
| "grad_norm": 2.127962589263916, | |
| "learning_rate": 1.89398085644954e-05, | |
| "loss": 1.138, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 0.7071024512884978, | |
| "grad_norm": 4.683898448944092, | |
| "learning_rate": 1.8923362971903932e-05, | |
| "loss": 1.2393, | |
| "step": 3375 | |
| }, | |
| { | |
| "epoch": 0.71024512884978, | |
| "grad_norm": 2.090588092803955, | |
| "learning_rate": 1.8906798061115654e-05, | |
| "loss": 1.1721, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 0.7133878064110623, | |
| "grad_norm": 2.328129768371582, | |
| "learning_rate": 1.8890114053627244e-05, | |
| "loss": 1.1351, | |
| "step": 3405 | |
| }, | |
| { | |
| "epoch": 0.7165304839723444, | |
| "grad_norm": 2.18684458732605, | |
| "learning_rate": 1.8873311172527883e-05, | |
| "loss": 1.1793, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 0.7196731615336267, | |
| "grad_norm": 3.865600109100342, | |
| "learning_rate": 1.885638964249625e-05, | |
| "loss": 1.1605, | |
| "step": 3435 | |
| }, | |
| { | |
| "epoch": 0.7228158390949089, | |
| "grad_norm": 1.9220081567764282, | |
| "learning_rate": 1.8839349689797542e-05, | |
| "loss": 1.1056, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.725958516656191, | |
| "grad_norm": 2.822507858276367, | |
| "learning_rate": 1.8822191542280433e-05, | |
| "loss": 1.2242, | |
| "step": 3465 | |
| }, | |
| { | |
| "epoch": 0.7291011942174733, | |
| "grad_norm": 2.1855289936065674, | |
| "learning_rate": 1.880491542937404e-05, | |
| "loss": 1.1582, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 0.7322438717787555, | |
| "grad_norm": 2.4138951301574707, | |
| "learning_rate": 1.8787521582084833e-05, | |
| "loss": 1.1889, | |
| "step": 3495 | |
| }, | |
| { | |
| "epoch": 0.7353865493400377, | |
| "grad_norm": 4.04121732711792, | |
| "learning_rate": 1.8770010232993562e-05, | |
| "loss": 1.1682, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 0.73852922690132, | |
| "grad_norm": 2.7555859088897705, | |
| "learning_rate": 1.875238161625216e-05, | |
| "loss": 1.1242, | |
| "step": 3525 | |
| }, | |
| { | |
| "epoch": 0.7416719044626021, | |
| "grad_norm": 2.1336309909820557, | |
| "learning_rate": 1.873463596758058e-05, | |
| "loss": 1.1823, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 0.7448145820238844, | |
| "grad_norm": 2.1928298473358154, | |
| "learning_rate": 1.871677352426367e-05, | |
| "loss": 1.1785, | |
| "step": 3555 | |
| }, | |
| { | |
| "epoch": 0.7479572595851666, | |
| "grad_norm": 1.8896750211715698, | |
| "learning_rate": 1.8698794525147988e-05, | |
| "loss": 1.1871, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 0.7510999371464487, | |
| "grad_norm": 2.354185104370117, | |
| "learning_rate": 1.8680699210638608e-05, | |
| "loss": 1.072, | |
| "step": 3585 | |
| }, | |
| { | |
| "epoch": 0.754242614707731, | |
| "grad_norm": 2.3728153705596924, | |
| "learning_rate": 1.8662487822695906e-05, | |
| "loss": 1.1607, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.7573852922690132, | |
| "grad_norm": 1.997661828994751, | |
| "learning_rate": 1.8644160604832344e-05, | |
| "loss": 1.1824, | |
| "step": 3615 | |
| }, | |
| { | |
| "epoch": 0.7605279698302954, | |
| "grad_norm": 2.4399378299713135, | |
| "learning_rate": 1.8625717802109173e-05, | |
| "loss": 1.1794, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 0.7636706473915776, | |
| "grad_norm": 2.1330130100250244, | |
| "learning_rate": 1.860715966113319e-05, | |
| "loss": 1.2458, | |
| "step": 3645 | |
| }, | |
| { | |
| "epoch": 0.7668133249528598, | |
| "grad_norm": 2.0650925636291504, | |
| "learning_rate": 1.8588486430053442e-05, | |
| "loss": 1.1785, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 0.7699560025141421, | |
| "grad_norm": 2.570436477661133, | |
| "learning_rate": 1.856969835855787e-05, | |
| "loss": 1.154, | |
| "step": 3675 | |
| }, | |
| { | |
| "epoch": 0.7730986800754243, | |
| "grad_norm": 4.064576625823975, | |
| "learning_rate": 1.855079569787002e-05, | |
| "loss": 1.1833, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 0.7762413576367064, | |
| "grad_norm": 2.199369430541992, | |
| "learning_rate": 1.8531778700745652e-05, | |
| "loss": 1.1681, | |
| "step": 3705 | |
| }, | |
| { | |
| "epoch": 0.7793840351979887, | |
| "grad_norm": 2.3457729816436768, | |
| "learning_rate": 1.851264762146937e-05, | |
| "loss": 1.1177, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 0.7825267127592709, | |
| "grad_norm": 1.790176272392273, | |
| "learning_rate": 1.849340271585121e-05, | |
| "loss": 1.1603, | |
| "step": 3735 | |
| }, | |
| { | |
| "epoch": 0.7856693903205532, | |
| "grad_norm": 1.7440158128738403, | |
| "learning_rate": 1.8474044241223247e-05, | |
| "loss": 1.1867, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.7888120678818353, | |
| "grad_norm": 2.208021640777588, | |
| "learning_rate": 1.8454572456436126e-05, | |
| "loss": 1.1381, | |
| "step": 3765 | |
| }, | |
| { | |
| "epoch": 0.7919547454431175, | |
| "grad_norm": 2.744593858718872, | |
| "learning_rate": 1.8434987621855615e-05, | |
| "loss": 1.1186, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 0.7950974230043998, | |
| "grad_norm": 2.6648318767547607, | |
| "learning_rate": 1.841528999935912e-05, | |
| "loss": 1.1345, | |
| "step": 3795 | |
| }, | |
| { | |
| "epoch": 0.798240100565682, | |
| "grad_norm": 2.1822752952575684, | |
| "learning_rate": 1.8395479852332185e-05, | |
| "loss": 1.2148, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 0.8013827781269641, | |
| "grad_norm": 2.2398853302001953, | |
| "learning_rate": 1.8375557445664962e-05, | |
| "loss": 1.1915, | |
| "step": 3825 | |
| }, | |
| { | |
| "epoch": 0.8045254556882464, | |
| "grad_norm": 3.89241886138916, | |
| "learning_rate": 1.835552304574869e-05, | |
| "loss": 1.2088, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 0.8076681332495286, | |
| "grad_norm": 2.0591068267822266, | |
| "learning_rate": 1.8335376920472098e-05, | |
| "loss": 1.2107, | |
| "step": 3855 | |
| }, | |
| { | |
| "epoch": 0.8108108108108109, | |
| "grad_norm": 2.421823501586914, | |
| "learning_rate": 1.831511933921787e-05, | |
| "loss": 1.1696, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 0.813953488372093, | |
| "grad_norm": 3.9787890911102295, | |
| "learning_rate": 1.829475057285899e-05, | |
| "loss": 1.1756, | |
| "step": 3885 | |
| }, | |
| { | |
| "epoch": 0.8170961659333752, | |
| "grad_norm": 2.6043264865875244, | |
| "learning_rate": 1.8274270893755166e-05, | |
| "loss": 1.1729, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.8202388434946575, | |
| "grad_norm": 2.131624221801758, | |
| "learning_rate": 1.825368057574916e-05, | |
| "loss": 1.1465, | |
| "step": 3915 | |
| }, | |
| { | |
| "epoch": 0.8233815210559396, | |
| "grad_norm": 2.2352840900421143, | |
| "learning_rate": 1.823297989416314e-05, | |
| "loss": 1.1126, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 0.8265241986172219, | |
| "grad_norm": 1.9127349853515625, | |
| "learning_rate": 1.821216912579499e-05, | |
| "loss": 1.1081, | |
| "step": 3945 | |
| }, | |
| { | |
| "epoch": 0.8296668761785041, | |
| "grad_norm": 1.8138941526412964, | |
| "learning_rate": 1.819124854891461e-05, | |
| "loss": 1.1369, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 0.8328095537397863, | |
| "grad_norm": 3.4601540565490723, | |
| "learning_rate": 1.817021844326021e-05, | |
| "loss": 1.1378, | |
| "step": 3975 | |
| }, | |
| { | |
| "epoch": 0.8359522313010685, | |
| "grad_norm": 2.171710968017578, | |
| "learning_rate": 1.814907909003454e-05, | |
| "loss": 1.146, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 0.8390949088623507, | |
| "grad_norm": 2.376568078994751, | |
| "learning_rate": 1.8127830771901157e-05, | |
| "loss": 1.1759, | |
| "step": 4005 | |
| }, | |
| { | |
| "epoch": 0.8422375864236329, | |
| "grad_norm": 4.197197437286377, | |
| "learning_rate": 1.810647377298064e-05, | |
| "loss": 1.1319, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 0.8453802639849152, | |
| "grad_norm": 1.8338217735290527, | |
| "learning_rate": 1.8085008378846777e-05, | |
| "loss": 1.126, | |
| "step": 4035 | |
| }, | |
| { | |
| "epoch": 0.8485229415461973, | |
| "grad_norm": 4.26235818862915, | |
| "learning_rate": 1.806343487652276e-05, | |
| "loss": 1.1889, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.8516656191074796, | |
| "grad_norm": 2.5566670894622803, | |
| "learning_rate": 1.8041753554477344e-05, | |
| "loss": 1.1377, | |
| "step": 4065 | |
| }, | |
| { | |
| "epoch": 0.8548082966687618, | |
| "grad_norm": 3.85239839553833, | |
| "learning_rate": 1.8019964702620987e-05, | |
| "loss": 1.1859, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 0.857950974230044, | |
| "grad_norm": 1.832269549369812, | |
| "learning_rate": 1.7998068612301976e-05, | |
| "loss": 1.1816, | |
| "step": 4095 | |
| }, | |
| { | |
| "epoch": 0.8610936517913262, | |
| "grad_norm": 2.670729160308838, | |
| "learning_rate": 1.7976065576302528e-05, | |
| "loss": 1.1686, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 0.8642363293526084, | |
| "grad_norm": 2.3271324634552, | |
| "learning_rate": 1.7953955888834885e-05, | |
| "loss": 1.1683, | |
| "step": 4125 | |
| }, | |
| { | |
| "epoch": 0.8673790069138906, | |
| "grad_norm": 2.2425904273986816, | |
| "learning_rate": 1.7931739845537365e-05, | |
| "loss": 1.182, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 0.8705216844751729, | |
| "grad_norm": 1.5735398530960083, | |
| "learning_rate": 1.7909417743470415e-05, | |
| "loss": 1.0888, | |
| "step": 4155 | |
| }, | |
| { | |
| "epoch": 0.873664362036455, | |
| "grad_norm": 2.4160172939300537, | |
| "learning_rate": 1.7886989881112647e-05, | |
| "loss": 1.1943, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 0.8768070395977373, | |
| "grad_norm": 2.265516519546509, | |
| "learning_rate": 1.786445655835684e-05, | |
| "loss": 1.1819, | |
| "step": 4185 | |
| }, | |
| { | |
| "epoch": 0.8799497171590195, | |
| "grad_norm": 3.3854482173919678, | |
| "learning_rate": 1.784181807650592e-05, | |
| "loss": 1.1299, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.8830923947203017, | |
| "grad_norm": 1.209488868713379, | |
| "learning_rate": 1.7819074738268956e-05, | |
| "loss": 1.0379, | |
| "step": 4215 | |
| }, | |
| { | |
| "epoch": 0.8862350722815839, | |
| "grad_norm": 2.617237091064453, | |
| "learning_rate": 1.7796226847757084e-05, | |
| "loss": 1.2145, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 0.8893777498428661, | |
| "grad_norm": 4.028698921203613, | |
| "learning_rate": 1.7773274710479473e-05, | |
| "loss": 1.117, | |
| "step": 4245 | |
| }, | |
| { | |
| "epoch": 0.8925204274041484, | |
| "grad_norm": 2.382817506790161, | |
| "learning_rate": 1.7750218633339194e-05, | |
| "loss": 1.1122, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 0.8956631049654306, | |
| "grad_norm": 2.063629150390625, | |
| "learning_rate": 1.7727058924629165e-05, | |
| "loss": 1.1553, | |
| "step": 4275 | |
| }, | |
| { | |
| "epoch": 0.8988057825267127, | |
| "grad_norm": 1.7411893606185913, | |
| "learning_rate": 1.7703795894028003e-05, | |
| "loss": 1.1544, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 0.901948460087995, | |
| "grad_norm": 1.9747669696807861, | |
| "learning_rate": 1.768042985259588e-05, | |
| "loss": 1.1875, | |
| "step": 4305 | |
| }, | |
| { | |
| "epoch": 0.9050911376492772, | |
| "grad_norm": 2.338653326034546, | |
| "learning_rate": 1.765696111277038e-05, | |
| "loss": 1.1609, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 0.9082338152105593, | |
| "grad_norm": 2.3144114017486572, | |
| "learning_rate": 1.7633389988362305e-05, | |
| "loss": 1.1843, | |
| "step": 4335 | |
| }, | |
| { | |
| "epoch": 0.9113764927718416, | |
| "grad_norm": 2.072577953338623, | |
| "learning_rate": 1.760971679455149e-05, | |
| "loss": 1.1109, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 0.9145191703331238, | |
| "grad_norm": 2.2952466011047363, | |
| "learning_rate": 1.758594184788258e-05, | |
| "loss": 1.1714, | |
| "step": 4365 | |
| }, | |
| { | |
| "epoch": 0.9176618478944061, | |
| "grad_norm": 2.254108428955078, | |
| "learning_rate": 1.7562065466260816e-05, | |
| "loss": 1.1627, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 0.9208045254556882, | |
| "grad_norm": 2.286694049835205, | |
| "learning_rate": 1.753808796894775e-05, | |
| "loss": 1.135, | |
| "step": 4395 | |
| }, | |
| { | |
| "epoch": 0.9239472030169704, | |
| "grad_norm": 3.5377025604248047, | |
| "learning_rate": 1.751400967655701e-05, | |
| "loss": 1.1172, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 0.9270898805782527, | |
| "grad_norm": 1.9915852546691895, | |
| "learning_rate": 1.7489830911049995e-05, | |
| "loss": 1.1235, | |
| "step": 4425 | |
| }, | |
| { | |
| "epoch": 0.9302325581395349, | |
| "grad_norm": 1.916908860206604, | |
| "learning_rate": 1.746555199573158e-05, | |
| "loss": 1.1604, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 0.933375235700817, | |
| "grad_norm": 2.122201442718506, | |
| "learning_rate": 1.7441173255245773e-05, | |
| "loss": 1.17, | |
| "step": 4455 | |
| }, | |
| { | |
| "epoch": 0.9365179132620993, | |
| "grad_norm": 2.4088730812072754, | |
| "learning_rate": 1.74166950155714e-05, | |
| "loss": 1.1222, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 0.9396605908233815, | |
| "grad_norm": 2.2933411598205566, | |
| "learning_rate": 1.7392117604017733e-05, | |
| "loss": 1.1322, | |
| "step": 4485 | |
| }, | |
| { | |
| "epoch": 0.9428032683846638, | |
| "grad_norm": 2.084245204925537, | |
| "learning_rate": 1.7367441349220108e-05, | |
| "loss": 1.139, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.9459459459459459, | |
| "grad_norm": 2.220381498336792, | |
| "learning_rate": 1.734266658113555e-05, | |
| "loss": 1.1533, | |
| "step": 4515 | |
| }, | |
| { | |
| "epoch": 0.9490886235072281, | |
| "grad_norm": 2.4368200302124023, | |
| "learning_rate": 1.731779363103833e-05, | |
| "loss": 1.1276, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 0.9522313010685104, | |
| "grad_norm": 3.9131979942321777, | |
| "learning_rate": 1.729282283151557e-05, | |
| "loss": 1.2134, | |
| "step": 4545 | |
| }, | |
| { | |
| "epoch": 0.9553739786297926, | |
| "grad_norm": 2.0956265926361084, | |
| "learning_rate": 1.7267754516462773e-05, | |
| "loss": 1.1336, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 0.9585166561910748, | |
| "grad_norm": 2.3662939071655273, | |
| "learning_rate": 1.724258902107936e-05, | |
| "loss": 1.1807, | |
| "step": 4575 | |
| }, | |
| { | |
| "epoch": 0.961659333752357, | |
| "grad_norm": 2.2511234283447266, | |
| "learning_rate": 1.721732668186421e-05, | |
| "loss": 1.1407, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 0.9648020113136392, | |
| "grad_norm": 2.2256953716278076, | |
| "learning_rate": 1.7191967836611115e-05, | |
| "loss": 1.1797, | |
| "step": 4605 | |
| }, | |
| { | |
| "epoch": 0.9679446888749215, | |
| "grad_norm": 1.6837149858474731, | |
| "learning_rate": 1.716651282440432e-05, | |
| "loss": 1.1586, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 0.9710873664362036, | |
| "grad_norm": 1.941137433052063, | |
| "learning_rate": 1.7140961985613952e-05, | |
| "loss": 1.1739, | |
| "step": 4635 | |
| }, | |
| { | |
| "epoch": 0.9742300439974858, | |
| "grad_norm": 3.3290293216705322, | |
| "learning_rate": 1.711531566189146e-05, | |
| "loss": 1.1511, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 0.9773727215587681, | |
| "grad_norm": 2.164433002471924, | |
| "learning_rate": 1.7089574196165087e-05, | |
| "loss": 1.1144, | |
| "step": 4665 | |
| }, | |
| { | |
| "epoch": 0.9805153991200503, | |
| "grad_norm": 2.117098569869995, | |
| "learning_rate": 1.706373793263525e-05, | |
| "loss": 1.0804, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 0.9836580766813325, | |
| "grad_norm": 2.2358627319335938, | |
| "learning_rate": 1.7037807216769952e-05, | |
| "loss": 1.158, | |
| "step": 4695 | |
| }, | |
| { | |
| "epoch": 0.9868007542426147, | |
| "grad_norm": 2.1659979820251465, | |
| "learning_rate": 1.7011782395300152e-05, | |
| "loss": 1.1633, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 0.9899434318038969, | |
| "grad_norm": 1.9542781114578247, | |
| "learning_rate": 1.6985663816215156e-05, | |
| "loss": 1.2018, | |
| "step": 4725 | |
| }, | |
| { | |
| "epoch": 0.9930861093651792, | |
| "grad_norm": 2.3751957416534424, | |
| "learning_rate": 1.6959451828757915e-05, | |
| "loss": 1.1924, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 0.9962287869264613, | |
| "grad_norm": 2.5773661136627197, | |
| "learning_rate": 1.6933146783420412e-05, | |
| "loss": 1.2159, | |
| "step": 4755 | |
| }, | |
| { | |
| "epoch": 0.9993714644877436, | |
| "grad_norm": 1.986519694328308, | |
| "learning_rate": 1.6906749031938932e-05, | |
| "loss": 1.1635, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 1.0025141420490258, | |
| "grad_norm": 2.238046646118164, | |
| "learning_rate": 1.688025892728938e-05, | |
| "loss": 1.1517, | |
| "step": 4785 | |
| }, | |
| { | |
| "epoch": 1.005656819610308, | |
| "grad_norm": 2.3779406547546387, | |
| "learning_rate": 1.685367682368255e-05, | |
| "loss": 1.1341, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 1.0087994971715901, | |
| "grad_norm": 2.1286470890045166, | |
| "learning_rate": 1.6827003076559403e-05, | |
| "loss": 1.1293, | |
| "step": 4815 | |
| }, | |
| { | |
| "epoch": 1.0119421747328725, | |
| "grad_norm": 3.604013442993164, | |
| "learning_rate": 1.68002380425863e-05, | |
| "loss": 1.1733, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 1.0150848522941547, | |
| "grad_norm": 1.7821085453033447, | |
| "learning_rate": 1.6773382079650237e-05, | |
| "loss": 1.1215, | |
| "step": 4845 | |
| }, | |
| { | |
| "epoch": 1.0182275298554369, | |
| "grad_norm": 2.2310407161712646, | |
| "learning_rate": 1.6746435546854068e-05, | |
| "loss": 1.1364, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 1.021370207416719, | |
| "grad_norm": 2.5142269134521484, | |
| "learning_rate": 1.6719398804511682e-05, | |
| "loss": 1.1585, | |
| "step": 4875 | |
| }, | |
| { | |
| "epoch": 1.0245128849780012, | |
| "grad_norm": 2.2651455402374268, | |
| "learning_rate": 1.6692272214143224e-05, | |
| "loss": 1.1408, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 1.0276555625392834, | |
| "grad_norm": 2.247772693634033, | |
| "learning_rate": 1.6665056138470212e-05, | |
| "loss": 1.1134, | |
| "step": 4905 | |
| }, | |
| { | |
| "epoch": 1.0307982401005658, | |
| "grad_norm": 2.3006253242492676, | |
| "learning_rate": 1.6637750941410725e-05, | |
| "loss": 1.1252, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 1.033940917661848, | |
| "grad_norm": 1.8263660669326782, | |
| "learning_rate": 1.6610356988074517e-05, | |
| "loss": 1.1427, | |
| "step": 4935 | |
| }, | |
| { | |
| "epoch": 1.03708359522313, | |
| "grad_norm": 2.199976921081543, | |
| "learning_rate": 1.658287464475814e-05, | |
| "loss": 1.1208, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 1.0402262727844123, | |
| "grad_norm": 2.1440887451171875, | |
| "learning_rate": 1.6555304278940053e-05, | |
| "loss": 1.1305, | |
| "step": 4965 | |
| }, | |
| { | |
| "epoch": 1.0433689503456944, | |
| "grad_norm": 2.2581751346588135, | |
| "learning_rate": 1.65276462592757e-05, | |
| "loss": 1.1683, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 1.0465116279069768, | |
| "grad_norm": 1.7286102771759033, | |
| "learning_rate": 1.6499900955592587e-05, | |
| "loss": 1.1857, | |
| "step": 4995 | |
| }, | |
| { | |
| "epoch": 1.049654305468259, | |
| "grad_norm": 1.8500277996063232, | |
| "learning_rate": 1.6472068738885322e-05, | |
| "loss": 1.0997, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 1.0527969830295412, | |
| "grad_norm": 2.4594027996063232, | |
| "learning_rate": 1.644414998131067e-05, | |
| "loss": 1.0588, | |
| "step": 5025 | |
| }, | |
| { | |
| "epoch": 1.0559396605908233, | |
| "grad_norm": 2.3058676719665527, | |
| "learning_rate": 1.6416145056182577e-05, | |
| "loss": 1.1304, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 1.0590823381521055, | |
| "grad_norm": 1.9614354372024536, | |
| "learning_rate": 1.6388054337967173e-05, | |
| "loss": 1.1542, | |
| "step": 5055 | |
| }, | |
| { | |
| "epoch": 1.062225015713388, | |
| "grad_norm": 2.4512290954589844, | |
| "learning_rate": 1.635987820227776e-05, | |
| "loss": 1.1455, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 1.06536769327467, | |
| "grad_norm": 3.3408472537994385, | |
| "learning_rate": 1.6331617025869787e-05, | |
| "loss": 1.1008, | |
| "step": 5085 | |
| }, | |
| { | |
| "epoch": 1.0685103708359522, | |
| "grad_norm": 3.3527655601501465, | |
| "learning_rate": 1.6303271186635837e-05, | |
| "loss": 1.1311, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 1.0716530483972344, | |
| "grad_norm": 2.1265969276428223, | |
| "learning_rate": 1.627484106360055e-05, | |
| "loss": 1.1328, | |
| "step": 5115 | |
| }, | |
| { | |
| "epoch": 1.0747957259585166, | |
| "grad_norm": 2.1224098205566406, | |
| "learning_rate": 1.624632703691556e-05, | |
| "loss": 1.1432, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 1.077938403519799, | |
| "grad_norm": 2.1932625770568848, | |
| "learning_rate": 1.6217729487854404e-05, | |
| "loss": 1.1336, | |
| "step": 5145 | |
| }, | |
| { | |
| "epoch": 1.0810810810810811, | |
| "grad_norm": 1.477196455001831, | |
| "learning_rate": 1.6189048798807453e-05, | |
| "loss": 1.0739, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 1.0842237586423633, | |
| "grad_norm": 2.2613472938537598, | |
| "learning_rate": 1.616028535327677e-05, | |
| "loss": 1.0342, | |
| "step": 5175 | |
| }, | |
| { | |
| "epoch": 1.0873664362036455, | |
| "grad_norm": 2.0612030029296875, | |
| "learning_rate": 1.6131439535870982e-05, | |
| "loss": 1.2227, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 1.0905091137649277, | |
| "grad_norm": 2.0832862854003906, | |
| "learning_rate": 1.6102511732300156e-05, | |
| "loss": 1.1417, | |
| "step": 5205 | |
| }, | |
| { | |
| "epoch": 1.0936517913262098, | |
| "grad_norm": 2.326160192489624, | |
| "learning_rate": 1.607350232937064e-05, | |
| "loss": 1.1104, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 1.0967944688874922, | |
| "grad_norm": 1.928688883781433, | |
| "learning_rate": 1.6044411714979866e-05, | |
| "loss": 1.1503, | |
| "step": 5235 | |
| }, | |
| { | |
| "epoch": 1.0999371464487744, | |
| "grad_norm": 2.209092617034912, | |
| "learning_rate": 1.6015240278111184e-05, | |
| "loss": 1.0927, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 1.1030798240100566, | |
| "grad_norm": 2.2068710327148438, | |
| "learning_rate": 1.5987941027952896e-05, | |
| "loss": 1.097, | |
| "step": 5265 | |
| }, | |
| { | |
| "epoch": 1.1062225015713387, | |
| "grad_norm": 2.1571450233459473, | |
| "learning_rate": 1.5958614441290784e-05, | |
| "loss": 1.0841, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 1.109365179132621, | |
| "grad_norm": 2.010093927383423, | |
| "learning_rate": 1.5929208179383714e-05, | |
| "loss": 1.1261, | |
| "step": 5295 | |
| }, | |
| { | |
| "epoch": 1.1125078566939033, | |
| "grad_norm": 2.4886395931243896, | |
| "learning_rate": 1.58997226354357e-05, | |
| "loss": 1.1357, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 1.1156505342551855, | |
| "grad_norm": 1.658639907836914, | |
| "learning_rate": 1.5870158203710876e-05, | |
| "loss": 1.0573, | |
| "step": 5325 | |
| }, | |
| { | |
| "epoch": 1.1187932118164676, | |
| "grad_norm": 2.1249542236328125, | |
| "learning_rate": 1.584051527952821e-05, | |
| "loss": 1.1161, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 1.1219358893777498, | |
| "grad_norm": 2.086566209793091, | |
| "learning_rate": 1.5810794259256238e-05, | |
| "loss": 1.1282, | |
| "step": 5355 | |
| }, | |
| { | |
| "epoch": 1.125078566939032, | |
| "grad_norm": 2.5150198936462402, | |
| "learning_rate": 1.5780995540307734e-05, | |
| "loss": 1.1408, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 1.1282212445003144, | |
| "grad_norm": 3.7854840755462646, | |
| "learning_rate": 1.575111952113444e-05, | |
| "loss": 1.1125, | |
| "step": 5385 | |
| }, | |
| { | |
| "epoch": 1.1313639220615965, | |
| "grad_norm": 2.2000720500946045, | |
| "learning_rate": 1.5721166601221697e-05, | |
| "loss": 1.099, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 1.1345065996228787, | |
| "grad_norm": 2.055802345275879, | |
| "learning_rate": 1.5691137181083126e-05, | |
| "loss": 1.1057, | |
| "step": 5415 | |
| }, | |
| { | |
| "epoch": 1.1376492771841609, | |
| "grad_norm": 1.9730473756790161, | |
| "learning_rate": 1.5661031662255254e-05, | |
| "loss": 1.0642, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 1.140791954745443, | |
| "grad_norm": 2.1444499492645264, | |
| "learning_rate": 1.563085044729218e-05, | |
| "loss": 1.1095, | |
| "step": 5445 | |
| }, | |
| { | |
| "epoch": 1.1439346323067254, | |
| "grad_norm": 3.2961201667785645, | |
| "learning_rate": 1.5600593939760146e-05, | |
| "loss": 1.1613, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 1.1470773098680076, | |
| "grad_norm": 2.405123710632324, | |
| "learning_rate": 1.557026254423218e-05, | |
| "loss": 1.1211, | |
| "step": 5475 | |
| }, | |
| { | |
| "epoch": 1.1502199874292898, | |
| "grad_norm": 3.308258533477783, | |
| "learning_rate": 1.5539856666282655e-05, | |
| "loss": 1.1381, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 1.153362664990572, | |
| "grad_norm": 2.2963814735412598, | |
| "learning_rate": 1.5509376712481898e-05, | |
| "loss": 1.1293, | |
| "step": 5505 | |
| }, | |
| { | |
| "epoch": 1.156505342551854, | |
| "grad_norm": 1.9797133207321167, | |
| "learning_rate": 1.5478823090390726e-05, | |
| "loss": 1.0968, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 1.1596480201131363, | |
| "grad_norm": 2.6032137870788574, | |
| "learning_rate": 1.5448196208555014e-05, | |
| "loss": 1.1462, | |
| "step": 5535 | |
| }, | |
| { | |
| "epoch": 1.1627906976744187, | |
| "grad_norm": 1.9197511672973633, | |
| "learning_rate": 1.5417496476500212e-05, | |
| "loss": 1.0224, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 1.1659333752357008, | |
| "grad_norm": 2.2676188945770264, | |
| "learning_rate": 1.538672430472589e-05, | |
| "loss": 1.0625, | |
| "step": 5565 | |
| }, | |
| { | |
| "epoch": 1.169076052796983, | |
| "grad_norm": 2.4030728340148926, | |
| "learning_rate": 1.535588010470026e-05, | |
| "loss": 1.1171, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 1.1722187303582652, | |
| "grad_norm": 1.387372374534607, | |
| "learning_rate": 1.5324964288854615e-05, | |
| "loss": 1.0836, | |
| "step": 5595 | |
| }, | |
| { | |
| "epoch": 1.1753614079195476, | |
| "grad_norm": 2.1212260723114014, | |
| "learning_rate": 1.5293977270577895e-05, | |
| "loss": 1.1069, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 1.1785040854808297, | |
| "grad_norm": 2.2747840881347656, | |
| "learning_rate": 1.526291946421109e-05, | |
| "loss": 1.134, | |
| "step": 5625 | |
| }, | |
| { | |
| "epoch": 1.181646763042112, | |
| "grad_norm": 1.7659993171691895, | |
| "learning_rate": 1.5231791285041741e-05, | |
| "loss": 1.1512, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 1.184789440603394, | |
| "grad_norm": 1.815944790840149, | |
| "learning_rate": 1.5200593149298375e-05, | |
| "loss": 1.0769, | |
| "step": 5655 | |
| }, | |
| { | |
| "epoch": 1.1879321181646763, | |
| "grad_norm": 2.154893636703491, | |
| "learning_rate": 1.5169325474144936e-05, | |
| "loss": 1.1198, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 1.1910747957259584, | |
| "grad_norm": 1.9412492513656616, | |
| "learning_rate": 1.5137988677675215e-05, | |
| "loss": 1.115, | |
| "step": 5685 | |
| }, | |
| { | |
| "epoch": 1.1942174732872408, | |
| "grad_norm": 1.8665941953659058, | |
| "learning_rate": 1.5106583178907244e-05, | |
| "loss": 1.1067, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 1.197360150848523, | |
| "grad_norm": 1.9800127744674683, | |
| "learning_rate": 1.5075109397777714e-05, | |
| "loss": 1.0744, | |
| "step": 5715 | |
| }, | |
| { | |
| "epoch": 1.2005028284098052, | |
| "grad_norm": 3.109827756881714, | |
| "learning_rate": 1.5043567755136346e-05, | |
| "loss": 1.0699, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 1.2036455059710873, | |
| "grad_norm": 1.9781428575515747, | |
| "learning_rate": 1.5011958672740261e-05, | |
| "loss": 1.1205, | |
| "step": 5745 | |
| }, | |
| { | |
| "epoch": 1.2067881835323695, | |
| "grad_norm": 2.4044976234436035, | |
| "learning_rate": 1.498028257324836e-05, | |
| "loss": 1.11, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 1.2099308610936519, | |
| "grad_norm": 2.0579380989074707, | |
| "learning_rate": 1.4948539880215642e-05, | |
| "loss": 1.1586, | |
| "step": 5775 | |
| }, | |
| { | |
| "epoch": 1.213073538654934, | |
| "grad_norm": 3.235210418701172, | |
| "learning_rate": 1.4916731018087576e-05, | |
| "loss": 1.1164, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 1.2162162162162162, | |
| "grad_norm": 2.258272647857666, | |
| "learning_rate": 1.4884856412194386e-05, | |
| "loss": 1.1057, | |
| "step": 5805 | |
| }, | |
| { | |
| "epoch": 1.2193588937774984, | |
| "grad_norm": 3.207092761993408, | |
| "learning_rate": 1.4852916488745409e-05, | |
| "loss": 1.1112, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 1.2225015713387806, | |
| "grad_norm": 2.802191972732544, | |
| "learning_rate": 1.4820911674823345e-05, | |
| "loss": 1.0854, | |
| "step": 5835 | |
| }, | |
| { | |
| "epoch": 1.2256442489000627, | |
| "grad_norm": 2.0806314945220947, | |
| "learning_rate": 1.4788842398378595e-05, | |
| "loss": 1.1596, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 1.2287869264613451, | |
| "grad_norm": 2.3855271339416504, | |
| "learning_rate": 1.4756709088223508e-05, | |
| "loss": 1.1108, | |
| "step": 5865 | |
| }, | |
| { | |
| "epoch": 1.2319296040226273, | |
| "grad_norm": 2.3936338424682617, | |
| "learning_rate": 1.4724512174026653e-05, | |
| "loss": 1.1988, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 1.2350722815839095, | |
| "grad_norm": 1.9991942644119263, | |
| "learning_rate": 1.4692252086307079e-05, | |
| "loss": 1.1588, | |
| "step": 5895 | |
| }, | |
| { | |
| "epoch": 1.2382149591451916, | |
| "grad_norm": 2.2678310871124268, | |
| "learning_rate": 1.465992925642856e-05, | |
| "loss": 1.084, | |
| "step": 5910 | |
| }, | |
| { | |
| "epoch": 1.241357636706474, | |
| "grad_norm": 2.767444610595703, | |
| "learning_rate": 1.4627544116593816e-05, | |
| "loss": 1.1242, | |
| "step": 5925 | |
| }, | |
| { | |
| "epoch": 1.2445003142677562, | |
| "grad_norm": 1.6328760385513306, | |
| "learning_rate": 1.4595097099838747e-05, | |
| "loss": 1.087, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 1.2476429918290384, | |
| "grad_norm": 1.8721544742584229, | |
| "learning_rate": 1.4562588640026629e-05, | |
| "loss": 1.1078, | |
| "step": 5955 | |
| }, | |
| { | |
| "epoch": 1.2507856693903205, | |
| "grad_norm": 2.0327460765838623, | |
| "learning_rate": 1.453001917184233e-05, | |
| "loss": 1.0601, | |
| "step": 5970 | |
| }, | |
| { | |
| "epoch": 1.2539283469516027, | |
| "grad_norm": 1.8650814294815063, | |
| "learning_rate": 1.4497389130786474e-05, | |
| "loss": 1.0553, | |
| "step": 5985 | |
| }, | |
| { | |
| "epoch": 1.2570710245128849, | |
| "grad_norm": 2.416186809539795, | |
| "learning_rate": 1.4464698953169649e-05, | |
| "loss": 1.1542, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 1.260213702074167, | |
| "grad_norm": 3.0162692070007324, | |
| "learning_rate": 1.443194907610654e-05, | |
| "loss": 1.1424, | |
| "step": 6015 | |
| }, | |
| { | |
| "epoch": 1.2633563796354494, | |
| "grad_norm": 2.833125352859497, | |
| "learning_rate": 1.4399139937510107e-05, | |
| "loss": 1.1691, | |
| "step": 6030 | |
| }, | |
| { | |
| "epoch": 1.2664990571967316, | |
| "grad_norm": 1.900730848312378, | |
| "learning_rate": 1.4366271976085719e-05, | |
| "loss": 1.0463, | |
| "step": 6045 | |
| }, | |
| { | |
| "epoch": 1.2696417347580138, | |
| "grad_norm": 2.243030071258545, | |
| "learning_rate": 1.4333345631325294e-05, | |
| "loss": 1.1671, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 1.2727844123192962, | |
| "grad_norm": 2.186521530151367, | |
| "learning_rate": 1.430036134350142e-05, | |
| "loss": 1.1653, | |
| "step": 6075 | |
| }, | |
| { | |
| "epoch": 1.2759270898805783, | |
| "grad_norm": 1.8847732543945312, | |
| "learning_rate": 1.4267319553661456e-05, | |
| "loss": 1.0842, | |
| "step": 6090 | |
| }, | |
| { | |
| "epoch": 1.2790697674418605, | |
| "grad_norm": 2.303593158721924, | |
| "learning_rate": 1.4234220703621658e-05, | |
| "loss": 1.0481, | |
| "step": 6105 | |
| }, | |
| { | |
| "epoch": 1.2822124450031427, | |
| "grad_norm": 3.213543653488159, | |
| "learning_rate": 1.4201065235961258e-05, | |
| "loss": 1.1446, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 1.2853551225644249, | |
| "grad_norm": 2.283177137374878, | |
| "learning_rate": 1.4167853594016539e-05, | |
| "loss": 1.114, | |
| "step": 6135 | |
| }, | |
| { | |
| "epoch": 1.288497800125707, | |
| "grad_norm": 3.1280677318573, | |
| "learning_rate": 1.4134586221874922e-05, | |
| "loss": 1.1568, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 1.2916404776869892, | |
| "grad_norm": 3.3405239582061768, | |
| "learning_rate": 1.410126356436902e-05, | |
| "loss": 1.0806, | |
| "step": 6165 | |
| }, | |
| { | |
| "epoch": 1.2947831552482716, | |
| "grad_norm": 2.9373440742492676, | |
| "learning_rate": 1.406788606707069e-05, | |
| "loss": 1.128, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 1.2979258328095538, | |
| "grad_norm": 2.1123921871185303, | |
| "learning_rate": 1.4034454176285073e-05, | |
| "loss": 1.017, | |
| "step": 6195 | |
| }, | |
| { | |
| "epoch": 1.301068510370836, | |
| "grad_norm": 2.298046827316284, | |
| "learning_rate": 1.4000968339044637e-05, | |
| "loss": 1.157, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 1.304211187932118, | |
| "grad_norm": 3.053687572479248, | |
| "learning_rate": 1.3967429003103175e-05, | |
| "loss": 1.1575, | |
| "step": 6225 | |
| }, | |
| { | |
| "epoch": 1.3073538654934005, | |
| "grad_norm": 2.9560487270355225, | |
| "learning_rate": 1.3933836616929856e-05, | |
| "loss": 1.1414, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 1.3104965430546827, | |
| "grad_norm": 2.401231050491333, | |
| "learning_rate": 1.3900191629703192e-05, | |
| "loss": 1.0565, | |
| "step": 6255 | |
| }, | |
| { | |
| "epoch": 1.3136392206159648, | |
| "grad_norm": 2.4497265815734863, | |
| "learning_rate": 1.3866494491305058e-05, | |
| "loss": 1.1296, | |
| "step": 6270 | |
| }, | |
| { | |
| "epoch": 1.316781898177247, | |
| "grad_norm": 2.0227253437042236, | |
| "learning_rate": 1.3832745652314652e-05, | |
| "loss": 1.1229, | |
| "step": 6285 | |
| }, | |
| { | |
| "epoch": 1.3199245757385292, | |
| "grad_norm": 1.9386115074157715, | |
| "learning_rate": 1.3798945564002493e-05, | |
| "loss": 1.0128, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 1.3230672532998113, | |
| "grad_norm": 1.967246413230896, | |
| "learning_rate": 1.376509467832437e-05, | |
| "loss": 1.1388, | |
| "step": 6315 | |
| }, | |
| { | |
| "epoch": 1.3262099308610937, | |
| "grad_norm": 3.249537706375122, | |
| "learning_rate": 1.3731193447915316e-05, | |
| "loss": 1.0584, | |
| "step": 6330 | |
| }, | |
| { | |
| "epoch": 1.329352608422376, | |
| "grad_norm": 2.7894890308380127, | |
| "learning_rate": 1.3697242326083536e-05, | |
| "loss": 1.1075, | |
| "step": 6345 | |
| }, | |
| { | |
| "epoch": 1.332495285983658, | |
| "grad_norm": 3.2315704822540283, | |
| "learning_rate": 1.3663241766804359e-05, | |
| "loss": 1.0972, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 1.3356379635449402, | |
| "grad_norm": 2.295436143875122, | |
| "learning_rate": 1.3629192224714163e-05, | |
| "loss": 1.1194, | |
| "step": 6375 | |
| }, | |
| { | |
| "epoch": 1.3387806411062226, | |
| "grad_norm": 1.2325416803359985, | |
| "learning_rate": 1.3595094155104297e-05, | |
| "loss": 1.099, | |
| "step": 6390 | |
| }, | |
| { | |
| "epoch": 1.3419233186675048, | |
| "grad_norm": 2.3862216472625732, | |
| "learning_rate": 1.3560948013914997e-05, | |
| "loss": 1.0684, | |
| "step": 6405 | |
| }, | |
| { | |
| "epoch": 1.345065996228787, | |
| "grad_norm": 2.1027228832244873, | |
| "learning_rate": 1.3526754257729273e-05, | |
| "loss": 1.0776, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 1.3482086737900691, | |
| "grad_norm": 2.421541213989258, | |
| "learning_rate": 1.3492513343766831e-05, | |
| "loss": 1.1078, | |
| "step": 6435 | |
| }, | |
| { | |
| "epoch": 1.3513513513513513, | |
| "grad_norm": 2.7943713665008545, | |
| "learning_rate": 1.3458225729877933e-05, | |
| "loss": 1.1076, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 1.3544940289126335, | |
| "grad_norm": 2.077890396118164, | |
| "learning_rate": 1.3423891874537289e-05, | |
| "loss": 1.1609, | |
| "step": 6465 | |
| }, | |
| { | |
| "epoch": 1.3576367064739157, | |
| "grad_norm": 2.0880138874053955, | |
| "learning_rate": 1.3389512236837931e-05, | |
| "loss": 1.0667, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 1.360779384035198, | |
| "grad_norm": 2.1045663356781006, | |
| "learning_rate": 1.3355087276485055e-05, | |
| "loss": 1.113, | |
| "step": 6495 | |
| }, | |
| { | |
| "epoch": 1.3639220615964802, | |
| "grad_norm": 2.125664472579956, | |
| "learning_rate": 1.3320617453789894e-05, | |
| "loss": 1.0619, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 1.3670647391577624, | |
| "grad_norm": 2.1109349727630615, | |
| "learning_rate": 1.3286103229663554e-05, | |
| "loss": 1.1235, | |
| "step": 6525 | |
| }, | |
| { | |
| "epoch": 1.3702074167190446, | |
| "grad_norm": 2.0411548614501953, | |
| "learning_rate": 1.3251545065610848e-05, | |
| "loss": 1.0872, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 1.373350094280327, | |
| "grad_norm": 3.083622694015503, | |
| "learning_rate": 1.3216943423724135e-05, | |
| "loss": 1.117, | |
| "step": 6555 | |
| }, | |
| { | |
| "epoch": 1.3764927718416091, | |
| "grad_norm": 1.8469839096069336, | |
| "learning_rate": 1.3182298766677136e-05, | |
| "loss": 1.1404, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 1.3796354494028913, | |
| "grad_norm": 2.121103525161743, | |
| "learning_rate": 1.3147611557718733e-05, | |
| "loss": 1.1007, | |
| "step": 6585 | |
| }, | |
| { | |
| "epoch": 1.3827781269641735, | |
| "grad_norm": 2.183732271194458, | |
| "learning_rate": 1.3112882260666805e-05, | |
| "loss": 1.1362, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 1.3859208045254556, | |
| "grad_norm": 2.0963168144226074, | |
| "learning_rate": 1.3078111339902004e-05, | |
| "loss": 1.1019, | |
| "step": 6615 | |
| }, | |
| { | |
| "epoch": 1.3890634820867378, | |
| "grad_norm": 3.269718885421753, | |
| "learning_rate": 1.3043299260361545e-05, | |
| "loss": 1.1073, | |
| "step": 6630 | |
| }, | |
| { | |
| "epoch": 1.3922061596480202, | |
| "grad_norm": 2.942408323287964, | |
| "learning_rate": 1.3008446487532998e-05, | |
| "loss": 1.0783, | |
| "step": 6645 | |
| }, | |
| { | |
| "epoch": 1.3953488372093024, | |
| "grad_norm": 2.6441168785095215, | |
| "learning_rate": 1.2973553487448068e-05, | |
| "loss": 1.1656, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 1.3984915147705845, | |
| "grad_norm": 2.102250099182129, | |
| "learning_rate": 1.2938620726676345e-05, | |
| "loss": 1.1408, | |
| "step": 6675 | |
| }, | |
| { | |
| "epoch": 1.4016341923318667, | |
| "grad_norm": 3.086590051651001, | |
| "learning_rate": 1.2903648672319084e-05, | |
| "loss": 1.1059, | |
| "step": 6690 | |
| }, | |
| { | |
| "epoch": 1.404776869893149, | |
| "grad_norm": 2.3277883529663086, | |
| "learning_rate": 1.2868637792002952e-05, | |
| "loss": 1.1665, | |
| "step": 6705 | |
| }, | |
| { | |
| "epoch": 1.4079195474544313, | |
| "grad_norm": 1.6698582172393799, | |
| "learning_rate": 1.2833588553873768e-05, | |
| "loss": 1.0649, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 1.4110622250157134, | |
| "grad_norm": 2.26218581199646, | |
| "learning_rate": 1.2798501426590262e-05, | |
| "loss": 1.1894, | |
| "step": 6735 | |
| }, | |
| { | |
| "epoch": 1.4142049025769956, | |
| "grad_norm": 1.8945990800857544, | |
| "learning_rate": 1.276337687931778e-05, | |
| "loss": 1.1507, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 1.4173475801382778, | |
| "grad_norm": 2.481565475463867, | |
| "learning_rate": 1.2728215381722044e-05, | |
| "loss": 1.0937, | |
| "step": 6765 | |
| }, | |
| { | |
| "epoch": 1.42049025769956, | |
| "grad_norm": 3.565307140350342, | |
| "learning_rate": 1.2693017403962842e-05, | |
| "loss": 1.197, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 1.4236329352608421, | |
| "grad_norm": 2.697265863418579, | |
| "learning_rate": 1.2657783416687763e-05, | |
| "loss": 1.0842, | |
| "step": 6795 | |
| }, | |
| { | |
| "epoch": 1.4267756128221245, | |
| "grad_norm": 2.244889736175537, | |
| "learning_rate": 1.2622513891025889e-05, | |
| "loss": 1.0942, | |
| "step": 6810 | |
| }, | |
| { | |
| "epoch": 1.4299182903834067, | |
| "grad_norm": 2.428379774093628, | |
| "learning_rate": 1.2587209298581506e-05, | |
| "loss": 1.131, | |
| "step": 6825 | |
| }, | |
| { | |
| "epoch": 1.4330609679446888, | |
| "grad_norm": 2.320680618286133, | |
| "learning_rate": 1.2551870111427791e-05, | |
| "loss": 1.1279, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 1.436203645505971, | |
| "grad_norm": 2.004258632659912, | |
| "learning_rate": 1.25164968021005e-05, | |
| "loss": 1.0718, | |
| "step": 6855 | |
| }, | |
| { | |
| "epoch": 1.4393463230672534, | |
| "grad_norm": 3.006464719772339, | |
| "learning_rate": 1.2481089843591655e-05, | |
| "loss": 1.1785, | |
| "step": 6870 | |
| }, | |
| { | |
| "epoch": 1.4424890006285356, | |
| "grad_norm": 2.481252670288086, | |
| "learning_rate": 1.2445649709343217e-05, | |
| "loss": 1.1382, | |
| "step": 6885 | |
| }, | |
| { | |
| "epoch": 1.4456316781898177, | |
| "grad_norm": 2.247004985809326, | |
| "learning_rate": 1.2410176873240751e-05, | |
| "loss": 1.1179, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 1.4487743557511, | |
| "grad_norm": 2.7122623920440674, | |
| "learning_rate": 1.237467180960709e-05, | |
| "loss": 1.0569, | |
| "step": 6915 | |
| }, | |
| { | |
| "epoch": 1.451917033312382, | |
| "grad_norm": 2.119673013687134, | |
| "learning_rate": 1.2339134993196007e-05, | |
| "loss": 1.0939, | |
| "step": 6930 | |
| }, | |
| { | |
| "epoch": 1.4550597108736643, | |
| "grad_norm": 2.4448978900909424, | |
| "learning_rate": 1.2303566899185842e-05, | |
| "loss": 1.0598, | |
| "step": 6945 | |
| }, | |
| { | |
| "epoch": 1.4582023884349467, | |
| "grad_norm": 2.211289405822754, | |
| "learning_rate": 1.2267968003173166e-05, | |
| "loss": 1.1679, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 1.4613450659962288, | |
| "grad_norm": 1.9416890144348145, | |
| "learning_rate": 1.2232338781166422e-05, | |
| "loss": 1.0983, | |
| "step": 6975 | |
| }, | |
| { | |
| "epoch": 1.464487743557511, | |
| "grad_norm": 2.0862271785736084, | |
| "learning_rate": 1.2196679709579546e-05, | |
| "loss": 1.0911, | |
| "step": 6990 | |
| }, | |
| { | |
| "epoch": 1.4676304211187932, | |
| "grad_norm": 2.4440407752990723, | |
| "learning_rate": 1.2160991265225621e-05, | |
| "loss": 1.1073, | |
| "step": 7005 | |
| }, | |
| { | |
| "epoch": 1.4707730986800756, | |
| "grad_norm": 2.1422340869903564, | |
| "learning_rate": 1.2125273925310465e-05, | |
| "loss": 1.0904, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 1.4739157762413577, | |
| "grad_norm": 2.2515573501586914, | |
| "learning_rate": 1.2089528167426294e-05, | |
| "loss": 1.1244, | |
| "step": 7035 | |
| }, | |
| { | |
| "epoch": 1.47705845380264, | |
| "grad_norm": 2.149007558822632, | |
| "learning_rate": 1.2053754469545291e-05, | |
| "loss": 1.0963, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 1.480201131363922, | |
| "grad_norm": 3.1129860877990723, | |
| "learning_rate": 1.2017953310013255e-05, | |
| "loss": 1.1655, | |
| "step": 7065 | |
| }, | |
| { | |
| "epoch": 1.4833438089252042, | |
| "grad_norm": 1.7863973379135132, | |
| "learning_rate": 1.1982125167543175e-05, | |
| "loss": 1.0844, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 1.4864864864864864, | |
| "grad_norm": 2.28974986076355, | |
| "learning_rate": 1.1946270521208845e-05, | |
| "loss": 1.0585, | |
| "step": 7095 | |
| }, | |
| { | |
| "epoch": 1.4896291640477686, | |
| "grad_norm": 2.3061442375183105, | |
| "learning_rate": 1.1910389850438457e-05, | |
| "loss": 1.1377, | |
| "step": 7110 | |
| }, | |
| { | |
| "epoch": 1.492771841609051, | |
| "grad_norm": 1.6387587785720825, | |
| "learning_rate": 1.1874483635008183e-05, | |
| "loss": 1.1109, | |
| "step": 7125 | |
| }, | |
| { | |
| "epoch": 1.4959145191703331, | |
| "grad_norm": 1.6554300785064697, | |
| "learning_rate": 1.1838552355035761e-05, | |
| "loss": 1.0686, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 1.4990571967316153, | |
| "grad_norm": 2.2356905937194824, | |
| "learning_rate": 1.1802596490974088e-05, | |
| "loss": 1.0877, | |
| "step": 7155 | |
| }, | |
| { | |
| "epoch": 1.5021998742928977, | |
| "grad_norm": 2.838385820388794, | |
| "learning_rate": 1.1766616523604774e-05, | |
| "loss": 1.1015, | |
| "step": 7170 | |
| }, | |
| { | |
| "epoch": 1.5053425518541799, | |
| "grad_norm": 1.8725031614303589, | |
| "learning_rate": 1.1730612934031737e-05, | |
| "loss": 1.0806, | |
| "step": 7185 | |
| }, | |
| { | |
| "epoch": 1.508485229415462, | |
| "grad_norm": 2.853379011154175, | |
| "learning_rate": 1.1694586203674742e-05, | |
| "loss": 1.0796, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 1.5116279069767442, | |
| "grad_norm": 2.5570309162139893, | |
| "learning_rate": 1.1658536814263e-05, | |
| "loss": 1.0773, | |
| "step": 7215 | |
| }, | |
| { | |
| "epoch": 1.5147705845380264, | |
| "grad_norm": 2.4062445163726807, | |
| "learning_rate": 1.1622465247828681e-05, | |
| "loss": 0.9971, | |
| "step": 7230 | |
| }, | |
| { | |
| "epoch": 1.5179132620993085, | |
| "grad_norm": 2.4074044227600098, | |
| "learning_rate": 1.1586371986700522e-05, | |
| "loss": 1.0774, | |
| "step": 7245 | |
| }, | |
| { | |
| "epoch": 1.5210559396605907, | |
| "grad_norm": 2.0330376625061035, | |
| "learning_rate": 1.155025751349732e-05, | |
| "loss": 1.1209, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 1.5241986172218729, | |
| "grad_norm": 2.229322671890259, | |
| "learning_rate": 1.1516531959834856e-05, | |
| "loss": 1.1011, | |
| "step": 7275 | |
| }, | |
| { | |
| "epoch": 1.5273412947831553, | |
| "grad_norm": 2.2459731101989746, | |
| "learning_rate": 1.1480377846161259e-05, | |
| "loss": 1.095, | |
| "step": 7290 | |
| }, | |
| { | |
| "epoch": 1.5304839723444374, | |
| "grad_norm": 1.9737571477890015, | |
| "learning_rate": 1.1444203937706724e-05, | |
| "loss": 1.0764, | |
| "step": 7305 | |
| }, | |
| { | |
| "epoch": 1.5336266499057196, | |
| "grad_norm": 2.0487782955169678, | |
| "learning_rate": 1.1408010718168447e-05, | |
| "loss": 1.1322, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 1.536769327467002, | |
| "grad_norm": 2.7282018661499023, | |
| "learning_rate": 1.1371798671501813e-05, | |
| "loss": 1.1322, | |
| "step": 7335 | |
| }, | |
| { | |
| "epoch": 1.5399120050282842, | |
| "grad_norm": 3.496860980987549, | |
| "learning_rate": 1.133556828191398e-05, | |
| "loss": 1.0829, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 1.5430546825895664, | |
| "grad_norm": 1.9811838865280151, | |
| "learning_rate": 1.1299320033857355e-05, | |
| "loss": 1.0813, | |
| "step": 7365 | |
| }, | |
| { | |
| "epoch": 1.5461973601508485, | |
| "grad_norm": 2.5440211296081543, | |
| "learning_rate": 1.1263054412023158e-05, | |
| "loss": 1.0347, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 1.5493400377121307, | |
| "grad_norm": 1.8516343832015991, | |
| "learning_rate": 1.1226771901334901e-05, | |
| "loss": 1.0719, | |
| "step": 7395 | |
| }, | |
| { | |
| "epoch": 1.5524827152734129, | |
| "grad_norm": 2.147573709487915, | |
| "learning_rate": 1.1190472986941939e-05, | |
| "loss": 1.068, | |
| "step": 7410 | |
| }, | |
| { | |
| "epoch": 1.555625392834695, | |
| "grad_norm": 1.999443531036377, | |
| "learning_rate": 1.1154158154212964e-05, | |
| "loss": 1.0674, | |
| "step": 7425 | |
| }, | |
| { | |
| "epoch": 1.5587680703959774, | |
| "grad_norm": 1.5025156736373901, | |
| "learning_rate": 1.111782788872952e-05, | |
| "loss": 1.1003, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 1.5619107479572596, | |
| "grad_norm": 2.162722587585449, | |
| "learning_rate": 1.1081482676279504e-05, | |
| "loss": 1.078, | |
| "step": 7455 | |
| }, | |
| { | |
| "epoch": 1.5650534255185418, | |
| "grad_norm": 2.049753427505493, | |
| "learning_rate": 1.1045123002850683e-05, | |
| "loss": 1.1825, | |
| "step": 7470 | |
| }, | |
| { | |
| "epoch": 1.5681961030798242, | |
| "grad_norm": 2.4640021324157715, | |
| "learning_rate": 1.1008749354624181e-05, | |
| "loss": 1.0942, | |
| "step": 7485 | |
| }, | |
| { | |
| "epoch": 1.5713387806411063, | |
| "grad_norm": 2.2149126529693604, | |
| "learning_rate": 1.0972362217967998e-05, | |
| "loss": 1.1415, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 1.5744814582023885, | |
| "grad_norm": 1.953383207321167, | |
| "learning_rate": 1.0935962079430468e-05, | |
| "loss": 1.0875, | |
| "step": 7515 | |
| }, | |
| { | |
| "epoch": 1.5776241357636707, | |
| "grad_norm": 2.371173620223999, | |
| "learning_rate": 1.0899549425733811e-05, | |
| "loss": 1.1068, | |
| "step": 7530 | |
| }, | |
| { | |
| "epoch": 1.5807668133249528, | |
| "grad_norm": 2.135591983795166, | |
| "learning_rate": 1.0863124743767566e-05, | |
| "loss": 1.1094, | |
| "step": 7545 | |
| }, | |
| { | |
| "epoch": 1.583909490886235, | |
| "grad_norm": 2.1090261936187744, | |
| "learning_rate": 1.0826688520582124e-05, | |
| "loss": 1.0459, | |
| "step": 7560 | |
| }, | |
| { | |
| "epoch": 1.5870521684475172, | |
| "grad_norm": 2.8963959217071533, | |
| "learning_rate": 1.0790241243382187e-05, | |
| "loss": 1.1019, | |
| "step": 7575 | |
| }, | |
| { | |
| "epoch": 1.5901948460087993, | |
| "grad_norm": 2.2832436561584473, | |
| "learning_rate": 1.0753783399520281e-05, | |
| "loss": 1.1125, | |
| "step": 7590 | |
| }, | |
| { | |
| "epoch": 1.5933375235700817, | |
| "grad_norm": 2.1328582763671875, | |
| "learning_rate": 1.0717315476490205e-05, | |
| "loss": 1.1088, | |
| "step": 7605 | |
| }, | |
| { | |
| "epoch": 1.596480201131364, | |
| "grad_norm": 1.7953342199325562, | |
| "learning_rate": 1.0680837961920538e-05, | |
| "loss": 1.0812, | |
| "step": 7620 | |
| }, | |
| { | |
| "epoch": 1.5996228786926463, | |
| "grad_norm": 2.0836551189422607, | |
| "learning_rate": 1.064435134356812e-05, | |
| "loss": 1.1015, | |
| "step": 7635 | |
| }, | |
| { | |
| "epoch": 1.6027655562539285, | |
| "grad_norm": 3.2132134437561035, | |
| "learning_rate": 1.0607856109311508e-05, | |
| "loss": 1.1377, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 1.6059082338152106, | |
| "grad_norm": 2.149661064147949, | |
| "learning_rate": 1.0571352747144477e-05, | |
| "loss": 1.0525, | |
| "step": 7665 | |
| }, | |
| { | |
| "epoch": 1.6090509113764928, | |
| "grad_norm": 1.8691046237945557, | |
| "learning_rate": 1.053484174516948e-05, | |
| "loss": 1.0954, | |
| "step": 7680 | |
| }, | |
| { | |
| "epoch": 1.612193588937775, | |
| "grad_norm": 1.9283403158187866, | |
| "learning_rate": 1.049832359159112e-05, | |
| "loss": 1.1199, | |
| "step": 7695 | |
| }, | |
| { | |
| "epoch": 1.6153362664990571, | |
| "grad_norm": 1.9671255350112915, | |
| "learning_rate": 1.0461798774709635e-05, | |
| "loss": 1.068, | |
| "step": 7710 | |
| }, | |
| { | |
| "epoch": 1.6184789440603393, | |
| "grad_norm": 2.721006393432617, | |
| "learning_rate": 1.0425267782914359e-05, | |
| "loss": 1.1274, | |
| "step": 7725 | |
| }, | |
| { | |
| "epoch": 1.6216216216216215, | |
| "grad_norm": 2.0059800148010254, | |
| "learning_rate": 1.0388731104677191e-05, | |
| "loss": 1.0832, | |
| "step": 7740 | |
| }, | |
| { | |
| "epoch": 1.6247642991829039, | |
| "grad_norm": 2.346212863922119, | |
| "learning_rate": 1.0352189228546068e-05, | |
| "loss": 1.1092, | |
| "step": 7755 | |
| }, | |
| { | |
| "epoch": 1.627906976744186, | |
| "grad_norm": 1.6566693782806396, | |
| "learning_rate": 1.031564264313843e-05, | |
| "loss": 1.033, | |
| "step": 7770 | |
| }, | |
| { | |
| "epoch": 1.6310496543054682, | |
| "grad_norm": 2.4089715480804443, | |
| "learning_rate": 1.0279091837134681e-05, | |
| "loss": 1.114, | |
| "step": 7785 | |
| }, | |
| { | |
| "epoch": 1.6341923318667506, | |
| "grad_norm": 2.9500551223754883, | |
| "learning_rate": 1.0242537299271675e-05, | |
| "loss": 1.1282, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 1.6373350094280328, | |
| "grad_norm": 1.8416279554367065, | |
| "learning_rate": 1.0205979518336148e-05, | |
| "loss": 1.0857, | |
| "step": 7815 | |
| }, | |
| { | |
| "epoch": 1.640477686989315, | |
| "grad_norm": 2.1354591846466064, | |
| "learning_rate": 1.0169418983158218e-05, | |
| "loss": 1.0807, | |
| "step": 7830 | |
| }, | |
| { | |
| "epoch": 1.6436203645505971, | |
| "grad_norm": 1.705767035484314, | |
| "learning_rate": 1.013285618260482e-05, | |
| "loss": 1.0273, | |
| "step": 7845 | |
| }, | |
| { | |
| "epoch": 1.6467630421118793, | |
| "grad_norm": 1.9959582090377808, | |
| "learning_rate": 1.009629160557318e-05, | |
| "loss": 1.0705, | |
| "step": 7860 | |
| }, | |
| { | |
| "epoch": 1.6499057196731615, | |
| "grad_norm": 2.004734754562378, | |
| "learning_rate": 1.0059725740984285e-05, | |
| "loss": 1.1216, | |
| "step": 7875 | |
| }, | |
| { | |
| "epoch": 1.6530483972344436, | |
| "grad_norm": 2.1259968280792236, | |
| "learning_rate": 1.0023159077776332e-05, | |
| "loss": 1.1618, | |
| "step": 7890 | |
| }, | |
| { | |
| "epoch": 1.6561910747957258, | |
| "grad_norm": 1.855248212814331, | |
| "learning_rate": 9.986592104898202e-06, | |
| "loss": 1.1007, | |
| "step": 7905 | |
| }, | |
| { | |
| "epoch": 1.6593337523570082, | |
| "grad_norm": 2.1948564052581787, | |
| "learning_rate": 9.950025311302914e-06, | |
| "loss": 1.1203, | |
| "step": 7920 | |
| }, | |
| { | |
| "epoch": 1.6624764299182904, | |
| "grad_norm": 2.140425205230713, | |
| "learning_rate": 9.913459185941085e-06, | |
| "loss": 1.0788, | |
| "step": 7935 | |
| }, | |
| { | |
| "epoch": 1.6656191074795728, | |
| "grad_norm": 3.158501148223877, | |
| "learning_rate": 9.87689421775441e-06, | |
| "loss": 1.0806, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 1.668761785040855, | |
| "grad_norm": 1.9126044511795044, | |
| "learning_rate": 9.840330895669096e-06, | |
| "loss": 1.086, | |
| "step": 7965 | |
| }, | |
| { | |
| "epoch": 1.671904462602137, | |
| "grad_norm": 2.586374044418335, | |
| "learning_rate": 9.803769708589352e-06, | |
| "loss": 1.0511, | |
| "step": 7980 | |
| }, | |
| { | |
| "epoch": 1.6750471401634193, | |
| "grad_norm": 2.192415475845337, | |
| "learning_rate": 9.767211145390827e-06, | |
| "loss": 1.1149, | |
| "step": 7995 | |
| }, | |
| { | |
| "epoch": 1.6781898177247014, | |
| "grad_norm": 2.12808895111084, | |
| "learning_rate": 9.73065569491409e-06, | |
| "loss": 1.1015, | |
| "step": 8010 | |
| }, | |
| { | |
| "epoch": 1.6813324952859836, | |
| "grad_norm": 2.3615734577178955, | |
| "learning_rate": 9.694103845958095e-06, | |
| "loss": 1.126, | |
| "step": 8025 | |
| }, | |
| { | |
| "epoch": 1.6844751728472658, | |
| "grad_norm": 1.6970425844192505, | |
| "learning_rate": 9.657556087273632e-06, | |
| "loss": 1.0458, | |
| "step": 8040 | |
| }, | |
| { | |
| "epoch": 1.687617850408548, | |
| "grad_norm": 2.4654476642608643, | |
| "learning_rate": 9.621012907556793e-06, | |
| "loss": 1.1316, | |
| "step": 8055 | |
| }, | |
| { | |
| "epoch": 1.6907605279698303, | |
| "grad_norm": 2.2715861797332764, | |
| "learning_rate": 9.584474795442454e-06, | |
| "loss": 1.0997, | |
| "step": 8070 | |
| }, | |
| { | |
| "epoch": 1.6939032055311125, | |
| "grad_norm": 2.117677688598633, | |
| "learning_rate": 9.547942239497716e-06, | |
| "loss": 1.0896, | |
| "step": 8085 | |
| }, | |
| { | |
| "epoch": 1.6970458830923947, | |
| "grad_norm": 1.9031733274459839, | |
| "learning_rate": 9.511415728215407e-06, | |
| "loss": 1.0468, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 1.700188560653677, | |
| "grad_norm": 3.147984504699707, | |
| "learning_rate": 9.474895750007508e-06, | |
| "loss": 1.0695, | |
| "step": 8115 | |
| }, | |
| { | |
| "epoch": 1.7033312382149592, | |
| "grad_norm": 2.6544809341430664, | |
| "learning_rate": 9.438382793198655e-06, | |
| "loss": 1.1083, | |
| "step": 8130 | |
| }, | |
| { | |
| "epoch": 1.7064739157762414, | |
| "grad_norm": 2.071148633956909, | |
| "learning_rate": 9.401877346019587e-06, | |
| "loss": 1.1407, | |
| "step": 8145 | |
| }, | |
| { | |
| "epoch": 1.7096165933375236, | |
| "grad_norm": 2.063387155532837, | |
| "learning_rate": 9.36537989660065e-06, | |
| "loss": 1.0749, | |
| "step": 8160 | |
| }, | |
| { | |
| "epoch": 1.7127592708988058, | |
| "grad_norm": 1.9773895740509033, | |
| "learning_rate": 9.328890932965225e-06, | |
| "loss": 1.0723, | |
| "step": 8175 | |
| }, | |
| { | |
| "epoch": 1.715901948460088, | |
| "grad_norm": 3.0571677684783936, | |
| "learning_rate": 9.29241094302324e-06, | |
| "loss": 1.1296, | |
| "step": 8190 | |
| }, | |
| { | |
| "epoch": 1.71904462602137, | |
| "grad_norm": 2.203171730041504, | |
| "learning_rate": 9.255940414564621e-06, | |
| "loss": 1.1401, | |
| "step": 8205 | |
| }, | |
| { | |
| "epoch": 1.7221873035826523, | |
| "grad_norm": 1.9974406957626343, | |
| "learning_rate": 9.21947983525279e-06, | |
| "loss": 1.1158, | |
| "step": 8220 | |
| }, | |
| { | |
| "epoch": 1.7253299811439347, | |
| "grad_norm": 2.110973834991455, | |
| "learning_rate": 9.183029692618134e-06, | |
| "loss": 1.0684, | |
| "step": 8235 | |
| }, | |
| { | |
| "epoch": 1.7284726587052168, | |
| "grad_norm": 1.8630388975143433, | |
| "learning_rate": 9.146590474051483e-06, | |
| "loss": 1.0376, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 1.7316153362664992, | |
| "grad_norm": 2.1341989040374756, | |
| "learning_rate": 9.110162666797595e-06, | |
| "loss": 1.0955, | |
| "step": 8265 | |
| }, | |
| { | |
| "epoch": 1.7347580138277814, | |
| "grad_norm": 2.5358388423919678, | |
| "learning_rate": 9.076174105269065e-06, | |
| "loss": 1.1651, | |
| "step": 8280 | |
| }, | |
| { | |
| "epoch": 1.7379006913890636, | |
| "grad_norm": 2.8575525283813477, | |
| "learning_rate": 9.039769740923183e-06, | |
| "loss": 1.1082, | |
| "step": 8295 | |
| }, | |
| { | |
| "epoch": 1.7410433689503457, | |
| "grad_norm": 2.287961959838867, | |
| "learning_rate": 9.003378216236902e-06, | |
| "loss": 1.1151, | |
| "step": 8310 | |
| }, | |
| { | |
| "epoch": 1.744186046511628, | |
| "grad_norm": 1.2058899402618408, | |
| "learning_rate": 8.967000017817247e-06, | |
| "loss": 1.0793, | |
| "step": 8325 | |
| }, | |
| { | |
| "epoch": 1.74732872407291, | |
| "grad_norm": 2.0018842220306396, | |
| "learning_rate": 8.930635632093046e-06, | |
| "loss": 1.0819, | |
| "step": 8340 | |
| }, | |
| { | |
| "epoch": 1.7504714016341922, | |
| "grad_norm": 1.9017311334609985, | |
| "learning_rate": 8.894285545308437e-06, | |
| "loss": 1.0658, | |
| "step": 8355 | |
| }, | |
| { | |
| "epoch": 1.7536140791954744, | |
| "grad_norm": 1.9596132040023804, | |
| "learning_rate": 8.857950243516364e-06, | |
| "loss": 1.1065, | |
| "step": 8370 | |
| }, | |
| { | |
| "epoch": 1.7567567567567568, | |
| "grad_norm": 1.979177713394165, | |
| "learning_rate": 8.821630212572074e-06, | |
| "loss": 1.0362, | |
| "step": 8385 | |
| }, | |
| { | |
| "epoch": 1.759899434318039, | |
| "grad_norm": 2.1770269870758057, | |
| "learning_rate": 8.785325938126608e-06, | |
| "loss": 1.0974, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 1.7630421118793211, | |
| "grad_norm": 2.125962734222412, | |
| "learning_rate": 8.749037905620334e-06, | |
| "loss": 1.1541, | |
| "step": 8415 | |
| }, | |
| { | |
| "epoch": 1.7661847894406035, | |
| "grad_norm": 2.0153768062591553, | |
| "learning_rate": 8.712766600276435e-06, | |
| "loss": 1.1233, | |
| "step": 8430 | |
| }, | |
| { | |
| "epoch": 1.7693274670018857, | |
| "grad_norm": 2.15311336517334, | |
| "learning_rate": 8.676512507094438e-06, | |
| "loss": 1.0341, | |
| "step": 8445 | |
| }, | |
| { | |
| "epoch": 1.7724701445631679, | |
| "grad_norm": 2.8032867908477783, | |
| "learning_rate": 8.640276110843702e-06, | |
| "loss": 1.0944, | |
| "step": 8460 | |
| }, | |
| { | |
| "epoch": 1.77561282212445, | |
| "grad_norm": 2.2355639934539795, | |
| "learning_rate": 8.604057896056965e-06, | |
| "loss": 1.1147, | |
| "step": 8475 | |
| }, | |
| { | |
| "epoch": 1.7787554996857322, | |
| "grad_norm": 2.222898006439209, | |
| "learning_rate": 8.567858347023843e-06, | |
| "loss": 1.1063, | |
| "step": 8490 | |
| }, | |
| { | |
| "epoch": 1.7818981772470144, | |
| "grad_norm": 2.043031692504883, | |
| "learning_rate": 8.531677947784383e-06, | |
| "loss": 1.1142, | |
| "step": 8505 | |
| }, | |
| { | |
| "epoch": 1.7850408548082966, | |
| "grad_norm": 2.3380000591278076, | |
| "learning_rate": 8.495517182122546e-06, | |
| "loss": 1.1379, | |
| "step": 8520 | |
| }, | |
| { | |
| "epoch": 1.7881835323695787, | |
| "grad_norm": 1.9827522039413452, | |
| "learning_rate": 8.459376533559784e-06, | |
| "loss": 1.1067, | |
| "step": 8535 | |
| }, | |
| { | |
| "epoch": 1.7913262099308611, | |
| "grad_norm": 2.2196788787841797, | |
| "learning_rate": 8.423256485348543e-06, | |
| "loss": 1.0563, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 1.7944688874921433, | |
| "grad_norm": 2.038891315460205, | |
| "learning_rate": 8.387157520465816e-06, | |
| "loss": 1.1151, | |
| "step": 8565 | |
| }, | |
| { | |
| "epoch": 1.7976115650534257, | |
| "grad_norm": 2.597698211669922, | |
| "learning_rate": 8.351080121606685e-06, | |
| "loss": 1.1025, | |
| "step": 8580 | |
| }, | |
| { | |
| "epoch": 1.8007542426147078, | |
| "grad_norm": 2.8073418140411377, | |
| "learning_rate": 8.315024771177862e-06, | |
| "loss": 1.136, | |
| "step": 8595 | |
| }, | |
| { | |
| "epoch": 1.80389692017599, | |
| "grad_norm": 2.143312692642212, | |
| "learning_rate": 8.278991951291234e-06, | |
| "loss": 1.0655, | |
| "step": 8610 | |
| }, | |
| { | |
| "epoch": 1.8070395977372722, | |
| "grad_norm": 1.5993940830230713, | |
| "learning_rate": 8.242982143757429e-06, | |
| "loss": 1.1023, | |
| "step": 8625 | |
| }, | |
| { | |
| "epoch": 1.8101822752985544, | |
| "grad_norm": 2.127690553665161, | |
| "learning_rate": 8.20699583007936e-06, | |
| "loss": 1.104, | |
| "step": 8640 | |
| }, | |
| { | |
| "epoch": 1.8133249528598365, | |
| "grad_norm": 2.57829213142395, | |
| "learning_rate": 8.171033491445806e-06, | |
| "loss": 1.0658, | |
| "step": 8655 | |
| }, | |
| { | |
| "epoch": 1.8164676304211187, | |
| "grad_norm": 2.4844679832458496, | |
| "learning_rate": 8.13509560872495e-06, | |
| "loss": 1.1207, | |
| "step": 8670 | |
| }, | |
| { | |
| "epoch": 1.8196103079824009, | |
| "grad_norm": 1.7117849588394165, | |
| "learning_rate": 8.099182662457972e-06, | |
| "loss": 1.1019, | |
| "step": 8685 | |
| }, | |
| { | |
| "epoch": 1.8227529855436833, | |
| "grad_norm": 1.6486339569091797, | |
| "learning_rate": 8.063295132852616e-06, | |
| "loss": 1.1566, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 1.8258956631049654, | |
| "grad_norm": 2.264791965484619, | |
| "learning_rate": 8.027433499776768e-06, | |
| "loss": 1.1269, | |
| "step": 8715 | |
| }, | |
| { | |
| "epoch": 1.8290383406662476, | |
| "grad_norm": 2.0826590061187744, | |
| "learning_rate": 7.99159824275204e-06, | |
| "loss": 1.1668, | |
| "step": 8730 | |
| }, | |
| { | |
| "epoch": 1.83218101822753, | |
| "grad_norm": 3.4813215732574463, | |
| "learning_rate": 7.955789840947357e-06, | |
| "loss": 1.0613, | |
| "step": 8745 | |
| }, | |
| { | |
| "epoch": 1.8353236957888122, | |
| "grad_norm": 1.950697898864746, | |
| "learning_rate": 7.920008773172549e-06, | |
| "loss": 1.1151, | |
| "step": 8760 | |
| }, | |
| { | |
| "epoch": 1.8384663733500943, | |
| "grad_norm": 2.761364459991455, | |
| "learning_rate": 7.884255517871955e-06, | |
| "loss": 1.0418, | |
| "step": 8775 | |
| }, | |
| { | |
| "epoch": 1.8416090509113765, | |
| "grad_norm": 2.035162925720215, | |
| "learning_rate": 7.848530553118024e-06, | |
| "loss": 1.1395, | |
| "step": 8790 | |
| }, | |
| { | |
| "epoch": 1.8447517284726587, | |
| "grad_norm": 2.061817169189453, | |
| "learning_rate": 7.812834356604913e-06, | |
| "loss": 1.1337, | |
| "step": 8805 | |
| }, | |
| { | |
| "epoch": 1.8478944060339408, | |
| "grad_norm": 2.214123010635376, | |
| "learning_rate": 7.777167405642109e-06, | |
| "loss": 1.09, | |
| "step": 8820 | |
| }, | |
| { | |
| "epoch": 1.851037083595223, | |
| "grad_norm": 2.1376426219940186, | |
| "learning_rate": 7.741530177148041e-06, | |
| "loss": 1.0908, | |
| "step": 8835 | |
| }, | |
| { | |
| "epoch": 1.8541797611565052, | |
| "grad_norm": 1.6116961240768433, | |
| "learning_rate": 7.70592314764371e-06, | |
| "loss": 1.1011, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 1.8573224387177876, | |
| "grad_norm": 2.253972291946411, | |
| "learning_rate": 7.670346793246317e-06, | |
| "loss": 1.0856, | |
| "step": 8865 | |
| }, | |
| { | |
| "epoch": 1.8604651162790697, | |
| "grad_norm": 1.9321953058242798, | |
| "learning_rate": 7.634801589662882e-06, | |
| "loss": 1.0851, | |
| "step": 8880 | |
| }, | |
| { | |
| "epoch": 1.8636077938403521, | |
| "grad_norm": 2.758145332336426, | |
| "learning_rate": 7.599288012183897e-06, | |
| "loss": 1.0952, | |
| "step": 8895 | |
| }, | |
| { | |
| "epoch": 1.8667504714016343, | |
| "grad_norm": 3.209368944168091, | |
| "learning_rate": 7.563806535676974e-06, | |
| "loss": 1.1081, | |
| "step": 8910 | |
| }, | |
| { | |
| "epoch": 1.8698931489629165, | |
| "grad_norm": 3.290693998336792, | |
| "learning_rate": 7.528357634580486e-06, | |
| "loss": 1.0403, | |
| "step": 8925 | |
| }, | |
| { | |
| "epoch": 1.8730358265241986, | |
| "grad_norm": 2.0125486850738525, | |
| "learning_rate": 7.4929417828972205e-06, | |
| "loss": 1.0801, | |
| "step": 8940 | |
| }, | |
| { | |
| "epoch": 1.8761785040854808, | |
| "grad_norm": 2.3519296646118164, | |
| "learning_rate": 7.457559454188052e-06, | |
| "loss": 1.1264, | |
| "step": 8955 | |
| }, | |
| { | |
| "epoch": 1.879321181646763, | |
| "grad_norm": 2.394660472869873, | |
| "learning_rate": 7.422211121565607e-06, | |
| "loss": 1.1271, | |
| "step": 8970 | |
| }, | |
| { | |
| "epoch": 1.8824638592080452, | |
| "grad_norm": 2.127760648727417, | |
| "learning_rate": 7.386897257687924e-06, | |
| "loss": 1.1308, | |
| "step": 8985 | |
| }, | |
| { | |
| "epoch": 1.8856065367693273, | |
| "grad_norm": 2.5212738513946533, | |
| "learning_rate": 7.3516183347521575e-06, | |
| "loss": 1.0758, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 1.8887492143306097, | |
| "grad_norm": 2.113893508911133, | |
| "learning_rate": 7.316374824488247e-06, | |
| "loss": 1.1463, | |
| "step": 9015 | |
| }, | |
| { | |
| "epoch": 1.8918918918918919, | |
| "grad_norm": 1.9831585884094238, | |
| "learning_rate": 7.281167198152615e-06, | |
| "loss": 1.1409, | |
| "step": 9030 | |
| }, | |
| { | |
| "epoch": 1.895034569453174, | |
| "grad_norm": 2.290170192718506, | |
| "learning_rate": 7.245995926521854e-06, | |
| "loss": 1.1156, | |
| "step": 9045 | |
| }, | |
| { | |
| "epoch": 1.8981772470144564, | |
| "grad_norm": 2.2289717197418213, | |
| "learning_rate": 7.2108614798864585e-06, | |
| "loss": 1.0908, | |
| "step": 9060 | |
| }, | |
| { | |
| "epoch": 1.9013199245757386, | |
| "grad_norm": 2.7847399711608887, | |
| "learning_rate": 7.17576432804451e-06, | |
| "loss": 1.1137, | |
| "step": 9075 | |
| }, | |
| { | |
| "epoch": 1.9044626021370208, | |
| "grad_norm": 2.2532174587249756, | |
| "learning_rate": 7.140704940295404e-06, | |
| "loss": 1.0993, | |
| "step": 9090 | |
| }, | |
| { | |
| "epoch": 1.907605279698303, | |
| "grad_norm": 3.097644329071045, | |
| "learning_rate": 7.1056837854335804e-06, | |
| "loss": 1.1322, | |
| "step": 9105 | |
| }, | |
| { | |
| "epoch": 1.9107479572595851, | |
| "grad_norm": 2.270334005355835, | |
| "learning_rate": 7.0707013317422446e-06, | |
| "loss": 1.0477, | |
| "step": 9120 | |
| }, | |
| { | |
| "epoch": 1.9138906348208673, | |
| "grad_norm": 2.698101282119751, | |
| "learning_rate": 7.035758046987122e-06, | |
| "loss": 1.0722, | |
| "step": 9135 | |
| }, | |
| { | |
| "epoch": 1.9170333123821495, | |
| "grad_norm": 2.2851641178131104, | |
| "learning_rate": 7.000854398410182e-06, | |
| "loss": 1.1588, | |
| "step": 9150 | |
| }, | |
| { | |
| "epoch": 1.9201759899434316, | |
| "grad_norm": 2.714940071105957, | |
| "learning_rate": 6.965990852723404e-06, | |
| "loss": 1.0664, | |
| "step": 9165 | |
| }, | |
| { | |
| "epoch": 1.923318667504714, | |
| "grad_norm": 2.079648971557617, | |
| "learning_rate": 6.9311678761025416e-06, | |
| "loss": 1.0701, | |
| "step": 9180 | |
| }, | |
| { | |
| "epoch": 1.9264613450659962, | |
| "grad_norm": 1.5704066753387451, | |
| "learning_rate": 6.896385934180867e-06, | |
| "loss": 1.085, | |
| "step": 9195 | |
| }, | |
| { | |
| "epoch": 1.9296040226272786, | |
| "grad_norm": 2.2583060264587402, | |
| "learning_rate": 6.861645492042977e-06, | |
| "loss": 1.1032, | |
| "step": 9210 | |
| }, | |
| { | |
| "epoch": 1.9327467001885608, | |
| "grad_norm": 2.0253028869628906, | |
| "learning_rate": 6.826947014218543e-06, | |
| "loss": 1.1232, | |
| "step": 9225 | |
| }, | |
| { | |
| "epoch": 1.935889377749843, | |
| "grad_norm": 3.0294480323791504, | |
| "learning_rate": 6.792290964676123e-06, | |
| "loss": 1.1583, | |
| "step": 9240 | |
| }, | |
| { | |
| "epoch": 1.939032055311125, | |
| "grad_norm": 2.480926036834717, | |
| "learning_rate": 6.7576778068169335e-06, | |
| "loss": 1.126, | |
| "step": 9255 | |
| }, | |
| { | |
| "epoch": 1.9421747328724073, | |
| "grad_norm": 2.393428087234497, | |
| "learning_rate": 6.725411298947324e-06, | |
| "loss": 1.0878, | |
| "step": 9270 | |
| }, | |
| { | |
| "epoch": 1.9453174104336894, | |
| "grad_norm": 2.0476040840148926, | |
| "learning_rate": 6.690882376871407e-06, | |
| "loss": 1.0794, | |
| "step": 9285 | |
| }, | |
| { | |
| "epoch": 1.9484600879949716, | |
| "grad_norm": 2.075193166732788, | |
| "learning_rate": 6.656397702457396e-06, | |
| "loss": 1.0825, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 1.9516027655562538, | |
| "grad_norm": 2.0196661949157715, | |
| "learning_rate": 6.621957736814983e-06, | |
| "loss": 1.0755, | |
| "step": 9315 | |
| }, | |
| { | |
| "epoch": 1.9547454431175362, | |
| "grad_norm": 2.4812753200531006, | |
| "learning_rate": 6.58756294045604e-06, | |
| "loss": 1.0883, | |
| "step": 9330 | |
| }, | |
| { | |
| "epoch": 1.9578881206788183, | |
| "grad_norm": 2.0828120708465576, | |
| "learning_rate": 6.553213773288465e-06, | |
| "loss": 1.1287, | |
| "step": 9345 | |
| }, | |
| { | |
| "epoch": 1.9610307982401005, | |
| "grad_norm": 2.273803949356079, | |
| "learning_rate": 6.518910694610023e-06, | |
| "loss": 1.0481, | |
| "step": 9360 | |
| }, | |
| { | |
| "epoch": 1.964173475801383, | |
| "grad_norm": 3.1020278930664062, | |
| "learning_rate": 6.4846541631022155e-06, | |
| "loss": 1.078, | |
| "step": 9375 | |
| }, | |
| { | |
| "epoch": 1.967316153362665, | |
| "grad_norm": 2.7231500148773193, | |
| "learning_rate": 6.4504446368241315e-06, | |
| "loss": 1.0389, | |
| "step": 9390 | |
| }, | |
| { | |
| "epoch": 1.9704588309239472, | |
| "grad_norm": 2.4498469829559326, | |
| "learning_rate": 6.416282573206341e-06, | |
| "loss": 1.1052, | |
| "step": 9405 | |
| }, | |
| { | |
| "epoch": 1.9736015084852294, | |
| "grad_norm": 2.266944408416748, | |
| "learning_rate": 6.382168429044769e-06, | |
| "loss": 1.1163, | |
| "step": 9420 | |
| }, | |
| { | |
| "epoch": 1.9767441860465116, | |
| "grad_norm": 2.877861976623535, | |
| "learning_rate": 6.34810266049459e-06, | |
| "loss": 1.1106, | |
| "step": 9435 | |
| }, | |
| { | |
| "epoch": 1.9798868636077938, | |
| "grad_norm": 2.256692886352539, | |
| "learning_rate": 6.314085723064118e-06, | |
| "loss": 1.0929, | |
| "step": 9450 | |
| }, | |
| { | |
| "epoch": 1.983029541169076, | |
| "grad_norm": 2.337369203567505, | |
| "learning_rate": 6.2801180716087315e-06, | |
| "loss": 1.1019, | |
| "step": 9465 | |
| }, | |
| { | |
| "epoch": 1.9861722187303583, | |
| "grad_norm": 3.803745746612549, | |
| "learning_rate": 6.246200160324789e-06, | |
| "loss": 1.0641, | |
| "step": 9480 | |
| }, | |
| { | |
| "epoch": 1.9893148962916405, | |
| "grad_norm": 2.354900598526001, | |
| "learning_rate": 6.21233244274354e-06, | |
| "loss": 1.1353, | |
| "step": 9495 | |
| }, | |
| { | |
| "epoch": 1.9924575738529227, | |
| "grad_norm": 2.0362517833709717, | |
| "learning_rate": 6.178515371725083e-06, | |
| "loss": 1.0638, | |
| "step": 9510 | |
| }, | |
| { | |
| "epoch": 1.995600251414205, | |
| "grad_norm": 1.7420244216918945, | |
| "learning_rate": 6.144749399452294e-06, | |
| "loss": 1.0539, | |
| "step": 9525 | |
| }, | |
| { | |
| "epoch": 1.9987429289754872, | |
| "grad_norm": 1.9186304807662964, | |
| "learning_rate": 6.11103497742478e-06, | |
| "loss": 1.0553, | |
| "step": 9540 | |
| }, | |
| { | |
| "epoch": 2.0018856065367694, | |
| "grad_norm": 1.9180148839950562, | |
| "learning_rate": 6.077372556452858e-06, | |
| "loss": 1.078, | |
| "step": 9555 | |
| }, | |
| { | |
| "epoch": 2.0050282840980516, | |
| "grad_norm": 1.8619567155838013, | |
| "learning_rate": 6.043762586651511e-06, | |
| "loss": 1.0238, | |
| "step": 9570 | |
| }, | |
| { | |
| "epoch": 2.0081709616593337, | |
| "grad_norm": 2.054584503173828, | |
| "learning_rate": 6.010205517434373e-06, | |
| "loss": 1.0361, | |
| "step": 9585 | |
| }, | |
| { | |
| "epoch": 2.011313639220616, | |
| "grad_norm": 1.6725273132324219, | |
| "learning_rate": 5.97670179750772e-06, | |
| "loss": 1.0048, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 2.014456316781898, | |
| "grad_norm": 2.8712847232818604, | |
| "learning_rate": 5.943251874864476e-06, | |
| "loss": 1.0199, | |
| "step": 9615 | |
| }, | |
| { | |
| "epoch": 2.0175989943431802, | |
| "grad_norm": 2.331249237060547, | |
| "learning_rate": 5.9098561967782165e-06, | |
| "loss": 1.0697, | |
| "step": 9630 | |
| }, | |
| { | |
| "epoch": 2.0207416719044624, | |
| "grad_norm": 2.359811782836914, | |
| "learning_rate": 5.876515209797189e-06, | |
| "loss": 1.0052, | |
| "step": 9645 | |
| }, | |
| { | |
| "epoch": 2.023884349465745, | |
| "grad_norm": 3.5132248401641846, | |
| "learning_rate": 5.843229359738336e-06, | |
| "loss": 1.0845, | |
| "step": 9660 | |
| }, | |
| { | |
| "epoch": 2.027027027027027, | |
| "grad_norm": 1.9716084003448486, | |
| "learning_rate": 5.809999091681349e-06, | |
| "loss": 1.1177, | |
| "step": 9675 | |
| }, | |
| { | |
| "epoch": 2.0301697045883094, | |
| "grad_norm": 2.3846986293792725, | |
| "learning_rate": 5.776824849962706e-06, | |
| "loss": 1.0848, | |
| "step": 9690 | |
| }, | |
| { | |
| "epoch": 2.0333123821495915, | |
| "grad_norm": 1.9413962364196777, | |
| "learning_rate": 5.743707078169731e-06, | |
| "loss": 1.0824, | |
| "step": 9705 | |
| }, | |
| { | |
| "epoch": 2.0364550597108737, | |
| "grad_norm": 1.9300832748413086, | |
| "learning_rate": 5.710646219134666e-06, | |
| "loss": 1.0474, | |
| "step": 9720 | |
| }, | |
| { | |
| "epoch": 2.039597737272156, | |
| "grad_norm": 2.1981208324432373, | |
| "learning_rate": 5.67764271492874e-06, | |
| "loss": 1.1053, | |
| "step": 9735 | |
| }, | |
| { | |
| "epoch": 2.042740414833438, | |
| "grad_norm": 2.165830612182617, | |
| "learning_rate": 5.644697006856268e-06, | |
| "loss": 1.0203, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 2.04588309239472, | |
| "grad_norm": 2.567518711090088, | |
| "learning_rate": 5.61180953544876e-06, | |
| "loss": 1.0509, | |
| "step": 9765 | |
| }, | |
| { | |
| "epoch": 2.0490257699560024, | |
| "grad_norm": 2.09709095954895, | |
| "learning_rate": 5.578980740459009e-06, | |
| "loss": 1.0906, | |
| "step": 9780 | |
| }, | |
| { | |
| "epoch": 2.0521684475172846, | |
| "grad_norm": 1.5784721374511719, | |
| "learning_rate": 5.546211060855211e-06, | |
| "loss": 1.0694, | |
| "step": 9795 | |
| }, | |
| { | |
| "epoch": 2.0553111250785667, | |
| "grad_norm": 3.4078545570373535, | |
| "learning_rate": 5.513500934815113e-06, | |
| "loss": 1.0352, | |
| "step": 9810 | |
| }, | |
| { | |
| "epoch": 2.0584538026398493, | |
| "grad_norm": 2.0774762630462646, | |
| "learning_rate": 5.4808507997201544e-06, | |
| "loss": 1.0513, | |
| "step": 9825 | |
| }, | |
| { | |
| "epoch": 2.0615964802011315, | |
| "grad_norm": 3.19948148727417, | |
| "learning_rate": 5.448261092149596e-06, | |
| "loss": 1.0432, | |
| "step": 9840 | |
| }, | |
| { | |
| "epoch": 2.0647391577624137, | |
| "grad_norm": 2.4168691635131836, | |
| "learning_rate": 5.415732247874696e-06, | |
| "loss": 1.0354, | |
| "step": 9855 | |
| }, | |
| { | |
| "epoch": 2.067881835323696, | |
| "grad_norm": 1.7410550117492676, | |
| "learning_rate": 5.383264701852886e-06, | |
| "loss": 1.0389, | |
| "step": 9870 | |
| }, | |
| { | |
| "epoch": 2.071024512884978, | |
| "grad_norm": 2.2156288623809814, | |
| "learning_rate": 5.350858888221957e-06, | |
| "loss": 1.0614, | |
| "step": 9885 | |
| }, | |
| { | |
| "epoch": 2.07416719044626, | |
| "grad_norm": 3.3456056118011475, | |
| "learning_rate": 5.318515240294241e-06, | |
| "loss": 1.099, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 2.0773098680075424, | |
| "grad_norm": 3.286379814147949, | |
| "learning_rate": 5.28623419055083e-06, | |
| "loss": 1.0667, | |
| "step": 9915 | |
| }, | |
| { | |
| "epoch": 2.0804525455688245, | |
| "grad_norm": 2.086221218109131, | |
| "learning_rate": 5.2540161706357855e-06, | |
| "loss": 1.0458, | |
| "step": 9930 | |
| }, | |
| { | |
| "epoch": 2.0835952231301067, | |
| "grad_norm": 3.3937273025512695, | |
| "learning_rate": 5.221861611350371e-06, | |
| "loss": 1.0696, | |
| "step": 9945 | |
| }, | |
| { | |
| "epoch": 2.086737900691389, | |
| "grad_norm": 2.162592887878418, | |
| "learning_rate": 5.189770942647291e-06, | |
| "loss": 1.034, | |
| "step": 9960 | |
| }, | |
| { | |
| "epoch": 2.0898805782526715, | |
| "grad_norm": 2.144193649291992, | |
| "learning_rate": 5.157744593624939e-06, | |
| "loss": 1.0997, | |
| "step": 9975 | |
| }, | |
| { | |
| "epoch": 2.0930232558139537, | |
| "grad_norm": 2.1846065521240234, | |
| "learning_rate": 5.125782992521664e-06, | |
| "loss": 1.034, | |
| "step": 9990 | |
| }, | |
| { | |
| "epoch": 2.0951183741881416, | |
| "eval_accuracy": 0.8874096274375916, | |
| "eval_loss": 1.1195236444473267, | |
| "eval_runtime": 3488.4257, | |
| "eval_samples_per_second": 1.372, | |
| "eval_steps_per_second": 0.343, | |
| "step": 10000 | |
| } | |
| ], | |
| "logging_steps": 15, | |
| "max_steps": 14319, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 10000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.1391882766264566e+18, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |