| { | |
| "best_metric": 1.3055968284606934, | |
| "best_model_checkpoint": "/Users/bbunzeck/Documents/lexdec-llamas/small-char/checkpoint-9063", | |
| "epoch": 0.9999130726340234, | |
| "eval_steps": 1007, | |
| "global_step": 10065, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0009934556111614737, | |
| "grad_norm": 3.579357862472534, | |
| "learning_rate": 1.4999999999999999e-05, | |
| "loss": 4.5444, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.0019869112223229474, | |
| "grad_norm": 2.9948267936706543, | |
| "learning_rate": 2.9999999999999997e-05, | |
| "loss": 4.4413, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.0029803668334844216, | |
| "grad_norm": 2.480799913406372, | |
| "learning_rate": 4.4999999999999996e-05, | |
| "loss": 4.2512, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.003973822444645895, | |
| "grad_norm": 2.1858479976654053, | |
| "learning_rate": 5.9999999999999995e-05, | |
| "loss": 4.051, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.004967278055807369, | |
| "grad_norm": 2.028580904006958, | |
| "learning_rate": 7.5e-05, | |
| "loss": 3.8934, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.005960733666968843, | |
| "grad_norm": 1.9097962379455566, | |
| "learning_rate": 8.999999999999999e-05, | |
| "loss": 3.7586, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.006954189278130317, | |
| "grad_norm": 1.6703391075134277, | |
| "learning_rate": 0.00010499999999999999, | |
| "loss": 3.6266, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.00794764488929179, | |
| "grad_norm": 1.4444409608840942, | |
| "learning_rate": 0.00011999999999999999, | |
| "loss": 3.4994, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.008941100500453264, | |
| "grad_norm": 1.2527669668197632, | |
| "learning_rate": 0.000135, | |
| "loss": 3.3532, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.009934556111614737, | |
| "grad_norm": 1.0899807214736938, | |
| "learning_rate": 0.00015, | |
| "loss": 3.2155, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.010928011722776212, | |
| "grad_norm": 0.961487889289856, | |
| "learning_rate": 0.000165, | |
| "loss": 3.0686, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.011921467333937686, | |
| "grad_norm": 1.0186645984649658, | |
| "learning_rate": 0.00017999999999999998, | |
| "loss": 2.9308, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.01291492294509916, | |
| "grad_norm": 0.7688561081886292, | |
| "learning_rate": 0.000195, | |
| "loss": 2.839, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.013908378556260634, | |
| "grad_norm": 0.829856276512146, | |
| "learning_rate": 0.00020999999999999998, | |
| "loss": 2.7612, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.014901834167422107, | |
| "grad_norm": 0.9920964241027832, | |
| "learning_rate": 0.000225, | |
| "loss": 2.6882, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.01589528977858358, | |
| "grad_norm": 1.155908226966858, | |
| "learning_rate": 0.00023999999999999998, | |
| "loss": 2.6164, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.016888745389745056, | |
| "grad_norm": 1.0534379482269287, | |
| "learning_rate": 0.00025499999999999996, | |
| "loss": 2.562, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.01788220100090653, | |
| "grad_norm": 1.0463899374008179, | |
| "learning_rate": 0.00027, | |
| "loss": 2.5073, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.018875656612068, | |
| "grad_norm": 1.2722645998001099, | |
| "learning_rate": 0.000285, | |
| "loss": 2.4548, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.019869112223229474, | |
| "grad_norm": 2.3002970218658447, | |
| "learning_rate": 0.0003, | |
| "loss": 2.3996, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.02086256783439095, | |
| "grad_norm": 2.3697524070739746, | |
| "learning_rate": 0.0002999992393822385, | |
| "loss": 2.3689, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.021856023445552424, | |
| "grad_norm": 1.4811162948608398, | |
| "learning_rate": 0.000299996957536668, | |
| "loss": 2.3283, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.022849479056713896, | |
| "grad_norm": 1.1468684673309326, | |
| "learning_rate": 0.0002999931544864298, | |
| "loss": 2.2929, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.023842934667875373, | |
| "grad_norm": 1.240814447402954, | |
| "learning_rate": 0.00029998783027009295, | |
| "loss": 2.2497, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.024836390279036846, | |
| "grad_norm": 1.659279704093933, | |
| "learning_rate": 0.0002999809849416534, | |
| "loss": 2.2341, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.02582984589019832, | |
| "grad_norm": 2.129495143890381, | |
| "learning_rate": 0.0002999726185705334, | |
| "loss": 2.2024, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.02682330150135979, | |
| "grad_norm": 1.3891750574111938, | |
| "learning_rate": 0.00029996273124158115, | |
| "loss": 2.1787, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.027816757112521268, | |
| "grad_norm": 2.0682430267333984, | |
| "learning_rate": 0.0002999513230550697, | |
| "loss": 2.1729, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.02881021272368274, | |
| "grad_norm": 1.329626202583313, | |
| "learning_rate": 0.00029993839412669593, | |
| "loss": 2.1258, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.029803668334844213, | |
| "grad_norm": 1.5232797861099243, | |
| "learning_rate": 0.0002999239445875795, | |
| "loss": 2.1091, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.030797123946005686, | |
| "grad_norm": 1.3861021995544434, | |
| "learning_rate": 0.00029990797458426146, | |
| "loss": 2.0744, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.03179057955716716, | |
| "grad_norm": 1.527296781539917, | |
| "learning_rate": 0.0002998904842787027, | |
| "loss": 2.063, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.03278403516832863, | |
| "grad_norm": 2.02929949760437, | |
| "learning_rate": 0.00029987147384828234, | |
| "loss": 2.0221, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.03377749077949011, | |
| "grad_norm": 1.675370454788208, | |
| "learning_rate": 0.000299850943485796, | |
| "loss": 2.0111, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.034770946390651584, | |
| "grad_norm": 1.679379940032959, | |
| "learning_rate": 0.00029982889339945384, | |
| "loss": 1.9831, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.03576440200181306, | |
| "grad_norm": 1.7419506311416626, | |
| "learning_rate": 0.00029980532381287836, | |
| "loss": 1.9748, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.03675785761297453, | |
| "grad_norm": 1.773558259010315, | |
| "learning_rate": 0.0002997802349651021, | |
| "loss": 1.9885, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.037751313224136, | |
| "grad_norm": 1.6703206300735474, | |
| "learning_rate": 0.00029975362711056545, | |
| "loss": 1.9377, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.038744768835297476, | |
| "grad_norm": 1.6859080791473389, | |
| "learning_rate": 0.00029972550051911377, | |
| "loss": 1.9489, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.03973822444645895, | |
| "grad_norm": 1.7479523420333862, | |
| "learning_rate": 0.00029969585547599494, | |
| "loss": 1.9195, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.04073168005762043, | |
| "grad_norm": 1.7081456184387207, | |
| "learning_rate": 0.00029966469228185613, | |
| "loss": 1.9051, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.0417251356687819, | |
| "grad_norm": 1.7588847875595093, | |
| "learning_rate": 0.0002996320112527412, | |
| "loss": 1.9003, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.042718591279943374, | |
| "grad_norm": 1.4162713289260864, | |
| "learning_rate": 0.00029959781272008695, | |
| "loss": 1.8973, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.04371204689110485, | |
| "grad_norm": 1.79458487033844, | |
| "learning_rate": 0.0002995620970307203, | |
| "loss": 1.8866, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.04470550250226632, | |
| "grad_norm": 2.2808878421783447, | |
| "learning_rate": 0.00029952486454685437, | |
| "loss": 1.8964, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.04569895811342779, | |
| "grad_norm": 1.9715750217437744, | |
| "learning_rate": 0.00029948611564608506, | |
| "loss": 1.8516, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.046692413724589266, | |
| "grad_norm": 1.7016363143920898, | |
| "learning_rate": 0.000299445850721387, | |
| "loss": 1.8659, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.047685869335750745, | |
| "grad_norm": 1.6106808185577393, | |
| "learning_rate": 0.0002994040701811098, | |
| "loss": 1.8435, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.04867932494691222, | |
| "grad_norm": 1.8425825834274292, | |
| "learning_rate": 0.00029936077444897373, | |
| "loss": 1.8523, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.04967278055807369, | |
| "grad_norm": 1.814981460571289, | |
| "learning_rate": 0.0002993159639640655, | |
| "loss": 1.8476, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.050666236169235164, | |
| "grad_norm": 1.938087821006775, | |
| "learning_rate": 0.00029926963918083376, | |
| "loss": 1.8236, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.05165969178039664, | |
| "grad_norm": 1.889202356338501, | |
| "learning_rate": 0.00029922180056908456, | |
| "loss": 1.8148, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.05265314739155811, | |
| "grad_norm": 1.7310645580291748, | |
| "learning_rate": 0.00029917244861397653, | |
| "loss": 1.8182, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.05364660300271958, | |
| "grad_norm": 1.5528088808059692, | |
| "learning_rate": 0.000299121583816016, | |
| "loss": 1.823, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.054640058613881055, | |
| "grad_norm": 1.7096353769302368, | |
| "learning_rate": 0.0002990692066910519, | |
| "loss": 1.7906, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.055633514225042535, | |
| "grad_norm": 1.835706353187561, | |
| "learning_rate": 0.00029901531777027046, | |
| "loss": 1.7938, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.05662696983620401, | |
| "grad_norm": 1.799190640449524, | |
| "learning_rate": 0.00029895991760019007, | |
| "loss": 1.7937, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.05762042544736548, | |
| "grad_norm": 1.7181247472763062, | |
| "learning_rate": 0.00029890300674265526, | |
| "loss": 1.7968, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.058613881058526954, | |
| "grad_norm": 1.6190788745880127, | |
| "learning_rate": 0.0002988445857748316, | |
| "loss": 1.7837, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.05960733666968843, | |
| "grad_norm": 1.9251811504364014, | |
| "learning_rate": 0.0002987846552891995, | |
| "loss": 1.757, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.0606007922808499, | |
| "grad_norm": 2.025871992111206, | |
| "learning_rate": 0.000298723215893548, | |
| "loss": 1.761, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.06159424789201137, | |
| "grad_norm": 2.0724892616271973, | |
| "learning_rate": 0.00029866026821096924, | |
| "loss": 1.7492, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.06258770350317285, | |
| "grad_norm": 1.5950913429260254, | |
| "learning_rate": 0.0002985958128798514, | |
| "loss": 1.7435, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.06358115911433432, | |
| "grad_norm": 1.7549567222595215, | |
| "learning_rate": 0.00029852985055387284, | |
| "loss": 1.7491, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.06457461472549579, | |
| "grad_norm": 1.7409313917160034, | |
| "learning_rate": 0.00029846238190199504, | |
| "loss": 1.7533, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.06556807033665726, | |
| "grad_norm": 1.6994491815567017, | |
| "learning_rate": 0.00029839340760845614, | |
| "loss": 1.7154, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.06656152594781875, | |
| "grad_norm": 1.740803837776184, | |
| "learning_rate": 0.00029832292837276376, | |
| "loss": 1.738, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.06755498155898022, | |
| "grad_norm": 1.9129489660263062, | |
| "learning_rate": 0.00029825094490968794, | |
| "loss": 1.7282, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.0685484371701417, | |
| "grad_norm": 1.713456630706787, | |
| "learning_rate": 0.00029817745794925415, | |
| "loss": 1.7422, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.06954189278130317, | |
| "grad_norm": 1.8211374282836914, | |
| "learning_rate": 0.00029810246823673537, | |
| "loss": 1.7305, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.07053534839246464, | |
| "grad_norm": 1.817071557044983, | |
| "learning_rate": 0.0002980259765326452, | |
| "loss": 1.7271, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.07152880400362611, | |
| "grad_norm": 2.049840211868286, | |
| "learning_rate": 0.00029794798361272964, | |
| "loss": 1.7244, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.07252225961478759, | |
| "grad_norm": 1.7831237316131592, | |
| "learning_rate": 0.0002978684902679592, | |
| "loss": 1.7225, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.07351571522594906, | |
| "grad_norm": 1.496175765991211, | |
| "learning_rate": 0.00029778749730452133, | |
| "loss": 1.6916, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.07450917083711053, | |
| "grad_norm": 1.6311109066009521, | |
| "learning_rate": 0.0002977050055438118, | |
| "loss": 1.7059, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.075502626448272, | |
| "grad_norm": 1.6398779153823853, | |
| "learning_rate": 0.0002976210158224266, | |
| "loss": 1.7218, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.07649608205943348, | |
| "grad_norm": 1.8606700897216797, | |
| "learning_rate": 0.00029753552899215344, | |
| "loss": 1.6729, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.07748953767059495, | |
| "grad_norm": 1.9344886541366577, | |
| "learning_rate": 0.0002974485459199629, | |
| "loss": 1.6932, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.07848299328175642, | |
| "grad_norm": 1.9203944206237793, | |
| "learning_rate": 0.000297360067488, | |
| "loss": 1.7095, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.0794764488929179, | |
| "grad_norm": 2.1540608406066895, | |
| "learning_rate": 0.00029727009459357487, | |
| "loss": 1.6914, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.08046990450407937, | |
| "grad_norm": 1.872130274772644, | |
| "learning_rate": 0.00029717862814915405, | |
| "loss": 1.6699, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.08146336011524086, | |
| "grad_norm": 1.784360647201538, | |
| "learning_rate": 0.00029708566908235077, | |
| "loss": 1.6692, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.08245681572640233, | |
| "grad_norm": 1.5963140726089478, | |
| "learning_rate": 0.00029699121833591605, | |
| "loss": 1.6919, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.0834502713375638, | |
| "grad_norm": 1.6963502168655396, | |
| "learning_rate": 0.0002968952768677287, | |
| "loss": 1.6787, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.08444372694872528, | |
| "grad_norm": 1.8380306959152222, | |
| "learning_rate": 0.00029679784565078584, | |
| "loss": 1.6945, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.08543718255988675, | |
| "grad_norm": 1.9056203365325928, | |
| "learning_rate": 0.00029669892567319307, | |
| "loss": 1.6809, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.08643063817104822, | |
| "grad_norm": 1.66342031955719, | |
| "learning_rate": 0.0002965985179381542, | |
| "loss": 1.6685, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.0874240937822097, | |
| "grad_norm": 1.6112103462219238, | |
| "learning_rate": 0.00029649662346396133, | |
| "loss": 1.6694, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.08841754939337117, | |
| "grad_norm": 1.6002370119094849, | |
| "learning_rate": 0.0002963932432839844, | |
| "loss": 1.6769, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.08941100500453264, | |
| "grad_norm": 1.7592132091522217, | |
| "learning_rate": 0.0002962883784466608, | |
| "loss": 1.6481, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.09040446061569411, | |
| "grad_norm": 1.7872523069381714, | |
| "learning_rate": 0.00029618203001548465, | |
| "loss": 1.6685, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.09139791622685559, | |
| "grad_norm": 1.5813547372817993, | |
| "learning_rate": 0.000296074199068996, | |
| "loss": 1.6594, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.09239137183801706, | |
| "grad_norm": 2.033475160598755, | |
| "learning_rate": 0.0002959648867007699, | |
| "loss": 1.6424, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.09338482744917853, | |
| "grad_norm": 1.5323055982589722, | |
| "learning_rate": 0.0002958540940194055, | |
| "loss": 1.6472, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.09437828306034, | |
| "grad_norm": 1.7794679403305054, | |
| "learning_rate": 0.0002957418221485145, | |
| "loss": 1.6469, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.09537173867150149, | |
| "grad_norm": 1.8037221431732178, | |
| "learning_rate": 0.00029562807222671, | |
| "loss": 1.6421, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.09636519428266296, | |
| "grad_norm": 1.8958500623703003, | |
| "learning_rate": 0.00029551284540759473, | |
| "loss": 1.6179, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.09735864989382444, | |
| "grad_norm": 1.9698679447174072, | |
| "learning_rate": 0.0002953961428597496, | |
| "loss": 1.6484, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.09835210550498591, | |
| "grad_norm": 1.8058844804763794, | |
| "learning_rate": 0.00029527796576672173, | |
| "loss": 1.656, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.09934556111614738, | |
| "grad_norm": 1.6842243671417236, | |
| "learning_rate": 0.00029515831532701234, | |
| "loss": 1.627, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.1000409800439604, | |
| "eval_loss": 1.5775400400161743, | |
| "eval_runtime": 37.5436, | |
| "eval_samples_per_second": 1331.784, | |
| "eval_steps_per_second": 166.473, | |
| "step": 1007 | |
| }, | |
| { | |
| "epoch": 0.10033901672730886, | |
| "grad_norm": 1.8795899152755737, | |
| "learning_rate": 0.00029503719275406477, | |
| "loss": 1.6432, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.10133247233847033, | |
| "grad_norm": 1.7589457035064697, | |
| "learning_rate": 0.0002949145992762521, | |
| "loss": 1.6343, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.1023259279496318, | |
| "grad_norm": 1.7681567668914795, | |
| "learning_rate": 0.0002947905361368647, | |
| "loss": 1.6273, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.10331938356079327, | |
| "grad_norm": 1.884647011756897, | |
| "learning_rate": 0.00029466500459409756, | |
| "loss": 1.6097, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.10431283917195475, | |
| "grad_norm": 1.6961677074432373, | |
| "learning_rate": 0.0002945380059210377, | |
| "loss": 1.6235, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.10530629478311622, | |
| "grad_norm": 1.8109626770019531, | |
| "learning_rate": 0.000294409541405651, | |
| "loss": 1.6122, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.10629975039427769, | |
| "grad_norm": 1.9362094402313232, | |
| "learning_rate": 0.0002942796123507693, | |
| "loss": 1.6103, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.10729320600543917, | |
| "grad_norm": 1.679543375968933, | |
| "learning_rate": 0.0002941482200740774, | |
| "loss": 1.6184, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.10828666161660064, | |
| "grad_norm": 1.771816372871399, | |
| "learning_rate": 0.00029401536590809913, | |
| "loss": 1.6059, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.10928011722776211, | |
| "grad_norm": 1.6501538753509521, | |
| "learning_rate": 0.00029388105120018446, | |
| "loss": 1.6103, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.1102735728389236, | |
| "grad_norm": 1.8216150999069214, | |
| "learning_rate": 0.00029374527731249533, | |
| "loss": 1.6056, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.11126702845008507, | |
| "grad_norm": 1.8488177061080933, | |
| "learning_rate": 0.00029360804562199224, | |
| "loss": 1.5924, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.11226048406124654, | |
| "grad_norm": 1.522996425628662, | |
| "learning_rate": 0.0002934693575204199, | |
| "loss": 1.5943, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.11325393967240802, | |
| "grad_norm": 1.7767702341079712, | |
| "learning_rate": 0.00029332921441429345, | |
| "loss": 1.6257, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.11424739528356949, | |
| "grad_norm": 1.6360911130905151, | |
| "learning_rate": 0.0002931876177248841, | |
| "loss": 1.6122, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.11524085089473096, | |
| "grad_norm": 1.769485592842102, | |
| "learning_rate": 0.0002930445688882045, | |
| "loss": 1.6042, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.11623430650589243, | |
| "grad_norm": 1.7624589204788208, | |
| "learning_rate": 0.00029290006935499453, | |
| "loss": 1.5807, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.11722776211705391, | |
| "grad_norm": 1.9215561151504517, | |
| "learning_rate": 0.00029275412059070636, | |
| "loss": 1.6109, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.11822121772821538, | |
| "grad_norm": 1.6967573165893555, | |
| "learning_rate": 0.0002926067240754896, | |
| "loss": 1.5595, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.11921467333937685, | |
| "grad_norm": 1.9382522106170654, | |
| "learning_rate": 0.0002924578813041764, | |
| "loss": 1.5937, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.12020812895053833, | |
| "grad_norm": 1.7853381633758545, | |
| "learning_rate": 0.000292307593786266, | |
| "loss": 1.5888, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.1212015845616998, | |
| "grad_norm": 1.6771260499954224, | |
| "learning_rate": 0.00029215586304590996, | |
| "loss": 1.5837, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.12219504017286127, | |
| "grad_norm": 1.7282408475875854, | |
| "learning_rate": 0.00029200269062189623, | |
| "loss": 1.5999, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.12318849578402274, | |
| "grad_norm": 1.7539353370666504, | |
| "learning_rate": 0.0002918480780676336, | |
| "loss": 1.5715, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.12418195139518422, | |
| "grad_norm": 1.855692744255066, | |
| "learning_rate": 0.00029169202695113617, | |
| "loss": 1.5913, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.1251754070063457, | |
| "grad_norm": 1.7310928106307983, | |
| "learning_rate": 0.00029153453885500736, | |
| "loss": 1.5903, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.12616886261750718, | |
| "grad_norm": 1.6173487901687622, | |
| "learning_rate": 0.00029137561537642367, | |
| "loss": 1.5711, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.12716231822866864, | |
| "grad_norm": 1.92397141456604, | |
| "learning_rate": 0.00029121525812711874, | |
| "loss": 1.5895, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.12815577383983012, | |
| "grad_norm": 1.7854608297348022, | |
| "learning_rate": 0.00029105346873336684, | |
| "loss": 1.5801, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.12914922945099158, | |
| "grad_norm": 1.7446389198303223, | |
| "learning_rate": 0.0002908902488359664, | |
| "loss": 1.6041, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.13014268506215307, | |
| "grad_norm": 1.8835922479629517, | |
| "learning_rate": 0.0002907256000902236, | |
| "loss": 1.5788, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.13113614067331453, | |
| "grad_norm": 1.7387545108795166, | |
| "learning_rate": 0.0002905595241659351, | |
| "loss": 1.584, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.13212959628447601, | |
| "grad_norm": 1.7039700746536255, | |
| "learning_rate": 0.0002903920227473717, | |
| "loss": 1.5661, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.1331230518956375, | |
| "grad_norm": 1.6951566934585571, | |
| "learning_rate": 0.00029022309753326065, | |
| "loss": 1.5845, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.13411650750679896, | |
| "grad_norm": 1.5891770124435425, | |
| "learning_rate": 0.0002900527502367689, | |
| "loss": 1.5666, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.13510996311796045, | |
| "grad_norm": 1.6581312417984009, | |
| "learning_rate": 0.0002898809825854855, | |
| "loss": 1.5737, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.1361034187291219, | |
| "grad_norm": 1.7027684450149536, | |
| "learning_rate": 0.00028970779632140416, | |
| "loss": 1.5584, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.1370968743402834, | |
| "grad_norm": 1.7661086320877075, | |
| "learning_rate": 0.0002895331932009055, | |
| "loss": 1.5489, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.13809032995144485, | |
| "grad_norm": 1.6079459190368652, | |
| "learning_rate": 0.0002893571749947393, | |
| "loss": 1.5583, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.13908378556260634, | |
| "grad_norm": 1.6080243587493896, | |
| "learning_rate": 0.0002891797434880066, | |
| "loss": 1.572, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.1400772411737678, | |
| "grad_norm": 1.8214818239212036, | |
| "learning_rate": 0.0002890009004801415, | |
| "loss": 1.5613, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.14107069678492928, | |
| "grad_norm": 1.7399227619171143, | |
| "learning_rate": 0.0002888206477848928, | |
| "loss": 1.565, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.14206415239609074, | |
| "grad_norm": 1.5903868675231934, | |
| "learning_rate": 0.00028863898723030594, | |
| "loss": 1.5699, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.14305760800725223, | |
| "grad_norm": 1.858116865158081, | |
| "learning_rate": 0.00028845592065870413, | |
| "loss": 1.5651, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.1440510636184137, | |
| "grad_norm": 1.6774224042892456, | |
| "learning_rate": 0.00028827144992666986, | |
| "loss": 1.567, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.14504451922957518, | |
| "grad_norm": 1.8893427848815918, | |
| "learning_rate": 0.000288085576905026, | |
| "loss": 1.5475, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.14603797484073663, | |
| "grad_norm": 1.637244701385498, | |
| "learning_rate": 0.0002878983034788169, | |
| "loss": 1.5681, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.14703143045189812, | |
| "grad_norm": 1.7454718351364136, | |
| "learning_rate": 0.00028770963154728905, | |
| "loss": 1.552, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.1480248860630596, | |
| "grad_norm": 1.6382954120635986, | |
| "learning_rate": 0.0002875195630238721, | |
| "loss": 1.5213, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.14901834167422107, | |
| "grad_norm": 1.7484924793243408, | |
| "learning_rate": 0.0002873280998361594, | |
| "loss": 1.5253, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.15001179728538255, | |
| "grad_norm": 1.732568383216858, | |
| "learning_rate": 0.00028713524392588827, | |
| "loss": 1.5459, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.151005252896544, | |
| "grad_norm": 1.6054543256759644, | |
| "learning_rate": 0.0002869409972489204, | |
| "loss": 1.5456, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.1519987085077055, | |
| "grad_norm": 1.680078387260437, | |
| "learning_rate": 0.00028674536177522206, | |
| "loss": 1.5417, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.15299216411886696, | |
| "grad_norm": 1.6200144290924072, | |
| "learning_rate": 0.00028654833948884423, | |
| "loss": 1.5469, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.15398561973002844, | |
| "grad_norm": 1.6205105781555176, | |
| "learning_rate": 0.0002863499323879022, | |
| "loss": 1.5298, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.1549790753411899, | |
| "grad_norm": 1.6521717309951782, | |
| "learning_rate": 0.0002861501424845555, | |
| "loss": 1.5356, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.1559725309523514, | |
| "grad_norm": 2.770686149597168, | |
| "learning_rate": 0.00028594897180498745, | |
| "loss": 1.5518, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.15696598656351285, | |
| "grad_norm": 1.694839596748352, | |
| "learning_rate": 0.00028574642238938467, | |
| "loss": 1.5274, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.15795944217467434, | |
| "grad_norm": 1.587899088859558, | |
| "learning_rate": 0.00028554249629191616, | |
| "loss": 1.5421, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.1589528977858358, | |
| "grad_norm": 1.58628249168396, | |
| "learning_rate": 0.00028533719558071285, | |
| "loss": 1.5334, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.15994635339699728, | |
| "grad_norm": 1.5966298580169678, | |
| "learning_rate": 0.0002851305223378462, | |
| "loss": 1.5323, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.16093980900815874, | |
| "grad_norm": 2.3015835285186768, | |
| "learning_rate": 0.0002849224786593075, | |
| "loss": 1.5278, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.16193326461932023, | |
| "grad_norm": 1.6424635648727417, | |
| "learning_rate": 0.00028471306665498624, | |
| "loss": 1.5472, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.16292672023048171, | |
| "grad_norm": 1.6421678066253662, | |
| "learning_rate": 0.000284502288448649, | |
| "loss": 1.5265, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.16392017584164317, | |
| "grad_norm": 1.5951546430587769, | |
| "learning_rate": 0.00028429014617791773, | |
| "loss": 1.5318, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.16491363145280466, | |
| "grad_norm": 1.7263596057891846, | |
| "learning_rate": 0.0002840766419942481, | |
| "loss": 1.5342, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.16590708706396612, | |
| "grad_norm": 1.579903483390808, | |
| "learning_rate": 0.0002838617780629079, | |
| "loss": 1.5258, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.1669005426751276, | |
| "grad_norm": 1.6390513181686401, | |
| "learning_rate": 0.0002836455565629546, | |
| "loss": 1.5428, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.16789399828628906, | |
| "grad_norm": 1.6977102756500244, | |
| "learning_rate": 0.0002834279796872138, | |
| "loss": 1.5407, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.16888745389745055, | |
| "grad_norm": 1.776199221611023, | |
| "learning_rate": 0.00028320904964225665, | |
| "loss": 1.513, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.169880909508612, | |
| "grad_norm": 1.759331464767456, | |
| "learning_rate": 0.0002829887686483775, | |
| "loss": 1.5347, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.1708743651197735, | |
| "grad_norm": 1.7496792078018188, | |
| "learning_rate": 0.0002827671389395716, | |
| "loss": 1.5262, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.17186782073093496, | |
| "grad_norm": 1.8374871015548706, | |
| "learning_rate": 0.0002825441627635121, | |
| "loss": 1.4986, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.17286127634209644, | |
| "grad_norm": 1.6566568613052368, | |
| "learning_rate": 0.00028231984238152766, | |
| "loss": 1.5167, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.1738547319532579, | |
| "grad_norm": 1.6925946474075317, | |
| "learning_rate": 0.000282094180068579, | |
| "loss": 1.5354, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.1748481875644194, | |
| "grad_norm": 1.5948251485824585, | |
| "learning_rate": 0.0002818671781132364, | |
| "loss": 1.5011, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.17584164317558085, | |
| "grad_norm": 1.6373355388641357, | |
| "learning_rate": 0.0002816388388176561, | |
| "loss": 1.5205, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.17683509878674233, | |
| "grad_norm": 1.7579413652420044, | |
| "learning_rate": 0.00028140916449755706, | |
| "loss": 1.5111, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.17782855439790382, | |
| "grad_norm": 1.6279116868972778, | |
| "learning_rate": 0.0002811781574821975, | |
| "loss": 1.5298, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.17882201000906528, | |
| "grad_norm": 1.7122087478637695, | |
| "learning_rate": 0.0002809458201143513, | |
| "loss": 1.5218, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.17981546562022677, | |
| "grad_norm": 1.4644914865493774, | |
| "learning_rate": 0.0002807121547502843, | |
| "loss": 1.5133, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.18080892123138823, | |
| "grad_norm": 1.5409467220306396, | |
| "learning_rate": 0.00028047716375972996, | |
| "loss": 1.5473, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.1818023768425497, | |
| "grad_norm": 1.8359103202819824, | |
| "learning_rate": 0.000280240849525866, | |
| "loss": 1.5307, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.18279583245371117, | |
| "grad_norm": 1.6710262298583984, | |
| "learning_rate": 0.0002800032144452898, | |
| "loss": 1.5165, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.18378928806487266, | |
| "grad_norm": 1.5416898727416992, | |
| "learning_rate": 0.00027976426092799416, | |
| "loss": 1.5178, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.18478274367603412, | |
| "grad_norm": 1.6697194576263428, | |
| "learning_rate": 0.000279523991397343, | |
| "loss": 1.5095, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.1857761992871956, | |
| "grad_norm": 1.5872129201889038, | |
| "learning_rate": 0.00027928240829004663, | |
| "loss": 1.5269, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.18676965489835706, | |
| "grad_norm": 1.5739141702651978, | |
| "learning_rate": 0.000279039514056137, | |
| "loss": 1.5065, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.18776311050951855, | |
| "grad_norm": 1.6365103721618652, | |
| "learning_rate": 0.0002787953111589431, | |
| "loss": 1.5011, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.18875656612068, | |
| "grad_norm": 1.501937747001648, | |
| "learning_rate": 0.00027854980207506566, | |
| "loss": 1.5226, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.1897500217318415, | |
| "grad_norm": 1.5829925537109375, | |
| "learning_rate": 0.00027830298929435235, | |
| "loss": 1.5068, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.19074347734300298, | |
| "grad_norm": 1.6055023670196533, | |
| "learning_rate": 0.0002780548753198723, | |
| "loss": 1.5166, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.19173693295416444, | |
| "grad_norm": 1.692913293838501, | |
| "learning_rate": 0.00027780546266789076, | |
| "loss": 1.5033, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.19273038856532593, | |
| "grad_norm": 1.5064820051193237, | |
| "learning_rate": 0.00027755475386784364, | |
| "loss": 1.5164, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.1937238441764874, | |
| "grad_norm": 1.5499558448791504, | |
| "learning_rate": 0.00027730275146231186, | |
| "loss": 1.5097, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.19471729978764887, | |
| "grad_norm": 1.6266753673553467, | |
| "learning_rate": 0.0002770494580069954, | |
| "loss": 1.5086, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.19571075539881033, | |
| "grad_norm": 1.5763877630233765, | |
| "learning_rate": 0.0002767948760706877, | |
| "loss": 1.507, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.19670421100997182, | |
| "grad_norm": 1.757982611656189, | |
| "learning_rate": 0.00027653900823524925, | |
| "loss": 1.5186, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.19769766662113328, | |
| "grad_norm": 1.6466922760009766, | |
| "learning_rate": 0.00027628185709558176, | |
| "loss": 1.5155, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.19869112223229476, | |
| "grad_norm": 1.635491132736206, | |
| "learning_rate": 0.0002760234252596015, | |
| "loss": 1.509, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.19968457784345622, | |
| "grad_norm": 1.699342131614685, | |
| "learning_rate": 0.000275763715348213, | |
| "loss": 1.4785, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.2000819600879208, | |
| "eval_loss": 1.4590226411819458, | |
| "eval_runtime": 37.214, | |
| "eval_samples_per_second": 1343.58, | |
| "eval_steps_per_second": 167.948, | |
| "step": 2014 | |
| }, | |
| { | |
| "epoch": 0.2006780334546177, | |
| "grad_norm": 1.736262321472168, | |
| "learning_rate": 0.00027550272999528265, | |
| "loss": 1.4976, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.20167148906577917, | |
| "grad_norm": 1.5499353408813477, | |
| "learning_rate": 0.0002752404718476116, | |
| "loss": 1.4941, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.20266494467694066, | |
| "grad_norm": 1.6107423305511475, | |
| "learning_rate": 0.0002749769435649094, | |
| "loss": 1.5041, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.20365840028810211, | |
| "grad_norm": 1.5725529193878174, | |
| "learning_rate": 0.00027471214781976645, | |
| "loss": 1.4969, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.2046518558992636, | |
| "grad_norm": 1.6998605728149414, | |
| "learning_rate": 0.00027444608729762754, | |
| "loss": 1.4947, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.2056453115104251, | |
| "grad_norm": 1.535861849784851, | |
| "learning_rate": 0.000274178764696764, | |
| "loss": 1.5, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.20663876712158655, | |
| "grad_norm": 1.66730535030365, | |
| "learning_rate": 0.0002739101827282468, | |
| "loss": 1.4687, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.20763222273274803, | |
| "grad_norm": 1.527089238166809, | |
| "learning_rate": 0.0002736403441159188, | |
| "loss": 1.497, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.2086256783439095, | |
| "grad_norm": 1.5996365547180176, | |
| "learning_rate": 0.0002733692515963672, | |
| "loss": 1.4976, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.20961913395507098, | |
| "grad_norm": 1.477732539176941, | |
| "learning_rate": 0.0002730969079188958, | |
| "loss": 1.4923, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.21061258956623244, | |
| "grad_norm": 1.6266570091247559, | |
| "learning_rate": 0.00027282331584549716, | |
| "loss": 1.4959, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.21160604517739393, | |
| "grad_norm": 1.7904537916183472, | |
| "learning_rate": 0.00027254847815082444, | |
| "loss": 1.4911, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.21259950078855538, | |
| "grad_norm": 1.6096490621566772, | |
| "learning_rate": 0.0002722723976221634, | |
| "loss": 1.4739, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.21359295639971687, | |
| "grad_norm": 1.4944965839385986, | |
| "learning_rate": 0.00027199507705940413, | |
| "loss": 1.5032, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.21458641201087833, | |
| "grad_norm": 1.7133740186691284, | |
| "learning_rate": 0.0002717165192750125, | |
| "loss": 1.4929, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.21557986762203982, | |
| "grad_norm": 1.7259196043014526, | |
| "learning_rate": 0.000271436727094002, | |
| "loss": 1.4936, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 0.21657332323320128, | |
| "grad_norm": 1.5825343132019043, | |
| "learning_rate": 0.0002711557033539044, | |
| "loss": 1.4755, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.21756677884436276, | |
| "grad_norm": 1.740415096282959, | |
| "learning_rate": 0.0002708734509047419, | |
| "loss": 1.4964, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.21856023445552422, | |
| "grad_norm": 1.7195324897766113, | |
| "learning_rate": 0.0002705899726089973, | |
| "loss": 1.4764, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.2195536900666857, | |
| "grad_norm": 1.5383018255233765, | |
| "learning_rate": 0.00027030527134158585, | |
| "loss": 1.4852, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 0.2205471456778472, | |
| "grad_norm": 1.6427196264266968, | |
| "learning_rate": 0.00027001934998982524, | |
| "loss": 1.4494, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.22154060128900865, | |
| "grad_norm": 1.5180752277374268, | |
| "learning_rate": 0.0002697322114534071, | |
| "loss": 1.4931, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 0.22253405690017014, | |
| "grad_norm": 1.6394000053405762, | |
| "learning_rate": 0.0002694438586443669, | |
| "loss": 1.4887, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.2235275125113316, | |
| "grad_norm": 1.5006839036941528, | |
| "learning_rate": 0.0002691542944870549, | |
| "loss": 1.4664, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.2245209681224931, | |
| "grad_norm": 1.6465349197387695, | |
| "learning_rate": 0.00026886352191810627, | |
| "loss": 1.4569, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.22551442373365455, | |
| "grad_norm": 1.639144778251648, | |
| "learning_rate": 0.0002685715438864116, | |
| "loss": 1.4649, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 0.22650787934481603, | |
| "grad_norm": 1.6679630279541016, | |
| "learning_rate": 0.00026827836335308644, | |
| "loss": 1.4718, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.2275013349559775, | |
| "grad_norm": 1.56136155128479, | |
| "learning_rate": 0.00026798398329144166, | |
| "loss": 1.4684, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 0.22849479056713898, | |
| "grad_norm": 1.733837604522705, | |
| "learning_rate": 0.0002676884066869535, | |
| "loss": 1.4958, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.22948824617830044, | |
| "grad_norm": 1.7151204347610474, | |
| "learning_rate": 0.0002673916365372327, | |
| "loss": 1.4493, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.23048170178946192, | |
| "grad_norm": 1.7269740104675293, | |
| "learning_rate": 0.0002670936758519946, | |
| "loss": 1.4855, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.23147515740062338, | |
| "grad_norm": 1.6267927885055542, | |
| "learning_rate": 0.0002667945276530284, | |
| "loss": 1.4506, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 0.23246861301178487, | |
| "grad_norm": 1.4621727466583252, | |
| "learning_rate": 0.0002664941949741665, | |
| "loss": 1.4817, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.23346206862294633, | |
| "grad_norm": 1.7759307622909546, | |
| "learning_rate": 0.0002661926808612539, | |
| "loss": 1.4833, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.23445552423410781, | |
| "grad_norm": 1.6150075197219849, | |
| "learning_rate": 0.000265889988372117, | |
| "loss": 1.4848, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.2354489798452693, | |
| "grad_norm": 1.367301106452942, | |
| "learning_rate": 0.0002655861205765331, | |
| "loss": 1.4886, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.23644243545643076, | |
| "grad_norm": 1.7386715412139893, | |
| "learning_rate": 0.00026528108055619856, | |
| "loss": 1.4671, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.23743589106759225, | |
| "grad_norm": 1.5642729997634888, | |
| "learning_rate": 0.0002649748714046983, | |
| "loss": 1.4597, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 0.2384293466787537, | |
| "grad_norm": 1.5864043235778809, | |
| "learning_rate": 0.0002646674962274739, | |
| "loss": 1.4701, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.2394228022899152, | |
| "grad_norm": 1.5328258275985718, | |
| "learning_rate": 0.0002643589581417922, | |
| "loss": 1.4673, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 0.24041625790107665, | |
| "grad_norm": 1.6110318899154663, | |
| "learning_rate": 0.0002640492602767139, | |
| "loss": 1.4844, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.24140971351223814, | |
| "grad_norm": 1.661152720451355, | |
| "learning_rate": 0.00026373840577306165, | |
| "loss": 1.4618, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.2424031691233996, | |
| "grad_norm": 1.7061209678649902, | |
| "learning_rate": 0.00026342639778338813, | |
| "loss": 1.4649, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.24339662473456108, | |
| "grad_norm": 1.5013798475265503, | |
| "learning_rate": 0.0002631132394719443, | |
| "loss": 1.4602, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.24439008034572254, | |
| "grad_norm": 1.715354323387146, | |
| "learning_rate": 0.0002627989340146471, | |
| "loss": 1.4979, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.24538353595688403, | |
| "grad_norm": 1.5487847328186035, | |
| "learning_rate": 0.0002624834845990475, | |
| "loss": 1.4673, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 0.2463769915680455, | |
| "grad_norm": 1.6842710971832275, | |
| "learning_rate": 0.00026216689442429774, | |
| "loss": 1.4615, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.24737044717920698, | |
| "grad_norm": 1.4814667701721191, | |
| "learning_rate": 0.0002618491667011193, | |
| "loss": 1.4622, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.24836390279036843, | |
| "grad_norm": 1.5666563510894775, | |
| "learning_rate": 0.0002615303046517703, | |
| "loss": 1.4669, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.24935735840152992, | |
| "grad_norm": 1.477264404296875, | |
| "learning_rate": 0.00026121031151001245, | |
| "loss": 1.4625, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 0.2503508140126914, | |
| "grad_norm": 1.46149742603302, | |
| "learning_rate": 0.00026088919052107867, | |
| "loss": 1.4802, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.25134426962385287, | |
| "grad_norm": 1.526932954788208, | |
| "learning_rate": 0.00026056694494164, | |
| "loss": 1.465, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 0.25233772523501435, | |
| "grad_norm": 1.5274999141693115, | |
| "learning_rate": 0.00026024357803977263, | |
| "loss": 1.4552, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.25333118084617584, | |
| "grad_norm": 1.5492175817489624, | |
| "learning_rate": 0.0002599190930949247, | |
| "loss": 1.4748, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.25432463645733727, | |
| "grad_norm": 1.6205475330352783, | |
| "learning_rate": 0.00025959349339788295, | |
| "loss": 1.4779, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.25531809206849876, | |
| "grad_norm": 1.511488437652588, | |
| "learning_rate": 0.00025926678225073963, | |
| "loss": 1.4579, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 0.25631154767966025, | |
| "grad_norm": 1.6094532012939453, | |
| "learning_rate": 0.00025893896296685866, | |
| "loss": 1.46, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.25730500329082173, | |
| "grad_norm": 1.566312313079834, | |
| "learning_rate": 0.00025861003887084246, | |
| "loss": 1.473, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 0.25829845890198316, | |
| "grad_norm": 1.6014615297317505, | |
| "learning_rate": 0.0002582800132984977, | |
| "loss": 1.4554, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.25929191451314465, | |
| "grad_norm": 1.5696370601654053, | |
| "learning_rate": 0.0002579488895968019, | |
| "loss": 1.4538, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.26028537012430614, | |
| "grad_norm": 1.4629592895507812, | |
| "learning_rate": 0.0002576166711238694, | |
| "loss": 1.4441, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.2612788257354676, | |
| "grad_norm": 1.5493783950805664, | |
| "learning_rate": 0.0002572833612489169, | |
| "loss": 1.4847, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 0.26227228134662905, | |
| "grad_norm": 1.576574444770813, | |
| "learning_rate": 0.00025694896335223024, | |
| "loss": 1.4624, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.26326573695779054, | |
| "grad_norm": 1.766044020652771, | |
| "learning_rate": 0.00025661348082512884, | |
| "loss": 1.4633, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.26425919256895203, | |
| "grad_norm": 1.5602974891662598, | |
| "learning_rate": 0.0002562769170699324, | |
| "loss": 1.4582, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.2652526481801135, | |
| "grad_norm": 1.6773608922958374, | |
| "learning_rate": 0.00025593927549992594, | |
| "loss": 1.4571, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.266246103791275, | |
| "grad_norm": 1.593564748764038, | |
| "learning_rate": 0.000255600559539325, | |
| "loss": 1.4568, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.26723955940243643, | |
| "grad_norm": 1.4533571004867554, | |
| "learning_rate": 0.0002552607726232413, | |
| "loss": 1.4544, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 0.2682330150135979, | |
| "grad_norm": 1.593860387802124, | |
| "learning_rate": 0.0002549199181976477, | |
| "loss": 1.4696, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.2692264706247594, | |
| "grad_norm": 1.5255321264266968, | |
| "learning_rate": 0.00025457799971934325, | |
| "loss": 1.4655, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 0.2702199262359209, | |
| "grad_norm": 1.5048162937164307, | |
| "learning_rate": 0.0002542350206559182, | |
| "loss": 1.457, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 0.2712133818470823, | |
| "grad_norm": 1.4930408000946045, | |
| "learning_rate": 0.0002538909844857187, | |
| "loss": 1.4824, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 0.2722068374582438, | |
| "grad_norm": 1.522271752357483, | |
| "learning_rate": 0.00025354589469781186, | |
| "loss": 1.4689, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 0.2732002930694053, | |
| "grad_norm": 1.4428378343582153, | |
| "learning_rate": 0.0002531997547919499, | |
| "loss": 1.4615, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.2741937486805668, | |
| "grad_norm": 1.5555633306503296, | |
| "learning_rate": 0.0002528525682785349, | |
| "loss": 1.4569, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.2751872042917282, | |
| "grad_norm": 1.7241359949111938, | |
| "learning_rate": 0.0002525043386785833, | |
| "loss": 1.4621, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 0.2761806599028897, | |
| "grad_norm": 1.477664589881897, | |
| "learning_rate": 0.00025215506952369003, | |
| "loss": 1.456, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 0.2771741155140512, | |
| "grad_norm": 1.759738802909851, | |
| "learning_rate": 0.0002518047643559927, | |
| "loss": 1.4691, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 0.2781675711252127, | |
| "grad_norm": 1.5456058979034424, | |
| "learning_rate": 0.00025145342672813574, | |
| "loss": 1.4633, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.27916102673637416, | |
| "grad_norm": 1.626283049583435, | |
| "learning_rate": 0.00025110106020323437, | |
| "loss": 1.4543, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 0.2801544823475356, | |
| "grad_norm": 1.467103362083435, | |
| "learning_rate": 0.0002507476683548384, | |
| "loss": 1.4542, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.2811479379586971, | |
| "grad_norm": 1.5302448272705078, | |
| "learning_rate": 0.00025039325476689607, | |
| "loss": 1.4496, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 0.28214139356985857, | |
| "grad_norm": 1.6114156246185303, | |
| "learning_rate": 0.00025003782303371767, | |
| "loss": 1.4559, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 0.28313484918102005, | |
| "grad_norm": 1.5678813457489014, | |
| "learning_rate": 0.00024968137675993897, | |
| "loss": 1.4585, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.2841283047921815, | |
| "grad_norm": 1.5041207075119019, | |
| "learning_rate": 0.00024932391956048497, | |
| "loss": 1.439, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 0.28512176040334297, | |
| "grad_norm": 1.435428500175476, | |
| "learning_rate": 0.00024896545506053285, | |
| "loss": 1.4741, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 0.28611521601450446, | |
| "grad_norm": 1.625834345817566, | |
| "learning_rate": 0.00024860598689547555, | |
| "loss": 1.4341, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.28710867162566595, | |
| "grad_norm": 1.6627609729766846, | |
| "learning_rate": 0.0002482455187108846, | |
| "loss": 1.4159, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 0.2881021272368274, | |
| "grad_norm": 1.7054451704025269, | |
| "learning_rate": 0.00024788405416247347, | |
| "loss": 1.451, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.28909558284798886, | |
| "grad_norm": 1.4866358041763306, | |
| "learning_rate": 0.00024752159691606015, | |
| "loss": 1.462, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 0.29008903845915035, | |
| "grad_norm": 1.543471336364746, | |
| "learning_rate": 0.0002471581506475303, | |
| "loss": 1.4559, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 0.29108249407031184, | |
| "grad_norm": 1.562407374382019, | |
| "learning_rate": 0.00024679371904279973, | |
| "loss": 1.4383, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 0.29207594968147327, | |
| "grad_norm": 1.6679459810256958, | |
| "learning_rate": 0.0002464283057977771, | |
| "loss": 1.4577, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.29306940529263475, | |
| "grad_norm": 1.5543076992034912, | |
| "learning_rate": 0.00024606191461832655, | |
| "loss": 1.4407, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.29406286090379624, | |
| "grad_norm": 1.552199125289917, | |
| "learning_rate": 0.00024569454922022987, | |
| "loss": 1.4241, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 0.29505631651495773, | |
| "grad_norm": 1.4755094051361084, | |
| "learning_rate": 0.00024532621332914897, | |
| "loss": 1.4418, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 0.2960497721261192, | |
| "grad_norm": 1.559299349784851, | |
| "learning_rate": 0.0002449569106805882, | |
| "loss": 1.4367, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 0.29704322773728065, | |
| "grad_norm": 1.5764288902282715, | |
| "learning_rate": 0.00024458664501985624, | |
| "loss": 1.4431, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 0.29803668334844213, | |
| "grad_norm": 1.5143660306930542, | |
| "learning_rate": 0.0002442154201020283, | |
| "loss": 1.4548, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.2990301389596036, | |
| "grad_norm": 1.6271705627441406, | |
| "learning_rate": 0.00024384323969190789, | |
| "loss": 1.4548, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 0.3000235945707651, | |
| "grad_norm": 1.4137121438980103, | |
| "learning_rate": 0.0002434701075639888, | |
| "loss": 1.4285, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 0.30012294013188123, | |
| "eval_loss": 1.404004693031311, | |
| "eval_runtime": 37.2633, | |
| "eval_samples_per_second": 1341.804, | |
| "eval_steps_per_second": 167.725, | |
| "step": 3021 | |
| }, | |
| { | |
| "epoch": 0.30101705018192654, | |
| "grad_norm": 1.473416805267334, | |
| "learning_rate": 0.0002430960275024166, | |
| "loss": 1.4298, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 0.302010505793088, | |
| "grad_norm": 1.5060755014419556, | |
| "learning_rate": 0.00024272100330095055, | |
| "loss": 1.4552, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 0.3030039614042495, | |
| "grad_norm": 1.4686816930770874, | |
| "learning_rate": 0.00024234503876292485, | |
| "loss": 1.4615, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.303997417015411, | |
| "grad_norm": 1.6299128532409668, | |
| "learning_rate": 0.00024196813770121026, | |
| "loss": 1.4505, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.30499087262657243, | |
| "grad_norm": 1.513812780380249, | |
| "learning_rate": 0.00024159030393817538, | |
| "loss": 1.4468, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 0.3059843282377339, | |
| "grad_norm": 1.6047786474227905, | |
| "learning_rate": 0.0002412115413056477, | |
| "loss": 1.4227, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.3069777838488954, | |
| "grad_norm": 1.518823266029358, | |
| "learning_rate": 0.00024083185364487522, | |
| "loss": 1.4478, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 0.3079712394600569, | |
| "grad_norm": 1.6182243824005127, | |
| "learning_rate": 0.0002404512448064868, | |
| "loss": 1.4088, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.3089646950712184, | |
| "grad_norm": 1.56973397731781, | |
| "learning_rate": 0.00024006971865045377, | |
| "loss": 1.457, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 0.3099581506823798, | |
| "grad_norm": 1.443355679512024, | |
| "learning_rate": 0.0002396872790460504, | |
| "loss": 1.4328, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.3109516062935413, | |
| "grad_norm": 1.4511569738388062, | |
| "learning_rate": 0.00023930392987181473, | |
| "loss": 1.4361, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 0.3119450619047028, | |
| "grad_norm": 1.6141853332519531, | |
| "learning_rate": 0.00023891967501550937, | |
| "loss": 1.4272, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 0.31293851751586427, | |
| "grad_norm": 1.4931188821792603, | |
| "learning_rate": 0.00023853451837408185, | |
| "loss": 1.4264, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.3139319731270257, | |
| "grad_norm": 1.3810125589370728, | |
| "learning_rate": 0.00023814846385362525, | |
| "loss": 1.438, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 0.3149254287381872, | |
| "grad_norm": 1.6664650440216064, | |
| "learning_rate": 0.00023776151536933864, | |
| "loss": 1.4212, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 0.31591888434934867, | |
| "grad_norm": 1.559873104095459, | |
| "learning_rate": 0.00023737367684548716, | |
| "loss": 1.4483, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 0.31691233996051016, | |
| "grad_norm": 1.6474990844726562, | |
| "learning_rate": 0.00023698495221536245, | |
| "loss": 1.4255, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 0.3179057955716716, | |
| "grad_norm": 1.645722508430481, | |
| "learning_rate": 0.00023659534542124261, | |
| "loss": 1.4177, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.3188992511828331, | |
| "grad_norm": 1.7545628547668457, | |
| "learning_rate": 0.0002362048604143522, | |
| "loss": 1.4439, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 0.31989270679399456, | |
| "grad_norm": 1.5791382789611816, | |
| "learning_rate": 0.00023581350115482245, | |
| "loss": 1.4342, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 0.32088616240515605, | |
| "grad_norm": 1.5567851066589355, | |
| "learning_rate": 0.00023542127161165063, | |
| "loss": 1.4476, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 0.3218796180163175, | |
| "grad_norm": 1.589906096458435, | |
| "learning_rate": 0.00023502817576266023, | |
| "loss": 1.4299, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 0.32287307362747897, | |
| "grad_norm": 1.5045454502105713, | |
| "learning_rate": 0.00023463421759446033, | |
| "loss": 1.4263, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.32386652923864045, | |
| "grad_norm": 1.5627963542938232, | |
| "learning_rate": 0.00023423940110240538, | |
| "loss": 1.4499, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 0.32485998484980194, | |
| "grad_norm": 1.6112831830978394, | |
| "learning_rate": 0.00023384373029055445, | |
| "loss": 1.425, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 0.32585344046096343, | |
| "grad_norm": 1.5887855291366577, | |
| "learning_rate": 0.00023344720917163093, | |
| "loss": 1.4211, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 0.32684689607212486, | |
| "grad_norm": 1.4299589395523071, | |
| "learning_rate": 0.00023304984176698157, | |
| "loss": 1.4253, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 0.32784035168328635, | |
| "grad_norm": 1.6069183349609375, | |
| "learning_rate": 0.00023265163210653569, | |
| "loss": 1.4384, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.32883380729444783, | |
| "grad_norm": 1.6113414764404297, | |
| "learning_rate": 0.00023225258422876455, | |
| "loss": 1.4312, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 0.3298272629056093, | |
| "grad_norm": 1.4784319400787354, | |
| "learning_rate": 0.00023185270218064021, | |
| "loss": 1.4143, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 0.33082071851677075, | |
| "grad_norm": 1.55092453956604, | |
| "learning_rate": 0.00023145199001759452, | |
| "loss": 1.4266, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 0.33181417412793224, | |
| "grad_norm": 1.6077501773834229, | |
| "learning_rate": 0.00023105045180347788, | |
| "loss": 1.4457, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 0.3328076297390937, | |
| "grad_norm": 1.6062836647033691, | |
| "learning_rate": 0.00023064809161051842, | |
| "loss": 1.4196, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.3338010853502552, | |
| "grad_norm": 1.673006534576416, | |
| "learning_rate": 0.00023024491351928019, | |
| "loss": 1.4168, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 0.33479454096141664, | |
| "grad_norm": 1.5694451332092285, | |
| "learning_rate": 0.00022984092161862206, | |
| "loss": 1.4368, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 0.33578799657257813, | |
| "grad_norm": 1.4904859066009521, | |
| "learning_rate": 0.00022943612000565625, | |
| "loss": 1.4033, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 0.3367814521837396, | |
| "grad_norm": 1.5374070405960083, | |
| "learning_rate": 0.00022903051278570677, | |
| "loss": 1.418, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 0.3377749077949011, | |
| "grad_norm": 1.479141116142273, | |
| "learning_rate": 0.00022862410407226762, | |
| "loss": 1.4379, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.3387683634060626, | |
| "grad_norm": 1.6345080137252808, | |
| "learning_rate": 0.00022821689798696132, | |
| "loss": 1.4369, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 0.339761819017224, | |
| "grad_norm": 1.5274293422698975, | |
| "learning_rate": 0.00022780889865949689, | |
| "loss": 1.4153, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 0.3407552746283855, | |
| "grad_norm": 1.5244622230529785, | |
| "learning_rate": 0.00022740011022762827, | |
| "loss": 1.4247, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 0.341748730239547, | |
| "grad_norm": 1.6338365077972412, | |
| "learning_rate": 0.00022699053683711187, | |
| "loss": 1.4411, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 0.3427421858507085, | |
| "grad_norm": 1.5007723569869995, | |
| "learning_rate": 0.00022658018264166502, | |
| "loss": 1.4353, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.3437356414618699, | |
| "grad_norm": 1.4929261207580566, | |
| "learning_rate": 0.00022616905180292363, | |
| "loss": 1.425, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 0.3447290970730314, | |
| "grad_norm": 1.5754997730255127, | |
| "learning_rate": 0.00022575714849039984, | |
| "loss": 1.4241, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 0.3457225526841929, | |
| "grad_norm": 1.5391665697097778, | |
| "learning_rate": 0.00022534447688144006, | |
| "loss": 1.4439, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 0.34671600829535437, | |
| "grad_norm": 1.4276331663131714, | |
| "learning_rate": 0.0002249310411611824, | |
| "loss": 1.3961, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 0.3477094639065158, | |
| "grad_norm": 1.5136425495147705, | |
| "learning_rate": 0.00022451684552251414, | |
| "loss": 1.4351, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.3487029195176773, | |
| "grad_norm": 1.4856855869293213, | |
| "learning_rate": 0.00022410189416602943, | |
| "loss": 1.4294, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 0.3496963751288388, | |
| "grad_norm": 1.7253488302230835, | |
| "learning_rate": 0.00022368619129998655, | |
| "loss": 1.4383, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 0.35068983074000026, | |
| "grad_norm": 1.5053879022598267, | |
| "learning_rate": 0.00022326974114026535, | |
| "loss": 1.4173, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 0.3516832863511617, | |
| "grad_norm": 1.4788745641708374, | |
| "learning_rate": 0.00022285254791032423, | |
| "loss": 1.4218, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 0.3526767419623232, | |
| "grad_norm": 1.6103191375732422, | |
| "learning_rate": 0.0002224346158411577, | |
| "loss": 1.4132, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.35367019757348467, | |
| "grad_norm": 1.5396193265914917, | |
| "learning_rate": 0.00022201594917125306, | |
| "loss": 1.4083, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 0.35466365318464615, | |
| "grad_norm": 1.558293104171753, | |
| "learning_rate": 0.00022159655214654787, | |
| "loss": 1.4126, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 0.35565710879580764, | |
| "grad_norm": 1.4636329412460327, | |
| "learning_rate": 0.00022117642902038629, | |
| "loss": 1.4115, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 0.3566505644069691, | |
| "grad_norm": 1.5806535482406616, | |
| "learning_rate": 0.00022075558405347659, | |
| "loss": 1.4362, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 0.35764402001813056, | |
| "grad_norm": 1.4940319061279297, | |
| "learning_rate": 0.0002203340215138475, | |
| "loss": 1.4417, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.35863747562929205, | |
| "grad_norm": 1.537243127822876, | |
| "learning_rate": 0.00021991174567680512, | |
| "loss": 1.4132, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 0.35963093124045353, | |
| "grad_norm": 1.643825888633728, | |
| "learning_rate": 0.00021948876082488936, | |
| "loss": 1.4182, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 0.36062438685161496, | |
| "grad_norm": 1.5917963981628418, | |
| "learning_rate": 0.0002190650712478309, | |
| "loss": 1.4225, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 0.36161784246277645, | |
| "grad_norm": 1.4540566205978394, | |
| "learning_rate": 0.0002186406812425073, | |
| "loss": 1.4064, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 0.36261129807393794, | |
| "grad_norm": 1.586795687675476, | |
| "learning_rate": 0.00021821559511289948, | |
| "loss": 1.423, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.3636047536850994, | |
| "grad_norm": 1.4627307653427124, | |
| "learning_rate": 0.0002177898171700483, | |
| "loss": 1.4106, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 0.36459820929626086, | |
| "grad_norm": 1.485191822052002, | |
| "learning_rate": 0.00021736335173201066, | |
| "loss": 1.3951, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 0.36559166490742234, | |
| "grad_norm": 1.6559752225875854, | |
| "learning_rate": 0.00021693620312381568, | |
| "loss": 1.4124, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 0.36658512051858383, | |
| "grad_norm": 1.6469494104385376, | |
| "learning_rate": 0.00021650837567742094, | |
| "loss": 1.4307, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 0.3675785761297453, | |
| "grad_norm": 1.6612638235092163, | |
| "learning_rate": 0.00021607987373166855, | |
| "loss": 1.4228, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.3685720317409068, | |
| "grad_norm": 1.5064705610275269, | |
| "learning_rate": 0.00021565070163224095, | |
| "loss": 1.4167, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 0.36956548735206823, | |
| "grad_norm": 1.5027666091918945, | |
| "learning_rate": 0.0002152208637316172, | |
| "loss": 1.441, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 0.3705589429632297, | |
| "grad_norm": 1.543009638786316, | |
| "learning_rate": 0.00021479036438902847, | |
| "loss": 1.4213, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 0.3715523985743912, | |
| "grad_norm": 1.655374526977539, | |
| "learning_rate": 0.00021435920797041404, | |
| "loss": 1.4, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 0.3725458541855527, | |
| "grad_norm": 1.5459624528884888, | |
| "learning_rate": 0.00021392739884837697, | |
| "loss": 1.4103, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.3735393097967141, | |
| "grad_norm": 1.5234006643295288, | |
| "learning_rate": 0.00021349494140213986, | |
| "loss": 1.4223, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 0.3745327654078756, | |
| "grad_norm": 1.538171648979187, | |
| "learning_rate": 0.00021306184001750012, | |
| "loss": 1.3941, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 0.3755262210190371, | |
| "grad_norm": 1.4911220073699951, | |
| "learning_rate": 0.00021262809908678584, | |
| "loss": 1.4091, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 0.3765196766301986, | |
| "grad_norm": 1.4538224935531616, | |
| "learning_rate": 0.00021219372300881117, | |
| "loss": 1.436, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 0.37751313224136, | |
| "grad_norm": 1.6953390836715698, | |
| "learning_rate": 0.0002117587161888315, | |
| "loss": 1.4128, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.3785065878525215, | |
| "grad_norm": 1.6637951135635376, | |
| "learning_rate": 0.00021132308303849906, | |
| "loss": 1.4266, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 0.379500043463683, | |
| "grad_norm": 1.4295332431793213, | |
| "learning_rate": 0.000210886827975818, | |
| "loss": 1.4222, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 0.3804934990748445, | |
| "grad_norm": 1.4962881803512573, | |
| "learning_rate": 0.0002104499554250996, | |
| "loss": 1.3955, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 0.38148695468600596, | |
| "grad_norm": 1.6462197303771973, | |
| "learning_rate": 0.00021001246981691758, | |
| "loss": 1.4074, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 0.3824804102971674, | |
| "grad_norm": 1.5675110816955566, | |
| "learning_rate": 0.0002095743755880628, | |
| "loss": 1.4375, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.3834738659083289, | |
| "grad_norm": 1.6916255950927734, | |
| "learning_rate": 0.0002091356771814987, | |
| "loss": 1.4191, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 0.38446732151949037, | |
| "grad_norm": 1.6852726936340332, | |
| "learning_rate": 0.0002086963790463159, | |
| "loss": 1.4366, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 0.38546077713065185, | |
| "grad_norm": 1.434536099433899, | |
| "learning_rate": 0.0002082564856376873, | |
| "loss": 1.4106, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 0.3864542327418133, | |
| "grad_norm": 1.6124340295791626, | |
| "learning_rate": 0.0002078160014168227, | |
| "loss": 1.4294, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 0.3874476883529748, | |
| "grad_norm": 1.6852911710739136, | |
| "learning_rate": 0.00020737493085092382, | |
| "loss": 1.4188, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.38844114396413626, | |
| "grad_norm": 1.4967029094696045, | |
| "learning_rate": 0.00020693327841313867, | |
| "loss": 1.4323, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 0.38943459957529775, | |
| "grad_norm": 1.5312552452087402, | |
| "learning_rate": 0.00020649104858251648, | |
| "loss": 1.428, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 0.3904280551864592, | |
| "grad_norm": 1.4007198810577393, | |
| "learning_rate": 0.00020604824584396204, | |
| "loss": 1.409, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 0.39142151079762066, | |
| "grad_norm": 1.6712150573730469, | |
| "learning_rate": 0.0002056048746881905, | |
| "loss": 1.4149, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 0.39241496640878215, | |
| "grad_norm": 1.5545682907104492, | |
| "learning_rate": 0.00020516093961168137, | |
| "loss": 1.419, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 0.39340842201994364, | |
| "grad_norm": 1.5229815244674683, | |
| "learning_rate": 0.00020471644511663352, | |
| "loss": 1.3922, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 0.39440187763110507, | |
| "grad_norm": 1.647059440612793, | |
| "learning_rate": 0.00020427139571091893, | |
| "loss": 1.4031, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 0.39539533324226656, | |
| "grad_norm": 1.5931661128997803, | |
| "learning_rate": 0.00020382579590803748, | |
| "loss": 1.3945, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 0.39638878885342804, | |
| "grad_norm": 1.548180103302002, | |
| "learning_rate": 0.00020337965022707084, | |
| "loss": 1.4058, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 0.39738224446458953, | |
| "grad_norm": 1.5571354627609253, | |
| "learning_rate": 0.00020293296319263664, | |
| "loss": 1.4119, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.398375700075751, | |
| "grad_norm": 1.4592036008834839, | |
| "learning_rate": 0.00020248573933484286, | |
| "loss": 1.408, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 0.39936915568691245, | |
| "grad_norm": 1.5266109704971313, | |
| "learning_rate": 0.00020203798318924152, | |
| "loss": 1.3989, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 0.4001639201758416, | |
| "eval_loss": 1.3744559288024902, | |
| "eval_runtime": 36.9341, | |
| "eval_samples_per_second": 1353.761, | |
| "eval_steps_per_second": 169.22, | |
| "step": 4028 | |
| }, | |
| { | |
| "epoch": 0.40036261129807393, | |
| "grad_norm": 1.7276591062545776, | |
| "learning_rate": 0.00020158969929678306, | |
| "loss": 1.3959, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 0.4013560669092354, | |
| "grad_norm": 1.509847640991211, | |
| "learning_rate": 0.00020114089220377005, | |
| "loss": 1.4173, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 0.4023495225203969, | |
| "grad_norm": 1.5896152257919312, | |
| "learning_rate": 0.00020069156646181107, | |
| "loss": 1.4212, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.40334297813155834, | |
| "grad_norm": 1.6238590478897095, | |
| "learning_rate": 0.0002002417266277746, | |
| "loss": 1.4191, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 0.4043364337427198, | |
| "grad_norm": 1.5905038118362427, | |
| "learning_rate": 0.00019979137726374295, | |
| "loss": 1.4013, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 0.4053298893538813, | |
| "grad_norm": 1.6326454877853394, | |
| "learning_rate": 0.0001993405229369658, | |
| "loss": 1.4189, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 0.4063233449650428, | |
| "grad_norm": 1.5920885801315308, | |
| "learning_rate": 0.0001988891682198139, | |
| "loss": 1.4212, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 0.40731680057620423, | |
| "grad_norm": 1.5048717260360718, | |
| "learning_rate": 0.00019843731768973274, | |
| "loss": 1.3998, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.4083102561873657, | |
| "grad_norm": 1.4494785070419312, | |
| "learning_rate": 0.00019798497592919624, | |
| "loss": 1.3922, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 0.4093037117985272, | |
| "grad_norm": 1.4546644687652588, | |
| "learning_rate": 0.00019753214752566006, | |
| "loss": 1.4163, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 0.4102971674096887, | |
| "grad_norm": 1.6149802207946777, | |
| "learning_rate": 0.0001970788370715152, | |
| "loss": 1.396, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 0.4112906230208502, | |
| "grad_norm": 1.4141395092010498, | |
| "learning_rate": 0.0001966250491640415, | |
| "loss": 1.4089, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 0.4122840786320116, | |
| "grad_norm": 1.6508771181106567, | |
| "learning_rate": 0.0001961707884053608, | |
| "loss": 1.3818, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 0.4132775342431731, | |
| "grad_norm": 1.5785223245620728, | |
| "learning_rate": 0.0001957160594023905, | |
| "loss": 1.3971, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 0.4142709898543346, | |
| "grad_norm": 1.416621446609497, | |
| "learning_rate": 0.00019526086676679663, | |
| "loss": 1.395, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 0.41526444546549607, | |
| "grad_norm": 1.430781364440918, | |
| "learning_rate": 0.00019480521511494727, | |
| "loss": 1.3915, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 0.4162579010766575, | |
| "grad_norm": 1.6651394367218018, | |
| "learning_rate": 0.0001943491090678656, | |
| "loss": 1.3797, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 0.417251356687819, | |
| "grad_norm": 1.4518263339996338, | |
| "learning_rate": 0.00019389255325118305, | |
| "loss": 1.403, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.4182448122989805, | |
| "grad_norm": 1.465232491493225, | |
| "learning_rate": 0.0001934355522950926, | |
| "loss": 1.412, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 0.41923826791014196, | |
| "grad_norm": 1.5731855630874634, | |
| "learning_rate": 0.00019297811083430138, | |
| "loss": 1.412, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 0.4202317235213034, | |
| "grad_norm": 1.6133655309677124, | |
| "learning_rate": 0.0001925202335079841, | |
| "loss": 1.3897, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 0.4212251791324649, | |
| "grad_norm": 1.4948188066482544, | |
| "learning_rate": 0.0001920619249597357, | |
| "loss": 1.4263, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 0.42221863474362636, | |
| "grad_norm": 1.6283082962036133, | |
| "learning_rate": 0.00019160318983752465, | |
| "loss": 1.4143, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 0.42321209035478785, | |
| "grad_norm": 1.5405102968215942, | |
| "learning_rate": 0.00019114403279364524, | |
| "loss": 1.3917, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 0.4242055459659493, | |
| "grad_norm": 1.5513877868652344, | |
| "learning_rate": 0.00019068445848467088, | |
| "loss": 1.3965, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 0.42519900157711077, | |
| "grad_norm": 1.6019749641418457, | |
| "learning_rate": 0.00019022447157140667, | |
| "loss": 1.3759, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 0.42619245718827226, | |
| "grad_norm": 1.4718899726867676, | |
| "learning_rate": 0.00018976407671884217, | |
| "loss": 1.3911, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 0.42718591279943374, | |
| "grad_norm": 1.4847612380981445, | |
| "learning_rate": 0.000189303278596104, | |
| "loss": 1.3731, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.42817936841059523, | |
| "grad_norm": 1.6331555843353271, | |
| "learning_rate": 0.00018884208187640875, | |
| "loss": 1.4175, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 0.42917282402175666, | |
| "grad_norm": 1.5743944644927979, | |
| "learning_rate": 0.00018838049123701524, | |
| "loss": 1.4041, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 0.43016627963291815, | |
| "grad_norm": 1.5514403581619263, | |
| "learning_rate": 0.00018791851135917739, | |
| "loss": 1.3995, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 0.43115973524407963, | |
| "grad_norm": 1.4812462329864502, | |
| "learning_rate": 0.00018745614692809644, | |
| "loss": 1.3994, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 0.4321531908552411, | |
| "grad_norm": 1.5028584003448486, | |
| "learning_rate": 0.0001869934026328738, | |
| "loss": 1.3845, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 0.43314664646640255, | |
| "grad_norm": 1.6334466934204102, | |
| "learning_rate": 0.00018653028316646317, | |
| "loss": 1.4243, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 0.43414010207756404, | |
| "grad_norm": 1.444973111152649, | |
| "learning_rate": 0.00018606679322562307, | |
| "loss": 1.3805, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 0.4351335576887255, | |
| "grad_norm": 1.5728076696395874, | |
| "learning_rate": 0.00018560293751086927, | |
| "loss": 1.3954, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 0.436127013299887, | |
| "grad_norm": 1.6659067869186401, | |
| "learning_rate": 0.00018513872072642708, | |
| "loss": 1.3877, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 0.43712046891104844, | |
| "grad_norm": 1.620434045791626, | |
| "learning_rate": 0.00018467414758018358, | |
| "loss": 1.3896, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.43811392452220993, | |
| "grad_norm": 1.4736016988754272, | |
| "learning_rate": 0.0001842092227836398, | |
| "loss": 1.3952, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 0.4391073801333714, | |
| "grad_norm": 1.4808287620544434, | |
| "learning_rate": 0.00018374395105186342, | |
| "loss": 1.4254, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 0.4401008357445329, | |
| "grad_norm": 1.6416784524917603, | |
| "learning_rate": 0.00018327833710344012, | |
| "loss": 1.3964, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 0.4410942913556944, | |
| "grad_norm": 1.6272752285003662, | |
| "learning_rate": 0.00018281238566042655, | |
| "loss": 1.4047, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 0.4420877469668558, | |
| "grad_norm": 1.5633876323699951, | |
| "learning_rate": 0.00018234610144830189, | |
| "loss": 1.3918, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 0.4430812025780173, | |
| "grad_norm": 1.547147512435913, | |
| "learning_rate": 0.00018187948919592023, | |
| "loss": 1.3854, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 0.4440746581891788, | |
| "grad_norm": 1.9166932106018066, | |
| "learning_rate": 0.00018141255363546244, | |
| "loss": 1.3926, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 0.4450681138003403, | |
| "grad_norm": 1.4858520030975342, | |
| "learning_rate": 0.00018094529950238829, | |
| "loss": 1.3788, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 0.4460615694115017, | |
| "grad_norm": 1.5069504976272583, | |
| "learning_rate": 0.00018047773153538835, | |
| "loss": 1.4112, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 0.4470550250226632, | |
| "grad_norm": 1.584439992904663, | |
| "learning_rate": 0.00018000985447633592, | |
| "loss": 1.3971, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.4480484806338247, | |
| "grad_norm": 1.5546822547912598, | |
| "learning_rate": 0.00017954167307023902, | |
| "loss": 1.4037, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 0.4490419362449862, | |
| "grad_norm": 1.5818507671356201, | |
| "learning_rate": 0.00017907319206519228, | |
| "loss": 1.3868, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 0.4500353918561476, | |
| "grad_norm": 1.5756354331970215, | |
| "learning_rate": 0.00017860441621232864, | |
| "loss": 1.399, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 0.4510288474673091, | |
| "grad_norm": 1.6583040952682495, | |
| "learning_rate": 0.00017813535026577128, | |
| "loss": 1.3802, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 0.4520223030784706, | |
| "grad_norm": 1.4960025548934937, | |
| "learning_rate": 0.0001776659989825854, | |
| "loss": 1.3994, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 0.45301575868963206, | |
| "grad_norm": 1.450750708580017, | |
| "learning_rate": 0.00017719636712273005, | |
| "loss": 1.3845, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 0.4540092143007935, | |
| "grad_norm": 1.7412645816802979, | |
| "learning_rate": 0.0001767264594490095, | |
| "loss": 1.3863, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 0.455002669911955, | |
| "grad_norm": 1.3514049053192139, | |
| "learning_rate": 0.00017625628072702562, | |
| "loss": 1.3675, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 0.45599612552311647, | |
| "grad_norm": 1.4116798639297485, | |
| "learning_rate": 0.00017578583572512877, | |
| "loss": 1.3907, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 0.45698958113427796, | |
| "grad_norm": 1.4500694274902344, | |
| "learning_rate": 0.00017531512921436997, | |
| "loss": 1.4047, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.45798303674543944, | |
| "grad_norm": 1.5988340377807617, | |
| "learning_rate": 0.00017484416596845226, | |
| "loss": 1.3995, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 0.4589764923566009, | |
| "grad_norm": 1.6010870933532715, | |
| "learning_rate": 0.00017437295076368252, | |
| "loss": 1.373, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 0.45996994796776236, | |
| "grad_norm": 1.5147273540496826, | |
| "learning_rate": 0.0001739014883789228, | |
| "loss": 1.3713, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 0.46096340357892385, | |
| "grad_norm": 1.5335299968719482, | |
| "learning_rate": 0.00017342978359554187, | |
| "loss": 1.3943, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 0.46195685919008533, | |
| "grad_norm": 1.6468431949615479, | |
| "learning_rate": 0.00017295784119736698, | |
| "loss": 1.4106, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 0.46295031480124677, | |
| "grad_norm": 1.6687713861465454, | |
| "learning_rate": 0.000172485665970635, | |
| "loss": 1.3766, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 0.46394377041240825, | |
| "grad_norm": 1.4923644065856934, | |
| "learning_rate": 0.00017201326270394418, | |
| "loss": 1.402, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 0.46493722602356974, | |
| "grad_norm": 1.5864273309707642, | |
| "learning_rate": 0.00017154063618820533, | |
| "loss": 1.4031, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 0.4659306816347312, | |
| "grad_norm": 1.5576728582382202, | |
| "learning_rate": 0.00017106779121659344, | |
| "loss": 1.3918, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 0.46692413724589266, | |
| "grad_norm": 1.516586184501648, | |
| "learning_rate": 0.00017059473258449894, | |
| "loss": 1.3903, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.46791759285705414, | |
| "grad_norm": 1.5140126943588257, | |
| "learning_rate": 0.00017012146508947916, | |
| "loss": 1.4046, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 0.46891104846821563, | |
| "grad_norm": 1.5689266920089722, | |
| "learning_rate": 0.00016964799353120963, | |
| "loss": 1.3804, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 0.4699045040793771, | |
| "grad_norm": 1.459067463874817, | |
| "learning_rate": 0.00016917432271143534, | |
| "loss": 1.4024, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 0.4708979596905386, | |
| "grad_norm": 1.4880290031433105, | |
| "learning_rate": 0.00016870045743392207, | |
| "loss": 1.3897, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 0.47189141530170003, | |
| "grad_norm": 1.619788408279419, | |
| "learning_rate": 0.0001682264025044079, | |
| "loss": 1.3494, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 0.4728848709128615, | |
| "grad_norm": 1.5221748352050781, | |
| "learning_rate": 0.00016775216273055406, | |
| "loss": 1.386, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 0.473878326524023, | |
| "grad_norm": 1.6048988103866577, | |
| "learning_rate": 0.00016727774292189657, | |
| "loss": 1.3615, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 0.4748717821351845, | |
| "grad_norm": 1.6888704299926758, | |
| "learning_rate": 0.0001668031478897971, | |
| "loss": 1.387, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 0.4758652377463459, | |
| "grad_norm": 1.657988429069519, | |
| "learning_rate": 0.00016632838244739454, | |
| "loss": 1.3829, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 0.4768586933575074, | |
| "grad_norm": 1.4432212114334106, | |
| "learning_rate": 0.00016585345140955593, | |
| "loss": 1.3703, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.4778521489686689, | |
| "grad_norm": 1.5761727094650269, | |
| "learning_rate": 0.00016537835959282767, | |
| "loss": 1.3943, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 0.4788456045798304, | |
| "grad_norm": 1.5926445722579956, | |
| "learning_rate": 0.00016490311181538672, | |
| "loss": 1.3769, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 0.4798390601909918, | |
| "grad_norm": 1.606236457824707, | |
| "learning_rate": 0.00016442771289699185, | |
| "loss": 1.3902, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 0.4808325158021533, | |
| "grad_norm": 1.6025710105895996, | |
| "learning_rate": 0.00016395216765893443, | |
| "loss": 1.3933, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 0.4818259714133148, | |
| "grad_norm": 1.5865193605422974, | |
| "learning_rate": 0.00016347648092398997, | |
| "loss": 1.3828, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 0.4828194270244763, | |
| "grad_norm": 1.588725209236145, | |
| "learning_rate": 0.00016300065751636878, | |
| "loss": 1.3841, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 0.4838128826356377, | |
| "grad_norm": 1.5232126712799072, | |
| "learning_rate": 0.00016252470226166735, | |
| "loss": 1.4022, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 0.4848063382467992, | |
| "grad_norm": 1.5612798929214478, | |
| "learning_rate": 0.0001620486199868193, | |
| "loss": 1.3964, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 0.4857997938579607, | |
| "grad_norm": 1.5628079175949097, | |
| "learning_rate": 0.0001615724155200464, | |
| "loss": 1.3962, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 0.48679324946912217, | |
| "grad_norm": 1.5569151639938354, | |
| "learning_rate": 0.00016109609369080964, | |
| "loss": 1.3772, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.48778670508028366, | |
| "grad_norm": 1.5461227893829346, | |
| "learning_rate": 0.00016061965932976033, | |
| "loss": 1.3846, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 0.4887801606914451, | |
| "grad_norm": 1.439954161643982, | |
| "learning_rate": 0.00016014311726869092, | |
| "loss": 1.3837, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 0.4897736163026066, | |
| "grad_norm": 1.5704083442687988, | |
| "learning_rate": 0.00015966647234048618, | |
| "loss": 1.3838, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 0.49076707191376806, | |
| "grad_norm": 1.4509100914001465, | |
| "learning_rate": 0.000159189729379074, | |
| "loss": 1.4002, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 0.49176052752492955, | |
| "grad_norm": 1.4755138158798218, | |
| "learning_rate": 0.0001587128932193767, | |
| "loss": 1.3691, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 0.492753983136091, | |
| "grad_norm": 1.579158902168274, | |
| "learning_rate": 0.00015823596869726152, | |
| "loss": 1.3796, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 0.49374743874725247, | |
| "grad_norm": 1.561621904373169, | |
| "learning_rate": 0.0001577589606494921, | |
| "loss": 1.3804, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 0.49474089435841395, | |
| "grad_norm": 1.5103092193603516, | |
| "learning_rate": 0.0001572818739136789, | |
| "loss": 1.3879, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 0.49573434996957544, | |
| "grad_norm": 1.571546196937561, | |
| "learning_rate": 0.00015680471332823055, | |
| "loss": 1.3838, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 0.49672780558073687, | |
| "grad_norm": 1.494307041168213, | |
| "learning_rate": 0.00015632748373230462, | |
| "loss": 1.3872, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.49772126119189836, | |
| "grad_norm": 1.575735330581665, | |
| "learning_rate": 0.00015585018996575853, | |
| "loss": 1.3639, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 0.49871471680305984, | |
| "grad_norm": 1.6904993057250977, | |
| "learning_rate": 0.00015537283686910047, | |
| "loss": 1.3772, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 0.49970817241422133, | |
| "grad_norm": 1.46591055393219, | |
| "learning_rate": 0.00015489542928344037, | |
| "loss": 1.3856, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 0.500204900219802, | |
| "eval_loss": 1.3503732681274414, | |
| "eval_runtime": 37.2894, | |
| "eval_samples_per_second": 1340.864, | |
| "eval_steps_per_second": 167.608, | |
| "step": 5035 | |
| }, | |
| { | |
| "epoch": 0.5007016280253828, | |
| "grad_norm": 1.698063850402832, | |
| "learning_rate": 0.00015441797205044083, | |
| "loss": 1.3936, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 0.5016950836365442, | |
| "grad_norm": 1.5648193359375, | |
| "learning_rate": 0.00015394047001226775, | |
| "loss": 1.3901, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 0.5026885392477057, | |
| "grad_norm": 1.543129324913025, | |
| "learning_rate": 0.0001534629280115416, | |
| "loss": 1.3873, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 0.5036819948588672, | |
| "grad_norm": 1.709431767463684, | |
| "learning_rate": 0.00015298535089128813, | |
| "loss": 1.3786, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 0.5046754504700287, | |
| "grad_norm": 1.4675122499465942, | |
| "learning_rate": 0.00015250774349488916, | |
| "loss": 1.3662, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 0.5056689060811902, | |
| "grad_norm": 1.4313937425613403, | |
| "learning_rate": 0.00015203011066603357, | |
| "loss": 1.3825, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 0.5066623616923517, | |
| "grad_norm": 1.5015852451324463, | |
| "learning_rate": 0.0001515524572486683, | |
| "loss": 1.3781, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.5076558173035131, | |
| "grad_norm": 1.491697907447815, | |
| "learning_rate": 0.00015107478808694887, | |
| "loss": 1.3855, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 0.5086492729146745, | |
| "grad_norm": 1.6736730337142944, | |
| "learning_rate": 0.0001505971080251907, | |
| "loss": 1.3746, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 0.509642728525836, | |
| "grad_norm": 1.6208219528198242, | |
| "learning_rate": 0.00015011942190781954, | |
| "loss": 1.372, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 0.5106361841369975, | |
| "grad_norm": 1.6461440324783325, | |
| "learning_rate": 0.0001496417345793228, | |
| "loss": 1.3778, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 0.511629639748159, | |
| "grad_norm": 1.450124740600586, | |
| "learning_rate": 0.00014916405088419998, | |
| "loss": 1.3954, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 0.5126230953593205, | |
| "grad_norm": 1.549975037574768, | |
| "learning_rate": 0.0001486863756669137, | |
| "loss": 1.367, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 0.513616550970482, | |
| "grad_norm": 1.5281715393066406, | |
| "learning_rate": 0.00014820871377184075, | |
| "loss": 1.3807, | |
| "step": 5170 | |
| }, | |
| { | |
| "epoch": 0.5146100065816435, | |
| "grad_norm": 1.666485071182251, | |
| "learning_rate": 0.00014773107004322278, | |
| "loss": 1.3732, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 0.515603462192805, | |
| "grad_norm": 1.5841244459152222, | |
| "learning_rate": 0.0001472534493251171, | |
| "loss": 1.3723, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 0.5165969178039663, | |
| "grad_norm": 1.6376242637634277, | |
| "learning_rate": 0.00014677585646134778, | |
| "loss": 1.3823, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.5175903734151278, | |
| "grad_norm": 1.4951201677322388, | |
| "learning_rate": 0.00014629829629545632, | |
| "loss": 1.3523, | |
| "step": 5210 | |
| }, | |
| { | |
| "epoch": 0.5185838290262893, | |
| "grad_norm": 1.6371444463729858, | |
| "learning_rate": 0.00014582077367065278, | |
| "loss": 1.3734, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 0.5195772846374508, | |
| "grad_norm": 1.564769983291626, | |
| "learning_rate": 0.00014534329342976612, | |
| "loss": 1.3774, | |
| "step": 5230 | |
| }, | |
| { | |
| "epoch": 0.5205707402486123, | |
| "grad_norm": 1.5347799062728882, | |
| "learning_rate": 0.00014486586041519587, | |
| "loss": 1.3859, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 0.5215641958597738, | |
| "grad_norm": 1.6608774662017822, | |
| "learning_rate": 0.00014438847946886236, | |
| "loss": 1.4043, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 0.5225576514709352, | |
| "grad_norm": 1.5226597785949707, | |
| "learning_rate": 0.00014391115543215801, | |
| "loss": 1.3695, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 0.5235511070820967, | |
| "grad_norm": 1.5455684661865234, | |
| "learning_rate": 0.00014343389314589797, | |
| "loss": 1.3732, | |
| "step": 5270 | |
| }, | |
| { | |
| "epoch": 0.5245445626932581, | |
| "grad_norm": 1.5501607656478882, | |
| "learning_rate": 0.00014295669745027125, | |
| "loss": 1.3734, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 0.5255380183044196, | |
| "grad_norm": 1.4892995357513428, | |
| "learning_rate": 0.00014247957318479148, | |
| "loss": 1.3737, | |
| "step": 5290 | |
| }, | |
| { | |
| "epoch": 0.5265314739155811, | |
| "grad_norm": 1.543616533279419, | |
| "learning_rate": 0.0001420025251882478, | |
| "loss": 1.3929, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.5275249295267426, | |
| "grad_norm": 1.5783209800720215, | |
| "learning_rate": 0.00014152555829865593, | |
| "loss": 1.3686, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 0.5285183851379041, | |
| "grad_norm": 1.6346206665039062, | |
| "learning_rate": 0.00014104867735320912, | |
| "loss": 1.3782, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 0.5295118407490655, | |
| "grad_norm": 1.4646329879760742, | |
| "learning_rate": 0.0001405718871882289, | |
| "loss": 1.3536, | |
| "step": 5330 | |
| }, | |
| { | |
| "epoch": 0.530505296360227, | |
| "grad_norm": 1.5457079410552979, | |
| "learning_rate": 0.00014009519263911616, | |
| "loss": 1.3764, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 0.5314987519713885, | |
| "grad_norm": 1.5181694030761719, | |
| "learning_rate": 0.00013961859854030208, | |
| "loss": 1.3815, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 0.53249220758255, | |
| "grad_norm": 1.6100943088531494, | |
| "learning_rate": 0.00013914210972519928, | |
| "loss": 1.3847, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 0.5334856631937114, | |
| "grad_norm": 1.4964196681976318, | |
| "learning_rate": 0.00013866573102615237, | |
| "loss": 1.3781, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 0.5344791188048729, | |
| "grad_norm": 1.548790454864502, | |
| "learning_rate": 0.0001381894672743894, | |
| "loss": 1.3806, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 0.5354725744160344, | |
| "grad_norm": 1.5508662462234497, | |
| "learning_rate": 0.0001377133232999726, | |
| "loss": 1.3621, | |
| "step": 5390 | |
| }, | |
| { | |
| "epoch": 0.5364660300271958, | |
| "grad_norm": 1.5982120037078857, | |
| "learning_rate": 0.0001372373039317495, | |
| "loss": 1.3805, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.5374594856383573, | |
| "grad_norm": 1.4170911312103271, | |
| "learning_rate": 0.0001367614139973039, | |
| "loss": 1.364, | |
| "step": 5410 | |
| }, | |
| { | |
| "epoch": 0.5384529412495188, | |
| "grad_norm": 1.7373018264770508, | |
| "learning_rate": 0.000136285658322907, | |
| "loss": 1.3839, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 0.5394463968606803, | |
| "grad_norm": 1.551241159439087, | |
| "learning_rate": 0.00013581004173346834, | |
| "loss": 1.3736, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 0.5404398524718418, | |
| "grad_norm": 1.4420864582061768, | |
| "learning_rate": 0.00013533456905248683, | |
| "loss": 1.3651, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 0.5414333080830032, | |
| "grad_norm": 1.5889432430267334, | |
| "learning_rate": 0.00013485924510200215, | |
| "loss": 1.3724, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 0.5424267636941646, | |
| "grad_norm": 1.5864027738571167, | |
| "learning_rate": 0.00013438407470254543, | |
| "loss": 1.3712, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 0.5434202193053261, | |
| "grad_norm": 1.521699070930481, | |
| "learning_rate": 0.0001339090626730906, | |
| "loss": 1.3796, | |
| "step": 5470 | |
| }, | |
| { | |
| "epoch": 0.5444136749164876, | |
| "grad_norm": 1.544364094734192, | |
| "learning_rate": 0.00013343421383100554, | |
| "loss": 1.3538, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 0.5454071305276491, | |
| "grad_norm": 1.4789470434188843, | |
| "learning_rate": 0.000132959532992003, | |
| "loss": 1.3563, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 0.5464005861388106, | |
| "grad_norm": 1.53242027759552, | |
| "learning_rate": 0.0001324850249700921, | |
| "loss": 1.3809, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.5473940417499721, | |
| "grad_norm": 1.5572789907455444, | |
| "learning_rate": 0.0001320106945775292, | |
| "loss": 1.3738, | |
| "step": 5510 | |
| }, | |
| { | |
| "epoch": 0.5483874973611336, | |
| "grad_norm": 1.5835766792297363, | |
| "learning_rate": 0.0001315365466247692, | |
| "loss": 1.3374, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 0.5493809529722951, | |
| "grad_norm": 1.5495179891586304, | |
| "learning_rate": 0.0001310625859204169, | |
| "loss": 1.3823, | |
| "step": 5530 | |
| }, | |
| { | |
| "epoch": 0.5503744085834564, | |
| "grad_norm": 1.623462200164795, | |
| "learning_rate": 0.00013058881727117798, | |
| "loss": 1.3615, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 0.5513678641946179, | |
| "grad_norm": 1.6338903903961182, | |
| "learning_rate": 0.00013011524548181048, | |
| "loss": 1.3755, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 0.5523613198057794, | |
| "grad_norm": 1.5701446533203125, | |
| "learning_rate": 0.00012964187535507594, | |
| "loss": 1.3463, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 0.5533547754169409, | |
| "grad_norm": 1.4608359336853027, | |
| "learning_rate": 0.00012916871169169061, | |
| "loss": 1.3778, | |
| "step": 5570 | |
| }, | |
| { | |
| "epoch": 0.5543482310281024, | |
| "grad_norm": 1.4946131706237793, | |
| "learning_rate": 0.00012869575929027704, | |
| "loss": 1.3697, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 0.5553416866392639, | |
| "grad_norm": 1.4983999729156494, | |
| "learning_rate": 0.0001282230229473152, | |
| "loss": 1.3803, | |
| "step": 5590 | |
| }, | |
| { | |
| "epoch": 0.5563351422504254, | |
| "grad_norm": 1.4606608152389526, | |
| "learning_rate": 0.00012775050745709383, | |
| "loss": 1.3684, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.5573285978615868, | |
| "grad_norm": 1.5483719110488892, | |
| "learning_rate": 0.00012727821761166196, | |
| "loss": 1.3661, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 0.5583220534727483, | |
| "grad_norm": 1.773607850074768, | |
| "learning_rate": 0.00012680615820078017, | |
| "loss": 1.3696, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 0.5593155090839097, | |
| "grad_norm": 1.53352952003479, | |
| "learning_rate": 0.0001263343340118722, | |
| "loss": 1.3649, | |
| "step": 5630 | |
| }, | |
| { | |
| "epoch": 0.5603089646950712, | |
| "grad_norm": 1.6477464437484741, | |
| "learning_rate": 0.00012586274982997596, | |
| "loss": 1.389, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 0.5613024203062327, | |
| "grad_norm": 1.4398772716522217, | |
| "learning_rate": 0.00012539141043769566, | |
| "loss": 1.3714, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 0.5622958759173942, | |
| "grad_norm": 1.477419376373291, | |
| "learning_rate": 0.00012492032061515274, | |
| "loss": 1.3503, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 0.5632893315285556, | |
| "grad_norm": 1.5293781757354736, | |
| "learning_rate": 0.00012444948513993774, | |
| "loss": 1.3646, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 0.5642827871397171, | |
| "grad_norm": 1.5651262998580933, | |
| "learning_rate": 0.00012397890878706162, | |
| "loss": 1.3635, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 0.5652762427508786, | |
| "grad_norm": 1.642146348953247, | |
| "learning_rate": 0.00012350859632890752, | |
| "loss": 1.3772, | |
| "step": 5690 | |
| }, | |
| { | |
| "epoch": 0.5662696983620401, | |
| "grad_norm": 1.7528361082077026, | |
| "learning_rate": 0.00012303855253518222, | |
| "loss": 1.3536, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.5672631539732015, | |
| "grad_norm": 1.6384177207946777, | |
| "learning_rate": 0.00012256878217286774, | |
| "loss": 1.373, | |
| "step": 5710 | |
| }, | |
| { | |
| "epoch": 0.568256609584363, | |
| "grad_norm": 1.6027686595916748, | |
| "learning_rate": 0.0001220992900061733, | |
| "loss": 1.3764, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 0.5692500651955245, | |
| "grad_norm": 1.4915357828140259, | |
| "learning_rate": 0.00012163008079648655, | |
| "loss": 1.3745, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 0.5702435208066859, | |
| "grad_norm": 1.6192013025283813, | |
| "learning_rate": 0.0001211611593023256, | |
| "loss": 1.3501, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 0.5712369764178474, | |
| "grad_norm": 1.615323543548584, | |
| "learning_rate": 0.00012069253027929076, | |
| "loss": 1.3537, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 0.5722304320290089, | |
| "grad_norm": 1.6354780197143555, | |
| "learning_rate": 0.00012022419848001607, | |
| "loss": 1.3627, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 0.5732238876401704, | |
| "grad_norm": 1.5796200037002563, | |
| "learning_rate": 0.0001197561686541214, | |
| "loss": 1.3535, | |
| "step": 5770 | |
| }, | |
| { | |
| "epoch": 0.5742173432513319, | |
| "grad_norm": 1.5657954216003418, | |
| "learning_rate": 0.00011928844554816401, | |
| "loss": 1.3741, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 0.5752107988624934, | |
| "grad_norm": 1.6198936700820923, | |
| "learning_rate": 0.00011882103390559055, | |
| "loss": 1.3703, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 0.5762042544736548, | |
| "grad_norm": 1.42495858669281, | |
| "learning_rate": 0.00011835393846668907, | |
| "loss": 1.3671, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.5771977100848162, | |
| "grad_norm": 1.5071784257888794, | |
| "learning_rate": 0.00011788716396854067, | |
| "loss": 1.3782, | |
| "step": 5810 | |
| }, | |
| { | |
| "epoch": 0.5781911656959777, | |
| "grad_norm": 1.4957716464996338, | |
| "learning_rate": 0.00011742071514497172, | |
| "loss": 1.3667, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 0.5791846213071392, | |
| "grad_norm": 1.5235507488250732, | |
| "learning_rate": 0.00011695459672650566, | |
| "loss": 1.379, | |
| "step": 5830 | |
| }, | |
| { | |
| "epoch": 0.5801780769183007, | |
| "grad_norm": 1.6448169946670532, | |
| "learning_rate": 0.00011648881344031516, | |
| "loss": 1.3635, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 0.5811715325294622, | |
| "grad_norm": 1.7475203275680542, | |
| "learning_rate": 0.00011602337001017399, | |
| "loss": 1.3737, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 0.5821649881406237, | |
| "grad_norm": 1.6699399948120117, | |
| "learning_rate": 0.00011555827115640947, | |
| "loss": 1.3568, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 0.5831584437517852, | |
| "grad_norm": 1.6669524908065796, | |
| "learning_rate": 0.00011509352159585417, | |
| "loss": 1.3423, | |
| "step": 5870 | |
| }, | |
| { | |
| "epoch": 0.5841518993629465, | |
| "grad_norm": 1.6885567903518677, | |
| "learning_rate": 0.00011462912604179843, | |
| "loss": 1.3659, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 0.585145354974108, | |
| "grad_norm": 1.560099720954895, | |
| "learning_rate": 0.00011416508920394228, | |
| "loss": 1.3462, | |
| "step": 5890 | |
| }, | |
| { | |
| "epoch": 0.5861388105852695, | |
| "grad_norm": 1.503902554512024, | |
| "learning_rate": 0.00011370141578834791, | |
| "loss": 1.356, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.587132266196431, | |
| "grad_norm": 1.594883680343628, | |
| "learning_rate": 0.00011323811049739184, | |
| "loss": 1.3652, | |
| "step": 5910 | |
| }, | |
| { | |
| "epoch": 0.5881257218075925, | |
| "grad_norm": 1.5127500295639038, | |
| "learning_rate": 0.00011277517802971704, | |
| "loss": 1.3622, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 0.589119177418754, | |
| "grad_norm": 1.5737817287445068, | |
| "learning_rate": 0.00011231262308018568, | |
| "loss": 1.3602, | |
| "step": 5930 | |
| }, | |
| { | |
| "epoch": 0.5901126330299155, | |
| "grad_norm": 1.6930840015411377, | |
| "learning_rate": 0.00011185045033983126, | |
| "loss": 1.3733, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 0.5911060886410769, | |
| "grad_norm": 1.6442506313323975, | |
| "learning_rate": 0.00011138866449581098, | |
| "loss": 1.3519, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 0.5920995442522384, | |
| "grad_norm": 1.5289831161499023, | |
| "learning_rate": 0.00011092727023135847, | |
| "loss": 1.3658, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 0.5930929998633998, | |
| "grad_norm": 1.6645240783691406, | |
| "learning_rate": 0.00011046627222573598, | |
| "loss": 1.3462, | |
| "step": 5970 | |
| }, | |
| { | |
| "epoch": 0.5940864554745613, | |
| "grad_norm": 1.6838901042938232, | |
| "learning_rate": 0.00011000567515418709, | |
| "loss": 1.3561, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 0.5950799110857228, | |
| "grad_norm": 1.6393985748291016, | |
| "learning_rate": 0.00010954548368788932, | |
| "loss": 1.3386, | |
| "step": 5990 | |
| }, | |
| { | |
| "epoch": 0.5960733666968843, | |
| "grad_norm": 1.5750504732131958, | |
| "learning_rate": 0.00010908570249390679, | |
| "loss": 1.3374, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.5970668223080458, | |
| "grad_norm": 1.6576467752456665, | |
| "learning_rate": 0.0001086263362351426, | |
| "loss": 1.3596, | |
| "step": 6010 | |
| }, | |
| { | |
| "epoch": 0.5980602779192072, | |
| "grad_norm": 1.7093911170959473, | |
| "learning_rate": 0.00010816738957029202, | |
| "loss": 1.3678, | |
| "step": 6020 | |
| }, | |
| { | |
| "epoch": 0.5990537335303687, | |
| "grad_norm": 1.5973155498504639, | |
| "learning_rate": 0.00010770886715379475, | |
| "loss": 1.3526, | |
| "step": 6030 | |
| }, | |
| { | |
| "epoch": 0.6000471891415302, | |
| "grad_norm": 1.5218548774719238, | |
| "learning_rate": 0.00010725077363578817, | |
| "loss": 1.3501, | |
| "step": 6040 | |
| }, | |
| { | |
| "epoch": 0.6002458802637625, | |
| "eval_loss": 1.3313369750976562, | |
| "eval_runtime": 37.2268, | |
| "eval_samples_per_second": 1343.12, | |
| "eval_steps_per_second": 167.89, | |
| "step": 6042 | |
| }, | |
| { | |
| "epoch": 0.6010406447526916, | |
| "grad_norm": 1.6207534074783325, | |
| "learning_rate": 0.00010679311366205971, | |
| "loss": 1.3678, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 0.6020341003638531, | |
| "grad_norm": 1.4012274742126465, | |
| "learning_rate": 0.00010633589187400009, | |
| "loss": 1.3619, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 0.6030275559750146, | |
| "grad_norm": 1.7136772871017456, | |
| "learning_rate": 0.00010587911290855625, | |
| "loss": 1.3547, | |
| "step": 6070 | |
| }, | |
| { | |
| "epoch": 0.604021011586176, | |
| "grad_norm": 1.6593294143676758, | |
| "learning_rate": 0.00010542278139818404, | |
| "loss": 1.3576, | |
| "step": 6080 | |
| }, | |
| { | |
| "epoch": 0.6050144671973375, | |
| "grad_norm": 1.5701771974563599, | |
| "learning_rate": 0.00010496690197080147, | |
| "loss": 1.3674, | |
| "step": 6090 | |
| }, | |
| { | |
| "epoch": 0.606007922808499, | |
| "grad_norm": 1.6840333938598633, | |
| "learning_rate": 0.00010451147924974182, | |
| "loss": 1.3482, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.6070013784196605, | |
| "grad_norm": 1.5875823497772217, | |
| "learning_rate": 0.00010405651785370654, | |
| "loss": 1.351, | |
| "step": 6110 | |
| }, | |
| { | |
| "epoch": 0.607994834030822, | |
| "grad_norm": 1.6702415943145752, | |
| "learning_rate": 0.00010360202239671843, | |
| "loss": 1.3532, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 0.6089882896419835, | |
| "grad_norm": 1.4088718891143799, | |
| "learning_rate": 0.0001031479974880752, | |
| "loss": 1.3694, | |
| "step": 6130 | |
| }, | |
| { | |
| "epoch": 0.6099817452531449, | |
| "grad_norm": 1.5418723821640015, | |
| "learning_rate": 0.0001026944477323022, | |
| "loss": 1.3584, | |
| "step": 6140 | |
| }, | |
| { | |
| "epoch": 0.6109752008643063, | |
| "grad_norm": 1.6395437717437744, | |
| "learning_rate": 0.00010224137772910618, | |
| "loss": 1.3572, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 0.6119686564754678, | |
| "grad_norm": 1.5876171588897705, | |
| "learning_rate": 0.00010178879207332834, | |
| "loss": 1.3526, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 0.6129621120866293, | |
| "grad_norm": 1.5576690435409546, | |
| "learning_rate": 0.0001013366953548978, | |
| "loss": 1.3624, | |
| "step": 6170 | |
| }, | |
| { | |
| "epoch": 0.6139555676977908, | |
| "grad_norm": 1.6883585453033447, | |
| "learning_rate": 0.00010088509215878528, | |
| "loss": 1.3603, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 0.6149490233089523, | |
| "grad_norm": 1.6483464241027832, | |
| "learning_rate": 0.00010043398706495613, | |
| "loss": 1.3605, | |
| "step": 6190 | |
| }, | |
| { | |
| "epoch": 0.6159424789201138, | |
| "grad_norm": 1.6271207332611084, | |
| "learning_rate": 9.998338464832434e-05, | |
| "loss": 1.3825, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.6169359345312753, | |
| "grad_norm": 1.5564217567443848, | |
| "learning_rate": 9.953328947870598e-05, | |
| "loss": 1.3467, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 0.6179293901424368, | |
| "grad_norm": 1.5181758403778076, | |
| "learning_rate": 9.908370612077272e-05, | |
| "loss": 1.353, | |
| "step": 6220 | |
| }, | |
| { | |
| "epoch": 0.6189228457535981, | |
| "grad_norm": 1.6605936288833618, | |
| "learning_rate": 9.863463913400582e-05, | |
| "loss": 1.3348, | |
| "step": 6230 | |
| }, | |
| { | |
| "epoch": 0.6199163013647596, | |
| "grad_norm": 1.678219199180603, | |
| "learning_rate": 9.818609307264953e-05, | |
| "loss": 1.3466, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 0.6209097569759211, | |
| "grad_norm": 1.5992830991744995, | |
| "learning_rate": 9.773807248566534e-05, | |
| "loss": 1.3562, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 0.6219032125870826, | |
| "grad_norm": 1.65514075756073, | |
| "learning_rate": 9.729058191668527e-05, | |
| "loss": 1.3466, | |
| "step": 6260 | |
| }, | |
| { | |
| "epoch": 0.6228966681982441, | |
| "grad_norm": 1.5987416505813599, | |
| "learning_rate": 9.684362590396643e-05, | |
| "loss": 1.3684, | |
| "step": 6270 | |
| }, | |
| { | |
| "epoch": 0.6238901238094056, | |
| "grad_norm": 1.5358765125274658, | |
| "learning_rate": 9.639720898034454e-05, | |
| "loss": 1.3716, | |
| "step": 6280 | |
| }, | |
| { | |
| "epoch": 0.624883579420567, | |
| "grad_norm": 1.4763336181640625, | |
| "learning_rate": 9.59513356731882e-05, | |
| "loss": 1.346, | |
| "step": 6290 | |
| }, | |
| { | |
| "epoch": 0.6258770350317285, | |
| "grad_norm": 1.6046220064163208, | |
| "learning_rate": 9.550601050435279e-05, | |
| "loss": 1.3653, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.6268704906428899, | |
| "grad_norm": 1.5814241170883179, | |
| "learning_rate": 9.506123799013476e-05, | |
| "loss": 1.3433, | |
| "step": 6310 | |
| }, | |
| { | |
| "epoch": 0.6278639462540514, | |
| "grad_norm": 1.6199915409088135, | |
| "learning_rate": 9.461702264122584e-05, | |
| "loss": 1.3399, | |
| "step": 6320 | |
| }, | |
| { | |
| "epoch": 0.6288574018652129, | |
| "grad_norm": 1.5651947259902954, | |
| "learning_rate": 9.417336896266701e-05, | |
| "loss": 1.3487, | |
| "step": 6330 | |
| }, | |
| { | |
| "epoch": 0.6298508574763744, | |
| "grad_norm": 1.6139787435531616, | |
| "learning_rate": 9.373028145380326e-05, | |
| "loss": 1.3561, | |
| "step": 6340 | |
| }, | |
| { | |
| "epoch": 0.6308443130875359, | |
| "grad_norm": 1.6268047094345093, | |
| "learning_rate": 9.328776460823764e-05, | |
| "loss": 1.3511, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 0.6318377686986973, | |
| "grad_norm": 1.6475530862808228, | |
| "learning_rate": 9.284582291378573e-05, | |
| "loss": 1.3597, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 0.6328312243098588, | |
| "grad_norm": 1.6050212383270264, | |
| "learning_rate": 9.240446085243033e-05, | |
| "loss": 1.37, | |
| "step": 6370 | |
| }, | |
| { | |
| "epoch": 0.6338246799210203, | |
| "grad_norm": 1.4539451599121094, | |
| "learning_rate": 9.196368290027572e-05, | |
| "loss": 1.3435, | |
| "step": 6380 | |
| }, | |
| { | |
| "epoch": 0.6348181355321818, | |
| "grad_norm": 1.6843843460083008, | |
| "learning_rate": 9.152349352750232e-05, | |
| "loss": 1.3837, | |
| "step": 6390 | |
| }, | |
| { | |
| "epoch": 0.6358115911433432, | |
| "grad_norm": 1.5121159553527832, | |
| "learning_rate": 9.108389719832165e-05, | |
| "loss": 1.3503, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.6368050467545047, | |
| "grad_norm": 1.670411467552185, | |
| "learning_rate": 9.064489837093069e-05, | |
| "loss": 1.3473, | |
| "step": 6410 | |
| }, | |
| { | |
| "epoch": 0.6377985023656662, | |
| "grad_norm": 1.771976113319397, | |
| "learning_rate": 9.020650149746679e-05, | |
| "loss": 1.3494, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 0.6387919579768276, | |
| "grad_norm": 1.6656084060668945, | |
| "learning_rate": 8.976871102396269e-05, | |
| "loss": 1.3669, | |
| "step": 6430 | |
| }, | |
| { | |
| "epoch": 0.6397854135879891, | |
| "grad_norm": 1.590765357017517, | |
| "learning_rate": 8.93315313903011e-05, | |
| "loss": 1.3527, | |
| "step": 6440 | |
| }, | |
| { | |
| "epoch": 0.6407788691991506, | |
| "grad_norm": 1.711053729057312, | |
| "learning_rate": 8.889496703017006e-05, | |
| "loss": 1.3679, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 0.6417723248103121, | |
| "grad_norm": 1.60274076461792, | |
| "learning_rate": 8.845902237101754e-05, | |
| "loss": 1.3559, | |
| "step": 6460 | |
| }, | |
| { | |
| "epoch": 0.6427657804214736, | |
| "grad_norm": 1.4877686500549316, | |
| "learning_rate": 8.80237018340069e-05, | |
| "loss": 1.3616, | |
| "step": 6470 | |
| }, | |
| { | |
| "epoch": 0.643759236032635, | |
| "grad_norm": 1.68259859085083, | |
| "learning_rate": 8.758900983397195e-05, | |
| "loss": 1.3526, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 0.6447526916437964, | |
| "grad_norm": 1.7352181673049927, | |
| "learning_rate": 8.71549507793721e-05, | |
| "loss": 1.3434, | |
| "step": 6490 | |
| }, | |
| { | |
| "epoch": 0.6457461472549579, | |
| "grad_norm": 1.5632843971252441, | |
| "learning_rate": 8.672152907224765e-05, | |
| "loss": 1.3502, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.6467396028661194, | |
| "grad_norm": 1.6187609434127808, | |
| "learning_rate": 8.628874910817528e-05, | |
| "loss": 1.3395, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 0.6477330584772809, | |
| "grad_norm": 1.6962559223175049, | |
| "learning_rate": 8.585661527622343e-05, | |
| "loss": 1.3767, | |
| "step": 6520 | |
| }, | |
| { | |
| "epoch": 0.6487265140884424, | |
| "grad_norm": 1.4766817092895508, | |
| "learning_rate": 8.542513195890758e-05, | |
| "loss": 1.3666, | |
| "step": 6530 | |
| }, | |
| { | |
| "epoch": 0.6497199696996039, | |
| "grad_norm": 1.4959608316421509, | |
| "learning_rate": 8.499430353214605e-05, | |
| "loss": 1.365, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 0.6507134253107654, | |
| "grad_norm": 1.6427513360977173, | |
| "learning_rate": 8.456413436521558e-05, | |
| "loss": 1.3654, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 0.6517068809219269, | |
| "grad_norm": 1.752276062965393, | |
| "learning_rate": 8.413462882070699e-05, | |
| "loss": 1.346, | |
| "step": 6560 | |
| }, | |
| { | |
| "epoch": 0.6527003365330882, | |
| "grad_norm": 1.6513816118240356, | |
| "learning_rate": 8.370579125448091e-05, | |
| "loss": 1.3439, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 0.6536937921442497, | |
| "grad_norm": 1.667937994003296, | |
| "learning_rate": 8.327762601562346e-05, | |
| "loss": 1.3755, | |
| "step": 6580 | |
| }, | |
| { | |
| "epoch": 0.6546872477554112, | |
| "grad_norm": 1.6565965414047241, | |
| "learning_rate": 8.285013744640266e-05, | |
| "loss": 1.3446, | |
| "step": 6590 | |
| }, | |
| { | |
| "epoch": 0.6556807033665727, | |
| "grad_norm": 1.620192050933838, | |
| "learning_rate": 8.24233298822236e-05, | |
| "loss": 1.3379, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.6566741589777342, | |
| "grad_norm": 1.5382760763168335, | |
| "learning_rate": 8.199720765158533e-05, | |
| "loss": 1.3639, | |
| "step": 6610 | |
| }, | |
| { | |
| "epoch": 0.6576676145888957, | |
| "grad_norm": 1.6407912969589233, | |
| "learning_rate": 8.15717750760362e-05, | |
| "loss": 1.338, | |
| "step": 6620 | |
| }, | |
| { | |
| "epoch": 0.6586610702000572, | |
| "grad_norm": 1.643744945526123, | |
| "learning_rate": 8.114703647013056e-05, | |
| "loss": 1.3393, | |
| "step": 6630 | |
| }, | |
| { | |
| "epoch": 0.6596545258112186, | |
| "grad_norm": 1.5156023502349854, | |
| "learning_rate": 8.072299614138482e-05, | |
| "loss": 1.3444, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 0.6606479814223801, | |
| "grad_norm": 1.6711229085922241, | |
| "learning_rate": 8.029965839023373e-05, | |
| "loss": 1.3543, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 0.6616414370335415, | |
| "grad_norm": 1.7823394536972046, | |
| "learning_rate": 7.98770275099868e-05, | |
| "loss": 1.3182, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 0.662634892644703, | |
| "grad_norm": 1.6579515933990479, | |
| "learning_rate": 7.94551077867846e-05, | |
| "loss": 1.3443, | |
| "step": 6670 | |
| }, | |
| { | |
| "epoch": 0.6636283482558645, | |
| "grad_norm": 1.795652985572815, | |
| "learning_rate": 7.903390349955574e-05, | |
| "loss": 1.3584, | |
| "step": 6680 | |
| }, | |
| { | |
| "epoch": 0.664621803867026, | |
| "grad_norm": 1.673013687133789, | |
| "learning_rate": 7.861341891997308e-05, | |
| "loss": 1.3704, | |
| "step": 6690 | |
| }, | |
| { | |
| "epoch": 0.6656152594781874, | |
| "grad_norm": 1.5517417192459106, | |
| "learning_rate": 7.819365831241036e-05, | |
| "loss": 1.3445, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.6666087150893489, | |
| "grad_norm": 1.5845626592636108, | |
| "learning_rate": 7.77746259338993e-05, | |
| "loss": 1.3378, | |
| "step": 6710 | |
| }, | |
| { | |
| "epoch": 0.6676021707005104, | |
| "grad_norm": 1.668227195739746, | |
| "learning_rate": 7.73563260340862e-05, | |
| "loss": 1.3395, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 0.6685956263116719, | |
| "grad_norm": 1.673973798751831, | |
| "learning_rate": 7.693876285518873e-05, | |
| "loss": 1.3546, | |
| "step": 6730 | |
| }, | |
| { | |
| "epoch": 0.6695890819228333, | |
| "grad_norm": 1.6369454860687256, | |
| "learning_rate": 7.652194063195329e-05, | |
| "loss": 1.3632, | |
| "step": 6740 | |
| }, | |
| { | |
| "epoch": 0.6705825375339948, | |
| "grad_norm": 1.7028038501739502, | |
| "learning_rate": 7.61058635916116e-05, | |
| "loss": 1.3382, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 0.6715759931451563, | |
| "grad_norm": 1.5395437479019165, | |
| "learning_rate": 7.569053595383814e-05, | |
| "loss": 1.3287, | |
| "step": 6760 | |
| }, | |
| { | |
| "epoch": 0.6725694487563177, | |
| "grad_norm": 1.741809606552124, | |
| "learning_rate": 7.527596193070742e-05, | |
| "loss": 1.3451, | |
| "step": 6770 | |
| }, | |
| { | |
| "epoch": 0.6735629043674792, | |
| "grad_norm": 1.5549176931381226, | |
| "learning_rate": 7.486214572665084e-05, | |
| "loss": 1.3395, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 0.6745563599786407, | |
| "grad_norm": 1.643896222114563, | |
| "learning_rate": 7.444909153841455e-05, | |
| "loss": 1.3439, | |
| "step": 6790 | |
| }, | |
| { | |
| "epoch": 0.6755498155898022, | |
| "grad_norm": 1.6707322597503662, | |
| "learning_rate": 7.40368035550165e-05, | |
| "loss": 1.3296, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.6765432712009637, | |
| "grad_norm": 1.596704363822937, | |
| "learning_rate": 7.362528595770423e-05, | |
| "loss": 1.3559, | |
| "step": 6810 | |
| }, | |
| { | |
| "epoch": 0.6775367268121252, | |
| "grad_norm": 1.8280783891677856, | |
| "learning_rate": 7.321454291991239e-05, | |
| "loss": 1.3582, | |
| "step": 6820 | |
| }, | |
| { | |
| "epoch": 0.6785301824232866, | |
| "grad_norm": 1.705898404121399, | |
| "learning_rate": 7.28045786072202e-05, | |
| "loss": 1.3358, | |
| "step": 6830 | |
| }, | |
| { | |
| "epoch": 0.679523638034448, | |
| "grad_norm": 1.5536643266677856, | |
| "learning_rate": 7.239539717730956e-05, | |
| "loss": 1.349, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 0.6805170936456095, | |
| "grad_norm": 1.6302812099456787, | |
| "learning_rate": 7.198700277992268e-05, | |
| "loss": 1.343, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 0.681510549256771, | |
| "grad_norm": 1.6454846858978271, | |
| "learning_rate": 7.157939955681996e-05, | |
| "loss": 1.3553, | |
| "step": 6860 | |
| }, | |
| { | |
| "epoch": 0.6825040048679325, | |
| "grad_norm": 1.5823239088058472, | |
| "learning_rate": 7.117259164173799e-05, | |
| "loss": 1.349, | |
| "step": 6870 | |
| }, | |
| { | |
| "epoch": 0.683497460479094, | |
| "grad_norm": 1.6325229406356812, | |
| "learning_rate": 7.07665831603479e-05, | |
| "loss": 1.3413, | |
| "step": 6880 | |
| }, | |
| { | |
| "epoch": 0.6844909160902555, | |
| "grad_norm": 1.7508260011672974, | |
| "learning_rate": 7.036137823021311e-05, | |
| "loss": 1.3747, | |
| "step": 6890 | |
| }, | |
| { | |
| "epoch": 0.685484371701417, | |
| "grad_norm": 1.728896141052246, | |
| "learning_rate": 6.995698096074797e-05, | |
| "loss": 1.3376, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.6864778273125783, | |
| "grad_norm": 1.6735886335372925, | |
| "learning_rate": 6.95533954531756e-05, | |
| "loss": 1.3289, | |
| "step": 6910 | |
| }, | |
| { | |
| "epoch": 0.6874712829237398, | |
| "grad_norm": 1.684324026107788, | |
| "learning_rate": 6.91506258004868e-05, | |
| "loss": 1.3403, | |
| "step": 6920 | |
| }, | |
| { | |
| "epoch": 0.6884647385349013, | |
| "grad_norm": 1.7846599817276, | |
| "learning_rate": 6.874867608739833e-05, | |
| "loss": 1.3457, | |
| "step": 6930 | |
| }, | |
| { | |
| "epoch": 0.6894581941460628, | |
| "grad_norm": 1.5779770612716675, | |
| "learning_rate": 6.834755039031133e-05, | |
| "loss": 1.3641, | |
| "step": 6940 | |
| }, | |
| { | |
| "epoch": 0.6904516497572243, | |
| "grad_norm": 1.5460431575775146, | |
| "learning_rate": 6.794725277727016e-05, | |
| "loss": 1.3453, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 0.6914451053683858, | |
| "grad_norm": 1.5156097412109375, | |
| "learning_rate": 6.75477873079212e-05, | |
| "loss": 1.3467, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 0.6924385609795473, | |
| "grad_norm": 1.4957499504089355, | |
| "learning_rate": 6.714915803347151e-05, | |
| "loss": 1.3346, | |
| "step": 6970 | |
| }, | |
| { | |
| "epoch": 0.6934320165907087, | |
| "grad_norm": 1.781427025794983, | |
| "learning_rate": 6.675136899664792e-05, | |
| "loss": 1.3396, | |
| "step": 6980 | |
| }, | |
| { | |
| "epoch": 0.6944254722018702, | |
| "grad_norm": 1.6415228843688965, | |
| "learning_rate": 6.635442423165571e-05, | |
| "loss": 1.3526, | |
| "step": 6990 | |
| }, | |
| { | |
| "epoch": 0.6954189278130316, | |
| "grad_norm": 1.6858388185501099, | |
| "learning_rate": 6.595832776413823e-05, | |
| "loss": 1.3601, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.6964123834241931, | |
| "grad_norm": 1.5956919193267822, | |
| "learning_rate": 6.55630836111354e-05, | |
| "loss": 1.3385, | |
| "step": 7010 | |
| }, | |
| { | |
| "epoch": 0.6974058390353546, | |
| "grad_norm": 1.7341642379760742, | |
| "learning_rate": 6.516869578104371e-05, | |
| "loss": 1.3331, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 0.6983992946465161, | |
| "grad_norm": 1.7775201797485352, | |
| "learning_rate": 6.477516827357486e-05, | |
| "loss": 1.334, | |
| "step": 7030 | |
| }, | |
| { | |
| "epoch": 0.6993927502576776, | |
| "grad_norm": 1.6503099203109741, | |
| "learning_rate": 6.438250507971574e-05, | |
| "loss": 1.3484, | |
| "step": 7040 | |
| }, | |
| { | |
| "epoch": 0.7002868603077229, | |
| "eval_loss": 1.320923089981079, | |
| "eval_runtime": 37.1653, | |
| "eval_samples_per_second": 1345.341, | |
| "eval_steps_per_second": 168.168, | |
| "step": 7049 | |
| }, | |
| { | |
| "epoch": 0.700386205868839, | |
| "grad_norm": 1.5613566637039185, | |
| "learning_rate": 6.399071018168768e-05, | |
| "loss": 1.3527, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 0.7013796614800005, | |
| "grad_norm": 1.6468534469604492, | |
| "learning_rate": 6.359978755290616e-05, | |
| "loss": 1.321, | |
| "step": 7060 | |
| }, | |
| { | |
| "epoch": 0.702373117091162, | |
| "grad_norm": 1.666261076927185, | |
| "learning_rate": 6.320974115794042e-05, | |
| "loss": 1.3341, | |
| "step": 7070 | |
| }, | |
| { | |
| "epoch": 0.7033665727023234, | |
| "grad_norm": 1.6668295860290527, | |
| "learning_rate": 6.282057495247326e-05, | |
| "loss": 1.3394, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 0.7043600283134849, | |
| "grad_norm": 1.7623732089996338, | |
| "learning_rate": 6.24322928832611e-05, | |
| "loss": 1.3441, | |
| "step": 7090 | |
| }, | |
| { | |
| "epoch": 0.7053534839246464, | |
| "grad_norm": 1.74466073513031, | |
| "learning_rate": 6.204489888809386e-05, | |
| "loss": 1.3414, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.7063469395358078, | |
| "grad_norm": 1.6080060005187988, | |
| "learning_rate": 6.16583968957548e-05, | |
| "loss": 1.3505, | |
| "step": 7110 | |
| }, | |
| { | |
| "epoch": 0.7073403951469693, | |
| "grad_norm": 1.5881328582763672, | |
| "learning_rate": 6.127279082598104e-05, | |
| "loss": 1.352, | |
| "step": 7120 | |
| }, | |
| { | |
| "epoch": 0.7083338507581308, | |
| "grad_norm": 1.6172865629196167, | |
| "learning_rate": 6.088808458942365e-05, | |
| "loss": 1.3501, | |
| "step": 7130 | |
| }, | |
| { | |
| "epoch": 0.7093273063692923, | |
| "grad_norm": 1.6604722738265991, | |
| "learning_rate": 6.050428208760781e-05, | |
| "loss": 1.3442, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 0.7103207619804538, | |
| "grad_norm": 1.7000654935836792, | |
| "learning_rate": 6.012138721289361e-05, | |
| "loss": 1.3384, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 0.7113142175916153, | |
| "grad_norm": 1.9108226299285889, | |
| "learning_rate": 5.97394038484362e-05, | |
| "loss": 1.3495, | |
| "step": 7160 | |
| }, | |
| { | |
| "epoch": 0.7123076732027767, | |
| "grad_norm": 1.5853825807571411, | |
| "learning_rate": 5.93583358681467e-05, | |
| "loss": 1.3293, | |
| "step": 7170 | |
| }, | |
| { | |
| "epoch": 0.7133011288139381, | |
| "grad_norm": 1.4958702325820923, | |
| "learning_rate": 5.897818713665282e-05, | |
| "loss": 1.3194, | |
| "step": 7180 | |
| }, | |
| { | |
| "epoch": 0.7142945844250996, | |
| "grad_norm": 1.7621620893478394, | |
| "learning_rate": 5.8598961509259504e-05, | |
| "loss": 1.3467, | |
| "step": 7190 | |
| }, | |
| { | |
| "epoch": 0.7152880400362611, | |
| "grad_norm": 1.687887191772461, | |
| "learning_rate": 5.8220662831910135e-05, | |
| "loss": 1.3522, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.7162814956474226, | |
| "grad_norm": 1.666365146636963, | |
| "learning_rate": 5.784329494114721e-05, | |
| "loss": 1.3419, | |
| "step": 7210 | |
| }, | |
| { | |
| "epoch": 0.7172749512585841, | |
| "grad_norm": 2.2286713123321533, | |
| "learning_rate": 5.746686166407369e-05, | |
| "loss": 1.3732, | |
| "step": 7220 | |
| }, | |
| { | |
| "epoch": 0.7182684068697456, | |
| "grad_norm": 1.7449884414672852, | |
| "learning_rate": 5.709136681831413e-05, | |
| "loss": 1.3444, | |
| "step": 7230 | |
| }, | |
| { | |
| "epoch": 0.7192618624809071, | |
| "grad_norm": 1.669190764427185, | |
| "learning_rate": 5.671681421197576e-05, | |
| "loss": 1.3512, | |
| "step": 7240 | |
| }, | |
| { | |
| "epoch": 0.7202553180920686, | |
| "grad_norm": 1.561073899269104, | |
| "learning_rate": 5.634320764361013e-05, | |
| "loss": 1.3334, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 0.7212487737032299, | |
| "grad_norm": 1.5225821733474731, | |
| "learning_rate": 5.5970550902174486e-05, | |
| "loss": 1.3395, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 0.7222422293143914, | |
| "grad_norm": 1.5557289123535156, | |
| "learning_rate": 5.559884776699341e-05, | |
| "loss": 1.3352, | |
| "step": 7270 | |
| }, | |
| { | |
| "epoch": 0.7232356849255529, | |
| "grad_norm": 1.691833257675171, | |
| "learning_rate": 5.522810200772013e-05, | |
| "loss": 1.3422, | |
| "step": 7280 | |
| }, | |
| { | |
| "epoch": 0.7242291405367144, | |
| "grad_norm": 1.6747978925704956, | |
| "learning_rate": 5.485831738429888e-05, | |
| "loss": 1.3483, | |
| "step": 7290 | |
| }, | |
| { | |
| "epoch": 0.7252225961478759, | |
| "grad_norm": 1.454331636428833, | |
| "learning_rate": 5.448949764692632e-05, | |
| "loss": 1.3432, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.7262160517590374, | |
| "grad_norm": 1.5704985857009888, | |
| "learning_rate": 5.412164653601378e-05, | |
| "loss": 1.35, | |
| "step": 7310 | |
| }, | |
| { | |
| "epoch": 0.7272095073701988, | |
| "grad_norm": 1.5449594259262085, | |
| "learning_rate": 5.375476778214899e-05, | |
| "loss": 1.3417, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 0.7282029629813603, | |
| "grad_norm": 1.7106105089187622, | |
| "learning_rate": 5.338886510605862e-05, | |
| "loss": 1.3389, | |
| "step": 7330 | |
| }, | |
| { | |
| "epoch": 0.7291964185925217, | |
| "grad_norm": 1.6524827480316162, | |
| "learning_rate": 5.30239422185704e-05, | |
| "loss": 1.3602, | |
| "step": 7340 | |
| }, | |
| { | |
| "epoch": 0.7301898742036832, | |
| "grad_norm": 1.7382267713546753, | |
| "learning_rate": 5.266000282057532e-05, | |
| "loss": 1.3389, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 0.7311833298148447, | |
| "grad_norm": 1.71548593044281, | |
| "learning_rate": 5.229705060299026e-05, | |
| "loss": 1.3417, | |
| "step": 7360 | |
| }, | |
| { | |
| "epoch": 0.7321767854260062, | |
| "grad_norm": 1.7593806982040405, | |
| "learning_rate": 5.193508924672067e-05, | |
| "loss": 1.3242, | |
| "step": 7370 | |
| }, | |
| { | |
| "epoch": 0.7331702410371677, | |
| "grad_norm": 1.7297049760818481, | |
| "learning_rate": 5.157412242262303e-05, | |
| "loss": 1.334, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 0.7341636966483291, | |
| "grad_norm": 1.685149908065796, | |
| "learning_rate": 5.1214153791467755e-05, | |
| "loss": 1.3399, | |
| "step": 7390 | |
| }, | |
| { | |
| "epoch": 0.7351571522594906, | |
| "grad_norm": 1.7318744659423828, | |
| "learning_rate": 5.085518700390187e-05, | |
| "loss": 1.3388, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 0.7361506078706521, | |
| "grad_norm": 1.7145448923110962, | |
| "learning_rate": 5.0497225700412334e-05, | |
| "loss": 1.3301, | |
| "step": 7410 | |
| }, | |
| { | |
| "epoch": 0.7371440634818136, | |
| "grad_norm": 1.7601321935653687, | |
| "learning_rate": 5.014027351128872e-05, | |
| "loss": 1.3395, | |
| "step": 7420 | |
| }, | |
| { | |
| "epoch": 0.738137519092975, | |
| "grad_norm": 1.7134897708892822, | |
| "learning_rate": 4.978433405658679e-05, | |
| "loss": 1.3223, | |
| "step": 7430 | |
| }, | |
| { | |
| "epoch": 0.7391309747041365, | |
| "grad_norm": 1.6700917482376099, | |
| "learning_rate": 4.9429410946091416e-05, | |
| "loss": 1.3352, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 0.740124430315298, | |
| "grad_norm": 1.7064131498336792, | |
| "learning_rate": 4.9075507779280246e-05, | |
| "loss": 1.3651, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 0.7411178859264594, | |
| "grad_norm": 1.5617907047271729, | |
| "learning_rate": 4.872262814528707e-05, | |
| "loss": 1.3385, | |
| "step": 7460 | |
| }, | |
| { | |
| "epoch": 0.7421113415376209, | |
| "grad_norm": 1.5047627687454224, | |
| "learning_rate": 4.8370775622865535e-05, | |
| "loss": 1.3487, | |
| "step": 7470 | |
| }, | |
| { | |
| "epoch": 0.7431047971487824, | |
| "grad_norm": 1.6555471420288086, | |
| "learning_rate": 4.801995378035259e-05, | |
| "loss": 1.3363, | |
| "step": 7480 | |
| }, | |
| { | |
| "epoch": 0.7440982527599439, | |
| "grad_norm": 1.5336803197860718, | |
| "learning_rate": 4.767016617563254e-05, | |
| "loss": 1.3524, | |
| "step": 7490 | |
| }, | |
| { | |
| "epoch": 0.7450917083711054, | |
| "grad_norm": 1.748270034790039, | |
| "learning_rate": 4.732141635610098e-05, | |
| "loss": 1.343, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.7460851639822668, | |
| "grad_norm": 1.7190279960632324, | |
| "learning_rate": 4.697370785862863e-05, | |
| "loss": 1.332, | |
| "step": 7510 | |
| }, | |
| { | |
| "epoch": 0.7470786195934283, | |
| "grad_norm": 1.575189232826233, | |
| "learning_rate": 4.6627044209525674e-05, | |
| "loss": 1.3406, | |
| "step": 7520 | |
| }, | |
| { | |
| "epoch": 0.7480720752045897, | |
| "grad_norm": 1.5859638452529907, | |
| "learning_rate": 4.628142892450575e-05, | |
| "loss": 1.3272, | |
| "step": 7530 | |
| }, | |
| { | |
| "epoch": 0.7490655308157512, | |
| "grad_norm": 1.6974937915802002, | |
| "learning_rate": 4.593686550865059e-05, | |
| "loss": 1.3389, | |
| "step": 7540 | |
| }, | |
| { | |
| "epoch": 0.7500589864269127, | |
| "grad_norm": 1.7742446660995483, | |
| "learning_rate": 4.5593357456374175e-05, | |
| "loss": 1.3612, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 0.7510524420380742, | |
| "grad_norm": 1.5838828086853027, | |
| "learning_rate": 4.5250908251387594e-05, | |
| "loss": 1.3328, | |
| "step": 7560 | |
| }, | |
| { | |
| "epoch": 0.7520458976492357, | |
| "grad_norm": 1.6167585849761963, | |
| "learning_rate": 4.490952136666338e-05, | |
| "loss": 1.3397, | |
| "step": 7570 | |
| }, | |
| { | |
| "epoch": 0.7530393532603972, | |
| "grad_norm": 1.8095159530639648, | |
| "learning_rate": 4.456920026440064e-05, | |
| "loss": 1.3389, | |
| "step": 7580 | |
| }, | |
| { | |
| "epoch": 0.7540328088715587, | |
| "grad_norm": 1.7143555879592896, | |
| "learning_rate": 4.422994839598969e-05, | |
| "loss": 1.3289, | |
| "step": 7590 | |
| }, | |
| { | |
| "epoch": 0.75502626448272, | |
| "grad_norm": 1.6431888341903687, | |
| "learning_rate": 4.389176920197723e-05, | |
| "loss": 1.3248, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 0.7560197200938815, | |
| "grad_norm": 1.726711392402649, | |
| "learning_rate": 4.355466611203119e-05, | |
| "loss": 1.3261, | |
| "step": 7610 | |
| }, | |
| { | |
| "epoch": 0.757013175705043, | |
| "grad_norm": 1.77463960647583, | |
| "learning_rate": 4.32186425449062e-05, | |
| "loss": 1.3323, | |
| "step": 7620 | |
| }, | |
| { | |
| "epoch": 0.7580066313162045, | |
| "grad_norm": 1.5855859518051147, | |
| "learning_rate": 4.288370190840887e-05, | |
| "loss": 1.3291, | |
| "step": 7630 | |
| }, | |
| { | |
| "epoch": 0.759000086927366, | |
| "grad_norm": 1.5605647563934326, | |
| "learning_rate": 4.2549847599363196e-05, | |
| "loss": 1.3345, | |
| "step": 7640 | |
| }, | |
| { | |
| "epoch": 0.7599935425385275, | |
| "grad_norm": 1.7549469470977783, | |
| "learning_rate": 4.221708300357598e-05, | |
| "loss": 1.3214, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 0.760986998149689, | |
| "grad_norm": 1.505378246307373, | |
| "learning_rate": 4.1885411495802794e-05, | |
| "loss": 1.3426, | |
| "step": 7660 | |
| }, | |
| { | |
| "epoch": 0.7619804537608504, | |
| "grad_norm": 1.701838493347168, | |
| "learning_rate": 4.155483643971349e-05, | |
| "loss": 1.3436, | |
| "step": 7670 | |
| }, | |
| { | |
| "epoch": 0.7629739093720119, | |
| "grad_norm": 1.7320512533187866, | |
| "learning_rate": 4.122536118785824e-05, | |
| "loss": 1.3368, | |
| "step": 7680 | |
| }, | |
| { | |
| "epoch": 0.7639673649831733, | |
| "grad_norm": 1.61899733543396, | |
| "learning_rate": 4.089698908163329e-05, | |
| "loss": 1.3625, | |
| "step": 7690 | |
| }, | |
| { | |
| "epoch": 0.7649608205943348, | |
| "grad_norm": 1.6827579736709595, | |
| "learning_rate": 4.056972345124746e-05, | |
| "loss": 1.3579, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 0.7659542762054963, | |
| "grad_norm": 1.4855307340621948, | |
| "learning_rate": 4.0243567615688116e-05, | |
| "loss": 1.3431, | |
| "step": 7710 | |
| }, | |
| { | |
| "epoch": 0.7669477318166578, | |
| "grad_norm": 1.6811317205429077, | |
| "learning_rate": 3.9918524882687566e-05, | |
| "loss": 1.3303, | |
| "step": 7720 | |
| }, | |
| { | |
| "epoch": 0.7679411874278192, | |
| "grad_norm": 1.7236701250076294, | |
| "learning_rate": 3.959459854868942e-05, | |
| "loss": 1.3465, | |
| "step": 7730 | |
| }, | |
| { | |
| "epoch": 0.7689346430389807, | |
| "grad_norm": 1.6243964433670044, | |
| "learning_rate": 3.9271791898815366e-05, | |
| "loss": 1.3538, | |
| "step": 7740 | |
| }, | |
| { | |
| "epoch": 0.7699280986501422, | |
| "grad_norm": 1.7215466499328613, | |
| "learning_rate": 3.895010820683172e-05, | |
| "loss": 1.324, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 0.7709215542613037, | |
| "grad_norm": 1.7503958940505981, | |
| "learning_rate": 3.8629550735116114e-05, | |
| "loss": 1.3493, | |
| "step": 7760 | |
| }, | |
| { | |
| "epoch": 0.7719150098724651, | |
| "grad_norm": 1.6961768865585327, | |
| "learning_rate": 3.8310122734624754e-05, | |
| "loss": 1.3485, | |
| "step": 7770 | |
| }, | |
| { | |
| "epoch": 0.7729084654836266, | |
| "grad_norm": 1.6427160501480103, | |
| "learning_rate": 3.7991827444859e-05, | |
| "loss": 1.3468, | |
| "step": 7780 | |
| }, | |
| { | |
| "epoch": 0.7739019210947881, | |
| "grad_norm": 1.7004997730255127, | |
| "learning_rate": 3.767466809383293e-05, | |
| "loss": 1.3571, | |
| "step": 7790 | |
| }, | |
| { | |
| "epoch": 0.7748953767059495, | |
| "grad_norm": 1.6564383506774902, | |
| "learning_rate": 3.7358647898040364e-05, | |
| "loss": 1.349, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 0.775888832317111, | |
| "grad_norm": 1.6278494596481323, | |
| "learning_rate": 3.7043770062422246e-05, | |
| "loss": 1.3374, | |
| "step": 7810 | |
| }, | |
| { | |
| "epoch": 0.7768822879282725, | |
| "grad_norm": 1.7883501052856445, | |
| "learning_rate": 3.6730037780334277e-05, | |
| "loss": 1.3381, | |
| "step": 7820 | |
| }, | |
| { | |
| "epoch": 0.777875743539434, | |
| "grad_norm": 1.643226146697998, | |
| "learning_rate": 3.6417454233514335e-05, | |
| "loss": 1.3372, | |
| "step": 7830 | |
| }, | |
| { | |
| "epoch": 0.7788691991505955, | |
| "grad_norm": 1.587995171546936, | |
| "learning_rate": 3.610602259205044e-05, | |
| "loss": 1.334, | |
| "step": 7840 | |
| }, | |
| { | |
| "epoch": 0.779862654761757, | |
| "grad_norm": 1.6008771657943726, | |
| "learning_rate": 3.5795746014348466e-05, | |
| "loss": 1.3272, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 0.7808561103729184, | |
| "grad_norm": 1.5943472385406494, | |
| "learning_rate": 3.54866276471e-05, | |
| "loss": 1.3391, | |
| "step": 7860 | |
| }, | |
| { | |
| "epoch": 0.7818495659840798, | |
| "grad_norm": 1.8545950651168823, | |
| "learning_rate": 3.517867062525072e-05, | |
| "loss": 1.3206, | |
| "step": 7870 | |
| }, | |
| { | |
| "epoch": 0.7828430215952413, | |
| "grad_norm": 1.729323387145996, | |
| "learning_rate": 3.487187807196838e-05, | |
| "loss": 1.3429, | |
| "step": 7880 | |
| }, | |
| { | |
| "epoch": 0.7838364772064028, | |
| "grad_norm": 1.581868052482605, | |
| "learning_rate": 3.456625309861116e-05, | |
| "loss": 1.3283, | |
| "step": 7890 | |
| }, | |
| { | |
| "epoch": 0.7848299328175643, | |
| "grad_norm": 1.7917063236236572, | |
| "learning_rate": 3.426179880469612e-05, | |
| "loss": 1.3327, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 0.7858233884287258, | |
| "grad_norm": 1.7997517585754395, | |
| "learning_rate": 3.3958518277867864e-05, | |
| "loss": 1.3484, | |
| "step": 7910 | |
| }, | |
| { | |
| "epoch": 0.7868168440398873, | |
| "grad_norm": 1.577675223350525, | |
| "learning_rate": 3.365641459386715e-05, | |
| "loss": 1.338, | |
| "step": 7920 | |
| }, | |
| { | |
| "epoch": 0.7878102996510488, | |
| "grad_norm": 1.6609553098678589, | |
| "learning_rate": 3.3355490816499716e-05, | |
| "loss": 1.3218, | |
| "step": 7930 | |
| }, | |
| { | |
| "epoch": 0.7888037552622101, | |
| "grad_norm": 1.7371225357055664, | |
| "learning_rate": 3.305574999760507e-05, | |
| "loss": 1.338, | |
| "step": 7940 | |
| }, | |
| { | |
| "epoch": 0.7897972108733716, | |
| "grad_norm": 1.6992961168289185, | |
| "learning_rate": 3.275719517702587e-05, | |
| "loss": 1.3378, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 0.7907906664845331, | |
| "grad_norm": 1.869924783706665, | |
| "learning_rate": 3.2459829382576664e-05, | |
| "loss": 1.3277, | |
| "step": 7960 | |
| }, | |
| { | |
| "epoch": 0.7917841220956946, | |
| "grad_norm": 1.9163962602615356, | |
| "learning_rate": 3.21636556300136e-05, | |
| "loss": 1.3261, | |
| "step": 7970 | |
| }, | |
| { | |
| "epoch": 0.7927775777068561, | |
| "grad_norm": 1.752663493156433, | |
| "learning_rate": 3.186867692300349e-05, | |
| "loss": 1.3421, | |
| "step": 7980 | |
| }, | |
| { | |
| "epoch": 0.7937710333180176, | |
| "grad_norm": 1.6545498371124268, | |
| "learning_rate": 3.157489625309363e-05, | |
| "loss": 1.3314, | |
| "step": 7990 | |
| }, | |
| { | |
| "epoch": 0.7947644889291791, | |
| "grad_norm": 1.5815753936767578, | |
| "learning_rate": 3.12823165996813e-05, | |
| "loss": 1.344, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.7957579445403405, | |
| "grad_norm": 1.5998033285140991, | |
| "learning_rate": 3.099094092998359e-05, | |
| "loss": 1.3211, | |
| "step": 8010 | |
| }, | |
| { | |
| "epoch": 0.796751400151502, | |
| "grad_norm": 1.6735488176345825, | |
| "learning_rate": 3.070077219900729e-05, | |
| "loss": 1.3383, | |
| "step": 8020 | |
| }, | |
| { | |
| "epoch": 0.7977448557626634, | |
| "grad_norm": 1.704650640487671, | |
| "learning_rate": 3.041181334951887e-05, | |
| "loss": 1.341, | |
| "step": 8030 | |
| }, | |
| { | |
| "epoch": 0.7987383113738249, | |
| "grad_norm": 1.551392912864685, | |
| "learning_rate": 3.012406731201485e-05, | |
| "loss": 1.319, | |
| "step": 8040 | |
| }, | |
| { | |
| "epoch": 0.7997317669849864, | |
| "grad_norm": 1.6002696752548218, | |
| "learning_rate": 2.9837537004691852e-05, | |
| "loss": 1.3297, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 0.8003278403516833, | |
| "eval_loss": 1.3116158246994019, | |
| "eval_runtime": 37.1076, | |
| "eval_samples_per_second": 1347.432, | |
| "eval_steps_per_second": 168.429, | |
| "step": 8056 | |
| }, | |
| { | |
| "epoch": 0.8007252225961479, | |
| "grad_norm": 1.6566636562347412, | |
| "learning_rate": 2.9552225333417047e-05, | |
| "loss": 1.3465, | |
| "step": 8060 | |
| }, | |
| { | |
| "epoch": 0.8017186782073094, | |
| "grad_norm": 1.7833751440048218, | |
| "learning_rate": 2.9268135191698776e-05, | |
| "loss": 1.3279, | |
| "step": 8070 | |
| }, | |
| { | |
| "epoch": 0.8027121338184708, | |
| "grad_norm": 1.8028894662857056, | |
| "learning_rate": 2.8985269460657175e-05, | |
| "loss": 1.3445, | |
| "step": 8080 | |
| }, | |
| { | |
| "epoch": 0.8037055894296323, | |
| "grad_norm": 1.7755157947540283, | |
| "learning_rate": 2.870363100899491e-05, | |
| "loss": 1.33, | |
| "step": 8090 | |
| }, | |
| { | |
| "epoch": 0.8046990450407938, | |
| "grad_norm": 1.7127573490142822, | |
| "learning_rate": 2.8423222692968052e-05, | |
| "loss": 1.3334, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 0.8056925006519552, | |
| "grad_norm": 1.7856546640396118, | |
| "learning_rate": 2.814404735635721e-05, | |
| "loss": 1.3495, | |
| "step": 8110 | |
| }, | |
| { | |
| "epoch": 0.8066859562631167, | |
| "grad_norm": 1.7438029050827026, | |
| "learning_rate": 2.7866107830438626e-05, | |
| "loss": 1.3504, | |
| "step": 8120 | |
| }, | |
| { | |
| "epoch": 0.8076794118742782, | |
| "grad_norm": 1.625188946723938, | |
| "learning_rate": 2.7589406933955597e-05, | |
| "loss": 1.3107, | |
| "step": 8130 | |
| }, | |
| { | |
| "epoch": 0.8086728674854397, | |
| "grad_norm": 1.7805283069610596, | |
| "learning_rate": 2.731394747308956e-05, | |
| "loss": 1.3669, | |
| "step": 8140 | |
| }, | |
| { | |
| "epoch": 0.8096663230966011, | |
| "grad_norm": 1.8052105903625488, | |
| "learning_rate": 2.7039732241432016e-05, | |
| "loss": 1.3415, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 0.8106597787077626, | |
| "grad_norm": 1.86104416847229, | |
| "learning_rate": 2.6766764019956006e-05, | |
| "loss": 1.3247, | |
| "step": 8160 | |
| }, | |
| { | |
| "epoch": 0.8116532343189241, | |
| "grad_norm": 1.8850599527359009, | |
| "learning_rate": 2.6495045576987828e-05, | |
| "loss": 1.3102, | |
| "step": 8170 | |
| }, | |
| { | |
| "epoch": 0.8126466899300856, | |
| "grad_norm": 1.777363657951355, | |
| "learning_rate": 2.622457966817921e-05, | |
| "loss": 1.3369, | |
| "step": 8180 | |
| }, | |
| { | |
| "epoch": 0.8136401455412471, | |
| "grad_norm": 1.8582979440689087, | |
| "learning_rate": 2.5955369036479047e-05, | |
| "loss": 1.3337, | |
| "step": 8190 | |
| }, | |
| { | |
| "epoch": 0.8146336011524085, | |
| "grad_norm": 1.7999472618103027, | |
| "learning_rate": 2.568741641210591e-05, | |
| "loss": 1.3277, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 0.81562705676357, | |
| "grad_norm": 1.7589097023010254, | |
| "learning_rate": 2.5420724512520142e-05, | |
| "loss": 1.3325, | |
| "step": 8210 | |
| }, | |
| { | |
| "epoch": 0.8166205123747314, | |
| "grad_norm": 1.5821574926376343, | |
| "learning_rate": 2.5155296042396305e-05, | |
| "loss": 1.3279, | |
| "step": 8220 | |
| }, | |
| { | |
| "epoch": 0.8176139679858929, | |
| "grad_norm": 1.8107210397720337, | |
| "learning_rate": 2.489113369359592e-05, | |
| "loss": 1.3369, | |
| "step": 8230 | |
| }, | |
| { | |
| "epoch": 0.8186074235970544, | |
| "grad_norm": 1.8852452039718628, | |
| "learning_rate": 2.462824014513991e-05, | |
| "loss": 1.3347, | |
| "step": 8240 | |
| }, | |
| { | |
| "epoch": 0.8196008792082159, | |
| "grad_norm": 1.6794813871383667, | |
| "learning_rate": 2.4366618063181652e-05, | |
| "loss": 1.3332, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 0.8205943348193774, | |
| "grad_norm": 1.729719877243042, | |
| "learning_rate": 2.4106270100979898e-05, | |
| "loss": 1.3264, | |
| "step": 8260 | |
| }, | |
| { | |
| "epoch": 0.8215877904305389, | |
| "grad_norm": 1.7096331119537354, | |
| "learning_rate": 2.3847198898871688e-05, | |
| "loss": 1.3375, | |
| "step": 8270 | |
| }, | |
| { | |
| "epoch": 0.8225812460417004, | |
| "grad_norm": 1.619379997253418, | |
| "learning_rate": 2.3589407084245854e-05, | |
| "loss": 1.3403, | |
| "step": 8280 | |
| }, | |
| { | |
| "epoch": 0.8235747016528617, | |
| "grad_norm": 1.5931566953659058, | |
| "learning_rate": 2.3332897271516177e-05, | |
| "loss": 1.3237, | |
| "step": 8290 | |
| }, | |
| { | |
| "epoch": 0.8245681572640232, | |
| "grad_norm": 1.6369714736938477, | |
| "learning_rate": 2.3077672062094892e-05, | |
| "loss": 1.3384, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 0.8255616128751847, | |
| "grad_norm": 1.5900071859359741, | |
| "learning_rate": 2.2823734044366325e-05, | |
| "loss": 1.3274, | |
| "step": 8310 | |
| }, | |
| { | |
| "epoch": 0.8265550684863462, | |
| "grad_norm": 1.8101078271865845, | |
| "learning_rate": 2.257108579366074e-05, | |
| "loss": 1.3278, | |
| "step": 8320 | |
| }, | |
| { | |
| "epoch": 0.8275485240975077, | |
| "grad_norm": 1.7749757766723633, | |
| "learning_rate": 2.2319729872228086e-05, | |
| "loss": 1.314, | |
| "step": 8330 | |
| }, | |
| { | |
| "epoch": 0.8285419797086692, | |
| "grad_norm": 1.490689754486084, | |
| "learning_rate": 2.2069668829212122e-05, | |
| "loss": 1.3425, | |
| "step": 8340 | |
| }, | |
| { | |
| "epoch": 0.8295354353198306, | |
| "grad_norm": 1.6246777772903442, | |
| "learning_rate": 2.1820905200624382e-05, | |
| "loss": 1.3443, | |
| "step": 8350 | |
| }, | |
| { | |
| "epoch": 0.8305288909309921, | |
| "grad_norm": 1.537866234779358, | |
| "learning_rate": 2.1573441509318727e-05, | |
| "loss": 1.3265, | |
| "step": 8360 | |
| }, | |
| { | |
| "epoch": 0.8315223465421535, | |
| "grad_norm": 1.6559864282608032, | |
| "learning_rate": 2.1327280264965462e-05, | |
| "loss": 1.3278, | |
| "step": 8370 | |
| }, | |
| { | |
| "epoch": 0.832515802153315, | |
| "grad_norm": 1.7555274963378906, | |
| "learning_rate": 2.1082423964026208e-05, | |
| "loss": 1.3248, | |
| "step": 8380 | |
| }, | |
| { | |
| "epoch": 0.8335092577644765, | |
| "grad_norm": 1.8152648210525513, | |
| "learning_rate": 2.0838875089728236e-05, | |
| "loss": 1.3332, | |
| "step": 8390 | |
| }, | |
| { | |
| "epoch": 0.834502713375638, | |
| "grad_norm": 1.8337817192077637, | |
| "learning_rate": 2.059663611203956e-05, | |
| "loss": 1.3385, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 0.8354961689867995, | |
| "grad_norm": 1.6960264444351196, | |
| "learning_rate": 2.0355709487643807e-05, | |
| "loss": 1.3318, | |
| "step": 8410 | |
| }, | |
| { | |
| "epoch": 0.836489624597961, | |
| "grad_norm": 1.7089858055114746, | |
| "learning_rate": 2.0116097659915225e-05, | |
| "loss": 1.3167, | |
| "step": 8420 | |
| }, | |
| { | |
| "epoch": 0.8374830802091224, | |
| "grad_norm": 1.6503387689590454, | |
| "learning_rate": 1.987780305889397e-05, | |
| "loss": 1.3178, | |
| "step": 8430 | |
| }, | |
| { | |
| "epoch": 0.8384765358202839, | |
| "grad_norm": 1.733608603477478, | |
| "learning_rate": 1.9640828101261413e-05, | |
| "loss": 1.3356, | |
| "step": 8440 | |
| }, | |
| { | |
| "epoch": 0.8394699914314454, | |
| "grad_norm": 1.7308598756790161, | |
| "learning_rate": 1.9405175190315747e-05, | |
| "loss": 1.3446, | |
| "step": 8450 | |
| }, | |
| { | |
| "epoch": 0.8404634470426068, | |
| "grad_norm": 1.6812063455581665, | |
| "learning_rate": 1.917084671594754e-05, | |
| "loss": 1.3326, | |
| "step": 8460 | |
| }, | |
| { | |
| "epoch": 0.8414569026537683, | |
| "grad_norm": 1.8042922019958496, | |
| "learning_rate": 1.8937845054615373e-05, | |
| "loss": 1.347, | |
| "step": 8470 | |
| }, | |
| { | |
| "epoch": 0.8424503582649298, | |
| "grad_norm": 1.7376409769058228, | |
| "learning_rate": 1.870617256932199e-05, | |
| "loss": 1.346, | |
| "step": 8480 | |
| }, | |
| { | |
| "epoch": 0.8434438138760912, | |
| "grad_norm": 1.7143795490264893, | |
| "learning_rate": 1.8475831609590135e-05, | |
| "loss": 1.3367, | |
| "step": 8490 | |
| }, | |
| { | |
| "epoch": 0.8444372694872527, | |
| "grad_norm": 1.726882815361023, | |
| "learning_rate": 1.8246824511438868e-05, | |
| "loss": 1.3301, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.8454307250984142, | |
| "grad_norm": 1.9756132364273071, | |
| "learning_rate": 1.801915359735967e-05, | |
| "loss": 1.316, | |
| "step": 8510 | |
| }, | |
| { | |
| "epoch": 0.8464241807095757, | |
| "grad_norm": 1.5057711601257324, | |
| "learning_rate": 1.7792821176293097e-05, | |
| "loss": 1.3339, | |
| "step": 8520 | |
| }, | |
| { | |
| "epoch": 0.8474176363207372, | |
| "grad_norm": 1.8790594339370728, | |
| "learning_rate": 1.7567829543605282e-05, | |
| "loss": 1.337, | |
| "step": 8530 | |
| }, | |
| { | |
| "epoch": 0.8484110919318986, | |
| "grad_norm": 1.7121455669403076, | |
| "learning_rate": 1.734418098106468e-05, | |
| "loss": 1.3268, | |
| "step": 8540 | |
| }, | |
| { | |
| "epoch": 0.84940454754306, | |
| "grad_norm": 1.6791712045669556, | |
| "learning_rate": 1.712187775681883e-05, | |
| "loss": 1.3269, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 0.8503980031542215, | |
| "grad_norm": 1.6731219291687012, | |
| "learning_rate": 1.6900922125371498e-05, | |
| "loss": 1.3183, | |
| "step": 8560 | |
| }, | |
| { | |
| "epoch": 0.851391458765383, | |
| "grad_norm": 1.7038770914077759, | |
| "learning_rate": 1.6681316327559763e-05, | |
| "loss": 1.3427, | |
| "step": 8570 | |
| }, | |
| { | |
| "epoch": 0.8523849143765445, | |
| "grad_norm": 1.7888379096984863, | |
| "learning_rate": 1.646306259053116e-05, | |
| "loss": 1.3113, | |
| "step": 8580 | |
| }, | |
| { | |
| "epoch": 0.853378369987706, | |
| "grad_norm": 1.6916706562042236, | |
| "learning_rate": 1.624616312772133e-05, | |
| "loss": 1.3222, | |
| "step": 8590 | |
| }, | |
| { | |
| "epoch": 0.8543718255988675, | |
| "grad_norm": 1.8067281246185303, | |
| "learning_rate": 1.6030620138831336e-05, | |
| "loss": 1.3189, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 0.855365281210029, | |
| "grad_norm": 1.754963994026184, | |
| "learning_rate": 1.5816435809805557e-05, | |
| "loss": 1.3234, | |
| "step": 8610 | |
| }, | |
| { | |
| "epoch": 0.8563587368211905, | |
| "grad_norm": 1.9376850128173828, | |
| "learning_rate": 1.5603612312809373e-05, | |
| "loss": 1.3195, | |
| "step": 8620 | |
| }, | |
| { | |
| "epoch": 0.8573521924323518, | |
| "grad_norm": 1.8017528057098389, | |
| "learning_rate": 1.5392151806207254e-05, | |
| "loss": 1.3332, | |
| "step": 8630 | |
| }, | |
| { | |
| "epoch": 0.8583456480435133, | |
| "grad_norm": 1.8763980865478516, | |
| "learning_rate": 1.5182056434540729e-05, | |
| "loss": 1.3332, | |
| "step": 8640 | |
| }, | |
| { | |
| "epoch": 0.8593391036546748, | |
| "grad_norm": 1.6288292407989502, | |
| "learning_rate": 1.4973328328506734e-05, | |
| "loss": 1.3178, | |
| "step": 8650 | |
| }, | |
| { | |
| "epoch": 0.8603325592658363, | |
| "grad_norm": 1.559564232826233, | |
| "learning_rate": 1.4765969604935995e-05, | |
| "loss": 1.3494, | |
| "step": 8660 | |
| }, | |
| { | |
| "epoch": 0.8613260148769978, | |
| "grad_norm": 1.6878244876861572, | |
| "learning_rate": 1.4559982366771616e-05, | |
| "loss": 1.3336, | |
| "step": 8670 | |
| }, | |
| { | |
| "epoch": 0.8623194704881593, | |
| "grad_norm": 1.7200804948806763, | |
| "learning_rate": 1.4355368703047543e-05, | |
| "loss": 1.3326, | |
| "step": 8680 | |
| }, | |
| { | |
| "epoch": 0.8633129260993208, | |
| "grad_norm": 1.622283697128296, | |
| "learning_rate": 1.4152130688867647e-05, | |
| "loss": 1.3038, | |
| "step": 8690 | |
| }, | |
| { | |
| "epoch": 0.8643063817104822, | |
| "grad_norm": 1.654721975326538, | |
| "learning_rate": 1.3950270385384526e-05, | |
| "loss": 1.3546, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 0.8652998373216437, | |
| "grad_norm": 1.587808609008789, | |
| "learning_rate": 1.3749789839778574e-05, | |
| "loss": 1.3256, | |
| "step": 8710 | |
| }, | |
| { | |
| "epoch": 0.8662932929328051, | |
| "grad_norm": 1.7595751285552979, | |
| "learning_rate": 1.3550691085237292e-05, | |
| "loss": 1.3159, | |
| "step": 8720 | |
| }, | |
| { | |
| "epoch": 0.8672867485439666, | |
| "grad_norm": 1.6317850351333618, | |
| "learning_rate": 1.335297614093469e-05, | |
| "loss": 1.3145, | |
| "step": 8730 | |
| }, | |
| { | |
| "epoch": 0.8682802041551281, | |
| "grad_norm": 1.696241855621338, | |
| "learning_rate": 1.315664701201074e-05, | |
| "loss": 1.3222, | |
| "step": 8740 | |
| }, | |
| { | |
| "epoch": 0.8692736597662896, | |
| "grad_norm": 1.7124111652374268, | |
| "learning_rate": 1.2961705689551122e-05, | |
| "loss": 1.3414, | |
| "step": 8750 | |
| }, | |
| { | |
| "epoch": 0.870267115377451, | |
| "grad_norm": 1.6316888332366943, | |
| "learning_rate": 1.2768154150566856e-05, | |
| "loss": 1.3431, | |
| "step": 8760 | |
| }, | |
| { | |
| "epoch": 0.8712605709886125, | |
| "grad_norm": 1.7555912733078003, | |
| "learning_rate": 1.2575994357974523e-05, | |
| "loss": 1.3345, | |
| "step": 8770 | |
| }, | |
| { | |
| "epoch": 0.872254026599774, | |
| "grad_norm": 1.6559699773788452, | |
| "learning_rate": 1.2385228260576091e-05, | |
| "loss": 1.334, | |
| "step": 8780 | |
| }, | |
| { | |
| "epoch": 0.8732474822109355, | |
| "grad_norm": 1.5741063356399536, | |
| "learning_rate": 1.2195857793039337e-05, | |
| "loss": 1.3551, | |
| "step": 8790 | |
| }, | |
| { | |
| "epoch": 0.8742409378220969, | |
| "grad_norm": 1.6742480993270874, | |
| "learning_rate": 1.2007884875878143e-05, | |
| "loss": 1.3087, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 0.8752343934332584, | |
| "grad_norm": 1.6519863605499268, | |
| "learning_rate": 1.182131141543301e-05, | |
| "loss": 1.347, | |
| "step": 8810 | |
| }, | |
| { | |
| "epoch": 0.8762278490444199, | |
| "grad_norm": 1.6424379348754883, | |
| "learning_rate": 1.163613930385181e-05, | |
| "loss": 1.3213, | |
| "step": 8820 | |
| }, | |
| { | |
| "epoch": 0.8772213046555813, | |
| "grad_norm": 1.7244398593902588, | |
| "learning_rate": 1.1452370419070517e-05, | |
| "loss": 1.3231, | |
| "step": 8830 | |
| }, | |
| { | |
| "epoch": 0.8782147602667428, | |
| "grad_norm": 1.7350554466247559, | |
| "learning_rate": 1.1270006624794154e-05, | |
| "loss": 1.3216, | |
| "step": 8840 | |
| }, | |
| { | |
| "epoch": 0.8792082158779043, | |
| "grad_norm": 1.5938142538070679, | |
| "learning_rate": 1.1089049770477892e-05, | |
| "loss": 1.3444, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 0.8802016714890658, | |
| "grad_norm": 1.6266322135925293, | |
| "learning_rate": 1.0909501691308387e-05, | |
| "loss": 1.3448, | |
| "step": 8860 | |
| }, | |
| { | |
| "epoch": 0.8811951271002273, | |
| "grad_norm": 1.8131626844406128, | |
| "learning_rate": 1.07313642081851e-05, | |
| "loss": 1.3156, | |
| "step": 8870 | |
| }, | |
| { | |
| "epoch": 0.8821885827113888, | |
| "grad_norm": 1.7156671285629272, | |
| "learning_rate": 1.0554639127701808e-05, | |
| "loss": 1.3393, | |
| "step": 8880 | |
| }, | |
| { | |
| "epoch": 0.8831820383225502, | |
| "grad_norm": 1.6699461936950684, | |
| "learning_rate": 1.0379328242128294e-05, | |
| "loss": 1.3191, | |
| "step": 8890 | |
| }, | |
| { | |
| "epoch": 0.8841754939337116, | |
| "grad_norm": 1.714072585105896, | |
| "learning_rate": 1.020543332939221e-05, | |
| "loss": 1.337, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 0.8851689495448731, | |
| "grad_norm": 1.6565090417861938, | |
| "learning_rate": 1.003295615306105e-05, | |
| "loss": 1.3319, | |
| "step": 8910 | |
| }, | |
| { | |
| "epoch": 0.8861624051560346, | |
| "grad_norm": 1.6230782270431519, | |
| "learning_rate": 9.861898462324158e-06, | |
| "loss": 1.3176, | |
| "step": 8920 | |
| }, | |
| { | |
| "epoch": 0.8871558607671961, | |
| "grad_norm": 1.8689581155776978, | |
| "learning_rate": 9.692261991975076e-06, | |
| "loss": 1.3128, | |
| "step": 8930 | |
| }, | |
| { | |
| "epoch": 0.8881493163783576, | |
| "grad_norm": 1.6196343898773193, | |
| "learning_rate": 9.524048462394024e-06, | |
| "loss": 1.3363, | |
| "step": 8940 | |
| }, | |
| { | |
| "epoch": 0.8891427719895191, | |
| "grad_norm": 1.6034671068191528, | |
| "learning_rate": 9.357259579530285e-06, | |
| "loss": 1.3402, | |
| "step": 8950 | |
| }, | |
| { | |
| "epoch": 0.8901362276006806, | |
| "grad_norm": 1.5668889284133911, | |
| "learning_rate": 9.191897034885033e-06, | |
| "loss": 1.2998, | |
| "step": 8960 | |
| }, | |
| { | |
| "epoch": 0.8911296832118419, | |
| "grad_norm": 1.5805506706237793, | |
| "learning_rate": 9.027962505494108e-06, | |
| "loss": 1.3235, | |
| "step": 8970 | |
| }, | |
| { | |
| "epoch": 0.8921231388230034, | |
| "grad_norm": 1.6430960893630981, | |
| "learning_rate": 8.86545765391104e-06, | |
| "loss": 1.337, | |
| "step": 8980 | |
| }, | |
| { | |
| "epoch": 0.8931165944341649, | |
| "grad_norm": 1.6430763006210327, | |
| "learning_rate": 8.704384128190156e-06, | |
| "loss": 1.3187, | |
| "step": 8990 | |
| }, | |
| { | |
| "epoch": 0.8941100500453264, | |
| "grad_norm": 1.7595213651657104, | |
| "learning_rate": 8.544743561869977e-06, | |
| "loss": 1.3271, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.8951035056564879, | |
| "grad_norm": 1.8313692808151245, | |
| "learning_rate": 8.386537573956431e-06, | |
| "loss": 1.3288, | |
| "step": 9010 | |
| }, | |
| { | |
| "epoch": 0.8960969612676494, | |
| "grad_norm": 1.7659707069396973, | |
| "learning_rate": 8.229767768906687e-06, | |
| "loss": 1.3396, | |
| "step": 9020 | |
| }, | |
| { | |
| "epoch": 0.8970904168788109, | |
| "grad_norm": 1.656742811203003, | |
| "learning_rate": 8.074435736612712e-06, | |
| "loss": 1.3118, | |
| "step": 9030 | |
| }, | |
| { | |
| "epoch": 0.8980838724899723, | |
| "grad_norm": 1.819999098777771, | |
| "learning_rate": 7.920543052385236e-06, | |
| "loss": 1.3511, | |
| "step": 9040 | |
| }, | |
| { | |
| "epoch": 0.8990773281011338, | |
| "grad_norm": 1.8005197048187256, | |
| "learning_rate": 7.768091276937688e-06, | |
| "loss": 1.322, | |
| "step": 9050 | |
| }, | |
| { | |
| "epoch": 0.9000707837122952, | |
| "grad_norm": 1.7250587940216064, | |
| "learning_rate": 7.617081956370413e-06, | |
| "loss": 1.3038, | |
| "step": 9060 | |
| }, | |
| { | |
| "epoch": 0.9003688203956437, | |
| "eval_loss": 1.3055968284606934, | |
| "eval_runtime": 37.2504, | |
| "eval_samples_per_second": 1342.267, | |
| "eval_steps_per_second": 167.783, | |
| "step": 9063 | |
| }, | |
| { | |
| "epoch": 0.9010642393234567, | |
| "grad_norm": 1.7743979692459106, | |
| "learning_rate": 7.467516622155079e-06, | |
| "loss": 1.3363, | |
| "step": 9070 | |
| }, | |
| { | |
| "epoch": 0.9020576949346182, | |
| "grad_norm": 1.5015814304351807, | |
| "learning_rate": 7.319396791118998e-06, | |
| "loss": 1.3241, | |
| "step": 9080 | |
| }, | |
| { | |
| "epoch": 0.9030511505457797, | |
| "grad_norm": 1.6709809303283691, | |
| "learning_rate": 7.172723965429811e-06, | |
| "loss": 1.3205, | |
| "step": 9090 | |
| }, | |
| { | |
| "epoch": 0.9040446061569412, | |
| "grad_norm": 1.7442833185195923, | |
| "learning_rate": 7.027499632580264e-06, | |
| "loss": 1.3056, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 0.9050380617681026, | |
| "grad_norm": 1.7080721855163574, | |
| "learning_rate": 6.883725265373185e-06, | |
| "loss": 1.3273, | |
| "step": 9110 | |
| }, | |
| { | |
| "epoch": 0.9060315173792641, | |
| "grad_norm": 1.6414085626602173, | |
| "learning_rate": 6.74140232190632e-06, | |
| "loss": 1.3046, | |
| "step": 9120 | |
| }, | |
| { | |
| "epoch": 0.9070249729904256, | |
| "grad_norm": 1.7125777006149292, | |
| "learning_rate": 6.600532245557833e-06, | |
| "loss": 1.3221, | |
| "step": 9130 | |
| }, | |
| { | |
| "epoch": 0.908018428601587, | |
| "grad_norm": 1.7101080417633057, | |
| "learning_rate": 6.461116464971461e-06, | |
| "loss": 1.3276, | |
| "step": 9140 | |
| }, | |
| { | |
| "epoch": 0.9090118842127485, | |
| "grad_norm": 1.7714415788650513, | |
| "learning_rate": 6.323156394042139e-06, | |
| "loss": 1.3399, | |
| "step": 9150 | |
| }, | |
| { | |
| "epoch": 0.91000533982391, | |
| "grad_norm": 1.6133977174758911, | |
| "learning_rate": 6.186653431901628e-06, | |
| "loss": 1.3275, | |
| "step": 9160 | |
| }, | |
| { | |
| "epoch": 0.9109987954350715, | |
| "grad_norm": 1.8453999757766724, | |
| "learning_rate": 6.051608962904225e-06, | |
| "loss": 1.3427, | |
| "step": 9170 | |
| }, | |
| { | |
| "epoch": 0.9119922510462329, | |
| "grad_norm": 1.8623254299163818, | |
| "learning_rate": 5.91802435661296e-06, | |
| "loss": 1.3472, | |
| "step": 9180 | |
| }, | |
| { | |
| "epoch": 0.9129857066573944, | |
| "grad_norm": 1.7393712997436523, | |
| "learning_rate": 5.78590096778544e-06, | |
| "loss": 1.3177, | |
| "step": 9190 | |
| }, | |
| { | |
| "epoch": 0.9139791622685559, | |
| "grad_norm": 1.656901478767395, | |
| "learning_rate": 5.655240136360306e-06, | |
| "loss": 1.3255, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 0.9149726178797174, | |
| "grad_norm": 1.8177196979522705, | |
| "learning_rate": 5.5260431874435515e-06, | |
| "loss": 1.3191, | |
| "step": 9210 | |
| }, | |
| { | |
| "epoch": 0.9159660734908789, | |
| "grad_norm": 1.7882544994354248, | |
| "learning_rate": 5.398311431295061e-06, | |
| "loss": 1.3258, | |
| "step": 9220 | |
| }, | |
| { | |
| "epoch": 0.9169595291020403, | |
| "grad_norm": 1.6114939451217651, | |
| "learning_rate": 5.2720461633154354e-06, | |
| "loss": 1.3436, | |
| "step": 9230 | |
| }, | |
| { | |
| "epoch": 0.9179529847132017, | |
| "grad_norm": 1.629774570465088, | |
| "learning_rate": 5.147248664032744e-06, | |
| "loss": 1.327, | |
| "step": 9240 | |
| }, | |
| { | |
| "epoch": 0.9189464403243632, | |
| "grad_norm": 1.725443720817566, | |
| "learning_rate": 5.023920199089576e-06, | |
| "loss": 1.317, | |
| "step": 9250 | |
| }, | |
| { | |
| "epoch": 0.9199398959355247, | |
| "grad_norm": 1.6908073425292969, | |
| "learning_rate": 4.902062019230191e-06, | |
| "loss": 1.3251, | |
| "step": 9260 | |
| }, | |
| { | |
| "epoch": 0.9209333515466862, | |
| "grad_norm": 1.8233212232589722, | |
| "learning_rate": 4.78167536028789e-06, | |
| "loss": 1.3097, | |
| "step": 9270 | |
| }, | |
| { | |
| "epoch": 0.9219268071578477, | |
| "grad_norm": 2.000664234161377, | |
| "learning_rate": 4.662761443172414e-06, | |
| "loss": 1.3351, | |
| "step": 9280 | |
| }, | |
| { | |
| "epoch": 0.9229202627690092, | |
| "grad_norm": 1.5021830797195435, | |
| "learning_rate": 4.545321473857616e-06, | |
| "loss": 1.3282, | |
| "step": 9290 | |
| }, | |
| { | |
| "epoch": 0.9239137183801707, | |
| "grad_norm": 1.816205620765686, | |
| "learning_rate": 4.429356643369159e-06, | |
| "loss": 1.3234, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 0.9249071739913322, | |
| "grad_norm": 1.569473147392273, | |
| "learning_rate": 4.314868127772503e-06, | |
| "loss": 1.3237, | |
| "step": 9310 | |
| }, | |
| { | |
| "epoch": 0.9259006296024935, | |
| "grad_norm": 1.6449724435806274, | |
| "learning_rate": 4.201857088160987e-06, | |
| "loss": 1.3258, | |
| "step": 9320 | |
| }, | |
| { | |
| "epoch": 0.926894085213655, | |
| "grad_norm": 1.769514560699463, | |
| "learning_rate": 4.090324670643968e-06, | |
| "loss": 1.3256, | |
| "step": 9330 | |
| }, | |
| { | |
| "epoch": 0.9278875408248165, | |
| "grad_norm": 1.6415711641311646, | |
| "learning_rate": 3.9802720063353e-06, | |
| "loss": 1.3347, | |
| "step": 9340 | |
| }, | |
| { | |
| "epoch": 0.928880996435978, | |
| "grad_norm": 1.612988829612732, | |
| "learning_rate": 3.8717002113417885e-06, | |
| "loss": 1.3247, | |
| "step": 9350 | |
| }, | |
| { | |
| "epoch": 0.9298744520471395, | |
| "grad_norm": 1.739996075630188, | |
| "learning_rate": 3.764610386751904e-06, | |
| "loss": 1.3199, | |
| "step": 9360 | |
| }, | |
| { | |
| "epoch": 0.930867907658301, | |
| "grad_norm": 1.7688846588134766, | |
| "learning_rate": 3.659003618624656e-06, | |
| "loss": 1.3274, | |
| "step": 9370 | |
| }, | |
| { | |
| "epoch": 0.9318613632694625, | |
| "grad_norm": 1.875583529472351, | |
| "learning_rate": 3.554880977978436e-06, | |
| "loss": 1.3354, | |
| "step": 9380 | |
| }, | |
| { | |
| "epoch": 0.9328548188806239, | |
| "grad_norm": 1.7558965682983398, | |
| "learning_rate": 3.4522435207803567e-06, | |
| "loss": 1.3252, | |
| "step": 9390 | |
| }, | |
| { | |
| "epoch": 0.9338482744917853, | |
| "grad_norm": 1.815974473953247, | |
| "learning_rate": 3.3510922879353476e-06, | |
| "loss": 1.3306, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 0.9348417301029468, | |
| "grad_norm": 1.6319866180419922, | |
| "learning_rate": 3.2514283052757605e-06, | |
| "loss": 1.3441, | |
| "step": 9410 | |
| }, | |
| { | |
| "epoch": 0.9358351857141083, | |
| "grad_norm": 1.7212560176849365, | |
| "learning_rate": 3.1532525835508636e-06, | |
| "loss": 1.326, | |
| "step": 9420 | |
| }, | |
| { | |
| "epoch": 0.9368286413252698, | |
| "grad_norm": 1.7307348251342773, | |
| "learning_rate": 3.0565661184166146e-06, | |
| "loss": 1.3189, | |
| "step": 9430 | |
| }, | |
| { | |
| "epoch": 0.9378220969364313, | |
| "grad_norm": 1.8027082681655884, | |
| "learning_rate": 2.9613698904255857e-06, | |
| "loss": 1.3363, | |
| "step": 9440 | |
| }, | |
| { | |
| "epoch": 0.9388155525475927, | |
| "grad_norm": 1.6482763290405273, | |
| "learning_rate": 2.867664865017022e-06, | |
| "loss": 1.3412, | |
| "step": 9450 | |
| }, | |
| { | |
| "epoch": 0.9398090081587542, | |
| "grad_norm": 1.5483489036560059, | |
| "learning_rate": 2.7754519925069995e-06, | |
| "loss": 1.3178, | |
| "step": 9460 | |
| }, | |
| { | |
| "epoch": 0.9408024637699157, | |
| "grad_norm": 1.7157379388809204, | |
| "learning_rate": 2.684732208078816e-06, | |
| "loss": 1.3171, | |
| "step": 9470 | |
| }, | |
| { | |
| "epoch": 0.9417959193810772, | |
| "grad_norm": 1.6473640203475952, | |
| "learning_rate": 2.5955064317735474e-06, | |
| "loss": 1.3474, | |
| "step": 9480 | |
| }, | |
| { | |
| "epoch": 0.9427893749922386, | |
| "grad_norm": 1.595017671585083, | |
| "learning_rate": 2.5077755684806745e-06, | |
| "loss": 1.3358, | |
| "step": 9490 | |
| }, | |
| { | |
| "epoch": 0.9437828306034001, | |
| "grad_norm": 1.73741614818573, | |
| "learning_rate": 2.421540507928871e-06, | |
| "loss": 1.3096, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.9447762862145616, | |
| "grad_norm": 1.6367441415786743, | |
| "learning_rate": 2.3368021246770786e-06, | |
| "loss": 1.3028, | |
| "step": 9510 | |
| }, | |
| { | |
| "epoch": 0.945769741825723, | |
| "grad_norm": 1.7460381984710693, | |
| "learning_rate": 2.2535612781055478e-06, | |
| "loss": 1.3221, | |
| "step": 9520 | |
| }, | |
| { | |
| "epoch": 0.9467631974368845, | |
| "grad_norm": 1.6745285987854004, | |
| "learning_rate": 2.1718188124071613e-06, | |
| "loss": 1.3258, | |
| "step": 9530 | |
| }, | |
| { | |
| "epoch": 0.947756653048046, | |
| "grad_norm": 1.6275012493133545, | |
| "learning_rate": 2.0915755565788728e-06, | |
| "loss": 1.348, | |
| "step": 9540 | |
| }, | |
| { | |
| "epoch": 0.9487501086592075, | |
| "grad_norm": 1.706201434135437, | |
| "learning_rate": 2.0128323244132826e-06, | |
| "loss": 1.3504, | |
| "step": 9550 | |
| }, | |
| { | |
| "epoch": 0.949743564270369, | |
| "grad_norm": 1.6678355932235718, | |
| "learning_rate": 1.9355899144904263e-06, | |
| "loss": 1.3339, | |
| "step": 9560 | |
| }, | |
| { | |
| "epoch": 0.9507370198815304, | |
| "grad_norm": 1.677817940711975, | |
| "learning_rate": 1.8598491101695978e-06, | |
| "loss": 1.3309, | |
| "step": 9570 | |
| }, | |
| { | |
| "epoch": 0.9517304754926919, | |
| "grad_norm": 1.6711294651031494, | |
| "learning_rate": 1.7856106795814729e-06, | |
| "loss": 1.3348, | |
| "step": 9580 | |
| }, | |
| { | |
| "epoch": 0.9527239311038533, | |
| "grad_norm": 1.6787000894546509, | |
| "learning_rate": 1.7128753756203484e-06, | |
| "loss": 1.3117, | |
| "step": 9590 | |
| }, | |
| { | |
| "epoch": 0.9537173867150148, | |
| "grad_norm": 1.6609556674957275, | |
| "learning_rate": 1.6416439359363487e-06, | |
| "loss": 1.3211, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 0.9547108423261763, | |
| "grad_norm": 1.8903698921203613, | |
| "learning_rate": 1.5719170829281646e-06, | |
| "loss": 1.3301, | |
| "step": 9610 | |
| }, | |
| { | |
| "epoch": 0.9557042979373378, | |
| "grad_norm": 1.721840500831604, | |
| "learning_rate": 1.503695523735543e-06, | |
| "loss": 1.3277, | |
| "step": 9620 | |
| }, | |
| { | |
| "epoch": 0.9566977535484993, | |
| "grad_norm": 1.7415471076965332, | |
| "learning_rate": 1.4369799502322254e-06, | |
| "loss": 1.3208, | |
| "step": 9630 | |
| }, | |
| { | |
| "epoch": 0.9576912091596608, | |
| "grad_norm": 1.875077486038208, | |
| "learning_rate": 1.3717710390188709e-06, | |
| "loss": 1.3144, | |
| "step": 9640 | |
| }, | |
| { | |
| "epoch": 0.9586846647708223, | |
| "grad_norm": 1.6852308511734009, | |
| "learning_rate": 1.3080694514162437e-06, | |
| "loss": 1.3215, | |
| "step": 9650 | |
| }, | |
| { | |
| "epoch": 0.9596781203819836, | |
| "grad_norm": 1.724203109741211, | |
| "learning_rate": 1.245875833458454e-06, | |
| "loss": 1.3095, | |
| "step": 9660 | |
| }, | |
| { | |
| "epoch": 0.9606715759931451, | |
| "grad_norm": 1.761801838874817, | |
| "learning_rate": 1.1851908158864443e-06, | |
| "loss": 1.3133, | |
| "step": 9670 | |
| }, | |
| { | |
| "epoch": 0.9616650316043066, | |
| "grad_norm": 1.667737364768982, | |
| "learning_rate": 1.126015014141546e-06, | |
| "loss": 1.3159, | |
| "step": 9680 | |
| }, | |
| { | |
| "epoch": 0.9626584872154681, | |
| "grad_norm": 1.7830792665481567, | |
| "learning_rate": 1.0683490283593499e-06, | |
| "loss": 1.3136, | |
| "step": 9690 | |
| }, | |
| { | |
| "epoch": 0.9636519428266296, | |
| "grad_norm": 1.5339853763580322, | |
| "learning_rate": 1.0121934433634793e-06, | |
| "loss": 1.3104, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 0.9646453984377911, | |
| "grad_norm": 1.7048654556274414, | |
| "learning_rate": 9.575488286597267e-07, | |
| "loss": 1.3209, | |
| "step": 9710 | |
| }, | |
| { | |
| "epoch": 0.9656388540489526, | |
| "grad_norm": 1.6286853551864624, | |
| "learning_rate": 9.044157384302919e-07, | |
| "loss": 1.3128, | |
| "step": 9720 | |
| }, | |
| { | |
| "epoch": 0.966632309660114, | |
| "grad_norm": 1.7102385759353638, | |
| "learning_rate": 8.527947115281542e-07, | |
| "loss": 1.3405, | |
| "step": 9730 | |
| }, | |
| { | |
| "epoch": 0.9676257652712754, | |
| "grad_norm": 1.6307735443115234, | |
| "learning_rate": 8.026862714715587e-07, | |
| "loss": 1.3347, | |
| "step": 9740 | |
| }, | |
| { | |
| "epoch": 0.9686192208824369, | |
| "grad_norm": 1.6709721088409424, | |
| "learning_rate": 7.540909264387718e-07, | |
| "loss": 1.3169, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 0.9696126764935984, | |
| "grad_norm": 1.8650215864181519, | |
| "learning_rate": 7.07009169262901e-07, | |
| "loss": 1.3148, | |
| "step": 9760 | |
| }, | |
| { | |
| "epoch": 0.9706061321047599, | |
| "grad_norm": 1.7589495182037354, | |
| "learning_rate": 6.614414774268661e-07, | |
| "loss": 1.3198, | |
| "step": 9770 | |
| }, | |
| { | |
| "epoch": 0.9715995877159214, | |
| "grad_norm": 1.8895137310028076, | |
| "learning_rate": 6.173883130586532e-07, | |
| "loss": 1.3295, | |
| "step": 9780 | |
| }, | |
| { | |
| "epoch": 0.9725930433270829, | |
| "grad_norm": 1.7110244035720825, | |
| "learning_rate": 5.748501229264679e-07, | |
| "loss": 1.3252, | |
| "step": 9790 | |
| }, | |
| { | |
| "epoch": 0.9735864989382443, | |
| "grad_norm": 1.6796895265579224, | |
| "learning_rate": 5.338273384343895e-07, | |
| "loss": 1.3245, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 0.9745799545494058, | |
| "grad_norm": 1.6769100427627563, | |
| "learning_rate": 4.94320375617857e-07, | |
| "loss": 1.3212, | |
| "step": 9810 | |
| }, | |
| { | |
| "epoch": 0.9755734101605673, | |
| "grad_norm": 1.7454545497894287, | |
| "learning_rate": 4.563296351394907e-07, | |
| "loss": 1.3277, | |
| "step": 9820 | |
| }, | |
| { | |
| "epoch": 0.9765668657717287, | |
| "grad_norm": 1.5692331790924072, | |
| "learning_rate": 4.1985550228506025e-07, | |
| "loss": 1.3272, | |
| "step": 9830 | |
| }, | |
| { | |
| "epoch": 0.9775603213828902, | |
| "grad_norm": 1.6089588403701782, | |
| "learning_rate": 3.8489834695953903e-07, | |
| "loss": 1.3095, | |
| "step": 9840 | |
| }, | |
| { | |
| "epoch": 0.9785537769940517, | |
| "grad_norm": 2.1445343494415283, | |
| "learning_rate": 3.5145852368337336e-07, | |
| "loss": 1.3406, | |
| "step": 9850 | |
| }, | |
| { | |
| "epoch": 0.9795472326052131, | |
| "grad_norm": 1.6898231506347656, | |
| "learning_rate": 3.195363715888688e-07, | |
| "loss": 1.3175, | |
| "step": 9860 | |
| }, | |
| { | |
| "epoch": 0.9805406882163746, | |
| "grad_norm": 1.6643315553665161, | |
| "learning_rate": 2.891322144167929e-07, | |
| "loss": 1.3262, | |
| "step": 9870 | |
| }, | |
| { | |
| "epoch": 0.9815341438275361, | |
| "grad_norm": 1.801221251487732, | |
| "learning_rate": 2.6024636051301116e-07, | |
| "loss": 1.3387, | |
| "step": 9880 | |
| }, | |
| { | |
| "epoch": 0.9825275994386976, | |
| "grad_norm": 1.5969946384429932, | |
| "learning_rate": 2.328791028254562e-07, | |
| "loss": 1.3256, | |
| "step": 9890 | |
| }, | |
| { | |
| "epoch": 0.9835210550498591, | |
| "grad_norm": 1.6810213327407837, | |
| "learning_rate": 2.0703071890108003e-07, | |
| "loss": 1.3432, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 0.9845145106610206, | |
| "grad_norm": 1.8331464529037476, | |
| "learning_rate": 1.8270147088308984e-07, | |
| "loss": 1.3039, | |
| "step": 9910 | |
| }, | |
| { | |
| "epoch": 0.985507966272182, | |
| "grad_norm": 1.6391011476516724, | |
| "learning_rate": 1.5989160550826663e-07, | |
| "loss": 1.3297, | |
| "step": 9920 | |
| }, | |
| { | |
| "epoch": 0.9865014218833434, | |
| "grad_norm": 1.7446280717849731, | |
| "learning_rate": 1.3860135410443395e-07, | |
| "loss": 1.3411, | |
| "step": 9930 | |
| }, | |
| { | |
| "epoch": 0.9874948774945049, | |
| "grad_norm": 1.5376980304718018, | |
| "learning_rate": 1.1883093258819309e-07, | |
| "loss": 1.3123, | |
| "step": 9940 | |
| }, | |
| { | |
| "epoch": 0.9884883331056664, | |
| "grad_norm": 1.6628799438476562, | |
| "learning_rate": 1.0058054146265815e-07, | |
| "loss": 1.3245, | |
| "step": 9950 | |
| }, | |
| { | |
| "epoch": 0.9894817887168279, | |
| "grad_norm": 1.7408899068832397, | |
| "learning_rate": 8.385036581545768e-08, | |
| "loss": 1.3133, | |
| "step": 9960 | |
| }, | |
| { | |
| "epoch": 0.9904752443279894, | |
| "grad_norm": 1.6350455284118652, | |
| "learning_rate": 6.864057531681955e-08, | |
| "loss": 1.329, | |
| "step": 9970 | |
| }, | |
| { | |
| "epoch": 0.9914686999391509, | |
| "grad_norm": 1.7738535404205322, | |
| "learning_rate": 5.49513242179056e-08, | |
| "loss": 1.3419, | |
| "step": 9980 | |
| }, | |
| { | |
| "epoch": 0.9924621555503124, | |
| "grad_norm": 1.6542763710021973, | |
| "learning_rate": 4.2782751349246246e-08, | |
| "loss": 1.3058, | |
| "step": 9990 | |
| }, | |
| { | |
| "epoch": 0.9934556111614737, | |
| "grad_norm": 1.6510471105575562, | |
| "learning_rate": 3.213498011924165e-08, | |
| "loss": 1.3377, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.9944490667726352, | |
| "grad_norm": 1.7229485511779785, | |
| "learning_rate": 2.3008118513012653e-08, | |
| "loss": 1.3331, | |
| "step": 10010 | |
| }, | |
| { | |
| "epoch": 0.9954425223837967, | |
| "grad_norm": 1.5858336687088013, | |
| "learning_rate": 1.540225909130166e-08, | |
| "loss": 1.3306, | |
| "step": 10020 | |
| }, | |
| { | |
| "epoch": 0.9964359779949582, | |
| "grad_norm": 1.8028407096862793, | |
| "learning_rate": 9.317478989423477e-09, | |
| "loss": 1.316, | |
| "step": 10030 | |
| }, | |
| { | |
| "epoch": 0.9974294336061197, | |
| "grad_norm": 2.418792486190796, | |
| "learning_rate": 4.753839916632474e-09, | |
| "loss": 1.3326, | |
| "step": 10040 | |
| }, | |
| { | |
| "epoch": 0.9984228892172812, | |
| "grad_norm": 1.748274326324463, | |
| "learning_rate": 1.7113881553731945e-09, | |
| "loss": 1.343, | |
| "step": 10050 | |
| }, | |
| { | |
| "epoch": 0.9994163448284427, | |
| "grad_norm": 1.8571064472198486, | |
| "learning_rate": 1.901545608973265e-10, | |
| "loss": 1.3276, | |
| "step": 10060 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 10065, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 1007, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 467960699289600.0, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |