{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 20.0, "eval_steps": 500, "global_step": 8400, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.023809523809523808, "grad_norm": 0.09493794292211533, "learning_rate": 5e-05, "loss": 0.2812, "step": 10 }, { "epoch": 0.047619047619047616, "grad_norm": 0.06558936089277267, "learning_rate": 5e-05, "loss": 0.0724, "step": 20 }, { "epoch": 0.07142857142857142, "grad_norm": 0.05934833735227585, "learning_rate": 5e-05, "loss": 0.0328, "step": 30 }, { "epoch": 0.09523809523809523, "grad_norm": 0.0350530743598938, "learning_rate": 5e-05, "loss": 0.0267, "step": 40 }, { "epoch": 0.11904761904761904, "grad_norm": 0.028712034225463867, "learning_rate": 5e-05, "loss": 0.0236, "step": 50 }, { "epoch": 0.14285714285714285, "grad_norm": 0.024718789383769035, "learning_rate": 5e-05, "loss": 0.023, "step": 60 }, { "epoch": 0.16666666666666666, "grad_norm": 0.022061526775360107, "learning_rate": 5e-05, "loss": 0.0226, "step": 70 }, { "epoch": 0.19047619047619047, "grad_norm": 0.025670066475868225, "learning_rate": 5e-05, "loss": 0.0218, "step": 80 }, { "epoch": 0.21428571428571427, "grad_norm": 0.02997678704559803, "learning_rate": 5e-05, "loss": 0.0209, "step": 90 }, { "epoch": 0.23809523809523808, "grad_norm": 0.0373968742787838, "learning_rate": 5e-05, "loss": 0.0221, "step": 100 }, { "epoch": 0.2619047619047619, "grad_norm": 0.02778824232518673, "learning_rate": 5e-05, "loss": 0.0207, "step": 110 }, { "epoch": 0.2857142857142857, "grad_norm": 0.018951816484332085, "learning_rate": 5e-05, "loss": 0.0212, "step": 120 }, { "epoch": 0.30952380952380953, "grad_norm": 0.03219084069132805, "learning_rate": 5e-05, "loss": 0.0197, "step": 130 }, { "epoch": 0.3333333333333333, "grad_norm": 0.035099539905786514, "learning_rate": 5e-05, "loss": 0.0212, "step": 140 }, { "epoch": 0.35714285714285715, "grad_norm": 0.022762970998883247, "learning_rate": 5e-05, "loss": 0.0205, "step": 150 }, { "epoch": 0.38095238095238093, "grad_norm": 0.021571127697825432, "learning_rate": 5e-05, "loss": 0.0192, "step": 160 }, { "epoch": 0.40476190476190477, "grad_norm": 0.028216874226927757, "learning_rate": 5e-05, "loss": 0.0194, "step": 170 }, { "epoch": 0.42857142857142855, "grad_norm": 0.018433215096592903, "learning_rate": 5e-05, "loss": 0.0201, "step": 180 }, { "epoch": 0.4523809523809524, "grad_norm": 0.03548741340637207, "learning_rate": 5e-05, "loss": 0.0193, "step": 190 }, { "epoch": 0.47619047619047616, "grad_norm": 0.027879556640982628, "learning_rate": 5e-05, "loss": 0.0187, "step": 200 }, { "epoch": 0.5, "grad_norm": 0.022470738738775253, "learning_rate": 5e-05, "loss": 0.0186, "step": 210 }, { "epoch": 0.5238095238095238, "grad_norm": 0.02206714078783989, "learning_rate": 5e-05, "loss": 0.0184, "step": 220 }, { "epoch": 0.5476190476190477, "grad_norm": 0.024714656174182892, "learning_rate": 5e-05, "loss": 0.0188, "step": 230 }, { "epoch": 0.5714285714285714, "grad_norm": 0.03632698208093643, "learning_rate": 5e-05, "loss": 0.0188, "step": 240 }, { "epoch": 0.5952380952380952, "grad_norm": 0.017954664304852486, "learning_rate": 5e-05, "loss": 0.0177, "step": 250 }, { "epoch": 0.6190476190476191, "grad_norm": 0.026338977739214897, "learning_rate": 5e-05, "loss": 0.0166, "step": 260 }, { "epoch": 0.6428571428571429, "grad_norm": 0.028515340760350227, "learning_rate": 5e-05, "loss": 0.0177, "step": 270 }, { "epoch": 0.6666666666666666, "grad_norm": 0.0308682881295681, "learning_rate": 5e-05, "loss": 0.0176, "step": 280 }, { "epoch": 0.6904761904761905, "grad_norm": 0.018636852502822876, "learning_rate": 5e-05, "loss": 0.0167, "step": 290 }, { "epoch": 0.7142857142857143, "grad_norm": 0.024639299139380455, "learning_rate": 5e-05, "loss": 0.0171, "step": 300 }, { "epoch": 0.7380952380952381, "grad_norm": 0.02328549325466156, "learning_rate": 5e-05, "loss": 0.018, "step": 310 }, { "epoch": 0.7619047619047619, "grad_norm": 0.02998117357492447, "learning_rate": 5e-05, "loss": 0.0166, "step": 320 }, { "epoch": 0.7857142857142857, "grad_norm": 0.022752949967980385, "learning_rate": 5e-05, "loss": 0.017, "step": 330 }, { "epoch": 0.8095238095238095, "grad_norm": 0.02922181785106659, "learning_rate": 5e-05, "loss": 0.0176, "step": 340 }, { "epoch": 0.8333333333333334, "grad_norm": 0.026469197124242783, "learning_rate": 5e-05, "loss": 0.0183, "step": 350 }, { "epoch": 0.8571428571428571, "grad_norm": 0.03358051925897598, "learning_rate": 5e-05, "loss": 0.0188, "step": 360 }, { "epoch": 0.8809523809523809, "grad_norm": 0.0185205340385437, "learning_rate": 5e-05, "loss": 0.0173, "step": 370 }, { "epoch": 0.9047619047619048, "grad_norm": 0.018949517980217934, "learning_rate": 5e-05, "loss": 0.0169, "step": 380 }, { "epoch": 0.9285714285714286, "grad_norm": 0.022263336926698685, "learning_rate": 5e-05, "loss": 0.017, "step": 390 }, { "epoch": 0.9523809523809523, "grad_norm": 0.023463713005185127, "learning_rate": 5e-05, "loss": 0.0172, "step": 400 }, { "epoch": 0.9761904761904762, "grad_norm": 0.023527899757027626, "learning_rate": 5e-05, "loss": 0.0156, "step": 410 }, { "epoch": 1.0, "grad_norm": 0.026089975610375404, "learning_rate": 5e-05, "loss": 0.0154, "step": 420 }, { "epoch": 1.0238095238095237, "grad_norm": 0.017531510442495346, "learning_rate": 5e-05, "loss": 0.0175, "step": 430 }, { "epoch": 1.0476190476190477, "grad_norm": 0.017702311277389526, "learning_rate": 5e-05, "loss": 0.0151, "step": 440 }, { "epoch": 1.0714285714285714, "grad_norm": 0.018016375601291656, "learning_rate": 5e-05, "loss": 0.0168, "step": 450 }, { "epoch": 1.0952380952380953, "grad_norm": 0.028256898745894432, "learning_rate": 5e-05, "loss": 0.0164, "step": 460 }, { "epoch": 1.119047619047619, "grad_norm": 0.017876937985420227, "learning_rate": 5e-05, "loss": 0.0148, "step": 470 }, { "epoch": 1.1428571428571428, "grad_norm": 0.018716128543019295, "learning_rate": 5e-05, "loss": 0.0144, "step": 480 }, { "epoch": 1.1666666666666667, "grad_norm": 0.02162998355925083, "learning_rate": 5e-05, "loss": 0.0148, "step": 490 }, { "epoch": 1.1904761904761905, "grad_norm": 0.023180391639471054, "learning_rate": 5e-05, "loss": 0.0156, "step": 500 }, { "epoch": 1.2142857142857142, "grad_norm": 0.022495009005069733, "learning_rate": 5e-05, "loss": 0.0149, "step": 510 }, { "epoch": 1.2380952380952381, "grad_norm": 0.024750931188464165, "learning_rate": 5e-05, "loss": 0.0158, "step": 520 }, { "epoch": 1.2619047619047619, "grad_norm": 0.027094174176454544, "learning_rate": 5e-05, "loss": 0.0149, "step": 530 }, { "epoch": 1.2857142857142856, "grad_norm": 0.021923832595348358, "learning_rate": 5e-05, "loss": 0.0146, "step": 540 }, { "epoch": 1.3095238095238095, "grad_norm": 0.025097332894802094, "learning_rate": 5e-05, "loss": 0.0136, "step": 550 }, { "epoch": 1.3333333333333333, "grad_norm": 0.034609246999025345, "learning_rate": 5e-05, "loss": 0.0152, "step": 560 }, { "epoch": 1.3571428571428572, "grad_norm": 0.02195250615477562, "learning_rate": 5e-05, "loss": 0.0141, "step": 570 }, { "epoch": 1.380952380952381, "grad_norm": 0.01835169643163681, "learning_rate": 5e-05, "loss": 0.0148, "step": 580 }, { "epoch": 1.4047619047619047, "grad_norm": 0.026779156178236008, "learning_rate": 5e-05, "loss": 0.0154, "step": 590 }, { "epoch": 1.4285714285714286, "grad_norm": 0.025482457131147385, "learning_rate": 5e-05, "loss": 0.0142, "step": 600 }, { "epoch": 1.4523809523809523, "grad_norm": 0.022315753623843193, "learning_rate": 5e-05, "loss": 0.0137, "step": 610 }, { "epoch": 1.4761904761904763, "grad_norm": 0.02147059701383114, "learning_rate": 5e-05, "loss": 0.0137, "step": 620 }, { "epoch": 1.5, "grad_norm": 0.021188456565141678, "learning_rate": 5e-05, "loss": 0.0135, "step": 630 }, { "epoch": 1.5238095238095237, "grad_norm": 0.018250903114676476, "learning_rate": 5e-05, "loss": 0.0137, "step": 640 }, { "epoch": 1.5476190476190477, "grad_norm": 0.02014111913740635, "learning_rate": 5e-05, "loss": 0.0141, "step": 650 }, { "epoch": 1.5714285714285714, "grad_norm": 0.031997792422771454, "learning_rate": 5e-05, "loss": 0.0145, "step": 660 }, { "epoch": 1.5952380952380953, "grad_norm": 0.02311546355485916, "learning_rate": 5e-05, "loss": 0.0139, "step": 670 }, { "epoch": 1.619047619047619, "grad_norm": 0.022159090265631676, "learning_rate": 5e-05, "loss": 0.0127, "step": 680 }, { "epoch": 1.6428571428571428, "grad_norm": 0.023639395833015442, "learning_rate": 5e-05, "loss": 0.0139, "step": 690 }, { "epoch": 1.6666666666666665, "grad_norm": 0.025874590501189232, "learning_rate": 5e-05, "loss": 0.0142, "step": 700 }, { "epoch": 1.6904761904761905, "grad_norm": 0.015242703258991241, "learning_rate": 5e-05, "loss": 0.0131, "step": 710 }, { "epoch": 1.7142857142857144, "grad_norm": 0.01896982081234455, "learning_rate": 5e-05, "loss": 0.0129, "step": 720 }, { "epoch": 1.7380952380952381, "grad_norm": 0.019098645076155663, "learning_rate": 5e-05, "loss": 0.014, "step": 730 }, { "epoch": 1.7619047619047619, "grad_norm": 0.020913679152727127, "learning_rate": 5e-05, "loss": 0.0139, "step": 740 }, { "epoch": 1.7857142857142856, "grad_norm": 0.02301224134862423, "learning_rate": 5e-05, "loss": 0.0134, "step": 750 }, { "epoch": 1.8095238095238095, "grad_norm": 0.020251275971531868, "learning_rate": 5e-05, "loss": 0.0126, "step": 760 }, { "epoch": 1.8333333333333335, "grad_norm": 0.023471834138035774, "learning_rate": 5e-05, "loss": 0.0136, "step": 770 }, { "epoch": 1.8571428571428572, "grad_norm": 0.024632664397358894, "learning_rate": 5e-05, "loss": 0.0145, "step": 780 }, { "epoch": 1.880952380952381, "grad_norm": 0.021192265674471855, "learning_rate": 5e-05, "loss": 0.0138, "step": 790 }, { "epoch": 1.9047619047619047, "grad_norm": 0.021170521154999733, "learning_rate": 5e-05, "loss": 0.0136, "step": 800 }, { "epoch": 1.9285714285714286, "grad_norm": 0.018560606986284256, "learning_rate": 5e-05, "loss": 0.0129, "step": 810 }, { "epoch": 1.9523809523809523, "grad_norm": 0.020409613847732544, "learning_rate": 5e-05, "loss": 0.0138, "step": 820 }, { "epoch": 1.9761904761904763, "grad_norm": 0.02413373813033104, "learning_rate": 5e-05, "loss": 0.0128, "step": 830 }, { "epoch": 2.0, "grad_norm": 0.022253163158893585, "learning_rate": 5e-05, "loss": 0.0127, "step": 840 }, { "epoch": 2.0238095238095237, "grad_norm": 0.017652267590165138, "learning_rate": 5e-05, "loss": 0.0143, "step": 850 }, { "epoch": 2.0476190476190474, "grad_norm": 0.018153948709368706, "learning_rate": 5e-05, "loss": 0.0128, "step": 860 }, { "epoch": 2.0714285714285716, "grad_norm": 0.01955491490662098, "learning_rate": 5e-05, "loss": 0.0132, "step": 870 }, { "epoch": 2.0952380952380953, "grad_norm": 0.021146738901734352, "learning_rate": 5e-05, "loss": 0.0124, "step": 880 }, { "epoch": 2.119047619047619, "grad_norm": 0.016339635476469994, "learning_rate": 5e-05, "loss": 0.0122, "step": 890 }, { "epoch": 2.142857142857143, "grad_norm": 0.017886130139231682, "learning_rate": 5e-05, "loss": 0.0113, "step": 900 }, { "epoch": 2.1666666666666665, "grad_norm": 0.02060827612876892, "learning_rate": 5e-05, "loss": 0.0117, "step": 910 }, { "epoch": 2.1904761904761907, "grad_norm": 0.016803547739982605, "learning_rate": 5e-05, "loss": 0.0132, "step": 920 }, { "epoch": 2.2142857142857144, "grad_norm": 0.018075264990329742, "learning_rate": 5e-05, "loss": 0.0119, "step": 930 }, { "epoch": 2.238095238095238, "grad_norm": 0.023032259196043015, "learning_rate": 5e-05, "loss": 0.0121, "step": 940 }, { "epoch": 2.261904761904762, "grad_norm": 0.028263242915272713, "learning_rate": 5e-05, "loss": 0.0114, "step": 950 }, { "epoch": 2.2857142857142856, "grad_norm": 0.02737823873758316, "learning_rate": 5e-05, "loss": 0.0112, "step": 960 }, { "epoch": 2.3095238095238093, "grad_norm": 0.021527152508497238, "learning_rate": 5e-05, "loss": 0.0111, "step": 970 }, { "epoch": 2.3333333333333335, "grad_norm": 0.027928918600082397, "learning_rate": 5e-05, "loss": 0.0123, "step": 980 }, { "epoch": 2.357142857142857, "grad_norm": 0.024291587993502617, "learning_rate": 5e-05, "loss": 0.0112, "step": 990 }, { "epoch": 2.380952380952381, "grad_norm": 0.024933788925409317, "learning_rate": 5e-05, "loss": 0.0122, "step": 1000 }, { "epoch": 2.4047619047619047, "grad_norm": 0.025925207883119583, "learning_rate": 5e-05, "loss": 0.0126, "step": 1010 }, { "epoch": 2.4285714285714284, "grad_norm": 0.021988635882735252, "learning_rate": 5e-05, "loss": 0.0111, "step": 1020 }, { "epoch": 2.4523809523809526, "grad_norm": 0.029331982135772705, "learning_rate": 5e-05, "loss": 0.0113, "step": 1030 }, { "epoch": 2.4761904761904763, "grad_norm": 0.023319706320762634, "learning_rate": 5e-05, "loss": 0.0111, "step": 1040 }, { "epoch": 2.5, "grad_norm": 0.016594231128692627, "learning_rate": 5e-05, "loss": 0.0106, "step": 1050 }, { "epoch": 2.5238095238095237, "grad_norm": 0.022670956328511238, "learning_rate": 5e-05, "loss": 0.0109, "step": 1060 }, { "epoch": 2.5476190476190474, "grad_norm": 0.02389667183160782, "learning_rate": 5e-05, "loss": 0.0116, "step": 1070 }, { "epoch": 2.571428571428571, "grad_norm": 0.03487304970622063, "learning_rate": 5e-05, "loss": 0.0114, "step": 1080 }, { "epoch": 2.5952380952380953, "grad_norm": 0.02180272527039051, "learning_rate": 5e-05, "loss": 0.0116, "step": 1090 }, { "epoch": 2.619047619047619, "grad_norm": 0.021106529980897903, "learning_rate": 5e-05, "loss": 0.0105, "step": 1100 }, { "epoch": 2.642857142857143, "grad_norm": 0.023611431941390038, "learning_rate": 5e-05, "loss": 0.0119, "step": 1110 }, { "epoch": 2.6666666666666665, "grad_norm": 0.024628831073641777, "learning_rate": 5e-05, "loss": 0.0114, "step": 1120 }, { "epoch": 2.6904761904761907, "grad_norm": 0.014470246620476246, "learning_rate": 5e-05, "loss": 0.0102, "step": 1130 }, { "epoch": 2.7142857142857144, "grad_norm": 0.022718733176589012, "learning_rate": 5e-05, "loss": 0.0103, "step": 1140 }, { "epoch": 2.738095238095238, "grad_norm": 0.020995931699872017, "learning_rate": 5e-05, "loss": 0.0106, "step": 1150 }, { "epoch": 2.761904761904762, "grad_norm": 0.024527721107006073, "learning_rate": 5e-05, "loss": 0.0107, "step": 1160 }, { "epoch": 2.7857142857142856, "grad_norm": 0.021473698318004608, "learning_rate": 5e-05, "loss": 0.0104, "step": 1170 }, { "epoch": 2.8095238095238093, "grad_norm": 0.018530316650867462, "learning_rate": 5e-05, "loss": 0.0101, "step": 1180 }, { "epoch": 2.8333333333333335, "grad_norm": 0.02302338182926178, "learning_rate": 5e-05, "loss": 0.0105, "step": 1190 }, { "epoch": 2.857142857142857, "grad_norm": 0.025742432102560997, "learning_rate": 5e-05, "loss": 0.0114, "step": 1200 }, { "epoch": 2.880952380952381, "grad_norm": 0.027784395962953568, "learning_rate": 5e-05, "loss": 0.0111, "step": 1210 }, { "epoch": 2.9047619047619047, "grad_norm": 0.021398941054940224, "learning_rate": 5e-05, "loss": 0.0102, "step": 1220 }, { "epoch": 2.928571428571429, "grad_norm": 0.027173219248652458, "learning_rate": 5e-05, "loss": 0.0101, "step": 1230 }, { "epoch": 2.9523809523809526, "grad_norm": 0.02305479347705841, "learning_rate": 5e-05, "loss": 0.0113, "step": 1240 }, { "epoch": 2.9761904761904763, "grad_norm": 0.024614600464701653, "learning_rate": 5e-05, "loss": 0.0105, "step": 1250 }, { "epoch": 3.0, "grad_norm": 0.01538562960922718, "learning_rate": 5e-05, "loss": 0.0103, "step": 1260 }, { "epoch": 3.0238095238095237, "grad_norm": 0.018940571695566177, "learning_rate": 5e-05, "loss": 0.0117, "step": 1270 }, { "epoch": 3.0476190476190474, "grad_norm": 0.014765114523470402, "learning_rate": 5e-05, "loss": 0.0104, "step": 1280 }, { "epoch": 3.0714285714285716, "grad_norm": 0.02444819174706936, "learning_rate": 5e-05, "loss": 0.0115, "step": 1290 }, { "epoch": 3.0952380952380953, "grad_norm": 0.02040857821702957, "learning_rate": 5e-05, "loss": 0.0102, "step": 1300 }, { "epoch": 3.119047619047619, "grad_norm": 0.01789598912000656, "learning_rate": 5e-05, "loss": 0.0095, "step": 1310 }, { "epoch": 3.142857142857143, "grad_norm": 0.020275596529245377, "learning_rate": 5e-05, "loss": 0.0084, "step": 1320 }, { "epoch": 3.1666666666666665, "grad_norm": 0.02351970411837101, "learning_rate": 5e-05, "loss": 0.0098, "step": 1330 }, { "epoch": 3.1904761904761907, "grad_norm": 0.017525261268019676, "learning_rate": 5e-05, "loss": 0.0103, "step": 1340 }, { "epoch": 3.2142857142857144, "grad_norm": 0.019003944471478462, "learning_rate": 5e-05, "loss": 0.0094, "step": 1350 }, { "epoch": 3.238095238095238, "grad_norm": 0.02266862988471985, "learning_rate": 5e-05, "loss": 0.0094, "step": 1360 }, { "epoch": 3.261904761904762, "grad_norm": 0.019711384549736977, "learning_rate": 5e-05, "loss": 0.0093, "step": 1370 }, { "epoch": 3.2857142857142856, "grad_norm": 0.02152351662516594, "learning_rate": 5e-05, "loss": 0.0082, "step": 1380 }, { "epoch": 3.3095238095238093, "grad_norm": 0.018401287496089935, "learning_rate": 5e-05, "loss": 0.0082, "step": 1390 }, { "epoch": 3.3333333333333335, "grad_norm": 0.030380433425307274, "learning_rate": 5e-05, "loss": 0.0095, "step": 1400 }, { "epoch": 3.357142857142857, "grad_norm": 0.018700776621699333, "learning_rate": 5e-05, "loss": 0.0087, "step": 1410 }, { "epoch": 3.380952380952381, "grad_norm": 0.01758432574570179, "learning_rate": 5e-05, "loss": 0.0092, "step": 1420 }, { "epoch": 3.4047619047619047, "grad_norm": 0.027525080367922783, "learning_rate": 5e-05, "loss": 0.0096, "step": 1430 }, { "epoch": 3.4285714285714284, "grad_norm": 0.023578746244311333, "learning_rate": 5e-05, "loss": 0.0087, "step": 1440 }, { "epoch": 3.4523809523809526, "grad_norm": 0.020561877638101578, "learning_rate": 5e-05, "loss": 0.0086, "step": 1450 }, { "epoch": 3.4761904761904763, "grad_norm": 0.024523900821805, "learning_rate": 5e-05, "loss": 0.0089, "step": 1460 }, { "epoch": 3.5, "grad_norm": 0.017780443653464317, "learning_rate": 5e-05, "loss": 0.0082, "step": 1470 }, { "epoch": 3.5238095238095237, "grad_norm": 0.025551529601216316, "learning_rate": 5e-05, "loss": 0.0083, "step": 1480 }, { "epoch": 3.5476190476190474, "grad_norm": 0.024370500817894936, "learning_rate": 5e-05, "loss": 0.0091, "step": 1490 }, { "epoch": 3.571428571428571, "grad_norm": 0.03639150410890579, "learning_rate": 5e-05, "loss": 0.0092, "step": 1500 }, { "epoch": 3.5952380952380953, "grad_norm": 0.02289510704576969, "learning_rate": 5e-05, "loss": 0.009, "step": 1510 }, { "epoch": 3.619047619047619, "grad_norm": 0.019606849178671837, "learning_rate": 5e-05, "loss": 0.0088, "step": 1520 }, { "epoch": 3.642857142857143, "grad_norm": 0.037268321961164474, "learning_rate": 5e-05, "loss": 0.0088, "step": 1530 }, { "epoch": 3.6666666666666665, "grad_norm": 0.028575578704476357, "learning_rate": 5e-05, "loss": 0.0099, "step": 1540 }, { "epoch": 3.6904761904761907, "grad_norm": 0.013424446806311607, "learning_rate": 5e-05, "loss": 0.0076, "step": 1550 }, { "epoch": 3.7142857142857144, "grad_norm": 0.01701328158378601, "learning_rate": 5e-05, "loss": 0.0084, "step": 1560 }, { "epoch": 3.738095238095238, "grad_norm": 0.025259124115109444, "learning_rate": 5e-05, "loss": 0.0081, "step": 1570 }, { "epoch": 3.761904761904762, "grad_norm": 0.024821562692523003, "learning_rate": 5e-05, "loss": 0.0086, "step": 1580 }, { "epoch": 3.7857142857142856, "grad_norm": 0.024605993181467056, "learning_rate": 5e-05, "loss": 0.0084, "step": 1590 }, { "epoch": 3.8095238095238093, "grad_norm": 0.02455594390630722, "learning_rate": 5e-05, "loss": 0.0075, "step": 1600 }, { "epoch": 3.8333333333333335, "grad_norm": 0.025864994153380394, "learning_rate": 5e-05, "loss": 0.008, "step": 1610 }, { "epoch": 3.857142857142857, "grad_norm": 0.02395598590373993, "learning_rate": 5e-05, "loss": 0.0087, "step": 1620 }, { "epoch": 3.880952380952381, "grad_norm": 0.015691502019762993, "learning_rate": 5e-05, "loss": 0.0084, "step": 1630 }, { "epoch": 3.9047619047619047, "grad_norm": 0.0223532821983099, "learning_rate": 5e-05, "loss": 0.0077, "step": 1640 }, { "epoch": 3.928571428571429, "grad_norm": 0.028256218880414963, "learning_rate": 5e-05, "loss": 0.0076, "step": 1650 }, { "epoch": 3.9523809523809526, "grad_norm": 0.023097924888134003, "learning_rate": 5e-05, "loss": 0.0091, "step": 1660 }, { "epoch": 3.9761904761904763, "grad_norm": 0.02783946879208088, "learning_rate": 5e-05, "loss": 0.0076, "step": 1670 }, { "epoch": 4.0, "grad_norm": 0.013011032715439796, "learning_rate": 5e-05, "loss": 0.008, "step": 1680 }, { "epoch": 4.023809523809524, "grad_norm": 0.02342393808066845, "learning_rate": 5e-05, "loss": 0.0092, "step": 1690 }, { "epoch": 4.0476190476190474, "grad_norm": 0.021954068914055824, "learning_rate": 5e-05, "loss": 0.009, "step": 1700 }, { "epoch": 4.071428571428571, "grad_norm": 0.0207274928689003, "learning_rate": 5e-05, "loss": 0.0092, "step": 1710 }, { "epoch": 4.095238095238095, "grad_norm": 0.03118881769478321, "learning_rate": 5e-05, "loss": 0.008, "step": 1720 }, { "epoch": 4.119047619047619, "grad_norm": 0.015900803729891777, "learning_rate": 5e-05, "loss": 0.0074, "step": 1730 }, { "epoch": 4.142857142857143, "grad_norm": 0.013303752988576889, "learning_rate": 5e-05, "loss": 0.0069, "step": 1740 }, { "epoch": 4.166666666666667, "grad_norm": 0.022115593776106834, "learning_rate": 5e-05, "loss": 0.0078, "step": 1750 }, { "epoch": 4.190476190476191, "grad_norm": 0.021266015246510506, "learning_rate": 5e-05, "loss": 0.008, "step": 1760 }, { "epoch": 4.214285714285714, "grad_norm": 0.023192718625068665, "learning_rate": 5e-05, "loss": 0.0075, "step": 1770 }, { "epoch": 4.238095238095238, "grad_norm": 0.02218669466674328, "learning_rate": 5e-05, "loss": 0.007, "step": 1780 }, { "epoch": 4.261904761904762, "grad_norm": 0.024710968136787415, "learning_rate": 5e-05, "loss": 0.0071, "step": 1790 }, { "epoch": 4.285714285714286, "grad_norm": 0.022398030385375023, "learning_rate": 5e-05, "loss": 0.0057, "step": 1800 }, { "epoch": 4.309523809523809, "grad_norm": 0.025463633239269257, "learning_rate": 5e-05, "loss": 0.0066, "step": 1810 }, { "epoch": 4.333333333333333, "grad_norm": 0.024359669536352158, "learning_rate": 5e-05, "loss": 0.0072, "step": 1820 }, { "epoch": 4.357142857142857, "grad_norm": 0.015460989437997341, "learning_rate": 5e-05, "loss": 0.0063, "step": 1830 }, { "epoch": 4.380952380952381, "grad_norm": 0.028968673199415207, "learning_rate": 5e-05, "loss": 0.0075, "step": 1840 }, { "epoch": 4.404761904761905, "grad_norm": 0.021456323564052582, "learning_rate": 5e-05, "loss": 0.0074, "step": 1850 }, { "epoch": 4.428571428571429, "grad_norm": 0.016484368592500687, "learning_rate": 5e-05, "loss": 0.0069, "step": 1860 }, { "epoch": 4.4523809523809526, "grad_norm": 0.03759278357028961, "learning_rate": 5e-05, "loss": 0.0073, "step": 1870 }, { "epoch": 4.476190476190476, "grad_norm": 0.02179703302681446, "learning_rate": 5e-05, "loss": 0.007, "step": 1880 }, { "epoch": 4.5, "grad_norm": 0.023886868730187416, "learning_rate": 5e-05, "loss": 0.0068, "step": 1890 }, { "epoch": 4.523809523809524, "grad_norm": 0.020535631105303764, "learning_rate": 5e-05, "loss": 0.0063, "step": 1900 }, { "epoch": 4.5476190476190474, "grad_norm": 0.024215754121541977, "learning_rate": 5e-05, "loss": 0.0065, "step": 1910 }, { "epoch": 4.571428571428571, "grad_norm": 0.02726290002465248, "learning_rate": 5e-05, "loss": 0.0068, "step": 1920 }, { "epoch": 4.595238095238095, "grad_norm": 0.022762540727853775, "learning_rate": 5e-05, "loss": 0.0066, "step": 1930 }, { "epoch": 4.619047619047619, "grad_norm": 0.03385463356971741, "learning_rate": 5e-05, "loss": 0.0066, "step": 1940 }, { "epoch": 4.642857142857143, "grad_norm": 0.029403358697891235, "learning_rate": 5e-05, "loss": 0.0065, "step": 1950 }, { "epoch": 4.666666666666667, "grad_norm": 0.03025682643055916, "learning_rate": 5e-05, "loss": 0.0078, "step": 1960 }, { "epoch": 4.690476190476191, "grad_norm": 0.012874845415353775, "learning_rate": 5e-05, "loss": 0.0063, "step": 1970 }, { "epoch": 4.714285714285714, "grad_norm": 0.021130386739969254, "learning_rate": 5e-05, "loss": 0.007, "step": 1980 }, { "epoch": 4.738095238095238, "grad_norm": 0.02132946252822876, "learning_rate": 5e-05, "loss": 0.0063, "step": 1990 }, { "epoch": 4.761904761904762, "grad_norm": 0.03074180707335472, "learning_rate": 5e-05, "loss": 0.0061, "step": 2000 }, { "epoch": 4.785714285714286, "grad_norm": 0.02292218618094921, "learning_rate": 5e-05, "loss": 0.0063, "step": 2010 }, { "epoch": 4.809523809523809, "grad_norm": 0.01976919360458851, "learning_rate": 5e-05, "loss": 0.0061, "step": 2020 }, { "epoch": 4.833333333333333, "grad_norm": 0.030889755114912987, "learning_rate": 5e-05, "loss": 0.0064, "step": 2030 }, { "epoch": 4.857142857142857, "grad_norm": 0.02626785635948181, "learning_rate": 5e-05, "loss": 0.0071, "step": 2040 }, { "epoch": 4.880952380952381, "grad_norm": 0.016065414994955063, "learning_rate": 5e-05, "loss": 0.0067, "step": 2050 }, { "epoch": 4.904761904761905, "grad_norm": 0.022564269602298737, "learning_rate": 5e-05, "loss": 0.0057, "step": 2060 }, { "epoch": 4.928571428571429, "grad_norm": 0.03417288884520531, "learning_rate": 5e-05, "loss": 0.0063, "step": 2070 }, { "epoch": 4.9523809523809526, "grad_norm": 0.02042423002421856, "learning_rate": 5e-05, "loss": 0.0067, "step": 2080 }, { "epoch": 4.976190476190476, "grad_norm": 0.038521867245435715, "learning_rate": 5e-05, "loss": 0.0064, "step": 2090 }, { "epoch": 5.0, "grad_norm": 0.026785830035805702, "learning_rate": 5e-05, "loss": 0.0061, "step": 2100 }, { "epoch": 5.023809523809524, "grad_norm": 0.01977149210870266, "learning_rate": 5e-05, "loss": 0.0067, "step": 2110 }, { "epoch": 5.0476190476190474, "grad_norm": 0.028995908796787262, "learning_rate": 5e-05, "loss": 0.0065, "step": 2120 }, { "epoch": 5.071428571428571, "grad_norm": 0.027958383783698082, "learning_rate": 5e-05, "loss": 0.0067, "step": 2130 }, { "epoch": 5.095238095238095, "grad_norm": 0.024838177487254143, "learning_rate": 5e-05, "loss": 0.0062, "step": 2140 }, { "epoch": 5.119047619047619, "grad_norm": 0.019940726459026337, "learning_rate": 5e-05, "loss": 0.0056, "step": 2150 }, { "epoch": 5.142857142857143, "grad_norm": 0.023204155266284943, "learning_rate": 5e-05, "loss": 0.0055, "step": 2160 }, { "epoch": 5.166666666666667, "grad_norm": 0.02349969372153282, "learning_rate": 5e-05, "loss": 0.006, "step": 2170 }, { "epoch": 5.190476190476191, "grad_norm": 0.030205531045794487, "learning_rate": 5e-05, "loss": 0.0065, "step": 2180 }, { "epoch": 5.214285714285714, "grad_norm": 0.0321480929851532, "learning_rate": 5e-05, "loss": 0.0074, "step": 2190 }, { "epoch": 5.238095238095238, "grad_norm": 0.020824981853365898, "learning_rate": 5e-05, "loss": 0.006, "step": 2200 }, { "epoch": 5.261904761904762, "grad_norm": 0.0227160956710577, "learning_rate": 5e-05, "loss": 0.0063, "step": 2210 }, { "epoch": 5.285714285714286, "grad_norm": 0.029231227934360504, "learning_rate": 5e-05, "loss": 0.005, "step": 2220 }, { "epoch": 5.309523809523809, "grad_norm": 0.02282622829079628, "learning_rate": 5e-05, "loss": 0.0045, "step": 2230 }, { "epoch": 5.333333333333333, "grad_norm": 0.03982138633728027, "learning_rate": 5e-05, "loss": 0.0061, "step": 2240 }, { "epoch": 5.357142857142857, "grad_norm": 0.018466891720891, "learning_rate": 5e-05, "loss": 0.0048, "step": 2250 }, { "epoch": 5.380952380952381, "grad_norm": 0.024992188438773155, "learning_rate": 5e-05, "loss": 0.0063, "step": 2260 }, { "epoch": 5.404761904761905, "grad_norm": 0.029194045811891556, "learning_rate": 5e-05, "loss": 0.0062, "step": 2270 }, { "epoch": 5.428571428571429, "grad_norm": 0.027634933590888977, "learning_rate": 5e-05, "loss": 0.0056, "step": 2280 }, { "epoch": 5.4523809523809526, "grad_norm": 0.03338102251291275, "learning_rate": 5e-05, "loss": 0.0061, "step": 2290 }, { "epoch": 5.476190476190476, "grad_norm": 0.020240414887666702, "learning_rate": 5e-05, "loss": 0.0053, "step": 2300 }, { "epoch": 5.5, "grad_norm": 0.01747560314834118, "learning_rate": 5e-05, "loss": 0.0053, "step": 2310 }, { "epoch": 5.523809523809524, "grad_norm": 0.02589523047208786, "learning_rate": 5e-05, "loss": 0.0052, "step": 2320 }, { "epoch": 5.5476190476190474, "grad_norm": 0.03360776975750923, "learning_rate": 5e-05, "loss": 0.0051, "step": 2330 }, { "epoch": 5.571428571428571, "grad_norm": 0.0385347418487072, "learning_rate": 5e-05, "loss": 0.0053, "step": 2340 }, { "epoch": 5.595238095238095, "grad_norm": 0.02222476154565811, "learning_rate": 5e-05, "loss": 0.0051, "step": 2350 }, { "epoch": 5.619047619047619, "grad_norm": 0.021084798499941826, "learning_rate": 5e-05, "loss": 0.0058, "step": 2360 }, { "epoch": 5.642857142857143, "grad_norm": 0.02974940650165081, "learning_rate": 5e-05, "loss": 0.0052, "step": 2370 }, { "epoch": 5.666666666666667, "grad_norm": 0.01862935908138752, "learning_rate": 5e-05, "loss": 0.0061, "step": 2380 }, { "epoch": 5.690476190476191, "grad_norm": 0.01248252298682928, "learning_rate": 5e-05, "loss": 0.0051, "step": 2390 }, { "epoch": 5.714285714285714, "grad_norm": 0.023803608492016792, "learning_rate": 5e-05, "loss": 0.0054, "step": 2400 }, { "epoch": 5.738095238095238, "grad_norm": 0.01642240583896637, "learning_rate": 5e-05, "loss": 0.0051, "step": 2410 }, { "epoch": 5.761904761904762, "grad_norm": 0.027454275637865067, "learning_rate": 5e-05, "loss": 0.0048, "step": 2420 }, { "epoch": 5.785714285714286, "grad_norm": 0.03252597153186798, "learning_rate": 5e-05, "loss": 0.0051, "step": 2430 }, { "epoch": 5.809523809523809, "grad_norm": 0.031316157430410385, "learning_rate": 5e-05, "loss": 0.0045, "step": 2440 }, { "epoch": 5.833333333333333, "grad_norm": 0.022797273471951485, "learning_rate": 5e-05, "loss": 0.0057, "step": 2450 }, { "epoch": 5.857142857142857, "grad_norm": 0.03269220516085625, "learning_rate": 5e-05, "loss": 0.0061, "step": 2460 }, { "epoch": 5.880952380952381, "grad_norm": 0.02592204324901104, "learning_rate": 5e-05, "loss": 0.0063, "step": 2470 }, { "epoch": 5.904761904761905, "grad_norm": 0.020954346284270287, "learning_rate": 5e-05, "loss": 0.005, "step": 2480 }, { "epoch": 5.928571428571429, "grad_norm": 0.03389604389667511, "learning_rate": 5e-05, "loss": 0.0044, "step": 2490 }, { "epoch": 5.9523809523809526, "grad_norm": 0.022789567708969116, "learning_rate": 5e-05, "loss": 0.005, "step": 2500 }, { "epoch": 5.976190476190476, "grad_norm": 0.021143564954400063, "learning_rate": 5e-05, "loss": 0.0052, "step": 2510 }, { "epoch": 6.0, "grad_norm": 0.01593031920492649, "learning_rate": 5e-05, "loss": 0.0048, "step": 2520 }, { "epoch": 6.023809523809524, "grad_norm": 0.024208232760429382, "learning_rate": 5e-05, "loss": 0.0054, "step": 2530 }, { "epoch": 6.0476190476190474, "grad_norm": 0.020461222156882286, "learning_rate": 5e-05, "loss": 0.0052, "step": 2540 }, { "epoch": 6.071428571428571, "grad_norm": 0.027278559282422066, "learning_rate": 5e-05, "loss": 0.0057, "step": 2550 }, { "epoch": 6.095238095238095, "grad_norm": 0.02183406613767147, "learning_rate": 5e-05, "loss": 0.005, "step": 2560 }, { "epoch": 6.119047619047619, "grad_norm": 0.024204745888710022, "learning_rate": 5e-05, "loss": 0.0047, "step": 2570 }, { "epoch": 6.142857142857143, "grad_norm": 0.020112114027142525, "learning_rate": 5e-05, "loss": 0.0042, "step": 2580 }, { "epoch": 6.166666666666667, "grad_norm": 0.03133062273263931, "learning_rate": 5e-05, "loss": 0.0043, "step": 2590 }, { "epoch": 6.190476190476191, "grad_norm": 0.016416160389780998, "learning_rate": 5e-05, "loss": 0.0051, "step": 2600 }, { "epoch": 6.214285714285714, "grad_norm": 0.02232821099460125, "learning_rate": 5e-05, "loss": 0.0049, "step": 2610 }, { "epoch": 6.238095238095238, "grad_norm": 0.023397671058773994, "learning_rate": 5e-05, "loss": 0.0044, "step": 2620 }, { "epoch": 6.261904761904762, "grad_norm": 0.021031634882092476, "learning_rate": 5e-05, "loss": 0.0048, "step": 2630 }, { "epoch": 6.285714285714286, "grad_norm": 0.05748249217867851, "learning_rate": 5e-05, "loss": 0.0043, "step": 2640 }, { "epoch": 6.309523809523809, "grad_norm": 0.030488435178995132, "learning_rate": 5e-05, "loss": 0.0039, "step": 2650 }, { "epoch": 6.333333333333333, "grad_norm": 0.019017908722162247, "learning_rate": 5e-05, "loss": 0.0041, "step": 2660 }, { "epoch": 6.357142857142857, "grad_norm": 0.021291224285960197, "learning_rate": 5e-05, "loss": 0.0046, "step": 2670 }, { "epoch": 6.380952380952381, "grad_norm": 0.02494419366121292, "learning_rate": 5e-05, "loss": 0.0047, "step": 2680 }, { "epoch": 6.404761904761905, "grad_norm": 0.02740258164703846, "learning_rate": 5e-05, "loss": 0.0051, "step": 2690 }, { "epoch": 6.428571428571429, "grad_norm": 0.02902950905263424, "learning_rate": 5e-05, "loss": 0.0046, "step": 2700 }, { "epoch": 6.4523809523809526, "grad_norm": 0.020637964829802513, "learning_rate": 5e-05, "loss": 0.0044, "step": 2710 }, { "epoch": 6.476190476190476, "grad_norm": 0.020837081596255302, "learning_rate": 5e-05, "loss": 0.0038, "step": 2720 }, { "epoch": 6.5, "grad_norm": 0.020413970574736595, "learning_rate": 5e-05, "loss": 0.0035, "step": 2730 }, { "epoch": 6.523809523809524, "grad_norm": 0.01466563530266285, "learning_rate": 5e-05, "loss": 0.004, "step": 2740 }, { "epoch": 6.5476190476190474, "grad_norm": 0.02238455042243004, "learning_rate": 5e-05, "loss": 0.0035, "step": 2750 }, { "epoch": 6.571428571428571, "grad_norm": 0.027074338868260384, "learning_rate": 5e-05, "loss": 0.0038, "step": 2760 }, { "epoch": 6.595238095238095, "grad_norm": 0.03330984711647034, "learning_rate": 5e-05, "loss": 0.0042, "step": 2770 }, { "epoch": 6.619047619047619, "grad_norm": 0.025337036699056625, "learning_rate": 5e-05, "loss": 0.0044, "step": 2780 }, { "epoch": 6.642857142857143, "grad_norm": 0.034126244485378265, "learning_rate": 5e-05, "loss": 0.0047, "step": 2790 }, { "epoch": 6.666666666666667, "grad_norm": 0.02013178914785385, "learning_rate": 5e-05, "loss": 0.0053, "step": 2800 }, { "epoch": 6.690476190476191, "grad_norm": 0.012780356220901012, "learning_rate": 5e-05, "loss": 0.0041, "step": 2810 }, { "epoch": 6.714285714285714, "grad_norm": 0.01535839308053255, "learning_rate": 5e-05, "loss": 0.0042, "step": 2820 }, { "epoch": 6.738095238095238, "grad_norm": 0.018024824559688568, "learning_rate": 5e-05, "loss": 0.0039, "step": 2830 }, { "epoch": 6.761904761904762, "grad_norm": 0.02146134339272976, "learning_rate": 5e-05, "loss": 0.0041, "step": 2840 }, { "epoch": 6.785714285714286, "grad_norm": 0.015317169018089771, "learning_rate": 5e-05, "loss": 0.0034, "step": 2850 }, { "epoch": 6.809523809523809, "grad_norm": 0.029330266639590263, "learning_rate": 5e-05, "loss": 0.0037, "step": 2860 }, { "epoch": 6.833333333333333, "grad_norm": 0.032067060470581055, "learning_rate": 5e-05, "loss": 0.0044, "step": 2870 }, { "epoch": 6.857142857142857, "grad_norm": 0.016328303143382072, "learning_rate": 5e-05, "loss": 0.0045, "step": 2880 }, { "epoch": 6.880952380952381, "grad_norm": 0.021490612998604774, "learning_rate": 5e-05, "loss": 0.0044, "step": 2890 }, { "epoch": 6.904761904761905, "grad_norm": 0.014762201346457005, "learning_rate": 5e-05, "loss": 0.0037, "step": 2900 }, { "epoch": 6.928571428571429, "grad_norm": 0.02523043379187584, "learning_rate": 5e-05, "loss": 0.0035, "step": 2910 }, { "epoch": 6.9523809523809526, "grad_norm": 0.01821116730570793, "learning_rate": 5e-05, "loss": 0.0037, "step": 2920 }, { "epoch": 6.976190476190476, "grad_norm": 0.025133155286312103, "learning_rate": 5e-05, "loss": 0.0039, "step": 2930 }, { "epoch": 7.0, "grad_norm": 0.01947784051299095, "learning_rate": 5e-05, "loss": 0.004, "step": 2940 }, { "epoch": 7.023809523809524, "grad_norm": 0.019346952438354492, "learning_rate": 5e-05, "loss": 0.0044, "step": 2950 }, { "epoch": 7.0476190476190474, "grad_norm": 0.02163790725171566, "learning_rate": 5e-05, "loss": 0.0038, "step": 2960 }, { "epoch": 7.071428571428571, "grad_norm": 0.015155022032558918, "learning_rate": 5e-05, "loss": 0.0044, "step": 2970 }, { "epoch": 7.095238095238095, "grad_norm": 0.02857312373816967, "learning_rate": 5e-05, "loss": 0.004, "step": 2980 }, { "epoch": 7.119047619047619, "grad_norm": 0.015818042680621147, "learning_rate": 5e-05, "loss": 0.0036, "step": 2990 }, { "epoch": 7.142857142857143, "grad_norm": 0.01129070483148098, "learning_rate": 5e-05, "loss": 0.0028, "step": 3000 }, { "epoch": 7.166666666666667, "grad_norm": 0.040307771414518356, "learning_rate": 5e-05, "loss": 0.0037, "step": 3010 }, { "epoch": 7.190476190476191, "grad_norm": 0.02898922935128212, "learning_rate": 5e-05, "loss": 0.0036, "step": 3020 }, { "epoch": 7.214285714285714, "grad_norm": 0.022514350712299347, "learning_rate": 5e-05, "loss": 0.0037, "step": 3030 }, { "epoch": 7.238095238095238, "grad_norm": 0.013335715979337692, "learning_rate": 5e-05, "loss": 0.0035, "step": 3040 }, { "epoch": 7.261904761904762, "grad_norm": 0.031238842755556107, "learning_rate": 5e-05, "loss": 0.0039, "step": 3050 }, { "epoch": 7.285714285714286, "grad_norm": 0.025638382881879807, "learning_rate": 5e-05, "loss": 0.0028, "step": 3060 }, { "epoch": 7.309523809523809, "grad_norm": 0.046518560498952866, "learning_rate": 5e-05, "loss": 0.003, "step": 3070 }, { "epoch": 7.333333333333333, "grad_norm": 0.024228127673268318, "learning_rate": 5e-05, "loss": 0.0038, "step": 3080 }, { "epoch": 7.357142857142857, "grad_norm": 0.02967258170247078, "learning_rate": 5e-05, "loss": 0.003, "step": 3090 }, { "epoch": 7.380952380952381, "grad_norm": 0.019939076155424118, "learning_rate": 5e-05, "loss": 0.004, "step": 3100 }, { "epoch": 7.404761904761905, "grad_norm": 0.02764073945581913, "learning_rate": 5e-05, "loss": 0.0037, "step": 3110 }, { "epoch": 7.428571428571429, "grad_norm": 0.02370736561715603, "learning_rate": 5e-05, "loss": 0.0033, "step": 3120 }, { "epoch": 7.4523809523809526, "grad_norm": 0.020530644804239273, "learning_rate": 5e-05, "loss": 0.0033, "step": 3130 }, { "epoch": 7.476190476190476, "grad_norm": 0.019800107926130295, "learning_rate": 5e-05, "loss": 0.0033, "step": 3140 }, { "epoch": 7.5, "grad_norm": 0.018785228952765465, "learning_rate": 5e-05, "loss": 0.0029, "step": 3150 }, { "epoch": 7.523809523809524, "grad_norm": 0.02633030340075493, "learning_rate": 5e-05, "loss": 0.0033, "step": 3160 }, { "epoch": 7.5476190476190474, "grad_norm": 0.08342000097036362, "learning_rate": 5e-05, "loss": 0.0031, "step": 3170 }, { "epoch": 7.571428571428571, "grad_norm": 0.008880950510501862, "learning_rate": 5e-05, "loss": 0.0026, "step": 3180 }, { "epoch": 7.595238095238095, "grad_norm": 0.02733513154089451, "learning_rate": 5e-05, "loss": 0.0028, "step": 3190 }, { "epoch": 7.619047619047619, "grad_norm": 0.03516070544719696, "learning_rate": 5e-05, "loss": 0.0032, "step": 3200 }, { "epoch": 7.642857142857143, "grad_norm": 0.03665764257311821, "learning_rate": 5e-05, "loss": 0.0035, "step": 3210 }, { "epoch": 7.666666666666667, "grad_norm": 0.03748982027173042, "learning_rate": 5e-05, "loss": 0.0039, "step": 3220 }, { "epoch": 7.690476190476191, "grad_norm": 0.014199526980519295, "learning_rate": 5e-05, "loss": 0.0036, "step": 3230 }, { "epoch": 7.714285714285714, "grad_norm": 0.01877681352198124, "learning_rate": 5e-05, "loss": 0.0036, "step": 3240 }, { "epoch": 7.738095238095238, "grad_norm": 0.013927435502409935, "learning_rate": 5e-05, "loss": 0.0033, "step": 3250 }, { "epoch": 7.761904761904762, "grad_norm": 0.012296963483095169, "learning_rate": 5e-05, "loss": 0.003, "step": 3260 }, { "epoch": 7.785714285714286, "grad_norm": 0.022298045456409454, "learning_rate": 5e-05, "loss": 0.0028, "step": 3270 }, { "epoch": 7.809523809523809, "grad_norm": 0.04318477213382721, "learning_rate": 5e-05, "loss": 0.0023, "step": 3280 }, { "epoch": 7.833333333333333, "grad_norm": 0.03864269703626633, "learning_rate": 5e-05, "loss": 0.0033, "step": 3290 }, { "epoch": 7.857142857142857, "grad_norm": 0.02180645242333412, "learning_rate": 5e-05, "loss": 0.0038, "step": 3300 }, { "epoch": 7.880952380952381, "grad_norm": 0.03274507075548172, "learning_rate": 5e-05, "loss": 0.0039, "step": 3310 }, { "epoch": 7.904761904761905, "grad_norm": 0.015056164003908634, "learning_rate": 5e-05, "loss": 0.0028, "step": 3320 }, { "epoch": 7.928571428571429, "grad_norm": 0.024232054129242897, "learning_rate": 5e-05, "loss": 0.003, "step": 3330 }, { "epoch": 7.9523809523809526, "grad_norm": 0.01742437854409218, "learning_rate": 5e-05, "loss": 0.0031, "step": 3340 }, { "epoch": 7.976190476190476, "grad_norm": 0.024546030908823013, "learning_rate": 5e-05, "loss": 0.0031, "step": 3350 }, { "epoch": 8.0, "grad_norm": 0.020465753972530365, "learning_rate": 5e-05, "loss": 0.0037, "step": 3360 }, { "epoch": 8.023809523809524, "grad_norm": 0.02927527017891407, "learning_rate": 5e-05, "loss": 0.0041, "step": 3370 }, { "epoch": 8.047619047619047, "grad_norm": 0.01250431314110756, "learning_rate": 5e-05, "loss": 0.0032, "step": 3380 }, { "epoch": 8.071428571428571, "grad_norm": 0.014103593304753304, "learning_rate": 5e-05, "loss": 0.0028, "step": 3390 }, { "epoch": 8.095238095238095, "grad_norm": 0.018405349925160408, "learning_rate": 5e-05, "loss": 0.0033, "step": 3400 }, { "epoch": 8.119047619047619, "grad_norm": 0.020473845303058624, "learning_rate": 5e-05, "loss": 0.0029, "step": 3410 }, { "epoch": 8.142857142857142, "grad_norm": 0.035039015114307404, "learning_rate": 5e-05, "loss": 0.0025, "step": 3420 }, { "epoch": 8.166666666666666, "grad_norm": 0.032077498733997345, "learning_rate": 5e-05, "loss": 0.0028, "step": 3430 }, { "epoch": 8.19047619047619, "grad_norm": 0.037181805819272995, "learning_rate": 5e-05, "loss": 0.003, "step": 3440 }, { "epoch": 8.214285714285714, "grad_norm": 0.018256695941090584, "learning_rate": 5e-05, "loss": 0.0029, "step": 3450 }, { "epoch": 8.238095238095237, "grad_norm": 0.01342601515352726, "learning_rate": 5e-05, "loss": 0.0034, "step": 3460 }, { "epoch": 8.261904761904763, "grad_norm": 0.03634987398982048, "learning_rate": 5e-05, "loss": 0.0032, "step": 3470 }, { "epoch": 8.285714285714286, "grad_norm": 0.050673890858888626, "learning_rate": 5e-05, "loss": 0.0021, "step": 3480 }, { "epoch": 8.30952380952381, "grad_norm": 0.013780270703136921, "learning_rate": 5e-05, "loss": 0.0019, "step": 3490 }, { "epoch": 8.333333333333334, "grad_norm": 0.028823478147387505, "learning_rate": 5e-05, "loss": 0.0027, "step": 3500 }, { "epoch": 8.357142857142858, "grad_norm": 0.024434154853224754, "learning_rate": 5e-05, "loss": 0.0031, "step": 3510 }, { "epoch": 8.380952380952381, "grad_norm": 0.025538841262459755, "learning_rate": 5e-05, "loss": 0.0024, "step": 3520 }, { "epoch": 8.404761904761905, "grad_norm": 0.02040352113544941, "learning_rate": 5e-05, "loss": 0.0029, "step": 3530 }, { "epoch": 8.428571428571429, "grad_norm": 0.031944241374731064, "learning_rate": 5e-05, "loss": 0.002, "step": 3540 }, { "epoch": 8.452380952380953, "grad_norm": 0.017712410539388657, "learning_rate": 5e-05, "loss": 0.0025, "step": 3550 }, { "epoch": 8.476190476190476, "grad_norm": 0.030166227370500565, "learning_rate": 5e-05, "loss": 0.0023, "step": 3560 }, { "epoch": 8.5, "grad_norm": 0.027017861604690552, "learning_rate": 5e-05, "loss": 0.0027, "step": 3570 }, { "epoch": 8.523809523809524, "grad_norm": 0.025174567475914955, "learning_rate": 5e-05, "loss": 0.0028, "step": 3580 }, { "epoch": 8.547619047619047, "grad_norm": 0.02370358444750309, "learning_rate": 5e-05, "loss": 0.0029, "step": 3590 }, { "epoch": 8.571428571428571, "grad_norm": 0.019355561584234238, "learning_rate": 5e-05, "loss": 0.0029, "step": 3600 }, { "epoch": 8.595238095238095, "grad_norm": 0.07427547127008438, "learning_rate": 5e-05, "loss": 0.0027, "step": 3610 }, { "epoch": 8.619047619047619, "grad_norm": 0.011759583838284016, "learning_rate": 5e-05, "loss": 0.0021, "step": 3620 }, { "epoch": 8.642857142857142, "grad_norm": 0.02567645162343979, "learning_rate": 5e-05, "loss": 0.0031, "step": 3630 }, { "epoch": 8.666666666666666, "grad_norm": 0.02148609049618244, "learning_rate": 5e-05, "loss": 0.0044, "step": 3640 }, { "epoch": 8.69047619047619, "grad_norm": 0.021046146750450134, "learning_rate": 5e-05, "loss": 0.0036, "step": 3650 }, { "epoch": 8.714285714285714, "grad_norm": 0.022530779242515564, "learning_rate": 5e-05, "loss": 0.0041, "step": 3660 }, { "epoch": 8.738095238095237, "grad_norm": 0.008970026858150959, "learning_rate": 5e-05, "loss": 0.0034, "step": 3670 }, { "epoch": 8.761904761904763, "grad_norm": 0.018136192113161087, "learning_rate": 5e-05, "loss": 0.0033, "step": 3680 }, { "epoch": 8.785714285714286, "grad_norm": 0.015567361377179623, "learning_rate": 5e-05, "loss": 0.0019, "step": 3690 }, { "epoch": 8.80952380952381, "grad_norm": 0.0197913758456707, "learning_rate": 5e-05, "loss": 0.0017, "step": 3700 }, { "epoch": 8.833333333333334, "grad_norm": 0.03988071531057358, "learning_rate": 5e-05, "loss": 0.0023, "step": 3710 }, { "epoch": 8.857142857142858, "grad_norm": 0.03059978410601616, "learning_rate": 5e-05, "loss": 0.0034, "step": 3720 }, { "epoch": 8.880952380952381, "grad_norm": 0.014758502133190632, "learning_rate": 5e-05, "loss": 0.0031, "step": 3730 }, { "epoch": 8.904761904761905, "grad_norm": 0.017123445868492126, "learning_rate": 5e-05, "loss": 0.0033, "step": 3740 }, { "epoch": 8.928571428571429, "grad_norm": 0.02551579289138317, "learning_rate": 5e-05, "loss": 0.0028, "step": 3750 }, { "epoch": 8.952380952380953, "grad_norm": 0.014496712014079094, "learning_rate": 5e-05, "loss": 0.0031, "step": 3760 }, { "epoch": 8.976190476190476, "grad_norm": 0.02045934647321701, "learning_rate": 5e-05, "loss": 0.0027, "step": 3770 }, { "epoch": 9.0, "grad_norm": 0.02440367080271244, "learning_rate": 5e-05, "loss": 0.0035, "step": 3780 }, { "epoch": 9.023809523809524, "grad_norm": 0.01835225522518158, "learning_rate": 5e-05, "loss": 0.0036, "step": 3790 }, { "epoch": 9.047619047619047, "grad_norm": 0.007453398313373327, "learning_rate": 5e-05, "loss": 0.0027, "step": 3800 }, { "epoch": 9.071428571428571, "grad_norm": 0.02707696706056595, "learning_rate": 5e-05, "loss": 0.0027, "step": 3810 }, { "epoch": 9.095238095238095, "grad_norm": 0.020089933648705482, "learning_rate": 5e-05, "loss": 0.0022, "step": 3820 }, { "epoch": 9.119047619047619, "grad_norm": 0.02362995594739914, "learning_rate": 5e-05, "loss": 0.002, "step": 3830 }, { "epoch": 9.142857142857142, "grad_norm": 0.00633473414927721, "learning_rate": 5e-05, "loss": 0.0019, "step": 3840 }, { "epoch": 9.166666666666666, "grad_norm": 0.013043406419456005, "learning_rate": 5e-05, "loss": 0.0017, "step": 3850 }, { "epoch": 9.19047619047619, "grad_norm": 0.010314088314771652, "learning_rate": 5e-05, "loss": 0.0025, "step": 3860 }, { "epoch": 9.214285714285714, "grad_norm": 0.024023326113820076, "learning_rate": 5e-05, "loss": 0.0025, "step": 3870 }, { "epoch": 9.238095238095237, "grad_norm": 0.033713992685079575, "learning_rate": 5e-05, "loss": 0.0019, "step": 3880 }, { "epoch": 9.261904761904763, "grad_norm": 0.024672770872712135, "learning_rate": 5e-05, "loss": 0.0028, "step": 3890 }, { "epoch": 9.285714285714286, "grad_norm": 0.017616739496588707, "learning_rate": 5e-05, "loss": 0.002, "step": 3900 }, { "epoch": 9.30952380952381, "grad_norm": 0.023528527468442917, "learning_rate": 5e-05, "loss": 0.0021, "step": 3910 }, { "epoch": 9.333333333333334, "grad_norm": 0.04894119128584862, "learning_rate": 5e-05, "loss": 0.0026, "step": 3920 }, { "epoch": 9.357142857142858, "grad_norm": 0.019267156720161438, "learning_rate": 5e-05, "loss": 0.0025, "step": 3930 }, { "epoch": 9.380952380952381, "grad_norm": 0.020611826330423355, "learning_rate": 5e-05, "loss": 0.0026, "step": 3940 }, { "epoch": 9.404761904761905, "grad_norm": 0.020003391429781914, "learning_rate": 5e-05, "loss": 0.002, "step": 3950 }, { "epoch": 9.428571428571429, "grad_norm": 0.017253611236810684, "learning_rate": 5e-05, "loss": 0.0015, "step": 3960 }, { "epoch": 9.452380952380953, "grad_norm": 0.045779161155223846, "learning_rate": 5e-05, "loss": 0.0019, "step": 3970 }, { "epoch": 9.476190476190476, "grad_norm": 0.024241678416728973, "learning_rate": 5e-05, "loss": 0.002, "step": 3980 }, { "epoch": 9.5, "grad_norm": 0.017647111788392067, "learning_rate": 5e-05, "loss": 0.0015, "step": 3990 }, { "epoch": 9.523809523809524, "grad_norm": 0.028641676530241966, "learning_rate": 5e-05, "loss": 0.0019, "step": 4000 }, { "epoch": 9.547619047619047, "grad_norm": 0.047451332211494446, "learning_rate": 5e-05, "loss": 0.0016, "step": 4010 }, { "epoch": 9.571428571428571, "grad_norm": 0.01619701273739338, "learning_rate": 5e-05, "loss": 0.0018, "step": 4020 }, { "epoch": 9.595238095238095, "grad_norm": 0.020237745717167854, "learning_rate": 5e-05, "loss": 0.0018, "step": 4030 }, { "epoch": 9.619047619047619, "grad_norm": 0.040860798209905624, "learning_rate": 5e-05, "loss": 0.0012, "step": 4040 }, { "epoch": 9.642857142857142, "grad_norm": 0.01412968896329403, "learning_rate": 5e-05, "loss": 0.0017, "step": 4050 }, { "epoch": 9.666666666666666, "grad_norm": 0.05102560669183731, "learning_rate": 5e-05, "loss": 0.0019, "step": 4060 }, { "epoch": 9.69047619047619, "grad_norm": 0.04887241870164871, "learning_rate": 5e-05, "loss": 0.0029, "step": 4070 }, { "epoch": 9.714285714285714, "grad_norm": 0.014414262026548386, "learning_rate": 5e-05, "loss": 0.0026, "step": 4080 }, { "epoch": 9.738095238095237, "grad_norm": 0.011788238771259785, "learning_rate": 5e-05, "loss": 0.0024, "step": 4090 }, { "epoch": 9.761904761904763, "grad_norm": 0.01252928376197815, "learning_rate": 5e-05, "loss": 0.0023, "step": 4100 }, { "epoch": 9.785714285714286, "grad_norm": 0.02384638413786888, "learning_rate": 5e-05, "loss": 0.0019, "step": 4110 }, { "epoch": 9.80952380952381, "grad_norm": 0.02536243200302124, "learning_rate": 5e-05, "loss": 0.0014, "step": 4120 }, { "epoch": 9.833333333333334, "grad_norm": 0.015807705000042915, "learning_rate": 5e-05, "loss": 0.0012, "step": 4130 }, { "epoch": 9.857142857142858, "grad_norm": 0.01973266527056694, "learning_rate": 5e-05, "loss": 0.0018, "step": 4140 }, { "epoch": 9.880952380952381, "grad_norm": 0.030758731067180634, "learning_rate": 5e-05, "loss": 0.0025, "step": 4150 }, { "epoch": 9.904761904761905, "grad_norm": 0.01584693230688572, "learning_rate": 5e-05, "loss": 0.0024, "step": 4160 }, { "epoch": 9.928571428571429, "grad_norm": 0.014015723019838333, "learning_rate": 5e-05, "loss": 0.002, "step": 4170 }, { "epoch": 9.952380952380953, "grad_norm": 0.03584891930222511, "learning_rate": 5e-05, "loss": 0.0025, "step": 4180 }, { "epoch": 9.976190476190476, "grad_norm": 0.047324661165475845, "learning_rate": 5e-05, "loss": 0.0028, "step": 4190 }, { "epoch": 10.0, "grad_norm": 0.06452364474534988, "learning_rate": 5e-05, "loss": 0.0023, "step": 4200 }, { "epoch": 10.023809523809524, "grad_norm": 0.044528111815452576, "learning_rate": 5e-05, "loss": 0.0025, "step": 4210 }, { "epoch": 10.047619047619047, "grad_norm": 0.02596374787390232, "learning_rate": 5e-05, "loss": 0.0023, "step": 4220 }, { "epoch": 10.071428571428571, "grad_norm": 0.01841367408633232, "learning_rate": 5e-05, "loss": 0.002, "step": 4230 }, { "epoch": 10.095238095238095, "grad_norm": 0.03294382989406586, "learning_rate": 5e-05, "loss": 0.0017, "step": 4240 }, { "epoch": 10.119047619047619, "grad_norm": 0.026590291410684586, "learning_rate": 5e-05, "loss": 0.0021, "step": 4250 }, { "epoch": 10.142857142857142, "grad_norm": 0.019535573199391365, "learning_rate": 5e-05, "loss": 0.0012, "step": 4260 }, { "epoch": 10.166666666666666, "grad_norm": 0.04901567101478577, "learning_rate": 5e-05, "loss": 0.0018, "step": 4270 }, { "epoch": 10.19047619047619, "grad_norm": 0.023335624486207962, "learning_rate": 5e-05, "loss": 0.002, "step": 4280 }, { "epoch": 10.214285714285714, "grad_norm": 0.02412569336593151, "learning_rate": 5e-05, "loss": 0.0021, "step": 4290 }, { "epoch": 10.238095238095237, "grad_norm": 0.015281541272997856, "learning_rate": 5e-05, "loss": 0.0018, "step": 4300 }, { "epoch": 10.261904761904763, "grad_norm": 0.037770889699459076, "learning_rate": 5e-05, "loss": 0.0017, "step": 4310 }, { "epoch": 10.285714285714286, "grad_norm": 0.02363361231982708, "learning_rate": 5e-05, "loss": 0.0017, "step": 4320 }, { "epoch": 10.30952380952381, "grad_norm": 0.019674105569720268, "learning_rate": 5e-05, "loss": 0.0015, "step": 4330 }, { "epoch": 10.333333333333334, "grad_norm": 0.02023392915725708, "learning_rate": 5e-05, "loss": 0.0032, "step": 4340 }, { "epoch": 10.357142857142858, "grad_norm": 0.01189426425844431, "learning_rate": 5e-05, "loss": 0.002, "step": 4350 }, { "epoch": 10.380952380952381, "grad_norm": 0.041460879147052765, "learning_rate": 5e-05, "loss": 0.0024, "step": 4360 }, { "epoch": 10.404761904761905, "grad_norm": 0.0068189771845936775, "learning_rate": 5e-05, "loss": 0.0014, "step": 4370 }, { "epoch": 10.428571428571429, "grad_norm": 0.019572466611862183, "learning_rate": 5e-05, "loss": 0.001, "step": 4380 }, { "epoch": 10.452380952380953, "grad_norm": 0.01345143560320139, "learning_rate": 5e-05, "loss": 0.0012, "step": 4390 }, { "epoch": 10.476190476190476, "grad_norm": 0.009295144118368626, "learning_rate": 5e-05, "loss": 0.0011, "step": 4400 }, { "epoch": 10.5, "grad_norm": 0.02156994678080082, "learning_rate": 5e-05, "loss": 0.0014, "step": 4410 }, { "epoch": 10.523809523809524, "grad_norm": 0.020932510495185852, "learning_rate": 5e-05, "loss": 0.0018, "step": 4420 }, { "epoch": 10.547619047619047, "grad_norm": 0.01585063897073269, "learning_rate": 5e-05, "loss": 0.0012, "step": 4430 }, { "epoch": 10.571428571428571, "grad_norm": 0.009516189806163311, "learning_rate": 5e-05, "loss": 0.0009, "step": 4440 }, { "epoch": 10.595238095238095, "grad_norm": 0.026152120903134346, "learning_rate": 5e-05, "loss": 0.0014, "step": 4450 }, { "epoch": 10.619047619047619, "grad_norm": 0.0138620026409626, "learning_rate": 5e-05, "loss": 0.0014, "step": 4460 }, { "epoch": 10.642857142857142, "grad_norm": 0.03089461848139763, "learning_rate": 5e-05, "loss": 0.0015, "step": 4470 }, { "epoch": 10.666666666666666, "grad_norm": 0.03294903039932251, "learning_rate": 5e-05, "loss": 0.0019, "step": 4480 }, { "epoch": 10.69047619047619, "grad_norm": 0.00928263645619154, "learning_rate": 5e-05, "loss": 0.002, "step": 4490 }, { "epoch": 10.714285714285714, "grad_norm": 0.029545562341809273, "learning_rate": 5e-05, "loss": 0.0017, "step": 4500 }, { "epoch": 10.738095238095237, "grad_norm": 0.009242654778063297, "learning_rate": 5e-05, "loss": 0.0018, "step": 4510 }, { "epoch": 10.761904761904763, "grad_norm": 0.014272899366915226, "learning_rate": 5e-05, "loss": 0.0019, "step": 4520 }, { "epoch": 10.785714285714286, "grad_norm": 0.02279002219438553, "learning_rate": 5e-05, "loss": 0.0013, "step": 4530 }, { "epoch": 10.80952380952381, "grad_norm": 0.03252875804901123, "learning_rate": 5e-05, "loss": 0.0015, "step": 4540 }, { "epoch": 10.833333333333334, "grad_norm": 0.009702946059405804, "learning_rate": 5e-05, "loss": 0.0011, "step": 4550 }, { "epoch": 10.857142857142858, "grad_norm": 0.03491836413741112, "learning_rate": 5e-05, "loss": 0.0014, "step": 4560 }, { "epoch": 10.880952380952381, "grad_norm": 0.03947118669748306, "learning_rate": 5e-05, "loss": 0.0024, "step": 4570 }, { "epoch": 10.904761904761905, "grad_norm": 0.025778034701943398, "learning_rate": 5e-05, "loss": 0.0019, "step": 4580 }, { "epoch": 10.928571428571429, "grad_norm": 0.02018180675804615, "learning_rate": 5e-05, "loss": 0.0015, "step": 4590 }, { "epoch": 10.952380952380953, "grad_norm": 0.005483316257596016, "learning_rate": 5e-05, "loss": 0.0014, "step": 4600 }, { "epoch": 10.976190476190476, "grad_norm": 0.017275342717766762, "learning_rate": 5e-05, "loss": 0.0015, "step": 4610 }, { "epoch": 11.0, "grad_norm": 0.019064828753471375, "learning_rate": 5e-05, "loss": 0.0022, "step": 4620 }, { "epoch": 11.023809523809524, "grad_norm": 0.013529101386666298, "learning_rate": 5e-05, "loss": 0.0018, "step": 4630 }, { "epoch": 11.047619047619047, "grad_norm": 0.01693836972117424, "learning_rate": 5e-05, "loss": 0.0022, "step": 4640 }, { "epoch": 11.071428571428571, "grad_norm": 0.02852066047489643, "learning_rate": 5e-05, "loss": 0.0018, "step": 4650 }, { "epoch": 11.095238095238095, "grad_norm": 0.018933508545160294, "learning_rate": 5e-05, "loss": 0.0018, "step": 4660 }, { "epoch": 11.119047619047619, "grad_norm": 0.021359335631132126, "learning_rate": 5e-05, "loss": 0.0015, "step": 4670 }, { "epoch": 11.142857142857142, "grad_norm": 0.019422901794314384, "learning_rate": 5e-05, "loss": 0.0014, "step": 4680 }, { "epoch": 11.166666666666666, "grad_norm": 0.013169987127184868, "learning_rate": 5e-05, "loss": 0.0015, "step": 4690 }, { "epoch": 11.19047619047619, "grad_norm": 0.03702198714017868, "learning_rate": 5e-05, "loss": 0.0018, "step": 4700 }, { "epoch": 11.214285714285714, "grad_norm": 0.030341317877173424, "learning_rate": 5e-05, "loss": 0.0015, "step": 4710 }, { "epoch": 11.238095238095237, "grad_norm": 0.020110273733735085, "learning_rate": 5e-05, "loss": 0.0014, "step": 4720 }, { "epoch": 11.261904761904763, "grad_norm": 0.02008882723748684, "learning_rate": 5e-05, "loss": 0.0013, "step": 4730 }, { "epoch": 11.285714285714286, "grad_norm": 0.06150029972195625, "learning_rate": 5e-05, "loss": 0.0013, "step": 4740 }, { "epoch": 11.30952380952381, "grad_norm": 0.0267238300293684, "learning_rate": 5e-05, "loss": 0.0012, "step": 4750 }, { "epoch": 11.333333333333334, "grad_norm": 0.008372998796403408, "learning_rate": 5e-05, "loss": 0.0012, "step": 4760 }, { "epoch": 11.357142857142858, "grad_norm": 0.03507944568991661, "learning_rate": 5e-05, "loss": 0.0015, "step": 4770 }, { "epoch": 11.380952380952381, "grad_norm": 0.018310125917196274, "learning_rate": 5e-05, "loss": 0.0017, "step": 4780 }, { "epoch": 11.404761904761905, "grad_norm": 0.009987925179302692, "learning_rate": 5e-05, "loss": 0.0011, "step": 4790 }, { "epoch": 11.428571428571429, "grad_norm": 0.0077149407006800175, "learning_rate": 5e-05, "loss": 0.0014, "step": 4800 }, { "epoch": 11.452380952380953, "grad_norm": 0.013563682325184345, "learning_rate": 5e-05, "loss": 0.0011, "step": 4810 }, { "epoch": 11.476190476190476, "grad_norm": 0.026644017547369003, "learning_rate": 5e-05, "loss": 0.0008, "step": 4820 }, { "epoch": 11.5, "grad_norm": 0.00553143722936511, "learning_rate": 5e-05, "loss": 0.0005, "step": 4830 }, { "epoch": 11.523809523809524, "grad_norm": 0.014055385254323483, "learning_rate": 5e-05, "loss": 0.0009, "step": 4840 }, { "epoch": 11.547619047619047, "grad_norm": 0.02219008095562458, "learning_rate": 5e-05, "loss": 0.0011, "step": 4850 }, { "epoch": 11.571428571428571, "grad_norm": 0.021185655146837234, "learning_rate": 5e-05, "loss": 0.0011, "step": 4860 }, { "epoch": 11.595238095238095, "grad_norm": 0.012521096505224705, "learning_rate": 5e-05, "loss": 0.0014, "step": 4870 }, { "epoch": 11.619047619047619, "grad_norm": 0.027117032557725906, "learning_rate": 5e-05, "loss": 0.0021, "step": 4880 }, { "epoch": 11.642857142857142, "grad_norm": 0.03251611068844795, "learning_rate": 5e-05, "loss": 0.0019, "step": 4890 }, { "epoch": 11.666666666666666, "grad_norm": 0.014405048452317715, "learning_rate": 5e-05, "loss": 0.001, "step": 4900 }, { "epoch": 11.69047619047619, "grad_norm": 0.037754759192466736, "learning_rate": 5e-05, "loss": 0.0013, "step": 4910 }, { "epoch": 11.714285714285714, "grad_norm": 0.008099643513560295, "learning_rate": 5e-05, "loss": 0.0013, "step": 4920 }, { "epoch": 11.738095238095237, "grad_norm": 0.01963508315384388, "learning_rate": 5e-05, "loss": 0.0015, "step": 4930 }, { "epoch": 11.761904761904763, "grad_norm": 0.017585638910531998, "learning_rate": 5e-05, "loss": 0.0015, "step": 4940 }, { "epoch": 11.785714285714286, "grad_norm": 0.030090732499957085, "learning_rate": 5e-05, "loss": 0.0011, "step": 4950 }, { "epoch": 11.80952380952381, "grad_norm": 0.018853867426514626, "learning_rate": 5e-05, "loss": 0.001, "step": 4960 }, { "epoch": 11.833333333333334, "grad_norm": 0.009620056487619877, "learning_rate": 5e-05, "loss": 0.0011, "step": 4970 }, { "epoch": 11.857142857142858, "grad_norm": 0.05379125103354454, "learning_rate": 5e-05, "loss": 0.0016, "step": 4980 }, { "epoch": 11.880952380952381, "grad_norm": 0.022303573787212372, "learning_rate": 5e-05, "loss": 0.0019, "step": 4990 }, { "epoch": 11.904761904761905, "grad_norm": 0.008755918592214584, "learning_rate": 5e-05, "loss": 0.0012, "step": 5000 }, { "epoch": 11.928571428571429, "grad_norm": 0.009491810575127602, "learning_rate": 5e-05, "loss": 0.0015, "step": 5010 }, { "epoch": 11.952380952380953, "grad_norm": 0.02040296606719494, "learning_rate": 5e-05, "loss": 0.0014, "step": 5020 }, { "epoch": 11.976190476190476, "grad_norm": 0.012556660920381546, "learning_rate": 5e-05, "loss": 0.0012, "step": 5030 }, { "epoch": 12.0, "grad_norm": 0.008164563216269016, "learning_rate": 5e-05, "loss": 0.0008, "step": 5040 }, { "epoch": 12.023809523809524, "grad_norm": 0.02336413785815239, "learning_rate": 5e-05, "loss": 0.0013, "step": 5050 }, { "epoch": 12.047619047619047, "grad_norm": 0.0061961328610777855, "learning_rate": 5e-05, "loss": 0.0009, "step": 5060 }, { "epoch": 12.071428571428571, "grad_norm": 0.012313502840697765, "learning_rate": 5e-05, "loss": 0.0011, "step": 5070 }, { "epoch": 12.095238095238095, "grad_norm": 0.010624333284795284, "learning_rate": 5e-05, "loss": 0.0009, "step": 5080 }, { "epoch": 12.119047619047619, "grad_norm": 0.01863931119441986, "learning_rate": 5e-05, "loss": 0.0012, "step": 5090 }, { "epoch": 12.142857142857142, "grad_norm": 0.01563103497028351, "learning_rate": 5e-05, "loss": 0.0015, "step": 5100 }, { "epoch": 12.166666666666666, "grad_norm": 0.014355426654219627, "learning_rate": 5e-05, "loss": 0.001, "step": 5110 }, { "epoch": 12.19047619047619, "grad_norm": 0.011460881680250168, "learning_rate": 5e-05, "loss": 0.0012, "step": 5120 }, { "epoch": 12.214285714285714, "grad_norm": 0.01539954636245966, "learning_rate": 5e-05, "loss": 0.0012, "step": 5130 }, { "epoch": 12.238095238095237, "grad_norm": 0.022988399490714073, "learning_rate": 5e-05, "loss": 0.0014, "step": 5140 }, { "epoch": 12.261904761904763, "grad_norm": 0.04906224086880684, "learning_rate": 5e-05, "loss": 0.0011, "step": 5150 }, { "epoch": 12.285714285714286, "grad_norm": 0.020332837477326393, "learning_rate": 5e-05, "loss": 0.0008, "step": 5160 }, { "epoch": 12.30952380952381, "grad_norm": 0.008848396129906178, "learning_rate": 5e-05, "loss": 0.0005, "step": 5170 }, { "epoch": 12.333333333333334, "grad_norm": 0.016721338033676147, "learning_rate": 5e-05, "loss": 0.0015, "step": 5180 }, { "epoch": 12.357142857142858, "grad_norm": 0.024787932634353638, "learning_rate": 5e-05, "loss": 0.0011, "step": 5190 }, { "epoch": 12.380952380952381, "grad_norm": 0.016609011217951775, "learning_rate": 5e-05, "loss": 0.0008, "step": 5200 }, { "epoch": 12.404761904761905, "grad_norm": 0.02383277751505375, "learning_rate": 5e-05, "loss": 0.001, "step": 5210 }, { "epoch": 12.428571428571429, "grad_norm": 0.009788766503334045, "learning_rate": 5e-05, "loss": 0.0009, "step": 5220 }, { "epoch": 12.452380952380953, "grad_norm": 0.008273842744529247, "learning_rate": 5e-05, "loss": 0.0007, "step": 5230 }, { "epoch": 12.476190476190476, "grad_norm": 0.0200930405408144, "learning_rate": 5e-05, "loss": 0.0006, "step": 5240 }, { "epoch": 12.5, "grad_norm": 0.004213349893689156, "learning_rate": 5e-05, "loss": 0.0007, "step": 5250 }, { "epoch": 12.523809523809524, "grad_norm": 0.010746212676167488, "learning_rate": 5e-05, "loss": 0.0008, "step": 5260 }, { "epoch": 12.547619047619047, "grad_norm": 0.009187485091388226, "learning_rate": 5e-05, "loss": 0.0015, "step": 5270 }, { "epoch": 12.571428571428571, "grad_norm": 0.011904279701411724, "learning_rate": 5e-05, "loss": 0.0006, "step": 5280 }, { "epoch": 12.595238095238095, "grad_norm": 0.011605814099311829, "learning_rate": 5e-05, "loss": 0.001, "step": 5290 }, { "epoch": 12.619047619047619, "grad_norm": 0.0053769913502037525, "learning_rate": 5e-05, "loss": 0.0014, "step": 5300 }, { "epoch": 12.642857142857142, "grad_norm": 0.028224363923072815, "learning_rate": 5e-05, "loss": 0.0015, "step": 5310 }, { "epoch": 12.666666666666666, "grad_norm": 0.009584328159689903, "learning_rate": 5e-05, "loss": 0.0007, "step": 5320 }, { "epoch": 12.69047619047619, "grad_norm": 0.012837843038141727, "learning_rate": 5e-05, "loss": 0.0008, "step": 5330 }, { "epoch": 12.714285714285714, "grad_norm": 0.013618317432701588, "learning_rate": 5e-05, "loss": 0.0012, "step": 5340 }, { "epoch": 12.738095238095237, "grad_norm": 0.010568618774414062, "learning_rate": 5e-05, "loss": 0.0012, "step": 5350 }, { "epoch": 12.761904761904763, "grad_norm": 0.015289129689335823, "learning_rate": 5e-05, "loss": 0.0014, "step": 5360 }, { "epoch": 12.785714285714286, "grad_norm": 0.015249863266944885, "learning_rate": 5e-05, "loss": 0.0009, "step": 5370 }, { "epoch": 12.80952380952381, "grad_norm": 0.012764991261065006, "learning_rate": 5e-05, "loss": 0.0008, "step": 5380 }, { "epoch": 12.833333333333334, "grad_norm": 0.018289143219590187, "learning_rate": 5e-05, "loss": 0.0009, "step": 5390 }, { "epoch": 12.857142857142858, "grad_norm": 0.006316890940070152, "learning_rate": 5e-05, "loss": 0.0007, "step": 5400 }, { "epoch": 12.880952380952381, "grad_norm": 0.017579294741153717, "learning_rate": 5e-05, "loss": 0.0008, "step": 5410 }, { "epoch": 12.904761904761905, "grad_norm": 0.01720363274216652, "learning_rate": 5e-05, "loss": 0.0011, "step": 5420 }, { "epoch": 12.928571428571429, "grad_norm": 0.012609842233359814, "learning_rate": 5e-05, "loss": 0.0011, "step": 5430 }, { "epoch": 12.952380952380953, "grad_norm": 0.006522722542285919, "learning_rate": 5e-05, "loss": 0.001, "step": 5440 }, { "epoch": 12.976190476190476, "grad_norm": 0.015605674125254154, "learning_rate": 5e-05, "loss": 0.0011, "step": 5450 }, { "epoch": 13.0, "grad_norm": 0.009054956026375294, "learning_rate": 5e-05, "loss": 0.0012, "step": 5460 }, { "epoch": 13.023809523809524, "grad_norm": 0.016910897567868233, "learning_rate": 5e-05, "loss": 0.0011, "step": 5470 }, { "epoch": 13.047619047619047, "grad_norm": 0.012559573166072369, "learning_rate": 5e-05, "loss": 0.0008, "step": 5480 }, { "epoch": 13.071428571428571, "grad_norm": 0.01342256274074316, "learning_rate": 5e-05, "loss": 0.0008, "step": 5490 }, { "epoch": 13.095238095238095, "grad_norm": 0.008237732574343681, "learning_rate": 5e-05, "loss": 0.0007, "step": 5500 }, { "epoch": 13.119047619047619, "grad_norm": 0.01400575041770935, "learning_rate": 5e-05, "loss": 0.0008, "step": 5510 }, { "epoch": 13.142857142857142, "grad_norm": 0.019757332280278206, "learning_rate": 5e-05, "loss": 0.0012, "step": 5520 }, { "epoch": 13.166666666666666, "grad_norm": 0.018210873007774353, "learning_rate": 5e-05, "loss": 0.0008, "step": 5530 }, { "epoch": 13.19047619047619, "grad_norm": 0.0070494236424565315, "learning_rate": 5e-05, "loss": 0.0011, "step": 5540 }, { "epoch": 13.214285714285714, "grad_norm": 0.0030856772791594267, "learning_rate": 5e-05, "loss": 0.0009, "step": 5550 }, { "epoch": 13.238095238095237, "grad_norm": 0.0075521948747336864, "learning_rate": 5e-05, "loss": 0.001, "step": 5560 }, { "epoch": 13.261904761904763, "grad_norm": 0.01915762387216091, "learning_rate": 5e-05, "loss": 0.0008, "step": 5570 }, { "epoch": 13.285714285714286, "grad_norm": 0.007310226559638977, "learning_rate": 5e-05, "loss": 0.0008, "step": 5580 }, { "epoch": 13.30952380952381, "grad_norm": 0.023431280627846718, "learning_rate": 5e-05, "loss": 0.0009, "step": 5590 }, { "epoch": 13.333333333333334, "grad_norm": 0.0055434429086744785, "learning_rate": 5e-05, "loss": 0.0007, "step": 5600 }, { "epoch": 13.357142857142858, "grad_norm": 0.051787495613098145, "learning_rate": 5e-05, "loss": 0.001, "step": 5610 }, { "epoch": 13.380952380952381, "grad_norm": 0.010900055058300495, "learning_rate": 5e-05, "loss": 0.0017, "step": 5620 }, { "epoch": 13.404761904761905, "grad_norm": 0.01772812381386757, "learning_rate": 5e-05, "loss": 0.0009, "step": 5630 }, { "epoch": 13.428571428571429, "grad_norm": 0.003741204971447587, "learning_rate": 5e-05, "loss": 0.0008, "step": 5640 }, { "epoch": 13.452380952380953, "grad_norm": 0.009689460508525372, "learning_rate": 5e-05, "loss": 0.0007, "step": 5650 }, { "epoch": 13.476190476190476, "grad_norm": 0.004325892776250839, "learning_rate": 5e-05, "loss": 0.0006, "step": 5660 }, { "epoch": 13.5, "grad_norm": 0.01864994689822197, "learning_rate": 5e-05, "loss": 0.0004, "step": 5670 }, { "epoch": 13.523809523809524, "grad_norm": 0.011471171863377094, "learning_rate": 5e-05, "loss": 0.0007, "step": 5680 }, { "epoch": 13.547619047619047, "grad_norm": 0.022726034745573997, "learning_rate": 5e-05, "loss": 0.0006, "step": 5690 }, { "epoch": 13.571428571428571, "grad_norm": 0.0025451872497797012, "learning_rate": 5e-05, "loss": 0.0005, "step": 5700 }, { "epoch": 13.595238095238095, "grad_norm": 0.0019215985666960478, "learning_rate": 5e-05, "loss": 0.0005, "step": 5710 }, { "epoch": 13.619047619047619, "grad_norm": 0.03238370642066002, "learning_rate": 5e-05, "loss": 0.0012, "step": 5720 }, { "epoch": 13.642857142857142, "grad_norm": 0.042818158864974976, "learning_rate": 5e-05, "loss": 0.0006, "step": 5730 }, { "epoch": 13.666666666666666, "grad_norm": 0.019736241549253464, "learning_rate": 5e-05, "loss": 0.0006, "step": 5740 }, { "epoch": 13.69047619047619, "grad_norm": 0.009512594901025295, "learning_rate": 5e-05, "loss": 0.0013, "step": 5750 }, { "epoch": 13.714285714285714, "grad_norm": 0.01680351048707962, "learning_rate": 5e-05, "loss": 0.0009, "step": 5760 }, { "epoch": 13.738095238095237, "grad_norm": 0.02108100987970829, "learning_rate": 5e-05, "loss": 0.0011, "step": 5770 }, { "epoch": 13.761904761904763, "grad_norm": 0.01012613344937563, "learning_rate": 5e-05, "loss": 0.001, "step": 5780 }, { "epoch": 13.785714285714286, "grad_norm": 0.012497921474277973, "learning_rate": 5e-05, "loss": 0.0014, "step": 5790 }, { "epoch": 13.80952380952381, "grad_norm": 0.006895515602082014, "learning_rate": 5e-05, "loss": 0.0013, "step": 5800 }, { "epoch": 13.833333333333334, "grad_norm": 0.026326004415750504, "learning_rate": 5e-05, "loss": 0.0009, "step": 5810 }, { "epoch": 13.857142857142858, "grad_norm": 0.02858036197721958, "learning_rate": 5e-05, "loss": 0.001, "step": 5820 }, { "epoch": 13.880952380952381, "grad_norm": 0.014894322492182255, "learning_rate": 5e-05, "loss": 0.0005, "step": 5830 }, { "epoch": 13.904761904761905, "grad_norm": 0.006153046153485775, "learning_rate": 5e-05, "loss": 0.0009, "step": 5840 }, { "epoch": 13.928571428571429, "grad_norm": 0.02046913094818592, "learning_rate": 5e-05, "loss": 0.0011, "step": 5850 }, { "epoch": 13.952380952380953, "grad_norm": 0.05420629680156708, "learning_rate": 5e-05, "loss": 0.0008, "step": 5860 }, { "epoch": 13.976190476190476, "grad_norm": 0.01419808715581894, "learning_rate": 5e-05, "loss": 0.0004, "step": 5870 }, { "epoch": 14.0, "grad_norm": 0.00912909209728241, "learning_rate": 5e-05, "loss": 0.0009, "step": 5880 }, { "epoch": 14.023809523809524, "grad_norm": 0.017824770882725716, "learning_rate": 5e-05, "loss": 0.0011, "step": 5890 }, { "epoch": 14.047619047619047, "grad_norm": 0.008159978315234184, "learning_rate": 5e-05, "loss": 0.0008, "step": 5900 }, { "epoch": 14.071428571428571, "grad_norm": 0.02015644870698452, "learning_rate": 5e-05, "loss": 0.001, "step": 5910 }, { "epoch": 14.095238095238095, "grad_norm": 0.020154140889644623, "learning_rate": 5e-05, "loss": 0.0008, "step": 5920 }, { "epoch": 14.119047619047619, "grad_norm": 0.008028664626181126, "learning_rate": 5e-05, "loss": 0.0009, "step": 5930 }, { "epoch": 14.142857142857142, "grad_norm": 0.01114033441990614, "learning_rate": 5e-05, "loss": 0.0006, "step": 5940 }, { "epoch": 14.166666666666666, "grad_norm": 0.009640121832489967, "learning_rate": 5e-05, "loss": 0.0005, "step": 5950 }, { "epoch": 14.19047619047619, "grad_norm": 0.009514660574495792, "learning_rate": 5e-05, "loss": 0.0008, "step": 5960 }, { "epoch": 14.214285714285714, "grad_norm": 0.010505247861146927, "learning_rate": 5e-05, "loss": 0.0007, "step": 5970 }, { "epoch": 14.238095238095237, "grad_norm": 0.0075341821648180485, "learning_rate": 5e-05, "loss": 0.0005, "step": 5980 }, { "epoch": 14.261904761904763, "grad_norm": 0.007736085448414087, "learning_rate": 5e-05, "loss": 0.0004, "step": 5990 }, { "epoch": 14.285714285714286, "grad_norm": 0.006500266492366791, "learning_rate": 5e-05, "loss": 0.0006, "step": 6000 }, { "epoch": 14.30952380952381, "grad_norm": 0.003755223471671343, "learning_rate": 5e-05, "loss": 0.0005, "step": 6010 }, { "epoch": 14.333333333333334, "grad_norm": 0.03438568115234375, "learning_rate": 5e-05, "loss": 0.0011, "step": 6020 }, { "epoch": 14.357142857142858, "grad_norm": 0.003875867696478963, "learning_rate": 5e-05, "loss": 0.0002, "step": 6030 }, { "epoch": 14.380952380952381, "grad_norm": 0.041189443320035934, "learning_rate": 5e-05, "loss": 0.0006, "step": 6040 }, { "epoch": 14.404761904761905, "grad_norm": 0.044477108865976334, "learning_rate": 5e-05, "loss": 0.001, "step": 6050 }, { "epoch": 14.428571428571429, "grad_norm": 0.021848198026418686, "learning_rate": 5e-05, "loss": 0.0013, "step": 6060 }, { "epoch": 14.452380952380953, "grad_norm": 0.009198046289384365, "learning_rate": 5e-05, "loss": 0.0006, "step": 6070 }, { "epoch": 14.476190476190476, "grad_norm": 0.00906557496637106, "learning_rate": 5e-05, "loss": 0.0005, "step": 6080 }, { "epoch": 14.5, "grad_norm": 0.027908943593502045, "learning_rate": 5e-05, "loss": 0.0004, "step": 6090 }, { "epoch": 14.523809523809524, "grad_norm": 0.010970670729875565, "learning_rate": 5e-05, "loss": 0.0005, "step": 6100 }, { "epoch": 14.547619047619047, "grad_norm": 0.02569267526268959, "learning_rate": 5e-05, "loss": 0.0006, "step": 6110 }, { "epoch": 14.571428571428571, "grad_norm": 0.00839224737137556, "learning_rate": 5e-05, "loss": 0.0007, "step": 6120 }, { "epoch": 14.595238095238095, "grad_norm": 0.00571605795994401, "learning_rate": 5e-05, "loss": 0.0005, "step": 6130 }, { "epoch": 14.619047619047619, "grad_norm": 0.01305799838155508, "learning_rate": 5e-05, "loss": 0.0004, "step": 6140 }, { "epoch": 14.642857142857142, "grad_norm": 0.033176738768815994, "learning_rate": 5e-05, "loss": 0.0008, "step": 6150 }, { "epoch": 14.666666666666666, "grad_norm": 0.08488498628139496, "learning_rate": 5e-05, "loss": 0.0011, "step": 6160 }, { "epoch": 14.69047619047619, "grad_norm": 0.008058450184762478, "learning_rate": 5e-05, "loss": 0.0005, "step": 6170 }, { "epoch": 14.714285714285714, "grad_norm": 0.028126804158091545, "learning_rate": 5e-05, "loss": 0.001, "step": 6180 }, { "epoch": 14.738095238095237, "grad_norm": 0.01536962017416954, "learning_rate": 5e-05, "loss": 0.0004, "step": 6190 }, { "epoch": 14.761904761904763, "grad_norm": 0.007554737385362387, "learning_rate": 5e-05, "loss": 0.001, "step": 6200 }, { "epoch": 14.785714285714286, "grad_norm": 0.009313613176345825, "learning_rate": 5e-05, "loss": 0.001, "step": 6210 }, { "epoch": 14.80952380952381, "grad_norm": 0.03818364813923836, "learning_rate": 5e-05, "loss": 0.0008, "step": 6220 }, { "epoch": 14.833333333333334, "grad_norm": 0.004172751680016518, "learning_rate": 5e-05, "loss": 0.0005, "step": 6230 }, { "epoch": 14.857142857142858, "grad_norm": 0.02367938496172428, "learning_rate": 5e-05, "loss": 0.001, "step": 6240 }, { "epoch": 14.880952380952381, "grad_norm": 0.029346376657485962, "learning_rate": 5e-05, "loss": 0.0013, "step": 6250 }, { "epoch": 14.904761904761905, "grad_norm": 0.008894797414541245, "learning_rate": 5e-05, "loss": 0.0006, "step": 6260 }, { "epoch": 14.928571428571429, "grad_norm": 0.0210129227489233, "learning_rate": 5e-05, "loss": 0.0006, "step": 6270 }, { "epoch": 14.952380952380953, "grad_norm": 0.006997206248342991, "learning_rate": 5e-05, "loss": 0.0008, "step": 6280 }, { "epoch": 14.976190476190476, "grad_norm": 0.027477620169520378, "learning_rate": 5e-05, "loss": 0.0007, "step": 6290 }, { "epoch": 15.0, "grad_norm": 0.01882118359208107, "learning_rate": 5e-05, "loss": 0.0009, "step": 6300 }, { "epoch": 15.023809523809524, "grad_norm": 0.004850170575082302, "learning_rate": 5e-05, "loss": 0.0006, "step": 6310 }, { "epoch": 15.047619047619047, "grad_norm": 0.03450813889503479, "learning_rate": 5e-05, "loss": 0.0011, "step": 6320 }, { "epoch": 15.071428571428571, "grad_norm": 0.012980910018086433, "learning_rate": 5e-05, "loss": 0.0008, "step": 6330 }, { "epoch": 15.095238095238095, "grad_norm": 0.01641819067299366, "learning_rate": 5e-05, "loss": 0.0006, "step": 6340 }, { "epoch": 15.119047619047619, "grad_norm": 0.0036320008803159, "learning_rate": 5e-05, "loss": 0.0004, "step": 6350 }, { "epoch": 15.142857142857142, "grad_norm": 0.005455750040709972, "learning_rate": 5e-05, "loss": 0.0004, "step": 6360 }, { "epoch": 15.166666666666666, "grad_norm": 0.007276155520230532, "learning_rate": 5e-05, "loss": 0.0005, "step": 6370 }, { "epoch": 15.19047619047619, "grad_norm": 0.02196861244738102, "learning_rate": 5e-05, "loss": 0.0009, "step": 6380 }, { "epoch": 15.214285714285714, "grad_norm": 0.004664530977606773, "learning_rate": 5e-05, "loss": 0.0003, "step": 6390 }, { "epoch": 15.238095238095237, "grad_norm": 0.008527121506631374, "learning_rate": 5e-05, "loss": 0.0004, "step": 6400 }, { "epoch": 15.261904761904763, "grad_norm": 0.015898073092103004, "learning_rate": 5e-05, "loss": 0.0006, "step": 6410 }, { "epoch": 15.285714285714286, "grad_norm": 0.007485649082809687, "learning_rate": 5e-05, "loss": 0.0005, "step": 6420 }, { "epoch": 15.30952380952381, "grad_norm": 0.011031672358512878, "learning_rate": 5e-05, "loss": 0.0006, "step": 6430 }, { "epoch": 15.333333333333334, "grad_norm": 0.017634835094213486, "learning_rate": 5e-05, "loss": 0.0006, "step": 6440 }, { "epoch": 15.357142857142858, "grad_norm": 0.010040491819381714, "learning_rate": 5e-05, "loss": 0.0008, "step": 6450 }, { "epoch": 15.380952380952381, "grad_norm": 0.00803917832672596, "learning_rate": 5e-05, "loss": 0.0006, "step": 6460 }, { "epoch": 15.404761904761905, "grad_norm": 0.002537789987400174, "learning_rate": 5e-05, "loss": 0.0008, "step": 6470 }, { "epoch": 15.428571428571429, "grad_norm": 0.004426058381795883, "learning_rate": 5e-05, "loss": 0.0009, "step": 6480 }, { "epoch": 15.452380952380953, "grad_norm": 0.004302819259464741, "learning_rate": 5e-05, "loss": 0.0006, "step": 6490 }, { "epoch": 15.476190476190476, "grad_norm": 0.006018012762069702, "learning_rate": 5e-05, "loss": 0.0006, "step": 6500 }, { "epoch": 15.5, "grad_norm": 0.00840548425912857, "learning_rate": 5e-05, "loss": 0.0005, "step": 6510 }, { "epoch": 15.523809523809524, "grad_norm": 0.011578329838812351, "learning_rate": 5e-05, "loss": 0.0004, "step": 6520 }, { "epoch": 15.547619047619047, "grad_norm": 0.013048551976680756, "learning_rate": 5e-05, "loss": 0.0003, "step": 6530 }, { "epoch": 15.571428571428571, "grad_norm": 0.004024627152830362, "learning_rate": 5e-05, "loss": 0.0003, "step": 6540 }, { "epoch": 15.595238095238095, "grad_norm": 0.007286364212632179, "learning_rate": 5e-05, "loss": 0.0004, "step": 6550 }, { "epoch": 15.619047619047619, "grad_norm": 0.014821533113718033, "learning_rate": 5e-05, "loss": 0.0004, "step": 6560 }, { "epoch": 15.642857142857142, "grad_norm": 0.002453850582242012, "learning_rate": 5e-05, "loss": 0.0005, "step": 6570 }, { "epoch": 15.666666666666666, "grad_norm": 0.03152019903063774, "learning_rate": 5e-05, "loss": 0.0011, "step": 6580 }, { "epoch": 15.69047619047619, "grad_norm": 0.046064406633377075, "learning_rate": 5e-05, "loss": 0.0014, "step": 6590 }, { "epoch": 15.714285714285714, "grad_norm": 0.019682571291923523, "learning_rate": 5e-05, "loss": 0.0008, "step": 6600 }, { "epoch": 15.738095238095237, "grad_norm": 0.0024483215529471636, "learning_rate": 5e-05, "loss": 0.0005, "step": 6610 }, { "epoch": 15.761904761904763, "grad_norm": 0.008708986453711987, "learning_rate": 5e-05, "loss": 0.0005, "step": 6620 }, { "epoch": 15.785714285714286, "grad_norm": 0.006394381169229746, "learning_rate": 5e-05, "loss": 0.0005, "step": 6630 }, { "epoch": 15.80952380952381, "grad_norm": 0.03494114428758621, "learning_rate": 5e-05, "loss": 0.0005, "step": 6640 }, { "epoch": 15.833333333333334, "grad_norm": 0.02163984254002571, "learning_rate": 5e-05, "loss": 0.0006, "step": 6650 }, { "epoch": 15.857142857142858, "grad_norm": 0.00985416304320097, "learning_rate": 5e-05, "loss": 0.0004, "step": 6660 }, { "epoch": 15.880952380952381, "grad_norm": 0.008056841790676117, "learning_rate": 5e-05, "loss": 0.0007, "step": 6670 }, { "epoch": 15.904761904761905, "grad_norm": 0.004157788585871458, "learning_rate": 5e-05, "loss": 0.0006, "step": 6680 }, { "epoch": 15.928571428571429, "grad_norm": 0.00597544526681304, "learning_rate": 5e-05, "loss": 0.0004, "step": 6690 }, { "epoch": 15.952380952380953, "grad_norm": 0.023443806916475296, "learning_rate": 5e-05, "loss": 0.001, "step": 6700 }, { "epoch": 15.976190476190476, "grad_norm": 0.012778760865330696, "learning_rate": 5e-05, "loss": 0.0007, "step": 6710 }, { "epoch": 16.0, "grad_norm": 0.015170056372880936, "learning_rate": 5e-05, "loss": 0.0006, "step": 6720 }, { "epoch": 16.023809523809526, "grad_norm": 0.01816493459045887, "learning_rate": 5e-05, "loss": 0.0007, "step": 6730 }, { "epoch": 16.047619047619047, "grad_norm": 0.00630978075787425, "learning_rate": 5e-05, "loss": 0.0006, "step": 6740 }, { "epoch": 16.071428571428573, "grad_norm": 0.004600324667990208, "learning_rate": 5e-05, "loss": 0.0009, "step": 6750 }, { "epoch": 16.095238095238095, "grad_norm": 0.03235204890370369, "learning_rate": 5e-05, "loss": 0.0008, "step": 6760 }, { "epoch": 16.11904761904762, "grad_norm": 0.012307102791965008, "learning_rate": 5e-05, "loss": 0.0006, "step": 6770 }, { "epoch": 16.142857142857142, "grad_norm": 0.0052837226539850235, "learning_rate": 5e-05, "loss": 0.0005, "step": 6780 }, { "epoch": 16.166666666666668, "grad_norm": 0.02251107059419155, "learning_rate": 5e-05, "loss": 0.0006, "step": 6790 }, { "epoch": 16.19047619047619, "grad_norm": 0.011189977638423443, "learning_rate": 5e-05, "loss": 0.0006, "step": 6800 }, { "epoch": 16.214285714285715, "grad_norm": 0.05799144506454468, "learning_rate": 5e-05, "loss": 0.001, "step": 6810 }, { "epoch": 16.238095238095237, "grad_norm": 0.023233134299516678, "learning_rate": 5e-05, "loss": 0.0006, "step": 6820 }, { "epoch": 16.261904761904763, "grad_norm": 0.005226644221693277, "learning_rate": 5e-05, "loss": 0.0004, "step": 6830 }, { "epoch": 16.285714285714285, "grad_norm": 0.022896727547049522, "learning_rate": 5e-05, "loss": 0.0005, "step": 6840 }, { "epoch": 16.30952380952381, "grad_norm": 0.009060421027243137, "learning_rate": 5e-05, "loss": 0.0003, "step": 6850 }, { "epoch": 16.333333333333332, "grad_norm": 0.019654234871268272, "learning_rate": 5e-05, "loss": 0.0004, "step": 6860 }, { "epoch": 16.357142857142858, "grad_norm": 0.01879570446908474, "learning_rate": 5e-05, "loss": 0.0005, "step": 6870 }, { "epoch": 16.38095238095238, "grad_norm": 0.015871930867433548, "learning_rate": 5e-05, "loss": 0.0009, "step": 6880 }, { "epoch": 16.404761904761905, "grad_norm": 0.005459831561893225, "learning_rate": 5e-05, "loss": 0.0006, "step": 6890 }, { "epoch": 16.428571428571427, "grad_norm": 0.002002556109800935, "learning_rate": 5e-05, "loss": 0.0007, "step": 6900 }, { "epoch": 16.452380952380953, "grad_norm": 0.00361634767614305, "learning_rate": 5e-05, "loss": 0.0005, "step": 6910 }, { "epoch": 16.476190476190474, "grad_norm": 0.008024908602237701, "learning_rate": 5e-05, "loss": 0.0006, "step": 6920 }, { "epoch": 16.5, "grad_norm": 0.0015706607373431325, "learning_rate": 5e-05, "loss": 0.0008, "step": 6930 }, { "epoch": 16.523809523809526, "grad_norm": 0.015266403555870056, "learning_rate": 5e-05, "loss": 0.0008, "step": 6940 }, { "epoch": 16.547619047619047, "grad_norm": 0.03144041448831558, "learning_rate": 5e-05, "loss": 0.0013, "step": 6950 }, { "epoch": 16.571428571428573, "grad_norm": 0.0020375235471874475, "learning_rate": 5e-05, "loss": 0.0006, "step": 6960 }, { "epoch": 16.595238095238095, "grad_norm": 0.03112536109983921, "learning_rate": 5e-05, "loss": 0.0006, "step": 6970 }, { "epoch": 16.61904761904762, "grad_norm": 0.005854692310094833, "learning_rate": 5e-05, "loss": 0.0005, "step": 6980 }, { "epoch": 16.642857142857142, "grad_norm": 0.01317546609789133, "learning_rate": 5e-05, "loss": 0.0003, "step": 6990 }, { "epoch": 16.666666666666668, "grad_norm": 0.008511519059538841, "learning_rate": 5e-05, "loss": 0.0007, "step": 7000 }, { "epoch": 16.69047619047619, "grad_norm": 0.006545855663716793, "learning_rate": 5e-05, "loss": 0.0003, "step": 7010 }, { "epoch": 16.714285714285715, "grad_norm": 0.010415365919470787, "learning_rate": 5e-05, "loss": 0.0006, "step": 7020 }, { "epoch": 16.738095238095237, "grad_norm": 0.01626661792397499, "learning_rate": 5e-05, "loss": 0.0007, "step": 7030 }, { "epoch": 16.761904761904763, "grad_norm": 0.005767187103629112, "learning_rate": 5e-05, "loss": 0.0006, "step": 7040 }, { "epoch": 16.785714285714285, "grad_norm": 0.01930147223174572, "learning_rate": 5e-05, "loss": 0.0004, "step": 7050 }, { "epoch": 16.80952380952381, "grad_norm": 0.022985486313700676, "learning_rate": 5e-05, "loss": 0.0009, "step": 7060 }, { "epoch": 16.833333333333332, "grad_norm": 0.015567641705274582, "learning_rate": 5e-05, "loss": 0.0005, "step": 7070 }, { "epoch": 16.857142857142858, "grad_norm": 0.023370876908302307, "learning_rate": 5e-05, "loss": 0.0005, "step": 7080 }, { "epoch": 16.88095238095238, "grad_norm": 0.007470586337149143, "learning_rate": 5e-05, "loss": 0.0005, "step": 7090 }, { "epoch": 16.904761904761905, "grad_norm": 0.006150704808533192, "learning_rate": 5e-05, "loss": 0.0007, "step": 7100 }, { "epoch": 16.928571428571427, "grad_norm": 0.0042198821902275085, "learning_rate": 5e-05, "loss": 0.0007, "step": 7110 }, { "epoch": 16.952380952380953, "grad_norm": 0.017077218741178513, "learning_rate": 5e-05, "loss": 0.0004, "step": 7120 }, { "epoch": 16.976190476190474, "grad_norm": 0.023657051846385002, "learning_rate": 5e-05, "loss": 0.0005, "step": 7130 }, { "epoch": 17.0, "grad_norm": 0.016245121136307716, "learning_rate": 5e-05, "loss": 0.0005, "step": 7140 }, { "epoch": 17.023809523809526, "grad_norm": 0.0030179405584931374, "learning_rate": 5e-05, "loss": 0.0007, "step": 7150 }, { "epoch": 17.047619047619047, "grad_norm": 0.025457536801695824, "learning_rate": 5e-05, "loss": 0.0006, "step": 7160 }, { "epoch": 17.071428571428573, "grad_norm": 0.03168899938464165, "learning_rate": 5e-05, "loss": 0.0007, "step": 7170 }, { "epoch": 17.095238095238095, "grad_norm": 0.036503903567790985, "learning_rate": 5e-05, "loss": 0.0009, "step": 7180 }, { "epoch": 17.11904761904762, "grad_norm": 0.0242981668561697, "learning_rate": 5e-05, "loss": 0.0008, "step": 7190 }, { "epoch": 17.142857142857142, "grad_norm": 0.014298885129392147, "learning_rate": 5e-05, "loss": 0.0005, "step": 7200 }, { "epoch": 17.166666666666668, "grad_norm": 0.02301887236535549, "learning_rate": 5e-05, "loss": 0.0004, "step": 7210 }, { "epoch": 17.19047619047619, "grad_norm": 0.03424462303519249, "learning_rate": 5e-05, "loss": 0.0008, "step": 7220 }, { "epoch": 17.214285714285715, "grad_norm": 0.023424234241247177, "learning_rate": 5e-05, "loss": 0.0007, "step": 7230 }, { "epoch": 17.238095238095237, "grad_norm": 0.023134572431445122, "learning_rate": 5e-05, "loss": 0.0009, "step": 7240 }, { "epoch": 17.261904761904763, "grad_norm": 0.025860372930765152, "learning_rate": 5e-05, "loss": 0.0008, "step": 7250 }, { "epoch": 17.285714285714285, "grad_norm": 0.01248745433986187, "learning_rate": 5e-05, "loss": 0.0004, "step": 7260 }, { "epoch": 17.30952380952381, "grad_norm": 0.018757270649075508, "learning_rate": 5e-05, "loss": 0.0005, "step": 7270 }, { "epoch": 17.333333333333332, "grad_norm": 0.011798521503806114, "learning_rate": 5e-05, "loss": 0.0006, "step": 7280 }, { "epoch": 17.357142857142858, "grad_norm": 0.020112033933401108, "learning_rate": 5e-05, "loss": 0.0007, "step": 7290 }, { "epoch": 17.38095238095238, "grad_norm": 0.009371660649776459, "learning_rate": 5e-05, "loss": 0.0006, "step": 7300 }, { "epoch": 17.404761904761905, "grad_norm": 0.010503299534320831, "learning_rate": 5e-05, "loss": 0.0005, "step": 7310 }, { "epoch": 17.428571428571427, "grad_norm": 0.02050066739320755, "learning_rate": 5e-05, "loss": 0.0005, "step": 7320 }, { "epoch": 17.452380952380953, "grad_norm": 0.034579671919345856, "learning_rate": 5e-05, "loss": 0.0011, "step": 7330 }, { "epoch": 17.476190476190474, "grad_norm": 0.02491345815360546, "learning_rate": 5e-05, "loss": 0.0006, "step": 7340 }, { "epoch": 17.5, "grad_norm": 0.0013771721860393882, "learning_rate": 5e-05, "loss": 0.0003, "step": 7350 }, { "epoch": 17.523809523809526, "grad_norm": 0.012594480067491531, "learning_rate": 5e-05, "loss": 0.0004, "step": 7360 }, { "epoch": 17.547619047619047, "grad_norm": 0.025901375338435173, "learning_rate": 5e-05, "loss": 0.0006, "step": 7370 }, { "epoch": 17.571428571428573, "grad_norm": 0.004503455013036728, "learning_rate": 5e-05, "loss": 0.0006, "step": 7380 }, { "epoch": 17.595238095238095, "grad_norm": 0.01678163930773735, "learning_rate": 5e-05, "loss": 0.0003, "step": 7390 }, { "epoch": 17.61904761904762, "grad_norm": 0.01972806081175804, "learning_rate": 5e-05, "loss": 0.0003, "step": 7400 }, { "epoch": 17.642857142857142, "grad_norm": 0.005659343674778938, "learning_rate": 5e-05, "loss": 0.0007, "step": 7410 }, { "epoch": 17.666666666666668, "grad_norm": 0.010966219939291477, "learning_rate": 5e-05, "loss": 0.0006, "step": 7420 }, { "epoch": 17.69047619047619, "grad_norm": 0.01521807536482811, "learning_rate": 5e-05, "loss": 0.0003, "step": 7430 }, { "epoch": 17.714285714285715, "grad_norm": 0.0045644501224160194, "learning_rate": 5e-05, "loss": 0.0005, "step": 7440 }, { "epoch": 17.738095238095237, "grad_norm": 0.0009657703340053558, "learning_rate": 5e-05, "loss": 0.0004, "step": 7450 }, { "epoch": 17.761904761904763, "grad_norm": 0.03967520594596863, "learning_rate": 5e-05, "loss": 0.0005, "step": 7460 }, { "epoch": 17.785714285714285, "grad_norm": 0.003580782562494278, "learning_rate": 5e-05, "loss": 0.0002, "step": 7470 }, { "epoch": 17.80952380952381, "grad_norm": 0.02645096555352211, "learning_rate": 5e-05, "loss": 0.0006, "step": 7480 }, { "epoch": 17.833333333333332, "grad_norm": 0.004722685553133488, "learning_rate": 5e-05, "loss": 0.0004, "step": 7490 }, { "epoch": 17.857142857142858, "grad_norm": 0.026108304038643837, "learning_rate": 5e-05, "loss": 0.0005, "step": 7500 }, { "epoch": 17.88095238095238, "grad_norm": 0.008777507580816746, "learning_rate": 5e-05, "loss": 0.0003, "step": 7510 }, { "epoch": 17.904761904761905, "grad_norm": 0.011482530273497105, "learning_rate": 5e-05, "loss": 0.0005, "step": 7520 }, { "epoch": 17.928571428571427, "grad_norm": 0.003905653487890959, "learning_rate": 5e-05, "loss": 0.0008, "step": 7530 }, { "epoch": 17.952380952380953, "grad_norm": 0.020037874579429626, "learning_rate": 5e-05, "loss": 0.001, "step": 7540 }, { "epoch": 17.976190476190474, "grad_norm": 0.017041349783539772, "learning_rate": 5e-05, "loss": 0.0004, "step": 7550 }, { "epoch": 18.0, "grad_norm": 0.0016616099746897817, "learning_rate": 5e-05, "loss": 0.0005, "step": 7560 }, { "epoch": 18.023809523809526, "grad_norm": 0.021135887131094933, "learning_rate": 5e-05, "loss": 0.0007, "step": 7570 }, { "epoch": 18.047619047619047, "grad_norm": 0.010796349495649338, "learning_rate": 5e-05, "loss": 0.0003, "step": 7580 }, { "epoch": 18.071428571428573, "grad_norm": 0.018032172694802284, "learning_rate": 5e-05, "loss": 0.0006, "step": 7590 }, { "epoch": 18.095238095238095, "grad_norm": 0.028129780665040016, "learning_rate": 5e-05, "loss": 0.0008, "step": 7600 }, { "epoch": 18.11904761904762, "grad_norm": 0.0026430708821862936, "learning_rate": 5e-05, "loss": 0.0007, "step": 7610 }, { "epoch": 18.142857142857142, "grad_norm": 0.00862652063369751, "learning_rate": 5e-05, "loss": 0.0003, "step": 7620 }, { "epoch": 18.166666666666668, "grad_norm": 0.0017266141949221492, "learning_rate": 5e-05, "loss": 0.0002, "step": 7630 }, { "epoch": 18.19047619047619, "grad_norm": 0.002414742251858115, "learning_rate": 5e-05, "loss": 0.0005, "step": 7640 }, { "epoch": 18.214285714285715, "grad_norm": 0.002050422364845872, "learning_rate": 5e-05, "loss": 0.0004, "step": 7650 }, { "epoch": 18.238095238095237, "grad_norm": 0.0194852314889431, "learning_rate": 5e-05, "loss": 0.0005, "step": 7660 }, { "epoch": 18.261904761904763, "grad_norm": 0.01717439480125904, "learning_rate": 5e-05, "loss": 0.0007, "step": 7670 }, { "epoch": 18.285714285714285, "grad_norm": 0.011901612393558025, "learning_rate": 5e-05, "loss": 0.0004, "step": 7680 }, { "epoch": 18.30952380952381, "grad_norm": 0.029567129909992218, "learning_rate": 5e-05, "loss": 0.0009, "step": 7690 }, { "epoch": 18.333333333333332, "grad_norm": 0.006910389289259911, "learning_rate": 5e-05, "loss": 0.0011, "step": 7700 }, { "epoch": 18.357142857142858, "grad_norm": 0.007212989963591099, "learning_rate": 5e-05, "loss": 0.0006, "step": 7710 }, { "epoch": 18.38095238095238, "grad_norm": 0.017439018934965134, "learning_rate": 5e-05, "loss": 0.0003, "step": 7720 }, { "epoch": 18.404761904761905, "grad_norm": 0.0020984190050512552, "learning_rate": 5e-05, "loss": 0.0004, "step": 7730 }, { "epoch": 18.428571428571427, "grad_norm": 0.011482183821499348, "learning_rate": 5e-05, "loss": 0.0009, "step": 7740 }, { "epoch": 18.452380952380953, "grad_norm": 0.006130084861069918, "learning_rate": 5e-05, "loss": 0.0003, "step": 7750 }, { "epoch": 18.476190476190474, "grad_norm": 0.018064623698592186, "learning_rate": 5e-05, "loss": 0.0011, "step": 7760 }, { "epoch": 18.5, "grad_norm": 0.007949289865791798, "learning_rate": 5e-05, "loss": 0.0005, "step": 7770 }, { "epoch": 18.523809523809526, "grad_norm": 0.0036398067604750395, "learning_rate": 5e-05, "loss": 0.0003, "step": 7780 }, { "epoch": 18.547619047619047, "grad_norm": 0.021670356392860413, "learning_rate": 5e-05, "loss": 0.0005, "step": 7790 }, { "epoch": 18.571428571428573, "grad_norm": 0.0020873206667602062, "learning_rate": 5e-05, "loss": 0.0006, "step": 7800 }, { "epoch": 18.595238095238095, "grad_norm": 0.007899454794824123, "learning_rate": 5e-05, "loss": 0.0003, "step": 7810 }, { "epoch": 18.61904761904762, "grad_norm": 0.03517717495560646, "learning_rate": 5e-05, "loss": 0.0003, "step": 7820 }, { "epoch": 18.642857142857142, "grad_norm": 0.010151904076337814, "learning_rate": 5e-05, "loss": 0.0002, "step": 7830 }, { "epoch": 18.666666666666668, "grad_norm": 0.001768286689184606, "learning_rate": 5e-05, "loss": 0.0005, "step": 7840 }, { "epoch": 18.69047619047619, "grad_norm": 0.005438578315079212, "learning_rate": 5e-05, "loss": 0.0003, "step": 7850 }, { "epoch": 18.714285714285715, "grad_norm": 0.01367899402976036, "learning_rate": 5e-05, "loss": 0.0003, "step": 7860 }, { "epoch": 18.738095238095237, "grad_norm": 0.005496785510331392, "learning_rate": 5e-05, "loss": 0.0005, "step": 7870 }, { "epoch": 18.761904761904763, "grad_norm": 0.010219060815870762, "learning_rate": 5e-05, "loss": 0.0003, "step": 7880 }, { "epoch": 18.785714285714285, "grad_norm": 0.005748805124312639, "learning_rate": 5e-05, "loss": 0.0004, "step": 7890 }, { "epoch": 18.80952380952381, "grad_norm": 0.0068986667320132256, "learning_rate": 5e-05, "loss": 0.0006, "step": 7900 }, { "epoch": 18.833333333333332, "grad_norm": 0.008341987617313862, "learning_rate": 5e-05, "loss": 0.0004, "step": 7910 }, { "epoch": 18.857142857142858, "grad_norm": 0.015715043991804123, "learning_rate": 5e-05, "loss": 0.0005, "step": 7920 }, { "epoch": 18.88095238095238, "grad_norm": 0.009235808625817299, "learning_rate": 5e-05, "loss": 0.0003, "step": 7930 }, { "epoch": 18.904761904761905, "grad_norm": 0.004791698418557644, "learning_rate": 5e-05, "loss": 0.0006, "step": 7940 }, { "epoch": 18.928571428571427, "grad_norm": 0.020587580278515816, "learning_rate": 5e-05, "loss": 0.0005, "step": 7950 }, { "epoch": 18.952380952380953, "grad_norm": 0.0022148280404508114, "learning_rate": 5e-05, "loss": 0.0003, "step": 7960 }, { "epoch": 18.976190476190474, "grad_norm": 0.0040055266581475735, "learning_rate": 5e-05, "loss": 0.0003, "step": 7970 }, { "epoch": 19.0, "grad_norm": 0.0041480157524347305, "learning_rate": 5e-05, "loss": 0.0007, "step": 7980 }, { "epoch": 19.023809523809526, "grad_norm": 0.009851422160863876, "learning_rate": 5e-05, "loss": 0.0003, "step": 7990 }, { "epoch": 19.047619047619047, "grad_norm": 0.029928863048553467, "learning_rate": 5e-05, "loss": 0.0004, "step": 8000 }, { "epoch": 19.071428571428573, "grad_norm": 0.005271303467452526, "learning_rate": 5e-05, "loss": 0.0004, "step": 8010 }, { "epoch": 19.095238095238095, "grad_norm": 0.011121473275125027, "learning_rate": 5e-05, "loss": 0.0004, "step": 8020 }, { "epoch": 19.11904761904762, "grad_norm": 0.002328911330550909, "learning_rate": 5e-05, "loss": 0.0005, "step": 8030 }, { "epoch": 19.142857142857142, "grad_norm": 0.0005802198429591954, "learning_rate": 5e-05, "loss": 0.0003, "step": 8040 }, { "epoch": 19.166666666666668, "grad_norm": 0.003331305691972375, "learning_rate": 5e-05, "loss": 0.0005, "step": 8050 }, { "epoch": 19.19047619047619, "grad_norm": 0.0036068561021238565, "learning_rate": 5e-05, "loss": 0.0003, "step": 8060 }, { "epoch": 19.214285714285715, "grad_norm": 0.008253911510109901, "learning_rate": 5e-05, "loss": 0.0003, "step": 8070 }, { "epoch": 19.238095238095237, "grad_norm": 0.023871315643191338, "learning_rate": 5e-05, "loss": 0.0004, "step": 8080 }, { "epoch": 19.261904761904763, "grad_norm": 0.004133463837206364, "learning_rate": 5e-05, "loss": 0.0005, "step": 8090 }, { "epoch": 19.285714285714285, "grad_norm": 0.0016440710751339793, "learning_rate": 5e-05, "loss": 0.0004, "step": 8100 }, { "epoch": 19.30952380952381, "grad_norm": 0.004808527417480946, "learning_rate": 5e-05, "loss": 0.0004, "step": 8110 }, { "epoch": 19.333333333333332, "grad_norm": 0.03961317241191864, "learning_rate": 5e-05, "loss": 0.0005, "step": 8120 }, { "epoch": 19.357142857142858, "grad_norm": 0.021009011194109917, "learning_rate": 5e-05, "loss": 0.0005, "step": 8130 }, { "epoch": 19.38095238095238, "grad_norm": 0.009781712666153908, "learning_rate": 5e-05, "loss": 0.0005, "step": 8140 }, { "epoch": 19.404761904761905, "grad_norm": 0.0019496449967846274, "learning_rate": 5e-05, "loss": 0.0002, "step": 8150 }, { "epoch": 19.428571428571427, "grad_norm": 0.00960745383054018, "learning_rate": 5e-05, "loss": 0.0008, "step": 8160 }, { "epoch": 19.452380952380953, "grad_norm": 0.008466621860861778, "learning_rate": 5e-05, "loss": 0.0005, "step": 8170 }, { "epoch": 19.476190476190474, "grad_norm": 0.014924910850822926, "learning_rate": 5e-05, "loss": 0.0006, "step": 8180 }, { "epoch": 19.5, "grad_norm": 0.006802048999816179, "learning_rate": 5e-05, "loss": 0.0006, "step": 8190 }, { "epoch": 19.523809523809526, "grad_norm": 0.0138113833963871, "learning_rate": 5e-05, "loss": 0.0004, "step": 8200 }, { "epoch": 19.547619047619047, "grad_norm": 0.01139379758387804, "learning_rate": 5e-05, "loss": 0.0004, "step": 8210 }, { "epoch": 19.571428571428573, "grad_norm": 0.03176883980631828, "learning_rate": 5e-05, "loss": 0.0008, "step": 8220 }, { "epoch": 19.595238095238095, "grad_norm": 0.007066241931170225, "learning_rate": 5e-05, "loss": 0.0003, "step": 8230 }, { "epoch": 19.61904761904762, "grad_norm": 0.001747833681292832, "learning_rate": 5e-05, "loss": 0.0002, "step": 8240 }, { "epoch": 19.642857142857142, "grad_norm": 0.004946041852235794, "learning_rate": 5e-05, "loss": 0.0005, "step": 8250 }, { "epoch": 19.666666666666668, "grad_norm": 0.0037244544364511967, "learning_rate": 5e-05, "loss": 0.0002, "step": 8260 }, { "epoch": 19.69047619047619, "grad_norm": 0.007341946009546518, "learning_rate": 5e-05, "loss": 0.0004, "step": 8270 }, { "epoch": 19.714285714285715, "grad_norm": 0.002960850251838565, "learning_rate": 5e-05, "loss": 0.0007, "step": 8280 }, { "epoch": 19.738095238095237, "grad_norm": 0.0019498078618198633, "learning_rate": 5e-05, "loss": 0.0003, "step": 8290 }, { "epoch": 19.761904761904763, "grad_norm": 0.017022594809532166, "learning_rate": 5e-05, "loss": 0.0006, "step": 8300 }, { "epoch": 19.785714285714285, "grad_norm": 0.001329585094936192, "learning_rate": 5e-05, "loss": 0.0003, "step": 8310 }, { "epoch": 19.80952380952381, "grad_norm": 0.003296843497082591, "learning_rate": 5e-05, "loss": 0.0003, "step": 8320 }, { "epoch": 19.833333333333332, "grad_norm": 0.015899814665317535, "learning_rate": 5e-05, "loss": 0.0004, "step": 8330 }, { "epoch": 19.857142857142858, "grad_norm": 0.010661507025361061, "learning_rate": 5e-05, "loss": 0.0003, "step": 8340 }, { "epoch": 19.88095238095238, "grad_norm": 0.0018181676277890801, "learning_rate": 5e-05, "loss": 0.0003, "step": 8350 }, { "epoch": 19.904761904761905, "grad_norm": 0.029869144782423973, "learning_rate": 5e-05, "loss": 0.0006, "step": 8360 }, { "epoch": 19.928571428571427, "grad_norm": 0.016890952363610268, "learning_rate": 5e-05, "loss": 0.0002, "step": 8370 }, { "epoch": 19.952380952380953, "grad_norm": 0.001414968748576939, "learning_rate": 5e-05, "loss": 0.0003, "step": 8380 }, { "epoch": 19.976190476190474, "grad_norm": 0.0023363372310996056, "learning_rate": 5e-05, "loss": 0.0004, "step": 8390 }, { "epoch": 20.0, "grad_norm": 0.001184182707220316, "learning_rate": 5e-05, "loss": 0.0005, "step": 8400 } ], "logging_steps": 10, "max_steps": 8400, "num_input_tokens_seen": 0, "num_train_epochs": 20, "save_steps": 2000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.2637663864989696e+18, "train_batch_size": 32, "trial_name": null, "trial_params": null }