| { | |
| "best_global_step": 1000, | |
| "best_metric": 2.343059778213501, | |
| "best_model_checkpoint": "saves/qwen2.5_3d/full/sft_7b_stage2/checkpoint-1000", | |
| "epoch": 2.9977671451355663, | |
| "eval_steps": 1000, | |
| "global_step": 1173, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.025518341307814992, | |
| "grad_norm": 41.38909683442534, | |
| "learning_rate": 1.1440677966101696e-06, | |
| "loss": 13.0588, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.051036682615629984, | |
| "grad_norm": 24.45025201744234, | |
| "learning_rate": 2.4152542372881355e-06, | |
| "loss": 11.6024, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.07655502392344497, | |
| "grad_norm": 8.918549090671922, | |
| "learning_rate": 3.686440677966102e-06, | |
| "loss": 9.677, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.10207336523125997, | |
| "grad_norm": 5.740463305990162, | |
| "learning_rate": 4.957627118644068e-06, | |
| "loss": 7.5451, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.12759170653907495, | |
| "grad_norm": 4.261996908905627, | |
| "learning_rate": 6.2288135593220344e-06, | |
| "loss": 6.2001, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.15311004784688995, | |
| "grad_norm": 2.9963531250158075, | |
| "learning_rate": 7.5e-06, | |
| "loss": 5.4325, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.17862838915470494, | |
| "grad_norm": 2.452719493622236, | |
| "learning_rate": 8.771186440677966e-06, | |
| "loss": 4.8224, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.20414673046251994, | |
| "grad_norm": 2.296924769718695, | |
| "learning_rate": 1.0042372881355933e-05, | |
| "loss": 4.3397, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.22966507177033493, | |
| "grad_norm": 2.02789705030659, | |
| "learning_rate": 1.1313559322033899e-05, | |
| "loss": 4.0637, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.2551834130781499, | |
| "grad_norm": 2.0102538064720865, | |
| "learning_rate": 1.2584745762711864e-05, | |
| "loss": 3.8254, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.2807017543859649, | |
| "grad_norm": 1.9988271290927235, | |
| "learning_rate": 1.385593220338983e-05, | |
| "loss": 3.6752, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.3062200956937799, | |
| "grad_norm": 1.9671546379874392, | |
| "learning_rate": 1.4999966747387387e-05, | |
| "loss": 3.5183, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.3317384370015949, | |
| "grad_norm": 2.3495671945712764, | |
| "learning_rate": 1.4995976790645575e-05, | |
| "loss": 3.4113, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.3572567783094099, | |
| "grad_norm": 2.4845114503377284, | |
| "learning_rate": 1.4985340365156019e-05, | |
| "loss": 3.3294, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.3827751196172249, | |
| "grad_norm": 2.2685356609887966, | |
| "learning_rate": 1.496806690193372e-05, | |
| "loss": 3.2443, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.4082934609250399, | |
| "grad_norm": 1.8552414857417485, | |
| "learning_rate": 1.494417171686568e-05, | |
| "loss": 3.1948, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.43381180223285487, | |
| "grad_norm": 2.5986726068440635, | |
| "learning_rate": 1.4913675997130732e-05, | |
| "loss": 3.1538, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.45933014354066987, | |
| "grad_norm": 2.3542672335324646, | |
| "learning_rate": 1.4876606782413468e-05, | |
| "loss": 3.0716, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.48484848484848486, | |
| "grad_norm": 2.1550658582198516, | |
| "learning_rate": 1.4832996940928936e-05, | |
| "loss": 3.038, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.5103668261562998, | |
| "grad_norm": 2.1753949765417206, | |
| "learning_rate": 1.4782885140279318e-05, | |
| "loss": 2.9772, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.5358851674641149, | |
| "grad_norm": 1.9692819011886733, | |
| "learning_rate": 1.4726315813168478e-05, | |
| "loss": 2.9469, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.5614035087719298, | |
| "grad_norm": 1.833984536566349, | |
| "learning_rate": 1.466333911800475e-05, | |
| "loss": 2.9032, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.5869218500797448, | |
| "grad_norm": 2.021073458068275, | |
| "learning_rate": 1.4594010894426907e-05, | |
| "loss": 2.8903, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.6124401913875598, | |
| "grad_norm": 1.8342599659729268, | |
| "learning_rate": 1.4518392613792747e-05, | |
| "loss": 2.8585, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.6379585326953748, | |
| "grad_norm": 1.8015476338333145, | |
| "learning_rate": 1.4436551324674196e-05, | |
| "loss": 2.8197, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.6634768740031898, | |
| "grad_norm": 1.774909204385973, | |
| "learning_rate": 1.4348559593407249e-05, | |
| "loss": 2.7953, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.6889952153110048, | |
| "grad_norm": 1.913402588745503, | |
| "learning_rate": 1.4254495439749472e-05, | |
| "loss": 2.7817, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.7145135566188198, | |
| "grad_norm": 1.8231676178000666, | |
| "learning_rate": 1.4154442267702106e-05, | |
| "loss": 2.7489, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.7400318979266348, | |
| "grad_norm": 1.8647896227394318, | |
| "learning_rate": 1.4048488791558127e-05, | |
| "loss": 2.7492, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.7655502392344498, | |
| "grad_norm": 1.7710269150416975, | |
| "learning_rate": 1.3936728957241805e-05, | |
| "loss": 2.7359, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.7910685805422647, | |
| "grad_norm": 1.7971354575124974, | |
| "learning_rate": 1.3819261859009536e-05, | |
| "loss": 2.682, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.8165869218500797, | |
| "grad_norm": 1.8543518497527265, | |
| "learning_rate": 1.3696191651585793e-05, | |
| "loss": 2.6693, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.8421052631578947, | |
| "grad_norm": 1.671172953201345, | |
| "learning_rate": 1.3567627457812107e-05, | |
| "loss": 2.6371, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.8676236044657097, | |
| "grad_norm": 1.793308965007255, | |
| "learning_rate": 1.3433683271890945e-05, | |
| "loss": 2.6105, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.8931419457735247, | |
| "grad_norm": 1.702891636275614, | |
| "learning_rate": 1.3294477858310318e-05, | |
| "loss": 2.5997, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.9186602870813397, | |
| "grad_norm": 1.8004220505203588, | |
| "learning_rate": 1.315013464653869e-05, | |
| "loss": 2.6038, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.9441786283891547, | |
| "grad_norm": 1.6762412987987798, | |
| "learning_rate": 1.300078162158359e-05, | |
| "loss": 2.5724, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.9696969696969697, | |
| "grad_norm": 1.8651248798134779, | |
| "learning_rate": 1.2846551210510956e-05, | |
| "loss": 2.572, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.9952153110047847, | |
| "grad_norm": 1.8166685018367434, | |
| "learning_rate": 1.2687580165025842e-05, | |
| "loss": 2.5526, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.0229665071770335, | |
| "grad_norm": 1.9595305580742408, | |
| "learning_rate": 1.2524009440218576e-05, | |
| "loss": 2.6489, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.0484848484848486, | |
| "grad_norm": 1.9448735030077777, | |
| "learning_rate": 1.2355984069583902e-05, | |
| "loss": 2.3534, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.0740031897926634, | |
| "grad_norm": 1.9568519341241308, | |
| "learning_rate": 1.218365303642392e-05, | |
| "loss": 2.3209, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.0995215311004785, | |
| "grad_norm": 1.97257201705927, | |
| "learning_rate": 1.200716914174885e-05, | |
| "loss": 2.3114, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.1250398724082935, | |
| "grad_norm": 2.0078778706593234, | |
| "learning_rate": 1.1826688868792723e-05, | |
| "loss": 2.3196, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.1505582137161086, | |
| "grad_norm": 1.9231486716564377, | |
| "learning_rate": 1.1642372244264167e-05, | |
| "loss": 2.2972, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.1760765550239234, | |
| "grad_norm": 1.9501054668769524, | |
| "learning_rate": 1.1454382696455303e-05, | |
| "loss": 2.2779, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.2015948963317384, | |
| "grad_norm": 1.8856831931474254, | |
| "learning_rate": 1.126288691033452e-05, | |
| "loss": 2.2627, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.2271132376395535, | |
| "grad_norm": 2.0213311293812573, | |
| "learning_rate": 1.106805467975168e-05, | |
| "loss": 2.2969, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.2526315789473683, | |
| "grad_norm": 1.9724633188826075, | |
| "learning_rate": 1.087005875688676e-05, | |
| "loss": 2.2616, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.2781499202551834, | |
| "grad_norm": 2.02716350744511, | |
| "learning_rate": 1.0669074699075414e-05, | |
| "loss": 2.2511, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.3036682615629984, | |
| "grad_norm": 2.0652621663353803, | |
| "learning_rate": 1.0465280713147303e-05, | |
| "loss": 2.2503, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.3291866028708135, | |
| "grad_norm": 2.033130261702073, | |
| "learning_rate": 1.0258857497415165e-05, | |
| "loss": 2.2372, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.3547049441786285, | |
| "grad_norm": 2.056888858914891, | |
| "learning_rate": 1.0049988081454802e-05, | |
| "loss": 2.2551, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.3802232854864434, | |
| "grad_norm": 2.0022262070955525, | |
| "learning_rate": 9.838857663817956e-06, | |
| "loss": 2.2482, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.4057416267942584, | |
| "grad_norm": 2.078167028605348, | |
| "learning_rate": 9.625653447822052e-06, | |
| "loss": 2.2707, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.4312599681020735, | |
| "grad_norm": 2.126752097583203, | |
| "learning_rate": 9.410564475562346e-06, | |
| "loss": 2.2295, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.4567783094098883, | |
| "grad_norm": 1.9806408754738973, | |
| "learning_rate": 9.193781460293695e-06, | |
| "loss": 2.1887, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.4822966507177033, | |
| "grad_norm": 2.0059818507445715, | |
| "learning_rate": 8.975496617330568e-06, | |
| "loss": 2.212, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.5078149920255184, | |
| "grad_norm": 1.958138514087202, | |
| "learning_rate": 8.755903493615193e-06, | |
| "loss": 2.2421, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.5333333333333332, | |
| "grad_norm": 1.9444870040248599, | |
| "learning_rate": 8.535196796105018e-06, | |
| "loss": 2.2001, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.5588516746411485, | |
| "grad_norm": 2.0790121588452677, | |
| "learning_rate": 8.3135722191316e-06, | |
| "loss": 2.1919, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.5843700159489633, | |
| "grad_norm": 2.09803077018995, | |
| "learning_rate": 8.09122627088403e-06, | |
| "loss": 2.181, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.6098883572567781, | |
| "grad_norm": 2.103389635373312, | |
| "learning_rate": 7.868356099170721e-06, | |
| "loss": 2.2188, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.6354066985645934, | |
| "grad_norm": 2.2123492409708, | |
| "learning_rate": 7.645159316614082e-06, | |
| "loss": 2.2191, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.6609250398724082, | |
| "grad_norm": 2.023742495950975, | |
| "learning_rate": 7.421833825433035e-06, | |
| "loss": 2.1925, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.6864433811802233, | |
| "grad_norm": 2.0082728157794567, | |
| "learning_rate": 7.198577641968766e-06, | |
| "loss": 2.1633, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.7119617224880384, | |
| "grad_norm": 2.1289624844092665, | |
| "learning_rate": 6.97558872110929e-06, | |
| "loss": 2.1864, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.7374800637958532, | |
| "grad_norm": 2.0363999925653395, | |
| "learning_rate": 6.753064780768488e-06, | |
| "loss": 2.1915, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.7629984051036682, | |
| "grad_norm": 2.1325211739700323, | |
| "learning_rate": 6.5312031265752946e-06, | |
| "loss": 2.1748, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.7885167464114833, | |
| "grad_norm": 2.073383778430727, | |
| "learning_rate": 6.310200476928403e-06, | |
| "loss": 2.1473, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.8140350877192981, | |
| "grad_norm": 2.0597448767291557, | |
| "learning_rate": 6.090252788571685e-06, | |
| "loss": 2.1668, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.8395534290271134, | |
| "grad_norm": 2.0432434933668033, | |
| "learning_rate": 5.87155508284494e-06, | |
| "loss": 2.1442, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.8650717703349282, | |
| "grad_norm": 2.0612494703787, | |
| "learning_rate": 5.654301272764029e-06, | |
| "loss": 2.1346, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.8905901116427433, | |
| "grad_norm": 2.0948535247185496, | |
| "learning_rate": 5.438683991083748e-06, | |
| "loss": 2.1436, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.9161084529505583, | |
| "grad_norm": 2.085105020269672, | |
| "learning_rate": 5.224894419495842e-06, | |
| "loss": 2.1523, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.9416267942583731, | |
| "grad_norm": 1.990712509803804, | |
| "learning_rate": 5.013122119113674e-06, | |
| "loss": 2.1595, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.9671451355661882, | |
| "grad_norm": 2.0676031547838565, | |
| "learning_rate": 4.803554862393779e-06, | |
| "loss": 2.1328, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.9926634768740032, | |
| "grad_norm": 2.0958105821364095, | |
| "learning_rate": 4.596378466643397e-06, | |
| "loss": 2.1694, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 2.0204146730462518, | |
| "grad_norm": 2.3390863338233348, | |
| "learning_rate": 4.391776629261555e-06, | |
| "loss": 2.126, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 2.045933014354067, | |
| "grad_norm": 2.168148536310652, | |
| "learning_rate": 4.1899307648598355e-06, | |
| "loss": 1.8096, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 2.071451355661882, | |
| "grad_norm": 2.107040319291286, | |
| "learning_rate": 3.991019844407206e-06, | |
| "loss": 1.7884, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 2.096969696969697, | |
| "grad_norm": 2.3752722791144096, | |
| "learning_rate": 3.7952202365415787e-06, | |
| "loss": 1.802, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 2.122488038277512, | |
| "grad_norm": 2.2622817947741636, | |
| "learning_rate": 3.602705551188764e-06, | |
| "loss": 1.7952, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 2.148006379585327, | |
| "grad_norm": 2.2685137603537058, | |
| "learning_rate": 3.4136464856275042e-06, | |
| "loss": 1.7794, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 2.173524720893142, | |
| "grad_norm": 2.3902445863895094, | |
| "learning_rate": 3.228210673137064e-06, | |
| "loss": 1.788, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 2.199043062200957, | |
| "grad_norm": 2.3611950118719314, | |
| "learning_rate": 3.0465625343615967e-06, | |
| "loss": 1.8184, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 2.2245614035087717, | |
| "grad_norm": 2.3628012509214322, | |
| "learning_rate": 2.8688631315230326e-06, | |
| "loss": 1.7802, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 2.250079744816587, | |
| "grad_norm": 2.274090723203016, | |
| "learning_rate": 2.6952700256118117e-06, | |
| "loss": 1.7664, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 2.275598086124402, | |
| "grad_norm": 2.4526876937461646, | |
| "learning_rate": 2.5259371366820507e-06, | |
| "loss": 1.8002, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 2.301116427432217, | |
| "grad_norm": 2.3336921883837616, | |
| "learning_rate": 2.3610146073750176e-06, | |
| "loss": 1.7763, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 2.326634768740032, | |
| "grad_norm": 2.271646264284567, | |
| "learning_rate": 2.200648669791939e-06, | |
| "loss": 1.7612, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 2.3521531100478468, | |
| "grad_norm": 2.4913302090031952, | |
| "learning_rate": 2.044981515834167e-06, | |
| "loss": 1.7946, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 2.377671451355662, | |
| "grad_norm": 2.439411059369649, | |
| "learning_rate": 1.8941511711256728e-06, | |
| "loss": 1.7869, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 2.403189792663477, | |
| "grad_norm": 2.3140599984104138, | |
| "learning_rate": 1.7482913726296778e-06, | |
| "loss": 1.7664, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 2.4287081339712917, | |
| "grad_norm": 2.500843401322422, | |
| "learning_rate": 1.607531450067886e-06, | |
| "loss": 1.7834, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 2.454226475279107, | |
| "grad_norm": 2.4192381369051033, | |
| "learning_rate": 1.4719962112475227e-06, | |
| "loss": 1.7711, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 2.479744816586922, | |
| "grad_norm": 2.381315270199767, | |
| "learning_rate": 1.3418058313978064e-06, | |
| "loss": 1.7727, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 2.5052631578947366, | |
| "grad_norm": 2.4411828126850454, | |
| "learning_rate": 1.2170757466140037e-06, | |
| "loss": 1.7536, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 2.530781499202552, | |
| "grad_norm": 2.4714371413125504, | |
| "learning_rate": 1.0979165515035402e-06, | |
| "loss": 1.7918, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 2.5562998405103667, | |
| "grad_norm": 2.3600482331288557, | |
| "learning_rate": 9.844339011249114e-07, | |
| "loss": 1.7683, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 2.5562998405103667, | |
| "eval_loss": 2.343059778213501, | |
| "eval_runtime": 20.7074, | |
| "eval_samples_per_second": 254.982, | |
| "eval_steps_per_second": 7.968, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 2.581818181818182, | |
| "grad_norm": 2.42847050890942, | |
| "learning_rate": 8.767284173063628e-07, | |
| "loss": 1.7602, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 2.607336523125997, | |
| "grad_norm": 2.495286470584022, | |
| "learning_rate": 7.748955994273793e-07, | |
| "loss": 1.7623, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 2.6328548644338117, | |
| "grad_norm": 2.380920410548109, | |
| "learning_rate": 6.790257397421221e-07, | |
| "loss": 1.7764, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 2.658373205741627, | |
| "grad_norm": 2.4026078375726305, | |
| "learning_rate": 5.892038433198463e-07, | |
| "loss": 1.7552, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 2.6838915470494418, | |
| "grad_norm": 2.4208779504436673, | |
| "learning_rate": 5.055095526733442e-07, | |
| "loss": 1.7624, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 2.709409888357257, | |
| "grad_norm": 2.349970978991061, | |
| "learning_rate": 4.280170771421957e-07, | |
| "loss": 1.7642, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 2.734928229665072, | |
| "grad_norm": 2.3526689883363248, | |
| "learning_rate": 3.5679512709346773e-07, | |
| "loss": 1.7577, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 2.7604465709728867, | |
| "grad_norm": 2.450274730203489, | |
| "learning_rate": 2.919068529981894e-07, | |
| "loss": 1.7654, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 2.7859649122807015, | |
| "grad_norm": 2.3651571766241246, | |
| "learning_rate": 2.3340978943763164e-07, | |
| "loss": 1.761, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 2.811483253588517, | |
| "grad_norm": 2.404642620983574, | |
| "learning_rate": 1.8135580408903674e-07, | |
| "loss": 1.7544, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 2.8370015948963316, | |
| "grad_norm": 2.473317448160841, | |
| "learning_rate": 1.357910517360264e-07, | |
| "loss": 1.7586, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 2.862519936204147, | |
| "grad_norm": 2.325829716379361, | |
| "learning_rate": 9.675593334447674e-08, | |
| "loss": 1.7541, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 2.8880382775119617, | |
| "grad_norm": 2.430438279086189, | |
| "learning_rate": 6.428506024012904e-08, | |
| "loss": 1.77, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 2.9135566188197766, | |
| "grad_norm": 2.3903497741918414, | |
| "learning_rate": 3.840722341971664e-08, | |
| "loss": 1.7565, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 2.939074960127592, | |
| "grad_norm": 2.3801298605080023, | |
| "learning_rate": 1.914536802281208e-08, | |
| "loss": 1.7433, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 2.9645933014354067, | |
| "grad_norm": 2.364114165792993, | |
| "learning_rate": 6.516572987022851e-09, | |
| "loss": 1.7701, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 2.990111642743222, | |
| "grad_norm": 2.2932824036528032, | |
| "learning_rate": 5.320359045882239e-10, | |
| "loss": 1.7531, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 2.9977671451355663, | |
| "step": 1173, | |
| "total_flos": 102874311622656.0, | |
| "train_loss": 2.6594049156169453, | |
| "train_runtime": 4231.9475, | |
| "train_samples_per_second": 71.111, | |
| "train_steps_per_second": 0.277 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 1173, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 1000, | |
| "stateful_callbacks": { | |
| "EarlyStoppingCallback": { | |
| "args": { | |
| "early_stopping_patience": 3000, | |
| "early_stopping_threshold": 0.0 | |
| }, | |
| "attributes": { | |
| "early_stopping_patience_counter": 0 | |
| } | |
| }, | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 102874311622656.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |