GENOME-gemma-2b-it / wizardlm /trainer_state.json
Estwld's picture
Upload 15 files
8156862 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 5.0,
"eval_steps": 1000,
"global_step": 4425,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.011299435028248588,
"grad_norm": 0.7552605867385864,
"learning_rate": 4.5146726862302485e-06,
"loss": 1.0465,
"step": 10
},
{
"epoch": 0.022598870056497175,
"grad_norm": 0.8854597210884094,
"learning_rate": 9.029345372460497e-06,
"loss": 1.0939,
"step": 20
},
{
"epoch": 0.03389830508474576,
"grad_norm": 0.7813611030578613,
"learning_rate": 1.3544018058690747e-05,
"loss": 1.0292,
"step": 30
},
{
"epoch": 0.04519774011299435,
"grad_norm": 0.6397517919540405,
"learning_rate": 1.8058690744920994e-05,
"loss": 0.995,
"step": 40
},
{
"epoch": 0.05649717514124294,
"grad_norm": 0.4727611839771271,
"learning_rate": 2.2573363431151244e-05,
"loss": 0.896,
"step": 50
},
{
"epoch": 0.06779661016949153,
"grad_norm": 0.37453365325927734,
"learning_rate": 2.7088036117381494e-05,
"loss": 0.8037,
"step": 60
},
{
"epoch": 0.07909604519774012,
"grad_norm": 0.5303831696510315,
"learning_rate": 3.1602708803611745e-05,
"loss": 0.7872,
"step": 70
},
{
"epoch": 0.0903954802259887,
"grad_norm": 0.3839394152164459,
"learning_rate": 3.611738148984199e-05,
"loss": 0.7697,
"step": 80
},
{
"epoch": 0.1016949152542373,
"grad_norm": 0.39426788687705994,
"learning_rate": 4.063205417607224e-05,
"loss": 0.7481,
"step": 90
},
{
"epoch": 0.11299435028248588,
"grad_norm": 0.29685771465301514,
"learning_rate": 4.514672686230249e-05,
"loss": 0.7166,
"step": 100
},
{
"epoch": 0.12429378531073447,
"grad_norm": 0.3550776243209839,
"learning_rate": 4.966139954853273e-05,
"loss": 0.7455,
"step": 110
},
{
"epoch": 0.13559322033898305,
"grad_norm": 0.49571502208709717,
"learning_rate": 5.417607223476299e-05,
"loss": 0.7451,
"step": 120
},
{
"epoch": 0.14689265536723164,
"grad_norm": 0.3369849622249603,
"learning_rate": 5.869074492099323e-05,
"loss": 0.7235,
"step": 130
},
{
"epoch": 0.15819209039548024,
"grad_norm": 0.37830305099487305,
"learning_rate": 6.320541760722349e-05,
"loss": 0.7201,
"step": 140
},
{
"epoch": 0.1694915254237288,
"grad_norm": 0.42834165692329407,
"learning_rate": 6.772009029345373e-05,
"loss": 0.7199,
"step": 150
},
{
"epoch": 0.1807909604519774,
"grad_norm": 0.3480307459831238,
"learning_rate": 7.223476297968398e-05,
"loss": 0.7134,
"step": 160
},
{
"epoch": 0.192090395480226,
"grad_norm": 0.47217118740081787,
"learning_rate": 7.674943566591423e-05,
"loss": 0.7193,
"step": 170
},
{
"epoch": 0.2033898305084746,
"grad_norm": 0.43827682733535767,
"learning_rate": 8.126410835214448e-05,
"loss": 0.7049,
"step": 180
},
{
"epoch": 0.21468926553672316,
"grad_norm": 0.38405266404151917,
"learning_rate": 8.577878103837473e-05,
"loss": 0.719,
"step": 190
},
{
"epoch": 0.22598870056497175,
"grad_norm": 0.48938581347465515,
"learning_rate": 9.029345372460498e-05,
"loss": 0.7077,
"step": 200
},
{
"epoch": 0.23728813559322035,
"grad_norm": 0.4330425560474396,
"learning_rate": 9.480812641083521e-05,
"loss": 0.7301,
"step": 210
},
{
"epoch": 0.24858757062146894,
"grad_norm": 0.3442066013813019,
"learning_rate": 9.932279909706546e-05,
"loss": 0.7174,
"step": 220
},
{
"epoch": 0.2598870056497175,
"grad_norm": 0.41479647159576416,
"learning_rate": 0.00010383747178329573,
"loss": 0.7163,
"step": 230
},
{
"epoch": 0.2711864406779661,
"grad_norm": 0.4269746243953705,
"learning_rate": 0.00010835214446952598,
"loss": 0.7206,
"step": 240
},
{
"epoch": 0.2824858757062147,
"grad_norm": 0.4595146179199219,
"learning_rate": 0.00011286681715575623,
"loss": 0.7198,
"step": 250
},
{
"epoch": 0.2937853107344633,
"grad_norm": 0.459258496761322,
"learning_rate": 0.00011738148984198646,
"loss": 0.6791,
"step": 260
},
{
"epoch": 0.3050847457627119,
"grad_norm": 0.40073147416114807,
"learning_rate": 0.00012189616252821671,
"loss": 0.7379,
"step": 270
},
{
"epoch": 0.3163841807909605,
"grad_norm": 0.4646053910255432,
"learning_rate": 0.00012641083521444698,
"loss": 0.7146,
"step": 280
},
{
"epoch": 0.327683615819209,
"grad_norm": 0.5096935629844666,
"learning_rate": 0.00013092550790067722,
"loss": 0.6855,
"step": 290
},
{
"epoch": 0.3389830508474576,
"grad_norm": 0.37696900963783264,
"learning_rate": 0.00013544018058690745,
"loss": 0.7169,
"step": 300
},
{
"epoch": 0.3502824858757062,
"grad_norm": 0.344178169965744,
"learning_rate": 0.00013995485327313772,
"loss": 0.7079,
"step": 310
},
{
"epoch": 0.3615819209039548,
"grad_norm": 0.33671554923057556,
"learning_rate": 0.00014446952595936795,
"loss": 0.68,
"step": 320
},
{
"epoch": 0.3728813559322034,
"grad_norm": 0.365024596452713,
"learning_rate": 0.00014898419864559822,
"loss": 0.7247,
"step": 330
},
{
"epoch": 0.384180790960452,
"grad_norm": 0.36559364199638367,
"learning_rate": 0.00015349887133182845,
"loss": 0.6753,
"step": 340
},
{
"epoch": 0.3954802259887006,
"grad_norm": 0.3757419288158417,
"learning_rate": 0.0001580135440180587,
"loss": 0.6616,
"step": 350
},
{
"epoch": 0.4067796610169492,
"grad_norm": 0.5549542307853699,
"learning_rate": 0.00016252821670428895,
"loss": 0.686,
"step": 360
},
{
"epoch": 0.4180790960451977,
"grad_norm": 0.3681774139404297,
"learning_rate": 0.0001670428893905192,
"loss": 0.7086,
"step": 370
},
{
"epoch": 0.4293785310734463,
"grad_norm": 0.3036465346813202,
"learning_rate": 0.00017155756207674945,
"loss": 0.6966,
"step": 380
},
{
"epoch": 0.4406779661016949,
"grad_norm": 0.32314741611480713,
"learning_rate": 0.0001760722347629797,
"loss": 0.7406,
"step": 390
},
{
"epoch": 0.4519774011299435,
"grad_norm": 0.41190987825393677,
"learning_rate": 0.00018058690744920995,
"loss": 0.6772,
"step": 400
},
{
"epoch": 0.4632768361581921,
"grad_norm": 0.28181225061416626,
"learning_rate": 0.0001851015801354402,
"loss": 0.6843,
"step": 410
},
{
"epoch": 0.4745762711864407,
"grad_norm": 0.2925048768520355,
"learning_rate": 0.00018961625282167043,
"loss": 0.7206,
"step": 420
},
{
"epoch": 0.4858757062146893,
"grad_norm": 0.33646613359451294,
"learning_rate": 0.0001941309255079007,
"loss": 0.6879,
"step": 430
},
{
"epoch": 0.4971751412429379,
"grad_norm": 0.40007829666137695,
"learning_rate": 0.00019864559819413093,
"loss": 0.7032,
"step": 440
},
{
"epoch": 0.5084745762711864,
"grad_norm": 0.29928356409072876,
"learning_rate": 0.00019999847502678901,
"loss": 0.7086,
"step": 450
},
{
"epoch": 0.519774011299435,
"grad_norm": 0.285306453704834,
"learning_rate": 0.00019999100588221217,
"loss": 0.6709,
"step": 460
},
{
"epoch": 0.5310734463276836,
"grad_norm": 0.5055399537086487,
"learning_rate": 0.00019997731293347473,
"loss": 0.6901,
"step": 470
},
{
"epoch": 0.5423728813559322,
"grad_norm": 0.3232295513153076,
"learning_rate": 0.00019995739703287558,
"loss": 0.6748,
"step": 480
},
{
"epoch": 0.5536723163841808,
"grad_norm": 0.40595105290412903,
"learning_rate": 0.00019993125942005297,
"loss": 0.6908,
"step": 490
},
{
"epoch": 0.5649717514124294,
"grad_norm": 0.29886409640312195,
"learning_rate": 0.00019989890172190697,
"loss": 0.6961,
"step": 500
},
{
"epoch": 0.576271186440678,
"grad_norm": 0.3396940231323242,
"learning_rate": 0.00019986032595249855,
"loss": 0.6741,
"step": 510
},
{
"epoch": 0.5875706214689266,
"grad_norm": 0.3772926926612854,
"learning_rate": 0.00019981553451292396,
"loss": 0.6681,
"step": 520
},
{
"epoch": 0.5988700564971752,
"grad_norm": 0.3138965964317322,
"learning_rate": 0.00019976453019116547,
"loss": 0.7027,
"step": 530
},
{
"epoch": 0.6101694915254238,
"grad_norm": 0.31275707483291626,
"learning_rate": 0.00019970731616191772,
"loss": 0.6955,
"step": 540
},
{
"epoch": 0.6214689265536724,
"grad_norm": 0.28057295083999634,
"learning_rate": 0.00019964389598639012,
"loss": 0.6617,
"step": 550
},
{
"epoch": 0.632768361581921,
"grad_norm": 0.5602757930755615,
"learning_rate": 0.00019957427361208522,
"loss": 0.6896,
"step": 560
},
{
"epoch": 0.6440677966101694,
"grad_norm": 0.43884241580963135,
"learning_rate": 0.00019949845337255306,
"loss": 0.6786,
"step": 570
},
{
"epoch": 0.655367231638418,
"grad_norm": 0.3063630759716034,
"learning_rate": 0.00019941643998712128,
"loss": 0.685,
"step": 580
},
{
"epoch": 0.6666666666666666,
"grad_norm": 0.2912139594554901,
"learning_rate": 0.00019932823856060153,
"loss": 0.691,
"step": 590
},
{
"epoch": 0.6779661016949152,
"grad_norm": 0.5503478646278381,
"learning_rate": 0.00019923385458297168,
"loss": 0.7076,
"step": 600
},
{
"epoch": 0.6892655367231638,
"grad_norm": 0.30678126215934753,
"learning_rate": 0.00019913329392903396,
"loss": 0.6628,
"step": 610
},
{
"epoch": 0.7005649717514124,
"grad_norm": 0.341691792011261,
"learning_rate": 0.00019902656285804954,
"loss": 0.6973,
"step": 620
},
{
"epoch": 0.711864406779661,
"grad_norm": 0.2795998454093933,
"learning_rate": 0.0001989136680133488,
"loss": 0.668,
"step": 630
},
{
"epoch": 0.7231638418079096,
"grad_norm": 0.3217742443084717,
"learning_rate": 0.00019879461642191774,
"loss": 0.6839,
"step": 640
},
{
"epoch": 0.7344632768361582,
"grad_norm": 0.2668191194534302,
"learning_rate": 0.00019866941549396075,
"loss": 0.6929,
"step": 650
},
{
"epoch": 0.7457627118644068,
"grad_norm": 0.8005201816558838,
"learning_rate": 0.0001985380730224393,
"loss": 0.701,
"step": 660
},
{
"epoch": 0.7570621468926554,
"grad_norm": 0.29280129075050354,
"learning_rate": 0.00019840059718258693,
"loss": 0.7222,
"step": 670
},
{
"epoch": 0.768361581920904,
"grad_norm": 0.3449852764606476,
"learning_rate": 0.00019825699653140032,
"loss": 0.6919,
"step": 680
},
{
"epoch": 0.7796610169491526,
"grad_norm": 0.2684054970741272,
"learning_rate": 0.0001981072800071066,
"loss": 0.6859,
"step": 690
},
{
"epoch": 0.7909604519774012,
"grad_norm": 0.3716861307621002,
"learning_rate": 0.00019795145692860726,
"loss": 0.666,
"step": 700
},
{
"epoch": 0.8022598870056498,
"grad_norm": 0.3001028597354889,
"learning_rate": 0.00019778953699489785,
"loss": 0.6587,
"step": 710
},
{
"epoch": 0.8135593220338984,
"grad_norm": 0.476034551858902,
"learning_rate": 0.0001976215302844644,
"loss": 0.6766,
"step": 720
},
{
"epoch": 0.8248587570621468,
"grad_norm": 0.3516661524772644,
"learning_rate": 0.00019744744725465617,
"loss": 0.6836,
"step": 730
},
{
"epoch": 0.8361581920903954,
"grad_norm": 0.275155246257782,
"learning_rate": 0.00019726729874103448,
"loss": 0.6729,
"step": 740
},
{
"epoch": 0.847457627118644,
"grad_norm": 0.31217560172080994,
"learning_rate": 0.00019708109595669858,
"loss": 0.6811,
"step": 750
},
{
"epoch": 0.8587570621468926,
"grad_norm": 0.3042563199996948,
"learning_rate": 0.00019688885049158752,
"loss": 0.7353,
"step": 760
},
{
"epoch": 0.8700564971751412,
"grad_norm": 0.3564906120300293,
"learning_rate": 0.0001966905743117588,
"loss": 0.6938,
"step": 770
},
{
"epoch": 0.8813559322033898,
"grad_norm": 0.3310820162296295,
"learning_rate": 0.00019648627975864355,
"loss": 0.6724,
"step": 780
},
{
"epoch": 0.8926553672316384,
"grad_norm": 0.4258658289909363,
"learning_rate": 0.00019627597954827833,
"loss": 0.6901,
"step": 790
},
{
"epoch": 0.903954802259887,
"grad_norm": 0.3146076202392578,
"learning_rate": 0.00019605968677051378,
"loss": 0.7008,
"step": 800
},
{
"epoch": 0.9152542372881356,
"grad_norm": 0.3338530957698822,
"learning_rate": 0.0001958374148881996,
"loss": 0.6567,
"step": 810
},
{
"epoch": 0.9265536723163842,
"grad_norm": 0.3822636008262634,
"learning_rate": 0.00019560917773634676,
"loss": 0.6756,
"step": 820
},
{
"epoch": 0.9378531073446328,
"grad_norm": 0.3074444830417633,
"learning_rate": 0.00019537498952126642,
"loss": 0.6958,
"step": 830
},
{
"epoch": 0.9491525423728814,
"grad_norm": 0.31832125782966614,
"learning_rate": 0.00019513486481968545,
"loss": 0.6871,
"step": 840
},
{
"epoch": 0.96045197740113,
"grad_norm": 0.3670668303966522,
"learning_rate": 0.00019488881857783935,
"loss": 0.6956,
"step": 850
},
{
"epoch": 0.9717514124293786,
"grad_norm": 0.2784234881401062,
"learning_rate": 0.00019463686611054172,
"loss": 0.6865,
"step": 860
},
{
"epoch": 0.9830508474576272,
"grad_norm": 0.3210557997226715,
"learning_rate": 0.00019437902310023126,
"loss": 0.6845,
"step": 870
},
{
"epoch": 0.9943502824858758,
"grad_norm": 0.3313763737678528,
"learning_rate": 0.0001941153055959954,
"loss": 0.6589,
"step": 880
},
{
"epoch": 1.0056497175141244,
"grad_norm": 0.35079240798950195,
"learning_rate": 0.00019384573001257155,
"loss": 0.671,
"step": 890
},
{
"epoch": 1.0169491525423728,
"grad_norm": 0.3715152442455292,
"learning_rate": 0.0001935703131293252,
"loss": 0.608,
"step": 900
},
{
"epoch": 1.0282485875706215,
"grad_norm": 0.3646804988384247,
"learning_rate": 0.00019328907208920567,
"loss": 0.6457,
"step": 910
},
{
"epoch": 1.03954802259887,
"grad_norm": 0.3106479048728943,
"learning_rate": 0.00019300202439767898,
"loss": 0.6105,
"step": 920
},
{
"epoch": 1.0508474576271187,
"grad_norm": 0.29659685492515564,
"learning_rate": 0.00019270918792163827,
"loss": 0.6221,
"step": 930
},
{
"epoch": 1.0621468926553672,
"grad_norm": 0.4973548948764801,
"learning_rate": 0.00019241058088829166,
"loss": 0.6465,
"step": 940
},
{
"epoch": 1.073446327683616,
"grad_norm": 0.4067288637161255,
"learning_rate": 0.00019210622188402788,
"loss": 0.6374,
"step": 950
},
{
"epoch": 1.0847457627118644,
"grad_norm": 0.4106499254703522,
"learning_rate": 0.00019179612985325908,
"loss": 0.6226,
"step": 960
},
{
"epoch": 1.0960451977401129,
"grad_norm": 0.31531280279159546,
"learning_rate": 0.000191480324097242,
"loss": 0.6604,
"step": 970
},
{
"epoch": 1.1073446327683616,
"grad_norm": 0.3855787515640259,
"learning_rate": 0.00019115882427287638,
"loss": 0.6466,
"step": 980
},
{
"epoch": 1.11864406779661,
"grad_norm": 0.36246171593666077,
"learning_rate": 0.00019083165039148153,
"loss": 0.6315,
"step": 990
},
{
"epoch": 1.1299435028248588,
"grad_norm": 0.4511447846889496,
"learning_rate": 0.00019049882281755066,
"loss": 0.6341,
"step": 1000
},
{
"epoch": 1.1299435028248588,
"eval_loss": 0.7100452184677124,
"eval_runtime": 16.7354,
"eval_samples_per_second": 89.093,
"eval_steps_per_second": 11.174,
"step": 1000
},
{
"epoch": 1.1412429378531073,
"grad_norm": 0.39097025990486145,
"learning_rate": 0.00019016036226748346,
"loss": 0.6144,
"step": 1010
},
{
"epoch": 1.152542372881356,
"grad_norm": 0.40734922885894775,
"learning_rate": 0.00018981628980829652,
"loss": 0.6458,
"step": 1020
},
{
"epoch": 1.1638418079096045,
"grad_norm": 0.3473665118217468,
"learning_rate": 0.00018946662685631203,
"loss": 0.6417,
"step": 1030
},
{
"epoch": 1.1751412429378532,
"grad_norm": 0.41758203506469727,
"learning_rate": 0.0001891113951758249,
"loss": 0.6355,
"step": 1040
},
{
"epoch": 1.1864406779661016,
"grad_norm": 0.6268119215965271,
"learning_rate": 0.0001887506168777479,
"loss": 0.6332,
"step": 1050
},
{
"epoch": 1.1977401129943503,
"grad_norm": 0.3336067795753479,
"learning_rate": 0.00018838431441823547,
"loss": 0.6445,
"step": 1060
},
{
"epoch": 1.2090395480225988,
"grad_norm": 0.3316787779331207,
"learning_rate": 0.00018801251059728604,
"loss": 0.6314,
"step": 1070
},
{
"epoch": 1.2203389830508475,
"grad_norm": 0.304683119058609,
"learning_rate": 0.00018763522855732266,
"loss": 0.6232,
"step": 1080
},
{
"epoch": 1.231638418079096,
"grad_norm": 0.34174299240112305,
"learning_rate": 0.00018725249178175286,
"loss": 0.6561,
"step": 1090
},
{
"epoch": 1.2429378531073447,
"grad_norm": 0.3202422559261322,
"learning_rate": 0.0001868643240935066,
"loss": 0.6066,
"step": 1100
},
{
"epoch": 1.2542372881355932,
"grad_norm": 0.3420744836330414,
"learning_rate": 0.0001864707496535537,
"loss": 0.6367,
"step": 1110
},
{
"epoch": 1.2655367231638417,
"grad_norm": 0.38407158851623535,
"learning_rate": 0.00018607179295939983,
"loss": 0.6278,
"step": 1120
},
{
"epoch": 1.2768361581920904,
"grad_norm": 0.37982526421546936,
"learning_rate": 0.00018566747884356182,
"loss": 0.6209,
"step": 1130
},
{
"epoch": 1.288135593220339,
"grad_norm": 0.33808642625808716,
"learning_rate": 0.00018525783247202188,
"loss": 0.6525,
"step": 1140
},
{
"epoch": 1.2994350282485876,
"grad_norm": 0.45425844192504883,
"learning_rate": 0.00018484287934266122,
"loss": 0.6367,
"step": 1150
},
{
"epoch": 1.310734463276836,
"grad_norm": 0.422880083322525,
"learning_rate": 0.00018442264528367304,
"loss": 0.6499,
"step": 1160
},
{
"epoch": 1.3220338983050848,
"grad_norm": 0.3406778573989868,
"learning_rate": 0.0001839971564519547,
"loss": 0.6478,
"step": 1170
},
{
"epoch": 1.3333333333333333,
"grad_norm": 0.37413087487220764,
"learning_rate": 0.00018356643933147986,
"loss": 0.6382,
"step": 1180
},
{
"epoch": 1.344632768361582,
"grad_norm": 0.3171542286872864,
"learning_rate": 0.00018313052073164982,
"loss": 0.6373,
"step": 1190
},
{
"epoch": 1.3559322033898304,
"grad_norm": 0.3208696246147156,
"learning_rate": 0.00018268942778562494,
"loss": 0.6231,
"step": 1200
},
{
"epoch": 1.3672316384180792,
"grad_norm": 0.37951213121414185,
"learning_rate": 0.0001822431879486357,
"loss": 0.6659,
"step": 1210
},
{
"epoch": 1.3785310734463276,
"grad_norm": 0.3654930591583252,
"learning_rate": 0.00018179182899627386,
"loss": 0.6467,
"step": 1220
},
{
"epoch": 1.3898305084745763,
"grad_norm": 0.5204548239707947,
"learning_rate": 0.00018133537902276342,
"loss": 0.6479,
"step": 1230
},
{
"epoch": 1.4011299435028248,
"grad_norm": 0.33258357644081116,
"learning_rate": 0.0001808738664392122,
"loss": 0.613,
"step": 1240
},
{
"epoch": 1.4124293785310735,
"grad_norm": 0.3374476134777069,
"learning_rate": 0.0001804073199718432,
"loss": 0.6308,
"step": 1250
},
{
"epoch": 1.423728813559322,
"grad_norm": 0.3677334189414978,
"learning_rate": 0.0001799357686602068,
"loss": 0.6412,
"step": 1260
},
{
"epoch": 1.4350282485875705,
"grad_norm": 0.503189742565155,
"learning_rate": 0.00017945924185537295,
"loss": 0.6242,
"step": 1270
},
{
"epoch": 1.4463276836158192,
"grad_norm": 0.44822338223457336,
"learning_rate": 0.00017897776921810453,
"loss": 0.6269,
"step": 1280
},
{
"epoch": 1.457627118644068,
"grad_norm": 0.38298189640045166,
"learning_rate": 0.00017849138071701092,
"loss": 0.6206,
"step": 1290
},
{
"epoch": 1.4689265536723164,
"grad_norm": 0.40119263529777527,
"learning_rate": 0.0001780001066266829,
"loss": 0.6398,
"step": 1300
},
{
"epoch": 1.4802259887005649,
"grad_norm": 0.5280100703239441,
"learning_rate": 0.00017750397752580796,
"loss": 0.6299,
"step": 1310
},
{
"epoch": 1.4915254237288136,
"grad_norm": 0.35528528690338135,
"learning_rate": 0.00017700302429526723,
"loss": 0.6511,
"step": 1320
},
{
"epoch": 1.5028248587570623,
"grad_norm": 0.3741821348667145,
"learning_rate": 0.00017649727811621315,
"loss": 0.6564,
"step": 1330
},
{
"epoch": 1.5141242937853108,
"grad_norm": 0.5371260046958923,
"learning_rate": 0.0001759867704681288,
"loss": 0.6186,
"step": 1340
},
{
"epoch": 1.5254237288135593,
"grad_norm": 0.3590337336063385,
"learning_rate": 0.00017547153312686832,
"loss": 0.6372,
"step": 1350
},
{
"epoch": 1.536723163841808,
"grad_norm": 0.34499940276145935,
"learning_rate": 0.00017495159816267929,
"loss": 0.6442,
"step": 1360
},
{
"epoch": 1.5480225988700564,
"grad_norm": 0.45834365487098694,
"learning_rate": 0.00017442699793820631,
"loss": 0.6363,
"step": 1370
},
{
"epoch": 1.559322033898305,
"grad_norm": 0.3389630615711212,
"learning_rate": 0.00017389776510647688,
"loss": 0.6399,
"step": 1380
},
{
"epoch": 1.5706214689265536,
"grad_norm": 1.0764904022216797,
"learning_rate": 0.0001733639326088688,
"loss": 0.6073,
"step": 1390
},
{
"epoch": 1.5819209039548023,
"grad_norm": 0.32130682468414307,
"learning_rate": 0.00017282553367305975,
"loss": 0.65,
"step": 1400
},
{
"epoch": 1.5932203389830508,
"grad_norm": 0.5010311603546143,
"learning_rate": 0.00017228260181095925,
"loss": 0.626,
"step": 1410
},
{
"epoch": 1.6045197740112993,
"grad_norm": 0.3811778128147125,
"learning_rate": 0.00017173517081662255,
"loss": 0.6565,
"step": 1420
},
{
"epoch": 1.615819209039548,
"grad_norm": 0.793699324131012,
"learning_rate": 0.00017118327476414735,
"loss": 0.6555,
"step": 1430
},
{
"epoch": 1.6271186440677967,
"grad_norm": 0.33635199069976807,
"learning_rate": 0.0001706269480055528,
"loss": 0.6431,
"step": 1440
},
{
"epoch": 1.6384180790960452,
"grad_norm": 0.46125972270965576,
"learning_rate": 0.00017006622516864127,
"loss": 0.6383,
"step": 1450
},
{
"epoch": 1.6497175141242937,
"grad_norm": 0.37351667881011963,
"learning_rate": 0.00016950114115484315,
"loss": 0.6304,
"step": 1460
},
{
"epoch": 1.6610169491525424,
"grad_norm": 0.35749056935310364,
"learning_rate": 0.0001689317311370442,
"loss": 0.6647,
"step": 1470
},
{
"epoch": 1.672316384180791,
"grad_norm": 0.3775118887424469,
"learning_rate": 0.0001683580305573966,
"loss": 0.6288,
"step": 1480
},
{
"epoch": 1.6836158192090396,
"grad_norm": 0.3228208124637604,
"learning_rate": 0.0001677800751251126,
"loss": 0.6067,
"step": 1490
},
{
"epoch": 1.694915254237288,
"grad_norm": 0.38675859570503235,
"learning_rate": 0.00016719790081424192,
"loss": 0.6438,
"step": 1500
},
{
"epoch": 1.7062146892655368,
"grad_norm": 0.3832201063632965,
"learning_rate": 0.0001666115438614328,
"loss": 0.6269,
"step": 1510
},
{
"epoch": 1.7175141242937855,
"grad_norm": 0.5429854989051819,
"learning_rate": 0.00016602104076367624,
"loss": 0.6397,
"step": 1520
},
{
"epoch": 1.7288135593220337,
"grad_norm": 0.4048645496368408,
"learning_rate": 0.0001654264282760343,
"loss": 0.6402,
"step": 1530
},
{
"epoch": 1.7401129943502824,
"grad_norm": 0.3748385012149811,
"learning_rate": 0.00016482774340935267,
"loss": 0.626,
"step": 1540
},
{
"epoch": 1.7514124293785311,
"grad_norm": 0.44016194343566895,
"learning_rate": 0.00016422502342795648,
"loss": 0.6542,
"step": 1550
},
{
"epoch": 1.7627118644067796,
"grad_norm": 0.34061527252197266,
"learning_rate": 0.00016361830584733124,
"loss": 0.6472,
"step": 1560
},
{
"epoch": 1.774011299435028,
"grad_norm": 0.3848430812358856,
"learning_rate": 0.00016300762843178754,
"loss": 0.6304,
"step": 1570
},
{
"epoch": 1.7853107344632768,
"grad_norm": 0.3756044805049896,
"learning_rate": 0.00016239302919211052,
"loss": 0.6161,
"step": 1580
},
{
"epoch": 1.7966101694915255,
"grad_norm": 0.40179935097694397,
"learning_rate": 0.0001617745463831939,
"loss": 0.6428,
"step": 1590
},
{
"epoch": 1.807909604519774,
"grad_norm": 0.4611279368400574,
"learning_rate": 0.00016115221850165892,
"loss": 0.6661,
"step": 1600
},
{
"epoch": 1.8192090395480225,
"grad_norm": 0.5279085636138916,
"learning_rate": 0.0001605260842834581,
"loss": 0.6293,
"step": 1610
},
{
"epoch": 1.8305084745762712,
"grad_norm": 0.5711485743522644,
"learning_rate": 0.00015989618270146423,
"loss": 0.6368,
"step": 1620
},
{
"epoch": 1.84180790960452,
"grad_norm": 0.37090200185775757,
"learning_rate": 0.00015926255296304454,
"loss": 0.6297,
"step": 1630
},
{
"epoch": 1.8531073446327684,
"grad_norm": 0.3836909830570221,
"learning_rate": 0.00015862523450762032,
"loss": 0.6817,
"step": 1640
},
{
"epoch": 1.8644067796610169,
"grad_norm": 0.4297361969947815,
"learning_rate": 0.00015798426700421196,
"loss": 0.6586,
"step": 1650
},
{
"epoch": 1.8757062146892656,
"grad_norm": 0.3696504235267639,
"learning_rate": 0.0001573396903489699,
"loss": 0.6466,
"step": 1660
},
{
"epoch": 1.8870056497175143,
"grad_norm": 0.39131200313568115,
"learning_rate": 0.00015669154466269137,
"loss": 0.6448,
"step": 1670
},
{
"epoch": 1.8983050847457628,
"grad_norm": 0.3583720624446869,
"learning_rate": 0.00015603987028832303,
"loss": 0.6385,
"step": 1680
},
{
"epoch": 1.9096045197740112,
"grad_norm": 0.40514451265335083,
"learning_rate": 0.00015538470778844995,
"loss": 0.6368,
"step": 1690
},
{
"epoch": 1.92090395480226,
"grad_norm": 0.4644288122653961,
"learning_rate": 0.00015472609794277083,
"loss": 0.6341,
"step": 1700
},
{
"epoch": 1.9322033898305084,
"grad_norm": 0.6364967823028564,
"learning_rate": 0.00015406408174555976,
"loss": 0.619,
"step": 1710
},
{
"epoch": 1.943502824858757,
"grad_norm": 0.3623166084289551,
"learning_rate": 0.00015339870040311448,
"loss": 0.6591,
"step": 1720
},
{
"epoch": 1.9548022598870056,
"grad_norm": 0.39671990275382996,
"learning_rate": 0.00015272999533119162,
"loss": 0.645,
"step": 1730
},
{
"epoch": 1.9661016949152543,
"grad_norm": 0.45107388496398926,
"learning_rate": 0.0001520580081524289,
"loss": 0.625,
"step": 1740
},
{
"epoch": 1.9774011299435028,
"grad_norm": 0.4473719596862793,
"learning_rate": 0.00015138278069375423,
"loss": 0.6481,
"step": 1750
},
{
"epoch": 1.9887005649717513,
"grad_norm": 0.390774130821228,
"learning_rate": 0.00015070435498378243,
"loss": 0.6482,
"step": 1760
},
{
"epoch": 2.0,
"grad_norm": 0.41401803493499756,
"learning_rate": 0.00015002277325019904,
"loss": 0.623,
"step": 1770
},
{
"epoch": 2.0112994350282487,
"grad_norm": 0.7757445573806763,
"learning_rate": 0.00014933807791713203,
"loss": 0.573,
"step": 1780
},
{
"epoch": 2.022598870056497,
"grad_norm": 0.44817861914634705,
"learning_rate": 0.00014865031160251116,
"loss": 0.6054,
"step": 1790
},
{
"epoch": 2.0338983050847457,
"grad_norm": 0.4326503574848175,
"learning_rate": 0.0001479595171154152,
"loss": 0.5724,
"step": 1800
},
{
"epoch": 2.0451977401129944,
"grad_norm": 0.4020352065563202,
"learning_rate": 0.00014726573745340747,
"loss": 0.5536,
"step": 1810
},
{
"epoch": 2.056497175141243,
"grad_norm": 0.3883920907974243,
"learning_rate": 0.00014656901579985934,
"loss": 0.5591,
"step": 1820
},
{
"epoch": 2.0677966101694913,
"grad_norm": 0.4672842025756836,
"learning_rate": 0.00014586939552126255,
"loss": 0.5467,
"step": 1830
},
{
"epoch": 2.07909604519774,
"grad_norm": 0.5037193894386292,
"learning_rate": 0.0001451669201645298,
"loss": 0.5548,
"step": 1840
},
{
"epoch": 2.0903954802259888,
"grad_norm": 0.46837809681892395,
"learning_rate": 0.00014446163345428411,
"loss": 0.5464,
"step": 1850
},
{
"epoch": 2.1016949152542375,
"grad_norm": 0.4075257480144501,
"learning_rate": 0.00014375357929013763,
"loss": 0.5658,
"step": 1860
},
{
"epoch": 2.1129943502824857,
"grad_norm": 0.5770477652549744,
"learning_rate": 0.00014304280174395862,
"loss": 0.5657,
"step": 1870
},
{
"epoch": 2.1242937853107344,
"grad_norm": 0.38950198888778687,
"learning_rate": 0.00014232934505712877,
"loss": 0.5488,
"step": 1880
},
{
"epoch": 2.135593220338983,
"grad_norm": 0.45963385701179504,
"learning_rate": 0.00014161325363778914,
"loss": 0.5536,
"step": 1890
},
{
"epoch": 2.146892655367232,
"grad_norm": 0.5212920904159546,
"learning_rate": 0.00014089457205807614,
"loss": 0.5518,
"step": 1900
},
{
"epoch": 2.15819209039548,
"grad_norm": 0.7819460034370422,
"learning_rate": 0.00014017334505134722,
"loss": 0.5493,
"step": 1910
},
{
"epoch": 2.169491525423729,
"grad_norm": 0.41791173815727234,
"learning_rate": 0.00013944961750939644,
"loss": 0.5685,
"step": 1920
},
{
"epoch": 2.1807909604519775,
"grad_norm": 0.40155965089797974,
"learning_rate": 0.00013872343447966033,
"loss": 0.5385,
"step": 1930
},
{
"epoch": 2.1920903954802258,
"grad_norm": 0.3540368974208832,
"learning_rate": 0.00013799484116241382,
"loss": 0.5601,
"step": 1940
},
{
"epoch": 2.2033898305084745,
"grad_norm": 0.45144760608673096,
"learning_rate": 0.00013726388290795697,
"loss": 0.5547,
"step": 1950
},
{
"epoch": 2.214689265536723,
"grad_norm": 0.4121219515800476,
"learning_rate": 0.00013653060521379216,
"loss": 0.5627,
"step": 1960
},
{
"epoch": 2.225988700564972,
"grad_norm": 0.5211389660835266,
"learning_rate": 0.0001357950537217921,
"loss": 0.5797,
"step": 1970
},
{
"epoch": 2.23728813559322,
"grad_norm": 0.5209491848945618,
"learning_rate": 0.00013505727421535892,
"loss": 0.5644,
"step": 1980
},
{
"epoch": 2.248587570621469,
"grad_norm": 0.40179020166397095,
"learning_rate": 0.00013431731261657462,
"loss": 0.5446,
"step": 1990
},
{
"epoch": 2.2598870056497176,
"grad_norm": 0.45757922530174255,
"learning_rate": 0.0001335752149833424,
"loss": 0.5528,
"step": 2000
},
{
"epoch": 2.2598870056497176,
"eval_loss": 0.731483519077301,
"eval_runtime": 16.8378,
"eval_samples_per_second": 88.551,
"eval_steps_per_second": 11.106,
"step": 2000
},
{
"epoch": 2.2711864406779663,
"grad_norm": 0.5437088012695312,
"learning_rate": 0.00013283102750652007,
"loss": 0.5476,
"step": 2010
},
{
"epoch": 2.2824858757062145,
"grad_norm": 0.5207261443138123,
"learning_rate": 0.00013208479650704503,
"loss": 0.5839,
"step": 2020
},
{
"epoch": 2.2937853107344632,
"grad_norm": 0.47627148032188416,
"learning_rate": 0.00013133656843305078,
"loss": 0.5791,
"step": 2030
},
{
"epoch": 2.305084745762712,
"grad_norm": 0.4785435199737549,
"learning_rate": 0.00013058638985697618,
"loss": 0.5389,
"step": 2040
},
{
"epoch": 2.3163841807909606,
"grad_norm": 0.40203753113746643,
"learning_rate": 0.0001298343074726663,
"loss": 0.5417,
"step": 2050
},
{
"epoch": 2.327683615819209,
"grad_norm": 0.6017877459526062,
"learning_rate": 0.00012908036809246623,
"loss": 0.544,
"step": 2060
},
{
"epoch": 2.3389830508474576,
"grad_norm": 0.5637748837471008,
"learning_rate": 0.0001283246186443073,
"loss": 0.5429,
"step": 2070
},
{
"epoch": 2.3502824858757063,
"grad_norm": 0.4815099835395813,
"learning_rate": 0.00012756710616878593,
"loss": 0.5466,
"step": 2080
},
{
"epoch": 2.361581920903955,
"grad_norm": 0.4496362805366516,
"learning_rate": 0.00012680787781623582,
"loss": 0.5894,
"step": 2090
},
{
"epoch": 2.3728813559322033,
"grad_norm": 0.4290124475955963,
"learning_rate": 0.00012604698084379322,
"loss": 0.5504,
"step": 2100
},
{
"epoch": 2.384180790960452,
"grad_norm": 0.5070333480834961,
"learning_rate": 0.00012528446261245515,
"loss": 0.5522,
"step": 2110
},
{
"epoch": 2.3954802259887007,
"grad_norm": 0.5735118985176086,
"learning_rate": 0.00012452037058413194,
"loss": 0.5674,
"step": 2120
},
{
"epoch": 2.406779661016949,
"grad_norm": 0.40572336316108704,
"learning_rate": 0.00012375475231869249,
"loss": 0.5814,
"step": 2130
},
{
"epoch": 2.4180790960451977,
"grad_norm": 0.5074964165687561,
"learning_rate": 0.00012298765547100445,
"loss": 0.5663,
"step": 2140
},
{
"epoch": 2.4293785310734464,
"grad_norm": 0.445563942193985,
"learning_rate": 0.0001222191277879678,
"loss": 0.5476,
"step": 2150
},
{
"epoch": 2.440677966101695,
"grad_norm": 0.5029497742652893,
"learning_rate": 0.00012144921710554278,
"loss": 0.5818,
"step": 2160
},
{
"epoch": 2.4519774011299433,
"grad_norm": 0.4790378212928772,
"learning_rate": 0.00012067797134577275,
"loss": 0.5878,
"step": 2170
},
{
"epoch": 2.463276836158192,
"grad_norm": 0.43805941939353943,
"learning_rate": 0.00011990543851380094,
"loss": 0.5602,
"step": 2180
},
{
"epoch": 2.4745762711864407,
"grad_norm": 0.49226316809654236,
"learning_rate": 0.00011913166669488278,
"loss": 0.5692,
"step": 2190
},
{
"epoch": 2.4858757062146895,
"grad_norm": 0.44978803396224976,
"learning_rate": 0.00011835670405139272,
"loss": 0.5698,
"step": 2200
},
{
"epoch": 2.4971751412429377,
"grad_norm": 0.4665353000164032,
"learning_rate": 0.00011758059881982639,
"loss": 0.567,
"step": 2210
},
{
"epoch": 2.5084745762711864,
"grad_norm": 0.3619357943534851,
"learning_rate": 0.00011680339930779843,
"loss": 0.5528,
"step": 2220
},
{
"epoch": 2.519774011299435,
"grad_norm": 0.5655903220176697,
"learning_rate": 0.00011602515389103529,
"loss": 0.57,
"step": 2230
},
{
"epoch": 2.5310734463276834,
"grad_norm": 0.5191211700439453,
"learning_rate": 0.00011524591101036446,
"loss": 0.5731,
"step": 2240
},
{
"epoch": 2.542372881355932,
"grad_norm": 0.4011457562446594,
"learning_rate": 0.00011446571916869917,
"loss": 0.5603,
"step": 2250
},
{
"epoch": 2.553672316384181,
"grad_norm": 0.4324322044849396,
"learning_rate": 0.00011368462692801944,
"loss": 0.5508,
"step": 2260
},
{
"epoch": 2.5649717514124295,
"grad_norm": 0.44678226113319397,
"learning_rate": 0.00011290268290634943,
"loss": 0.5695,
"step": 2270
},
{
"epoch": 2.576271186440678,
"grad_norm": 0.7076367139816284,
"learning_rate": 0.00011211993577473121,
"loss": 0.5812,
"step": 2280
},
{
"epoch": 2.5875706214689265,
"grad_norm": 0.36905479431152344,
"learning_rate": 0.00011133643425419531,
"loss": 0.5775,
"step": 2290
},
{
"epoch": 2.598870056497175,
"grad_norm": 0.4640151560306549,
"learning_rate": 0.0001105522271127283,
"loss": 0.5932,
"step": 2300
},
{
"epoch": 2.610169491525424,
"grad_norm": 0.541059672832489,
"learning_rate": 0.00010976736316223698,
"loss": 0.5803,
"step": 2310
},
{
"epoch": 2.621468926553672,
"grad_norm": 0.46233227849006653,
"learning_rate": 0.00010898189125551052,
"loss": 0.5939,
"step": 2320
},
{
"epoch": 2.632768361581921,
"grad_norm": 0.49569135904312134,
"learning_rate": 0.0001081958602831794,
"loss": 0.5762,
"step": 2330
},
{
"epoch": 2.6440677966101696,
"grad_norm": 0.46241089701652527,
"learning_rate": 0.00010740931917067231,
"loss": 0.5455,
"step": 2340
},
{
"epoch": 2.655367231638418,
"grad_norm": 0.4825233221054077,
"learning_rate": 0.00010662231687517112,
"loss": 0.5697,
"step": 2350
},
{
"epoch": 2.6666666666666665,
"grad_norm": 0.3972626328468323,
"learning_rate": 0.00010583490238256322,
"loss": 0.5544,
"step": 2360
},
{
"epoch": 2.6779661016949152,
"grad_norm": 0.458835244178772,
"learning_rate": 0.00010504712470439266,
"loss": 0.5603,
"step": 2370
},
{
"epoch": 2.689265536723164,
"grad_norm": 0.5480552911758423,
"learning_rate": 0.00010425903287480969,
"loss": 0.5555,
"step": 2380
},
{
"epoch": 2.7005649717514126,
"grad_norm": 0.5089293122291565,
"learning_rate": 0.0001034706759475182,
"loss": 0.5374,
"step": 2390
},
{
"epoch": 2.711864406779661,
"grad_norm": 0.39211511611938477,
"learning_rate": 0.00010268210299272294,
"loss": 0.5499,
"step": 2400
},
{
"epoch": 2.7231638418079096,
"grad_norm": 0.4466584324836731,
"learning_rate": 0.00010189336309407492,
"loss": 0.5692,
"step": 2410
},
{
"epoch": 2.7344632768361583,
"grad_norm": 0.4684606194496155,
"learning_rate": 0.00010110450534561629,
"loss": 0.5858,
"step": 2420
},
{
"epoch": 2.7457627118644066,
"grad_norm": 0.4925701320171356,
"learning_rate": 0.00010031557884872471,
"loss": 0.5577,
"step": 2430
},
{
"epoch": 2.7570621468926553,
"grad_norm": 0.5621470212936401,
"learning_rate": 9.952663270905691e-05,
"loss": 0.543,
"step": 2440
},
{
"epoch": 2.768361581920904,
"grad_norm": 0.5334658622741699,
"learning_rate": 9.873771603349229e-05,
"loss": 0.5673,
"step": 2450
},
{
"epoch": 2.7796610169491527,
"grad_norm": 0.4116109311580658,
"learning_rate": 9.79488779270763e-05,
"loss": 0.5759,
"step": 2460
},
{
"epoch": 2.7909604519774014,
"grad_norm": 0.579428493976593,
"learning_rate": 9.716016748996397e-05,
"loss": 0.5804,
"step": 2470
},
{
"epoch": 2.8022598870056497,
"grad_norm": 0.6294684410095215,
"learning_rate": 9.637163381436368e-05,
"loss": 0.573,
"step": 2480
},
{
"epoch": 2.8135593220338984,
"grad_norm": 0.4215356409549713,
"learning_rate": 9.558332598148156e-05,
"loss": 0.5702,
"step": 2490
},
{
"epoch": 2.824858757062147,
"grad_norm": 0.6607437133789062,
"learning_rate": 9.479529305846652e-05,
"loss": 0.5556,
"step": 2500
},
{
"epoch": 2.8361581920903953,
"grad_norm": 0.4144829213619232,
"learning_rate": 9.400758409535597e-05,
"loss": 0.5417,
"step": 2510
},
{
"epoch": 2.847457627118644,
"grad_norm": 0.47953999042510986,
"learning_rate": 9.322024812202295e-05,
"loss": 0.5531,
"step": 2520
},
{
"epoch": 2.8587570621468927,
"grad_norm": 0.4800129532814026,
"learning_rate": 9.243333414512428e-05,
"loss": 0.5692,
"step": 2530
},
{
"epoch": 2.870056497175141,
"grad_norm": 0.6145460605621338,
"learning_rate": 9.164689114505007e-05,
"loss": 0.556,
"step": 2540
},
{
"epoch": 2.8813559322033897,
"grad_norm": 0.45880794525146484,
"learning_rate": 9.086096807287525e-05,
"loss": 0.547,
"step": 2550
},
{
"epoch": 2.8926553672316384,
"grad_norm": 0.4852420389652252,
"learning_rate": 9.007561384731249e-05,
"loss": 0.5843,
"step": 2560
},
{
"epoch": 2.903954802259887,
"grad_norm": 0.4079766571521759,
"learning_rate": 8.929087735166738e-05,
"loss": 0.5642,
"step": 2570
},
{
"epoch": 2.915254237288136,
"grad_norm": 0.41701099276542664,
"learning_rate": 8.850680743079576e-05,
"loss": 0.5191,
"step": 2580
},
{
"epoch": 2.926553672316384,
"grad_norm": 0.51291823387146,
"learning_rate": 8.772345288806338e-05,
"loss": 0.5823,
"step": 2590
},
{
"epoch": 2.937853107344633,
"grad_norm": 0.3969896137714386,
"learning_rate": 8.694086248230836e-05,
"loss": 0.5673,
"step": 2600
},
{
"epoch": 2.9491525423728815,
"grad_norm": 0.5091503858566284,
"learning_rate": 8.615908492480598e-05,
"loss": 0.5555,
"step": 2610
},
{
"epoch": 2.9604519774011298,
"grad_norm": 0.5286157131195068,
"learning_rate": 8.537816887623706e-05,
"loss": 0.5883,
"step": 2620
},
{
"epoch": 2.9717514124293785,
"grad_norm": 0.42479345202445984,
"learning_rate": 8.459816294365892e-05,
"loss": 0.6002,
"step": 2630
},
{
"epoch": 2.983050847457627,
"grad_norm": 0.6383501291275024,
"learning_rate": 8.381911567747986e-05,
"loss": 0.5751,
"step": 2640
},
{
"epoch": 2.994350282485876,
"grad_norm": 0.4221038520336151,
"learning_rate": 8.304107556843739e-05,
"loss": 0.5583,
"step": 2650
},
{
"epoch": 3.005649717514124,
"grad_norm": 0.46783480048179626,
"learning_rate": 8.22640910445799e-05,
"loss": 0.5075,
"step": 2660
},
{
"epoch": 3.016949152542373,
"grad_norm": 0.48267218470573425,
"learning_rate": 8.14882104682522e-05,
"loss": 0.4768,
"step": 2670
},
{
"epoch": 3.0282485875706215,
"grad_norm": 0.3863345682621002,
"learning_rate": 8.07134821330855e-05,
"loss": 0.4959,
"step": 2680
},
{
"epoch": 3.0395480225988702,
"grad_norm": 0.5208786129951477,
"learning_rate": 7.993995426099132e-05,
"loss": 0.4765,
"step": 2690
},
{
"epoch": 3.0508474576271185,
"grad_norm": 0.4619675874710083,
"learning_rate": 7.916767499915998e-05,
"loss": 0.4615,
"step": 2700
},
{
"epoch": 3.062146892655367,
"grad_norm": 0.4720297157764435,
"learning_rate": 7.839669241706368e-05,
"loss": 0.495,
"step": 2710
},
{
"epoch": 3.073446327683616,
"grad_norm": 0.5332727432250977,
"learning_rate": 7.762705450346462e-05,
"loss": 0.486,
"step": 2720
},
{
"epoch": 3.084745762711864,
"grad_norm": 0.5170637965202332,
"learning_rate": 7.685880916342795e-05,
"loss": 0.4816,
"step": 2730
},
{
"epoch": 3.096045197740113,
"grad_norm": 0.47557196021080017,
"learning_rate": 7.609200421533991e-05,
"loss": 0.4706,
"step": 2740
},
{
"epoch": 3.1073446327683616,
"grad_norm": 0.7031615376472473,
"learning_rate": 7.532668738793159e-05,
"loss": 0.479,
"step": 2750
},
{
"epoch": 3.1186440677966103,
"grad_norm": 0.4733956456184387,
"learning_rate": 7.456290631730795e-05,
"loss": 0.5023,
"step": 2760
},
{
"epoch": 3.1299435028248586,
"grad_norm": 0.5425024032592773,
"learning_rate": 7.380070854398292e-05,
"loss": 0.477,
"step": 2770
},
{
"epoch": 3.1412429378531073,
"grad_norm": 0.4106122851371765,
"learning_rate": 7.304014150992016e-05,
"loss": 0.511,
"step": 2780
},
{
"epoch": 3.152542372881356,
"grad_norm": 0.580517053604126,
"learning_rate": 7.228125255558026e-05,
"loss": 0.4797,
"step": 2790
},
{
"epoch": 3.1638418079096047,
"grad_norm": 0.6407458186149597,
"learning_rate": 7.152408891697406e-05,
"loss": 0.4908,
"step": 2800
},
{
"epoch": 3.175141242937853,
"grad_norm": 0.6218848824501038,
"learning_rate": 7.076869772272231e-05,
"loss": 0.4944,
"step": 2810
},
{
"epoch": 3.1864406779661016,
"grad_norm": 0.49968206882476807,
"learning_rate": 7.001512599112254e-05,
"loss": 0.4764,
"step": 2820
},
{
"epoch": 3.1977401129943503,
"grad_norm": 0.465017706155777,
"learning_rate": 6.926342062722223e-05,
"loss": 0.4917,
"step": 2830
},
{
"epoch": 3.209039548022599,
"grad_norm": 0.49751996994018555,
"learning_rate": 6.851362841989926e-05,
"loss": 0.5105,
"step": 2840
},
{
"epoch": 3.2203389830508473,
"grad_norm": 0.5394355654716492,
"learning_rate": 6.776579603894983e-05,
"loss": 0.4888,
"step": 2850
},
{
"epoch": 3.231638418079096,
"grad_norm": 0.628495454788208,
"learning_rate": 6.701997003218334e-05,
"loss": 0.4597,
"step": 2860
},
{
"epoch": 3.2429378531073447,
"grad_norm": 0.544737696647644,
"learning_rate": 6.627619682252514e-05,
"loss": 0.5198,
"step": 2870
},
{
"epoch": 3.2542372881355934,
"grad_norm": 0.3959389626979828,
"learning_rate": 6.553452270512708e-05,
"loss": 0.4881,
"step": 2880
},
{
"epoch": 3.2655367231638417,
"grad_norm": 0.7966346740722656,
"learning_rate": 6.47949938444858e-05,
"loss": 0.4839,
"step": 2890
},
{
"epoch": 3.2768361581920904,
"grad_norm": 0.5471003651618958,
"learning_rate": 6.405765627156934e-05,
"loss": 0.4953,
"step": 2900
},
{
"epoch": 3.288135593220339,
"grad_norm": 0.6175327301025391,
"learning_rate": 6.332255588095211e-05,
"loss": 0.4936,
"step": 2910
},
{
"epoch": 3.2994350282485874,
"grad_norm": 0.5062096118927002,
"learning_rate": 6.258973842795803e-05,
"loss": 0.4927,
"step": 2920
},
{
"epoch": 3.310734463276836,
"grad_norm": 0.5980800986289978,
"learning_rate": 6.185924952581286e-05,
"loss": 0.4908,
"step": 2930
},
{
"epoch": 3.3220338983050848,
"grad_norm": 0.47181573510169983,
"learning_rate": 6.11311346428046e-05,
"loss": 0.4938,
"step": 2940
},
{
"epoch": 3.3333333333333335,
"grad_norm": 0.4729149341583252,
"learning_rate": 6.040543909945393e-05,
"loss": 0.4795,
"step": 2950
},
{
"epoch": 3.3446327683615817,
"grad_norm": 0.4721032977104187,
"learning_rate": 5.968220806569289e-05,
"loss": 0.4821,
"step": 2960
},
{
"epoch": 3.3559322033898304,
"grad_norm": 0.49077048897743225,
"learning_rate": 5.89614865580534e-05,
"loss": 0.4765,
"step": 2970
},
{
"epoch": 3.367231638418079,
"grad_norm": 0.4721141755580902,
"learning_rate": 5.8243319436865554e-05,
"loss": 0.4622,
"step": 2980
},
{
"epoch": 3.378531073446328,
"grad_norm": 0.5188248157501221,
"learning_rate": 5.7527751403464935e-05,
"loss": 0.4966,
"step": 2990
},
{
"epoch": 3.389830508474576,
"grad_norm": 0.5185617208480835,
"learning_rate": 5.681482699741052e-05,
"loss": 0.4847,
"step": 3000
},
{
"epoch": 3.389830508474576,
"eval_loss": 0.7709711194038391,
"eval_runtime": 16.7631,
"eval_samples_per_second": 88.945,
"eval_steps_per_second": 11.155,
"step": 3000
},
{
"epoch": 3.401129943502825,
"grad_norm": 0.44520846009254456,
"learning_rate": 5.610459059371229e-05,
"loss": 0.4904,
"step": 3010
},
{
"epoch": 3.4124293785310735,
"grad_norm": 0.4945251941680908,
"learning_rate": 5.53970864000693e-05,
"loss": 0.486,
"step": 3020
},
{
"epoch": 3.423728813559322,
"grad_norm": 0.45569783449172974,
"learning_rate": 5.4692358454117596e-05,
"loss": 0.485,
"step": 3030
},
{
"epoch": 3.4350282485875705,
"grad_norm": 0.6237002015113831,
"learning_rate": 5.399045062068977e-05,
"loss": 0.5055,
"step": 3040
},
{
"epoch": 3.446327683615819,
"grad_norm": 0.4695291519165039,
"learning_rate": 5.329140658908423e-05,
"loss": 0.5047,
"step": 3050
},
{
"epoch": 3.457627118644068,
"grad_norm": 0.5210557579994202,
"learning_rate": 5.259526987034593e-05,
"loss": 0.5181,
"step": 3060
},
{
"epoch": 3.4689265536723166,
"grad_norm": 0.5110927224159241,
"learning_rate": 5.190208379455802e-05,
"loss": 0.4901,
"step": 3070
},
{
"epoch": 3.480225988700565,
"grad_norm": 0.5741479396820068,
"learning_rate": 5.121189150814507e-05,
"loss": 0.4919,
"step": 3080
},
{
"epoch": 3.4915254237288136,
"grad_norm": 0.49151936173439026,
"learning_rate": 5.052473597118715e-05,
"loss": 0.4954,
"step": 3090
},
{
"epoch": 3.5028248587570623,
"grad_norm": 0.5301800966262817,
"learning_rate": 4.984065995474601e-05,
"loss": 0.4966,
"step": 3100
},
{
"epoch": 3.5141242937853105,
"grad_norm": 0.5711139440536499,
"learning_rate": 4.915970603820293e-05,
"loss": 0.4907,
"step": 3110
},
{
"epoch": 3.5254237288135593,
"grad_norm": 0.6174890995025635,
"learning_rate": 4.84819166066082e-05,
"loss": 0.4957,
"step": 3120
},
{
"epoch": 3.536723163841808,
"grad_norm": 0.5370630621910095,
"learning_rate": 4.780733384804312e-05,
"loss": 0.5131,
"step": 3130
},
{
"epoch": 3.548022598870056,
"grad_norm": 0.4994024634361267,
"learning_rate": 4.713599975099399e-05,
"loss": 0.5214,
"step": 3140
},
{
"epoch": 3.559322033898305,
"grad_norm": 0.8335681557655334,
"learning_rate": 4.646795610173864e-05,
"loss": 0.498,
"step": 3150
},
{
"epoch": 3.5706214689265536,
"grad_norm": 0.5772273540496826,
"learning_rate": 4.5803244481745275e-05,
"loss": 0.4889,
"step": 3160
},
{
"epoch": 3.5819209039548023,
"grad_norm": 0.725974977016449,
"learning_rate": 4.514190626508465e-05,
"loss": 0.4889,
"step": 3170
},
{
"epoch": 3.593220338983051,
"grad_norm": 0.4802244305610657,
"learning_rate": 4.448398261585459e-05,
"loss": 0.4688,
"step": 3180
},
{
"epoch": 3.6045197740112993,
"grad_norm": 1.078092098236084,
"learning_rate": 4.3829514485617754e-05,
"loss": 0.4719,
"step": 3190
},
{
"epoch": 3.615819209039548,
"grad_norm": 0.5553079843521118,
"learning_rate": 4.31785426108527e-05,
"loss": 0.4816,
"step": 3200
},
{
"epoch": 3.6271186440677967,
"grad_norm": 0.48087623715400696,
"learning_rate": 4.253110751041846e-05,
"loss": 0.4894,
"step": 3210
},
{
"epoch": 3.638418079096045,
"grad_norm": 0.6121565103530884,
"learning_rate": 4.1887249483032195e-05,
"loss": 0.4978,
"step": 3220
},
{
"epoch": 3.6497175141242937,
"grad_norm": 0.616296648979187,
"learning_rate": 4.1247008604761096e-05,
"loss": 0.4834,
"step": 3230
},
{
"epoch": 3.6610169491525424,
"grad_norm": 0.5769548416137695,
"learning_rate": 4.061042472652786e-05,
"loss": 0.5177,
"step": 3240
},
{
"epoch": 3.672316384180791,
"grad_norm": 0.5508149862289429,
"learning_rate": 3.997753747163014e-05,
"loss": 0.5206,
"step": 3250
},
{
"epoch": 3.68361581920904,
"grad_norm": 0.4783056080341339,
"learning_rate": 3.934838623327427e-05,
"loss": 0.4996,
"step": 3260
},
{
"epoch": 3.694915254237288,
"grad_norm": 0.9651548266410828,
"learning_rate": 3.872301017212337e-05,
"loss": 0.4741,
"step": 3270
},
{
"epoch": 3.7062146892655368,
"grad_norm": 0.5215980410575867,
"learning_rate": 3.8101448213859846e-05,
"loss": 0.4757,
"step": 3280
},
{
"epoch": 3.7175141242937855,
"grad_norm": 0.4626060724258423,
"learning_rate": 3.7483739046762326e-05,
"loss": 0.4887,
"step": 3290
},
{
"epoch": 3.7288135593220337,
"grad_norm": 0.5218005180358887,
"learning_rate": 3.6869921119297714e-05,
"loss": 0.4603,
"step": 3300
},
{
"epoch": 3.7401129943502824,
"grad_norm": 0.557963490486145,
"learning_rate": 3.626003263772813e-05,
"loss": 0.5009,
"step": 3310
},
{
"epoch": 3.751412429378531,
"grad_norm": 0.7049193382263184,
"learning_rate": 3.565411156373252e-05,
"loss": 0.4982,
"step": 3320
},
{
"epoch": 3.7627118644067794,
"grad_norm": 0.5267760753631592,
"learning_rate": 3.505219561204398e-05,
"loss": 0.4716,
"step": 3330
},
{
"epoch": 3.774011299435028,
"grad_norm": 0.5169788002967834,
"learning_rate": 3.445432224810232e-05,
"loss": 0.4992,
"step": 3340
},
{
"epoch": 3.785310734463277,
"grad_norm": 0.5268625020980835,
"learning_rate": 3.386052868572186e-05,
"loss": 0.5063,
"step": 3350
},
{
"epoch": 3.7966101694915255,
"grad_norm": 0.5163342356681824,
"learning_rate": 3.3270851884775203e-05,
"loss": 0.5045,
"step": 3360
},
{
"epoch": 3.8079096045197742,
"grad_norm": 0.6581404805183411,
"learning_rate": 3.268532854889287e-05,
"loss": 0.4818,
"step": 3370
},
{
"epoch": 3.8192090395480225,
"grad_norm": 0.5500733256340027,
"learning_rate": 3.210399512317849e-05,
"loss": 0.5064,
"step": 3380
},
{
"epoch": 3.830508474576271,
"grad_norm": 0.530192494392395,
"learning_rate": 3.1526887791940395e-05,
"loss": 0.4761,
"step": 3390
},
{
"epoch": 3.84180790960452,
"grad_norm": 0.5225520133972168,
"learning_rate": 3.095404247643955e-05,
"loss": 0.4874,
"step": 3400
},
{
"epoch": 3.853107344632768,
"grad_norm": 0.52415531873703,
"learning_rate": 3.038549483265348e-05,
"loss": 0.4748,
"step": 3410
},
{
"epoch": 3.864406779661017,
"grad_norm": 0.6530227661132812,
"learning_rate": 2.982128024905697e-05,
"loss": 0.4841,
"step": 3420
},
{
"epoch": 3.8757062146892656,
"grad_norm": 0.5648587346076965,
"learning_rate": 2.9261433844419306e-05,
"loss": 0.4687,
"step": 3430
},
{
"epoch": 3.8870056497175143,
"grad_norm": 0.957992672920227,
"learning_rate": 2.870599046561858e-05,
"loss": 0.489,
"step": 3440
},
{
"epoch": 3.898305084745763,
"grad_norm": 0.5714943408966064,
"learning_rate": 2.8154984685472373e-05,
"loss": 0.4886,
"step": 3450
},
{
"epoch": 3.9096045197740112,
"grad_norm": 0.48912012577056885,
"learning_rate": 2.7608450800586016e-05,
"loss": 0.5028,
"step": 3460
},
{
"epoch": 3.92090395480226,
"grad_norm": 0.5882663726806641,
"learning_rate": 2.7066422829217875e-05,
"loss": 0.5216,
"step": 3470
},
{
"epoch": 3.9322033898305087,
"grad_norm": 0.5730474591255188,
"learning_rate": 2.6528934509161763e-05,
"loss": 0.4596,
"step": 3480
},
{
"epoch": 3.943502824858757,
"grad_norm": 0.6839239001274109,
"learning_rate": 2.599601929564709e-05,
"loss": 0.4944,
"step": 3490
},
{
"epoch": 3.9548022598870056,
"grad_norm": 0.5317196249961853,
"learning_rate": 2.5467710359256557e-05,
"loss": 0.4779,
"step": 3500
},
{
"epoch": 3.9661016949152543,
"grad_norm": 0.5707146525382996,
"learning_rate": 2.4944040583861284e-05,
"loss": 0.4878,
"step": 3510
},
{
"epoch": 3.9774011299435026,
"grad_norm": 0.781765341758728,
"learning_rate": 2.4425042564574184e-05,
"loss": 0.4719,
"step": 3520
},
{
"epoch": 3.9887005649717513,
"grad_norm": 0.6404000520706177,
"learning_rate": 2.3910748605721066e-05,
"loss": 0.4983,
"step": 3530
},
{
"epoch": 4.0,
"grad_norm": 0.6482042074203491,
"learning_rate": 2.3401190718829945e-05,
"loss": 0.4916,
"step": 3540
},
{
"epoch": 4.011299435028248,
"grad_norm": 0.4203549027442932,
"learning_rate": 2.289640062063839e-05,
"loss": 0.4653,
"step": 3550
},
{
"epoch": 4.022598870056497,
"grad_norm": 0.48745134472846985,
"learning_rate": 2.239640973111945e-05,
"loss": 0.4595,
"step": 3560
},
{
"epoch": 4.033898305084746,
"grad_norm": 0.5251093506813049,
"learning_rate": 2.1901249171526063e-05,
"loss": 0.4471,
"step": 3570
},
{
"epoch": 4.045197740112994,
"grad_norm": 0.6273601651191711,
"learning_rate": 2.1410949762453735e-05,
"loss": 0.437,
"step": 3580
},
{
"epoch": 4.056497175141243,
"grad_norm": 0.5651510953903198,
"learning_rate": 2.0925542021922283e-05,
"loss": 0.4194,
"step": 3590
},
{
"epoch": 4.067796610169491,
"grad_norm": 0.7425692081451416,
"learning_rate": 2.0445056163476374e-05,
"loss": 0.4568,
"step": 3600
},
{
"epoch": 4.0790960451977405,
"grad_norm": 0.49514123797416687,
"learning_rate": 1.9969522094304704e-05,
"loss": 0.4673,
"step": 3610
},
{
"epoch": 4.090395480225989,
"grad_norm": 0.7164167761802673,
"learning_rate": 1.949896941337861e-05,
"loss": 0.4507,
"step": 3620
},
{
"epoch": 4.101694915254237,
"grad_norm": 0.5990516543388367,
"learning_rate": 1.9033427409609738e-05,
"loss": 0.4271,
"step": 3630
},
{
"epoch": 4.112994350282486,
"grad_norm": 0.4122953414916992,
"learning_rate": 1.8572925060026868e-05,
"loss": 0.4331,
"step": 3640
},
{
"epoch": 4.124293785310734,
"grad_norm": 0.5930763483047485,
"learning_rate": 1.8117491027972476e-05,
"loss": 0.4388,
"step": 3650
},
{
"epoch": 4.135593220338983,
"grad_norm": 0.5768269300460815,
"learning_rate": 1.766715366131837e-05,
"loss": 0.4432,
"step": 3660
},
{
"epoch": 4.146892655367232,
"grad_norm": 1.2128398418426514,
"learning_rate": 1.722194099070148e-05,
"loss": 0.4445,
"step": 3670
},
{
"epoch": 4.15819209039548,
"grad_norm": 0.5718667507171631,
"learning_rate": 1.678188072777891e-05,
"loss": 0.4348,
"step": 3680
},
{
"epoch": 4.169491525423728,
"grad_norm": 0.6063762903213501,
"learning_rate": 1.6347000263503176e-05,
"loss": 0.4293,
"step": 3690
},
{
"epoch": 4.1807909604519775,
"grad_norm": 0.5244751572608948,
"learning_rate": 1.591732666641732e-05,
"loss": 0.4469,
"step": 3700
},
{
"epoch": 4.192090395480226,
"grad_norm": 0.6261086463928223,
"learning_rate": 1.5492886680969963e-05,
"loss": 0.4415,
"step": 3710
},
{
"epoch": 4.203389830508475,
"grad_norm": 1.0888699293136597,
"learning_rate": 1.5073706725850679e-05,
"loss": 0.4365,
"step": 3720
},
{
"epoch": 4.214689265536723,
"grad_norm": 0.5183214545249939,
"learning_rate": 1.4659812892345692e-05,
"loss": 0.4717,
"step": 3730
},
{
"epoch": 4.2259887005649714,
"grad_norm": 0.5120430588722229,
"learning_rate": 1.425123094271369e-05,
"loss": 0.4484,
"step": 3740
},
{
"epoch": 4.237288135593221,
"grad_norm": 0.4748518764972687,
"learning_rate": 1.3847986308582384e-05,
"loss": 0.4566,
"step": 3750
},
{
"epoch": 4.248587570621469,
"grad_norm": 0.5438849329948425,
"learning_rate": 1.3450104089365611e-05,
"loss": 0.4768,
"step": 3760
},
{
"epoch": 4.259887005649717,
"grad_norm": 0.4719776213169098,
"learning_rate": 1.3057609050700881e-05,
"loss": 0.4657,
"step": 3770
},
{
"epoch": 4.271186440677966,
"grad_norm": 0.5300427675247192,
"learning_rate": 1.267052562290808e-05,
"loss": 0.4412,
"step": 3780
},
{
"epoch": 4.2824858757062145,
"grad_norm": 0.6340104937553406,
"learning_rate": 1.2288877899468653e-05,
"loss": 0.4528,
"step": 3790
},
{
"epoch": 4.293785310734464,
"grad_norm": 0.5274319648742676,
"learning_rate": 1.1912689635526086e-05,
"loss": 0.4466,
"step": 3800
},
{
"epoch": 4.305084745762712,
"grad_norm": 0.5659282803535461,
"learning_rate": 1.1541984246407189e-05,
"loss": 0.4332,
"step": 3810
},
{
"epoch": 4.31638418079096,
"grad_norm": 0.5424573421478271,
"learning_rate": 1.1176784806164676e-05,
"loss": 0.4223,
"step": 3820
},
{
"epoch": 4.327683615819209,
"grad_norm": 0.5446699261665344,
"learning_rate": 1.081711404614104e-05,
"loss": 0.4335,
"step": 3830
},
{
"epoch": 4.338983050847458,
"grad_norm": 0.4954063594341278,
"learning_rate": 1.0462994353553512e-05,
"loss": 0.4221,
"step": 3840
},
{
"epoch": 4.350282485875706,
"grad_norm": 0.6320489048957825,
"learning_rate": 1.0114447770100688e-05,
"loss": 0.43,
"step": 3850
},
{
"epoch": 4.361581920903955,
"grad_norm": 0.5482968688011169,
"learning_rate": 9.77149599059063e-06,
"loss": 0.4407,
"step": 3860
},
{
"epoch": 4.372881355932203,
"grad_norm": 0.6075878739356995,
"learning_rate": 9.434160361590372e-06,
"loss": 0.4434,
"step": 3870
},
{
"epoch": 4.3841807909604515,
"grad_norm": 0.6049032211303711,
"learning_rate": 9.102461880097302e-06,
"loss": 0.4411,
"step": 3880
},
{
"epoch": 4.395480225988701,
"grad_norm": 0.5798004865646362,
"learning_rate": 8.776421192232298e-06,
"loss": 0.431,
"step": 3890
},
{
"epoch": 4.406779661016949,
"grad_norm": 0.5278260707855225,
"learning_rate": 8.456058591954463e-06,
"loss": 0.4422,
"step": 3900
},
{
"epoch": 4.418079096045198,
"grad_norm": 0.568495512008667,
"learning_rate": 8.141394019798177e-06,
"loss": 0.4413,
"step": 3910
},
{
"epoch": 4.429378531073446,
"grad_norm": 0.5533962845802307,
"learning_rate": 7.832447061631688e-06,
"loss": 0.4191,
"step": 3920
},
{
"epoch": 4.440677966101695,
"grad_norm": 0.5724579691886902,
"learning_rate": 7.529236947438256e-06,
"loss": 0.4428,
"step": 3930
},
{
"epoch": 4.451977401129944,
"grad_norm": 0.5295354723930359,
"learning_rate": 7.231782550119015e-06,
"loss": 0.4363,
"step": 3940
},
{
"epoch": 4.463276836158192,
"grad_norm": 0.6352728605270386,
"learning_rate": 6.940102384318314e-06,
"loss": 0.4462,
"step": 3950
},
{
"epoch": 4.47457627118644,
"grad_norm": 0.4653913080692291,
"learning_rate": 6.654214605271414e-06,
"loss": 0.4463,
"step": 3960
},
{
"epoch": 4.4858757062146895,
"grad_norm": 0.5769869685173035,
"learning_rate": 6.37413700767423e-06,
"loss": 0.4388,
"step": 3970
},
{
"epoch": 4.497175141242938,
"grad_norm": 0.4576704204082489,
"learning_rate": 6.099887024575901e-06,
"loss": 0.4466,
"step": 3980
},
{
"epoch": 4.508474576271187,
"grad_norm": 0.5665444135665894,
"learning_rate": 5.831481726293609e-06,
"loss": 0.4628,
"step": 3990
},
{
"epoch": 4.519774011299435,
"grad_norm": 0.527612566947937,
"learning_rate": 5.568937819350084e-06,
"loss": 0.4261,
"step": 4000
},
{
"epoch": 4.519774011299435,
"eval_loss": 0.8213748335838318,
"eval_runtime": 16.8138,
"eval_samples_per_second": 88.677,
"eval_steps_per_second": 11.122,
"step": 4000
},
{
"epoch": 4.531073446327683,
"grad_norm": 0.668229877948761,
"learning_rate": 5.312271645433697e-06,
"loss": 0.4298,
"step": 4010
},
{
"epoch": 4.5423728813559325,
"grad_norm": 0.5199172496795654,
"learning_rate": 5.061499180381391e-06,
"loss": 0.4547,
"step": 4020
},
{
"epoch": 4.553672316384181,
"grad_norm": 0.5515559911727905,
"learning_rate": 4.816636033184119e-06,
"loss": 0.4172,
"step": 4030
},
{
"epoch": 4.564971751412429,
"grad_norm": 0.648692786693573,
"learning_rate": 4.577697445015472e-06,
"loss": 0.4664,
"step": 4040
},
{
"epoch": 4.576271186440678,
"grad_norm": 0.5000047087669373,
"learning_rate": 4.3446982882828555e-06,
"loss": 0.4401,
"step": 4050
},
{
"epoch": 4.5875706214689265,
"grad_norm": 0.51460862159729,
"learning_rate": 4.117653065701899e-06,
"loss": 0.4297,
"step": 4060
},
{
"epoch": 4.598870056497175,
"grad_norm": 0.5801824331283569,
"learning_rate": 3.896575909393663e-06,
"loss": 0.453,
"step": 4070
},
{
"epoch": 4.610169491525424,
"grad_norm": 0.5613316297531128,
"learning_rate": 3.6814805800050255e-06,
"loss": 0.445,
"step": 4080
},
{
"epoch": 4.621468926553672,
"grad_norm": 0.556430459022522,
"learning_rate": 3.4723804658522318e-06,
"loss": 0.4332,
"step": 4090
},
{
"epoch": 4.632768361581921,
"grad_norm": 0.7290319204330444,
"learning_rate": 3.269288582087493e-06,
"loss": 0.4667,
"step": 4100
},
{
"epoch": 4.6440677966101696,
"grad_norm": 0.5109266638755798,
"learning_rate": 3.072217569888847e-06,
"loss": 0.4484,
"step": 4110
},
{
"epoch": 4.655367231638418,
"grad_norm": 0.5184862017631531,
"learning_rate": 2.8811796956734418e-06,
"loss": 0.4476,
"step": 4120
},
{
"epoch": 4.666666666666667,
"grad_norm": 0.45906636118888855,
"learning_rate": 2.696186850333893e-06,
"loss": 0.4556,
"step": 4130
},
{
"epoch": 4.677966101694915,
"grad_norm": 0.6825990080833435,
"learning_rate": 2.5172505484982733e-06,
"loss": 0.4361,
"step": 4140
},
{
"epoch": 4.6892655367231635,
"grad_norm": 0.6751702427864075,
"learning_rate": 2.3443819278132996e-06,
"loss": 0.454,
"step": 4150
},
{
"epoch": 4.700564971751413,
"grad_norm": 0.6289049983024597,
"learning_rate": 2.17759174825114e-06,
"loss": 0.4438,
"step": 4160
},
{
"epoch": 4.711864406779661,
"grad_norm": 0.6290619969367981,
"learning_rate": 2.016890391439674e-06,
"loss": 0.4665,
"step": 4170
},
{
"epoch": 4.72316384180791,
"grad_norm": 0.5864591002464294,
"learning_rate": 1.8622878600162763e-06,
"loss": 0.4477,
"step": 4180
},
{
"epoch": 4.734463276836158,
"grad_norm": 0.558210015296936,
"learning_rate": 1.7137937770052349e-06,
"loss": 0.4605,
"step": 4190
},
{
"epoch": 4.745762711864407,
"grad_norm": 0.6048043966293335,
"learning_rate": 1.5714173852187875e-06,
"loss": 0.4458,
"step": 4200
},
{
"epoch": 4.757062146892656,
"grad_norm": 0.6476958990097046,
"learning_rate": 1.4351675466817705e-06,
"loss": 0.4311,
"step": 4210
},
{
"epoch": 4.768361581920904,
"grad_norm": 0.6173336505889893,
"learning_rate": 1.3050527420800928e-06,
"loss": 0.4526,
"step": 4220
},
{
"epoch": 4.779661016949152,
"grad_norm": 0.5486857295036316,
"learning_rate": 1.1810810702327701e-06,
"loss": 0.4407,
"step": 4230
},
{
"epoch": 4.790960451977401,
"grad_norm": 0.4905402362346649,
"learning_rate": 1.063260247587905e-06,
"loss": 0.4367,
"step": 4240
},
{
"epoch": 4.80225988700565,
"grad_norm": 0.6765434741973877,
"learning_rate": 9.515976077423716e-07,
"loss": 0.4438,
"step": 4250
},
{
"epoch": 4.813559322033898,
"grad_norm": 0.5942672491073608,
"learning_rate": 8.461001009852809e-07,
"loss": 0.4326,
"step": 4260
},
{
"epoch": 4.824858757062147,
"grad_norm": 0.5058199167251587,
"learning_rate": 7.46774293865471e-07,
"loss": 0.4258,
"step": 4270
},
{
"epoch": 4.836158192090395,
"grad_norm": 0.5257803797721863,
"learning_rate": 6.536263687827005e-07,
"loss": 0.4377,
"step": 4280
},
{
"epoch": 4.847457627118644,
"grad_norm": 0.5546150207519531,
"learning_rate": 5.66662123602879e-07,
"loss": 0.447,
"step": 4290
},
{
"epoch": 4.858757062146893,
"grad_norm": 0.5133059024810791,
"learning_rate": 4.858869712971669e-07,
"loss": 0.4536,
"step": 4300
},
{
"epoch": 4.870056497175141,
"grad_norm": 0.6438310742378235,
"learning_rate": 4.113059396050445e-07,
"loss": 0.4328,
"step": 4310
},
{
"epoch": 4.88135593220339,
"grad_norm": 0.5640572309494019,
"learning_rate": 3.4292367072140677e-07,
"loss": 0.4629,
"step": 4320
},
{
"epoch": 4.892655367231638,
"grad_norm": 0.6676624417304993,
"learning_rate": 2.807444210075616e-07,
"loss": 0.4347,
"step": 4330
},
{
"epoch": 4.903954802259887,
"grad_norm": 0.45671963691711426,
"learning_rate": 2.2477206072633038e-07,
"loss": 0.4536,
"step": 4340
},
{
"epoch": 4.915254237288136,
"grad_norm": 0.6115604043006897,
"learning_rate": 1.7501007380115174e-07,
"loss": 0.4453,
"step": 4350
},
{
"epoch": 4.926553672316384,
"grad_norm": 0.7250502109527588,
"learning_rate": 1.3146155759923285e-07,
"loss": 0.4454,
"step": 4360
},
{
"epoch": 4.937853107344633,
"grad_norm": 0.5707434415817261,
"learning_rate": 9.412922273871471e-08,
"loss": 0.4107,
"step": 4370
},
{
"epoch": 4.9491525423728815,
"grad_norm": 0.6125264763832092,
"learning_rate": 6.301539292001834e-08,
"loss": 0.4355,
"step": 4380
},
{
"epoch": 4.96045197740113,
"grad_norm": 0.6177554130554199,
"learning_rate": 3.8122004781149246e-08,
"loss": 0.4502,
"step": 4390
},
{
"epoch": 4.971751412429379,
"grad_norm": 0.7295418381690979,
"learning_rate": 1.9450607777204978e-08,
"loss": 0.4426,
"step": 4400
},
{
"epoch": 4.983050847457627,
"grad_norm": 0.5891022682189941,
"learning_rate": 7.00236408388566e-09,
"loss": 0.449,
"step": 4410
},
{
"epoch": 4.994350282485875,
"grad_norm": 0.5898358225822449,
"learning_rate": 7.780485252073533e-10,
"loss": 0.4552,
"step": 4420
},
{
"epoch": 5.0,
"step": 4425,
"total_flos": 1.2071933436588196e+18,
"train_loss": 0.5702266787943867,
"train_runtime": 4287.601,
"train_samples_per_second": 33.024,
"train_steps_per_second": 1.032
}
],
"logging_steps": 10,
"max_steps": 4425,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.2071933436588196e+18,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}