hp_ablations_mistral_epoch5 / trainer_state.json
sedrickkeh's picture
End of training
7f2031e verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 5.0,
"eval_steps": 500,
"global_step": 2470,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.020242914979757085,
"grad_norm": 2.8825272538542044,
"learning_rate": 5e-06,
"loss": 0.7569,
"step": 10
},
{
"epoch": 0.04048582995951417,
"grad_norm": 2.3606485385931157,
"learning_rate": 5e-06,
"loss": 0.6504,
"step": 20
},
{
"epoch": 0.06072874493927125,
"grad_norm": 1.426935919321139,
"learning_rate": 5e-06,
"loss": 0.6287,
"step": 30
},
{
"epoch": 0.08097165991902834,
"grad_norm": 1.9896158325770887,
"learning_rate": 5e-06,
"loss": 0.6136,
"step": 40
},
{
"epoch": 0.10121457489878542,
"grad_norm": 1.2993488018928139,
"learning_rate": 5e-06,
"loss": 0.6036,
"step": 50
},
{
"epoch": 0.1214574898785425,
"grad_norm": 1.3692905065095424,
"learning_rate": 5e-06,
"loss": 0.5974,
"step": 60
},
{
"epoch": 0.1417004048582996,
"grad_norm": 1.817904836385203,
"learning_rate": 5e-06,
"loss": 0.5925,
"step": 70
},
{
"epoch": 0.16194331983805668,
"grad_norm": 1.37930069536892,
"learning_rate": 5e-06,
"loss": 0.5878,
"step": 80
},
{
"epoch": 0.18218623481781376,
"grad_norm": 1.9672481173378518,
"learning_rate": 5e-06,
"loss": 0.5869,
"step": 90
},
{
"epoch": 0.20242914979757085,
"grad_norm": 1.5190200995442689,
"learning_rate": 5e-06,
"loss": 0.5895,
"step": 100
},
{
"epoch": 0.22267206477732793,
"grad_norm": 1.467909545843528,
"learning_rate": 5e-06,
"loss": 0.5814,
"step": 110
},
{
"epoch": 0.242914979757085,
"grad_norm": 1.2624328705735892,
"learning_rate": 5e-06,
"loss": 0.579,
"step": 120
},
{
"epoch": 0.2631578947368421,
"grad_norm": 2.8330967598290644,
"learning_rate": 5e-06,
"loss": 0.5783,
"step": 130
},
{
"epoch": 0.2834008097165992,
"grad_norm": 2.606189250371495,
"learning_rate": 5e-06,
"loss": 0.5692,
"step": 140
},
{
"epoch": 0.30364372469635625,
"grad_norm": 2.0755945998925616,
"learning_rate": 5e-06,
"loss": 0.5778,
"step": 150
},
{
"epoch": 0.32388663967611336,
"grad_norm": 2.112489158829737,
"learning_rate": 5e-06,
"loss": 0.5757,
"step": 160
},
{
"epoch": 0.3441295546558704,
"grad_norm": 1.773147820799013,
"learning_rate": 5e-06,
"loss": 0.5757,
"step": 170
},
{
"epoch": 0.3643724696356275,
"grad_norm": 1.9389286348778036,
"learning_rate": 5e-06,
"loss": 0.5699,
"step": 180
},
{
"epoch": 0.38461538461538464,
"grad_norm": 2.0257261841165506,
"learning_rate": 5e-06,
"loss": 0.5708,
"step": 190
},
{
"epoch": 0.4048582995951417,
"grad_norm": 1.6235586407175255,
"learning_rate": 5e-06,
"loss": 0.5664,
"step": 200
},
{
"epoch": 0.4251012145748988,
"grad_norm": 1.3600120185742202,
"learning_rate": 5e-06,
"loss": 0.5624,
"step": 210
},
{
"epoch": 0.44534412955465585,
"grad_norm": 1.5782442409144959,
"learning_rate": 5e-06,
"loss": 0.5686,
"step": 220
},
{
"epoch": 0.46558704453441296,
"grad_norm": 1.9084679996506322,
"learning_rate": 5e-06,
"loss": 0.5681,
"step": 230
},
{
"epoch": 0.48582995951417,
"grad_norm": 1.562642052863713,
"learning_rate": 5e-06,
"loss": 0.5643,
"step": 240
},
{
"epoch": 0.5060728744939271,
"grad_norm": 1.9619840593224824,
"learning_rate": 5e-06,
"loss": 0.5666,
"step": 250
},
{
"epoch": 0.5263157894736842,
"grad_norm": 1.5861604962625888,
"learning_rate": 5e-06,
"loss": 0.5518,
"step": 260
},
{
"epoch": 0.5465587044534413,
"grad_norm": 1.7006302815881864,
"learning_rate": 5e-06,
"loss": 0.5605,
"step": 270
},
{
"epoch": 0.5668016194331984,
"grad_norm": 1.6651788716675646,
"learning_rate": 5e-06,
"loss": 0.5568,
"step": 280
},
{
"epoch": 0.5870445344129555,
"grad_norm": 2.2785089542841708,
"learning_rate": 5e-06,
"loss": 0.5565,
"step": 290
},
{
"epoch": 0.6072874493927125,
"grad_norm": 1.5770155712986973,
"learning_rate": 5e-06,
"loss": 0.5575,
"step": 300
},
{
"epoch": 0.6275303643724697,
"grad_norm": 1.3220149177446632,
"learning_rate": 5e-06,
"loss": 0.5622,
"step": 310
},
{
"epoch": 0.6477732793522267,
"grad_norm": 1.3472535369102203,
"learning_rate": 5e-06,
"loss": 0.5594,
"step": 320
},
{
"epoch": 0.6680161943319838,
"grad_norm": 1.3309997590990947,
"learning_rate": 5e-06,
"loss": 0.5601,
"step": 330
},
{
"epoch": 0.6882591093117408,
"grad_norm": 1.4273218646615096,
"learning_rate": 5e-06,
"loss": 0.5601,
"step": 340
},
{
"epoch": 0.708502024291498,
"grad_norm": 1.2085847860457979,
"learning_rate": 5e-06,
"loss": 0.5619,
"step": 350
},
{
"epoch": 0.728744939271255,
"grad_norm": 1.122941072415825,
"learning_rate": 5e-06,
"loss": 0.564,
"step": 360
},
{
"epoch": 0.7489878542510121,
"grad_norm": 1.3973216531402086,
"learning_rate": 5e-06,
"loss": 0.5571,
"step": 370
},
{
"epoch": 0.7692307692307693,
"grad_norm": 1.18082932151255,
"learning_rate": 5e-06,
"loss": 0.5538,
"step": 380
},
{
"epoch": 0.7894736842105263,
"grad_norm": 1.1780579152037594,
"learning_rate": 5e-06,
"loss": 0.557,
"step": 390
},
{
"epoch": 0.8097165991902834,
"grad_norm": 1.3785246959799171,
"learning_rate": 5e-06,
"loss": 0.5576,
"step": 400
},
{
"epoch": 0.8299595141700404,
"grad_norm": 1.2948034196229,
"learning_rate": 5e-06,
"loss": 0.5524,
"step": 410
},
{
"epoch": 0.8502024291497976,
"grad_norm": 1.186346298990548,
"learning_rate": 5e-06,
"loss": 0.5598,
"step": 420
},
{
"epoch": 0.8704453441295547,
"grad_norm": 1.1531795830071445,
"learning_rate": 5e-06,
"loss": 0.5532,
"step": 430
},
{
"epoch": 0.8906882591093117,
"grad_norm": 1.238207775770585,
"learning_rate": 5e-06,
"loss": 0.5587,
"step": 440
},
{
"epoch": 0.9109311740890689,
"grad_norm": 1.579785932754029,
"learning_rate": 5e-06,
"loss": 0.5499,
"step": 450
},
{
"epoch": 0.9311740890688259,
"grad_norm": 1.0932455184252672,
"learning_rate": 5e-06,
"loss": 0.5547,
"step": 460
},
{
"epoch": 0.951417004048583,
"grad_norm": 1.1455184663362503,
"learning_rate": 5e-06,
"loss": 0.5548,
"step": 470
},
{
"epoch": 0.97165991902834,
"grad_norm": 1.0783335686803248,
"learning_rate": 5e-06,
"loss": 0.5515,
"step": 480
},
{
"epoch": 0.9919028340080972,
"grad_norm": 1.221913718764988,
"learning_rate": 5e-06,
"loss": 0.5486,
"step": 490
},
{
"epoch": 1.0,
"eval_loss": 0.27600792050361633,
"eval_runtime": 127.9056,
"eval_samples_per_second": 104.014,
"eval_steps_per_second": 0.407,
"step": 494
},
{
"epoch": 1.0121457489878543,
"grad_norm": 1.946031690795471,
"learning_rate": 5e-06,
"loss": 0.5099,
"step": 500
},
{
"epoch": 1.0323886639676114,
"grad_norm": 1.9289680946263923,
"learning_rate": 5e-06,
"loss": 0.4748,
"step": 510
},
{
"epoch": 1.0526315789473684,
"grad_norm": 1.353788705122084,
"learning_rate": 5e-06,
"loss": 0.468,
"step": 520
},
{
"epoch": 1.0728744939271255,
"grad_norm": 1.205205720059485,
"learning_rate": 5e-06,
"loss": 0.4735,
"step": 530
},
{
"epoch": 1.0931174089068827,
"grad_norm": 1.282155131195443,
"learning_rate": 5e-06,
"loss": 0.4657,
"step": 540
},
{
"epoch": 1.1133603238866396,
"grad_norm": 1.5080942345199266,
"learning_rate": 5e-06,
"loss": 0.4699,
"step": 550
},
{
"epoch": 1.1336032388663968,
"grad_norm": 2.00023442913857,
"learning_rate": 5e-06,
"loss": 0.4659,
"step": 560
},
{
"epoch": 1.1538461538461537,
"grad_norm": 1.7909520333071323,
"learning_rate": 5e-06,
"loss": 0.4715,
"step": 570
},
{
"epoch": 1.174089068825911,
"grad_norm": 1.6662712618465199,
"learning_rate": 5e-06,
"loss": 0.4705,
"step": 580
},
{
"epoch": 1.194331983805668,
"grad_norm": 1.3259873704215552,
"learning_rate": 5e-06,
"loss": 0.4633,
"step": 590
},
{
"epoch": 1.214574898785425,
"grad_norm": 1.6651495485618664,
"learning_rate": 5e-06,
"loss": 0.4706,
"step": 600
},
{
"epoch": 1.2348178137651822,
"grad_norm": 1.477115453248806,
"learning_rate": 5e-06,
"loss": 0.4674,
"step": 610
},
{
"epoch": 1.2550607287449393,
"grad_norm": 1.263440522189552,
"learning_rate": 5e-06,
"loss": 0.4715,
"step": 620
},
{
"epoch": 1.2753036437246963,
"grad_norm": 1.4025183176779765,
"learning_rate": 5e-06,
"loss": 0.4715,
"step": 630
},
{
"epoch": 1.2955465587044535,
"grad_norm": 1.3060745387835275,
"learning_rate": 5e-06,
"loss": 0.473,
"step": 640
},
{
"epoch": 1.3157894736842106,
"grad_norm": 1.2936497262200077,
"learning_rate": 5e-06,
"loss": 0.4775,
"step": 650
},
{
"epoch": 1.3360323886639676,
"grad_norm": 1.2889809170808153,
"learning_rate": 5e-06,
"loss": 0.4671,
"step": 660
},
{
"epoch": 1.3562753036437247,
"grad_norm": 1.3063275232148253,
"learning_rate": 5e-06,
"loss": 0.4808,
"step": 670
},
{
"epoch": 1.376518218623482,
"grad_norm": 1.1392335054785994,
"learning_rate": 5e-06,
"loss": 0.4765,
"step": 680
},
{
"epoch": 1.3967611336032388,
"grad_norm": 1.1028492143829693,
"learning_rate": 5e-06,
"loss": 0.4759,
"step": 690
},
{
"epoch": 1.417004048582996,
"grad_norm": 1.1765124226702088,
"learning_rate": 5e-06,
"loss": 0.4739,
"step": 700
},
{
"epoch": 1.4372469635627532,
"grad_norm": 1.2359278437898007,
"learning_rate": 5e-06,
"loss": 0.4799,
"step": 710
},
{
"epoch": 1.45748987854251,
"grad_norm": 1.206361807359358,
"learning_rate": 5e-06,
"loss": 0.4792,
"step": 720
},
{
"epoch": 1.4777327935222673,
"grad_norm": 1.2584288631713865,
"learning_rate": 5e-06,
"loss": 0.4829,
"step": 730
},
{
"epoch": 1.4979757085020242,
"grad_norm": 1.3266851287176349,
"learning_rate": 5e-06,
"loss": 0.4757,
"step": 740
},
{
"epoch": 1.5182186234817814,
"grad_norm": 1.174062885309594,
"learning_rate": 5e-06,
"loss": 0.4786,
"step": 750
},
{
"epoch": 1.5384615384615383,
"grad_norm": 1.217286859959994,
"learning_rate": 5e-06,
"loss": 0.48,
"step": 760
},
{
"epoch": 1.5587044534412957,
"grad_norm": 1.2963071104766335,
"learning_rate": 5e-06,
"loss": 0.4771,
"step": 770
},
{
"epoch": 1.5789473684210527,
"grad_norm": 1.1301017979019057,
"learning_rate": 5e-06,
"loss": 0.4788,
"step": 780
},
{
"epoch": 1.5991902834008096,
"grad_norm": 1.2142771905592493,
"learning_rate": 5e-06,
"loss": 0.4813,
"step": 790
},
{
"epoch": 1.6194331983805668,
"grad_norm": 1.1156710660198195,
"learning_rate": 5e-06,
"loss": 0.478,
"step": 800
},
{
"epoch": 1.639676113360324,
"grad_norm": 1.3426333112263065,
"learning_rate": 5e-06,
"loss": 0.4712,
"step": 810
},
{
"epoch": 1.6599190283400809,
"grad_norm": 1.3351933873992157,
"learning_rate": 5e-06,
"loss": 0.4782,
"step": 820
},
{
"epoch": 1.680161943319838,
"grad_norm": 1.1798934266276804,
"learning_rate": 5e-06,
"loss": 0.4759,
"step": 830
},
{
"epoch": 1.7004048582995952,
"grad_norm": 1.0719853459322617,
"learning_rate": 5e-06,
"loss": 0.4751,
"step": 840
},
{
"epoch": 1.7206477732793521,
"grad_norm": 1.1539682879539461,
"learning_rate": 5e-06,
"loss": 0.4787,
"step": 850
},
{
"epoch": 1.7408906882591093,
"grad_norm": 1.1331656389881875,
"learning_rate": 5e-06,
"loss": 0.4782,
"step": 860
},
{
"epoch": 1.7611336032388665,
"grad_norm": 1.2305656250383992,
"learning_rate": 5e-06,
"loss": 0.4808,
"step": 870
},
{
"epoch": 1.7813765182186234,
"grad_norm": 1.484161550142479,
"learning_rate": 5e-06,
"loss": 0.4829,
"step": 880
},
{
"epoch": 1.8016194331983806,
"grad_norm": 1.180007815990233,
"learning_rate": 5e-06,
"loss": 0.4781,
"step": 890
},
{
"epoch": 1.8218623481781377,
"grad_norm": 1.200692474334057,
"learning_rate": 5e-06,
"loss": 0.4779,
"step": 900
},
{
"epoch": 1.8421052631578947,
"grad_norm": 1.1652057043804007,
"learning_rate": 5e-06,
"loss": 0.48,
"step": 910
},
{
"epoch": 1.8623481781376519,
"grad_norm": 1.2546941772227673,
"learning_rate": 5e-06,
"loss": 0.4796,
"step": 920
},
{
"epoch": 1.882591093117409,
"grad_norm": 1.1919644772038955,
"learning_rate": 5e-06,
"loss": 0.4754,
"step": 930
},
{
"epoch": 1.902834008097166,
"grad_norm": 1.086719187935772,
"learning_rate": 5e-06,
"loss": 0.4798,
"step": 940
},
{
"epoch": 1.9230769230769231,
"grad_norm": 1.1603089251078402,
"learning_rate": 5e-06,
"loss": 0.4816,
"step": 950
},
{
"epoch": 1.9433198380566803,
"grad_norm": 1.082753063279857,
"learning_rate": 5e-06,
"loss": 0.4842,
"step": 960
},
{
"epoch": 1.9635627530364372,
"grad_norm": 1.1470965026545579,
"learning_rate": 5e-06,
"loss": 0.4826,
"step": 970
},
{
"epoch": 1.9838056680161942,
"grad_norm": 1.317035146950926,
"learning_rate": 5e-06,
"loss": 0.4834,
"step": 980
},
{
"epoch": 2.0,
"eval_loss": 0.2760300040245056,
"eval_runtime": 128.3279,
"eval_samples_per_second": 103.672,
"eval_steps_per_second": 0.405,
"step": 988
},
{
"epoch": 2.0040485829959516,
"grad_norm": 2.423406809168664,
"learning_rate": 5e-06,
"loss": 0.4596,
"step": 990
},
{
"epoch": 2.0242914979757085,
"grad_norm": 1.8171058614039022,
"learning_rate": 5e-06,
"loss": 0.3883,
"step": 1000
},
{
"epoch": 2.0445344129554655,
"grad_norm": 1.5419534841689435,
"learning_rate": 5e-06,
"loss": 0.3836,
"step": 1010
},
{
"epoch": 2.064777327935223,
"grad_norm": 1.476614885030913,
"learning_rate": 5e-06,
"loss": 0.3801,
"step": 1020
},
{
"epoch": 2.08502024291498,
"grad_norm": 1.379440361306421,
"learning_rate": 5e-06,
"loss": 0.3784,
"step": 1030
},
{
"epoch": 2.1052631578947367,
"grad_norm": 1.4083322026387224,
"learning_rate": 5e-06,
"loss": 0.3778,
"step": 1040
},
{
"epoch": 2.125506072874494,
"grad_norm": 1.4696113259366608,
"learning_rate": 5e-06,
"loss": 0.3879,
"step": 1050
},
{
"epoch": 2.145748987854251,
"grad_norm": 1.2713132554208233,
"learning_rate": 5e-06,
"loss": 0.3806,
"step": 1060
},
{
"epoch": 2.165991902834008,
"grad_norm": 1.3917546218353265,
"learning_rate": 5e-06,
"loss": 0.386,
"step": 1070
},
{
"epoch": 2.1862348178137654,
"grad_norm": 1.4708546369587632,
"learning_rate": 5e-06,
"loss": 0.3845,
"step": 1080
},
{
"epoch": 2.2064777327935223,
"grad_norm": 1.595163601660772,
"learning_rate": 5e-06,
"loss": 0.3828,
"step": 1090
},
{
"epoch": 2.2267206477732793,
"grad_norm": 1.4877074059133855,
"learning_rate": 5e-06,
"loss": 0.3897,
"step": 1100
},
{
"epoch": 2.246963562753036,
"grad_norm": 1.7370364288636082,
"learning_rate": 5e-06,
"loss": 0.3875,
"step": 1110
},
{
"epoch": 2.2672064777327936,
"grad_norm": 1.8541525908514815,
"learning_rate": 5e-06,
"loss": 0.3884,
"step": 1120
},
{
"epoch": 2.2874493927125505,
"grad_norm": 1.5781496384158298,
"learning_rate": 5e-06,
"loss": 0.3894,
"step": 1130
},
{
"epoch": 2.3076923076923075,
"grad_norm": 1.7708762236924664,
"learning_rate": 5e-06,
"loss": 0.3859,
"step": 1140
},
{
"epoch": 2.327935222672065,
"grad_norm": 1.4475591252123796,
"learning_rate": 5e-06,
"loss": 0.3887,
"step": 1150
},
{
"epoch": 2.348178137651822,
"grad_norm": 1.2226909125336796,
"learning_rate": 5e-06,
"loss": 0.3885,
"step": 1160
},
{
"epoch": 2.3684210526315788,
"grad_norm": 1.3657056897603896,
"learning_rate": 5e-06,
"loss": 0.392,
"step": 1170
},
{
"epoch": 2.388663967611336,
"grad_norm": 1.314590138508193,
"learning_rate": 5e-06,
"loss": 0.3926,
"step": 1180
},
{
"epoch": 2.408906882591093,
"grad_norm": 1.4093248309707875,
"learning_rate": 5e-06,
"loss": 0.3929,
"step": 1190
},
{
"epoch": 2.42914979757085,
"grad_norm": 1.3473340294150489,
"learning_rate": 5e-06,
"loss": 0.39,
"step": 1200
},
{
"epoch": 2.4493927125506074,
"grad_norm": 1.4512860257939517,
"learning_rate": 5e-06,
"loss": 0.3902,
"step": 1210
},
{
"epoch": 2.4696356275303644,
"grad_norm": 1.4529803502641332,
"learning_rate": 5e-06,
"loss": 0.3947,
"step": 1220
},
{
"epoch": 2.4898785425101213,
"grad_norm": 1.384452479045095,
"learning_rate": 5e-06,
"loss": 0.3953,
"step": 1230
},
{
"epoch": 2.5101214574898787,
"grad_norm": 1.2935497884972742,
"learning_rate": 5e-06,
"loss": 0.3986,
"step": 1240
},
{
"epoch": 2.5303643724696356,
"grad_norm": 1.190834719805313,
"learning_rate": 5e-06,
"loss": 0.3914,
"step": 1250
},
{
"epoch": 2.5506072874493926,
"grad_norm": 1.5623940633492222,
"learning_rate": 5e-06,
"loss": 0.3887,
"step": 1260
},
{
"epoch": 2.57085020242915,
"grad_norm": 1.4071250725201772,
"learning_rate": 5e-06,
"loss": 0.3941,
"step": 1270
},
{
"epoch": 2.591093117408907,
"grad_norm": 1.4686855669517231,
"learning_rate": 5e-06,
"loss": 0.3956,
"step": 1280
},
{
"epoch": 2.611336032388664,
"grad_norm": 1.29119112691104,
"learning_rate": 5e-06,
"loss": 0.3998,
"step": 1290
},
{
"epoch": 2.6315789473684212,
"grad_norm": 1.2673323070448081,
"learning_rate": 5e-06,
"loss": 0.3961,
"step": 1300
},
{
"epoch": 2.651821862348178,
"grad_norm": 1.2450186102009178,
"learning_rate": 5e-06,
"loss": 0.395,
"step": 1310
},
{
"epoch": 2.672064777327935,
"grad_norm": 1.3598218086289704,
"learning_rate": 5e-06,
"loss": 0.3969,
"step": 1320
},
{
"epoch": 2.6923076923076925,
"grad_norm": 1.476188415686639,
"learning_rate": 5e-06,
"loss": 0.401,
"step": 1330
},
{
"epoch": 2.7125506072874495,
"grad_norm": 1.2772087881864984,
"learning_rate": 5e-06,
"loss": 0.3981,
"step": 1340
},
{
"epoch": 2.7327935222672064,
"grad_norm": 1.3901040643475941,
"learning_rate": 5e-06,
"loss": 0.399,
"step": 1350
},
{
"epoch": 2.753036437246964,
"grad_norm": 1.2669956962100783,
"learning_rate": 5e-06,
"loss": 0.3936,
"step": 1360
},
{
"epoch": 2.7732793522267207,
"grad_norm": 1.3605832094262316,
"learning_rate": 5e-06,
"loss": 0.3996,
"step": 1370
},
{
"epoch": 2.7935222672064777,
"grad_norm": 1.4625604256084817,
"learning_rate": 5e-06,
"loss": 0.3997,
"step": 1380
},
{
"epoch": 2.813765182186235,
"grad_norm": 1.784848571739763,
"learning_rate": 5e-06,
"loss": 0.4012,
"step": 1390
},
{
"epoch": 2.834008097165992,
"grad_norm": 1.6211701201270445,
"learning_rate": 5e-06,
"loss": 0.3948,
"step": 1400
},
{
"epoch": 2.854251012145749,
"grad_norm": 1.5362470053479813,
"learning_rate": 5e-06,
"loss": 0.4017,
"step": 1410
},
{
"epoch": 2.8744939271255063,
"grad_norm": 1.4029392342160418,
"learning_rate": 5e-06,
"loss": 0.3992,
"step": 1420
},
{
"epoch": 2.8947368421052633,
"grad_norm": 1.4342688190560124,
"learning_rate": 5e-06,
"loss": 0.395,
"step": 1430
},
{
"epoch": 2.91497975708502,
"grad_norm": 1.2695118103828797,
"learning_rate": 5e-06,
"loss": 0.4008,
"step": 1440
},
{
"epoch": 2.9352226720647776,
"grad_norm": 1.2741085284215934,
"learning_rate": 5e-06,
"loss": 0.4111,
"step": 1450
},
{
"epoch": 2.9554655870445345,
"grad_norm": 1.2910941516817205,
"learning_rate": 5e-06,
"loss": 0.3975,
"step": 1460
},
{
"epoch": 2.9757085020242915,
"grad_norm": 1.4853937254356013,
"learning_rate": 5e-06,
"loss": 0.4012,
"step": 1470
},
{
"epoch": 2.9959514170040484,
"grad_norm": 1.4022248360408927,
"learning_rate": 5e-06,
"loss": 0.4034,
"step": 1480
},
{
"epoch": 3.0,
"eval_loss": 0.29170867800712585,
"eval_runtime": 128.4123,
"eval_samples_per_second": 103.604,
"eval_steps_per_second": 0.405,
"step": 1482
},
{
"epoch": 3.016194331983806,
"grad_norm": 2.450301362827223,
"learning_rate": 5e-06,
"loss": 0.3253,
"step": 1490
},
{
"epoch": 3.0364372469635628,
"grad_norm": 1.7227064520999624,
"learning_rate": 5e-06,
"loss": 0.2917,
"step": 1500
},
{
"epoch": 3.0566801619433197,
"grad_norm": 1.708297211407951,
"learning_rate": 5e-06,
"loss": 0.2951,
"step": 1510
},
{
"epoch": 3.076923076923077,
"grad_norm": 1.6773762742996041,
"learning_rate": 5e-06,
"loss": 0.2916,
"step": 1520
},
{
"epoch": 3.097165991902834,
"grad_norm": 1.5835044005502448,
"learning_rate": 5e-06,
"loss": 0.2949,
"step": 1530
},
{
"epoch": 3.117408906882591,
"grad_norm": 1.6991756969105973,
"learning_rate": 5e-06,
"loss": 0.2906,
"step": 1540
},
{
"epoch": 3.1376518218623484,
"grad_norm": 1.6309085248071211,
"learning_rate": 5e-06,
"loss": 0.2887,
"step": 1550
},
{
"epoch": 3.1578947368421053,
"grad_norm": 1.467082791239925,
"learning_rate": 5e-06,
"loss": 0.2958,
"step": 1560
},
{
"epoch": 3.1781376518218623,
"grad_norm": 1.5977415982902385,
"learning_rate": 5e-06,
"loss": 0.2958,
"step": 1570
},
{
"epoch": 3.1983805668016196,
"grad_norm": 1.794838585622722,
"learning_rate": 5e-06,
"loss": 0.2985,
"step": 1580
},
{
"epoch": 3.2186234817813766,
"grad_norm": 1.6188047839576813,
"learning_rate": 5e-06,
"loss": 0.3009,
"step": 1590
},
{
"epoch": 3.2388663967611335,
"grad_norm": 1.5996247856636512,
"learning_rate": 5e-06,
"loss": 0.2993,
"step": 1600
},
{
"epoch": 3.2591093117408905,
"grad_norm": 1.5799607329692924,
"learning_rate": 5e-06,
"loss": 0.2981,
"step": 1610
},
{
"epoch": 3.279352226720648,
"grad_norm": 1.6462393560491397,
"learning_rate": 5e-06,
"loss": 0.3014,
"step": 1620
},
{
"epoch": 3.299595141700405,
"grad_norm": 1.743995264000295,
"learning_rate": 5e-06,
"loss": 0.3034,
"step": 1630
},
{
"epoch": 3.3198380566801617,
"grad_norm": 1.6858202723993232,
"learning_rate": 5e-06,
"loss": 0.303,
"step": 1640
},
{
"epoch": 3.340080971659919,
"grad_norm": 1.8191610290572566,
"learning_rate": 5e-06,
"loss": 0.3003,
"step": 1650
},
{
"epoch": 3.360323886639676,
"grad_norm": 1.601413073895001,
"learning_rate": 5e-06,
"loss": 0.3051,
"step": 1660
},
{
"epoch": 3.380566801619433,
"grad_norm": 1.7563790440673033,
"learning_rate": 5e-06,
"loss": 0.2999,
"step": 1670
},
{
"epoch": 3.4008097165991904,
"grad_norm": 1.5579994158970751,
"learning_rate": 5e-06,
"loss": 0.2989,
"step": 1680
},
{
"epoch": 3.4210526315789473,
"grad_norm": 1.516213960669368,
"learning_rate": 5e-06,
"loss": 0.3011,
"step": 1690
},
{
"epoch": 3.4412955465587043,
"grad_norm": 1.658051304756316,
"learning_rate": 5e-06,
"loss": 0.3024,
"step": 1700
},
{
"epoch": 3.4615384615384617,
"grad_norm": 1.5918375939252065,
"learning_rate": 5e-06,
"loss": 0.3056,
"step": 1710
},
{
"epoch": 3.4817813765182186,
"grad_norm": 1.6167160427713403,
"learning_rate": 5e-06,
"loss": 0.3059,
"step": 1720
},
{
"epoch": 3.5020242914979756,
"grad_norm": 1.633353659275257,
"learning_rate": 5e-06,
"loss": 0.3103,
"step": 1730
},
{
"epoch": 3.522267206477733,
"grad_norm": 1.6652424042294116,
"learning_rate": 5e-06,
"loss": 0.3047,
"step": 1740
},
{
"epoch": 3.54251012145749,
"grad_norm": 1.5945478465514338,
"learning_rate": 5e-06,
"loss": 0.3097,
"step": 1750
},
{
"epoch": 3.562753036437247,
"grad_norm": 1.4266070240229398,
"learning_rate": 5e-06,
"loss": 0.3053,
"step": 1760
},
{
"epoch": 3.582995951417004,
"grad_norm": 1.5417916459202972,
"learning_rate": 5e-06,
"loss": 0.3107,
"step": 1770
},
{
"epoch": 3.603238866396761,
"grad_norm": 1.6806576746631992,
"learning_rate": 5e-06,
"loss": 0.3077,
"step": 1780
},
{
"epoch": 3.623481781376518,
"grad_norm": 1.6366680075861002,
"learning_rate": 5e-06,
"loss": 0.3096,
"step": 1790
},
{
"epoch": 3.6437246963562755,
"grad_norm": 1.8008610671747132,
"learning_rate": 5e-06,
"loss": 0.3095,
"step": 1800
},
{
"epoch": 3.6639676113360324,
"grad_norm": 1.7218050046753175,
"learning_rate": 5e-06,
"loss": 0.3129,
"step": 1810
},
{
"epoch": 3.6842105263157894,
"grad_norm": 1.54966815167092,
"learning_rate": 5e-06,
"loss": 0.3089,
"step": 1820
},
{
"epoch": 3.7044534412955468,
"grad_norm": 1.5907699518804959,
"learning_rate": 5e-06,
"loss": 0.3115,
"step": 1830
},
{
"epoch": 3.7246963562753037,
"grad_norm": 1.5950760981279428,
"learning_rate": 5e-06,
"loss": 0.3132,
"step": 1840
},
{
"epoch": 3.7449392712550607,
"grad_norm": 1.6178462612967588,
"learning_rate": 5e-06,
"loss": 0.3112,
"step": 1850
},
{
"epoch": 3.765182186234818,
"grad_norm": 1.6595052047937346,
"learning_rate": 5e-06,
"loss": 0.3144,
"step": 1860
},
{
"epoch": 3.785425101214575,
"grad_norm": 1.5960921622659012,
"learning_rate": 5e-06,
"loss": 0.3181,
"step": 1870
},
{
"epoch": 3.805668016194332,
"grad_norm": 1.8789757609659636,
"learning_rate": 5e-06,
"loss": 0.3128,
"step": 1880
},
{
"epoch": 3.8259109311740893,
"grad_norm": 1.6017653115735,
"learning_rate": 5e-06,
"loss": 0.318,
"step": 1890
},
{
"epoch": 3.8461538461538463,
"grad_norm": 1.4883871110556444,
"learning_rate": 5e-06,
"loss": 0.3168,
"step": 1900
},
{
"epoch": 3.866396761133603,
"grad_norm": 1.7783804818718214,
"learning_rate": 5e-06,
"loss": 0.3114,
"step": 1910
},
{
"epoch": 3.8866396761133606,
"grad_norm": 1.7097670030385954,
"learning_rate": 5e-06,
"loss": 0.3196,
"step": 1920
},
{
"epoch": 3.9068825910931175,
"grad_norm": 1.519954953217676,
"learning_rate": 5e-06,
"loss": 0.3161,
"step": 1930
},
{
"epoch": 3.9271255060728745,
"grad_norm": 1.5786075983313086,
"learning_rate": 5e-06,
"loss": 0.3142,
"step": 1940
},
{
"epoch": 3.9473684210526314,
"grad_norm": 1.8339039956017331,
"learning_rate": 5e-06,
"loss": 0.3156,
"step": 1950
},
{
"epoch": 3.967611336032389,
"grad_norm": 1.504586213513185,
"learning_rate": 5e-06,
"loss": 0.3173,
"step": 1960
},
{
"epoch": 3.9878542510121457,
"grad_norm": 1.603597507259475,
"learning_rate": 5e-06,
"loss": 0.3182,
"step": 1970
},
{
"epoch": 4.0,
"eval_loss": 0.3246242105960846,
"eval_runtime": 128.1021,
"eval_samples_per_second": 103.855,
"eval_steps_per_second": 0.406,
"step": 1976
},
{
"epoch": 4.008097165991903,
"grad_norm": 5.133136743732594,
"learning_rate": 5e-06,
"loss": 0.2774,
"step": 1980
},
{
"epoch": 4.02834008097166,
"grad_norm": 2.5228879416463936,
"learning_rate": 5e-06,
"loss": 0.2123,
"step": 1990
},
{
"epoch": 4.048582995951417,
"grad_norm": 2.0169504998470833,
"learning_rate": 5e-06,
"loss": 0.2057,
"step": 2000
},
{
"epoch": 4.068825910931174,
"grad_norm": 1.9085723632427691,
"learning_rate": 5e-06,
"loss": 0.2061,
"step": 2010
},
{
"epoch": 4.089068825910931,
"grad_norm": 1.8551214473550242,
"learning_rate": 5e-06,
"loss": 0.2058,
"step": 2020
},
{
"epoch": 4.109311740890688,
"grad_norm": 1.8804505495475972,
"learning_rate": 5e-06,
"loss": 0.2076,
"step": 2030
},
{
"epoch": 4.129554655870446,
"grad_norm": 1.8325377033435544,
"learning_rate": 5e-06,
"loss": 0.2055,
"step": 2040
},
{
"epoch": 4.149797570850202,
"grad_norm": 1.892921271335565,
"learning_rate": 5e-06,
"loss": 0.2073,
"step": 2050
},
{
"epoch": 4.17004048582996,
"grad_norm": 2.00985346837552,
"learning_rate": 5e-06,
"loss": 0.2084,
"step": 2060
},
{
"epoch": 4.190283400809717,
"grad_norm": 1.820158031557001,
"learning_rate": 5e-06,
"loss": 0.209,
"step": 2070
},
{
"epoch": 4.2105263157894735,
"grad_norm": 1.8327459405866524,
"learning_rate": 5e-06,
"loss": 0.2127,
"step": 2080
},
{
"epoch": 4.230769230769231,
"grad_norm": 1.9394580410768052,
"learning_rate": 5e-06,
"loss": 0.2123,
"step": 2090
},
{
"epoch": 4.251012145748988,
"grad_norm": 1.9657450113344472,
"learning_rate": 5e-06,
"loss": 0.2108,
"step": 2100
},
{
"epoch": 4.271255060728745,
"grad_norm": 1.949924211787456,
"learning_rate": 5e-06,
"loss": 0.2139,
"step": 2110
},
{
"epoch": 4.291497975708502,
"grad_norm": 1.8700037789000614,
"learning_rate": 5e-06,
"loss": 0.2146,
"step": 2120
},
{
"epoch": 4.3117408906882595,
"grad_norm": 1.8948956761363887,
"learning_rate": 5e-06,
"loss": 0.2145,
"step": 2130
},
{
"epoch": 4.331983805668016,
"grad_norm": 2.183456888675007,
"learning_rate": 5e-06,
"loss": 0.2168,
"step": 2140
},
{
"epoch": 4.352226720647773,
"grad_norm": 2.01397562728079,
"learning_rate": 5e-06,
"loss": 0.2151,
"step": 2150
},
{
"epoch": 4.372469635627531,
"grad_norm": 2.1466988238739697,
"learning_rate": 5e-06,
"loss": 0.2148,
"step": 2160
},
{
"epoch": 4.392712550607287,
"grad_norm": 1.8790514185213776,
"learning_rate": 5e-06,
"loss": 0.2159,
"step": 2170
},
{
"epoch": 4.412955465587045,
"grad_norm": 1.9845491426747,
"learning_rate": 5e-06,
"loss": 0.2179,
"step": 2180
},
{
"epoch": 4.433198380566802,
"grad_norm": 1.9657898682355717,
"learning_rate": 5e-06,
"loss": 0.2173,
"step": 2190
},
{
"epoch": 4.4534412955465585,
"grad_norm": 1.8874931048307386,
"learning_rate": 5e-06,
"loss": 0.2156,
"step": 2200
},
{
"epoch": 4.473684210526316,
"grad_norm": 1.845963836099471,
"learning_rate": 5e-06,
"loss": 0.2188,
"step": 2210
},
{
"epoch": 4.493927125506072,
"grad_norm": 1.9372215749139772,
"learning_rate": 5e-06,
"loss": 0.2191,
"step": 2220
},
{
"epoch": 4.51417004048583,
"grad_norm": 2.150693124788377,
"learning_rate": 5e-06,
"loss": 0.2201,
"step": 2230
},
{
"epoch": 4.534412955465587,
"grad_norm": 2.0442818699824317,
"learning_rate": 5e-06,
"loss": 0.22,
"step": 2240
},
{
"epoch": 4.554655870445345,
"grad_norm": 1.9514620437414862,
"learning_rate": 5e-06,
"loss": 0.2185,
"step": 2250
},
{
"epoch": 4.574898785425101,
"grad_norm": 1.8963967573781124,
"learning_rate": 5e-06,
"loss": 0.2213,
"step": 2260
},
{
"epoch": 4.5951417004048585,
"grad_norm": 1.8526961706560494,
"learning_rate": 5e-06,
"loss": 0.2221,
"step": 2270
},
{
"epoch": 4.615384615384615,
"grad_norm": 1.9821664496420446,
"learning_rate": 5e-06,
"loss": 0.2214,
"step": 2280
},
{
"epoch": 4.635627530364372,
"grad_norm": 1.915513929466144,
"learning_rate": 5e-06,
"loss": 0.2237,
"step": 2290
},
{
"epoch": 4.65587044534413,
"grad_norm": 1.8762006846363004,
"learning_rate": 5e-06,
"loss": 0.2249,
"step": 2300
},
{
"epoch": 4.676113360323887,
"grad_norm": 1.9227849786208417,
"learning_rate": 5e-06,
"loss": 0.2265,
"step": 2310
},
{
"epoch": 4.696356275303644,
"grad_norm": 1.9572380827157467,
"learning_rate": 5e-06,
"loss": 0.2265,
"step": 2320
},
{
"epoch": 4.716599190283401,
"grad_norm": 1.8826393602367635,
"learning_rate": 5e-06,
"loss": 0.2252,
"step": 2330
},
{
"epoch": 4.7368421052631575,
"grad_norm": 1.825743956937918,
"learning_rate": 5e-06,
"loss": 0.2234,
"step": 2340
},
{
"epoch": 4.757085020242915,
"grad_norm": 2.194102804306542,
"learning_rate": 5e-06,
"loss": 0.2243,
"step": 2350
},
{
"epoch": 4.777327935222672,
"grad_norm": 1.8765617982862395,
"learning_rate": 5e-06,
"loss": 0.2246,
"step": 2360
},
{
"epoch": 4.797570850202429,
"grad_norm": 1.8687466226351828,
"learning_rate": 5e-06,
"loss": 0.2276,
"step": 2370
},
{
"epoch": 4.817813765182186,
"grad_norm": 1.7759017848087686,
"learning_rate": 5e-06,
"loss": 0.2257,
"step": 2380
},
{
"epoch": 4.838056680161944,
"grad_norm": 1.8906492602012772,
"learning_rate": 5e-06,
"loss": 0.2253,
"step": 2390
},
{
"epoch": 4.8582995951417,
"grad_norm": 2.050008885412097,
"learning_rate": 5e-06,
"loss": 0.2294,
"step": 2400
},
{
"epoch": 4.8785425101214575,
"grad_norm": 1.9449950754235916,
"learning_rate": 5e-06,
"loss": 0.2307,
"step": 2410
},
{
"epoch": 4.898785425101215,
"grad_norm": 1.866407529076546,
"learning_rate": 5e-06,
"loss": 0.2293,
"step": 2420
},
{
"epoch": 4.919028340080971,
"grad_norm": 1.9434490467739143,
"learning_rate": 5e-06,
"loss": 0.2285,
"step": 2430
},
{
"epoch": 4.939271255060729,
"grad_norm": 1.9646396229024239,
"learning_rate": 5e-06,
"loss": 0.2306,
"step": 2440
},
{
"epoch": 4.959514170040486,
"grad_norm": 1.8683726614276728,
"learning_rate": 5e-06,
"loss": 0.2277,
"step": 2450
},
{
"epoch": 4.979757085020243,
"grad_norm": 1.984199504770925,
"learning_rate": 5e-06,
"loss": 0.2312,
"step": 2460
},
{
"epoch": 5.0,
"grad_norm": 1.8362584513440876,
"learning_rate": 5e-06,
"loss": 0.2292,
"step": 2470
},
{
"epoch": 5.0,
"eval_loss": 0.37886306643486023,
"eval_runtime": 128.1269,
"eval_samples_per_second": 103.835,
"eval_steps_per_second": 0.406,
"step": 2470
},
{
"epoch": 5.0,
"step": 2470,
"total_flos": 4136504477614080.0,
"train_loss": 0.39363470135430095,
"train_runtime": 36289.8965,
"train_samples_per_second": 34.825,
"train_steps_per_second": 0.068
}
],
"logging_steps": 10,
"max_steps": 2470,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 4136504477614080.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}