GX-XinGao's picture
Initial release v1.0
4d227ce verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 1925,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.005194805194805195,
"grad_norm": 1.3003292727167528,
"learning_rate": 2.8735632183908046e-06,
"loss": 0.5747,
"step": 10
},
{
"epoch": 0.01038961038961039,
"grad_norm": 0.8164205391130951,
"learning_rate": 5.747126436781609e-06,
"loss": 0.5181,
"step": 20
},
{
"epoch": 0.015584415584415584,
"grad_norm": 0.488946468114722,
"learning_rate": 8.620689655172414e-06,
"loss": 0.4525,
"step": 30
},
{
"epoch": 0.02077922077922078,
"grad_norm": 0.5240257767228009,
"learning_rate": 1.1494252873563218e-05,
"loss": 0.4153,
"step": 40
},
{
"epoch": 0.025974025974025976,
"grad_norm": 0.3878833187137561,
"learning_rate": 1.4367816091954022e-05,
"loss": 0.3917,
"step": 50
},
{
"epoch": 0.03116883116883117,
"grad_norm": 0.3932438541306065,
"learning_rate": 1.7241379310344828e-05,
"loss": 0.3707,
"step": 60
},
{
"epoch": 0.03636363636363636,
"grad_norm": 0.4466114932768719,
"learning_rate": 2.0114942528735632e-05,
"loss": 0.3605,
"step": 70
},
{
"epoch": 0.04155844155844156,
"grad_norm": 0.8032738854038463,
"learning_rate": 2.2988505747126437e-05,
"loss": 0.3591,
"step": 80
},
{
"epoch": 0.046753246753246755,
"grad_norm": 0.543812557370986,
"learning_rate": 2.5862068965517244e-05,
"loss": 0.3472,
"step": 90
},
{
"epoch": 0.05194805194805195,
"grad_norm": 0.5733743111451876,
"learning_rate": 2.8735632183908045e-05,
"loss": 0.3341,
"step": 100
},
{
"epoch": 0.05714285714285714,
"grad_norm": 0.9136032757030011,
"learning_rate": 3.160919540229885e-05,
"loss": 0.34,
"step": 110
},
{
"epoch": 0.06233766233766234,
"grad_norm": 0.8015940010159968,
"learning_rate": 3.4482758620689657e-05,
"loss": 0.3346,
"step": 120
},
{
"epoch": 0.06753246753246753,
"grad_norm": 0.5317651402718742,
"learning_rate": 3.735632183908046e-05,
"loss": 0.3205,
"step": 130
},
{
"epoch": 0.07272727272727272,
"grad_norm": 0.9794116224334949,
"learning_rate": 4.0229885057471265e-05,
"loss": 0.3247,
"step": 140
},
{
"epoch": 0.07792207792207792,
"grad_norm": 0.5926607382301553,
"learning_rate": 4.3103448275862066e-05,
"loss": 0.3137,
"step": 150
},
{
"epoch": 0.08311688311688312,
"grad_norm": 0.501504811182817,
"learning_rate": 4.597701149425287e-05,
"loss": 0.3196,
"step": 160
},
{
"epoch": 0.08831168831168831,
"grad_norm": 0.4844466938932139,
"learning_rate": 4.885057471264368e-05,
"loss": 0.3185,
"step": 170
},
{
"epoch": 0.09350649350649351,
"grad_norm": 0.8792222600261355,
"learning_rate": 4.999985842691236e-05,
"loss": 0.317,
"step": 180
},
{
"epoch": 0.0987012987012987,
"grad_norm": 0.538260729358137,
"learning_rate": 4.999899326385009e-05,
"loss": 0.3122,
"step": 190
},
{
"epoch": 0.1038961038961039,
"grad_norm": 0.5532332343125024,
"learning_rate": 4.99973416166265e-05,
"loss": 0.308,
"step": 200
},
{
"epoch": 0.10909090909090909,
"grad_norm": 0.5065931571721304,
"learning_rate": 4.999490353720347e-05,
"loss": 0.305,
"step": 210
},
{
"epoch": 0.11428571428571428,
"grad_norm": 0.4661004020519144,
"learning_rate": 4.9991679102284494e-05,
"loss": 0.3031,
"step": 220
},
{
"epoch": 0.11948051948051948,
"grad_norm": 0.4350447913864305,
"learning_rate": 4.998766841331236e-05,
"loss": 0.2979,
"step": 230
},
{
"epoch": 0.12467532467532468,
"grad_norm": 0.3782265622233307,
"learning_rate": 4.998287159646586e-05,
"loss": 0.3035,
"step": 240
},
{
"epoch": 0.12987012987012986,
"grad_norm": 0.3863507226733708,
"learning_rate": 4.997728880265592e-05,
"loss": 0.3024,
"step": 250
},
{
"epoch": 0.13506493506493505,
"grad_norm": 0.45465564793802865,
"learning_rate": 4.9970920207520756e-05,
"loss": 0.2984,
"step": 260
},
{
"epoch": 0.14025974025974025,
"grad_norm": 0.331617769411873,
"learning_rate": 4.9963766011420394e-05,
"loss": 0.2947,
"step": 270
},
{
"epoch": 0.14545454545454545,
"grad_norm": 0.4323281120205803,
"learning_rate": 4.9955826439430384e-05,
"loss": 0.2885,
"step": 280
},
{
"epoch": 0.15064935064935064,
"grad_norm": 0.3828139513983012,
"learning_rate": 4.994710174133469e-05,
"loss": 0.2957,
"step": 290
},
{
"epoch": 0.15584415584415584,
"grad_norm": 0.45442672024966796,
"learning_rate": 4.9937592191617846e-05,
"loss": 0.2929,
"step": 300
},
{
"epoch": 0.16103896103896104,
"grad_norm": 0.4324048943234123,
"learning_rate": 4.992729808945629e-05,
"loss": 0.287,
"step": 310
},
{
"epoch": 0.16623376623376623,
"grad_norm": 0.3358736775493111,
"learning_rate": 4.991621975870901e-05,
"loss": 0.2831,
"step": 320
},
{
"epoch": 0.17142857142857143,
"grad_norm": 0.33113790181444713,
"learning_rate": 4.990435754790731e-05,
"loss": 0.2868,
"step": 330
},
{
"epoch": 0.17662337662337663,
"grad_norm": 0.3207965425558816,
"learning_rate": 4.9891711830243845e-05,
"loss": 0.2911,
"step": 340
},
{
"epoch": 0.18181818181818182,
"grad_norm": 0.361708114933415,
"learning_rate": 4.987828300356091e-05,
"loss": 0.2857,
"step": 350
},
{
"epoch": 0.18701298701298702,
"grad_norm": 0.34168626805695096,
"learning_rate": 4.9864071490337896e-05,
"loss": 0.2849,
"step": 360
},
{
"epoch": 0.19220779220779222,
"grad_norm": 0.2908926746764898,
"learning_rate": 4.9849077737678e-05,
"loss": 0.2794,
"step": 370
},
{
"epoch": 0.1974025974025974,
"grad_norm": 0.5087944109212423,
"learning_rate": 4.983330221729419e-05,
"loss": 0.2787,
"step": 380
},
{
"epoch": 0.2025974025974026,
"grad_norm": 0.27804001121817934,
"learning_rate": 4.9816745425494326e-05,
"loss": 0.2759,
"step": 390
},
{
"epoch": 0.2077922077922078,
"grad_norm": 0.4013635528070106,
"learning_rate": 4.979940788316556e-05,
"loss": 0.2817,
"step": 400
},
{
"epoch": 0.21298701298701297,
"grad_norm": 0.25511672135848257,
"learning_rate": 4.978129013575796e-05,
"loss": 0.2785,
"step": 410
},
{
"epoch": 0.21818181818181817,
"grad_norm": 0.298813003536876,
"learning_rate": 4.976239275326733e-05,
"loss": 0.2803,
"step": 420
},
{
"epoch": 0.22337662337662337,
"grad_norm": 0.25617082862388485,
"learning_rate": 4.974271633021729e-05,
"loss": 0.2736,
"step": 430
},
{
"epoch": 0.22857142857142856,
"grad_norm": 0.29247425322362797,
"learning_rate": 4.9722261485640584e-05,
"loss": 0.2767,
"step": 440
},
{
"epoch": 0.23376623376623376,
"grad_norm": 0.33775737625365165,
"learning_rate": 4.9701028863059563e-05,
"loss": 0.2753,
"step": 450
},
{
"epoch": 0.23896103896103896,
"grad_norm": 0.38447368527387143,
"learning_rate": 4.967901913046598e-05,
"loss": 0.2805,
"step": 460
},
{
"epoch": 0.24415584415584415,
"grad_norm": 0.36418471782114614,
"learning_rate": 4.9656232980299976e-05,
"loss": 0.2707,
"step": 470
},
{
"epoch": 0.24935064935064935,
"grad_norm": 0.2826049309528878,
"learning_rate": 4.963267112942826e-05,
"loss": 0.2775,
"step": 480
},
{
"epoch": 0.2545454545454545,
"grad_norm": 0.3501209948268494,
"learning_rate": 4.9608334319121584e-05,
"loss": 0.2731,
"step": 490
},
{
"epoch": 0.2597402597402597,
"grad_norm": 0.3608819312798831,
"learning_rate": 4.958322331503141e-05,
"loss": 0.2707,
"step": 500
},
{
"epoch": 0.2649350649350649,
"grad_norm": 0.299020142675044,
"learning_rate": 4.9557338907165833e-05,
"loss": 0.2732,
"step": 510
},
{
"epoch": 0.2701298701298701,
"grad_norm": 0.2685575579055106,
"learning_rate": 4.9530681909864724e-05,
"loss": 0.2689,
"step": 520
},
{
"epoch": 0.2753246753246753,
"grad_norm": 0.23478660501034596,
"learning_rate": 4.950325316177409e-05,
"loss": 0.2726,
"step": 530
},
{
"epoch": 0.2805194805194805,
"grad_norm": 0.26267550311986976,
"learning_rate": 4.947505352581974e-05,
"loss": 0.2688,
"step": 540
},
{
"epoch": 0.2857142857142857,
"grad_norm": 0.2804831055299143,
"learning_rate": 4.944608388918005e-05,
"loss": 0.2724,
"step": 550
},
{
"epoch": 0.2909090909090909,
"grad_norm": 0.36597728037439853,
"learning_rate": 4.941634516325816e-05,
"loss": 0.2674,
"step": 560
},
{
"epoch": 0.2961038961038961,
"grad_norm": 0.27054673370836463,
"learning_rate": 4.9385838283653216e-05,
"loss": 0.2649,
"step": 570
},
{
"epoch": 0.3012987012987013,
"grad_norm": 0.3066834081735395,
"learning_rate": 4.9354564210130976e-05,
"loss": 0.2677,
"step": 580
},
{
"epoch": 0.3064935064935065,
"grad_norm": 0.2950401672059928,
"learning_rate": 4.93225239265936e-05,
"loss": 0.2622,
"step": 590
},
{
"epoch": 0.3116883116883117,
"grad_norm": 0.31772179112234966,
"learning_rate": 4.928971844104868e-05,
"loss": 0.2641,
"step": 600
},
{
"epoch": 0.3168831168831169,
"grad_norm": 0.27924556453889027,
"learning_rate": 4.9256148785577606e-05,
"loss": 0.2647,
"step": 610
},
{
"epoch": 0.3220779220779221,
"grad_norm": 0.28644663252200886,
"learning_rate": 4.9221816016302966e-05,
"loss": 0.2625,
"step": 620
},
{
"epoch": 0.32727272727272727,
"grad_norm": 0.2606246292323375,
"learning_rate": 4.9186721213355455e-05,
"loss": 0.2636,
"step": 630
},
{
"epoch": 0.33246753246753247,
"grad_norm": 0.3172496714001626,
"learning_rate": 4.915086548083978e-05,
"loss": 0.2683,
"step": 640
},
{
"epoch": 0.33766233766233766,
"grad_norm": 0.24927905040341644,
"learning_rate": 4.9114249946800003e-05,
"loss": 0.2654,
"step": 650
},
{
"epoch": 0.34285714285714286,
"grad_norm": 0.32250447729794757,
"learning_rate": 4.907687576318401e-05,
"loss": 0.2636,
"step": 660
},
{
"epoch": 0.34805194805194806,
"grad_norm": 0.32565974721616914,
"learning_rate": 4.903874410580731e-05,
"loss": 0.2601,
"step": 670
},
{
"epoch": 0.35324675324675325,
"grad_norm": 0.23517482221948124,
"learning_rate": 4.899985617431597e-05,
"loss": 0.2611,
"step": 680
},
{
"epoch": 0.35844155844155845,
"grad_norm": 0.2438523561922534,
"learning_rate": 4.896021319214895e-05,
"loss": 0.2601,
"step": 690
},
{
"epoch": 0.36363636363636365,
"grad_norm": 0.2668670929832916,
"learning_rate": 4.8919816406499584e-05,
"loss": 0.2696,
"step": 700
},
{
"epoch": 0.36883116883116884,
"grad_norm": 0.24937135360115686,
"learning_rate": 4.887866708827633e-05,
"loss": 0.2602,
"step": 710
},
{
"epoch": 0.37402597402597404,
"grad_norm": 0.23999629542679116,
"learning_rate": 4.8836766532062804e-05,
"loss": 0.2588,
"step": 720
},
{
"epoch": 0.37922077922077924,
"grad_norm": 0.23334349462758497,
"learning_rate": 4.879411605607704e-05,
"loss": 0.2606,
"step": 730
},
{
"epoch": 0.38441558441558443,
"grad_norm": 0.2215145938323352,
"learning_rate": 4.8750717002130024e-05,
"loss": 0.2567,
"step": 740
},
{
"epoch": 0.38961038961038963,
"grad_norm": 0.3218548674660387,
"learning_rate": 4.870657073558349e-05,
"loss": 0.2627,
"step": 750
},
{
"epoch": 0.3948051948051948,
"grad_norm": 0.29152150655446074,
"learning_rate": 4.866167864530693e-05,
"loss": 0.2561,
"step": 760
},
{
"epoch": 0.4,
"grad_norm": 0.29823420885887736,
"learning_rate": 4.8616042143633937e-05,
"loss": 0.2594,
"step": 770
},
{
"epoch": 0.4051948051948052,
"grad_norm": 0.256792339442467,
"learning_rate": 4.856966266631777e-05,
"loss": 0.2565,
"step": 780
},
{
"epoch": 0.4103896103896104,
"grad_norm": 0.25804885977573755,
"learning_rate": 4.8522541672486156e-05,
"loss": 0.2577,
"step": 790
},
{
"epoch": 0.4155844155844156,
"grad_norm": 0.2903609751193798,
"learning_rate": 4.84746806445954e-05,
"loss": 0.2534,
"step": 800
},
{
"epoch": 0.42077922077922075,
"grad_norm": 0.26400859462593973,
"learning_rate": 4.8426081088383756e-05,
"loss": 0.2586,
"step": 810
},
{
"epoch": 0.42597402597402595,
"grad_norm": 0.2960587838599708,
"learning_rate": 4.837674453282404e-05,
"loss": 0.261,
"step": 820
},
{
"epoch": 0.43116883116883115,
"grad_norm": 0.23331561495605277,
"learning_rate": 4.832667253007554e-05,
"loss": 0.2536,
"step": 830
},
{
"epoch": 0.43636363636363634,
"grad_norm": 0.24020811230350025,
"learning_rate": 4.8275866655435175e-05,
"loss": 0.2564,
"step": 840
},
{
"epoch": 0.44155844155844154,
"grad_norm": 0.22023402453548904,
"learning_rate": 4.8224328507287946e-05,
"loss": 0.2562,
"step": 850
},
{
"epoch": 0.44675324675324674,
"grad_norm": 0.293317498213313,
"learning_rate": 4.8172059707056626e-05,
"loss": 0.2565,
"step": 860
},
{
"epoch": 0.45194805194805193,
"grad_norm": 0.2692215437341758,
"learning_rate": 4.811906189915078e-05,
"loss": 0.2506,
"step": 870
},
{
"epoch": 0.45714285714285713,
"grad_norm": 0.23694698773474526,
"learning_rate": 4.806533675091501e-05,
"loss": 0.2518,
"step": 880
},
{
"epoch": 0.4623376623376623,
"grad_norm": 0.22885916720084376,
"learning_rate": 4.80108859525765e-05,
"loss": 0.252,
"step": 890
},
{
"epoch": 0.4675324675324675,
"grad_norm": 0.24916135308130166,
"learning_rate": 4.795571121719187e-05,
"loss": 0.253,
"step": 900
},
{
"epoch": 0.4727272727272727,
"grad_norm": 0.24894984701102493,
"learning_rate": 4.7899814280593226e-05,
"loss": 0.2529,
"step": 910
},
{
"epoch": 0.4779220779220779,
"grad_norm": 0.2723728137565129,
"learning_rate": 4.78431969013336e-05,
"loss": 0.2555,
"step": 920
},
{
"epoch": 0.4831168831168831,
"grad_norm": 0.3183305552275493,
"learning_rate": 4.778586086063159e-05,
"loss": 0.2514,
"step": 930
},
{
"epoch": 0.4883116883116883,
"grad_norm": 0.2414413013327865,
"learning_rate": 4.772780796231537e-05,
"loss": 0.2484,
"step": 940
},
{
"epoch": 0.4935064935064935,
"grad_norm": 0.22563943539011178,
"learning_rate": 4.766904003276589e-05,
"loss": 0.2539,
"step": 950
},
{
"epoch": 0.4987012987012987,
"grad_norm": 0.30205872999506944,
"learning_rate": 4.760955892085942e-05,
"loss": 0.2527,
"step": 960
},
{
"epoch": 0.5038961038961038,
"grad_norm": 0.2860786126383834,
"learning_rate": 4.754936649790942e-05,
"loss": 0.2516,
"step": 970
},
{
"epoch": 0.509090909090909,
"grad_norm": 0.27572406640999436,
"learning_rate": 4.7488464657607635e-05,
"loss": 0.2498,
"step": 980
},
{
"epoch": 0.5142857142857142,
"grad_norm": 0.24673507061013106,
"learning_rate": 4.7426855315964535e-05,
"loss": 0.2531,
"step": 990
},
{
"epoch": 0.5194805194805194,
"grad_norm": 0.21033978842271397,
"learning_rate": 4.736454041124904e-05,
"loss": 0.2504,
"step": 1000
},
{
"epoch": 0.5246753246753246,
"grad_norm": 0.24381830272002009,
"learning_rate": 4.7301521903927505e-05,
"loss": 0.2428,
"step": 1010
},
{
"epoch": 0.5298701298701298,
"grad_norm": 0.2334282560856222,
"learning_rate": 4.723780177660209e-05,
"loss": 0.2501,
"step": 1020
},
{
"epoch": 0.535064935064935,
"grad_norm": 0.2751808654527514,
"learning_rate": 4.717338203394836e-05,
"loss": 0.2507,
"step": 1030
},
{
"epoch": 0.5402597402597402,
"grad_norm": 0.2697637268340861,
"learning_rate": 4.71082647026522e-05,
"loss": 0.2503,
"step": 1040
},
{
"epoch": 0.5454545454545454,
"grad_norm": 0.22403492548323756,
"learning_rate": 4.7042451831346136e-05,
"loss": 0.2495,
"step": 1050
},
{
"epoch": 0.5506493506493506,
"grad_norm": 0.22425594786040917,
"learning_rate": 4.697594549054474e-05,
"loss": 0.2475,
"step": 1060
},
{
"epoch": 0.5558441558441558,
"grad_norm": 0.24265650046282458,
"learning_rate": 4.690874777257964e-05,
"loss": 0.2491,
"step": 1070
},
{
"epoch": 0.561038961038961,
"grad_norm": 0.22077426247794457,
"learning_rate": 4.684086079153359e-05,
"loss": 0.2449,
"step": 1080
},
{
"epoch": 0.5662337662337662,
"grad_norm": 0.30484192484636535,
"learning_rate": 4.6772286683174025e-05,
"loss": 0.245,
"step": 1090
},
{
"epoch": 0.5714285714285714,
"grad_norm": 0.22182045074526108,
"learning_rate": 4.670302760488582e-05,
"loss": 0.2477,
"step": 1100
},
{
"epoch": 0.5766233766233766,
"grad_norm": 0.2200234503444333,
"learning_rate": 4.663308573560343e-05,
"loss": 0.2506,
"step": 1110
},
{
"epoch": 0.5818181818181818,
"grad_norm": 0.21994287683589026,
"learning_rate": 4.656246327574238e-05,
"loss": 0.2421,
"step": 1120
},
{
"epoch": 0.587012987012987,
"grad_norm": 0.272334351757034,
"learning_rate": 4.649116244712998e-05,
"loss": 0.2476,
"step": 1130
},
{
"epoch": 0.5922077922077922,
"grad_norm": 0.2236806239622702,
"learning_rate": 4.641918549293545e-05,
"loss": 0.2454,
"step": 1140
},
{
"epoch": 0.5974025974025974,
"grad_norm": 0.23546182291174625,
"learning_rate": 4.634653467759936e-05,
"loss": 0.2477,
"step": 1150
},
{
"epoch": 0.6025974025974026,
"grad_norm": 0.2505748637769869,
"learning_rate": 4.6273212286762376e-05,
"loss": 0.2449,
"step": 1160
},
{
"epoch": 0.6077922077922078,
"grad_norm": 0.18468403137918,
"learning_rate": 4.619922062719335e-05,
"loss": 0.2432,
"step": 1170
},
{
"epoch": 0.612987012987013,
"grad_norm": 0.23235694779454488,
"learning_rate": 4.6124562026716766e-05,
"loss": 0.2457,
"step": 1180
},
{
"epoch": 0.6181818181818182,
"grad_norm": 0.2587545121058708,
"learning_rate": 4.604923883413946e-05,
"loss": 0.2467,
"step": 1190
},
{
"epoch": 0.6233766233766234,
"grad_norm": 0.23524218849591322,
"learning_rate": 4.59732534191768e-05,
"loss": 0.2425,
"step": 1200
},
{
"epoch": 0.6285714285714286,
"grad_norm": 0.20482111091171828,
"learning_rate": 4.589660817237805e-05,
"loss": 0.2446,
"step": 1210
},
{
"epoch": 0.6337662337662338,
"grad_norm": 0.23935891812153473,
"learning_rate": 4.581930550505122e-05,
"loss": 0.2359,
"step": 1220
},
{
"epoch": 0.638961038961039,
"grad_norm": 0.2088944848253035,
"learning_rate": 4.5741347849187186e-05,
"loss": 0.2435,
"step": 1230
},
{
"epoch": 0.6441558441558441,
"grad_norm": 0.19769218239953817,
"learning_rate": 4.566273765738318e-05,
"loss": 0.2429,
"step": 1240
},
{
"epoch": 0.6493506493506493,
"grad_norm": 0.21104513975336958,
"learning_rate": 4.558347740276562e-05,
"loss": 0.2414,
"step": 1250
},
{
"epoch": 0.6545454545454545,
"grad_norm": 0.22299730594468617,
"learning_rate": 4.550356957891232e-05,
"loss": 0.2405,
"step": 1260
},
{
"epoch": 0.6597402597402597,
"grad_norm": 0.2553717349606562,
"learning_rate": 4.5423016699774025e-05,
"loss": 0.242,
"step": 1270
},
{
"epoch": 0.6649350649350649,
"grad_norm": 0.184621184174687,
"learning_rate": 4.5341821299595334e-05,
"loss": 0.2377,
"step": 1280
},
{
"epoch": 0.6701298701298701,
"grad_norm": 0.1931019421949112,
"learning_rate": 4.525998593283496e-05,
"loss": 0.2401,
"step": 1290
},
{
"epoch": 0.6753246753246753,
"grad_norm": 0.2327800605150306,
"learning_rate": 4.517751317408537e-05,
"loss": 0.2405,
"step": 1300
},
{
"epoch": 0.6805194805194805,
"grad_norm": 0.21060753158636902,
"learning_rate": 4.5094405617991796e-05,
"loss": 0.2363,
"step": 1310
},
{
"epoch": 0.6857142857142857,
"grad_norm": 0.18780386461582757,
"learning_rate": 4.501066587917058e-05,
"loss": 0.2437,
"step": 1320
},
{
"epoch": 0.6909090909090909,
"grad_norm": 0.2671736773226871,
"learning_rate": 4.4926296592126946e-05,
"loss": 0.2431,
"step": 1330
},
{
"epoch": 0.6961038961038961,
"grad_norm": 0.23810395896846592,
"learning_rate": 4.484130041117211e-05,
"loss": 0.243,
"step": 1340
},
{
"epoch": 0.7012987012987013,
"grad_norm": 0.18025186403432847,
"learning_rate": 4.475568001033974e-05,
"loss": 0.2457,
"step": 1350
},
{
"epoch": 0.7064935064935065,
"grad_norm": 0.21115305940327297,
"learning_rate": 4.466943808330189e-05,
"loss": 0.2415,
"step": 1360
},
{
"epoch": 0.7116883116883117,
"grad_norm": 0.25256979094205834,
"learning_rate": 4.45825773432842e-05,
"loss": 0.2407,
"step": 1370
},
{
"epoch": 0.7168831168831169,
"grad_norm": 0.22014008453128092,
"learning_rate": 4.449510052298056e-05,
"loss": 0.2357,
"step": 1380
},
{
"epoch": 0.7220779220779221,
"grad_norm": 0.20062628753000003,
"learning_rate": 4.440701037446714e-05,
"loss": 0.2396,
"step": 1390
},
{
"epoch": 0.7272727272727273,
"grad_norm": 0.2083485579826855,
"learning_rate": 4.431830966911582e-05,
"loss": 0.2391,
"step": 1400
},
{
"epoch": 0.7324675324675325,
"grad_norm": 0.18184305857175756,
"learning_rate": 4.422900119750695e-05,
"loss": 0.2355,
"step": 1410
},
{
"epoch": 0.7376623376623377,
"grad_norm": 0.2048232593631597,
"learning_rate": 4.4139087769341625e-05,
"loss": 0.2332,
"step": 1420
},
{
"epoch": 0.7428571428571429,
"grad_norm": 0.21132211989073768,
"learning_rate": 4.4048572213353234e-05,
"loss": 0.2422,
"step": 1430
},
{
"epoch": 0.7480519480519481,
"grad_norm": 0.19792895604344352,
"learning_rate": 4.39574573772185e-05,
"loss": 0.2334,
"step": 1440
},
{
"epoch": 0.7532467532467533,
"grad_norm": 0.22178836320367148,
"learning_rate": 4.3865746127467876e-05,
"loss": 0.2423,
"step": 1450
},
{
"epoch": 0.7584415584415585,
"grad_norm": 0.21967650568135474,
"learning_rate": 4.3773441349395374e-05,
"loss": 0.2357,
"step": 1460
},
{
"epoch": 0.7636363636363637,
"grad_norm": 0.1917556477695145,
"learning_rate": 4.368054594696775e-05,
"loss": 0.2443,
"step": 1470
},
{
"epoch": 0.7688311688311689,
"grad_norm": 0.20969861600848638,
"learning_rate": 4.3587062842733216e-05,
"loss": 0.2341,
"step": 1480
},
{
"epoch": 0.7740259740259741,
"grad_norm": 0.19871375239851857,
"learning_rate": 4.349299497772945e-05,
"loss": 0.2361,
"step": 1490
},
{
"epoch": 0.7792207792207793,
"grad_norm": 0.19082750332598916,
"learning_rate": 4.339834531139104e-05,
"loss": 0.2316,
"step": 1500
},
{
"epoch": 0.7844155844155845,
"grad_norm": 0.2177029161255871,
"learning_rate": 4.330311682145645e-05,
"loss": 0.2343,
"step": 1510
},
{
"epoch": 0.7896103896103897,
"grad_norm": 0.20562958726540304,
"learning_rate": 4.320731250387429e-05,
"loss": 0.2401,
"step": 1520
},
{
"epoch": 0.7948051948051948,
"grad_norm": 0.2070237852219627,
"learning_rate": 4.311093537270905e-05,
"loss": 0.2374,
"step": 1530
},
{
"epoch": 0.8,
"grad_norm": 0.18967395002327114,
"learning_rate": 4.301398846004634e-05,
"loss": 0.2363,
"step": 1540
},
{
"epoch": 0.8051948051948052,
"grad_norm": 0.1970271386066234,
"learning_rate": 4.291647481589742e-05,
"loss": 0.2302,
"step": 1550
},
{
"epoch": 0.8103896103896104,
"grad_norm": 0.18673676151020974,
"learning_rate": 4.28183975081033e-05,
"loss": 0.2416,
"step": 1560
},
{
"epoch": 0.8155844155844156,
"grad_norm": 0.2183111540642943,
"learning_rate": 4.271975962223821e-05,
"loss": 0.2342,
"step": 1570
},
{
"epoch": 0.8207792207792208,
"grad_norm": 0.1792298886397136,
"learning_rate": 4.2620564261512496e-05,
"loss": 0.2388,
"step": 1580
},
{
"epoch": 0.825974025974026,
"grad_norm": 0.21429193275126804,
"learning_rate": 4.2520814546675037e-05,
"loss": 0.2323,
"step": 1590
},
{
"epoch": 0.8311688311688312,
"grad_norm": 0.1923357673969473,
"learning_rate": 4.242051361591505e-05,
"loss": 0.2398,
"step": 1600
},
{
"epoch": 0.8363636363636363,
"grad_norm": 0.1825902322292911,
"learning_rate": 4.2319664624763325e-05,
"loss": 0.2355,
"step": 1610
},
{
"epoch": 0.8415584415584415,
"grad_norm": 0.1708452665847616,
"learning_rate": 4.2218270745993016e-05,
"loss": 0.2361,
"step": 1620
},
{
"epoch": 0.8467532467532467,
"grad_norm": 0.2003983431936864,
"learning_rate": 4.211633516951975e-05,
"loss": 0.237,
"step": 1630
},
{
"epoch": 0.8519480519480519,
"grad_norm": 0.1809948763155965,
"learning_rate": 4.201386110230134e-05,
"loss": 0.2291,
"step": 1640
},
{
"epoch": 0.8571428571428571,
"grad_norm": 0.19621979591943875,
"learning_rate": 4.1910851768236825e-05,
"loss": 0.2284,
"step": 1650
},
{
"epoch": 0.8623376623376623,
"grad_norm": 0.20784597945629102,
"learning_rate": 4.180731040806511e-05,
"loss": 0.2359,
"step": 1660
},
{
"epoch": 0.8675324675324675,
"grad_norm": 0.22581034014160772,
"learning_rate": 4.170324027926297e-05,
"loss": 0.2329,
"step": 1670
},
{
"epoch": 0.8727272727272727,
"grad_norm": 0.1789163119753752,
"learning_rate": 4.159864465594255e-05,
"loss": 0.2338,
"step": 1680
},
{
"epoch": 0.8779220779220779,
"grad_norm": 0.1949206924337472,
"learning_rate": 4.1493526828748416e-05,
"loss": 0.2392,
"step": 1690
},
{
"epoch": 0.8831168831168831,
"grad_norm": 0.20147429000086556,
"learning_rate": 4.1387890104754004e-05,
"loss": 0.233,
"step": 1700
},
{
"epoch": 0.8883116883116883,
"grad_norm": 0.1537005161376695,
"learning_rate": 4.128173780735753e-05,
"loss": 0.2291,
"step": 1710
},
{
"epoch": 0.8935064935064935,
"grad_norm": 0.17777763693741433,
"learning_rate": 4.117507327617751e-05,
"loss": 0.2291,
"step": 1720
},
{
"epoch": 0.8987012987012987,
"grad_norm": 0.174198062693491,
"learning_rate": 4.1067899866947665e-05,
"loss": 0.2294,
"step": 1730
},
{
"epoch": 0.9038961038961039,
"grad_norm": 0.1884364748511166,
"learning_rate": 4.096022095141132e-05,
"loss": 0.235,
"step": 1740
},
{
"epoch": 0.9090909090909091,
"grad_norm": 0.1912652069094164,
"learning_rate": 4.085203991721535e-05,
"loss": 0.2318,
"step": 1750
},
{
"epoch": 0.9142857142857143,
"grad_norm": 0.206558739242339,
"learning_rate": 4.0743360167803614e-05,
"loss": 0.2317,
"step": 1760
},
{
"epoch": 0.9194805194805195,
"grad_norm": 0.18252166114267931,
"learning_rate": 4.063418512230987e-05,
"loss": 0.2346,
"step": 1770
},
{
"epoch": 0.9246753246753247,
"grad_norm": 0.18463778266166328,
"learning_rate": 4.0524518215450166e-05,
"loss": 0.2306,
"step": 1780
},
{
"epoch": 0.9298701298701298,
"grad_norm": 0.207923278938462,
"learning_rate": 4.041436289741489e-05,
"loss": 0.2301,
"step": 1790
},
{
"epoch": 0.935064935064935,
"grad_norm": 0.25335695776490813,
"learning_rate": 4.0303722633760085e-05,
"loss": 0.2258,
"step": 1800
},
{
"epoch": 0.9402597402597402,
"grad_norm": 0.16805426564943104,
"learning_rate": 4.019260090529854e-05,
"loss": 0.2284,
"step": 1810
},
{
"epoch": 0.9454545454545454,
"grad_norm": 0.1884837989936669,
"learning_rate": 4.008100120799019e-05,
"loss": 0.2285,
"step": 1820
},
{
"epoch": 0.9506493506493506,
"grad_norm": 0.19643081968195814,
"learning_rate": 3.996892705283222e-05,
"loss": 0.2354,
"step": 1830
},
{
"epoch": 0.9558441558441558,
"grad_norm": 0.18949399067442121,
"learning_rate": 3.9856381965748506e-05,
"loss": 0.234,
"step": 1840
},
{
"epoch": 0.961038961038961,
"grad_norm": 0.18422150470243814,
"learning_rate": 3.974336948747879e-05,
"loss": 0.228,
"step": 1850
},
{
"epoch": 0.9662337662337662,
"grad_norm": 0.179837212988977,
"learning_rate": 3.962989317346722e-05,
"loss": 0.2283,
"step": 1860
},
{
"epoch": 0.9714285714285714,
"grad_norm": 0.18190735457463206,
"learning_rate": 3.951595659375048e-05,
"loss": 0.2337,
"step": 1870
},
{
"epoch": 0.9766233766233766,
"grad_norm": 0.17903757429753223,
"learning_rate": 3.9401563332845545e-05,
"loss": 0.2225,
"step": 1880
},
{
"epoch": 0.9818181818181818,
"grad_norm": 0.18775008592274955,
"learning_rate": 3.928671698963686e-05,
"loss": 0.226,
"step": 1890
},
{
"epoch": 0.987012987012987,
"grad_norm": 0.16101556678112905,
"learning_rate": 3.917142117726312e-05,
"loss": 0.2312,
"step": 1900
},
{
"epoch": 0.9922077922077922,
"grad_norm": 0.18133929144854413,
"learning_rate": 3.90556795230036e-05,
"loss": 0.2264,
"step": 1910
},
{
"epoch": 0.9974025974025974,
"grad_norm": 0.21238218367672568,
"learning_rate": 3.893949566816404e-05,
"loss": 0.2273,
"step": 1920
}
],
"logging_steps": 10,
"max_steps": 5775,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 6056617202155520.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}