Nemo-Recent2 / checkpoint-600 /trainer_state.json
zaddyzaddy's picture
Upload folder using huggingface_hub
c55123d verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.8391608391608392,
"eval_steps": 500,
"global_step": 600,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"clip_ratio": 0.0,
"completion_length": 94.14286041259766,
"epoch": 0.0013986013986013986,
"grad_norm": 0.09098726750251449,
"kl": 0.0003147125244140625,
"learning_rate": 3.496503496503497e-07,
"loss": -0.0001,
"num_tokens": 19338.0,
"reward": 1.3853150606155396,
"reward_std": 0.4827312231063843,
"rewards/check_gptzero_func": 0.3214285671710968,
"rewards/check_perplexity_diff_func": 0.6428571343421936,
"rewards/check_winston_local_func": 0.42102929949760437,
"step": 1
},
{
"clip_ratio": 0.0,
"epoch": 0.002797202797202797,
"grad_norm": 0.09100260462019598,
"kl": 0.0003147125244140625,
"learning_rate": 6.993006993006994e-07,
"loss": -0.0001,
"step": 2
},
{
"clip_ratio": 0.0031990089919418097,
"completion_length": 97.73214721679688,
"epoch": 0.004195804195804196,
"grad_norm": 0.16978190242842756,
"kl": 0.0003662109375,
"learning_rate": 1.0489510489510491e-06,
"loss": -0.0017,
"num_tokens": 39487.0,
"reward": 0.9710169434547424,
"reward_std": 0.6376833319664001,
"rewards/check_gptzero_func": 0.125,
"rewards/check_perplexity_diff_func": 0.5,
"rewards/check_winston_local_func": 0.34601688385009766,
"step": 3
},
{
"clip_ratio": 0.0018373647471889853,
"epoch": 0.005594405594405594,
"grad_norm": 0.17663660241053314,
"kl": 0.000423431396484375,
"learning_rate": 1.3986013986013987e-06,
"loss": -0.0013,
"step": 4
},
{
"clip_ratio": 0.0016871786210685968,
"completion_length": 72.83928680419922,
"epoch": 0.006993006993006993,
"grad_norm": 0.09466520543424979,
"kl": 0.00034332275390625,
"learning_rate": 1.7482517482517485e-06,
"loss": 0.0008,
"num_tokens": 56082.0,
"reward": 1.0736479759216309,
"reward_std": 0.6862176656723022,
"rewards/check_gptzero_func": 0.2321428507566452,
"rewards/check_perplexity_diff_func": 0.3928571343421936,
"rewards/check_winston_local_func": 0.4486479163169861,
"step": 5
},
{
"clip_ratio": 0.0010098000057041645,
"epoch": 0.008391608391608392,
"grad_norm": 0.09539550614656178,
"kl": 0.0003376007080078125,
"learning_rate": 2.0979020979020983e-06,
"loss": 0.001,
"step": 6
},
{
"clip_ratio": 0.00184189947322011,
"completion_length": 95.92857360839844,
"epoch": 0.009790209790209791,
"grad_norm": 0.10573631695411624,
"kl": 0.0003757476806640625,
"learning_rate": 2.4475524475524477e-06,
"loss": 0.001,
"num_tokens": 75906.0,
"reward": 1.135968804359436,
"reward_std": 0.48974505066871643,
"rewards/check_gptzero_func": 0.2142857164144516,
"rewards/check_perplexity_diff_func": 0.5,
"rewards/check_winston_local_func": 0.42168304324150085,
"step": 7
},
{
"clip_ratio": 0.0018530809320509434,
"epoch": 0.011188811188811189,
"grad_norm": 0.10614532740192534,
"kl": 0.0004119873046875,
"learning_rate": 2.7972027972027974e-06,
"loss": 0.0011,
"step": 8
},
{
"clip_ratio": 0.0017129909247159958,
"completion_length": 106.76786041259766,
"epoch": 0.012587412587412588,
"grad_norm": 0.07916450383171113,
"kl": 0.000377655029296875,
"learning_rate": 3.1468531468531472e-06,
"loss": -0.0012,
"num_tokens": 96989.0,
"reward": 0.9590713977813721,
"reward_std": 0.5206417441368103,
"rewards/check_gptzero_func": 0.2321428507566452,
"rewards/check_perplexity_diff_func": 0.3928571343421936,
"rewards/check_winston_local_func": 0.3340713679790497,
"step": 9
},
{
"clip_ratio": 0.0034840807784348726,
"epoch": 0.013986013986013986,
"grad_norm": 0.08061029037170311,
"kl": 0.00037384033203125,
"learning_rate": 3.496503496503497e-06,
"loss": -0.0008,
"step": 10
},
{
"clip_ratio": 0.0017933619674295187,
"completion_length": 102.85714721679688,
"epoch": 0.015384615384615385,
"grad_norm": 0.07930375667041344,
"kl": 0.00030517578125,
"learning_rate": 3.846153846153847e-06,
"loss": -0.0038,
"num_tokens": 117271.0,
"reward": 1.1903148889541626,
"reward_std": 0.4722224473953247,
"rewards/check_gptzero_func": 0.2857142984867096,
"rewards/check_perplexity_diff_func": 0.5357142686843872,
"rewards/check_winston_local_func": 0.368886262178421,
"step": 11
},
{
"clip_ratio": 0.0015239757485687733,
"epoch": 0.016783216783216783,
"grad_norm": 0.07898151410343454,
"kl": 0.0003871917724609375,
"learning_rate": 4.195804195804197e-06,
"loss": -0.0041,
"step": 12
},
{
"clip_ratio": 0.0029670600779354572,
"completion_length": 104.37500762939453,
"epoch": 0.01818181818181818,
"grad_norm": 0.09247854491588844,
"kl": 0.0003986358642578125,
"learning_rate": 4.5454545454545455e-06,
"loss": -0.0049,
"num_tokens": 138174.0,
"reward": 0.9787324070930481,
"reward_std": 0.63657546043396,
"rewards/check_gptzero_func": 0.1964285671710968,
"rewards/check_perplexity_diff_func": 0.5714285969734192,
"rewards/check_winston_local_func": 0.21087531745433807,
"step": 13
},
{
"clip_ratio": 0.003217793768271804,
"epoch": 0.019580419580419582,
"grad_norm": 0.09293749933034297,
"kl": 0.0003204345703125,
"learning_rate": 4.895104895104895e-06,
"loss": -0.0045,
"step": 14
},
{
"clip_ratio": 0.002677519340068102,
"completion_length": 67.41072082519531,
"epoch": 0.02097902097902098,
"grad_norm": 0.14020673218962584,
"kl": 0.000499725341796875,
"learning_rate": 5.244755244755245e-06,
"loss": -0.008,
"num_tokens": 154385.0,
"reward": 1.0573337078094482,
"reward_std": 0.5388314127922058,
"rewards/check_gptzero_func": 0.2857142984867096,
"rewards/check_perplexity_diff_func": 0.3571428656578064,
"rewards/check_winston_local_func": 0.41447654366493225,
"step": 15
},
{
"clip_ratio": 0.0021408216562122107,
"epoch": 0.022377622377622378,
"grad_norm": 0.16110555979669544,
"kl": 0.000530242919921875,
"learning_rate": 5.594405594405595e-06,
"loss": -0.0081,
"step": 16
},
{
"clip_ratio": 0.002852825215086341,
"completion_length": 89.83928680419922,
"epoch": 0.023776223776223775,
"grad_norm": 0.1066932843002634,
"kl": 0.000415802001953125,
"learning_rate": 5.944055944055945e-06,
"loss": 0.0166,
"num_tokens": 173158.0,
"reward": 1.2226747274398804,
"reward_std": 0.45578521490097046,
"rewards/check_gptzero_func": 0.3392857015132904,
"rewards/check_perplexity_diff_func": 0.5,
"rewards/check_winston_local_func": 0.3833889663219452,
"step": 17
},
{
"clip_ratio": 0.002018069615587592,
"epoch": 0.025174825174825177,
"grad_norm": 0.10578254537451186,
"kl": 0.0003833770751953125,
"learning_rate": 6.2937062937062944e-06,
"loss": 0.0164,
"step": 18
},
{
"clip_ratio": 0.0032663617748767138,
"completion_length": 105.83928680419922,
"epoch": 0.026573426573426574,
"grad_norm": 0.11554707163597037,
"kl": 0.0003719329833984375,
"learning_rate": 6.643356643356643e-06,
"loss": -0.0005,
"num_tokens": 194649.0,
"reward": 0.8957912921905518,
"reward_std": 0.47702187299728394,
"rewards/check_gptzero_func": 0.1785714328289032,
"rewards/check_perplexity_diff_func": 0.3571428656578064,
"rewards/check_winston_local_func": 0.3600769340991974,
"step": 19
},
{
"clip_ratio": 0.0024422537535429,
"epoch": 0.027972027972027972,
"grad_norm": 0.11417666172406592,
"kl": 0.0003814697265625,
"learning_rate": 6.993006993006994e-06,
"loss": -0.0008,
"step": 20
},
{
"clip_ratio": 0.0013639701064676046,
"completion_length": 107.39286041259766,
"epoch": 0.02937062937062937,
"grad_norm": 0.09634196153318929,
"kl": 0.000396728515625,
"learning_rate": 7.342657342657343e-06,
"loss": -0.0011,
"num_tokens": 215987.0,
"reward": 0.9197577238082886,
"reward_std": 0.39100831747055054,
"rewards/check_gptzero_func": 0.2321428507566452,
"rewards/check_perplexity_diff_func": 0.3571428656578064,
"rewards/check_winston_local_func": 0.33047202229499817,
"step": 21
},
{
"clip_ratio": 0.003595889313146472,
"epoch": 0.03076923076923077,
"grad_norm": 0.09779857975187953,
"kl": 0.00052642822265625,
"learning_rate": 7.692307692307694e-06,
"loss": -0.0013,
"step": 22
},
{
"clip_ratio": 0.0024316231720149517,
"completion_length": 89.08928680419922,
"epoch": 0.032167832167832165,
"grad_norm": 0.13189477135999747,
"kl": 0.0003414154052734375,
"learning_rate": 8.041958041958042e-06,
"loss": -0.0009,
"num_tokens": 234678.0,
"reward": 0.9593304395675659,
"reward_std": 0.5972030758857727,
"rewards/check_gptzero_func": 0.1964285671710968,
"rewards/check_perplexity_diff_func": 0.4285714328289032,
"rewards/check_winston_local_func": 0.33433040976524353,
"step": 23
},
{
"clip_ratio": 0.000728283659555018,
"epoch": 0.033566433566433566,
"grad_norm": 0.12953291972672934,
"kl": 0.000335693359375,
"learning_rate": 8.391608391608393e-06,
"loss": -0.0011,
"step": 24
},
{
"clip_ratio": 0.0021546650677919388,
"completion_length": 90.98214721679688,
"epoch": 0.03496503496503497,
"grad_norm": 0.07682474826469807,
"kl": 0.0004558563232421875,
"learning_rate": 8.741258741258741e-06,
"loss": 0.0077,
"num_tokens": 253833.0,
"reward": 0.8255766034126282,
"reward_std": 0.3988816440105438,
"rewards/check_gptzero_func": 0.2678571343421936,
"rewards/check_perplexity_diff_func": 0.2857142984867096,
"rewards/check_winston_local_func": 0.272005170583725,
"step": 25
},
{
"clip_ratio": 0.0019250252516940236,
"epoch": 0.03636363636363636,
"grad_norm": 0.07873511079065725,
"kl": 0.0003490447998046875,
"learning_rate": 9.090909090909091e-06,
"loss": 0.0075,
"step": 26
},
{
"clip_ratio": 0.001886485842987895,
"completion_length": 74.4464340209961,
"epoch": 0.03776223776223776,
"grad_norm": 0.06765277577538054,
"kl": 0.000354766845703125,
"learning_rate": 9.44055944055944e-06,
"loss": -0.0003,
"num_tokens": 270908.0,
"reward": 1.3086735010147095,
"reward_std": 0.3881511986255646,
"rewards/check_gptzero_func": 0.4107142984867096,
"rewards/check_perplexity_diff_func": 0.4642857015132904,
"rewards/check_winston_local_func": 0.4336733818054199,
"step": 27
},
{
"clip_ratio": 0.0020898371003568172,
"epoch": 0.039160839160839164,
"grad_norm": 0.06722581226675682,
"kl": 0.000453948974609375,
"learning_rate": 9.79020979020979e-06,
"loss": -0.0006,
"step": 28
},
{
"clip_ratio": 0.002855924190953374,
"completion_length": 103.64286041259766,
"epoch": 0.04055944055944056,
"grad_norm": 0.16301390061045865,
"kl": 0.0003948211669921875,
"learning_rate": 1.013986013986014e-05,
"loss": 0.0063,
"num_tokens": 292060.0,
"reward": 1.22488272190094,
"reward_std": 0.8093323111534119,
"rewards/check_gptzero_func": 0.2678571343421936,
"rewards/check_perplexity_diff_func": 0.6071428656578064,
"rewards/check_winston_local_func": 0.34988275170326233,
"step": 29
},
{
"clip_ratio": 0.0016208746237680316,
"epoch": 0.04195804195804196,
"grad_norm": 0.17335125351231856,
"kl": 0.000507354736328125,
"learning_rate": 1.048951048951049e-05,
"loss": 0.0059,
"step": 30
},
{
"clip_ratio": 0.0011572305811569095,
"completion_length": 82.91072082519531,
"epoch": 0.043356643356643354,
"grad_norm": 0.11830097508590764,
"kl": 0.0006256103515625,
"learning_rate": 1.0839160839160838e-05,
"loss": 0.002,
"num_tokens": 310163.0,
"reward": 0.9069231748580933,
"reward_std": 0.4621748924255371,
"rewards/check_gptzero_func": 0.25,
"rewards/check_perplexity_diff_func": 0.3571428656578064,
"rewards/check_winston_local_func": 0.29978030920028687,
"step": 31
},
{
"clip_ratio": 0.00402362085878849,
"epoch": 0.044755244755244755,
"grad_norm": 0.11861334422671065,
"kl": 0.000762939453125,
"learning_rate": 1.118881118881119e-05,
"loss": 0.0019,
"step": 32
},
{
"clip_ratio": 0.0017474278574809432,
"completion_length": 90.9464340209961,
"epoch": 0.046153846153846156,
"grad_norm": 0.1025496804751534,
"kl": 0.000701904296875,
"learning_rate": 1.153846153846154e-05,
"loss": -0.0014,
"num_tokens": 328922.0,
"reward": 1.1198338270187378,
"reward_std": 0.5045111775398254,
"rewards/check_gptzero_func": 0.0892857164144516,
"rewards/check_perplexity_diff_func": 0.7857142686843872,
"rewards/check_winston_local_func": 0.24483375251293182,
"step": 33
},
{
"clip_ratio": 0.0012307984288781881,
"epoch": 0.04755244755244755,
"grad_norm": 0.10274008519956918,
"kl": 0.00087738037109375,
"learning_rate": 1.188811188811189e-05,
"loss": -0.0015,
"step": 34
},
{
"clip_ratio": 0.0029886537231504917,
"completion_length": 93.00000762939453,
"epoch": 0.04895104895104895,
"grad_norm": 0.1311802256012859,
"kl": 0.001190185546875,
"learning_rate": 1.2237762237762239e-05,
"loss": -0.0141,
"num_tokens": 348852.0,
"reward": 0.9456299543380737,
"reward_std": 0.6374980807304382,
"rewards/check_gptzero_func": 0.2321428507566452,
"rewards/check_perplexity_diff_func": 0.4285714328289032,
"rewards/check_winston_local_func": 0.28491565585136414,
"step": 35
},
{
"clip_ratio": 0.0028984802775084972,
"epoch": 0.05034965034965035,
"grad_norm": 0.13267305607921692,
"kl": 0.00145721435546875,
"learning_rate": 1.2587412587412589e-05,
"loss": -0.0147,
"step": 36
},
{
"clip_ratio": 0.0018746532732620835,
"completion_length": 89.85714721679688,
"epoch": 0.05174825174825175,
"grad_norm": 0.1497182815907709,
"kl": 0.002044677734375,
"learning_rate": 1.2937062937062939e-05,
"loss": -0.0253,
"num_tokens": 367482.0,
"reward": 1.2605940103530884,
"reward_std": 0.66156405210495,
"rewards/check_gptzero_func": 0.1785714328289032,
"rewards/check_perplexity_diff_func": 0.75,
"rewards/check_winston_local_func": 0.3320225179195404,
"step": 37
},
{
"clip_ratio": 0.0020914783235639334,
"epoch": 0.05314685314685315,
"grad_norm": 0.15027543310666272,
"kl": 0.0026092529296875,
"learning_rate": 1.3286713286713287e-05,
"loss": -0.0257,
"step": 38
},
{
"clip_ratio": 0.0017028081929311156,
"completion_length": 101.58928680419922,
"epoch": 0.05454545454545454,
"grad_norm": 0.1306748742984494,
"kl": 0.0027008056640625,
"learning_rate": 1.3636363636363637e-05,
"loss": -0.0091,
"num_tokens": 388365.0,
"reward": 1.1187876462936401,
"reward_std": 0.4044211804866791,
"rewards/check_gptzero_func": 0.2321428507566452,
"rewards/check_perplexity_diff_func": 0.5357142686843872,
"rewards/check_winston_local_func": 0.3509305417537689,
"step": 39
},
{
"clip_ratio": 0.003386714030057192,
"epoch": 0.055944055944055944,
"grad_norm": 0.12880289464580882,
"kl": 0.003265380859375,
"learning_rate": 1.3986013986013988e-05,
"loss": -0.0086,
"step": 40
},
{
"clip_ratio": 0.0024523374158889055,
"completion_length": 100.41072082519531,
"epoch": 0.057342657342657345,
"grad_norm": 0.12385816200639352,
"kl": 0.004547119140625,
"learning_rate": 1.4335664335664336e-05,
"loss": -0.0161,
"num_tokens": 409074.0,
"reward": 0.804709792137146,
"reward_std": 0.5129754543304443,
"rewards/check_gptzero_func": 0.1428571492433548,
"rewards/check_perplexity_diff_func": 0.3928571343421936,
"rewards/check_winston_local_func": 0.2689954936504364,
"step": 41
},
{
"clip_ratio": 0.004267544951289892,
"epoch": 0.05874125874125874,
"grad_norm": 0.12301160152684054,
"kl": 0.005218505859375,
"learning_rate": 1.4685314685314686e-05,
"loss": -0.0161,
"step": 42
},
{
"clip_ratio": 0.0019341636216267943,
"completion_length": 90.46428680419922,
"epoch": 0.06013986013986014,
"grad_norm": 0.1390908713431379,
"kl": 0.0048828125,
"learning_rate": 1.5034965034965034e-05,
"loss": 0.0068,
"num_tokens": 428262.0,
"reward": 1.087609887123108,
"reward_std": 0.6448312997817993,
"rewards/check_gptzero_func": 0.2321428507566452,
"rewards/check_perplexity_diff_func": 0.3928571343421936,
"rewards/check_winston_local_func": 0.46260982751846313,
"step": 43
},
{
"clip_ratio": 0.0016832423862069845,
"epoch": 0.06153846153846154,
"grad_norm": 0.13873080950458352,
"kl": 0.00567626953125,
"learning_rate": 1.5384615384615387e-05,
"loss": 0.0065,
"step": 44
},
{
"clip_ratio": 0.0019954824820160866,
"completion_length": 96.92857360839844,
"epoch": 0.06293706293706294,
"grad_norm": 0.11630124487143968,
"kl": 0.00897216796875,
"learning_rate": 1.5734265734265734e-05,
"loss": 0.008,
"num_tokens": 448480.0,
"reward": 1.0604994297027588,
"reward_std": 0.5378546118736267,
"rewards/check_gptzero_func": 0.2321428507566452,
"rewards/check_perplexity_diff_func": 0.4642857015132904,
"rewards/check_winston_local_func": 0.36407074332237244,
"step": 45
},
{
"clip_ratio": 0.002663462422788143,
"epoch": 0.06433566433566433,
"grad_norm": 0.1149796219701467,
"kl": 0.01007080078125,
"learning_rate": 1.6083916083916083e-05,
"loss": 0.0081,
"step": 46
},
{
"clip_ratio": 0.0018039485439658165,
"completion_length": 79.51786041259766,
"epoch": 0.06573426573426573,
"grad_norm": 0.12721969778213826,
"kl": 0.01123046875,
"learning_rate": 1.6433566433566433e-05,
"loss": 0.0112,
"num_tokens": 465671.0,
"reward": 0.9921315312385559,
"reward_std": 0.38579052686691284,
"rewards/check_gptzero_func": 0.1607142835855484,
"rewards/check_perplexity_diff_func": 0.3928571343421936,
"rewards/check_winston_local_func": 0.43856000900268555,
"step": 47
},
{
"clip_ratio": 0.0027954555116593838,
"epoch": 0.06713286713286713,
"grad_norm": 0.12930272448754576,
"kl": 0.01300048828125,
"learning_rate": 1.6783216783216786e-05,
"loss": 0.0107,
"step": 48
},
{
"clip_ratio": 0.0028269642498344183,
"completion_length": 116.5714340209961,
"epoch": 0.06853146853146853,
"grad_norm": 0.12430052891449103,
"kl": 0.01226806640625,
"learning_rate": 1.7132867132867133e-05,
"loss": 0.025,
"num_tokens": 488367.0,
"reward": 0.8830849528312683,
"reward_std": 0.5276607871055603,
"rewards/check_gptzero_func": 0.25,
"rewards/check_perplexity_diff_func": 0.3214285671710968,
"rewards/check_winston_local_func": 0.31165632605552673,
"step": 49
},
{
"clip_ratio": 0.0023628019262105227,
"epoch": 0.06993006993006994,
"grad_norm": 0.12293590613717971,
"kl": 0.0142822265625,
"learning_rate": 1.7482517482517483e-05,
"loss": 0.0249,
"step": 50
},
{
"clip_ratio": 0.0033425339497625828,
"completion_length": 83.39286041259766,
"epoch": 0.07132867132867132,
"grad_norm": 0.11106847343390151,
"kl": 0.01177978515625,
"learning_rate": 1.7832167832167836e-05,
"loss": -0.0041,
"num_tokens": 506551.0,
"reward": 1.029296875,
"reward_std": 0.4901208281517029,
"rewards/check_gptzero_func": 0.2857142984867096,
"rewards/check_perplexity_diff_func": 0.3214285671710968,
"rewards/check_winston_local_func": 0.42215394973754883,
"step": 51
},
{
"clip_ratio": 0.0022109400015324354,
"epoch": 0.07272727272727272,
"grad_norm": 0.11017795132415774,
"kl": 0.013671875,
"learning_rate": 1.8181818181818182e-05,
"loss": -0.0043,
"step": 52
},
{
"clip_ratio": 0.0019256824161857367,
"completion_length": 86.5714340209961,
"epoch": 0.07412587412587412,
"grad_norm": 0.11010712395827207,
"kl": 0.0208740234375,
"learning_rate": 1.8531468531468532e-05,
"loss": -0.0032,
"num_tokens": 524863.0,
"reward": 0.962996244430542,
"reward_std": 0.6607940196990967,
"rewards/check_gptzero_func": 0.1071428582072258,
"rewards/check_perplexity_diff_func": 0.5,
"rewards/check_winston_local_func": 0.355853408575058,
"step": 53
},
{
"clip_ratio": 0.0025832760147750378,
"epoch": 0.07552447552447553,
"grad_norm": 0.10695378216267973,
"kl": 0.0250244140625,
"learning_rate": 1.888111888111888e-05,
"loss": -0.0034,
"step": 54
},
{
"clip_ratio": 0.0017639577854424715,
"completion_length": 96.08928680419922,
"epoch": 0.07692307692307693,
"grad_norm": 0.17472941592839772,
"kl": 0.0322265625,
"learning_rate": 1.923076923076923e-05,
"loss": -0.0141,
"num_tokens": 544736.0,
"reward": 1.3478963375091553,
"reward_std": 0.8002303242683411,
"rewards/check_gptzero_func": 0.2857142984867096,
"rewards/check_perplexity_diff_func": 0.7142857313156128,
"rewards/check_winston_local_func": 0.34789615869522095,
"step": 55
},
{
"clip_ratio": 0.0034089265391230583,
"epoch": 0.07832167832167833,
"grad_norm": 0.14666884899599142,
"kl": 0.041748046875,
"learning_rate": 1.958041958041958e-05,
"loss": -0.0155,
"step": 56
},
{
"clip_ratio": 0.0026615143287926912,
"completion_length": 96.91072082519531,
"epoch": 0.07972027972027972,
"grad_norm": 0.1400337476406121,
"kl": 0.035400390625,
"learning_rate": 1.993006993006993e-05,
"loss": -0.027,
"num_tokens": 564803.0,
"reward": 1.0594887733459473,
"reward_std": 0.6912091970443726,
"rewards/check_gptzero_func": 0.2857142984867096,
"rewards/check_perplexity_diff_func": 0.4285714328289032,
"rewards/check_winston_local_func": 0.34520307183265686,
"step": 57
},
{
"clip_ratio": 0.003153084311634302,
"epoch": 0.08111888111888112,
"grad_norm": 0.138816359725178,
"kl": 0.042236328125,
"learning_rate": 2.027972027972028e-05,
"loss": -0.0274,
"step": 58
},
{
"clip_ratio": 0.0012626759707927704,
"completion_length": 76.35714721679688,
"epoch": 0.08251748251748252,
"grad_norm": 0.15940476459351663,
"kl": 0.0556640625,
"learning_rate": 2.062937062937063e-05,
"loss": -0.0057,
"num_tokens": 582265.0,
"reward": 0.992591381072998,
"reward_std": 0.5938137769699097,
"rewards/check_gptzero_func": 0.125,
"rewards/check_perplexity_diff_func": 0.4642857015132904,
"rewards/check_winston_local_func": 0.4033055901527405,
"step": 59
},
{
"clip_ratio": 0.004643784370273352,
"epoch": 0.08391608391608392,
"grad_norm": 0.15305834024157844,
"kl": 0.07421875,
"learning_rate": 2.097902097902098e-05,
"loss": -0.0064,
"step": 60
},
{
"clip_ratio": 0.0032869603019207716,
"completion_length": 103.33928680419922,
"epoch": 0.08531468531468532,
"grad_norm": 0.12260760902895611,
"kl": 0.0810546875,
"learning_rate": 2.132867132867133e-05,
"loss": 0.0066,
"num_tokens": 603454.0,
"reward": 1.191124439239502,
"reward_std": 0.5557684898376465,
"rewards/check_gptzero_func": 0.2321428507566452,
"rewards/check_perplexity_diff_func": 0.5357142686843872,
"rewards/check_winston_local_func": 0.4232672154903412,
"step": 61
},
{
"clip_ratio": 0.008058370091021061,
"epoch": 0.08671328671328671,
"grad_norm": 0.10224499878092017,
"kl": 0.1162109375,
"learning_rate": 2.1678321678321677e-05,
"loss": 0.0058,
"step": 62
},
{
"clip_ratio": 0.002202474046498537,
"completion_length": 91.4464340209961,
"epoch": 0.08811188811188811,
"grad_norm": 0.09776882919444682,
"kl": 0.2138671875,
"learning_rate": 2.202797202797203e-05,
"loss": 0.0021,
"num_tokens": 622705.0,
"reward": 1.1320958137512207,
"reward_std": 0.4060821831226349,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 0.3214285671710968,
"rewards/check_winston_local_func": 0.41781002283096313,
"step": 63
},
{
"clip_ratio": 0.0025903189089149237,
"epoch": 0.08951048951048951,
"grad_norm": 0.0963149975780195,
"kl": 0.2109375,
"learning_rate": 2.237762237762238e-05,
"loss": 0.0014,
"step": 64
},
{
"clip_ratio": 0.0016666523879393935,
"completion_length": 100.00000762939453,
"epoch": 0.09090909090909091,
"grad_norm": 0.12490001498293682,
"kl": 0.1220703125,
"learning_rate": 2.272727272727273e-05,
"loss": 0.0081,
"num_tokens": 643191.0,
"reward": 1.274580478668213,
"reward_std": 0.5277792811393738,
"rewards/check_gptzero_func": 0.2321428507566452,
"rewards/check_perplexity_diff_func": 0.5357142686843872,
"rewards/check_winston_local_func": 0.5067232847213745,
"step": 65
},
{
"clip_ratio": 0.0032784033101052046,
"epoch": 0.09230769230769231,
"grad_norm": 0.12114304348028067,
"kl": 0.134765625,
"learning_rate": 2.307692307692308e-05,
"loss": 0.0074,
"step": 66
},
{
"clip_ratio": 0.0017520035617053509,
"completion_length": 95.16072082519531,
"epoch": 0.0937062937062937,
"grad_norm": 0.14292853522837679,
"kl": 0.216796875,
"learning_rate": 2.342657342657343e-05,
"loss": -0.0018,
"num_tokens": 662736.0,
"reward": 1.1007850170135498,
"reward_std": 0.45985397696495056,
"rewards/check_gptzero_func": 0.25,
"rewards/check_perplexity_diff_func": 0.4642857015132904,
"rewards/check_winston_local_func": 0.38649922609329224,
"step": 67
},
{
"clip_ratio": 0.00579434959217906,
"epoch": 0.0951048951048951,
"grad_norm": 0.14340586759573087,
"kl": 0.2431640625,
"learning_rate": 2.377622377622378e-05,
"loss": -0.0029,
"step": 68
},
{
"clip_ratio": 0.0022546499967575073,
"completion_length": 86.17857360839844,
"epoch": 0.0965034965034965,
"grad_norm": 0.17287042335415018,
"kl": 0.26171875,
"learning_rate": 2.4125874125874125e-05,
"loss": 0.0059,
"num_tokens": 681030.0,
"reward": 1.4391711950302124,
"reward_std": 0.6516181230545044,
"rewards/check_gptzero_func": 0.5178571343421936,
"rewards/check_perplexity_diff_func": 0.4642857015132904,
"rewards/check_winston_local_func": 0.45702821016311646,
"step": 69
},
{
"clip_ratio": 0.004240955226123333,
"epoch": 0.0979020979020979,
"grad_norm": 0.16371494613893736,
"kl": 0.28125,
"learning_rate": 2.4475524475524478e-05,
"loss": 0.0048,
"step": 70
},
{
"clip_ratio": 0.003548440057784319,
"completion_length": 102.42857360839844,
"epoch": 0.0993006993006993,
"grad_norm": 0.1615418523026911,
"kl": 0.1708984375,
"learning_rate": 2.4825174825174828e-05,
"loss": 0.0015,
"num_tokens": 701874.0,
"reward": 1.4929367303848267,
"reward_std": 0.6083499193191528,
"rewards/check_gptzero_func": 0.3214285671710968,
"rewards/check_perplexity_diff_func": 0.7142857313156128,
"rewards/check_winston_local_func": 0.4572224020957947,
"step": 71
},
{
"clip_ratio": 0.004331245087087154,
"epoch": 0.1006993006993007,
"grad_norm": 0.12876017762018707,
"kl": 0.216796875,
"learning_rate": 2.5174825174825178e-05,
"loss": 0.0,
"step": 72
},
{
"clip_ratio": 0.003951852675527334,
"completion_length": 73.46428680419922,
"epoch": 0.1020979020979021,
"grad_norm": 0.20791748542634808,
"kl": 0.365234375,
"learning_rate": 2.5524475524475528e-05,
"loss": -0.0037,
"num_tokens": 718488.0,
"reward": 1.7953797578811646,
"reward_std": 0.5393027067184448,
"rewards/check_gptzero_func": 0.4464285671710968,
"rewards/check_perplexity_diff_func": 0.8571428656578064,
"rewards/check_winston_local_func": 0.49180838465690613,
"step": 73
},
{
"clip_ratio": 0.008312534540891647,
"epoch": 0.1034965034965035,
"grad_norm": 0.19232465302245624,
"kl": 0.5390625,
"learning_rate": 2.5874125874125877e-05,
"loss": -0.0053,
"step": 74
},
{
"clip_ratio": 0.0021736263297498226,
"completion_length": 97.73214721679688,
"epoch": 0.1048951048951049,
"grad_norm": 0.18504738647098173,
"kl": 0.2421875,
"learning_rate": 2.6223776223776224e-05,
"loss": -0.0023,
"num_tokens": 738619.0,
"reward": 1.284185767173767,
"reward_std": 0.6959513425827026,
"rewards/check_gptzero_func": 0.4464285671710968,
"rewards/check_perplexity_diff_func": 0.3928571343421936,
"rewards/check_winston_local_func": 0.4449000358581543,
"step": 75
},
{
"clip_ratio": 0.0033756059128791094,
"epoch": 0.1062937062937063,
"grad_norm": 0.17585639043073228,
"kl": 0.255859375,
"learning_rate": 2.6573426573426574e-05,
"loss": -0.0043,
"step": 76
},
{
"clip_ratio": 0.003740633837878704,
"completion_length": 95.67857360839844,
"epoch": 0.1076923076923077,
"grad_norm": 0.21573775826044048,
"kl": 0.2451171875,
"learning_rate": 2.6923076923076923e-05,
"loss": 0.0105,
"num_tokens": 758111.0,
"reward": 1.4672702550888062,
"reward_std": 0.85318922996521,
"rewards/check_gptzero_func": 0.25,
"rewards/check_perplexity_diff_func": 0.7857142686843872,
"rewards/check_winston_local_func": 0.43155592679977417,
"step": 77
},
{
"clip_ratio": 0.004517109598964453,
"epoch": 0.10909090909090909,
"grad_norm": 0.2071235680791843,
"kl": 0.302734375,
"learning_rate": 2.7272727272727273e-05,
"loss": 0.009,
"step": 78
},
{
"clip_ratio": 0.004641580395400524,
"completion_length": 117.75000762939453,
"epoch": 0.11048951048951049,
"grad_norm": 0.19278645376555428,
"kl": 0.259765625,
"learning_rate": 2.762237762237762e-05,
"loss": -0.0004,
"num_tokens": 780403.0,
"reward": 1.4974778890609741,
"reward_std": 0.59557044506073,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 0.6785714030265808,
"rewards/check_winston_local_func": 0.39033493399620056,
"step": 79
},
{
"clip_ratio": 0.006305322516709566,
"epoch": 0.11188811188811189,
"grad_norm": 0.13786850726930885,
"kl": 0.3125,
"learning_rate": 2.7972027972027976e-05,
"loss": -0.0019,
"step": 80
},
{
"clip_ratio": 0.0017778041074052453,
"completion_length": 101.03572082519531,
"epoch": 0.11328671328671329,
"grad_norm": 0.18099911272050398,
"kl": 1.7578125,
"learning_rate": 2.8321678321678326e-05,
"loss": 0.0053,
"num_tokens": 800231.0,
"reward": 1.420699954032898,
"reward_std": 0.5104399919509888,
"rewards/check_gptzero_func": 0.2678571343421936,
"rewards/check_perplexity_diff_func": 0.6428571343421936,
"rewards/check_winston_local_func": 0.5099858045578003,
"step": 81
},
{
"clip_ratio": 0.0030425102449953556,
"epoch": 0.11468531468531469,
"grad_norm": 0.10725025403922342,
"kl": 0.9921875,
"learning_rate": 2.8671328671328672e-05,
"loss": 0.0043,
"step": 82
},
{
"clip_ratio": 0.0030319676734507084,
"completion_length": 93.3214340209961,
"epoch": 0.11608391608391608,
"grad_norm": 0.22858957476463732,
"kl": 0.35546875,
"learning_rate": 2.9020979020979022e-05,
"loss": 0.0021,
"num_tokens": 819773.0,
"reward": 1.4201573133468628,
"reward_std": 0.761705756187439,
"rewards/check_gptzero_func": 0.3392857015132904,
"rewards/check_perplexity_diff_func": 0.5714285969734192,
"rewards/check_winston_local_func": 0.5094431042671204,
"step": 83
},
{
"clip_ratio": 0.0027519434224814177,
"epoch": 0.11748251748251748,
"grad_norm": 0.180287994009811,
"kl": 0.392578125,
"learning_rate": 2.9370629370629372e-05,
"loss": -0.0002,
"step": 84
},
{
"clip_ratio": 0.0034523813519626856,
"completion_length": 102.66072082519531,
"epoch": 0.11888111888111888,
"grad_norm": 0.25571068617285075,
"kl": 0.494140625,
"learning_rate": 2.972027972027972e-05,
"loss": -0.0103,
"num_tokens": 840280.0,
"reward": 1.7226576805114746,
"reward_std": 0.7545516490936279,
"rewards/check_gptzero_func": 0.5714285969734192,
"rewards/check_perplexity_diff_func": 0.6785714030265808,
"rewards/check_winston_local_func": 0.4726576805114746,
"step": 85
},
{
"clip_ratio": 0.00493992306292057,
"epoch": 0.12027972027972028,
"grad_norm": 0.2172932526081326,
"kl": 0.490234375,
"learning_rate": 3.0069930069930068e-05,
"loss": -0.013,
"step": 86
},
{
"clip_ratio": 0.003154533449560404,
"completion_length": 98.96428680419922,
"epoch": 0.12167832167832168,
"grad_norm": 0.3651883135439141,
"kl": 0.5859375,
"learning_rate": 3.0419580419580425e-05,
"loss": -0.0129,
"num_tokens": 860422.0,
"reward": 1.6449967622756958,
"reward_std": 0.7163441777229309,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 0.7142857313156128,
"rewards/check_winston_local_func": 0.5378537178039551,
"step": 87
},
{
"clip_ratio": 0.007247431669384241,
"epoch": 0.12307692307692308,
"grad_norm": 0.2167770735292921,
"kl": 0.61328125,
"learning_rate": 3.0769230769230774e-05,
"loss": -0.0168,
"step": 88
},
{
"clip_ratio": 0.0027036736719310284,
"completion_length": 108.03572082519531,
"epoch": 0.12447552447552447,
"grad_norm": 0.12604084880729077,
"kl": 0.291015625,
"learning_rate": 3.111888111888112e-05,
"loss": -0.0039,
"num_tokens": 881364.0,
"reward": 1.3799673318862915,
"reward_std": 0.6881024837493896,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 0.5714285969734192,
"rewards/check_winston_local_func": 0.4156815707683563,
"step": 89
},
{
"clip_ratio": 0.005238416139036417,
"epoch": 0.1258741258741259,
"grad_norm": 0.12103303780018655,
"kl": 0.30859375,
"learning_rate": 3.146853146853147e-05,
"loss": -0.005,
"step": 90
},
{
"clip_ratio": 0.003768512513488531,
"completion_length": 122.80357360839844,
"epoch": 0.12727272727272726,
"grad_norm": 0.18423815092289067,
"kl": 0.67578125,
"learning_rate": 3.181818181818182e-05,
"loss": -0.0171,
"num_tokens": 904487.0,
"reward": 1.6101170778274536,
"reward_std": 0.6241902709007263,
"rewards/check_gptzero_func": 0.3392857015132904,
"rewards/check_perplexity_diff_func": 0.7857142686843872,
"rewards/check_winston_local_func": 0.48511695861816406,
"step": 91
},
{
"clip_ratio": 0.005407070741057396,
"epoch": 0.12867132867132866,
"grad_norm": 0.15457101650134836,
"kl": 0.66796875,
"learning_rate": 3.216783216783217e-05,
"loss": -0.0182,
"step": 92
},
{
"clip_ratio": 0.002042532665655017,
"completion_length": 108.17857360839844,
"epoch": 0.13006993006993006,
"grad_norm": 0.19588538280583437,
"kl": 1.75,
"learning_rate": 3.251748251748252e-05,
"loss": -0.0023,
"num_tokens": 925633.0,
"reward": 1.3932582139968872,
"reward_std": 0.5087055563926697,
"rewards/check_gptzero_func": 0.4107142984867096,
"rewards/check_perplexity_diff_func": 0.4642857015132904,
"rewards/check_winston_local_func": 0.5182580351829529,
"step": 93
},
{
"clip_ratio": 0.00876330491155386,
"epoch": 0.13146853146853146,
"grad_norm": 0.1506256243597478,
"kl": 0.80078125,
"learning_rate": 3.2867132867132866e-05,
"loss": -0.0035,
"step": 94
},
{
"clip_ratio": 0.003118924330919981,
"completion_length": 77.4464340209961,
"epoch": 0.13286713286713286,
"grad_norm": 0.34784170699852607,
"kl": 0.72265625,
"learning_rate": 3.321678321678322e-05,
"loss": -0.0134,
"num_tokens": 942892.0,
"reward": 1.832355260848999,
"reward_std": 1.064102292060852,
"rewards/check_gptzero_func": 0.4821428656578064,
"rewards/check_perplexity_diff_func": 0.8214285969734192,
"rewards/check_winston_local_func": 0.5287837982177734,
"step": 95
},
{
"clip_ratio": 0.00509096821770072,
"epoch": 0.13426573426573427,
"grad_norm": 0.25968424971577175,
"kl": 0.96875,
"learning_rate": 3.356643356643357e-05,
"loss": -0.0176,
"step": 96
},
{
"clip_ratio": 0.0037272910121828318,
"completion_length": 95.26786041259766,
"epoch": 0.13566433566433567,
"grad_norm": 0.16914051199341906,
"kl": 0.349609375,
"learning_rate": 3.391608391608392e-05,
"loss": 0.0037,
"num_tokens": 962317.0,
"reward": 1.805726170539856,
"reward_std": 0.831261396408081,
"rewards/check_gptzero_func": 0.4107142984867096,
"rewards/check_perplexity_diff_func": 0.9285714030265808,
"rewards/check_winston_local_func": 0.46644020080566406,
"step": 97
},
{
"clip_ratio": 0.0034601751249283552,
"epoch": 0.13706293706293707,
"grad_norm": 0.16538870434198547,
"kl": 0.388671875,
"learning_rate": 3.4265734265734265e-05,
"loss": 0.0019,
"step": 98
},
{
"clip_ratio": 0.004487877711653709,
"completion_length": 111.8214340209961,
"epoch": 0.13846153846153847,
"grad_norm": 0.12415078751808917,
"kl": 0.384765625,
"learning_rate": 3.461538461538462e-05,
"loss": -0.001,
"num_tokens": 983513.0,
"reward": 1.2062333822250366,
"reward_std": 0.591974139213562,
"rewards/check_gptzero_func": 0.3214285671710968,
"rewards/check_perplexity_diff_func": 0.3571428656578064,
"rewards/check_winston_local_func": 0.5276618599891663,
"step": 99
},
{
"clip_ratio": 0.00284260674379766,
"epoch": 0.13986013986013987,
"grad_norm": 0.1197068572237121,
"kl": 0.376953125,
"learning_rate": 3.4965034965034965e-05,
"loss": -0.0017,
"step": 100
},
{
"clip_ratio": 0.0023206709884107113,
"completion_length": 101.08928680419922,
"epoch": 0.14125874125874127,
"grad_norm": 0.16400691003183243,
"kl": 0.30078125,
"learning_rate": 3.531468531468531e-05,
"loss": 0.0033,
"num_tokens": 1003878.0,
"reward": 1.631854772567749,
"reward_std": 0.5223387479782104,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 0.7142857313156128,
"rewards/check_winston_local_func": 0.5247119665145874,
"step": 101
},
{
"clip_ratio": 0.0016618981026113033,
"epoch": 0.14265734265734265,
"grad_norm": 0.1573929247787687,
"kl": 0.326171875,
"learning_rate": 3.566433566433567e-05,
"loss": 0.0012,
"step": 102
},
{
"clip_ratio": 0.0022201864048838615,
"completion_length": 124.46429443359375,
"epoch": 0.14405594405594405,
"grad_norm": 0.1353861256020412,
"kl": 0.3828125,
"learning_rate": 3.601398601398602e-05,
"loss": 0.005,
"num_tokens": 1027066.0,
"reward": 1.6920486688613892,
"reward_std": 0.48560601472854614,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 0.75,
"rewards/check_winston_local_func": 0.5491914749145508,
"step": 103
},
{
"clip_ratio": 0.0029075751081109047,
"epoch": 0.14545454545454545,
"grad_norm": 0.13328591024866146,
"kl": 0.39453125,
"learning_rate": 3.6363636363636364e-05,
"loss": 0.0034,
"step": 104
},
{
"clip_ratio": 0.0019614642951637506,
"completion_length": 85.26786041259766,
"epoch": 0.14685314685314685,
"grad_norm": 0.19674478002774554,
"kl": 0.333984375,
"learning_rate": 3.671328671328672e-05,
"loss": -0.0188,
"num_tokens": 1045567.0,
"reward": 1.6822034120559692,
"reward_std": 0.7559517621994019,
"rewards/check_gptzero_func": 0.5357142686843872,
"rewards/check_perplexity_diff_func": 0.6071428656578064,
"rewards/check_winston_local_func": 0.5393460988998413,
"step": 105
},
{
"clip_ratio": 0.004534203093498945,
"epoch": 0.14825174825174825,
"grad_norm": 0.16787611299305724,
"kl": 0.3671875,
"learning_rate": 3.7062937062937064e-05,
"loss": -0.0217,
"step": 106
},
{
"clip_ratio": 0.003009920008480549,
"completion_length": 106.55357360839844,
"epoch": 0.14965034965034965,
"grad_norm": 0.19751132540015684,
"kl": 0.455078125,
"learning_rate": 3.741258741258741e-05,
"loss": 0.0111,
"num_tokens": 1066466.0,
"reward": 2.178619861602783,
"reward_std": 0.7016831040382385,
"rewards/check_gptzero_func": 0.6071428656578064,
"rewards/check_perplexity_diff_func": 0.8928571343421936,
"rewards/check_winston_local_func": 0.6786197423934937,
"step": 107
},
{
"clip_ratio": 0.004473494831472635,
"epoch": 0.15104895104895105,
"grad_norm": 0.1431470008572649,
"kl": 0.4375,
"learning_rate": 3.776223776223776e-05,
"loss": 0.0095,
"step": 108
},
{
"clip_ratio": 0.0026541000697761774,
"completion_length": 100.6964340209961,
"epoch": 0.15244755244755245,
"grad_norm": 0.33135604940901486,
"kl": 0.310546875,
"learning_rate": 3.811188811188811e-05,
"loss": 0.0138,
"num_tokens": 1086751.0,
"reward": 1.6315226554870605,
"reward_std": 0.7806248068809509,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 0.6071428656578064,
"rewards/check_winston_local_func": 0.5600939393043518,
"step": 109
},
{
"clip_ratio": 0.0028892713598906994,
"epoch": 0.15384615384615385,
"grad_norm": 0.24217207616877234,
"kl": 0.361328125,
"learning_rate": 3.846153846153846e-05,
"loss": 0.0074,
"step": 110
},
{
"clip_ratio": 0.0021747422870248556,
"completion_length": 84.76786041259766,
"epoch": 0.15524475524475526,
"grad_norm": 0.15803622964815373,
"kl": 0.66796875,
"learning_rate": 3.8811188811188816e-05,
"loss": -0.0045,
"num_tokens": 1104518.0,
"reward": 1.6692500114440918,
"reward_std": 0.5965060591697693,
"rewards/check_gptzero_func": 0.5535714030265808,
"rewards/check_perplexity_diff_func": 0.6071428656578064,
"rewards/check_winston_local_func": 0.5085356831550598,
"step": 111
},
{
"clip_ratio": 0.004626331850886345,
"epoch": 0.15664335664335666,
"grad_norm": 0.14920406573741435,
"kl": 0.7578125,
"learning_rate": 3.916083916083916e-05,
"loss": -0.0064,
"step": 112
},
{
"clip_ratio": 0.002694516209885478,
"completion_length": 92.17857360839844,
"epoch": 0.15804195804195803,
"grad_norm": 0.161427063551978,
"kl": 0.361328125,
"learning_rate": 3.9510489510489516e-05,
"loss": 0.0058,
"num_tokens": 1123466.0,
"reward": 1.353637456893921,
"reward_std": 0.5352396965026855,
"rewards/check_gptzero_func": 0.4821428656578064,
"rewards/check_perplexity_diff_func": 0.3571428656578064,
"rewards/check_winston_local_func": 0.5143517255783081,
"step": 113
},
{
"clip_ratio": 0.00391918933019042,
"epoch": 0.15944055944055943,
"grad_norm": 0.14561571783883442,
"kl": 0.39453125,
"learning_rate": 3.986013986013986e-05,
"loss": 0.0036,
"step": 114
},
{
"clip_ratio": 0.001278667594306171,
"completion_length": 107.12500762939453,
"epoch": 0.16083916083916083,
"grad_norm": 0.1805624176716803,
"kl": 0.3046875,
"learning_rate": 4.020979020979021e-05,
"loss": -0.001,
"num_tokens": 1144619.0,
"reward": 1.6282455921173096,
"reward_std": 0.7325619459152222,
"rewards/check_gptzero_func": 0.4464285671710968,
"rewards/check_perplexity_diff_func": 0.5357142686843872,
"rewards/check_winston_local_func": 0.6461027264595032,
"step": 115
},
{
"clip_ratio": 0.0017050534952431917,
"epoch": 0.16223776223776223,
"grad_norm": 0.16942268791902212,
"kl": 0.337890625,
"learning_rate": 4.055944055944056e-05,
"loss": -0.0027,
"step": 116
},
{
"clip_ratio": 0.0015462420415133238,
"completion_length": 114.8214340209961,
"epoch": 0.16363636363636364,
"grad_norm": 0.1777133538888581,
"kl": 0.322265625,
"learning_rate": 4.0909090909090915e-05,
"loss": -0.0028,
"num_tokens": 1166185.0,
"reward": 1.566156268119812,
"reward_std": 0.38663557171821594,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 0.5,
"rewards/check_winston_local_func": 0.6018705368041992,
"step": 117
},
{
"clip_ratio": 0.0030793240293860435,
"epoch": 0.16503496503496504,
"grad_norm": 0.15084856022622706,
"kl": 0.2890625,
"learning_rate": 4.125874125874126e-05,
"loss": -0.0048,
"step": 118
},
{
"clip_ratio": 0.0018360918620601296,
"completion_length": 100.25000762939453,
"epoch": 0.16643356643356644,
"grad_norm": 0.10943003413749244,
"kl": 0.83984375,
"learning_rate": 4.1608391608391614e-05,
"loss": 0.0043,
"num_tokens": 1186317.0,
"reward": 1.7977957725524902,
"reward_std": 0.5172301530838013,
"rewards/check_gptzero_func": 0.5714285969734192,
"rewards/check_perplexity_diff_func": 0.7142857313156128,
"rewards/check_winston_local_func": 0.5120813250541687,
"step": 119
},
{
"clip_ratio": 0.0025589358992874622,
"epoch": 0.16783216783216784,
"grad_norm": 0.10431032316313237,
"kl": 0.68359375,
"learning_rate": 4.195804195804196e-05,
"loss": 0.0029,
"step": 120
},
{
"clip_ratio": 0.002800893737003207,
"completion_length": 120.91072082519531,
"epoch": 0.16923076923076924,
"grad_norm": 0.1567960287886374,
"kl": 0.37890625,
"learning_rate": 4.230769230769231e-05,
"loss": -0.0016,
"num_tokens": 1208530.0,
"reward": 1.5369055271148682,
"reward_std": 0.7073518633842468,
"rewards/check_gptzero_func": 0.3214285671710968,
"rewards/check_perplexity_diff_func": 0.6785714030265808,
"rewards/check_winston_local_func": 0.5369054079055786,
"step": 121
},
{
"clip_ratio": 0.0026409339625388384,
"epoch": 0.17062937062937064,
"grad_norm": 0.15013150065673506,
"kl": 0.37890625,
"learning_rate": 4.265734265734266e-05,
"loss": -0.0035,
"step": 122
},
{
"clip_ratio": 0.002211854327470064,
"completion_length": 110.66072082519531,
"epoch": 0.17202797202797201,
"grad_norm": 0.17274409069851862,
"kl": 0.380859375,
"learning_rate": 4.300699300699301e-05,
"loss": -0.0062,
"num_tokens": 1229171.0,
"reward": 1.6140996217727661,
"reward_std": 0.7221139669418335,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 0.5357142686843872,
"rewards/check_winston_local_func": 0.6140995025634766,
"step": 123
},
{
"clip_ratio": 0.005159804597496986,
"epoch": 0.17342657342657342,
"grad_norm": 0.13306053466541726,
"kl": 0.443359375,
"learning_rate": 4.335664335664335e-05,
"loss": -0.009,
"step": 124
},
{
"clip_ratio": 0.0031610180158168077,
"completion_length": 104.30357360839844,
"epoch": 0.17482517482517482,
"grad_norm": 0.19261275892706695,
"kl": 0.30078125,
"learning_rate": 4.370629370629371e-05,
"loss": -0.0129,
"num_tokens": 1249350.0,
"reward": 1.9047484397888184,
"reward_std": 0.6767317652702332,
"rewards/check_gptzero_func": 0.6071428656578064,
"rewards/check_perplexity_diff_func": 0.6071428656578064,
"rewards/check_winston_local_func": 0.6904626488685608,
"step": 125
},
{
"clip_ratio": 0.004231306724250317,
"epoch": 0.17622377622377622,
"grad_norm": 0.1752737652506695,
"kl": 0.33203125,
"learning_rate": 4.405594405594406e-05,
"loss": -0.0162,
"step": 126
},
{
"clip_ratio": 0.003981932066380978,
"completion_length": 109.53572082519531,
"epoch": 0.17762237762237762,
"grad_norm": 0.1423250496935692,
"kl": 0.48828125,
"learning_rate": 4.4405594405594406e-05,
"loss": -0.0102,
"num_tokens": 1269848.0,
"reward": 1.7591207027435303,
"reward_std": 0.5321380496025085,
"rewards/check_gptzero_func": 0.5,
"rewards/check_perplexity_diff_func": 0.6785714030265808,
"rewards/check_winston_local_func": 0.5805493593215942,
"step": 127
},
{
"clip_ratio": 0.005287667270749807,
"epoch": 0.17902097902097902,
"grad_norm": 0.13255848380623775,
"kl": 0.498046875,
"learning_rate": 4.475524475524476e-05,
"loss": -0.0121,
"step": 128
},
{
"clip_ratio": 0.0025025398936122656,
"completion_length": 105.30357360839844,
"epoch": 0.18041958041958042,
"grad_norm": 0.1177341259986552,
"kl": 0.283203125,
"learning_rate": 4.5104895104895105e-05,
"loss": -0.0016,
"num_tokens": 1290033.0,
"reward": 1.7174798250198364,
"reward_std": 0.5096268057823181,
"rewards/check_gptzero_func": 0.5357142686843872,
"rewards/check_perplexity_diff_func": 0.5357142686843872,
"rewards/check_winston_local_func": 0.6460510492324829,
"step": 129
},
{
"clip_ratio": 0.0033291254658252,
"epoch": 0.18181818181818182,
"grad_norm": 0.11141937591016414,
"kl": 0.294921875,
"learning_rate": 4.545454545454546e-05,
"loss": -0.0032,
"step": 130
},
{
"clip_ratio": 0.0024425899609923363,
"completion_length": 96.30357360839844,
"epoch": 0.18321678321678322,
"grad_norm": 0.23593816935965856,
"kl": 0.341796875,
"learning_rate": 4.5804195804195805e-05,
"loss": -0.006,
"num_tokens": 1309356.0,
"reward": 1.7568891048431396,
"reward_std": 0.7255779504776001,
"rewards/check_gptzero_func": 0.5892857313156128,
"rewards/check_perplexity_diff_func": 0.5357142686843872,
"rewards/check_winston_local_func": 0.6318890452384949,
"step": 131
},
{
"clip_ratio": 0.004331678152084351,
"epoch": 0.18461538461538463,
"grad_norm": 0.2135627059179365,
"kl": 0.349609375,
"learning_rate": 4.615384615384616e-05,
"loss": -0.0089,
"step": 132
},
{
"clip_ratio": 0.0028827630449086428,
"completion_length": 111.6964340209961,
"epoch": 0.18601398601398603,
"grad_norm": 0.1622976058137894,
"kl": 0.369140625,
"learning_rate": 4.6503496503496505e-05,
"loss": 0.014,
"num_tokens": 1330017.0,
"reward": 1.6692737340927124,
"reward_std": 0.628279983997345,
"rewards/check_gptzero_func": 0.4464285671710968,
"rewards/check_perplexity_diff_func": 0.6428571343421936,
"rewards/check_winston_local_func": 0.5799878835678101,
"step": 133
},
{
"clip_ratio": 0.004608546383678913,
"epoch": 0.1874125874125874,
"grad_norm": 0.16503328062657166,
"kl": 0.4140625,
"learning_rate": 4.685314685314686e-05,
"loss": 0.0121,
"step": 134
},
{
"clip_ratio": 0.0027426625601947308,
"completion_length": 107.58928680419922,
"epoch": 0.1888111888111888,
"grad_norm": 0.27926954484170907,
"kl": 0.5859375,
"learning_rate": 4.7202797202797204e-05,
"loss": 0.0139,
"num_tokens": 1350164.0,
"reward": 1.9428951740264893,
"reward_std": 0.6565932035446167,
"rewards/check_gptzero_func": 0.6785714030265808,
"rewards/check_perplexity_diff_func": 0.5714285969734192,
"rewards/check_winston_local_func": 0.6928950548171997,
"step": 135
},
{
"clip_ratio": 0.00439854059368372,
"epoch": 0.1902097902097902,
"grad_norm": 0.1925514269864402,
"kl": 0.5546875,
"learning_rate": 4.755244755244756e-05,
"loss": 0.0074,
"step": 136
},
{
"clip_ratio": 0.0018738384824246168,
"completion_length": 116.4464340209961,
"epoch": 0.1916083916083916,
"grad_norm": 0.14918417137816656,
"kl": 0.58984375,
"learning_rate": 4.7902097902097904e-05,
"loss": 0.0037,
"num_tokens": 1371507.0,
"reward": 2.0137577056884766,
"reward_std": 0.5378711819648743,
"rewards/check_gptzero_func": 0.6071428656578064,
"rewards/check_perplexity_diff_func": 0.6785714030265808,
"rewards/check_winston_local_func": 0.7280434370040894,
"step": 137
},
{
"clip_ratio": 0.003529219189658761,
"epoch": 0.193006993006993,
"grad_norm": 0.1301060608041501,
"kl": 0.6171875,
"learning_rate": 4.825174825174825e-05,
"loss": 0.0016,
"step": 138
},
{
"clip_ratio": 0.0011985624441877007,
"completion_length": 112.6964340209961,
"epoch": 0.1944055944055944,
"grad_norm": 0.13034934146859325,
"kl": 0.322265625,
"learning_rate": 4.86013986013986e-05,
"loss": -0.0024,
"num_tokens": 1392414.0,
"reward": 1.9355616569519043,
"reward_std": 0.4726971983909607,
"rewards/check_gptzero_func": 0.5892857313156128,
"rewards/check_perplexity_diff_func": 0.75,
"rewards/check_winston_local_func": 0.596275806427002,
"step": 139
},
{
"clip_ratio": 0.0032699662260711193,
"epoch": 0.1958041958041958,
"grad_norm": 0.12091032490165694,
"kl": 0.318359375,
"learning_rate": 4.8951048951048956e-05,
"loss": -0.0039,
"step": 140
},
{
"clip_ratio": 0.0013384540798142552,
"completion_length": 132.05357360839844,
"epoch": 0.1972027972027972,
"grad_norm": 0.12728573210952063,
"kl": 0.59765625,
"learning_rate": 4.93006993006993e-05,
"loss": 0.0045,
"num_tokens": 1416177.0,
"reward": 1.7924094200134277,
"reward_std": 0.7202263474464417,
"rewards/check_gptzero_func": 0.5714285969734192,
"rewards/check_perplexity_diff_func": 0.7142857313156128,
"rewards/check_winston_local_func": 0.5066950917243958,
"step": 141
},
{
"clip_ratio": 0.0017050639726221561,
"epoch": 0.1986013986013986,
"grad_norm": 0.12231361158637834,
"kl": 0.65234375,
"learning_rate": 4.9650349650349656e-05,
"loss": 0.0025,
"step": 142
},
{
"clip_ratio": 0.0015860958956182003,
"completion_length": 99.73214721679688,
"epoch": 0.2,
"grad_norm": 0.1853946177277763,
"kl": 0.58984375,
"learning_rate": 5e-05,
"loss": 0.0021,
"num_tokens": 1435732.0,
"reward": 2.0557591915130615,
"reward_std": 0.503829836845398,
"rewards/check_gptzero_func": 0.6964285969734192,
"rewards/check_perplexity_diff_func": 0.75,
"rewards/check_winston_local_func": 0.6093305945396423,
"step": 143
},
{
"clip_ratio": 0.002123563550412655,
"epoch": 0.2013986013986014,
"grad_norm": 0.13470126319919157,
"kl": 0.423828125,
"learning_rate": 5.0349650349650356e-05,
"loss": -0.0009,
"step": 144
},
{
"clip_ratio": 0.0018757216166704893,
"completion_length": 95.73214721679688,
"epoch": 0.20279720279720279,
"grad_norm": 0.14930190515465938,
"kl": 0.388671875,
"learning_rate": 5.06993006993007e-05,
"loss": -0.0026,
"num_tokens": 1454705.0,
"reward": 2.0560998916625977,
"reward_std": 0.5541732311248779,
"rewards/check_gptzero_func": 0.6964285969734192,
"rewards/check_perplexity_diff_func": 0.5714285969734192,
"rewards/check_winston_local_func": 0.788242518901825,
"step": 145
},
{
"clip_ratio": 0.003540371311828494,
"epoch": 0.2041958041958042,
"grad_norm": 0.1304251509197979,
"kl": 0.41796875,
"learning_rate": 5.1048951048951055e-05,
"loss": -0.0046,
"step": 146
},
{
"clip_ratio": 0.0031265008728951216,
"completion_length": 80.9464340209961,
"epoch": 0.2055944055944056,
"grad_norm": 0.2087929657840581,
"kl": 1.1328125,
"learning_rate": 5.1398601398601395e-05,
"loss": 0.0026,
"num_tokens": 1472182.0,
"reward": 1.9076076745986938,
"reward_std": 0.46633273363113403,
"rewards/check_gptzero_func": 0.6428571343421936,
"rewards/check_perplexity_diff_func": 0.5,
"rewards/check_winston_local_func": 0.7647504210472107,
"step": 147
},
{
"clip_ratio": 0.0029583375435322523,
"epoch": 0.206993006993007,
"grad_norm": 0.178151823438605,
"kl": 0.66015625,
"learning_rate": 5.1748251748251755e-05,
"loss": -0.0014,
"step": 148
},
{
"clip_ratio": 0.0020277961157262325,
"completion_length": 126.58929443359375,
"epoch": 0.2083916083916084,
"grad_norm": 0.13736153026044381,
"kl": 0.373046875,
"learning_rate": 5.2097902097902094e-05,
"loss": 0.0127,
"num_tokens": 1494581.0,
"reward": 1.870171308517456,
"reward_std": 0.4412032663822174,
"rewards/check_gptzero_func": 0.4464285671710968,
"rewards/check_perplexity_diff_func": 0.8214285969734192,
"rewards/check_winston_local_func": 0.6023141741752625,
"step": 149
},
{
"clip_ratio": 0.0022643795236945152,
"epoch": 0.2097902097902098,
"grad_norm": 0.12942677801199462,
"kl": 0.40625,
"learning_rate": 5.244755244755245e-05,
"loss": 0.0106,
"step": 150
},
{
"clip_ratio": 0.001886948710307479,
"completion_length": 95.14286041259766,
"epoch": 0.2111888111888112,
"grad_norm": 0.15555312890089615,
"kl": 0.62890625,
"learning_rate": 5.279720279720281e-05,
"loss": 0.0013,
"num_tokens": 1513403.0,
"reward": 1.9735006093978882,
"reward_std": 0.5218394994735718,
"rewards/check_gptzero_func": 0.5535714030265808,
"rewards/check_perplexity_diff_func": 0.6071428656578064,
"rewards/check_winston_local_func": 0.8127861022949219,
"step": 151
},
{
"clip_ratio": 0.003476199461147189,
"epoch": 0.2125874125874126,
"grad_norm": 0.13833602394582134,
"kl": 0.5390625,
"learning_rate": 5.314685314685315e-05,
"loss": -0.0015,
"step": 152
},
{
"clip_ratio": 0.0029247915372252464,
"completion_length": 104.1964340209961,
"epoch": 0.213986013986014,
"grad_norm": 0.18332277821774023,
"kl": 0.451171875,
"learning_rate": 5.34965034965035e-05,
"loss": -0.0047,
"num_tokens": 1533154.0,
"reward": 1.977178931236267,
"reward_std": 0.5758928656578064,
"rewards/check_gptzero_func": 0.6785714030265808,
"rewards/check_perplexity_diff_func": 0.6785714030265808,
"rewards/check_winston_local_func": 0.6200361251831055,
"step": 153
},
{
"clip_ratio": 0.002894646255299449,
"epoch": 0.2153846153846154,
"grad_norm": 0.12417162413759981,
"kl": 0.48828125,
"learning_rate": 5.384615384615385e-05,
"loss": -0.0082,
"step": 154
},
{
"clip_ratio": 0.002533347113057971,
"completion_length": 106.6964340209961,
"epoch": 0.21678321678321677,
"grad_norm": 0.2165608004438418,
"kl": 0.4140625,
"learning_rate": 5.41958041958042e-05,
"loss": 0.0367,
"num_tokens": 1554005.0,
"reward": 2.254620313644409,
"reward_std": 0.6860859394073486,
"rewards/check_gptzero_func": 0.7857142686843872,
"rewards/check_perplexity_diff_func": 0.7142857313156128,
"rewards/check_winston_local_func": 0.7546200752258301,
"step": 155
},
{
"clip_ratio": 0.005156368017196655,
"epoch": 0.21818181818181817,
"grad_norm": 0.16185524517286934,
"kl": 0.4609375,
"learning_rate": 5.4545454545454546e-05,
"loss": 0.0322,
"step": 156
},
{
"clip_ratio": 0.002192203886806965,
"completion_length": 129.55357360839844,
"epoch": 0.21958041958041957,
"grad_norm": 0.13686573639431518,
"kl": 0.421875,
"learning_rate": 5.48951048951049e-05,
"loss": 0.0095,
"num_tokens": 1576988.0,
"reward": 2.202296495437622,
"reward_std": 0.4840867817401886,
"rewards/check_gptzero_func": 0.6428571343421936,
"rewards/check_perplexity_diff_func": 0.8928571343421936,
"rewards/check_winston_local_func": 0.6665821075439453,
"step": 157
},
{
"clip_ratio": 0.0037054666317999363,
"epoch": 0.22097902097902097,
"grad_norm": 0.12040805378810834,
"kl": 0.45703125,
"learning_rate": 5.524475524475524e-05,
"loss": 0.0077,
"step": 158
},
{
"clip_ratio": 0.002324200002476573,
"completion_length": 141.10714721679688,
"epoch": 0.22237762237762237,
"grad_norm": 0.1722210877266639,
"kl": 0.9765625,
"learning_rate": 5.55944055944056e-05,
"loss": -0.0049,
"num_tokens": 1600826.0,
"reward": 2.0245378017425537,
"reward_std": 0.45877185463905334,
"rewards/check_gptzero_func": 0.5892857313156128,
"rewards/check_perplexity_diff_func": 0.75,
"rewards/check_winston_local_func": 0.6852518916130066,
"step": 159
},
{
"clip_ratio": 0.003470373572781682,
"epoch": 0.22377622377622378,
"grad_norm": 0.13356363193434526,
"kl": 0.65234375,
"learning_rate": 5.594405594405595e-05,
"loss": -0.0085,
"step": 160
},
{
"clip_ratio": 0.0026535126380622387,
"completion_length": 109.76786041259766,
"epoch": 0.22517482517482518,
"grad_norm": 0.409404915148595,
"kl": 0.515625,
"learning_rate": 5.629370629370629e-05,
"loss": -0.0051,
"num_tokens": 1621103.0,
"reward": 2.0040969848632812,
"reward_std": 0.5942196249961853,
"rewards/check_gptzero_func": 0.5178571343421936,
"rewards/check_perplexity_diff_func": 0.7857142686843872,
"rewards/check_winston_local_func": 0.7005256414413452,
"step": 161
},
{
"clip_ratio": 0.0058495416305959225,
"epoch": 0.22657342657342658,
"grad_norm": 0.16376328147020022,
"kl": 0.5078125,
"learning_rate": 5.664335664335665e-05,
"loss": -0.0129,
"step": 162
},
{
"clip_ratio": 0.002924378262832761,
"completion_length": 114.3214340209961,
"epoch": 0.22797202797202798,
"grad_norm": 0.17698045766822795,
"kl": 0.62109375,
"learning_rate": 5.699300699300699e-05,
"loss": -0.0053,
"num_tokens": 1642629.0,
"reward": 2.0913121700286865,
"reward_std": 0.533393144607544,
"rewards/check_gptzero_func": 0.6785714030265808,
"rewards/check_perplexity_diff_func": 0.6785714030265808,
"rewards/check_winston_local_func": 0.7341693639755249,
"step": 163
},
{
"clip_ratio": 0.006533453240990639,
"epoch": 0.22937062937062938,
"grad_norm": 0.15566114152311913,
"kl": 0.66796875,
"learning_rate": 5.7342657342657345e-05,
"loss": -0.0085,
"step": 164
},
{
"clip_ratio": 0.002033352619037032,
"completion_length": 110.48214721679688,
"epoch": 0.23076923076923078,
"grad_norm": 0.19340316318251602,
"kl": 0.4453125,
"learning_rate": 5.769230769230769e-05,
"loss": 0.0054,
"num_tokens": 1663000.0,
"reward": 1.809409499168396,
"reward_std": 0.5209853053092957,
"rewards/check_gptzero_func": 0.6428571343421936,
"rewards/check_perplexity_diff_func": 0.3571428656578064,
"rewards/check_winston_local_func": 0.809409499168396,
"step": 165
},
{
"clip_ratio": 0.005329853855073452,
"epoch": 0.23216783216783216,
"grad_norm": 0.15299329983895626,
"kl": 0.55078125,
"learning_rate": 5.8041958041958044e-05,
"loss": 0.0018,
"step": 166
},
{
"clip_ratio": 0.0025733087677508593,
"completion_length": 95.73214721679688,
"epoch": 0.23356643356643356,
"grad_norm": 0.16102285972903455,
"kl": 0.458984375,
"learning_rate": 5.83916083916084e-05,
"loss": -0.0008,
"num_tokens": 1682123.0,
"reward": 1.9705681800842285,
"reward_std": 0.5465096235275269,
"rewards/check_gptzero_func": 0.6428571343421936,
"rewards/check_perplexity_diff_func": 0.6428571343421936,
"rewards/check_winston_local_func": 0.6848538517951965,
"step": 167
},
{
"clip_ratio": 0.0034487086813896894,
"epoch": 0.23496503496503496,
"grad_norm": 0.12896214559568192,
"kl": 0.474609375,
"learning_rate": 5.8741258741258744e-05,
"loss": -0.0039,
"step": 168
},
{
"clip_ratio": 0.0012721805833280087,
"completion_length": 72.125,
"epoch": 0.23636363636363636,
"grad_norm": 0.19976827264055144,
"kl": 0.5703125,
"learning_rate": 5.90909090909091e-05,
"loss": -0.0046,
"num_tokens": 1698078.0,
"reward": 1.8698246479034424,
"reward_std": 0.4430284798145294,
"rewards/check_gptzero_func": 0.7142857313156128,
"rewards/check_perplexity_diff_func": 0.3571428656578064,
"rewards/check_winston_local_func": 0.7983959913253784,
"step": 169
},
{
"clip_ratio": 0.005269515328109264,
"epoch": 0.23776223776223776,
"grad_norm": 0.14855282442633314,
"kl": 0.61328125,
"learning_rate": 5.944055944055944e-05,
"loss": -0.0079,
"step": 170
},
{
"clip_ratio": 0.0027731256559491158,
"completion_length": 117.83929443359375,
"epoch": 0.23916083916083916,
"grad_norm": 0.23058249563979205,
"kl": 0.412109375,
"learning_rate": 5.9790209790209796e-05,
"loss": 0.0055,
"num_tokens": 1719673.0,
"reward": 2.2663450241088867,
"reward_std": 0.46630793809890747,
"rewards/check_gptzero_func": 0.6785714030265808,
"rewards/check_perplexity_diff_func": 0.7142857313156128,
"rewards/check_winston_local_func": 0.8734878301620483,
"step": 171
},
{
"clip_ratio": 0.0034511894918978214,
"epoch": 0.24055944055944056,
"grad_norm": 0.17368207545699044,
"kl": 0.373046875,
"learning_rate": 6.0139860139860136e-05,
"loss": -0.0007,
"step": 172
},
{
"clip_ratio": 0.0032431341242045164,
"completion_length": 126.85714721679688,
"epoch": 0.24195804195804196,
"grad_norm": 0.17995690811350934,
"kl": 0.353515625,
"learning_rate": 6.048951048951049e-05,
"loss": -0.0011,
"num_tokens": 1742569.0,
"reward": 2.258990526199341,
"reward_std": 0.6008436679840088,
"rewards/check_gptzero_func": 0.75,
"rewards/check_perplexity_diff_func": 0.75,
"rewards/check_winston_local_func": 0.7589904069900513,
"step": 173
},
{
"clip_ratio": 0.006950656417757273,
"epoch": 0.24335664335664337,
"grad_norm": 0.13892848528506246,
"kl": 0.359375,
"learning_rate": 6.083916083916085e-05,
"loss": -0.0033,
"step": 174
},
{
"clip_ratio": 0.0027662119828164577,
"completion_length": 130.8928680419922,
"epoch": 0.24475524475524477,
"grad_norm": 0.10040261340554028,
"kl": 0.345703125,
"learning_rate": 6.118881118881119e-05,
"loss": -0.0032,
"num_tokens": 1765643.0,
"reward": 2.074061632156372,
"reward_std": 0.3745954632759094,
"rewards/check_gptzero_func": 0.6607142686843872,
"rewards/check_perplexity_diff_func": 0.6785714030265808,
"rewards/check_winston_local_func": 0.7347758412361145,
"step": 175
},
{
"clip_ratio": 0.0030742601957172155,
"epoch": 0.24615384615384617,
"grad_norm": 0.0921134439221548,
"kl": 0.3359375,
"learning_rate": 6.153846153846155e-05,
"loss": -0.0044,
"step": 176
},
{
"clip_ratio": 0.002027927665039897,
"completion_length": 89.73214721679688,
"epoch": 0.24755244755244754,
"grad_norm": 0.1709976566266231,
"kl": 0.50390625,
"learning_rate": 6.188811188811188e-05,
"loss": 0.0051,
"num_tokens": 1783674.0,
"reward": 1.7810758352279663,
"reward_std": 0.4943030774593353,
"rewards/check_gptzero_func": 0.6428571343421936,
"rewards/check_perplexity_diff_func": 0.4285714328289032,
"rewards/check_winston_local_func": 0.7096471786499023,
"step": 177
},
{
"clip_ratio": 0.0020612890366464853,
"epoch": 0.24895104895104894,
"grad_norm": 0.1477275348028994,
"kl": 0.50390625,
"learning_rate": 6.223776223776224e-05,
"loss": 0.0011,
"step": 178
},
{
"clip_ratio": 0.0021104796323925257,
"completion_length": 132.10714721679688,
"epoch": 0.25034965034965034,
"grad_norm": 0.08842848405697774,
"kl": 0.39453125,
"learning_rate": 6.258741258741259e-05,
"loss": 0.0005,
"num_tokens": 1806754.0,
"reward": 2.107767105102539,
"reward_std": 0.269815057516098,
"rewards/check_gptzero_func": 0.75,
"rewards/check_perplexity_diff_func": 0.5357142686843872,
"rewards/check_winston_local_func": 0.8220529556274414,
"step": 179
},
{
"clip_ratio": 0.004407648928463459,
"epoch": 0.2517482517482518,
"grad_norm": 0.07542595290297384,
"kl": 0.416015625,
"learning_rate": 6.293706293706293e-05,
"loss": -0.0006,
"step": 180
},
{
"clip_ratio": 0.0018615310546010733,
"completion_length": 89.30357360839844,
"epoch": 0.25314685314685315,
"grad_norm": 0.2895549842864223,
"kl": 0.498046875,
"learning_rate": 6.32867132867133e-05,
"loss": -0.0082,
"num_tokens": 1824795.0,
"reward": 1.8801069259643555,
"reward_std": 0.5523244738578796,
"rewards/check_gptzero_func": 0.6071428656578064,
"rewards/check_perplexity_diff_func": 0.5,
"rewards/check_winston_local_func": 0.7729640603065491,
"step": 181
},
{
"clip_ratio": 0.007555535528808832,
"epoch": 0.2545454545454545,
"grad_norm": 0.20874610098022578,
"kl": 0.59375,
"learning_rate": 6.363636363636364e-05,
"loss": -0.016,
"step": 182
},
{
"clip_ratio": 0.0016618422232568264,
"completion_length": 84.46428680419922,
"epoch": 0.25594405594405595,
"grad_norm": 0.20653753743630154,
"kl": 0.5625,
"learning_rate": 6.398601398601399e-05,
"loss": 0.0058,
"num_tokens": 1842573.0,
"reward": 2.1319751739501953,
"reward_std": 0.5030335187911987,
"rewards/check_gptzero_func": 0.7321428656578064,
"rewards/check_perplexity_diff_func": 0.5714285969734192,
"rewards/check_winston_local_func": 0.8284037709236145,
"step": 183
},
{
"clip_ratio": 0.00579726742580533,
"epoch": 0.2573426573426573,
"grad_norm": 0.15782001907156346,
"kl": 0.578125,
"learning_rate": 6.433566433566433e-05,
"loss": 0.0001,
"step": 184
},
{
"clip_ratio": 0.0016176491044461727,
"completion_length": 125.39286041259766,
"epoch": 0.25874125874125875,
"grad_norm": 0.12385736188741701,
"kl": 0.4921875,
"learning_rate": 6.46853146853147e-05,
"loss": -0.0027,
"num_tokens": 1864297.0,
"reward": 1.7732529640197754,
"reward_std": 0.3774341642856598,
"rewards/check_gptzero_func": 0.625,
"rewards/check_perplexity_diff_func": 0.4285714328289032,
"rewards/check_winston_local_func": 0.7196813821792603,
"step": 185
},
{
"clip_ratio": 0.00623862212523818,
"epoch": 0.2601398601398601,
"grad_norm": 0.10877531810926387,
"kl": 0.53125,
"learning_rate": 6.503496503496504e-05,
"loss": -0.0048,
"step": 186
},
{
"clip_ratio": 0.0031279984395951033,
"completion_length": 98.00000762939453,
"epoch": 0.26153846153846155,
"grad_norm": 0.3661429387824046,
"kl": 0.6328125,
"learning_rate": 6.538461538461539e-05,
"loss": -0.0063,
"num_tokens": 1882931.0,
"reward": 2.0694050788879395,
"reward_std": 0.6090536117553711,
"rewards/check_gptzero_func": 0.6785714030265808,
"rewards/check_perplexity_diff_func": 0.5714285969734192,
"rewards/check_winston_local_func": 0.819405198097229,
"step": 187
},
{
"clip_ratio": 0.010017934255301952,
"epoch": 0.2629370629370629,
"grad_norm": 0.34977841113727764,
"kl": 0.65234375,
"learning_rate": 6.573426573426573e-05,
"loss": -0.0185,
"step": 188
},
{
"clip_ratio": 0.00222155568189919,
"completion_length": 133.44644165039062,
"epoch": 0.26433566433566436,
"grad_norm": 0.18715243199492715,
"kl": 0.46484375,
"learning_rate": 6.608391608391609e-05,
"loss": 0.0062,
"num_tokens": 1906084.0,
"reward": 2.364468812942505,
"reward_std": 0.5593475699424744,
"rewards/check_gptzero_func": 0.6964285969734192,
"rewards/check_perplexity_diff_func": 0.8571428656578064,
"rewards/check_winston_local_func": 0.8108974099159241,
"step": 189
},
{
"clip_ratio": 0.008103223517537117,
"epoch": 0.26573426573426573,
"grad_norm": 0.14058389303467,
"kl": 0.45703125,
"learning_rate": 6.643356643356644e-05,
"loss": 0.0025,
"step": 190
},
{
"clip_ratio": 0.0019668787717819214,
"completion_length": 142.07144165039062,
"epoch": 0.26713286713286716,
"grad_norm": 0.14399812178023874,
"kl": 0.43359375,
"learning_rate": 6.678321678321679e-05,
"loss": -0.0031,
"num_tokens": 1930224.0,
"reward": 2.1098926067352295,
"reward_std": 0.5390018820762634,
"rewards/check_gptzero_func": 0.8214285969734192,
"rewards/check_perplexity_diff_func": 0.4642857015132904,
"rewards/check_winston_local_func": 0.8241782784461975,
"step": 191
},
{
"clip_ratio": 0.005082913674414158,
"epoch": 0.26853146853146853,
"grad_norm": 0.12024703070308183,
"kl": 0.447265625,
"learning_rate": 6.713286713286715e-05,
"loss": -0.0055,
"step": 192
},
{
"clip_ratio": 0.0032251765951514244,
"completion_length": 84.98214721679688,
"epoch": 0.2699300699300699,
"grad_norm": 0.32527529059153654,
"kl": 0.6640625,
"learning_rate": 6.748251748251748e-05,
"loss": 0.0186,
"num_tokens": 1947399.0,
"reward": 1.9443087577819824,
"reward_std": 0.6417390704154968,
"rewards/check_gptzero_func": 0.7678571343421936,
"rewards/check_perplexity_diff_func": 0.2857142984867096,
"rewards/check_winston_local_func": 0.8907372355461121,
"step": 193
},
{
"clip_ratio": 0.008085422217845917,
"epoch": 0.27132867132867133,
"grad_norm": 0.21806984416977268,
"kl": 0.69140625,
"learning_rate": 6.783216783216784e-05,
"loss": 0.0068,
"step": 194
},
{
"clip_ratio": 0.0025381618179380894,
"completion_length": 91.66072082519531,
"epoch": 0.2727272727272727,
"grad_norm": 0.1831840025015104,
"kl": 0.53125,
"learning_rate": 6.818181818181818e-05,
"loss": 0.0014,
"num_tokens": 1965402.0,
"reward": 2.3979477882385254,
"reward_std": 0.45264866948127747,
"rewards/check_gptzero_func": 0.7678571343421936,
"rewards/check_perplexity_diff_func": 0.7857142686843872,
"rewards/check_winston_local_func": 0.8443759083747864,
"step": 195
},
{
"clip_ratio": 0.005082620773464441,
"epoch": 0.27412587412587414,
"grad_norm": 0.14399845631168656,
"kl": 0.515625,
"learning_rate": 6.853146853146853e-05,
"loss": -0.0023,
"step": 196
},
{
"clip_ratio": 0.0025359569117426872,
"completion_length": 118.80357360839844,
"epoch": 0.2755244755244755,
"grad_norm": 0.1617404225919446,
"kl": 0.5625,
"learning_rate": 6.888111888111889e-05,
"loss": 0.0146,
"num_tokens": 1986731.0,
"reward": 2.078895330429077,
"reward_std": 0.31289052963256836,
"rewards/check_gptzero_func": 0.8035714030265808,
"rewards/check_perplexity_diff_func": 0.3928571343421936,
"rewards/check_winston_local_func": 0.882466733455658,
"step": 197
},
{
"clip_ratio": 0.0062335156835615635,
"epoch": 0.27692307692307694,
"grad_norm": 0.12006362566688389,
"kl": 0.69921875,
"learning_rate": 6.923076923076924e-05,
"loss": 0.0125,
"step": 198
},
{
"clip_ratio": 0.0023628328926861286,
"completion_length": 113.76786041259766,
"epoch": 0.2783216783216783,
"grad_norm": 0.19235514876641358,
"kl": 0.58984375,
"learning_rate": 6.958041958041958e-05,
"loss": -0.0069,
"num_tokens": 2007822.0,
"reward": 2.0453529357910156,
"reward_std": 0.576248288154602,
"rewards/check_gptzero_func": 0.7321428656578064,
"rewards/check_perplexity_diff_func": 0.5714285969734192,
"rewards/check_winston_local_func": 0.7417814135551453,
"step": 199
},
{
"clip_ratio": 0.0030245708767324686,
"epoch": 0.27972027972027974,
"grad_norm": 0.15320753759426498,
"kl": 0.578125,
"learning_rate": 6.993006993006993e-05,
"loss": -0.0115,
"step": 200
},
{
"clip_ratio": 0.0028001146856695414,
"completion_length": 121.51786041259766,
"epoch": 0.2811188811188811,
"grad_norm": 0.2770839119590159,
"kl": 0.5859375,
"learning_rate": 7.027972027972029e-05,
"loss": -0.0321,
"num_tokens": 2030153.0,
"reward": 2.223505735397339,
"reward_std": 0.580276370048523,
"rewards/check_gptzero_func": 0.8392857313156128,
"rewards/check_perplexity_diff_func": 0.5,
"rewards/check_winston_local_func": 0.8842198252677917,
"step": 201
},
{
"clip_ratio": 0.004435424692928791,
"epoch": 0.28251748251748254,
"grad_norm": 0.1752123363810394,
"kl": 0.61328125,
"learning_rate": 7.062937062937062e-05,
"loss": -0.0398,
"step": 202
},
{
"clip_ratio": 0.003775001736357808,
"completion_length": 132.96429443359375,
"epoch": 0.2839160839160839,
"grad_norm": 0.20528144857768493,
"kl": 0.5390625,
"learning_rate": 7.097902097902098e-05,
"loss": -0.0054,
"num_tokens": 2053611.0,
"reward": 2.4398131370544434,
"reward_std": 0.6120996475219727,
"rewards/check_gptzero_func": 0.8214285969734192,
"rewards/check_perplexity_diff_func": 0.8571428656578064,
"rewards/check_winston_local_func": 0.761241614818573,
"step": 203
},
{
"clip_ratio": 0.0037133130244910717,
"epoch": 0.2853146853146853,
"grad_norm": 0.14916355666745199,
"kl": 0.55859375,
"learning_rate": 7.132867132867134e-05,
"loss": -0.0112,
"step": 204
},
{
"clip_ratio": 0.002573356730863452,
"completion_length": 109.51786041259766,
"epoch": 0.2867132867132867,
"grad_norm": 0.15672915743695726,
"kl": 1.4921875,
"learning_rate": 7.167832167832168e-05,
"loss": 0.0032,
"num_tokens": 2073846.0,
"reward": 2.221467971801758,
"reward_std": 0.492183119058609,
"rewards/check_gptzero_func": 0.6964285969734192,
"rewards/check_perplexity_diff_func": 0.6428571343421936,
"rewards/check_winston_local_func": 0.8821821212768555,
"step": 205
},
{
"clip_ratio": 0.0036338225472718477,
"epoch": 0.2881118881118881,
"grad_norm": 0.27785877634425765,
"kl": 0.74609375,
"learning_rate": 7.202797202797204e-05,
"loss": 0.0013,
"step": 206
},
{
"clip_ratio": 0.004580673761665821,
"completion_length": 95.05357360839844,
"epoch": 0.2895104895104895,
"grad_norm": 0.1746155777114636,
"kl": 0.671875,
"learning_rate": 7.237762237762238e-05,
"loss": -0.0068,
"num_tokens": 2092355.0,
"reward": 2.0429205894470215,
"reward_std": 0.36572587490081787,
"rewards/check_gptzero_func": 0.6964285969734192,
"rewards/check_perplexity_diff_func": 0.4642857015132904,
"rewards/check_winston_local_func": 0.8822061419487,
"step": 207
},
{
"clip_ratio": 0.007289381232112646,
"epoch": 0.2909090909090909,
"grad_norm": 0.12704093759252294,
"kl": 0.75390625,
"learning_rate": 7.272727272727273e-05,
"loss": -0.0111,
"step": 208
},
{
"clip_ratio": 0.0038105440326035023,
"completion_length": 113.4464340209961,
"epoch": 0.2923076923076923,
"grad_norm": 0.20958397036994925,
"kl": 0.6015625,
"learning_rate": 7.307692307692307e-05,
"loss": 0.0082,
"num_tokens": 2113542.0,
"reward": 2.162785530090332,
"reward_std": 0.49807849526405334,
"rewards/check_gptzero_func": 0.75,
"rewards/check_perplexity_diff_func": 0.5714285969734192,
"rewards/check_winston_local_func": 0.8413568139076233,
"step": 209
},
{
"clip_ratio": 0.007606918923556805,
"epoch": 0.2937062937062937,
"grad_norm": 0.1766466418371376,
"kl": 0.63671875,
"learning_rate": 7.342657342657343e-05,
"loss": 0.0025,
"step": 210
},
{
"clip_ratio": 0.0019986790139228106,
"completion_length": 92.73214721679688,
"epoch": 0.2951048951048951,
"grad_norm": 0.2716166901527731,
"kl": 0.6171875,
"learning_rate": 7.377622377622378e-05,
"loss": -0.0005,
"num_tokens": 2131873.0,
"reward": 2.099912405014038,
"reward_std": 0.5661742687225342,
"rewards/check_gptzero_func": 0.8214285969734192,
"rewards/check_perplexity_diff_func": 0.4285714328289032,
"rewards/check_winston_local_func": 0.8499122858047485,
"step": 211
},
{
"clip_ratio": 0.005917501635849476,
"epoch": 0.2965034965034965,
"grad_norm": 0.19270405158731238,
"kl": 0.62890625,
"learning_rate": 7.412587412587413e-05,
"loss": -0.0079,
"step": 212
},
{
"clip_ratio": 0.002530448604375124,
"completion_length": 96.51786041259766,
"epoch": 0.29790209790209793,
"grad_norm": 0.22723256705117592,
"kl": 0.69921875,
"learning_rate": 7.447552447552449e-05,
"loss": -0.009,
"num_tokens": 2150462.0,
"reward": 2.1355700492858887,
"reward_std": 0.3738899827003479,
"rewards/check_gptzero_func": 0.8571428656578064,
"rewards/check_perplexity_diff_func": 0.3571428656578064,
"rewards/check_winston_local_func": 0.9212842583656311,
"step": 213
},
{
"clip_ratio": 0.011397747322916985,
"epoch": 0.2993006993006993,
"grad_norm": 0.18189402035580596,
"kl": 0.7265625,
"learning_rate": 7.482517482517482e-05,
"loss": -0.0139,
"step": 214
},
{
"clip_ratio": 0.004513947293162346,
"completion_length": 94.66072082519531,
"epoch": 0.3006993006993007,
"grad_norm": 0.2675295151019955,
"kl": 0.67578125,
"learning_rate": 7.517482517482518e-05,
"loss": 0.0014,
"num_tokens": 2169243.0,
"reward": 2.2817916870117188,
"reward_std": 0.4118366539478302,
"rewards/check_gptzero_func": 0.8571428656578064,
"rewards/check_perplexity_diff_func": 0.4642857015132904,
"rewards/check_winston_local_func": 0.9603630900382996,
"step": 215
},
{
"clip_ratio": 0.010964194312691689,
"epoch": 0.3020979020979021,
"grad_norm": 0.17031623307626814,
"kl": 0.875,
"learning_rate": 7.552447552447553e-05,
"loss": -0.0057,
"step": 216
},
{
"clip_ratio": 0.003371638245880604,
"completion_length": 118.17857360839844,
"epoch": 0.3034965034965035,
"grad_norm": 0.1847364030092435,
"kl": 0.60546875,
"learning_rate": 7.587412587412587e-05,
"loss": -0.0022,
"num_tokens": 2190571.0,
"reward": 2.1225900650024414,
"reward_std": 0.5403507947921753,
"rewards/check_gptzero_func": 0.75,
"rewards/check_perplexity_diff_func": 0.4642857015132904,
"rewards/check_winston_local_func": 0.9083043336868286,
"step": 217
},
{
"clip_ratio": 0.01002263929694891,
"epoch": 0.3048951048951049,
"grad_norm": 0.1449002041728082,
"kl": 0.61328125,
"learning_rate": 7.622377622377622e-05,
"loss": -0.0067,
"step": 218
},
{
"clip_ratio": 0.0013253266224637628,
"completion_length": 129.33929443359375,
"epoch": 0.3062937062937063,
"grad_norm": 0.1504069021149342,
"kl": 0.578125,
"learning_rate": 7.657342657342658e-05,
"loss": -0.0075,
"num_tokens": 2213408.0,
"reward": 2.1974618434906006,
"reward_std": 0.4602964520454407,
"rewards/check_gptzero_func": 0.7857142686843872,
"rewards/check_perplexity_diff_func": 0.5,
"rewards/check_winston_local_func": 0.911747395992279,
"step": 219
},
{
"clip_ratio": 0.006876418832689524,
"epoch": 0.3076923076923077,
"grad_norm": 0.11992971275284438,
"kl": 0.57421875,
"learning_rate": 7.692307692307693e-05,
"loss": -0.0101,
"step": 220
},
{
"clip_ratio": 0.0024474726524204016,
"completion_length": 89.67857360839844,
"epoch": 0.3090909090909091,
"grad_norm": 0.46915411380362765,
"kl": 0.73046875,
"learning_rate": 7.727272727272727e-05,
"loss": 0.0333,
"num_tokens": 2231248.0,
"reward": 2.161350727081299,
"reward_std": 0.47136664390563965,
"rewards/check_gptzero_func": 0.8214285969734192,
"rewards/check_perplexity_diff_func": 0.4642857015132904,
"rewards/check_winston_local_func": 0.8756363987922668,
"step": 221
},
{
"clip_ratio": 0.02464432455599308,
"epoch": 0.3104895104895105,
"grad_norm": 0.30204889343007607,
"kl": 0.8046875,
"learning_rate": 7.762237762237763e-05,
"loss": 0.0211,
"step": 222
},
{
"clip_ratio": 0.0017527465242892504,
"completion_length": 134.17857360839844,
"epoch": 0.3118881118881119,
"grad_norm": 0.18948229022832336,
"kl": 0.6328125,
"learning_rate": 7.797202797202798e-05,
"loss": -0.0049,
"num_tokens": 2255162.0,
"reward": 2.4446465969085693,
"reward_std": 0.5171672105789185,
"rewards/check_gptzero_func": 0.7678571343421936,
"rewards/check_perplexity_diff_func": 0.7857142686843872,
"rewards/check_winston_local_func": 0.8910752534866333,
"step": 223
},
{
"clip_ratio": 0.006593752186745405,
"epoch": 0.3132867132867133,
"grad_norm": 0.15013602627394823,
"kl": 0.640625,
"learning_rate": 7.832167832167832e-05,
"loss": -0.0097,
"step": 224
},
{
"clip_ratio": 0.0019926591776311398,
"completion_length": 127.00000762939453,
"epoch": 0.3146853146853147,
"grad_norm": 0.22875481051435598,
"kl": 0.6875,
"learning_rate": 7.867132867132867e-05,
"loss": 0.0021,
"num_tokens": 2277320.0,
"reward": 2.2040371894836426,
"reward_std": 0.5528277158737183,
"rewards/check_gptzero_func": 0.7321428656578064,
"rewards/check_perplexity_diff_func": 0.5357142686843872,
"rewards/check_winston_local_func": 0.936180055141449,
"step": 225
},
{
"clip_ratio": 0.005839359946548939,
"epoch": 0.31608391608391606,
"grad_norm": 0.17829200080123941,
"kl": 0.68359375,
"learning_rate": 7.902097902097903e-05,
"loss": -0.0053,
"step": 226
},
{
"clip_ratio": 0.0026252593379467726,
"completion_length": 130.0357208251953,
"epoch": 0.3174825174825175,
"grad_norm": 0.16076101957107508,
"kl": 0.66015625,
"learning_rate": 7.937062937062938e-05,
"loss": -0.0033,
"num_tokens": 2299818.0,
"reward": 2.3278117179870605,
"reward_std": 0.48680880665779114,
"rewards/check_gptzero_func": 0.8035714030265808,
"rewards/check_perplexity_diff_func": 0.6428571343421936,
"rewards/check_winston_local_func": 0.8813830614089966,
"step": 227
},
{
"clip_ratio": 0.004954828415066004,
"epoch": 0.31888111888111886,
"grad_norm": 0.12734264227867195,
"kl": 0.6875,
"learning_rate": 7.972027972027972e-05,
"loss": -0.0071,
"step": 228
},
{
"clip_ratio": 0.0034341278951615095,
"completion_length": 133.7857208251953,
"epoch": 0.3202797202797203,
"grad_norm": 0.24065024754984402,
"kl": 0.6875,
"learning_rate": 8.006993006993007e-05,
"loss": -0.0052,
"num_tokens": 2322948.0,
"reward": 2.5632615089416504,
"reward_std": 0.6206622123718262,
"rewards/check_gptzero_func": 0.9642857313156128,
"rewards/check_perplexity_diff_func": 0.6785714030265808,
"rewards/check_winston_local_func": 0.9204041361808777,
"step": 229
},
{
"clip_ratio": 0.006521092262119055,
"epoch": 0.32167832167832167,
"grad_norm": 0.18602361781154253,
"kl": 0.6796875,
"learning_rate": 8.041958041958042e-05,
"loss": -0.013,
"step": 230
},
{
"clip_ratio": 0.004207184072583914,
"completion_length": 116.53572082519531,
"epoch": 0.3230769230769231,
"grad_norm": 0.36704569865090053,
"kl": 0.69921875,
"learning_rate": 8.076923076923078e-05,
"loss": 0.0055,
"num_tokens": 2344106.0,
"reward": 2.4917171001434326,
"reward_std": 0.6073458790779114,
"rewards/check_gptzero_func": 0.8392857313156128,
"rewards/check_perplexity_diff_func": 0.75,
"rewards/check_winston_local_func": 0.902431309223175,
"step": 231
},
{
"clip_ratio": 0.013317321427166462,
"epoch": 0.32447552447552447,
"grad_norm": 0.13844634287775134,
"kl": 0.73046875,
"learning_rate": 8.111888111888112e-05,
"loss": -0.0039,
"step": 232
},
{
"clip_ratio": 0.004024635534733534,
"completion_length": 134.92857360839844,
"epoch": 0.3258741258741259,
"grad_norm": 0.34168858248313216,
"kl": 0.64453125,
"learning_rate": 8.146853146853147e-05,
"loss": -0.0057,
"num_tokens": 2367724.0,
"reward": 2.20210599899292,
"reward_std": 0.5888614654541016,
"rewards/check_gptzero_func": 0.7321428656578064,
"rewards/check_perplexity_diff_func": 0.6071428656578064,
"rewards/check_winston_local_func": 0.8628200888633728,
"step": 233
},
{
"clip_ratio": 0.009708845987915993,
"epoch": 0.32727272727272727,
"grad_norm": 0.19135959991877635,
"kl": 0.67578125,
"learning_rate": 8.181818181818183e-05,
"loss": -0.0146,
"step": 234
},
{
"clip_ratio": 0.005571329966187477,
"completion_length": 98.75000762939453,
"epoch": 0.32867132867132864,
"grad_norm": 0.32328761096657005,
"kl": 0.77734375,
"learning_rate": 8.216783216783218e-05,
"loss": 0.0128,
"num_tokens": 2386636.0,
"reward": 2.2358410358428955,
"reward_std": 0.599101185798645,
"rewards/check_gptzero_func": 0.8214285969734192,
"rewards/check_perplexity_diff_func": 0.5357142686843872,
"rewards/check_winston_local_func": 0.8786982297897339,
"step": 235
},
{
"clip_ratio": 0.009436835534870625,
"epoch": 0.3300699300699301,
"grad_norm": 0.19557085227388898,
"kl": 0.8515625,
"learning_rate": 8.251748251748252e-05,
"loss": 0.0022,
"step": 236
},
{
"clip_ratio": 0.006164635997265577,
"completion_length": 79.75,
"epoch": 0.33146853146853145,
"grad_norm": 0.3250433111655094,
"kl": 0.98828125,
"learning_rate": 8.286713286713287e-05,
"loss": -0.0126,
"num_tokens": 2403400.0,
"reward": 2.1074206829071045,
"reward_std": 0.41642776131629944,
"rewards/check_gptzero_func": 0.9107142686843872,
"rewards/check_perplexity_diff_func": 0.25,
"rewards/check_winston_local_func": 0.946706235408783,
"step": 237
},
{
"clip_ratio": 0.0066314926370978355,
"epoch": 0.3328671328671329,
"grad_norm": 0.1612667131666411,
"kl": 0.98828125,
"learning_rate": 8.321678321678323e-05,
"loss": -0.0225,
"step": 238
},
{
"clip_ratio": 0.003079179208725691,
"completion_length": 117.10714721679688,
"epoch": 0.33426573426573425,
"grad_norm": 0.2505534626607256,
"kl": 0.6796875,
"learning_rate": 8.356643356643356e-05,
"loss": -0.0103,
"num_tokens": 2425354.0,
"reward": 2.3166799545288086,
"reward_std": 0.5270359516143799,
"rewards/check_gptzero_func": 0.8214285969734192,
"rewards/check_perplexity_diff_func": 0.5714285969734192,
"rewards/check_winston_local_func": 0.9238227605819702,
"step": 239
},
{
"clip_ratio": 0.006801524665206671,
"epoch": 0.3356643356643357,
"grad_norm": 0.16826419365505382,
"kl": 0.671875,
"learning_rate": 8.391608391608392e-05,
"loss": -0.0182,
"step": 240
},
{
"clip_ratio": 0.004480496048927307,
"completion_length": 125.05357360839844,
"epoch": 0.33706293706293705,
"grad_norm": 0.26807661638826913,
"kl": 0.60546875,
"learning_rate": 8.426573426573428e-05,
"loss": -0.0086,
"num_tokens": 2447333.0,
"reward": 2.2939677238464355,
"reward_std": 0.39558398723602295,
"rewards/check_gptzero_func": 0.9107142686843872,
"rewards/check_perplexity_diff_func": 0.5357142686843872,
"rewards/check_winston_local_func": 0.8475390672683716,
"step": 241
},
{
"clip_ratio": 0.007587091531604528,
"epoch": 0.3384615384615385,
"grad_norm": 0.13602673124742348,
"kl": 0.57421875,
"learning_rate": 8.461538461538461e-05,
"loss": -0.0152,
"step": 242
},
{
"clip_ratio": 0.0034923183266073465,
"completion_length": 98.71428680419922,
"epoch": 0.33986013986013985,
"grad_norm": 0.7773422467135167,
"kl": 6.5,
"learning_rate": 8.496503496503497e-05,
"loss": 0.0071,
"num_tokens": 2466435.0,
"reward": 2.4026196002960205,
"reward_std": 0.4264836013317108,
"rewards/check_gptzero_func": 0.8571428656578064,
"rewards/check_perplexity_diff_func": 0.6785714030265808,
"rewards/check_winston_local_func": 0.8669052124023438,
"step": 243
},
{
"clip_ratio": 0.026189187541604042,
"epoch": 0.3412587412587413,
"grad_norm": 130.9971047023429,
"kl": 0.70703125,
"learning_rate": 8.531468531468532e-05,
"loss": 1.8998,
"step": 244
},
{
"clip_ratio": 0.004780410788953304,
"completion_length": 135.0,
"epoch": 0.34265734265734266,
"grad_norm": 0.2329479243774761,
"kl": 0.87109375,
"learning_rate": 8.566433566433567e-05,
"loss": 0.0011,
"num_tokens": 2489651.0,
"reward": 2.2166569232940674,
"reward_std": 0.5797281265258789,
"rewards/check_gptzero_func": 0.9107142686843872,
"rewards/check_perplexity_diff_func": 0.4285714328289032,
"rewards/check_winston_local_func": 0.8773713111877441,
"step": 245
},
{
"clip_ratio": 0.021273447200655937,
"epoch": 0.34405594405594403,
"grad_norm": 0.22672502397660207,
"kl": 0.9609375,
"learning_rate": 8.601398601398601e-05,
"loss": -0.0045,
"step": 246
},
{
"clip_ratio": 0.002856873208656907,
"completion_length": 122.66072082519531,
"epoch": 0.34545454545454546,
"grad_norm": 0.19772681964626543,
"kl": 0.984375,
"learning_rate": 8.636363636363637e-05,
"loss": 0.0028,
"num_tokens": 2510506.0,
"reward": 2.323245048522949,
"reward_std": 0.46726977825164795,
"rewards/check_gptzero_func": 0.9107142686843872,
"rewards/check_perplexity_diff_func": 0.5,
"rewards/check_winston_local_func": 0.9125306010246277,
"step": 247
},
{
"clip_ratio": 0.02299562282860279,
"epoch": 0.34685314685314683,
"grad_norm": 0.1812005085005865,
"kl": 0.87890625,
"learning_rate": 8.67132867132867e-05,
"loss": -0.001,
"step": 248
},
{
"clip_ratio": 0.0026335662696510553,
"completion_length": 116.85714721679688,
"epoch": 0.34825174825174826,
"grad_norm": 0.28731547808119623,
"kl": 0.76953125,
"learning_rate": 8.706293706293707e-05,
"loss": 0.024,
"num_tokens": 2531254.0,
"reward": 2.5312695503234863,
"reward_std": 0.5851892828941345,
"rewards/check_gptzero_func": 0.8571428656578064,
"rewards/check_perplexity_diff_func": 0.8214285969734192,
"rewards/check_winston_local_func": 0.8526979088783264,
"step": 249
},
{
"clip_ratio": 0.012148472480475903,
"epoch": 0.34965034965034963,
"grad_norm": 0.21250398409595742,
"kl": 0.734375,
"learning_rate": 8.741258741258743e-05,
"loss": 0.0156,
"step": 250
},
{
"clip_ratio": 0.0036881309933960438,
"completion_length": 109.12500762939453,
"epoch": 0.35104895104895106,
"grad_norm": 0.278787459589393,
"kl": 1.3671875,
"learning_rate": 8.776223776223776e-05,
"loss": -0.0127,
"num_tokens": 2551265.0,
"reward": 2.24423885345459,
"reward_std": 0.6221095323562622,
"rewards/check_gptzero_func": 0.875,
"rewards/check_perplexity_diff_func": 0.5,
"rewards/check_winston_local_func": 0.8692389130592346,
"step": 251
},
{
"clip_ratio": 0.011438491754233837,
"epoch": 0.35244755244755244,
"grad_norm": 0.21324998809887244,
"kl": 1.0859375,
"learning_rate": 8.811188811188812e-05,
"loss": -0.023,
"step": 252
},
{
"clip_ratio": 0.0039367591962218285,
"completion_length": 112.14286041259766,
"epoch": 0.35384615384615387,
"grad_norm": 0.31069817037713615,
"kl": 0.91015625,
"learning_rate": 8.846153846153847e-05,
"loss": 0.0167,
"num_tokens": 2572119.0,
"reward": 2.492161989212036,
"reward_std": 0.6407585144042969,
"rewards/check_gptzero_func": 0.875,
"rewards/check_perplexity_diff_func": 0.75,
"rewards/check_winston_local_func": 0.8671619296073914,
"step": 253
},
{
"clip_ratio": 0.005880096461623907,
"epoch": 0.35524475524475524,
"grad_norm": 0.19008007229653096,
"kl": 0.90234375,
"learning_rate": 8.881118881118881e-05,
"loss": 0.0057,
"step": 254
},
{
"clip_ratio": 0.003151817014440894,
"completion_length": 120.78572082519531,
"epoch": 0.35664335664335667,
"grad_norm": 0.4113945542101958,
"kl": 0.890625,
"learning_rate": 8.916083916083916e-05,
"loss": 0.0079,
"num_tokens": 2593081.0,
"reward": 2.3228745460510254,
"reward_std": 0.5329480171203613,
"rewards/check_gptzero_func": 0.875,
"rewards/check_perplexity_diff_func": 0.6428571343421936,
"rewards/check_winston_local_func": 0.8050170540809631,
"step": 255
},
{
"clip_ratio": 0.01224872563034296,
"epoch": 0.35804195804195804,
"grad_norm": 0.24677625747594412,
"kl": 1.0078125,
"learning_rate": 8.951048951048952e-05,
"loss": 0.0006,
"step": 256
},
{
"clip_ratio": 0.003666934324428439,
"completion_length": 146.5178680419922,
"epoch": 0.3594405594405594,
"grad_norm": 0.21537926236506497,
"kl": 0.484375,
"learning_rate": 8.986013986013986e-05,
"loss": 0.0068,
"num_tokens": 2618758.0,
"reward": 2.198406934738159,
"reward_std": 0.3875991106033325,
"rewards/check_gptzero_func": 0.9285714030265808,
"rewards/check_perplexity_diff_func": 0.4285714328289032,
"rewards/check_winston_local_func": 0.8412641882896423,
"step": 257
},
{
"clip_ratio": 0.012731654569506645,
"epoch": 0.36083916083916084,
"grad_norm": 0.1543455831571161,
"kl": 0.478515625,
"learning_rate": 9.020979020979021e-05,
"loss": 0.0025,
"step": 258
},
{
"clip_ratio": 0.0012050624936819077,
"completion_length": 129.375,
"epoch": 0.3622377622377622,
"grad_norm": 0.16107873823673136,
"kl": 1.6640625,
"learning_rate": 9.055944055944057e-05,
"loss": 0.0054,
"num_tokens": 2641815.0,
"reward": 2.4805288314819336,
"reward_std": 0.3510296940803528,
"rewards/check_gptzero_func": 0.8392857313156128,
"rewards/check_perplexity_diff_func": 0.8214285969734192,
"rewards/check_winston_local_func": 0.8198142647743225,
"step": 259
},
{
"clip_ratio": 0.006208478473126888,
"epoch": 0.36363636363636365,
"grad_norm": 0.10546346547579268,
"kl": 0.8046875,
"learning_rate": 9.090909090909092e-05,
"loss": 0.0019,
"step": 260
},
{
"clip_ratio": 0.0025593352038413286,
"completion_length": 128.57144165039062,
"epoch": 0.365034965034965,
"grad_norm": 0.1986264685447053,
"kl": 0.6171875,
"learning_rate": 9.125874125874126e-05,
"loss": -0.0025,
"num_tokens": 2664343.0,
"reward": 2.261420249938965,
"reward_std": 0.5704914927482605,
"rewards/check_gptzero_func": 0.7857142686843872,
"rewards/check_perplexity_diff_func": 0.6428571343421936,
"rewards/check_winston_local_func": 0.8328486680984497,
"step": 261
},
{
"clip_ratio": 0.002767725382000208,
"epoch": 0.36643356643356645,
"grad_norm": 0.14267481856078115,
"kl": 0.66015625,
"learning_rate": 9.160839160839161e-05,
"loss": -0.0071,
"step": 262
},
{
"clip_ratio": 0.0036019759718328714,
"completion_length": 108.30357360839844,
"epoch": 0.3678321678321678,
"grad_norm": 0.25429899311591847,
"kl": 0.61328125,
"learning_rate": 9.195804195804196e-05,
"loss": 0.0114,
"num_tokens": 2684460.0,
"reward": 2.1853349208831787,
"reward_std": 0.5545978546142578,
"rewards/check_gptzero_func": 0.7142857313156128,
"rewards/check_perplexity_diff_func": 0.6428571343421936,
"rewards/check_winston_local_func": 0.828191876411438,
"step": 263
},
{
"clip_ratio": 0.002995865885168314,
"epoch": 0.36923076923076925,
"grad_norm": 0.17454926732818832,
"kl": 0.62109375,
"learning_rate": 9.230769230769232e-05,
"loss": 0.0028,
"step": 264
},
{
"clip_ratio": 0.002558046253398061,
"completion_length": 102.80357360839844,
"epoch": 0.3706293706293706,
"grad_norm": 0.18701691067416304,
"kl": 0.734375,
"learning_rate": 9.265734265734266e-05,
"loss": 0.0401,
"num_tokens": 2704225.0,
"reward": 2.1770143508911133,
"reward_std": 0.573693037033081,
"rewards/check_gptzero_func": 0.8035714030265808,
"rewards/check_perplexity_diff_func": 0.5357142686843872,
"rewards/check_winston_local_func": 0.8377286195755005,
"step": 265
},
{
"clip_ratio": 0.006838452070951462,
"epoch": 0.37202797202797205,
"grad_norm": 0.13942096443359253,
"kl": 0.71875,
"learning_rate": 9.300699300699301e-05,
"loss": 0.0354,
"step": 266
},
{
"clip_ratio": 0.00375689216889441,
"completion_length": 106.37500762939453,
"epoch": 0.3734265734265734,
"grad_norm": 0.25266992310982356,
"kl": 0.7421875,
"learning_rate": 9.335664335664336e-05,
"loss": -0.0156,
"num_tokens": 2724530.0,
"reward": 2.11887526512146,
"reward_std": 0.5366143584251404,
"rewards/check_gptzero_func": 0.875,
"rewards/check_perplexity_diff_func": 0.4642857015132904,
"rewards/check_winston_local_func": 0.7795897126197815,
"step": 267
},
{
"clip_ratio": 0.006593361962586641,
"epoch": 0.3748251748251748,
"grad_norm": 0.18497903526945206,
"kl": 0.7265625,
"learning_rate": 9.370629370629372e-05,
"loss": -0.0243,
"step": 268
},
{
"clip_ratio": 0.004569682292640209,
"completion_length": 124.39286041259766,
"epoch": 0.37622377622377623,
"grad_norm": 0.21331332062489702,
"kl": 0.953125,
"learning_rate": 9.405594405594406e-05,
"loss": 0.0346,
"num_tokens": 2746246.0,
"reward": 2.403602123260498,
"reward_std": 0.5621734857559204,
"rewards/check_gptzero_func": 0.7857142686843872,
"rewards/check_perplexity_diff_func": 0.75,
"rewards/check_winston_local_func": 0.8678879141807556,
"step": 269
},
{
"clip_ratio": 0.006006812676787376,
"epoch": 0.3776223776223776,
"grad_norm": 0.15420322253654764,
"kl": 0.8671875,
"learning_rate": 9.440559440559441e-05,
"loss": 0.0283,
"step": 270
},
{
"clip_ratio": 0.002397142816334963,
"completion_length": 102.41072082519531,
"epoch": 0.37902097902097903,
"grad_norm": 0.22567114655173642,
"kl": 0.609375,
"learning_rate": 9.475524475524477e-05,
"loss": 0.0231,
"num_tokens": 2765203.0,
"reward": 2.649076223373413,
"reward_std": 0.47521141171455383,
"rewards/check_gptzero_func": 0.875,
"rewards/check_perplexity_diff_func": 1.0,
"rewards/check_winston_local_func": 0.7740762829780579,
"step": 271
},
{
"clip_ratio": 0.007639365270733833,
"epoch": 0.3804195804195804,
"grad_norm": 0.18662930190408594,
"kl": 0.578125,
"learning_rate": 9.510489510489511e-05,
"loss": 0.0165,
"step": 272
},
{
"clip_ratio": 0.0010056205792352557,
"completion_length": 115.28572082519531,
"epoch": 0.38181818181818183,
"grad_norm": 0.14954423121719349,
"kl": 0.625,
"learning_rate": 9.545454545454546e-05,
"loss": 0.0237,
"num_tokens": 2785985.0,
"reward": 2.3541975021362305,
"reward_std": 0.5445494651794434,
"rewards/check_gptzero_func": 0.8392857313156128,
"rewards/check_perplexity_diff_func": 0.75,
"rewards/check_winston_local_func": 0.7649118304252625,
"step": 273
},
{
"clip_ratio": 0.003297280054539442,
"epoch": 0.3832167832167832,
"grad_norm": 0.11713806873969078,
"kl": 0.640625,
"learning_rate": 9.580419580419581e-05,
"loss": 0.0206,
"step": 274
},
{
"clip_ratio": 0.0016172031173482537,
"completion_length": 128.73214721679688,
"epoch": 0.38461538461538464,
"grad_norm": 0.155009547701029,
"kl": 0.51171875,
"learning_rate": 9.615384615384617e-05,
"loss": 0.0191,
"num_tokens": 2808414.0,
"reward": 2.376694679260254,
"reward_std": 0.33215656876564026,
"rewards/check_gptzero_func": 0.8392857313156128,
"rewards/check_perplexity_diff_func": 0.8571428656578064,
"rewards/check_winston_local_func": 0.6802659630775452,
"step": 275
},
{
"clip_ratio": 0.004294655751436949,
"epoch": 0.386013986013986,
"grad_norm": 0.11409067239204185,
"kl": 0.515625,
"learning_rate": 9.65034965034965e-05,
"loss": 0.0155,
"step": 276
},
{
"clip_ratio": 0.002925862092524767,
"completion_length": 116.41072082519531,
"epoch": 0.38741258741258744,
"grad_norm": 0.2764503095198425,
"kl": 0.578125,
"learning_rate": 9.685314685314686e-05,
"loss": 0.01,
"num_tokens": 2829285.0,
"reward": 2.3352210521698,
"reward_std": 0.7207943201065063,
"rewards/check_gptzero_func": 0.6607142686843872,
"rewards/check_perplexity_diff_func": 0.9642857313156128,
"rewards/check_winston_local_func": 0.7102211713790894,
"step": 277
},
{
"clip_ratio": 0.005557883996516466,
"epoch": 0.3888111888111888,
"grad_norm": 0.2140511595370052,
"kl": 0.5859375,
"learning_rate": 9.72027972027972e-05,
"loss": -0.0004,
"step": 278
},
{
"clip_ratio": 0.0029403052758425474,
"completion_length": 114.25000762939453,
"epoch": 0.3902097902097902,
"grad_norm": 0.21363596109825803,
"kl": 0.63671875,
"learning_rate": 9.755244755244755e-05,
"loss": 0.0259,
"num_tokens": 2849989.0,
"reward": 2.6570937633514404,
"reward_std": 0.5007120370864868,
"rewards/check_gptzero_func": 0.75,
"rewards/check_perplexity_diff_func": 1.2142857313156128,
"rewards/check_winston_local_func": 0.6928080320358276,
"step": 279
},
{
"clip_ratio": 0.00384224159643054,
"epoch": 0.3916083916083916,
"grad_norm": 0.15102706915256958,
"kl": 0.6875,
"learning_rate": 9.790209790209791e-05,
"loss": 0.0185,
"step": 280
},
{
"clip_ratio": 0.0027283646631985903,
"completion_length": 90.25000762939453,
"epoch": 0.393006993006993,
"grad_norm": 0.3585134615134873,
"kl": 0.73046875,
"learning_rate": 9.825174825174826e-05,
"loss": -0.006,
"num_tokens": 2867567.0,
"reward": 2.4528698921203613,
"reward_std": 0.6383817791938782,
"rewards/check_gptzero_func": 0.6071428656578064,
"rewards/check_perplexity_diff_func": 1.1428571939468384,
"rewards/check_winston_local_func": 0.7028695940971375,
"step": 281
},
{
"clip_ratio": 0.016295742243528366,
"epoch": 0.3944055944055944,
"grad_norm": 0.2358766583499307,
"kl": 0.73046875,
"learning_rate": 9.86013986013986e-05,
"loss": -0.0151,
"step": 282
},
{
"clip_ratio": 0.0025735762901604176,
"completion_length": 132.125,
"epoch": 0.3958041958041958,
"grad_norm": 0.12427534341566092,
"kl": 0.515625,
"learning_rate": 9.895104895104895e-05,
"loss": 0.006,
"num_tokens": 2890954.0,
"reward": 2.172079086303711,
"reward_std": 0.31105130910873413,
"rewards/check_gptzero_func": 0.6607142686843872,
"rewards/check_perplexity_diff_func": 0.7857142686843872,
"rewards/check_winston_local_func": 0.7256504893302917,
"step": 283
},
{
"clip_ratio": 0.008568070828914642,
"epoch": 0.3972027972027972,
"grad_norm": 0.10751149742872082,
"kl": 0.5234375,
"learning_rate": 9.930069930069931e-05,
"loss": 0.0038,
"step": 284
},
{
"clip_ratio": 0.0013303400482982397,
"completion_length": 115.50000762939453,
"epoch": 0.3986013986013986,
"grad_norm": 0.11155145645226237,
"kl": 0.427734375,
"learning_rate": 9.965034965034964e-05,
"loss": 0.0136,
"num_tokens": 2912320.0,
"reward": 2.499699354171753,
"reward_std": 0.22870054841041565,
"rewards/check_gptzero_func": 0.6071428656578064,
"rewards/check_perplexity_diff_func": 1.3214285373687744,
"rewards/check_winston_local_func": 0.5711276531219482,
"step": 285
},
{
"clip_ratio": 0.004690885543823242,
"epoch": 0.4,
"grad_norm": 0.08246302090980155,
"kl": 0.443359375,
"learning_rate": 0.0001,
"loss": 0.0117,
"step": 286
},
{
"clip_ratio": 0.0013457606546580791,
"completion_length": 153.60714721679688,
"epoch": 0.4013986013986014,
"grad_norm": 0.0959414180998777,
"kl": 0.458984375,
"learning_rate": 9.999996275889018e-05,
"loss": -0.0037,
"num_tokens": 2937286.0,
"reward": 2.60290789604187,
"reward_std": 0.2643657624721527,
"rewards/check_gptzero_func": 0.625,
"rewards/check_perplexity_diff_func": 1.3214285373687744,
"rewards/check_winston_local_func": 0.6564791798591614,
"step": 287
},
{
"clip_ratio": 0.0025680752005428076,
"epoch": 0.4027972027972028,
"grad_norm": 0.08295696905412306,
"kl": 0.453125,
"learning_rate": 9.999985103561615e-05,
"loss": -0.0053,
"step": 288
},
{
"clip_ratio": 0.0,
"completion_length": 106.39286041259766,
"epoch": 0.4041958041958042,
"grad_norm": 0.12524695486629228,
"kl": 0.3671875,
"learning_rate": 9.999966483034437e-05,
"loss": -0.0031,
"num_tokens": 2957572.0,
"reward": 2.607825994491577,
"reward_std": 0.2113000899553299,
"rewards/check_gptzero_func": 0.5535714030265808,
"rewards/check_perplexity_diff_func": 1.4642857313156128,
"rewards/check_winston_local_func": 0.5899689793586731,
"step": 289
},
{
"clip_ratio": 0.0015743181575089693,
"epoch": 0.40559440559440557,
"grad_norm": 0.08958540244405325,
"kl": 0.36328125,
"learning_rate": 9.999940414335222e-05,
"loss": -0.0046,
"step": 290
},
{
"clip_ratio": 0.0017556955572217703,
"completion_length": 96.0714340209961,
"epoch": 0.406993006993007,
"grad_norm": 0.19185238369967877,
"kl": 2.109375,
"learning_rate": 9.999906897502803e-05,
"loss": 0.008,
"num_tokens": 2976076.0,
"reward": 2.6148557662963867,
"reward_std": 0.24822747707366943,
"rewards/check_gptzero_func": 0.5178571343421936,
"rewards/check_perplexity_diff_func": 1.4285714626312256,
"rewards/check_winston_local_func": 0.6684269309043884,
"step": 291
},
{
"clip_ratio": 0.00337741756811738,
"epoch": 0.4083916083916084,
"grad_norm": 0.08769186584743054,
"kl": 0.671875,
"learning_rate": 9.999865932587107e-05,
"loss": 0.0046,
"step": 292
},
{
"clip_ratio": 0.0012849332997575402,
"completion_length": 91.87500762939453,
"epoch": 0.4097902097902098,
"grad_norm": 0.13036265265723118,
"kl": 0.4453125,
"learning_rate": 9.999817519649158e-05,
"loss": 0.0127,
"num_tokens": 2994773.0,
"reward": 2.552300214767456,
"reward_std": 0.14500018954277039,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 1.5,
"rewards/check_winston_local_func": 0.6237286329269409,
"step": 293
},
{
"clip_ratio": 0.0027297178748995066,
"epoch": 0.4111888111888112,
"grad_norm": 0.0976002240930887,
"kl": 0.439453125,
"learning_rate": 9.99976165876107e-05,
"loss": 0.0101,
"step": 294
},
{
"clip_ratio": 0.0,
"completion_length": 113.35714721679688,
"epoch": 0.4125874125874126,
"grad_norm": 0.02609480806956132,
"kl": 0.1591796875,
"learning_rate": 9.999698350006063e-05,
"loss": 0.0012,
"num_tokens": 3015993.0,
"reward": 2.874926805496216,
"reward_std": 0.0874238833785057,
"rewards/check_gptzero_func": 0.4821428656578064,
"rewards/check_perplexity_diff_func": 1.8928571939468384,
"rewards/check_winston_local_func": 0.4999266266822815,
"step": 295
},
{
"clip_ratio": 9.294544724980369e-05,
"epoch": 0.413986013986014,
"grad_norm": 0.02050573548257188,
"kl": 0.16015625,
"learning_rate": 9.999627593478442e-05,
"loss": 0.0011,
"step": 296
},
{
"clip_ratio": 0.00012664640962611884,
"completion_length": 98.5714340209961,
"epoch": 0.4153846153846154,
"grad_norm": 0.01680753751461254,
"kl": 0.201171875,
"learning_rate": 9.999549389283606e-05,
"loss": -0.0003,
"num_tokens": 3035155.0,
"reward": 3.0224545001983643,
"reward_std": 0.016296973451972008,
"rewards/check_gptzero_func": 0.4107142984867096,
"rewards/check_perplexity_diff_func": 1.9642857313156128,
"rewards/check_winston_local_func": 0.647454559803009,
"step": 297
},
{
"clip_ratio": 0.0,
"epoch": 0.4167832167832168,
"grad_norm": 0.012884367061070623,
"kl": 0.203125,
"learning_rate": 9.999463737538053e-05,
"loss": -0.0003,
"step": 298
},
{
"clip_ratio": 0.0003541912592481822,
"completion_length": 95.76786041259766,
"epoch": 0.41818181818181815,
"grad_norm": 0.04526732622569909,
"kl": 0.291015625,
"learning_rate": 9.999370638369377e-05,
"loss": 0.0002,
"num_tokens": 3054174.0,
"reward": 2.3303959369659424,
"reward_std": 0.08971218019723892,
"rewards/check_gptzero_func": 0.3035714328289032,
"rewards/check_perplexity_diff_func": 1.6071428060531616,
"rewards/check_winston_local_func": 0.4196813404560089,
"step": 299
},
{
"clip_ratio": 0.00011806376278400421,
"epoch": 0.4195804195804196,
"grad_norm": 0.03862986955837794,
"kl": 0.296875,
"learning_rate": 9.999270091916257e-05,
"loss": -0.0002,
"step": 300
},
{
"clip_ratio": 0.0012040403671562672,
"completion_length": 101.73214721679688,
"epoch": 0.42097902097902096,
"grad_norm": 0.12616717395678975,
"kl": 0.35546875,
"learning_rate": 9.999162098328474e-05,
"loss": 0.0033,
"num_tokens": 3073401.0,
"reward": 2.8153316974639893,
"reward_std": 0.18918544054031372,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 1.7142857313156128,
"rewards/check_winston_local_func": 0.6724745631217957,
"step": 301
},
{
"clip_ratio": 0.0013847867958247662,
"epoch": 0.4223776223776224,
"grad_norm": 0.06929991057149441,
"kl": 0.412109375,
"learning_rate": 9.999046657766903e-05,
"loss": 0.0012,
"step": 302
},
{
"clip_ratio": 0.0,
"completion_length": 81.3214340209961,
"epoch": 0.42377622377622376,
"grad_norm": 0.06730614630696258,
"kl": 0.2099609375,
"learning_rate": 9.998923770403505e-05,
"loss": -0.0019,
"num_tokens": 3090625.0,
"reward": 2.802220106124878,
"reward_std": 0.03781326860189438,
"rewards/check_gptzero_func": 0.375,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.42722010612487793,
"step": 303
},
{
"clip_ratio": 0.0,
"epoch": 0.4251748251748252,
"grad_norm": 0.022175315902565728,
"kl": 0.2109375,
"learning_rate": 9.99879343642134e-05,
"loss": -0.0024,
"step": 304
},
{
"clip_ratio": 0.00013683634460903704,
"completion_length": 105.53572082519531,
"epoch": 0.42657342657342656,
"grad_norm": 0.04867704636152166,
"kl": 0.173828125,
"learning_rate": 9.998655656014561e-05,
"loss": 0.0117,
"num_tokens": 3110877.0,
"reward": 2.6083197593688965,
"reward_std": 0.13502858579158783,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 1.8928571939468384,
"rewards/check_winston_local_func": 0.35831958055496216,
"step": 305
},
{
"clip_ratio": 0.0,
"epoch": 0.427972027972028,
"grad_norm": 0.043987709695075865,
"kl": 0.1728515625,
"learning_rate": 9.99851042938841e-05,
"loss": 0.0113,
"step": 306
},
{
"clip_ratio": 0.0004155792994424701,
"completion_length": 124.03572082519531,
"epoch": 0.42937062937062936,
"grad_norm": 0.03684760048760383,
"kl": 0.2080078125,
"learning_rate": 9.998357756759222e-05,
"loss": -0.0258,
"num_tokens": 3134097.0,
"reward": 2.7439539432525635,
"reward_std": 0.11075940728187561,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 1.7857142686843872,
"rewards/check_winston_local_func": 0.4939536452293396,
"step": 307
},
{
"clip_ratio": 0.0,
"epoch": 0.4307692307692308,
"grad_norm": 0.013648247572273129,
"kl": 0.20703125,
"learning_rate": 9.998197638354428e-05,
"loss": -0.026,
"step": 308
},
{
"clip_ratio": 0.0,
"completion_length": 89.12500762939453,
"epoch": 0.43216783216783217,
"grad_norm": 0.02465210474674629,
"kl": 0.443359375,
"learning_rate": 9.998030074412545e-05,
"loss": 0.0014,
"num_tokens": 3152358.0,
"reward": 2.420307159423828,
"reward_std": 0.11704181134700775,
"rewards/check_gptzero_func": 0.2321428507566452,
"rewards/check_perplexity_diff_func": 1.75,
"rewards/check_winston_local_func": 0.4381641745567322,
"step": 309
},
{
"clip_ratio": 0.0,
"epoch": 0.43356643356643354,
"grad_norm": 0.021184461362474653,
"kl": 0.48828125,
"learning_rate": 9.997855065183184e-05,
"loss": 0.0014,
"step": 310
},
{
"clip_ratio": 0.0,
"completion_length": 115.73214721679688,
"epoch": 0.43496503496503497,
"grad_norm": 0.0049134080725686995,
"kl": 0.162109375,
"learning_rate": 9.997672610927047e-05,
"loss": -0.0047,
"num_tokens": 3173997.0,
"reward": 2.61786150932312,
"reward_std": 0.013112460263073444,
"rewards/check_gptzero_func": 0.375,
"rewards/check_perplexity_diff_func": 1.8214285373687744,
"rewards/check_winston_local_func": 0.4214327931404114,
"step": 311
},
{
"clip_ratio": 0.0,
"epoch": 0.43636363636363634,
"grad_norm": 0.00595537475288129,
"kl": 0.173828125,
"learning_rate": 9.997482711915927e-05,
"loss": -0.0047,
"step": 312
},
{
"clip_ratio": 0.0001360544265480712,
"completion_length": 90.42857360839844,
"epoch": 0.43776223776223777,
"grad_norm": 0.007409687643468917,
"kl": 0.2275390625,
"learning_rate": 9.997285368432703e-05,
"loss": 0.0006,
"num_tokens": 3192417.0,
"reward": 2.682650566101074,
"reward_std": 0.02417343109846115,
"rewards/check_gptzero_func": 0.375,
"rewards/check_perplexity_diff_func": 1.8214285373687744,
"rewards/check_winston_local_func": 0.4862218499183655,
"step": 313
},
{
"clip_ratio": 0.0,
"epoch": 0.43916083916083914,
"grad_norm": 0.007129921917131849,
"kl": 0.2275390625,
"learning_rate": 9.997080580771349e-05,
"loss": 0.0006,
"step": 314
},
{
"clip_ratio": 0.0,
"completion_length": 83.14286041259766,
"epoch": 0.4405594405594406,
"grad_norm": 1.3606081670938062e-05,
"kl": 0.2255859375,
"learning_rate": 9.996868349236927e-05,
"loss": 0.0002,
"num_tokens": 3209875.0,
"reward": 2.671295404434204,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.45700937509536743,
"step": 315
},
{
"clip_ratio": 0.0,
"epoch": 0.44195804195804195,
"grad_norm": 1.3537787639105168e-05,
"kl": 0.2255859375,
"learning_rate": 9.996648674145583e-05,
"loss": 0.0002,
"step": 316
},
{
"clip_ratio": 0.0,
"completion_length": 126.21429443359375,
"epoch": 0.4433566433566434,
"grad_norm": 0.00991208286904361,
"kl": 0.15625,
"learning_rate": 9.996421555824558e-05,
"loss": 0.0103,
"num_tokens": 3232963.0,
"reward": 2.6654393672943115,
"reward_std": 0.0725974440574646,
"rewards/check_gptzero_func": 0.5357142686843872,
"rewards/check_perplexity_diff_func": 1.6071428060531616,
"rewards/check_winston_local_func": 0.5225821733474731,
"step": 317
},
{
"clip_ratio": 0.0,
"epoch": 0.44475524475524475,
"grad_norm": 0.010677817125212348,
"kl": 0.154296875,
"learning_rate": 9.996186994612176e-05,
"loss": 0.0102,
"step": 318
},
{
"clip_ratio": 0.0,
"completion_length": 103.4464340209961,
"epoch": 0.4461538461538462,
"grad_norm": 5.275581365835224e-05,
"kl": 0.16796875,
"learning_rate": 9.995944990857849e-05,
"loss": 0.0002,
"num_tokens": 3252898.0,
"reward": 3.0420854091644287,
"reward_std": 0.00039901022682897747,
"rewards/check_gptzero_func": 0.5714285969734192,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.5420854687690735,
"step": 319
},
{
"clip_ratio": 0.0,
"epoch": 0.44755244755244755,
"grad_norm": 6.10762506510971e-05,
"kl": 0.16796875,
"learning_rate": 9.995695544922077e-05,
"loss": 0.0002,
"step": 320
},
{
"clip_ratio": 0.0,
"completion_length": 102.28572082519531,
"epoch": 0.4489510489510489,
"grad_norm": 0.010696195320132886,
"kl": 0.1669921875,
"learning_rate": 9.995438657176448e-05,
"loss": -0.0177,
"num_tokens": 3272574.0,
"reward": 2.7592146396636963,
"reward_std": 0.03917063772678375,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 1.9642857313156128,
"rewards/check_winston_local_func": 0.437785804271698,
"step": 321
},
{
"clip_ratio": 0.0,
"epoch": 0.45034965034965035,
"grad_norm": 0.01083841197384593,
"kl": 0.1669921875,
"learning_rate": 9.995174328003631e-05,
"loss": -0.0177,
"step": 322
},
{
"clip_ratio": 8.473140769638121e-05,
"completion_length": 117.41072082519531,
"epoch": 0.45174825174825173,
"grad_norm": 0.004522641361347185,
"kl": 0.1640625,
"learning_rate": 9.994902557797382e-05,
"loss": -0.0002,
"num_tokens": 3294025.0,
"reward": 2.92221736907959,
"reward_std": 0.025547320023179054,
"rewards/check_gptzero_func": 0.5535714030265808,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.36864587664604187,
"step": 323
},
{
"clip_ratio": 0.0,
"epoch": 0.45314685314685316,
"grad_norm": 0.004737684573095642,
"kl": 0.1650390625,
"learning_rate": 9.994623346962544e-05,
"loss": -0.0002,
"step": 324
},
{
"clip_ratio": 0.0,
"completion_length": 95.26786041259766,
"epoch": 0.45454545454545453,
"grad_norm": 3.372088632491828e-06,
"kl": 0.1796875,
"learning_rate": 9.99433669591504e-05,
"loss": 0.0002,
"num_tokens": 3312902.0,
"reward": 2.595499038696289,
"reward_std": 4.388691013446078e-05,
"rewards/check_gptzero_func": 0.25,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.34549909830093384,
"step": 325
},
{
"clip_ratio": 0.0,
"epoch": 0.45594405594405596,
"grad_norm": 3.349405120445447e-06,
"kl": 0.1796875,
"learning_rate": 9.994042605081879e-05,
"loss": 0.0002,
"step": 326
},
{
"clip_ratio": 0.0,
"completion_length": 97.80357360839844,
"epoch": 0.45734265734265733,
"grad_norm": 0.010175912326793437,
"kl": 0.177734375,
"learning_rate": 9.993741074901153e-05,
"loss": 0.0003,
"num_tokens": 3331815.0,
"reward": 2.496483325958252,
"reward_std": 0.04960284009575844,
"rewards/check_gptzero_func": 0.2142857164144516,
"rewards/check_perplexity_diff_func": 1.9642857313156128,
"rewards/check_winston_local_func": 0.31791168451309204,
"step": 327
},
{
"clip_ratio": 0.0,
"epoch": 0.45874125874125876,
"grad_norm": 0.01017493586702021,
"kl": 0.177734375,
"learning_rate": 9.993432105822034e-05,
"loss": 0.0003,
"step": 328
},
{
"clip_ratio": 0.0,
"completion_length": 96.30357360839844,
"epoch": 0.46013986013986014,
"grad_norm": 0.0165654292436505,
"kl": 0.1572265625,
"learning_rate": 9.993115698304774e-05,
"loss": 0.0003,
"num_tokens": 3351210.0,
"reward": 2.8074374198913574,
"reward_std": 0.07505974173545837,
"rewards/check_gptzero_func": 0.5178571343421936,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.4324372708797455,
"step": 329
},
{
"clip_ratio": 0.0,
"epoch": 0.46153846153846156,
"grad_norm": 0.014211382809754432,
"kl": 0.16015625,
"learning_rate": 9.992791852820709e-05,
"loss": 0.0003,
"step": 330
},
{
"clip_ratio": 0.0,
"completion_length": 86.1964340209961,
"epoch": 0.46293706293706294,
"grad_norm": 0.0580909978469744,
"kl": 0.2119140625,
"learning_rate": 9.992460569852256e-05,
"loss": 0.0013,
"num_tokens": 3369025.0,
"reward": 2.986760377883911,
"reward_std": 0.10425443947315216,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.6653318405151367,
"step": 331
},
{
"clip_ratio": 0.0,
"epoch": 0.4643356643356643,
"grad_norm": 0.0012843504186470114,
"kl": 0.2109375,
"learning_rate": 9.992121849892904e-05,
"loss": 0.0012,
"step": 332
},
{
"clip_ratio": 0.0,
"completion_length": 113.42857360839844,
"epoch": 0.46573426573426574,
"grad_norm": 0.03054414776989089,
"kl": 0.1494140625,
"learning_rate": 9.99177569344723e-05,
"loss": -0.0,
"num_tokens": 3390093.0,
"reward": 2.5890934467315674,
"reward_std": 0.1023683100938797,
"rewards/check_gptzero_func": 0.25,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.41052165627479553,
"step": 333
},
{
"clip_ratio": 0.00022784233442507684,
"epoch": 0.4671328671328671,
"grad_norm": 0.006141110067038759,
"kl": 0.150390625,
"learning_rate": 9.99142210103088e-05,
"loss": -0.0001,
"step": 334
},
{
"clip_ratio": 0.0,
"completion_length": 109.00000762939453,
"epoch": 0.46853146853146854,
"grad_norm": 0.011702000814979486,
"kl": 0.158203125,
"learning_rate": 9.991061073170585e-05,
"loss": 0.0001,
"num_tokens": 3410511.0,
"reward": 2.601181745529175,
"reward_std": 0.04859553650021553,
"rewards/check_gptzero_func": 0.2857142984867096,
"rewards/check_perplexity_diff_func": 1.8928571939468384,
"rewards/check_winston_local_func": 0.42261001467704773,
"step": 335
},
{
"clip_ratio": 0.0,
"epoch": 0.4699300699300699,
"grad_norm": 0.004380253375033637,
"kl": 0.158203125,
"learning_rate": 9.990692610404145e-05,
"loss": 0.0001,
"step": 336
},
{
"clip_ratio": 0.00015908369095996022,
"completion_length": 124.37500762939453,
"epoch": 0.47132867132867134,
"grad_norm": 0.011635533606345248,
"kl": 0.1318359375,
"learning_rate": 9.99031671328044e-05,
"loss": 0.0002,
"num_tokens": 3433250.0,
"reward": 2.719287395477295,
"reward_std": 0.05448899790644646,
"rewards/check_gptzero_func": 0.5,
"rewards/check_perplexity_diff_func": 1.75,
"rewards/check_winston_local_func": 0.4692873954772949,
"step": 337
},
{
"clip_ratio": 0.0,
"epoch": 0.4727272727272727,
"grad_norm": 0.009393665715625162,
"kl": 0.1318359375,
"learning_rate": 9.989933382359422e-05,
"loss": 0.0001,
"step": 338
},
{
"clip_ratio": 0.0,
"completion_length": 110.8214340209961,
"epoch": 0.47412587412587415,
"grad_norm": 6.954019260271828e-07,
"kl": 0.15625,
"learning_rate": 9.98954261821212e-05,
"loss": 0.0002,
"num_tokens": 3454202.0,
"reward": 2.5842461585998535,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.7857142686843872,
"rewards/check_winston_local_func": 0.40567442774772644,
"step": 339
},
{
"clip_ratio": 0.0,
"epoch": 0.4755244755244755,
"grad_norm": 6.417345751201852e-07,
"kl": 0.15625,
"learning_rate": 9.98914442142063e-05,
"loss": 0.0002,
"step": 340
},
{
"clip_ratio": 0.0,
"completion_length": 76.55357360839844,
"epoch": 0.47692307692307695,
"grad_norm": 0.008322229771779063,
"kl": 0.263671875,
"learning_rate": 9.988738792578126e-05,
"loss": 0.0005,
"num_tokens": 3471041.0,
"reward": 2.916919231414795,
"reward_std": 0.05185602605342865,
"rewards/check_gptzero_func": 0.3214285671710968,
"rewards/check_perplexity_diff_func": 1.9642857313156128,
"rewards/check_winston_local_func": 0.631205141544342,
"step": 341
},
{
"clip_ratio": 0.0,
"epoch": 0.4783216783216783,
"grad_norm": 0.010116392524351476,
"kl": 0.263671875,
"learning_rate": 9.988325732288851e-05,
"loss": 0.0004,
"step": 342
},
{
"clip_ratio": 0.0,
"completion_length": 122.21429443359375,
"epoch": 0.4797202797202797,
"grad_norm": 1.4137707959903356e-05,
"kl": 0.134765625,
"learning_rate": 9.987905241168117e-05,
"loss": 0.0001,
"num_tokens": 3493483.0,
"reward": 2.7157416343688965,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.46574172377586365,
"step": 343
},
{
"clip_ratio": 0.0,
"epoch": 0.4811188811188811,
"grad_norm": 1.3729789968238485e-05,
"kl": 0.134765625,
"learning_rate": 9.987477319842307e-05,
"loss": 0.0001,
"step": 344
},
{
"clip_ratio": 0.0,
"completion_length": 93.53572082519531,
"epoch": 0.4825174825174825,
"grad_norm": 7.468121696306164e-07,
"kl": 0.1845703125,
"learning_rate": 9.987041968948869e-05,
"loss": 0.0002,
"num_tokens": 3512009.0,
"reward": 2.5505218505859375,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.1785714328289032,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.44337886571884155,
"step": 345
},
{
"clip_ratio": 0.0,
"epoch": 0.48391608391608393,
"grad_norm": 7.137709602226676e-07,
"kl": 0.1845703125,
"learning_rate": 9.986599189136325e-05,
"loss": 0.0002,
"step": 346
},
{
"clip_ratio": 0.00031959093757905066,
"completion_length": 102.58928680419922,
"epoch": 0.4853146853146853,
"grad_norm": 0.04707770778424934,
"kl": 0.1982421875,
"learning_rate": 9.986148981064258e-05,
"loss": 0.0004,
"num_tokens": 3531574.0,
"reward": 2.7341363430023193,
"reward_std": 0.0231277234852314,
"rewards/check_gptzero_func": 0.4107142984867096,
"rewards/check_perplexity_diff_func": 1.8928571939468384,
"rewards/check_winston_local_func": 0.430564820766449,
"step": 347
},
{
"clip_ratio": 0.0,
"epoch": 0.48671328671328673,
"grad_norm": 0.020919799034125892,
"kl": 0.2060546875,
"learning_rate": 9.985691345403316e-05,
"loss": 0.0002,
"step": 348
},
{
"clip_ratio": 0.0,
"completion_length": 127.67857360839844,
"epoch": 0.4881118881118881,
"grad_norm": 6.517786151476589e-05,
"kl": 0.1328125,
"learning_rate": 9.985226282835216e-05,
"loss": 0.0001,
"num_tokens": 3554440.0,
"reward": 2.926541566848755,
"reward_std": 0.0005768488626927137,
"rewards/check_gptzero_func": 0.5357142686843872,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.5336844325065613,
"step": 349
},
{
"clip_ratio": 0.0,
"epoch": 0.48951048951048953,
"grad_norm": 6.370254129427888e-05,
"kl": 0.1328125,
"learning_rate": 9.984753794052735e-05,
"loss": 0.0001,
"step": 350
},
{
"clip_ratio": 0.0,
"completion_length": 98.75000762939453,
"epoch": 0.4909090909090909,
"grad_norm": 2.085540464425244e-05,
"kl": 0.205078125,
"learning_rate": 9.984273879759713e-05,
"loss": 0.0002,
"num_tokens": 3573976.0,
"reward": 2.8204286098480225,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.4989997446537018,
"step": 351
},
{
"clip_ratio": 0.0,
"epoch": 0.49230769230769234,
"grad_norm": 2.009578048518353e-05,
"kl": 0.205078125,
"learning_rate": 9.983786540671051e-05,
"loss": 0.0002,
"step": 352
},
{
"clip_ratio": 0.0,
"completion_length": 85.21428680419922,
"epoch": 0.4937062937062937,
"grad_norm": 2.1499404490472988e-06,
"kl": 0.1826171875,
"learning_rate": 9.983291777512711e-05,
"loss": 0.0002,
"num_tokens": 3591644.0,
"reward": 2.8764142990112305,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.44784265756607056,
"step": 353
},
{
"clip_ratio": 0.0,
"epoch": 0.4951048951048951,
"grad_norm": 2.160124410750654e-06,
"kl": 0.1826171875,
"learning_rate": 9.982789591021715e-05,
"loss": 0.0002,
"step": 354
},
{
"clip_ratio": 0.0,
"completion_length": 112.00000762939453,
"epoch": 0.4965034965034965,
"grad_norm": 0.00014814082463704463,
"kl": 0.1669921875,
"learning_rate": 9.982279981946143e-05,
"loss": 0.0002,
"num_tokens": 3612596.0,
"reward": 2.8398191928863525,
"reward_std": 0.0006443771999329329,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.5183902382850647,
"step": 355
},
{
"clip_ratio": 0.0,
"epoch": 0.4979020979020979,
"grad_norm": 0.00014606763289714295,
"kl": 0.1669921875,
"learning_rate": 9.98176295104513e-05,
"loss": 0.0002,
"step": 356
},
{
"clip_ratio": 0.0,
"completion_length": 88.50000762939453,
"epoch": 0.4993006993006993,
"grad_norm": 3.33550016470454e-07,
"kl": 0.1669921875,
"learning_rate": 9.98123849908887e-05,
"loss": 0.0002,
"num_tokens": 3630862.0,
"reward": 2.655492067337036,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3214285671710968,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.47692039608955383,
"step": 357
},
{
"clip_ratio": 0.0,
"epoch": 0.5006993006993007,
"grad_norm": 3.3462192763241657e-07,
"kl": 0.1669921875,
"learning_rate": 9.980706626858607e-05,
"loss": 0.0002,
"step": 358
},
{
"clip_ratio": 0.0,
"completion_length": 126.39286041259766,
"epoch": 0.5020979020979021,
"grad_norm": 3.4860786526471938e-06,
"kl": 0.1416015625,
"learning_rate": 9.980167335146642e-05,
"loss": 0.0001,
"num_tokens": 3653738.0,
"reward": 2.765826463699341,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.5,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.4086834490299225,
"step": 359
},
{
"clip_ratio": 0.0,
"epoch": 0.5034965034965035,
"grad_norm": 3.71619859471226e-06,
"kl": 0.1416015625,
"learning_rate": 9.979620624756329e-05,
"loss": 0.0001,
"step": 360
},
{
"clip_ratio": 0.0,
"completion_length": 129.92857360839844,
"epoch": 0.5048951048951049,
"grad_norm": 8.195642286342138e-07,
"kl": 0.1376953125,
"learning_rate": 9.979066496502074e-05,
"loss": 0.0001,
"num_tokens": 3677274.0,
"reward": 2.8696835041046143,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.5357142686843872,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.4053979218006134,
"step": 361
},
{
"clip_ratio": 0.0,
"epoch": 0.5062937062937063,
"grad_norm": 8.760444459307852e-07,
"kl": 0.1376953125,
"learning_rate": 9.978504951209327e-05,
"loss": 0.0001,
"step": 362
},
{
"clip_ratio": 0.0,
"completion_length": 118.0714340209961,
"epoch": 0.5076923076923077,
"grad_norm": 1.595860947799347e-06,
"kl": 0.162109375,
"learning_rate": 9.977935989714595e-05,
"loss": 0.0002,
"num_tokens": 3698896.0,
"reward": 2.6664552688598633,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.2857142984867096,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.45216941833496094,
"step": 363
},
{
"clip_ratio": 0.0,
"epoch": 0.509090909090909,
"grad_norm": 1.6260194368178069e-06,
"kl": 0.162109375,
"learning_rate": 9.977359612865423e-05,
"loss": 0.0002,
"step": 364
},
{
"clip_ratio": 0.0,
"completion_length": 85.96428680419922,
"epoch": 0.5104895104895105,
"grad_norm": 4.357301744161908e-06,
"kl": 0.2138671875,
"learning_rate": 9.976775821520412e-05,
"loss": 0.0002,
"num_tokens": 3716536.0,
"reward": 2.8877956867218018,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.566366970539093,
"step": 365
},
{
"clip_ratio": 0.0,
"epoch": 0.5118881118881119,
"grad_norm": 4.353902612581284e-06,
"kl": 0.2138671875,
"learning_rate": 9.976184616549203e-05,
"loss": 0.0002,
"step": 366
},
{
"clip_ratio": 0.0,
"completion_length": 116.96429443359375,
"epoch": 0.5132867132867133,
"grad_norm": 1.5148419096170459e-06,
"kl": 0.146484375,
"learning_rate": 9.97558599883248e-05,
"loss": 0.0001,
"num_tokens": 3737950.0,
"reward": 2.687312602996826,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.36588403582572937,
"step": 367
},
{
"clip_ratio": 0.0,
"epoch": 0.5146853146853146,
"grad_norm": 1.4470246412763422e-06,
"kl": 0.146484375,
"learning_rate": 9.97497996926197e-05,
"loss": 0.0001,
"step": 368
},
{
"clip_ratio": 0.0,
"completion_length": 94.83928680419922,
"epoch": 0.5160839160839161,
"grad_norm": 0.0004647591690002515,
"kl": 0.2138671875,
"learning_rate": 9.974366528740441e-05,
"loss": 0.0002,
"num_tokens": 3756935.0,
"reward": 2.634523391723633,
"reward_std": 0.0005499019753187895,
"rewards/check_gptzero_func": 0.2857142984867096,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.4202377200126648,
"step": 369
},
{
"clip_ratio": 0.0,
"epoch": 0.5174825174825175,
"grad_norm": 0.0004885464229234562,
"kl": 0.2138671875,
"learning_rate": 9.973745678181705e-05,
"loss": 0.0002,
"step": 370
},
{
"clip_ratio": 0.0,
"completion_length": 100.17857360839844,
"epoch": 0.5188811188811189,
"grad_norm": 3.0418176986251483e-06,
"kl": 0.1640625,
"learning_rate": 9.973117418510605e-05,
"loss": 0.0002,
"num_tokens": 3776087.0,
"reward": 2.5216457843780518,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.2857142984867096,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.3787885308265686,
"step": 371
},
{
"clip_ratio": 0.0,
"epoch": 0.5202797202797202,
"grad_norm": 3.0723635480561103e-06,
"kl": 0.1640625,
"learning_rate": 9.972481750663026e-05,
"loss": 0.0002,
"step": 372
},
{
"clip_ratio": 0.0,
"completion_length": 89.66072082519531,
"epoch": 0.5216783216783217,
"grad_norm": 0.008590656277289149,
"kl": 0.40234375,
"learning_rate": 9.971838675585888e-05,
"loss": 0.0007,
"num_tokens": 3794550.0,
"reward": 2.5267839431762695,
"reward_std": 0.05061079189181328,
"rewards/check_gptzero_func": 0.3214285671710968,
"rewards/check_perplexity_diff_func": 1.8214285373687744,
"rewards/check_winston_local_func": 0.383926659822464,
"step": 373
},
{
"clip_ratio": 0.0,
"epoch": 0.5230769230769231,
"grad_norm": 0.003770576536622864,
"kl": 0.4296875,
"learning_rate": 9.97118819423714e-05,
"loss": 0.0007,
"step": 374
},
{
"clip_ratio": 0.0,
"completion_length": 80.17857360839844,
"epoch": 0.5244755244755245,
"grad_norm": 3.560813902520289e-07,
"kl": 0.2216796875,
"learning_rate": 9.970530307585773e-05,
"loss": 0.0002,
"num_tokens": 3811756.0,
"reward": 2.883481740951538,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.5263389348983765,
"step": 375
},
{
"clip_ratio": 0.0,
"epoch": 0.5258741258741259,
"grad_norm": 3.5463275242856317e-07,
"kl": 0.2216796875,
"learning_rate": 9.9698650166118e-05,
"loss": 0.0002,
"step": 376
},
{
"clip_ratio": 0.0,
"completion_length": 86.75000762939453,
"epoch": 0.5272727272727272,
"grad_norm": 4.776682547251814e-07,
"kl": 0.205078125,
"learning_rate": 9.969192322306271e-05,
"loss": 0.0002,
"num_tokens": 3829616.0,
"reward": 2.9529268741607666,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.5600695610046387,
"step": 377
},
{
"clip_ratio": 0.0,
"epoch": 0.5286713286713287,
"grad_norm": 4.804658823710114e-07,
"kl": 0.205078125,
"learning_rate": 9.96851222567126e-05,
"loss": 0.0002,
"step": 378
},
{
"clip_ratio": 0.0,
"completion_length": 117.89286041259766,
"epoch": 0.5300699300699301,
"grad_norm": 9.143102926453055e-06,
"kl": 0.158203125,
"learning_rate": 9.96782472771987e-05,
"loss": 0.0002,
"num_tokens": 3851170.0,
"reward": 2.7617037296295166,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.40456071496009827,
"step": 379
},
{
"clip_ratio": 0.0,
"epoch": 0.5314685314685315,
"grad_norm": 9.41817239216292e-06,
"kl": 0.158203125,
"learning_rate": 9.967129829476228e-05,
"loss": 0.0002,
"step": 380
},
{
"clip_ratio": 0.000286861730273813,
"completion_length": 80.8214340209961,
"epoch": 0.5328671328671328,
"grad_norm": 0.03337399461775748,
"kl": 0.224609375,
"learning_rate": 9.966427531975483e-05,
"loss": -0.0402,
"num_tokens": 3868508.0,
"reward": 2.820319652557373,
"reward_std": 0.0946396142244339,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.5346053838729858,
"step": 381
},
{
"clip_ratio": 0.0,
"epoch": 0.5342657342657343,
"grad_norm": 0.00034351439093039966,
"kl": 0.224609375,
"learning_rate": 9.965717836263812e-05,
"loss": -0.0403,
"step": 382
},
{
"clip_ratio": 0.0,
"completion_length": 95.03572082519531,
"epoch": 0.5356643356643357,
"grad_norm": 3.3945579906577137e-07,
"kl": 0.216796875,
"learning_rate": 9.965000743398408e-05,
"loss": 0.0002,
"num_tokens": 3887044.0,
"reward": 2.8529021739959717,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.4243304431438446,
"step": 383
},
{
"clip_ratio": 0.0,
"epoch": 0.5370629370629371,
"grad_norm": 2.8390305863820515e-07,
"kl": 0.216796875,
"learning_rate": 9.964276254447484e-05,
"loss": 0.0002,
"step": 384
},
{
"clip_ratio": 0.0,
"completion_length": 140.1428680419922,
"epoch": 0.5384615384615384,
"grad_norm": 8.521451122880038e-07,
"kl": 0.134765625,
"learning_rate": 9.96354437049027e-05,
"loss": 0.0001,
"num_tokens": 3911264.0,
"reward": 2.4770619869232178,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.2857142984867096,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.33420488238334656,
"step": 385
},
{
"clip_ratio": 0.0,
"epoch": 0.5398601398601398,
"grad_norm": 6.753306754213034e-07,
"kl": 0.134765625,
"learning_rate": 9.962805092617016e-05,
"loss": 0.0001,
"step": 386
},
{
"clip_ratio": 0.0,
"completion_length": 97.64286041259766,
"epoch": 0.5412587412587413,
"grad_norm": 0.08784199050747857,
"kl": 2.28125,
"learning_rate": 9.962058421928979e-05,
"loss": 0.0023,
"num_tokens": 3930478.0,
"reward": 2.897810459136963,
"reward_std": 0.002231778111308813,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.4692386984825134,
"step": 387
},
{
"clip_ratio": 0.0,
"epoch": 0.5426573426573427,
"grad_norm": 0.03620083942117924,
"kl": 0.98828125,
"learning_rate": 9.961304359538437e-05,
"loss": 0.001,
"step": 388
},
{
"clip_ratio": 0.0,
"completion_length": 102.21428680419922,
"epoch": 0.544055944055944,
"grad_norm": 0.025447754609912356,
"kl": 0.1806640625,
"learning_rate": 9.96054290656867e-05,
"loss": 0.0002,
"num_tokens": 3950362.0,
"reward": 2.8110532760620117,
"reward_std": 0.04895726963877678,
"rewards/check_gptzero_func": 0.5357142686843872,
"rewards/check_perplexity_diff_func": 1.8214285373687744,
"rewards/check_winston_local_func": 0.45391014218330383,
"step": 389
},
{
"clip_ratio": 0.0,
"epoch": 0.5454545454545454,
"grad_norm": 5.1964811610508515e-05,
"kl": 0.1806640625,
"learning_rate": 9.959774064153977e-05,
"loss": 0.0001,
"step": 390
},
{
"clip_ratio": 0.0,
"completion_length": 103.5714340209961,
"epoch": 0.5468531468531469,
"grad_norm": 3.816406536402076e-07,
"kl": 0.173828125,
"learning_rate": 9.958997833439657e-05,
"loss": 0.0002,
"num_tokens": 3970164.0,
"reward": 2.5441431999206543,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3214285671710968,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.3655718266963959,
"step": 391
},
{
"clip_ratio": 0.0,
"epoch": 0.5482517482517483,
"grad_norm": 4.199010025491445e-07,
"kl": 0.173828125,
"learning_rate": 9.958214215582018e-05,
"loss": 0.0002,
"step": 392
},
{
"clip_ratio": 0.0,
"completion_length": 126.8214340209961,
"epoch": 0.5496503496503496,
"grad_norm": 6.309179959293235e-06,
"kl": 0.1533203125,
"learning_rate": 9.957423211748374e-05,
"loss": 0.0002,
"num_tokens": 3992994.0,
"reward": 2.7432734966278076,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.42184481024742126,
"step": 393
},
{
"clip_ratio": 0.0,
"epoch": 0.551048951048951,
"grad_norm": 5.547540566387994e-06,
"kl": 0.1533203125,
"learning_rate": 9.956624823117036e-05,
"loss": 0.0002,
"step": 394
},
{
"clip_ratio": 0.0,
"completion_length": 125.62500762939453,
"epoch": 0.5524475524475524,
"grad_norm": 0.007987942179452279,
"kl": 0.142578125,
"learning_rate": 9.955819050877321e-05,
"loss": 0.0001,
"num_tokens": 4015669.0,
"reward": 3.091567039489746,
"reward_std": 0.018998777493834496,
"rewards/check_gptzero_func": 0.5535714030265808,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.6094242334365845,
"step": 395
},
{
"clip_ratio": 0.0,
"epoch": 0.5538461538461539,
"grad_norm": 0.0034758785486018436,
"kl": 0.142578125,
"learning_rate": 9.955005896229543e-05,
"loss": 0.0,
"step": 396
},
{
"clip_ratio": 0.0,
"completion_length": 96.75000762939453,
"epoch": 0.5552447552447553,
"grad_norm": 0.011495641299277456,
"kl": 0.283203125,
"learning_rate": 9.954185360385013e-05,
"loss": 0.0006,
"num_tokens": 4035041.0,
"reward": 2.950004816055298,
"reward_std": 0.01496803853660822,
"rewards/check_gptzero_func": 0.4821428656578064,
"rewards/check_perplexity_diff_func": 1.8928571939468384,
"rewards/check_winston_local_func": 0.5750047564506531,
"step": 397
},
{
"clip_ratio": 0.0,
"epoch": 0.5566433566433566,
"grad_norm": 0.010082500068390739,
"kl": 0.283203125,
"learning_rate": 9.953357444566039e-05,
"loss": 0.0006,
"step": 398
},
{
"clip_ratio": 0.0,
"completion_length": 110.8214340209961,
"epoch": 0.558041958041958,
"grad_norm": 8.489376613416e-06,
"kl": 0.1787109375,
"learning_rate": 9.952522150005919e-05,
"loss": 0.0002,
"num_tokens": 4055855.0,
"reward": 2.913240909576416,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.5203836560249329,
"step": 399
},
{
"clip_ratio": 0.0,
"epoch": 0.5594405594405595,
"grad_norm": 8.766485288799225e-06,
"kl": 0.1787109375,
"learning_rate": 9.951679477948947e-05,
"loss": 0.0002,
"step": 400
},
{
"clip_ratio": 0.0,
"completion_length": 105.46428680419922,
"epoch": 0.5608391608391609,
"grad_norm": 8.652849000461684e-06,
"kl": 0.150390625,
"learning_rate": 9.9508294296504e-05,
"loss": 0.0002,
"num_tokens": 4076051.0,
"reward": 2.871354818344116,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.44278329610824585,
"step": 401
},
{
"clip_ratio": 0.0,
"epoch": 0.5622377622377622,
"grad_norm": 8.18246821524729e-06,
"kl": 0.150390625,
"learning_rate": 9.949972006376556e-05,
"loss": 0.0002,
"step": 402
},
{
"clip_ratio": 0.0,
"completion_length": 75.25,
"epoch": 0.5636363636363636,
"grad_norm": 7.351342876079361e-07,
"kl": 0.1953125,
"learning_rate": 9.949107209404665e-05,
"loss": 0.0002,
"num_tokens": 4092527.0,
"reward": 2.9322519302368164,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.5751089453697205,
"step": 403
},
{
"clip_ratio": 0.0,
"epoch": 0.5650349650349651,
"grad_norm": 7.136344464618312e-07,
"kl": 0.1953125,
"learning_rate": 9.948235040022966e-05,
"loss": 0.0002,
"step": 404
},
{
"clip_ratio": 0.0,
"completion_length": 108.08928680419922,
"epoch": 0.5664335664335665,
"grad_norm": 0.007956878335048736,
"kl": 0.169921875,
"learning_rate": 9.947355499530683e-05,
"loss": 0.0049,
"num_tokens": 4113282.0,
"reward": 3.1798436641693115,
"reward_std": 0.06495096534490585,
"rewards/check_gptzero_func": 0.6785714030265808,
"rewards/check_perplexity_diff_func": 1.9642857313156128,
"rewards/check_winston_local_func": 0.5369864702224731,
"step": 405
},
{
"clip_ratio": 0.0,
"epoch": 0.5678321678321678,
"grad_norm": 0.006640329404323574,
"kl": 0.1728515625,
"learning_rate": 9.946468589238021e-05,
"loss": 0.0049,
"step": 406
},
{
"clip_ratio": 0.0,
"completion_length": 121.39286041259766,
"epoch": 0.5692307692307692,
"grad_norm": 1.5620797981026132e-06,
"kl": 0.1396484375,
"learning_rate": 9.945574310466159e-05,
"loss": 0.0001,
"num_tokens": 4135506.0,
"reward": 2.573127269744873,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.2857142984867096,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.3588416278362274,
"step": 407
},
{
"clip_ratio": 0.0,
"epoch": 0.5706293706293706,
"grad_norm": 1.51593967204227e-06,
"kl": 0.1396484375,
"learning_rate": 9.944672664547252e-05,
"loss": 0.0001,
"step": 408
},
{
"clip_ratio": 0.0,
"completion_length": 107.8214340209961,
"epoch": 0.5720279720279721,
"grad_norm": 0.0007775272826913423,
"kl": 0.17578125,
"learning_rate": 9.943763652824436e-05,
"loss": 0.0002,
"num_tokens": 4156060.0,
"reward": 2.55350661277771,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 1.7142857313156128,
"rewards/check_winston_local_func": 0.48207801580429077,
"step": 409
},
{
"clip_ratio": 0.0,
"epoch": 0.5734265734265734,
"grad_norm": 0.0008914625567854122,
"kl": 0.1767578125,
"learning_rate": 9.942847276651811e-05,
"loss": 0.0002,
"step": 410
},
{
"clip_ratio": 0.0,
"completion_length": 101.62500762939453,
"epoch": 0.5748251748251748,
"grad_norm": 4.9539161428254635e-05,
"kl": 0.1474609375,
"learning_rate": 9.941923537394456e-05,
"loss": 0.0001,
"num_tokens": 4176019.0,
"reward": 2.8906424045562744,
"reward_std": 0.00012528452498372644,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.5334992408752441,
"step": 411
},
{
"clip_ratio": 0.0,
"epoch": 0.5762237762237762,
"grad_norm": 5.844422779033556e-05,
"kl": 0.1474609375,
"learning_rate": 9.940992436428409e-05,
"loss": 0.0001,
"step": 412
},
{
"clip_ratio": 0.0,
"completion_length": 105.28572082519531,
"epoch": 0.5776223776223777,
"grad_norm": 0.0006008878616346907,
"kl": 0.2216796875,
"learning_rate": 9.940053975140684e-05,
"loss": 0.0002,
"num_tokens": 4196667.0,
"reward": 2.943253755569458,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.5357142686843872,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.5503963828086853,
"step": 413
},
{
"clip_ratio": 0.0,
"epoch": 0.579020979020979,
"grad_norm": 0.0003365492175069851,
"kl": 0.220703125,
"learning_rate": 9.939108154929252e-05,
"loss": 0.0002,
"step": 414
},
{
"clip_ratio": 0.0,
"completion_length": 114.60714721679688,
"epoch": 0.5804195804195804,
"grad_norm": 0.008680207230787525,
"kl": 0.171875,
"learning_rate": 9.938154977203049e-05,
"loss": -0.0005,
"num_tokens": 4217727.0,
"reward": 2.7773444652557373,
"reward_std": 0.051937274634838104,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 1.8214285373687744,
"rewards/check_winston_local_func": 0.5273441076278687,
"step": 415
},
{
"clip_ratio": 0.0,
"epoch": 0.5818181818181818,
"grad_norm": 0.00033864647249595475,
"kl": 0.171875,
"learning_rate": 9.937194443381972e-05,
"loss": -0.0005,
"step": 416
},
{
"clip_ratio": 0.0,
"completion_length": 100.60714721679688,
"epoch": 0.5832167832167832,
"grad_norm": 4.017405087628847e-06,
"kl": 0.193359375,
"learning_rate": 9.936226554896875e-05,
"loss": 0.0002,
"num_tokens": 4237189.0,
"reward": 2.7684807777404785,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.44705215096473694,
"step": 417
},
{
"clip_ratio": 0.0,
"epoch": 0.5846153846153846,
"grad_norm": 3.880139616189675e-06,
"kl": 0.193359375,
"learning_rate": 9.935251313189564e-05,
"loss": 0.0002,
"step": 418
},
{
"clip_ratio": 0.0,
"completion_length": 107.75000762939453,
"epoch": 0.586013986013986,
"grad_norm": 0.00032680437267053027,
"kl": 0.1826171875,
"learning_rate": 9.934268719712807e-05,
"loss": 0.0002,
"num_tokens": 4258213.0,
"reward": 2.577641487121582,
"reward_std": 0.0015505586052313447,
"rewards/check_gptzero_func": 0.2857142984867096,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.4347843825817108,
"step": 419
},
{
"clip_ratio": 0.0,
"epoch": 0.5874125874125874,
"grad_norm": 0.0003238465648274268,
"kl": 0.1826171875,
"learning_rate": 9.933278775930317e-05,
"loss": 0.0002,
"step": 420
},
{
"clip_ratio": 0.0,
"completion_length": 108.67857360839844,
"epoch": 0.5888111888111888,
"grad_norm": 0.015591092439371165,
"kl": 0.16015625,
"learning_rate": 9.932281483316758e-05,
"loss": 0.0004,
"num_tokens": 4278847.0,
"reward": 2.615447998046875,
"reward_std": 0.05428782477974892,
"rewards/check_gptzero_func": 0.3214285671710968,
"rewards/check_perplexity_diff_func": 1.8214285373687744,
"rewards/check_winston_local_func": 0.47259071469306946,
"step": 421
},
{
"clip_ratio": 0.0,
"epoch": 0.5902097902097903,
"grad_norm": 0.00016911180587449477,
"kl": 0.16015625,
"learning_rate": 9.931276843357742e-05,
"loss": 0.0003,
"step": 422
},
{
"clip_ratio": 0.0,
"completion_length": 111.16072082519531,
"epoch": 0.5916083916083916,
"grad_norm": 1.6847767080478394e-05,
"kl": 0.1630859375,
"learning_rate": 9.930264857549825e-05,
"loss": 0.0002,
"num_tokens": 4300096.0,
"reward": 2.857067823410034,
"reward_std": 0.00101040443405509,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.6070676445960999,
"step": 423
},
{
"clip_ratio": 0.0,
"epoch": 0.593006993006993,
"grad_norm": 2.453025078669727e-05,
"kl": 0.1630859375,
"learning_rate": 9.929245527400503e-05,
"loss": 0.0002,
"step": 424
},
{
"clip_ratio": 0.0,
"completion_length": 102.50000762939453,
"epoch": 0.5944055944055944,
"grad_norm": 0.0015183266942611472,
"kl": 0.173828125,
"learning_rate": 9.928218854428221e-05,
"loss": 0.0002,
"num_tokens": 4320064.0,
"reward": 2.6430578231811523,
"reward_std": 5.020291791879572e-05,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 1.7857142686843872,
"rewards/check_winston_local_func": 0.5002006888389587,
"step": 425
},
{
"clip_ratio": 0.0,
"epoch": 0.5958041958041959,
"grad_norm": 0.0013596061771866538,
"kl": 0.1708984375,
"learning_rate": 9.927184840162354e-05,
"loss": 0.0002,
"step": 426
},
{
"clip_ratio": 0.0,
"completion_length": 110.66072082519531,
"epoch": 0.5972027972027972,
"grad_norm": 0.12076121626694686,
"kl": 0.189453125,
"learning_rate": 9.926143486143214e-05,
"loss": -0.001,
"num_tokens": 4341145.0,
"reward": 2.6579177379608154,
"reward_std": 0.05154726281762123,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 1.8214285373687744,
"rewards/check_winston_local_func": 0.4079175889492035,
"step": 427
},
{
"clip_ratio": 0.0,
"epoch": 0.5986013986013986,
"grad_norm": 0.0008289218843548021,
"kl": 0.19140625,
"learning_rate": 9.92509479392205e-05,
"loss": -0.0013,
"step": 428
},
{
"clip_ratio": 0.0,
"completion_length": 89.67857360839844,
"epoch": 0.6,
"grad_norm": 0.005379412774279031,
"kl": 0.1787109375,
"learning_rate": 9.924038765061042e-05,
"loss": 0.0001,
"num_tokens": 4359277.0,
"reward": 2.633777141571045,
"reward_std": 0.05039219558238983,
"rewards/check_gptzero_func": 0.2857142984867096,
"rewards/check_perplexity_diff_func": 1.9642857313156128,
"rewards/check_winston_local_func": 0.38377735018730164,
"step": 429
},
{
"clip_ratio": 0.0,
"epoch": 0.6013986013986014,
"grad_norm": 0.010439668945298157,
"kl": 0.1787109375,
"learning_rate": 9.922975401133293e-05,
"loss": 0.0002,
"step": 430
},
{
"clip_ratio": 0.0,
"completion_length": 102.67857360839844,
"epoch": 0.6027972027972028,
"grad_norm": 0.06317955454404899,
"kl": 0.201171875,
"learning_rate": 9.92190470372284e-05,
"loss": 0.0068,
"num_tokens": 4379109.0,
"reward": 2.6650938987731934,
"reward_std": 0.05930115655064583,
"rewards/check_gptzero_func": 0.25,
"rewards/check_perplexity_diff_func": 1.9642857313156128,
"rewards/check_winston_local_func": 0.4508081078529358,
"step": 431
},
{
"clip_ratio": 0.0002587991766631603,
"epoch": 0.6041958041958042,
"grad_norm": 0.05122842630547611,
"kl": 0.1865234375,
"learning_rate": 9.920826674424642e-05,
"loss": 0.0063,
"step": 432
},
{
"clip_ratio": 0.0,
"completion_length": 80.35714721679688,
"epoch": 0.6055944055944056,
"grad_norm": 7.162658137250354e-07,
"kl": 0.1875,
"learning_rate": 9.919741314844577e-05,
"loss": 0.0002,
"num_tokens": 4396693.0,
"reward": 3.0637807846069336,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.5357142686843872,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.6709235310554504,
"step": 433
},
{
"clip_ratio": 0.0,
"epoch": 0.606993006993007,
"grad_norm": 8.451796450955314e-07,
"kl": 0.1875,
"learning_rate": 9.918648626599447e-05,
"loss": 0.0002,
"step": 434
},
{
"clip_ratio": 0.0,
"completion_length": 124.98214721679688,
"epoch": 0.6083916083916084,
"grad_norm": 0.016099687268680437,
"kl": 0.1455078125,
"learning_rate": 9.91754861131697e-05,
"loss": -0.0001,
"num_tokens": 4418950.0,
"reward": 2.94807505607605,
"reward_std": 0.050821539014577866,
"rewards/check_gptzero_func": 0.5357142686843872,
"rewards/check_perplexity_diff_func": 1.8928571939468384,
"rewards/check_winston_local_func": 0.5195035338401794,
"step": 435
},
{
"clip_ratio": 0.0,
"epoch": 0.6097902097902098,
"grad_norm": 0.014390989758817311,
"kl": 0.1455078125,
"learning_rate": 9.916441270635772e-05,
"loss": -0.0001,
"step": 436
},
{
"clip_ratio": 0.0,
"completion_length": 106.64286041259766,
"epoch": 0.6111888111888112,
"grad_norm": 1.968332582426962e-06,
"kl": 0.14453125,
"learning_rate": 9.915326606205404e-05,
"loss": 0.0001,
"num_tokens": 4439018.0,
"reward": 2.77024507522583,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.4131018817424774,
"step": 437
},
{
"clip_ratio": 0.0,
"epoch": 0.6125874125874126,
"grad_norm": 1.8961430119120163e-06,
"kl": 0.14453125,
"learning_rate": 9.914204619686314e-05,
"loss": 0.0001,
"step": 438
},
{
"clip_ratio": 0.0,
"completion_length": 102.21428680419922,
"epoch": 0.6139860139860139,
"grad_norm": 0.000668351376631818,
"kl": 0.1728515625,
"learning_rate": 9.913075312749866e-05,
"loss": 0.0002,
"num_tokens": 4458806.0,
"reward": 2.8182172775268555,
"reward_std": 0.0025242711417376995,
"rewards/check_gptzero_func": 0.5357142686843872,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.42535990476608276,
"step": 439
},
{
"clip_ratio": 0.0,
"epoch": 0.6153846153846154,
"grad_norm": 0.0006751756209535994,
"kl": 0.1728515625,
"learning_rate": 9.911938687078324e-05,
"loss": 0.0002,
"step": 440
},
{
"clip_ratio": 0.0,
"completion_length": 81.25,
"epoch": 0.6167832167832168,
"grad_norm": 0.007904133813646444,
"kl": 0.240234375,
"learning_rate": 9.910794744364857e-05,
"loss": -0.0001,
"num_tokens": 4475982.0,
"reward": 2.8269970417022705,
"reward_std": 0.05014092102646828,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.8928571939468384,
"rewards/check_winston_local_func": 0.5412828326225281,
"step": 441
},
{
"clip_ratio": 0.0,
"epoch": 0.6181818181818182,
"grad_norm": 0.00827597712943806,
"kl": 0.240234375,
"learning_rate": 9.909643486313533e-05,
"loss": -0.0001,
"step": 442
},
{
"clip_ratio": 0.0,
"completion_length": 113.00000762939453,
"epoch": 0.6195804195804195,
"grad_norm": 6.016032313066646e-06,
"kl": 0.16796875,
"learning_rate": 9.908484914639318e-05,
"loss": 0.0002,
"num_tokens": 4497200.0,
"reward": 2.9243876934051514,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.5315303206443787,
"step": 443
},
{
"clip_ratio": 0.0,
"epoch": 0.620979020979021,
"grad_norm": 7.198804696892823e-06,
"kl": 0.16796875,
"learning_rate": 9.90731903106807e-05,
"loss": 0.0002,
"step": 444
},
{
"clip_ratio": 0.0,
"completion_length": 129.375,
"epoch": 0.6223776223776224,
"grad_norm": 0.02075011817403209,
"kl": 0.158203125,
"learning_rate": 9.90614583733654e-05,
"loss": -0.0001,
"num_tokens": 4519807.0,
"reward": 2.866011619567871,
"reward_std": 0.10101933032274246,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.47315436601638794,
"step": 445
},
{
"clip_ratio": 0.0,
"epoch": 0.6237762237762238,
"grad_norm": 0.0034619333956659956,
"kl": 0.1591796875,
"learning_rate": 9.904965335192373e-05,
"loss": -0.0002,
"step": 446
},
{
"clip_ratio": 0.0,
"completion_length": 98.92857360839844,
"epoch": 0.6251748251748251,
"grad_norm": 0.010057352537761423,
"kl": 0.212890625,
"learning_rate": 9.903777526394094e-05,
"loss": -0.0086,
"num_tokens": 4539039.0,
"reward": 2.980358362197876,
"reward_std": 0.05100167542695999,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 1.9642857313156128,
"rewards/check_winston_local_func": 0.587501049041748,
"step": 447
},
{
"clip_ratio": 0.0,
"epoch": 0.6265734265734266,
"grad_norm": 0.006958390301115274,
"kl": 0.21875,
"learning_rate": 9.90258241271112e-05,
"loss": -0.0087,
"step": 448
},
{
"clip_ratio": 0.0,
"completion_length": 111.71429443359375,
"epoch": 0.627972027972028,
"grad_norm": 1.6536770511152556e-06,
"kl": 0.142578125,
"learning_rate": 9.901379995923738e-05,
"loss": 0.0001,
"num_tokens": 4559675.0,
"reward": 2.681791067123413,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.2857142984867096,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.4675052762031555,
"step": 449
},
{
"clip_ratio": 0.0,
"epoch": 0.6293706293706294,
"grad_norm": 1.6773170531853214e-06,
"kl": 0.142578125,
"learning_rate": 9.900170277823129e-05,
"loss": 0.0001,
"step": 450
},
{
"clip_ratio": 0.0,
"completion_length": 107.35714721679688,
"epoch": 0.6307692307692307,
"grad_norm": 2.0958023033337787e-06,
"kl": 0.2109375,
"learning_rate": 9.898953260211338e-05,
"loss": 0.0002,
"num_tokens": 4579995.0,
"reward": 2.7400293350219727,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.4186006188392639,
"step": 451
},
{
"clip_ratio": 0.0,
"epoch": 0.6321678321678321,
"grad_norm": 1.982971590769392e-06,
"kl": 0.2109375,
"learning_rate": 9.897728944901292e-05,
"loss": 0.0002,
"step": 452
},
{
"clip_ratio": 0.0,
"completion_length": 118.26786041259766,
"epoch": 0.6335664335664336,
"grad_norm": 0.009450375112303724,
"kl": 0.1416015625,
"learning_rate": 9.896497333716783e-05,
"loss": -0.0051,
"num_tokens": 4602048.0,
"reward": 2.640944004058838,
"reward_std": 0.043768420815467834,
"rewards/check_gptzero_func": 0.5357142686843872,
"rewards/check_perplexity_diff_func": 1.5357142686843872,
"rewards/check_winston_local_func": 0.5695151090621948,
"step": 453
},
{
"clip_ratio": 0.0,
"epoch": 0.634965034965035,
"grad_norm": 0.0017713963386990425,
"kl": 0.1416015625,
"learning_rate": 9.895258428492475e-05,
"loss": -0.0051,
"step": 454
},
{
"clip_ratio": 0.0,
"completion_length": 98.42857360839844,
"epoch": 0.6363636363636364,
"grad_norm": 0.014431904681184264,
"kl": 0.1943359375,
"learning_rate": 9.894012231073894e-05,
"loss": 0.0002,
"num_tokens": 4621654.0,
"reward": 2.7700040340423584,
"reward_std": 0.03728308901190758,
"rewards/check_gptzero_func": 0.4107142984867096,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.43071839213371277,
"step": 455
},
{
"clip_ratio": 0.0,
"epoch": 0.6377622377622377,
"grad_norm": 0.008239966094546339,
"kl": 0.1943359375,
"learning_rate": 9.892758743317434e-05,
"loss": 0.0002,
"step": 456
},
{
"clip_ratio": 0.0,
"completion_length": 90.50000762939453,
"epoch": 0.6391608391608392,
"grad_norm": 2.579298856684113e-06,
"kl": 0.1923828125,
"learning_rate": 9.891497967090344e-05,
"loss": 0.0002,
"num_tokens": 4640140.0,
"reward": 2.6383018493652344,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.25,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.4597306251525879,
"step": 457
},
{
"clip_ratio": 0.0,
"epoch": 0.6405594405594406,
"grad_norm": 2.6625863420088633e-06,
"kl": 0.1923828125,
"learning_rate": 9.890229904270731e-05,
"loss": 0.0002,
"step": 458
},
{
"clip_ratio": 0.0,
"completion_length": 113.03572082519531,
"epoch": 0.641958041958042,
"grad_norm": 1.7932858894376123e-05,
"kl": 0.1748046875,
"learning_rate": 9.888954556747563e-05,
"loss": 0.0002,
"num_tokens": 4661276.0,
"reward": 2.74210524559021,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.3849623203277588,
"step": 459
},
{
"clip_ratio": 0.0,
"epoch": 0.6433566433566433,
"grad_norm": 1.7905922488331998e-05,
"kl": 0.1748046875,
"learning_rate": 9.887671926420648e-05,
"loss": 0.0002,
"step": 460
},
{
"clip_ratio": 0.0,
"completion_length": 112.42857360839844,
"epoch": 0.6447552447552447,
"grad_norm": 0.014077582219597583,
"kl": 0.173828125,
"learning_rate": 9.886382015200652e-05,
"loss": 0.0002,
"num_tokens": 4682450.0,
"reward": 2.8680295944213867,
"reward_std": 0.05049533396959305,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 1.9642857313156128,
"rewards/check_winston_local_func": 0.43945807218551636,
"step": 461
},
{
"clip_ratio": 0.0,
"epoch": 0.6461538461538462,
"grad_norm": 0.010796297977350395,
"kl": 0.173828125,
"learning_rate": 9.885084825009086e-05,
"loss": 0.0001,
"step": 462
},
{
"clip_ratio": 0.0,
"completion_length": 105.14286041259766,
"epoch": 0.6475524475524476,
"grad_norm": 0.02303930816984521,
"kl": 0.17578125,
"learning_rate": 9.883780357778299e-05,
"loss": 0.0003,
"num_tokens": 4702710.0,
"reward": 2.885434150695801,
"reward_std": 0.06019994616508484,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 1.75,
"rewards/check_winston_local_func": 0.7068625092506409,
"step": 463
},
{
"clip_ratio": 0.0,
"epoch": 0.6489510489510489,
"grad_norm": 0.0015050223064125838,
"kl": 0.1767578125,
"learning_rate": 9.882468615451484e-05,
"loss": 0.0002,
"step": 464
},
{
"clip_ratio": 0.0,
"completion_length": 127.5714340209961,
"epoch": 0.6503496503496503,
"grad_norm": 3.0460982652821775e-06,
"kl": 0.1650390625,
"learning_rate": 9.881149599982671e-05,
"loss": 0.0002,
"num_tokens": 4725654.0,
"reward": 2.407341957092285,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.2857142984867096,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.2644847333431244,
"step": 465
},
{
"clip_ratio": 0.0,
"epoch": 0.6517482517482518,
"grad_norm": 2.8166666383288673e-06,
"kl": 0.1650390625,
"learning_rate": 9.879823313336722e-05,
"loss": 0.0002,
"step": 466
},
{
"clip_ratio": 0.0,
"completion_length": 122.16072082519531,
"epoch": 0.6531468531468532,
"grad_norm": 0.0001021700279542065,
"kl": 0.140625,
"learning_rate": 9.878489757489337e-05,
"loss": 0.0001,
"num_tokens": 4748375.0,
"reward": 2.6119117736816406,
"reward_std": 2.7804879209725186e-05,
"rewards/check_gptzero_func": 0.5357142686843872,
"rewards/check_perplexity_diff_func": 1.5714285373687744,
"rewards/check_winston_local_func": 0.5047687888145447,
"step": 467
},
{
"clip_ratio": 0.0,
"epoch": 0.6545454545454545,
"grad_norm": 9.33433928584176e-05,
"kl": 0.140625,
"learning_rate": 9.877148934427037e-05,
"loss": 0.0001,
"step": 468
},
{
"clip_ratio": 0.0006711409660056233,
"completion_length": 95.64286041259766,
"epoch": 0.6559440559440559,
"grad_norm": 0.0004684591425632982,
"kl": 0.171875,
"learning_rate": 9.87580084614717e-05,
"loss": 0.0002,
"num_tokens": 4767439.0,
"reward": 2.3873257637023926,
"reward_std": 0.0017833748133853078,
"rewards/check_gptzero_func": 0.25,
"rewards/check_perplexity_diff_func": 1.7142857313156128,
"rewards/check_winston_local_func": 0.42303988337516785,
"step": 469
},
{
"clip_ratio": 0.0004793864209204912,
"epoch": 0.6573426573426573,
"grad_norm": 0.00047799981287926325,
"kl": 0.171875,
"learning_rate": 9.874445494657911e-05,
"loss": 0.0002,
"step": 470
},
{
"clip_ratio": 0.0,
"completion_length": 79.39286041259766,
"epoch": 0.6587412587412588,
"grad_norm": 0.011441113949782547,
"kl": 0.216796875,
"learning_rate": 9.873082881978251e-05,
"loss": -0.0064,
"num_tokens": 4784715.0,
"reward": 3.037661075592041,
"reward_std": 0.023328183218836784,
"rewards/check_gptzero_func": 0.4821428656578064,
"rewards/check_perplexity_diff_func": 1.9642857313156128,
"rewards/check_winston_local_func": 0.5912323594093323,
"step": 471
},
{
"clip_ratio": 0.0,
"epoch": 0.6601398601398601,
"grad_norm": 0.008433986682608557,
"kl": 0.216796875,
"learning_rate": 9.871713010137997e-05,
"loss": -0.0064,
"step": 472
},
{
"clip_ratio": 0.0,
"completion_length": 119.42857360839844,
"epoch": 0.6615384615384615,
"grad_norm": 7.638914800437301e-06,
"kl": 0.1435546875,
"learning_rate": 9.870335881177774e-05,
"loss": 0.0001,
"num_tokens": 4806207.0,
"reward": 2.9706830978393555,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.5063972473144531,
"step": 473
},
{
"clip_ratio": 0.0,
"epoch": 0.6629370629370629,
"grad_norm": 7.505737589947172e-06,
"kl": 0.1435546875,
"learning_rate": 9.868951497149011e-05,
"loss": 0.0001,
"step": 474
},
{
"clip_ratio": 0.00019073051225859672,
"completion_length": 89.87500762939453,
"epoch": 0.6643356643356644,
"grad_norm": 0.03662860311910791,
"kl": 0.2060546875,
"learning_rate": 9.86755986011395e-05,
"loss": 0.0065,
"num_tokens": 4824332.0,
"reward": 2.643022060394287,
"reward_std": 0.04776443541049957,
"rewards/check_gptzero_func": 0.3214285671710968,
"rewards/check_perplexity_diff_func": 1.8928571939468384,
"rewards/check_winston_local_func": 0.42873620986938477,
"step": 475
},
{
"clip_ratio": 0.0,
"epoch": 0.6657342657342658,
"grad_norm": 0.03126231173733144,
"kl": 0.2060546875,
"learning_rate": 9.866160972145634e-05,
"loss": 0.0062,
"step": 476
},
{
"clip_ratio": 0.0,
"completion_length": 123.10714721679688,
"epoch": 0.6671328671328671,
"grad_norm": 0.0002559996362709361,
"kl": 0.1357421875,
"learning_rate": 9.864754835327909e-05,
"loss": 0.0001,
"num_tokens": 4847094.0,
"reward": 2.767104387283325,
"reward_std": 0.0007265785825438797,
"rewards/check_gptzero_func": 0.5,
"rewards/check_perplexity_diff_func": 1.7142857313156128,
"rewards/check_winston_local_func": 0.5528185963630676,
"step": 477
},
{
"clip_ratio": 8.97343925316818e-05,
"epoch": 0.6685314685314685,
"grad_norm": 0.00042633622080268514,
"kl": 0.1357421875,
"learning_rate": 9.86334145175542e-05,
"loss": 0.0001,
"step": 478
},
{
"clip_ratio": 0.0,
"completion_length": 96.67857360839844,
"epoch": 0.66993006993007,
"grad_norm": 0.0008693689003033054,
"kl": 0.16015625,
"learning_rate": 9.861920823533606e-05,
"loss": 0.0002,
"num_tokens": 4866394.0,
"reward": 2.5761330127716064,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 1.7142857313156128,
"rewards/check_winston_local_func": 0.504704475402832,
"step": 479
},
{
"clip_ratio": 0.0,
"epoch": 0.6713286713286714,
"grad_norm": 0.0008435837049153919,
"kl": 0.16015625,
"learning_rate": 9.860492952778696e-05,
"loss": 0.0002,
"step": 480
},
{
"clip_ratio": 0.0,
"completion_length": 94.39286041259766,
"epoch": 0.6727272727272727,
"grad_norm": 3.349433978828943e-06,
"kl": 0.177734375,
"learning_rate": 9.859057841617709e-05,
"loss": 0.0002,
"num_tokens": 4885318.0,
"reward": 2.8612334728240967,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.5755191445350647,
"step": 481
},
{
"clip_ratio": 0.0,
"epoch": 0.6741258741258741,
"grad_norm": 3.5095106255593373e-06,
"kl": 0.177734375,
"learning_rate": 9.857615492188452e-05,
"loss": 0.0002,
"step": 482
},
{
"clip_ratio": 0.0,
"completion_length": 87.98214721679688,
"epoch": 0.6755244755244755,
"grad_norm": 0.00016582191324414645,
"kl": 0.208984375,
"learning_rate": 9.856165906639513e-05,
"loss": 0.0002,
"num_tokens": 4903401.0,
"reward": 2.797102928161621,
"reward_std": 0.0008326892857439816,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.4756740629673004,
"step": 483
},
{
"clip_ratio": 0.0,
"epoch": 0.676923076923077,
"grad_norm": 0.00017592290286556547,
"kl": 0.208984375,
"learning_rate": 9.85470908713026e-05,
"loss": 0.0002,
"step": 484
},
{
"clip_ratio": 0.0,
"completion_length": 122.14286041259766,
"epoch": 0.6783216783216783,
"grad_norm": 3.840262302649335e-05,
"kl": 0.140625,
"learning_rate": 9.853245035830834e-05,
"loss": 0.0001,
"num_tokens": 4925777.0,
"reward": 2.80380916595459,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.4823804199695587,
"step": 485
},
{
"clip_ratio": 0.0,
"epoch": 0.6797202797202797,
"grad_norm": 3.855750270963149e-05,
"kl": 0.140625,
"learning_rate": 9.851773754922152e-05,
"loss": 0.0001,
"step": 486
},
{
"clip_ratio": 0.0,
"completion_length": 110.5714340209961,
"epoch": 0.6811188811188811,
"grad_norm": 1.0000521495087345e-06,
"kl": 0.1572265625,
"learning_rate": 9.850295246595898e-05,
"loss": 0.0002,
"num_tokens": 4946515.0,
"reward": 2.9591448307037354,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.5,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.5305731296539307,
"step": 487
},
{
"clip_ratio": 0.0,
"epoch": 0.6825174825174826,
"grad_norm": 9.579988148024946e-07,
"kl": 0.1572265625,
"learning_rate": 9.848809513054523e-05,
"loss": 0.0002,
"step": 488
},
{
"clip_ratio": 0.0,
"completion_length": 104.67857360839844,
"epoch": 0.6839160839160839,
"grad_norm": 4.4290499511336513e-07,
"kl": 0.1728515625,
"learning_rate": 9.847316556511245e-05,
"loss": 0.0002,
"num_tokens": 4966537.0,
"reward": 3.0352413654327393,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.5709553956985474,
"step": 489
},
{
"clip_ratio": 0.0,
"epoch": 0.6853146853146853,
"grad_norm": 4.228282290104698e-07,
"kl": 0.1728515625,
"learning_rate": 9.845816379190036e-05,
"loss": 0.0002,
"step": 490
},
{
"clip_ratio": 0.0,
"completion_length": 106.92857360839844,
"epoch": 0.6867132867132867,
"grad_norm": 7.161264654013381e-07,
"kl": 0.1865234375,
"learning_rate": 9.844308983325625e-05,
"loss": 0.0002,
"num_tokens": 4986837.0,
"reward": 2.8011527061462402,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.40829533338546753,
"step": 491
},
{
"clip_ratio": 0.0,
"epoch": 0.6881118881118881,
"grad_norm": 6.606409864369399e-07,
"kl": 0.1865234375,
"learning_rate": 9.842794371163501e-05,
"loss": 0.0002,
"step": 492
},
{
"clip_ratio": 0.0,
"completion_length": 127.3214340209961,
"epoch": 0.6895104895104895,
"grad_norm": 3.5056088482557826e-05,
"kl": 0.11767578125,
"learning_rate": 9.841272544959892e-05,
"loss": 0.0001,
"num_tokens": 5009783.0,
"reward": 2.6788315773010254,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.5,
"rewards/check_perplexity_diff_func": 1.7857142686843872,
"rewards/check_winston_local_func": 0.393117219209671,
"step": 493
},
{
"clip_ratio": 0.0,
"epoch": 0.6909090909090909,
"grad_norm": 3.5363245519599165e-05,
"kl": 0.11767578125,
"learning_rate": 9.839743506981782e-05,
"loss": 0.0001,
"step": 494
},
{
"clip_ratio": 0.0004093328316230327,
"completion_length": 99.08928680419922,
"epoch": 0.6923076923076923,
"grad_norm": 0.0028759704706919223,
"kl": 0.1650390625,
"learning_rate": 9.838207259506891e-05,
"loss": 0.0002,
"num_tokens": 5029238.0,
"reward": 2.5216493606567383,
"reward_std": 0.0011310166446492076,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 1.7142857313156128,
"rewards/check_winston_local_func": 0.45022064447402954,
"step": 495
},
{
"clip_ratio": 0.0004093328316230327,
"epoch": 0.6937062937062937,
"grad_norm": 0.0035050811312865673,
"kl": 0.1650390625,
"learning_rate": 9.836663804823683e-05,
"loss": 0.0002,
"step": 496
},
{
"clip_ratio": 0.0,
"completion_length": 124.85714721679688,
"epoch": 0.6951048951048951,
"grad_norm": 3.8191875557546916e-05,
"kl": 0.158203125,
"learning_rate": 9.835113145231356e-05,
"loss": 0.0002,
"num_tokens": 5051408.0,
"reward": 2.8089945316314697,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.4161372184753418,
"step": 497
},
{
"clip_ratio": 0.0,
"epoch": 0.6965034965034965,
"grad_norm": 3.1854181850026e-05,
"kl": 0.158203125,
"learning_rate": 9.833555283039842e-05,
"loss": 0.0002,
"step": 498
},
{
"clip_ratio": 0.0,
"completion_length": 115.17857360839844,
"epoch": 0.6979020979020979,
"grad_norm": 3.735234931957505e-07,
"kl": 0.15234375,
"learning_rate": 9.831990220569801e-05,
"loss": 0.0002,
"num_tokens": 5072486.0,
"reward": 2.7535111904144287,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.46779707074165344,
"step": 499
},
{
"clip_ratio": 0.0,
"epoch": 0.6993006993006993,
"grad_norm": 3.749561048042854e-07,
"kl": 0.15234375,
"learning_rate": 9.83041796015262e-05,
"loss": 0.0002,
"step": 500
},
{
"clip_ratio": 0.0,
"completion_length": 137.85714721679688,
"epoch": 0.7006993006993008,
"grad_norm": 0.004581651252064326,
"kl": 0.1357421875,
"learning_rate": 9.828838504130406e-05,
"loss": 0.0001,
"num_tokens": 24128.0,
"reward": 2.5751242637634277,
"reward_std": 0.050465863198041916,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.8214285373687744,
"rewards/check_winston_local_func": 0.36083847284317017,
"step": 501
},
{
"clip_ratio": 0.0,
"epoch": 0.7020979020979021,
"grad_norm": 0.004623783518082634,
"kl": 0.1357421875,
"learning_rate": 9.827251854855991e-05,
"loss": 0.0001,
"step": 502
},
{
"clip_ratio": 0.0,
"completion_length": 84.0714340209961,
"epoch": 0.7034965034965035,
"grad_norm": 5.4224474596128455e-06,
"kl": 0.1962890625,
"learning_rate": 9.825658014692914e-05,
"loss": 0.0002,
"num_tokens": 41660.0,
"reward": 2.881941556930542,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.48908427357673645,
"step": 503
},
{
"clip_ratio": 0.0,
"epoch": 0.7048951048951049,
"grad_norm": 5.774418529053615e-06,
"kl": 0.1962890625,
"learning_rate": 9.824056986015433e-05,
"loss": 0.0002,
"step": 504
},
{
"clip_ratio": 0.0,
"completion_length": 125.76786041259766,
"epoch": 0.7062937062937062,
"grad_norm": 0.012190639094338367,
"kl": 0.158203125,
"learning_rate": 9.82244877120851e-05,
"loss": -0.0,
"num_tokens": 63809.0,
"reward": 2.9153220653533936,
"reward_std": 0.05151599273085594,
"rewards/check_gptzero_func": 0.5,
"rewards/check_perplexity_diff_func": 1.9642857313156128,
"rewards/check_winston_local_func": 0.45103612542152405,
"step": 505
},
{
"clip_ratio": 0.0,
"epoch": 0.7076923076923077,
"grad_norm": 0.0008706312902749691,
"kl": 0.1572265625,
"learning_rate": 9.820833372667812e-05,
"loss": -0.0001,
"step": 506
},
{
"clip_ratio": 0.0,
"completion_length": 106.67857360839844,
"epoch": 0.7090909090909091,
"grad_norm": 7.4245877265908374e-06,
"kl": 0.171875,
"learning_rate": 9.819210792799712e-05,
"loss": 0.0002,
"num_tokens": 83877.0,
"reward": 2.742607831954956,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.3854646682739258,
"step": 507
},
{
"clip_ratio": 0.0,
"epoch": 0.7104895104895105,
"grad_norm": 5.9863315776137925e-06,
"kl": 0.171875,
"learning_rate": 9.817581034021272e-05,
"loss": 0.0002,
"step": 508
},
{
"clip_ratio": 0.0,
"completion_length": 112.21429443359375,
"epoch": 0.7118881118881119,
"grad_norm": 1.5219106062929897e-05,
"kl": 0.154296875,
"learning_rate": 9.815944098760257e-05,
"loss": 0.0002,
"num_tokens": 105225.0,
"reward": 2.886364698410034,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.5649359226226807,
"step": 509
},
{
"clip_ratio": 0.0,
"epoch": 0.7132867132867133,
"grad_norm": 3.826854189322478e-05,
"kl": 0.154296875,
"learning_rate": 9.814299989455117e-05,
"loss": 0.0002,
"step": 510
},
{
"clip_ratio": 0.0,
"completion_length": 93.50000762939453,
"epoch": 0.7146853146853147,
"grad_norm": 3.6634275067756628e-06,
"kl": 0.18359375,
"learning_rate": 9.81264870855499e-05,
"loss": 0.0002,
"num_tokens": 124237.0,
"reward": 2.9750092029571533,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.5,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.5464377403259277,
"step": 511
},
{
"clip_ratio": 0.0,
"epoch": 0.7160839160839161,
"grad_norm": 3.3120330925968346e-06,
"kl": 0.18359375,
"learning_rate": 9.810990258519699e-05,
"loss": 0.0002,
"step": 512
},
{
"clip_ratio": 0.0,
"completion_length": 114.5714340209961,
"epoch": 0.7174825174825175,
"grad_norm": 1.5887563588226216e-07,
"kl": 0.1474609375,
"learning_rate": 9.809324641819741e-05,
"loss": 0.0001,
"num_tokens": 145557.0,
"reward": 2.767850637435913,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.4107076823711395,
"step": 513
},
{
"clip_ratio": 0.0,
"epoch": 0.7188811188811188,
"grad_norm": 1.5359302478120885e-07,
"kl": 0.1474609375,
"learning_rate": 9.807651860936297e-05,
"loss": 0.0001,
"step": 514
},
{
"clip_ratio": 0.0,
"completion_length": 110.3214340209961,
"epoch": 0.7202797202797203,
"grad_norm": 1.620447571811333e-07,
"kl": 0.1630859375,
"learning_rate": 9.805971918361214e-05,
"loss": 0.0002,
"num_tokens": 165997.0,
"reward": 2.9055941104888916,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.5127367377281189,
"step": 515
},
{
"clip_ratio": 0.0,
"epoch": 0.7216783216783217,
"grad_norm": 1.472040658632873e-07,
"kl": 0.1630859375,
"learning_rate": 9.804284816597008e-05,
"loss": 0.0002,
"step": 516
},
{
"clip_ratio": 0.0,
"completion_length": 91.67857360839844,
"epoch": 0.7230769230769231,
"grad_norm": 2.4623250489986615e-06,
"kl": 0.142578125,
"learning_rate": 9.802590558156862e-05,
"loss": 0.0001,
"num_tokens": 184617.0,
"reward": 2.853912830352783,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.46105554699897766,
"step": 517
},
{
"clip_ratio": 0.0,
"epoch": 0.7244755244755244,
"grad_norm": 2.0559549189734124e-06,
"kl": 0.142578125,
"learning_rate": 9.800889145564617e-05,
"loss": 0.0001,
"step": 518
},
{
"clip_ratio": 0.0,
"completion_length": 127.21429443359375,
"epoch": 0.7258741258741259,
"grad_norm": 2.3530034340292295e-07,
"kl": 0.1552734375,
"learning_rate": 9.799180581354774e-05,
"loss": 0.0002,
"num_tokens": 207223.0,
"reward": 2.7364511489868164,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.45073673129081726,
"step": 519
},
{
"clip_ratio": 0.0,
"epoch": 0.7272727272727273,
"grad_norm": 2.2322343800833895e-07,
"kl": 0.1552734375,
"learning_rate": 9.797464868072488e-05,
"loss": 0.0002,
"step": 520
},
{
"clip_ratio": 0.0,
"completion_length": 104.5714340209961,
"epoch": 0.7286713286713287,
"grad_norm": 2.968346809067501e-07,
"kl": 0.177734375,
"learning_rate": 9.795742008273558e-05,
"loss": 0.0002,
"num_tokens": 227059.0,
"reward": 2.6932995319366455,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.3361565172672272,
"step": 521
},
{
"clip_ratio": 0.0,
"epoch": 0.73006993006993,
"grad_norm": 2.852980195095417e-07,
"kl": 0.177734375,
"learning_rate": 9.794012004524434e-05,
"loss": 0.0002,
"step": 522
},
{
"clip_ratio": 0.0,
"completion_length": 105.42857360839844,
"epoch": 0.7314685314685314,
"grad_norm": 3.3796985539550523e-06,
"kl": 0.1640625,
"learning_rate": 9.792274859402205e-05,
"loss": 0.0002,
"num_tokens": 246949.0,
"reward": 2.6483211517333984,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3214285671710968,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.3983212113380432,
"step": 523
},
{
"clip_ratio": 0.0,
"epoch": 0.7328671328671329,
"grad_norm": 3.1476330793325015e-06,
"kl": 0.1640625,
"learning_rate": 9.790530575494603e-05,
"loss": 0.0002,
"step": 524
},
{
"clip_ratio": 0.0,
"completion_length": 125.03572082519531,
"epoch": 0.7342657342657343,
"grad_norm": 0.0060337949589422705,
"kl": 0.16796875,
"learning_rate": 9.788779155399987e-05,
"loss": -0.0001,
"num_tokens": 269441.0,
"reward": 2.768662214279175,
"reward_std": 0.049707408994436264,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.9642857313156128,
"rewards/check_winston_local_func": 0.4115191400051117,
"step": 525
},
{
"clip_ratio": 0.0,
"epoch": 0.7356643356643356,
"grad_norm": 0.0068588182961818464,
"kl": 0.16796875,
"learning_rate": 9.787020601727352e-05,
"loss": -0.0,
"step": 526
},
{
"clip_ratio": 0.0,
"completion_length": 78.39286041259766,
"epoch": 0.737062937062937,
"grad_norm": 1.3153821115529096e-05,
"kl": 0.2353515625,
"learning_rate": 9.785254917096318e-05,
"loss": 0.0002,
"num_tokens": 286345.0,
"reward": 3.074557304382324,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.5,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.5745573043823242,
"step": 527
},
{
"clip_ratio": 0.0,
"epoch": 0.7384615384615385,
"grad_norm": 1.256427796379324e-05,
"kl": 0.2353515625,
"learning_rate": 9.783482104137127e-05,
"loss": 0.0002,
"step": 528
},
{
"clip_ratio": 0.0,
"completion_length": 124.39286041259766,
"epoch": 0.7398601398601399,
"grad_norm": 1.3517164643713064e-05,
"kl": 0.12353515625,
"learning_rate": 9.781702165490639e-05,
"loss": 0.0001,
"num_tokens": 309195.0,
"reward": 2.4804210662841797,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.6428571939468384,
"rewards/check_winston_local_func": 0.44470664858818054,
"step": 529
},
{
"clip_ratio": 0.0,
"epoch": 0.7412587412587412,
"grad_norm": 1.3131407312669822e-05,
"kl": 0.12353515625,
"learning_rate": 9.779915103808328e-05,
"loss": 0.0001,
"step": 530
},
{
"clip_ratio": 0.0,
"completion_length": 111.53572082519531,
"epoch": 0.7426573426573426,
"grad_norm": 7.637416736333039e-07,
"kl": 0.15625,
"learning_rate": 9.778120921752285e-05,
"loss": 0.0002,
"num_tokens": 330505.0,
"reward": 3.070021390914917,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.5357142686843872,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.5343068838119507,
"step": 531
},
{
"clip_ratio": 0.0,
"epoch": 0.7440559440559441,
"grad_norm": 7.316834780823122e-07,
"kl": 0.15625,
"learning_rate": 9.776319621995201e-05,
"loss": 0.0002,
"step": 532
},
{
"clip_ratio": 0.0,
"completion_length": 106.35714721679688,
"epoch": 0.7454545454545455,
"grad_norm": 3.663080215251399e-07,
"kl": 0.193359375,
"learning_rate": 9.77451120722037e-05,
"loss": 0.0002,
"num_tokens": 350813.0,
"reward": 2.63775634765625,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3214285671710968,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.3877563774585724,
"step": 533
},
{
"clip_ratio": 0.0,
"epoch": 0.7468531468531469,
"grad_norm": 3.8678973306591443e-07,
"kl": 0.193359375,
"learning_rate": 9.77269568012169e-05,
"loss": 0.0002,
"step": 534
},
{
"clip_ratio": 0.0,
"completion_length": 106.67857360839844,
"epoch": 0.7482517482517482,
"grad_norm": 6.123173544504598e-06,
"kl": 0.2216796875,
"learning_rate": 9.770873043403648e-05,
"loss": 0.0002,
"num_tokens": 371585.0,
"reward": 2.874614953994751,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.5174719095230103,
"step": 535
},
{
"clip_ratio": 0.0,
"epoch": 0.7496503496503496,
"grad_norm": 6.141198966103098e-06,
"kl": 0.2216796875,
"learning_rate": 9.769043299781327e-05,
"loss": 0.0002,
"step": 536
},
{
"clip_ratio": 0.0,
"completion_length": 104.71428680419922,
"epoch": 0.7510489510489511,
"grad_norm": 6.024782598667822e-07,
"kl": 0.212890625,
"learning_rate": 9.767206451980394e-05,
"loss": 0.0002,
"num_tokens": 391201.0,
"reward": 2.666369676589966,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.25,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.4163695275783539,
"step": 537
},
{
"clip_ratio": 0.0,
"epoch": 0.7524475524475525,
"grad_norm": 5.651815000751003e-07,
"kl": 0.212890625,
"learning_rate": 9.765362502737097e-05,
"loss": 0.0002,
"step": 538
},
{
"clip_ratio": 0.0,
"completion_length": 107.21428680419922,
"epoch": 0.7538461538461538,
"grad_norm": 4.5359076717402366e-07,
"kl": 0.19140625,
"learning_rate": 9.763511454798268e-05,
"loss": 0.0002,
"num_tokens": 412087.0,
"reward": 2.660142660140991,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.25,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.4815710484981537,
"step": 539
},
{
"clip_ratio": 0.0,
"epoch": 0.7552447552447552,
"grad_norm": 7.72545386297213e-07,
"kl": 0.19140625,
"learning_rate": 9.761653310921307e-05,
"loss": 0.0002,
"step": 540
},
{
"clip_ratio": 0.0,
"completion_length": 133.85714721679688,
"epoch": 0.7566433566433567,
"grad_norm": 8.430058973039463e-08,
"kl": 0.146484375,
"learning_rate": 9.759788073874189e-05,
"loss": 0.0001,
"num_tokens": 435535.0,
"reward": 2.720244884490967,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.3631021976470947,
"step": 541
},
{
"clip_ratio": 0.0,
"epoch": 0.7580419580419581,
"grad_norm": 8.327469845321351e-08,
"kl": 0.146484375,
"learning_rate": 9.757915746435453e-05,
"loss": 0.0001,
"step": 542
},
{
"clip_ratio": 0.0002040816325461492,
"completion_length": 102.89286041259766,
"epoch": 0.7594405594405594,
"grad_norm": 0.000356312443076848,
"kl": 0.203125,
"learning_rate": 9.756036331394202e-05,
"loss": 0.0002,
"num_tokens": 455137.0,
"reward": 2.6731011867523193,
"reward_std": 1.646135569899343e-05,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.351672500371933,
"step": 543
},
{
"clip_ratio": 0.0,
"epoch": 0.7608391608391608,
"grad_norm": 0.00031516713281171804,
"kl": 0.203125,
"learning_rate": 9.754149831550098e-05,
"loss": 0.0002,
"step": 544
},
{
"clip_ratio": 0.0,
"completion_length": 89.35714721679688,
"epoch": 0.7622377622377622,
"grad_norm": 2.70311309473727e-07,
"kl": 0.2080078125,
"learning_rate": 9.752256249713351e-05,
"loss": 0.0002,
"num_tokens": 473297.0,
"reward": 2.869070291519165,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.5119272470474243,
"step": 545
},
{
"clip_ratio": 0.0,
"epoch": 0.7636363636363637,
"grad_norm": 2.830814999627113e-07,
"kl": 0.2080078125,
"learning_rate": 9.750355588704727e-05,
"loss": 0.0002,
"step": 546
},
{
"clip_ratio": 0.0,
"completion_length": 132.44644165039062,
"epoch": 0.765034965034965,
"grad_norm": 0.0009688545197921594,
"kl": 0.1806640625,
"learning_rate": 9.748447851355535e-05,
"loss": 0.0002,
"num_tokens": 496926.0,
"reward": 2.7388832569122314,
"reward_std": 0.0023702967446297407,
"rewards/check_gptzero_func": 0.5357142686843872,
"rewards/check_perplexity_diff_func": 1.7142857313156128,
"rewards/check_winston_local_func": 0.4888834059238434,
"step": 547
},
{
"clip_ratio": 0.0,
"epoch": 0.7664335664335664,
"grad_norm": 0.0016734864168510978,
"kl": 0.1806640625,
"learning_rate": 9.746533040507624e-05,
"loss": 0.0002,
"step": 548
},
{
"clip_ratio": 0.0,
"completion_length": 97.03572082519531,
"epoch": 0.7678321678321678,
"grad_norm": 1.9163803223436476e-06,
"kl": 0.177734375,
"learning_rate": 9.744611159013381e-05,
"loss": 0.0002,
"num_tokens": 516080.0,
"reward": 2.9889118671417236,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.5,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.48891177773475647,
"step": 549
},
{
"clip_ratio": 0.0,
"epoch": 0.7692307692307693,
"grad_norm": 1.2761616262245509e-06,
"kl": 0.177734375,
"learning_rate": 9.742682209735727e-05,
"loss": 0.0002,
"step": 550
},
{
"clip_ratio": 0.0,
"completion_length": 92.39286041259766,
"epoch": 0.7706293706293706,
"grad_norm": 1.896094459987994e-07,
"kl": 0.1923828125,
"learning_rate": 9.740746195548112e-05,
"loss": 0.0002,
"num_tokens": 534554.0,
"reward": 2.943678855895996,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.5151072144508362,
"step": 551
},
{
"clip_ratio": 0.0,
"epoch": 0.772027972027972,
"grad_norm": 2.1137044966091938e-07,
"kl": 0.1923828125,
"learning_rate": 9.738803119334506e-05,
"loss": 0.0002,
"step": 552
},
{
"clip_ratio": 0.0,
"completion_length": 103.21428680419922,
"epoch": 0.7734265734265734,
"grad_norm": 4.705318347406964e-07,
"kl": 0.1826171875,
"learning_rate": 9.736852983989404e-05,
"loss": 0.0002,
"num_tokens": 554220.0,
"reward": 2.9958415031433105,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.5,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.4958415627479553,
"step": 553
},
{
"clip_ratio": 0.0,
"epoch": 0.7748251748251749,
"grad_norm": 5.24917119837944e-07,
"kl": 0.1826171875,
"learning_rate": 9.734895792417811e-05,
"loss": 0.0002,
"step": 554
},
{
"clip_ratio": 0.0,
"completion_length": 102.53572082519531,
"epoch": 0.7762237762237763,
"grad_norm": 0.0005595837692618245,
"kl": 0.1611328125,
"learning_rate": 9.73293154753525e-05,
"loss": -0.0001,
"num_tokens": 574520.0,
"reward": 2.660358190536499,
"reward_std": 0.05061452463269234,
"rewards/check_gptzero_func": 0.2857142984867096,
"rewards/check_perplexity_diff_func": 1.75,
"rewards/check_winston_local_func": 0.6246438026428223,
"step": 555
},
{
"clip_ratio": 0.0,
"epoch": 0.7776223776223776,
"grad_norm": 0.0005476655329061088,
"kl": 0.1611328125,
"learning_rate": 9.730960252267743e-05,
"loss": -0.0001,
"step": 556
},
{
"clip_ratio": 0.0,
"completion_length": 110.28572082519531,
"epoch": 0.779020979020979,
"grad_norm": 1.9994241055330475e-07,
"kl": 0.162109375,
"learning_rate": 9.728981909551824e-05,
"loss": 0.0002,
"num_tokens": 595056.0,
"reward": 2.7659287452697754,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.40878555178642273,
"step": 557
},
{
"clip_ratio": 0.0,
"epoch": 0.7804195804195804,
"grad_norm": 2.102720872162732e-07,
"kl": 0.162109375,
"learning_rate": 9.726996522334516e-05,
"loss": 0.0002,
"step": 558
},
{
"clip_ratio": 0.0,
"completion_length": 116.3214340209961,
"epoch": 0.7818181818181819,
"grad_norm": 1.4418577967265551e-06,
"kl": 0.197265625,
"learning_rate": 9.725004093573342e-05,
"loss": 0.0002,
"num_tokens": 616218.0,
"reward": 2.90523099899292,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.4766596257686615,
"step": 559
},
{
"clip_ratio": 0.0,
"epoch": 0.7832167832167832,
"grad_norm": 1.4761501812795077e-06,
"kl": 0.197265625,
"learning_rate": 9.723004626236314e-05,
"loss": 0.0002,
"step": 560
},
{
"clip_ratio": 0.0,
"completion_length": 92.28572082519531,
"epoch": 0.7846153846153846,
"grad_norm": 6.640556191873379e-07,
"kl": 0.1904296875,
"learning_rate": 9.720998123301923e-05,
"loss": 0.0002,
"num_tokens": 634450.0,
"reward": 2.910466432571411,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.4461804926395416,
"step": 561
},
{
"clip_ratio": 0.0,
"epoch": 0.786013986013986,
"grad_norm": 6.785626739636823e-07,
"kl": 0.1904296875,
"learning_rate": 9.718984587759148e-05,
"loss": 0.0002,
"step": 562
},
{
"clip_ratio": 0.0,
"completion_length": 97.67857360839844,
"epoch": 0.7874125874125875,
"grad_norm": 0.0002364011375932314,
"kl": 0.1953125,
"learning_rate": 9.71696402260744e-05,
"loss": 0.0002,
"num_tokens": 653730.0,
"reward": 3.09153151512146,
"reward_std": 4.5663102355320007e-05,
"rewards/check_gptzero_func": 0.5714285969734192,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.5201030373573303,
"step": 563
},
{
"clip_ratio": 0.00014228800137061626,
"epoch": 0.7888111888111888,
"grad_norm": 0.00020388090420570156,
"kl": 0.1953125,
"learning_rate": 9.714936430856723e-05,
"loss": 0.0002,
"step": 564
},
{
"clip_ratio": 0.0,
"completion_length": 120.66072082519531,
"epoch": 0.7902097902097902,
"grad_norm": 8.089432221384178e-05,
"kl": 0.1357421875,
"learning_rate": 9.712901815527386e-05,
"loss": 0.0001,
"num_tokens": 675657.0,
"reward": 2.674098253250122,
"reward_std": 0.0001108883589040488,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.38838380575180054,
"step": 565
},
{
"clip_ratio": 0.0,
"epoch": 0.7916083916083916,
"grad_norm": 8.889658368646956e-05,
"kl": 0.1357421875,
"learning_rate": 9.710860179650287e-05,
"loss": 0.0001,
"step": 566
},
{
"clip_ratio": 0.0,
"completion_length": 107.3214340209961,
"epoch": 0.793006993006993,
"grad_norm": 0.0001037228042730343,
"kl": 0.17578125,
"learning_rate": 9.70881152626673e-05,
"loss": 0.0002,
"num_tokens": 695807.0,
"reward": 2.615325927734375,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.25,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.4367544949054718,
"step": 567
},
{
"clip_ratio": 0.0,
"epoch": 0.7944055944055944,
"grad_norm": 9.62410190276379e-05,
"kl": 0.17578125,
"learning_rate": 9.706755858428486e-05,
"loss": 0.0002,
"step": 568
},
{
"clip_ratio": 0.0,
"completion_length": 128.3928680419922,
"epoch": 0.7958041958041958,
"grad_norm": 3.2052372287941215e-05,
"kl": 0.12890625,
"learning_rate": 9.704693179197767e-05,
"loss": 0.0001,
"num_tokens": 719201.0,
"reward": 2.616921901702881,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 1.7857142686843872,
"rewards/check_winston_local_func": 0.40263620018959045,
"step": 569
},
{
"clip_ratio": 0.0,
"epoch": 0.7972027972027972,
"grad_norm": 3.1589569913845654e-05,
"kl": 0.12890625,
"learning_rate": 9.702623491647233e-05,
"loss": 0.0001,
"step": 570
},
{
"clip_ratio": 0.0,
"completion_length": 97.3214340209961,
"epoch": 0.7986013986013986,
"grad_norm": 2.966365002312992e-07,
"kl": 0.1572265625,
"learning_rate": 9.70054679885998e-05,
"loss": 0.0002,
"num_tokens": 738947.0,
"reward": 2.835333824157715,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.47819074988365173,
"step": 571
},
{
"clip_ratio": 0.0,
"epoch": 0.8,
"grad_norm": 4.1573650424054196e-07,
"kl": 0.1572265625,
"learning_rate": 9.698463103929542e-05,
"loss": 0.0002,
"step": 572
},
{
"clip_ratio": 0.0,
"completion_length": 105.78572082519531,
"epoch": 0.8013986013986014,
"grad_norm": 0.000588551803293507,
"kl": 0.1611328125,
"learning_rate": 9.696372409959886e-05,
"loss": 0.0002,
"num_tokens": 759359.0,
"reward": 2.967468738555908,
"reward_std": 0.0020603849552571774,
"rewards/check_gptzero_func": 0.5,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.5388973355293274,
"step": 573
},
{
"clip_ratio": 0.0,
"epoch": 0.8027972027972028,
"grad_norm": 0.0006215594211925717,
"kl": 0.1611328125,
"learning_rate": 9.694274720065399e-05,
"loss": 0.0002,
"step": 574
},
{
"clip_ratio": 0.0,
"completion_length": 97.14286041259766,
"epoch": 0.8041958041958042,
"grad_norm": 3.3967306751168834e-06,
"kl": 0.17578125,
"learning_rate": 9.692170037370898e-05,
"loss": 0.0002,
"num_tokens": 778637.0,
"reward": 2.9381115436553955,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.509539783000946,
"step": 575
},
{
"clip_ratio": 0.0,
"epoch": 0.8055944055944056,
"grad_norm": 3.2036118908588705e-06,
"kl": 0.17578125,
"learning_rate": 9.690058365011607e-05,
"loss": 0.0002,
"step": 576
},
{
"clip_ratio": 0.0,
"completion_length": 102.64286041259766,
"epoch": 0.806993006993007,
"grad_norm": 8.125129903430204e-06,
"kl": 0.1630859375,
"learning_rate": 9.687939706133168e-05,
"loss": 0.0002,
"num_tokens": 798473.0,
"reward": 2.781018018722534,
"reward_std": 0.025253813713788986,
"rewards/check_gptzero_func": 0.4107142984867096,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.4417320787906647,
"step": 577
},
{
"clip_ratio": 0.0,
"epoch": 0.8083916083916084,
"grad_norm": 7.98774787404577e-06,
"kl": 0.1630859375,
"learning_rate": 9.685814063891631e-05,
"loss": 0.0002,
"step": 578
},
{
"clip_ratio": 0.0,
"completion_length": 126.85714721679688,
"epoch": 0.8097902097902098,
"grad_norm": 3.679599498892351e-07,
"kl": 0.138671875,
"learning_rate": 9.683681441453445e-05,
"loss": 0.0001,
"num_tokens": 821081.0,
"reward": 2.8629186153411865,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.43434715270996094,
"step": 579
},
{
"clip_ratio": 0.0,
"epoch": 0.8111888111888111,
"grad_norm": 3.93663211553172e-07,
"kl": 0.138671875,
"learning_rate": 9.681541841995461e-05,
"loss": 0.0001,
"step": 580
},
{
"clip_ratio": 0.0,
"completion_length": 106.00000762939453,
"epoch": 0.8125874125874126,
"grad_norm": 8.022029708518222e-07,
"kl": 0.162109375,
"learning_rate": 9.67939526870492e-05,
"loss": 0.0002,
"num_tokens": 841529.0,
"reward": 2.8752760887145996,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.55384761095047,
"step": 581
},
{
"clip_ratio": 0.0,
"epoch": 0.813986013986014,
"grad_norm": 8.501682817147847e-07,
"kl": 0.162109375,
"learning_rate": 9.677241724779453e-05,
"loss": 0.0002,
"step": 582
},
{
"clip_ratio": 0.0,
"completion_length": 109.85714721679688,
"epoch": 0.8153846153846154,
"grad_norm": 1.7930012533014129e-06,
"kl": 0.1689453125,
"learning_rate": 9.675081213427076e-05,
"loss": 0.0002,
"num_tokens": 862213.0,
"reward": 2.9139628410339355,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.48539119958877563,
"step": 583
},
{
"clip_ratio": 0.0,
"epoch": 0.8167832167832167,
"grad_norm": 1.6519465345892093e-06,
"kl": 0.1689453125,
"learning_rate": 9.672913737866179e-05,
"loss": 0.0002,
"step": 584
},
{
"clip_ratio": 0.0,
"completion_length": 98.66072082519531,
"epoch": 0.8181818181818182,
"grad_norm": 0.002787236207669909,
"kl": 0.25,
"learning_rate": 9.670739301325534e-05,
"loss": 0.0002,
"num_tokens": 881842.0,
"reward": 2.6230576038360596,
"reward_std": 0.006216124631464481,
"rewards/check_gptzero_func": 0.3571428656578064,
"rewards/check_perplexity_diff_func": 1.7857142686843872,
"rewards/check_winston_local_func": 0.48020049929618835,
"step": 585
},
{
"clip_ratio": 0.0,
"epoch": 0.8195804195804196,
"grad_norm": 0.002885109231591189,
"kl": 0.2490234375,
"learning_rate": 9.668557907044276e-05,
"loss": 0.0002,
"step": 586
},
{
"clip_ratio": 0.0,
"completion_length": 79.14286041259766,
"epoch": 0.820979020979021,
"grad_norm": 4.462250819797621e-06,
"kl": 0.1923828125,
"learning_rate": 9.666369558271909e-05,
"loss": 0.0002,
"num_tokens": 898716.0,
"reward": 2.900649309158325,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.3928571343421936,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.5077921152114868,
"step": 587
},
{
"clip_ratio": 0.0,
"epoch": 0.8223776223776224,
"grad_norm": 4.488075066816524e-06,
"kl": 0.1923828125,
"learning_rate": 9.66417425826829e-05,
"loss": 0.0002,
"step": 588
},
{
"clip_ratio": 0.0,
"completion_length": 112.03572082519531,
"epoch": 0.8237762237762237,
"grad_norm": 1.3821264354883407e-07,
"kl": 0.14453125,
"learning_rate": 9.661972010303641e-05,
"loss": 0.0001,
"num_tokens": 919636.0,
"reward": 2.7668612003326416,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.4454323649406433,
"step": 589
},
{
"clip_ratio": 0.0,
"epoch": 0.8251748251748252,
"grad_norm": 1.4174011116816794e-07,
"kl": 0.14453125,
"learning_rate": 9.659762817658524e-05,
"loss": 0.0001,
"step": 590
},
{
"clip_ratio": 0.0,
"completion_length": 107.92857360839844,
"epoch": 0.8265734265734266,
"grad_norm": 8.773734381605369e-07,
"kl": 0.1572265625,
"learning_rate": 9.65754668362385e-05,
"loss": 0.0002,
"num_tokens": 940096.0,
"reward": 2.8289589881896973,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4642857015132904,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.4361015856266022,
"step": 591
},
{
"clip_ratio": 0.0,
"epoch": 0.827972027972028,
"grad_norm": 1.0023469469060766e-06,
"kl": 0.1572265625,
"learning_rate": 9.655323611500875e-05,
"loss": 0.0002,
"step": 592
},
{
"clip_ratio": 0.0,
"completion_length": 113.85714721679688,
"epoch": 0.8293706293706293,
"grad_norm": 6.1541718025779055e-06,
"kl": 0.166015625,
"learning_rate": 9.653093604601183e-05,
"loss": 0.0002,
"num_tokens": 961340.0,
"reward": 2.9606730937957764,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.5357142686843872,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.4249587059020996,
"step": 593
},
{
"clip_ratio": 0.0,
"epoch": 0.8307692307692308,
"grad_norm": 6.512813773208274e-06,
"kl": 0.166015625,
"learning_rate": 9.650856666246693e-05,
"loss": 0.0002,
"step": 594
},
{
"clip_ratio": 0.0,
"completion_length": 91.60714721679688,
"epoch": 0.8321678321678322,
"grad_norm": 3.2190340652283214e-06,
"kl": 0.1767578125,
"learning_rate": 9.648612799769644e-05,
"loss": 0.0002,
"num_tokens": 979916.0,
"reward": 3.0096070766448975,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 2.0,
"rewards/check_winston_local_func": 0.5810357332229614,
"step": 595
},
{
"clip_ratio": 0.0,
"epoch": 0.8335664335664336,
"grad_norm": 3.1137752081162388e-06,
"kl": 0.1767578125,
"learning_rate": 9.646362008512602e-05,
"loss": 0.0002,
"step": 596
},
{
"clip_ratio": 0.0,
"completion_length": 106.53572082519531,
"epoch": 0.8349650349650349,
"grad_norm": 6.812051267845749e-06,
"kl": 0.185546875,
"learning_rate": 9.644104295828447e-05,
"loss": 0.0002,
"num_tokens": 1000300.0,
"reward": 2.7739078998565674,
"reward_std": 0.0,
"rewards/check_gptzero_func": 0.4285714328289032,
"rewards/check_perplexity_diff_func": 1.9285714626312256,
"rewards/check_winston_local_func": 0.4167649447917938,
"step": 597
},
{
"clip_ratio": 0.0,
"epoch": 0.8363636363636363,
"grad_norm": 2.398683493432653e-06,
"kl": 0.185546875,
"learning_rate": 9.641839665080363e-05,
"loss": 0.0002,
"step": 598
},
{
"clip_ratio": 0.0,
"completion_length": 107.41072082519531,
"epoch": 0.8377622377622378,
"grad_norm": 0.0008785473557125962,
"kl": 0.1796875,
"learning_rate": 9.63956811964185e-05,
"loss": 0.0002,
"num_tokens": 1021105.0,
"reward": 2.91171932220459,
"reward_std": 0.00400555832311511,
"rewards/check_gptzero_func": 0.5357142686843872,
"rewards/check_perplexity_diff_func": 1.8571428060531616,
"rewards/check_winston_local_func": 0.518862247467041,
"step": 599
},
{
"clip_ratio": 0.0,
"epoch": 0.8391608391608392,
"grad_norm": 0.000892784560673435,
"kl": 0.1796875,
"learning_rate": 9.6372896628967e-05,
"loss": 0.0002,
"step": 600
}
],
"logging_steps": 1,
"max_steps": 2860,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}