TempDPO / trainer_state.json
Milian's picture
Upload folder using huggingface_hub
ff7cfac verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 44,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.022727272727272728,
"grad_norm": 16.625,
"learning_rate": 0.0,
"logits/chosen": 0.4483538866043091,
"logits/rejected": 0.7762327194213867,
"logps/chosen": -34.067535400390625,
"logps/rejected": -26.805654525756836,
"loss": 0.6781,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.017476718872785568,
"rewards/margins": 0.03066336363554001,
"rewards/rejected": -0.01318664662539959,
"step": 1
},
{
"epoch": 0.045454545454545456,
"grad_norm": 11.875,
"learning_rate": 1.2e-05,
"logits/chosen": 0.5961519479751587,
"logits/rejected": 0.6054960489273071,
"logps/chosen": -20.344480514526367,
"logps/rejected": -23.291126251220703,
"loss": 0.696,
"rewards/accuracies": 0.3333333432674408,
"rewards/chosen": 0.002745183650404215,
"rewards/margins": -0.00570220872759819,
"rewards/rejected": 0.008447392843663692,
"step": 2
},
{
"epoch": 0.06818181818181818,
"grad_norm": 12.4375,
"learning_rate": 2.4e-05,
"logits/chosen": -0.2268427461385727,
"logits/rejected": -0.08561667054891586,
"logps/chosen": -26.2761287689209,
"logps/rejected": -23.514366149902344,
"loss": 0.6965,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": -0.0071416860446333885,
"rewards/margins": -0.0062666586600244045,
"rewards/rejected": -0.0008750278502702713,
"step": 3
},
{
"epoch": 0.09090909090909091,
"grad_norm": 13.0,
"learning_rate": 3.6e-05,
"logits/chosen": 0.5151932239532471,
"logits/rejected": 0.5493708848953247,
"logps/chosen": -29.416748046875,
"logps/rejected": -21.80704116821289,
"loss": 0.6664,
"rewards/accuracies": 1.0,
"rewards/chosen": 0.04196351394057274,
"rewards/margins": 0.05446663126349449,
"rewards/rejected": -0.012503115460276604,
"step": 4
},
{
"epoch": 0.11363636363636363,
"grad_norm": 12.1875,
"learning_rate": 4.8e-05,
"logits/chosen": 0.3497011661529541,
"logits/rejected": 0.31218263506889343,
"logps/chosen": -31.662017822265625,
"logps/rejected": -24.33977508544922,
"loss": 0.6927,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.007583046332001686,
"rewards/margins": 0.0009033204987645149,
"rewards/rejected": 0.0066797262988984585,
"step": 5
},
{
"epoch": 0.13636363636363635,
"grad_norm": 12.3125,
"learning_rate": 6e-05,
"logits/chosen": 0.5960320234298706,
"logits/rejected": 1.0259536504745483,
"logps/chosen": -30.316076278686523,
"logps/rejected": -21.38100242614746,
"loss": 0.6842,
"rewards/accuracies": 1.0,
"rewards/chosen": 0.012160556390881538,
"rewards/margins": 0.01807880401611328,
"rewards/rejected": -0.0059182485565543175,
"step": 6
},
{
"epoch": 0.1590909090909091,
"grad_norm": 12.6875,
"learning_rate": 5.846153846153846e-05,
"logits/chosen": 0.3294275999069214,
"logits/rejected": 0.2885156273841858,
"logps/chosen": -30.754199981689453,
"logps/rejected": -21.946483612060547,
"loss": 0.6911,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.020008979365229607,
"rewards/margins": 0.004138438496738672,
"rewards/rejected": 0.015870539471507072,
"step": 7
},
{
"epoch": 0.18181818181818182,
"grad_norm": 11.75,
"learning_rate": 5.692307692307692e-05,
"logits/chosen": 0.4179474115371704,
"logits/rejected": 0.6303722262382507,
"logps/chosen": -25.757431030273438,
"logps/rejected": -22.029760360717773,
"loss": 0.6991,
"rewards/accuracies": 0.3333333432674408,
"rewards/chosen": -0.005146535113453865,
"rewards/margins": -0.011694017797708511,
"rewards/rejected": 0.006547483615577221,
"step": 8
},
{
"epoch": 0.20454545454545456,
"grad_norm": 11.125,
"learning_rate": 5.538461538461539e-05,
"logits/chosen": 0.6332809329032898,
"logits/rejected": 0.7613773345947266,
"logps/chosen": -28.643695831298828,
"logps/rejected": -22.429039001464844,
"loss": 0.686,
"rewards/accuracies": 1.0,
"rewards/chosen": 0.019583575427532196,
"rewards/margins": 0.014355978928506374,
"rewards/rejected": 0.005227597430348396,
"step": 9
},
{
"epoch": 0.22727272727272727,
"grad_norm": 8.875,
"learning_rate": 5.384615384615385e-05,
"logits/chosen": 0.4406185746192932,
"logits/rejected": 0.44449326395988464,
"logps/chosen": -23.943906784057617,
"logps/rejected": -20.787511825561523,
"loss": 0.6815,
"rewards/accuracies": 1.0,
"rewards/chosen": 0.06105436012148857,
"rewards/margins": 0.023596446961164474,
"rewards/rejected": 0.0374579131603241,
"step": 10
},
{
"epoch": 0.25,
"grad_norm": 18.375,
"learning_rate": 5.230769230769231e-05,
"logits/chosen": 1.2527503967285156,
"logits/rejected": 1.176261067390442,
"logps/chosen": -42.38333511352539,
"logps/rejected": -26.054731369018555,
"loss": 0.6936,
"rewards/accuracies": 0.3333333432674408,
"rewards/chosen": 0.026292739436030388,
"rewards/margins": -0.0008817669004201889,
"rewards/rejected": 0.027174506336450577,
"step": 11
},
{
"epoch": 0.2727272727272727,
"grad_norm": 9.625,
"learning_rate": 5.076923076923077e-05,
"logits/chosen": 0.37028029561042786,
"logits/rejected": 0.5766373872756958,
"logps/chosen": -22.245540618896484,
"logps/rejected": -21.667922973632812,
"loss": 0.663,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.05539780110120773,
"rewards/margins": 0.06313318014144897,
"rewards/rejected": -0.007735380437225103,
"step": 12
},
{
"epoch": 0.29545454545454547,
"grad_norm": 9.8125,
"learning_rate": 4.923076923076923e-05,
"logits/chosen": -0.09108446538448334,
"logits/rejected": 0.24406792223453522,
"logps/chosen": -28.064685821533203,
"logps/rejected": -21.54052734375,
"loss": 0.6802,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.05286191403865814,
"rewards/margins": 0.026534145697951317,
"rewards/rejected": 0.026327770203351974,
"step": 13
},
{
"epoch": 0.3181818181818182,
"grad_norm": 11.8125,
"learning_rate": 4.769230769230769e-05,
"logits/chosen": 0.48153504729270935,
"logits/rejected": 0.533014714717865,
"logps/chosen": -25.028493881225586,
"logps/rejected": -24.618581771850586,
"loss": 0.7134,
"rewards/accuracies": 0.3333333432674408,
"rewards/chosen": -0.03459123149514198,
"rewards/margins": -0.038723185658454895,
"rewards/rejected": 0.004131953231990337,
"step": 14
},
{
"epoch": 0.3409090909090909,
"grad_norm": 12.4375,
"learning_rate": 4.615384615384616e-05,
"logits/chosen": 0.32169950008392334,
"logits/rejected": 0.5244829058647156,
"logps/chosen": -26.596710205078125,
"logps/rejected": -23.023845672607422,
"loss": 0.6797,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.047678254544734955,
"rewards/margins": 0.02770627662539482,
"rewards/rejected": 0.019971976056694984,
"step": 15
},
{
"epoch": 0.36363636363636365,
"grad_norm": 11.8125,
"learning_rate": 4.461538461538462e-05,
"logits/chosen": -1.4551410675048828,
"logits/rejected": -1.4342135190963745,
"logps/chosen": -29.413467407226562,
"logps/rejected": -25.22298240661621,
"loss": 0.6804,
"rewards/accuracies": 1.0,
"rewards/chosen": 0.03757172077894211,
"rewards/margins": 0.02564137801527977,
"rewards/rejected": 0.01193033903837204,
"step": 16
},
{
"epoch": 0.38636363636363635,
"grad_norm": 12.6875,
"learning_rate": 4.307692307692308e-05,
"logits/chosen": 1.1149061918258667,
"logits/rejected": 1.3534132242202759,
"logps/chosen": -28.74739646911621,
"logps/rejected": -23.868566513061523,
"loss": 0.6526,
"rewards/accuracies": 1.0,
"rewards/chosen": 0.031014060601592064,
"rewards/margins": 0.08419768512248993,
"rewards/rejected": -0.05318361520767212,
"step": 17
},
{
"epoch": 0.4090909090909091,
"grad_norm": 13.6875,
"learning_rate": 4.153846153846154e-05,
"logits/chosen": 0.43228524923324585,
"logits/rejected": 0.44759899377822876,
"logps/chosen": -29.343727111816406,
"logps/rejected": -20.927490234375,
"loss": 0.6779,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.0534362867474556,
"rewards/margins": 0.03123900294303894,
"rewards/rejected": 0.022197280079126358,
"step": 18
},
{
"epoch": 0.4318181818181818,
"grad_norm": 11.0,
"learning_rate": 3.9999999999999996e-05,
"logits/chosen": 0.9617244601249695,
"logits/rejected": 1.036171317100525,
"logps/chosen": -22.17565155029297,
"logps/rejected": -23.60016632080078,
"loss": 0.6727,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.015546607784926891,
"rewards/margins": 0.0422033928334713,
"rewards/rejected": -0.02665678784251213,
"step": 19
},
{
"epoch": 0.45454545454545453,
"grad_norm": 15.6875,
"learning_rate": 3.846153846153846e-05,
"logits/chosen": 0.6454523205757141,
"logits/rejected": 1.2390544414520264,
"logps/chosen": -26.925029754638672,
"logps/rejected": -24.303573608398438,
"loss": 0.6238,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.10690931975841522,
"rewards/margins": 0.15301425755023956,
"rewards/rejected": -0.04610493779182434,
"step": 20
},
{
"epoch": 0.4772727272727273,
"grad_norm": 14.6875,
"learning_rate": 3.692307692307693e-05,
"logits/chosen": 0.8735722303390503,
"logits/rejected": 1.108435034751892,
"logps/chosen": -33.20803451538086,
"logps/rejected": -23.885366439819336,
"loss": 0.7071,
"rewards/accuracies": 0.3333333432674408,
"rewards/chosen": -0.025695737451314926,
"rewards/margins": -0.026847710832953453,
"rewards/rejected": 0.0011519752442836761,
"step": 21
},
{
"epoch": 0.5,
"grad_norm": 11.125,
"learning_rate": 3.538461538461539e-05,
"logits/chosen": 0.3912545144557953,
"logits/rejected": 0.8124356269836426,
"logps/chosen": -28.345073699951172,
"logps/rejected": -29.91925048828125,
"loss": 0.6494,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.0346221923828125,
"rewards/margins": 0.09574788063764572,
"rewards/rejected": -0.06112569198012352,
"step": 22
},
{
"epoch": 0.5227272727272727,
"grad_norm": 12.25,
"learning_rate": 3.384615384615385e-05,
"logits/chosen": 0.8518146872520447,
"logits/rejected": 0.7960999608039856,
"logps/chosen": -30.20541000366211,
"logps/rejected": -25.19312858581543,
"loss": 0.673,
"rewards/accuracies": 0.3333333432674408,
"rewards/chosen": -0.005412420257925987,
"rewards/margins": 0.04422512277960777,
"rewards/rejected": -0.04963754117488861,
"step": 23
},
{
"epoch": 0.5454545454545454,
"grad_norm": 11.1875,
"learning_rate": 3.230769230769231e-05,
"logits/chosen": 0.5866941213607788,
"logits/rejected": 0.9357667565345764,
"logps/chosen": -26.213180541992188,
"logps/rejected": -22.639881134033203,
"loss": 0.6633,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.04306405782699585,
"rewards/margins": 0.0637446790933609,
"rewards/rejected": -0.02068062126636505,
"step": 24
},
{
"epoch": 0.5681818181818182,
"grad_norm": 12.625,
"learning_rate": 3.076923076923077e-05,
"logits/chosen": 0.4800880551338196,
"logits/rejected": 0.47039034962654114,
"logps/chosen": -34.4006233215332,
"logps/rejected": -22.132925033569336,
"loss": 0.6946,
"rewards/accuracies": 0.3333333432674408,
"rewards/chosen": 0.020339712500572205,
"rewards/margins": -0.002547518815845251,
"rewards/rejected": 0.022887229919433594,
"step": 25
},
{
"epoch": 0.5909090909090909,
"grad_norm": 9.375,
"learning_rate": 2.923076923076923e-05,
"logits/chosen": 0.09336434304714203,
"logits/rejected": 0.05750793218612671,
"logps/chosen": -23.79607391357422,
"logps/rejected": -21.185327529907227,
"loss": 0.6685,
"rewards/accuracies": 1.0,
"rewards/chosen": 0.053456246852874756,
"rewards/margins": 0.050235815346241,
"rewards/rejected": 0.0032204315066337585,
"step": 26
},
{
"epoch": 0.6136363636363636,
"grad_norm": 12.125,
"learning_rate": 2.7692307692307694e-05,
"logits/chosen": 0.4014180600643158,
"logits/rejected": 1.0060770511627197,
"logps/chosen": -24.400341033935547,
"logps/rejected": -26.51728630065918,
"loss": 0.6725,
"rewards/accuracies": 1.0,
"rewards/chosen": 0.04896443709731102,
"rewards/margins": 0.04188614338636398,
"rewards/rejected": 0.007078297436237335,
"step": 27
},
{
"epoch": 0.6363636363636364,
"grad_norm": 16.875,
"learning_rate": 2.6153846153846157e-05,
"logits/chosen": 0.24801835417747498,
"logits/rejected": 0.2863054573535919,
"logps/chosen": -26.580617904663086,
"logps/rejected": -22.799705505371094,
"loss": 0.7299,
"rewards/accuracies": 0.3333333432674408,
"rewards/chosen": 0.022428005933761597,
"rewards/margins": -0.06923790276050568,
"rewards/rejected": 0.09166590869426727,
"step": 28
},
{
"epoch": 0.6590909090909091,
"grad_norm": 11.0625,
"learning_rate": 2.4615384615384616e-05,
"logits/chosen": 0.31185609102249146,
"logits/rejected": 0.16100364923477173,
"logps/chosen": -27.79303741455078,
"logps/rejected": -22.794635772705078,
"loss": 0.6317,
"rewards/accuracies": 1.0,
"rewards/chosen": 0.08723494410514832,
"rewards/margins": 0.12933623790740967,
"rewards/rejected": -0.042101286351680756,
"step": 29
},
{
"epoch": 0.6818181818181818,
"grad_norm": 14.9375,
"learning_rate": 2.307692307692308e-05,
"logits/chosen": 0.3959487974643707,
"logits/rejected": 0.688706636428833,
"logps/chosen": -30.851974487304688,
"logps/rejected": -22.691082000732422,
"loss": 0.644,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.1131121963262558,
"rewards/margins": 0.10812320560216904,
"rewards/rejected": 0.004988989792764187,
"step": 30
},
{
"epoch": 0.7045454545454546,
"grad_norm": 8.625,
"learning_rate": 2.153846153846154e-05,
"logits/chosen": 0.37772199511528015,
"logits/rejected": 0.6661175489425659,
"logps/chosen": -24.510616302490234,
"logps/rejected": -20.977294921875,
"loss": 0.6215,
"rewards/accuracies": 1.0,
"rewards/chosen": 0.1579880565404892,
"rewards/margins": 0.15024060010910034,
"rewards/rejected": 0.0077474601566791534,
"step": 31
},
{
"epoch": 0.7272727272727273,
"grad_norm": 11.625,
"learning_rate": 1.9999999999999998e-05,
"logits/chosen": 0.8124392032623291,
"logits/rejected": 0.7364901900291443,
"logps/chosen": -24.003555297851562,
"logps/rejected": -22.78453826904297,
"loss": 0.5977,
"rewards/accuracies": 1.0,
"rewards/chosen": 0.13395456969738007,
"rewards/margins": 0.21094201505184174,
"rewards/rejected": -0.07698746025562286,
"step": 32
},
{
"epoch": 0.75,
"grad_norm": 11.5,
"learning_rate": 1.8461538461538465e-05,
"logits/chosen": 0.4202606678009033,
"logits/rejected": 0.8325386643409729,
"logps/chosen": -20.655643463134766,
"logps/rejected": -22.911235809326172,
"loss": 0.6293,
"rewards/accuracies": 1.0,
"rewards/chosen": 0.09568717330694199,
"rewards/margins": 0.1334756314754486,
"rewards/rejected": -0.03778845816850662,
"step": 33
},
{
"epoch": 0.7727272727272727,
"grad_norm": 9.25,
"learning_rate": 1.6923076923076924e-05,
"logits/chosen": 0.7410422563552856,
"logits/rejected": 0.9519510269165039,
"logps/chosen": -20.877716064453125,
"logps/rejected": -21.208219528198242,
"loss": 0.6509,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.10315850377082825,
"rewards/margins": 0.09082140028476715,
"rewards/rejected": 0.012337110936641693,
"step": 34
},
{
"epoch": 0.7954545454545454,
"grad_norm": 10.6875,
"learning_rate": 1.5384615384615384e-05,
"logits/chosen": 0.2906338572502136,
"logits/rejected": 0.2896530032157898,
"logps/chosen": -27.066699981689453,
"logps/rejected": -20.529428482055664,
"loss": 0.7098,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.016274897381663322,
"rewards/margins": -0.03218403086066246,
"rewards/rejected": 0.04845893010497093,
"step": 35
},
{
"epoch": 0.8181818181818182,
"grad_norm": 12.0625,
"learning_rate": 1.3846153846153847e-05,
"logits/chosen": 0.3423420190811157,
"logits/rejected": 0.4290481507778168,
"logps/chosen": -27.17491912841797,
"logps/rejected": -21.299049377441406,
"loss": 0.7005,
"rewards/accuracies": 0.3333333432674408,
"rewards/chosen": 0.0788511335849762,
"rewards/margins": -0.012380982749164104,
"rewards/rejected": 0.09123211354017258,
"step": 36
},
{
"epoch": 0.8409090909090909,
"grad_norm": 12.125,
"learning_rate": 1.2307692307692308e-05,
"logits/chosen": 0.43100249767303467,
"logits/rejected": 0.45868536829948425,
"logps/chosen": -26.59796142578125,
"logps/rejected": -23.339920043945312,
"loss": 0.6465,
"rewards/accuracies": 1.0,
"rewards/chosen": 0.13345324993133545,
"rewards/margins": 0.09718997776508331,
"rewards/rejected": 0.036263275891542435,
"step": 37
},
{
"epoch": 0.8636363636363636,
"grad_norm": 18.25,
"learning_rate": 1.076923076923077e-05,
"logits/chosen": 1.6746655702590942,
"logits/rejected": 1.4878392219543457,
"logps/chosen": -34.438411712646484,
"logps/rejected": -21.078689575195312,
"loss": 0.6676,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.02436860464513302,
"rewards/margins": 0.056867726147174835,
"rewards/rejected": -0.032499127089977264,
"step": 38
},
{
"epoch": 0.8863636363636364,
"grad_norm": 10.375,
"learning_rate": 9.230769230769232e-06,
"logits/chosen": 0.42151492834091187,
"logits/rejected": 0.5604749917984009,
"logps/chosen": -25.84881591796875,
"logps/rejected": -23.173572540283203,
"loss": 0.6422,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.09398200362920761,
"rewards/margins": 0.11083710193634033,
"rewards/rejected": -0.016855113208293915,
"step": 39
},
{
"epoch": 0.9090909090909091,
"grad_norm": 10.625,
"learning_rate": 7.692307692307692e-06,
"logits/chosen": 0.7431029677391052,
"logits/rejected": 0.671082079410553,
"logps/chosen": -20.48221206665039,
"logps/rejected": -21.603567123413086,
"loss": 0.6616,
"rewards/accuracies": 1.0,
"rewards/chosen": 0.08623453974723816,
"rewards/margins": 0.06446807086467743,
"rewards/rejected": 0.02176647260785103,
"step": 40
},
{
"epoch": 0.9318181818181818,
"grad_norm": 9.5625,
"learning_rate": 6.153846153846154e-06,
"logits/chosen": 0.4396713674068451,
"logits/rejected": 0.5854610800743103,
"logps/chosen": -22.81239891052246,
"logps/rejected": -22.814224243164062,
"loss": 0.6006,
"rewards/accuracies": 1.0,
"rewards/chosen": 0.12625077366828918,
"rewards/margins": 0.2127111405134201,
"rewards/rejected": -0.08646037429571152,
"step": 41
},
{
"epoch": 0.9545454545454546,
"grad_norm": 15.625,
"learning_rate": 4.615384615384616e-06,
"logits/chosen": 0.7838262319564819,
"logits/rejected": 0.7330495715141296,
"logps/chosen": -32.488609313964844,
"logps/rejected": -21.370939254760742,
"loss": 0.6928,
"rewards/accuracies": 0.3333333432674408,
"rewards/chosen": 0.08423754572868347,
"rewards/margins": 0.009891888126730919,
"rewards/rejected": 0.0743456482887268,
"step": 42
},
{
"epoch": 0.9772727272727273,
"grad_norm": 10.0,
"learning_rate": 3.076923076923077e-06,
"logits/chosen": 0.8008474111557007,
"logits/rejected": 0.8280413150787354,
"logps/chosen": -24.55078125,
"logps/rejected": -23.09003257751465,
"loss": 0.6579,
"rewards/accuracies": 0.6666666865348816,
"rewards/chosen": 0.06935082376003265,
"rewards/margins": 0.07333603501319885,
"rewards/rejected": -0.003985214047133923,
"step": 43
},
{
"epoch": 1.0,
"grad_norm": 27.5,
"learning_rate": 1.5384615384615385e-06,
"logits/chosen": 0.68901127576828,
"logits/rejected": 0.41714319586753845,
"logps/chosen": -38.19911193847656,
"logps/rejected": -22.654001235961914,
"loss": 0.7624,
"rewards/accuracies": 0.0,
"rewards/chosen": -0.0440673828125,
"rewards/margins": -0.13395901024341583,
"rewards/rejected": 0.08989162743091583,
"step": 44
}
],
"logging_steps": 1,
"max_steps": 44,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 3,
"trial_name": null,
"trial_params": null
}