File size: 8,744 Bytes
c847b29
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 5.381165919282511,
  "eval_steps": 500,
  "global_step": 150,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.35874439461883406,
      "grad_norm": 0.5037005543708801,
      "learning_rate": 9.259259259259259e-07,
      "logits/chosen": 1.851180076599121,
      "logits/rejected": 1.8187497854232788,
      "logps/chosen": -71.38922119140625,
      "logps/rejected": -77.1716537475586,
      "loss": 0.6947,
      "rewards/accuracies": 0.36250001192092896,
      "rewards/chosen": -0.0033944465685635805,
      "rewards/margins": -0.004582419525831938,
      "rewards/rejected": 0.0011879729572683573,
      "step": 10
    },
    {
      "epoch": 0.7174887892376681,
      "grad_norm": 0.4899679124355316,
      "learning_rate": 1.8518518518518519e-06,
      "logits/chosen": 1.804616928100586,
      "logits/rejected": 1.8522075414657593,
      "logps/chosen": -68.71460723876953,
      "logps/rejected": -92.10017395019531,
      "loss": 0.6935,
      "rewards/accuracies": 0.5,
      "rewards/chosen": 0.0006334683275781572,
      "rewards/margins": 0.002181120216846466,
      "rewards/rejected": -0.0015476513653993607,
      "step": 20
    },
    {
      "epoch": 1.0762331838565022,
      "grad_norm": 0.4610757529735565,
      "learning_rate": 2.7777777777777783e-06,
      "logits/chosen": 1.7368440628051758,
      "logits/rejected": 1.747293472290039,
      "logps/chosen": -86.10456848144531,
      "logps/rejected": -94.18408203125,
      "loss": 0.6923,
      "rewards/accuracies": 0.5249999761581421,
      "rewards/chosen": -0.004750942811369896,
      "rewards/margins": -3.7298166716936976e-05,
      "rewards/rejected": -0.0047136456705629826,
      "step": 30
    },
    {
      "epoch": 1.4349775784753362,
      "grad_norm": 0.5546987056732178,
      "learning_rate": 3.7037037037037037e-06,
      "logits/chosen": 1.760947585105896,
      "logits/rejected": 1.7774131298065186,
      "logps/chosen": -72.0299301147461,
      "logps/rejected": -75.53373718261719,
      "loss": 0.6938,
      "rewards/accuracies": 0.550000011920929,
      "rewards/chosen": 0.004539580084383488,
      "rewards/margins": 0.004153046756982803,
      "rewards/rejected": 0.0003865335020236671,
      "step": 40
    },
    {
      "epoch": 1.7937219730941703,
      "grad_norm": 0.5019991397857666,
      "learning_rate": 4.62962962962963e-06,
      "logits/chosen": 1.6594927310943604,
      "logits/rejected": 1.7026996612548828,
      "logps/chosen": -74.72074890136719,
      "logps/rejected": -80.34657287597656,
      "loss": 0.694,
      "rewards/accuracies": 0.5874999761581421,
      "rewards/chosen": 0.0014741375343874097,
      "rewards/margins": 0.00692129647359252,
      "rewards/rejected": -0.00544715765863657,
      "step": 50
    },
    {
      "epoch": 2.1524663677130045,
      "grad_norm": 0.5497386455535889,
      "learning_rate": 4.998119881260576e-06,
      "logits/chosen": 1.845393180847168,
      "logits/rejected": 1.8288724422454834,
      "logps/chosen": -85.8925552368164,
      "logps/rejected": -79.84625244140625,
      "loss": 0.6939,
      "rewards/accuracies": 0.5249999761581421,
      "rewards/chosen": 0.0007343249162659049,
      "rewards/margins": 0.0027797420043498278,
      "rewards/rejected": -0.0020454167388379574,
      "step": 60
    },
    {
      "epoch": 2.5112107623318387,
      "grad_norm": 0.53052818775177,
      "learning_rate": 4.9866405060165044e-06,
      "logits/chosen": 1.7219512462615967,
      "logits/rejected": 1.739393949508667,
      "logps/chosen": -82.69403076171875,
      "logps/rejected": -85.24137878417969,
      "loss": 0.6912,
      "rewards/accuracies": 0.512499988079071,
      "rewards/chosen": 0.004084458574652672,
      "rewards/margins": 8.91583040356636e-05,
      "rewards/rejected": 0.003995300270617008,
      "step": 70
    },
    {
      "epoch": 2.8699551569506725,
      "grad_norm": 0.5697340965270996,
      "learning_rate": 4.964774158361991e-06,
      "logits/chosen": 1.688643217086792,
      "logits/rejected": 1.7375280857086182,
      "logps/chosen": -77.02912902832031,
      "logps/rejected": -84.49137115478516,
      "loss": 0.6923,
      "rewards/accuracies": 0.4375,
      "rewards/chosen": -0.0026642463635653257,
      "rewards/margins": -0.0012511021923273802,
      "rewards/rejected": -0.0014131448697298765,
      "step": 80
    },
    {
      "epoch": 3.2286995515695067,
      "grad_norm": 0.6337418556213379,
      "learning_rate": 4.93261217644956e-06,
      "logits/chosen": 1.7215973138809204,
      "logits/rejected": 1.6856582164764404,
      "logps/chosen": -81.94705963134766,
      "logps/rejected": -79.82647705078125,
      "loss": 0.6916,
      "rewards/accuracies": 0.4749999940395355,
      "rewards/chosen": 0.004837460815906525,
      "rewards/margins": 0.0014790964778512716,
      "rewards/rejected": 0.003358363639563322,
      "step": 90
    },
    {
      "epoch": 3.587443946188341,
      "grad_norm": 0.6216289401054382,
      "learning_rate": 4.8902889044347e-06,
      "logits/chosen": 1.8092563152313232,
      "logits/rejected": 1.7624406814575195,
      "logps/chosen": -91.27610778808594,
      "logps/rejected": -76.58167266845703,
      "loss": 0.6879,
      "rewards/accuracies": 0.612500011920929,
      "rewards/chosen": 0.004790940321981907,
      "rewards/margins": 0.018742190673947334,
      "rewards/rejected": -0.013951249420642853,
      "step": 100
    },
    {
      "epoch": 3.9461883408071747,
      "grad_norm": 0.6686798930168152,
      "learning_rate": 4.837981131305475e-06,
      "logits/chosen": 1.8108171224594116,
      "logits/rejected": 1.8439037799835205,
      "logps/chosen": -68.89733123779297,
      "logps/rejected": -90.71812438964844,
      "loss": 0.686,
      "rewards/accuracies": 0.5375000238418579,
      "rewards/chosen": 0.0019698950927704573,
      "rewards/margins": 0.011438245885074139,
      "rewards/rejected": -0.009468351490795612,
      "step": 110
    },
    {
      "epoch": 4.304932735426009,
      "grad_norm": 0.690719723701477,
      "learning_rate": 4.775907352415367e-06,
      "logits/chosen": 1.7892429828643799,
      "logits/rejected": 1.7746537923812866,
      "logps/chosen": -76.5858383178711,
      "logps/rejected": -70.5009765625,
      "loss": 0.6841,
      "rewards/accuracies": 0.625,
      "rewards/chosen": 0.00402367627248168,
      "rewards/margins": 0.02823176048696041,
      "rewards/rejected": -0.024208081886172295,
      "step": 120
    },
    {
      "epoch": 4.663677130044843,
      "grad_norm": 0.6445876955986023,
      "learning_rate": 4.70432685680402e-06,
      "logits/chosen": 1.7731869220733643,
      "logits/rejected": 1.70940363407135,
      "logps/chosen": -80.2589111328125,
      "logps/rejected": -69.99195861816406,
      "loss": 0.6807,
      "rewards/accuracies": 0.612500011920929,
      "rewards/chosen": 0.0028332143556326628,
      "rewards/margins": 0.027303647249937057,
      "rewards/rejected": -0.024470435455441475,
      "step": 130
    },
    {
      "epoch": 5.022421524663677,
      "grad_norm": 0.7485262751579285,
      "learning_rate": 4.623538644118244e-06,
      "logits/chosen": 1.707404375076294,
      "logits/rejected": 1.7899534702301025,
      "logps/chosen": -85.12259674072266,
      "logps/rejected": -96.5855712890625,
      "loss": 0.6717,
      "rewards/accuracies": 0.6000000238418579,
      "rewards/chosen": 0.0005208252114243805,
      "rewards/margins": 0.046183306723833084,
      "rewards/rejected": -0.04566247761249542,
      "step": 140
    },
    {
      "epoch": 5.381165919282511,
      "grad_norm": 0.7476359009742737,
      "learning_rate": 4.533880175657419e-06,
      "logits/chosen": 1.8374496698379517,
      "logits/rejected": 1.8710906505584717,
      "logps/chosen": -91.67071533203125,
      "logps/rejected": -89.70133209228516,
      "loss": 0.6683,
      "rewards/accuracies": 0.6499999761581421,
      "rewards/chosen": 0.004734884016215801,
      "rewards/margins": 0.06081797555088997,
      "rewards/rejected": -0.05608309060335159,
      "step": 150
    }
  ],
  "logging_steps": 10,
  "max_steps": 540,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 20,
  "save_steps": 50,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": false
      },
      "attributes": {}
    }
  },
  "total_flos": 5.1052954900863386e+17,
  "train_batch_size": 1,
  "trial_name": null,
  "trial_params": null
}