yingyingzhang commited on
Commit
57c0b1f
·
verified ·
1 Parent(s): b362d3c

Upload my_tdpo_trainer.py

Browse files
Files changed (1) hide show
  1. my_tdpo_trainer.py +309 -0
my_tdpo_trainer.py ADDED
@@ -0,0 +1,309 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+ from collections import defaultdict
4
+ from contextlib import nullcontext
5
+ from typing import Dict, Literal, Optional, Tuple, Union
6
+
7
+ import torch
8
+ from transformers import BatchEncoding, Trainer
9
+ from trl import DPOTrainer
10
+ from trl.trainer.utils import disable_dropout_in_model
11
+ from transformers import PreTrainedModel
12
+ from copy import deepcopy
13
+ from accelerate import Accelerator
14
+ from accelerate.state import AcceleratorState
15
+ import deepspeed
16
+ import torch.nn.functional as F
17
+ IGNORE_INDEX = -100
18
+
19
+ def get_eval_ds_config(deepspeed_states, offload=None, stage=3):
20
+ deepspeed_states = AcceleratorState().deepspeed_plugin
21
+
22
+ device = "cpu" if offload else "none"
23
+ zero_opt_dict = {
24
+ "stage": stage,
25
+ "stage3_param_persistence_threshold": 1e4,
26
+ "offload_param": {
27
+ "device": device
28
+ }
29
+ }
30
+ return {
31
+ "train_micro_batch_size_per_gpu": deepspeed_states.deepspeed_config['train_micro_batch_size_per_gpu'],
32
+ "steps_per_print": 10,
33
+ "zero_optimization": zero_opt_dict,
34
+ "bf16": {
35
+ "enabled": True
36
+ },
37
+ "gradient_clipping": 1.0,
38
+ "prescale_gradients": False,
39
+ "wall_clock_breakdown": False
40
+ }
41
+
42
+
43
+ class CustomTDPOTrainer(DPOTrainer):
44
+ def __init__(
45
+ self,
46
+ beta: float,
47
+ alpha: float,
48
+ if_tdpo2: bool,
49
+ model: Union["PreTrainedModel", torch.nn.Module],
50
+ ref_model: Optional[Union["PreTrainedModel", torch.nn.Module]] = None,
51
+ disable_dropout: Optional[bool] = True,
52
+ **kwargs,
53
+ ):
54
+ if disable_dropout:
55
+ disable_dropout_in_model(model)
56
+ if ref_model is not None:
57
+ disable_dropout_in_model(ref_model)
58
+
59
+ self.use_dpo_data_collator = True # hack to avoid warning
60
+ self.generate_during_eval = False # disable at evaluation
61
+ self.label_pad_token_id = IGNORE_INDEX
62
+ self.padding_value = 0
63
+ self.is_encoder_decoder = model.config.is_encoder_decoder
64
+ self.precompute_ref_log_probs = False
65
+ self._precomputed_train_ref_log_probs = False
66
+ self._precomputed_eval_ref_log_probs = False
67
+ self._peft_has_been_casted_to_bf16 = False
68
+ self.reference_free = False
69
+
70
+ self.ref_model = ref_model
71
+ self.label_smoothing = 0
72
+ self._stored_metrics = defaultdict(lambda: defaultdict(list))
73
+
74
+ self.beta = beta
75
+ self.alpha = alpha
76
+ self.if_tdpo2 = if_tdpo2
77
+
78
+ Trainer.__init__(self, model=model, **kwargs)
79
+
80
+ if not hasattr(self, "accelerator"):
81
+ raise AttributeError("Please update `transformers`.")
82
+
83
+ print('====prepare ref_model====')
84
+ if ref_model is not None:
85
+ if self.is_deepspeed_enabled:
86
+ if not (
87
+ getattr(ref_model, "is_loaded_in_8bit", False) or getattr(ref_model, "is_loaded_in_4bit", False)
88
+ ): # quantized models are already set on the correct device
89
+ self.ref_model = self.my_prepare_deepspeed(self.ref_model)
90
+ else:
91
+ self.ref_model = self.accelerator.prepare_model(self.ref_model,
92
+ evaluation_mode=True)
93
+ # self.ref_model = self.accelerator.prepare_model(self.ref_model,
94
+ # evaluation_mode=True)
95
+
96
+ def my_prepare_deepspeed(self, model):
97
+ deepspeed_states = self.accelerator.state.deepspeed_plugin
98
+ config_kwargs = deepspeed_states.deepspeed_config
99
+ if config_kwargs["zero_optimization"]["stage"] != 3:
100
+ stage = 0
101
+ else:
102
+ stage = 3
103
+ ds_config = get_eval_ds_config(deepspeed_states, offload=True, stage=stage)
104
+ model, *_ = deepspeed.initialize(model=model, config=config_kwargs)
105
+ model.eval()
106
+ return model
107
+
108
+ def _prepare_deepspeed(self, model):
109
+ # config_kwargs = get_eval_ds_config(offload=True)
110
+ # model, *_ = deepspeed.initialize(model=model, config=config_kwargs)
111
+ # model.eval()
112
+ deepspeed_plugin = self.accelerator.state.deepspeed_plugin
113
+ config_kwargs = deepspeed_plugin.deepspeed_config
114
+ if model is not None:
115
+ if hasattr(model, "config"):
116
+ hidden_size = (
117
+ max(model.config.hidden_sizes)
118
+ if getattr(model.config, "hidden_sizes", None)
119
+ else getattr(model.config, "hidden_size", None)
120
+ )
121
+ if hidden_size is not None and config_kwargs["zero_optimization"]["stage"] == 3:
122
+ # Note that `stage3_prefetch_bucket_size` can produce DeepSpeed messages like: `Invalidate trace cache @ step 0: expected module 1, but got module 0`
123
+ # This is expected and is not an error, see: https://github.com/microsoft/DeepSpeed/discussions/4081
124
+ config_kwargs.update(
125
+ {
126
+ "zero_optimization.reduce_bucket_size": hidden_size * hidden_size,
127
+ "zero_optimization.stage3_param_persistence_threshold": 10 * hidden_size,
128
+ "zero_optimization.stage3_prefetch_bucket_size": 0.9 * hidden_size * hidden_size,
129
+ }
130
+ )
131
+
132
+ # If ZeRO-3 is used, we shard both the active and reference model.
133
+ # Otherwise, we assume the reference model fits in memory and is initialized on each device with ZeRO disabled (stage 0)
134
+ if config_kwargs["zero_optimization"]["stage"] != 3:
135
+ config_kwargs["zero_optimization"]["stage"] = 0
136
+ model, *_ = deepspeed.initialize(model=model, config=config_kwargs)
137
+ model.eval()
138
+ return model
139
+
140
+ def concatenated_forward(
141
+ self, model: "PreTrainedModel", batch: Dict[str, torch.Tensor],
142
+ average_log_prob: bool = False
143
+ ) -> Tuple[torch.FloatTensor, torch.FloatTensor, torch.FloatTensor, torch.FloatTensor]:
144
+ batch_copied = BatchEncoding({k: v.detach().clone() for k, v in batch.items()}) # avoid error
145
+
146
+ all_logits = model(
147
+ input_ids=batch_copied["input_ids"], attention_mask=batch_copied["attention_mask"], return_dict=True
148
+ ).logits.to(torch.float32)
149
+
150
+ return all_logits
151
+
152
+ def _tdpo_get_batch_logps(self, logits: torch.FloatTensor,
153
+ reference_logits: torch.FloatTensor,
154
+ labels: torch.LongTensor,
155
+ average_log_prob: bool = False):
156
+ """Compute the kl divergence/log probabilities of the given labels under the given logits.
157
+
158
+ Args:
159
+ logits: Logits of the model (unnormalized). Shape: (batch_size, sequence_length, vocab_size)
160
+ reference_logits: Logits of the reference model (unnormalized). Shape: (batch_size, sequence_length, vocab_size)
161
+ labels: Labels for which to compute the log probabilities. Label tokens with a value of -100 are ignored. Shape: (batch_size, sequence_length)
162
+ average_log_prob: If True, return the average log probability per (non-masked) token. Otherwise, return the sum of the log probabilities of the (non-masked) tokens.
163
+
164
+ Returns:
165
+ Several tensors of shape (batch_size,) containing the average/sum kl divergence/log probabilities of the given labels under the given logits.
166
+ """
167
+ assert logits.shape[:-1] == labels.shape
168
+ assert reference_logits.shape[:-1] == labels.shape
169
+
170
+ labels = labels[:, 1:].clone()
171
+ logits = logits[:, :-1, :]
172
+ reference_logits = reference_logits[:, :-1, :]
173
+
174
+ loss_mask = (labels != -100)
175
+
176
+ # dummy token; we'll ignore the losses on these tokens later
177
+ labels[labels == -100] = 0
178
+
179
+ vocab_logps = logits.log_softmax(-1)
180
+
181
+ reference_vocab_ps = reference_logits.softmax(-1)
182
+ # reference_vocab_logps = reference_vocab_ps.log()
183
+ reference_vocab_logps = reference_logits.log_softmax(-1)
184
+
185
+ per_position_kl = (reference_vocab_ps * (reference_vocab_logps - vocab_logps)).sum(-1)
186
+ per_token_logps = torch.gather(vocab_logps, dim=2, index=labels.unsqueeze(2)).squeeze(2)
187
+ per_reference_token_logps = torch.gather(reference_vocab_logps, dim=2, index=labels.unsqueeze(2)).squeeze(2)
188
+
189
+ logps_margin = per_token_logps - per_reference_token_logps
190
+
191
+ if average_log_prob:
192
+ return (logps_margin * loss_mask).sum(-1) / loss_mask.sum(-1), \
193
+ (per_position_kl * loss_mask).sum(-1) / loss_mask.sum(-1), \
194
+ (per_token_logps * loss_mask).sum(-1) / loss_mask.sum(-1)
195
+ else:
196
+ return (logps_margin * loss_mask).sum(-1), \
197
+ (per_position_kl * loss_mask).sum(-1), \
198
+ (per_token_logps * loss_mask).sum(-1)
199
+
200
+ def tdpo_loss(self, chosen_logps_margin: torch.FloatTensor,
201
+ rejected_logps_margin: torch.FloatTensor,
202
+ chosen_position_kl: torch.FloatTensor,
203
+ rejected_position_kl: torch.FloatTensor,
204
+ beta: float, alpha: float = 0.5, if_tdpo2: bool = True) -> Tuple[torch.FloatTensor, torch.FloatTensor, torch.FloatTensor]:
205
+ """Compute the TDPO loss for a batch of policy and reference model log probabilities.
206
+
207
+ Args:
208
+ chosen_logps_margin: The difference of log probabilities between the policy model and the reference model for the chosen responses. Shape: (batch_size,)
209
+ rejected_logps_margin: The difference of log probabilities between the policy model and the reference model for the rejected responses. Shape: (batch_size,)
210
+ chosen_position_kl: The difference of sequential kl divergence between the policy model and the reference model for the chosen responses. Shape: (batch_size,)
211
+ rejected_position_kl: The difference of sequential kl divergence between the policy model and the reference model for the rejected responses. Shape: (batch_size,)
212
+ beta: Temperature parameter for the TDPO loss, typically something in the range of 0.1 to 0.5. We ignore the reference model as beta -> 0.
213
+ alpha: Temperature parameter for the TDPO loss, used to adjust the impact of sequential kl divergence.
214
+ if_tdpo2: Determine whether to use method TDPO2, default is True; if False, then use method TDPO1.
215
+
216
+ Returns:
217
+ A tuple of two tensors: (losses, rewards).
218
+ The losses tensor contains the TDPO loss for each example in the batch.
219
+ The rewards tensors contain the rewards for response pair.
220
+ """
221
+
222
+ chosen_values = chosen_logps_margin + chosen_position_kl
223
+ rejected_values = rejected_logps_margin + rejected_position_kl
224
+
225
+ chosen_rejected_logps_margin = chosen_logps_margin - rejected_logps_margin
226
+
227
+ if not if_tdpo2:
228
+ logits = chosen_rejected_logps_margin - (rejected_position_kl - chosen_position_kl) # tdpo1
229
+ else:
230
+ logits = chosen_rejected_logps_margin - alpha * (rejected_position_kl - chosen_position_kl.detach()) # tdpo2
231
+ losses = -F.logsigmoid(beta * logits)
232
+
233
+ chosen_rewards = beta * chosen_values.detach()
234
+ rejected_rewards = beta * rejected_values.detach()
235
+
236
+ return losses, chosen_rewards, rejected_rewards
237
+
238
+
239
+ def get_batch_loss_metrics(
240
+ self,
241
+ model: "PreTrainedModel",
242
+ batch: Dict[str, torch.Tensor],
243
+ train_eval: Optional[Literal["train", "eval"]] = "train",
244
+ ) -> Tuple[torch.Tensor, Dict[str, torch.Tensor]]:
245
+ r"""
246
+ Computes the DPO loss and other metrics for the given batch of inputs for train or test.
247
+ """
248
+ metrics = {}
249
+ (
250
+ chosen_logits
251
+ ) = self.concatenated_forward(model, batch)
252
+ with torch.no_grad():
253
+ if self.ref_model is None:
254
+ ref_model = self.model
255
+ ref_context = self.accelerator.unwrap_model(self.model).disable_adapter()
256
+ else:
257
+ ref_model = self.ref_model
258
+ ref_context = nullcontext()
259
+
260
+ with ref_context:
261
+ (
262
+ reference_logits
263
+ ) = self.concatenated_forward(ref_model, batch)
264
+
265
+ [all_logps_margin,
266
+ all_position_kl,
267
+ all_logps] = self._tdpo_get_batch_logps(chosen_logits,
268
+ reference_logits,
269
+ batch['labels'],
270
+ average_log_prob=False)
271
+
272
+ batch_size = batch["input_ids"].size(0) // 2
273
+
274
+ chosen_logps_margin = all_logps_margin[:batch_size]
275
+ rejected_logps_margin = all_logps_margin[batch_size:]
276
+ chosen_position_kl = all_position_kl[:batch_size]
277
+ rejected_position_kl = all_position_kl[batch_size:]
278
+
279
+ chosen_logps = all_logps[:batch_size].detach()
280
+ rejected_logps = all_logps[batch_size:].detach()
281
+
282
+ losses, chosen_rewards, rejected_rewards = self.tdpo_loss(
283
+ chosen_logps_margin, rejected_logps_margin,
284
+ chosen_position_kl, rejected_position_kl,
285
+ self.beta, self.alpha, self.if_tdpo2
286
+ )
287
+
288
+ reward_accuracies = (chosen_rewards > rejected_rewards).float()
289
+ chosen_rewards_accuracies = (chosen_logps_margin > torch.zeros_like(chosen_logps_margin)).float()
290
+
291
+ prefix = "eval_" if train_eval == "eval" else ""
292
+ metrics[f"{prefix}rewards/chosen"] = chosen_rewards.cpu().mean()
293
+ metrics[f"{prefix}rewards/rejected"] = rejected_rewards.cpu().mean()
294
+ metrics[f"{prefix}rewards/accuracies"] = reward_accuracies.cpu().mean()
295
+ metrics[f"{prefix}rewards/margins"] = (chosen_rewards - rejected_rewards).cpu().mean()
296
+ metrics[f"{prefix}logps/rejected"] = rejected_logps.detach().cpu().mean()
297
+ metrics[f"{prefix}logps/chosen"] = chosen_logps.detach().cpu().mean()
298
+ metrics[f"{prefix}logits/logp_chosen>logp_ref_chosen"] = chosen_rewards_accuracies.detach().cpu().mean()
299
+
300
+ all_device_chosen_position_kl = chosen_position_kl.detach()
301
+ all_device_rejected_position_kl = rejected_position_kl.detach()
302
+
303
+ metrics[f'{prefix}kl/chosen'] = all_device_chosen_position_kl.cpu().numpy().tolist()
304
+ metrics[f'{prefix}kl/rejected'] = all_device_rejected_position_kl.cpu().numpy().tolist()
305
+ metrics[f'{prefix}kl/margin'] = (all_device_chosen_position_kl - all_device_rejected_position_kl).cpu().numpy().tolist()
306
+
307
+ return losses.mean(), metrics
308
+
309
+